The $1 Trillion Question
The $1 Trillion Question, Mortality in the Age of Generative Ghosts, Your Mind and AI, How to Design AI Tutors for Learning, The Imagining Summit Preview: Adam Cutler, and Helen's Book of the Week.
A review of research by Phil Tetlock and other experts on crafting better prompts by investigating if human forecasting can be improved through the use of a large language model.
Our focus is deeply rooted in leveraging AI to enhance decision-making, so you can imagine my enthusiasm when I discovered that Phil Tetlock, the author of Superforecasters and a pivotal figure behind many strategies in our book, Make Better Decisions, has collaborated with experts from the Federal Reserve, MIT, and the London School of Economics. Their goal? To investigate if human forecasting can be improved through the use of a large language model.
The study, titled LLM Assistants Improve Human Forecasting Accuracy, explores how AI augmentation interacts with human cognitive biases. The findings reveal that collaborating with GPT-4 on forecasting tasks can markedly improve decision-making processes, particularly by addressing issues of overconfidence and enhancing forecasting precision.
Overconfidence in decision-making is characterized by an inflated belief in one's own judgment, capabilities, or control over outcomes. It's been called the mother of all biases and most of us are overconfident most of the time. This cognitive bias often leads to a misjudgment of risks and an overestimation of one’s predictive accuracy, resulting in decisions that fall short of optimal, especially in complex scenarios where uncertainty is high and information is incomplete.
The paper finds an innovative approach to mitigating the adverse effects of overconfidence through the strategic use of LLMs, including an experiment involving a "biased" LLM designed to neglect base rates and exhibit overconfidence.
The design of the study is especially intriguing for its structured approach. It allowed participants the option to consult with two distinct types of AI advisors: one being a high-quality "superforecasting" prompt designed to provide optimized forecasting guidance, and the other, a biased advisor characterized by overconfidence and a disregard for base rates. Conversely, the control group was only equipped with a less sophisticated model that lacked any direct forecasting assistance.
Participants leveraging the advanced models—regardless of whether they opted for the superforecaster or the biased forecaster—demonstrated a 23% improvement in performance over those in the control group.
The Artificiality Weekend Briefing: About AI, Not Written by AI