Published by Princeton University Press in 2005, Expert Political Judgment: How Good Is It? How Can We Know? presented the results of Tetlock's twenty-year study tracking 28,000 predictions from 284 political scientists, economists, and intelligence analysts. The central finding — that expert forecasts were no more accurate than random guessing — challenged the authority of expertise itself. More consequentially, the book identified a small minority of forecasters who consistently outperformed everyone else, and documented the cognitive habits distinguishing them: thinking in probabilities, updating frequently, seeking disconfirmation, and resisting identity-protective reasoning. The book introduced the fox-hedgehog framework as an empirical predictor of forecasting accuracy.
The study's design was elegant in its rigor. Tetlock asked experts to make specific, falsifiable predictions about geopolitical events: Would the Soviet Union use force to retain its Baltic states? Would Quebec secede from Canada? Would the European Union expand eastward? Each prediction required a probability estimate and a time horizon. Each was scored against what actually happened using a Brier score — a quadratic measure of the distance between predicted probability and observed outcome. The scoring was merciless: there was no room for post-hoc reinterpretation, no opportunity to claim that the prediction was 'directionally correct' when the specific event failed to materialize. The experts were held to the standard they implicitly claimed to meet: better-than-chance accuracy.
The book's second finding was more actionable than its first. A subset of forecasters — roughly fifteen percent of the sample — consistently beat both the group average and statistical extrapolation models. These 'superforecasters' (the term would be formalized later) shared cognitive habits: they thought in granular probabilities rather than verbal estimates, updated their beliefs proportionally as evidence accumulated, actively sought information that could disconfirm their hypotheses, and resisted the pull of identity-protective cognition. They were, in Berlin's taxonomy, foxes: eclectic, self-critical, comfortable with ambiguity. The hedgehogs — confident, theory-driven, committed to grand narratives — were the worst forecasters in the study, yet they were also the most prominent in public discourse.
The book's reception split along predictable lines. The media focused on the dart-throwing chimpanzee comparison, which became a punchline. Academic reviewers engaged the deeper methodological questions: whether Tetlock's scoring system was fair, whether the sample was representative, whether twenty years was sufficient to test forecasts whose time horizons extended decades. Practitioners — intelligence analysts, risk managers, policy advisors — absorbed the operational lessons: probabilistic thinking works, confidence is a liability, and the institutional structures surrounding expertise systematically reward the wrong cognitive habits. The book won the Woodrow Wilson Foundation Award and the Alexander George Award, but its practical impact remained limited because organizations continued to reward confident hedgehogs over calibrated foxes.
The study originated in Tetlock's 1984 observation that political and economic experts appeared on television constantly, making predictions with apparent authority, and those predictions were never systematically evaluated. The absence of accountability created a perverse selection pressure: the experts who sounded most confident got invited back, regardless of whether their previous predictions had been accurate. Tetlock proposed a natural experiment: collect the predictions, wait for the outcomes, score the correspondence, and publish the results. The design required patience — some predictions had five- or ten-year time horizons — and institutional continuity across two decades of data collection. The National Science Foundation and other funders supported the research, and Tetlock's move from Berkeley to Penn in 2000 did not interrupt it. The accumulation of 28,000 predictions provided statistical power no previous forecasting study had achieved.
Inverse confidence-accuracy correlation. The experts who appeared most frequently on television and spoke with greatest certainty were the least accurate forecasters — confidence is a social performance, not an epistemic signal.
Base-rate neglect. Experts systematically ignored the prior probability of events (the outside view) in favor of case-specific narratives (the inside view), producing predictable overconfidence.
Belief perseverance. Disconfirming evidence rarely changed expert minds — instead, anomalies were explained away through auxiliary hypotheses that preserved the core framework.
Cognitive style trumps credentials. Whether a forecaster was a full professor or an assistant, had a PhD or a master's, worked at a think tank or a university mattered less than their cognitive style.
Feedback vacuum. Experts operated in low-feedback environments where predictions were rarely scored, enabling overconfidence to persist uncorrected across entire careers.