"Prediction is very difficult, especially about the future." But why, and how can we do better? Nate Silver unpacks the challenges of prediction and offers a roadmap for sifting valuable signals from overwhelming noise.
1. Economists Struggle with Prediction and Overconfidence
Economic forecasting often fails—not because of lack of data, but due to flawed methods and misplaced confidence. Economists regularly make specific predictions, such as estimating GDP growth down to a decimal point, but these numbers are misleading. They come from broader intervals, which are themselves frequently inaccurate.
The models used by economists are often too optimistic about their predictive accuracy. If a forecast claims a 90% confidence interval, one would expect actual GDP outcomes to fall outside this range only 10% of the time. In reality, they fall outside the range about half the time. This discrepancy indicates a fundamental issue with economic assumptions and models.
Consider depressions: between 1990 and 2000, economists aimed to predict global downturns. Yet they managed to forecast only two out of sixty depressions one year in advance. This ineffectiveness shows that complex economic systems elude even well-trained experts.
Examples
- GDP growth forecasts frequently fall outside their intervals, showing errors in foundational assumptions.
- During the 1990s, economists predicted only two out of 60 global depressions despite analyzing numerous indicators.
- Historical data of inaccurate forecasts reveals economists' overconfidence in their own models.
2. The Economy Is a Complicated and Shifting Web
Predicting the economy is daunting because countless interconnected factors influence outcomes. For example, a natural disaster in one part of the world might ripple into another, affecting unrelated industries or jobs.
The interdependencies create feedback loops, like high sales driving job growth, which in turn fuels more spending. These loops make it hard to determine what causes what. External factors, like government interference, may also distort indicators. Rising housing prices usually suggest economic health, but the housing bubble of the early 2000s proved otherwise.
Even economic theories and data themselves are fluid. Indicators are frequently revised, meaning economists base predictions on flawed or outdated information. For example, the US government initially reported a 3.8% GDP drop for late 2008, only to later revise it to a much grimmer 9%.
Examples
- Tsunamis in Asia can disrupt factory supply chains that affect hiring in the United States.
- Feedback loops, like increased hiring boosting consumer spending, mask causation.
- The 2008 financial collapse exposed inaccurate preliminary economic data.
3. Statistical Analysis Alone Is Insufficient
Economic forecasts often rely heavily on detecting patterns in data. However, this reliance on statistics can lead to errors from coincidences mistaken for meaningful trends. For example, from 1967 to 1997, a peculiar correlation emerged: the Super Bowl winner seemed to predict stock market success. This connection, though statistically improbable, was pure chance.
With vast data sources—tracking more than four million economic indicators—some patterns will inevitably appear random. Blindly depending on such "false signals" can cause predictions to go awry. Human judgment must accompany statistical tools, ensuring that plausible causality backs any correlation.
Instead of approaching data with caution, some analysts assume adding more variables will improve accuracy. But more data only leads to more irrelevant information (noise), making it harder to isolate the real signals driving outcomes.
Examples
- The Super Bowl stock market trend failed after decades of seeming correlation.
- Over 4 million tracked indicators inevitably generate coincidental trends.
- Larger datasets don’t automatically improve predictions; they increase noise.
4. The Housing Bubble of 2008 Highlighted Predictive Failures
The 2008 US financial crisis exposed several layers of flawed prediction. Many assumed housing prices would continue rising. Homeowners, brokers, and lenders embraced this trend, despite historical evidence that skyrocketing housing markets often end in crashes.
Financial institutions poorly rated mortgage-based securities, or CDOs, oblivious to large-scale risks. Statistical models assessed risk solely on individual homeowners defaulting but missed how a housing crash could amplify defaults industry-wide. Their miscalculations had colossal consequences.
Even after the recession began, predictions failed. Analysts assumed this was a standard, short-term downturn, ignoring evidence that financial crises typically cause prolonged unemployment. This oversight led to inadequate stimulus measures.
Examples
- Home prices rose despite warning signs of a bubble, creating false confidence among lenders and brokers.
- Standard & Poor's gave AAA-ratings to financial instruments, yet 28% of those defaulted.
- Policymakers underestimated the long recovery from financial recessions, deepening economic damage.
5. Bayes’ Theorem Improves Probabilities with Each New Data Point
Bayesian thinking is invaluable in refining predictions. This approach begins with a "prior probability," or initial assumption, and adjusts it as new evidence emerges. Far from intuitive, it provides clarity about balancing prior knowledge against current observations.
Take breast cancer predictions, for example. A woman in her forties knows her base risk is about 1.4%. After a positive mammogram, her probability only rises to 10%, given mammograms have high false-positive rates. Without Bayes’ logic, many mistakenly assume the chance would be much higher.
Relying solely on the latest piece of information overstates its role in overall probability. Bayesian updating anchors beliefs in both historical data and present observations—a balance many predictors lack.
Examples
- Bayes’ theorem reveals a positive mammogram raises breast cancer likelihood only to 10%, not much higher as assumed.
- Historical incidence outweighs the mammogram’s imperfect accuracy.
- Bayesian probabilities reconcile previous trends with emerging evidence.
6. Forward-Thinking Experts Use Fox-Like Mentality
Philip Tetlock’s studies revealed predictors who succeeded relied on versatility, not rigid ideologies. By consuming data from many sources and remaining flexible, "foxes" outperformed "hedgehogs," who fiercely clung to one core idea.
Hedgehogs, often confident and appealing to the media, seldom fared better than random guesses. Foxes, though cautious and less headline-grabbing, leveraged multiple perspectives to refine their forecasts and settle on productive middle grounds.
Tetlock’s distinctions emphasize the value of humility and adaptability. Relying on one overarching theory about the world usually misleads, while collaborating across disciplines provides nuanced and practical forecasts.
Examples
- Hedgehogs like Freud fixated on single ideas, often exaggerating their importance.
- Fox-like predictors balanced competing theories, improving accuracy.
- Media prefers confident predictions, despite their lack of reliability.
7. The Stock Market Resists Prediction Due to Efficiency
Short-term stock movements are infamously hard to predict because markets self-correct quickly. Any over- or under-pricing gets resolved as investors react promptly. This efficiency ensures few opportunities for consistent gains by any one trader.
Studies reinforce this. Hedge funds rarely repeat outperforming years—it’s usually just luck. Even collective forecasts from economists rarely outperform basic market averages. Inside information offers the only clear advantage, as seen with Congress members achieving unusually high returns using their privileged knowledge.
In short, stock market efficiency neutralizes most strategies, limiting opportunities for reliable prediction.
Examples
- Stock market forecasts by economists underperform collective averages.
- A study found hedge funds’ best-performing years were coincidence, not skill.
- Congress members' portfolios benefit from insider knowledge.
8. Climate Forecasting Works Best with Simple Models
Despite climate complexity, simpler models often outperform overengineered predictions. Research relying mainly on CO2 levels, for example, provided more precise projections of temperature changes over decades compared to multi-factor models.
Even as global warming is unequivocal, its far-reaching consequences remain harder to predict. The simpler models’ success lies their focus on key signals—such as CO2 trapping heat in the atmosphere—while excluding noise like cyclical phenomena.
Recognizing such clear cause-and-effect relationships strengthens forecasting reliability, prioritizing clarity over complication.
Examples
- The IPCC’s 1990 climate model overpredicted temperature growth.
- Scientists vastly disagreed on their models’ reliability for sea-level rise.
- Models focusing on CO2 correlations accurately predicted long-term trends.
9. Preventing Terrorist Attacks Demands Signal Over Noise
The US government faced vast noise in analyzing risks before the 9/11 attacks. Thousands of small leads and suspicious behaviors clouded their ability to detect credible threats. Only in hindsight did key warnings seem obvious.
Yet a pattern exists: large-scale attacks like 9/11, based on Clauset’s curve, tend to appear every eight decades. Countries like Israel focus resources on preventing such large attacks, treating smaller incidents as criminal rather than forming intelligence hierarchies.
Following this model could empower countries to spot meaningful signals while avoiding distractions by minor threats.
Examples
- Numerous indicators suggested risk before 9/11 but weren’t prioritized.
- Clauset’s curve predicts large-scale attacks with long frequencies.
- Israel’s disproportionate focus on major threats prevents mass-casualty events.
Takeaways
- Apply Bayesian reasoning in daily decisions by combining past experience with new inputs.
- Treat experts’ predictions with skepticism unless they show versatility and engagement with diverse viewpoints.
- Focus on isolating key signals from irrelevant noise, especially in complex systems like finance, climate, or security.