Skip to content
Longterm Wiki
Back

Credibility Rating

3/5
Good(3)

Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.

Rating inherited from publication venue: Good Judgment

Tetlock's superforecasting work is frequently cited in AI safety communities as a methodological foundation for rigorous reasoning about AI timelines, risk estimates, and capability predictions under deep uncertainty.

Metadata

Importance: 55/100homepageeducational

Summary

Philip Tetlock's superforecasting research demonstrates that trained individuals using systematic probabilistic thinking can significantly outperform experts and prediction markets on a wide range of forecasting questions. The approach emphasizes calibration, updating beliefs on new evidence, and aggregating diverse forecaster perspectives. These methods are directly applicable to forecasting AI development timelines and risks.

Key Points

  • Superforecasters consistently outperform experts by using calibrated probabilistic reasoning rather than vague qualitative predictions.
  • Key practices include breaking problems into components, actively updating on new evidence, and avoiding overconfidence or underconfidence.
  • Aggregating forecasts from diverse, well-calibrated individuals (crowd wisdom) improves accuracy beyond any single forecaster.
  • The methodology has direct applications to AI capabilities forecasting, risk estimation, and evaluating safety-relevant timelines.
  • Superforecasting principles can reduce uncertainty in domains critical to AI safety planning, such as predicting transformative AI development windows.

Review

Tetlock's work on superforecasting provides a groundbreaking approach to improving predictive accuracy by emphasizing disciplined, probabilistic reasoning over traditional expert intuition. By studying individuals who consistently outperform expectations in forecasting complex events, he reveals that effective prediction requires breaking down complex problems, updating beliefs based on new evidence, and avoiding cognitive biases. The methodology centers on training forecasters to think in probabilities, actively update their views, and maintain intellectual humility. While the approach has shown remarkable success in geopolitical and economic predictions, its application to emerging domains like pandemic forecasting demonstrates its potential for addressing high-stakes uncertainty. However, the method is not without limitations, as it requires significant cognitive effort, ongoing training, and may not always capture black swan events or fundamental paradigm shifts.

Cited by 1 page

PageTypeQuality
AI-Augmented ForecastingApproach54.0

Cached Content Preview

HTTP 200Fetched Apr 7, 20261 KB
The Superforecasting Guide to Coronavirus - Good Judgment 
 
 
 
 
 
 
 
 
 
 

 
 
 
 
 
 
 

 

 
 

 
 
 
 
 

 

 
 
 
 
 
 
 
 
 
 
 

 

 

 
 
 
 
 
 
 The Superforecasting Guide to Coronavirus 
 

 

 
 
 
 
 

 
 
 
 
 Intrigued?

 
 Stop guessing. Start Superforecasting. 

 Schedule a consultation to learn how our FutureFirst monitoring tool, custom Superforecasts, and training services can help your organization make better decisions.

 
 
 
 Comments This field is for validation purposes and should be left unchanged. 
 First Name * 
 Last Name * 
 Company Name * 
 Email * 
 
 
 How can we help? 
 * 
 
 I consent to the terms described in Good Judgment's Privacy Policy . 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 Δ 

 
 
 
 
 
 
 
 
 
 
 
 
 
 Keep up with Superforecasting news. Discover training opportunities. Make better decisions. Sign up for Good Judgment’s newsletter.

 
 
 
 
 
 Name This field is for validation purposes and should be left unchanged. 
 First Name 
 Last Name 
 Email * 
 
 
 Organization 
 GDPR * 
 
 I consent to the terms described in Good Judgment's Privacy Policy . 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 Δ
Resource ID: 55e4c8653a8ad2d2 | Stable ID: Mzc3ZjI5Mz