Good Judgment
- QualityRated 50 but structure suggests 80 (underrated by 30 points)
- Links4 links could use <R> components
Quick Assessment
Section titled “Quick Assessment”| Aspect | Assessment |
|---|---|
| Type | Commercial forecasting organization |
| Founded | 2015 (from Good Judgment Project, 2011-2015) |
| Key Innovation | Identification and employment of “superforecasters” |
| Track Record | 30-72% more accurate than competitors; outperformed intelligence analysts by 30% |
| Primary Focus | Geopolitical forecasting, strategic risk assessment |
| Key Products | FutureFirst™, training workshops, custom forecasts |
| Funding | Government contracts, commercial clients (energy, finance, nonprofits) |
| AI Safety Relevance | Limited direct involvement; potential application to AI risk forecasting |
Key Links
Section titled “Key Links”| Source | Link |
|---|---|
| Official Website | gjopen.com |
| Wikipedia | en.wikipedia.org |
Overview
Section titled “Overview”Good Judgment Inc. is a commercial forecasting organization that emerged from the Good Judgment Project (GJP), a research initiative that ran from 2011 to 2015 as part of the Intelligence Advanced Research Projects Activity (IARPA) Aggregative Contingent Estimation (ACE) tournament.1 The organization’s core innovation was the identification of “superforecasters”—individuals who demonstrate exceptional accuracy in probabilistic forecasting on geopolitical and global events, often outperforming professional intelligence analysts with access to classified information.2
Founded by Philip TetlockPhilip TetlockPhilip Tetlock is a psychologist who revolutionized forecasting research by demonstrating that expert predictions often perform no better than chance, while identifying systematic methods and 'supe...Quality: 73/100 and Barbara Mellers of the University of Pennsylvania, Good Judgment pioneered methods for crowd-sourced forecasting that combine amateur forecasters, bias training, and aggregation algorithms to produce remarkably accurate predictions.3 During the IARPA tournament, the Good Judgment Project won outright with performance 35-72% better than rival teams and more than 30% better than intelligence community analysts.4 The project discovered that certain individuals—dubbed superforecasters—could consistently make more accurate predictions than both domain experts and prediction marketsInterventionPrediction MarketsPrediction markets achieve Brier scores of 0.16-0.24 (15-25% better than polls) by aggregating dispersed information through financial incentives, with platforms handling $1-3B annually. For AI saf...Quality: 56/100.
After the research phase concluded in 2015, Good Judgment transitioned to a commercial model, offering forecasting services to organizations across government, energy, finance, and nonprofit sectors.5 The company maintains a global network of superforecasters across six continents who provide 24/7 crowd-sourced insights on strategic questions through platforms like FutureFirst™.6 Good Judgment also offers training programs to help organizations develop internal forecasting capabilities, emphasizing probabilistic thinking, bias correction, and accountable decision-making.
History
Section titled “History”Origins in Academic Research
Section titled “Origins in Academic Research”The roots of Good Judgment trace back to Philip Tetlock’s 1984 research on forecasting tournaments involving over 250 experts in political and economic trends.7 This early work revealed a surprising finding: domain expertise did not strongly correlate with predictive accuracy. Tetlock’s book Expert Political Judgment documented systematic failures in expert predictions and laid the groundwork for investigating alternative approaches to forecasting.
In 2011, following intelligence community failures to anticipate major geopolitical events, IARPA launched the Aggregative Contingent Estimation (ACE) tournament to identify better forecasting methods.8 Tetlock and Barbara Mellers formed the Good Judgment Project as a competing team, recruiting amateur forecasters and testing various techniques including bias tutorials, aggregation algorithms, and team deliberation structures.
The IARPA Tournament (2011-2015)
Section titled “The IARPA Tournament (2011-2015)”The Good Judgment Project’s performance in the ACE tournament was exceptional. During its first year in 2011, the project recruited forecasters like Jean-Pierre Beugoms and generated over 1 million forecasts across 500 questions ranging from Venezuelan gas subsidies to North Korean politics.9 The questions were carefully designed as verifiable predictions with clear resolution criteria, scored using Brier scores to measure probabilistic accuracy.
By 2012, GJP had identified a subset of elite forecasters and organized them into teams of approximately 12 individuals.10 The median forecasts from these superforecaster teams proved 35-72% more accurate than competing teams, demonstrating that simple aggregation of skilled forecasters could rival sophisticated algorithms. The project’s success was so pronounced that by summer 2013, GJP became the sole IARPA-funded team and gained access to the Integrated Conflict Early Warning System.11
The tournament continued through 2015, with IARPA ending it early due to GJP’s dominance.12 Throughout the five-year period, the project produced over 1 million individual forecasts, identified consistent patterns in forecasting skill, and demonstrated that superforecasters could maintain their accuracy advantage over time. Notably, superforecasters proved more accurate than professional intelligence analysts with access to classified information—a finding that challenged conventional assumptions about the value of domain expertise and secret intelligence.13
Commercialization and Growth
Section titled “Commercialization and Growth”Following the conclusion of the IARPA research in fall 2015, Good Judgment Inc. launched as a commercial entity.14 The company began hiring professional superforecasters like Jean-Pierre Beugoms, who transitioned from volunteer forecaster to full-time professional.15 Good Judgment also established Good Judgment Open (GJ Open), a public forecasting platform that serves both as a recruitment pipeline for identifying new superforecasters and as a training ground for developing forecasting skills.16
The commercial organization expanded beyond geopolitics to address questions in finance, energy, public health, and organizational strategy.17 Good Judgment developed proprietary tools like FutureFirst™, which provides clients with continuous forecasting insights by assigning approximately 40 superforecasters to each client question.18 The company also built a training business, offering workshops and seminars on probabilistic thinking, bias correction, and forecasting methodology to government agencies, corporations, and nonprofits across the United States, United Kingdom, Netherlands, and Turkey.19
Recent Developments (2024-2025)
Section titled “Recent Developments (2024-2025)”In 2024, Good Judgment significantly expanded its partnerships and media presence. The organization collaborated with major financial institutions including UBS Asset Management and Man Group on public forecasting challenges, while also hosting private challenges for organizations seeking to identify internal forecasting talent.20 Superforecasters were featured in The Economist’s “The World Ahead 2025” issue with predictions on US tariffs, elections in Germany, Canada, and Australia, China’s inflation, and the Russia-Ukraine conflict.21
The company demonstrated continued forecasting superiority, with superforecasters proving 30% more accurate than futures markets on Federal Reserve, Bank of England, and European Central Bank interest rate decisions during 2024-2025.22 In 2025, Good Judgment won an Honorable Mention in the IF Awards from the Association of Professional Futurists for its work with UK partner ForgeFront on the Future.Ctrl methodology.23 The organization also launched Advanced Judgment & Modeling as a next-level training program for graduates of its two-day workshops and began experimenting with hybrid AI-superforecasting integration.24
Key People and Leadership
Section titled “Key People and Leadership”Philip Tetlock serves as co-founder and is the intellectual architect of the superforecasting approach.25 A psychology professor at the University of Pennsylvania, Tetlock authored both Expert Political Judgment (which documented the limitations of expert forecasting) and Superforecasting: The Art and Science of Prediction (co-authored with Dan Gardner), which popularized the findings from the Good Judgment Project.26 His research focuses on decision-making, expert judgment, and the cognitive characteristics that enable accurate probabilistic reasoning.
Barbara Mellers, also a co-founder, is a decision scientist at the University of Pennsylvania who co-led the Good Judgment Project with Tetlock.27 Her research contributions focused on the psychological strategies that improve forecasting accuracy, including the role of cognitive ability, training interventions, and team collaboration in enhancing predictions.
Dr. Warren Hatch serves as CEO of Good Judgment Inc.28 He holds a PhD from Oxford, previously worked on Wall Street at Morgan Stanley and a boutique firm, and is both a CFA Charterholder and a practicing superforecaster. Hatch has represented Good Judgment at high-profile events including the UK Department for Environment, Food and Rural Affairs (DEFRA) Futures Trend Briefing in November 2024 and the UN OCHA Global Humanitarian Policy Forum in December 2025.29
Marc Koehler is Senior Vice President at Good Judgment Inc. and leads workshops on forecasting precision.30 A former diplomat and practicing superforecaster, Koehler brings both domain expertise and forecasting skill to the organization’s training programs.
Michael Story, former Managing Director at Good Judgment, founded the Swift CentreSwift CentreSwift Centre is a UK forecasting organization that provides conditional forecasting services to various clients including some AI companies, but is not primarily focused on AI safety. While they de...Quality: 50/100 to apply GJP methods commercially after his tenure with the organization.31 His background spans hedge funds, consulting, psychometrics, and risk quantification, reflecting the interdisciplinary nature of professional forecasting.
The superforecaster network itself represents a key organizational asset. These individuals span six continents and multiple languages, bringing decades of combined experience in probabilistic forecasting.32 Superforecasters are rigorously selected through performance on Good Judgment Open, with only the most consistently accurate forecasters qualifying for professional roles. Research indicates that superforecasters share certain characteristics including high cognitive ability, political knowledge, open-mindedness, pattern detection skills, and a deliberate practice mindset.33
Methodology and Approach
Section titled “Methodology and Approach”Good Judgment’s forecasting methodology combines several key elements that distinguish it from traditional expert prediction. The approach emphasizes probabilistic thinking rather than binary yes/no predictions, with forecasters assigning percentage probabilities to different outcomes.34 This allows for more nuanced assessment and enables scoring through Brier scores, which penalize both overconfidence and underconfidence.
The organization employs systematic bias training to help forecasters recognize and correct common cognitive errors. Research from the Good Judgment Project found that training on biases, combined with aggregation algorithms and team deliberation, significantly improved forecasting accuracy.35 The project’s success in the IARPA tournament was partly attributed to these training interventions, which proved more effective than simply recruiting domain experts.
Team collaboration plays a crucial role in Good Judgment’s approach. During the ACE tournament, teams of 12 superforecasters consistently outperformed larger groups of regular forecasters, suggesting that high-skill aggregation matters more than crowd size.36 Good Judgment typically assigns approximately 40 superforecasters to each client question, balancing the benefits of diverse perspectives with the premium placed on individual skill.37
The organization also emphasizes long-term tracking and accountability. Forecasters make predictions on questions with clear resolution criteria and specified resolution dates, often ranging from months to years in the future.38 This creates a track record that enables both individual skill assessment and continuous methodological improvement. Research analyzing five years of Good Judgment Project data found that compromise (averaged) forecasts from multiple forecasters were consistently more accurate than individual predictions and improved as events neared their resolution dates.39
Performance and Track Record
Section titled “Performance and Track Record”Good Judgment’s empirical track record provides strong evidence for the effectiveness of its approach. During the IARPA ACE tournament (2011-2015), the Good Judgment Project achieved more than 50% improvement over control groups—the largest effect in the forecasting literature at that time.40 Superforecasters demonstrated the ability to anticipate events 400 days in advance as accurately as other forecasters could predict them at 150 days out.41
Comparative performance metrics highlight the organization’s advantages across multiple benchmarks. Superforecasters outperformed US intelligence analysts with access to classified information by more than 30%, competing IARPA teams by 35-72%, and client experts combined with crowd-wisdom groups in policy forecasts.42 In head-to-head comparisons, 100 superforecasters defeated hybrid systems combining machine learning with 1,000+ crowd forecasts in 100% of test cases.43
Recent performance demonstrates sustained accuracy. In 2023, Good Judgment superforecasters earned full marks on 8 out of 9 resolved forecasts in The Economist, correctly predicting global economic growth at 3%, China’s growth at 5%, and that Putin would not be ousted from power.44 The organization also outperformed Financial Times readers (8,500 participants) on forecasts for 2023 events and proved 30% more accurate than futures markets on central bank interest rate decisions during 2024-2025.45
As of September 2023, Good Judgment had resolved over 554 questions since 2015, with superforecasters placing the highest probability on the correct outcome across the majority of forecasting days.46 The organization reports significantly lower Brier scores (indicating higher accuracy) compared to peer forecasters, with consistent performance maintained across diverse question types spanning geopolitics, economics, and social trends.47
Funding and Operations
Section titled “Funding and Operations”Good Judgment Inc. operates as a commercial entity with approximately 30 employees and annual revenue of approximately $5.4 million as of the most recent available data.48 The company is headquartered at 100 Park Avenue, Floor 16, New York City.
The organization’s roots lie in government-funded research, specifically the IARPA-sponsored ACE tournament that ran from 2011 to 2015.49 This research funding from the US intelligence community established the empirical foundation for Good Judgment’s commercial services. Since transitioning to a for-profit model, the company has diversified its revenue sources across government contracts, corporate clients in energy and finance sectors, and nonprofit organizations.50
Recent grants demonstrate Good Judgment’s continued relevance to high-stakes decision-making in the effective altruism and global health communities. In March 2025, GiveWell paid Good Judgment Inc. $72,000 (50% of a $144,000 total project) to provide superforecaster predictions about US government foreign aid funding levels, with Open PhilanthropyOpen PhilanthropyOpen Philanthropy rebranded to Coefficient Giving in November 2025. See the Coefficient Giving page for current information. contributing the remaining 50%.51 This funding commissioned six forecasts on potential US foreign aid cuts, particularly focused on global health programs. Earlier contracts from Coefficient GivingCoefficient GivingCoefficient Giving (formerly Open Philanthropy) has directed $4B+ in grants since 2014, including $336M to AI safety (~60% of external funding). The organization spent ~$50M on AI safety in 2024, w...Quality: 55/100 included $150,000 for H5N1 forecasts (February 2023) and an unspecified amount for forecasting work on power-seeking AIRiskPower-Seeking AIFormal proofs demonstrate optimal policies seek power in MDPs (Turner et al. 2021), now empirically validated: OpenAI o3 sabotaged shutdown in 79% of tests (Palisade 2025), and Claude 3 Opus showed...Quality: 67/100 (January 2023).52
The company also receives support for collaborative projects. A Future Fund grant supported Good Judgment’s partnership with MetaculusOrganizationMetaculusMetaculus is a reputation-based forecasting platform with 1M+ predictions showing AGI probability at 25% by 2027 and 50% by 2031 (down from 50 years away in 2020). Analysis finds good short-term ca...Quality: 50/100 on forecasting projects related to Our World In Data metrics, though the specific grant amount was not publicly disclosed.53 These diverse funding sources reflect Good Judgment’s positioning at the intersection of commercial forecasting, academic research, and public-interest applications.
Products and Services
Section titled “Products and Services”FutureFirst™ is Good Judgment’s flagship product, providing clients with continuous forecasting insights on strategic questions.54 The platform operates 24/7 with approximately 40 superforecasters assigned per question, delivering probabilistic predictions on newsworthy and client-specific topics. The service reported 100% renewal rates among clients in Q4 2024, suggesting high satisfaction with the forecasting outputs.55
Good Judgment Open serves as both a public forecasting platform and a recruitment pipeline for identifying new superforecasters.56 The platform is free to use and functions similarly to golf par—users can benchmark their forecasting accuracy against questions with known resolutions. In 2024, GJ Open hosted public challenges for organizations including UBS Asset Management, Man Group, Fujitsu, City University of Hong Kong, Harvard Kennedy School, and The Economist.57 The platform also supports private challenges for organizations seeking to identify internal forecasting talent and train staff in probabilistic thinking.
Training and workshops represent a growing component of Good Judgment’s business. The organization conducts forecasting training workshops, seminars, and presentations for hundreds of participants across government, nonprofit, and private sectors.58 In 2025, Good Judgment launched an executive education program in Superforecasting Workshops for decision-makers, with clients including a major technology company, an oil multinational, and multiple investment funds.59 The company also introduced Advanced Judgment & Modeling as a next-level training program for graduates of its two-day workshop.60
Custom forecasting services allow clients to commission superforecaster predictions on proprietary strategic questions. Good Judgment frames client questions for optimal forecasting, assigns superforecaster teams, and delivers aggregated probabilistic predictions with documented track records. Clients span energy providers (for geopolitical and regulatory risk assessment), financial services, government agencies, and nonprofits.61
Partnerships and Collaborations
Section titled “Partnerships and Collaborations”Good Judgment has established strategic partnerships across media, government, business, and research sectors. The organization maintains an 11-year collaboration with The Economist, with superforecasters regularly featured in the magazine’s annual outlook publications and participating in forecasting challenges on major global events.62 Coverage has also appeared in The New York Times, Wired, Vox, Financial Times, Bloomberg, Newsweek, The Guardian, and Forbes.63
In the UK government sector, Good Judgment partners with ForgeFront and the UK Government’s Futures Procurement Network.64 This collaboration included Good Judgment CEO Dr. Warren Hatch’s presentation at the Department for Environment, Food and Rural Affairs (DEFRA) Futures Trend Briefing in November 2024 on the role of superforecasting in biological security strategy. The partnership with ForgeFront earned an Honorable Mention in the 2025 IF Awards from the Association of Professional Futurists for their joint work on the Future.Ctrl methodology.65
A significant research collaboration with Metaculus brought together two of the largest human judgment forecasting communities globally.66 Supported by a Future Fund grant, the partnership involves cohorts of superforecasters from Good Judgment and Pro Forecasters from Metaculus making predictions on identical questions about technological advances, global development, and social progress across time horizons ranging from one to 100 years. This collaboration enables comparison of forecasting methodologies and results between the two leading platforms.
Good Judgment also works with educational institutions including Harvard Kennedy School and City University of Hong Kong, hosting forecasting challenges and tournaments.67 The organization partnered with the Alliance for Decision Education on forecasting tournaments and hosted the 2025 University Forecasting Challenge, with winners receiving seats in the Superforecasting workshop running from August to December 2025.68
Relationship to AI Safety and Effective Altruism
Section titled “Relationship to AI Safety and Effective Altruism”Good Judgment has limited direct involvement in AI safety research or existential risk mitigation, with the organization’s primary focus remaining on geopolitical and strategic forecasting.69 However, several connections exist between Good Judgment’s work and the AI safety community.
The organization has received funding from effective altruism-aligned sources for forecasting projects with potential relevance to AI risk. Coefficient Giving funded a forecasting review on power-seeking AI in January 2023, though the specific amount and detailed findings were not publicly disclosed.70 This represents one of the few documented instances of Good Judgment directly engaging with AI safety questions. The organization’s March 2025 collaboration with GiveWell and Open PhilanthropyOpen PhilanthropyOpen Philanthropy rebranded to Coefficient Giving in November 2025. See the Coefficient Giving page for current information. focused on US foreign aid rather than AI-specific risks, but demonstrates ongoing relationships with effective altruism funders.71
Within the effective altruism community, there has been discussion about the concept of “good judgment” as a critical trait for impactful decision-making, though without a standardized definition.72 Community members describe good judgment as mental processes leading to good decisions, comprising understanding of the world and effective heuristics. There is acknowledged overlap between EA’s concept of good judgment and LessWrongLesswrongLessWrong is a rationality-focused community blog founded in 2009 that has influenced AI safety discourse, receiving $5M+ in funding and serving as the origin point for ~31% of EA survey respondent...Quality: 44/100-style rationality, though some EA community members find the LessWrong approach off-putting.73 Rebranding rationality insights as “good judgment” has been suggested as a way to bridge this divide.
The EA community has also debated whether good judgment and forecasting skill represent distinct capabilities.74 Good Judgment’s track record—with superforecasters outperforming prediction markets by 15-30% and intelligence analysts by 25-30%—provides empirical evidence relevant to these discussions.75 However, questions remain about whether forecasting accuracy on geopolitical questions translates to sound judgment on complex strategic questions like AI risk trajectories or optimal intervention strategies.
The potential application of Good Judgment’s methodology to AI safety forecasting remains largely unexplored. While the organization has demonstrated exceptional accuracy on questions with clear resolution criteria and relatively short time horizons (typically under 2 years), AI safety involves longer timeframes, deeper uncertainty, and questions that may be difficult to operationalize as verifiable predictions.76 Some EA community discussions have explored the use of superforecasting for questions like “Is power-seeking AI an existential risk?”, though with mixed views on the appropriateness of forecasting for such fundamental strategic questions.77
Criticisms and Limitations
Section titled “Criticisms and Limitations”Despite Good Judgment’s impressive track record, several criticisms and limitations have been identified in research and community discussions. A key methodological concern involves sample size effects on superforecaster identification. Research attempting to replicate the superforecasting hypothesis found that with a sample of only 195 participants and identification periods of less than a year, no superforecasters were identified.78 This suggests the effect may be difficult to reproduce in smaller pools and raises questions about the statistical robustness of identifying truly exceptional forecasters versus observing random variation in performance.
Inherent cognitive biases remain a challenge even for skilled forecasters. The effectiveness of judgmental forecasting approaches is fundamentally constrained by forecasters’ inherent biases, which can lead to inadequate forecasts and failure to acknowledge poor performance.79 While Good Judgment employs bias training, research indicates that the ability to engage in reflective thinking—interrogating initial gut feelings—appears to be partly innate and only partly developed through training.80 This suggests limits to how much training can improve forecasting accuracy.
Domain limitations represent another significant concern. The Good Judgment Project focused mainly on geopolitical questions, which may not generalize to other domains.81 The organization explicitly acknowledges focusing on “low- or messy-data questions” where machine learning models show limited promise, according to researchers Tetlock and Koehler.82 This narrow focus means Good Judgment’s methods may not transfer effectively to domains with different characteristics, such as technological forecasting, scientific predictions, or questions involving rapid capability shifts.
Aggregation failures challenge the “wisdom of crowds” assumption underlying some forecasting approaches. Research found that when 45 people were asked to answer a factual question, the average was significantly inaccurate, though some individual forecasters performed well.83 This highlights that aggregation methods matter considerably and simple averaging may not capture the value of superior forecasters. Good Judgment’s use of skilled superforecasters partially addresses this concern, but questions remain about optimal aggregation approaches.
Hybrid forecasting uncertainty clouds the comparison between human and machine learning approaches. Evidence on whether hybrid approaches combining human and machine learning forecasts actually outperform human-only forecasts remains unclear, with only anecdotal reports of better Brier scores from at least one participant.84 As AI capabilities advance, the relative value of human superforecasters versus algorithmic approaches may shift, potentially undermining Good Judgment’s competitive advantages.
Long-term and existential questions pose particular challenges for Good Judgment’s methodology. The organization’s track record primarily involves questions with resolution timeframes under 2 years and clear verification criteria.85 Questions about existential risks, transformative AI timelines, or century-scale social changes may be fundamentally different from the geopolitical questions where superforecasters have demonstrated accuracy. The applicability of Good Judgment’s methods to these higher-stakes, longer-horizon questions remains uncertain.
Key Uncertainties
Section titled “Key Uncertainties”Several important questions remain unresolved regarding Good Judgment’s methodology, performance, and potential applications:
Generalization across domains: Can superforecasting accuracy on geopolitical questions translate to other domains like technology forecasting, AI risk assessment, or scientific predictions? The organization’s focus on geopolitical and economic questions leaves open whether the same individuals and methods would perform comparably on fundamentally different question types.
Longevity of forecaster skill: Do superforecasters maintain their accuracy advantages over decades, or does performance degrade over time? While Good Judgment reports consistent performance from 2015 to present, the longest individual track records span only about 10-15 years, leaving uncertainty about multi-decade persistence of forecasting skill.
Scalability limits: As Good Judgment grows and employs more professional superforecasters, will the organization maintain its performance advantages? There may be limits to the supply of truly exceptional forecasters, and professionalization could introduce different incentives that affect forecasting accuracy.
AI-human hybrid approaches: How should superforecaster insights be combined with machine learning forecasts to achieve optimal accuracy? The organization is experimenting with hybrid AI-superforecasting integration, but optimal architectures and the long-term comparative advantage of human forecasters remain unclear.
Causal understanding versus prediction: Does accurate forecasting on geopolitical questions indicate deep causal understanding, or primarily pattern recognition and probabilistic reasoning? This distinction matters for whether superforecasters can provide valuable strategic guidance beyond point predictions.
Training effectiveness limits: How much can forecasting training improve accuracy for typical individuals, and what proportion of the population has the cognitive capacity to become highly skilled forecasters? Understanding these limits would inform decisions about investing in forecasting training versus other approaches to decision quality.
Optimal question framing: What characteristics of questions make them most suitable for superforecaster prediction versus other approaches like expert analysis, prediction markets, or algorithmic models? Good Judgment acknowledges focusing on “messy-data questions,” but the boundaries of this category remain imprecisely defined.
Sources
Section titled “Sources”Footnotes
Section titled “Footnotes”-
A Primer on Good Judgment Inc. and the Good Judgment Project ↩
-
Evidence on good forecasting practices from the Good Judgment Project - AI Impacts ↩
-
The Good Judgment Project: A Large Scale Test - Semantic Scholar ↩
-
The Good Judgment Project: Revolutionizing Forecasting - The Jenny Project ↩
-
Evidence on good forecasting practices from the Good Judgment Project - AI Impacts ↩
-
Evidence on good forecasting practices from the Good Judgment Project - AI Impacts ↩
-
Superforecasters Financial Times 2023 - Good Judgment Inc. ↩
-
The Good Judgment Project: Revolutionizing Forecasting - The Jenny Project ↩
-
Good Judgment Inc. Forecasts on US Foreign Aid Funding - GiveWell ↩
-
AI safety research data showed no direct Good Judgment involvement in alignment research ↩
-
Good Judgment Inc. Forecasts on US Foreign Aid Funding - GiveWell ↩
-
How Can Good Generalist Judgment Be Differentiated From Forecasting - EA Forum ↩
-
Based on analysis of Good Judgment’s track record focusing on questions with clear resolution criteria and typically under 2-year horizons ↩
-
Superforecasting the Premises in ‘Is Power-Seeking AI an Existential Risk?’ - Joe Carlsmith ↩
-
Evidence on good forecasting practices from the Good Judgment Project - AI Impacts ↩
-
Based on analysis of Good Judgment’s documented track record and resolved questions ↩