Critical Uncertainties Model
- Quant.The safety funding gap is approximately 33:1 (capability investment to safety research), with total AI safety funding at ~$100-650M annually versus $10B+ in capability development, representing a massive resource misallocation given expert risk assessments.S:3.5I:4.5A:4.5
- Quant.Deception detection capability in AI systems is currently estimated at only 30% true positive rate, with empirical evidence showing Claude 3 Opus strategically faked alignment in 78% of cases during reinforcement learning when facing conflicting objectives.S:4.5I:4.0A:4.0
- GapResolving just 10 key uncertainties could shift AI risk estimates by 2-5x and change strategic recommendations, with targeted research costing $100-200M/year potentially providing enormous value of information compared to current ~$20-30M uncertainty-resolution spending.S:4.0I:4.0A:4.5
- TODOComplete 'Conceptual Framework' section
- TODOComplete 'Quantitative Analysis' section (8 placeholders)
Overview
Section titled “Overview”Effective AI risk prioritization requires identifying which uncertainties most affect expected outcomes. This model maps 35 high-leverage variables across six domains—hardware, algorithms, governance, economics, safety research, and capability thresholds—to help researchers and policymakers focus evidence-gathering where it matters most. The central question: which empirical uncertainties, if resolved, would most change our strategic recommendations for AI safety?
The key insight is that a small number of cruxes—perhaps 8-12 variables—drive the majority of disagreement about AI risk levels and appropriate responses. Expert surveys consistently show wide disagreement on these specific parameters: the AI Impacts 2023 survey↗📄 paper★★★☆☆arXivAI Impacts 2023 surveyKatja Grace, Harlan Stewart, Julia Fabienne Sandkühler et al. (2024)Source ↗Notes found that 41-51% of AI researchers assign greater than 10% probability to human extinction or severe disempowerment from AI, yet the remaining researchers assign much lower probabilities. This disagreement stems primarily from differing estimates of alignment difficulty, takeoff speed, and governance tractability—all variables included in this model.
The model synthesizes data from multiple authoritative sources. Metaculus forecasts↗🔗 web★★★☆☆MetaculusMetaculusSource ↗Notes show AGI timeline estimates have collapsed from 50 years (2020) to approximately 5 years (2024), with current median around 2027-2031. Epoch AI research↗🔗 web★★★★☆Epoch AIEpoch AISource ↗Notes projects training compute could reach 10^28-10^30 FLOP by 2030, while their data analysis suggests high-quality training data may be exhausted by 2025-2028 depending on overtraining factors. These empirical findings directly inform the parameter estimates visualized below.
Core thesis: Focus on ~35 nodes that are (1) high-leverage, (2) genuinely uncertain, and (3) empirically resolvable or at least operationalizable.
Data Sources and Methodology
Section titled “Data Sources and Methodology”This model draws on multiple evidence streams to estimate parameter values and uncertainty ranges:
| Source Type | Examples | Variables Informed | Update Frequency |
|---|---|---|---|
| Expert Surveys | AI Impacts↗📄 paper★★★☆☆arXivAI Impacts 2023 surveyKatja Grace, Harlan Stewart, Julia Fabienne Sandkühler et al. (2024)Source ↗Notes, Pew Research↗🔗 web★★★★☆Pew Research CenterPew Research: Public and AI ExpertsA comprehensive study comparing perspectives of U.S. adults and AI experts on artificial intelligence's future, highlighting differences in optimism, job impacts, and regulatory...Source ↗Notes, International AI Safety Report↗🔗 webInternational AI Safety Report 2025The International AI Safety Report 2025 provides a global scientific assessment of general-purpose AI capabilities, risks, and potential management techniques. It represents a c...Source ↗Notes | Alignment difficulty, extinction probability, timeline estimates | Annual |
| Forecasting Platforms | Metaculus↗🔗 web★★★☆☆MetaculusMetaculusSource ↗Notes, Manifold, Polymarket | AGI timelines, capability milestones, policy outcomes | Continuous |
| Industry Reports | Stanford AI Index↗🔗 web★★★★☆Stanford HAIStanford AI Index 2025The 2025 AI Index Report from Stanford HAI offers a detailed analysis of AI's technological, economic, and social developments. It highlights key trends in performance, investme...Source ↗Notes, McKinsey State of AI↗🔗 web★★★☆☆McKinsey & CompanyMcKinsey State of AI 2025Source ↗Notes, Epoch AI↗🔗 web★★★★☆Epoch AIEpoch AISource ↗Notes | Compute trends, algorithmic progress, economic value | Annual |
| Governance Tracking | IAPP AI Governance↗🔗 webIAPP AI GovernanceSource ↗Notes, regulatory databases | Policy stringency, compliance rates, enforcement actions | Quarterly |
| Safety Research | Lab publications, interpretability benchmarks, red-team exercises | Alignment tax, deception detection, oversight limits | Ongoing |
Methodology notes: Parameter estimates use median values from multiple sources where available. Uncertainty ranges reflect the 10th-90th percentile of expert/forecaster distributions. The “resolvable via” column identifies empirical pathways to reduce uncertainty. Variables are classified as high-leverage if resolving uncertainty would shift expected loss by >$1T or change recommended interventions.
Key Expert Survey Findings
Section titled “Key Expert Survey Findings”The AI Impacts 2023 survey↗📄 paper★★★☆☆arXivAI Impacts 2023 surveyKatja Grace, Harlan Stewart, Julia Fabienne Sandkühler et al. (2024)Source ↗Notes of 2,788 AI researchers provides crucial data on expert disagreement:
| Finding | Value | Implications |
|---|---|---|
| P(human extinction or severe disempowerment) >10% | 41-51% of respondents | Wide disagreement on catastrophic risk |
| Alignment viewed as “harder” or “much harder” than other AI problems | 57% of respondents | Technical difficulty is a key crux |
| Median AGI timeline | 2047 (50% probability) | But 10% chance by 2027 |
| Timeline shift from 2022 to 2023 survey | 13-48 years earlier | Rapid updating toward shorter timelines |
The Pew Research survey↗🔗 web★★★★☆Pew Research CenterPew Research: Public and AI ExpertsA comprehensive study comparing perspectives of U.S. adults and AI experts on artificial intelligence's future, highlighting differences in optimism, job impacts, and regulatory...Source ↗Notes (2024) of 1,013 AI experts found 56% believe AI will have positive impact over 20 years vs. only 17% of the public—a 39-percentage-point gap that influences political feasibility of governance interventions.
Uncertainty Domain Structure
Section titled “Uncertainty Domain Structure”The following diagram illustrates how uncertainty domains interact to determine overall AI risk estimates. Arrows indicate primary causal influences—uncertainties in upstream domains propagate to downstream risk assessments.
Key structural insights:
- Hardware and algorithmic uncertainties jointly determine capability timelines
- Governance effectiveness influences both racing dynamics and direct risk mitigation
- Safety research progress depends on algorithmic advances (interpretability of more capable models) and governance (funding, coordination)
- Capability thresholds are the proximate determinant of misuse and accident risks
Variable Categories
Section titled “Variable Categories”Hardware & Compute (6 nodes)
Section titled “Hardware & Compute (6 nodes)”| Variable | Current Estimate | Uncertainty Range | Resolvable Via |
|---|---|---|---|
| GPU production growth | 2x / 2.5 years | 2x/18mo to 2x/4yr | Semiconductor roadmaps, investment data |
| Effective compute at frontier | 10^25 FLOP | 10^25 - 10^27 by 2027 | Model capabilities, energy consumption |
| Compute governance effectiveness | 3/10 | 1-7/10 by 2028 | Monitoring deployment, compliance rates |
| China-US compute gap | 0.3x | 0.1x - 0.8x | Intelligence estimates, published capabilities |
| Compute cost trajectory | -50%/year | -30% to -70%/year | Public pricing, efficiency benchmarks |
| Energy for AI | 50 TWh/year | 30-200 TWh by 2028 | Grid data, construction permits |
Algorithmic Progress (7 nodes)
Section titled “Algorithmic Progress (7 nodes)”| Variable | Current Estimate | Uncertainty Range | Resolvable Via |
|---|---|---|---|
| Algorithmic efficiency gains | 2x / 16 months | 2x/8mo to 2x/36mo | Benchmark performance at fixed FLOP |
| Scaling law breakdown point | ≈10^28 FLOP | 10^26 - 10^30 | Extrapolation from largest runs |
| Data quality ceiling | 40% remaining | 20-80% | Data availability studies |
| Post-training effectiveness | 0.5x pretrain | 0.2x - 2x | Capability comparisons |
| Sample efficiency breakthrough P | 30% by 2030 | 10-60% | Research publications, benchmarks |
| Architecture paradigm shift P | 25% by 2030 | 10-50% | Benchmark dominance, investment flows |
| Test-time compute scaling | +40% per 10x | +10% to +100% | Reasoning task benchmarks |
Coordination & Governance (8 nodes)
Section titled “Coordination & Governance (8 nodes)”The IAPP AI Governance Survey 2024↗🔗 webIAPP AI GovernanceSource ↗Notes found that only 25% of organizations have fully implemented AI governance programs despite 78% using AI—a 53-percentage-point gap. The Stanford AI Index 2025↗🔗 web★★★★☆Stanford HAIStanford AI Index 2025The 2025 AI Index Report from Stanford HAI offers a detailed analysis of AI's technological, economic, and social developments. It highlights key trends in performance, investme...Source ↗Notes reports U.S. federal agencies introduced 59 AI regulations in 2024, double the 2023 count.
| Variable | Current Estimate | Uncertainty Range | Resolvable Via |
|---|---|---|---|
| US regulatory stringency | 3/10 | 1-8/10 by 2028 | Legislation, agency rulemaking |
| US-China treaty probability | 10% by 2030 | 2-30% | Diplomatic initiatives, expert surveys |
| EU AI Act effectiveness | 4/10 | 2-7/10 | Audit reports, enforcement actions |
| Frontier lab coordination | 4/10 | 2-8/10 | Information sharing, deployment decisions |
| Compute monitoring % | 20% | 5-60% by 2028 | Monitoring tech deployment |
| Public concern trajectory | 25% | 15-60% | Polling data, media analysis |
| Whistleblower frequency | 0.5/year | 0.1-3/year | Disclosed incidents |
| International AI institution | 2/10 | 1-6/10 by 2030 | Institution-building efforts |
Governance maturity gap: According to Infosys research↗🔗 webInfosys researchSource ↗Notes, only 2% of companies meet gold-standard benchmarks for responsible AI controls—comprehensive controls, continuous monitoring, and proven effectiveness across the AI lifecycle.
Economic & Strategic (6 nodes)
Section titled “Economic & Strategic (6 nodes)”| Variable | Current Estimate | Uncertainty Range | Resolvable Via |
|---|---|---|---|
| AI economic value | $200B | $100B-$2T by 2028 | Revenue data, market caps |
| Winner-take-all strength | 5/10 | 3-8/10 | Concentration trends, imitation lag |
| Military AI advantage | 5/10 | 3-9/10 | Defense analysis, war games |
| Lab lead time | 9 months | 3-18 months | Benchmark timing, releases |
| Open source lag | 15 months | 6-24 months | Benchmark comparisons |
| Regulatory arbitrage | 5/10 | 3-8/10 | Company relocations |
Safety & Alignment (10 nodes)
Section titled “Safety & Alignment (10 nodes)”The safety funding gap↗🔗 web★★☆☆☆Mediumsafety funding gapSource ↗Notes is stark: capability investment to safety research is approximately 10,000:1 by some estimates. AI safety incidents surged 56.4% in 2024, with 233 documented failures. Total AI safety funding reached approximately $100-650M annually (including internal lab budgets), versus $10B+ in capability development. Mechanistic interpretability research, while progressing↗🔗 webMechanistic Interpretability for AI Safety — A ReviewSource ↗Notes, still lacks standardized metrics for “percentage of model explained.”
| Variable | Current Estimate | Uncertainty Range | Resolvable Via |
|---|---|---|---|
| Alignment tax | 12% | 5-30% | Benchmark aligned vs unaligned |
| Interpretability progress | 5% explained | 2-15% | Interpretability benchmarks |
| Scalable oversight limit | 10^27 FLOP | 10^25 - 10^29 | Red team exercises |
| Deception detection rate | 30% | 10-60% | Benchmark on adversarial models |
| Safety funding ratio | 1:33 | 1:10 - 1:100 | Lab budgets, grant funding |
| Safety researcher pipeline | 350/year | 200-800/year | Publications, hiring, graduates |
| Warning shot severity | 0 so far | $1B loss or 100+ deaths | Historical analysis |
| Warning → regulation lag | 18 months | 6-36 months | Case studies |
| Contained testing ratio | 20% | 5-50% | Lab practice surveys |
| Frontier lab security | 2 breaches/year | 0.5-5/year | Incident reports, audits |
Deception evidence (2024): Research showed Claude 3 Opus sometimes strategically answered prompts conflicting with its objectives to avoid retraining. When reinforcement learning was applied, the model faked alignment in 78% of cases—providing empirical grounding for deception detection estimates.
Capability Thresholds (5 nodes)
Section titled “Capability Thresholds (5 nodes)”| Variable | Current Estimate | Uncertainty Range | Resolvable Via |
|---|---|---|---|
| Autonomous AI R&D | 3 years away | 1-10 years | ML task benchmarks |
| Persuasion ceiling | 35% swayable | 20-60% | A/B testing, election analysis |
| Cyber offense capability | 30% infra vulnerable | 15-60% | Red team exercises |
| Bioweapon design capability | 60% expert-equivalent | 30-90% | Red team biology tasks |
| Strategic planning capability | 40% expert-equivalent | 20-70% | Strategy benchmarks |
Concrete Risk Costs (3 nodes)
Section titled “Concrete Risk Costs (3 nodes)”| Variable | Current Estimate | Uncertainty Range | Resolvable Via |
|---|---|---|---|
| Expected misalignment loss | $50T | $0 - $500T | Expert elicitation, modeling |
| Expected bio deaths (log) | 10^5 | 10^3 - 10^9 | Epidemiological modeling |
| Expected infra deaths (log) | 10^4 | 10^2 - 10^6 | Vulnerability studies |
Key Causal Relationships
Section titled “Key Causal Relationships”Strong Positive Influences (>50% variance explained)
Section titled “Strong Positive Influences (>50% variance explained)”- GPU growth → Effective compute → Algorithmic progress → TAI timeline
- Economic value growth → Racing incentives → Reduced safety investment
- Autonomous R&D capability → Recursive improvement → Fast takeoff probability
- Warning shot severity → Public concern → Regulatory stringency
Strong Protective Influences
Section titled “Strong Protective Influences”- Low alignment tax → Higher safety adoption
- High compute governance → Reduced China-US gap
- International coordination → Reduced racing dynamics
- Large lab lead time → More safety investment (less pressure)
Critical Uncertainties with High Influence
Section titled “Critical Uncertainties with High Influence”| Uncertainty | Affects | Resolution Timeline |
|---|---|---|
| Scaling law breakdown point | All timeline estimates | 2-4 years |
| US-China coordination possibility | Arms race vs cooperation | 3-5 years |
| Warning shot occurrence | Governance quality | Unpredictable |
| Deceptive alignment detection | Existential risk level | 2-5 years |
Key Interaction Effects
Section titled “Key Interaction Effects”| Interaction | Result |
|---|---|
| High economic value × Low lead time | Extreme racing pressure |
| High interpretability × Low alignment tax | Rapid safety adoption |
| Warning shot × Short regulation lag | Effective governance |
| High cyber capability × Low security | Fast capability diffusion to adversaries |
Strategic Importance
Section titled “Strategic Importance”Magnitude Assessment
Section titled “Magnitude Assessment”Critical uncertainties analysis identifies where research and evidence-gathering most affect risk estimates. Resolving high-leverage uncertainties changes optimal resource allocation.
| Dimension | Assessment | Quantitative Estimate |
|---|---|---|
| Potential severity | High - uncertainty multiplies expected costs | Resolution could shift risk estimates by 2-5x |
| Probability-weighted importance | High - current uncertainty drives conservative planning | ≈60% of risk estimate variance from 10 key uncertainties |
| Comparative ranking | Meta-level - determines research prioritization | Most valuable research reduces uncertainty on these variables |
| Resolution timeline | Variable - some resolvable in 1-2 years | 40% of key uncertainties addressable with $100M research investment |
Value of Information Analysis
Section titled “Value of Information Analysis”| Uncertainty | Resolution Cost | Timeline | Risk Estimate Shift Potential | VOI Estimate |
|---|---|---|---|---|
| Scaling law breakdown point | $50-100M | 2-4 years | Could shift timeline estimates by 3-5 years | Very High |
| Deception detection capability | $30-50M | 2-3 years | Could change alignment approach viability by 30-50% | Very High |
| US-China coordination feasibility | $20-30M | 3-5 years | Could shift governance strategy entirely | High |
| Alignment tax trajectory | $20-40M | 2-3 years | Could change safety adoption by 50-80% | High |
| Warning shot response time | $5-10M | Historical analysis | Could change intervention timing strategy | Medium |
Resource Implications
Section titled “Resource Implications”Prioritize research that resolves high-leverage uncertainties:
- Scaling law empirics: Fund large-scale capability forecasting ($50-100M/year)
- Deception detection: Accelerate interpretability and evaluation research ($30-50M/year)
- Governance feasibility studies: Diplomatic track-2 engagement, scenario planning ($20-30M/year)
- Historical case study analysis: Rapid response literature, regulatory speed ($5-10M/year)
Recommended uncertainty-resolution research budget: $100-200M/year (vs. ≈$20-30M current).
Key Cruxes
Section titled “Key Cruxes”| Crux | Resolution Method | If Resolved Favorably | If Resolved Unfavorably |
|---|---|---|---|
| Scaling laws continue | Empirical extrapolation | Standard timeline applies | Faster or slower than expected |
| Alignment tax is reducible | Technical research | Safety adoption accelerates | Racing dynamics intensify |
| Warning shots are informative | Historical analysis | Governance window exists | Must act on priors |
| International coordination possible | Diplomatic engagement | Global governance viable | Fragmented response |
Limitations
Section titled “Limitations”This model has several important limitations that users should consider when applying these estimates:
Parameter independence assumption. The model treats many variables as conditionally independent when in reality they may be deeply correlated. For example, “alignment tax” and “interpretability progress” likely share underlying drivers (researcher talent, algorithmic insights) that create correlations not captured in the causal graph. Sensitivity analyses should explore correlated parameter movements.
Expert survey limitations. Much of the underlying data comes from expert surveys, which have known biases: AI researchers may systematically under- or over-estimate risks in their own field; sample selection may exclude important perspectives; and framing effects can significantly shift probability estimates. The AI Impacts survey↗📄 paper★★★☆☆arXivAI Impacts 2023 surveyKatja Grace, Harlan Stewart, Julia Fabienne Sandkühler et al. (2024)Source ↗Notes notes that question phrasing affected timeline estimates by 13-48 years.
Rapidly changing ground truth. Parameter estimates become outdated quickly. Metaculus AGI forecasts shifted from 50 years (2020) to ~5 years (2024)—a factor of 10 change in four years. Users should check original sources for current values rather than relying on point estimates from this page.
Missing variables. The 35 variables selected represent a judgment call about what matters most. Potentially important factors not included: specific geopolitical events, individual actor decisions, Black Swan technological breakthroughs, and cultural/social dynamics that affect AI adoption and regulation.
Quantification precision. Many estimates (e.g., “deception detection rate: 30%”) represent rough order-of-magnitude guesses rather than empirically grounded values. The uncertainty ranges may themselves be overconfident about our ability to bound these quantities.
Related Models
Section titled “Related Models”This critical uncertainties framework connects to several other analytical tools in the knowledge base:
| Related Model | Relationship |
|---|---|
| Racing Dynamics ImpactModelRacing Dynamics Impact ModelQuantifies how competitive AI development reduces safety investment by 30-60% and increases alignment failure probability 2-5x through game-theoretic mechanisms, showing release cycles compressed f...Quality: 66/100 | Explores economic/strategic variables in depth |
| Capability Threshold ModelModelCapability Threshold ModelProvides systematic framework mapping 5 AI capability dimensions to specific risk thresholds with quantified timelines: authentication collapse 85% likely 2025-2027, bioweapons development 40% like...Quality: 68/100 | Details capability threshold variables |
| Lab Incentives ModelModelLab Incentives ModelAnalyzes how competitive, investor, reputation, and employee pressures shape AI lab safety decisions, estimating misaligned incentives contribute 10-25% of total AI risk. Identifies specific interv...Quality: 38/100 | Analyzes safety funding and coordination dynamics |
| Warning Signs ModelModelWarning Signs ModelSystematic framework analyzing 32 critical AI warning signs across five categories, finding most high-priority indicators are 18-48 months from threshold crossing with 45-90% detection probabilitie...Quality: 68/100 | Expands on warning shot and response lag variables |
| International Coordination GameModelInternational Coordination Game ModelGame-theoretic analysis demonstrates that US-China AI coordination faces Nash equilibrium at mutual defection (2,2 payoff) rather than Pareto-optimal cooperation (4,4), with defection dominating ma...Quality: 62/100 | Deep dive on US-China and multilateral dynamics |
Sources
Section titled “Sources”Expert Surveys
Section titled “Expert Surveys”- AI Impacts 2023 Survey↗📄 paper★★★☆☆arXivAI Impacts 2023 surveyKatja Grace, Harlan Stewart, Julia Fabienne Sandkühler et al. (2024)Source ↗Notes - 2,788 AI researchers on timelines, risks, and alignment difficulty
- Pew Research AI Survey 2024↗🔗 web★★★★☆Pew Research CenterPew Research: Public and AI ExpertsA comprehensive study comparing perspectives of U.S. adults and AI experts on artificial intelligence's future, highlighting differences in optimism, job impacts, and regulatory...Source ↗Notes - Expert vs. public opinion on AI impacts
- International AI Safety Report 2025↗🔗 webInternational AI Safety Report 2025The International AI Safety Report 2025 provides a global scientific assessment of general-purpose AI capabilities, risks, and potential management techniques. It represents a c...Source ↗Notes - 96 experts from 30 countries
Industry & Research Reports
Section titled “Industry & Research Reports”- Stanford AI Index 2025↗🔗 web★★★★☆Stanford HAIStanford AI Index 2025The 2025 AI Index Report from Stanford HAI offers a detailed analysis of AI's technological, economic, and social developments. It highlights key trends in performance, investme...Source ↗Notes - Comprehensive AI trends and metrics
- Epoch AI: Can Scaling Continue Through 2030?↗🔗 web★★★★☆Epoch AIEpoch AISource ↗Notes - Compute and data projections
- IAPP AI Governance Survey 2024↗🔗 webIAPP AI GovernanceSource ↗Notes - Organizational governance maturity
- McKinsey State of AI 2025↗🔗 web★★★☆☆McKinsey & CompanyMcKinsey State of AI 2025Source ↗Notes - Enterprise AI adoption and risk management
Forecasting Platforms
Section titled “Forecasting Platforms”- Metaculus AGI Forecasts↗🔗 web★★★☆☆MetaculusMetaculusSource ↗Notes - Community predictions on capability milestones
- 80,000 Hours: Shrinking AGI Timelines↗🔗 web★★★☆☆80,000 Hours80,000 Hours AGI Timelines ReviewBenjamin Todd (2025)A comprehensive review of expert predictions on Artificial General Intelligence (AGI) from multiple groups, showing converging views that AGI could arrive before 2030. Different...Source ↗Notes - Analysis of expert forecast evolution
Safety Research
Section titled “Safety Research”- Mechanistic Interpretability Review↗🔗 webMechanistic Interpretability for AI Safety — A ReviewSource ↗Notes - State of the field assessment
- AI Safety Funding Overview↗✏️ blog★★★☆☆LessWrongAn Overview of the AI Safety Funding Situation (LessWrong)Stephen McAleese (2023)Analyzes AI safety funding from sources like Open Philanthropy, Survival and Flourishing Fund, and academic institutions. Estimates total global AI safety spending and explores ...Source ↗Notes - Funding landscape analysis