Longterm Wiki
Navigation
Updated 2026-03-13HistoryData
Citations verified1 accurate1 unchecked
Page StatusContent
Edited today2.6k words4 backlinksUpdated every 6 monthsDue in 26 weeks
55QualityAdequate •5.5ImportancePeripheral5.5ResearchMinimal
Summary

Models what AI safety spending could accomplish at different budget levels from \$1B to \$50B+/year. Current global safety spending (~\$500M-1B/year) is 100-600x below capabilities investment. At \$5B/year, could fund 5,000+ dedicated safety researchers, comprehensive interpretability programs, and independent evaluation infrastructure. Key finding: absorptive capacity is the binding constraint below \$10B/year—the field cannot productively absorb unlimited funding without growing the researcher pipeline (current ~1,000 qualified safety researchers globally). Above \$10B/year, institutional capacity and research direction clarity become primary constraints. Provides concrete portfolio recommendations at each funding level.

Content8/13
LLM summaryScheduleEntityEdit history2Overview
Tables12/ ~11Diagrams2/ ~1Int. links22/ ~21Ext. links0/ ~13Footnotes0/ ~8References0/ ~8Quotes1/2Accuracy1/2RatingsN:7.5 R:5 A:8 C:7.5Backlinks4
Change History2
Remove legacy pageTemplate frontmatter3 weeks ago

Removed the legacy `pageTemplate` frontmatter field from 15 MDX files. This field was carried over from the Astro/Starlight era and is not used by the Next.js application.

opus-4-6 · ~10min

Migrate CAIRN pre-TAI capital pages#1554 weeks ago

Migrated 6 new model pages from CAIRN PR #11 to longterm-wiki, adapting from Astro/Starlight to Next.js MDX format. Created entity definitions (E700-E705). Fixed technical issues (orphaned footnotes, extra ratings fields, swapped refs). Ran Crux improve --tier=polish on all 6 pages for better sourcing, hedged language, and numeric EntityLink IDs. Added cross-links from 4 existing pages (safety-research-value, winner-take-all-concentration, racing-dynamics-impact, anthropic-impact).

Issues1
QualityRated 55 but structure suggests 80 (underrated by 25 points)

Safety Spending at Scale

Analysis

Safety Spending at Scale

Models what AI safety spending could accomplish at different budget levels from \$1B to \$50B+/year. Current global safety spending (~\$500M-1B/year) is 100-600x below capabilities investment. At \$5B/year, could fund 5,000+ dedicated safety researchers, comprehensive interpretability programs, and independent evaluation infrastructure. Key finding: absorptive capacity is the binding constraint below \$10B/year—the field cannot productively absorb unlimited funding without growing the researcher pipeline (current ~1,000 qualified safety researchers globally). Above \$10B/year, institutional capacity and research direction clarity become primary constraints. Provides concrete portfolio recommendations at each funding level.

Related
Analyses
Pre-TAI Capital Deployment: $100B-$300B+ Spending AnalysisAI Safety Research Value ModelAI Talent Market Dynamics
Approaches
AI Safety Field Building Analysis
2.6k words · 4 backlinks
InfoBox requires type or entityId

Overview

Current global AI safety spending is approximately $700M-1.25B per year, while AI capabilities investment exceeds $100B annually—a ratio of roughly 100:1 to 150:1, though these figures depend heavily on how "safety spending" is defined.1 The AI Safety Research Value Model suggests scaling safety funding 3-10x to $2-5B/year. The Pre-TAI Capital Deployment analysis shows that frontier labs alone may spend $100-300B+ over the coming years. This page examines what different levels of safety spending—$5B, $10B, or $50B annually—could accomplish given current field constraints.

This analysis models the absorptive capacity problem: the gap between theoretical safety spending and what the field can productively deploy. It provides portfolio recommendations at different funding levels and identifies bottlenecks that constrain effective scaling.

Central finding: Under these assumptions, below approximately $10B/year, the primary constraint is the researcher talent pipeline. Above that level, institutional capacity, research direction clarity, and the fundamental difficulty of the alignment problem become dominant constraints. This suggests that expanding safety spending requires parallel investment in talent development rather than funding alone.

Current Baseline

Global AI Safety Spending (2025 Estimates)

SourceAnnual SpendingGrowth RatePrimary Focus
Frontier AI labs (internal)$400-700M+40-50%/yearInterpretability, RLHF, evaluations, red-teaming
Philanthropic funders$150-250M+20-30%/yearAcademic research, field building, policy
Government$100-200M+50-100%/yearStandards (NIST), evaluation (UK AISI), military
Academia (dedicated safety)$50-100M+15-20%/yearTheoretical alignment, robustness, fairness
Total$700M-1.25B+35-50%/year

Current Safety Research Workforce

LevelEstimated CountAverage CompensationTotal CostNotes
Senior safety researchers150-300$600K-1.5M$200-400MAt frontier labs + top academic positions
Mid-level safety researchers500-1,000$300K-600K$200-500MLabs, research orgs, academia
Junior/entry-level1,000-2,000$100K-300K$150-400MPhD students, postdocs, early career
Adjacent researchers (ML safety-relevant)2,000-5,000$200K-500KNot countedWorking on safety-adjacent problems
Total dedicated safety workforce≈2,000-3,500$550M-1.3B

Approximately 2,000-3,500 people globally work full-time on AI safety research as of 2025.2

Absorptive Capacity Analysis

What Limits Safety Spending?

The concept of "absorptive capacity" models how much funding can be productively deployed at different scales. The binding constraint varies by funding level:

Loading diagram...

Binding Constraints by Funding Level

Funding LevelBinding ConstraintEstimated Absorptive RateKey Action Required
$1-2B/yrTalent (not enough researchers)70-85%Build pipeline; expand training programs
$2-5B/yrTalent + infrastructure50-70%Create new labs; fund compute access
$5-10B/yrInstitutional capacity40-60%Build organizations; develop research agendas
$10-30B/yrResearch direction clarity30-50%Need fundamental breakthroughs to guide resource allocation
$30-50B+/yrProblem difficulty20-40%Alignment difficulty may limit returns to additional resources

Note: "Absorptive rate" estimates represent the approximate fraction of funding that could be deployed on high-value safety research under current field constraints, based on researcher availability, institutional capacity, and research direction clarity. These are model estimates rather than empirical measurements.

Portfolio Recommendations by Funding Level

Level 1: $1-2B/year (2-4x Current)

Primary objective: Expand the researcher pipeline while maintaining research quality.

This funding level could be absorbed with moderate efficiency given current field constraints.

CategoryAllocationAnnual SpendWhat It Funds
Expand existing orgs30%$300-600MDouble safety teams at Anthropic, DeepMind, OpenAI; expand ARC Evaluations, MIRI, Redwood Research
Talent pipeline25%$250-500MPhD fellowships (500+), postdoc programs, bootcamps, career transitions
Academic programs20%$200-400M20-30 university safety research centers; endowed chairs
Evaluations & red-teaming15%$150-300MIndependent eval orgs; bug bounties; adversarial testing infrastructure
Governance research10%$100-200MPolicy research; regulatory frameworks; international coordination

Projected outcomes (3-5 year horizon):

  • Safety researcher workforce: 2,500-3,500 → 5,000-7,000
  • Independent evaluation capacity: 5-10x current levels
  • Academic safety programs: 10-20 → 50-80 universities
  • Capabilities-to-safety spending ratio: 100:1 → 50:1

Level 2: $5-10B/year (5-10x Current)

Primary objective: Build institutional safety research capacity independent of capability-focused labs.

At this level, safety research could develop organizational infrastructure comparable to other mature research disciplines.

CategoryAllocationAnnual SpendWhat It Funds
Dedicated safety labs25%$1.25-2.5B3-5 new independent safety research institutes at scale
Safety compute20%$1-2BDedicated compute clusters for safety research; model access
Expanded talent pipeline15%$750M-1.5B1,000+ new PhD positions; global training programs
Interpretability at scale15%$750M-1.5BLarge-scale mechanistic interpretability; automated tools
Evaluation infrastructure10%$500M-1BNational-scale evaluation facilities; continuous monitoring
Governance & institutions10%$500M-1BInternational coordination mechanisms; regulatory capacity building
Exploratory research5%$250-500MHigh-variance fundamental research; novel approaches

Projected outcomes (3-5 year horizon):

  • Safety researcher workforce: 10,000-15,000
  • Multiple independent labs with frontier model access
  • Continuous safety monitoring systems for deployed models
  • Established academic discipline with standardized curricula
  • Capabilities-to-safety spending ratio: 20:1 → 10:1

Level 3: $10-30B/year (Capabilities-Scale Funding)

Primary objective: Achieve resource parity with capabilities research in key safety domains.

At this scale, efficiency concerns increase substantially. The constraint shifts from researcher availability to clarity about productive research directions.

CategoryAllocationAnnual SpendWhat It Funds
Alignment R&D25%$2.5-7.5BLarge-scale alignment experiments; parallel research programs
Safety infrastructure20%$2-6BDedicated safety compute; monitoring systems; secure facilities
Human capital15%$1.5-4.5BGlobal researcher pipeline; competitive compensation
Interpretability15%$1.5-4.5BAutomated interpretability tools; comprehensive model understanding
Governance & policy10%$1-3BInternational institutions; regulatory implementation; standards
Applied safety10%$1-3BDeployment safety; incident response; societal resilience
Exploratory approaches5%$500M-1.5BHigh-risk/high-reward approaches; formal verification at scale

Level 4: $30-50B+/year (Maximum Plausible Scale)

At this level, safety investment approaches capabilities investment levels. The primary uncertainty shifts to whether additional resources yield proportional safety improvements.

Key question: If alignment is fundamentally difficult—requiring conceptual breakthroughs rather than engineering effort—then $50B may not provide substantially more value than $10B. Additional funding is most valuable if:

  1. Safety research benefits from parallelizable investigations across multiple approaches
  2. Compute-intensive experiments are resource-constrained rather than idea-constrained
  3. Applied safety engineering (monitoring, evaluation, deployment) can scale with model deployment

Highest-value applications at this scale:

  • Comprehensive safety infrastructure (monitoring, evaluation, incident response) scaling with deployment
  • Large-scale interpretability programs comparable to genome sequencing projects
  • Redundant, independent safety research programs to increase probability of breakthroughs
  • International safety institutions with enforcement capacity

The Talent Pipeline Problem

Loading diagram...

The current pipeline produces approximately 200-500 new safety researchers per year (see AI Talent Market Dynamics for detailed pipeline analysis). To productively absorb $5-10B/year would require 10,000-15,000 researchers—implying a 3-5x increase in the pipeline sustained over several years.

Pipeline Expansion Strategies

StrategyAnnual CostTimelineEstimated Scale ImpactQuality Considerations
PhD fellowships ($100K/year each)$500M for 1,000 positions4-6 yearsHighLow if selective
Postdoc bridge programs$200M/year1-3 yearsMediumLow
Industry → safety career transitions$100M/year1-2 yearsMediumMedium
University center creation$50M each, 20+ centers3-5 yearsHighLow-Medium
Bootcamps/intensive programs$10M/year for 500 seats3-6 monthsLow-MediumMedium-High
International programs (non-US/UK)$200M/year2-4 yearsHighMedium
Competitive compensation matching$300M+/year premiumImmediateMediumLow

The Compensation Gap

RoleIndustry (Lab)Safety Research OrgAcademicGap
Senior Researcher$800K-2M$250K-600K$120K-250K2-8x
Mid-level$400K-800K$150K-300K$80K-150K2-5x
Junior$200K-400K$80K-150K$40K-80K2-5x

Reducing this compensation gap by 50% for safety researchers would cost $300-500M/year across the field, and could increase talent flow into safety roles.

Historical Analogies for Scaling Safety Research

Nuclear Safety (1950s-1980s)

Nuclear weapons and power development provides the most direct historical parallel for scaling safety research in response to catastrophic risk.

PeriodAnnual Safety Spend (2024 $)ResearchersKey DevelopmentRatio to Capabilities Spend
1945-1955$50-200MHundredsBasic radiation protection standards≈1:100
1955-1970$500M-2BThousandsReactor safety frameworks; test ban treaty≈1:20
1970-1985$2-5B≈10,000NRC establishment; comprehensive regulatory standards≈1:10
Post-TMI (1979+)$5-10B15,000+Containment standards; probabilistic risk assessment≈1:5

Key patterns:

  • Safety investment scaled substantially after the Three Mile Island incident (1979)
  • The safety research apparatus required 20-30 years to develop institutional maturity
  • Safety spending reached 10-20% of capabilities investment but never exceeded it
  • Institutional capacity (NRC, IAEA) development was as important as research funding

Pharmaceutical Safety

MetricPharma SafetyAI Safety (Current)AI Safety (Modeled)
Safety as % of R&D15-25%1-3%5-10% (proposed)
Regulatory bodiesFDA, EMA, many national agenciesNIST AISI, UK AISI (nascent)To be determined
Independent testingRequired pre-deploymentVoluntary, inconsistentTo be determined
Workforce≈50,000 in safety/regulatory≈3,00010,000-30,000 (projected)
Time to maturity≈50 years (1938-1990s)≈5 years so farUnknown

Categories of Safety Work

Different types of safety work address different categories of risk. Understanding this distribution is important for evaluating spending effectiveness:

Technical Alignment Research

Focus: Ensuring AI systems reliably pursue intended objectives

Examples: Scalable oversight, interpretability, value learning

Primary risk addressed: Deceptive alignment, scheming, goal misgeneralization

Estimated current spending: $300-500M/year

Deployment Safety

Focus: Preventing harm from deployed systems

Examples: Content moderation, robustness testing, red-teaming

Primary risk addressed: Misuse, accidents, unintended consequences

Estimated current spending: $200-400M/year

Evaluation and Monitoring

Focus: Detecting dangerous capabilities and behaviors

Examples: ARC Evaluations, METR, Responsible Scaling Policies

Primary risk addressed: Surprise capability gains, deployment of unsafe systems

Estimated current spending: $100-200M/year

Governance and Policy

Focus: Institutional mechanisms for managing AI risk

Examples: International coordination, regulatory frameworks, standards development

Primary risk addressed: Race dynamics, inadequate oversight, coordination failures

Estimated current spending: $100-200M/year

Public Communication and Trust

Focus: Building public understanding and institutional legitimacy

Examples: Safety announcements, transparency reports, public engagement

Primary risk addressed: Loss of public trust, regulatory overreach, insufficient public input

Estimated current spending: $50-150M/year

Some work announced as "safety" primarily serves branding or regulatory compliance rather than reducing technical or catastrophic risk. Indicators that spending may not reduce existential risk include:

  • Safety announcements timed with product launches rather than research milestones
  • Focus on content moderation or PR-relevant metrics rather than catastrophic risk
  • Lack of independent oversight or publication of negative results
  • Evaluation programs that test user-facing features rather than dangerous capabilities

The OpenAI Foundation page documents related concerns in the philanthropic context. Whether public communication and trust-building constitute legitimate safety investments depends on whether institutional legitimacy is necessary for effective safety regulation—a question with reasonable disagreement.

Spending Efficiency Concerns

At larger scales, several failure modes become more likely:

Failure ModeRisk LevelExampleMitigation
Duplicated effortHigh at >$5BMultiple teams solving identical problemsCoordination bodies; shared infrastructure
Hiring for headcount targetsVery High at >$2BLowering standards to fill positionsMaintain hiring standards; accept slower growth
Infrastructure without directionHigh at >$10BCompute purchases without clear research questionsFund researchers first, compute second
Institutional overheadMedium at all levelsAdministrative costs consuming 30-40% of budgetLean organizations; direct grants
Misdirected researchHigh at all levelsFunding approaches unlikely to address core risksDiverse portfolio; regular evaluation

Counterarguments and Limitations

Cases Where More Spending Might Not Help

  1. Fundamental difficulty: If alignment requires conceptual breakthroughs rather than engineering effort, scaling spending may not proportionally reduce risk
  2. Information hazards: Publishing certain safety research could accelerate capabilities or provide blueprints for misuse
  3. Researcher quality dilution: Rapid expansion may lower average researcher quality below productivity thresholds
  4. Opportunity costs: Safety spending competes with other existential risk reduction (biosecurity, nuclear security)

Measurement Challenges

Evaluating safety research effectiveness faces fundamental difficulties:

  • Long feedback loops (decades until deployment outcomes are known)
  • Counterfactual dependence (success means preventing events that don't occur)
  • Dual-use research (safety work that also advances capabilities)
  • Publication bias (positive results published more than negative)

Geographic Distribution

Current safety research is concentrated in the US (≈50%), UK (≈20%), and Western Europe (≈15%), with limited presence in Asia, Latin America, and Africa. This concentration creates:

  • Single-points-of-failure if one jurisdiction restricts safety research
  • Limited cultural and technical diversity in safety approaches
  • Coordination challenges for international AI governance
  • Workforce constraints as non-US researchers face visa limitations

Timeline Sensitivity

The analysis above assumes 5-15 year timelines to transformative AI. If timelines are shorter (≈3 years), the emphasis shifts toward:

  • Deployment safety and monitoring over long-term alignment research
  • Existing researcher productivity over pipeline building
  • Government coordination over academic institution building

If timelines are longer (≈20+ years), emphasis shifts toward:

  • Fundamental theoretical research
  • Building sustainable academic institutions
  • International norm development

Recommendations

For Frontier AI Labs

  1. Publish safety spending — Verifiable figures in annual reports would enable external assessment
  2. Fund independent safety research — External researchers provide complementary perspectives
  3. Provide model access to safety researchers — Safety research requires access to frontier systems
  4. Consider minimum safety allocation commitments — Some labs (e.g., Anthropic) allocate 5-8% of budgets to safety; whether this should be an industry norm remains debated

For Philanthropic Funders

  1. Prioritize pipeline over projects at current funding levels—the bottleneck is researchers, not ideas
  2. Fund at competitive salaries to reduce talent flow from safety to capabilities roles
  3. Support institutional development — Safety research requires organizations, not just individual grants
  4. Invest in research direction clarity through field surveys, theory development, and agenda-setting

For Governments

  1. Fund public safety evaluation infrastructure — Independent testing capacity comparable to FDA drug approval
  2. Invest in academic safety programs through research grants, center funding, and fellowship programs
  3. Consider safety spending disclosure requirements — Transparency as precondition for informed regulation
  4. Support international coordination mechanisms to enable consistent standards

Key Uncertainties

UncertaintyImpact on AnalysisResolution Timeline
Is alignment fundamentally difficult?If yes, diminishing returns above $5-10B; if no, resources can solve itUnknown
Will labs voluntarily increase safety spend?Determines need for external mandates1-3 years
Can the talent pipeline scale 5-10x?Determines achievable scale of safety research3-5 years
Will safety insights transfer across architectures?Affects long-term value of current investmentOngoing
How much safety compute is needed?Determines whether infrastructure or talent is primary bottleneck2-3 years

Sources

Footnotes

  1. Estimates based on published safety team sizes from <EntityLink id="E22" name="anthropic">Anthropic</EntityLink>, <En... — Estimates based on published safety team sizes from <EntityLink id="E22" name="anthropic">Anthropic</EntityLink>, <EntityLink id="E218" name="openai">OpenAI</EntityLink>, and <EntityLink id="E98" name="deepmind">Google DeepMind</EntityLink> public reporting, and philanthropic grant databases including <EntityLink id="E552" name="open-philanthropy">Open Philanthropy</EntityLink> grants database (2024).

  2. Citation rc-138d (data unavailable — rebuild with wiki-server access)

Related Pages

Top Related Pages

Risks

SchemingDeceptive Alignment

Approaches

Mechanistic InterpretabilityAI Safety Training Programs

Analysis

Capability-Alignment Race ModelFrontier Lab Cost StructureAI Safety Researcher Gap Model

Safety Research

Scalable OversightInterpretability

Organizations

OpenAI FoundationRedwood ResearchAnthropicMETROpenAIMachine Intelligence Research Institute

Policy

International Coordination MechanismsResponsible Scaling Policies (RSPs)

Concepts

RLHF