Multi-Actor Strategic Landscape
AI Safety Multi-Actor Strategic Landscape
Analyzes AI risk through the lens of which actors develop TAI, finding actor identity may account for 40-60% of total risk variance, with detailed quantitative data on US-China capability convergence (benchmark gap narrowed from 9.26% to 1.70% in ~13 months), investment asymmetries ($222B US vs $98B China VC), and four risk pathways summing to ~25% combined x-risk. The framework is well-populated with 2025-2026 data but the core model parameters (risk estimates, variance attribution) are explicitly acknowledged as illustrative rather than empirically derived.
Core thesis: Risk is primarily determined by which actors develop TAI and their incentive structures. The strategic landscape of competition and cooperation shapes outcomes.
Overview
This model analyzes how AI existential risk depends on which actors—US frontier labs, Chinese developers, open-source communities, or malicious actors—develop Transformative AI first, and under what competitive conditions. The core insight is that actor identity and incentive structures may matter as much as technical alignment progress in determining outcomes. For a relational network analysis of how these actors connect and influence each other, see the AI Power and Influence Map.
The strategic landscape shifted substantially in 2024-2025. According to Recorded Future analysis↗🔗 webRecorded Future - US-China AI Gap 2025 AnalysisUseful for understanding the geopolitical and technical landscape of US-China AI competition, particularly relevant to governance and compute policy discussions around export controls and AI leadership.Recorded Future's intelligence analysis concludes that China is unlikely to sustainably surpass the US in AI capabilities by 2030, examining competitive dynamics across governme...capabilitiesgovernancepolicycompute+1Source ↗, the gap in overall model performance between the best US and Chinese models narrowed from 9.26% in January 2024 to just 1.70% by February 2025. AI Impacts documented the MMLU benchmark gap narrowing from 17.5 percentage points (end 2023) to 0.3 percentage points (end 2024).1 This convergence was catalyzed by DeepSeek's R1 release in January 2025, which matched OpenAI's o1 performance↗🔗 webmatched OpenAI's o1 performancePublished early 2025, this article provides accessible analysis of DeepSeek's significance for AI competition and US export control policy, relevant to debates about compute governance and geopolitical AI dynamics.This Boston University article examines DeepSeek, a Chinese AI model that reportedly matched OpenAI's o1 performance at a fraction of the cost, analyzing its implications for AI...capabilitiesgovernancepolicycompute+2Source ↗ while training for just $1.6 million. By April 2026, the top six Chatbot Arena Elo models were separated by only 20 Elo points, with Chinese models (Moonshot's Kimi K2.5, DeepSeek V3.2) occupying top positions.2 Similarly, open-source models closed to within 1.70%↗🔗 webopen-source models closed to within 1.70%Relevant to AI governance discussions around open-weight model release policies, export controls, and the geopolitical dimensions of AI development; provides empirical data on the shifting open-model ecosystem as of late 2025.A 2025 year-end analysis of open-model AI trends showing China surpassing the US in Hugging Face downloads for the first time, with Chinese models like Qwen and DeepSeek gaining...governancecapabilitiespolicyopen-source+4Source ↗ of frontier closed models on Chatbot Arena by early 2025, and the gap had effectively reached near-zero on knowledge benchmarks by early 2026,3 fundamentally changing proliferation dynamics.
Despite narrowing capability gaps, structural asymmetries persist in investment and compute infrastructure. US AI/ML venture capital investment reached approximately $222 billion in 2025—65.6% of total US VC deal value4—while China's AI capital spending reached approximately $98 billion (including government investment).5 On compute infrastructure, the US holds approximately 74.5% of global GPU cluster performance versus China's 14.1%, according to Epoch AI analysis covering ~417 GPU clusters (May 2025).6 China leads in deployment: it installed approximately 295,000 industrial robots in 2024—more than the rest of the world combined—and accounts for 69.7% of all AI patents.
Conceptual Framework
The Multi-Actor Strategic Landscape model structures AI risk analysis around four actor categories, three competitive pressure mechanisms, and four risk pathways. Unlike purely technical alignment models, it treats actor identity and incentive structures as primary determinants of AI risk outcomes.
Four actor categories:
- US Frontier Labs (OpenAI, Anthropic, Google DeepMind, Meta): Currently at the capability frontier; highest identified investment in both capabilities and safety research; subject to democratic accountability, though institutional oversight mechanisms remain underdeveloped (model parameter: ~0.35/1.0).
- Chinese Labs (DeepSeek, Baidu, Alibaba, ByteDance, Zhipu AI): Reached near-parity with US frontier labs on standard benchmarks by early 2026; development priorities shaped by government strategic objectives and different alignment research contexts.
- Open-Source Ecosystem (Llama, Qwen, DeepSeek open weights, Mistral): Near-zero capability gap on knowledge benchmarks as of early 2026; enables broad proliferation but also distributed safety research and external verification.
- Malicious Actors (cybercriminals, non-state armed groups, rogue states): Access capabilities primarily through open-source releases, jailbreaks, and API access; currently estimated at 40–60% of frontier capability.
Three competitive pressure mechanisms:
- Racing dynamics: Strategic competition between US and Chinese labs (model intensity parameter: 0.75/1.0) creates pressure to accelerate development timelines, potentially reducing safety investment.
- Capability diffusion: Open-source releases and technology transfer spread frontier capabilities to a broader set of actors, including potential malicious users.
- Safety investment tradeoffs: Profit pressure (model parameter: ~0.80/1.0) competes with safety investment, creating tension between deployment speed and risk reduction.
Four risk pathways (model estimates; see Limitations for caveats):
The four pathways—Unaligned Singleton (8%), Multi-Agent Conflict (6%), Authoritarian Lock-in (5%), and Catastrophic Misuse (7%)—sum to approximately 25% combined x-risk. These are model parameters rather than empirically derived estimates; they are illustrative of how actor identity affects pathway probabilities rather than precise forecasts.
Actor identity is estimated to account for 40–60% of total risk variance across these pathways, meaning differences in which actors lead AI development may matter comparably to technical alignment quality in determining outcomes. For governance frameworks relevant to these pathways, see the AI Governance Effectiveness Analysis.
Capability Gap Estimates (2025-2026)
The following table synthesizes publicly available data on relative AI capabilities across actor categories, updated with 2026 data. Estimates draw from benchmark performance, investment levels, and expert assessments.
| Actor Category | Capability vs Frontier | Trend | Key Evidence | Source |
|---|---|---|---|---|
| US Frontier Labs | 100% (reference) | Stable | GPT-4.5, Claude 3.5, Gemini 2.0 define frontier; top Chatbot Arena positions | Industry consensus |
| Chinese Labs (aggregate) | ≥99% on standard benchmarks | Converged | Gap narrowed from 9.26% (Jan 2024) to ≈1.70% (Feb 2025); Chinese models in top-6 Chatbot Arena (Apr 2026) | Recorded Future↗🔗 webRecorded Future - US-China AI Gap 2025 AnalysisUseful for understanding the geopolitical and technical landscape of US-China AI competition, particularly relevant to governance and compute policy discussions around export controls and AI leadership.Recorded Future's intelligence analysis concludes that China is unlikely to sustainably surpass the US in AI capabilities by 2030, examining competitive dynamics across governme...capabilitiesgovernancepolicycompute+1Source ↗;2 |
| DeepSeek specifically | ≈100% on standard benchmarks | At frontier | R1 matched o1 at $1.6M training cost; V3.2-Speciale gold at IMO 2025 and IOI 2025; V3.2 third-highest Arena Elo (≈1,421) | CSIS↗🔗 web★★★★☆CSISDeepSeek, Huawei, Export Controls, and the Future of the U.S.-China AI RaceRelevant for understanding how geopolitical compute restrictions interact with AI capabilities development, and whether hardware-focused governance strategies remain viable as algorithmic efficiency improves.This CSIS analysis examines how DeepSeek's emergence and Huawei's chip development challenge the effectiveness of U.S. export controls on advanced semiconductors. It assesses wh...governancepolicycomputecapabilities+2Source ↗;7 |
| Open-Source (Llama, Qwen, GLM-5) | ≥99% on knowledge benchmarks | Near-zero gap | GLM-5 top open-weight Arena Elo (≈1,451); Qwen3.5 leads open-weight GPQA at 88.4% (Feb 2026) | State of Open-Source AI↗🔗 webopen-source models closed to within 1.70%Relevant to AI governance discussions around open-weight model release policies, export controls, and the geopolitical dimensions of AI development; provides empirical data on the shifting open-model ecosystem as of late 2025.A 2025 year-end analysis of open-model AI trends showing China surpassing the US in Hugging Face downloads for the first time, with Chinese models like Qwen and DeepSeek gaining...governancecapabilitiespolicyopen-source+4Source ↗;38 |
| Malicious Actor Access | ≈40-60% | Increasing | Access via open-source, jailbreaks, or theft | Expert estimate |
Investment and Infrastructure Asymmetries (2024-2025)
| Dimension | United States | China | Notes |
|---|---|---|---|
| Private AI investment (2024) | $109 billion | ≈$9.3 billion | Stanford HAI 2025 AI Index; US figure is ≈12× China1 |
| AI/ML VC investment (2025) | ≈$222 billion | — | 65.6% of $339B total US VC; PitchBook Q4 20254 |
| AI capital spending (2025) | — | ≈$98 billion | Includes government investment; 48% YoY increase5 |
| Stargate / infrastructure commitment | $500B (4-year) | ¥380B (≈$52B, tech giants) | OpenAI Jan 20259; includes immediate $100B deployment |
| US hyperscaler data center capex (2025-2026) | >$350B / ≈$400B | — | Alphabet, Amazon, Microsoft, Meta, Oracle combined10 |
| GPU cluster performance share | 74.5% of global | 14.1% of global | Epoch AI, May 2025; covers ≈417 identified clusters6 |
| Data Centers | 4,049 | 379 | ≈11:1; RAND↗🔗 web★★★★☆RAND Corporationlargest single advantagePublished by RAND in May 2025, this commentary by Lennart Heim provides a strategic framing for the U.S.-China AI competition centered on compute infrastructure rather than model benchmarks, with detailed analysis of export control failures.RAND analyst Lennart Heim argues that while China is rapidly closing the gap on AI model capabilities, the U.S. retains a decisive advantage in total compute capacity and advanc...governancecomputepolicycapabilities+2Source ↗ |
| Industrial robot installations (2024) | 34,000 | 295,000 | 1:9; China leads deployment |
| AI patents (2023) | 13% of global | 69.7% of global | China dominates IP filings |
| AI research citations (2023) | 13% of global | 22.6% of global | China leads academic output |
Sources: CFR↗🔗 web★★★★☆Council on Foreign Relations\$109 billion in 2024A CFR policy analysis relevant to AI governance researchers tracking geopolitical competition in AI; useful background for understanding how US-China rivalry shapes global AI development and safety coordination prospects.A Council on Foreign Relations analysis examining the competitive dynamics between China and the United States in artificial intelligence development, with the $109 billion figu...governancepolicycapabilitiescompute+2Source ↗, RAND↗🔗 web★★★★☆RAND Corporationlargest single advantagePublished by RAND in May 2025, this commentary by Lennart Heim provides a strategic framing for the U.S.-China AI competition centered on compute infrastructure rather than model benchmarks, with detailed analysis of export control failures.RAND analyst Lennart Heim argues that while China is rapidly closing the gap on AI model capabilities, the U.S. retains a decisive advantage in total compute capacity and advanc...governancecomputepolicycapabilities+2Source ↗, Stanford HAI↗🔗 webChina AI vs. US in 2024: Key Findings from the Stanford HAI AI Index ReportThis is a third-party summary of the Stanford HAI AI Index Report focused on US-China AI competition; users seeking primary data should consult the full Stanford HAI report directly for comprehensive methodology and findings.This resource summarizes findings from the Stanford Human-Centered AI (HAI) Institute's AI Index Report comparing the United States and China across key AI metrics in 2024. It c...governancepolicycapabilitiescompute+2Source ↗, Epoch AI, PitchBook
Note: 2024 and 2025 investment figures use different methodologies (private investment vs. VC deal value vs. capital spending including government) and are not directly comparable across years or countries. China's 2025 figure includes substantial government-directed investment absent from the US VC measure.
Key Dynamics
The following diagram illustrates how actor competition dynamics flow through to risk outcomes:
Diagram (loading…)
flowchart TD
subgraph Actors["Actor Competition"]
USL[US Frontier Labs]
CNL[Chinese Labs]
OSS[Open-Source]
MAL[Malicious Actors]
end
subgraph Dynamics["Competitive Pressures"]
RACE[Racing Dynamics]
DIFF[Capability Diffusion]
SAFE[Safety Investment]
end
subgraph Outcomes["Risk Pathways"]
SING[Unaligned Singleton]
MULT[Multi-Agent Conflict]
AUTH[Authoritarian Lock-in]
MISUSE[Catastrophic Misuse]
end
USL -->|competes with| CNL
CNL -->|closes gap| RACE
USL -->|pressure to lead| RACE
OSS -->|enables access| DIFF
DIFF -->|reaches| MAL
RACE -->|reduces| SAFE
SAFE -->|insufficient| SING
CNL -->|if wins| AUTH
MAL -->|enables| MISUSE
RACE -->|increases| MULT
style USL fill:#e6f3ff
style CNL fill:#ffe6e6
style OSS fill:#e6ffe6
style MAL fill:#ffcccc
style SING fill:#ffdddd
style MULT fill:#ffdddd
style AUTH fill:#ffdddd
style MISUSE fill:#ffddddThe key mechanisms are:
-
Competition intensity → Safety shortcuts → Misalignment risk: As US-China competition intensifies (currently 0.75 on normalized scale), labs face pressure to accelerate timelines, potentially cutting safety corners.
-
Capability diffusion → Malicious access → Misuse risk: Open-source releases (now near-zero gap from frontier on knowledge benchmarks) enable rapid proliferation to actors who may lack safety constraints or beneficial intent.
-
First-mover advantage → Winner-take-all → Reduced caution: If decisive strategic advantage exists for first-mover, actors rationally accept higher alignment risk to capture it.
-
Democratic oversight → Deployment delays → Capability gaps: Strong oversight in democratic nations may create windows where authoritarian actors gain advantages, creating perverse incentives against regulation.
-
Transparency → Better coordination → Reduced racing: Conversely, capability transparency and safety research sharing (currently ~0.6 openness) can reduce competitive pressure.
Quantitative Analysis
US-China Benchmark Gap: Trajectory (2023-2026)
Benchmark gaps between US and Chinese frontier models narrowed substantially across all major evaluations.1
| Benchmark | US–China Gap (End 2023) | US–China Gap (End 2024) | Direction |
|---|---|---|---|
| MMLU | 17.5 pp | 0.3 pp | Near-parity |
| MMMU | 13.5 pp | 8.1 pp | Closing |
| MATH | 24.3 pp | 1.6 pp | Near-parity |
| HumanEval | 31.6 pp | 3.7 pp | Near-parity |
| Composite benchmark | 9.26% (Jan 2024) | 1.70% (Feb 2025) | Near-parity |
By April 2026, the top six Chatbot Arena models were separated by only 20 Elo points, with Chinese models (Kimi K2.5, DeepSeek V3.2 at Arena Elo ~1,421) occupying top-tier positions alongside Western models.2 DeepSeek V3.2-Speciale achieved gold-level results at IMO 2025, CMO, ICPC World Finals, and IOI 2025.7
Open-Source vs. Frontier Gap (2025-2026)
By early 2026, the gap between open-source and proprietary AI had effectively closed on knowledge benchmarks:3
| Model | Organization | GPQA Diamond | Arena Elo | Status |
|---|---|---|---|---|
| GLM-5 (Reasoning) | Zhipu AI | 94% | ≈1,451 (top open-weight) | Open-weight |
| Qwen3.5-397B-A17B | Alibaba | 88.4% | Top open-weight tier | Open-weight |
| DeepSeek V3.2 | DeepSeek | — | ≈1,421 (3rd overall) | Open-weight |
| Top proprietary (reference) | Various | ≈90-92% | ≈1,440-1,460 | Proprietary |
Single-digit gaps remain on frontier reasoning tasks (e.g., OpenAI o3-class models on competition mathematics).8
AI Safety and Alignment Research Funding (2024-2025)
| Funding Source | Amount | Year | Notes |
|---|---|---|---|
| Open Philanthropy (technical AI safety) | ≈$50M committed | 2024 | Largest single philanthropic funder; grants to CAIS, Redwood Research, Machine Intelligence Research Institute (MIRI) |
| Open Philanthropy (RFP) | $40M available | 2025 | Dedicated request for proposals |
| Survival and Flourishing Fund (SFF) (SFF) | $34.33M distributed | 2025 | 86% to AI safety |
| Long-Term Future Fund (LTFF) (LTFF) | $8.4M total | 2024 | Over $5M to AI safety |
| Total philanthropic AI safety | ≈$83.6M | 2024 | Open Philanthropy ≈76%; individual donors (Jaan Tallinn ≈$20M) ~24% |
| Alignment research funding (all sources) | $150-170M | Mid-2025 | 36% increase from $110-130M in 2024 |
| Broader AI safety market | >$1 billion | 2025 | Includes commercial/VC safety-focused startups; Safe Superintelligence raised $2B in 2025 |
Source: Stephen McAleese, LessWrong, January 202511; AI Safety Market Funding Trends 2022-202612
AI Governance Institutional Capacity (2025-2026)
| Institution | Funding | Year | Notes |
|---|---|---|---|
| NIST AI Safety Institute (US) | ≈$47.7M | FY2025 | AI research, standards, and testing |
| NIST AI Safety Institute (US) | ≈$55M | FY2026 | Includes AI Standards and Innovation Center ($10M) |
| UK AI Security Institute (AISI) research programme | £15M+ | 2025 | Initial £4M pot expanding to £8.5M; ranges from £50K to £1M per grant |
| UK DSIT/AISI foundation investment | £100M | 2023-2024 | Original Frontier AI Taskforce commitment |
| EU AI Office | Operational | Aug 2025 | 125+ staff; no public budget figure available |
| Global public AI governance (US + UK + EU combined est.) | ≈$72.3M (projected) | 2025 | Federal agencies; does not include EU AI Office |
Sources: NIST FY2026 budget announcement13; UK AISI research programme14; EU AI Office15
Safety Spend as Share of Total AI Investment
Identified alignment research funding ($150-170M in mid-2025) and public governance budgets (≈$72M) together represent approximately 0.06-0.09% of global AI VC investment ($243.9B in 2025).16 Actor-focused governance specifically—covering capability intelligence, game-theoretic analysis of lab incentives, and policy engagement—is estimated as a smaller subset of this total. Lab-level safety investment percentages are not publicly disclosed; estimates in the literature (e.g., Anthropic ≈30%, OpenAI ≈15%) are inferred from organizational statements and cannot be independently verified.
Risk Pathways
| Pathway | Description | Model Estimate |
|---|---|---|
| Unaligned Singleton | One misaligned AI gains decisive advantage | 8% |
| Multi-Agent Conflict | Multiple powerful AI systems in conflict | 6% |
| Authoritarian Lock-in | AI enables permanent authoritarian control | 5% |
| Catastrophic Misuse | Intentional misuse causes catastrophe | 7% |
| Combined X-Risk | Total from all pathways | ≈25% |
These are model parameters illustrating how actor identity affects pathway probabilities. They are not empirically derived forecasts.
Actor Categories
| Category | Key Actors |
|---|---|
| Leading US | OpenAI, Anthropic, Google DeepMind, Meta |
| Leading China | DeepSeek, Baidu, Alibaba, ByteDance, Zhipu AI |
| Open-Source | Meta (Llama), Mistral, Hugging Face ecosystem, Alibaba (Qwen) |
| Malicious | Cybercriminals, terrorists, rogue states |
| Governments | US (NSA, DARPA), China (PLA, MSS), EU |
Full Variable List
This diagram simplifies the full model. The complete Multi-Actor Strategic Landscape includes:
Actor Capabilities (15 variables): Leading US lab, leading Chinese lab, US government AI, Chinese government AI, open-source ecosystem, second-tier corporate labs, academic research, cybercriminal AI, terrorist access, authoritarian regime AI, democratic allies AI, corporate espionage, state IP theft, insider threat, supply chain security.
Actor Incentives (12 variables): US-China competition, profit pressure, academic openness, classification levels, democratic accountability, authoritarian control, geopolitical crises, economic desperation, military doctrine, regulatory arbitrage, talent mobility, public-private partnerships.
Information & Transparency (7 variables): Capability disclosure, safety sharing, incident reporting, capability intelligence, dual-use publication norms, evaluation standards, third-party verification.
Alignment & Control (8 variables): US actor alignment, China actor alignment, Constitutional AI effectiveness, human oversight scalability, kill switch reliability, containment protocols, red-teaming, post-deployment monitoring.
Strategic Outcomes (8 variables): First-mover advantage, winner-take-all dynamics, diffusion speed, multipolar vs bipolar, offense-defense balance, escalation control, governance lock-in, misuse probability.
Existential Risk Paths (5 variables): Unaligned singleton, multi-agent conflict, authoritarian lock-in, economic/social collapse, combined risk.
Strategic Importance
Magnitude Assessment
The multi-actor landscape determines whether AI development is coordinated or conflictual. Actor heterogeneity creates both risks (racing, proliferation) and opportunities (diverse approaches).
| Dimension | Assessment | Quantitative Estimate |
|---|---|---|
| Potential severity | Multipolar dynamics drive racing and proliferation | Actor landscape estimated to contribute 40-60% of total risk variance (model parameter) |
| Probability-weighted importance | Currently in competitive multipolar phase | 75% model-estimated probability of continued multipolar competition through 2030 |
| Comparative ranking | Essential context for governance and coordination strategies | Priority comparable to technical alignment in governance-focused analyses; ordering varies by framework |
| Malleability | Actor incentives partially shiftable via policy | 20-30% of racing dynamics estimated as addressable via policy (model parameter; unverified benchmark) |
Actor Safety Assessment
Lab-level safety investment percentages are not publicly disclosed. The following table describes observable indicators rather than inferred grades. For governance effectiveness context, see the AI Governance Effectiveness Analysis.
| Actor Category | Observable Safety Indicators | External Transparency | Notes |
|---|---|---|---|
| Anthropic | Responsible Scaling Policy published; Constitutional AI methodology; regular system cards | High | Organization states safety as core mission; ≈1,500 employees (early 2026); valuation grew to $380B (2025 Series G) |
| OpenAI | Safety council established 2025; Head of Preparedness role created (late 2025); Preparedness team active | Medium | Safety team structure changed during 2024-2025; targeting ≈8,000 employees by end 2026 |
| Google DeepMind | Published AGI safety and security strategy (2025); multiple alignment papers; ≈6,000 employees (Aug 2025) | Medium | DeepMind and Google Brain merged 2023; safety research is published |
| Meta AI | Open-source model releases with safety evaluations; FAIR research published | High (open-source) | Reduced AI safety research headcount Oct 2025; ≈3,400 lab staff (Jul 2025) |
| Chinese Labs | Model cards and safety system cards published (DeepSeek, Alibaba Qwen); affiliated university alignment research | Lower relative to Western labs in Western-accessible venues | Have published safety research; Chinese labs have lower external transparency to Western research community |
| Open-Source Ecosystem | Distributed responsibility; community red-teaming; open safety research | Very High | Variable implementation across projects; no central coordinating body |
Note: Safety investment percentages (e.g., "30% of budget") circulate in AI governance literature but are not drawn from public financial disclosures. No lab has confirmed specific safety budget percentages.
Diffusion Timeline Estimates
| Capability Level | US Labs | Chinese Labs | Open-Source | Malicious Actors |
|---|---|---|---|---|
| GPT-4 class | 2023 | 2024-2025 ✓ | 2024-2025 ✓ | 2025-2026 |
| Frontier reasoning (o1/R1 class) | 2024-2025 | 2025 ✓ (DeepSeek R1) | 2025-2026 | 2027-2028 |
| GPT-5 class (projected) | 2025 | 2026 (approaching) | 2027-2028 | 2028-2030 |
| Autonomous agents (dangerous capability threshold) | 2025-2026 | 2026-2027 | 2027-2028 | 2028-2029 |
Key Finding: The open-source and Chinese lab lags have collapsed on standard benchmarks. As of late 2025, the center of gravity for open-weight models has shifted toward China↗🔗 webopen-source models closed to within 1.70%Relevant to AI governance discussions around open-weight model release policies, export controls, and the geopolitical dimensions of AI development; provides empirical data on the shifting open-model ecosystem as of late 2025.A 2025 year-end analysis of open-model AI trends showing China surpassing the US in Hugging Face downloads for the first time, with Chinese models like Qwen and DeepSeek gaining...governancecapabilitiespolicyopen-source+4Source ↗, with DeepSeek and Qwen becoming leading providers. US firms released fewer open-weight models citing commercial and safety constraints, while Chinese labs treated open-weight leadership as a deliberate strategy. Meta—long a champion of frontier open models—has delayed release of Llama Behemoth↗🔗 web★★★☆☆CNBCdelayed release of Llama BehemothRelevant to AI safety discussions around open-source vs. proprietary model release strategies, as Meta's potential move toward closed frontier models may affect AI access, safety oversight, and competitive dynamics in the industry.CNBC reports that Meta is pursuing a new proprietary frontier AI model codenamed 'Avocado,' marking a significant strategic shift away from its open-source Llama models. The com...capabilitiesopen-sourcedeploymentgovernance+2Source ↗ and indicated it may keep future frontier models behind paywalls.
The "GPT-5 class" row should be interpreted cautiously: Chinese labs achieved near-parity with GPT-4o-comparable models in 2025, making the "GPT-5 class" timeline dependent on where proprietary labs define the next capability threshold.
First-Mover Advantage: Evidence Assessment
The model's risk estimates depend critically on the magnitude of first-mover advantage. Strong first-mover advantages create racing incentives; weak ones reduce them. Current evidence suggests first-mover advantages are significant but not overwhelming:
| Evidence Type | Finding | Implication for FMA |
|---|---|---|
| Historical analysis | First movers have 47% failure rate; only 11% become market leaders (Golder & Tellis↗🔗 webWhat Is a First Mover? The Complete Guide to First-Mover AdvantageThis is a general business strategy resource with no direct AI safety content; it may be tangentially relevant for understanding competitive dynamics around AI deployment or racing incentives in AI development contexts.A business strategy guide explaining first-mover advantage, distinguishing between being 'first-to-market' (technical achievement) and being a true 'first mover' (establishing a...capabilitiescoordinationdeploymentSource ↗) | Suggests weak FMA |
| AI competitive landscape | 2,011 companies in 2024 ML/AI landscape, 578 new entrants since 2023 | Suggests weak FMA |
| Model replication | 11 different developers globally achieved GPT-4-level models in 2024 | Suggests weak FMA |
| Cloud market | AWS, Azure, and Google Cloud trading leadership position; "more than one winner" possible | Suggests moderate FMA |
| Network effects | AI systems less network-effect-driven than social platforms | Suggests weak FMA |
| TAI-specific dynamics | Decisive strategic advantage at TAI level remains uncertain | Unknown |
Evidence from the Abundance Institute↗🔗 webVibrant AI Competitive Landscape - Abundance InstitutePublished by the Abundance Institute, this article examines the competitive AI landscape; content could not be fully retrieved, so metadata reflects reasonable inference from URL and title. Verify before citing.An analysis of the competitive dynamics in the AI industry, examining the landscape of companies, capabilities, and market forces shaping AI development. The piece likely covers...capabilitiesgovernanceai-safetydeployment+2Source ↗ suggests "no signs of winner-take-all dynamics" in the current AI ecosystem. However, TAI (transformative AI) may differ qualitatively if it enables rapid capability improvements or strategic advantages not available to followers. The model's 0.7 first-mover advantage estimate may be calibrated too high based on current evidence, but TAI-level dynamics remain highly uncertain.
Resource Implications
Understanding the actor landscape enables targeted governance strategies:
- Targeted engagement with highest-leverage actors: Focus on top 3-4 US labs could cover 70% of frontier capability
- Coalition-building for safety standards: Anthropic-OpenAI-DeepMind coalition would set de facto standards
- Monitoring of capability diffusion: $50-100M/year for comprehensive capability intelligence
- Anticipation of strategic behavior: Game-theoretic modeling estimated at ≈$10-20M/year
Estimated investment required for comprehensive coverage: Analyses estimate that actor-focused governance—capability intelligence, lab engagement, diffusion monitoring—may require $100-200M/year. Current philanthropic alignment research funding reached approximately $150-170M by mid-2025, though most is oriented toward technical safety rather than actor-focused governance specifically.11 The $20-30M estimate for current actor-focused governance work specifically is an analytical estimate without a publicly verified benchmark.
Key Cruxes
These probability estimates are model author estimates; they are not drawn from formal elicitation surveys.
| Crux | If True | If False | Model Estimate |
|---|---|---|---|
| Leading coalition is stable | Top 3 can set norms | Racing to bottom | 45% |
| Safety can be coordination point | Voluntary standards viable | Regulation required | 35% |
| China is engageable on safety | Global coordination possible | Bifurcated governance | 30% |
| Diffusion to malicious actors is slow | Window for governance | Proliferation dominates | 50% |
Multipolar vs Unipolar Governance Considerations
A crucial variable in this model is whether AI development converges toward unipolar (single dominant actor or coalition) or multipolar (distributed power among multiple actors) outcomes. Each presents distinct risk profiles:
| Governance Structure | Key Risks | Key Advantages |
|---|---|---|
| Unipolar (single dominant actor) | Value lock-in, institutional stagnation, internal corruption, single points of failure | Coordination easier, racing reduced, unified safety standards |
| Multipolar (distributed power) | Unchecked proliferation, system instability, coordination failures, racing dynamics | Diversity of approaches, no single point of failure, competitive pressure for safety |
Current research from AI Impacts↗🔗 web★★★☆☆AI ImpactsMultipolar Research Projects - AI ImpactsPart of AI Impacts' broader research agenda; useful for those studying competitive AI development scenarios and coordination problems as alternatives to single-actor AI takeover models.This AI Impacts page outlines proposed research projects focused on multipolar AI scenarios, where multiple powerful AI systems or actors compete rather than a single dominant s...ai-safetycoordinationexistential-riskgovernance+2Source ↗ identifies key research questions: What "considerations might tip us between multipolar and unipolar scenarios"? What "risks [are] distinctive to a multipolar scenario"? The CO/AI analysis↗🔗 webAI Multipolarity Gains Importance in Global Tech LandscapeThis is a CO/AI news digest summarizing a LessWrong post on multipolar vs. unipolar AI governance; useful as an entry point to decentralized AI governance debates but secondary to the original LessWrong source.This analysis summarizes a LessWrong post arguing that multipolar AI governance frameworks—featuring decentralized control among diverse AI agents and human actors—offer a compe...governancecoordinationai-safetyexistential-risk+3Source ↗ notes that while current AI safety discussions often default to unipolar frameworks, "exploring decentralized governance structures could address key risks like value lock-in and institutional stagnation."
Current assessment: The model estimates 55% probability of continued multipolar development, with the US-China bifurcation appearing increasingly stable. Geopolitical tensions, divergent regulatory approaches, and the collapse of open-source lags all point toward a world with multiple competing AI powers rather than a single dominant actor.
Limitations
-
Capability estimates rapidly outdating: The 2025-2026 data showing near-parity may not persist; breakthrough capabilities could restore gaps. The near-zero benchmark gap reflects standard evaluations and may not extend to capabilities at the frontier of reasoning, agentic tasks, or future benchmark domains.
-
Safety investment data opaque: Lab safety budgets are not publicly disclosed; estimates in circulation are inferential and unverified. The Actor Safety Assessment table reflects observable indicators rather than confirmed investment levels.
-
TAI dynamics uncertain: Current competitive patterns may not predict TAI-level dynamics where decisive advantages could differ fundamentally.
-
Geopolitical volatility: US-China relations, export control effectiveness, and regulatory trajectories are highly uncertain. Export controls on advanced chips may slow Chinese compute accumulation in ways not yet reflected in 2026 data.
-
Malicious actor access hard to estimate: Underground markets and state-sponsored theft create significant uncertainty in capability diffusion.
-
Model parameters are illustrative: The causal graph confidence values and risk pathway percentages are analytical estimates to illustrate model structure, not empirically derived posterior probabilities. The 40-60% risk variance attribution figure is a model claim without a published methodology or independent verification.
Sources
- Recorded Future: US-China AI Gap Analysis (2025)↗🔗 webRecorded Future - US-China AI Gap 2025 AnalysisUseful for understanding the geopolitical and technical landscape of US-China AI competition, particularly relevant to governance and compute policy discussions around export controls and AI leadership.Recorded Future's intelligence analysis concludes that China is unlikely to sustainably surpass the US in AI capabilities by 2030, examining competitive dynamics across governme...capabilitiesgovernancepolicycompute+1Source ↗
- RAND: China's AI Models Closing the Gap (2025)↗🔗 web★★★★☆RAND Corporationlargest single advantagePublished by RAND in May 2025, this commentary by Lennart Heim provides a strategic framing for the U.S.-China AI competition centered on compute infrastructure rather than model benchmarks, with detailed analysis of export control failures.RAND analyst Lennart Heim argues that while China is rapidly closing the gap on AI model capabilities, the U.S. retains a decisive advantage in total compute capacity and advanc...governancecomputepolicycapabilities+2Source ↗
- Council on Foreign Relations: China, the United States, and the AI Race↗🔗 web★★★★☆Council on Foreign Relations\$109 billion in 2024A CFR policy analysis relevant to AI governance researchers tracking geopolitical competition in AI; useful background for understanding how US-China rivalry shapes global AI development and safety coordination prospects.A Council on Foreign Relations analysis examining the competitive dynamics between China and the United States in artificial intelligence development, with the $109 billion figu...governancepolicycapabilitiescompute+2Source ↗
- Boston University: DeepSeek and AI Frontier (2025)↗🔗 webmatched OpenAI's o1 performancePublished early 2025, this article provides accessible analysis of DeepSeek's significance for AI competition and US export control policy, relevant to debates about compute governance and geopolitical AI dynamics.This Boston University article examines DeepSeek, a Chinese AI model that reportedly matched OpenAI's o1 performance at a fraction of the cost, analyzing its implications for AI...capabilitiesgovernancepolicycompute+2Source ↗
- State of Open-Source AI 2025↗🔗 webopen-source models closed to within 1.70%Relevant to AI governance discussions around open-weight model release policies, export controls, and the geopolitical dimensions of AI development; provides empirical data on the shifting open-model ecosystem as of late 2025.A 2025 year-end analysis of open-model AI trends showing China surpassing the US in Hugging Face downloads for the first time, with Chinese models like Qwen and DeepSeek gaining...governancecapabilitiespolicyopen-source+4Source ↗
- CSIS: DeepSeek, Huawei, and US-China AI Race↗🔗 web★★★★☆CSISDeepSeek, Huawei, Export Controls, and the Future of the U.S.-China AI RaceRelevant for understanding how geopolitical compute restrictions interact with AI capabilities development, and whether hardware-focused governance strategies remain viable as algorithmic efficiency improves.This CSIS analysis examines how DeepSeek's emergence and Huawei's chip development challenge the effectiveness of U.S. export controls on advanced semiconductors. It assesses wh...governancepolicycomputecapabilities+2Source ↗
- Abundance Institute: AI Competitive Landscape↗🔗 webVibrant AI Competitive Landscape - Abundance InstitutePublished by the Abundance Institute, this article examines the competitive AI landscape; content could not be fully retrieved, so metadata reflects reasonable inference from URL and title. Verify before citing.An analysis of the competitive dynamics in the AI industry, examining the landscape of companies, capabilities, and market forces shaping AI development. The piece likely covers...capabilitiesgovernanceai-safetydeployment+2Source ↗
- AI Impacts: Multipolar Research Projects↗🔗 web★★★☆☆AI ImpactsMultipolar Research Projects - AI ImpactsPart of AI Impacts' broader research agenda; useful for those studying competitive AI development scenarios and coordination problems as alternatives to single-actor AI takeover models.This AI Impacts page outlines proposed research projects focused on multipolar AI scenarios, where multiple powerful AI systems or actors compete rather than a single dominant s...ai-safetycoordinationexistential-riskgovernance+2Source ↗
- Frontier Model Forum: Progress Update 2024↗🔗 web★★★☆☆Frontier Model ForumFrontier Model ForumThe Frontier Model Forum is an industry body founded by Anthropic, Google, Microsoft, and OpenAI to coordinate on AI safety; this update reflects industry-led self-governance efforts as of mid-2024.The Frontier Model Forum (FMF), an industry consortium of leading AI labs, provides a 2024 progress update on its AI safety initiatives, including workstreams addressing biosecu...ai-safetygovernanceevaluationpolicy+4Source ↗
Footnotes
-
Stanford HAI, 2025 AI Index Report: Technical Performance, April 2025. MMLU gap narrowed from 17.5pp (end 2023) to 0.3pp (end 2024); composite benchmark gap from 9.26% (Jan 2024) to 1.70% (Feb 2025). ↩ ↩2 ↩3
-
LMSYS Chatbot Arena leaderboard, April 2026. Top six models separated by only 20 Elo points; Moonshot's Kimi K2.5 debuted at #15; Chinese models occupy multiple top-tier positions. ↩ ↩2 ↩3
-
futurehumanism.co, "Open Source vs Closed AI: Where the Battle Stands in 2026," 2026. By early 2026, gap between open-source and proprietary AI effectively zero on knowledge benchmarks; GLM-5 holds #1 open-weight Arena Elo (~1,451). ↩ ↩2 ↩3
-
PitchBook-NVCA Q4 2025 Venture Monitor, January 2026. AI/ML deals captured 65.6% of all US VC deal value in 2025 ($222B of $339B total US VC); global AI VC investment $243.9B. ↩ ↩2
-
TechWire Asia, "China AI investment hits $98B in 2025 as tech war with US intensifies," June 2025. Up 48% from 2024. ↩ ↩2
-
Epoch AI, "The US hosts the majority of GPU cluster performance," May 2025. Dataset covers ~417 GPU clusters with known country locations; represents 10-20% of total AI computing capacity. ↩ ↩2
-
DeepSeek, V3.2-Speciale release notes, September 29, 2025. Gold-level results at IMO 2025, CMO, ICPC World Finals, and IOI 2025. V3.2 Chatbot Arena rating ~1,421 (third-highest overall). ↩ ↩2
-
Alibaba Qwen Team, "Qwen3.5: Benchmarks and Capabilities," DataCamp, February-March 2026. Qwen3.5-397B-A17B GPQA Diamond 88.4%; AIME 2026 91.3%. ↩ ↩2
-
OpenAI, "Announcing the Stargate Project," January 21, 2025. $500B planned over 4 years; $100B deployed immediately; initial partners SoftBank, OpenAI, Oracle, MGX. ↩
-
Next Big Future, "China AI Chip and AI Data Centers Versus US AI Data Centers," October 2025. US hyperscalers (Alphabet, Amazon, Microsoft, Meta) plan >$350B in data centers in 2025 and ≈$400B in 2026. ↩
-
Stephen McAleese, "An Overview of the AI Safety Funding Situation," LessWrong, January 2025. Philanthropic AI safety ≈$83.6M in 2024 (Open Philanthropy $63.6M); alignment research funding increased 36% to $150-170M by mid-2025. ↩ ↩2
-
AI Safety Market Funding Trends (2022-2026), newmarketpitch.com, 2026. Core alignment research $110-130M in 2024; Safe Superintelligence raised $2B in 2025; broader AI safety market >$1B including commercial safety startups. ↩
-
Granted AI, "NIST Secures $55 Million for AI Standards and Safety Research," 2026. NIST FY2025: $47.7M for AI research, standards, testing; FY2026: $55M including $10M AI Standards and Innovation Center. ↩
-
myresearchconnect.com, "AI Security Institute Opens £15 Million Funding Programme," 2025. Initial £4M pot expanding to £8.5M; original Frontier AI Taskforce backed by £100M. ↩
-
Wikipedia, "European Artificial Intelligence Office." Became operational August 2, 2025; 125+ staff; established by European Commission decision January 24, 2024. ↩
-
PitchBook Q4 2025 Venture Monitor: global AI VC investment $243.9B in 2025. Safety and governance spend as percentage calculated from combined identified funding (≈$150-170M alignment research + ≈$72M public governance) divided by $243.9B global AI VC. ↩
References
This resource summarizes findings from the Stanford Human-Centered AI (HAI) Institute's AI Index Report comparing the United States and China across key AI metrics in 2024. It covers competitiveness in research output, talent, investment, and technological capabilities between the two leading AI powers. The analysis highlights areas where each country leads and the implications for global AI development.
A business strategy guide explaining first-mover advantage, distinguishing between being 'first-to-market' (technical achievement) and being a true 'first mover' (establishing a market category and shaping expectations). Uses examples like Amazon and Uber to illustrate how transformative market positioning differs from simply releasing a product first.
RAND analyst Lennart Heim argues that while China is rapidly closing the gap on AI model capabilities, the U.S. retains a decisive advantage in total compute capacity and advanced chip infrastructure. The commentary highlights critical failures in U.S. export control enforcement—including TSMC producing chips for Huawei via a proxy—and warns that U.S. policymakers risk squandering the compute advantage by fixating on benchmark comparisons rather than strategic infrastructure leverage.
A 2025 year-end analysis of open-model AI trends showing China surpassing the US in Hugging Face downloads for the first time, with Chinese models like Qwen and DeepSeek gaining significant ground. The piece examines shifts in open-weight vs. open-source dynamics, the rise of small language models, and geopolitical implications for AI governance and export controls.
CNBC reports that Meta is pursuing a new proprietary frontier AI model codenamed 'Avocado,' marking a significant strategic shift away from its open-source Llama models. The company spent $14.3 billion acquiring Scale AI's founder and top researchers to compete with OpenAI and Google, but the rapid pivot has created internal culture clashes and confusion. The delayed release of Llama Behemoth and the possible move away from open-source signals a fundamental rethinking of Meta's AI positioning.
The Frontier Model Forum (FMF), an industry consortium of leading AI labs, provides a 2024 progress update on its AI safety initiatives, including workstreams addressing biosecurity, cybersecurity, model security, safety evaluations, and an AI Safety Fund. The update details early best practices development, expert workshops, and participation in international AI safety governance events like the AI Seoul Summit.
This analysis summarizes a LessWrong post arguing that multipolar AI governance frameworks—featuring decentralized control among diverse AI agents and human actors—offer a compelling alternative to unipolar (centralized) AI control models. It outlines both the risks of decentralized approaches (instability, coordination failures) and the risks of centralized ones (value lock-in, corruption), while proposing pathways like modular AI services and cooperative AI research.
An analysis of the competitive dynamics in the AI industry, examining the landscape of companies, capabilities, and market forces shaping AI development. The piece likely covers key players, investment trends, and implications for the pace and direction of AI progress.
This CSIS analysis examines how DeepSeek's emergence and Huawei's chip development challenge the effectiveness of U.S. export controls on advanced semiconductors. It assesses whether restricting China's access to cutting-edge chips can sustainably constrain Chinese AI capabilities, and considers implications for the broader U.S.-China AI competition.
Recorded Future's intelligence analysis concludes that China is unlikely to sustainably surpass the US in AI capabilities by 2030, examining competitive dynamics across government funding, talent pipelines, technology development, and semiconductor supply chains. The report provides a structured comparison of the two nations' AI ecosystems and identifies key chokepoints in China's development trajectory.
This Boston University article examines DeepSeek, a Chinese AI model that reportedly matched OpenAI's o1 performance at a fraction of the cost, analyzing its implications for AI competition between the US and China. It explores how DeepSeek achieved high performance with limited compute resources, potentially undermining US export controls on advanced chips. The piece discusses broader geopolitical and safety implications of this development.
A Council on Foreign Relations analysis examining the competitive dynamics between China and the United States in artificial intelligence development, with the $109 billion figure referring to U.S. AI investment in 2024. The piece explores how the two superpowers are competing across AI capabilities, infrastructure, and policy.
This AI Impacts page outlines proposed research projects focused on multipolar AI scenarios, where multiple powerful AI systems or actors compete rather than a single dominant system emerging. It identifies open questions and research directions relevant to understanding coordination failures, competitive dynamics, and safety implications of multipolar futures.
Wikipedia article covering the European AI Office, an EU body established in 2024 to oversee implementation and enforcement of the EU AI Act, particularly for general-purpose AI models. It serves as the central regulatory authority coordinating AI governance across EU member states.