Skip to content
Longterm Wiki
Navigation
Updated 2026-04-12HistoryData
Page StatusContent
Edited 2 weeks ago4.1k words7 backlinksUpdated quarterlyDue in 11 weeks
79QualityGood •62.5ImportanceUseful86.5ResearchHigh
Content9/13
SummaryScheduleEntityEdit history1Overview
Tables14/ ~16Diagrams2/ ~2Int. links54/ ~32Ext. links5/ ~20Footnotes16/ ~12References14/ ~12Quotes0Accuracy0RatingsN:5 R:6 A:5.5 C:7.5Backlinks7
Change History1
Auto-improve (standard): Multi-Actor Strategic Landscape2 weeks ago

Improved "Multi-Actor Strategic Landscape" via standard pipeline (1293.3s). Quality score: 77. Issues resolved: Malformed EntityLink in 'First-Mover Advantage' table: `<Ent; Missing `name` attribute on second E2940 reference: `<Entity; Content error in 'First-Mover Advantage' table: the Cloud ma.

1293.3s · $5-8

Issues2
QualityRated 79 but structure suggests 100 (underrated by 21 points)
Links2 links could use <R> components
TODOs2
Complete 'Conceptual Framework' section
Complete 'Quantitative Analysis' section (8 placeholders)

Multi-Actor Strategic Landscape

Analysis

AI Safety Multi-Actor Strategic Landscape

Analyzes AI risk through the lens of which actors develop TAI, finding actor identity may account for 40-60% of total risk variance, with detailed quantitative data on US-China capability convergence (benchmark gap narrowed from 9.26% to 1.70% in ~13 months), investment asymmetries ($222B US vs $98B China VC), and four risk pathways summing to ~25% combined x-risk. The framework is well-populated with 2025-2026 data but the core model parameters (risk estimates, variance attribution) are explicitly acknowledged as illustrative rather than empirically derived.

Related
Analyses
Capability-Alignment Race Model
Organizations
OpenAIAnthropic
4.1k words · 7 backlinks

Core thesis: Risk is primarily determined by which actors develop TAI and their incentive structures. The strategic landscape of competition and cooperation shapes outcomes.

List View
Computing layout...
Legend
Node Types
Leaf Nodes
Causes
Intermediate
Effects
Arrow Strength
Strong
Medium
Weak

Overview

This model analyzes how AI existential risk depends on which actors—US frontier labs, Chinese developers, open-source communities, or malicious actors—develop Transformative AI first, and under what competitive conditions. The core insight is that actor identity and incentive structures may matter as much as technical alignment progress in determining outcomes. For a relational network analysis of how these actors connect and influence each other, see the AI Power and Influence Map.

The strategic landscape shifted substantially in 2024-2025. According to Recorded Future analysis, the gap in overall model performance between the best US and Chinese models narrowed from 9.26% in January 2024 to just 1.70% by February 2025. AI Impacts documented the MMLU benchmark gap narrowing from 17.5 percentage points (end 2023) to 0.3 percentage points (end 2024).1 This convergence was catalyzed by DeepSeek's R1 release in January 2025, which matched OpenAI's o1 performance while training for just $1.6 million. By April 2026, the top six Chatbot Arena Elo models were separated by only 20 Elo points, with Chinese models (Moonshot's Kimi K2.5, DeepSeek V3.2) occupying top positions.2 Similarly, open-source models closed to within 1.70% of frontier closed models on Chatbot Arena by early 2025, and the gap had effectively reached near-zero on knowledge benchmarks by early 2026,3 fundamentally changing proliferation dynamics.

Despite narrowing capability gaps, structural asymmetries persist in investment and compute infrastructure. US AI/ML venture capital investment reached approximately $222 billion in 2025—65.6% of total US VC deal value4—while China's AI capital spending reached approximately $98 billion (including government investment).5 On compute infrastructure, the US holds approximately 74.5% of global GPU cluster performance versus China's 14.1%, according to Epoch AI analysis covering ~417 GPU clusters (May 2025).6 China leads in deployment: it installed approximately 295,000 industrial robots in 2024—more than the rest of the world combined—and accounts for 69.7% of all AI patents.

Conceptual Framework

The Multi-Actor Strategic Landscape model structures AI risk analysis around four actor categories, three competitive pressure mechanisms, and four risk pathways. Unlike purely technical alignment models, it treats actor identity and incentive structures as primary determinants of AI risk outcomes.

Four actor categories:

  • US Frontier Labs (OpenAI, Anthropic, Google DeepMind, Meta): Currently at the capability frontier; highest identified investment in both capabilities and safety research; subject to democratic accountability, though institutional oversight mechanisms remain underdeveloped (model parameter: ~0.35/1.0).
  • Chinese Labs (DeepSeek, Baidu, Alibaba, ByteDance, Zhipu AI): Reached near-parity with US frontier labs on standard benchmarks by early 2026; development priorities shaped by government strategic objectives and different alignment research contexts.
  • Open-Source Ecosystem (Llama, Qwen, DeepSeek open weights, Mistral): Near-zero capability gap on knowledge benchmarks as of early 2026; enables broad proliferation but also distributed safety research and external verification.
  • Malicious Actors (cybercriminals, non-state armed groups, rogue states): Access capabilities primarily through open-source releases, jailbreaks, and API access; currently estimated at 40–60% of frontier capability.

Three competitive pressure mechanisms:

  1. Racing dynamics: Strategic competition between US and Chinese labs (model intensity parameter: 0.75/1.0) creates pressure to accelerate development timelines, potentially reducing safety investment.
  2. Capability diffusion: Open-source releases and technology transfer spread frontier capabilities to a broader set of actors, including potential malicious users.
  3. Safety investment tradeoffs: Profit pressure (model parameter: ~0.80/1.0) competes with safety investment, creating tension between deployment speed and risk reduction.

Four risk pathways (model estimates; see Limitations for caveats):

The four pathways—Unaligned Singleton (8%), Multi-Agent Conflict (6%), Authoritarian Lock-in (5%), and Catastrophic Misuse (7%)—sum to approximately 25% combined x-risk. These are model parameters rather than empirically derived estimates; they are illustrative of how actor identity affects pathway probabilities rather than precise forecasts.

Actor identity is estimated to account for 40–60% of total risk variance across these pathways, meaning differences in which actors lead AI development may matter comparably to technical alignment quality in determining outcomes. For governance frameworks relevant to these pathways, see the AI Governance Effectiveness Analysis.

Capability Gap Estimates (2025-2026)

The following table synthesizes publicly available data on relative AI capabilities across actor categories, updated with 2026 data. Estimates draw from benchmark performance, investment levels, and expert assessments.

Actor CategoryCapability vs FrontierTrendKey EvidenceSource
US Frontier Labs100% (reference)StableGPT-4.5, Claude 3.5, Gemini 2.0 define frontier; top Chatbot Arena positionsIndustry consensus
Chinese Labs (aggregate)≥99% on standard benchmarksConvergedGap narrowed from 9.26% (Jan 2024) to ≈1.70% (Feb 2025); Chinese models in top-6 Chatbot Arena (Apr 2026)Recorded Future;2
DeepSeek specifically≈100% on standard benchmarksAt frontierR1 matched o1 at $1.6M training cost; V3.2-Speciale gold at IMO 2025 and IOI 2025; V3.2 third-highest Arena Elo (≈1,421)CSIS;7
Open-Source (Llama, Qwen, GLM-5)≥99% on knowledge benchmarksNear-zero gapGLM-5 top open-weight Arena Elo (≈1,451); Qwen3.5 leads open-weight GPQA at 88.4% (Feb 2026)State of Open-Source AI;38
Malicious Actor Access≈40-60%IncreasingAccess via open-source, jailbreaks, or theftExpert estimate

Investment and Infrastructure Asymmetries (2024-2025)

DimensionUnited StatesChinaNotes
Private AI investment (2024)$109 billion≈$9.3 billionStanford HAI 2025 AI Index; US figure is ≈12× China1
AI/ML VC investment (2025)≈$222 billion65.6% of $339B total US VC; PitchBook Q4 20254
AI capital spending (2025)≈$98 billionIncludes government investment; 48% YoY increase5
Stargate / infrastructure commitment$500B (4-year)¥380B (≈$52B, tech giants)OpenAI Jan 20259; includes immediate $100B deployment
US hyperscaler data center capex (2025-2026)>$350B / ≈$400BAlphabet, Amazon, Microsoft, Meta, Oracle combined10
GPU cluster performance share74.5% of global14.1% of globalEpoch AI, May 2025; covers ≈417 identified clusters6
Data Centers4,049379≈11:1; RAND
Industrial robot installations (2024)34,000295,0001:9; China leads deployment
AI patents (2023)13% of global69.7% of globalChina dominates IP filings
AI research citations (2023)13% of global22.6% of globalChina leads academic output

Sources: CFR, RAND, Stanford HAI, Epoch AI, PitchBook

Note: 2024 and 2025 investment figures use different methodologies (private investment vs. VC deal value vs. capital spending including government) and are not directly comparable across years or countries. China's 2025 figure includes substantial government-directed investment absent from the US VC measure.

Key Dynamics

The following diagram illustrates how actor competition dynamics flow through to risk outcomes:

Diagram (loading…)
flowchart TD
  subgraph Actors["Actor Competition"]
      USL[US Frontier Labs]
      CNL[Chinese Labs]
      OSS[Open-Source]
      MAL[Malicious Actors]
  end

  subgraph Dynamics["Competitive Pressures"]
      RACE[Racing Dynamics]
      DIFF[Capability Diffusion]
      SAFE[Safety Investment]
  end

  subgraph Outcomes["Risk Pathways"]
      SING[Unaligned Singleton]
      MULT[Multi-Agent Conflict]
      AUTH[Authoritarian Lock-in]
      MISUSE[Catastrophic Misuse]
  end

  USL -->|competes with| CNL
  CNL -->|closes gap| RACE
  USL -->|pressure to lead| RACE

  OSS -->|enables access| DIFF
  DIFF -->|reaches| MAL

  RACE -->|reduces| SAFE
  SAFE -->|insufficient| SING

  CNL -->|if wins| AUTH
  MAL -->|enables| MISUSE
  RACE -->|increases| MULT

  style USL fill:#e6f3ff
  style CNL fill:#ffe6e6
  style OSS fill:#e6ffe6
  style MAL fill:#ffcccc
  style SING fill:#ffdddd
  style MULT fill:#ffdddd
  style AUTH fill:#ffdddd
  style MISUSE fill:#ffdddd

The key mechanisms are:

  1. Competition intensity → Safety shortcuts → Misalignment risk: As US-China competition intensifies (currently 0.75 on normalized scale), labs face pressure to accelerate timelines, potentially cutting safety corners.

  2. Capability diffusion → Malicious access → Misuse risk: Open-source releases (now near-zero gap from frontier on knowledge benchmarks) enable rapid proliferation to actors who may lack safety constraints or beneficial intent.

  3. First-mover advantage → Winner-take-all → Reduced caution: If decisive strategic advantage exists for first-mover, actors rationally accept higher alignment risk to capture it.

  4. Democratic oversight → Deployment delays → Capability gaps: Strong oversight in democratic nations may create windows where authoritarian actors gain advantages, creating perverse incentives against regulation.

  5. Transparency → Better coordination → Reduced racing: Conversely, capability transparency and safety research sharing (currently ~0.6 openness) can reduce competitive pressure.

Quantitative Analysis

US-China Benchmark Gap: Trajectory (2023-2026)

Benchmark gaps between US and Chinese frontier models narrowed substantially across all major evaluations.1

BenchmarkUS–China Gap (End 2023)US–China Gap (End 2024)Direction
MMLU17.5 pp0.3 ppNear-parity
MMMU13.5 pp8.1 ppClosing
MATH24.3 pp1.6 ppNear-parity
HumanEval31.6 pp3.7 ppNear-parity
Composite benchmark9.26% (Jan 2024)1.70% (Feb 2025)Near-parity

By April 2026, the top six Chatbot Arena models were separated by only 20 Elo points, with Chinese models (Kimi K2.5, DeepSeek V3.2 at Arena Elo ~1,421) occupying top-tier positions alongside Western models.2 DeepSeek V3.2-Speciale achieved gold-level results at IMO 2025, CMO, ICPC World Finals, and IOI 2025.7

Open-Source vs. Frontier Gap (2025-2026)

By early 2026, the gap between open-source and proprietary AI had effectively closed on knowledge benchmarks:3

ModelOrganizationGPQA DiamondArena EloStatus
GLM-5 (Reasoning)Zhipu AI94%≈1,451 (top open-weight)Open-weight
Qwen3.5-397B-A17BAlibaba88.4%Top open-weight tierOpen-weight
DeepSeek V3.2DeepSeek≈1,421 (3rd overall)Open-weight
Top proprietary (reference)Various≈90-92%≈1,440-1,460Proprietary

Single-digit gaps remain on frontier reasoning tasks (e.g., OpenAI o3-class models on competition mathematics).8

AI Safety and Alignment Research Funding (2024-2025)

Funding SourceAmountYearNotes
Open Philanthropy (technical AI safety)≈$50M committed2024Largest single philanthropic funder; grants to CAIS, Redwood Research, Machine Intelligence Research Institute (MIRI)
Open Philanthropy (RFP)$40M available2025Dedicated request for proposals
Survival and Flourishing Fund (SFF) (SFF)$34.33M distributed202586% to AI safety
Long-Term Future Fund (LTFF) (LTFF)$8.4M total2024Over $5M to AI safety
Total philanthropic AI safety≈$83.6M2024Open Philanthropy ≈76%; individual donors (Jaan Tallinn ≈$20M) ~24%
Alignment research funding (all sources)$150-170MMid-202536% increase from $110-130M in 2024
Broader AI safety market>$1 billion2025Includes commercial/VC safety-focused startups; Safe Superintelligence raised $2B in 2025

Source: Stephen McAleese, LessWrong, January 202511; AI Safety Market Funding Trends 2022-202612

AI Governance Institutional Capacity (2025-2026)

InstitutionFundingYearNotes
NIST AI Safety Institute (US)≈$47.7MFY2025AI research, standards, and testing
NIST AI Safety Institute (US)≈$55MFY2026Includes AI Standards and Innovation Center ($10M)
UK AI Security Institute (AISI) research programme£15M+2025Initial £4M pot expanding to £8.5M; ranges from £50K to £1M per grant
UK DSIT/AISI foundation investment£100M2023-2024Original Frontier AI Taskforce commitment
EU AI OfficeOperationalAug 2025125+ staff; no public budget figure available
Global public AI governance (US + UK + EU combined est.)≈$72.3M (projected)2025Federal agencies; does not include EU AI Office

Sources: NIST FY2026 budget announcement13; UK AISI research programme14; EU AI Office15

Safety Spend as Share of Total AI Investment

Identified alignment research funding ($150-170M in mid-2025) and public governance budgets (≈$72M) together represent approximately 0.06-0.09% of global AI VC investment ($243.9B in 2025).16 Actor-focused governance specifically—covering capability intelligence, game-theoretic analysis of lab incentives, and policy engagement—is estimated as a smaller subset of this total. Lab-level safety investment percentages are not publicly disclosed; estimates in the literature (e.g., Anthropic ≈30%, OpenAI ≈15%) are inferred from organizational statements and cannot be independently verified.

Risk Pathways

PathwayDescriptionModel Estimate
Unaligned SingletonOne misaligned AI gains decisive advantage8%
Multi-Agent ConflictMultiple powerful AI systems in conflict6%
Authoritarian Lock-inAI enables permanent authoritarian control5%
Catastrophic MisuseIntentional misuse causes catastrophe7%
Combined X-RiskTotal from all pathways≈25%

These are model parameters illustrating how actor identity affects pathway probabilities. They are not empirically derived forecasts.

Actor Categories

CategoryKey Actors
Leading USOpenAI, Anthropic, Google DeepMind, Meta
Leading ChinaDeepSeek, Baidu, Alibaba, ByteDance, Zhipu AI
Open-SourceMeta (Llama), Mistral, Hugging Face ecosystem, Alibaba (Qwen)
MaliciousCybercriminals, terrorists, rogue states
GovernmentsUS (NSA, DARPA), China (PLA, MSS), EU

Full Variable List

This diagram simplifies the full model. The complete Multi-Actor Strategic Landscape includes:

Actor Capabilities (15 variables): Leading US lab, leading Chinese lab, US government AI, Chinese government AI, open-source ecosystem, second-tier corporate labs, academic research, cybercriminal AI, terrorist access, authoritarian regime AI, democratic allies AI, corporate espionage, state IP theft, insider threat, supply chain security.

Actor Incentives (12 variables): US-China competition, profit pressure, academic openness, classification levels, democratic accountability, authoritarian control, geopolitical crises, economic desperation, military doctrine, regulatory arbitrage, talent mobility, public-private partnerships.

Information & Transparency (7 variables): Capability disclosure, safety sharing, incident reporting, capability intelligence, dual-use publication norms, evaluation standards, third-party verification.

Alignment & Control (8 variables): US actor alignment, China actor alignment, Constitutional AI effectiveness, human oversight scalability, kill switch reliability, containment protocols, red-teaming, post-deployment monitoring.

Strategic Outcomes (8 variables): First-mover advantage, winner-take-all dynamics, diffusion speed, multipolar vs bipolar, offense-defense balance, escalation control, governance lock-in, misuse probability.

Existential Risk Paths (5 variables): Unaligned singleton, multi-agent conflict, authoritarian lock-in, economic/social collapse, combined risk.

Strategic Importance

Magnitude Assessment

The multi-actor landscape determines whether AI development is coordinated or conflictual. Actor heterogeneity creates both risks (racing, proliferation) and opportunities (diverse approaches).

DimensionAssessmentQuantitative Estimate
Potential severityMultipolar dynamics drive racing and proliferationActor landscape estimated to contribute 40-60% of total risk variance (model parameter)
Probability-weighted importanceCurrently in competitive multipolar phase75% model-estimated probability of continued multipolar competition through 2030
Comparative rankingEssential context for governance and coordination strategiesPriority comparable to technical alignment in governance-focused analyses; ordering varies by framework
MalleabilityActor incentives partially shiftable via policy20-30% of racing dynamics estimated as addressable via policy (model parameter; unverified benchmark)

Actor Safety Assessment

Lab-level safety investment percentages are not publicly disclosed. The following table describes observable indicators rather than inferred grades. For governance effectiveness context, see the AI Governance Effectiveness Analysis.

Actor CategoryObservable Safety IndicatorsExternal TransparencyNotes
AnthropicResponsible Scaling Policy published; Constitutional AI methodology; regular system cardsHighOrganization states safety as core mission; ≈1,500 employees (early 2026); valuation grew to $380B (2025 Series G)
OpenAISafety council established 2025; Head of Preparedness role created (late 2025); Preparedness team activeMediumSafety team structure changed during 2024-2025; targeting ≈8,000 employees by end 2026
Google DeepMindPublished AGI safety and security strategy (2025); multiple alignment papers; ≈6,000 employees (Aug 2025)MediumDeepMind and Google Brain merged 2023; safety research is published
Meta AIOpen-source model releases with safety evaluations; FAIR research publishedHigh (open-source)Reduced AI safety research headcount Oct 2025; ≈3,400 lab staff (Jul 2025)
Chinese LabsModel cards and safety system cards published (DeepSeek, Alibaba Qwen); affiliated university alignment researchLower relative to Western labs in Western-accessible venuesHave published safety research; Chinese labs have lower external transparency to Western research community
Open-Source EcosystemDistributed responsibility; community red-teaming; open safety researchVery HighVariable implementation across projects; no central coordinating body

Note: Safety investment percentages (e.g., "30% of budget") circulate in AI governance literature but are not drawn from public financial disclosures. No lab has confirmed specific safety budget percentages.

Diffusion Timeline Estimates

Capability LevelUS LabsChinese LabsOpen-SourceMalicious Actors
GPT-4 class20232024-2025 ✓2024-2025 ✓2025-2026
Frontier reasoning (o1/R1 class)2024-20252025 ✓ (DeepSeek R1)2025-20262027-2028
GPT-5 class (projected)20252026 (approaching)2027-20282028-2030
Autonomous agents (dangerous capability threshold)2025-20262026-20272027-20282028-2029

Key Finding: The open-source and Chinese lab lags have collapsed on standard benchmarks. As of late 2025, the center of gravity for open-weight models has shifted toward China, with DeepSeek and Qwen becoming leading providers. US firms released fewer open-weight models citing commercial and safety constraints, while Chinese labs treated open-weight leadership as a deliberate strategy. Meta—long a champion of frontier open models—has delayed release of Llama Behemoth and indicated it may keep future frontier models behind paywalls.

The "GPT-5 class" row should be interpreted cautiously: Chinese labs achieved near-parity with GPT-4o-comparable models in 2025, making the "GPT-5 class" timeline dependent on where proprietary labs define the next capability threshold.

First-Mover Advantage: Evidence Assessment

The model's risk estimates depend critically on the magnitude of first-mover advantage. Strong first-mover advantages create racing incentives; weak ones reduce them. Current evidence suggests first-mover advantages are significant but not overwhelming:

Evidence TypeFindingImplication for FMA
Historical analysisFirst movers have 47% failure rate; only 11% become market leaders (Golder & Tellis)Suggests weak FMA
AI competitive landscape2,011 companies in 2024 ML/AI landscape, 578 new entrants since 2023Suggests weak FMA
Model replication11 different developers globally achieved GPT-4-level models in 2024Suggests weak FMA
Cloud marketAWS, Azure, and Google Cloud trading leadership position; "more than one winner" possibleSuggests moderate FMA
Network effectsAI systems less network-effect-driven than social platformsSuggests weak FMA
TAI-specific dynamicsDecisive strategic advantage at TAI level remains uncertainUnknown

Evidence from the Abundance Institute suggests "no signs of winner-take-all dynamics" in the current AI ecosystem. However, TAI (transformative AI) may differ qualitatively if it enables rapid capability improvements or strategic advantages not available to followers. The model's 0.7 first-mover advantage estimate may be calibrated too high based on current evidence, but TAI-level dynamics remain highly uncertain.

Resource Implications

Understanding the actor landscape enables targeted governance strategies:

  • Targeted engagement with highest-leverage actors: Focus on top 3-4 US labs could cover 70% of frontier capability
  • Coalition-building for safety standards: Anthropic-OpenAI-DeepMind coalition would set de facto standards
  • Monitoring of capability diffusion: $50-100M/year for comprehensive capability intelligence
  • Anticipation of strategic behavior: Game-theoretic modeling estimated at ≈$10-20M/year

Estimated investment required for comprehensive coverage: Analyses estimate that actor-focused governance—capability intelligence, lab engagement, diffusion monitoring—may require $100-200M/year. Current philanthropic alignment research funding reached approximately $150-170M by mid-2025, though most is oriented toward technical safety rather than actor-focused governance specifically.11 The $20-30M estimate for current actor-focused governance work specifically is an analytical estimate without a publicly verified benchmark.

Key Cruxes

These probability estimates are model author estimates; they are not drawn from formal elicitation surveys.

CruxIf TrueIf FalseModel Estimate
Leading coalition is stableTop 3 can set normsRacing to bottom45%
Safety can be coordination pointVoluntary standards viableRegulation required35%
China is engageable on safetyGlobal coordination possibleBifurcated governance30%
Diffusion to malicious actors is slowWindow for governanceProliferation dominates50%

Multipolar vs Unipolar Governance Considerations

A crucial variable in this model is whether AI development converges toward unipolar (single dominant actor or coalition) or multipolar (distributed power among multiple actors) outcomes. Each presents distinct risk profiles:

Governance StructureKey RisksKey Advantages
Unipolar (single dominant actor)Value lock-in, institutional stagnation, internal corruption, single points of failureCoordination easier, racing reduced, unified safety standards
Multipolar (distributed power)Unchecked proliferation, system instability, coordination failures, racing dynamicsDiversity of approaches, no single point of failure, competitive pressure for safety

Current research from AI Impacts identifies key research questions: What "considerations might tip us between multipolar and unipolar scenarios"? What "risks [are] distinctive to a multipolar scenario"? The CO/AI analysis notes that while current AI safety discussions often default to unipolar frameworks, "exploring decentralized governance structures could address key risks like value lock-in and institutional stagnation."

Current assessment: The model estimates 55% probability of continued multipolar development, with the US-China bifurcation appearing increasingly stable. Geopolitical tensions, divergent regulatory approaches, and the collapse of open-source lags all point toward a world with multiple competing AI powers rather than a single dominant actor.

Limitations

  1. Capability estimates rapidly outdating: The 2025-2026 data showing near-parity may not persist; breakthrough capabilities could restore gaps. The near-zero benchmark gap reflects standard evaluations and may not extend to capabilities at the frontier of reasoning, agentic tasks, or future benchmark domains.

  2. Safety investment data opaque: Lab safety budgets are not publicly disclosed; estimates in circulation are inferential and unverified. The Actor Safety Assessment table reflects observable indicators rather than confirmed investment levels.

  3. TAI dynamics uncertain: Current competitive patterns may not predict TAI-level dynamics where decisive advantages could differ fundamentally.

  4. Geopolitical volatility: US-China relations, export control effectiveness, and regulatory trajectories are highly uncertain. Export controls on advanced chips may slow Chinese compute accumulation in ways not yet reflected in 2026 data.

  5. Malicious actor access hard to estimate: Underground markets and state-sponsored theft create significant uncertainty in capability diffusion.

  6. Model parameters are illustrative: The causal graph confidence values and risk pathway percentages are analytical estimates to illustrate model structure, not empirically derived posterior probabilities. The 40-60% risk variance attribution figure is a model claim without a published methodology or independent verification.

Sources

  • Recorded Future: US-China AI Gap Analysis (2025)
  • RAND: China's AI Models Closing the Gap (2025)
  • Council on Foreign Relations: China, the United States, and the AI Race
  • Boston University: DeepSeek and AI Frontier (2025)
  • State of Open-Source AI 2025
  • CSIS: DeepSeek, Huawei, and US-China AI Race
  • Abundance Institute: AI Competitive Landscape
  • AI Impacts: Multipolar Research Projects
  • Frontier Model Forum: Progress Update 2024

Footnotes

  1. Stanford HAI, 2025 AI Index Report: Technical Performance, April 2025. MMLU gap narrowed from 17.5pp (end 2023) to 0.3pp (end 2024); composite benchmark gap from 9.26% (Jan 2024) to 1.70% (Feb 2025). 2 3

  2. LMSYS Chatbot Arena leaderboard, April 2026. Top six models separated by only 20 Elo points; Moonshot's Kimi K2.5 debuted at #15; Chinese models occupy multiple top-tier positions. 2 3

  3. futurehumanism.co, "Open Source vs Closed AI: Where the Battle Stands in 2026," 2026. By early 2026, gap between open-source and proprietary AI effectively zero on knowledge benchmarks; GLM-5 holds #1 open-weight Arena Elo (~1,451). 2 3

  4. PitchBook-NVCA Q4 2025 Venture Monitor, January 2026. AI/ML deals captured 65.6% of all US VC deal value in 2025 ($222B of $339B total US VC); global AI VC investment $243.9B. 2

  5. TechWire Asia, "China AI investment hits $98B in 2025 as tech war with US intensifies," June 2025. Up 48% from 2024. 2

  6. Epoch AI, "The US hosts the majority of GPU cluster performance," May 2025. Dataset covers ~417 GPU clusters with known country locations; represents 10-20% of total AI computing capacity. 2

  7. DeepSeek, V3.2-Speciale release notes, September 29, 2025. Gold-level results at IMO 2025, CMO, ICPC World Finals, and IOI 2025. V3.2 Chatbot Arena rating ~1,421 (third-highest overall). 2

  8. Alibaba Qwen Team, "Qwen3.5: Benchmarks and Capabilities," DataCamp, February-March 2026. Qwen3.5-397B-A17B GPQA Diamond 88.4%; AIME 2026 91.3%. 2

  9. OpenAI, "Announcing the Stargate Project," January 21, 2025. $500B planned over 4 years; $100B deployed immediately; initial partners SoftBank, OpenAI, Oracle, MGX.

  10. Next Big Future, "China AI Chip and AI Data Centers Versus US AI Data Centers," October 2025. US hyperscalers (Alphabet, Amazon, Microsoft, Meta) plan >$350B in data centers in 2025 and ≈$400B in 2026.

  11. Stephen McAleese, "An Overview of the AI Safety Funding Situation," LessWrong, January 2025. Philanthropic AI safety ≈$83.6M in 2024 (Open Philanthropy $63.6M); alignment research funding increased 36% to $150-170M by mid-2025. 2

  12. AI Safety Market Funding Trends (2022-2026), newmarketpitch.com, 2026. Core alignment research $110-130M in 2024; Safe Superintelligence raised $2B in 2025; broader AI safety market >$1B including commercial safety startups.

  13. Granted AI, "NIST Secures $55 Million for AI Standards and Safety Research," 2026. NIST FY2025: $47.7M for AI research, standards, testing; FY2026: $55M including $10M AI Standards and Innovation Center.

  14. myresearchconnect.com, "AI Security Institute Opens £15 Million Funding Programme," 2025. Initial £4M pot expanding to £8.5M; original Frontier AI Taskforce backed by £100M.

  15. Wikipedia, "European Artificial Intelligence Office." Became operational August 2, 2025; 125+ staff; established by European Commission decision January 24, 2024.

  16. PitchBook Q4 2025 Venture Monitor: global AI VC investment $243.9B in 2025. Safety and governance spend as percentage calculated from combined identified funding (≈$150-170M alignment research + ≈$72M public governance) divided by $243.9B global AI VC.

References

This resource summarizes findings from the Stanford Human-Centered AI (HAI) Institute's AI Index Report comparing the United States and China across key AI metrics in 2024. It covers competitiveness in research output, talent, investment, and technological capabilities between the two leading AI powers. The analysis highlights areas where each country leads and the implications for global AI development.

A business strategy guide explaining first-mover advantage, distinguishing between being 'first-to-market' (technical achievement) and being a true 'first mover' (establishing a market category and shaping expectations). Uses examples like Amazon and Uber to illustrate how transformative market positioning differs from simply releasing a product first.

3largest single advantageRAND Corporation·2025

RAND analyst Lennart Heim argues that while China is rapidly closing the gap on AI model capabilities, the U.S. retains a decisive advantage in total compute capacity and advanced chip infrastructure. The commentary highlights critical failures in U.S. export control enforcement—including TSMC producing chips for Huawei via a proxy—and warns that U.S. policymakers risk squandering the compute advantage by fixating on benchmark comparisons rather than strategic infrastructure leverage.

★★★★☆

A 2025 year-end analysis of open-model AI trends showing China surpassing the US in Hugging Face downloads for the first time, with Chinese models like Qwen and DeepSeek gaining significant ground. The piece examines shifts in open-weight vs. open-source dynamics, the rise of small language models, and geopolitical implications for AI governance and export controls.

CNBC reports that Meta is pursuing a new proprietary frontier AI model codenamed 'Avocado,' marking a significant strategic shift away from its open-source Llama models. The company spent $14.3 billion acquiring Scale AI's founder and top researchers to compete with OpenAI and Google, but the rapid pivot has created internal culture clashes and confusion. The delayed release of Llama Behemoth and the possible move away from open-source signals a fundamental rethinking of Meta's AI positioning.

★★★☆☆
6Frontier Model ForumFrontier Model Forum

The Frontier Model Forum (FMF), an industry consortium of leading AI labs, provides a 2024 progress update on its AI safety initiatives, including workstreams addressing biosecurity, cybersecurity, model security, safety evaluations, and an AI Safety Fund. The update details early best practices development, expert workshops, and participation in international AI safety governance events like the AI Seoul Summit.

★★★☆☆

This analysis summarizes a LessWrong post arguing that multipolar AI governance frameworks—featuring decentralized control among diverse AI agents and human actors—offer a compelling alternative to unipolar (centralized) AI control models. It outlines both the risks of decentralized approaches (instability, coordination failures) and the risks of centralized ones (value lock-in, corruption), while proposing pathways like modular AI services and cooperative AI research.

An analysis of the competitive dynamics in the AI industry, examining the landscape of companies, capabilities, and market forces shaping AI development. The piece likely covers key players, investment trends, and implications for the pace and direction of AI progress.

This CSIS analysis examines how DeepSeek's emergence and Huawei's chip development challenge the effectiveness of U.S. export controls on advanced semiconductors. It assesses whether restricting China's access to cutting-edge chips can sustainably constrain Chinese AI capabilities, and considers implications for the broader U.S.-China AI competition.

★★★★☆

Recorded Future's intelligence analysis concludes that China is unlikely to sustainably surpass the US in AI capabilities by 2030, examining competitive dynamics across government funding, talent pipelines, technology development, and semiconductor supply chains. The report provides a structured comparison of the two nations' AI ecosystems and identifies key chokepoints in China's development trajectory.

This Boston University article examines DeepSeek, a Chinese AI model that reportedly matched OpenAI's o1 performance at a fraction of the cost, analyzing its implications for AI competition between the US and China. It explores how DeepSeek achieved high performance with limited compute resources, potentially undermining US export controls on advanced chips. The piece discusses broader geopolitical and safety implications of this development.

12\$109 billion in 2024Council on Foreign Relations

A Council on Foreign Relations analysis examining the competitive dynamics between China and the United States in artificial intelligence development, with the $109 billion figure referring to U.S. AI investment in 2024. The piece explores how the two superpowers are competing across AI capabilities, infrastructure, and policy.

★★★★☆

This AI Impacts page outlines proposed research projects focused on multipolar AI scenarios, where multiple powerful AI systems or actors compete rather than a single dominant system emerging. It identifies open questions and research directions relevant to understanding coordination failures, competitive dynamics, and safety implications of multipolar futures.

★★★☆☆

Wikipedia article covering the European AI Office, an EU body established in 2024 to oversee implementation and enforcement of the EU AI Act, particularly for general-purpose AI models. It serves as the central regulatory authority coordinating AI governance across EU member states.

★★★☆☆

Related Wiki Pages

Top Related Pages

Risks

AI ProliferationMultipolar Trap (AI Development)

Approaches

MAIM (Mutually Assured AI Malfunction)

Analysis

AI Power and Influence Map

Organizations

AI ImpactsSurvival and Flourishing Fund (SFF)Epoch AILong-Term Future Fund (LTFF)Open PhilanthropyMachine Intelligence Research Institute (MIRI)

Other

Jaan TallinnGPQA DiamondLlamaHumanEvalMistralStella Biderman

Concepts

Actor Power ScorecardTransformative AI

Policy

China AI Regulatory Framework

Historical

International AI Safety Summit Series