Contributes to: Governance Capacity
Primary outcomes affected:
- Existential Catastrophe ↓↓ — Effective regulation can slow dangerous development and enforce safety
- Transition Smoothness ↓↓ — Adaptive governance manages economic and social disruption
Regulatory Capacity measures the ability of governments to effectively understand, evaluate, and regulate AI systems. Higher regulatory capacity is better—it enables evidence-based oversight that can actually keep pace with AI development. This parameter encompasses technical expertise within regulatory agencies, institutional resources for enforcement, and the capability to keep pace with rapidly advancing AI technology. Unlike international coordination, which focuses on cooperation between nations, regulatory capacity addresses the fundamental question of whether any government—acting alone—can meaningfully oversee AI development.
Institutional investments, talent flows, and political priorities all shape whether regulatory capacity grows or declines. High capacity enables evidence-based regulation and credible enforcement; low capacity results in either ineffective oversight or innovation-stifling rules that fail to address actual risks.
This parameter underpins:
Contributes to: Governance Capacity
Primary outcomes affected:
| Metric | Current Value | Comparison | Trend | Source |
|---|---|---|---|---|
| Combined AISI budgets | ~$150M annually | 0.15% of industry R&D | Constrained | UK/US/EU AISI budgets |
| Industry AI investment | $100B+ annually (US alone) | 600:1 vs. regulators | Growing rapidly | Industry reports |
| NIST AI RMF adoption | 40-60% Fortune 500 | Voluntary framework | Growing | NIST |
| Federal AI regulations | 59 (2024) | 25 (2023) | +136% YoY | Stanford HAI |
| State AI bills passed | 131 (2024) | ~50 (2023) | +162% YoY | State legislatures |
| Federal AI talent hired | 200+ (2024) | Target: 500 by FY2025 | +100% YoY | White House AI Task Force |
| Government AI readiness | US: #1, China: #2 (2025) | 195 countries assessed | Bipolar leadership | Oxford Insights Index |
| AISI network size | 11 countries + EU | Nov 2023: 1 (UK) | +1100% growth | International AI Safety Report |
| Institution | Annual Budget | Staff | Primary Focus |
|---|---|---|---|
| UK AI Security Institute | ~$65M (50M GBP) | ~100+ | Model evaluations, red-teaming |
| US CAISI (formerly AISI) | ~$10M | ~50 | Standards, innovation (refocused 2025) |
| EU AI Office | ~$8M | Growing | AI Act enforcement |
| OpenAI (for comparison) | ~$5B+ | 2,000+ | AI development |
| Anthropic (for comparison) | ~$2B+ | 1,000+ | AI development |
The resource asymmetry is stark: a single frontier AI lab spends 30-50x more than the entire global network of AI Safety Institutes combined.
Healthy regulatory capacity would enable governments to understand AI systems at a technical level sufficient to evaluate safety claims, enforce requirements, and adapt frameworks as technology evolves.
| Characteristic | Current Status | Gap |
|---|---|---|
| Technical expertise | Building via AISIs; still limited | Large—industry expertise 10-100x greater |
| Competitive compensation | Government salaries 50-80% below industry | Very large |
| Independent evaluation | First joint evaluations in 2024 | Large—capacity limited to ~2-3 models/year |
| Enforcement resources | Minimal for AI-specific violations | Very large |
| Adaptive processes | EU AI Act: 2-3 year implementation | Medium—improving but still slow |
| Threat | Mechanism | Evidence | Probability Range |
|---|---|---|---|
| Budget disparity | Industry outspends regulators 600:1 | $100B+ vs. $150M | 95-99% likelihood gap persists through 2027 |
| Talent competition | Top AI researchers choose industry salaries | Google pays $1M+; government pays $150-250K; federal hiring surge reached 200/500 target by mid-2024 | 70-85% of top talent chooses industry |
| Information asymmetry | Companies know more about their systems than regulators | Model evaluations require company cooperation; voluntary access agreements with OpenAI, Anthropic, DeepMind | 80-90% of evaluation data comes from labs |
| Expertise gap widening | AI capabilities advance faster than regulatory learning | UK AISI evaluations show models now complete expert-level cyber tasks (10+ years experience equivalent) | 60-75% chance gap widens 2025-2027 |
| Threat | Mechanism | Evidence |
|---|---|---|
| Mission reversal | New administrations can redirect agencies | AISI renamed CAISI; refocused from safety to innovation (June 2025) |
| Leadership turnover | Key officials depart with administration changes | Elizabeth Kelly (AISI director) resigned February 2025 |
| Budget cuts | Regulatory funding depends on political priorities | Congressional appropriators cut AISI funding requests |
| Threat | Mechanism | Evidence |
|---|---|---|
| Capability outpacing | AI advances faster than regulatory adaptation | AI capabilities advance weekly; rules take years |
| Model opacity | Even developers cannot fully explain model behavior | Interpretability covers ~10% of frontier model capacity |
| Evaluation complexity | Assessing safety requires sophisticated technical infrastructure | UK AISI evaluation of o1 took months with dedicated resources |
| Factor | Mechanism | Status | Growth Trajectory |
|---|---|---|---|
| AISI network development | Building dedicated evaluation expertise | 11 countries + EU (2024-2025); inaugural network meeting November 2024 | From 1 institute (Nov 2023) to 11+ (Dec 2024); 15-20 institutes projected by 2026 |
| Academic partnerships | Universities provide research capacity | NIST AI RMF community of 6,500+ participants | Growing 30-40% annually |
| Industry cooperation | Voluntary testing agreements expand access | Anthropic, OpenAI, DeepMind signed pre-deployment access agreements (2024) | Fragile—depends on continued voluntary participation |
| Federal talent recruitment | Specialized hiring programs for AI experts | 200+ hired in 2024; target 500 by FY2025 via AI Corps, US Digital Corps | 40-60% of target achieved mid-2024; uncertain post-administration change |
| Factor | Mechanism | Status | Implementation Details |
|---|---|---|---|
| EU AI Act | Creates mandatory compliance obligations with penalties up to €35M/7% revenue | Implementation timeline: entered force August 2024; GPAI obligations active August 2025; full enforcement August 2026 | Only 3 of 27 member states designated authorities by August 2025 deadline—severe implementation capacity gap |
| NIST AI RMF | Provides structured assessment methodology | 40-60% Fortune 500 adoption; voluntary framework limits enforcement | 70-75% adoption in financial services (existing regulatory culture); 25-35% in retail |
| State legislation | Creates enforcement opportunities | 131 state AI bills passed (2024); over 1,000 bills introduced in 2025 legislative session | Fragmentation risk—federal preemption efforts may override state capacity building |
| Factor | Mechanism | Status |
|---|---|---|
| Interpretability research | Better understanding of model behavior | 70% of Claude 3 Sonnet features interpretable |
| Evaluation tools | Open-source frameworks for safety assessment | UK AISI Inspect framework released May 2024 |
| Automated auditing | AI-assisted oversight could reduce resource needs | Research stage |
| Domain | Impact | Severity |
|---|---|---|
| Compliance theater | Companies perform safety rituals without substantive risk reduction | High |
| Reactive governance | Regulation only after harms materialize | High |
| Credibility gap | Industry ignores regulations it knows cannot be enforced | Critical |
| Innovation harm | Poorly designed rules burden companies without improving safety | Medium |
| Democratic accountability | Citizens cannot hold companies accountable through government | High |
Regulatory capacity affects existential risk through several mechanisms:
Pre-deployment evaluation: If regulators cannot assess frontier AI systems before deployment, safety depends entirely on company self-governance. The ~$150M combined AISI budget versus $100B+ industry spending suggests current capacity is insufficient for meaningful pre-deployment oversight. The UK AISI's Frontier AI Trends Report documents evaluation capacity of 2-3 major models per year—insufficient when labs release models quarterly or monthly.
Enforcement credibility: Without enforcement capability, even well-designed rules become voluntary. The EU AI Act establishes penalties up to €35M or 7% of global revenue, but only 3 of 27 member states designated enforcement authorities by the August 2025 deadline. This 11% compliance rate with basic administrative requirements suggests severe capacity constraints for actual enforcement. The US has zero federal AI-specific enforcement actions as of December 2025.
Adaptive governance: Transformative AI may require rapid regulatory response—potentially within weeks of capability emergence. Current regulatory processes operate on multi-year timelines: the EU AI Act took 3 years to pass (2021-2024) and requires 2 more years for full implementation (2024-2026). The OECD's research on AI in regulatory design finds governments must shift from "regulate-and-forget" to "adapt-and-learn" approaches, but 70% of countries still lack capacity for AI-enhanced policy implementation as of 2023.
Capability-regulation race dynamics: Academic research documents "regulatory inertia" where lack of technical capabilities prevents timely response despite urgent need. Nature's 2024 analysis identifies information asymmetry, pacing problems, and risk of regulatory capture as fundamental challenges requiring new approaches—yet most jurisdictions continue traditional frameworks. The probability of meaningful catastrophic risk regulation before transformative AI arrival is estimated at 15-30% given current trajectories.
| Timeframe | Key Developments | Capacity Impact |
|---|---|---|
| 2025-2026 | EU AI Act enforcement begins; CAISI mission unclear; state legislation proliferates | Mixed—EU capacity growing; US uncertain |
| 2027-2028 | Next-gen frontier models deployed; AISI network matures | Capacity gap may widen if models advance faster than institutions |
| 2029-2030 | Potential new frameworks; enforcement track record emerges | Depends on political commitments and incident history |
| Scenario | Probability | Outcome | Key Drivers | Timeline |
|---|---|---|---|---|
| Capacity catch-up | 15-20% | Major incident or political shift drives significant regulatory investment (5-10x budget increases); capacity begins closing gap with industry | Catastrophic AI incident, bipartisan legislative action, international coordination breakthrough | 2026-2028 window; requires sustained 3-5 year commitment |
| Muddle through | 45-55% | AISI network grows modestly (15-20 institutes by 2027); EU enforcement proceeds with gaps; US capacity stagnates; industry remains 80-90% self-governing | Status quo political dynamics, incremental funding increases, continued voluntary cooperation | 2025-2030; baseline trajectory |
| Capacity decline | 20-25% | Budget cuts (30-50% reductions), talent drain (net negative hiring), and political deprioritization reduce regulatory capability; safety depends 95%+ on industry self-governance | Economic recession, anti-regulation political shift, US-China competition prioritizes speed over safety | 2025-2027; accelerated by administration changes |
| Regulatory innovation | 10-15% | AI-assisted oversight, novel funding models (industry levies), or international pooling dramatically improve capacity efficiency (3-5x multiplier effect) | Technical breakthroughs in automated evaluation, new governance models (e.g., AI Safety Institutes gain enforcement authority) | 2026-2029; requires both technical and political innovation |
To provide meaningful oversight of frontier AI development, regulators would need capacity to evaluate major model releases before deployment. Current capacity falls far short:
| Metric | Current State | Required for Adequate Oversight | Gap Magnitude |
|---|---|---|---|
| Models evaluated per year | 2-3 (UK AISI, 2024) | 12-24 (quarterly releases from 4-6 frontier labs) | 4-8x shortage |
| Evaluation time per model | 8-12 weeks | 2-4 weeks (to avoid deployment delays) | 2-3x too slow |
| Technical staff per evaluation | 10-15 researchers | 20-30 (to match lab eval teams) | 2x shortage |
| Budget per evaluation | $500K-1M (estimated) | $2-5M (comprehensive red-teaming) | 2-5x underfunded |
| Annual evaluation capacity | $2-3M total | $30-60M (if all frontier labs evaluated) | 10-20x shortfall |
Implication: Current AISI network capacity would need to grow 10-20x to provide pre-deployment evaluation of all frontier models. At current growth rates (doubling every 18-24 months), adequate capacity would require 5-7 years—likely longer than the timeline to transformative AI systems.
The salary differential creates structural barriers to regulatory capacity:
| Position Level | Industry Compensation | Government Compensation | Multiplier | Annual Talent Loss Estimate |
|---|---|---|---|---|
| Entry-level ML engineer | $180-250K total comp | $80-120K | 1.5-2x | 60-70% choose industry |
| Senior researcher | $400-800K total comp | $150-200K | 2.5-4x | 75-85% choose industry |
| Principal/Staff level | $800K-2M total comp | $180-250K | 3-8x | 85-95% choose industry |
| Top 1% talent | $2-5M+ (equity-heavy) | $200-280K (GS-15 max) | 7-20x | 95-99% choose industry |
The 2024 federal AI hiring initiative offers recruitment incentives up to 25% of base pay (plus relocation, retention bonuses, and $60K student loan repayment). This improves the situation at entry levels but leaves senior/principal gaps unchanged:
Implication: Government can potentially hire entry-level talent with aggressive incentives, but acquiring senior expertise required to lead evaluations faces near-insurmountable compensation barriers. Estimates suggest 70-85% of regulatory technical leadership comes from individuals unable to secure equivalent industry positions, not from top-tier talent choosing public service.
The EU AI Act provides a test case for enforcement capacity needs. With 27 member states and an estimated 500-2,000 high-risk AI systems requiring compliance:
| Enforcement Function | Estimated Annual Cost per Member State | Total EU Cost (27 states) | Current Budget Allocation |
|---|---|---|---|
| Authority setup | $2-5M (one-time) | $54-135M | Unknown—only 3 states compliant |
| Market surveillance | $5-10M annually | $135-270M | Severely underfunded |
| Conformity assessment | $10-20M annually | $270-540M | Mostly delegated to private notified bodies |
| Incident investigation | $3-8M annually | $81-216M | Not yet established |
| Penalty enforcement | $2-5M annually | $54-135M | Zero enforcement actions to date |
| Total annual requirement | $20-43M | $540-1,160M | $8M EU AI Office (2024) |
Gap assessment: The EU AI Office budget of ~$8M represents 0.7-1.5% of estimated enforcement requirements. Even if member states collectively spend 10x the EU Office budget ($80M total), this reaches only 7-15% of required capacity. The 11% compliance rate (3 of 27 states designated authorities by deadline) suggests many states lack resources for even basic administrative setup.
Arguments for feasibility:
Arguments against:
Arguments for voluntary (NIST AI RMF approach):
Arguments against:
The trajectory of the US AI Safety Institute illustrates both the potential and fragility of regulatory capacity:
| Phase | Date | Development |
|---|---|---|
| Founding | November 2023 | AISI established at NIST; $10M initial budget |
| Momentum | 2024 | Director appointed; agreements signed with Anthropic, OpenAI |
| Demonstrated value | November 2024 | Joint evaluation of Claude 3.5 Sonnet published |
| Political shift | January 2025 | EO 14110 revoked; AISI future uncertain |
| Transformation | June 2025 | Renamed CAISI; mission shifted from safety to innovation |
Key lesson: Regulatory capacity built over 18 months was effectively redirected in weeks, demonstrating the fragility of government capacity without legislative foundation.
NIST AI RMF adoption shows uneven capacity effects across sectors:
| Sector | Adoption Rate | Implementation Depth | Capacity Effect |
|---|---|---|---|
| Financial services | 70-75% | High (full four-function) | Significant |
| Healthcare | 60-65% | Medium-High | Moderate |
| Technology | 45-70% | Variable | Mixed |
| Government | 30-40% (rising) | Growing | Building |
| Retail | 25-35% | Low | Minimal |
Key lesson: Voluntary frameworks achieve highest adoption where existing regulatory culture (finance, healthcare) creates implementation incentives.
Auto-generated from the master graph. Shows key relationships.