Edited today2.0k words2 backlinksUpdated every 3 weeksDue in 3 weeks
72QualityGood •Quality: 72/100LLM-assigned rating of overall page quality, considering depth, accuracy, and completeness.Structure suggests 10078.5ImportanceHighImportance: 78.5/100How central this topic is to AI safety. Higher scores mean greater relevance to understanding or mitigating AI risk.67ResearchModerateResearch Value: 67/100How much value deeper investigation of this topic could yield. Higher scores indicate under-explored topics with high insight potential.
Summary
Comprehensive analysis of pause/moratorium proposals finding they would provide very high safety benefits if implemented (buying time for safety research to close the growing capability-safety gap) but face critical enforcement and coordination challenges with zero current adoption by major labs. The FLI 2023 open letter garnered 30,000+ signatures but resulted in no actual slowdown, highlighting severe tractability issues despite theoretical effectiveness.
Content8/13
LLM summaryLLM summaryBasic text summary used in search results, entity link tooltips, info boxes, and related page cards.ScheduleScheduleHow often the page should be refreshed. Drives the overdue tracking system.EntityEntityYAML entity definition with type, description, and related entries.Edit historyEdit historyTracked changes from improve pipeline runs and manual edits.crux edit-log view <id>OverviewOverviewA ## Overview heading section that orients readers. Helps with search and AI summaries.
Tables19/ ~8TablesData tables for structured comparisons and reference material.Diagrams2/ ~1DiagramsVisual content — Mermaid diagrams, charts, or Squiggle estimate models.–Int. links7/ ~16Int. linksLinks to other wiki pages. More internal links = better graph connectivity.Add links to other wiki pagesExt. links26/ ~10Ext. linksLinks to external websites, papers, and resources outside the wiki.Footnotes0/ ~6FootnotesFootnote citations [^N] with source references at the bottom of the page.Add [^N] footnote citationsReferences8/ ~6ReferencesCurated external resources linked via <R> components or cited_by in YAML.Quotes0QuotesSupporting quotes extracted from cited sources to back up page claims.crux citations extract-quotes <id>Accuracy0AccuracyCitations verified against their sources for factual accuracy.crux citations verify <id>RatingsN:4.2 R:6.8 A:5.5 C:7.5RatingsSub-quality ratings: Novelty, Rigor, Actionability, Completeness (0-10 scale).Backlinks2BacklinksNumber of other wiki pages that link to this page. Higher backlink count means better integration into the knowledge graph.
Issues2
QualityRated 72 but structure suggests 100 (underrated by 28 points)
Links15 links could use <R> components
Pause / Moratorium
Policy
Pause / Moratorium
Comprehensive analysis of pause/moratorium proposals finding they would provide very high safety benefits if implemented (buying time for safety research to close the growing capability-safety gap) but face critical enforcement and coordination challenges with zero current adoption by major labs. The FLI 2023 open letter garnered 30,000+ signatures but resulted in no actual slowdown, highlighting severe tractability issues despite theoretical effectiveness.
Future of Life Institute (FLI)OrganizationFuture of Life Institute (FLI)Comprehensive profile of FLI documenting \$25M+ in grants distributed (2015: \$7M to 37 projects, 2021: \$25M program), major public campaigns (Asilomar Principles with 5,700+ signatories, 2023 Pau...Quality: 46/100
People
Stuart RussellPersonStuart RussellStuart Russell (born 1962) is a British computer scientist and UC Berkeley professor who co-authored the dominant AI textbook 'Artificial Intelligence: A Modern Approach' (used in over 1,500 univer...Quality: 30/100
Risks
AI Development Racing DynamicsRiskAI Development Racing DynamicsRacing dynamics analysis shows competitive pressure has shortened safety evaluation timelines by 40-60% since ChatGPT's launch, with commercial labs reducing safety work from 12 weeks to 4-6 weeks....Quality: 72/100
Approaches
Pause AdvocacyApproachPause AdvocacyComprehensive analysis of pause advocacy as an AI safety intervention, estimating 15-40% probability of meaningful policy implementation by 2030 with potential to provide 2-5 years of additional sa...Quality: 91/100
2k words · 2 backlinks
Overview
Pause and moratorium proposals represent the most direct governance intervention for AI safety: deliberately slowing or halting frontier AI development to allow safety research, governance frameworks, and societal preparation to catch up with rapidly advancing capabilities. These proposals range from targeted pauses on specific capability thresholds to comprehensive moratoria on all advanced AI development, with proponents arguing that the current pace of development may be outstripping humanity's ability to ensure safe deployment.
The most prominent call for a pause came in March 2023, when the Future of Life InstituteOrganizationFuture of Life Institute (FLI)Comprehensive profile of FLI documenting \$25M+ in grants distributed (2015: \$7M to 37 projects, 2021: \$25M program), major public campaigns (Asilomar Principles with 5,700+ signatories, 2023 Pau...Quality: 46/100 (FLI) published an open letter calling for a six-month pause on training AI systems more powerful than GPT-4. Released just one week after GPT-4's launch, the letter garnered over 30,000 signatures, including prominent AI researchers such as Yoshua Bengio and Stuart RussellPersonStuart RussellStuart Russell (born 1962) is a British computer scientist and UC Berkeley professor who co-authored the dominant AI textbook 'Artificial Intelligence: A Modern Approach' (used in over 1,500 univer...Quality: 30/100, as well as technology leaders like Elon MuskPersonElon MuskComprehensive profile of Elon Musk's role in AI, documenting his early safety warnings (2014-2017), OpenAI founding and contentious departure, xAI launch and funding history, Neuralink BCI developm...Quality: 38/100 and Steve Wozniak. The letter cited risks including AI-generated propaganda, extreme automation of jobs, and a society-wide loss of control. However, no major AI laboratory implemented a voluntary pause, and the letter's six-month timeline passed without meaningful slowdown in frontier development. As MIT Technology Review noted six months later, AI companies instead directed "vast investments in infrastructure to train ever-more giant AI systems."
The fundamental logic behind pause proposals is straightforward: if AI development is proceeding faster than our ability to make it safe, slowing development provides time for safety work. As Bengio et al. wrote in Science in May 2024, "downside artificial intelligence risks must be managed effectively and urgently if posited AI benefits are to be realized safely." However, implementation faces severe challenges including competitive dynamics between nations and companies, enforcement difficulties, and concerns that pauses might push development underground or to jurisdictions with fewer safety constraints. These proposals remain controversial even within the AI safety community, with some arguing they are essential for survival and others viewing them as impractical or counterproductive.
Risk Assessment & Impact
Dimension
Assessment
Rationale
Confidence
Safety Uplift
High (if implemented)
Would buy time for safety research
High
Capability Uplift
Negative
Explicitly slows capability development
High
Net World Safety
Unclear
Could help if coordinated; could backfire if unilateral
Medium
Lab Incentive
Negative
Labs strongly opposed; competitive dynamics
High
Research Investment
$1-5M/yr
Advocacy organizations; FLI, PauseAI
Medium
Current Adoption
None
Advocacy only; no major labs paused
High
Core Arguments
Case for Pausing
Loading diagram...
Argument
Description
Strength
Safety-Capability Gap
Safety research not keeping pace with capabilities
Strong if gap is real
Irreversibility
Some AI risks may be impossible to reverse once realized
Strong for existential risks
Precautionary Principle
Burden of proof should be on developers to show safety
Philosophically contested
Coordination Signal
Demonstrates seriousness; creates space for governance
Moderate
Research Time
Enables catch-up on interpretability, alignment
Strong
Case Against Pausing
Argument
Description
Strength
Enforcement
Unenforceable without international agreement
Strong
Displacement
Development moves to less cautious actors
Moderate-Strong
Lost Benefits
Delays positive AI applications
Moderate
Talent Dispersion
Safety researchers may leave paused organizations
Moderate
False Security
Pause without progress creates complacency
Moderate
Definition Problems
Hard to define what to pause
Strong
Pause Proposals Analyzed
FLI Open Letter (2023)
Aspect
Detail
Scope
Training systems more powerful than GPT-4
Duration
Six months (renewable)
Signatories
30,000+ including Yoshua BengioPersonYoshua BengioComprehensive biographical overview of Yoshua Bengio's transition from deep learning pioneer (Turing Award 2018) to AI safety advocate, documenting his 2020 pivot at Mila toward safety research, co...Quality: 39/100, Elon MuskPersonElon MuskComprehensive profile of Elon Musk's role in AI, documenting his early safety warnings (2014-2017), OpenAI founding and contentious departure, xAI launch and funding history, Neuralink BCI developm...Quality: 38/100, Stuart RussellPersonStuart RussellStuart Russell (born 1962) is a British computer scientist and UC Berkeley professor who co-authored the dominant AI textbook 'Artificial Intelligence: A Modern Approach' (used in over 1,500 univer...Quality: 30/100, Steve Wozniak, Yuval Noah Harari
Notable critiques: AI researcher Andrew Ng argued that "there is no realistic way to implement a moratorium" without government intervention, which would be "anti-competitive" and "awful innovation policy." Reid Hoffman criticized the letter as "virtue signaling" that would hurt the cause by alienating the AI developer community needed to achieve safety goals.
PauseAI Movement
Aspect
Detail
Founded
May 2023 in Utrecht, Netherlands by software entrepreneur Joep Meindertsma
Network of local organizations; US chapter led by Holly Elmore, UK by Joseph Miller (Oxford PhD)
Approach
Grassroots activism, protests at AI labs (OpenAI Feb 2024, Anthropic Nov 2024), policy advocacy
Policy Asks
Global pause enforced through international treaty; democratic control over AI development
Key Actions
International protests in May 2024 timed to Seoul AI Safety Summit; protests held in San Francisco, New York, Berlin, Rome, Ottawa, London
Academic Proposals
Proposal
Scope
Mechanism
Compute Caps
Limit training compute
Hardware governance
Capability Gates
Pause at defined capability thresholds
Eval-based triggers
Conditional Pause
Pause if safety benchmarks not met
RSP-like framework
Research Moratoria
Pause specific capability research
Targeted restrictions
Implementation Challenges
Coordination Problems
Challenge
Description
Severity
Potential Solution
International Competition
US-China dynamics; neither wants to pause first
Critical
Treaty with verification
Corporate Competition
First-mover advantages; defection incentives
High
Regulatory mandate
Verification
How to confirm compliance
High
Compute monitoring
Definition
What counts as "frontier" AI
High
Clear technical thresholds
Enforcement Mechanisms
Mechanism
Feasibility
Effectiveness
Notes
Voluntary Compliance
Low
Very Low
No incentive to comply
National Regulation
Medium
Medium
Jurisdictional limits
International Treaty
Low-Medium
High if achieved
Requires major power agreement
Compute Restrictions
Medium
Medium-High
Physical infrastructure trackable
Social Pressure
Medium
Low
Insufficient against strong incentives
Unintended Consequences
Consequence
Likelihood
Severity
Mitigation
Development Displacement
High
High
International coordination
Underground Development
Medium
Very High
Compute monitoring
Safety Researcher Exodus
Medium
Medium
Continued safety funding
Competitive Disadvantage
High
Variable
Coordinated action
Delayed Benefits
High
Medium
Risk-benefit analysis
Historical Precedents
Domain
Intervention
Outcome
Lessons
Nuclear Weapons
Various moratoria and treaties
Partial success; proliferation continued
Verification essential
Human Cloning
Research moratoria
Generally effective
Narrow scope helps
Gain-of-Function
Research pause (2014-2017)
Temporary; research resumed
Pressure to resume
Recombinant DNA
Asilomar conference (1975)
Self-regulation worked initially
Community buy-in crucial
CFCs
Montreal Protocol
Highly successful
Clear harm identification
Lessons for AI Pauses
Narrow scope is more enforceable than broad moratoria
Verification mechanisms are essential for compliance
International coordination requires identifying mutual interests
Community buy-in from researchers enables voluntary compliance
Clear triggering conditions help define when restrictions apply
Scalability Assessment
Dimension
Assessment
Rationale
International Scalability
Unknown
Depends on coordination
Enforcement Scalability
Partial
Compute monitoring possible
SI Readiness
Yes (if works)
Would prevent reaching SI until prepared
Deception Robustness
N/A
External policy; doesn't address model behavior
Strategic Considerations
When Pauses Might Work
Condition
Importance
Current Status
International Agreement
Critical
Very limited
Clear Triggers
High
Undefined
Verification Methods
High
Underdeveloped
Alternative Pathway
Medium
Safety research ongoing
Industry Buy-In
Medium-High
Very low
Alternative Approaches
Alternative
Relationship to Pause
Tradeoffs
Differential Progress
Accelerate safety, not slow capabilities
Competitive with capabilities
Responsible Scaling Policies
Conditional pauses at thresholds
Voluntary; lab-controlled
Compute Governance
Indirect slowdown through resource control
More enforceable
International Coordination
Framework for coordinated pause
Slower to achieve
Quick Assessment
Dimension
Assessment
Evidence
Tractability
Low
Severe coordination and enforcement challenges; no major lab has voluntarily paused
Effectiveness
Very High (if implemented)
Would directly address timeline concerns by buying time for safety research
Neglectedness
Medium
Active advocacy (FLI, PauseAI); major gap in implementation and enforcement mechanisms
Current Maturity
Early Advocacy
FLI letter catalyzed debate but no binding commitments achieved
Time Horizon
Immediate-Long Term
Could theoretically be implemented quickly but requires international coordination
Key Proponents
FLI, PauseAI, Yoshua Bengio
Grassroots movements and prominent AI researchers
Key Opponents
Major AI Labs, Andrew Ng
Competitive dynamics and concerns about practicality
Risks Addressed
If implemented effectively, pause/moratorium would address:
Risk
Mechanism
Effectiveness
AI Development Racing DynamicsRiskAI Development Racing DynamicsRacing dynamics analysis shows competitive pressure has shortened safety evaluation timelines by 40-60% since ChatGPT's launch, with commercial labs reducing safety work from 12 weeks to 4-6 weeks....Quality: 72/100
Eliminates competitive pressure
Very High
Safety-Capability Gap
Time for safety research
Very High
Governance Lag
Time for policy development
High
Societal Preparation
Time for adaptation
High
****
Prevents deployment of unaligned systems
Very High (during pause)
Limitations
Enforcement Infeasibility: No mechanism to enforce global compliance
The Future of Life Institute works to guide transformative technologies like AI towards beneficial outcomes and away from large-scale risks. They engage in policy advocacy, research, education, and grantmaking to promote safe and responsible technological development.
A research organization focused on understanding AI's societal impacts, governance challenges, and policy implications across various domains like workforce, infrastructure, and public perception.
AnthropicOrganizationAnthropicComprehensive reference page on Anthropic covering financials (\$380B valuation, \$19B ARR), safety research (Constitutional AI, mechanistic interpretability, model welfare), governance (LTBT struc...Quality: 74/100OpenAIOrganizationOpenAIComprehensive organizational profile of OpenAI documenting evolution from 2015 non-profit to Public Benefit Corporation, with detailed analysis of governance crisis, 2024-2025 ownership restructuri...Quality: 62/100xAIOrganizationxAIComprehensive profile of xAI covering its founding by Elon Musk in 2023, rapid growth to \$230B valuation and \$3.8B revenue, development of Grok models, and controversial 'truth-seeking' safety ap...Quality: 48/100
Risks
Multipolar Trap (AI Development)RiskMultipolar Trap (AI Development)Analysis of coordination failures in AI development using game theory, documenting how competitive dynamics between nations (US \$109B vs China \$9.3B investment in 2024 per Stanford HAI 2025) and ...Quality: 91/100AI Trust Cascade FailureRiskAI Trust Cascade FailureAnalysis of how declining institutional trust (media 31%, federal government 17% per 2024-2025 Gallup/Pew data) could create self-reinforcing collapse where no trusted entity can validate others, p...Quality: 55/100
Approaches
Multi-Agent SafetyApproachMulti-Agent SafetyMulti-agent safety addresses coordination failures, conflict, and collusion risks when AI systems interact. A 2025 report from 50+ researchers identifies seven key risk factors; empirical studies s...Quality: 68/100Corporate AI Safety ResponsesApproachCorporate AI Safety ResponsesMajor AI labs invest \$300-500M annually in safety (5-10% of R&D) through responsible scaling policies and dedicated teams, but face 30-40% safety team turnover and significant implementation gaps ...Quality: 68/100
Analysis
Multipolar Trap Dynamics ModelAnalysisMultipolar Trap Dynamics ModelGame-theoretic analysis of AI competition traps showing universal cooperation probability drops from 81% (2 actors) to 21% (15 actors), with 5-10% catastrophic lock-in risk and 20-35% partial coord...Quality: 61/100Anthropic Impact Assessment ModelAnalysisAnthropic Impact Assessment ModelModels Anthropic's net impact on AI safety by weighing positive contributions (safety research \$100-200M/year, Constitutional AI as industry standard, largest interpretability team globally, RSP f...Quality: 55/100Racing Dynamics Game Theory ModelAnalysisRacing Dynamics Game Theory ModelGame-theoretic analysis of competitive pressures in AI development, modeling safety-capability tradeoffs as prisoner's dilemma with asymmetric payoffs.AI Lab Incentives ModelAnalysisAI Lab Incentives ModelThis model analyzes competitive and reputational pressures on lab safety decisions. It identifies conditions where market dynamics systematically underweight safety investment.
Other
Elon MuskPersonElon MuskComprehensive profile of Elon Musk's role in AI, documenting his early safety warnings (2014-2017), OpenAI founding and contentious departure, xAI launch and funding history, Neuralink BCI developm...Quality: 38/100
Concepts
Alignment Policy OverviewAlignment Policy OverviewThis is a stub overview page that lists four policy/governance topic areas (RSPs, model specs, evaluation governance, pause/moratorium) with one-line descriptions and links to deeper pages. It cont...Quality: 21/100
Key Debates
AI Safety Solution CruxesCruxAI Safety Solution CruxesA comprehensive structured mapping of AI safety solution uncertainties across technical, alignment, governance, and agentic domains, using probability-weighted crux frameworks with specific estimat...Quality: 65/100AI Structural Risk CruxesCruxAI Structural Risk CruxesAnalyzes 12 key uncertainties about AI structural risks across power concentration, coordination feasibility, and institutional adaptation. Provides quantified probability ranges: US-China coordina...Quality: 66/100Should We Pause AI Development?CruxShould We Pause AI Development?Comprehensive synthesis of the AI pause debate showing moderate expert support (35-40% of 2,778 researchers) and high public support (72%) but very low implementation feasibility, with all major la...Quality: 47/100
Policy
International Compute RegimesPolicyInternational Compute RegimesComprehensive analysis of international AI compute governance finds 10-25% chance of meaningful regimes by 2035, but potential for 30-60% reduction in racing dynamics if achieved. First binding tre...Quality: 67/100