LLM summaryLLM summaryBasic text summary used in search results, entity link tooltips, info boxes, and related page cards.crux content improve <id>ScheduleScheduleHow often the page should be refreshed. Drives the overdue tracking system.Set updateFrequency in frontmatterEntityEntityYAML entity definition with type, description, and related entries.Add entity YAML in data/entities/Edit historyEdit historyTracked changes from improve pipeline runs and manual edits.crux edit-log view <id>OverviewOverviewA ## Overview heading section that orients readers. Helps with search and AI summaries.
Tables0/ ~1TablesData tables for structured comparisons and reference material.Add data tables to the pageDiagrams0DiagramsVisual content — Mermaid diagrams, charts, or Squiggle estimate models.Add Mermaid diagrams or Squiggle modelsInt. links11/ ~3Int. linksLinks to other wiki pages. More internal links = better graph connectivity.Ext. links0/ ~1Ext. linksLinks to external websites, papers, and resources outside the wiki.Add links to external sourcesFootnotes0/ ~2FootnotesFootnote citations [^N] with source references at the bottom of the page.Add [^N] footnote citationsReferences0/ ~1ReferencesCurated external resources linked via <R> components or cited_by in YAML.Add <R> resource linksQuotes0QuotesSupporting quotes extracted from cited sources to back up page claims.crux citations extract-quotes <id>Accuracy0AccuracyCitations verified against their sources for factual accuracy.crux citations verify <id>
Issues1
StructureNo tables or diagrams - consider adding visual content
Key Debates
Overview
This section presents structured analyses of major debates in AI safety. For each debate, we present the strongest arguments on multiple sides, identify key cruxes, and track how expert opinion has evolved.
Major Debates
Foundational Questions
Is AI Existential Risk Real?CruxIs AI Existential Risk Real?Presents two core cruxes in the AI x-risk debate: whether advanced AI would develop dangerous goals (instrumental convergence vs. trainable safety) and whether we'll get warning signs (gradual fail...Quality: 12/100 - The core question of whether AI poses genuine existential risk
AGI Timeline DebateCruxWhen Will AGI Arrive?Comprehensive survey of AGI timeline predictions ranging from 2025-2027 (ultra-short) to never with current approaches, with median expert estimates around 2032-2037. Key cruxes include whether sca...Quality: 33/100 - When might transformative AI arrive?
Scaling DebateCruxIs Scaling All You Need?Comprehensive survey of the 2024-2025 scaling debate, documenting the shift from pure pretraining to 'scaling-plus' approaches after o3 achieved 87.5% on ARC-AGI-1 but GPT-5 faced 2-year delays. Ex...Quality: 42/100 - Will scaling alone produce AGI?
Alignment Approach
Why Alignment Might Be HardArgumentWhy Alignment Might Be HardA comprehensive taxonomy of alignment difficulty arguments spanning specification problems, inner alignment failures, verification limits, and adversarial dynamics, with expert p(doom) estimates ra...Quality: 69/100 - Arguments for difficulty
Why Alignment Might Be EasyArgumentWhy Alignment Might Be EasySynthesizes empirical evidence that alignment is tractable, citing 29-41% RLHF improvements, Constitutional AI reducing bias across 9 dimensions, millions of interpretable features from Claude 3, a...Quality: 53/100 - Arguments for tractability
Interpretability Sufficient?CruxIs Interpretability Sufficient for Safety?Comprehensive survey of the interpretability sufficiency debate with 2024-2025 empirical progress: Anthropic extracted 34M features from Claude 3 Sonnet (70% interpretable), but scaling requires bi...Quality: 49/100 - Can we rely on understanding models?
Strategy & Policy
Pause DebateCruxShould We Pause AI Development?Comprehensive synthesis of the AI pause debate showing moderate expert support (35-40% of 2,778 researchers) and high public support (72%) but very low implementation feasibility, with all major la...Quality: 47/100 - Should AI development be paused?
Open vs ClosedCruxOpen vs Closed Source AIComprehensive analysis of open vs closed source AI debate, documenting that open model performance gap narrowed from 8% to 1.7% in 2024, with 1.2B+ Llama downloads by April 2025 and DeepSeek R1 dem...Quality: 60/100 - Open-source AI tradeoffs
Regulation DebateCruxGovernment Regulation vs Industry Self-GovernanceComprehensive comparison of government regulation versus industry self-governance for AI, documenting that US federal AI regulations doubled to 59 in 2024 while industry lobbying surged 141% to 648...Quality: 54/100 - Government intervention approaches
Risk Assessment
The Case FOR AI Existential RiskArgumentThe Case For AI Existential RiskComprehensive formal argument that AI poses 5-14% median extinction risk by 2100 (per 2,788 researcher survey), structured around four premises: capabilities will advance, alignment is hard (with d...Quality: 66/100 - Strongest arguments for concern
The Case AGAINST AI Existential RiskArgumentThe Case Against AI Existential RiskComprehensive synthesis of skeptical arguments against AI x-risk from prominent researchers (LeCun, Marcus, Ng, Brooks), concluding x-risk probability is <5% (likely ~2%) based on challenges to sca...Quality: 58/100 - Strongest counterarguments
How Debates Are Structured
Each debate page includes:
Steelmanned positions - Best versions of each argument
Key cruxes - What would change minds
Evidence and citations - Supporting data for each side
Expert distribution - Where informed people fall
Implications - How resolution would affect priorities
Why Alignment Might Be EasyArgumentWhy Alignment Might Be EasySynthesizes empirical evidence that alignment is tractable, citing 29-41% RLHF improvements, Constitutional AI reducing bias across 9 dimensions, millions of interpretable features from Claude 3, a...Quality: 53/100Is Interpretability Sufficient for Safety?CruxIs Interpretability Sufficient for Safety?Comprehensive survey of the interpretability sufficiency debate with 2024-2025 empirical progress: Anthropic extracted 34M features from Claude 3 Sonnet (70% interpretable), but scaling requires bi...Quality: 49/100Should We Pause AI Development?CruxShould We Pause AI Development?Comprehensive synthesis of the AI pause debate showing moderate expert support (35-40% of 2,778 researchers) and high public support (72%) but very low implementation feasibility, with all major la...Quality: 47/100Is Scaling All You Need?CruxIs Scaling All You Need?Comprehensive survey of the 2024-2025 scaling debate, documenting the shift from pure pretraining to 'scaling-plus' approaches after o3 achieved 87.5% on ARC-AGI-1 but GPT-5 faced 2-year delays. Ex...Quality: 42/100When Will AGI Arrive?CruxWhen Will AGI Arrive?Comprehensive survey of AGI timeline predictions ranging from 2025-2027 (ultra-short) to never with current approaches, with median expert estimates around 2032-2037. Key cruxes include whether sca...Quality: 33/100Is AI Existential Risk Real?CruxIs AI Existential Risk Real?Presents two core cruxes in the AI x-risk debate: whether advanced AI would develop dangerous goals (instrumental convergence vs. trainable safety) and whether we'll get warning signs (gradual fail...Quality: 12/100