Longterm Wiki
Navigation
Updated 2026-03-13HistoryData
Page StatusContent
Edited today349 words
Content2/13
LLM summaryScheduleEntityEdit historyOverview
Tables0/ ~1Diagrams0Int. links33/ ~3Ext. links0/ ~2Footnotes0/ ~2References0/ ~1Quotes0Accuracy0
Issues1
StructureNo tables or diagrams - consider adding visual content

Knowledge Base

Overview

The LongtermWiki Knowledge Base provides structured documentation of the AI safety landscape, covering risks, interventions, organizations, and key debates. Content is organized to help researchers, funders, and policymakers understand the current state of AI safety and make informed decisions about resource allocation.

Content Categories

Risks

Documentation of potential failure modes and hazards from advanced AI systems, organized by type:

  • Accident Risks - Unintended failures like scheming, deceptive alignment, mesa-optimization
  • Misuse Risks - Deliberate harmful applications like bioweapons, cyberweapons, disinformation
  • Structural Risks - Systemic issues like racing dynamics, concentration of power, lock-in
  • Epistemic Risks - Threats to knowledge and truth like authentication collapse, trust erosion

Responses

Interventions and approaches to address AI risks:

  • Technical Alignment - Interpretability, RLHF, constitutional AI, AI control
  • Governance - Compute governance, international coordination, legislation
  • Institutional - AI safety institutes, standards bodies
  • Epistemic Tools - Prediction markets, content authentication, coordination technologies

Models

Analytical frameworks for understanding AI risk dynamics:

  • Framework Models - Carlsmith's six premises, instrumental convergence
  • Risk Models - Deceptive alignment decomposition, scheming likelihood
  • Dynamics Models - Racing dynamics impact, feedback loops
  • Societal Models - Trust erosion, lock-in mechanisms

Organizations

Profiles of key actors in AI development and safety:

  • AI Labs - OpenAI, Anthropic, DeepMind, xAI
  • Safety Research Orgs - MIRI, ARC, Redwood, Apollo Research
  • Government Bodies - US AISI, UK AISI

People

Profiles of influential researchers and leaders in AI safety.

Capabilities

Documentation of AI capability domains and their safety implications.

Debates

Structured analysis of key disagreements in the field.

Cruxes

Key uncertainties that drive disagreement and prioritization decisions.

How to Use This Knowledge Base

  1. Exploring risks: Start with the scheming page for the most discussed risk, then browse related accident risks
  2. Understanding responses: See interpretability for a well-documented technical approach
  3. Analytical depth: The Carlsmith six-premise model provides a rigorous framework for AI risk estimation
  4. Browse everything: Use the Browse page to search and filter all entries

Quality Indicators

Pages include quality and importance ratings:

  • Quality (0-100): How well-developed and accurate the content is
  • Importance (0-100): How significant the topic is for AI safety decisions

High-priority pages (quality < importance) are actively being improved.

Related Pages

Top Related Pages

Organizations

AnthropicOpenAI

Risks

Deceptive AlignmentAI Development Racing DynamicsInstrumental ConvergenceAI-Driven Concentration of PowerAI Value Lock-inMesa-Optimization

Key Debates

AI Accident Risk CruxesIs AI Existential Risk Real?

Policy

AI Safety Institutes (AISIs)

Concepts

Agentic AIRLHF

Approaches

Constitutional AI

Safety Research

Interpretability

Analysis

Carlsmith's Six-Premise ArgumentAI Risk Activation Timeline Model

Other

Paul Christiano