Skip to content

Longterm Wiki

📋Page Status
Page Type:ResponseStyle Guide →Intervention/response page
Quality:63 (Good)⚠️
Importance:12 (Peripheral)
Words:2.2k
Structure:
📊 24📈 7🔗 24📚 134%Score: 15/15
LLM Summary:A self-referential documentation page describing the Longterm Wiki platform itself—a strategic intelligence tool with ~550 pages, crux mapping of ~50 uncertainties, and quality scoring across 6 dimensions. Features include entity cross-linking, interactive causal diagrams, and structured YAML databases tracking expert positions on key AI safety cruxes.
Issues (2):
  • QualityRated 63 but structure suggests 100 (underrated by 37 points)
  • Links4 links could use <R> components
DimensionAssessmentEvidence
ScopeAI safety focused≈550 pages covering risks, interventions, organizations, cruxes
Content ModelCurated synthesisEditorial control, quality scoring, not community wiki
Unique ValueCrux mappingExplicit uncertainty tracking, worldview→priority linkages
TechnicalModern stackAstro/Starlight, React, interactive causal diagrams
Open SourceFullyMIT licensed, GitHub repository
StatusActive developmentLaunched 2025, ongoing content expansion
AttributeDetails
NameLongterm Wiki (LongtermWiki)
OrganizationQURI (Quantified Uncertainty Research Institute)
LeadOzzie Gooen
Launched2025
Websitelongtermwiki.com
GitHubgithub.com/quantified-uncertainty/cairn
LicenseMIT
PlatformStatic site (Astro/Starlight)

The Longterm Wiki is a strategic intelligence platform for AI safety prioritization. Unlike general encyclopedias or community wikis, it serves as a decision-support tool for funders, researchers, and policymakers asking: “Where should the next marginal dollar or researcher-hour go?”

Loading diagram...

The project addresses four problems in the AI safety field:

ProblemHow the Wiki Addresses It
Fragmented knowledgeConsolidated, cross-linked knowledge base with ≈550 pages
Unclear cruxesExplicit mapping of key uncertainties and expert disagreements
Poor prioritization legibilityWorldview → intervention mapping showing how assumptions lead to priorities
Slow information synthesisQuality tracking, staleness monitoring, regular updates

The wiki is deliberately opinionated about importance and uncertainty—it rates content quality, tracks expert positions on cruxes, and makes prioritization implications explicit. This distinguishes it from neutral reference works like Wikipedia or discussion platforms like LessWrong.

Content is editorially curated rather than community-contributed, ensuring consistency and quality control. Each page goes through a grading pipeline that scores novelty, rigor, actionability, and completeness.

The wiki has four interconnected layers of content:

Loading diagram...
LayerContentsScale
Factual FoundationRisks, interventions, organizations, people, concepts≈350 pages
Causal ModelsRisk pathways, intervention effects, transition dynamics≈80 models
Uncertainty StructureKey cruxes, expert positions, worldview clusters≈50 cruxes, ~30 experts
Strategic ImplicationsPriority rankings, robust interventions, high-VOI researchDerived from above
SectionContentPage CountExample Pages
Knowledge BaseRisks, interventions, organizations, people≈350Deceptive Alignment, AI Safety Institutes
AI Transition ModelComprehensive factor network with outcomes and scenarios≈80Factors, scenarios, quantitative estimates
Analytical ModelsRisk models, cascade models, governance dynamics≈80Cascade models, dynamics models
ProjectVision, strategy, similar projects analysis≈10Vision, Similar Projects
InternalStyle guides, automation tools, technical docs≈30Automation Tools
TypePurposeQuality Scored?Validation
ContentSubstantive knowledge base pagesYes (0-100 scale)Full (schema, links, escaping, quality)
StubRedirects, brief profilesNoMinimal (compilation, links)
DocumentationStyle guides, internal referenceNoCompilation only
OverviewNavigation hub pages (index.mdx)NoSidebar structure

The wiki explicitly tracks key uncertainties that drive disagreement and prioritization:

Loading diagram...
Crux CategoryExamplesTracked In
TechnicalP(deceptive alignment), interpretability tractability, mesa-optimization riskcruxes.yaml
TimelineYears to transformative AI, takeoff speed, capability jumpsestimates.yaml
GovernanceCoordination tractability, warning shot likelihood, regulatory capacitycruxes.yaml
ValueWeight on near-term vs. long-term, risk tolerance, discount ratesexperts.yaml positions

Expert positions on cruxes are tracked in structured YAML databases, enabling:

  • Disagreement decomposition: Breaking “Is alignment hard?” into component empirical claims
  • Worldview clustering: Grouping experts by crux positions (e.g., “short-timeline technical doomer”)
  • Priority implications: Showing how different assumptions lead to different resource allocations

Stable ID-based linking system that survives URL reorganization:

import {EntityLink} from '@components/wiki';
The <EntityLink id="scheming">scheming</EntityLink> risk relates to
<EntityLink id="deceptive-alignment">deceptive alignment</EntityLink>.
FeatureBenefit
Stable IDsLinks survive path reorganization
Auto title lookupComponent fetches current title from database
Backlink trackingEvery page shows what links to it
CI validationBroken links caught before deployment
Type iconsVisual distinction between risks, responses, orgs

ReactFlow-based cause-effect graphs for complex causal models:

Loading diagram...
Node TypePurposeVisual
LeafInput factorsLight blue
IntermediateMediating factorsWhite
EffectOutcomesGreen/Red
GroupClusters of related nodesDashed border
Loading diagram...

Six-dimension rating system (0-10 scale, harsh—7+ is exceptional):

DimensionWhat It MeasuresWeight by Page Type
FocusDoes it answer the title’s promise?All types
NoveltyValue beyond obvious sourcesHigher for analysis
RigorEvidence quality and precisionHigher for reference
CompletenessThorough coverage of claimed topicHigher for reference
ConcretenessSpecific vs. abstract recommendationsHigher for analysis
ActionabilityCan readers make different decisions?Higher for analysis

Quality is set only through automated grading pipeline using Claude, never manually. The quality-source validation rule enforces this.

Loading diagram...

Structured YAML databases enable data-aware components:

DatabaseContentsSizeExample Usage
experts.yamlAI safety researchers with positions on cruxes≈30 entries<DataInfoBox expertId="paul-christiano" />
organizations.yamlLabs, research orgs, funders≈15 entries<DataInfoBox orgId="anthropic" />
cruxes.yamlKey uncertainties with expert positions≈50 entries<DataCrux dataId="deceptive-alignment-likelihood" />
estimates.yamlProbability distributions for key variables≈35 entries<DataEstimateBox dataId="p-doom" />
external-links.yamlCurated external resources with metadata≈200 entries<R id="hash">Link text</R>
LayerTechnologyPurpose
FrameworkAstro 5 + StarlightStatic site generation, documentation theme
ComponentsReact 19Interactive UI components
StylingTailwind CSS 4Utility-first styling
Type SafetyTypeScript + ZodCompile-time and runtime validation
GraphsReactFlow + Dagre/ELKInteractive causal diagrams
DiagramsMermaid 11Declarative flowcharts
MathKaTeXLaTeX rendering
DataYAML → JSONStructured entity databases
UI Componentsshadcn/ui + RadixAccessible component primitives

Unified CLI (crux) provides comprehensive tooling:

Terminal window
npm run crux -- --help # Show all domains
npm run crux -- validate # Run all validators
npm run crux -- analyze # Analysis and reporting
npm run crux -- fix # Auto-fix common issues
npm run crux -- content # Page management
npm run crux -- generate # Content generation
DomainCommandsExample
Validationcompile, frontmatter, entity-links, escaping, mermaidcrux validate compile --quick
Analysismentions, entity-links, quality, redundancycrux analyze entity-links sam-altman
Contentgrade, improve, regrade, createcrux content grade --page scheming
Fixesescaping, entity-links, markdowncrux fix escaping
Generatesummaries, content, diagramscrux generate summaries --batch 50
RuleWhat It ChecksFailure Mode
compileMDX syntax and React componentsBuild failure
frontmatter-schemaYAML frontmatter against Zod schemaCI failure
dollar-signsLaTeX escaping (\$100 not $100)Rendering issues
comparison-operatorsJSX escaping (\<100ms not <100ms)Build failure
entitylink-idsAll EntityLink references exist in databaseBroken links
quality-sourceQuality set by pipeline, not manuallyData integrity
mermaidDiagram syntax validationRendering issues
markdown-listsConsistent list formattingMarkdown errors
Use CaseWiki FeatureExample
Intervention comparisonSide-by-side pages with explicit uncertaintyCompare interpretability vs governance approaches
Crux identificationCrux mapping shows which uncertainties matter mostWhich assumptions drive different funding priorities?
Expert landscapeExpert profiles with positionsWho believes what about deceptive alignment?
Gap analysisQuality scores reveal under-developed areasWhich important topics lack quality coverage?
Use CaseWiki FeatureExample
Literature synthesisConsolidated coverage with citationsFind all sources on a specific risk
Gap identificationCoverage analysis, importance vs qualityWhat important topics need more research?
Position mappingDisagreement visualizationWhere do Yudkowsky and Christiano diverge?
Model buildingCausal diagrams as starting pointsUse wiki models as research scaffolding
Use CaseWiki FeatureExample
Risk taxonomyStructured hierarchy with assessmentsNavigate from high-level categories to specific risks
Response optionsComprehensive intervention catalogWhat governance tools exist?
Expert consensusCrux tracking shows agreement/disagreementWhere do experts converge vs diverge?
Quick briefingsLLM summaries on every pageOne-sentence takeaways for busy readers
ResourceFocusUpdate ModelLongterm Wiki Differentiator
WikipediaGeneral referenceCommunity editsLW is opinionated about importance and uncertainty
LessWrong / AI Alignment ForumDiscussionContinuous postsLW is curated synthesis, not discussion platform
80K Problem ProfilesCareer guidancePeriodic updatesLW goes deeper on cruxes, covers more interventions
Stampy / AISafety.infoFAQ for newcomersCommunity + fellowshipLW is strategic (prioritization), not educational
MIT AI Risk RepositoryAcademic taxonomyResearch updatesLW adds expert positions, crux mapping, intervention analysis
EA Forum WikiIntegrated tagsTag-basedLW is standalone with dedicated editorial

The Similar Projects Analysis identified key success factors:

FactorFailed ProjectsLW Approach
OwnershipArbital (committee), EA Wiki v1 (volunteer)Dedicated editorial lead
FundingMost wikis (volunteer-only)Paid core team
ScopeArbital (everything)AI safety prioritization only
IntegrationStandalone wikis (low traffic)Links to LW/EAF, doesn’t replace
UpdatesSurvey papers (stale in 1-2 years)Quality tracking, staleness monitoring
StrengthEvidenceImpact
Crux-first design50+ tracked uncertainties, expert positionsEnables disagreement decomposition
Quality controlAutomated grading, 6-dimension scoringConsistent content standards
Cross-linking550+ pages with stable entity referencesKnowledge graph navigation
Open sourceMIT license, public GitHubTransparency, reproducibility
Modern stackAstro, React, TypeScriptFast, maintainable, accessible
Interactive visualizationsReactFlow graphs, Mermaid diagramsComplex relationships made legible
LimitationImpactMitigation
Single editorial voiceMay reflect particular worldviewExplicit about opinions, tracks disagreement
Staleness riskContent can become outdatedQuality tracking, freshness monitoring, dated pages
Limited community contributionSlower content growthFocus on quality over volume, open for feedback
AI safety focusNarrow scopeIntentional limitation, links to broader resources
Early stageIncomplete coverageActive development, prioritized expansion
No real-time dataStatic forecastsLinks to Metaforecast for live data
Loading diagram...
ToolRelationship to Longterm Wiki
SquiggleLW documents probabilistic concepts; Squiggle enables quantification
SquiggleAILW models could be converted to executable Squiggle estimates
MetaforecastLW links to relevant forecasts as evidence for claims
Squiggle HubPotential future integration for interactive models embedded in pages

Based on the Vision Document:

PhaseGoalsStatus
FoundationCore knowledge base structure, 30+ high-quality pages, initial crux taxonomyComplete
Depth80+ quality pages, full crux graph, worldview archetypesIn progress
PolishExpert review, interactive worldview→priority tool, public launchPlanned
MaintenanceQuarterly review cycle, community contribution guidelinesPlanned

The wiki is open source and welcomes contributions:

Contribution TypeHow to Help
Content feedbackReport issues or suggest improvements via GitHub Issues
Technical contributionsPRs welcome for components, validation, automation
Expert inputContact for crux positions, disagreement mapping
Bug reportsFile issues for broken links, rendering problems

See Internal Documentation for: