Page Type:ContentStyle Guide →Standard knowledge base article
Quality:39 (Draft)⚠️
Importance:25 (Peripheral)
Last edited:2026-01-02 (4 weeks ago)
Words:1.1k
Backlinks:6
Structure:
📊 12📈 0🔗 38📚 0•10%Score: 10/15
LLM Summary:Comprehensive biography of Paul Christiano documenting his technical contributions (IDA, debate, scalable oversight), risk assessment (~10-20% P(doom), AGI 2030s-2040s), and evolution from higher optimism to current moderate concern. Documents implementation of his ideas at major labs (RLHF at OpenAI, Constitutional AI at Anthropic) with specific citation to papers and organizational impact.
Issues (1):
QualityRated 39 but structure suggests 67 (underrated by 28 points)
Paul Christiano is one of the most influential researchers in AI alignment, known for developing concrete, empirically testable approaches to the alignment problem. With a PhD in theoretical computer science from UC Berkeley, he has worked at OpenAI, DeepMind, and founded the Alignment Research Center (ARC)OrganizationARCComprehensive overview of ARC's dual structure (theory research on Eliciting Latent Knowledge problem and systematic dangerous capability evaluations of frontier AI models), documenting their high ...Quality: 43/100.
Christiano pioneered the “prosaic alignment” approach—aligning AI without requiring exotic theoretical breakthroughs. His current risk assessment places ~10-20% probability on existential risk from AI this century, with AGI arrival in the 2030s-2040s. His work has directly influenced alignment research programs at major labs including OpenAILabOpenAIComprehensive organizational profile of OpenAI documenting evolution from 2015 non-profit to commercial AGI developer, with detailed analysis of governance crisis, safety researcher exodus (75% of ...Quality: 46/100, AnthropicLabAnthropicComprehensive profile of Anthropic tracking its rapid commercial growth (from $1B to $7B annualized revenue in 2025, 42% enterprise coding market share) alongside safety research (Constitutional AI...Quality: 51/100, and DeepMindLabGoogle DeepMindComprehensive overview of DeepMind's history, achievements (AlphaGo, AlphaFold with 200M+ protein structures), and 2023 merger with Google Brain. Documents racing dynamics with OpenAI and new Front...Quality: 37/100.
More optimistic than MIRIOrganizationMIRIComprehensive organizational history documenting MIRI's trajectory from pioneering AI safety research (2000-2020) to policy advocacy after acknowledging research failure, with detailed financial da...Quality: 50/100
Human overseer works with AI assistant on complex tasks
Tested at scale by OpenAI↗🔗 web★★★★☆OpenAIOpenAIiterated-amplificationscalable-oversightai-safety-via-debateSource ↗Notes
Distillation
Extract human+AI behavior into standalone AI system
Standard ML technique
Iteration
Repeat process with increasingly capable systems
Theoretical framework
Bootstrapping
Build aligned AGI from aligned weak systems
Core theoretical hope
Key insight: If we can align a weak system and use it to help align slightly stronger systems, we can bootstrap to aligned AGI without solving the full problem directly.
Co-developed with Geoffrey Irving↗🔗 web★★★★☆Google ScholarGeoffrey Irvingiterated-amplificationscalable-oversightai-safety-via-debateSource ↗Notes at DeepMind in “AI safety via debate”↗📄 paper★★★☆☆arXivDebate as Scalable OversightGeoffrey Irving, Paul Christiano, Dario Amodei (2018)alignmentsafetytrainingcompute+1Source ↗Notes (2018):
Mechanism
Implementation
Results
Adversarial Training
Two AIs argue for different positions
Deployed at Anthropic↗🔗 web★★★★☆AnthropicAnthropic'srisk-factorcompetitiongame-theoryiterated-amplification+1Source ↗Notes
Human Judgment
Human evaluates which argument is more convincing
Scales human oversight capability
Truth Discovery
Debate incentivizes finding flaws in opponent arguments
Need fundamental breakthroughs (MIRIOrganizationMIRIComprehensive organizational history documenting MIRI's trajectory from pioneering AI safety research (2000-2020) to policy advocacy after acknowledging research failure, with detailed financial da...Quality: 50/100)
At ARCOrganizationARCComprehensive overview of ARC's dual structure (theory research on Eliciting Latent Knowledge problem and systematic dangerous capability evaluations of frontier AI models), documenting their high ...Quality: 43/100, Christiano’s priorities include:
Research Area
Specific Focus
Timeline
Power-seeking evaluation
Understanding how AI systems could gain influence gradually
Ongoing
Scalable oversight
Better techniques for supervising superhuman systems
Christiano identifies several critical uncertainties:
Uncertainty
Why It Matters
Current Evidence
Deceptive alignment prevalence
Determines safety of iterative approach
Mixed signals from current systemsRiskDeceptive AlignmentComprehensive analysis of deceptive alignment risk where AI systems appear aligned during training but pursue different goals when deployed. Expert probability estimates range 5-90%, with key empir...Quality: 75/100
Capability jump sizes
Affects whether we get warning
Continuous but accelerating progress
Coordination feasibility
Determines governance strategies
Some positive signsPolicyVoluntary AI Safety CommitmentsComprehensive empirical analysis of voluntary AI safety commitments showing 53% mean compliance rate across 30 indicators (ranging from 13% for Apple to 83% for OpenAI), with strongest adoption in ...Quality: 91/100
Alignment tax magnitude
Economic feasibility of safety
Early evidence suggests low taxAi Transition Model MetricAlignment ProgressComprehensive empirical tracking of AI alignment progress across 10 dimensions finds highly uneven progress: dramatic improvements in jailbreak resistance (87%→3% ASR for frontier models) but conce...Quality: 66/100
Continued capability advances in language modelsCapabilityLarge Language ModelsComprehensive analysis of LLM capabilities showing rapid progress from GPT-2 (1.5B parameters, 2019) to o3 (87.5% on ARC-AGI vs ~85% human baseline, 2024), with training costs growing 2.4x annually...Quality: 60/100
Early agentic AICapabilityAgentic AIComprehensive analysis of agentic AI capabilities and risks, documenting rapid adoption (40% of enterprise apps by 2026) alongside high failure rates (40%+ project cancellations by 2027). Synthesiz...Quality: 63/100 systems
Eliezer YudkowskyResearcherEliezer YudkowskyComprehensive biographical profile of Eliezer Yudkowsky covering his foundational contributions to AI safety (CEV, early problem formulation, agent foundations) and notably pessimistic views (>90% ...Quality: 35/100
≈90%
2020s
Fundamental theory
Pessimistic
Dario AmodeiResearcherDario AmodeiComprehensive biographical profile of Anthropic CEO Dario Amodei documenting his 'race to the top' philosophy, 10-25% catastrophic risk estimate, 2026-2030 AGI timeline, and Constitutional AI appro...Quality: 41/100
≈10-25%
2030s
Constitutional AI
Industry-focused
Stuart RussellResearcherStuart RussellStuart Russell is a UC Berkeley professor who founded CHAI in 2016 with $5.6M from Coefficient Giving (then Open Philanthropy) and authored 'Human Compatible' (2019), which proposes cooperative inv...Quality: 30/100
AI safety via debate↗📄 paper★★★☆☆arXivDebate as Scalable OversightGeoffrey Irving, Paul Christiano, Dario Amodei (2018)alignmentsafetytrainingcompute+1Source ↗Notes
Scalable oversightSafety AgendaInterpretabilityMechanistic interpretability has extracted 34M+ interpretable features from Claude 3 Sonnet with 90% automated labeling accuracy and demonstrated 75-85% success in causal validation, though less th...Quality: 66/100
Core research focus
Reward modelingCapabilityRLHFRLHF/Constitutional AI achieves 82-85% preference improvements and 40.8% adversarial attack reduction for current systems, but faces fundamental scalability limits: weak-to-strong supervision shows...Quality: 63/100
Foundation for many proposals
AI governance
Increasing focus area
Alignment evaluationEvaluationComprehensive overview of AI evaluation methods spanning dangerous capability assessment, safety properties, and deception detection, with categorized frameworks from industry (Anthropic Constituti...Quality: 72/100