Lock-in Probability Model
- ClaimValue lock-in has the shortest reversibility window (3-7 years during development phase) despite being one of the most likely scenarios, creating urgent prioritization needs for AI development governance.S:3.5I:4.5A:4.5
- Quant.Expert assessments estimate a 10-30% cumulative probability of significant AI-enabled lock-in by 2050, with value lock-in via AI training (10-20%) and economic power concentration (15-25%) being the most likely scenarios.S:4.0I:4.5A:3.5
- Quant.The IMD AI Safety Clock advanced 9 minutes in one year (from 29 to 20 minutes to midnight by September 2025), indicating rapidly compressing decision timelines for preventing lock-in scenarios.S:3.5I:4.0A:4.0
- TODOComplete 'Quantitative Analysis' section (8 placeholders)
- TODOComplete 'Strategic Importance' section
- TODOComplete 'Limitations' section (6 placeholders)
Overview
Section titled βOverviewβThis model provides a quantitative framework for assessing AI-enabled lock-in risk, complementing the comprehensive mechanism analysis in Lock-in RiskRiskLock-inComprehensive analysis of AI lock-in scenarios where values, systems, or power structures become permanently entrenched. Documents evidence including Big Tech's 66-70% cloud control, AI surveillanc...Quality: 64/100.
For detailed coverage of mechanisms, current trends, and responses, see Lock-in RiskRiskLock-inComprehensive analysis of AI lock-in scenarios where values, systems, or power structures become permanently entrenched. Documents evidence including Big Tech's 66-70% cloud control, AI surveillanc...Quality: 64/100.
Probability Estimates by Scenario
Section titled βProbability Estimates by ScenarioβThe following probability estimates draw on expert assessments from Future of Life Instituteβπ webβ β β ββTIMEFuture of Life InstituteSource βNotes, EA Forum analysesβπ webβ β β ββEA ForumEA Forum80000_Hours, poppinfresh (2024)Source βNotes, and Future of Humanity Instituteβπ webTheoretical workSource βNotes research:
| Scenario | Probability by 2050 | Duration if Realized | Key Drivers | Reversibility Window |
|---|---|---|---|---|
| Totalitarian surveillance state | 5-15% | Potentially indefinite | AI-enhanced monitoring, predictive policing, autonomous enforcement | 5-10 years before fully entrenched |
| Value lock-in via AI training | 10-20% | Centuries to millennia | Constitutional AI approaches, training data choices, RLHF value embeddingβπ paperβ β β β βAnthropicConstitutional AI: Harmlessness from AI FeedbackAnthropic introduces a novel approach to AI training called Constitutional AI, which uses self-critique and AI feedback to develop safer, more principled AI systems without exte...Source βNotes | 3-7 years during development phase |
| Economic power concentration | 15-25% | Decades to centuries | Network effects, compute monopolyβπ webFive tech companies control over 80%Source βNotes, data advantages | 10-20 years with antitrust action |
| Geopolitical lock-in | 10-20% | Decades to centuries | First-mover AI advantages, regulatory capture | Uncertain, depends on coordination |
| Aligned singleton (positive) | 5-10% | Indefinite | Successful alignment, beneficial governance | N/A (desirable outcome) |
| Misaligned AI takeover | 2-10% | Permanent | Deceptive alignment, capability overhang | Days to weeks at critical juncture |
Note: These ranges reflect significant uncertainty. The stable totalitarianism analysisβπ webβ β β ββEA ForumEA Forum80000_Hours, poppinfresh (2024)Source βNotes suggests extreme scenarios may be below 1%, while other researchers place combined lock-in risk at 10-30%.
Risk Factor Framework
Section titled βRisk Factor Frameworkβ| Risk Factor | Mechanism | AI Amplification | Reversibility |
|---|---|---|---|
| Enforcement capability | Autonomous systems maintain control | 10-100x more comprehensive surveillance; no human defection risk | Very Low |
| Path dependence | Early choices constrain future options | Faster deployment cycles compress decision windows | Low |
| Network effects | Systems become more valuable as adoption grows | AI models compound advantages via data and compute | Low-Medium |
| Value embedding | Preferences encoded during development persist | Constitutional AI approaches embed values during training | Medium |
| Complexity barriers | System understanding requires specialized expertise | AI systems may become inscrutable even to developers | Very Low |
Timeline Indicators
Section titled βTimeline IndicatorsβThe IMD AI Safety Clockβπ webIMD AI Safety ClockSource βNotes tracks lock-in urgency:
| Date | Clock Position | Key Developments |
|---|---|---|
| September 2024 | 29 minutes to midnight | Clock launched |
| December 2024 | 26 minutes | AGI timeline acceleration |
| February 2025 | 24 minutes | California SB 1047 vetoed |
| September 2025 | 20 minutes | Agentic AI proliferation |
The nine-minute advance in one year reflects compressed decision timelines.
Model Limitations
Section titled βModel LimitationsβThis framework cannot capture:
- Novel lock-in pathways not yet identified
- Interaction effects between scenarios
- Tail risks from capability discontinuities
- Political feasibility of interventions
For intervention analysis, see Lock-in Risk: ResponsesRiskLock-inComprehensive analysis of AI lock-in scenarios where values, systems, or power structures become permanently entrenched. Documents evidence including Big Tech's 66-70% cloud control, AI surveillanc...Quality: 64/100.
Sources
Section titled βSourcesβ- Stable Totalitarianism: An Overviewβπ webβ β β ββEA ForumEA Forum80000_Hours, poppinfresh (2024)Source βNotes - EA Forum analysis
- IMD AI Safety Clockβπ webIMD AI Safety ClockSource βNotes - Real-time risk tracking
- Bostrom on permanent lock-in scenariosβπ webNick Bostrom has arguedSource βNotes
- Finnveden, Riedel, and Shulman on AI-enabled dictatorshipβπ webβ β β ββ80,000 HoursCarl Shulman and colleaguesSource βNotes