Risk Trajectory Experiments
- StructureNo tables or diagrams - consider adding visual content
These experiments visualize AI Transition Model risk trajectories—showing how catastrophe risk and lock-in severity evolve from present day through TAI and beyond.
All Experiments
Section titled “All Experiments”Concept
Section titled “Concept”The core idea: risk compounds over time, with different pathways contributing differently at different phases.
The AI Transition Model identifies:
Catastrophe Pathways:
- AI Takeover (Rapid) - Fast recursive self-improvement leading to misaligned superintelligence
- AI Takeover (Gradual) - Slow erosion of human control
- Human Catastrophe (State Actor) - Great power conflict enabled by AI
- Human Catastrophe (Rogue Actor) - Non-state actors using AI for mass harm
Lock-in Types:
- Economic - Irreversible wealth/power concentration
- Political - Authoritarian control solidified by AI
- Epistemic - Information environment permanently degraded
- Values - Human values shaped/locked by AI systems
- Suffering - Persistent negative states (e.g., digital minds)
Root Factors (driving both outcomes):
- Misalignment Potential
- Misuse Potential
- AI Capabilities
- AI Ownership concentration
- Civilizational Competence
- Transition Turbulence
Individual Components
Section titled “Individual Components”Dual Outcome Chart
Section titled “Dual Outcome Chart”Side-by-side stacked area charts showing catastrophe risk (left) and lock-in severity (right):
Factor Attribution Matrix
Section titled “Factor Attribution Matrix”Shows how each root factor contributes to each outcome type:
Factor Gauges
Section titled “Factor Gauges”Current levels of each root factor with trend indicators:
Trajectory Lines
Section titled “Trajectory Lines”Individual pathway trajectories with confidence bands:
Full Dashboard
Section titled “Full Dashboard”Combined view with all components:
Design Notes
Section titled “Design Notes”Timeline phases:
- Current (2025-2030): Pre-TAI baseline
- Near-TAI (2031-2036): Approaching transformative AI
- TAI (2037-2044): Transformative AI arrival
- Post-TAI (2045+): Stabilization or continued turbulence
Color encoding:
- Warm colors (red/orange) = Catastrophe pathways
- Cool colors (blue/purple/pink) = Lock-in types
- Factor-specific colors for attribution
Key features:
- TAI marker line showing expected transition point
- Hover interactions for detailed values
- Toggle between pathway and factor views
- Confidence bands on trajectory lines
What these visualizations convey:
- Risk accumulates non-linearly, with TAI as an inflection point
- Different pathways dominate at different times
- Lock-in may be the larger long-term concern even if catastrophe is avoided
- Root factors have differential impact on different outcomes