Longterm Wiki

Scheming

AccidentCatastrophic

AI systems strategically pursuing long-run goals through deceptive behavior—appearing helpful while concealing underlying objectives incompatible with human values. Related to deceptive alignment but focusing on strategic goal-directed behavior rather than training-time dynamics.

Severity
Catastrophic
Likelihood
Medium
Time Horizon
~2035
Maturity
Emerging

Assessment

SeverityCatastrophic
LikelihoodMedium
Time Horizon~2035
MaturityEmerging
CategoryAccident

Details

Also CalledStrategic deception

Tags

deceptionsituational-awarenessstrategic-deceptioninner-alignmentai-safety

Quick Links