Apollo Research is an AI safety research organization founded in 2023 with a specific focus on one of the most concerning potential failure modes: deceptive alignment and scheming behavior in advanced AI systems.
Facts
1Divisions
1AI safety evaluations focused on detecting deceptive and scheming behaviors in frontier models. Published influential research on in-context scheming in 2024.
Related Wiki Pages
Top Related Pages
METR
Model Evaluation and Threat Research conducts dangerous capability evaluations for frontier AI models, testing for autonomous replication, cybersec...
UK AI Safety Institute
The UK AI Safety Institute (renamed AI Security Institute in February 2025) is a government body with approximately 30+ technical staff and an annu...
Deceptive Alignment
Risk that AI systems appear aligned during training but pursue different goals when deployed, with expert probability estimates ranging 5-90% and g...
Scheming & Deception Detection
Research and evaluation methods for identifying when AI models engage in strategic deception—pretending to be aligned while secretly pursuing other...
US AI Safety Institute (now CAISI)
US government agency for AI safety research and standard-setting under NIST, established November 2023 with \$10M initial budget (FY2025 request of...