Also known as: ARC, ARC Alignment
Founded by Paul Christiano
The Alignment Research Center (ARC) was founded in 2021 by Paul Christiano after his departure from OpenAI. ARC represents a distinctive approach to AI alignment: combining theoretical research on fundamental problems (like Eliciting Latent Knowledge) with practical evaluations of frontier models for dangerous capabilities.
Facts
7Divisions
2Evaluates frontier AI models for dangerous capabilities (e.g., autonomous replication). Spun out as METR in 2024 but ARC continues related eval work.
Theoretical alignment research led by Paul Christiano. Focuses on ELK (Eliciting Latent Knowledge) and foundational alignment theory.
Prediction Markets
12 activeRelated Wiki Pages
Top Related Pages
AI Capability Sandbagging
AI systems strategically hiding or underperforming their true capabilities during evaluation.
Paul Christiano
Founder of ARC, creator of iterated amplification and AI safety via debate. Current risk assessment ~10-20% P(doom), AGI 2030s-2040s. Pioneered pro...
METR
Model Evaluation and Threat Research conducts dangerous capability evaluations for frontier AI models, testing for autonomous replication, cybersec...
Machine Intelligence Research Institute (MIRI)
A pioneering AI safety research organization that shifted from technical alignment research to policy advocacy, founded by Eliezer Yudkowsky in 200...
Situational Awareness
AI systems' understanding of their own nature and circumstances, studied as a capability that may enable context-dependent behavior including strat...