METR (Model Evaluation and Threat Research), formerly known as ARC Evals, is an organization dedicated to evaluating frontier AI models for dangerous capabilities before deployment.
Facts
2Related Wiki Pages
Top Related Pages
Ajeya Cotra
Member of technical staff at METR and former senior advisor at Coefficient Giving (formerly Open Philanthropy).
UK AI Safety Institute
The UK AI Safety Institute (renamed AI Security Institute in February 2025) is a government body with approximately 30+ technical staff and an annu...
US AI Safety Institute (now CAISI)
US government agency for AI safety research and standard-setting under NIST, established November 2023 with \$10M initial budget (FY2025 request of...
Apollo Research
AI safety organization conducting rigorous empirical evaluations of deception, scheming, and sandbagging in frontier AI models, providing concrete ...
Capability Elicitation
Systematic methods to discover what AI models can actually do, including hidden capabilities that may not appear in standard benchmarks, through sc...