Evals-Based Deployment Gates
in-effectInternationalEvals-based deployment gates require AI models to pass safety evaluations before deployment or capability scaling. The EU AI Act mandates conformity assessments for high-risk systems with fines up to EUR 35M or 7% global turnover, while UK AISI has evaluated 30+ frontier models.
Related Legislation
| Name | Status |
|---|---|
| EU AI Act | in-effect |
Related Topics
Related Pages
Top Related Pages
METR
Model Evaluation and Threat Research conducts dangerous capability evaluations for frontier AI models, testing for autonomous replication, cybersec...
Anthropic
An AI safety company founded by former OpenAI researchers that develops frontier AI models while pursuing safety research, including the Claude mod...
EU AI Act
The world's first comprehensive AI regulation, adopting a risk-based approach to regulate foundation models and general-purpose AI systems
Responsible Scaling Policies
Industry self-regulation frameworks establishing capability thresholds that trigger safety evaluations. Anthropic's ASL-3 requires 30%+ bioweapon d...
AI Evaluation
Methods and frameworks for evaluating AI system safety, capabilities, and alignment properties before deployment, including dangerous capability de...
Safety Research
Historical
Key Debates
Other
Quick Facts
- Introduced
- 2023
- Status
- Active; EU binding, UK/Lab voluntary
- Scope
- International