Longterm Wiki

Third-Party Model Auditing

External organizations independently assess AI models for safety and dangerous capabilities. METR, Apollo Research, and government AI Safety Institutes now conduct pre-deployment evaluations of all major frontier models, with the field evolving from voluntary arrangements to EU AI Act mandatory requirements.

Related

Related Pages

Top Related Pages

Risks

Multipolar Trap (AI Development)

Analysis

OpenAI Foundation Governance ParadoxLong-Term Benefit Trust (Anthropic)

Approaches

Capability ElicitationDangerous Capability EvaluationsAI EvaluationAlignment EvaluationsAI Governance Coordination Technologies

Organizations

AnthropicUK AI Safety Institute

Policy

AI Safety Institutes (AISIs)Evals-Based Deployment GatesUS Executive Order on Safe, Secure, and Trustworthy AI

Concepts

Alignment Evaluation OverviewGovernance-Focused Worldview

Other

Beth BarnesAI EvaluationsYoshua Bengio

Key Debates

Open vs Closed Source AIGovernment Regulation vs Industry Self-Governance

Tags

third-party-auditingindependent-evaluationgovernancedeployment-oversightregulatory-compliance