Responsible Scaling Policies
Responsible Scaling Policies (RSPs) are voluntary commitments by AI labs to pause scaling when capability or safety thresholds are crossed. As of December 2025, 20 companies have published policies, though SaferAI grades the three major frameworks 1.9-2.2/5 for specificity.
Related
Related Wiki Pages
Top Related Pages
METR
Model Evaluation and Threat Research conducts dangerous capability evaluations for frontier AI models, testing for autonomous replication, cybersec...
AI Safety Intervention Portfolio
Strategic overview of AI safety interventions analyzing ~\\$650M annual investment across 1,100 FTEs. Maps 13+ interventions against 4 risk categor...
Pause Advocacy
Advocacy for slowing or halting frontier AI development until adequate safety measures are in place. Analysis suggests 15-40% probability of meanin...
Corporate AI Safety Responses
How major AI companies are responding to safety concerns through internal policies, responsible scaling frameworks, safety teams, and disclosure pr...
Anthropic
An AI safety company founded by former OpenAI researchers that develops frontier AI models while pursuing safety research, including the Claude mod...