Longterm Wiki

Multi-Agent Safety

Multi-agent safety research addresses coordination failures, conflict, and collusion risks when multiple AI systems interact. A 2025 report from 50+ researchers across DeepMind, Anthropic, and academia identifies seven key risk factors and finds that even individually safe systems may contribute to harm through interaction.

Related

Related Pages

Top Related Pages

Risks

Multipolar Trap (AI Development)AI Development Racing Dynamics

Analysis

AI Safety Research Allocation ModelAI Risk Activation Timeline ModelMultipolar Trap Dynamics ModelAI Safety Culture Equilibrium ModelRacing Dynamics Impact ModelInternational AI Coordination Game Model

Approaches

Sandboxing / ContainmentTool-Use RestrictionsAI Governance Coordination Technologies

Concepts

Agentic AICooperative Funding MechanismsAlignment Deployment OverviewReasoning and Planning

Organizations

AnthropicCoefficient Giving

Policy

Pause / Moratorium

Key Debates

AI Safety Solution Cruxes

Tags

multi-agent-systemscoordinationcollusion-riskgame-theoryagent-safety