Longterm Wiki
Updated 2026-03-13HistoryData
Page StatusContent
Edited today952 words
48QualityAdequate52.3ImportanceUseful
Summary

A well-organized reference overview of ~20 AI safety organizations categorized by function (alignment research, policy, field-building), with a comparative budget/headcount table showing estimated annual budgets of \$3-10M and cost-per-researcher of \$143K-\$400K across nine major orgs, all primarily funded by Coefficient Giving (formerly Open Philanthropy). The page is a competent compilation with useful quantitative estimates but offers little original analysis beyond organizing publicly available information.

Content4/13
LLM summaryScheduleEntityEdit history3Overview
Tables1/ ~4Diagrams0Int. links67/ ~8Ext. links0/ ~5Footnotes0/ ~3References0/ ~3Quotes0Accuracy0RatingsN:3.5 R:4.5 A:4.5 C:6
Change History3
Add concrete shareable data tables to high-value pages3 weeks ago

Added three concrete, screenshot-worthy data tables to high-value wiki pages: (1) OpenAI ownership/stakeholder table to openai.mdx showing the 2024-2025 PBC restructuring with Foundation ~26%, Microsoft transitioning from 49% profit share to ~2.5% equity, and Sam Altman's proposed 7% grant; (2) Budget and headcount comparison table to safety-orgs-overview.mdx covering MIRI, ARC, METR, Redwood Research, CAIS, Apollo Research, GovAI, Conjecture, and FAR AI with annual budgets, headcounts, and cost-per-researcher; (3) Per-company compensation comparison table to ai-talent-market-dynamics.mdx comparing Anthropic, OpenAI, Google DeepMind, xAI, Meta AI, and Microsoft Research by total comp range, base salary, equity type, and benefits including Anthropic's unique DAF matching program.

sonnet-4 · ~45min

Clarify overview pages with new entity type3 weeks ago

Added `overview` as a proper entity type throughout the system, migrated all 36 overview pages to `entityType: overview`, built overview-specific InfoBox rendering with child page links, created an OverviewBanner component, and added a knowledge-base-overview page template to Crux.

Fix conflicting numeric IDs + add integrity checks#1684 weeks ago

Fixed all 9 overview pages from PR #118 which had numeric IDs (E687-E695) that conflicted with existing YAML entities. Reassigned to E710-E718. Then hardened the system to prevent recurrence: 1. Added page-level numericId conflict detection to `build-data.mjs` (build now fails on conflicts) 2. Created `numeric-id-integrity` global validation rule (cross-page uniqueness, format validation, entity conflict detection) 3. Added `numericId` and `subcategory` to frontmatter Zod schema with format regex

AI Safety Organizations (Overview)

Overview

The AI safety organizational landscape spans dedicated alignment research labs, policy think tanks, advocacy groups, and field-building institutions. These organizations aim to reduce catastrophic and existential risks from advanced AI systems through technical research, governance advocacy, talent development, and public engagement.

Funding is heavily concentrated through a small number of major funders, most prominently Coefficient Giving, which has provided grants to the majority of organizations listed on this page. This concentration produces a relatively coordinated funding environment, with most grantees sharing compatible research agendas and norms, while also reducing diversification of funding sources across the field.

Note: The AI safety organizational landscape evolves rapidly. Headcount, budget, and focus area descriptions reflect available information as of mid-2025 and may not capture recent changes. Check individual entity pages for the most current details.

Alignment Research Labs

Dedicated organizations conducting technical AI safety research:

  • ARC (Alignment Research Center): Founded by Paul Christiano; focuses on alignment evaluation and theoretical alignment research
  • METR: Evaluates dangerous capabilities in frontier AI models; spun out of ARC
  • Apollo Research: Focuses on detecting and understanding deceptive AI behavior, including scheming evaluations
  • Redwood Research: Alignment research lab working on interpretability, adversarial training, and AI control
  • Conjecture: Alignment research and product company based in London
  • FAR AI: Researches robustness, adversarial attacks, and alignment failures in AI systems
  • Palisade Research: Focuses on practical AI safety evaluation and red-teaming
  • Seldon Lab: Works on alignment approaches and safety evaluations
  • Goodfire: Interpretability-focused startup building tools for understanding neural networks
  • MIRI (Machine Intelligence Research Institute): Pioneer in AI alignment theory; founded 2000

Policy and Governance Organizations

Think tanks and research centers focused on AI governance and policy:

  • GovAI: Research center focused on AI governance based at Oxford
  • CSET (Center for Security and Emerging Technology): Georgetown think tank producing policy-relevant research on AI and emerging technologies
  • CSER (Centre for the Study of Existential Risk): Cambridge-based research center studying existential risks including from AI
  • Secure AI Project: Advocacy organization focused on AI safety policy
  • ControlAI: Advocacy organization pushing for stronger AI regulation and safety standards
  • Pause AI: Grassroots advocacy movement calling for a pause on frontier AI development
  • Frontier Model Forum: Industry-led consortium for frontier AI safety, founded by Anthropic, Google DeepMind, Microsoft, and OpenAI. The forum's stated mission centers on safety research and best-practice sharing; observers differ on the extent to which it functions as a coordination body versus an industry advocacy vehicle

Field-Building and Talent Development

Organizations supporting the growth of the AI safety field:

  • 80,000 Hours: Career advisory organization directing talent toward high-impact careers including AI safety
  • MATS (ML Alignment Theory Scholars): Training program connecting aspiring alignment researchers with mentors
  • Lightning Rod Labs: Works on AI safety infrastructure and tooling
  • AI Futures Project: Research and analysis on AI development trajectories and safety considerations

Research and Analysis

Organizations focused on understanding AI progress and risks:

  • Epoch AI: Tracks AI compute trends, model capabilities, and training data
  • CAIS (Center for AI Safety): Conducts safety research and field-building for AI safety; hosts a compute cluster for safety research
  • CHAI (Center for Human-Compatible AI): UC Berkeley research center founded by Stuart Russell focusing on human-compatible AI

Budget and Headcount Comparison

For funders and researchers evaluating organizational capacity and capital efficiency, comparative budget and headcount data can help identify where additional resources may be most impactful and how different organizations structure their research operations. The table below aggregates publicly available estimates across nine prominent independent AI safety organizations.

All figures are estimates derived from IRS Form 990 filings (via ProPublica Nonprofit Explorer), Coefficient Giving (formerly Open Philanthropy) grant disclosures, LinkedIn headcount data, and news reports. Figures are approximate, may lag actual values by one to two years, and should be treated as indicative rather than authoritative. The "Est. Budget per Staff Member/year" column is calculated using the midpoint of the headcount range and counts all staff, not researchers only.

OrganizationAnnual Budget (Est.)Headcount (Est.)Est. Budget per Staff Member/year (Est.)Primary FunderFocus Area
MIRI≈$5M10–15≈$400KCoefficient GivingAlignment theory
ARC≈$8M20–30≈$320KCoefficient GivingAlignment research & evaluation
METR≈$5M20–30≈$200KCoefficient GivingDangerous capability evaluation
CAIS≈$5M15–20≈$286KCoefficient GivingResearch & field-building
Redwood Research≈$10M30–40≈$286KCoefficient GivingInterpretability & AI control
Apollo Research≈$4M15–20≈$229KCoefficient GivingDeceptive alignment & scheming
Conjecture≈$5M30–40≈$143KMixed (VC + grants)Alignment research & products
FAR AI≈$3M10–15≈$240KCoefficient GivingRobustness & adversarial ML
GovAI≈$5M20–30≈$200KCoefficient GivingAI governance & policy

The budget-per-staff figures reflect meaningful variation in organizational structure. Organizations with lower ratios (e.g., Conjecture) typically employ a higher proportion of non-researcher staff or operate hybrid research-product models, whereas those with higher ratios (e.g., MIRI) tend toward smaller, senior-heavy research teams. These figures should not be interpreted as proxies for research quality or output volume.

Key Patterns

Specialization trend: The field has moved from generalist safety organizations—such as MIRI and the Future of Humanity Institute (FHI, which closed in 2024)—toward more specialized roles: dedicated evaluation labs (METR, Apollo Research), interpretability startups (Goodfire), policy research centers (ControlAI, GovAI), and talent pipelines (MATS, 80,000 Hours).

Industry-adjacent positioning: Organizations in this landscape occupy a range of positions relative to frontier AI developers. Some—such as the Frontier Model Forum, Redwood Research, and Apollo Research—maintain active collaborative relationships with frontier labs. Others, including Pause AI and ControlAI, advocate for regulatory constraints on AI development and position themselves independently of industry partnerships. Proponents of each approach offer different accounts of how safety outcomes are best achieved.

Funding concentration: As illustrated in the budget table above, most organizations in this cluster report Coefficient Giving as their primary funder. This pattern is visible across alignment research, governance research, and field-building organizations alike.

Related Pages

Top Related Pages

Organizations

AnthropicGoodfireMETRPalisade ResearchOpenAIPause AI

Safety Research

Interpretability

Approaches

Adversarial Training

Other

Paul ChristianoStuart Russell

Analysis

AI Safety Research Allocation Model