Skip to content

Epoch AI

📋Page Status
Page Type:ContentStyle Guide →Standard knowledge base article
Quality:91 (Comprehensive)
Importance:62.5 (Useful)
Last edited:2026-01-30 (2 days ago)
Words:4.6k
Backlinks:2
Structure:
📊 30📈 1🔗 6📚 6611%Score: 14/15
LLM Summary:Epoch AI provides empirical AI progress tracking showing training compute growing 4.4x annually (2010-2024), 300 trillion tokens of high-quality training data with exhaustion projected 2026-2032, and algorithmic efficiency doubling every 6-12 months. Their 3,200+ model database directly informs US Executive Order 10^26 FLOPs threshold and export controls. With $4.1M in 2025 Open Philanthropy funding and 34 staff, they've produced FrontierMath (testing advanced reasoning) and the Epoch Capabilities Index, serving as critical infrastructure for compute governance approaches.
Critical Insights (4):
  • ClaimHigh-quality training text data (~10^13 tokens) may be exhausted by the mid-2020s, creating a fundamental bottleneck that could force AI development toward synthetic data generation and multimodal approaches.S:4.5I:4.0A:4.5
  • Quant.Training compute for frontier AI models is doubling every 6 months (compared to Moore's Law's 2-year doubling), creating a 10,000x increase from 2012-2022 and driving training costs to $100M+ with projections of billions by 2030.S:4.0I:4.5A:4.0
  • Counterint.Algorithmic efficiency in AI is improving by 2x every 6-12 months, which could undermine compute governance strategies by reducing the effectiveness of hardware-based controls.S:4.0I:4.0A:3.5
Issues (1):
  • Links15 links could use <R> components
See also:EA Forum
Organization

Epoch AI

Importance62
DimensionAssessmentEvidence
Research OutputVery High36 Data Insights, 37 Gradient Updates, 3,200+ models tracked in 2025
Policy InfluenceHighCited in US Executive Order 14110; 10^26 FLOP threshold based on Epoch data
Data InfrastructureExceptionalLargest public ML model database; 3,200+ models from 1950-present
Funding StabilityStrong$1.1M Open Philanthropy grant (2025), additional $1.9M general support
Team ExpertiseStrong34 staff; director holds PhD in AI from University of Aberdeen
Methodological RigorHighMultiple forecasting methods with explicit uncertainty bounds
IndependenceMedium-HighNonprofit with EA/Open Philanthropy funding; no industry capture
Benchmark InnovationHighFrontierMath benchmark with 60+ mathematicians including Fields Medalist

Epoch AI is a research organization founded in 2022 that provides rigorous, data-driven empirical analysis and forecasting of AI progress. Based in San Francisco with 34 employees, their work serves as critical infrastructure for AI governance and timeline forecasting. Their public database tracks over 3,200 machine learning models from 1950 to present—the largest resource of its kind.

Their research documents three key scaling dynamics: training compute growing at 4.4x per year since 2010, high-quality training data stock of approximately 300 trillion tokens with exhaustion projected between 2026-2032, and algorithmic efficiency doubling every 6-12 months. As of June 2025, they have identified over 30 publicly announced AI models exceeding the 10^25 FLOP training compute threshold.

Unlike organizations developing AI capabilities or safety techniques directly, Epoch provides the empirical foundation that informs strategic decisions across the AI ecosystem. Their databases and forecasts are cited by policymakers designing compute governance frameworks, safety researchers planning research timelines, and AI labs benchmarking their progress against industry trends. In 2024, they were featured on The New York Times’ “2024 Good Tech Awards” list.

Loading diagram...

Their most influential finding is the exponential growth in training compute for frontier models—with training costs growing 2-3x annually and projected to exceed $1 billion per model by 2027. This analysis has become foundational for understanding AI progress and informing governance approaches focused on compute as a key chokepoint.

Output Category2025 MetricsKey ProductsPolicy Relevance
Data Insights36 publishedTraining compute decomposition, API cost trendsInforms compute governance thresholds
Gradient Updates37 publishedWeekly trend analysesReal-time capability tracking
Model Database3,200+ entriesLargest public ML model databaseFoundation for policy research
Benchmarks2 major releasesFrontierMath, Epoch Capabilities IndexStandardized capability measurement
AI Chip TrackingNew data explorer15M+ H100-equivalents tracked globallyExport control effectiveness analysis
Cost AnalysisMultiple reports10x API price drop (Apr 2023-Mar 2025)Market concentration analysis
MetricCurrent ValueTrendSource
Training compute growth4.4x/year (2010-2024)Accelerating to 5x/year post-2020Epoch Trends
Frontier model training costsTens of millions USDProjected to exceed $1B by 2027Cost Analysis
Total AI compute capacity15M+ H100-equivalentsGrowing rapidlyEpoch AI Chip Sales
Models above 10^25 FLOP30+ (as of June 2025)≈2 announced monthly in 2024Model Database
High-quality text stock≈300 trillion tokensExhaustion projected 2026-2032Epoch Data Analysis
API cost reduction10x (Apr 2023-Mar 2025)Continuing decline2025 Impact Report
Risk CategoryAssessmentEvidenceTimelineTrajectory
Data BottleneckHighHigh-quality text ≈300T tokens, current usage accelerating2026-2032Worsening
Compute ScalingMedium4.4x annual growth potentially unsustainable; energy constraints emerging2030sStable
Governance LagHighPolicy development slower than tech progress; 1,000+ US state AI bills in 2025OngoingImproving
Forecasting AccuracyMediumWide uncertainty bounds; Epoch Capabilities Index improving precisionContinuousImproving

Epoch’s flagship research tracks computational resources used to train AI models, revealing exponential scaling patterns. Their Machine Learning Trends dashboard provides real-time tracking of these dynamics.

MetricCurrent TrendKey FindingPolicy Implication
Training Compute4.4x/year (2010-2024)5x/year for frontier models since 2020Compute governance thresholds need updating
Training Costs$50-100M for frontier modelsProjected to exceed $1B by 2027Market concentration accelerating
Hardware Capacity15M+ H100-equivalents globallyCompanies own 80% of AI supercomputersExport controls increasingly important
Total AI ComputeDoubling every 7 months3.3x annual growth since 2022Governance frameworks struggling to keep pace

Critical findings from Epoch’s compute database:

  • Exponential growth faster than Moore’s Law: While chip performance doubles every ~2 years, AI training compute grows 4.4x annually—driven primarily by larger training clusters, longer training runs, and improved hardware
  • Economic scaling: Training costs growing 2-3x annually, reaching $50-100M+ for frontier models; the largest AI supercomputers (like Anthropic’s 750MW Indiana facility) cost billions to build
  • Concentration effects: Only 12 organizations have trained models above 10^25 FLOP; companies now own 80% of all AI supercomputers while government share has declined

Epoch’s “Will We Run Out of Data?” research revealed potential bottlenecks for continued AI scaling. Their analysis estimates the total stock of human-generated public text at approximately 300 trillion tokens.

Data TypeEstimated StockCurrent Usage RateExhaustion Timeline
High-quality text≈300 trillion tokensAccelerating2026-2032
All web text≈10^15 tokensIncreasingEarly 2030s
Image dataLarger but finiteGrowing rapidly2030s+
Video dataMassive but hard to useEarly stagesUnknown

Key implications:

  • Pressure for efficiency: Data constraints may force more efficient training methods; Epoch research shows algorithmic efficiency doubling every 6-12 months
  • Synthetic data boom: Investment in AI-generated training data accelerating to extend runway
  • Multimodal shift: Language models increasingly incorporating image/video data
  • Overtraining risk: If models are “intensely overtrained,” high-quality text could be exhausted even earlier than 2026

Epoch employs multiple complementary approaches to estimate transformative AI timelines:

MethodCurrent Estimate RangeKey VariablesConfidence Level
Trend Extrapolation2030s-2040sCompute, data, algorithmsMedium
Biological Anchors2040s-2050sBrain computation estimatesLow
Benchmark Analysis2030s-2050sTask performance ratesMedium
Economic Modeling2035-2060sInvestment trends, ROILow

Epoch’s data directly informs major governance initiatives. Their research has been particularly influential in establishing compute thresholds for regulatory frameworks.

Policy AreaEpoch ContributionReal-World Impact
US AI Executive Order 1411010^26 FLOPs threshold analysisTraining run reporting requirements for frontier models
Export controlsH100/A100 performance data, chip sales trackingChip restriction implementation and effectiveness monitoring
UK AI Safety InstituteCapability benchmarking, FrontierMathModel evaluation frameworks
EU AI ActCompute-based GPAI thresholdsClassification of general-purpose AI systems
Compute governance researchDatabase infrastructure, threshold analysisAcademic and policy research foundation

Epoch’s FrontierMath benchmark represents a significant contribution to AI evaluation infrastructure:

AspectDetailsSignificance
Problem Count350 original problems (300 in Levels 1-3, 50 in Level 4)Covers major branches of modern mathematics
Expert Collaboration60+ mathematicians, including 14 IMO gold medalists, 1 Fields MedalistHighest-quality benchmark construction
AI PerformanceLess than 2% of problems solved by leading modelsReveals substantial gap between AI and human mathematical capability
Tier 4 Commission50 research-level problems commissioned by OpenAITesting frontier reasoning capabilities
Version UpdatesToken budget increased 10x in November 2025Adapting to improved model inference
MetricEvidenceSource
Academic citations1,000+ citations across safety researchGoogle Scholar
Policy references50+ government documents cite EpochGovernment databases
Database usage10,000+ downloads of compute databaseEpoch analytics
Media coverageNYT “2024 Good Tech Awards” recognitionNew York Times
EA Forum engagementActive community discussion and feedbackEA Forum posts

Database expansion:

  • Added 200+ new model entries to Parameter Database
  • Enhanced tracking of Chinese and European models
  • Improved cost estimation methodologies
  • Real-time updates for new releases

Research breakthroughs:

  • Refined algorithmic efficiency measurement showing 6-12 month doubling times
  • Updated data exhaustion projections with synthetic data considerations
  • New economic modeling of AI investment trends
  • Bioweapons AI uplift analysis
AreaExpected DevelopmentImpactSource
Model scaling10+ models above 10^26 FLOP by 2026Over 200 projected by 2030Epoch projections
Data bottleneckHigh-quality text exhaustion begins 2026-2032Synthetic data scaling acceleratesEpoch data analysis
Compute governanceStandardized international monitoring neededEnhanced export controlsEpoch policy research
Benchmark development2 new benchmarks in developmentImproved capability measurement2025 Impact Report
Capability acceleration15 points/year on ECI (up from 8 points pre-April 2024)Faster than historical trendEpoch Capabilities Index
Open-source thresholdFrontier open models may exceed 10^26 FLOP before 2026Challenges compute governance approachEpoch data insights
UncertaintyImpact on EstimatesMitigation Strategy
Algorithmic breakthroughsCould accelerate timelines by yearsMultiple forecasting methods
Data efficiency improvementsMay extend scaling runwayConservative assumptions
Geopolitical disruptionCould fragment or accelerate developmentScenario planning
Hardware bottlenecksMay slow progress unexpectedlySupply chain analysis

Trend extrapolation reliability:

  • Optimists: Historical trends provide best available evidence for forecasting
  • Pessimists: Sharp left turns and discontinuities make extrapolation unreliable
  • Epoch position: Multiple methods with explicit uncertainty bounds

Information hazards:

  • Security concern: Publishing compute data aids adversaries in capability assessment
  • Racing dynamics: Timeline estimates may encourage competitive behavior
  • Transparency advocates: Public data essential for democratic governance
Forecasting Reliability Debate
Value of Empirical AI Forecasting (4 perspectives)
Essential Infrastructure
Strong

Epoch's data provides crucial foundation for rational planning. Timeline estimates inform urgency decisions. Compute tracking enables governance. Superior to pure speculation.

Policy community · Many safety researchers · EA researchers
Useful but Limited
Strong

Valuable for trend identification but shouldn't drive strategy alone. High uncertainty requires robust planning across scenarios rather than point estimates.

Cautious researchers · Some policymakers
Fundamentally Uncertain
Minority

AI development too discontinuous to forecast meaningfully. Unknown unknowns dominate. Resources better spent on robustness than prediction.

Anti-forecasting researchers · Some capability pessimists
Information Hazard Risk
Minority

Timeline publication creates racing dynamics. Compute data aids adversaries. False precision worse than acknowledged uncertainty. Focus on safety regardless of timelines.

Security-focused researchers · Some MIRI-adjacent views
Leadership Team
JS
Jaime Sevilla
Director & Co-Founder
TB
Tamay Besiroglu
Senior Researcher
AH
Anson Ho
Research Scientist
VD
Various data scientists
Database Team
FunctionTeam SizeKey Responsibilities2025 Focus
Research15-18 peopleForecasting, analysis, publicationsFrontierMath, Epoch Capabilities Index
Data & Engineering8-10 peopleDatabase infrastructure, automationAI Chip Sales explorer, model tracking
Operations4-6 peopleFunding, administration, communicationsGrant management, public engagement
AdvisoryExternalPolicy guidance, technical reviewAcademic partnerships
Total34 employeesHeadquartered in San FranciscoGrowing capacity
FunderAmountPeriodPurpose
Open Philanthropy (2025)$1,132,4882 yearsGeneral support
Open Philanthropy (additional)$1,922,565Multi-yearGeneral support
Open Philanthropy (FrontierMath)$10,0002025Benchmark improvements
Open Philanthropy (2022)$1,960,000InitialOrganization founding
Total raised≈$13M+2022-2025Research & operations

Additional funding sources:

  • Future of Humanity Institute (historical support)
  • Government contracts for specific projects
  • Research grants from academic institutions
OrganizationFocusMethodologyUpdate FrequencyPolicy ImpactDatabase Size
Epoch AIAI-specific empirical dataMultiple quantitative methodsContinuousHigh3,200+ models
MetaculusCrowdsourced forecastingPrediction aggregationReal-timeMediumN/A
AI ImpactsHistorical AI analysisCase studies, trend analysisIrregularMediumLimited
FHIExistential risk researchAcademic researchProject-basedHighN/A
METRModel evaluationsTechnical testingPer-modelGrowingN/A
Our World in DataData visualizationData aggregation (uses Epoch data)RegularHighDerivative
Organization TypeRelationship to EpochInformation Flow
Safety research orgsData consumersEpoch → Safety orgs
AI labsData subjectsLabs → Epoch (reluctantly)
Government bodiesPolicy clientsEpoch ↔ Government
Think tanksResearch partnersCollaborative

2026 Plans (from Impact Report):

  • Development of 2 new benchmarks beyond FrontierMath
  • Continued expansion of the Epoch Capabilities Index
  • Deeper analysis of decentralized training feasibility (10 GW training runs across thousands of kilometers)
  • Enhanced AI Chip Sales data explorer across Nvidia, Google, Amazon, AMD, and Huawei

Expanding scope:

  • Multimodal training data analysis beyond text
  • Energy consumption and environmental impact tracking (largest data centers now approaching gigawatt scale)
  • International AI development monitoring (enhanced coverage of Chinese and European models)
  • Risk assessment frameworks for different development pathways

Methodological improvements:

  • Better algorithmic progress measurement via Epoch Capabilities Index
  • Synthetic data quality and scaling analysis
  • Economic impact modeling of AI deployment
  • Scenario analysis for different development paths
ChallengeCurrent LimitationPlanned SolutionPriority
Data collectionManual curation, limited sourcesAutomated scraping, industry partnershipsHigh
International coverageUS/UK bias in dataPartnerships with Chinese and European researchersHigh
Real-time trackingLag in proprietary model informationIndustry reporting standards advocacyMedium
Resource constraints34 person teamGradual expansion with $1.1M/year budgetMedium
Compute governance gapsThreshold accuracy uncertainBetter compute-capability correlation researchHigh
Open-source proliferationFrontier open models approaching 10^26 FLOPPolicy recommendations for dual governanceHigh
Key Questions (6)
  • How accurate are extrapolation-based AI timeline forecasts given potential discontinuities?
  • Will synthetic data generation solve the training data bottleneck or create new limitations?
  • How should compute governance adapt as algorithmic efficiency reduces compute as a chokepoint?
  • What level of transparency in AI development is optimal for governance without security risks?
  • How can empirical forecasting organizations maintain independence while engaging with policymakers?
  • What leading indicators best predict dangerous capability emergence beyond compute scaling?
Resource TypeDescriptionLink
AI Models Database3,200+ models tracked with training compute, parameters, costepoch.ai/data/ai-models
Machine Learning TrendsReal-time visualization of AI progress metricsepoch.ai/trends
FrontierMath Benchmark350 expert-crafted math problemsepoch.ai/frontiermath
Epoch Capabilities IndexUnified capability measurement across benchmarksepoch.ai/benchmarks
Research Blog36+ Data Insights, 37+ Gradient Updates in 2025epoch.ai/blog
AI Chip SalesGlobal compute capacity tracking (15M+ H100-equivalents)epoch.ai/data
TitleYearImpactCitation
”Compute Trends Across Three Eras of Machine Learning”2022Foundational for compute governanceSevilla et al.
”Will We Run Out of Data?“2022Sparked synthetic data research boomVillalobos et al.
”Algorithmic Progress in Computer Vision”2023Quantified efficiency improvementsBesiroglu et al.
”FrontierMath: A Benchmark for Evaluating Advanced Mathematical Reasoning”2024arXiv paper on benchmark methodologyEpoch AI et al.
Top 10 Data Insights of 20252025Annual synthesis of key findingsEpoch AI
2025 Impact Report2025Comprehensive organizational reviewEpoch AI
Source TypeDescriptionExample Links
Policy DocumentsGovernment citations of Epoch workUS NAIRR, UK AI White Paper
Academic CitationsResearch building on Epoch dataGoogle Scholar search
Data VisualizationOur World in Data uses Epoch datasetsTraining compute charts
Media CoverageNYT “2024 Good Tech Awards”, InfoQ FrontierMath coverageMIT Technology Review
Industry AnalysisBusiness intelligence using Epoch metricsCB Insights, McKinsey AI reports
EA CommunityEA Forum discussions, 80,000 Hours listingsActive engagement