Skip to content
Longterm Wiki
Back

Training Compute of Frontier AI Models Grows by 4-5x Per Year (Epoch AI)

web

Credibility Rating

4/5
High(4)

High quality. Established institution or organization with editorial oversight and accountability.

Rating inherited from publication venue: Epoch AI

This Epoch AI analysis is frequently cited in AI governance and safety contexts to justify compute-based thresholds in policy proposals, such as executive orders requiring reporting above certain FLOP counts.

Metadata

Importance: 72/100blog postanalysis

Summary

Epoch AI analyzes historical trends in the training compute used for frontier AI models, finding that compute has grown approximately 4-5x per year. This rapid scaling has significant implications for AI capabilities trajectories, resource requirements, and safety planning horizons.

Key Points

  • Frontier AI model training compute has grown at roughly 4-5x per year, a faster rate than Moore's Law.
  • This trend implies dramatic increases in compute requirements and costs for state-of-the-art models over short timeframes.
  • The analysis provides empirical grounding for forecasting when compute thresholds relevant to risk assessments might be crossed.
  • Rapid compute scaling informs governance discussions around compute-based monitoring and regulatory trigger points.
  • Historical data on compute growth is essential for calibrating timelines for transformative or dangerous AI capabilities.

Cited by 1 page

PageTypeQuality
AI Capability Threshold ModelAnalysis72.0

Cached Content Preview

HTTP 200Fetched Apr 9, 202632 KB
Training compute of frontier AI models grows by 4-5x per year | Epoch AI 

 
 
 
 

 

 
 

 
 

 Introduction

 Over the last ten years, we have witnessed a dramatic increase in the computational resources dedicated to training state-of-the-art AI models. This strategy has been incredibly productive, translating into large gains in generality and performance . For example, we estimate that about two-thirds of the improvements in performance in language models in the last decade have been due to increases in model scale.

 Given the central role of scaling, it is important to track how the computational resources (‘compute’) used to train models have grown in recent years. In this short article, we provide an updated view of the trends so far, having collected three times more data since our last analysis .

 We tentatively conclude that compute growth in recent years is currently best described as increasing by a factor of 4-5x/year. We find consistent growth between recent notable models, the running top 10 of models by compute, recent large language models, and top models released by OpenAI, Google DeepMind and Meta AI.

 There are some unresolved uncertainties. We cannot rule out that the overall trend of compute might have accelerated. We also find evidence of a slowdown of growth in the frontier around 2018, which complicates the interpretation, and recent frontier growth since 2018 is better described as a 4x/year trend. We also find a significantly faster trend for notable language models overall, which have grown as fast as 9x/year between June 2017 and May 2024. However, when focusing on the frontier of language models, we see that the trend slows down to a ~5x/year pace after the largest language models catch up with the overall frontier in AI around mid-2020.

 All in all, we recommend summarizing the recent trend of compute growth for notable and frontier models with the 4-5x/year figure. This should also be used as a baseline for expectations of growth in the future, before taking into account additional considerations such as possible bottlenecks or speed-ups.

 Figure 1: Summary of the compute growth trends we found for overall notable models (top left), frontier models (top right), top language models (bottom left) and top models within leading companies (bottom right). All point to a recent trend of 4-5x/year growth.

 The overall trend of training compute growth has held

 We have previously investigated the trend of growing training compute. In 2022, we found that the amount of compute used to train notable ML models had grown about 4x per year from 2010 to 2022. 1 Many notable models have been released since, and we have expanded our database by tripling the number of compute estimates , 2 so an update is in order.

 In short, we find that the amount of compute used to train notable models has grown about 4.1x/year (90% CI: 3.7x to 4.6x) between 2010 to May 2024. 3 If we look at the trend since our last update in Feb 2022, we

... (truncated, 32 KB total)
Resource ID: 7d0515f6079d8beb | Stable ID: sid_n83bbTc9MB