Longterm Wiki

Is Scaling All You Need?

scaling-debatecruxPath: /knowledge-base/debates/scaling-debate/
E272Entity ID (EID)
← Back to page5 backlinksQuality: 42Updated: 2026-03-13
Page Recorddatabase.json — merged from MDX frontmatter + Entity YAML + computed metrics at build time
{
  "id": "scaling-debate",
  "numericId": null,
  "path": "/knowledge-base/debates/scaling-debate/",
  "filePath": "knowledge-base/debates/scaling-debate.mdx",
  "title": "Is Scaling All You Need?",
  "quality": 42,
  "readerImportance": 55,
  "researchImportance": 81.5,
  "tacticalValue": null,
  "contentFormat": "article",
  "tractability": null,
  "neglectedness": null,
  "uncertainty": null,
  "causalLevel": null,
  "lastUpdated": "2026-03-13",
  "dateCreated": "2026-02-15",
  "llmSummary": "Comprehensive survey of the 2024-2025 scaling debate, documenting the shift from pure pretraining to 'scaling-plus' approaches after o3 achieved 87.5% on ARC-AGI-1 but GPT-5 faced 2-year delays. Expert consensus has moved to ~45% favoring hybrid approaches, with data wall projected 2026-2030 and AGI timelines spanning 5-30+ years depending on paradigm.",
  "description": "The scaling debate examines whether current AI approaches will reach AGI through more compute and data, or require new paradigms. By 2025, evidence is mixed: o3 achieved 87.5% on ARC-AGI-1, but GPT-5 took 2 years longer than expected and ARC-AGI-2 remains unsolved by all models. The emerging consensus favors 'scaling-plus'—combining pretraining with reasoning via test-time compute.",
  "ratings": {
    "novelty": 2.5,
    "rigor": 4,
    "actionability": 3,
    "completeness": 5.5
  },
  "category": "debates",
  "subcategory": "policy-debates",
  "clusters": [
    "ai-safety"
  ],
  "metrics": {
    "wordCount": 1024,
    "tableCount": 7,
    "diagramCount": 1,
    "internalLinks": 10,
    "externalLinks": 36,
    "footnoteCount": 0,
    "bulletRatio": 0.16,
    "sectionCount": 19,
    "hasOverview": false,
    "structuralScore": 14
  },
  "suggestedQuality": 93,
  "updateFrequency": 45,
  "evergreen": true,
  "wordCount": 1024,
  "unconvertedLinks": [
    {
      "text": "Stanford AI Index 2025",
      "url": "https://hai.stanford.edu/ai-index/2025-ai-index-report/technical-performance",
      "resourceId": "1a26f870e37dcc68",
      "resourceTitle": "Technical Performance - 2025 AI Index Report"
    },
    {
      "text": "ARC Prize Technical Report",
      "url": "https://arcprize.org/blog/oai-o3-pub-breakthrough",
      "resourceId": "457fa3b0b79d8812",
      "resourceTitle": "o3 scores 87.5% on ARC-AGI"
    },
    {
      "text": "Stanford AI Index",
      "url": "https://hai.stanford.edu/ai-index/2025-ai-index-report/technical-performance",
      "resourceId": "1a26f870e37dcc68",
      "resourceTitle": "Technical Performance - 2025 AI Index Report"
    },
    {
      "text": "Epoch AI",
      "url": "https://epoch.ai/blog/can-ai-scaling-continue-through-2030",
      "resourceId": "9587b65b1192289d",
      "resourceTitle": "Epoch AI"
    },
    {
      "text": "o1/o3 reasoning paradigm",
      "url": "https://openai.com/index/introducing-o3-and-o4-mini/",
      "resourceId": "bf92f3d905c3de0d",
      "resourceTitle": "announced December 2024"
    },
    {
      "text": "ARC Prize",
      "url": "https://arcprize.org/blog/oai-o3-pub-breakthrough",
      "resourceId": "457fa3b0b79d8812",
      "resourceTitle": "o3 scores 87.5% on ARC-AGI"
    },
    {
      "text": "OpenAI observed",
      "url": "https://www.interconnects.ai/p/openais-o3-the-2024-finale-of-ai",
      "resourceId": "3c8e4281a140e1cd",
      "resourceTitle": "GPQA Diamond"
    },
    {
      "text": "Introducing o3 and o4-mini",
      "url": "https://openai.com/index/introducing-o3-and-o4-mini/",
      "resourceId": "bf92f3d905c3de0d",
      "resourceTitle": "announced December 2024"
    },
    {
      "text": "Can AI scaling continue through 2030?",
      "url": "https://epoch.ai/blog/can-ai-scaling-continue-through-2030",
      "resourceId": "9587b65b1192289d",
      "resourceTitle": "Epoch AI"
    },
    {
      "text": "AI Index 2025",
      "url": "https://hai.stanford.edu/ai-index/2025-ai-index-report/technical-performance",
      "resourceId": "1a26f870e37dcc68",
      "resourceTitle": "Technical Performance - 2025 AI Index Report"
    },
    {
      "text": "o3: The grand finale of AI in 2024",
      "url": "https://www.interconnects.ai/p/openais-o3-the-2024-finale-of-ai",
      "resourceId": "3c8e4281a140e1cd",
      "resourceTitle": "GPQA Diamond"
    },
    {
      "text": "Scaling Laws for LLMs",
      "url": "https://cameronrwolfe.substack.com/p/llm-scaling-laws",
      "resourceId": "056c40c4515292c5",
      "resourceTitle": "AIME 2024"
    },
    {
      "text": "AI Beyond the Scaling Laws",
      "url": "https://www.hec.edu/en/dare/tech-ai/ai-beyond-scaling-laws",
      "resourceId": "40560014cfc7663d",
      "resourceTitle": "some researchers note"
    }
  ],
  "unconvertedLinkCount": 13,
  "convertedLinkCount": 0,
  "backlinkCount": 5,
  "hallucinationRisk": {
    "level": "medium",
    "score": 45,
    "factors": [
      "no-citations",
      "conceptual-content"
    ]
  },
  "entityType": "crux",
  "redundancy": {
    "maxSimilarity": 14,
    "similarPages": [
      {
        "id": "large-language-models",
        "title": "Large Language Models",
        "path": "/knowledge-base/capabilities/large-language-models/",
        "similarity": 14
      },
      {
        "id": "agi-timeline-debate",
        "title": "When Will AGI Arrive?",
        "path": "/knowledge-base/debates/agi-timeline-debate/",
        "similarity": 13
      },
      {
        "id": "agi-timeline",
        "title": "AGI Timeline",
        "path": "/knowledge-base/forecasting/agi-timeline/",
        "similarity": 13
      },
      {
        "id": "dense-transformers",
        "title": "Dense Transformers",
        "path": "/knowledge-base/intelligence-paradigms/dense-transformers/",
        "similarity": 13
      },
      {
        "id": "coding",
        "title": "Autonomous Coding",
        "path": "/knowledge-base/capabilities/coding/",
        "similarity": 12
      }
    ]
  },
  "coverage": {
    "passing": 8,
    "total": 13,
    "targets": {
      "tables": 4,
      "diagrams": 0,
      "internalLinks": 8,
      "externalLinks": 5,
      "footnotes": 3,
      "references": 3
    },
    "actuals": {
      "tables": 7,
      "diagrams": 1,
      "internalLinks": 10,
      "externalLinks": 36,
      "footnotes": 0,
      "references": 7,
      "quotesWithQuotes": 0,
      "quotesTotal": 0,
      "accuracyChecked": 0,
      "accuracyTotal": 0
    },
    "items": {
      "llmSummary": "green",
      "schedule": "green",
      "entity": "green",
      "editHistory": "red",
      "overview": "red",
      "tables": "green",
      "diagrams": "green",
      "internalLinks": "green",
      "externalLinks": "green",
      "footnotes": "red",
      "references": "green",
      "quotes": "red",
      "accuracy": "red"
    },
    "ratingsString": "N:2.5 R:4 A:3 C:5.5"
  },
  "readerRank": 265,
  "researchRank": 79,
  "recommendedScore": 132.87
}
External Links
{
  "lesswrong": "https://www.lesswrong.com/tag/scaling-laws"
}
Backlinks (5)
idtitletyperelationship
ai-compute-scaling-metricsAI Compute Scaling Metricsanalysis
__index__/knowledge-base/debatesKey Debatesconcept
agi-timelineAGI Timelineconcept
compounding-risks-analysisCompounding Risks Analysisanalysis
toby-ordToby Ordperson
Longterm Wiki