Intervention Timing Windows
intervention-timing-windowsanalysisPath: /knowledge-base/models/intervention-timing-windows/
E178Entity ID (EID)
Page Recorddatabase.json — merged from MDX frontmatter + Entity YAML + computed metrics at build time
{
"id": "intervention-timing-windows",
"numericId": null,
"path": "/knowledge-base/models/intervention-timing-windows/",
"filePath": "knowledge-base/models/intervention-timing-windows.mdx",
"title": "Intervention Timing Windows",
"quality": 72,
"readerImportance": 90,
"researchImportance": 61.5,
"tacticalValue": 60,
"contentFormat": "article",
"tractability": null,
"neglectedness": null,
"uncertainty": null,
"causalLevel": null,
"lastUpdated": "2026-03-13",
"dateCreated": "2026-02-15",
"llmSummary": "Framework for prioritizing AI safety interventions by temporal urgency rather than impact alone, identifying four critical closing windows (2024-2028): compute governance (70% closure by 2027), international coordination (60% by 2028), lab safety culture (80% by 2026), and regulatory precedent (75% by 2027). Recommends reallocating 20-30% of resources from stable-window work to closing-window interventions, with specific funding increases (triple compute governance, double international coordination) and quantified timelines with uncertainty ranges.",
"description": "Strategic model categorizing AI safety interventions by temporal urgency. Identifies compute governance (70% closure by 2027), international coordination (60% closure by 2028), lab safety culture (80% closure by 2026), and regulatory precedent (75% closure by 2027) as closing windows requiring immediate action. Recommends shifting 20-30% of resources toward closing-window interventions, with quantified timelines and uncertainty ranges for each window.",
"ratings": {
"focus": 8.5,
"novelty": 6.5,
"rigor": 7,
"completeness": 8,
"concreteness": 8.5,
"actionability": 8
},
"category": "models",
"subcategory": "timeline-models",
"clusters": [
"ai-safety",
"governance"
],
"metrics": {
"wordCount": 4401,
"tableCount": 30,
"diagramCount": 3,
"internalLinks": 59,
"externalLinks": 28,
"footnoteCount": 0,
"bulletRatio": 0.12,
"sectionCount": 43,
"hasOverview": true,
"structuralScore": 15
},
"suggestedQuality": 100,
"updateFrequency": 90,
"evergreen": true,
"wordCount": 4401,
"unconvertedLinks": [
{
"text": "GovAI",
"url": "https://www.governance.ai/",
"resourceId": "f35c467b353f990f",
"resourceTitle": "GovAI"
},
{
"text": "CSET Georgetown",
"url": "https://cset.georgetown.edu/",
"resourceId": "f0d95954b449240a",
"resourceTitle": "CSET: AI Market Dynamics"
},
{
"text": "Institute for Law & AI research",
"url": "https://law-ai.org/the-role-of-compute-thresholds-for-ai-governance/",
"resourceId": "510c42bfa643b8de",
"resourceTitle": "EU AI Act"
},
{
"text": "Atlantic Council analysis",
"url": "https://www.atlanticcouncil.org/blogs/new-atlanticist/reading-between-the-lines-of-the-dueling-us-and-chinese-ai-action-plans/",
"resourceId": "7629a035e7e22ee1",
"resourceTitle": "Paris AI Summit divergence"
},
{
"text": "White House executive order analysis",
"url": "https://www.whitehouse.gov/presidential-actions/2025/12/eliminating-state-law-obstruction-of-national-artificial-intelligence-policy/",
"resourceId": "07bbfdc7df05b11a",
"resourceTitle": "Ensuring a National Policy Framework for Artificial Intelligence (The White House, December 11, 2025)"
},
{
"text": "Institute for Law & AI",
"url": "https://law-ai.org/the-role-of-compute-thresholds-for-ai-governance/",
"resourceId": "510c42bfa643b8de",
"resourceTitle": "EU AI Act"
},
{
"text": "GovAI Research",
"url": "https://www.governance.ai/research",
"resourceId": "571cb6299c6d27cf",
"resourceTitle": "Governance research"
},
{
"text": "FLI AI Safety Index 2024",
"url": "https://futureoflife.org/document/fli-ai-safety-index-2024/",
"resourceId": "f7ea8fb78f67f717",
"resourceTitle": "Future of Life Institute: AI Safety Index 2024"
}
],
"unconvertedLinkCount": 8,
"convertedLinkCount": 41,
"backlinkCount": 1,
"hallucinationRisk": {
"level": "medium",
"score": 40,
"factors": [
"no-citations",
"high-rigor"
]
},
"entityType": "analysis",
"redundancy": {
"maxSimilarity": 18,
"similarPages": [
{
"id": "international-coordination-game",
"title": "International AI Coordination Game",
"path": "/knowledge-base/models/international-coordination-game/",
"similarity": 18
},
{
"id": "us-aisi",
"title": "US AI Safety Institute",
"path": "/knowledge-base/organizations/us-aisi/",
"similarity": 17
},
{
"id": "coordination-mechanisms",
"title": "International Coordination Mechanisms",
"path": "/knowledge-base/responses/coordination-mechanisms/",
"similarity": 17
},
{
"id": "export-controls",
"title": "AI Chip Export Controls",
"path": "/knowledge-base/responses/export-controls/",
"similarity": 17
},
{
"id": "governance-policy",
"title": "AI Governance and Policy",
"path": "/knowledge-base/responses/governance-policy/",
"similarity": 17
}
]
},
"coverage": {
"passing": 9,
"total": 13,
"targets": {
"tables": 18,
"diagrams": 2,
"internalLinks": 35,
"externalLinks": 22,
"footnotes": 13,
"references": 13
},
"actuals": {
"tables": 30,
"diagrams": 3,
"internalLinks": 59,
"externalLinks": 28,
"footnotes": 0,
"references": 34,
"quotesWithQuotes": 0,
"quotesTotal": 0,
"accuracyChecked": 0,
"accuracyTotal": 0
},
"items": {
"llmSummary": "green",
"schedule": "green",
"entity": "green",
"editHistory": "red",
"overview": "green",
"tables": "green",
"diagrams": "green",
"internalLinks": "green",
"externalLinks": "green",
"footnotes": "red",
"references": "green",
"quotes": "red",
"accuracy": "red"
},
"ratingsString": "N:6.5 R:7 A:8 C:8"
},
"readerRank": 20,
"researchRank": 212,
"recommendedScore": 210.86
}External Links
No external links
Backlinks (1)
| id | title | type | relationship |
|---|---|---|---|
| ai-acceleration-tradeoff | AI Acceleration Tradeoff Model | analysis | related |