Longterm Wiki

Redwood Research

redwood-researchorganizationPath: /knowledge-base/organizations/redwood-research/
E557Entity ID (EID)
← Back to page49 backlinksQuality: 78Updated: 2026-03-13
Page Recorddatabase.json — merged from MDX frontmatter + Entity YAML + computed metrics at build time
{
  "id": "redwood-research",
  "numericId": null,
  "path": "/knowledge-base/organizations/redwood-research/",
  "filePath": "knowledge-base/organizations/redwood-research.mdx",
  "title": "Redwood Research",
  "quality": 78,
  "readerImportance": 31.5,
  "researchImportance": 45,
  "tacticalValue": null,
  "contentFormat": "article",
  "tractability": null,
  "neglectedness": null,
  "uncertainty": null,
  "causalLevel": null,
  "lastUpdated": "2026-03-13",
  "dateCreated": "2026-02-15",
  "llmSummary": null,
  "description": "A nonprofit AI safety and security research organization founded in 2021, known for pioneering AI Control research, developing causal scrubbing interpretability methods, and conducting landmark alignment faking studies with Anthropic.",
  "ratings": {
    "novelty": 7,
    "rigor": 7,
    "actionability": 5,
    "completeness": 7
  },
  "category": "organizations",
  "subcategory": "safety-orgs",
  "clusters": [
    "ai-safety",
    "community"
  ],
  "metrics": {
    "wordCount": 1497,
    "tableCount": 8,
    "diagramCount": 0,
    "internalLinks": 22,
    "externalLinks": 22,
    "footnoteCount": 0,
    "bulletRatio": 0,
    "sectionCount": 24,
    "hasOverview": true,
    "structuralScore": 14
  },
  "suggestedQuality": 93,
  "updateFrequency": 21,
  "evergreen": true,
  "wordCount": 1497,
  "unconvertedLinks": [
    {
      "text": "Anthropic",
      "url": "https://www.anthropic.com/research/alignment-faking",
      "resourceId": "c2cfd72baafd64a9",
      "resourceTitle": "Anthropic's 2024 alignment faking study"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/pdf/2312.06942",
      "resourceId": "cc80ab28579c5794",
      "resourceTitle": "Redwood Research's AI Control paper (December 2023)"
    },
    {
      "text": "Anthropic",
      "url": "https://www.anthropic.com/research/alignment-faking",
      "resourceId": "c2cfd72baafd64a9",
      "resourceTitle": "Anthropic's 2024 alignment faking study"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/abs/2405.19550",
      "resourceId": "2bb15dbd965ba994",
      "resourceTitle": "Stress-Testing Capability Elicitation With Password-Locked Models - arXiv"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/abs/2501.17315",
      "resourceId": "56bae0415b228464",
      "resourceTitle": "A Sketch of an AI Control Safety Case - arXiv"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/abs/2504.10374",
      "resourceId": "1cbfd6cb09ce2b50",
      "resourceTitle": "Ctrl-Z: Controlling AI Agents via Resampling - arXiv"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/abs/2512.15688",
      "resourceId": "a4efd028ec6d9614",
      "resourceTitle": "BashArena: A Control Setting for Highly Privileged AI Agents - arXiv"
    },
    {
      "text": "OP Grants",
      "url": "https://www.openphilanthropy.org/grants/redwood-research-general-support/",
      "resourceId": "8c79e00bab007a63",
      "resourceTitle": "over $9.4 million from Open Philanthropy"
    },
    {
      "text": "OP Grants",
      "url": "https://www.openphilanthropy.org/grants/",
      "resourceId": "2fcdf851ed57384c",
      "resourceTitle": "Open Philanthropy Grants Database"
    }
  ],
  "unconvertedLinkCount": 9,
  "convertedLinkCount": 0,
  "backlinkCount": 49,
  "citationHealth": {
    "total": 45,
    "withQuotes": 30,
    "verified": 30,
    "accuracyChecked": 30,
    "accurate": 18,
    "inaccurate": 0,
    "avgScore": 0.9502938191095988
  },
  "hallucinationRisk": {
    "level": "medium",
    "score": 60,
    "factors": [
      "biographical-claims",
      "no-citations",
      "high-rigor"
    ]
  },
  "entityType": "organization",
  "redundancy": {
    "maxSimilarity": 13,
    "similarPages": [
      {
        "id": "ai-control",
        "title": "AI Control",
        "path": "/knowledge-base/responses/ai-control/",
        "similarity": 13
      },
      {
        "id": "apollo-research",
        "title": "Apollo Research",
        "path": "/knowledge-base/organizations/apollo-research/",
        "similarity": 12
      },
      {
        "id": "secure-ai-project",
        "title": "Secure AI Project",
        "path": "/knowledge-base/organizations/secure-ai-project/",
        "similarity": 12
      },
      {
        "id": "ssi",
        "title": "Safe Superintelligence Inc (SSI)",
        "path": "/knowledge-base/organizations/ssi/",
        "similarity": 12
      },
      {
        "id": "ilya-sutskever",
        "title": "Ilya Sutskever",
        "path": "/knowledge-base/people/ilya-sutskever/",
        "similarity": 12
      }
    ]
  },
  "coverage": {
    "passing": 7,
    "total": 13,
    "targets": {
      "tables": 6,
      "diagrams": 1,
      "internalLinks": 12,
      "externalLinks": 7,
      "footnotes": 4,
      "references": 4
    },
    "actuals": {
      "tables": 8,
      "diagrams": 0,
      "internalLinks": 22,
      "externalLinks": 22,
      "footnotes": 0,
      "references": 9,
      "quotesWithQuotes": 30,
      "quotesTotal": 45,
      "accuracyChecked": 30,
      "accuracyTotal": 45
    },
    "items": {
      "llmSummary": "red",
      "schedule": "green",
      "entity": "green",
      "editHistory": "red",
      "overview": "green",
      "tables": "green",
      "diagrams": "red",
      "internalLinks": "green",
      "externalLinks": "green",
      "footnotes": "red",
      "references": "green",
      "quotes": "amber",
      "accuracy": "amber"
    },
    "ratingsString": "N:7 R:7 A:5 C:7"
  },
  "readerRank": 438,
  "researchRank": 306,
  "recommendedScore": 193.28
}
External Links
{
  "lesswrong": "https://www.lesswrong.com/tag/redwood-research",
  "eaForum": "https://forum.effectivealtruism.org/topics/redwood-research",
  "grokipedia": "https://grokipedia.com/page/Redwood_Research"
}
Backlinks (49)
idtitletyperelationship
field-buildingAI Safety Field Building and Communitycrux
research-agendasAI Alignment Research Agendascrux
technical-researchTechnical AI Safety Researchcrux
conjectureConjectureorganization
ai-controlAI Controlsafety-agenda
interpretabilityInterpretabilitysafety-agenda
language-modelsLarge Language Modelscapability
large-language-modelsLarge Language Modelsconcept
why-alignment-easyWhy Alignment Might Be Easyargument
ea-epistemic-failures-in-the-ftx-eraEA Epistemic Failures in the FTX Eraconcept
ea-longtermist-wins-lossesEA and Longtermist Wins and Lossesconcept
ai-talent-market-dynamicsAI Talent Market Dynamicsanalysis
corrigibility-failure-pathwaysCorrigibility Failure Pathwaysanalysis
deceptive-alignment-decompositionDeceptive Alignment Decomposition Modelanalysis
instrumental-convergence-frameworkInstrumental Convergence Frameworkanalysis
intervention-effectiveness-matrixIntervention Effectiveness Matrixanalysis
model-organisms-of-misalignmentModel Organisms of Misalignmentanalysis
planning-for-frontier-lab-scalingPlanning for Frontier Lab Scalinganalysis
power-seeking-conditionsPower-Seeking Emergence Conditions Modelanalysis
safety-spending-at-scaleSafety Spending at Scaleanalysis
scheming-likelihood-modelScheming Likelihood Assessmentanalysis
worldview-intervention-mappingWorldview-Intervention Mappinganalysis
arcARC (Alignment Research Center)organization
bridgewater-aia-labsBridgewater AIA Labsorganization
caisCAIS (Center for AI Safety)organization
controlaiControlAIorganization
ea-funding-absorption-capacityEA Funding Absorption Capacityconcept
ea-globalEA Globalorganization
fliFuture of Life Institute (FLI)organization
founders-fundFounders Fundorganization
frontier-model-forumFrontier Model Forumorganization
ftx-future-fundFTX Future Fundorganization
funders-overviewLongtermist Funders (Overview)concept
giving-pledgeGiving Pledgeorganization
__index__/knowledge-base/organizationsOrganizationsconcept
matsMATS ML Alignment Theory Scholars programorganization
palisade-researchPalisade Researchorganization
safety-orgs-overviewAI Safety Organizations (Overview)concept
seldon-labSeldon Laborganization
sentinelSentinel (Catastrophic Risk Foresight)organization
swift-centreSwift Centreorganization
the-foundation-layerThe Foundation Layerorganization
ajeya-cotraAjeya Cotraperson
connor-leahyConnor Leahyperson
holden-karnofskyHolden Karnofskyperson
jan-leikeJan Leikeperson
nick-bostromNick Bostromperson
alignmentAI Alignmentapproach
deceptive-alignmentDeceptive Alignmentrisk
Longterm Wiki