Skip to content
Longterm Wiki

Redwood Research

redwood-researchorganizationPath: /knowledge-base/organizations/redwood-research/
E557Entity ID (EID)
← Back to page54 backlinksQuality: 78Updated: 2026-01-31
Page Recorddatabase.json — merged from MDX frontmatter + Entity YAML + computed metrics at build time
{
  "id": "redwood-research",
  "wikiId": "E557",
  "path": "/knowledge-base/organizations/redwood-research/",
  "filePath": "knowledge-base/organizations/redwood-research.mdx",
  "title": "Redwood Research",
  "quality": 78,
  "readerImportance": 31.5,
  "researchImportance": 45,
  "tacticalValue": null,
  "contentFormat": "article",
  "causalLevel": null,
  "lastUpdated": "2026-01-31",
  "dateCreated": "2026-02-15",
  "summary": null,
  "description": "A nonprofit AI safety and security research organization founded in 2021, known for pioneering AI Control research, developing causal scrubbing interpretability methods, and conducting landmark alignment faking studies with Anthropic.",
  "ratings": {
    "novelty": 7,
    "rigor": 7,
    "completeness": 7,
    "actionability": 5
  },
  "category": "organizations",
  "subcategory": "safety-orgs",
  "clusters": [
    "ai-safety",
    "community"
  ],
  "metrics": {
    "wordCount": 1489,
    "tableCount": 8,
    "diagramCount": 0,
    "internalLinks": 26,
    "externalLinks": 22,
    "footnoteCount": 32,
    "bulletRatio": 0,
    "sectionCount": 24,
    "hasOverview": true,
    "structuralScore": 14
  },
  "suggestedQuality": 93,
  "updateFrequency": 21,
  "evergreen": true,
  "wordCount": 1489,
  "unconvertedLinks": [
    {
      "text": "LessWrong",
      "url": "https://www.lesswrong.com/posts/JvZhhzycHu2Yd57RN/causal-scrubbing-a-method-for-rigorously-testing",
      "resourceId": "018b403483a001b9",
      "resourceTitle": "Causal Scrubbing: a method for rigorously testing interpretability hypotheses [Redwood Research]"
    },
    {
      "text": "EA Forum",
      "url": "https://forum.effectivealtruism.org/posts/MGbdhjgd2v6cg3vjv/apply-to-the-redwood-research-mechanistic-interpretability",
      "resourceId": "58b17dcb71ac7212",
      "resourceTitle": "Apply to the Redwood Research Mechanistic Interpretability Experiment (REMIX), a research program in Berkeley"
    },
    {
      "text": "Anthropic",
      "url": "https://www.anthropic.com/research/alignment-faking",
      "resourceId": "c2cfd72baafd64a9",
      "resourceTitle": "Anthropic's 2024 alignment faking study"
    },
    {
      "text": "EA Forum",
      "url": "https://forum.effectivealtruism.org/posts/MGbdhjgd2v6cg3vjv/apply-to-the-redwood-research-mechanistic-interpretability",
      "resourceId": "58b17dcb71ac7212",
      "resourceTitle": "Apply to the Redwood Research Mechanistic Interpretability Experiment (REMIX), a research program in Berkeley"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/pdf/2312.06942",
      "resourceId": "cc80ab28579c5794",
      "resourceTitle": "Redwood Research's AI Control paper (December 2023)"
    },
    {
      "text": "Anthropic",
      "url": "https://www.anthropic.com/research/alignment-faking",
      "resourceId": "c2cfd72baafd64a9",
      "resourceTitle": "Anthropic's 2024 alignment faking study"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/abs/2405.19550",
      "resourceId": "2bb15dbd965ba994",
      "resourceTitle": "Stress-Testing Capability Elicitation With Password-Locked Models - arXiv"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/abs/2501.17315",
      "resourceId": "56bae0415b228464",
      "resourceTitle": "A Sketch of an AI Control Safety Case - arXiv"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/abs/2504.10374",
      "resourceId": "1cbfd6cb09ce2b50",
      "resourceTitle": "Ctrl-Z: Controlling AI Agents via Resampling - arXiv"
    },
    {
      "text": "arXiv",
      "url": "https://arxiv.org/abs/2512.15688",
      "resourceId": "a4efd028ec6d9614",
      "resourceTitle": "BashArena: A Control Setting for Highly Privileged AI Agents - arXiv"
    },
    {
      "text": "OP Grants",
      "url": "https://www.openphilanthropy.org/grants/redwood-research-general-support/",
      "resourceId": "8c79e00bab007a63",
      "resourceTitle": "over $9.4 million from Open Philanthropy"
    },
    {
      "text": "OP Grants",
      "url": "https://www.openphilanthropy.org/grants/",
      "resourceId": "2fcdf851ed57384c",
      "resourceTitle": "Open Philanthropy Grants Database"
    }
  ],
  "unconvertedLinkCount": 12,
  "convertedLinkCount": 0,
  "backlinkCount": 54,
  "citationHealth": {
    "total": 45,
    "withQuotes": 30,
    "verified": 30,
    "accuracyChecked": 30,
    "accurate": 18,
    "inaccurate": 0,
    "avgScore": 0.9502938191095988
  },
  "hallucinationRisk": {
    "level": "low",
    "score": 30,
    "factors": [
      "biographical-claims",
      "well-cited",
      "high-rigor"
    ]
  },
  "entityType": "organization",
  "redundancy": {
    "maxSimilarity": 13,
    "similarPages": [
      {
        "id": "ai-control",
        "title": "AI Control",
        "path": "/knowledge-base/responses/ai-control/",
        "similarity": 13
      },
      {
        "id": "apollo-research",
        "title": "Apollo Research",
        "path": "/knowledge-base/organizations/apollo-research/",
        "similarity": 12
      },
      {
        "id": "secure-ai-project",
        "title": "Secure AI Project",
        "path": "/knowledge-base/organizations/secure-ai-project/",
        "similarity": 12
      },
      {
        "id": "ssi",
        "title": "Safe Superintelligence Inc. (SSI)",
        "path": "/knowledge-base/organizations/ssi/",
        "similarity": 12
      },
      {
        "id": "ilya-sutskever",
        "title": "Ilya Sutskever",
        "path": "/knowledge-base/people/ilya-sutskever/",
        "similarity": 12
      }
    ]
  },
  "coverage": {
    "passing": 8,
    "total": 13,
    "targets": {
      "tables": 6,
      "diagrams": 1,
      "internalLinks": 12,
      "externalLinks": 7,
      "footnotes": 4,
      "references": 4
    },
    "actuals": {
      "tables": 8,
      "diagrams": 0,
      "internalLinks": 26,
      "externalLinks": 22,
      "footnotes": 32,
      "references": 9,
      "quotesWithQuotes": 30,
      "quotesTotal": 45,
      "accuracyChecked": 30,
      "accuracyTotal": 45
    },
    "items": {
      "summary": "red",
      "schedule": "green",
      "entity": "green",
      "editHistory": "red",
      "overview": "green",
      "tables": "green",
      "diagrams": "red",
      "internalLinks": "green",
      "externalLinks": "green",
      "footnotes": "green",
      "references": "green",
      "quotes": "amber",
      "accuracy": "amber"
    },
    "ratingsString": "N:7 R:7 A:5 C:7"
  },
  "readerRank": 438,
  "researchRank": 305,
  "recommendedScore": 185.14
}
External Links
{
  "lesswrong": "https://www.lesswrong.com/tag/redwood-research",
  "eaForum": "https://forum.effectivealtruism.org/topics/redwood-research",
  "grokipedia": "https://grokipedia.com/page/Redwood_Research"
}
Backlinks (54)
idtitletyperelationship
field-buildingAI Safety Field Building and Communitycrux
research-agendasAI Alignment Research Agendascrux
technical-researchTechnical AI Safety Researchcrux
conjectureConjectureorganization
apart-researchApart Researchorganizationrelated
ai-controlAI Controlresearch-arearesearch
large-language-modelsLarge Language Modelsconcept
accident-risksAI Accident Risk Cruxescrux
intra-coalition-engagement-strategiesIntra-Coalition Engagement Strategiesconcept
is-ai-xrisk-realIs AI Existential Risk Real?crux
why-alignment-easyWhy Alignment Might Be Easyargument
ea-epistemic-failures-in-the-ftx-eraEA Epistemic Failures in the FTX Eraconcept
ea-longtermist-wins-lossesEA and Longtermist Wins and Lossesconcept
ai-talent-market-dynamicsAI Talent Market Dynamicsanalysis
corrigibility-failure-pathwaysCorrigibility Failure Pathwaysanalysis
deceptive-alignment-decompositionDeceptive Alignment Decomposition Modelanalysis
instrumental-convergence-frameworkInstrumental Convergence Frameworkanalysis
intervention-effectiveness-matrixIntervention Effectiveness Matrixanalysis
model-organisms-of-misalignmentModel Organisms of Misalignmentanalysis
planning-for-frontier-lab-scalingPlanning for Frontier Lab Scalinganalysis
power-seeking-conditionsPower-Seeking Emergence Conditions Modelanalysis
safety-spending-at-scaleSafety Spending at Scaleanalysis
scheming-likelihood-modelScheming Likelihood Assessmentanalysis
worldview-intervention-mappingWorldview-Intervention Mappinganalysis
arcAlignment Research Center (ARC)organization
bridgewater-aia-labsBridgewater AIA Labsorganization
caisCenter for AI Safety (CAIS)organization
controlaiControlAIorganization
ea-funding-absorption-capacityEA Funding Absorption Capacityconcept
ea-globalEA Globalorganization
fliFuture of Life Institute (FLI)organization
founders-fundFounders Fundorganization
frontier-model-forumFrontier Model Forumorganization
ftx-future-fundFTX Future Fundorganization
funders-overviewLongtermist Funders (Overview)concept
giving-pledgeGiving Pledgeorganization
__index__/knowledge-base/organizationsOrganizationsconcept
matsMATS ML Alignment Theory Scholars programorganization
palisade-researchPalisade Researchorganization
safety-orgs-overviewAI Safety Organizations (Overview)concept
seldon-labSeldon Laborganization
sentinelSentinel (Catastrophic Risk Foresight)organization
swift-centreSwift Centreorganization
the-foundation-layerThe Foundation Layerorganization
ajeya-cotraAjeya Cotraperson
connor-leahyConnor Leahyperson
david-dalrympleDavid Dalrympleperson
dustin-moskovitzDustin Moskovitzperson
holden-karnofskyHolden Karnofskyperson
jan-leikeJan Leikeperson
neel-nandaNeel Nandaperson
nick-bostromNick Bostromperson
alignmentAI Alignmentapproach
deceptive-alignmentDeceptive Alignmentrisk
Longterm Wiki