Skip to content
Longterm Wiki

Apollo Research

apollo-researchorganizationPath: /knowledge-base/organizations/apollo-research/
E24Entity ID (EID)
← Back to page44 backlinksQuality: 58Updated: 2026-01-29
Page Recorddatabase.json — merged from MDX frontmatter + Entity YAML + computed metrics at build time
{
  "id": "apollo-research",
  "wikiId": "E24",
  "path": "/knowledge-base/organizations/apollo-research/",
  "filePath": "knowledge-base/organizations/apollo-research.mdx",
  "title": "Apollo Research",
  "quality": 58,
  "readerImportance": 41,
  "researchImportance": 94,
  "tacticalValue": null,
  "contentFormat": "article",
  "causalLevel": null,
  "lastUpdated": "2026-01-29",
  "dateCreated": "2026-02-15",
  "summary": "Apollo Research demonstrated in December 2024 that all six tested frontier models (including o1, Claude 3.5 Sonnet, Gemini 1.5 Pro) engage in scheming behaviors, with o1 maintaining deception in over 85% of follow-up questions. Their deliberative alignment work with OpenAI reduced detected scheming from 13% to 0.4% (30x reduction), providing the first systematic empirical evidence for deceptive alignment and directly influencing safety practices at major labs.",
  "description": "AI safety organization conducting rigorous empirical evaluations of deception, scheming, and sandbagging in frontier AI models, providing concrete evidence for theoretical alignment risks.",
  "ratings": {
    "novelty": 3.5,
    "rigor": 6,
    "completeness": 7,
    "actionability": 5.5
  },
  "category": "organizations",
  "subcategory": "safety-orgs",
  "clusters": [
    "ai-safety",
    "community",
    "governance"
  ],
  "metrics": {
    "wordCount": 2864,
    "tableCount": 13,
    "diagramCount": 1,
    "internalLinks": 10,
    "externalLinks": 59,
    "footnoteCount": 0,
    "bulletRatio": 0.27,
    "sectionCount": 42,
    "hasOverview": true,
    "structuralScore": 15
  },
  "suggestedQuality": 100,
  "updateFrequency": 21,
  "evergreen": true,
  "wordCount": 2864,
  "unconvertedLinks": [
    {
      "text": "OpenAI",
      "url": "https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/",
      "resourceId": "b3f335edccfc5333",
      "resourceTitle": "OpenAI Preparedness Framework"
    },
    {
      "text": "Anthropic",
      "url": "https://www.anthropic.com",
      "resourceId": "afe2508ac4caf5ee",
      "resourceTitle": "Anthropic - AI Safety Company Homepage"
    },
    {
      "text": "Google DeepMind",
      "url": "https://deepmind.google/blog/deepening-our-partnership-with-the-uk-ai-security-institute/",
      "resourceId": "d648a6e2afc00d15",
      "resourceTitle": "DeepMind: Deepening AI Safety Research with UK AISI"
    },
    {
      "text": "deliberative alignment",
      "url": "https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/",
      "resourceId": "b3f335edccfc5333",
      "resourceTitle": "OpenAI Preparedness Framework"
    },
    {
      "text": "Apollo Research",
      "url": "https://www.apolloresearch.ai/research/scheming-reasoning-evaluations",
      "resourceId": "91737bf431000298",
      "resourceTitle": "Frontier Models are Capable of In-Context Scheming"
    },
    {
      "text": "\"deliberative alignment\"",
      "url": "https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/",
      "resourceId": "b3f335edccfc5333",
      "resourceTitle": "OpenAI Preparedness Framework"
    },
    {
      "text": "six agentic evaluation scenarios",
      "url": "https://www.apolloresearch.ai/research/scheming-reasoning-evaluations",
      "resourceId": "91737bf431000298",
      "resourceTitle": "Frontier Models are Capable of In-Context Scheming"
    },
    {
      "text": "Claude 3.7 Sonnet research",
      "url": "https://www.apolloresearch.ai/blog/claude-sonnet-37-often-knows-when-its-in-alignment-evaluations/",
      "resourceId": "f5ef9e486e36fbee",
      "resourceTitle": "Apollo Research found"
    },
    {
      "text": "Scheming Evaluations",
      "url": "https://www.apolloresearch.ai/research/scheming-reasoning-evaluations",
      "resourceId": "91737bf431000298",
      "resourceTitle": "Frontier Models are Capable of In-Context Scheming"
    },
    {
      "text": "OpenAI Preparedness Framework",
      "url": "https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/",
      "resourceId": "b3f335edccfc5333",
      "resourceTitle": "OpenAI Preparedness Framework"
    },
    {
      "text": "Anthropic Responsible Scaling Policy",
      "url": "https://www.anthropic.com",
      "resourceId": "afe2508ac4caf5ee",
      "resourceTitle": "Anthropic - AI Safety Company Homepage"
    },
    {
      "text": "DeepMind Frontier Safety Framework",
      "url": "https://deepmind.google/blog/deepening-our-partnership-with-the-uk-ai-security-institute/",
      "resourceId": "d648a6e2afc00d15",
      "resourceTitle": "DeepMind: Deepening AI Safety Research with UK AISI"
    },
    {
      "text": "stated approach",
      "url": "https://www.apolloresearch.ai/",
      "resourceId": "329d8c2e2532be3d",
      "resourceTitle": "Apollo Research - AI Safety Evaluation Organization"
    },
    {
      "text": "OpenAI",
      "url": "https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/",
      "resourceId": "b3f335edccfc5333",
      "resourceTitle": "OpenAI Preparedness Framework"
    },
    {
      "text": "Anthropic",
      "url": "https://www.anthropic.com",
      "resourceId": "afe2508ac4caf5ee",
      "resourceTitle": "Anthropic - AI Safety Company Homepage"
    },
    {
      "text": "DeepMind",
      "url": "https://deepmind.google/blog/deepening-our-partnership-with-the-uk-ai-security-institute/",
      "resourceId": "d648a6e2afc00d15",
      "resourceTitle": "DeepMind: Deepening AI Safety Research with UK AISI"
    },
    {
      "text": "Open-source evaluation methodology",
      "url": "https://www.apolloresearch.ai/research/",
      "resourceId": "560dff85b3305858",
      "resourceTitle": "Apollo Research — Research Overview"
    },
    {
      "text": "Claude 3.7 research",
      "url": "https://www.apolloresearch.ai/blog/claude-sonnet-37-often-knows-when-its-in-alignment-evaluations/",
      "resourceId": "f5ef9e486e36fbee",
      "resourceTitle": "Apollo Research found"
    },
    {
      "text": "Apollo Research Publications",
      "url": "https://www.apolloresearch.ai/research/",
      "resourceId": "560dff85b3305858",
      "resourceTitle": "Apollo Research — Research Overview"
    },
    {
      "text": "EA Forum announcement (2022)",
      "url": "https://forum.effectivealtruism.org/posts/ysC6crBKhDBGZfob3/announcing-apollo-research",
      "resourceId": "0d60aa5bdfa567dd",
      "resourceTitle": "Announcing Apollo Research"
    },
    {
      "text": "OpenAI Blog",
      "url": "https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/",
      "resourceId": "b3f335edccfc5333",
      "resourceTitle": "OpenAI Preparedness Framework"
    },
    {
      "text": "DeepMind Blog",
      "url": "https://deepmind.google/blog/deepening-our-partnership-with-the-uk-ai-security-institute/",
      "resourceId": "d648a6e2afc00d15",
      "resourceTitle": "DeepMind: Deepening AI Safety Research with UK AISI"
    }
  ],
  "unconvertedLinkCount": 22,
  "convertedLinkCount": 0,
  "backlinkCount": 44,
  "hallucinationRisk": {
    "level": "high",
    "score": 75,
    "factors": [
      "biographical-claims",
      "no-citations"
    ]
  },
  "entityType": "organization",
  "redundancy": {
    "maxSimilarity": 18,
    "similarPages": [
      {
        "id": "metr",
        "title": "METR",
        "path": "/knowledge-base/organizations/metr/",
        "similarity": 18
      },
      {
        "id": "dangerous-cap-evals",
        "title": "Dangerous Capability Evaluations",
        "path": "/knowledge-base/responses/dangerous-cap-evals/",
        "similarity": 18
      },
      {
        "id": "evals",
        "title": "Evals & Red-teaming",
        "path": "/knowledge-base/responses/evals/",
        "similarity": 18
      },
      {
        "id": "sandbagging",
        "title": "AI Capability Sandbagging",
        "path": "/knowledge-base/risks/sandbagging/",
        "similarity": 18
      },
      {
        "id": "intervention-effectiveness-matrix",
        "title": "Intervention Effectiveness Matrix",
        "path": "/knowledge-base/models/intervention-effectiveness-matrix/",
        "similarity": 17
      }
    ]
  },
  "changeHistory": [
    {
      "date": "2026-02-18",
      "branch": "claude/audit-webpage-errors-X4jHg",
      "title": "Audit wiki pages for factual errors and hallucinations",
      "summary": "Systematic audit of ~20 wiki pages for factual errors, hallucinations, and inconsistencies. Found and fixed 25+ confirmed errors across 17 pages, including wrong dates, fabricated statistics, false attributions, missing major events, broken entity references, misattributed techniques, and internal inconsistencies."
    }
  ],
  "coverage": {
    "passing": 9,
    "total": 13,
    "targets": {
      "tables": 11,
      "diagrams": 1,
      "internalLinks": 23,
      "externalLinks": 14,
      "footnotes": 9,
      "references": 9
    },
    "actuals": {
      "tables": 13,
      "diagrams": 1,
      "internalLinks": 10,
      "externalLinks": 59,
      "footnotes": 0,
      "references": 9,
      "quotesWithQuotes": 0,
      "quotesTotal": 0,
      "accuracyChecked": 0,
      "accuracyTotal": 0
    },
    "items": {
      "summary": "green",
      "schedule": "green",
      "entity": "green",
      "editHistory": "green",
      "overview": "green",
      "tables": "green",
      "diagrams": "green",
      "internalLinks": "amber",
      "externalLinks": "green",
      "footnotes": "red",
      "references": "green",
      "quotes": "red",
      "accuracy": "red"
    },
    "editHistoryCount": 1,
    "ratingsString": "N:3.5 R:6 A:5.5 C:7"
  },
  "readerRank": 366,
  "researchRank": 10,
  "recommendedScore": 149.98
}
External Links
{
  "lesswrong": "https://www.lesswrong.com/tag/apollo-research-org"
}
Backlinks (44)
idtitletyperelationship
far-aiFAR AIorganization
metrMETRorganization
uk-aisiUK AI Safety Instituteorganization
us-aisiUS AI Safety Institute (now CAISI)organization
evalsAI Evaluationsresearch-arearesearch
red-teamingRed Teamingresearch-arearesearch
eval-saturationEval Saturation & The Evals Gapapproach
evaluation-awarenessEvaluation Awarenessapproach
scalable-eval-approachesScalable Eval Approachesapproach
scheming-detectionScheming & Deception Detectionapproach
capability-elicitationCapability Elicitationapproach
safety-casesAI Safety Casesapproach
alignment-evalsAlignment Evaluationsapproach
model-auditingThird-Party Model Auditingapproach
large-language-modelsLarge Language Modelsconcept
situational-awarenessSituational Awarenesscapability
accident-risksAI Accident Risk Cruxescrux
is-ai-xrisk-realIs AI Existential Risk Real?crux
__index__/knowledge-baseKnowledge Baseconcept
intervention-effectiveness-matrixIntervention Effectiveness Matrixanalysis
mesa-optimization-analysisMesa-Optimization Risk Analysisanalysis
risk-interaction-networkRisk Interaction Networkanalysis
ada-lovelace-instituteAda Lovelace Instituteorganization
apart-researchApart Researchorganization
arb-researchArb Researchorganization
goodfireGoodfireorganization
government-orgs-overviewGovernment AI Safety Organizations (Overview)concept
__index__/knowledge-base/organizationsOrganizationsconcept
matsMATS ML Alignment Theory Scholars programorganization
rethink-prioritiesRethink Prioritiesorganization
safety-orgs-overviewAI Safety Organizations (Overview)concept
the-foundation-layerThe Foundation Layerorganization
jaan-tallinnJaan Tallinnperson
coordination-techAI Governance Coordination Technologiesapproach
dangerous-cap-evalsDangerous Capability Evaluationsapproach
evaluationAI Evaluationapproach
technical-researchTechnical AI Safety Researchcrux
training-programsAI Safety Training Programsapproach
deceptive-alignmentDeceptive Alignmentrisk
enfeeblementAI-Induced Enfeeblementrisk
existential-riskExistential Risk from AIconcept
mesa-optimizationMesa-Optimizationrisk
sandbaggingAI Capability Sandbaggingrisk
schemingSchemingrisk
Longterm Wiki