Longterm Wiki

Misalignment Potential

misalignment-potential (E205)
← Back to pagePath: /ai-transition-model/factors-misalignment-potential-overview/
Page Metadata

No page data found for "misalignment-potential"

Entity Data
{
  "id": "misalignment-potential",
  "type": "ai-transition-model-factor",
  "title": "Misalignment Potential",
  "description": "The aggregate risk that AI systems pursue goals misaligned with human values—combining technical alignment challenges, interpretability gaps, and oversight limitations.",
  "tags": [
    "ai-transition-model",
    "factor",
    "technical",
    "alignment"
  ],
  "relatedEntries": [
    {
      "id": "existential-catastrophe",
      "type": "ai-transition-model-scenario",
      "relationship": "drives"
    },
    {
      "id": "ai-takeover",
      "type": "ai-transition-model-scenario",
      "relationship": "enables"
    },
    {
      "id": "alignment-robustness",
      "type": "ai-transition-model-parameter",
      "relationship": "composed-of"
    },
    {
      "id": "interpretability-coverage",
      "type": "ai-transition-model-parameter",
      "relationship": "composed-of"
    },
    {
      "id": "human-oversight-quality",
      "type": "ai-transition-model-parameter",
      "relationship": "composed-of"
    },
    {
      "id": "safety-capability-gap",
      "type": "ai-transition-model-parameter",
      "relationship": "composed-of"
    },
    {
      "id": "safety-culture-strength",
      "type": "ai-transition-model-parameter",
      "relationship": "composed-of"
    }
  ],
  "sources": [],
  "lastUpdated": "2026-01",
  "customFields": [
    {
      "label": "Model Role",
      "value": "Root Factor (AI System)"
    },
    {
      "label": "Key Parameters",
      "value": "Alignment Robustness, Interpretability Coverage, Human Oversight Quality"
    },
    {
      "label": "Primary Outcome",
      "value": "Existential Catastrophe"
    }
  ]
}
Canonical Facts (0)

No facts for this entity

External Links

No external links

Backlinks (3)
idtitletyperelationship
ai-capabilitiesAI Capabilitiesai-transition-model-factoramplifies
existential-catastropheExistential Catastropheai-transition-model-scenariodriver
ai-takeoverAI Takeoverai-transition-model-scenariodriven-by
Frontmatter
{
  "numericId": "E665",
  "title": "Misalignment Potential",
  "description": "Root factor measuring the likelihood AI systems pursue unintended goals. Primary driver of AI Takeover scenarios.",
  "sidebar": {
    "label": "Overview",
    "order": 0
  },
  "lastEdited": "2026-01-03",
  "subcategory": "factors-misalignment-potential"
}
Raw MDX Source
---
numericId: E665
title: Misalignment Potential
description: Root factor measuring the likelihood AI systems pursue unintended goals. Primary driver of AI Takeover scenarios.
sidebar:
  label: Overview
  order: 0
lastEdited: "2026-01-03"
subcategory: factors-misalignment-potential
---
import {DataInfoBox, FactorSubItemsList, FactorRelationshipDiagram, ImpactList, PageCauseEffectGraph, EntityLink} from '@components/wiki';

<DataInfoBox entityId="E205" />

Misalignment Potential measures the likelihood that AI systems will pursue goals other than what we intend. This aggregate combines the technical and organizational factors that determine whether advanced AI systems might behave in harmful ways despite our efforts.

**Primary outcome affected:** <EntityLink id="E130">Existential Catastrophe</EntityLink> ↑↑↑

When misalignment potential is high, catastrophic loss of control, accidents at scale, and goal divergence become more likely. Reducing this potential is the most direct lever for reducing existential and catastrophic AI risk.

---

## Component Parameters

<FactorRelationshipDiagram nodeId="misalignment-potential" direction="outgoing" showSubItems={true} />

<FactorSubItemsList factorId="misalignment-potential" />

---

## Internal Dynamics

These components interact:

- **Interpretability enables alignment verification**: We can only confirm alignment if we understand model internals
- **Safety culture sustains investment**: Without organizational commitment, safety research loses funding to capabilities
- **Oversight requires interpretability**: Human overseers need tools to understand what systems are doing
- **Gap closure requires all components**: No single factor is sufficient; safety capacity emerges from their combination

---

## How This Affects Scenarios

<ImpactList nodeId="misalignment-potential" direction="from" />