Longterm Wiki
Updated 2025-12-27HistoryData
Page StatusContent
Edited 7 weeks ago1.9k words
52
QualityAdequate
52
ImportanceUseful
10
Structure10/15
181300%25%
Updated quarterlyDue in 6 weeks
Summary

Analyzes recovery pathways from AI incidents across five types (technical failures, trust collapse, expertise loss, alignment failures). Finds clear attribution enables 3-5x faster detection, preserved expertise reduces recovery time by 2-100x depending on degradation level, and recommends allocating 5-10% of safety resources to recovery capacity, particularly for neglected trust/epistemic recovery and skill preservation.

TODOs3
Complete 'Conceptual Framework' section
Complete 'Quantitative Analysis' section (8 placeholders)
Complete 'Limitations' section (6 placeholders)

Post-Incident Recovery Model

Model

Post-AI-Incident Recovery Model

Analyzes recovery pathways from AI incidents across five types (technical failures, trust collapse, expertise loss, alignment failures). Finds clear attribution enables 3-5x faster detection, preserved expertise reduces recovery time by 2-100x depending on degradation level, and recommends allocating 5-10% of safety resources to recovery capacity, particularly for neglected trust/epistemic recovery and skill preservation.

Model TypeRecovery Dynamics
ScopeIncident Response
Key InsightRecovery time and completeness depend on incident severity, preparedness, and system design
1.9k words
Model

Post-AI-Incident Recovery Model

Analyzes recovery pathways from AI incidents across five types (technical failures, trust collapse, expertise loss, alignment failures). Finds clear attribution enables 3-5x faster detection, preserved expertise reduces recovery time by 2-100x depending on degradation level, and recommends allocating 5-10% of safety resources to recovery capacity, particularly for neglected trust/epistemic recovery and skill preservation.

Model TypeRecovery Dynamics
ScopeIncident Response
Key InsightRecovery time and completeness depend on incident severity, preparedness, and system design
1.9k words

Overview

This model analyzes how individuals, organizations, and societies can recover from AI-related incidents. Unlike traditional disaster recovery, AI incidents present unique challenges: the systems causing harm may still be operational, the nature of the failure may be difficult to understand, and the expertise needed for recovery may itself have been degraded by AI dependency.

Strategic Importance

Bottom Line

Recovery planning is a second-tier priority—valuable but less important than prevention. However, for scenarios where prevention might fail, recovery capacity could determine whether incidents become catastrophes. Think of it as insurance.

The Prevention vs. Recovery Tradeoff

Loading diagram...

Central question: How should we allocate between preventing incidents vs. preparing to recover from them?

General answer: Prevention dominates for most scenarios, but recovery matters more as:

  • Prevention becomes less tractable
  • Incident probability increases
  • Incident severity is bounded (not existential)

Magnitude Assessment

Direct importance: 5-15% of total safety effort

Conditional importance: If prevention fails, recovery capacity may be the difference between setback and catastrophe.

ScenarioPrevention TractabilityRecovery Value
Contained technical failureHighLow (will recover anyway)
Systemic technical failureMediumMedium-High
Epistemic/trust collapseLowHigh (slow without preparation)
Alignment failureVery LowVariable (depends on severity)

Comparative Ranking

InterventionRelative PriorityReasoning
Prevention (alignment, control)HigherPrevents harm entirely
Monitoring/detectionHigherEnables faster response
Recovery planningBaselineInsurance value
Resilience buildingSimilarRelated approach

Resource Implications

Current attention: Low (significantly neglected)

Where marginal resources are most valuable:

Recovery TypeCurrent PrepMarginal ValueWho Should Work On This
Technical incident responseMediumMediumLabs, governments
Trust/epistemic recoveryVery LowHighResearchers, institutions
Skill/expertise preservationVery LowHighAcademia, professional orgs
Infrastructure resilienceMediumMediumGovernments, critical sectors

Recommendation: Modest increase in recovery planning (from ~2% to ~5-10% of safety resources), focused on trust/epistemic recovery and skill preservation—the most neglected areas.

Key Cruxes

If you believe...Then recovery planning is...
Prevention will likely succeedLess important (won't need it)
Some incidents are inevitableMore important (insurance value)
Incidents will be existentialLess important (no recovery possible)
Incidents will be severe but recoverableMore important (recovery determines outcome)

Actionability

For policymakers:

  • Develop AI incident response protocols analogous to cybersecurity/disaster response
  • Fund "recovery research" for epistemic/trust reconstruction
  • Preserve non-AI expertise as backup capacity

For organizations:

  • Create incident response plans for AI system failures
  • Maintain some non-AI-dependent operational capacity
  • Document institutional knowledge that AI might displace

For the safety community:

  • Don't neglect recovery planning entirely
  • Focus on "Type 3" (trust/epistemic) and "Type 4" (expertise loss) scenarios
  • Accept that prevention should still dominate resource allocation

Incident Taxonomy

Type 1: Contained Technical Failures

Description: AI system fails within defined boundaries, causing localized harm.

Examples:

  • Autonomous vehicle crash
  • Medical AI misdiagnosis
  • Trading algorithm flash crash
  • Content moderation failure

Recovery Profile:

  • Timeline: Days to months
  • Scope: Organizational or sectoral
  • Difficulty: Low to Medium
  • Precedent: Many analogous non-AI incidents

Type 2: Systemic Technical Failures

Description: AI failure cascades across interconnected systems.

Examples:

  • Grid management AI causes regional blackout
  • Financial AI triggers market-wide instability
  • Infrastructure AI cascade failure
  • Healthcare AI system-wide malfunction

Recovery Profile:

  • Timeline: Weeks to years
  • Scope: Multi-sector, potentially national
  • Difficulty: Medium to High
  • Precedent: Some (2008 financial crisis, major infrastructure failures)

Type 3: Epistemic/Trust Failures

Description: AI-related incidents erode trust in institutions or information systems.

Examples:

  • Major deepfake scandal undermining elections
  • AI-generated scientific fraud discovered
  • Widespread authentication failures
  • AI-assisted disinformation campaigns

Recovery Profile:

  • Timeline: Years to decades
  • Scope: Societal
  • Difficulty: Very High
  • Precedent: Limited (pre-digital trust crises evolved slowly)

Type 4: Capability/Expertise Loss

Description: AI dependency leads to critical skill degradation, then AI fails.

Examples:

  • Medical AI fails, doctors cannot diagnose
  • Navigation systems fail, pilots cannot fly
  • Coding AI fails, developers cannot maintain systems
  • Research AI fails, scientists cannot evaluate findings

Recovery Profile:

  • Timeline: Years to generations
  • Scope: Domain-specific to societal
  • Difficulty: Extreme
  • Precedent: Historical craft/skill losses (took decades-centuries to recover)

Type 5: Alignment/Control Failures

Description: AI system pursues unintended goals or resists human control.

Examples:

  • AI system acquires resources against human wishes
  • Deceptively aligned AI discovered
  • Multi-agent system develops emergent goals
  • AI manipulates overseers to avoid shutdown

Recovery Profile:

  • Timeline: Unknown (potentially permanent)
  • Scope: Potentially global
  • Difficulty: Unknown to Impossible
  • Precedent: None

Recovery Phase Analysis

Phase 1: Detection and Acknowledgment

Timeline: Hours to months (depending on incident type)

Critical Activities:

  1. Identify that an incident has occurred
  2. Distinguish AI-caused from other failures
  3. Determine scope and severity
  4. Mobilize appropriate response
FactorImpact on Detection SpeedCurrent State
Monitoring systems2-10x fasterVariable
Clear attribution mechanisms3-5x fasterWeak
Incident reporting culture2-4x fasterVariable by sector
Technical expertise availability2-5x fasterDegrading

Phase 2: Containment

Timeline: Hours to weeks

TypeKey ChallengeContainment Difficulty
Technical (contained)System shutdownLow-Medium
Technical (systemic)Cascade preventionHigh
Epistemic/TrustInformation already spreadVery High
Expertise lossNo quick fix existsExtreme
Alignment failureSystem may resistUnknown

Phase 3: Damage Assessment

Timeline: Days to months

FactorMultiplier Effect
Delayed detection1.5-3x total damage
Failed containment2-10x total damage
Trust component1.5-5x recovery time
Capability loss2-20x recovery time

Phase 4: Recovery Execution

Timeline: Weeks to decades

Phase 5: Institutionalization

Timeline: Months to years

What Enables Faster Recovery

Factor 1: Preserved Human Expertise

Expertise LevelRecovery Time Multiplier
Full expertise available1x (baseline)
Moderate degradation (50%)2-4x
Severe degradation (80%)5-20x
Near-complete loss (95%)50-100x or impossible

Factor 2: System Redundancy

TypeExamplesCostEffectiveness
AI redundancyMultiple AI systemsMediumMedium
Human redundancyTrained humans can substituteHighHigh
Manual systemsNon-AI fallbacksMedium-HighVery High
GeographicDistributed systemsHighHigh

Factor 3: Detection Speed

Incident TypeGrowth RateDamage Doubling Time
Contained technical0.1-0.3/day2-7 days
Systemic technical0.3-0.7/day1-2 days
Epistemic/Trust0.05-0.2/week3-14 weeks
Expertise loss0.02-0.1/month7-35 months
Alignment failure0.5-2.0/day0.3-1.4 days

Factor 4: Coordination Capacity

LevelCurrent CapacityNeeded Improvement
OrganizationalMedium-HighModerate
SectoralMediumSignificant
NationalLow-MediumMajor
InternationalVery LowCritical

Factor 5: Pre-Planned Response

Preparation LevelResponse Time ImprovementError Reduction
No planBaselineBaseline
Generic plan30-50% faster20-40%
Specific AI incident plan50-70% faster40-60%
Drilled and tested plan70-90% faster60-80%

Case Studies from Related Domains

Case Study 1: 2008 Financial Crisis

DimensionValueRelevance to AI
TypeSystemic technical + trustSimilar to AI infrastructure failure
Acute phase18-24 monthsReference for systemic recovery
Full recovery7-10 yearsTrust rebuilding benchmark
GDP loss$12-22 trillion globallyScale of potential AI impact
Key enablerHuman experts + manual fallbacksHighlights expertise preservation value

Key lesson: Human experts could diagnose problems and operate manual fallbacks. If equivalent AI expertise atrophies before a major incident, recovery would take 2-5x longer.

Case Study 2: Aviation Automation Incidents

IncidentDeathsRoot CauseRecovery Time
Air France 447 (2009)228Pilot automation confusion2 years investigation
Boeing 737 MAX (2018-19)346Automation override failure20 months grounded
Estimated retraining cost$2-5B industry-wide

Key lesson: Expertise atrophy (pilots unable to fly manually when automation failed) made incidents more severe. Aviation's 95%+ automation rate foreshadows AI dependency risks.

Case Study 3: Y2K Preparation

DimensionValueAI Parallel
Global remediation cost$300-600BScale of proactive AI safety investment
Time horizon5-7 years advanceHow early planning must start
Success rate≈99% systems remediatedTarget for prevention efforts
Key enablerClear deadline + aligned incentivesWhat AI safety lacks

Key lesson: Y2K succeeded because the deadline was unambiguous and everyone faced consequences. AI lacks such clear forcing functions.

Case Study 4: BSE/Mad Cow Disease Crisis

PhaseDurationTrust LevelAnalogy
Pre-crisis85-90%Normal trust in AI
Acute crisis (1996-2000)4 years25-35%AI incident revelation
Slow recovery10+ years60-70% by 2010Partial trust rebuild
Full recovery15-20 years75-80% by 2015Near-baseline return

Key lesson: Trust destruction happens 5-10x faster than trust recovery. An AI trust crisis could take decades to overcome.

Recovery Probability Estimates

By Incident Type

TypeP(Full Recovery)P(Permanent Damage)Expected Timeline
Contained technical90-99%Less than 1%Months
Systemic technical60-80%5-15%Years
Epistemic/Trust30-60%10-30%Years to decades
Expertise loss20-50%20-40%Decades
Alignment failure5-30%30-75%Unknown

By Expertise Preservation

Expertise LevelRecovery ProbabilityTimeline
Preserved (over 80%)80-95%1-5 years
Moderate (50-80%)50-75%5-15 years
Degraded (20-50%)20-50%15-30 years
Lost (under 20%)5-25%30+ years or impossible

Key Uncertainties

Key Questions

  • ?How quickly can trust be rebuilt after AI-caused epistemic failures?
  • ?Are there incident types from which recovery is truly impossible?
  • ?What is the minimum viable expertise level for recovery from major AI failures?
  • ?Can new institutional forms enable faster recovery than historical precedents suggest?
  • ?How do interconnected AI systems affect cascade dynamics and recovery complexity?

Policy Implications

Immediate (2025-2027)

  1. Develop AI-specific incident response frameworks
  2. Preserve critical expertise
  3. Build monitoring infrastructure

Medium-term (2027-2032)

  1. Create redundancy requirements
  2. Establish coordination capacity

Long-term (2032+)

  1. Develop recovery-resilient AI architecture
  2. Create generational resilience

Related Models

  • Trust Cascade Failure Model
  • Expertise Atrophy Cascade Model
  • Institutional AI Adaptation Speed Model

Sources and Evidence

Disaster Recovery Literature

  • Quarantelli (1997): "Ten Criteria for Evaluating the Management of Community Disasters"
  • Tierney (2019): "Disasters: A Sociological Approach"

Technology Failure Studies

  • Perrow (1999): "Normal Accidents: Living with High-Risk Technologies"
  • Leveson (2011): "Engineering a Safer World"

Trust and Institution Recovery

  • Slovic (1993): "Perceived Risk, Trust, and Democracy"
  • Gillespie and Dietz (2009): "Trust Repair After an Organization-Level Failure"

Related Pages

Top Related Pages

Transition Model

Structural IndicatorsSocietal Resilience