Skip to content
Longterm Wiki
Navigation
Updated 2025-12-26HistoryData
Page StatusContent
Edited 3 months ago1.9k words3 backlinksUpdated quarterlyOverdue by 10 days
65QualityGood85ImportanceHigh79.5ResearchHigh
Content8/13
SummaryScheduleEntityEdit historyOverview
Tables13/ ~7Diagrams2/ ~1Int. links28/ ~15Ext. links0/ ~9Footnotes0/ ~6References19/ ~6Quotes0Accuracy0RatingsN:5 R:6.5 A:7 C:7.5Backlinks3
Issues1
StaleLast edited 100 days ago - may need review
TODOs3
Complete 'Quantitative Analysis' section (8 placeholders)
Complete 'Strategic Importance' section
Complete 'Limitations' section (6 placeholders)

AI Proliferation Risk Model

Analysis

AI Proliferation Risk Model

Quantitative model of AI capability diffusion across 5 actor tiers, documenting compression from 24-36 months (2020) to 12-18 months (2024) with projections of 6-12 months by 2025-2026. Identifies compute governance (70-85% effectiveness) and pre-deployment gates (60-80%) as highest-leverage interventions before irreversible open-source proliferation, with specific actor-level risk calculations showing 5,000 expected misuse events at Tier 4-5 proliferation.

Model TypeDiffusion Analysis
Target FactorAI Proliferation
Related
Risks
AI ProliferationAI Development Racing Dynamics
1.9k words · 3 backlinks

Overview

This model analyzes the diffusion of AI capabilities from frontier laboratories to progressively broader populations of actors. It examines proliferation mechanisms, control points, and the relationship between diffusion speed and risk accumulation. The central question: How fast do dangerous AI capabilities spread from frontier labs to millions of users, and which intervention points offer meaningful leverage?

Key findings show proliferation follows predictable tier-based patterns, but time constants are compressing dramatically. Capabilities that took 24-36 months to diffuse from Tier 1 (frontier labs) to Tier 4 (open source) in 2020 now spread in 12-18 months. Projections suggest 6-12 month cycles by 2025-2026, fundamentally changing governance calculus.

The model identifies an "irreversibility threshold" where proliferation cannot be reversed once capabilities reach open source. This threshold is crossed earlier than commonly appreciated—often before policymakers recognize capabilities as dangerous. High-leverage interventions must occur pre-proliferation; post-proliferation controls offer diminishing returns as diffusion accelerates.

Risk Assessment Framework

Risk DimensionCurrent Assessment2025-2026 ProjectionEvidenceTrend
Diffusion SpeedHighVery High50% reduction in proliferation timelines since 2020Accelerating
Control WindowMediumLow12-18 month average control periodsShrinking
Actor ProliferationHighVery HighTier 4 access growing exponentiallyExpanding
Irreversibility RiskHighExtremeMultiple capabilities already irreversibly proliferatedIncreasing

Proliferation Tier Analysis

Actor Tier Classification

The proliferation cascade operates through five distinct actor tiers, each with different access mechanisms, resource requirements, and risk profiles.

TierActor TypeCountAccess MechanismDiffusion TimeControl Feasibility
1Frontier Labs5-10Original development-High (concentrated)
2Major Tech50-100API/Partnerships6-18 monthsMedium-High
3Well-Resourced Orgs1K-10KFine-tuning/Replication12-24 monthsMedium
4Open SourceMillionsPublic weights18-36 monthsVery Low
5IndividualsBillionsConsumer apps24-48 monthsNone
Diagram (loading…)
flowchart TD
  T1[Tier 1: Frontier Labs<br/>OpenAI, Anthropic, Google, etc.<br/>~10 actors] --> T2[Tier 2: Major Tech<br/>Microsoft, Amazon, Meta<br/>~100 actors]
  T2 --> T3[Tier 3: Well-Resourced Orgs<br/>Large corps, governments<br/>~10,000 actors]
  T3 --> T4[Tier 4: Open Source<br/>Public model weights<br/>Millions of actors]
  T4 --> T5[Tier 5: Consumer Access<br/>Apps and services<br/>Billions of users]

  style T1 fill:#ff9999
  style T2 fill:#ffcc99
  style T3 fill:#fff4cc
  style T4 fill:#99ff99
  style T5 fill:#99ccff

Historical Diffusion Data

Analysis of actual proliferation timelines reveals accelerating diffusion across multiple capability domains:

CapabilityTier 1 DateTier 4 DateTotal TimeKey Events
GPT-3 levelMay 2020Jul 202226 monthsOpenAI → HuggingFace release
DALL-E levelJan 2021Aug 202219 monthsOpenAI → Stable Diffusion
GPT-4 levelMar 2023Jan 202522 monthsOpenAI → DeepSeek-R1
Code generationAug 2021Dec 202216 monthsCodex → StarCoder
Protein foldingNov 2020Jul 20218 monthsAlphaFold → ColabFold

Mathematical Model

Core Risk Equation

Total proliferation risk combines actor count, capability level, and misuse probability:

Rtotal(t)=i=15Ni(t)Ci(t)Pmisuse,iR_{\text{total}}(t) = \sum_{i=1}^{5} N_i(t) \cdot C_i(t) \cdot P_{\text{misuse},i}

Where:

  • Ni(t)N_i(t) = Number of actors in tier ii with access at time tt
  • Ci(t)C_i(t) = Capability level accessible to tier ii at time tt
  • Pmisuse,iP_{\text{misuse},i} = Per-actor misuse probability for tier ii

Diffusion Dynamics

Each tier transition follows modified logistic growth with accelerating rates:

Ni(t)=Ni,max1+eki(tt0,i)N_i(t) = \frac{N_{i,\max}}{1 + e^{-k_i(t - t_{0,i})}}

The acceleration factor captures increasing diffusion speed:

ki(t)=ki,0eαtk_i(t) = k_{i,0} \cdot e^{\alpha t}

With α0.15\alpha \approx 0.15 per year, implying diffusion rates double every ~5 years. This matches observed compression from 24-36 month cycles (2020) to 12-18 months (2024).

Control Point Effectiveness

High-Leverage Interventions

Control PointEffectivenessDurabilityImplementation DifficultyCurrent Status
Compute governance70-85%5-15 yearsHighPartial (US export controls)
Pre-deployment gates60-80%UnknownVery HighVoluntary only
Weight security50-70%FragileMediumIndustry standard emerging
International coordination40-70%MediumVery HighEarly stages

Medium-Leverage Interventions

Control PointCurrent EffectivenessKey LimitationExample Implementation
API controls40-60%Continuous bypass developmentOpenAI usage policies
Capability evaluation50-70%May miss emergent capabilitiesARC Evals
Publication norms30-50%Competitive pressure to publishFHI publication guidelines
Talent restrictions20-40%Limited in free societiesCFIUS review process

Proliferation Scenarios

2025-2030 Trajectory Analysis

ScenarioProbabilityTier 1-4 TimeKey DriversRisk Level
Accelerating openness35%3-6 monthsOpen-source ideology, regulation failureVery High
Current trajectory40%6-12 monthsMixed open/closed, partial regulationHigh
Managed deceleration15%12-24 monthsInternational coordination, major incidentMedium
Effective control10%24+ monthsStrong compute governance, industry agreementLow-Medium

Threshold Analysis

Critical proliferation thresholds mark qualitative shifts in control feasibility:

ThresholdDescriptionControl StatusResponse Window
ContainedTier 1-2 onlyControl possibleMonths
OrganizationalTier 3 accessState/criminal access likelyWeeks
IndividualTier 4/5 accessMonitoring overwhelmedDays
IrreversibleOpen source + common knowledgeControl impossibleN/A
Diagram (loading…)
graph LR
  A[Contained<br/>Tier 1-2] --> B[Organizational<br/>Tier 3]
  B --> C[Individual<br/>Tier 4-5]
  C --> D[Irreversible<br/>Open Source]
  
  A --> A1[Control possible<br/>Months to act]
  B --> B1[State actor access<br/>Weeks to act]
  C --> C1[Mass access<br/>Days to act]
  D --> D1[No control<br/>Focus on defense]
  
  style A fill:#ccffcc
  style B fill:#fff4cc
  style C fill:#ffcc99
  style D fill:#ff9999

Risk by Actor Type

Misuse Probability Assessment

Different actor types present distinct risk profiles based on capability access and motivation:

Actor TypeEstimated CountCapability AccessP(Access)P(Misuse|Access)Risk Weight
Hostile state programs5-15Frontier0.950.15-0.40Very High
Major criminal orgs50-200Near-frontier0.70-0.850.30-0.60High
Terrorist groups100-500Moderate0.40-0.700.50-0.80High
Ideological groups1K-10KModerate0.50-0.800.10-0.30Medium
Malicious individuals10K-100KBasic-Moderate0.60-0.900.01-0.10Medium (scale)

Expected Misuse Events

Even low individual misuse probabilities become concerning at scale:

E[misuse events]=iNiP(access)iP(misuseaccess)iE[\text{misuse events}] = \sum_i N_i \cdot P(\text{access})_i \cdot P(\text{misuse}|\text{access})_i

For Tier 4-5 proliferation with 100,000 capable actors and 5% misuse probability, expected annual misuse events: 5,000.

Current State & Trajectory

Recent Developments

The proliferation landscape has shifted dramatically since 2023:

2023 Developments:

  • LLaMA leak demonstrated fragility of controlled releases
  • LLaMA 2 open release established new norm for frontier model sharing
  • U.S. export controls on advanced semiconductors implemented

2024-2025 Developments:

  • DeepSeek R1 release achieved GPT-4 level performance with open weights
  • Qwen 2.5 and Mistral continued aggressive open-source strategy
  • Chinese labs increasingly releasing frontier capabilities openly

2025-2030 Projections

Accelerating Factors:

  • Algorithmic efficiency reducing compute requirements ~2x annually
  • China developing domestic chip capabilities to circumvent controls
  • Open-source ideology gaining ground in AI community
  • Economic incentives for ecosystem building through open models

Decelerating Factors:

  • Growing awareness of proliferation risks among frontier labs
  • Potential regulatory intervention following AI incidents
  • Voluntary industry agreements on responsible disclosure
  • Technical barriers to replicating frontier training runs

Critical Unknown Parameters

UncertaintyImpact on ModelCurrent StateResolution Timeline
Chinese chip developmentVery High2-3 generations behind3-7 years
Algorithmic efficiency gainsHigh≈2x annual improvementOngoing
Open vs closed normsVery HighTrending toward open1-3 years
Regulatory interventionHighMinimal but increasing2-5 years
Major AI incidentVery HighNone yetUnpredictable

Model Sensitivity Analysis

The model is most sensitive to three parameters:

Diffusion Rate Acceleration (α): 10% change in α yields 25-40% change in risk estimates over 5-year horizon. This parameter depends heavily on continued algorithmic progress and open-source community growth.

Tier 4/5 Misuse Probability: Uncertainty ranges from 1-15% create order-of-magnitude differences in expected incidents. Better empirical data on malicious actor populations is critical.

Compute Control Durability: Estimates ranging from 3-15 years until circumvention dramatically affect intervention value. China's semiconductor progress is the key uncertainty.

Policy Implications

Immediate Actions (0-18 months)

Strengthen Compute Governance:

  • Expand semiconductor export controls to cover training and inference chips
  • Implement cloud provider monitoring for large training runs
  • Establish international coordination on chip supply chain security

Establish Evaluation Frameworks:

  • Define dangerous capability thresholds with measurable criteria
  • Create mandatory pre-deployment evaluation requirements
  • Build verification infrastructure for model capabilities

Medium-Term Priorities (18 months-5 years)

International Coordination:

  • Negotiate binding agreements on proliferation control
  • Establish verification mechanisms for training run detection
  • Create sanctions framework for violating proliferation norms

Industry Standards:

  • Implement weight security requirements for frontier models
  • Establish differential access policies based on actor verification
  • Create liability frameworks for irresponsible proliferation

Long-Term Structural Changes (5+ years)

Governance Architecture:

  • Build adaptive regulatory systems that evolve with technology
  • Establish international AI safety organization with enforcement powers
  • Create sustainable funding for proliferation monitoring infrastructure

Research Priorities:

  • Develop better offensive-defensive balance understanding
  • Create empirical measurement systems for proliferation tracking
  • Build tools for post-proliferation risk mitigation

Research Gaps

Several critical uncertainties limit model precision and policy effectiveness:

Empirical Proliferation Tracking: Systematic measurement of capability diffusion timelines across domains remains limited. Most analysis relies on high-profile case studies rather than comprehensive data collection.

Reverse Engineering Difficulty: Time and resources required to replicate capabilities from limited information varies dramatically across capability types. Better understanding could inform targeted protection strategies.

Actor Intent Modeling: Current misuse probability estimates rely on theoretical analysis rather than empirical study of malicious actor populations and motivations.

Control Mechanism Effectiveness: Rigorous testing of governance interventions is lacking. Most effectiveness estimates derive from analogies to other domains rather than AI-specific validation.

Defensive Capability Development: The model focuses on capability proliferation while ignoring parallel development of defensive tools that could partially offset risks.

Sources & Resources

Academic Research

SourceFocusKey FindingsLink
Heim et al. (2023)Compute governanceExport controls 60-80% effective short-termCSET Georgetown
Anderljung et al. (2023)Model securityWeight protection reduces proliferation 50-70%arXiv
Shavit et al. (2023)Capability evaluationCurrent evals miss 30-50% of dangerous capabilitiesarXiv

Policy Documents

DocumentOrganizationKey RecommendationsYear
AI Executive OrderWhite HouseMandatory reporting, evaluation requirements2023
UK AI Safety SummitUK GovernmentInternational coordination framework2023
EU AI ActEuropean UnionRisk-based regulatory approach2024

Technical Resources

ResourceTypeDescriptionAccess
Model weight leaderboardsDataOpen-source capability trackingHuggingFace
Compute trend analysisAnalysisTraining cost trends over timeEpoch AI
Export control guidancePolicyCurrent semiconductor restrictionsBIS Commerce
ModelFocusRelationship
Racing DynamicsCompetitive pressuresExplains drivers of open release
Multipolar TrapCoordination failuresModels governance challenges
Winner-Take-AllMarket structureAlternative to proliferation scenario

References

METR (formerly ARC Evals) conducts research and evaluations to assess the capabilities and risks of frontier AI systems, focusing on autonomous capabilities, AI R&D acceleration potential, and evaluation integrity. They are notable for developing the 'time horizon' metric measuring how long AI agents can complete tasks, and for conducting pre-deployment evaluations for major AI labs.

2EU AI Act – Official Resource Hubartificialintelligenceact.eu

The EU AI Act is the world's first comprehensive legal framework for artificial intelligence, establishing a risk-based classification system for AI applications. It imposes varying obligations on developers and deployers depending on the risk level of their AI systems, from minimal-risk to unacceptable-risk categories. The act sets precedents for global AI governance and compliance requirements.

3Industry standard emergingarXiv·D. Estevez-Moya, E. Estevez-Rams & H. Kantz·2023·Paper

This paper investigates coupled non-linear oscillators with local (nearest-neighbor) coupling using phase approximation under weak coupling assumptions. The study focuses on characterizing the 'needle region' in parameter space for Adler-type oscillators, where computation enhancement at the edge of chaos has been previously reported. The authors identify diverse dynamical behaviors within this region, including wave-like spatiotemporal patterns and heterogeneous dynamics revealed through entropic measures. The research demonstrates that spatial correlations emerge locally at the onset of chaos, with coherent oscillator clusters separated by disordered boundaries.

★★★☆☆

The Open LLM Leaderboard is a HuggingFace-hosted benchmarking platform that compares open-source large language models across standardized evaluations in a transparent and reproducible manner. It allows researchers and practitioners to filter, search, and rank models by performance metrics, providing a community reference for tracking AI capabilities progress. The leaderboard has since been archived, reflecting the rapid pace of LLM development.

Executive Order 14110, signed by President Biden on October 30, 2023, established comprehensive federal directives for AI safety, security, and governance in the United States. It required safety testing and reporting for frontier AI models, directed agencies to address AI risks across sectors including national security and civil rights, and aimed to position the US as a global leader in responsible AI development. The page content is currently unavailable, but the order is a landmark AI governance document.

★★★★☆

This is the official UK government Chair's statement from the inaugural AI Safety Summit held at Bletchley Park on 1-2 November 2023. The summit brought together international stakeholders to identify next steps for the safe development of frontier AI. It represents a landmark moment in international AI governance coordination, resulting in the Bletchley Declaration.

★★★★☆

Meta's Llama is a family of open-source large language models including Llama 3 and Llama 4 variants, offering multimodal capabilities, extended context windows, and various model sizes for deployment across diverse use cases. The latest Llama 4 models feature native multimodality with early fusion architecture, supporting up to 10M token context windows. Models are freely downloadable and fine-tunable, positioning Llama as a major open-source alternative to proprietary AI systems.

★★★★☆

Qwen 2.5 is Alibaba's latest series of large language models, representing significant capability advances across language understanding, coding, mathematics, and multimodal tasks. The series includes models of various sizes designed for both research and commercial deployment. It represents a major frontier model release from a leading Chinese AI lab.

9Shavit et al. (2023)arXiv·Lewis Ho et al.·2023·Paper

This paper proposes four complementary international institutional models for governing advanced AI: a Commission on Frontier AI for expert consensus, an Advanced AI Governance Organization for safety standards, a Frontier AI Collaborative for equitable access, and an AI Safety Project for coordinated research. The framework draws on precedents from existing international organizations and aims to balance AI's benefits against global risks from powerful systems.

★★★☆☆
10FHI publication guidelinesFuture of Humanity Institute

This resource from Oxford's Future of Humanity Institute (FHI) and Centre for the Governance of AI outlines recommended publication norms for machine learning researchers, addressing how and when to publish potentially dangerous AI capabilities research. It proposes frameworks for assessing dual-use risks and considering staged or restricted disclosure. The guidelines aim to balance scientific openness with responsible stewardship of potentially harmful information.

★★★★☆

Mistral AI is a European AI company developing frontier large language models, assistants, and AI services. They offer both open-weight models and commercial API products, positioning themselves as a competitive alternative to US-based AI labs. Their work is relevant to AI safety discussions around model diffusion, open-source risks, and governance.

DeepSeek-R1 is an open-source large language model from DeepSeek-AI that achieves strong reasoning capabilities through reinforcement learning, reportedly matching or approaching OpenAI's o1 performance on reasoning benchmarks. The release includes model weights, technical details, and distilled smaller variants, representing a significant open-source milestone in frontier reasoning AI. Its release demonstrated that high-capability reasoning models can be developed at lower cost and made openly available.

★★★☆☆
13Partial (US export controls)Bureau of Industry and Security·Government

This U.S. Bureau of Industry and Security (BIS) page provides regulatory guidance on export controls relevant to semiconductors and advanced technologies, administered in the interest of national security. It serves as a reference point for understanding how U.S. policy restricts the diffusion of critical technologies, including AI-relevant compute hardware, to adversarial or controlled entities.

★★★★☆

Meta's LLaMA large language model, initially released only to approved researchers, was leaked publicly on 4chan and spread across the internet. The incident raised significant concerns about the ability to control access to powerful AI models once released, even in restricted form, and highlighted tensions between open research access and preventing misuse.

The White House announced voluntary commitments from major AI companies (including Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI) to manage AI risks, covering safety testing, information sharing, and transparency measures. These non-binding pledges represent the Biden administration's early governance approach before formal regulation, focusing on watermarking AI-generated content, red-teaming, and vulnerability reporting. Critics and analysts noted the limited enforceability of voluntary frameworks.

★★★★☆

This Epoch AI analysis tracks historical trends in the monetary cost of training machine learning systems, examining how dollar costs have evolved alongside compute scaling. It provides empirical data on training cost trajectories to inform forecasts about future AI development economics and accessibility.

★★★★☆
17CFIUS review processhome.treasury.gov·Government

CFIUS is a U.S. interagency committee that reviews foreign investment transactions and real estate purchases for national security implications under the Defense Production Act of 1950. The Treasury Department is developing a Known Investor Program to streamline review for vetted investors from allied nations. A public Request for Information is open through March 2026 to gather feedback on process improvements.

OpenAI's official usage policies outline the rules and restrictions governing how its AI models and APIs may be used, including prohibited use cases and safety guidelines. The policies cover disallowed activities such as generating disinformation, facilitating influence operations, creating harmful content, and misusing AI for deceptive or dangerous purposes. These policies serve as a practical governance framework for responsible deployment of OpenAI's systems.

★★★★☆

This CSET report examines the geopolitical dimensions of AI chip supply chains, analyzing how control over semiconductor manufacturing and export creates strategic leverage in AI competition between nations. It explores how chip restrictions and export controls shape the global diffusion of AI capabilities and influence international AI development trajectories.

★★★★☆

Related Wiki Pages

Top Related Pages

Risks

AI-Induced IrreversibilityAI Winner-Take-All DynamicsAI-Powered Investigation Risks

Analysis

Irreversibility Threshold ModelRacing Dynamics Impact ModelLAWS Proliferation ModelMultipolar Trap Dynamics ModelFlash Dynamics Threshold ModelExpertise Atrophy Progression Model