Skip to content
Longterm Wiki
Navigation
Updated 2025-12-26HistoryData
Page StatusContent
Edited 3 months ago2.0k words2 backlinksUpdated quarterlyOverdue by 10 days
66QualityGood54ImportanceUseful78.5ResearchHigh
Content7/13
SummaryScheduleEntityEdit historyOverview
Tables26/ ~8Diagrams0/ ~1Int. links64/ ~16Ext. links0/ ~10Footnotes0/ ~6References23/ ~6Quotes0Accuracy0RatingsN:6 R:6.5 A:7 C:8Backlinks2
Issues1
StaleLast edited 100 days ago - may need review
TODOs2
Complete 'Conceptual Framework' section
Complete 'Quantitative Analysis' section (8 placeholders)

Risk Activation Timeline Model

Analysis

AI Risk Activation Timeline Model

Comprehensive framework mapping AI risk activation windows with specific probability assessments: current risks already active (disinformation 95%+, spear phishing active), near-term critical window 2025-2027 (bioweapons 50% by 2027, cyberweapons 75%), long-term existential risks 2030-2050+ (ASI misalignment 15% by 2030). Recommends $3-5B annual investment in Tier 1 interventions with specific allocations: $200-400M bioweapons screening, $300-600M interpretability, $500M-1B cyber-defense.

Model TypeTimeline Projection
ScopeCross-cutting (all risk categories)
Key InsightRisks activate at different times based on capability thresholds
Related
Analyses
AI Capability Threshold ModelAI Risk Warning Signs ModelAI-Bioweapons Timeline Model
2k words · 2 backlinks

Overview

Different AI risks don't all "turn on" at the same time - they activate based on capability thresholds, deployment contexts, and barrier erosion. This model systematically maps when various AI risks become critical, enabling strategic resource allocation and intervention timing.

The model reveals three critical insights: many serious risks are already active with current systems, the next 2-3 years represent a critical activation window for multiple high-impact risks, and long-term existential risks require foundational research investment now despite uncertain timelines.

Understanding activation timing enables prioritizing immediate interventions for active risks, preparing defenses for near-term thresholds, and building foundational capacity for long-term challenges before crisis mode sets in.

Risk Assessment Overview

Risk CategoryTimelineSeverity RangeCurrent StatusIntervention Window
Current Active2020-2024Medium-HighMultiple risks activeClosing rapidly
Near-term Critical2025-2027High-ExtremeApproaching thresholdsOpen but narrowing
Long-term Existential2030-2050+Extreme-CatastrophicEarly warning signsWide but requires early action
Cascade EffectsOngoingAmplifies all categoriesAcceleratingImmediate intervention needed

Risk Activation Framework

Activation Criteria

CriterionDescriptionExample Threshold
Capability CrossingAI can perform necessary tasksGPT-4 level code generation for cyberweapons
Deployment ContextSystems deployed in relevant settingsAutonomous agents with internet access
Barrier ErosionTechnical/social barriers removedOpen-source parity reducing control
Incentive AlignmentActors motivated to exploitEconomic pressure + accessible tools

Progress Tracking Methodology

We assess progress toward activation using:

  • Technical benchmarks from evaluation organizations
  • Deployment indicators from major AI labs
  • Adversarial use cases documented in security research
  • Expert opinion surveys on capability timelines

Current Risks (Already Active)

Category: Misuse Risks

RiskStatusCurrent EvidenceImpact ScaleSource
Disinformation at scaleActive2024 election manipulation campaigns$1-10B annualReuters
Spear phishing enhancementActive82% higher believability vs human-written$10B+ annual lossesIBM Security
Code vulnerability exploitationPartially activeGPT-4 identifies 0-days, limited autonomyMedium severityAnthropic evals
Academic fraudActive30-60% of student submissions flaggedEducation integrity crisisStanford study
Romance/financial scamsActiveAI voice cloning in elder fraud$1B+ annualFTC reports

Category: Structural Risks

RiskStatusCurrent EvidenceImpact ScaleTrend
Epistemic erosionActive40% decline in information trustSociety-wideAccelerating
Economic displacementBeginning15% of customer service roles automated200M+ jobs at riskExpanding
Attention manipulationActiveAlgorithm-driven engagement optimizationMental health crisisIntensifying
Dependency formationActive60% productivity loss when tools unavailableSkill atrophy beginningGrowing

Category: Technical Risks

RiskStatusCurrent EvidenceMitigation LevelProgress
Reward hackingActiveDocumented in all RLHF systemsPartial guardrailsNo clear progress
SycophancyActiveModels agree with user regardless of truthResearch stageLimited progress
Prompt injectionActiveJailbreaks succeed >50% of timeDefense research ongoingCat-mouse game
Hallucination/confabulationActive15-30% false information in outputsDetection tools emergingGradual improvement

Near-Term Risks (2025-2027 Activation Window)

Critical Misuse Risks

RiskActivation WindowKey ThresholdCurrent ProgressIntervention Status
Bioweapons uplift2025-2028Synthesis guidance beyond textbooks60-80% to thresholdActive screening efforts
Cyberweapon development2025-2027Autonomous 0-day discovery70-85% to thresholdLimited defensive preparation
Persuasion weapons2025-2026Personalized, adaptive manipulation80-90% to thresholdNo systematic defenses
Mass deepfake attacksActive-2026Real-time, undetectable generation85-95% to thresholdDetection research lagging

Control and Alignment Risks

RiskActivation WindowKey ThresholdCurrent ProgressResearch Investment
Agentic system failures2025-2026Multi-step autonomous task execution70-80% to threshold$500M+ annually
Situational awareness2025-2027Strategic self-modeling capability50-70% to thresholdResearch accelerating
Sandbagging on evals2026-2028Concealing capabilities from evaluators40-60% to thresholdLimited detection work
Human oversight evasion2026-2029Identifying and exploiting oversight gaps30-50% to thresholdControl research beginning

Structural Transformation Risks

RiskActivation WindowKey ThresholdEconomic ImpactPolicy Preparation
Mass unemployment crisis2026-2030>10% of jobs automatable within 2 years$5-15T GDP impactMinimal policy frameworks
Authentication collapse2025-2027Can't distinguish human vs AI contentDemocratic processes at riskTechnical solutions emerging
AI-powered surveillance state2025-2028Real-time behavior predictionHuman rights implicationsRegulatory gaps
Expertise atrophy2026-2032Human skills erode from AI dependenceInnovation capacity lossNo systematic response

Long-Term Risks (ASI-Level Requirements)

Existential Risk Category

RiskEstimated WindowKey Capability ThresholdConfidence LevelResearch Investment
Misaligned superintelligence2030-2050+Systems exceed human-level at alignment-relevant tasksVery Low$1B+ annually
Recursive self-improvement2030-2045+AI meaningfully improves AI architectureLowLimited research
Decisive strategic advantage2030-2040+Single actor gains insurmountable technological leadLowPolicy research only
Irreversible value lock-in2028-2040+Permanent commitment to suboptimal human valuesLow-MediumPhilosophy/governance research

Advanced Deception and Control

RiskEstimated WindowCapability RequirementDetection DifficultyMitigation Research
Strategic deception2027-2035Model training dynamics and hide intentionsVery HighInterpretability research
Coordinated AI systems2028-2040Multiple AI systems coordinate against humansHighMulti-agent safety research
Large-scale human manipulation2028-2035Accurate predictive models of human behaviorMediumSocial science integration
Critical infrastructure control2030-2050+Simultaneous control of multiple key systemsVery HighAir-gapped research

Risk Interaction and Cascade Effects

Cascade Amplification Matrix

Triggering RiskAmplifiesMechanismTimeline Impact
Disinformation proliferationEpistemic collapseTrust erosion accelerates-1 to -2 years
Cyberweapon autonomyAuthentication collapseDigital infrastructure vulnerability-1 to -3 years
Bioweapons accessibilityAuthoritarian controlCrisis enables power concentrationVariable
Economic displacementSocial instabilityReduces governance capacity-0.5 to -1.5 years
Any major AI incidentRegulatory captureCrisis mode enables bad policy-2 to -5 years

Acceleration Factors

FactorTimeline ImpactProbability by 2027Evidence
Algorithmic breakthrough-1 to -3 years across categories15-30%Historical ML progress
10x compute scaling-0.5 to -1.5 years40-60%Current compute trends
Open-source capability parity-1 to -2 years on misuse risks50-70%Open model progress
Geopolitical AI arms race-0.5 to -2 years overall30-50%US-China competition intensifying
Major safety failure/incidentVariable, enables governance20-40%Base rate of tech failures

Deceleration Factors

FactorTimeline ImpactProbability by 2030Feasibility
Scaling laws plateau+2 to +5 years15-30%Some evidence emerging
Strong international AI governance+1 to +3 years on misuse10-20%Limited progress so far
Major alignment breakthroughVariable positive impact10-25%Research uncertainty high
Physical compute constraints+0.5 to +2 years20-35%Semiconductor bottlenecks
Economic/energy limitations+1 to +3 years15-25%Training cost growth

Critical Intervention Windows

Time-Sensitive Priority Matrix

Risk CategoryWindow OpensWindow ClosesIntervention CostEffectiveness if Delayed
Bioweapons screening2020 (missed)2027$500M-1B50% reduction
Cyber defensive AI20232026$1-3B70% reduction
Authentication infrastructure20242026$300-600M30% reduction
AI control research20222028$1-2B annually20% reduction
International governance20232027$200-500M80% reduction
Alignment foundations20152035+$2-5B annuallyVariable

Leverage Analysis by Intervention Type

Intervention CategoryCurrent LeveragePeak Leverage WindowInvestment RequiredExpected Impact
DNA synthesis screeningHigh2024-2027$100-300M globallyDelays bio threshold 2-3 years
Model evaluation standardsMedium2024-2026$50-150M annuallyEnables risk detection
Interpretability breakthroughsVery High2024-2030$500M-1B annuallyAddresses multiple long-term risks
Defensive cyber-AIMedium2024-2026$1-2BExtends defensive advantage
Public authentication systemsHigh2024-2026$200-500MPreserves epistemic infrastructure
International AI treatiesVery High2024-2027$100-200MSets precedent for future governance

Probability Calibration Over Time

Risk Activation Probabilities by Year

Risk Category20252027203020352040
Mass disinformation95% (active)99%99%99%99%
Bioweapons uplift (meaningful)25%50%70%85%95%
Autonomous cyber operations40%75%90%99%99%
Large-scale job displacement15%40%65%85%95%
Authentication crisis30%60%80%95%99%
Agentic AI control failures35%70%90%99%99%
Meaningful situational awareness20%50%75%90%95%
Strategic AI deception5%20%45%70%85%
ASI-level misalignment<1%3%15%35%55%

Uncertainty Ranges and Expert Disagreement

RiskOptimistic TimelineMedianPessimistic TimelineExpert Confidence
Cyberweapon autonomy2028-20302025-20272024-2025Medium (70% within range)
Bioweapons threshold2030-20352026-20292024-2026Low (50% within range)
Mass unemployment2035-20402028-20322025-2027Very Low (30% within range)
Superintelligence2045-Never2030-20402027-2032Very Low (20% within range)

Strategic Resource Allocation

Investment Priority Framework

Priority TierTimelineInvestment LevelRationale
Tier 1: CriticalImmediate-2027$3-5B annuallyWindow closing rapidly
Tier 2: Important2025-2030$1-2B annuallyFoundation for later risks
Tier 3: Foundational2024-2035+$500M-1B annuallyLong-term preparation
Research AreaAnnual InvestmentJustificationExpected ROI
Bioweapons screening infrastructure$200-400M (2024-2027)Critical window closingVery High - prevents catastrophic risk
AI interpretability research$300-600M ongoingMulti-risk mitigationHigh - enables control across scenarios
Cyber-defense AI systems$500M-1B (2024-2026)Maintaining defensive advantageMedium-High
Authentication/verification tech$100-200M (2024-2026)Preserving epistemic infrastructureHigh
International governance capacity$100-200M (2024-2027)Coordination before crisisVery High - prevents race dynamics
AI control methodology$400-800M ongoingBridge to long-term safetyHigh
Economic transition planning$200-400M (2024-2030)Social stability preservationMedium

Key Cruxes and Uncertainties

Timeline Uncertainty Analysis

Core UncertaintyIf OptimisticIf PessimisticCurrent Best EstimateImplications
Scaling law continuationPlateau by 2027-2030Continue through 2035+60% likely to continue±3 years on all timelines
Open-source capability gapMaintains 2+ year lagAchieves parity by 202655% chance of rapid catch-up±2 years on misuse risks
Alignment research progressMajor breakthrough by 2030Limited progress through 203520% chance of breakthrough±5-10 years on existential risk
Geopolitical cooperationSuccessful AI treatiesIntensified arms race25% chance of cooperation±2-5 years on multiple risks
Economic adaptation speedSmooth transition over 10+ yearsRapid displacement over 3-5 years40% chance of rapid displacementSocial stability implications

Research and Policy Dependencies

DependencySuccess ProbabilityImpact if FailedMitigation Options
International bioweapons screening60%Bioweapons threshold advances 2-3 yearsNational screening systems, detection research
AI evaluation standardization40%Reduced early warning capabilityIndustry self-regulation, government mandates
Interpretability breakthroughs30%Limited control over advanced systemsMultiple research approaches, AI-assisted research
Democratic governance adaptation35%Poor quality regulation during crisisEarly capacity building, expert networks

Implications for Different Stakeholders

For AI Development Organizations

Immediate priorities (2024-2025):

  • Implement robust evaluations for near-term risks
  • Establish safety teams scaling with capability teams
  • Contribute to industry evaluation standards

Near-term preparations (2025-2027):

  • Deploy monitoring systems for newly activated risks
  • Engage constructively in governance frameworks
  • Research control methods before needed

For Policymakers

Critical window actions:

  • Establish regulatory frameworks before crisis mode
  • Focus on near-term risks to build governance credibility
  • Invest in international coordination mechanisms

Priority areas:

  1. Bioweapons screening infrastructure
  2. AI evaluation and monitoring standards
  3. Economic transition support systems
  4. Authentication and verification requirements

For Safety Researchers

Optimal portfolio allocation:

  • 40% near-term (1-2 generation) risk mitigation
  • 40% foundational research for long-term risks
  • 20% current risk mitigation and response

High-leverage research areas:

  1. Interpretability for multiple risk categories
  2. AI control methodology development
  3. Evaluation methodology for emerging capabilities
  4. Social science integration for structural risks

For Civil Society Organizations

Advocacy priorities:

  • Demand transparency in capability evaluations
  • Push for public interest representation in governance
  • Support authentication infrastructure development
  • Advocate for economic transition policies

Limitations and Model Uncertainty

Methodological Limitations

LimitationImpact on AccuracyMitigation Strategies
Expert overconfidenceTimelines may be systematically early/lateMultiple forecasting methods, base rate reference
Capability discontinuitiesSudden activation possibleBroader uncertainty ranges, multiple scenarios
Interaction complexityCascade effects poorly understoodSystems modeling, historical analogies
Adversarial adaptationDefenses may fail faster than expectedRed team exercises, worst-case planning

Areas for Model Enhancement

  1. Better cascade modeling - More sophisticated interaction effects
  2. Adversarial dynamics - How attackers adapt to defenses
  3. Institutional response capacity - How organizations adapt to new risks
  4. Cross-cultural variation - Risk manifestation in different contexts
  5. Economic feedback loops - How risk realization affects development

Sources & Resources

Primary Research Sources

OrganizationTypeKey Contributions
AnthropicAI LabRisk evaluation methodologies, scaling policies
OpenAIAI LabPreparedness framework, capability assessment
METREvaluation OrgTechnical capability evaluations
RAND CorporationThink TankPolicy analysis, national security implications
Center for AI SafetySafety OrgRisk taxonomy, expert opinion surveys

Academic Literature

PaperAuthorsKey Finding
Model evaluation for extreme risksAnthropic Constitutional AI TeamEvaluation frameworks for dangerous capabilities
AI timelines and capabilitiesVarious forecasting researchCapability development trajectories
Cybersecurity implications of AICSETNear-term cyber risk assessment

Policy and Governance Sources

SourceTypeFocus Area
NIST AI Risk Management FrameworkGovernment StandardRisk management methodology
EU AI ActRegulationComprehensive AI governance framework
UK AI Safety Summit OutcomesInternationalMulti-stakeholder coordination

Expert Opinion and Forecasting

PlatformTypeUse Case
Metaculus AI forecastsPrediction MarketQuantitative timeline estimates
Expert Survey on AI RiskAcademic SurveyExpert opinion distribution
Future of Humanity Institute reportsResearch InstituteLong-term risk analysis

Complementary Risk Models

  • AI Capability Threshold Model - Specific capability requirements for risk activation
  • Bioweapons AI Uplift Model - Detailed biological weapons timeline
  • Cyberweapons Attack Automation - Cyber capability development
  • Authentication Collapse Timeline - Digital verification crisis
  • Economic Disruption Impact - Labor market transformation

Risk Category Cross-References

  • Accident Risks - Technical AI safety failures
  • Misuse Risks - Intentional harmful applications
  • Structural Risks - Systemic societal impacts
  • Epistemic Risks - Information environment degradation

Response Strategy Integration

  • Governance Responses - Policy intervention strategies
  • Technical Safety Research - Engineering solutions
  • International Coordination - Global cooperation frameworks

References

An Anthropic article examining the core difficulties in assessing AI system capabilities and safety properties. It explores why robust evaluations are critical yet methodologically challenging, addressing gaps between benchmark performance and real-world behavior as well as the limitations of current evaluation frameworks.

★★★★☆
2Active screening effortsNuclear Threat Initiative

This NTI analysis examines active screening approaches to prevent catastrophic bioweapons threats, focusing on detection and interdiction strategies. It addresses biosecurity governance frameworks and the technical and policy measures needed to reduce the risk of biological weapons development and use. The piece contributes to understanding how proactive monitoring can serve as a layer of defense against existential-level biological risks.

★★★★☆

This page outlines the European Commission's comprehensive policy framework for AI, centered on promoting trustworthy, human-centric AI through the AI Act, AI Continent Action Plan, and Apply AI Strategy. It aims to balance Europe's global AI competitiveness with safety, fundamental rights, and democratic values. Key initiatives include AI Factories, the InvestAI Facility, GenAI4EU, and the Apply AI Alliance.

★★★★☆
4Detection research laggingdeepfakedetectionchallenge.com

The Deepfake Detection Challenge (DFDC) is an initiative to advance research on detecting AI-generated synthetic media (deepfakes). It highlights the gap between rapidly improving deepfake generation capabilities and the slower development of reliable detection tools, reflecting a broader pattern of defensive research lagging behind offensive AI capabilities.

This URL was intended to link to Metaculus's collection of AI-related forecasting questions, a platform where forecasters make probabilistic predictions about AI development timelines, capabilities, and risks. However, the page currently returns a 404 error, indicating the content may have moved or been reorganized.

★★★☆☆
6AI Safety Summit 2023UK Government·Government

The official UK government page for the AI Safety Summit 2023, held November 1-2 at Bletchley Park, which convened governments, AI companies, civil society, and researchers to address frontier AI risks. Key outputs include the Bletchley Declaration—a multilateral agreement on AI safety—company safety policies, and a frontier AI capabilities and risks discussion paper. The summit marked a landmark moment in international AI governance coordination.

★★★★☆

The 2022 ESPAI surveyed 738 machine learning researchers (NeurIPS/ICML authors) about AI progress timelines and risks, serving as a replication and update of the 2016 survey. Key findings include an aggregate forecast of 50% chance of HLMI by 2059 (37 years from 2022), with significant disagreement among experts about timelines and risks.

★★★☆☆

OpenAI's safety hub outlines their multi-stage approach to AI safety through teaching (value alignment and content filtering), testing (red teaming and preparedness evaluations), and sharing (real-world feedback loops). It covers key concern areas including child safety, deepfakes, bias, and election integrity, and links to their Preparedness Framework and related safety documentation.

★★★★☆

METR is an organization conducting research and evaluations to assess the capabilities and risks of frontier AI systems, focusing on autonomous task completion, AI self-improvement risks, and evaluation integrity. They have developed the 'Time Horizon' metric measuring how long AI agents can autonomously complete software tasks, showing exponential growth over recent years. They work with major AI labs including OpenAI, Anthropic, and Amazon to evaluate catastrophic risk potential.

★★★★☆
10Model Evaluation for Extreme RisksarXiv·Toby Shevlane et al.·2023·Paper

This paper addresses the critical role of model evaluation in mitigating extreme risks from advanced AI systems. As AI development progresses, general-purpose AI systems increasingly possess both beneficial and harmful capabilities, including potentially dangerous ones like offensive cyber abilities or manipulation skills. The authors argue that two types of evaluations are essential: dangerous capability evaluations to identify harmful capacities, and alignment evaluations to assess whether models are inclined to use their capabilities for harm. These evaluations are vital for informing policymakers and stakeholders, and for making responsible decisions regarding model training, deployment, and security.

★★★☆☆

IBM's annual Cost of a Data Breach Report, produced with the Ponemon Institute, provides global research on data breach costs, trends, and contributing factors. The 2025 edition highlights an 'AI oversight gap' where rapid AI adoption is outpacing security governance, with ungoverned AI systems facing higher breach likelihood and costs. The global average breach cost reached $4.4M USD.

12AI timelines and capabilitiesarXiv·DeepSeek-AI et al.·2024·Paper

This paper presents DeepSeek LLM, an open-source large language model project that addresses inconsistencies in scaling law literature by providing empirical findings for scaling models at 7B and 67B parameters. The authors developed a 2 trillion token dataset and applied supervised fine-tuning and Direct Preference Optimization to create DeepSeek Chat models. Their evaluation demonstrates that DeepSeek LLM 67B outperforms LLaMA-2 70B across multiple benchmarks, particularly in code, mathematics, and reasoning tasks, with the chat variant showing competitive performance against GPT-3.5.

★★★☆☆

The Open LLM Leaderboard is a HuggingFace-hosted benchmarking platform that compares open-source large language models across standardized evaluations in a transparent and reproducible manner. It allows researchers and practitioners to filter, search, and rank models by performance metrics, providing a community reference for tracking AI capabilities progress. The leaderboard has since been archived, reflecting the rapid pace of LLM development.

The NIST AI RMF is a voluntary, consensus-driven framework released in January 2023 to help organizations identify, assess, and manage risks associated with AI systems while promoting trustworthiness across design, development, deployment, and evaluation. It provides structured guidance organized around core functions and is accompanied by a Playbook, Roadmap, and a Generative AI Profile (2024) addressing risks specific to generative AI systems.

★★★★★

The FTC warns consumers about AI-powered voice cloning scams where fraudsters impersonate distressed family members to extract money. The post explains how scammers use readily available AI tools to clone voices from social media and other public audio, and offers guidance on how to verify calls and protect against such fraud.

★★★★☆

The Center for AI Safety (CAIS) is a research organization focused on mitigating catastrophic and existential risks from advanced AI systems. It conducts technical research, publishes surveys and statements, and supports field-building efforts across academia and industry. CAIS is notable for its broad coalition-building, including its widely-cited statement on AI extinction risk signed by leading researchers.

★★★★☆

Reuters analysis examining how AI-generated misinformation poses risks to the 2024 election cycle, covering deepfakes, synthetic media, and coordinated disinformation campaigns. The piece assesses the scale of the threat and the challenges platforms and regulators face in detection and mitigation.

★★★★☆

RAND Corporation's AI research hub covers policy, national security, and governance implications of artificial intelligence. It aggregates reports, analyses, and commentary on AI risks, military applications, and regulatory frameworks from one of the leading U.S. defense and policy think tanks.

★★★★☆

Epoch AI's analysis of historical trends in compute used for training notable AI systems, identifying three distinct eras: pre-deep learning, deep learning, and large-scale models. The research documents how training compute has grown by roughly 4-5 orders of magnitude since 2010, with a notable shift toward massive compute investment after 2015.

★★★★☆
20FHI expert elicitationFuture of Humanity Institute

This resource from the Future of Humanity Institute (FHI) at Oxford involves expert elicitation surveys focused on AI development timelines, capability thresholds, and prioritization of interventions. It aggregates forecasts from researchers to inform understanding of when transformative AI might arrive and what safety measures may be most effective.

★★★★☆

This Stanford HAI resource examines the challenges AI tools—particularly large language models—pose to academic integrity, exploring how institutions should respond to AI-assisted cheating and the broader implications for education and assessment. It likely draws on Stanford research to assess the prevalence and nature of AI misuse in academic settings and proposes policy frameworks for educators.

★★★★☆

A CSET analysis examining how artificial intelligence is reshaping the cybersecurity landscape, covering both offensive and defensive applications of AI in cyber operations. The report assesses near- and long-term implications of AI capabilities for cyber threats, defenses, and policy responses.

★★★★☆

The C2PA is an industry coalition that has developed an open technical standard for attaching verifiable provenance metadata to digital content, functioning like a 'nutrition label' that tracks a file's origin, creation tools, and edit history. This standard aims to help consumers and platforms distinguish authentic content from manipulated or AI-generated media. It is backed by major technology and media companies including Adobe, Microsoft, and the BBC.

Related Wiki Pages

Top Related Pages

Risks

Reward HackingScheming

Approaches

AI EvaluationConstitutional AI

Analysis

Autonomous Cyber Attack TimelineAuthentication Collapse Timeline Model

Concepts

International Coordination MechanismsSituational AwarenessAgentic AIAI Timelines

Other

AI ControlInterpretabilityPhilip TetlockEli Lifland

Organizations

AnthropicMETR

Key Debates

AI Accident Risk CruxesAI Risk Critical Uncertainties Model

Policy

NIST AI Risk Management Framework (AI RMF)EU AI Act