Skip to content
Longterm Wiki
Navigation
Updated 2026-03-17HistoryData
Page StatusContent
Edited 3 weeks ago3.9k words3 backlinksUpdated quarterlyDue in 10 weeks
69QualityGood72ImportanceHigh64ResearchModerate
Content8/13
SummaryScheduleEntityEdit history1Overview
Tables16/ ~16Diagrams1/ ~2Int. links41/ ~31Ext. links0/ ~20Footnotes2/ ~12References17/ ~12Quotes0Accuracy0RatingsN:5.8 R:7.2 A:7.5 C:8Backlinks3
Change History1
Auto-improve (standard): Bioweapons Attack Chain Model3 weeks ago

Improved "Bioweapons Attack Chain Model" via standard pipeline (590.5s). Quality score: 84. Issues resolved: Bare URL in Technical Resources table: 'cdc.gov/biowatch' is; Bare URL in Technical Resources table: 'who.int/emergencies'; Bare URL in Technical Resources table: 'nist.gov/cyberframew.

590.5s · $5-8

TODOs3
Complete 'Quantitative Analysis' section (8 placeholders)
Complete 'Limitations' section (6 placeholders)
Update AI model comparison table as new benchmark evaluations become available

Bioweapons Attack Chain Model

Analysis

Bioweapons Attack Chain Model

Multiplicative attack chain model estimates catastrophic bioweapons probability at 0.02-3.6%, with state actors (3.0%) showing highest estimated risk due to lab access. DNA synthesis screening offers highest modeled cost-effectiveness at $7-20M per 1% risk reduction, with defense-in-depth providing 5-25% total reduction through targeting multiple bottlenecks. Recent expert virology benchmarks and competing biosecurity assessments suggest the AI uplift parameter carries greater uncertainty than earlier RAND findings indicated.

Model TypeProbability Decomposition
Target RiskBioweapons
Related
Risks
Bioweapons Risk
3.9k words · 3 backlinks

Overview

This model decomposes AI-assisted bioweapons attacks into seven sequential bottlenecks, revealing that catastrophic biological terrorism requires success across multiple independent failure modes. The framework draws on RAND Corporation's 2024 red-team study finding no statistically significant AI uplift for biological attacks, combined with historical analysis of state bioweapons programs and terrorist attempts.

Key findings: Overall attack probability ranges from 0.02–3.6%, with state actors posing the highest estimated risk (3.0%) due to superior laboratory access. The multiplicative probability structure means moderate interventions at any step provide substantial protection. DNA synthesis screening offers 5–15% modeled risk reduction for $1–20M annually, while metagenomic surveillance provides 15–25% reduction for $500M. This mathematical structure is consistent with defense-in-depth strategies against bioweapons risks.

The model incorporates a key empirical claim that information is not capability—even complete knowledge cannot alone overcome the synthesis bottleneck, where tacit knowledge and laboratory skills create barriers that are not fully addressed by AI-assisted information retrieval. This claim reflects current empirical evidence but carries substantial uncertainty as AI capabilities evolve; the conditions under which it could cease to hold are discussed in the AI Uplift and Limitations sections below.

Expert opinion on biosecurity trajectories is divided. Some analysts emphasize persistent technical barriers and improving defensive capabilities as grounds for measured optimism; others point to accelerating AI capabilities, dual-use research proliferation, and governance gaps as reasons for concern.1 The sections that follow present both perspectives where relevant, rather than a single projected trajectory.

Risk Assessment

Risk DimensionAssessmentConfidence LevelEvidence Base
SeverityExtremeHighHistorical pandemic impacts, WMD classification
LikelihoodVery Low (0.02–3.6%)LowHigh uncertainty across all parameters
Timeline5–10 yearsMediumAI capability trajectory, countermeasure development
TrendSlowly increasing (contested)LowAI advancement vs. biosecurity improvements; expert disagreement on net direction
ReversibilityNoneHighPermanent knowledge proliferation
PrecedentLimitedHighFew historical bioterror attempts, no confirmed AI-assisted case

Attack Chain Architecture

Sequential Bottleneck Model

A successful attack requires traversing all seven stages sequentially. Failure at any stage terminates the attack chain with multiplicative probability reduction.

Diagram (loading…)
flowchart TD
  subgraph phase1["Phase 1: Preparation"]
      A[Motivated Actor<br/>P = 0.95] --> B[AI Access<br/>P = 0.7-0.9]
      B --> C[AI Provides Uplift<br/>P = 0.2-0.5]
  end

  subgraph phase2["Phase 2: Development"]
      C --> D[Lab Access<br/>P = 0.3-0.6]
      D --> E[Successful Synthesis<br/>P = 0.1-0.4]
  end

  subgraph phase3["Phase 3: Execution"]
      E --> F[Effective Deployment<br/>P = 0.2-0.5]
      F --> G[Evades Countermeasures<br/>P = 0.3-0.7]
  end

  G --> H[Catastrophic Attack<br/>P = 0.02-3.6%]

  style A fill:#fee
  style H fill:#fcc
  style E fill:#ffe

Mathematical Framework

The compound probability follows a multiplicative model with independence assumption:

P(catastrophic attack)=i=17PiP(\text{catastrophic attack}) = \prod_{i=1}^{7} P_i

where each PiP_i represents conditional success probability at step ii. This structure creates a defense multiplier effect: reducing any single parameter by 50% reduces overall risk by 50%, regardless of which step is targeted.

The independence assumption is a modeling convenience rather than an empirical claim. As noted in the Limitations section, positive correlation across steps—plausible for sophisticated actors who tend to succeed in multiple domains—would increase compound probability substantially. The correlation sensitivity analysis provides estimates under different correlation assumptions.

Parameter Estimates

Actor-Specific Risk Profiles

Different actor types face distinct bottleneck patterns based on resource access and operational constraints. The figures below reflect model-derived estimates under the independence assumption; they should be interpreted as illustrative ordinal rankings rather than precise probabilities.

Actor TypeResourcesLab AccessSynthesis CapabilityCompound RiskPrimary Bottleneck
State programUnlimitedHigh (0.90)High (0.50)≈3.0%Attribution/deterrence
Well-funded terroristHighMedium (0.40)Medium (0.25)≈0.6%Laboratory acquisition
Lone actorLowVery Low (0.15)Very Low (0.10)≈0.06%Technical execution
Criminal organizationMediumLow (0.30)Low (0.15)≈0.15%Motivation sustainability

Under the model's assumptions, state actors account for a large share of estimated catastrophic risk, primarily due to unrestricted laboratory access and scientific expertise. The specific figure of 80% cited in earlier versions of this page is model-derived under independence assumptions and should not be treated as an empirical finding; the true share depends heavily on correlation structure and parameter values that carry substantial uncertainty.

Parameter Uncertainty Analysis

The 180x range in compound probability reflects deep (Knightian) uncertainty rather than a conventional statistical confidence interval. No empirical study has directly estimated these joint probabilities; the ranges derive from expert elicitation and boundary-case reasoning. Readers should treat the compound estimate as an order-of-magnitude heuristic rather than a calibrated forecast.

StepLow EstimateCentralHigh EstimateUncertainty FactorKey Variable
Motivated actor0.900.950.991.1xState program count
AI access0.700.800.901.3xOpen-source proliferation
AI uplift0.200.350.502.5xCapability trajectory
Lab access0.300.450.602.0xImprovised lab viability
Synthesis success0.100.250.404.0xTacit knowledge barrier
Deployment0.200.350.502.5xDelivery mechanism reliability
Countermeasures0.300.500.702.3xResponse capability variation
Compound0.02%0.5%3.6%180xCompounding uncertainty

Critical Bottleneck Analysis

Step 3: AI Uplift Assessment

The most uncertain parameter in the model is whether AI systems provide meaningful uplift to actors seeking to develop biological weapons. Current evidence is mixed and rapidly evolving.

The RAND Corporation's 2024 study compared AI-assisted versus internet-only groups for biological attack planning, finding no statistically significant difference in information quality or actionability. However, this study has known design limitations: it tested a specific task framing, used particular model versions, and did not assess laboratory execution capability directly. Its null result should be understood as evidence against one specific form of uplift (information retrieval quality), not a comprehensive assessment of AI's biosecurity implications.

More recent capability evaluations suggest a different picture for specialized technical tasks. The AI Safety Newsletter issue 52 reported on an expert virology benchmark designed to assess whether frontier AI systems could answer questions at or above the level of trained virologists.2 Unlike the RAND study's focus on information retrieval, this benchmark targets domain-specific reasoning and problem-solving of the kind that could reduce the tacit knowledge barrier. The benchmark methodology, the specific systems evaluated, and the precise performance levels achieved are discussed in the AI Uplift Assessment Model; the implication for this model is that the AI uplift probability range (0.20–0.50) may require upward revision as frontier models are evaluated on more demanding virology tasks.

The table below shows performance data from the RAND study for the models evaluated at that time. It should be noted that these models (GPT-4, Claude-3) have since been superseded by more capable successors, and that the evaluation methodology differs from the virology benchmark approach. Direct comparison across rows is therefore limited.

Information SourceAccuracy ScoreCompleteness ScoreActionability ScoreUplift Factor
Internet search only7.2/106.8/105.9/10Baseline
GPT-4 assisted7.4/107.1/106.2/101.04x
Claude-3 assisted7.1/106.9/106.0/101.01x
Expert consultation9.1/108.7/108.2/101.35x

Note: These scores reflect the RAND 2024 study design and the specific model versions tested. Frontier models released in 2024–2025 (GPT-4o, Claude 3.5/3.7, Gemini 1.5/2.0) have not been evaluated under the same protocol. The virology benchmark referenced in AI Safety Newsletter #52 uses a different methodology and is not directly comparable to these scores.

The critical unresolved question is whether future AI systems will bridge the gap between information access and laboratory execution—specifically, whether they can convey tacit procedural knowledge in ways that reduce the synthesis bottleneck. This remains an active area of research and policy debate, with implications for the AI uplift parameter range in this model. The AI Uplift Assessment Model provides more granular analysis of this question.

Step 5: Synthesis Bottleneck

The synthesis step represents the strongest persistent barrier in the current model. Even with complete genetic sequences and theoretical protocols, wet-lab execution requires tacit knowledge that transfers poorly through text-based AI interaction under current capabilities.

Synthesis ChallengeInformation AvailabilityTacit Knowledge RequirementAI Assistability
DNA synthesisHighLowHigh
Protein expressionHighMediumMedium
Virus assemblyMediumHighLow
Virulence optimizationLowVery HighVery Low
Environmental stabilityLowVery HighVery Low

Historical evidence is consistent with high synthesis failure rates. The Soviet Biopreparat program, despite substantial resources and expert personnel, required years to develop reliable production methods for select agents. Aum Shinrikyo's biological weapons program did not achieve effective biological weapons deployment despite investment in laboratory facilities.

AI-assisted Interpretability: A notable development not reflected in the original synthesis challenge table is the emergence of structure-prediction and protein design systems such as AlphaFold3, RFDiffusion, and ESM3. These tools can accelerate certain steps in protein engineering—predicting how sequence changes affect structure and function—in ways that may partially reduce tacit knowledge requirements for actors with some laboratory background. However, their relevance varies substantially by synthesis challenge: they are more applicable to protein expression problems than to viral assembly or environmental stability, and they require wet-lab validation that remains a significant barrier. The AI Assistability ratings in the table above reflect current capabilities; this column warrants revision as protein design tool capabilities are evaluated against specific synthesis tasks. This potential pathway is discussed further in the AI Uplift Assessment Model.

Step 7: Countermeasure Effectiveness

Modern biosurveillance capabilities vary by region and pathogen type, creating geographic risk differentials.

Countermeasure TypeDetection TimeCoverageEffectiveness vs. Novel Agents
Syndromic surveillance3–7 daysUrban areasMedium
Laboratory networks5–14 daysDeveloped countriesHigh
Genomic sequencing1–5 daysMajor citiesVery High
Medical countermeasuresImmediate–YearsVariableLow (for novel agents)

The CDC's BioWatch program provides continuous aerosol monitoring in major U.S. cities, while WHO's Disease Outbreak News coordinates global surveillance. Coverage remains more limited in lower-resource regions, creating geographic asymmetries in detection probability.

AI as a defensive tool: The countermeasure landscape is not static, and AI systems are being developed for defensive biosecurity applications as well as representing a potential offensive risk factor. Relevant defensive applications include AI-assisted analysis of syndromic surveillance data to identify anomalous outbreak patterns, accelerated genomic sequencing interpretation for novel pathogen identification, and AI-assisted countermeasure development pipelines. The current model treats AI exclusively as a variable affecting offensive capability (Step 3); a more complete analysis would incorporate AI's potential effects on countermeasure effectiveness (Step 7) as well. Proponents of AI-assisted biosurveillance argue it could meaningfully reduce detection time for novel agents; critics note that the same capability improvements benefit both attacker and defender, and that defensive AI tools require significant infrastructure investment to deploy at scale.

Intervention Cost-Effectiveness

High-Leverage Interventions

Defense-in-depth strategies exploit the multiplicative probability structure, where moderate improvements across multiple steps compound to substantial risk reduction. The cost estimates below are model-derived and illustrative; empirical grounding for specific figures is limited. See also: AI Safety Defense in Depth Model for the general framework.

InterventionAnnual CostRisk ReductionCost per % ReductionImplementation Difficulty
Metagenomic surveillance$500M15–25%$20–33MMedium
DNA synthesis screening$100M5–15%$7–20MLow
BSL facility security$200M5–10%$20–40MMedium
AI model guardrails$50M2–8%$6–25MHigh
Universal flu vaccine$2B20–40%$50–100MVery High
International monitoring$300M3–8%$38–100MVery High

DNA synthesis screening appears cost-effective in the model's framework, requiring minimal international coordination while providing risk reduction across actor types. However, this intervention involves tradeoffs: mandatory screening requirements add costs and delays for legitimate research, and the effectiveness of screening against novel synthesis routes (rather than known sequences) is uncertain. International coordination for monitoring raises sovereignty considerations that have slowed similar efforts in other domains. These tradeoffs are not reflected in the cost-per-percent-reduction figures above.

Intervention Targeting by Actor Type

Different actor types require tailored intervention strategies based on their specific capability profiles and bottlenecks.

Actor TypePrimary BottleneckMost Effective InterventionSecondary Intervention
State programAttribution costsInternational monitoringDiplomatic deterrence
Funded terroristLaboratory accessBSL security screeningFinancial monitoring
Lone actorSynthesis capabilityDNA synthesis screeningTechnical education controls
Criminal orgSustained motivationLaw enforcement intelligenceSupply chain monitoring

Timeline and Trajectory Analysis

Expert Views: Pessimistic and Optimistic Scenarios

Expert opinion on biosecurity trajectories is substantively divided. The framing below draws on competing analytical positions in the literature rather than presenting a single projected scenario.1

Pessimistic perspective: Analysts concerned about near-term risk deterioration point to: accelerating AI capability development that may outpace current threat models; proliferation of open-source biological research tools and methods; weakening of international biosecurity governance norms; dual-use potential of protein design tools reducing synthesis barriers; and historically slow defensive infrastructure investment relative to offensive capability development. From this perspective, the window for effective preventive governance may be narrowing.1

Optimistic perspective: Analysts who see grounds for longer-term improvement note: the synthesis bottleneck has persisted despite decades of information proliferation; biosurveillance infrastructure is expanding with declining sequencing costs; major AI laboratories have adopted pre-deployment biosecurity evaluations as an emerging norm; the RAND findings suggest near-term AI uplift may be more limited than feared; and historical bioterrorism attempts have had low success rates even for well-resourced actors.1

Assessment: Neither perspective fully captures the uncertainty. The model's trajectory table below presents a single-scenario projection that reflects one plausible path; alternative trajectories consistent with the pessimistic or optimistic framings would show materially different risk evolution.

Capability Evolution Scenarios

The figures in this table represent one illustrative scenario under central-case assumptions. Alternative trajectories could diverge substantially depending on AI capability progression rates, biosecurity investment levels, and governance developments.

TimeframeAI Uplift TrendBiosecurity InvestmentNet Risk (central case)Key Developments
2025–2027+20%+10%+8%Open-source model proliferation
2027–2030+50%+25%+15%AI-lab tool integration
2030–2035+100%+75%−5%Universal vaccine platforms

The 2030–2035 net-risk reduction is contingent on sustained biosecurity governance investment and successful development of universal vaccine platforms—neither of which is assured. Under a pessimistic scenario where AI capability acceleration outpaces defensive investment, net risk in that period could remain positive.

Critical Decision Points

YearDecision PointRisk ImpactPolicy Window
2025Open-source AI regulationMediumCurrent
2026DNA synthesis screening mandateHigh12–18 months
2028International bioweapons verificationVery High3–5 years
2030Universal vaccine platformExtreme5–10 years

Regarding AI laboratory pre-deployment biosecurity evaluations: several major AI developers have incorporated biosecurity-specific red-teaming into their deployment processes, though the scope, methodology, and third-party auditability of these evaluations vary. Standardization and independent verification of pre-deployment biosecurity assessments remain open governance questions.2

Model Limitations and Uncertainties

Structural Assumptions

The multiplicative independence model embeds several simplifying assumptions that may not reflect reality:

AssumptionValidityImpact if ViolatedEvidence
Step independenceLow2–5x higher riskSophisticated actors may succeed at multiple steps simultaneously
Single attempt onlyMedium1.5–3x higher riskPersistent actors make multiple tries
Binary outcomesLowUnderestimates full harm distributionSmaller attacks causing significant (non-catastrophic) harm are not captured
Static probabilitiesLowDynamic risk evolutionAI and countermeasures co-evolve; probabilities change over time

Binary outcome limitation: The model is calibrated for catastrophic attack scenarios (mass-casualty bioweapons). This framing excludes a range of harmful outcomes that fall short of catastrophe: targeted assassinations, limited-scale attacks causing dozens or hundreds of casualties, economic disruption from credible threats, and public health burdens from partially effective agents. The full harm distribution from biological threats is broader than the catastrophic tail, and the model's policy implications may not generalize to the sub-catastrophic range.

Dynamic probability evolution: The model treats each step's probability as fixed. In practice, AI capabilities, biosecurity countermeasures, regulatory frameworks, and actor capabilities co-evolve. A more complete model would treat probabilities as time-varying functions with feedback between attacker adaptation and defender response. The timeline table above provides a partial approximation of this dynamic, but does not capture strategic adaptation by sophisticated actors in response to specific interventions.

Insider threat scenarios: The model does not separately address insider threats—actors with existing laboratory access and technical expertise who seek to misuse their position. Insider scenarios have materially different bottleneck profiles: Steps 3 (AI uplift) and 4 (lab access) are largely bypassed, making synthesis capability and countermeasure evasion the primary limiting factors. The intervention priorities for insider threats differ correspondingly, with personnel security and behavioral monitoring taking precedence over synthesis screening.

Permanent knowledge proliferation: As noted in the Risk Assessment table, knowledge proliferation is treated as irreversible. This means interventions that reduce synthesis barrier height provide diminishing returns over time as laboratory techniques become more widely documented. The long-term policy implication is that early investment in detection and response capabilities may have higher durable value than early investment in synthesis access barriers, though both contribute in the near term.

Correlation Sensitivity Analysis

If attack steps are positively correlated rather than independent, overall risk increases substantially. The estimates below are model-derived; no empirical study has directly measured inter-step correlations for bioweapons attack scenarios. The moderate-correlation scenario (r=0.4) may be plausible for sophisticated actors, but this is a judgment rather than an empirical finding.

Correlation LevelEffective Independent StepsCompound ProbabilityRisk Multiplier
None (r=0)70.5%1.0x
Weak (r=0.2)≈60.8%1.6x
Moderate (r=0.4)≈4.51.8%3.6x
Strong (r=0.7)≈2.56.2%12.4x

The compound uncertainty range (0.02–3.6%) presented throughout this model assumes zero correlation. Under moderate positive correlation, the upper bound would rise to approximately 6–7%, and the central estimate to approximately 1.8%. This sensitivity analysis suggests the model's headline figures may understate risk for sophisticated, well-resourced actors.

Key Insights and Policy Implications

Strategic Insights

  • Defense multiplier effect: The multiplicative structure means moderate barriers at multiple steps provide compounding protection
  • Information ≠ capability (contingent): The synthesis bottleneck has persisted despite information proliferation under current AI capabilities; this may change if AI systems develop stronger tacit-knowledge transfer for laboratory procedures
  • Geographic risk concentration: Detection and response capability is unevenly distributed, with lower-resource regions having substantially longer average detection times
  • Actor type heterogeneity: Different actor categories have materially different bottleneck profiles, requiring differentiated intervention strategies

High-Priority Research Questions

Research PriorityUncertainty ReductionPolicy RelevanceTimeline
AI uplift measurement (advanced models)HighVery High1–2 years
Synthesis barrier persistence with protein design toolsHighHigh2–3 years
Correlation structure across attack stepsMediumMedium3–5 years
Countermeasure effectiveness vs. novel agentsHighVery High1–3 years
AI-assisted biosurveillance effectivenessMediumHigh2–4 years

Policy Recommendations

The model supports defense-in-depth approaches over single-point solutions, given the multiplicative protection structure and robustness to parameter uncertainty. The recommendations below are ordered by modeled cost-effectiveness; each involves implementation tradeoffs that policymakers would need to weigh.

  1. Immediate (2025–2026): Expand DNA synthesis screening programs with international coordination. Tradeoff: Adds compliance costs and latency for legitimate research; effectiveness against novel synthesis routes not yet demonstrated; international coordination has historically been slow.

  2. Short-term (2026–2028): Expand metagenomic surveillance to major population centers globally. Tradeoff: High infrastructure cost; raises data sovereignty and privacy considerations; coverage gaps in lower-resource regions may persist.

  3. Medium-term (2028–2032): Develop international bioweapons verification protocols with enforcement mechanisms. Tradeoff: Enforcement mechanisms require state compliance that may not be achievable; verification of biological programs is technically more difficult than nuclear verification; diplomatic and sovereignty constraints are substantial.

  4. Long-term (2030+): Invest in universal vaccine platforms and rapid response capabilities. Tradeoff: Very high development cost; long lead times; platform approaches may not generalize to all novel agents.

  5. Cross-cutting: Standardize and independently audit pre-deployment biosecurity evaluations by AI developers.2 Tradeoff: Adds development costs; auditing methodology not yet established; competitive pressures may create incentives to minimize scope.

The model's cost-effectiveness estimates carry the same deep uncertainty as the underlying probability estimates. The ranking of interventions could shift under different parameter assumptions, particularly if AI uplift probabilities are revised upward based on more recent capability evaluations.

This model connects to several related risk assessments and intervention frameworks:

  • AI Uplift Assessment Model — Detailed quantitative analysis of Step 3 parameters, including the virology benchmark findings and protein design tool implications
  • AI-Bioweapons Timeline Model — Temporal evolution of attack chain probabilities
  • AI Safety Defense in Depth Model — General framework for layered security strategies applied in the Intervention Cost-Effectiveness section
  • Misuse Risks — Broader category including bioweapons and autonomous weapons

Sources & Resources

Primary Research

Source TypeCitationKey FindingsAccess
Government studyRAND Corporation Bioweapons Red Team (2024)No significant AI uplift detected for information retrieval tasks in study designPublic
Academic analysisEsvelt - Delay, Detect, Defend (2022)Synthesis barriers persist; layered defense frameworkOpen access
Industry reportAnthropic Frontier Threats Assessment (2023)Current model limitations; biosecurity evaluation methodologyPublic
Policy analysisNTI Synthetic Biology Report (2024)Governance recommendations for synthetic biologyPublic

Technical Resources

OrganizationResourceFocus AreaURL
CDCBioWatch ProgramAerosol detectioncdc.gov/biowatch
WHODisease Outbreak NewsGlobal surveillancewho.int/emergencies
NISTCybersecurity FrameworkCritical infrastructurenist.gov/cyberframework
HarvardGlobal Health InstituteBiosecurity researchglobalhealth.harvard.edu

Expert Organizations

TypeOrganizationSpecialtyContact
ResearchNuclear Threat InitiativeWMD preventionnti.org
AcademicJohns Hopkins Center for Health SecurityBiosecurity researchcenterforhealthsecurity.org
GovernmentCISACritical infrastructurecisa.gov
InternationalAustralia GroupExport controlsaustraliagroup.net

Footnotes

  1. The framing of pessimistic and optimistic perspectives on biosecurity trajectories draws on competing analytical positions in the biosecurity literature. Sources presenting reasons for concern emphasize AI capability acceleration, governance gaps, and dual-use research proliferation. Sources presenting grounds for measured optimism emphasize the persistence of the synthesis bottleneck, expanding biosurveillance infrastructure, and historically low bioterrorism success rates. Both framings reflect genuine expert disagreement rather than settled consensus. 2 3 4

  2. AI Safety Newsletter Issue 52 reported on an expert virology benchmark evaluating whether frontier AI systems could perform at or above trained virologist level on domain-specific reasoning tasks. The benchmark methodology and findings are discussed in the AI Uplift Assessment Model. The emergence of such evaluations also reflects a growing norm among AI developers toward pre-deployment biosecurity assessment, though standardization and independent auditing of these processes remain open questions. 2 3

References

1RAND Corporation studyRAND Corporation·2024

This RAND Corporation research report examines the risk of AI systems providing meaningful uplift to actors seeking to develop biological weapons, focusing on how to assess capability thresholds and decompose the problem for evaluation purposes. It likely provides a framework for analyzing when AI crosses dangerous capability boundaries in the bioweapons domain and how to structure risk assessments accordingly.

★★★★☆
2DNA synthesis screeningNuclear Threat Initiative

This NTI resource appears to address biosecurity risks from synthetic biology, specifically focusing on DNA synthesis screening as a mechanism to prevent misuse of biological technologies for creating dangerous pathogens or bioweapons. The page is currently unavailable (404), limiting full content analysis.

★★★★☆
3Cybersecurity FrameworkNIST·Government

The NIST Cybersecurity Framework (CSF 2.0) provides structured guidance to help organizations across industry and government understand, manage, and reduce cybersecurity risk. It includes quick-start guides, community profiles, and mappings to other NIST standards, with ongoing extensions into AI cybersecurity and sector-specific applications.

★★★★★
4Harvard Global Health Instituteglobalhealth.harvard.edu

The Harvard Global Health Institute (HGHI) is an interfaculty initiative advancing global health equity through transdisciplinary research, education, and partnerships. It addresses major health challenges including infectious diseases, climate change, and increasingly AI governance in global health contexts.

5NTI Synthetic Biology Report (2024)Nuclear Threat Initiative

This NTI report addresses governance frameworks and risk mitigation strategies for preventing the misuse of synthetic biology for harmful purposes, including bioweapons development. The page appears to be inaccessible (404 error), so full content cannot be verified. Based on NTI's known work, it likely covers biosecurity policy recommendations relevant to dual-use research concerns.

★★★★☆

The Australia Group is an informal multilateral forum of countries working to harmonize export controls on materials, equipment, and technologies that could be used to develop chemical or biological weapons. It maintains common control lists and guidelines to prevent proliferation, complementing the Chemical Weapons Convention (CWC) and Biological Weapons Convention (BWC). The group coordinates member nations' export licensing policies to reduce the risk of dual-use materials contributing to WMD development.

7Johns Hopkins Center for Health Securitycenterforhealthsecurity.org

The Johns Hopkins Center for Health Security is a leading research and policy institution focused on health security, pandemic preparedness, biosecurity, and emerging biological threats. It conducts research, advises policymakers, and trains the next generation of health security professionals. Key focus areas include deliberate biological threats, emerging infectious diseases, biotechnology risks, and global health security.

8Esvelt - Delay, Detect, Defend (2022)PubMed Central (peer-reviewed)·Paper

This resource appears to be misidentified. The URL and content correspond to a 2022 environmental science paper about using aerial and underwater drones to monitor marine litter in coastal waters, not Esvelt's 'Delay, Detect, Defend' biosecurity paper. The actual content is unrelated to AI safety or biosecurity.

★★★★☆

This CDC MMWR report presents preliminary FoodNet surveillance data on nine foodborne illnesses across eight U.S. sites in 2000, covering 29.5 million people. It documents incidence rates for pathogens including Campylobacter, Salmonella, E. coli O157, and Listeria, comparing trends from 1996–1999. The data supports public health monitoring and intervention evaluation for foodborne disease reduction.

The WHO Disease Outbreak News is an official public health surveillance feed providing real-time updates on infectious disease outbreaks globally. It serves as a primary reference for tracking emerging biological threats, pandemic risks, and novel pathogen events. This resource is relevant to biosecurity and existential risk analysis by providing ground-truth data on natural outbreak dynamics.

Anthropic describes their approach to red teaming frontier AI models for catastrophic risks, with particular focus on biological, chemical, nuclear, and radiological (CBRN) threats. The piece outlines methodologies for assessing whether advanced AI could provide meaningful 'uplift' to bad actors seeking to cause mass harm. It represents an early industry effort to systematize safety evaluation for the most severe potential misuse scenarios.

★★★★☆
12CDC BioWatch Programcdc.gov·Government

The CDC BioWatch Program is a U.S. government environmental biosurveillance initiative designed to provide early detection of biological attacks in major metropolitan areas. It monitors air samples for potential bioterrorism agents to enable rapid public health response. The program represents a key component of U.S. biodefense infrastructure.

13Nuclear Threat InitiativeNuclear Threat Initiative

NTI is a nonprofit global security organization focused on reducing risks from nuclear, biological, chemical, and radiological weapons. It conducts policy research, advocates for international security frameworks, and publishes assessments of biosecurity and nuclear security threats. NTI is particularly notable for its Global Health Security Index and nuclear/biosecurity risk analyses relevant to catastrophic and existential risk.

★★★★☆

CISA is the U.S. federal agency responsible for cybersecurity and critical infrastructure protection. It coordinates national efforts to defend against cyber threats, shares threat intelligence, and sets security standards for government and private sector systems. Relevant to AI safety through its work on securing AI-enabled infrastructure and emerging technology risks.

★★★★☆

This Federation of American Scientists resource documents the Soviet Union's clandestine Biopreparat biological weapons program, one of the largest and most sophisticated bioweapons programs in history. It covers the program's scope, organizational structure, pathogens developed, and the risks posed by its legacy after the Soviet collapse.

16Biosecurity Surveillanceglobalhealth.harvard.edu

This Harvard Global Health Institute resource focuses on biosecurity surveillance systems, likely covering monitoring frameworks for biological threats including naturally occurring outbreaks and potential bioweapons. It addresses the intersection of public health infrastructure and national/global security through surveillance and early warning systems.

17CDC's BioWatch programdhs.gov·Government

BioWatch is the U.S. Department of Homeland Security's early-warning biosurveillance system designed to detect the release of biological agents in major cities. The program operates environmental sensors in urban areas to provide advance warning of potential bioterrorism attacks, enabling faster public health response. It represents a key component of U.S. biodefense infrastructure.

Related Wiki Pages

Top Related Pages

Risks

AI ProliferationAutonomous Weapons

Approaches

Capability Unlearning / RemovalDangerous Capability Evaluations

Analysis

Deceptive Alignment Decomposition ModelCarlsmith's Six-Premise ArgumentAI Talent Market DynamicsGoal Misgeneralization Probability ModelScheming Likelihood AssessmentCapabilities-to-Safety Pipeline Model

Other

InterpretabilityGPT-4GPT-4oGeminiClaude

Organizations

Johns Hopkins Center for Health Security

Key Debates

AI Misuse Risk CruxesAI Safety Solution Cruxes

Concepts

AI Timelines