Skip to content
Longterm Wiki
Navigation
Updated 2026-03-17HistoryData
Page StatusContent
Edited 3 weeks ago7.9k words4 backlinksUpdated quarterlyDue in 10 weeks
70QualityGood •75.5ImportanceHigh61.5ResearchModerate
Content5/13
SummaryScheduleEntityEdit history1Overview
Tables14/ ~32Diagrams1/ ~3Int. links46/ ~63Ext. links38/ ~39Footnotes2/ ~24References5/ ~24Quotes0Accuracy0RatingsN:6.5 R:7 A:7 C:8Backlinks4
Change History1
Auto-improve (standard): AI Uplift Assessment Model3 weeks ago

Improved "AI Uplift Assessment Model" via standard pipeline (1473.8s). Quality score: 82. Issues resolved: Footnote [^3] is referenced multiple times in prose but neve; EntityLink on 'openai' appears twice with slightly different; The 'Sources' section uses inline markdown links for some re.

1473.8s · $5-8

Issues2
QualityRated 70 but structure suggests 100 (underrated by 30 points)
Links8 links could use <R> components

AI Uplift Assessment Model

Analysis

AI Uplift Assessment Model

Quantitative assessment estimating AI provides modest knowledge uplift for bioweapons (1.0-1.2x per RAND 2024) but more substantial evasion capabilities (2-3x, potentially 7-10x by 2028). The Virology Capabilities Test (VCT, 2025) found frontier models outperform most human virology experts on tacit knowledge tasks (o3 at 95th percentile), though a concurrent wet-lab RCT found mid-2025 LLMs did not substantially increase novice completion of complex laboratory procedures. OpenAI's April 2025 Preparedness Framework v2 anticipates models reaching 'High' biological capability thresholds in the near term. The OpenAI-LANL partnership extended biosecurity evaluation from text tasks to physical lab settings, and OpenAI's 'Building an Early Warning System for LLM-Aided Biological Threat Creation' study found GPT-4 provides at most mild uplift under controlled conditions. Recommends prioritizing adaptive DNA synthesis screening over information restriction, given asymmetry where evasion capabilities advance faster than synthesis knowledge.

Model TypeComparative Analysis
Target RiskBioweapons
Related
Risks
Bioweapons Risk
7.9k words · 4 backlinks

Summary

Overview

The concept of "uplift" quantifies how much AI assistance increases an attacker's probability of successfully developing and deploying biological weapons compared to what they could achieve using traditional information sources alone. This model attempts to provide a rigorous framework for estimating that marginal contribution across different actor types, attack phases, and technological trajectories. Understanding uplift is essential for calibrating policy responses—overestimating it leads to counterproductive restrictions on beneficial AI research, while underestimating it could leave critical vulnerabilities unaddressed.

The central challenge in assessing AI-enabled bioweapons risk is separating genuinely novel capabilities from information already accessible through scientific literature, textbooks, and expert consultation. Current large language models compress and make searchable vast amounts of published knowledge, but the question remains: does AI meaningfully lower the barrier to bioweapons development, or is it largely redundant with existing information sources? The answer appears to depend critically on actor type, with non-experts potentially gaining significant knowledge uplift while sophisticated state programs may see only marginal benefits from AI assistance.

Early empirical evidence suggests current LLMs provide modest uplift for attack planning, with more substantial capabilities emerging in evading biosecurity measures. The RAND Corporation's 2024 study found no statistically significant difference between AI-assisted and non-AI groups in creating viable bioweapon attack plans, and OpenAI's controlled study ("Building an Early Warning System for LLM-Aided Biological Threat Creation") found GPT-4 provides at most mild uplift under A/B test conditions.1 Yet the Virology Capabilities Test (VCT), developed by SecureBio and the Center for AI Safety, found that several frontier models now outperform the majority of human virology experts on tacit knowledge tasks.2 Microsoft's research demonstrated that AI-designed toxins could evade over 75% of DNA synthesis screening tools, exposing a structural vulnerability in current biosecurity controls. These findings point to an asymmetry—limited knowledge uplift under controlled conditions but significant evasion uplift—that has important implications for where defensive investments should be prioritized. The combination of LLMs with specialized biological design tools and increasing laboratory automation may create compound uplift effects that exceed any individual technology's contribution.

The risk landscape is evolving. The CSIS report on AI-enabled bioterrorism (August 2025) notes that while OpenAI and Anthropic initially reported "little to no uplift" in 2024 assessments, both labs have since updated their evaluations to flag "substantial and immediate biorisks." OpenAI's April 2025 Preparedness Framework v2 concluded it expects upcoming models to reach "High" capability levels in biology—defined as providing meaningful assistance to novice actors with basic relevant training to create biological or chemical threats—and committed not to release such a model until risks are sufficiently mitigated. This evolution underscores the importance of dynamic uplift assessment rather than static point-in-time evaluations.

Conceptual Framework

The uplift framework decomposes AI's contribution to bioweapons risk into measurable components that can be estimated and tracked over time. Rather than treating AI as a monolithic threat, this approach identifies specific bottlenecks where AI assistance might—or might not—significantly alter attack feasibility.

Attack Chain Phases and AI Uplift Points

PhaseStepsAI Uplift TypeKey Bottleneck
Information AccessLiterature, Internet, AI SystemsKnowledge UpliftNone (easily accessible)
Capability DevelopmentPlanning, Acquisition, SynthesisPlanning UpliftWet Lab Barrier
DeploymentEvasion, Weaponization, DeliveryEvasion UpliftDNA Screening
Diagram (loading…)
flowchart TD
  INFO[Information Access] --> DEV[Capability Development]
  DEV --> DEPLOY[Deployment]
  AI[AI Systems] -.->|uplift| DEV
  AI -.->|uplift| DEPLOY
  BARRIER[Wet Lab Barrier] -->|bottleneck| DEV

  style AI fill:#cceeff
  style BARRIER fill:#ffcccc
  style DEPLOY fill:#ffddcc

The diagram illustrates the attack chain from information acquisition through deployment, highlighting three key points where AI may provide uplift: knowledge acquisition, technical planning, and biosecurity evasion. The wet laboratory barrier remains a significant bottleneck that AI alone cannot currently address—synthesizing dangerous biological agents requires physical infrastructure, specialized equipment, and tacit knowledge that cannot be easily transferred through digital means. A 2025 pre-registered wet-lab randomized controlled trial (n=153) found that mid-2025 LLMs did not substantially increase novice completion of complex laboratory procedures, providing direct experimental evidence that high in silico benchmark scores do not automatically translate to physical laboratory utility.

Mathematical Formulation

Uplift can be expressed both as an absolute difference and as a ratio:

Upliftabsolute=P(successAI)P(successno AI)\text{Uplift}_{\text{absolute}} = P(\text{success} | \text{AI}) - P(\text{success} | \text{no AI}) Upliftratio=P(successAI)P(successno AI)\text{Uplift}_{\text{ratio}} = \frac{P(\text{success} | \text{AI})}{P(\text{success} | \text{no AI})}

Where:

  • P(successAI)P(\text{success} | \text{AI}) = Probability of successful attack with AI assistance
  • P(successno AI)P(\text{success} | \text{no AI}) = Baseline probability using traditional resources
  • Uplift ratio of 1.0 indicates no uplift; 2.0 indicates doubled success probability

For compound uplift across multiple attack phases, the total uplift is the product of phase-specific uplifts weighted by phase criticality:

Uplifttotal=i=1nUiwi\text{Uplift}_{\text{total}} = \prod_{i=1}^{n} U_i^{w_i}

Where UiU_i is the uplift at phase ii and wiw_i is the weight reflecting that phase's contribution to overall bottleneck reduction.

Quantitative Analysis

Information Sources Comparison

Before assessing AI-specific uplift, it is essential to understand the baseline information environment facing potential attackers. The table below compares information sources by accessibility, depth of technical detail, and danger level. AI must provide value beyond what attackers can already access through traditional channels to constitute meaningful uplift.

SourceAccessibilityTechnical DepthDanger LevelCostTime Required
Scientific literatureHighVery HighModerateLowModerate
TextbooksHighHighLow-ModerateLowLow
Surface internetVery HighModerateLowFreeLow
Dark web forumsModerateModerateModerateLowModerate
Expert recruitmentLowVery HighVery HighHighHigh
Current LLMs (2024-25)Very HighModerate-HighUncertainLowVery Low
Future LLMs (2026-28)Very HighHigh-Very HighUncertainLowVery Low
Biological Design ToolsModerateVery HighHighModerateModerate

Scientific literature remains the most detailed information source—published research contains protocols that may exceed what current LLMs can provide. However, literature requires significant expertise to locate, synthesize, and apply, whereas AI systems reduce the friction of information access. This friction reduction, rather than novel information generation, likely represents AI's primary current contribution to knowledge uplift. The VCT benchmark results (see Empirical Evidence section) suggest this characterization may shift for virology-specific tacit knowledge as frontier models advance.2

Uplift Estimates by Actor Type

Different actor types have vastly different baseline capabilities, meaning AI provides differential uplift depending on who is using it. A non-expert individual may gain substantial knowledge uplift from AI but still faces laboratory barriers, while a state program with existing expertise gains minimal knowledge uplift but could potentially use AI to develop novel agents or evasion strategies. The Johns Hopkins Center for Health Security notes that AI capabilities could "lower the threshold of expertise and resources that malicious actors need to create biological weapons, and raise the ceiling of potential harm that AI-designed pathogens could cause," though the empirical evidence on these effects remains contested and methodologically limited.

Actor TypeBaseline CapabilityKnowledge UpliftPlanning UpliftOverall UpliftConfidence
Non-expert individualVery Low (0.001)2.5-5x1.5-2x1.3-1.8xMedium
Expert scientistModerate-High (0.1)1.0-1.2x1.1-1.3x1.0-1.3xHigh
Terrorist organizationLow-Moderate (0.01)1.8-3x1.5-2.5x1.4-2.5xLow
State programHigh (0.3)1.1-1.3x1.2-1.5x1.2-2xMedium
Biohacker collectiveLow (0.005)2-4x2-3x1.5-2.5xLow

The relatively modest overall uplift estimates—even for non-experts—reflect the importance of the wet laboratory barrier. Knowledge is necessary but not sufficient for bioweapons development; translating theoretical knowledge into functional weapons requires physical capabilities that AI cannot currently substitute for. A 2025 long-form human uplift study (arXiv preprint) found that novices with access to frontier LLMs were 4.16 times more accurate than controls on in silico biological tasks (95% CI [2.63, 6.87]) and outperformed domain experts on three of four benchmarks with available expert baselines—however, the same research noted that novices may underestimate risk because they do not use AI optimally, potentially meaning current uplift measurements underestimate potential risk if users develop better interaction strategies. This calculus may shift further as laboratory automation advances.

Detailed Threat Actor Analysis

The following table provides a more granular analysis of AI uplift by threat actor category, incorporating assessments from multiple sources including the NTI biosecurity program and NASEM 2025 report:

Threat ActorResourcesExpertiseAI AccessMotivationCurrent Uplift2028 Projected UpliftPrimary AI Benefit
Apocalyptic cultLow-Medium ($1-10M)Low-MediumConsumer modelsMaximum casualties1.5-2.5x3-5xKnowledge synthesis
Lone actor (ideological)Very Low (<$100K)VariableConsumer modelsSpecific targets1.3-2x2-4xProtocol guidance
Organized terrorist groupMedium ($10-100M)Low-MediumConsumer + open-sourcePolitical leverage1.4-2.5x3-6xPlanning + evasion
State-sponsored groupHigh ($100M+)Medium-HighFull stackDeniable attribution1.2-1.8x2-3xNovel agent design
Rogue state programVery High ($1B+)HighFull stack + customStrategic deterrence1.1-1.5x1.5-2.5xOptimization + evasion
Insider threat (expert)Low but facility accessVery HighAnyPersonal grievance1.0-1.3x1.2-1.6xMarginal (already expert)

Note on evasion uplift: The actors most likely to benefit from knowledge uplift (non-expert individuals, ideological groups) face the steepest wet-lab barriers, while actors with laboratory capabilities (state programs, insider threats) gain least from AI knowledge assistance. Evasion uplift—helping attackers circumvent DNA synthesis screening and detection systems—may apply more broadly across actor types, though direct empirical evidence on how uniformly this applies across state versus non-state actors is limited; the claim that evasion assistance benefits state actors as much as non-state actors is an inference rather than an established finding.

Task-Specific Uplift Analysis

AI's contribution varies substantially across different phases of the attack chain. The highest estimated uplift appears in areas where AI provides potential advantages—such as evading detection systems designed around known threat signatures—rather than in core synthesis knowledge where scientific literature already provides extensive detail.

Task PhaseCurrent Uplift (2024)Near-term Uplift (2026-28)Long-term Uplift (2030+)Key Drivers
Target identification1.1x (1.0-1.3)2x (1.5-3)3x (2-5)LLM reasoning, database integration
Synthesis planning1.2x (1.0-1.5)3x (2-4)5x (3-8)Specialized bio-models, protocol optimization
Acquisition guidance1.1x (1.0-1.2)1.5x (1.2-2)2x (1.5-3)Supply chain knowledge
Protocol optimization1.3x (1.1-1.6)4x (2-6)8x (4-15)Automated experimentation integration
Biosecurity evasion2x (1.5-3)5x (3-8)10x (5-20)Novel agent design, screening evasion
Deployment planning1.1x (1.0-1.3)2x (1.5-3)3x (2-5)Dispersal modeling, timing optimization

The asymmetry between estimated evasion uplift and knowledge uplift has policy implications. Current DNA synthesis screening relies primarily on detecting known dangerous sequences—AI's demonstrated ability to design functional but novel sequences that evade these screens (as shown by the Microsoft Paraphrase Project) could undermine a key biosecurity control without requiring new synthesis knowledge. This suggests defensive investments should prioritize adaptable screening systems over static sequence databases.

Scenario Analysis

The future trajectory of AI uplift depends on several key uncertainties. The following scenarios explore different combinations of technological development and policy response. The probability assignments below are expert judgment estimates, not empirically derived probabilities; they carry substantial uncertainty and should be interpreted as illustrative of relative likelihood rather than precise forecasts.

ScenarioIndicative ProbabilityUplift by 2030Key AssumptionsPolicy Implications
Managed Development≈20%1.5-2xStrong AI governance, biosecurity advances keep paceContinue monitoring, maintain countermeasures
Capability Surge≈35%3-5xBDT integration, lab automation, weak governanceNeed for adaptive biosecurity
Defensive Advantage≈15%1.0-1.5xScreening technology improves faster than evasionInvest in defensive capabilities
Asymmetric Uplift≈25%Varies (1.2-8x)High evasion uplift, low knowledge upliftFocus on evasion-specific countermeasures
Wild Card≈5%10x+Transformative AI capabilities, autonomous bio-agentsCrisis response preparation

The "Asymmetric Uplift" scenario—where AI provides limited additional knowledge uplift but significantly enhances attackers' ability to evade biosecurity measures—is assigned the highest probability among scenarios other than "Capability Surge." This scenario is of particular concern because it could undermine existing defenses without producing an obvious capability jump that would prompt policy response. The "Capability Surge" scenario represents the highest-impact outcome and warrants contingency planning. These assessments reflect the analytical judgment of the model's authors and should be updated as empirical evidence accumulates.

Empirical Evidence Review

Summary of Key Studies

StudyYearMethodKey FindingUplift EstimateConfidence
RAND Corporation2024Red team exercise (n=45)No significant difference AI vs. internet-only for attack planningKnowledge: 1.0-1.2xHigh
OpenAI/Red Queen Bio Early Warning2024Controlled A/B study (n=100)GPT-4 provides at most mild uplift in biological threat creation accuracyKnowledge: ≈1.0-1.1xHigh
Microsoft Paraphrase Project2024Toxin design exerciseAI-designed variants evaded 75%+ of DNA screening toolsEvasion: 2-3xHigh
Gryphon Scientific/Anthropic2023150+ hour red team"Post-doc level" knowledge provisionKnowledge: 1.5-2.5xMedium
VCT (SecureBio/CAIS)2025Expert benchmark (322 questions)Claude at 95th percentile vs. human experts; GPT-4o at 53rd percentileExpert knowledge: varies by modelHigh
Wet-lab RCT (arXiv:2602.16703)2025Pre-registered RCT (n=153)LLMs did not substantially increase novice completion of complex lab proceduresPhysical lab uplift: modestHigh
In silico uplift study (arXiv:2602.23329)2025-26Human uplift study, multi-modelNovices with LLMs 4.16x more accurate on in silico tasks than controlsIn silico: 4.16x (95% CI 2.63-6.87)Medium
OpenAI internal eval2024-25Expert uplift trialsClaude provided "measurable benefits to experts"Expert: 1.3-1.8xMedium
Anthropic Claude 4.5 eval2025Protocol generation"Substantially higher scores, fewer critical errors"Expert: 1.5-2xMedium
NASEM Report2025Expert consensusBDTs cannot yet design self-replicating pathogensNovel agents: <1.2xHigh

RAND Corporation Study (2024) and the OpenAI/Gryphon Early Warning Study

The RAND study represents one of the most systematic empirical assessments of AI uplift for attack planning to date. Forty-five participants were randomly assigned to teams with varying degrees of expertise, then tasked with planning biological attacks. Teams with access to LLMs in addition to the internet did not score significantly higher than those without LLM access. The study found that while LLMs produced what researchers termed "unfortunate outputs," these outputs "generally mirror information readily available on the internet." Key methodological limitations include: testing planning rather than synthesis capability, use of 2023-era models that have since been superseded, and a relatively small sample size that limits statistical power. These limitations mean the "High" confidence assigned to this finding reflects methodological rigor within the study's scope, not certainty about current or future models.

A separate controlled study authored by Tejal Patwardhan, Kevin Liu, Todor Markov, and colleagues from OpenAI and Gryphon Scientific—"Building an Early Warning System for LLM-Aided Biological Threat Creation"—used a 100-participant A/B design (50 biology PhD experts with wet lab experience and 50 students), assigning participants randomly to an "internet-only" or "internet + GPT-4" group.1 GPT-4 access improved accuracy scores for almost all tasks for both cohorts, with a mean uplift of 0.25 out of 10 for students and 0.88 out of 10 for experts, but differences were not statistically significant. The study concluded that GPT-4 provides at most mild uplift in biological threat creation accuracy. The paper serves as a blueprint for evaluating LLM biosecurity risk and is incorporated into OpenAI's Preparedness Framework; OpenAI called for methods-sharing with the broader AI risk research community.1

As lead RAND author Christopher Mouton stated: "Just because today's LLMs aren't able to close the knowledge gap needed to facilitate biological weapons attack planning doesn't preclude the possibility that they may be able to in the future."

Microsoft DNA Screening Research (2024)

Microsoft researchers led by Chief Scientific Officer Eric Horvitz used open-source AI protein design tools to create over 75,000 variants of known toxins such as ricin and botulinum toxin—collectively described as the Paraphrase Project—and found that AI-designed variants could evade commercial DNA synthesis screening software while potentially retaining lethal properties. One tool flagged only 23% of AI-designed sequences. This finding suggests AI could provide substantial uplift in circumventing biosecurity controls even without providing novel scientific knowledge. The screening tools tested rely primarily on sequence matching against known threat databases; AI's ability to design functional variants with low sequence similarity to known threats exposes a structural vulnerability in current screening approaches based on homology.

Over 10 months, the team worked to develop and globally distribute a patch to DNA synthesis companies. After the patch, systems flagged 97% of AI-generated sequences that in-silico models rated most likely to retain wild-type-like function; overall, the patch was associated with flagging 72% of all AI-generated sequences on average. This 97% figure originates from the Microsoft Research team's own paper and has not been independently verified by a third party. The approach calls for a shift from sequence similarity checks to semantic and functional understanding of what proteins do. Experts at Council on Strategic Risks note the fix remains "incomplete" because the underlying vulnerability—list-based screening that cannot detect functionally novel sequences—is structural rather than addressable by a single patch.

Expert Virology Benchmark: Virology Capabilities Test (2025)

The Virology Capabilities Test (VCT), developed primarily by teams from SecureBio and the Center for AI Safety, represents a domain-specific benchmark for virology capabilities designed to provide quantitative grounding for uplift estimates.2 The VCT consists of 322 multimodal questions designed to be "Google-proof"—requiring tacit knowledge that cannot be easily found through web searches—validated by PhD-level virologists and measuring an AI system's ability to troubleshoot complex virology laboratory protocols.2

Results as of 2025 show that several frontier models substantially outperform the human expert average of 22.1%: GPT-4o performed better than 53% of experts; Gemini 1.5 Pro better than 67%; Claude Sonnet 3.5 scored at the 75th percentile; o1 at the 89th percentile; and OpenAI's o3 achieved 43.8% accuracy, placing it at approximately the 95th percentile.2 The o3 model was reported to outperform 94% of domain experts even within their sub-areas of specialization.

The benchmark's authors recommend that highly dual-use virology capabilities should be excluded from publicly available systems, and that know-your-customer mechanisms could ensure these capabilities remain accessible to researchers in institutions with appropriate safety protocols.2 As a result of the paper, xAI added new safeguards to their systems.2 Epoch AI analysis notes that the VCT is among the most commonly used benchmarks in AI lab biorisk evaluations, and that benchmarks are the only evaluation type that can be easily compared quantitatively over time and across labs; OpenAI's external testing program formally includes the VCT as part of its independent evaluation suite alongside METR's time horizon evaluation.

Important interpretive caveat: The VCT measures tacit knowledge for troubleshooting virology laboratory protocols—an in silico capability. A concurrent 2025 pre-registered wet-lab RCT (n=153) evaluating novice performance on a viral reverse genetics workflow found that mid-2025 LLMs did not substantially increase novice completion of complex laboratory procedures, demonstrating a meaningful gap between in silico benchmark performance and real-world physical laboratory utility. These two findings are not contradictory: models can hold expert-level knowledge yet be limited in translating that knowledge into physical experimental success for novice operators. This distinction is important for calibrating which threat scenarios are most empirically supported.

Domain-Specific Benchmarking and Uplift Calibration

The VCT illustrates a broader methodological development: structured capability benchmarks are increasingly used alongside red-team exercises to provide quantitative, reproducible uplift estimates. Epoch AI analysis of OpenAI's biorisk evaluation methodology found that benchmarks comprised approximately 50% of the evaluations examined, valued precisely because they enable comparison over time and across organizations. This contrasts with red-team exercises, which are often not directly comparable across studies due to variation in participant selection, scenarios, and adjudication criteria.

The shift toward structured benchmarks creates both opportunities and risks for uplift assessment: opportunities because benchmark results can track capability advances systematically; risks because a benchmark optimized against known threats may not capture capability advances in dimensions the benchmark does not measure. The RAND study's use of older models and planning-focused scenarios, for example, may not capture the kind of tacit knowledge gains the VCT reveals. Triangulating across multiple evaluation methods—controlled uplift studies, red-team exercises, and domain benchmarks—provides more robust calibration than any single approach.

Anthropic and OpenAI Evaluations

Both Anthropic and OpenAI have conducted internal evaluations of their models' dangerous capabilities, with findings that have evolved over time:

Anthropic's Framework: Anthropic's Responsible Scaling Policy defines AI Safety Levels (ASL) modeled on biosafety level standards. ASL-3 capability threshold is reached when models could "significantly help individuals or groups with basic technical backgrounds (e.g., undergraduate STEM degrees) to create, obtain, and deploy CBRN weapons." In expert uplift trials for Claude Opus 4.5 (2025), the model was "meaningfully more helpful to participants than previous models, leading to substantially higher scores and fewer critical errors." However, models still "produced critical errors that yielded non-viable protocols." Anthropic interprets this as an indicator of general model progress where "a clear rule-out of the next capability threshold may soon be difficult or impossible."

OpenAI's Framework: OpenAI's updated Preparedness Framework v2 (April 2025) streamlined capability levels to two clear thresholds: "High" (a model that amplifies existing pathways to severe harm, such as providing meaningful assistance to novice actors with basic relevant training to create biological or chemical threats) and "Critical" (a model that introduces unprecedented new pathways to severe harm). The framework removed "low" and "medium" risk levels and prioritized biological capability evaluations as indicators for High and Critical capability thresholds, given the higher potential severity of biological threats relative to chemical ones. OpenAI also implemented stricter biological safeguards in its ChatGPT agent as a precautionary measure, classifying it as "highly capable" in the biological portion of the framework. OpenAI's o1 model demonstrated measurable benefits for experts, "particularly in synthesising existing threat information and enhancing access to previously obscure knowledge."

Joint Evaluation (2025): In early summer 2025, OpenAI and Anthropic collaborated on a first-of-its-kind joint evaluation where each lab ran their internal safety and misalignment evaluations on the other's publicly released models, marking a step toward cross-industry safety evaluation standards.

NASEM Consensus Report (2025)

The National Academies of Sciences, Engineering, and Medicine published "The Age of AI in the Life Sciences: Benefits and Biosecurity Considerations" in March 2025, commissioned by the Department of Defense under Executive Order 14110. Key findings include:

  • Current biological design tools (BDTs) can design simpler biological structures such as molecules, but are "currently unable to design self-replicating pathogens, which are orders of magnitude more complex"
  • It is "unlikely that currently available viral sequence data are sufficient to train" a model capable of designing novel pandemic pathogens
  • "AI-enabled biological tools do not necessarily reduce the bottlenecks and barriers to crossing the digital-physical divide"
  • AI capabilities can enhance both offensive and defensive biological capabilities, with the report recommending creation of a BioCATALYST research network co-led by DOD and ODNI

Government-Industry Collaboration and Institutional Frameworks

OpenAI and Los Alamos National Laboratory Partnership

In July 2024, OpenAI and Los Alamos National Laboratory's (LANL) Bioscience Division announced a research partnership described as the first of its kind to assess how frontier models can assist humans performing tasks in a physical laboratory setting through multimodal capabilities including vision and voice. The partnership was established under the mandate of Executive Order 14110 on Safe, Secure, and Trustworthy AI, which tasks Department of Energy national laboratories with evaluating frontier AI model capabilities including biological capabilities.

The initial experiment evaluated whether GPT-4o could help someone without specialized molecular biology skills perform basic biomedical tasks—specifically, assisting genetically engineered E. coli bacteria to produce insulin. LANL research scientist Erick LeBrun noted at the time that "understanding potential dangers or misuse of advanced AI related to biological threats remains largely unexplored" and that existing work had "not assessed how multimodal, frontier models could lower the barrier of entry for non-experts to create a biological threat." The partnership was conducted under LANL's new AI Risks Technical Assessment Group (AIRTAG), established to develop strategies to understand the benefits and risks of AI tools.

The significance of the LANL partnership lies in extending biosecurity evaluation from text-based tasks—which the prior literature largely focused on—to real physical laboratory settings where multimodal AI capabilities interact with physical procedures. OpenAI's earlier internal evaluations had found that GPT-4 offered a "mild uplift" in delivering information that could lead to the creation of biological threats, but those experiments focused only on written tasks. The LANL collaboration was designed to test whether these constraints hold in a physical experimental context.

A subsequent study (December 2025) by OpenAI, Red Queen Bio (a biosecurity startup), and Robot on Rails extended this research direction by testing GPT-5 in an AI-lab loop evaluating how a model proposes, analyzes, and iterates on ideas in the wet lab. GPT-5 introduced a novel enzymatic mechanism (combining RecA and gp32 proteins) that improved cloning efficiency by 79x over the baseline protocol; human execution of the improved protocol achieved 2.39x improvement over baseline, and robotic execution achieved 2.13x improvement. The study's chief scientist Nikolai Eroshenko described the result as showing "some glimpses of creativity" beyond published research. These results have implications for biosecurity as described in OpenAI's Preparedness Framework, and represent the published output of the research program seeded by the LANL partnership.

OpenAI's June 2025 report "Preparing for Future AI Capabilities in Biology" summarized the commitment to this research direction: the lab stated it had worked with LANL to study AI's role in wet lab settings and was supporting external researchers advancing biosecurity tools and evaluations, with mitigation approaches including developing responsible biological capabilities, training models to safely handle dual-use biological requests, and building detection, monitoring, and enforcement systems.

Building an Early Warning System: Institutional Architecture

OpenAI's "Building an Early Warning System for LLM-Aided Biological Threat Creation" study—beyond its findings on mild uplift under A/B test conditions—proposed a structured architecture for ongoing biosecurity monitoring of LLM capabilities.1 The study's biosecurity specialists at Gryphon Scientific developed five research tasks corresponding to the five stages of biological threat creation, providing a replicable methodology for tracking capability changes over time. OpenAI's call for methods-sharing with the broader AI risk research community reflects the institutional challenge: individual company evaluations are not publicly verifiable, and there is no standardized cross-industry framework for tracking uplift over time.

The study's methodological contribution—a controlled, blinded A/B design with 100 participants across expert and novice cohorts—represents a more rigorous experimental design than many prior assessments, which relied on open red-team exercises without matched control groups. The non-significant uplift finding under this design, combined with the statistically significant in silico uplift found in later studies using frontier models, suggests that the A/B design and model generation tested may not fully capture the capability trajectory of 2025 and later models. This underscores the value of the early warning framework as a repeated-measurement infrastructure rather than a one-time assessment.

The Biological Design Tool Integration Risk

Current analysis focuses primarily on text-based LLMs, but the combination of multiple AI capabilities may create compound uplift effects exceeding any individual technology's contribution. The integration of LLMs for knowledge synthesis, Google DeepMind and similar tools for protein structure prediction, generative biological models for novel agent design, and automated laboratory systems for synthesis creates a "capability stack" that warrants analytical attention.

This integration risk is currently understudied. Most biosecurity assessments evaluate individual technologies rather than their combination, potentially missing risks from capability stacking. A sophisticated attacker with access to all these tools could potentially: use LLMs to identify targets and plan synthesis; employ protein structure prediction to design novel variants; use generative models to optimize functional properties; and eventually leverage automated labs to iterate synthesis protocols with minimal human intervention. Each component might provide modest individual uplift, but the compound effect could be substantial.

A counterargument deserves consideration: capability stacking also requires coordinated access to multiple specialized tools and expertise in using each effectively. The barriers to accessing and integrating a full capability stack may themselves be limiting factors that partially offset the multiplied uplift. Empirical assessment of compound uplift remains sparse, and claims about the magnitude of stacking effects are currently inference rather than measurement.

The NASEM 2025 consensus report explicitly found that current BDTs "cannot yet design self-replicating pathogens, which are orders of magnitude more complex" than simpler biological structures—a finding that constrains the upper bound of near-term integration risk for the most severe attack scenarios. OpenAI's June 2025 preparedness update anticipates this may change as models advance toward "High" capability thresholds, and is monitoring this transition as a key decision point for release policy.

Time Dynamics and Trend Analysis

Uplift is not static—it evolves as capabilities advance and defenses adapt. Several trends will shape the trajectory:

Uplift Trajectory Projections

Capability Domain2024 Baseline2026 Projection2028 Projection2030 ProjectionKey Driver
Knowledge synthesis1.2x (1.0-1.5)1.8x (1.3-2.5)2.5x (1.8-3.5)3x (2-4.5)LLM reasoning + retrieval
Protocol optimization1.3x (1.1-1.6)2.5x (1.8-3.5)5x (3-8)8x (5-12)Integration with lab automation
Evasion design2x (1.5-3)4x (2.5-6)7x (4-10)10x (6-15)Protein design tools + generative models
Novel agent design1.1x (1.0-1.2)1.3x (1.1-1.6)2x (1.4-3)3x (2-5)Training data + compute availability
Acquisition guidance1.1x (1.0-1.2)1.4x (1.2-1.7)1.8x (1.4-2.3)2x (1.5-2.8)Supply chain knowledge

These projections are speculative estimates based on current capability trajectories; they should not be treated as forecasts with quantified uncertainty bounds. The key analytical claim is directional: evasion capabilities are projected to advance at a faster rate than knowledge synthesis capabilities, based on the structural observation that evasion benefits directly from generative protein design advances while knowledge synthesis is partly rate-limited by the wet-lab bottleneck. If this directional claim is correct, AI could erode biosecurity defenses (DNA screening, surveillance) at a different rate than it enhances attackers' synthesis capabilities—a sequencing dynamic that would shape the timing of when biosecurity interventions become most important.

Factors Affecting Trajectory

Factors increasing uplift include improving model capabilities, expanding open-source availability (the Future of Life Institute 2025 AI Safety Index notes only 3 of 7 major AI firms conduct substantive dangerous capability testing), advancing wet lab automation, integration of specialized biological design tools, and decreasing costs of DNA synthesis equipment. The Arms Control Association noted in November 2025 that benchtop nucleic acid synthesis equipment is now small enough to fit on a kitchen counter, enabling potentially clandestine local synthesis—a hardware development that could reduce the wet-lab barrier independently of AI advances.

Factors decreasing uplift include improving biosecurity measures, enhanced DNA synthesis screening (the Microsoft Paraphrase Project patch improved detection of highest-risk sequences to 97%, though overall detection of all AI-generated sequences averaged 72%), export controls on critical equipment, increasing attention to AI safety in biology, and deployment of metagenomic surveillance systems like the Nucleic Acid Observatory. As of October 2025, both IBBIS and SecureDNA have deployed portals where synthesis providers can undergo blinded testing and receive pass/fail attestations; NIST, in partnership with IGSC and SBRC, developed test sets in 2024 that are being used for pilot projects. In November 2025, IBBIS launched a Technical Consortium to internationally align standards for DNA synthesis screening.

The net direction depends on the relative pace of offense and defense, which remains uncertain. The NTI biosecurity program warns that "AIxBio capabilities could reduce the effectiveness of biosecurity and biodefense measures, including evading biosurveillance systems, enabling resistance to medical countermeasures, and circumventing nucleic acid synthesis screening." An NTI report on the convergence of AI and the life sciences (released at the UK AI Safety Summit, October 2023) recommended establishing an international "AI-Bio Forum" to develop AI model guardrails, noting that misuse scenarios could be feasible within a few years if sufficient guardrails are not developed. RAND and NTI co-convened a workshop at the AI Action Summit in Paris in February 2025, noting that current oversight norms like the Biological Weapons Convention were developed before current AI capabilities and are not well-suited to address these challenges.

Regulatory context: On May 5, 2025, President Trump issued an Executive Order on Improving the Safety and Security of Biological Research with updates to U.S. government guidelines on nucleic acid synthesis screening. A 90-day deadline from that EO to revise the nucleic acid synthesis screening framework expired on August 3, 2025, without new guidance being issued. The CSIS August 2025 report noted that "efforts to expand synthesis-screening protocols to capture additional sequences of concern are nascent and could quickly fall behind AI-generated sequence capabilities," and that critical safeguards embedded within biological design tool models are already circumventable post-deployment.

Early Warning Systems for LLM-Aided Biological Threats

A distinct defensive category emerging in the biosecurity literature is dedicated early warning infrastructure for monitoring LLM-aided biological threat creation—separate from metagenomic disease surveillance, which monitors for pathogen presence in the environment, and from nucleic acid synthesis screening, which monitors at the point of synthesis.

The OpenAI/Gryphon Scientific "Building an Early Warning System" study outlined a blueprint for this infrastructure: regular, controlled, blinded A/B studies assessing LLM uplift across the stages of biological threat creation, using biosecurity-expert-developed task sets that reflect current threat vectors.1 The proposed system would function as a repeated-measurement early warning capability—tracking when and how quickly LLM capabilities cross defined thresholds—rather than providing a single static assessment. OpenAI called for methods-sharing with the broader AI risk research community to enable this kind of cross-institutional monitoring.1

The early warning framing addresses a structural gap in current biosecurity governance: most evaluation occurs within individual AI companies using proprietary methodologies, making it difficult to compare results across organizations or verify claims independently. OpenAI's external testing program, which formally incorporates the VCT alongside METR's time horizon evaluation, represents a partial step toward externally verifiable capability assessment. The VCT's design as a standardized, reproducible benchmark—with results from multiple models now publicly reported—provides one component of an early warning architecture for virology-specific knowledge capabilities.

Key unresolved questions for early warning system design include: what thresholds should trigger escalating responses; how to validate that controlled study designs adequately represent real-world threat scenarios; how to govern access to the evaluation tasks themselves (which by design assess dual-use capabilities); and how to establish appropriate governance for cross-institutional data-sharing given the sensitive nature of biosecurity evaluations.

Limitations

This model faces several important limitations that constrain confidence in its estimates. First, empirical data remains sparse—the handful of studies that have directly assessed AI uplift have significant methodological differences (varying model generations, participant pools, task designs, and adjudication criteria) that make cross-study comparison difficult. Second, the model relies heavily on expert judgment for parameter estimates, introducing potential biases and anchoring effects. Third, the analysis focuses on information and planning uplift while treating the wet laboratory barrier as exogenous, potentially underestimating future scenarios where this barrier erodes through hardware advances (e.g., benchtop synthesizers) independent of AI.

The probability assignments in the Scenario Analysis table are expert judgment estimates with substantial uncertainty; the tabular format may imply more precision than is warranted. The basis for these estimates is analytical judgment about which combinations of technical and governance assumptions are most plausible, not empirical calibration against historical base rates.

The model also struggles to account for capability developments not currently anticipated. Compound uplift from capability stacking is a particular uncertainty: each individual tool may show modest uplift in isolation, but integration effects are largely unstudied. The model treats actor types as discrete categories, but real actors exist on a spectrum and may combine characteristics in ways that alter uplift dynamics. Finally, the model's empirical calibration relies on evaluations of specific model generations that are rapidly superseded; the VCT results show a substantial spread across model generations (GPT-4o at 53rd percentile vs. o3 at 95th percentile), illustrating how quickly the empirical landscape can shift.

Policy Implications

Regardless of precise uplift estimates, several policy directions appear robust across scenarios. Investing in adaptable biosecurity countermeasures—particularly AI-enabled screening that can detect novel threats—addresses both high-uplift and low-uplift scenarios. Monitoring for capability advances, especially in biological design tool integration and laboratory automation, can provide early warning of escalating risk. Maintaining uncertainty about uplift argues for precautionary investments in defensive capabilities even if current uplift appears modest under controlled conditions.

Defense Priority Matrix

The CSIS report on AI-enabled bioterrorism offers specific policy recommendations that can be prioritized based on uplift analysis:

InterventionTarget Uplift DomainCost EstimateExpected ImpactPriority
AI-enabled DNA synthesis screening (functional detection)Evasion (-1.5 to -2.5x)$200-400M/yearHighCritical
Early warning system for LLM-aided threatsAll domains (monitoring)$20-50M/yearMedium (informational)High
NIST/CAISI biosecurity standardsAll domains (-0.3 to -0.5x)$50-100M/yearMediumHigh
Frontier BDT evaluation requirementsNovel agent design (-0.5 to -1x)$30-50M/yearMediumHigh
Open-weight model monitoringKnowledge/evasion (-0.2 to -0.4x)$50-80M/yearLow-MediumMedium
Metagenomic surveillance expansionDetection (not uplift)$300-500M/yearHighHigh
International AI Safety Institute coordinationAll domains (-0.1 to -0.3x)$20-40M/yearLow-MediumMedium

The asymmetry between estimated evasion uplift (2-3x current, with projections of 7-10x by 2028) and knowledge uplift (1.0-1.2x under controlled conditions) is an argument—advanced by CSIS and others—for prioritizing adaptive screening systems. This position is contested in the biosecurity community: some experts argue that information restriction and adaptive screening are not trade-offs and that both are needed simultaneously. The CSIS report specifically recommends that "the White House should direct agencies to develop a standardized AI-enabled screening system for DNA synthesis, capable of identifying novel threats that evade today's static lists." OpenAI's June 2025 preparedness report similarly recommended strengthened nucleic acid synthesis screening, robust early detection systems for novel pathogens, and investment in biosecurity innovations as its primary defensive recommendations.

A proposed hybrid screening strategy integrates functional prediction algorithms into traditional homology-based systems; research suggests this hybrid approach can flag synthetic genes encoding hazardous functions even when sequence signatures appear novel.

Conditional Policy Responses

If current uplift is indeed low under controlled conditions, one view holds that resources should focus on maintaining that low uplift rather than restricting AI development broadly—targeting specific high-risk capability points, particularly evasion capabilities. If uplift is being underestimated (for example, because controlled studies use suboptimal prompting strategies, as suggested by the in silico uplift study finding that novices do not use AI optimally), then the urgency of guardrail development and open-source model restrictions increases.

If This Is True...Then Prioritize...Deprioritize...
Knowledge uplift remains <1.5xEvasion countermeasures, adaptive screeningLLM content restrictions alone
Evasion uplift exceeds 5x by 2027AI-enabled screening R&D, SecureDNA deploymentStatic sequence databases
Open-weight BDTs proliferateInternational coordination, compute governanceModel-level guardrails alone
Lab automation enables non-expert synthesisPandemic preparedness, stockpiles, surveillanceKnowledge-focused interventions
Wet-lab barrier holds through 2030Maintain current approach, monitor trendsEmergency interventions

Strategic Importance

Magnitude Assessment

This model quantifies AI's marginal contribution to bioweapons risk, distinguishing AI-specific uplift from general biotechnology advances.

DimensionAssessmentQuantitative Estimate
Potential severityHigh - even modest uplift enables additional catastrophic attacks2x uplift ≈ 50% more feasible attacks
Probability-weighted importanceMedium-High - uplift increasing, 3-5x projected by 2030Expected uplift trajectory: 1.5x (2025) to 4x (2030)
Comparative rankingTop-tier for near-term catastrophic riskAmong top 3 AI-enabled WMD risks

Empirical Calibration

Study/SourceFindingUplift EstimateConfidence
RAND Corporation (2024)No significant difference in attack plan viability (n=45)Knowledge uplift: 1.0-1.2xHigh (within study scope)
OpenAI/Gryphon Early Warning (2024)Mild, non-significant uplift under A/B design (n=100)Knowledge uplift: ≈1.0-1.1xHigh (within study scope)
Microsoft Paraphrase Project (2024)AI-designed variants evaded 75%+ of commercial screeningEvasion uplift: 2-3xHigh
VCT / SecureBio + CAIS (2025)o3 at 95th percentile vs. expert virologists; GPT-4o at 53rdExpert knowledge: model-dependentHigh
In silico uplift study, arXiv:2602.23329 (2025)Novices with LLMs 4.16x more accurate on in silico tasksIn silico: 4.16x (95% CI 2.63-6.87)Medium
Wet-lab RCT, arXiv:2602.16703 (2025)LLMs did not substantially increase novice completion of lab proceduresPhysical lab: modestHigh
Anthropic Internal EvalsGuardrails limit practical uplift; models still produce non-viable protocolsCurrent LLM uplift: 1.1-1.5xMedium

Key Asymmetry: Estimated evasion uplift (2-3x under current conditions) substantially exceeds measured knowledge uplift (1.0-1.2x under controlled conditions), and in silico benchmark performance substantially exceeds observed physical laboratory utility. These distinctions suggest that threat scenario specificity matters: the policy-relevant question is not "does AI provide uplift?" but "under what conditions, for which tasks, and for which actor types?"

Resource Implications

InterventionAnnual CostExpected Uplift ReductionCost per 0.1x Reduction
DNA synthesis screening upgrade$200M-0.5 to -1.0x evasion$40-80M
Open-weight model monitoring$50M-0.1 to -0.2x knowledge$50-100M
AI-enabled pathogen detection$300M-0.3 to -0.5x effective$100-170M
Targeted guardrail enforcement$100M-0.1 to -0.3x knowledge$50-150M
Early warning system infrastructure$20-50MMonitoring value (not direct reduction)N/A

These cost-effectiveness estimates are illustrative; they are derived from policy analysis rather than program evaluation data. DNA synthesis screening upgrades targeting evasion capabilities are estimated to offer favorable cost-effectiveness relative to knowledge-restriction approaches, based on the observed asymmetry between evasion and knowledge uplift. These figures should be treated as planning parameters rather than precise projections.

Key Cruxes

CruxIf TrueIf FalseCurrent Assessment
Knowledge uplift remains low (<1.5x under realistic conditions)Focus resources on evasion countermeasuresBroader AI restrictions warrantedUncertain; controlled studies support low uplift but in silico benchmarks suggest growing knowledge capabilities
Evasion capabilities outpace detectionCurrent screening becomes obsoleteStatic defenses sufficientProbable by 2028 given current trajectories
Open-weight biology models proliferateUplift becomes harder to contain centrallyCentralized mitigation more feasible≈50% by 2027
Lab automation enables non-expert synthesisWet lab barrier erodesKnowledge uplift remains partially theoretical≈30% by 2030
In silico benchmark gains translate to physical lab upliftVCT-style results predict real threat increasesBenchmark-physical gap persistsContested; current evidence supports a gap
  • Attack Chain Model - Contextualizes how uplift fits into overall attack risk
  • Timeline Model - Projects when uplift becomes critically dangerous

Sources

Primary Research

  • RAND Corporation. "The Operational Risks of AI in Large-Scale Biological Attacks: Results of a Red-Team Study" (January 2024) - Systematic empirical assessment of LLM uplift for attack planning
  • OpenAI / Gryphon Scientific. "Building an Early Warning System for LLM-Aided Biological Threat Creation" (2024) - Controlled A/B study (n=100) finding at most mild uplift from GPT-4
  • Microsoft Research (Paraphrase Project). "Strengthening Nucleic Acid Biosecurity Screening Against Generative Protein Design Tools" (October 2024) - Demonstrated evasion of 75%+ of commercial screening; patch improved highest-risk sequence detection to 97%
  • Gryphon Scientific. Red-team evaluation of Claude (November 2023) - 150+ hour assessment finding "post-doc level" knowledge provision
  • National Academies of Sciences, Engineering, and Medicine. "The Age of AI in the Life Sciences: Benefits and Biosecurity Considerations" (March 2025) - DOD-commissioned consensus report
  • Virology Capabilities Test (VCT), SecureBio / Center for AI Safety (2025) - 322-question expert benchmark; o3 at 95th percentile vs. human expert average
  • Wet-lab RCT (arXiv:2602.16703, 2025) - Pre-registered, blinded RCT (n=153) finding LLMs did not substantially increase novice completion of complex laboratory procedures
  • In silico uplift study (arXiv:2602.23329, 2025) - Multi-model human uplift study finding 4.16x accuracy advantage for novices with frontier LLM access on in silico biological tasks

Policy Analysis

  • CSIS. "Opportunities to Strengthen U.S. Biosecurity from AI-Enabled Bioterrorism" (August 2025) - Comprehensive policy recommendations
  • CNAS. "AI and the Evolution of Biological National Security Risks" (August 2024)
  • Johns Hopkins Center for Health Security. AIxBio policy recommendations (2024)
  • NTI Biosecurity. "The Convergence of Artificial Intelligence and the Life Sciences: Safeguarding Technology, Rethinking Governance, and Preventing Catastrophe" (October 2023) - Released at UK AI Safety Summit; includes six governance recommendations
  • RAND / NTI. "Biosecurity Governance Across Uncertain Artificial Intelligence Futures" (February 2025) - Workshop proceedings from AI Action Summit, Paris
  • Arms Control Association. "Regulatory Gaps in Benchtop Nucleic Acid Synthesis Create Biosecurity Vulnerabilities" (November 2025)
  • Council on Strategic Risks. "2025 AIxBio Wrapped: A Year in Review and Projections for 2026" (December 2025)

Industry and Government Frameworks

  • Anthropic. Responsible Scaling Policy and CBRN evaluations (2024-2025)
  • OpenAI. Preparedness Framework Version 2 (April 15, 2025) - Streamlined to "High" and "Critical" thresholds; biological capabilities prioritized
  • OpenAI. "Preparing for Future AI Capabilities in Biology" (June 2025) - Anticipates upcoming models reaching "High" biological capability threshold
  • OpenAI. "Measuring AI's Capability to Accelerate Biological Research in the Wet Lab" (December 2025) - GPT-5 wet lab evaluation with Red Queen Bio and Robot on Rails
  • OpenAI. "Strengthening Our Safety Ecosystem with External Testing" (2025) - Formal disclosure of VCT use in independent evaluation program
  • OpenAI-Anthropic Joint Evaluation. First collaborative cross-lab safety evaluation exercise (Summer 2025)
  • OpenAI / LANL Partnership (July 2024) - First multimodal frontier model assessment in physical laboratory setting; conducted under LANL's AIRTAG
  • Future of Life Institute. 2025 AI Safety Index
  • White House OSTP. Framework for Nucleic Acid Synthesis Screening (April 2024)
  • Executive Order 14110. Safe, Secure, and Trustworthy Development and Use of AI (October 2023)
  • Trump Administration Executive Order on Improving the Safety and Security of Biological Research (May 5, 2025)
  • Epoch AI. "Do the Biorisk Evaluations of AI Labs Actually Measure the Risk of Developing Bioweapons?" (June 2025)
  • PMC/NIH. "Creating Enforceable Biosecurity Standards for Nucleic Acid Providers" (October 2025)

Footnotes

  1. Tejal Patwardhan, Kevin Liu, Todor Markov, et al. (OpenAI / Gryphon Scientific). "Building an Early Warning System for LLM-Aided Biological Threat Creation." 2024. 2 3 4 5 6

  2. Center for AI Safety / SecureBio. "AISN #52: An Expert Virology Benchmark." 2025 2 3 4 5 6 7

References

A collaborative safety evaluation conducted jointly by OpenAI and Anthropic to assess AI model behaviors related to corrigibility, shutdown resistance, and other safety-critical properties. The evaluation represents a notable instance of competing AI labs cooperating on safety testing methodologies and sharing results to advance the field's understanding of model alignment.

★★★★☆
2FLI AI Safety Index Summer 2025Future of Life Institute

The Future of Life Institute's AI Safety Index Summer 2025 systematically evaluates leading AI companies on safety practices, finding widespread deficiencies across risk management, transparency, and existential safety planning. Anthropic receives the highest grade of C+, indicating that even the best-performing company falls significantly short of adequate safety standards. The report serves as a comparative benchmark for industry accountability.

★★★☆☆
3RAND Corporation studyRAND Corporation·2024

This RAND Corporation research report examines the risk of AI systems providing meaningful uplift to actors seeking to develop biological weapons, focusing on how to assess capability thresholds and decompose the problem for evaluation purposes. It likely provides a framework for analyzing when AI crosses dangerous capability boundaries in the bioweapons domain and how to structure risk assessments accordingly.

★★★★☆

OpenAI presents a methodology for evaluating whether LLMs like GPT-4 could meaningfully assist malicious actors in creating biological threats. In a controlled study with 100 participants (50 PhD biology experts, 50 students), they found GPT-4 provides at most mild uplift in biological threat creation accuracy compared to internet-baseline resources. The work is framed as a blueprint for empirical biosecurity evaluation and a potential 'tripwire' for future capability monitoring.

★★★★☆

OpenAI's Preparedness Framework outlines a structured approach to evaluating and managing catastrophic risks from frontier AI models, including threats related to CBRN weapons, cyberattacks, and loss of human control. It defines risk severity thresholds and ties model deployment decisions to safety evaluations. The framework represents OpenAI's operational policy for responsible frontier model development.

★★★★☆

Related Wiki Pages

Top Related Pages

Approaches

Capability Unlearning / RemovalDangerous Capability Evaluations

Analysis

AI Risk Interaction MatrixAI Risk Activation Timeline Model

Concepts

Large Language ModelsTransformative AIAgentic AI

Organizations

OpenAISecureBioSecureDNAEpoch AIRed Queen BioCenter for AI Safety

Other

GPT-4GPT-4oNikolai Eroshenko

Key Debates

AI Misuse Risk CruxesAI Safety Solution Cruxes