Longterm Wiki
Updated 2026-01-31HistoryData
Page StatusContent
Edited 13 days ago1.4k words1 backlinks
65
QualityGood
48
ImportanceReference
13
Structure13/15
30121414%10%
Updated every 6 weeksDue in 5 weeks
Issues2
QualityRated 65 but structure suggests 87 (underrated by 22 points)
Links1 link could use <R> components

SecureBio

Organization

SecureBio

A biosecurity nonprofit applying the Delay/Detect/Defend framework to protect against catastrophic pandemics, including AI-enabled biological threats, through wastewater surveillance (Nucleic Acid Observatory) and AI capability evaluations (Virology Capabilities Test). Co-founded by Kevin Esvelt, who also co-founded the legally separate SecureDNA synthesis screening initiative.

1.4k words · 1 backlinks

Quick Assessment

DimensionAssessmentEvidence
Focus AreaAI-bio risks + pathogen surveillanceTwo divisions: AI team (capability evals) and NAO (wastewater monitoring)1
FundingWell-funded by EA sources≈$9.4M from Coefficient Giving across multiple grants234
Policy InfluenceGrowingVirology Capabilities Test adopted by major AI labs; NIST AI Safety Consortium member5
Team≈25 staffExperienced leadership from MIT, biosecurity, and operations backgrounds6
Key ConcernInformation hazard balanceBiosecurity research inherently involves dual-use knowledge7

Overview

SecureBio is a 501(c)(3) nonprofit founded in 2022 by Kevin Esvelt, an MIT professor who invented CRISPR-based gene drive technology.1 The organization works to protect against catastrophic pandemics—both natural and engineered—through a three-part framework: Delay dangerous capability proliferation, Detect novel pathogens early, and Defend through physical and institutional interventions.8

The organization operates two main divisions. The AI & Biotechnology Risks team, led by Seth Donoughe, evaluates how AI systems might accelerate biological threats and develops benchmarks for AI labs. The Nucleic Acid Observatory (NAO), led by Jeff Kaufman, pioneers wastewater-based metagenomic surveillance to detect novel pathogens before they spread widely.6

SecureBio occupies a distinctive position in the biosecurity landscape by explicitly focusing on catastrophic and existential biological risks rather than routine public health threats. The organization also bridges the AI safety and biosecurity communities, recognizing that advances in AI could significantly lower barriers to creating engineered pathogens.5 See the Biosecurity Interventions overview for how SecureBio fits within the broader intervention portfolio.

History

Founding and Early Development (2022-2024)

SecureBio was founded by Kevin Esvelt, who brought credibility from his work on gene drive technology and his security-minded approach to biotechnology. Esvelt had previously argued that the "security mindset was not sufficiently present in most of the bioengineering industry."9

The organization quickly attracted funding from Coefficient Giving, receiving multiple grants:

  • $4,000,000 for general support over three years2
  • $3,430,000 for the Nucleic Acid Observatory program3
  • $1,420,937 for biosecurity research over three years4
  • $570,000 for pathogen early warning systems10

Leadership Transition (2024-Present)

In September 2024, Kevin Esvelt stepped back from SecureBio's Board to focus on his MIT professorship and academic research, though he remains involved as a co-founder.11 Dr. Benjamin Mueller, formerly COO, became Executive Director and Chairman of the Board.

The board was restructured with new members bringing diverse expertise:

  • Christine Parthemore - CEO of Council on Strategic Risks, national security expert on WMD/CBRN threats
  • Michael Specter - Staff writer at The New Yorker, science and public health journalist
  • Liv Boeree - Science communicator and philanthropist11

Core Framework: Delay, Detect, Defend

Delay

The Delay pillar focuses on slowing the spread of dangerous biological capabilities to potential bad actors. Key initiatives include:

SecureDNA: A legally separate Swiss nonprofit co-founded by Esvelt, providing free DNA synthesis screening software to prevent dangerous pathogen creation. The system can detect hazardous genetic sequences down to 30 base pairs.11

AI Capability Restrictions: Working with AI labs to implement appropriate access controls for models that could assist with biological weapon development.5

Detect

The Detect pillar centers on early warning systems for novel biological threats:

Nucleic Acid Observatory (NAO): Uses untargeted metagenomic sequencing of wastewater to identify new and unknown pathogens—including those with extended incubation periods that evade symptom-based surveillance. As of November 2025, NAO monitors 31 sampling sites across 19 US cities and 5 US military facilities in the Indo-Pacific region, sequencing approximately 60 billion read pairs weekly.1112

The approach is deliberately pathogen-agnostic, meaning it can detect threats regardless of whether they are natural, accidentally released, or deliberately engineered.1

Defend

The Defend pillar develops physical and institutional measures to prevent infection and maintain societal resilience:

Far-UVC Research: Previously investigated germicidal ultraviolet light as a passive defense against airborne pathogens (currently inactive).8

Pandemic-Proof PPE: Advocated for improved personal protective equipment standards.8

AI & Biotechnology Risks Program

SecureBio's AI team has become increasingly prominent as concerns about AI-enabled bioterrorism have grown.

Virology Capabilities Test

The team developed the Virology Capabilities Test (VCT), a benchmark measuring AI models' ability to assist with complex virology tasks. A 2025 study using the VCT found that OpenAI's o3 model "outperformed 94 percent of expert virologists" on troubleshooting complex lab protocols, demonstrating an "urgent need for thoughtful access controls."13

The VCT is now used by major AI laboratories for pre-release safety evaluations.11

NIST Engagement

SecureBio participates in the NIST US AI Safety Consortium and submitted formal recommendations including:5

  • Ensuring the AI Risk Management Framework addresses biosecurity risks from foundation models
  • Developing evaluations for CBRN risks including static benchmarks, model-graded evaluations, and task-based evaluations
  • Assessing biological design tools (BDTs) that could lower barriers for non-experts

Collaboration with AI Labs

SecureBio collaborates with frontier AI companies including Anthropic to build evaluation tools and mitigation strategies. Notably, former SecureBio Research Scientist Anjali Gopal, who co-led the AI project, moved to Anthropic's technical staff.6

Team and Organization

Current Leadership

Leadership Team

Ben Mueller
Executive Director & Board Chair
Former COO; fintech strategy/operations background
Seth Donoughe
Director of AI
Leads AI & Biotechnology Risks division
Jeff Kaufman
Director of NAO
Leads Nucleic Acid Observatory; pathogen early warning
Alvaro Morales
Director of Operations
Administrative and operational leadership

Team Structure

DivisionHeadTeam SizeFocus
AI & Biotechnology RisksSeth Donoughe≈12AI capability evaluations, policy recommendations
Nucleic Acid ObservatoryJeff Kaufman≈15Wastewater surveillance, metagenomic sequencing
OperationsAlvaro Morales≈3Administration, finance, HR

Key researchers include Jasper Götting (Head of Research, AI team), Will Bradshaw (Head of Computational Programs, NAO), and James Kremer (Head of Laboratory Science, NAO).6

Notable Departures

  • Kevin Esvelt - Stepped back September 2024 to focus on MIT professorship11
  • Anjali Gopal - Former AI Project Co-Lead, now at Anthropic6

Funding

Major Grants

FunderAmountPurposeYear
Coefficient Giving$4,000,000General support (3 years)2023
Coefficient Giving$3,430,000Nucleic Acid Observatory2023
Coefficient Giving$1,420,937Biosecurity research (3 years)2022
Coefficient Giving$570,000Pathogen early warning2022
Total identified≈$9,420,937

Operating Costs

The AI-Bio evaluation project alone costs approximately $700,000 per 6 months, covering team leads, policy scientists, contractors, and research assistants.14 As of early 2024, SecureBio was actively seeking additional funding to sustain operations.14

Criticisms and Concerns

Information Hazard Dilemma

SecureBio's work inherently involves dual-use knowledge—understanding how to defend against biological threats requires understanding the threats themselves. The broader biosecurity community remains divided on how to handle such information hazards.7

Some experts argue that emphasizing certain risks could "fuel threats rather than mitigate them"—citing how Al Qaeda's bioweapons efforts reportedly began after the United States publicly highlighted biosecurity vulnerabilities.7

SecureBio has been praised for being "mindful of information hazards" in its approach,9 but the tension between transparency and security remains unresolved in the field.

Scope Limitations

Evaluators have noted that SecureBio's expertise "concentrates on biological threats, not financial systems or weapons infrastructure vulnerabilities."14 The organization is not positioned to address all catastrophic risks.

Funding Dependence

Like many EA-aligned organizations, SecureBio relies heavily on Coefficient Giving funding. This concentration creates sustainability risks and potential influence concerns, though no specific issues have been raised.

Key Uncertainties

Key Questions

  • ?Can DNA synthesis screening scale globally before advanced AI models make biological weapon creation significantly easier?
  • ?Will wastewater surveillance provide sufficient warning time to prevent catastrophic pandemic spread?
  • ?How should AI labs balance capability advancement with biosecurity concerns?
  • ?Can information hazard frameworks be developed that enable defensive research without enabling attacks?
  • ?Will SecureBio's AI capability evaluations keep pace with rapidly advancing AI systems?

Relationship to AI Safety

SecureBio represents an important bridge between the biosecurity and AI safety communities. The organization's work on AI capability evaluations directly addresses concerns about AI misuse risks, specifically the potential for AI to lower barriers to biological weapon creation.

The fact that SecureBio staff have moved to AI labs (Anjali Gopal to Anthropic) and that AI labs use SecureBio's evaluation tools suggests meaningful knowledge transfer between these communities.

Sources

Footnotes

  1. SecureBio Official Website - Organization overview and mission 2 3

  2. Coefficient Giving - SecureBio General Support - $4M grant 2

  3. Coefficient Giving - SecureBio NAO - $3.43M grant 2

  4. Coefficient Giving - Biosecurity Research - $1.42M grant 2

  5. SecureBio NIST RFI Submission - AI policy recommendations 2 3 4

  6. SecureBio Team Page - Staff and organizational structure 2 3 4 5

  7. Bulletin of the Atomic Scientists - Biosecurity Community Divided - Information hazard debates 2 3

  8. Founders Pledge - SecureBio Assessment - Impact evaluation 2 3

  9. EA Forum - SecureBio Notes from SoGive - Independent assessment 2

  10. Coefficient Giving - Pathogen Early Warning - $570K grant

  11. Letter from the Executive Director - SecureBio Substack - Leadership transition and current status 2 3 4 5 6

  12. NAO Updates November 2025 - Current scale and detections

  13. Virology Capabilities Test Paper - AI virology benchmark results

  14. SoGive Notes on SecureBio - Funding and operational details 2 3

Related Pages

Top Related Pages

Concepts

Open PhilanthropySocietal ResilienceAI Misuse Risk CruxesSecureDNAResponses/biosecurity

Organizations

Blueprint Biosecurity1Day SoonerCentre for Long-Term ResilienceJohns Hopkins Center for Health Security