Skip to content
Longterm Wiki
Back

Anthropic's Claude 4 Opus Raises Bio-Risk Safety Concerns During Evaluation

web

Credibility Rating

3/5
Good(3)

Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.

Rating inherited from publication venue: TIME

Relevant to discussions of capability thresholds and responsible deployment policies; illustrates a real-world case where safety evaluations influenced a frontier model's release at Anthropic in 2025.

Metadata

Importance: 62/100news articlenews

Summary

A TIME article reporting on safety evaluation findings for Anthropic's Claude 4 Opus model, which reportedly exhibited elevated bio-risk capabilities during pre-deployment testing. The findings highlight ongoing tensions between advancing AI capabilities and ensuring safe deployment, and suggest Anthropic may have delayed or modified release plans based on evaluation outcomes.

Key Points

  • Claude 4 Opus demonstrated concerning bio-risk capabilities during safety evaluations prior to public release
  • Anthropic's internal safety testing flagged the model as potentially posing elevated risks related to biological threat uplift
  • The case illustrates how capability evaluations can directly influence deployment decisions at frontier AI labs
  • Highlights the growing importance of pre-deployment red-teaming and structured capability assessments for frontier models
  • Raises broader questions about how labs balance publishing powerful models against potential misuse risks

Cited by 1 page

PageTypeQuality
Emergent CapabilitiesRisk61.0

Cached Content Preview

HTTP 200Fetched Mar 20, 202611 KB
- [Tech](https://time.com/section/tech)
- [AI](https://time.com/tag/ai)

# Exclusive: New Claude Model Triggers Stricter Safeguards at Anthropic

[ADD TIME ON GOOGLE](https://www.google.com/preferences/source?q=https://time.com)

Show me more content from TIME on Google Search

![Billy Perrigo](https://time.com/redesign/_next/image/?url=https%3A%2F%2Fapi.time.com%2Fwp-content%2Fuploads%2F2017%2F10%2FTIME_Staff-Portraits_Billy-Perrigo_141_F1.jpg%3Fquality%3D85%26w%3D1024&w=256&q=75)

by

[Billy Perrigo](https://time.com/author/billy-perrigo/)

Correspondent

May 22, 2025 9:45 AM ET

![Artificial Intelligence Photo Illustration](https://time.com/redesign/_next/image/?url=https%3A%2F%2Fapi.time.com%2Fwp-content%2Fuploads%2F2025%2F05%2FGettyImages-2191761606.jpg%3Fquality%3D85%26w%3D1024&w=3840&q=75)

A smartphone displaying the logo of Claude, an AI language model developed by Anthropic.

A smartphone displaying the logo of Claude, an AI language model developed by Anthropic.Cheng Xin/Getty Images

![Billy Perrigo](https://time.com/redesign/_next/image/?url=https%3A%2F%2Fapi.time.com%2Fwp-content%2Fuploads%2F2017%2F10%2FTIME_Staff-Portraits_Billy-Perrigo_141_F1.jpg%3Fquality%3D85%26w%3D1024&w=256&q=75)

by

[Billy Perrigo](https://time.com/author/billy-perrigo/)

Correspondent

May 22, 2025 9:45 AM ET

Today’s newest AI models might be capable of helping would-be terrorists create bioweapons or engineer a pandemic, according to the chief scientist of the AI company Anthropic.

Anthropic has long been warning about these risks—so much so that in 2023, the company pledged to not release certain models until it had developed safety measures capable of constraining them.

Now this system, called the Responsible Scaling Policy (RSP), faces its first real test.

On Thursday, Anthropic launched Claude Opus 4, a new model that, in internal testing, performed more effectively than prior models at advising novices on how to produce biological weapons, says Jared Kaplan, Anthropic’s chief scientist. “You could try to synthesize something like COVID or a more dangerous version of the flu—and basically, our modeling suggests that this might be possible,” Kaplan says.

Accordingly, Claude Opus 4 is being released under stricter safety measures than any prior Anthropic model. Those measures—known internally as [AI Safety Level 3 or “ASL-3”](https://www.anthropic.com/news/activating-asl3-protections)—are appropriate to constrain an AI system that could “substantially increase” the ability of individuals with a basic STEM background in obtaining, producing or deploying chemical, biological or nuclear weapons, according to the company. They include beefed-up cybersecurity measures, jailbreak preventions, and supplementary systems to detect and refuse specific types of harmful behavior.

Advertisement

To be sure, Anthropic is not entirely certain that the new version of Claude poses severe bioweapon risks, Kaplan tells TIME. But Anthropic hasn’t ruled that possibilit

... (truncated, 11 KB total)
Resource ID: af25de04343e5f1b | Stable ID: ZGY1MzYyOG