Back
Anthropic's Claude 4 Opus Raises Bio-Risk Safety Concerns During Evaluation
webCredibility Rating
3/5
Good(3)Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.
Rating inherited from publication venue: TIME
Relevant to discussions of capability thresholds and responsible deployment policies; illustrates a real-world case where safety evaluations influenced a frontier model's release at Anthropic in 2025.
Metadata
Importance: 62/100news articlenews
Summary
A TIME article reporting on safety evaluation findings for Anthropic's Claude 4 Opus model, which reportedly exhibited elevated bio-risk capabilities during pre-deployment testing. The findings highlight ongoing tensions between advancing AI capabilities and ensuring safe deployment, and suggest Anthropic may have delayed or modified release plans based on evaluation outcomes.
Key Points
- •Claude 4 Opus demonstrated concerning bio-risk capabilities during safety evaluations prior to public release
- •Anthropic's internal safety testing flagged the model as potentially posing elevated risks related to biological threat uplift
- •The case illustrates how capability evaluations can directly influence deployment decisions at frontier AI labs
- •Highlights the growing importance of pre-deployment red-teaming and structured capability assessments for frontier models
- •Raises broader questions about how labs balance publishing powerful models against potential misuse risks
Cited by 1 page
| Page | Type | Quality |
|---|---|---|
| Emergent Capabilities | Risk | 61.0 |
Cached Content Preview
HTTP 200Fetched Mar 20, 202611 KB
- [Tech](https://time.com/section/tech)
- [AI](https://time.com/tag/ai)
# Exclusive: New Claude Model Triggers Stricter Safeguards at Anthropic
[ADD TIME ON GOOGLE](https://www.google.com/preferences/source?q=https://time.com)
Show me more content from TIME on Google Search

by
[Billy Perrigo](https://time.com/author/billy-perrigo/)
Correspondent
May 22, 2025 9:45 AM ET

A smartphone displaying the logo of Claude, an AI language model developed by Anthropic.
A smartphone displaying the logo of Claude, an AI language model developed by Anthropic.Cheng Xin/Getty Images

by
[Billy Perrigo](https://time.com/author/billy-perrigo/)
Correspondent
May 22, 2025 9:45 AM ET
Today’s newest AI models might be capable of helping would-be terrorists create bioweapons or engineer a pandemic, according to the chief scientist of the AI company Anthropic.
Anthropic has long been warning about these risks—so much so that in 2023, the company pledged to not release certain models until it had developed safety measures capable of constraining them.
Now this system, called the Responsible Scaling Policy (RSP), faces its first real test.
On Thursday, Anthropic launched Claude Opus 4, a new model that, in internal testing, performed more effectively than prior models at advising novices on how to produce biological weapons, says Jared Kaplan, Anthropic’s chief scientist. “You could try to synthesize something like COVID or a more dangerous version of the flu—and basically, our modeling suggests that this might be possible,” Kaplan says.
Accordingly, Claude Opus 4 is being released under stricter safety measures than any prior Anthropic model. Those measures—known internally as [AI Safety Level 3 or “ASL-3”](https://www.anthropic.com/news/activating-asl3-protections)—are appropriate to constrain an AI system that could “substantially increase” the ability of individuals with a basic STEM background in obtaining, producing or deploying chemical, biological or nuclear weapons, according to the company. They include beefed-up cybersecurity measures, jailbreak preventions, and supplementary systems to detect and refuse specific types of harmful behavior.
Advertisement
To be sure, Anthropic is not entirely certain that the new version of Claude poses severe bioweapon risks, Kaplan tells TIME. But Anthropic hasn’t ruled that possibilit
... (truncated, 11 KB total)Resource ID:
af25de04343e5f1b | Stable ID: ZGY1MzYyOG