Preparedness Framework
webCredibility Rating
High quality. Established institution or organization with editorial oversight and accountability.
Rating inherited from publication venue: OpenAI
OpenAI's official institutional framework for catastrophic risk evaluation; relevant for understanding how leading AI labs operationalize safety policies and set deployment guardrails for frontier models.
Metadata
Summary
OpenAI's Preparedness Framework outlines a structured approach to evaluating and managing catastrophic risks from frontier AI models, including threats related to CBRN weapons, cyberattacks, and loss of human control. It defines risk severity thresholds and ties model deployment decisions to safety evaluations. The framework represents OpenAI's operational policy for responsible frontier model development.
Key Points
- •Defines 'Preparedness' as the function responsible for tracking, evaluating, and forecasting catastrophic risks from frontier AI models.
- •Establishes risk categories including CBRN (chemical, biological, radiological, nuclear), cybersecurity, model autonomy, and societal disruption.
- •Sets deployment thresholds: models rated 'critical' risk cannot be deployed; 'high' risk models require safeguards before release.
- •Introduces a Safety Advisory Group and oversight structure to review evaluations and recommend deployment decisions to leadership.
- •Represents a living policy document subject to revision as capabilities and understanding evolve.
Cited by 6 pages
| Page | Type | Quality |
|---|---|---|
| AI Uplift Assessment Model | Analysis | 70.0 |
| AI Risk Critical Uncertainties Model | Crux | 71.0 |
| Alignment Evaluations | Approach | 65.0 |
| Corporate AI Safety Responses | Approach | 68.0 |
| Dangerous Capability Evaluations | Approach | 64.0 |
| Bioweapons Risk | Risk | 91.0 |
Cached Content Preview
OpenAI
April 15, 2025
[Publication](https://openai.com/research/index/publication/) [Safety](https://openai.com/news/safety-alignment/)
# Our updated Preparedness Framework
Sharing our updated framework for measuring and protecting against severe harm from frontier AI capabilities.
[Read full document](https://cdn.openai.com/pdf/18a02b5d-6b67-4cec-ab64-68cdfbddebcd/preparedness-framework-v2.pdf)

Share
We’re releasing an update to our Preparedness Framework, our process for tracking and preparing for advanced AI capabilities that could introduce new risks of severe harm. As our models continue to get [more capable](https://openai.com/index/deep-research-system-card/), safety will increasingly depend on having the right real-world safeguards in place.
This update introduces a sharper focus on the specific risks that matter most, stronger requirements for what it means to “sufficiently minimize” those risks in practice, and clearer operational guidance on how we evaluate, govern, and disclose our safeguards. Additionally, we introduce future-facing research categories that allow us to remain at the forefront of understanding emerging capabilities to keep pace with where the technology is headed. We will continue investing deeply in this process by making our preparedness work more actionable, rigorous, and transparent as the technology advances.
We’ve learned a great deal from our own testing, insights from external experts, and lessons from the field. This update reflects that progress. In line with [our core safety principles](https://openai.com/safety/how-we-think-about-safety-alignment/#our-core-principles), it makes targeted improvements that include:
- **Clear criteria for prioritizing high-risk capabilities.** We use a structured risk assessment process to evaluate whether a frontier capability could lead to severe harm and we assign it to a category based on defined criteria. We track capabilities that meet five key criteria that make it a priority for us to prepare in advance: the risk should be plausible, measurable, severe, net new, and instantaneous or irremediable. We measure progress on these capabilities, and build safeguards against the risks that these capabilities create.
- **Sharper capability categories.** We've updated our categorization of capabilities to apply these criteria and reflect our current understanding.
- **Tracked Categories:** These are established areas where we have mature evaluations and ongoing safeguards. They are Biological and Chemical capabilities, Cybersecurity capabilities, and AI Self-improvement capabilities. We continue to believe some of the most transformative benefits from AI will come from its use in science, engineering, and research - including from capabilities in our Tracked Categories. Investing
... (truncated, 7 KB total)ded0b05862511312 | Stable ID: OTU2ZjU2MD