Skip to content
Longterm Wiki
Back

Can Preparedness Frameworks Pull Their Weight?

web

Published by the Federation of American Scientists, this policy-oriented analysis is relevant for those tracking how AI labs' internal safety frameworks (e.g., OpenAI's Preparedness Framework) are being scrutinized and whether external governance mechanisms are needed to ensure accountability at scale.

Metadata

Importance: 62/100organizational reportanalysis

Summary

This Federation of American Scientists publication examines whether current AI preparedness frameworks—such as those adopted by major AI labs—are adequate for managing risks as AI systems scale. It analyzes the strengths and limitations of existing evaluation and red-teaming approaches and offers policy recommendations for more robust safety infrastructure.

Key Points

  • Evaluates whether voluntary preparedness frameworks from AI labs are sufficient to address risks from increasingly capable AI systems.
  • Identifies gaps in current capability assessments and red-teaming methodologies that may leave critical risks undetected.
  • Argues that scaling AI capabilities requires commensurately scaling safety evaluation rigor and institutional oversight.
  • Offers policy recommendations for government and industry to strengthen pre-deployment safety requirements.
  • Highlights coordination challenges between labs, regulators, and policymakers in establishing enforceable safety standards.

Cited by 1 page

PageTypeQuality
AI EvaluationsResearch Area72.0

Cached Content Preview

HTTP 200Fetched Mar 20, 202642 KB
A new class of risk mitigation [policies](https://fas.org/accelerator/ai-legislation/) has recently come into vogue for frontier AI developers. Known alternately as [Responsible Scaling Policies](https://metr.org/blog/2023-09-26-rsp/) or Preparedness Frameworks, these policies outline commitments to risk mitigations that developers of the most advanced AI models will implement as their models display increasingly risky capabilities. While the idea for these policies is less than a year old, already two of the most advanced AI developers, Anthropic and OpenAI, have published initial versions of these policies. The U.K. AI Safety Institute asked frontier AI developers about their “Responsible Capability Scaling” policies ahead of the November 2023 [UK AI Safety Summit](https://www.aisafetysummit.gov.uk/policy-updates/#company-policies). It seems that these policies are here to stay.

The National Institute of Standards & Technology (NIST) recently [sought public input](https://www.federalregister.gov/documents/2023/12/21/2023-28232/request-for-information-rfi-related-to-nists-assignments-under-sections-41-45-and-11-of-the) on its assignments regarding generative AI risk management, AI evaluation, and red-teaming. The Federation of American Scientists was happy to provide input; this is the [full text of our response](https://drive.google.com/file/d/1ZWrH2ofx1tfxibcBnxL8CMqEzaPpD4R3/view?usp=sharing). NIST’s request for information (RFI) highlighted several potential risks and impacts of potentially dual-use foundation models, including: “Negative effects of system interaction and tool use…chemical, biological, radiological, and nuclear (CBRN) risks…\[e\]nhancing or otherwise affecting malign cyber actors’ capabilities…\[and i\]mpacts to individuals and society.” This RFI presented a good opportunity for us to discuss the benefits and drawbacks of these new risk mitigation policies.

This report will provide some background on this class of risk mitigation policies (we use the term Preparedness Framework, for reasons to be described below). We outline suggested criteria for robust Preparedness Frameworks (PFs) and evaluate two key documents, Anthropic’s [Responsible Scaling Policy](https://www-cdn.anthropic.com/files/4zrzovbb/website/1adf000c8f675958c2ee23805d91aaade1cd4613.pdf) and OpenAI’s [Preparedness Framework](https://cdn.openai.com/openai-preparedness-framework-beta.pdf), against these criteria. We claim that these policies are net-positive and should be encouraged. At the same time, we identify shortcomings of current PFs, chiefly that they are underspecified, insufficiently conservative, and address structural risks poorly. Improvement in the state of the art of risk evaluation for frontier AI models is a prerequisite for a meaningfully binding PF. Most importantly, PFs, as unilateral commitments by private actors, cannot replace public policy.

## Motivation for Preparedness Frameworks

As AI labs develop potentially dual-use foundation mo

... (truncated, 42 KB total)
Resource ID: bf534eeba9c14113 | Stable ID: ZDc1NmRhZj