Skip to content
Longterm Wiki
Back

Credibility Rating

3/5
Good(3)

Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.

Rating inherited from publication venue: Future of Life Institute

A structured industry-wide safety benchmarking report from FLI; useful for governance discussions and tracking whether leading AI labs are meeting their stated safety commitments over successive index editions.

Metadata

Importance: 72/100organizational reportanalysis

Summary

The Future of Life Institute evaluated eight major AI companies across 35 safety indicators, finding widespread deficiencies in risk management and existential safety practices. Even top performers Anthropic and OpenAI received only marginal passing grades, highlighting systemic gaps across the industry in preparedness for advanced AI risks.

Key Points

  • Eight AI companies were assessed on 35 safety indicators covering risk management, transparency, and existential safety practices.
  • No company performed strongly overall; Anthropic and OpenAI led but only marginally, indicating industry-wide safety shortfalls.
  • Substantial gaps were identified in existential risk preparedness and long-term safety governance across all evaluated firms.
  • The index provides a comparative benchmark enabling accountability and tracking of safety commitments over time.
  • Results suggest current voluntary safety practices are insufficient relative to the pace of AI capability development.

Review

The AI Safety Index represents a critical effort to systematically evaluate the safety practices of leading AI companies, highlighting significant structural weaknesses in how frontier AI systems are being developed and deployed. The study reveals a clear divide between top performers like Anthropic, OpenAI, and Google DeepMind, and the rest of the companies, with substantial gaps particularly in risk assessment, safety frameworks, and information sharing. The index's most significant finding is the universal lack of credible existential safety strategies among all evaluated companies. Despite public commitments, none of the companies presented explicit, actionable plans for controlling or aligning potentially superintelligent AI systems. The expert panel, comprising distinguished AI researchers, emphasized the urgent need for more rigorous, measurable, and transparent safety practices that go beyond high-level statements and incorporate meaningful external oversight and independent testing.

Cited by 11 pages

1 FactBase fact citing this source

Cached Content Preview

HTTP 200Fetched Apr 4, 202698 KB
AI Safety Index Winter 2025 - Future of Life Institute 
 Skip to content DECEMBER 2025 AI Safety Index

 Winter 2025 Edition

 AI experts rate leading AI companies on key safety and security domains. How to access this content: Full report PDF Two-page summary Explore the index Coverage Contents: Scorecard Key Findings Panellists Indicators Recommendations Methodology Scorecard Company Company grade & score Anthropic C+ 2.67 OpenAI C+ 2.31 Google DeepMind C 2.08 xAI D 1.17 Z.ai D 1.12 Meta D 1.10 DeepSeek D 1.02 Alibaba Cloud D- 0.98 Overall Grade Score Anthropic C+ 2.67 OpenAI C+ 2.31 Google DeepMind C 2.08 xAI D 1.17 Z.ai D 1.12 Meta D 1.10 DeepSeek D 1.02 Alibaba Cloud D- 0.98 Domains Hint : Click on a domain to inspect Risk Assessment 6 indicators Current Harms 7 indicators Safety Frameworks 4 indicators Existential Safety 4 indicators Governance & Accountability 4 indicators Information Sharing 10 indicators Survey Responses B C+ C+ D B- A- B C- C+ D C+ B C+ C C+ D C- C D F D+ F D C D+ D D- F D C- D D+ D+ F D D- - D D+ F F D C- - D D+ F F D+ D+ - Grading : We use the US GPA system for grade boundaries: A, B, C, D, F letter values correspond to numerical values 4.0, 3.0, 2.0, 1.0, 0. How are scores calculated? Chinese Regulatory Context Save Domains Breakdown Hint: View this webpage on desktop for a visual overview of scores across all domains Risk Assessment 6 indicators ↗
 Current Harms 7 indicators ↗
 Safety Frameworks 4 indicators ↗
 Existential Safety 4 indicators ↗
 Governance & Accountability 4 indicators ↗
 Information Sharing 10 indicators ↗
 Survey Responses Antropic OpenAI xAI Google DeepMind Z.ai Index Content

 Full report PDF View the full report in PDF format, including extended content. View file Two-page Summary A quick, printable summary of the report scorecard, key findings, and methodology. View file Watch: Max Tegmark on the Winter 2025 Results FLI’s President Max Tegmark joins AI Safety Investigator Sabina Nong to discuss the importance of driving a ‘race to the top’ on safety amongst AI companies. Video • 01:48 Key findings

 Top takeaways from the index findings: A clear divide persists between top performers and the rest
 A clear divide persists between the top performers (Anthropic, OpenAI, and Google DeepMind) and the rest of the companies reviewed (Z.ai, xAI, Meta, Alibaba Cloud, DeepSeek). The most substantial gaps exist in the domains of risk assessment, safety framework, and information sharing, caused by limited disclosure, weak evidence of systematic safety processes, and uneven adoption of robust evaluation practices.
 Existential safety remains the industry’s core structural weakness All of the companies reviewed are racing toward AGI/superintelligence without presenting any explicit plans for controlling or aligning such smarter-than-human technology, thus leaving the most consequential risks effectively unaddressed.
 Despite public commitments, companies’ safety practices continue to fall short of emerging

... (truncated, 98 KB total)
Resource ID: 97185b28d68545b4 | Stable ID: MDQ5NTdmZW