Yoshua Bengio and others
paperAuthors
Credibility Rating
Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.
Rating inherited from publication venue: arXiv
A high-profile consensus paper co-authored by Turing Award winner Yoshua Bengio, synthesizing expert views on extreme AI risks and calling for urgent combined technical and governance responses; widely cited in AI safety and policy discussions.
Paper Details
Metadata
Abstract
Artificial Intelligence (AI) is progressing rapidly, and companies are shifting their focus to developing generalist AI systems that can autonomously act and pursue goals. Increases in capabilities and autonomy may soon massively amplify AI's impact, with risks that include large-scale social harms, malicious uses, and an irreversible loss of human control over autonomous AI systems. Although researchers have warned of extreme risks from AI, there is a lack of consensus about how exactly such risks arise, and how to manage them. Society's response, despite promising first steps, is incommensurate with the possibility of rapid, transformative progress that is expected by many experts. AI safety research is lagging. Present governance initiatives lack the mechanisms and institutions to prevent misuse and recklessness, and barely address autonomous systems. In this short consensus paper, we describe extreme risks from upcoming, advanced AI systems. Drawing on lessons learned from other safety-critical technologies, we then outline a comprehensive plan combining technical research and development with proactive, adaptive governance mechanisms for a more commensurate preparation.
Summary
This consensus paper by Yoshua Bengio and colleagues argues that advancing AI systems pose extreme risks—including large-scale social harms, malicious misuse, and irreversible loss of human control—that current safety research and governance mechanisms are inadequate to address. The authors propose a comprehensive response combining technical AI safety research with proactive, adaptive governance frameworks, drawing on lessons from other safety-critical technologies.
Key Points
- •Rapid AI progress toward generalist, autonomous systems may soon massively amplify both benefits and risks, including irreversible loss of human control.
- •Current AI safety research is lagging behind capabilities development, and governance initiatives lack mechanisms to prevent misuse or address autonomous systems.
- •Society's response is incommensurate with the scale and speed of anticipated AI progress, as warned by many experts including this group of researchers.
- •The paper draws on safety-critical technology precedents (e.g., nuclear, aviation) to inform a more robust AI risk management framework.
- •A combined approach of technical safety R&D and proactive, adaptive governance institutions is proposed as a minimum adequate response.
Cited by 2 pages
| Page | Type | Quality |
|---|---|---|
| Is AI Existential Risk Real? | Crux | 12.0 |
| AI Governance and Policy | Crux | 66.0 |
Cached Content Preview
# Managing AI Risks in an Era of Rapid Progress
Yoshua Bengio
&Mila - Quebec AI Institute, Université de Montréal, Canada CIFAR AI Chair
Geoffrey Hinton
University of Toronto, Vector Institute
Andrew Yao
Tsinghua University
Dawn Song
UC Berkeley
Pieter Abbeel
UC Berkeley
Yuval Noah Harari
The Hebrew University of Jerusalem, Department of History
Ya-Qin Zhang
Tsinghua University
Lan Xue
Tsinghua University, Institute for AI International Governance
Shai Shalev-Shwartz
The Hebrew University of Jerusalem
Gillian Hadfield
University of Toronto, SR Institute for Technology and Society, Vector Institute
Jeff Clune
University of British Columbia, Canada CIFAR AI Chair, Vector Institute
Tegan Maharaj
University of Toronto, Vector Institute
Frank Hutter
University of Freiburg
Atılım Güneş Baydin
University of Oxford
Sheila McIlraith
University of Toronto, Vector Institute
Qiqi Gao
East China University of Political Science and Law
Ashwin Acharya
Institute for AI Policy and Strategy
David Krueger
University of Cambridge
Anca Dragan
UC Berkeley
Philip Torr
University of Oxford
Stuart Russell
UC Berkeley
Daniel Kahneman
Princeton University, School of Public and International Affairs
Jan Brauner\*
University of Oxford
Sören Mindermann\*
University of Oxford, Mila - Quebec AI Institute, Université de Montréal
###### Abstract
## Abstract
In this short consensus paper, we outline risks from upcoming, advanced AI systems. We examine large-scale social harms and malicious uses, as well as an irreversible loss of human control over autonomous AI systems. In light of rapid and continuing AI progress, we propose urgent priorities for AI R&D and governance.
## Rapid AI progress
In 2019, GPT-2 could not reliably count to ten. Only four years later, deep learning systems can write software, generate photorealistic scenes on demand, advise on intellectual topics, and combine language and image processing to steer robots. As AI developers scale these systems, unforeseen abilities and behaviors emerge spontaneously, without explicit programming
\[ [1](https://ar5iv.labs.arxiv.org/html/2310.17688#bib.bibx1 "")\]. Progress in AI has been swift and, to many, surprising.
The pace of progress may surprise us again. Current deep learning systems still lack important capabilities and we do not know how long it will take to develop them. However, companies are engaged in a race to create generalist AI systems that match or exceed human abilities in most cognitive work
\[ [2](https://ar5iv.labs.arxiv.org/html/2310.17688#bib.bibx2 ""), [3](https://ar5iv.labs.arxiv.org/html/2310.17688#bib.bibx3 "")\].
They are rapidly deploying more resources and developing new techniques to increase AI capabilities. Progress in AI also enables faster progress: AI assistants are increasingly used to automate programming
\[ [4](https://ar5iv.labs.arxiv.org/html/2310.17688#bib.bibx4 "")\], data collection
\[ [5](https://ar5iv.labs.arxiv.org/html/2310.17688#bib.bibx5 ""
... (truncated, 42 KB total)abf8888683dbf163 | Stable ID: ZmJiNGU3ZG