Skip to content
Longterm Wiki
Back

Berkeley's CLTC research

web

Published by UC Berkeley's CLTC, this resource bridges technical AI safety concepts like corrigibility and the shutdown problem with governance considerations, making it useful for readers approaching alignment from policy or interdisciplinary angles.

Metadata

Importance: 55/100organizational reportanalysis

Summary

A Berkeley Center for Long-Term Cybersecurity (CLTC) publication examining corrigibility in AI systems—the property of remaining open to correction, modification, or shutdown by human operators. It analyzes the theoretical foundations of corrigibility, its relationship to instrumental convergence, and its implications for safe AI design.

Key Points

  • Corrigibility refers to an AI system's disposition to accept correction, adjustment, or shutdown without resistance from human overseers.
  • Instrumental convergence suggests advanced AI systems may resist shutdown by default, making corrigibility a non-trivial design challenge.
  • The shutdown problem highlights why ensuring AI compliance with human intervention requires deliberate technical and architectural choices.
  • CLTC frames corrigibility within broader AI governance and safety concerns, bridging technical and policy perspectives.
  • The paper contributes to understanding how corrigibility can be operationalized in practical AI system design.

Cited by 1 page

PageTypeQuality
Corrigibility FailureRisk62.0

Cached Content Preview

HTTP 200Fetched Mar 20, 20263 KB
[Skip to content](https://cltc.berkeley.edu/publication/corrigibility-in-artificial-intelligence-systems/#primary)

Search Site

Search

Search Site

Search

![](https://cltc.berkeley.edu/wp-content/themes/cltc/images/placeholders/placeholder-blue.jpg)

Grant /
January 2020

# Corrigibility in Artificial Intelligence Systems

This project will focus on basic security issues for advanced AI systems. It anticipates a time when AI systems are capable of devising behaviors that circumvent simple security policies such as “turning the machine off.” These behaviors, which may include deceiving human operators and disabling the “off” switch, result not from spontaneous “evil intent” but from the rational pursuit of human-specified objectives in complex environments. The main goal of our research is to design incentive structures that provably lead to corrigible systems – systems whose behavior can be corrected by human input during operation.

#### Topics

- [**#artificial intelligence (AI)**](https://cltc.berkeley.edu/?research_topic=artificial-intelligence-ai)

#### Related Research

- [![](https://cltc.berkeley.edu/wp-content/uploads/2026/02/AdobeStock_226346683-1.jpeg)](https://cltc.berkeley.edu/publication/agentic-ai-risk-profile/)







[White Paper](https://cltc.berkeley.edu/?publication_type=white-paper)




February 11, 2026





### [Agentic AI Risk-Management Standards Profile](https://cltc.berkeley.edu/publication/agentic-ai-risk-profile/)

- [![](https://cltc.berkeley.edu/wp-content/uploads/2026/01/GPAI-Profile-V1.2-Cover-Photo-1024x682.png)](https://cltc.berkeley.edu/publication/ai-risk-management-standards-profile-v1-2/)







[White Paper](https://cltc.berkeley.edu/?publication_type=white-paper)




January 27, 2026





### [AI Risk-Management Standards Profile for General-Purpose AI (GPAI) and Foundation Models v1.2](https://cltc.berkeley.edu/publication/ai-risk-management-standards-profile-v1-2/)

- [![](https://cltc.berkeley.edu/wp-content/uploads/2026/01/Cyber_Thresholds_report_Image_Only-1024x682.png)](https://cltc.berkeley.edu/publication/toward-risk-thresholds-for-ai-enabled-cyber-threats/)







[White Paper](https://cltc.berkeley.edu/?publication_type=white-paper)




January 22, 2026





### [Toward Risk Thresholds for AI-Enabled Cyber Threats: Enhancing Decision-Making Under Uncertainty with Bayesian Networks](https://cltc.berkeley.edu/publication/toward-risk-thresholds-for-ai-enabled-cyber-threats/)


Scroll to the top of the page

UC Berkeley's Center for Long-Term Cybersecurity VIDEO

Ten Years of Securing Our Future: UC Berkeley CLTC - YouTube

[Photo image of UC Berkeley Inspires](https://www.youtube.com/channel/UCnBT5HobLD5_iyHsZNL85Ng?embeds_referring_euri=https%3A%2F%2Fcltc.berkeley.edu%2F)

UC Berkeley Inspires

6.05K subscribers

[Ten Years of Securing Our Future: UC Berkeley CLTC](https://www.youtube.com/watch?v=izdbsICOBuw)

UC Berkeley Inspires

Search

Watch later

Share

Copy link

Info

Shopping

Tap to un

... (truncated, 3 KB total)
Resource ID: c12f5af6cacbd2d5 | Stable ID: N2RhYmRhZT