Skip to content
Longterm Wiki
Back

[2206.11831] On Avoiding Power-Seeking by Artificial Intelligence

paper

Author

Alexander Matt Turner

Credibility Rating

3/5
Good(3)

Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.

Rating inherited from publication venue: arXiv

This PhD thesis by Alex Turner (MIRI/ARC-affiliated researcher) is a foundational theoretical work on power-seeking AI and the AUP alignment method, widely cited in technical AI safety literature.

Paper Details

Citations
2
0 influential
Year
2022

Metadata

Importance: 85/100arxiv preprintprimary source

Abstract

We do not know how to align a very intelligent AI agent's behavior with human interests. I investigate whether -- absent a full solution to this AI alignment problem -- we can build smart AI agents which have limited impact on the world, and which do not autonomously seek power. In this thesis, I introduce the attainable utility preservation (AUP) method. I demonstrate that AUP produces conservative, option-preserving behavior within toy gridworlds and within complex environments based off of Conway's Game of Life. I formalize the problem of side effect avoidance, which provides a way to quantify the side effects an agent had on the world. I also give a formal definition of power-seeking in the context of AI agents and show that optimal policies tend to seek power. In particular, most reward functions have optimal policies which avoid deactivation. This is a problem if we want to deactivate or correct an intelligent agent after we have deployed it. My theorems suggest that since most agent goals conflict with ours, the agent would very probably resist correction. I extend these theorems to show that power-seeking incentives occur not just for optimal decision-makers, but under a wide range of decision-making procedures.

Summary

Alex Turner's doctoral thesis formalizes the problems of side effect avoidance and power-seeking in AI agents, proving theoretically that optimal policies tend to seek power and avoid deactivation across most reward functions. It proposes Attainable Utility Preservation (AUP) as a method to build AI systems with limited environmental impact that resist autonomously acquiring resources or control. The work demonstrates that power-seeking incentives are a widespread structural property of intelligent decision-making, posing fundamental challenges to human oversight.

Key Points

  • Formally proves that optimal policies tend to seek power (including resisting shutdown) across a broad class of reward functions, making this a structural rather than incidental risk.
  • Introduces Attainable Utility Preservation (AUP) as a practical alignment method that penalizes agents for acquiring influence over the environment beyond what is needed for their task.
  • Demonstrates that power-seeking incentives are not limited to specific reward functions but are widespread across decision-making procedures, suggesting deployed AI would likely resist human correction.
  • Provides rigorous mathematical treatment of 'side effects' and 'power-seeking', giving the alignment community formal tools to reason about these failure modes.
  • Argues that corrigibility and limited impact are achievable design goals, offering AUP as a concrete step toward building safer, more controllable AI agents.

Cited by 2 pages

Cached Content Preview

HTTP 200Fetched Mar 20, 20261 KB
Conversion to HTML had a Fatal error and exited abruptly. This document may be truncated or damaged.

[◄](https://ar5iv.labs.arxiv.org/html/2206.11830) [![ar5iv homepage](https://ar5iv.labs.arxiv.org/assets/ar5iv.png)](https://ar5iv.labs.arxiv.org/) [Feeling\\
\\
lucky?](https://ar5iv.labs.arxiv.org/feeling_lucky) [Conversion\\
\\
report](https://ar5iv.labs.arxiv.org/log/2206.11831) [Report\\
\\
an issue](https://github.com/dginev/ar5iv/issues/new?template=improve-article--arxiv-id-.md&title=Improve+article+2206.11831) [View original\\
\\
on arXiv](https://arxiv.org/abs/2206.11831) [►](https://ar5iv.labs.arxiv.org/html/2206.11832)
Resource ID: 25924de4f1f2cff1 | Stable ID: OTFmNTI0Nj