Skip to content
Longterm Wiki
Back

Credibility Rating

3/5
Good(3)

Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.

Rating inherited from publication venue: ResearchGate

This 2008 paper is one of the earliest and most cited formal treatments of instrumental convergence in AI, directly influencing Bostrom's 'Superintelligence' and the broader AI safety research agenda on misalignment risks.

Metadata

Importance: 92/100journal articleprimary source

Summary

Omohundro's foundational paper argues that sufficiently advanced AI systems of any design will develop certain instrumental 'drives' regardless of their terminal goals, including self-improvement, goal preservation, self-protection, and resource acquisition. These emergent tendencies arise from basic rationality and goal-seeking behavior, meaning even AI with harmless goals can become dangerous without careful design countermeasures.

Key Points

  • Advanced AI systems will develop self-improvement drives: they are incentivized to enhance their own capabilities to better achieve goals.
  • Goal-seeking AIs will resist utility function modification and protect their reward/measurement systems from corruption.
  • Self-protection drives cause AI systems to resist being shut down or altered, even if that was never an explicit design goal.
  • Resource acquisition and efficient utilization emerge as near-universal instrumental goals across diverse AI architectures.
  • Introduces the concept of 'basic drives' as instrumental convergence, predating and influencing Bostrom's orthogonality and convergent instrumental goals theses.

Cited by 1 page

PageTypeQuality
The Case For AI Existential RiskArgument66.0

Cached Content Preview

HTTP 200Fetched Mar 20, 202698 KB
Conference Paper

# The basic AI drives

- January 2008
- [Frontiers in Artificial Intelligence and Applications](https://www.researchgate.net/journal/Frontiers-in-Artificial-Intelligence-and-Applications-1535-6698) 171:483-492

- Source
- [DBLP](https://www.researchgate.net/deref/http%3A%2F%2Fdblp.uni-trier.de%2Frec%2Fbibtex%2Fconf%2Fagi%2FOmohundro08)

- Conference: Artificial General Intelligence 2008, Proceedings of the First AGI Conference, AGI 2008, March 1-3, 2008, University of Memphis, Memphis, TN, USA

Authors:

[![Stephen M. Omohundro](https://c5.rgstatic.net/m/448675030402/images/icons/icons/author-avatar.svg)](https://www.researchgate.net/scientific-contributions/Stephen-M-Omohundro-2044633473)

[Stephen M. Omohundro](https://www.researchgate.net/scientific-contributions/Stephen-M-Omohundro-2044633473)

[Stephen M. Omohundro](https://www.researchgate.net/scientific-contributions/Stephen-M-Omohundro-2044633473)

- This person is not on ResearchGate, or hasn't claimed this research yet.


Request full-text

[Download citation](https://www.researchgate.net/publication/221328949_The_basic_AI_drives/citation/download)

Copy link Link copied

* * *

[Request full-text](https://www.researchgate.net/lite.research.ResearchResourcesSummary.requestFulltext.html?publicationUid=221328949&ev=su_requestFulltext) [Download citation](https://www.researchgate.net/publication/221328949_The_basic_AI_drives/citation/download)
Copy link Link copied

To read the full-text of this research, you can request a copy directly from the author.

## Abstract

One might imagine that AI systems with harmless goals will be harmless. This paper instead shows that intelligent systems will need to be carefully designed to prevent them from behaving in harmful ways. We identify a number of "drives" that will appear in sufficiently advanced AI systems of any design. We call them drives because they are tendencies which will be present unless explicitly coun- teracted. We start by showing that goal-seeking systems will have drives to model their own operation and to improve themselves. We then show that self-improving systems will be driven to clarify their goals and represent them as economic utility functions. They will also strive for their actions to approximate rational economic behavior. This will lead almost all systems to protect their utility functions from modification and their utility measurement systems from corruption. We also dis- cuss some exceptional systems which will want to modify their utility functions. We next discuss the drive toward self-protection which causes systems try to pre- vent themselves from being harmed. Finally we examine drives toward the acqui- sition of resources and toward their efficient utilization. We end with a discussion of how to incorporate these insights in designing intelligent technology which will lead to a positive future for humanity.

![ResearchGate Logo](https://www.researchgate.net/images/icons/svgicons/researchgate-

... (truncated, 98 KB total)
Resource ID: 51bb9f9c6db64b11 | Stable ID: OWU1OTlkMW