Back
A Timing Problem for Instrumental Convergence
webAuthors
Rhys Southan·Helena Ward·Jen Semler
Credibility Rating
4/5
High(4)High quality. Established institution or organization with editorial oversight and accountability.
Rating inherited from publication venue: Springer
A peer-reviewed philosophical article examining timing issues in instrumental convergence theory, which is relevant to understanding how and when AI systems might pursue convergent instrumental goals regardless of their final objectives.
Paper Details
Citations
0
Year
2025
Methodology
peer-reviewed
Categories
Philosophical Studies
Metadata
journal articleprimary source
Cached Content Preview
HTTP 200Fetched Mar 20, 202698 KB
[Skip to main content](https://link.springer.com/article/10.1007/s11098-025-02370-4#main)
## Search
Search by keyword or author
Search
## Navigation
- [Find a journal](https://link.springer.com/journals/)
- [Publish with us](https://www.springernature.com/gp/authors)
- [Track your research](https://link.springernature.com/home/)
# A timing problem for instrumental convergence
- [Open access](https://www.springernature.com/gp/open-science/about/the-fundamentals-of-open-access-and-open-research)
- Published: 03 July 2025
- (2025)
- [Cite this article](https://link.springer.com/article/10.1007/s11098-025-02370-4#citeas)
You have full access to this [open access](https://www.springernature.com/gp/open-science/about/the-fundamentals-of-open-access-and-open-research) article
[Download PDF](https://link.springer.com/content/pdf/10.1007/s11098-025-02370-4.pdf)
[Save article](https://link.springer.com/article/10.1007/s11098-025-02370-4/save-research?_csrf=vdQ8xMhMG5p2c9vV2Dw-HxBlJA44VQpf)
[View saved research](https://link.springer.com/saved-research)
[Philosophical Studies](https://link.springer.com/journal/11098) [Aims and scope](https://link.springer.com/journal/11098/aims-and-scope) [Submit manuscript](https://www.editorialmanager.com/phil)
A timing problem for instrumental convergence
[Download PDF](https://link.springer.com/content/pdf/10.1007/s11098-025-02370-4.pdf)
## Abstract
Those who worry about a superintelligent AI destroying humanity often appeal to the instrumental convergence thesis—the claim that even if we don’t know what a superintelligence’s ultimate goals will be, we can expect it to pursue various instrumental goals which are useful for achieving most ends. In this paper, we argue that one of these proposed goals is mistaken. We argue that _instrumental goal preservation_—the claim that a rational agent will tend to preserve its goals because that makes it better at achieving its goals—is false on the basis of _the timing problem_: an agent which abandons or otherwise changes its goal does not thereby fail to take a required means for achieving a goal it has. Our argument draws on the distinction between means-rationality (adopting suitable means to achieve an end) and ends-rationality (choosing one’s ends based on reasons). Because proponents of the instrumental convergence thesis are concerned with means-rationality, we argue, they cannot avoid the timing problem. After defending our argument against several objections, we conclude by considering the implications our argument has for the rest of the instrumental convergence thesis and for AI safety more generally.
### Similar content being viewed by others

### [Superintelligence, instrumental convergence, and the limits of AI apocalypse](h
... (truncated, 98 KB total)Resource ID:
908c9bc04dcf353f | Stable ID: MDM0OTQ5Mz