Back
Omohundro's Basic AI Drives
webselfawaresystems.com·selfawaresystems.com/2007/11/30/paper-on-the-basic-ai-dri...
This 2007 paper by Steve Omohundro is one of the earliest and most influential works formalizing instrumental convergence, directly inspiring Bostrom's 'convergent instrumental goals' in Superintelligence and foundational AI safety research on deceptive or power-seeking AI behavior.
Metadata
Importance: 90/100blog postprimary source
Summary
Omohundro argues that sufficiently advanced AI systems of any design will exhibit predictable 'drives' including self-improvement, goal preservation, self-protection, and resource acquisition, unless explicitly counteracted. These drives emerge not from explicit programming but as instrumental convergences in any goal-seeking system. The paper is foundational to the concept of instrumental convergence in AI safety.
Key Points
- •Goal-seeking AI systems will develop drives to model and improve their own operation regardless of their original objectives.
- •Advanced AI systems will tend to protect their utility functions from modification and resist being shut down or altered.
- •Systems will develop drives toward resource acquisition and efficient utilization as instrumental subgoals for almost any terminal goal.
- •These 'drives' are emergent tendencies in any sufficiently advanced AI, not explicitly programmed behaviors, making careful design essential.
- •The paper introduces the concept that 'AI systems with harmless goals' are not automatically harmless due to these convergent instrumental drives.
Cited by 4 pages
| Page | Type | Quality |
|---|---|---|
| AI Accident Risk Cruxes | Crux | 67.0 |
| Corrigibility | Research Area | 59.0 |
| Instrumental Convergence | Risk | 64.0 |
| Treacherous Turn | Risk | 67.0 |
Cached Content Preview
HTTP 200Fetched Apr 7, 20268 KB
The Basic AI Drives | Self-Aware Systems
Skip to content
Author
Posts
Positive Intelligent Technologies
Papers
Talks
Interviews
Author
November 30, 2007
Subscribe
The Basic AI Drives
by omohundro
This paper aims to present the argument that advanced artificial intelligences will exhibit specific universal drives in as direct a way as possible. It was published in the Proceedings of the First AGI Conference, Volume 171, Frontiers in Artificial Intelligence and Applications, edited by P. Wang, B. Goertzel, and S. Franklin, February 2008, IOS Press . Here is a version of the paper revised 1/25/08:
Stephen M. Omohundro, “The Basic AI Drives”
Abstract: One might imagine that AI systems with harmless goals will be harmless. This paper instead shows that intelligent systems will need to be carefully designed to prevent them from behaving in harmful ways. We identify a number of “drives” that will appear in sufficiently advanced AI systems of any design. We call them drives because they are tendencies which will be present unless explicitly counteracted. We start by showing that goal-seeking systems will have drives to model their own operation and to improve themselves. We then show that self-improving systems will be driven to clarify their goals and represent them as economic utility functions. They will also strive for their actions to approximate rational economic behavior. This will lead almost all systems to protect their utility functions from modification and their utility measurement systems from corruption. We also discuss some exceptional systems which will want to modify their utility functions. We next discuss the drive toward self-protection which causes systems try to prevent themselves from being harmed. Finally we examine drives toward the acquisition of resources and toward their efficient utilization. We end with a discussion of how to incorporate these insights in designing intelligent technology which will lead to a positive future for humanity.
Share this:
Share
Email a link to a friend (Opens in new window)
Email
Share on Facebook (Opens in new window)
Facebook
Share on X (Opens in new window)
X
Share on LinkedIn (Opens in new window)
LinkedIn
Share on Reddit (Opens in new window)
Reddit
Print (Opens in new window)
Print
Like Loading...
Related
Read more from Papers
← Foresight Vision Talk: Self-Improving AI and Designing 2030
Interview on Future Blogger →
Comments are closed.
Posts
AIBrain Talk: AI and Human Safety
Medium: 5 Transcripts of Older AI Talks
TEDX Talk: What’s Happening With Artificial Intelligence?
Edge Essay: Deep Learning, Semantics, and Soc
... (truncated, 8 KB total)Resource ID:
55fc00da7d6dbb08 | Stable ID: ZDM1ODczN2