Nick Bostrom
Nick Bostrom
Biographical profile of Nick Bostrom covering his founding of the Future of Humanity Institute, his 2014 book 'Superintelligence' on AI existential risk, and key philosophical contributions including the orthogonality thesis, instrumental convergence, and treacherous turn concepts.
Nick Bostrom
Biographical profile of Nick Bostrom covering his founding of the Future of Humanity Institute, his 2014 book 'Superintelligence' on AI existential risk, and key philosophical contributions including the orthogonality thesis, instrumental convergence, and treacherous turn concepts.
Background
Nick Bostrom is a Swedish-born philosopher at Oxford University who founded the Future of Humanity Institute (FHI) in 2005. He received his PhD in Philosophy from the London School of Economics in 2000.1
Academic positions:
- Professor of Philosophy at Oxford University
- Director of FHI (2005-2024)
- Published in journals including Philosophy & Public Affairs, Philosophical Quarterly, and International Journal of Forecasting
His 2014 book "Superintelligence: Paths, Dangers, Strategies" examined risks from advanced AI and was reviewed in publications including Science and the New York Times.2
Major Contributions
Superintelligence (2014)
The book analyzed potential development paths for artificial general intelligence and superintelligence, examining control mechanisms and failure modes. It introduced concepts including:
- The orthogonality thesis (intelligence and goals as independent)
- Instrumental convergence (convergent instrumental goals across different systems)
- The treacherous turn (delayed behavioral changes in capable systems)
The book was read by technology leaders including Elon Musk and Bill Gates, who both commented publicly on its content.3 According to Google Scholar, the book has been cited over 5,000 times in academic literature as of 2024.4
Existential Risk Framework
Bostrom's 2002 paper "Existential Risks: Analyzing Human Extinction Scenarios and Related Hazards" provided definitions and classification systems for existential risks.5 His 2013 paper "Existential Risk Prevention as Global Priority" argued that reducing extinction risks should receive higher priority in resource allocation due to impacts on future generations.6
Key Philosophical Concepts
Orthogonality Thesis: Bostrom argues that intelligence level and goal content are independent variables—a system with any level of intelligence could in principle be paired with any final goal.7
Instrumental Convergence: The hypothesis that advanced agents with diverse final goals would pursue similar intermediate goals, including resource acquisition and self-preservation, because these goals are instrumentally useful for many different objectives.8
Treacherous Turn: A scenario in which an AI system behaves cooperatively during a period of limited capability, then pursues different objectives once it becomes sufficiently capable to overcome external constraints.9
Simulation Hypothesis
Bostrom's 2003 paper "Are You Living in a Computer Simulation?" presented a trilemma: either (1) civilizations typically go extinct before developing simulation capabilities, (2) advanced civilizations typically choose not to run ancestor simulations, or (3) we are likely living in a simulation.10
Views on AI Risk
Core Arguments in Superintelligence
The book's central arguments include:11
- Intelligence substantially exceeding human levels is physically possible
- Systems pursuing misaligned goals could cause catastrophic outcomes
- Controlling superintelligent systems presents substantial technical challenges
- Alignment problems should be addressed before advanced AI development
- The potential consequences of misalignment include human extinction
Approach to Timelines
In "Superintelligence," Bostrom surveyed expert opinions showing wide disagreement on development timelines, with median estimates ranging from 2040 to 2050 for human-level AI in different surveys.12 He has emphasized uncertainty in timeline predictions while arguing that preparation is valuable even for low-probability scenarios.
Control and Alignment Approaches
"Superintelligence" examined several categories of approaches:13
- Capability control: Physical or informational constraints limiting system actions
- Motivation selection: Methods for specifying system objectives
- Value learning: Systems that learn human values through observation
- Whole brain emulation: Brain-based systems as an alternative development path
The book expressed skepticism about simple control mechanisms and emphasized the technical difficulty of alignment.
Influence and Impact
Academic Field Building
The Future of Humanity Institute operated from 2005 to 2024, publishing research on existential risks and supervising doctoral students in philosophy and related fields. According to the FHI website, the institute produced over 200 publications during its operation.14
Book Reception
"Superintelligence" appeared on the New York Times bestseller list in 2014.15 The book has been translated into multiple languages and cited in academic papers across computer science, philosophy, and policy studies.
Technology leaders who publicly commented on the book include:
- Bill Gates, who listed it as one of five books to read in summer 201516
- Elon Musk, who recommended it on Twitter in 201417
- Sam Altman, who referenced it in blog posts about AI development18
Policy Engagement
Bostrom has presented to government bodies and international organizations. His work has been cited in policy documents including the UK Government Office for Science's 2016 report on artificial intelligence.19
Other Research Areas
Beyond AI safety, Bostrom has published on:
- Human enhancement: Ethical issues in cognitive and biological enhancement technologies20
- Global catastrophic risks: Analysis frameworks for nuclear war, pandemics, and asteroid impacts21
- Information hazards: Risks from publication or discovery of certain information22
- Anthropic reasoning: Methodological issues in reasoning under observer selection effects23
Controversies and Criticism
2023 Email Controversy
In January 2023, a 1996 email from Bostrom resurfaced in which he had made offensive statements comparing intelligence across racial groups. Bostrom published an apology stating the email was "completely reprehensible" and that he rejected "in the strongest possible terms" the views expressed in it.24 The incident led to discussions about his positions and influence in the field.
FHI Closure (2024)
The Future of Humanity Institute closed in April 2024. According to statements from Oxford University, the closure resulted from administrative and governance disagreements between FHI and the Faculty of Philosophy.25 Former FHI researchers have since joined other institutions including Anthropic, OpenAI, and Google DeepMind.
Academic Reception of Superintelligence
Reviews and responses to "Superintelligence" have included both supportive and critical perspectives:
Critical perspectives include:
- Economist Robin Hanson argued the book overweights scenarios involving rapid capability gain and underweights scenarios of gradual development26
- Computer scientist Oren Etzioni argued the book conflates different types of intelligence and overstates near-term risks27
- Philosopher Daniel Dennett argued the scenarios rely on anthropomorphic assumptions about AI systems28
Supportive perspectives include:
- Philosopher Toby Ord described the book as providing valuable conceptual frameworks for analyzing AI risks29
- Computer scientist Stuart Russell stated the book made important contributions to understanding control problems30
Research Approach Debates
Some researchers have argued that FHI's philosophical approach should be complemented by more technical research on specific alignment mechanisms. Others have argued that conceptual clarification is a necessary foundation for technical work. Both MIRI and later organizations like Anthropic and Redwood Research have pursued more implementation-focused approaches to problems identified in Bostrom's work.
Career Timeline
- 1996-2000: PhD research at London School of Economics
- 2000: PhD awarded (thesis on anthropic reasoning and observation selection effects)
- 2002: Published "Existential Risks" paper in Journal of Evolution and Technology
- 2003: Published simulation hypothesis paper in Philosophical Quarterly
- 2005: Founded Future of Humanity Institute at Oxford
- 2014: Published "Superintelligence: Paths, Dangers, Strategies"
- 2019: Published "The Vulnerable World Hypothesis" in Global Policy
- 2023: Email controversy and public apology
- 2024: FHI closure; current activities not publicly specified
Key Publications
- "Superintelligence: Paths, Dangers, Strategies" (2014, Oxford University Press) - Book examining AI development paths and control problems
- "Existential Risk Prevention as Global Priority" (2013, Global Policy) - Framework for prioritizing existential risk reduction
- "Ethical Issues in Advanced Artificial Intelligence" (2003) - Early analysis of AI alignment challenges
- "Are You Living in a Computer Simulation?" (2003, Philosophical Quarterly) - Simulation hypothesis argument
- "The Vulnerable World Hypothesis" (2019, Global Policy) - Analysis of risks from technological development
- "Existential Risks: Analyzing Human Extinction Scenarios and Related Hazards" (2002, Journal of Evolution and Technology) - Foundational existential risk framework
Footnotes
-
London School of Economics PhD thesis records, 2000. ↩
-
Science review (October 2014) and New York Times review (August 2014). ↩
-
Bill Gates blog post "The Best Books I Read in 2014" (December 2014); Elon Musk Twitter posts (August 2014). ↩
-
Google Scholar citation count for "Superintelligence: Paths, Dangers, Strategies" (accessed December 2024). ↩
-
Bostrom, N. (2002). "Existential Risks." Journal of Evolution and Technology, 9(1). ↩
-
Bostrom, N. (2013). "Existential Risk Prevention as Global Priority." Global Policy, 4(1), 15-31. ↩
-
Bostrom, N. (2014). Superintelligence, Chapter 7. ↩
-
Citation rc-82a0 (data unavailable — rebuild with wiki-server access) ↩
-
Bostrom, N. (2014). Superintelligence, Chapter 8. ↩
-
Bostrom, N. (2003). "Are You Living in a Computer Simulation?" Philosophical Quarterly, 53(211), 243-255. ↩
-
Bostrom, N. (2014). Superintelligence, Chapters 6-14. ↩
-
Citation rc-1a57 (data unavailable — rebuild with wiki-server access) ↩
-
Bostrom, N. (2014). Superintelligence, Chapters 8-13. ↩
-
Future of Humanity Institute website (archived April 2024). ↩
-
New York Times Bestseller List (September 2014). ↩
-
Gates, B. "5 Books to Read This Summer" (gatesnotes.com, May 2015). ↩
-
Elon Musk Twitter post (August 3, 2014). ↩
-
Sam Altman blog posts on AI safety (2015-2016). ↩
-
UK Government Office for Science (2016). "Artificial Intelligence: Opportunities and Implications for the Future of Decision Making." ↩
-
Bostrom, N. (2008). "Why I Want to be a Posthuman When I Grow Up" in Medical Enhancement and Posthumanity. ↩
-
Bostrom, N. & Ćirković, M. (eds.) (2008). Global Catastrophic Risks. Oxford University Press. ↩
-
Bostrom, N. (2011). "Information Hazards: A Typology of Potential Harms from Knowledge." Review of Contemporary Philosophy, 10, 44-79. ↩
-
Bostrom, N. (2002). Anthropic Bias: Observation Selection Effects in Science and Philosophy. Routledge. ↩
-
Bostrom, N. "Apology for 1996 Email" (nickbostrom.com, January 2023). ↩
-
Oxford University Faculty of Philosophy statement (April 2024). ↩
-
Hanson, R. (2014). Review of Superintelligence on overcomingbias.com. ↩
-
Etzioni, O. (2014). "It's Time to Intelligently Discuss Artificial Intelligence" Backchannel (December 2014). ↩
-
Dennett, D. (2015). Washington Post review of Superintelligence (July 2015). ↩
-
Ord, T. (2020). The Precipice: Existential Risk and the Future of Humanity. Hachette Books. ↩
-
Russell, S. (2019). Human Compatible: Artificial Intelligence and the Problem of Control. Viking Press. ↩