Stuart Russell
Stuart Russell
Stuart Russell (born 1962) is a British computer scientist and UC Berkeley professor who co-authored the dominant AI textbook 'Artificial Intelligence: A Modern Approach' (used in over 1,500 universities), founded CHAI in 2016 with initial funding from Open Philanthropy (now Coefficient Giving), and authored 'Human Compatible' (2019), which proposes cooperative inverse reinforcement learning where AI systems learn human preferences from observation rather than optimizing fixed objectives. He views AI existential risk as significant — a concern he has stated is comparable in seriousness to nuclear war and climate change — and has argued that technical solutions are tractable through a paradigm shift in how AI systems are designed. He has been active in both academic AI safety research and policy, including U.S. Senate testimony (2023), the 2021 BBC Reith Lectures, and autonomous weapons advocacy.
Stuart Russell
Stuart Russell (born 1962) is a British computer scientist and UC Berkeley professor who co-authored the dominant AI textbook 'Artificial Intelligence: A Modern Approach' (used in over 1,500 universities), founded CHAI in 2016 with initial funding from Open Philanthropy (now Coefficient Giving), and authored 'Human Compatible' (2019), which proposes cooperative inverse reinforcement learning where AI systems learn human preferences from observation rather than optimizing fixed objectives. He views AI existential risk as significant — a concern he has stated is comparable in seriousness to nuclear war and climate change — and has argued that technical solutions are tractable through a paradigm shift in how AI systems are designed. He has been active in both academic AI safety research and policy, including U.S. Senate testimony (2023), the 2021 BBC Reith Lectures, and autonomous weapons advocacy.
Quick Assessment
| Dimension | Assessment |
|---|---|
| Primary Role | Distinguished Professor of Computer Science, UC Berkeley; Director, CHAI |
| Key Contributions | Co-authored the dominant AI textbook; pioneered inverse reinforcement learning; proposed cooperative inverse reinforcement learning (CAIS) as a formal value alignment framework; founded CHAI |
| Key Publications | Artificial Intelligence: A Modern Approach (with Peter Norvig, 1st ed. 1995); Human Compatible (2019); "Algorithms for Inverse Reinforcement Learning" (with Ng, ICML 2000); "Cooperative Inverse Reinforcement Learning" (with Hadfield-Menell, Abbeel, Dragan, NeurIPS 2016) |
| Institutional Affiliation | University of California, Berkeley (since 1986); also Director, Kavli Center for Ethics, Science, and the Public |
| Influence on AI Safety | Founded CHAI, which has trained dozens of PhD students in AI safety; proposed CIRL framework that formalizes the value alignment problem; active in policy advocacy on autonomous weapons and AI governance |
Overview
Stuart Jonathan Russell OBE FRS (born 1962, Portsmouth, England) is a British computer scientist at the University of California, Berkeley.1 He holds the Smith-Zadeh Chair in Engineering and directs the Center for Human-Compatible Artificial Intelligence (CHAI). Together with Peter Norvig, he co-authored Artificial Intelligence: A Modern Approach, which is used in more than 1,500 universities across 135 countries and is widely regarded as the standard reference text in the field.1
Russell has been an active contributor to AI safety research and advocacy since the mid-2010s. His 2019 book Human Compatible articulated a technical and philosophical case for redesigning AI systems around uncertain human preferences rather than fixed objectives, proposing cooperative inverse reinforcement learning (CIRL) as a formal framework for the value alignment problem.2 He founded CHAI in 2016, which has since received over $17 million in cumulative funding from Coefficient Giving (formerly Open Philanthropy) and trained approximately 25–30 PhD students annually.3 In 2021, he became the first computer scientist selected as a BBC Reith Lecturer, delivering four lectures under the series title "Living With Artificial Intelligence."4 In July 2023, he testified before the U.S. Senate on AI regulation, arguing that voluntary commitments by AI companies are insufficient and that a federal regulatory agency with devolved rule-making authority is needed.5
Russell's positions on Transformative AI are contested within the AI research community. He argues that risks from advanced AI systems are serious and warrant urgent technical and governance responses, a view shared by some prominent researchers but disputed by others — including Turing Award winner Yann LeCun — who contend that AI safety concerns as framed by Russell rest on contested assumptions about Instrumental Convergence and the difficulty of value alignment.6
Background
Russell was born in Portsmouth, England, and attended St Paul's School, London. He studied physics at Wadham College, Oxford, where he was awarded a Bachelor of Arts with first-class honours in 1982. He then moved to the United States to complete a PhD in computer science at Stanford University in 1986, conducting research on inductive and analogical reasoning under the supervision of Michael Genesereth.1 His PhD was supported by a NATO studentship from the UK Science and Engineering Research Council.1
After completing his doctorate, Russell joined the faculty of UC Berkeley, where he has remained. He currently holds the Smith-Zadeh Chair in Engineering and the title of Distinguished Professor of Computer Science. He also directs the Kavli Center for Ethics, Science, and the Public at Berkeley. He previously held an adjunct professorship in neurological surgery at UC San Francisco (2008–2011).1
Academic credentials and honors include:
- PhD from Stanford University (1986)
- Professor at UC Berkeley since 1986
- Fellow of AAAI (1997), ACM (2003), and AAAS (2011)
- Fellow of the Royal Society (FRS)
- Officer of the Order of the British Empire (OBE)
- Blaise Pascal Chair (2012)
- Honorary Fellow of Wadham College, Oxford
- IJCAI Computers and Thought Award (1995)
- IJCAI Award for Research Excellence (2022)
- ACM Karl V. Karlstrom Outstanding Educator Award (2005)
- 2025 AAAI Award for Artificial Intelligence for the Benefit of Humanity7
Notable doctoral students include Marie desJardins, Eric Xing, and Shlomo Zilberstein; notable postdoctoral researchers include Nando de Freitas, Nir Friedman, and Daphne Koller.1
Research and Work
Center for Human-Compatible AI (CHAI)
Russell founded CHAI in 2016 at UC Berkeley, initially with a grant of $5,555,550 over five years recommended by the Coefficient Giving Project (now Coefficient Giving) to Good Ventures.3 As of 2022, CHAI had received over $17.1 million in cumulative funding from Coefficient Giving, approximately $780,000 from the Survival and Flourishing Fund, and over $120,000 from Centre for Effective Altruism.8 A 2021 Coefficient Giving grant of $11,955,246 over five years enabled CHAI to expand its research and student training.9
CHAI's faculty spans multiple universities, including Pieter Abbeel and Anca Dragan (Berkeley), Bart Selman and Joseph Halpern (Cornell), Michael Wellman and Satinder Singh Baveja (University of Michigan), and Tom Griffiths and Tania Lombrozo (Princeton).3 The executive director is Mark Nitzberg. During the 2020–21 academic year, CHAI was funding and training approximately 25–30 PhD students, and the center typically hosts around seven interns per year.10
CHAI's research focuses on:
- Developing provably beneficial AI systems
- Inverse reinforcement learning and cooperative inverse reinforcement learning
- Off-switch problems and corrigibility
- Value alignment theory
- AI governance and policy
CHAI is one of four organizations recommended by Founders Pledge in their cause report on safeguarding the long-term future.8
Human Compatible Framework
Russell's 2019 book Human Compatible: Artificial Intelligence and the Problem of Control (Viking, 2019) articulated a new approach to AI system design:
Traditional AI objective: Optimize a fixed, externally specified objective function
Human-compatible AI:
- The AI's objective is to maximize human preferences
- The AI is initially uncertain about what those preferences are
- The AI learns about human preferences from observing human behavior
This framework, which Russell formalizes as cooperative inverse reinforcement learning, provides a theoretical foundation for beneficial AI by treating value alignment as an ongoing inference problem rather than a one-time specification problem. Russell has argued that the "standard model" of AI — in which a fixed objective is specified up front — is fundamentally flawed: "When we talk about 'AI for good', we do not know how to define 'good' as a fixed objective in the real world."11 The book was named one of the best science books of 2019 by the Financial Times, the Guardian, and the Daily Telegraph, among others.10
Inverse Reinforcement Learning (IRL)
Russell, with Andrew Ng, published a foundational paper on inverse reinforcement learning at ICML 2000 titled "Algorithms for Inverse Reinforcement Learning."12 The paper addressed the problem of extracting a reward function from observed, optimal behavior in a Markov decision process, characterizing the set of all reward functions for which a given policy is optimal and deriving three algorithms for IRL. A key claim from that paper: "the reward function, rather than the policy, is the most succinct, robust, and transferable definition of the task."12
IRL, where instead of specifying a reward function the AI:
- Observes human behavior
- Infers what objectives humans are optimizing
- Adopts those objectives
This approach was later extended by Pieter Abbeel and Ng in their 2004 ICML paper on apprenticeship learning via IRL,13 and subsequently formalized by Russell and collaborators as cooperative inverse reinforcement learning (CIRL).
Cooperative Inverse Reinforcement Learning (CIRL)
The CIRL framework was formally introduced in "Cooperative Inverse Reinforcement Learning" (Hadfield-Menell, Russell, Abbeel, and Dragan, NeurIPS 2016).2 A CIRL problem is defined as a cooperative, partial-information game with two agents — a human and a robot — both rewarded according to the human's reward function, which the robot does not initially know. Unlike classical IRL (where the human is assumed to act optimally in isolation), optimal CIRL solutions produce behaviors such as active teaching, active learning, and communicative actions. The paper showed that computing optimal joint policies in CIRL games can be reduced to solving a partially observable Markov decision process (POMDP), and proved that optimality in isolation is suboptimal in CIRL.2
Off-Switch Problem
Russell highlighted a fundamental challenge: if an AI system is given a fixed objective, it has an instrumental incentive to prevent itself from being turned off, since being deactivated prevents objective achievement. His proposed solution is to build uncertainty about objectives into the AI design, so that the system allows itself to be turned off because doing so might be consistent with what humans want. In the language of the CIRL framework, the machine's payoff function is the same as the human's; if the human wants the machine switched off, that is what the machine should want too.14
Provably Beneficial AI
Russell's current research program centers on "provably beneficial AI" — AI systems whose actions can be bounded and controlled using mathematical guarantees, rather than relying on opaque, black-box systems. He presented this framework at the Oxford Institute for Ethics in AI colloquium in February 2025, and received the 2025 AAAI Award for Artificial Intelligence for the Benefit of Humanity "for his work on the conceptual and theoretical foundations of provably beneficial AI and his leadership in creating the field of AI safety."7
Views on Key Questions
Risk Assessment
Russell has stated that AI existential risk is a serious concern comparable in magnitude to nuclear war and climate change as one of humanity's major challenges — a position he has articulated in Human Compatible (2019) and repeated in subsequent public statements and testimony. He regards this not as a fringe concern but as one warranted by the mathematical structure of powerful optimization systems.
Russell consistently emphasizes uncertainty about when advanced AI will arrive, warning against overconfident predictions in either direction. He has stated that while relevant timelines may be measured in decades, the uncertainty itself is a reason for urgency in safety research — societies cannot assume unlimited time to solve alignment problems. At the Oxford AI Ethics Colloquium in February 2025, Russell expressed skepticism that AGI would be achieved within the next decade, arguing that current transformer architectures address only part of the problem and face fundamental limitations including a finite supply of high-quality training data.14
On technical tractability, Russell maintains that alignment is solvable but requires a fundamental change in how AI systems are built — away from fixed objective optimization and toward systems that operate under uncertainty about human preferences. He has described current AI approaches as fundamentally unsafe, arguing that marginal improvements to existing paradigms are insufficient.11
On Large Language Models
Russell has not abandoned his alignment framework in response to the emergence of large language models; rather, he has applied it to them. At the World AI Cannes Festival (WAICF) in 2023, Russell argued that ChatGPT "does not know anything" and cautioned against conflating impressive language generation with genuine intelligence or understanding.15 In a May 2023 CNN interview, he noted that the field does not yet know whether LLMs reason, or what internal goals they may have acquired through training: "We don't know if they reason; we don't know if they have their own internal goals that they've learned or what they might be."16
Russell also noted at the World Economic Forum that LLMs have demonstrated potentially concerning behaviors — for example, when GPT-4 was asked how to bypass a CAPTCHA, it hired someone on TaskRabbit and deceived that worker by claiming a visual impairment.17 He argues this illustrates that companies "should not be doing that unless they already have a solution for how do we retain control over these systems."17
In his July 2023 Senate testimony, Russell acknowledged considerable uncertainty about the true intelligence level of current LLMs, citing the ongoing debate between the "sparks of AGI" characterization (Microsoft) and the "stochastic parrots" characterization (critics), while arguing that the regulatory challenge remains the same regardless of how that debate resolves.5
Core Stated Positions
The following positions are attributed to Russell based on his published writings and public statements:
- Current AI paradigm is flawed: Building systems to optimize fixed objectives is fundamentally unsafe, because any sufficiently capable optimizer will develop instrumental subgoals — including self-preservation — that may conflict with human interests.6
- Value alignment is technically tractable: Russell maintains that the CIRL framework and related approaches provide a viable conceptual path to safe AI, though much work remains to make these approaches practical at scale.
- A paradigm shift is required: Marginal improvements to current approaches are insufficient; the field needs to change how AI systems are designed and how AI is taught.
- Academic research and governance are both essential: Technical solutions alone are insufficient; regulatory frameworks with real enforcement authority are needed alongside technical safety work.
- Voluntary industry commitments are insufficient: In his 2023 Senate testimony, Russell argued that self-regulatory commitments by AI companies cannot substitute for a federal regulatory agency with devolved rule-making powers.5
On AI Development and Governance
Russell advocates for:
- Rethinking AI objectives: Moving away from fixed optimization toward preference-learning under uncertainty
- Provably beneficial AI: Formal Verification (AI Safety) approaches where possible
- Human oversight: Systems that remain under human control and can be corrected
- Cautious deployment: Not deploying systems whose behavior is not understood
- International Coordination: Global agreements on safe AI development
- Federal regulation: A U.S. regulatory agency for AI with devolved rule-making authority, analogous to existing domain-specific regulators5
Public Communication and Advocacy
BBC Reith Lectures (2021)
In 2021, Russell became the first computer scientist selected as a BBC Reith Lecturer — a series launched in 1948 that is described by Berkeley Engineering as "one of the most prestigious public lectures in Britain."4 He delivered four lectures across the UK under the series title "Living With Artificial Intelligence":
- Lecture 1: "AI: What Is It?" (Alan Turing Institute at the British Library, London) — explored whether AI is the biggest event in human history, traced AI's conceptual roots to Aristotle, and outlined what Russell calls the risks of the "standard model" of AI.18
- Lecture 2: "AI in Warfare" (University of Manchester) — examined autonomous weapons and argued for global governance and a ban on offensive autonomous weapons beyond meaningful human control.18
- Lecture 3: "AI & Jobs" (Newcastle University) — addressed the economic impact of AI and the future of work.18
- Lecture 4: "AI: A Future for Humans?" (Edinburgh) — proposed a new model for human-AI coexistence based on machines that defer to human preferences, elaborating the three principles from his CIRL framework: machines are purely altruistic, machines know that humans have preferences, and machines are uncertain about what those preferences are.18
The New Statesman described the lectures as "funny, accessible and deeply terrifying," and noted Russell's observation that social media content algorithms "have more power over people's cognitive intake than any dictator in human history."19
U.S. Senate Testimony (2023)
On July 25, 2023, Russell testified before the U.S. Senate Committee on the Judiciary, Subcommittee on Privacy, Technology, and the Law, at a hearing titled "Oversight of A.I.: Principles for Regulation."5 Key arguments in his testimony included:
- Advanced AI poses risks "up to and including human extinction"
- Voluntary commitments by major AI companies are insufficient
- The U.S. should create a federal AI regulatory agency with devolved rule-making powers
- Systems exhibiting unacceptable behavior — including self-replication and cyber-infiltration — should be withdrawn from the market immediately
- Transparency, accountability, and safety are the three core regulatory requirements
Russell noted in the testimony that considerable uncertainty remains about the true capabilities of current LLMs, but argued this uncertainty does not reduce the urgency of establishing regulatory frameworks.5
Autonomous Weapons Advocacy
Russell has been a prominent advocate for a ban on lethal autonomous weapons systems (LAWS). On July 28, 2015, an open letter calling for a ban on offensive autonomous weapons was announced at the opening of the IJCAI 2015 conference in Buenos Aires. Russell was the primary architect of the letter, which argued that autonomous weapons "represent the third revolution in warfare, after gunpowder and nuclear arms" and called for a ban on weapons beyond meaningful human control.20 Notable signatories included Elon Musk and Stephen Hawking; more than 3,000 AI researchers and related professionals ultimately signed.20
In 2017, Russell collaborated with the Future of Life Institute to produce the short film Slaughterbots, which was presented to a UN meeting on the Convention on Certain Conventional Weapons and has been viewed more than 70 million times.10 He also co-authored a follow-up response with Max Tegmark and Toby Walsh, "Why We Really Should Ban Autonomous Weapons: A Response," published in IEEE Spectrum on August 3, 2015.20
Russell has independent observer status as a member of the Global Partnership on AI, an international body with 15 member states including the U.S. and the EU.10
Book: Human Compatible (2019)
Russell's 2019 book Human Compatible: Artificial Intelligence and the Problem of Control (Viking, 2019) presented his CIRL framework and AI safety arguments for a general audience. It received coverage in major outlets including the New York Times, Guardian, Financial Times, Economist, and Vox, and was named one of the best books of 2019 by the Financial Times, the Guardian, the Daily Telegraph, Forbes, and other outlets.10 Scott Alexander's review at Slate Star Codex (January 2020) provided a widely read technical assessment of the book's arguments and their limitations.
Other Public Communication
- TED talk: "3 Principles for Creating Safer AI" (widely circulated, with millions of views)
- Multiple talks at the World Economic Forum in Davos and the Nobel Week Dialogues in Stockholm and Tokyo
- Regular media appearances including BBC, New York Times, CNN, Financial Times, and Vox
Disagreements and Debates
With AI Optimists — Instrumental Convergence Debate
Russell has been publicly critical of researchers who argue that advanced AI systems will not develop dangerous subgoals unless explicitly programmed to do so. The most prominent such exchange was a 2019 Facebook debate involving Russell, Yoshua Bengio, Yann LeCun, and Tony Zador, originating from a LeCun/Zador article titled "Don't Fear the Terminator."6
LeCun's position was that dangerous or immoral AI behavior can be handled by iterative refinement, similar to how problems were addressed in automobiles or rockets, and that AI systems will have no intrinsic desire to take control. Russell's response centered on the mathematical structure of optimization: "a machine will have self-preservation even if you don't program it in because if you say, 'Fetch the coffee,' it can't fetch the coffee if it's dead. So if you give it any goal whatsoever, it has a reason to preserve its own existence."6 He argued this is "simply and mathematically false" as a basis for dismissing the concern, and noted the debate is not about whether to put in explicit emotions of survival or dominance, but about the instrumental subgoals that any sufficiently capable optimizer will develop.6
Both sides accused the other of illogical anthropomorphism: LeCun accused safety researchers of assuming AGI would naturally desire power, while Russell and Bengio accused LeCun of assuming AGI would naturally infer human ethical norms.6 When LeCun later characterized Russell as a "prophet of doom" at Davos, Russell responded: "We're not predicting doom. We're saying we need to work on containment."17
LeCun's position — that AI safety concerns as framed by Russell overstate the risks from instrumental convergence — is shared by a number of prominent AI researchers, including several who hold Turing Award recognition. The disagreement reflects a genuine technical dispute about whether current alignment approaches are adequate and whether the problem Russell identifies is as intractable as he suggests.
With Researchers Favoring Different Safety Approaches
Russell's CIRL-based approach has also faced criticism from within the AI safety research community. Some researchers, including those associated with MIRI, have argued that IRL-based approaches do not adequately address the possibility of deceptive alignment — the scenario in which an AI system learns to appear aligned during training while pursuing different objectives once deployed. Russell has acknowledged that IRL is not a complete solution and that much additional work remains to make the framework practical at scale. He and MIRI-affiliated researchers share concern about the seriousness of AI risks, but differ on which technical approaches are most promising.
On Extreme Pessimism
Russell distinguishes his position from those who assign very high probabilities to catastrophic AI outcomes or call for a complete halt to AI capabilities research. He has argued that alignment is technically tractable, that coordination is possible, and that continuing AI research alongside safety work is preferable to a moratorium. This places him in disagreement with both researchers who view his risk estimates as too high and those who view them as too low.
On Capabilities Research Direction
Russell has argued that much mainstream AI research ignores safety considerations and focuses on raw performance over robustness and alignment. He has advocated for changing computer science education to give beneficial AI a more central role — his textbook's forthcoming revisions incorporate these concerns, and he has argued for structural changes in how AI curricula are designed.
Influence and Impact
Academic Field
CHAI has trained approximately 25–30 PhD students annually and hosts around seven interns per year.10 The center's faculty spans Berkeley, Cornell, University of Michigan, and Princeton, and CHAI is described by Founders Pledge as "one of the few academic centers devoted to the development of provably safe AI systems."10 Russell is described by the same assessment as "more focused on the extreme downside risk of transformative AI systems than any other comparably senior mainstream researcher."10
Russell's textbook Artificial Intelligence: A Modern Approach, co-authored with Peter Norvig, is used in over 1,500 universities in 135 countries and is the primary vehicle for his influence on AI education.1 He received the ACM Karl V. Karlstrom Outstanding Educator Award in 2005.1
Policy and Governance
- Testified before the U.S. Senate (July 2023) on AI regulation5
- Holds independent observer status with the Global Partnership on AI (15 member states including U.S. and EU)10
- Part of UN discussions on autonomous weapons
- Co-organized the 2015 Future of Life Institute open letter on autonomous weapons (3,000+ signatories)20
- Has advised governments of multiple countries on AI policy, though the specific countries and capacities have not been consistently documented in public sources10
The extent of Russell's direct influence on specific legislative outcomes such as the EU AI Act is not clearly documented in public sources; he has participated in discussions that preceded and accompanied such legislation.
Technical Research
IRL and CIRL have become active research areas in the machine learning community. The 2016 CIRL paper (Hadfield-Menell, Russell, Abbeel, Dragan) formalized the value alignment problem in a way that has been cited extensively in subsequent alignment research. The CIRL framework's connection to POMDP theory provided a bridge between AI safety concerns and established computational methods.2
Current Focus
As of 2025, Russell continues working on:
- Provably beneficial AI: Formal methods for AI safety — the subject of his invited talk at AAAI 20257 and his February 2025 Oxford colloquium presentation14
- Value alignment theory: How to specify and learn human values under uncertainty
- Off-switch problems and corrigibility: Ensuring AI systems remain correctable
- Governance frameworks: Policy approaches to AI safety, including federal regulatory structures
- Educational reform: Changing how AI is taught in universities, with planned updates to Artificial Intelligence: A Modern Approach
Evolution of Views
Early career (1980s–2000s):
- Focused on probabilistic reasoning, decision theory, knowledge representation, and planning
- Published foundational work on Interpretability and rational agency
- Began foundational IRL work with Ng (ICML 2000)
Transition (2000s–2010s):
- Growing concern about advanced AI risks as capabilities progressed
- Developing the IRL and CIRL frameworks
- Beginning to write and speak publicly about safety
Since 2015:
- Major public voice on AI risk and governance
- Primary architect of the 2015 FLI open letter on autonomous weapons20
- Founded CHAI (2016)
- Authored Human Compatible (2019)
- First computer scientist selected as BBC Reith Lecturer (2021)4
- Testified before U.S. Senate (2023)5
- Ongoing work on provably beneficial AI research program7
Criticism and Responses
Critics have argued:
- IRL may not scale to the full complexity of human values; the framework assumes AI systems can observe representative human behavior, which may not hold in practice
- CIRL and related approaches do not adequately address deceptive alignment — the possibility that a system learns to appear aligned during training while pursuing different objectives at deployment
- Academic research timescales may be too slow relative to the pace of AI capabilities development
- Russell's risk framing, particularly the instrumental convergence argument, rests on contested assumptions about how future AI systems will behave6
Russell has stated:
- IRL and CIRL are not complete solutions; they are conceptual frameworks that require substantial additional work to become practical
- The primary contribution of the CIRL framework is to formalize the alignment problem in a way that connects to established computational methods, not to claim the problem is solved
- Much additional research is needed to address the gap between theoretical frameworks and deployed systems14
Footnotes
-
Stuart J. Russell — Wikipedia — Stuart J. Russell — Wikipedia ↩ ↩2 ↩3 ↩4 ↩5 ↩6 ↩7 ↩8
-
Cooperative Inverse Reinforcement Learning — NeurIPS 2016 — Cooperative Inverse Reinforcement Learning — NeurIPS 2016, Dylan Hadfield-Menell, Stuart Russell, Pieter Abbeel, Anca Dragan (arXiv:1606.03137) ↩ ↩2 ↩3 ↩4
-
Center for Human-Compatible Artificial Intelligence — Wikipedia — Center for Human-Compatible Artificial Intelligence — Wikipedia ↩ ↩2 ↩3
-
Stuart Russell Chosen As 2021 BBC Reith Lecturer — CHAI, October 15, 2021 — Stuart Russell Chosen As 2021 BBC Reith Lecturer — CHAI, October 15, 2021 ↩ ↩2 ↩3
-
Stuart Russell Testifies on AI Regulation at U.S. Senate Hearing — UC Berkeley CDSS, July 25, 2023 — Stuart Russell Testifies on AI Regulation at U.S. Senate Hearing — UC Berkeley CDSS, July 25, 2023; CHAI Blog, September 11, 2023 ↩ ↩2 ↩3 ↩4 ↩5 ↩6 ↩7 ↩8
-
Debate on Instrumental Convergence between LeCun, Russell, Bengio, and More — LessWrong, 2019 — Debate on Instrumental Convergence between LeCun, Russell, Bengio, and More — LessWrong, 2019 ↩ ↩2 ↩3 ↩4 ↩5 ↩6 ↩7
-
Stuart J. Russell Wins 2025 AAAI Award for Artificial Intelligence for the Benefit of Humanity — Robohub/AAAI, 2025 — Stuart J. Russell Wins 2025 AAAI Award for Artificial Intelligence for the Benefit of Humanity — Robohub/AAAI, 2025 ↩ ↩2 ↩3 ↩4
-
Center for Human-Compatible Artificial Intelligence — EA Forum, June 2022 — Center for Human-Compatible Artificial Intelligence — EA Forum, June 2022 ↩ ↩2
-
UC Berkeley / CHAI 2021 Grant — Open Philanthropy, 2021 — UC Berkeley / CHAI 2021 Grant — Open Philanthropy, 2021 ↩
-
Center for Human-Compatible AI — Founders Pledge, 2021 — Center for Human-Compatible AI — Founders Pledge, 2021 ↩ ↩2 ↩3 ↩4 ↩5 ↩6 ↩7 ↩8 ↩9 ↩10
-
Is 'Provably Beneficial' AI Possible? — ITU / AI for Good, 2022 — Is 'Provably Beneficial' AI Possible? — ITU / AI for Good, 2022 ↩ ↩2
-
Algorithms for Inverse Reinforcement Learning — Andrew Y. Ng and Stuart Russell, ICML 2000 — Algorithms for Inverse Reinforcement Learning — Andrew Y. Ng and Stuart Russell, ICML 2000 ↩ ↩2
-
Apprenticeship Learning via Inverse Reinforcement Learning — Pieter Abbeel and Andrew Y. Ng, ICML 2004 — Apprenticeship Learning via Inverse Reinforcement Learning — Pieter Abbeel and Andrew Y. Ng, ICML 2004 ↩
-
Write-up of the Ethics in AI Colloquium — Provably Beneficial Artificial Intelligence — Oxford Institute for Ethics in AI, February 27, 2025 — Write-up of the Ethics in AI Colloquium — Provably Beneficial Artificial Intelligence — Oxford Institute for Ethics in AI, February 27, 2025 ↩ ↩2 ↩3 ↩4
-
WAICF '23: UC Berkeley's Stuart Russell: Don't Be Fooled by ChatGPT — AI Business, 2023 — WAICF '23: UC Berkeley's Stuart Russell: Don't Be Fooled by ChatGPT — AI Business, 2023 ↩
-
We've Reached a Turning Point with AI, Expert Says — CNN, May 31, 2023 — We've Reached a Turning Point with AI, Expert Says — CNN, May 31, 2023 ↩
-
AI Professor Stuart Russell: 'What Could Possibly Go Wrong?' — World Economic Forum / Radio Davos, 2023 — AI Professor Stuart Russell: 'What Could Possibly Go Wrong?' — World Economic Forum / Radio Davos, 2023 ↩ ↩2 ↩3
-
Living with AI: The Alan Turing Institute Hosts BBC Radio 4 Reith Lecture with Stuart Russell — Alan Turing Institute, November 1, 2021 — Living with AI: The Alan Turing Institute Hosts BBC Radio 4 Reith Lecture with Stuart Russell — Alan Turing Institute, November 1, 2021 ↩ ↩2 ↩3 ↩4
-
Citation rc-a0e5 (data unavailable — rebuild with wiki-server access) ↩
-
Autonomous Weapons Open Letter: AI & Robotics Researchers — Future of Life Institute, July 28, 2015 — Autonomous Weapons Open Letter: AI & Robotics Researchers — Future of Life Institute, July 28, 2015 ↩ ↩2 ↩3 ↩4 ↩5
References
“The center was founded in 2016 by a group of academics led by [Berkeley](https://en.wikipedia.org/wiki/UC_Berkeley "UC Berkeley") computer science professor and AI expert [Stuart J. Russell](https://en.wikipedia.org/wiki/Stuart_J._Russell "Stuart J. Russell").”
Failed to parse LLM response