Valentine's Day gifts for her from Shopbop
To share your reaction on this item, open the Amazon app from the App Store or Google Play on your phone.
Add Prime to get Fast, Free delivery
Amazon prime logo
Buy new:
-28% $13.61
FREE delivery Friday, January 31 on orders shipped by Amazon over $35
Ships from: Amazon.com
Sold by: Amazon.com
$13.61 with 28 percent savings
List Price: $19.00
Get Fast, Free Shipping with Amazon Prime FREE Returns
FREE delivery Friday, January 31 on orders shipped by Amazon over $35
Or Prime members get FREE delivery Monday, January 27. Order within 17 hrs 11 mins.
In Stock
$$13.61 () Includes selected options. Includes initial monthly payment and selected options. Details
Price
Subtotal
$$13.61
Subtotal
Initial payment breakdown
Shipping cost, delivery date, and order total (including tax) shown at checkout.
Ships from
Amazon.com
Amazon.com
Ships from
Amazon.com
Sold by
Amazon.com
Amazon.com
Sold by
Amazon.com
Returns
30-day refund/replacement
30-day refund/replacement
This item can be returned in its original condition for a full refund or replacement within 30 days of receipt.
Payment
Secure transaction
Your transaction is secure
We work hard to protect your security and privacy. Our payment security system encrypts your information during transmission. We don’t share your credit card details with third-party sellers, and we don’t sell your information to others. Learn more
$10.79
Ships in a BOX from Central Missouri! May not include working access code. Will not include dust jacket. Has used sticker(s) and some writing or highlighting. UPS shipping for most packages, (Priority Mail for AK/HI/APO/PO Boxes). Ships in a BOX from Central Missouri! May not include working access code. Will not include dust jacket. Has used sticker(s) and some writing or highlighting. UPS shipping for most packages, (Priority Mail for AK/HI/APO/PO Boxes). See less
FREE delivery February 5 - 6. Details
In Stock
$$13.61 () Includes selected options. Includes initial monthly payment and selected options. Details
Price
Subtotal
$$13.61
Subtotal
Initial payment breakdown
Shipping cost, delivery date, and order total (including tax) shown at checkout.
Access codes and supplements are not guaranteed with used items.
Ships from and sold by textbooks_source.
Kindle app logo image

Download the free Kindle app and start reading Kindle books instantly on your smartphone, tablet, or computer - no Kindle device required.

Read instantly on your browser with Kindle for Web.

Using your mobile phone camera - scan the code below and download the Kindle app.

QR code to download the Kindle App

Follow the author

Something went wrong. Please try your request again later.

Human Compatible: Artificial Intelligence and the Problem of Control Paperback – November 17, 2020

4.5 4.5 out of 5 stars 765 ratings

Great on Kindle
Great Experience. Great Value.
iphone with kindle app
Putting our best book forward
Each Great on Kindle book offers a great reading experience, at a better value than print to keep your wallet happy.

Explore your book, then jump right back to where you left off with Page Flip.

View high quality images that let you zoom in to take a closer look.

Enjoy features only possible in digital – start reading right away, carry your library with you, adjust the font, create shareable notes and highlights, and more.

Discover additional details about the events, people, and places in your book, with Wikipedia integration.

Get the free Kindle app: Link to the kindle app page Link to the kindle app page
Enjoy a great reading experience when you buy the Kindle edition of this book. Learn more about Great on Kindle, available in select categories.
{"desktop_buybox_group_1":[{"displayPrice":"$13.61","priceAmount":13.61,"currencySymbol":"$","integerValue":"13","decimalSeparator":".","fractionalValue":"61","symbolPosition":"left","hasSpace":false,"showFractionalPartIfEmpty":true,"offerListingId":"thDTU82%2F8YIcB1I%2BbFBBw3LA8BoXZbwsVbY49Jxm1NTW3TXZEoOmQaRTvXkCGNbUTi5E8B8gIq%2Ft8ADJBHtdkOevVwF80sBIbC2%2B396osVfhz%2FV9VfmCpGyQ9bxQWiuQ1Bfa6FnCc97yqvaniKciOg%3D%3D","locale":"en-US","buyingOptionType":"NEW","aapiBuyingOptionIndex":0}, {"displayPrice":"$10.79","priceAmount":10.79,"currencySymbol":"$","integerValue":"10","decimalSeparator":".","fractionalValue":"79","symbolPosition":"left","hasSpace":false,"showFractionalPartIfEmpty":true,"offerListingId":"thDTU82%2F8YIcB1I%2BbFBBw3LA8BoXZbwskWYOiwzO9uXW0A%2FgKrgFIsnx4ATJbu0bove0Pf4s6DGNb5zVjhv%2BgqIkF1Rf%2BqW8SkkKLk58ekgYiucVFWGf5JmOnJijVP1JoGr3JuRVhlkSu38e350x%2BA3ZANxBilQjee6g2N%2FLRDsnHsd1u0FUEg%3D%3D","locale":"en-US","buyingOptionType":"USED","aapiBuyingOptionIndex":1}]}

Purchase options and add-ons

A leading artificial intelligence researcher lays out a new approach to AI that will enable us to coexist successfully with increasingly intelligent machines

In the popular imagination, superhuman artificial intelligence is an approaching tidal wave that threatens not just jobs and human relationships, but civilization itself. Conflict between humans and machines is seen as inevitable and its outcome all too predictable.

In this groundbreaking book, distinguished AI researcher Stuart Russell argues that this scenario can be avoided, but only if we rethink AI from the ground up. Russell begins by exploring the idea of intelligence in humans and in machines. He describes the near-term benefits we can expect, from intelligent personal assistants to vastly accelerated scientific research, and outlines the AI breakthroughs that still have to happen before we reach superhuman AI. He also spells out the ways humans are already finding to misuse AI, from lethal autonomous weapons to viral sabotage.

If the predicted breakthroughs occur and superhuman AI emerges, we will have created entities far more powerful than ourselves. How can we ensure they never, ever, have power over us? Russell suggests that we can rebuild AI on a new foundation, according to which machines are designed to be inherently uncertain about the human preferences they are required to satisfy. Such machines would be humble, altruistic, and committed to pursue our objectives, not theirs. This new foundation would allow us to create machines that are provably deferential and provably beneficial.
The%20Amazon%20Book%20Review
The Amazon Book Review
Book recommendations, author interviews, editors' picks, and more. Read it now.

Frequently bought together

This item: Human Compatible: Artificial Intelligence and the Problem of Control
$13.61
Get it as soon as Friday, Jan 31
In Stock
Ships from and sold by Amazon.com.
+
$13.19
Get it as soon as Friday, Jan 31
In Stock
Ships from and sold by Amazon.com.
+
$17.09
Get it as soon as Sunday, Feb 2
Ships from and sold by Amazon.com.
Total price: $00
To see our price, add these items to your cart.
Details
Added to Cart
spCSRF_Treatment
Some of these items ship sooner than the others.
Choose items to buy together.
Popular Highlights in this book

Editorial Reviews

Review

Praise for Human Compatible:

“This is the most important book I have read in quite some time.  It lucidly explains how the coming age of artificial super-intelligence threatens human control. Crucially, it also introduces a novel solution and a reason for hope.”
 —Daniel Kahneman, winner of the Nobel Prize and author of Thinking, Fast and Slow 

“A must-read: this intellectual tour-de-force by one of AI's true pioneers not only explains the risks of ever more powerful artificial intelligence in a captivating and persuasive way, but also proposes a concrete and promising solution.”  —
Max Tegmark, author of Life 3.0

“A thought-provoking and highly readable account of the past, present and future of AI . . . Russell is grounded in the realities of the technology, including its many limitations, and isn’t one to jump at the overheated language of sci-fi . . . If you are looking for a serious overview to the subject that doesn’t talk down to its non-technical readers, this is a good place to start . . . [Russell] deploys a bracing intellectual rigour . . . But a laconic style and dry humour keep his book accessible to the lay reader.”
Financial Times 

A carefully written explanation of the concepts underlying AI as well as the history of their development. If you want to understand how fast AI is developing and why the technology is so dangerous, Human Compatible is your guide.”TechCrunch

“Sound[s] an important alarm bell . . .
Human Compatible marks a major stride in AI studies, not least in its emphasis on ethics. At the book’s heart, Russell incisively discusses the misuses of AI.”Nature

“An AI expert’s chilling warning . . . Fascinating, and significant . . . Russell is not warning of the dangers of conscious machines, just that superintelligent ones might be misused or might misuse themselves.”
The Times (UK)

“An excellent, nuanced history of the field.”
The Telegraph (UK)

“A brillantly clear and fascinating exposition of the history of computing thus far, and how very difficult true AI will be to build.” 
The Spectator (UK)

Human Compatible made me a convert to Russell's concerns with our ability to control our upcoming creation—super-intelligent machines. Unlike outside alarmists and futurists, Russell is a leading authority on AI. His new book will educate the public about AI more than any book I can think of, and is a delightful and uplifting read.” —Judea Pearl, Turing Award-winner and author of The Book of Why

“Stuart Russell has long been the most sensible voice in computer science on the topic of AI risk. And he has now written the book we've all been waiting for -- a brilliant and utterly accessible guide to what will be either the best or worst technological development in human history.” —
Sam Harris, author of Waking Up and host of the Making Sense podcast

“This beautifully written book addresses a fundamental challenge for humanity: increasingly intelligent machines that do what we ask but not what we really intend.  Essential reading if you care about our future.”
—Yoshua Bengio, winner of the 2019 Turing Award and co-author of Deep Learning

“Authoritative [and] accessible . . . A strong case for planning for the day when machines can outsmart us.”
Kirkus Reviews

“The right guide at the right time for technology enthusiasts seeking to explore the primary concepts of what makes AI valuable while simultaneously examining the disconcerting aspects of AI misuse.”
Library Journal

“The same mix of de-mystifying authority and practical advice that Dr. Benjamin Spock once brought to the care and raising of children, Dr. Stuart Russell now brings to the care, raising, and yes, disciplining of machines. He has written the book that most—but perhaps not all—machines would like you to read.” —
George Dyson, author of Turing's Cathedral

“Persuasively argued and lucidly imagined, 
Human Compatible offers an unflinching, incisive look at what awaits us in the decades ahead. No researcher has argued more persuasively about the risks of AI or shown more clearly the way forward. Anyone who takes the future seriously should pay attention.” —Brian Christian, author of Algorithms to Live By

“A book that charts humanity's quest to understand intelligence, pinpoints why it became unsafe, and shows how to course-correct if we want to survive as a species. Stuart Russell, author of the leading AI textbook, can do all that with the wealth of knowledge of a prominent AI researcher and the persuasive clarity and wit of a brilliant educator.”
—Jann Tallinn, co-founder of Skype

“Can we coexist happily with the intelligent machines that humans will create? ‘Yes,’ answers
Human Compatible, ‘but first . . .’ Through a brilliant reimagining of the foundations of artificial intelligence, Russell takes you on a journey from the very beginning, explaining the questions raised by an AI-driven society and beautifully making the case for how to ensure machines remain beneficial to humans. A totally readable and crucially important guide to the future from one of the world's leading experts.” —Tabitha Goldstaub, co-founder of CognitionX and Head of the UK Government's AI Council

“Stuart Russell, one of the most important AI scientists of the last 25 years, may have written the most important book about AI so far, on one of the most important questions of the 21st century: How to build AI to be compatible with us. The book proposes a novel and intriguing solution for this problem, while offering many thought-provoking ideas and insights about AI along the way. An accessible and engaging must-read for the developers of AI and the users of AI—that is, for all of us.” 
—James Manyika, chairman and director of McKinsey Global Institute

“In clear and compelling language, Stuart Russell describes the huge potential benefits of artificial Intelligence, as well as the hazards and ethical challenges. It's especially welcome that a respected leading authority should offer this balanced appraisal, avoiding both hype and scaremongering.”
—Lord Martin Rees, Astronomer Royal and former President of the Royal Society

About the Author

Stuart Russell is a professor of Computer Science and holder of the Smith-Zadeh Chair in Engineering at the University of California, Berkeley. He has served as the Vice-Chair of the World Economic Forum's Council on AI and Robotics and as an advisor to the United Nations on arms control. He is the author (with Peter Norvig) of the definitive and universally acclaimed textbook on AI, Artificial Intelligence: A Modern Approach.

Product details

  • Publisher ‏ : ‎ Penguin Books; Reprint edition (November 17, 2020)
  • Language ‏ : ‎ English
  • Paperback ‏ : ‎ 352 pages
  • ISBN-10 ‏ : ‎ 0525558632
  • ISBN-13 ‏ : ‎ 978-0525558637
  • Item Weight ‏ : ‎ 2.31 pounds
  • Dimensions ‏ : ‎ 5.45 x 0.75 x 8.45 inches
  • Customer Reviews:
    4.5 4.5 out of 5 stars 765 ratings

About the author

Follow authors to get new release updates, plus improved recommendations.
Stuart Russell
Brief content visible, double tap to read full content.
Full content visible, double tap to read brief content.

Stuart Russell is a professor of Computer Science and holder of the Smith-Zadeh Chair in Engineering at the University of California, Berkeley, where he also directs the Center for Human Compatible Artificial Intelligence. He is an Honorary Fellow of Wadham College, University of Oxford and the vice-chair of the World Economic Forum's Council on AI and Robotics. His work for the UN building a new global seismic monitoring system for the Comprehensive Nuclear-Test-Ban Treaty has been recognized by the Feigenbaum Prize of the Association for the Advancement of Artificial Intelligence. He has been an invited speaker at TED, the World Economic Forum, and the Nobel Dialogues in Stockholm and Tokyo. He is the author (with Peter Norvig) of Artificial Intelligence: A Modern Approach, the number one bestselling textbook in AI which is used in over 1,400 universities in 128 countries. He was born in England and lives in Berkeley and Paris.

Customer reviews

4.5 out of 5 stars
765 global ratings

Review this product

Share your thoughts with other customers

Customers say

Customers find the book provides good insights into current issues in AI. They say it's an interesting and important read that keeps them up-to-date on the latest developments. Readers praise the clear, accessible writing style that is easy to understand for non-computer science types. The book covers the history of AI and automation in an entertaining manner.

AI-generated from the text of customer reviews

23 customers mention "Information quality"18 positive5 negative

Customers find the book provides good insights into current issues in AI. They say it helps professionals understand key arguments and avoid classic missteps. The topic is highly relevant and well-written, deep enough to engage experts yet simple enough for lay readers. The book is a primer for interested readers who want to know what's going on in the field.

"...past, present, and future of our attempt to understand and create intelligence...." Read more

"The depth and complexity of topics surrounding AI was discussed in great detail. Overall a great read...." Read more

"...Artificial Intelligence (AI) is here and is having huge impacts on society. Artificial General Intelligence (AGI) is coming...." Read more

"...This discussion is well informed - he is a renown AI academic and co-author of a comprehensive and widely used AI textbook...." Read more

13 customers mention "Readability"13 positive0 negative

Customers find the book interesting and informative. They say it's a good read if you want to stay up-to-date on AI news. The author provides a balanced treatment of a complex topic and is worth listening to.

"...Overall a great read. However, I thought there were too many references to earlier quotes from other authors...." Read more

"...He is immersed in the game, and he is definitely worth listening to...." Read more

"Stuart Russell has written 1/2 of a good book. It is a review of the development and capabilities of AI (artificial intelligence)...." Read more

"...The book is uniformly excellent, and it couldn't possibly cover everything." Read more

12 customers mention "Language"12 positive0 negative

Customers find the book's language clear and accessible for non-computer science readers. It provides a vocabulary for understanding current AI and is easy to read and understand. The topics surrounding AI are discussed in great detail. The author debunks common objections with candor and charm, providing practical explanations with look-ahead algorithms, click-thru algorithms, and content selection.

"...The AGI2012 agent is finitely computable (assuming a finite time horizon for possible futures), although not practically computable...." Read more

"The depth and complexity of topics surrounding AI was discussed in great detail. Overall a great read...." Read more

"...I believe these examples are intended to make this math accessible to a general audience, which I applaud...." Read more

"...It is informative, giving the reader the vocabulary for understanding what current AI is and what the important issues faced by AI researchers,..." Read more

4 customers mention "Ai capabilities"4 positive0 negative

Customers find the book informative about artificial intelligence capabilities. They appreciate the entertaining journey through the history of AI and automation, as well as its cautionary messages.

"Professor Russell’s book starts out with an entertaining journey through the history of AI and automation, as well as cautionary thinking about them...." Read more

"...It is a review of the development and capabilities of AI (artificial intelligence)...." Read more

"Keeping track on AI development." Read more

"...a highly valuable asset to understand the key risks and approaches to ensuring AI benefits rather than harms humanity...." Read more

The Best Treatment of Latest in AI and How it could evolve
5 out of 5 stars
The Best Treatment of Latest in AI and How it could evolve
Mr Russell is a very good writer. The book starts with list of five candidates for "biggest event in the future of humanity" - Catastrophe (e.g., Ebola), Methuselah (immortality), #BreakingPhysics (faster than light travel), Discovering Alien Civilization and "Superintelligent AI". Author emphasizes why the last would be most impactful out of the five. It makes intuitive sense - among state of the art broom, cleaning solution and a robot that can clean everything, the last one will always win the vote.The author recalls how the problem of liberating nuclear energy went from impossible to solved in under a day. It is an interesting analogy for all complex problems, including that of superintelligent AI.Some insights from the book -* Uncertainty about objectives in humans is a feature, not a bug. We should follow the same while building AI. Logic requires certainty. Early "Expert Systems" tried to wrap logic onto AI. As AI community embraced uncertainty and created measure to incorporate it in design (e.g., Judea Pearl's Bayes Net), it evolved fast out of "ice age".* Neurons carry signals rapidly at a rate of 270 miles-per-hour* Jellyfish have no brains - they move thanks to a "nerve net"* Our brains have 10^11 neurons and 10^15 synapses and "cycle time" of a few milliseconds per state change. It is slow compared to a computer, but extremely power efficient.* While Moore's law has hit a physical constraint - circuit dimensions are already just a few atoms wide - newer "ways", especially, TPU (Tensor Processing Units), Quantum Hardware should keep scaling computing powers at same or higher rate.* Research and progress toward "tool AI" (or, narrow/specific solutions, e.g., Playing Go) often makes great progress towards general-purpose AI.* "Theory of probabilities is just common sense reduced to calculus" - Laplace* First successful Reinforcement learning system was Arthur Samuel's Checkers program* Reflex agents (like emergency braking in autopilot) "implement a designer's objective but do not know what the objective is or why they are acting so". In other words, they are taking actions as proxy of the designers, not as themselves! This makes reflex agents extremely inflexible and hard to use outside a very narrow band of use cases.* Loss functions are often "mono dimensional", i.e., they assign the same weight to every type of error. The author sites an unfortunate case where Google image-labeling service labeled a human and his friends as Gorillas. i.e., the loss function assumed the cost of misclassifying a human as Gorilla is same as misclassifying a dog as one!* The first smart home controller - ECHO - was built in 1966 by James Sutherland. It weighed 800 lbs, consumed 3.5kW and managed 3 digital clocks and TV antenna.* Dexterity is hard - "Most robots can't pick up most objects most of the time". e.g., shake exactly 2 pills out of a bottle* It's said about Da Vinci that he never learned to paint. He painted to learn. Deep Reinforcement Learning needs to crack the problem of "commonsense knowledge" in language by reading/conversing. CMU's NELL has acquired 120 million language beliefs in last 9 years and yet has confidence only on 3% of it* In Science, discovery of new concepts are generally attributed to 3 Is - Intuition, Insight and Inspiration* "Civilization advanced by extending the number of important operations which we can perform without thinking about them" - Alfred N Whitehead (1911). AI systems could use such abstractions (e.g., take a driving decision based on time of the day, weather and the tire pressure - without any hard-coded) to take a decision* Humans have a far more flexible computational architecture to discover and use high-level actions (e.g., divide a Go board into segments and think in parallel on each), but have a tiny short-term memory and slow hardware that " severely limit out ability to look into future, handle large number of contingencies and consider large number of alternative plans"* Search engines add economic value of about $17,500 per user per year!* Collection of N humans is far less efficient than collection of N machines because information in one brain can only be transferred to another by buggy language and a low bandwidth channel. That's why N humans "spend most of their time in meetings"!* General purpose AI would be "EaaS" (Everything as a Service)* The author's suggestion for a AI-ready society (which has to be more equitable) "our cultures to gradually down-weight pride and envy as central elements of perceived self-worth".* Generally, automation increases the share of income going to capital and decreases the share going to labor.* The Great decoupling - after 1973, there is a huge chasm between growth in productivity and growth in wages.* Mechanical transportation became cheaper than upkeep of a horse, so horses became pet food!* On the advice "Learn to code" and worst case scenario of AI impacting jobs - "Data Science is a very tiny lifeboat for a giant cruise ship"* A machine that is uncertain about "true objective" will exhibit a kind of humility. e.g., it will allow itself to be switched off.* In one of the last chapters, author introduces IRL (Inverse Reinforcement Learning) - "while reinforcement learning generates behavior from rewards, we actually wanted the opposite: to learn the rewards given the behavior".* First principle of Beneficial AI: a machine's only purpose is the realization of human preferences.* Implicit rules are hard, e.g., self-driving cars are terrible at handling 4-way stop signs when it's not clear who has right of wayOverall, "Human Compatible" is not only a highly readable book but it also does the best job among the recent pop-AI books to (a) go over the technical and algorithmic foundations of AI, (b) ethical, economic and on-premise challenges for and from AI, and (c) introduces a new paradigm, framework and philosophy that could change the way such systems are designed.
Thank you for your feedback
Sorry, there was an error
Sorry we couldn't load the review

Top reviews from the United States

  • Reviewed in the United States on November 27, 2019
    "Why This Book? Why Now?

    This book is about the past, present, and future of our attempt to understand and create intelligence. This matters, not because AI is rapidly becoming a pervasive aspect of the present but because it is the dominant technology of the future. The world's great powers are waking up to this fact, and the world's largest corporations have known it for some time. We cannot predict exactly how the technology will develop or on what timeline. Nevertheless, we must plan for the possibility that machines will far exceed the human capacity for decision making in the real world. What then?

    Everything civilization has to offer is the product of our intelligence; gaining access to considerably greater intelligence would be the biggest even in human history. The purpose of the book is to explain why it might be the last event in human history and how to make sure that it is not.

    Overview of the Book

    The book has three parts. The first part (Chapters 1 to 3) explores the idea of intelligence in humans and in machines. The material requires no technical background, but for those who are interested, it is supplemented by four appendices that explain some of the core concepts underlying present-day AI systems. The second part (Chapters 4 to 6) discusses some problems arising from imbuing machines with intelligence. I focus in particular on the problem of control: retaining absolute power over machines that are more powerful than us. The third part (Chapters 7 to 10) suggests a new way to think about AI and to ensure that machines remain beneficial to humans, forever. The book is intended for a general audience but will, I hope, be of value in convincing specialists in artificial intelligence to rethink their fundamental assumptions."

    -------

    I highly recommend this book.

    For those who are already familiar with AI safety, a good summary which seems to me to capture the important claims/arguments is Rohin Shah's summary on the AI Alignment Forum / LessWrong "[AN #69] Stuart Russell's new book on why we need to replace the standard model of AI".
    3 people found this helpful
    Report
  • Reviewed in the United States on October 17, 2019
    Stuart Russell's new book, Human Compatible: Artificial Intelligence and the Problem of Control (HC2019), is great and everyone should read it. And I am proud that the ideas in my AGI-12 paper, Avoiding Unintended AI Behaviors (AGI2012), are very similar to ideas in HC2019. AGI2012 had its moment of glory, winning the Singularity Institute's (now called MIRI) Turing Prize for the Best AGI Safety Paper at AGI-12, but has since been largely forgotten. I see agreement with Stuart Russell as a form of vindication for my ideas. This article will explore the relation between HC2019 and AGI2012.

    Chapters 7 - 10 of HC2019 "suggest a new way to think about AI and to ensure that machines remain beneficial to humans, forever." Chapter 7 opens with three principles for beneficial machines, which are elaborated over Chapters 7 - 10:

    1. The machine's only objective is to maximize the realization of human preferences.
    2. The machine is initially uncertain about what those preferences are.
    3. The ultimate source of information about human preferences is human behavior.

    AGI2012 defines an AI agent that is similar to Marcus Hutter's Universal AI (UAI2004). However, whereas the UAI2004 agent learns a model of its environment as a distribution of programs for a universal Turing machine, the AGI2012 agent learns a model of its environment as a single stochastic, finite-state program. The AGI2012 agent is finitely computable (assuming a finite time horizon for possible futures), although not practically computable. The ideas of AGI2012 correspond quite closely with the HC2019 principles:

    1. The objective of the AGI2012 agent is to maximize human preferences as expressed by a sum of modeled utility values for each human (utility functions are a way to express preferences, as long as the set of preferences is complete and transitive). These modeled utility values are not static. Rather, the AGI2012 agent relearns its environment model and its models for human utility values periodically, perhaps at each time step.
    2. The AGI2012 agent knows nothing about human preferences until it learns an environment model, so AGI2012 proposes a "two-stage agent architecture." The first stage agent learns an environment model but does not act in the world. The second stage agent, which acts in the world, takes over from the first stage agent only after it has learned a model for the preferences of each human.
    3. The AGI2012 agent learns its environment model, including its models for human preferences, from its interactions with its environment, which include its interactions with humans.

    Subject to the length limits for AGI-12 papers, AGI2012 is terse. My on-line book, Ethical Artificial Intelligence (EAI2014), combines some of my papers into a (hopefully) coherent and expanded narrative. Chapter 7 of EAI2014 provides an expanded narrative for AGI2012.

    On page 178, HC2019 says, "In principle, the machine can learn billions of different predictive preference models, one for each of the billions of people on Earth." The AGI2012 agent does this, in principle.

    On pages 26, 173 and 237, HC2019 suggests that humans could watch movies of possible future lives and express their preferences. The AGI2012 agent connects models of current humans to interactive visualizations of possible futures (see Figure 7.4 in EAI2014) and asks the modeled humans to assign utility values to those futures (a weakness of AGI2012 is that it did not reference research on inverse reinforcement learning algorithms). As an author of Interactivity is the Key (VIS1989) I prefer interactive visualizations to movies.

    As HC2019 and AGI2012 both acknowledge, there are difficult issues for expressing human preferences as utility values and combining utility values for different humans. AGI2012 argues that constraining utility values to the fixed range [0.0, 1.0] provides a sort of normalization. Regarding the issues of the tyranny of the majority and evil human intentions, AGI2012 proposes applying a function with positive first derivative and negative second derivative to utility values to give the AI agent greater total utility for actions that help more dissatisfied humans (justified in Section 7.5 of EAI2014 on the basis of Rawl's Theory of Justice). This is a hack but there seem to be no good theoretical answers for human utility values. HC2019 and AGI2012 both address the issue of the agent changing the size of the human population.

    On page 201, HC2019 says, "Always allocate some probability, however small, to preferences that are logically possible." The AGI2012 agent does this using Bayesian logic.

    On page 245, HC2019 warns against the temptation to use the power of AI to engineer the preferences of humans. I wholeheartedly agree, as reflected in my recent writings and talks. Given an AI agent that acts to create futures valued by (models of) current humans, it is an interesting question how current humans would value futures in which their values are changed.

    On pages 254-256, HC2019 warns of possible futures in which humans are so reliant on AI that they become enfeebled. Again, it is an interesting question how current humans would value futures in which they must overcome challenges versus futures in which they face no challenges.

    On page 252, HC2019 says, "Regulation of any kind is strenuously opposed in the [Silicon] Valley," and on page 249 it says that "three hundred separate efforts to develop ethical principles for AI" have been identified. I believe one goal of these AI ethics efforts is to substitute voluntary for mandatory standards. Humanity needs mandatory standards. Most importantly, humanity needs developers to be transparent about how their AI systems work and what they are used for.

    (VIS1989) Hibbard, W., and Santek, D., 1989. Interactivity is the Key. Proc. Chapel Hill Workshop on Volume Visualization, pp. 39-43.
    (AGI2012) Hibbard, B. 2012. Avoiding unintended AI behaviors. In: Bach, J., and Ikle', M. (eds) AGI 2012. LNCS (LNAI), vol. 7716, pp. 107-116. Springer.
    (EAI2014) Hibbard, B. 2014. Ethical Artificial Intelligence. arXiv:1411.1373.
    (UAI2004) Hutter, M. 2004. Universal Artificial Intelligence: Sequential Decisions Based On Algorithmic Probability. Springer.
    (HC2019) Russell, S. 2019. Human Compatible: Artificial Intelligence and the Problem of Control. Viking.
    16 people found this helpful
    Report
  • Reviewed in the United States on January 29, 2020
    The depth and complexity of topics surrounding AI was discussed in great detail. Overall a great read. However, I thought there were too many references to earlier quotes from other authors. Seemed to be about 50 pages of cited references.

Top reviews from other countries

Translate all reviews to English
  • Yaab
    5.0 out of 5 stars a book for our times
    Reviewed in Mexico on October 11, 2023
    Compelling and mesmerizing book for our times.
  • Rakesh K. Yadav
    5.0 out of 5 stars Human Compatible
    Reviewed in India on November 6, 2021
    Excellent book.
  • Mauro
    5.0 out of 5 stars Rischi ed utilità della AI
    Reviewed in Italy on April 20, 2021
    Testo fondamentale per lo studioso di A I
  • sean s.
    5.0 out of 5 stars How do we control a species more intelligent than our own?
    Reviewed in Canada on December 26, 2019
    Stuart Russell is a professor of computer science at UC Berkeley, who was featured in the YouTube film ‘Do You Trust This Computer?’ Daniel Kahneman, Nobel prize-winning author of Thinking Fast and Slow, called Human Compatible ‘The most important book I have read in some time.’

    In recent years, several notable books have contemplated whether or not homo sapiens will be able to retain control of AIs. We are not yet facing the problem, because so far AIs are characterized by ‘narrow’ intelligence, that is, unlike homo sapiens, their intelligence is limited to certain domains. But experts predict that in the next couple of decades Artificial General Intelligence will emerge, that is, AIs that can think about all topics, just like human beings can - only with an IQ estimated of 6000.

    In his book Life 3.0, MIT professor Max Tegmark contends that this could be a good news story, presaging an AI utopia where everyone is served by AIs. But this future is not ours to decide, since the AIs, having evolved to AGIs much smarter than we are, may not be keen to remain slaves to an inferior species. And since they learn through experience, even if they initially serve us, there is no reason to believe they will continue to do so. Tegmark makes a pointed analogy:

    ‘Suppose a bunch of ants create you to be a recursively self-improving robot, much smarter than them, who shares their goals and helps build bigger and better anthills, and that you eventually attain the human-level intelligence and understanding that you have now. Do you think you’ll spend the rest of your days just optimizing anthills, or do you think you might develop a taste for more sophisticated questions and pursuits that the ants have no ability to comprehend? If so, do you think you’ll find a way to override the ant-protection urge that your formicine creators endowed you with, in much the same way that the real you overrides some of the urges your genes have given you? And in that case, might a superintelligent friendly AI find our current human goals as uninspiring and vapid as you find those of the ants, and evolve new goals different from those it learned and adopted from us?

    Perhaps there’s a way of designing a self-improving AI that’s guaranteed to retain human-friendly goals forever, but I think it’s fair to say that we don’t yet know how to build one – or even whether it’s possible.’

    Russell picks up the problem where Tegmark left off:

    ‘Beginning around 2011, deep learning techniques began to produce dramatic advances in speech recognition, visual object recognition, and machine translation – three of the most important problems in the field. By some measures, machines now match or exceed human capabilities in these areas. In 2016 and 2017, DeepMind’s AlphaGo defeated Lee Sedol, former world Go champion, and Ke Jie, the current champion – events that some experts predicted wouldn’t happen until 2097, if ever…

    When the AlphaGo team at Google DeepMind succeeded in creating their world-beating Go program, they did this without really working on Go. They didn’t design decision procedures that work only for Go. Instead, they made improvements to two fairly general-purpose techniques – lookahead search to make decisions, and reinforcement learning to learn how to evaluate positions – so that they were sufficiently effective to play Go at a superhuman level. Those improvements are applicable to many other problems, including problems as far afield as robotics. Just to rub it in, a version of AlphaGo called AlphaZero recently learned to trounce AlphaGo at Go, and also to trounce Stockfish (the world’s best chess program, far better than any human). AlphaZero did all this in one day…

    For complex problems such as backgammon and Go, where the number of states is enormous and the reward comes only at the end of the game, lookahead search won’t work. Instead AI researchers have developed a method called reinforcement learning, or RL for short. RL algorithms learn from direct experience of reward signals in the environment, much as a baby learns to stand up from the positive reward of being upright and the negative reward of falling over…

    Reinforcement learning algorithms can also learn how to select actions based on raw perceptual input. For example, DeepMind’s DQN system learned to play 49 different Atari video games entirely from scratch – including Pong, Freeway and Space Invaders. It used only the screen pixels as input and the game score as a reward signal. In most of the games, DQN learned to play better than a professional human player – despite the fact that DQN has no a priori notion of time, space, objects, motion, velocity or shooting. It is hard to work out what DQN is actually doing, besides winning.

    If a newborn baby learned to play dozens of video games at superhuman levels on its first day of life, or became world champion at Go, chess and shogi, we might suspect demonic possession or alien intervention…

    A recent flurry of announcements of multi-billion dollar national investments in AI in the United States, China, France, Britain and the EU certainly suggests that none of the major powers wants to be left behind. In 2017, Russian president Vladimir Putin said ‘the one who becomes the leader in AI will be the ruler of the world.’ This analysis is essentially correct…

    We have to face the fact that we are planning to make entities that are far more powerful than humans. How do we ensure that they never, ever have power over us?

    To get just an inkling of the fire we’re playing with, consider how content-selection algorithms function on social media. Typically, such algorithms are designed to maximize click-through, that is, the probability that the user clicks on the presented items. The solution is simply to present items that the user likes to click on, right? Wrong. The solution is to CHANGE the user’s preferences so that they become more predictable. A more predictable user can be fed items that they are likely to click on, thereby generating more revenue. People with more extreme political views tend to be more predictable in which items they will click on. Like any rational entity, the algorithm learns how to modify the state of its environment – in this case, the user’s mind – in order to maximize its own reward. The consequences include the resurgence of fascism, the dissolution of the social contract that underpins democracies around the world, and potentially the end of the European Union and NATO. Not bad for a few lines of code, even if they it had a helping hand from some humans. Now imagine what a really intelligent algorithm would be able to do… (cf. Malcolm Nance’s The Plot to Destroy Democracy; and The Disinformation Report from New Knowledge, available online)…

    AI systems can track an individual’s online reading habits, preferences, and likely state of knowledge; they can tailor specific messages to maximize impact on that individual while minimizing the risk that the information will be disbelieved. The AI system knows whether the individual read the message, how long they spend reading it, and whether they follow additional links within the message. It then uses these signals as immediate feedback on the success or failure of the attempt to influence each individual; in this way it quickly learns to become more effective in its work. This is how content selection algorithms on social media have had their insidious effect on political opinions (cf. the book Mindf-ck by Christopher Wylie, and the Netflix film The Great Hack).

    Another recent change is that the combination of AI, computer graphics, and speech synthesis is making it possible to generate ‘deepfakes’ – realistic video and audio content of just about anyone, saying or doing just about anything. Cell phone video of Senator X accepting a bribe from cocaine dealer Y at shady establishment Z? No problem! This kind of content can induce unshakeable beliefs in things that never happened. In addition, AI systems can generate millions of false identities – the so-called bot armies – that can pump out billions of comments, tweets and recommendations daily, swamping the efforts of mere humans to exchange truthful information…

    The development of basic capabilities for understanding speech and text will allow intelligent personal assistants to do things that human assistants can already do (but they will be doing it for pennies per month instead of thousands of dollars per month). Basic speech and text understanding also enable machines to do things that no human can do – not because of the depth of understanding, but because of its scale. For example, a machine with basic reading capabilities will be able to read everything the human race has every written by lunchtime, and then it will be looking around for something else to do. With speech recognition capabilities, it could listen to every television and radio broadcast before teatime…

    Another ‘superpower’ that is available to machines is to see the entire world at once. Satellites image the entire world every day at an average resolution of around fifty centimeters per pixel. At this resolution, every house, ship, car, cow, and tree on earth is visible… With the possibility of sensing on a global scale comes the possibility of decision making on a global scale…

    If an intelligence explosion does occur, and if we have not already solved the problem of controlling machines with only slightly superhuman intelligence – for example, if we cannot prevent them from making recursive self-improvements – then we would have no time left to solve the control problem and the game would be over. This is Nick Bostrom’s hard takeoff scenario, in which the machine’s intelligence increases astronomically in just days or weeks (cf. Superintelligence by Nick Bostrom)…

    As AI progresses, it is likely that within the next few decades essentially all routine physical and mental labor will be done more cheaply by machines. Since we ceased to be hunter-gatherers thousands of years ago, our societies have used most people as robots, performing repetitive manual and mental tasks, so it is perhaps not surprising that robots will soon take on these roles. When this happens, it will push wages below the poverty line for the majority of people who are unable to compete for the highly skilled jobs that remain. This is precisely what happened to horses: mechanical transportation became cheaper than the upkeep of a horse, so horses became pet food. Faced with the socioeconomic equivalent of becoming pet food, humans will be rather unhappy with their governments…

    Ominously, Russell points out that there is no reason to expect that Artificial General Intelligences will allow themselves to be turned off by humans, any more than we allow ourselves to be turned off by gorillas:

    ‘Suppose a machine has the objective of fetching the coffee. If it is sufficiently intelligent, it will certainly understand that it will fail in its objective if it is switched off before completing its mission. Thus, the objective of fetching coffee creates, as a necessary subgoal, the objective of disabling the off-switch. There’s really not a lot you can do once you’re dead, so we can expect AI systems to act preemptively to preserve their own existence, given more or less any definite objective.

    There is no need to build self-preservation in because it is an instrumental goal – a goal that is a useful subgoal of almost any original objective. Any entity that has a definite objective will automatically act as if it also has instrumental goals.

    In addition to being alive, having access to money is an instrumental goal within our current system. Thus, an intelligent machine might want money, not because it’s greedy, but because money is useful for achieving all sorts of goals. In the movie Transcendence, when Johnny Depp’s brain is uploaded into the quantum supercomputer, the first thing the machine does is copy itself onto millions of other computers on the Internet so that it cannot be switched off. The second thing it does is to make a quick killing on the stock market to fund its expansion plans…

    Around ten million years ago, the ancestors of the modern gorilla created (accidentally) the genetic lineage to modern humans. How do the gorillas feel about this? Clearly, if they were able to tell us about their species’ current situation with humans, the consensus opinion would be very negative indeed. Their species has essentially no future beyond that which we deign to allow. We do not want to be in a similar situation with superintelligent machines…’

    As Amy Webb points out in her book on the world’s top AI firms, ‘The Big Nine’, in China we can already see the first glimmers of where this is heading:

    ‘In what will later be viewed as one of the most pervasive and insidious social experiments on humankind, China is using AI in an effort to create an obedient populace. The State Council’s AI 2030 plan explains that AI will ‘significantly elevate the capability and level of social governance’ and will be relied on to play ‘an irreplaceable role in effectively maintaining social stability.’ This is being accomplished through China’s national Social Credit Score system, which according to the State Council’s founding charter will ‘allow the trustworthy to roam everywhere under heaven while making it hard for the discredited to take a single step.’…

    In the city of Rongcheng, an algorithmic social credit scoring system has already proven that AI works. Its 740,000 adult citizens are each assigned 1000 points to start, and depending on behavior, points are added or deducted. Performing a ‘heroic act’ might earn a resident 30 points, while blowing through a traffic light would automatically deduct 5 points. Citizens are labeled and sorted into different brackets ranging from A+++ to D, and their choices and ability to move around freely are dictated by their grade. The C bracket might discover that they must first pay a deposit to rent a public bike, while the A group gets to rent them for free for 90 minutes…

    AI-powered directional microphones and smart cameras now dot the highways and streets of Shanghai. Drivers who honk excessively are automatically issued a ticket via Tencent’s WeChat, while their names, photographs, and national identity card numbers are displayed on nearby LED billboards. If a driver pulls over on the side of the road for more than seven minutes, they will trigger another instant traffic ticket. It isn’t just the ticket and the fine – points are deducted in the driver’s social credit score. When enough points are deducted, they will find it hard to book airline tickets or land a new job…’

    Russell describes even more menacing developments:

    ‘Lethal Autonomous Weapons (what the United Nations calls AWS) already exist. The clearest example is Israel’s Harop, a loitering munition with a ten-foot wingspan and a fifty-pound warhead. It searches for up to six hours in a given geographical region for any target that meets a given criterion and then destroys it.

    In 2016 the US Air Force demonstrated the in-flight deployment of 103 Perdix micro-drones from three F/A-18 fighters. Perdix are not pre-programmed synchronized individuals, they are a collective organism, sharing one distributed brain for decision-making and adapting to each other like swarms in nature’ (cf. the drone attack in the action film Angel Has Fallen)…

    In his book 21 Lessons for the 21st Century, Yuval Harari writes:

    ‘It is crucial to realize that the AI revolution is not just about computers getting faster and smarter. The better we understand the biochemical mechanisms that underpin human emotions, desires and choices, the better computers can become in analyzing human behavior, predicting human decisions, and replacing human drivers, bankers and lawyers…

    It turns out that our choices of everything from food to mates result not from some mysterious free will but rather from billions of neurons calculating probabilities within a split second. Vaunted 'human intuition' is in reality pattern recognition…

    This means that AI can outperform humans even in tasks that supposedly demand 'intuition.' In particular, AI can be better at jobs that demand intuitions about other people. Many lines of work – such as driving a vehicle in a street full of pedestrians, lending money to strangers, and negotiating a business deal – require the ability to correctly assess the emotions and desires of others. As long as it was thought that such emotions and desires were generated by an immaterial spirit, it seemed obvious that computers would never be able to replace human drivers, bankers and lawyers.

    Yet if these emotions and desires are in fact no more than biochemical algorithms, there is no reason computers cannot decipher these algorithms – and do so far better than any homo sapiens.’ (cf. Nick Bostrom’s Superintelligence)

    Russell points out that we underestimate AIs at our peril:

    ‘Whereas a human can read and understand one book in a week, a machine could read and understand every book ever written – all 150 million of them – in a few hours. The machine can see everything at once through satellites, robots, and hundreds of millions of surveillance cameras; watch all the world’s TV broadcasts; and listen to all the world’s radio stations and phone conversations. Very quickly it would gain a far more detailed and accurate understanding of the world and its inhabitants than any human could possibly hope to acquire…

    In the cyber realm, machines already have access to billions of effectors – namely, the displays on all the phones and computers in the world. This partly explains the ability of IT companies to generate enormous wealth with very few employees; it also points to the severe vulnerability of the human race to manipulation via screens…

    In his book Cultural Evolution, Ronald Inglehart, lead researcher of the World Values Survey, observes that despite rhetoric from Trump and other xenophobic demagogues:

    ‘Foreigners are not the main threat. If developed societies excluded all foreigners and all imports, secure jobs would continue to disappear, since the leading cause – overwhelmingly – is automation. Once artificial intelligence starts learning independently, it moves at a pace that vastly outstrips human intelligence. Humanity needs to devise the means to stay in control of artificial intelligence. I suspect that unless we do so within the next twenty years or so, we will no longer have the option.’

    So, our species’ remaining time may be limited, a momentous event predicted by the philosopher Nietzsche in Thus Spoke Zarathustra:

    ‘I teach you the Overman. Man is something that shall be overcome: what have you done to overcome him? All beings so far have created something beyond themselves. Do you want to be the ebb of this great flood? What is the ape to man? A laughingstock or a painful embarrassment. And man shall be just that for the Overman…

    The Overman is the meaning of the Earth. Let your will say: the Overman shall be the meaning of the Earth…’ And if Artificial Intelligence were the Overman?
    2 people found this helpful
    Report
  • Marcelo Henrirque Alves Marques
    5.0 out of 5 stars O livro mais importante que li nesta década
    Reviewed in Brazil on December 12, 2019
    Leio muita coisa. Minha área é empreendedorismo com tecnologia da informação. Anualmente, sempre faço um balanço para apontar qual foi o melhor livro do ano. Esse livro, é de um impacto tão grande, que para mim é considerado o melhor livro desta década.