Amodei Machines of
Amodei Machines of
I think and talk a lot about the risks of powerful AI. The company I’m the
CEO of, Anthropic, does a lot of research on how to reduce these risks.
Because of this, people sometimes draw the conclusion that I’m a
pessimist or “doomer” who thinks AI will be mostly bad or dangerous. I
don’t think that at all. In fact, one of my main reasons for focusing on risks
is that they’re the only thing standing between us and what I see as a
fundamentally positive future. I think that most people are
underestimating just how radical the upside of AI could be, just as I
think most people are underestimating how bad the risks could be.
In this essay I try to sketch out what that upside might look like—what a
world with powerful AI might look like if everything goes right. Of course
no one can know the future with any certainty or precision, and the effects
of powerful AI are likely to be even more unpredictable than past
technological changes, so all of this is unavoidably going to consist of
guesses. But I am aiming for at least educated and useful guesses, which
capture the flavor of what will happen even if most details end up being
wrong. I’m including lots of details mainly because I think a concrete
vision does more to advance discussion than a highly hedged and
abstract one.
Yet despite all of the concerns above, I really do think it’s important to
discuss what a good world with powerful AI could look like, while doing
our best to avoid the above pitfalls. In fact I think it is critical to have a
genuinely inspiring vision of the future, and not just a plan to fight fires.
Many of the implications of powerful AI are adversarial or dangerous, but
at the end of it all, there has to be something we’re fighting for, some
positive-sum outcome where everyone is better off, something to rally
people to rise above their squabbles and confront the challenges ahead.
Fear is one kind of motivator, but it’s not enough: we need hope as well.
What powerful AI (I dislike the term AGI)3 will look like, and when (or if) it
will arrive, is a huge topic in itself. It’s one I’ve discussed publicly and
could write a completely separate essay on (I probably will at some point).
Obviously, many people are skeptical that powerful AI will be built soon
and some are skeptical that it will ever be built at all. I think it could come
as early as 2026, though there are also ways it could take much longer.
But for the purposes of this essay, I’d like to put these issues aside,
assume it will come reasonably soon, and focus on what happens in the
5-10 years after that. I also want to assume a definition of what such a
system will look like, what its capabilities are and how it interacts, even
though there is room for disagreement on this.
I think the truth is likely to be some messy admixture of these two extreme
pictures, something that varies by task and field and is very subtle in its
details. I believe we need new frameworks to think about these details in a
productive way.
Economists often talk about “factors of production”: things like labor, land,
and capital. The phrase “marginal returns to labor/land/capital” captures
the idea that in a given situation, a given factor may or may not be the
limiting one – for example, an air force needs both planes and pilots, and
hiring more pilots doesn’t help much if you’re out of planes. I believe that
in the AI age, we should be talking about the marginal returns to
intelligence7, and trying to figure out what the other factors are that are
complementary to intelligence and that become limiting factors when
intelligence is very high. We are not used to thinking in this way—to asking
“how much does being smarter help with this task, and on what
timescale?”—but it seems like the right way to conceptualize a world with
very powerful AI.
With the above framework in mind, I’ll try to answer that question for the
five areas mentioned in the introduction.
Given all this, many biologists have long been skeptical of the value of AI
and “big data” more generally in biology. Historically, mathematicians,
computer scientists, and physicists who have applied their skills to biology
over the last 30 years have been quite successful, but have not had the
truly transformative impact initially hoped for. Some of the skepticism has
been reduced by major and revolutionary breakthroughs like AlphaFold
(which has just deservedly won its creators the Nobel Prize in Chemistry)
and AlphaProteo11, but there’s still a perception that AI is (and will
continue to be) useful in only a limited set of circumstances. A common
formulation is “AI can do a better job analyzing your data, but it can’t
produce more data or improve the quality of the data. Garbage in,
garbage out”.
But I think that pessimistic perspective is thinking about AI in the wrong
way. If our core hypothesis about AI progress is correct, then the right way
to think of AI is not as a method of data analysis, but as a virtual biologist
who performs all the tasks biologists do, including designing and running
experiments in the real world (by controlling lab robots or simply telling
humans which experiments to run – as a Principal Investigator would to
their graduate students), inventing new biological methods or
measurement techniques, and so on. It is by speeding up the whole
research process that AI can truly accelerate biology. I want to repeat
this because it’s the most common misconception that comes up
when I talk about AI’s ability to transform biology: I am not talking
about AI as merely a tool to analyze data. In line with the definition of
powerful AI at the beginning of this essay, I’m talking about using AI
to perform, direct, and improve upon nearly everything biologists do.
I’m going to the trouble of listing all these technologies because I want to
make a crucial claim about them: I think their rate of discovery could be
increased by 10x or more if there were a lot more talented, creative
researchers. Or, put another way, I think the returns to intelligence are
high for these discoveries, and that everything else in biology and
medicine mostly follows from them.
Thus, it’s my guess that powerful AI could at least 10x the rate of these
discoveries, giving us the next 50-100 years of biological progress in 5-10
years.14 Why not 100x? Perhaps it is possible, but here both serial
dependence and experiment times become important: getting 100 years
of progress in 1 year requires a lot of things to go right the first time,
including animal experiments and things like designing microscopes or
expensive lab facilities. I’m actually open to the (perhaps absurd-
sounding) idea that we could get 1000 years of progress in 5-10 years,
but very skeptical that we can get 100 years in 1 year. Another way to put
it is I think there’s an unavoidable constant delay: experiments and
hardware design have a certain “latency” and need to be iterated upon a
certain “irreducible” number of times in order to learn things that can’t be
deduced logically. But massive parallelism may be possible on top of
that15.
What about clinical trials? Although there is a lot of bureaucracy and
slowdown associated with them, the truth is that a lot (though by no
means all!) of their slowness ultimately derives from the need to rigorously
evaluate drugs that barely work or ambiguously work. This is sadly true of
most therapies today: the average cancer drug increases survival by a few
months while having significant side effects that need to be carefully
measured (there’s a similar story for Alzheimer’s drugs). This leads to
huge studies (in order to achieve statistical power) and difficult tradeoffs
which regulatory agencies generally aren’t great at making, again because
of bureaucracy and the complexity of competing interests.
Below I try to make a list of what we might expect. This is not based on
any rigorous methodology, and will almost certainly prove wrong in the
details, but it’s trying to get across the general level of radicalism we
should expect:
It is worth looking at this list and reflecting on how different the world will
be if all of it is achieved 7-12 years from now (which would be in line with
an aggressive AI timeline). It goes without saying that it would be an
unimaginable humanitarian triumph, the elimination all at once of most of
the scourges that have haunted humanity for millennia. Many of my
friends and colleagues are raising children, and when those children grow
up, I hope that any mention of disease will sound to them the way scurvy,
smallpox, or bubonic plague sounds to us. That generation will also
benefit from increased biological freedom and self-expression, and with
luck may also be able to live as long as they want.
The basic framework that I laid out for biology applies equally to
neuroscience. The field is propelled forward by a small number of
discoveries often related to tools for measurement or precise intervention
– in the list of those above, optogenetics was a neuroscience discovery,
and more recently CLARITY and expansion microscopy are advances in
the same vein, in addition to many of the general cell biology methods
directly carrying over to neuroscience. I think the rate of these advances
will be similarly accelerated by AI and therefore that the framework of
“100 years of progress in 5-10 years” applies to neuroscience in the same
way it does to biology and for the same reasons. As in biology, the
progress in 20th century neuroscience was enormous – for example we
didn’t even understand how or why neurons fired until the 1950’s. Thus, it
seems reasonable to expect AI-accelerated neuroscience to produce
rapid progress over a few years.
There is one thing we should add to this basic picture, which is that some
of the things we’ve learned (or are learning) about AI itself in the last few
years are likely to help advance neuroscience, even if it continues to be
done only by humans. Interpretability is an obvious example: although
biological neurons superficially operate in a completely different manner
from artificial neurons (they communicate via spikes and often spike rates,
so there is a time element not present in artificial neurons, and a bunch of
details relating to cell physiology and neurotransmitters modifies their
operation substantially), the basic question of “how do distributed, trained
networks of simple units that perform combined linear/non-linear
operations work together to perform important computations” is the
same, and I strongly suspect the details of individual neuron
communication will be abstracted away in most of the interesting
questions about computation and circuits22. As just one example of this, a
computational mechanism discovered by interpretability researchers in AI
systems was recently rediscovered in the brains of mice.
It’s my guess that these four routes of progress working together would,
as with physical disease, be on track to lead to the cure or prevention of
most mental illness in the next 100 years even if AI was not involved – and
thus might reasonably be completed in 5-10 AI-accelerated years.
Concretely my guess at what will happen is something like:
One topic that often comes up in sci-fi depictions of AI, but that I
intentionally haven’t discussed here, is “mind uploading”, the idea of
capturing the pattern and dynamics of a human brain and instantiating
them in software. This topic could be the subject of an essay all by itself,
but suffice it to say that while I think uploading is almost certainly possible
in principle, in practice it faces significant technological and societal
challenges, even with powerful AI, that likely put it outside the 5-10 year
window we are discussing.
The challenges facing the developing world are made even more
complicated by pervasive corruption in both private and public sectors.
Corruption creates a vicious cycle: it exacerbates poverty, and poverty in
turn breeds more corruption. AI-driven plans for economic development
need to reckon with corruption, weak institutions, and other very human
challenges.
Below I make some guesses about how I think things may go in the
developing world over the 5-10 years after powerful AI is developed:
I think of the issue as having two parts: international conflict, and the
internal structure of nations. On the international side, it seems very
important that democracies have the upper hand on the world stage when
powerful AI is created. AI-powered authoritarianism seems too terrible to
contemplate, so democracies need to be able to set the terms by which
powerful AI is brought into the world, both to avoid being overpowered by
authoritarians and to prevent human rights abuses within authoritarian
countries.
If we can do all this, we will have a world in which democracies lead on the
world stage and have the economic and military strength to avoid being
undermined, conquered, or sabotaged by autocracies, and may be able to
parlay their AI superiority into a durable advantage. This could
optimistically lead to an “eternal 1991”—a world where democracies have
the upper hand and Fukuyama’s dreams are realized. Again, this will be
very difficult to achieve, and will in particular require close cooperation
between private AI companies and democratic governments, as well as
extraordinarily wise decisions about the balance between carrot and stick.
Even if all that goes well, it leaves the question of the fight between
democracy and autocracy within each country. It is obviously hard to
predict what will happen here, but I do have some optimism that given a
global environment in which democracies control the most powerful AI,
then AI may actually structurally favor democracy everywhere. In
particular, in this environment democratic governments can use their
superior AI to win the information war: they can counter influence and
propaganda operations by autocracies and may even be able to create a
globally free information environment by providing channels of
information and AI services in a way that autocracies lack the technical
ability to block or monitor. It probably isn’t necessary to deliver
propaganda, only to counter malicious attacks and unblock the free flow
of information. Although not immediate, a level playing field like this
stands a good chance of gradually tilting global governance towards
democracy, for several reasons.
First, the increases in quality of life in Sections 1-3 should, all things equal,
promote democracy: historically they have, to at least some extent. In
particular I expect improvements in mental health, well-being, and
education to increase democracy, as all three are negatively correlated
with support for authoritarian leaders. In general people want more self-
expression when their other needs are met, and democracy is among
other things a form of self-expression. Conversely, authoritarianism
thrives on fear and resentment.
As with neuroscience and biology, we can also ask how things could be
“better than normal”—not just how to avoid autocracy, but how to make
democracies better than they are today. Even within democracies,
injustices happen all the time. Rule-of-law societies make a promise to
their citizens that everyone will be equal under the law and everyone is
entitled to basic human rights, but obviously people do not always receive
those rights in practice. That this promise is even partially fulfilled makes
it something to be proud of, but can AI help us do better?
For example, could AI improve our legal and judicial system by making
decisions and processes more impartial? Today people mostly worry in
legal or judicial contexts that AI systems will be a cause of discrimination,
and these worries are important and need to be defended against. At the
same time, the vitality of democracy depends on harnessing new
technologies to improve democratic institutions, not just responding to
risks. A truly mature and successful implementation of AI has the potential
to reduce bias and be fairer for everyone.
For centuries, legal systems have faced the dilemma that the law aims to
be impartial, but is inherently subjective and thus must be interpreted by
biased humans. Trying to make the law fully mechanical hasn’t worked
because the real world is messy and can’t always be captured in
mathematical formulas. Instead legal systems rely on notoriously
imprecise criteria like “cruel and unusual punishment” or “utterly without
redeeming social importance”, which humans then interpret—and often do
so in a manner that displays bias, favoritism, or arbitrariness. “Smart
contracts” in cryptocurrencies haven’t revolutionized law because
ordinary code isn’t smart enough to adjudicate all that much of interest.
But AI might be smart enough for this: it is the first technology capable of
making broad, fuzzy judgements in a repeatable and mechanical way.
All of these are somewhat vague ideas, and as I said at the beginning of
this section, I am not nearly as confident in their feasibility as I am in the
advances in biology, neuroscience, and poverty alleviation. They may be
unrealistically utopian. But the important thing is to have an ambitious
vision, to be willing to dream big and try things out. The vision of AI as a
guarantor of liberty, individual rights, and equality under the law is too
powerful a vision not to fight for. A 21st century, AI-enabled polity could
be both a stronger protector of individual freedom, and a beacon of hope
that helps make liberal democracy the form of government that the whole
world wants to adopt.
I think this question is more difficult than the others. I don’t mean that I am
necessarily more pessimistic about it than I am about the other questions
(although I do see challenges). I mean that it is fuzzier and harder to
predict in advance, because it relates to macroscopic questions about
how society is organized that tend to resolve themselves only over time
and in a decentralized manner. For example, historical hunter-gatherer
societies might have imagined that life is meaningless without hunting and
various kinds of hunting-related religious rituals, and would have imagined
that our well-fed technological society is devoid of purpose. They might
also have not understood how our economy can provide for everyone, or
what function people can usefully service in a mechanized society.
Nevertheless, it’s worth saying at least a few words, while keeping in mind
that the brevity of this section is not at all to be taken as a sign that I don’t
take these issues seriously—on the contrary, it is a sign of a lack of clear
answers.
First of all, in the short term I agree with arguments that comparative
advantage will continue to keep humans relevant and in fact increase their
productivity, and may even in some ways level the playing field between
humans. As long as AI is only better at 90% of a given job, the other 10%
will cause humans to become highly leveraged, increasing compensation
and in fact creating a bunch of new human jobs complementing and
amplifying what AI is good at, such that the “10%” expands to continue to
employ almost everyone. In fact, even if AI can do 100% of things better
than humans, but it remains inefficient or expensive at some tasks, or if
the resource inputs to humans and AI’s are meaningfully different, then
the logic of comparative advantage continues to apply. One area humans
are likely to maintain a relative (or even absolute) advantage for a
significant time is the physical world. Thus, I think that the human
economy may continue to make sense even a little past the point where
we reach “a country of geniuses in a datacenter”.
However, I do think in the long run AI will become so broadly effective and
so cheap that this will no longer apply. At that point our current economic
setup will no longer make sense, and there will be a need for a broader
societal conversation about how the economy should be organized.
While that might sound crazy, the fact is that civilization has successfully
navigated major economic shifts in the past: from hunter-gathering to
farming, farming to feudalism, and feudalism to industrialism. I suspect
that some new and stranger thing will be needed, and that it’s something
no one today has done a good job of envisioning. It could be as simple as
a large universal basic income for everyone, although I suspect that will
only be a small part of a solution. It could be a capitalist economy of AI
systems, which then give out resources (huge amounts of them, since the
overall economic pie will be gigantic) to humans based on some
secondary economy of what the AI systems think makes sense to reward
in humans (based on some judgment ultimately derived from human
values). Perhaps the economy runs on Whuffie points. Or perhaps
humans will continue to be economically valuable after all, in some way
not anticipated by the usual economic models. All of these solutions have
tons of possible problems, and it’s not possible to know whether they will
make sense without lots of iteration and experimentation. And as with
some of the other challenges, we will likely have to fight to get a good
outcome here: exploitative or dystopian directions are clearly also
possible and have to be prevented. Much more could be written about
these questions and I hope to do so at some later time.
Taking stock
Through the varied topics above, I’ve tried to lay out a vision of a world
that is both plausible if everything goes right with AI, and much better
than the world today. I don’t know if this world is realistic, and even if it is,
it will not be achieved without a huge amount of effort and struggle by
many brave and dedicated people. Everyone (including AI companies!) will
need to do their part both to prevent risks and to fully realize the benefits.
But it is a world worth fighting for. If all of this really does happen over 5 to
10 years—the defeat of most diseases, the growth in biological and
cognitive freedom, the lifting of billions of people out of poverty to share
in the new technologies, a renaissance of liberal democracy and human
rights—I suspect everyone watching it will be surprised by the effect it has
on them. I don’t mean the experience of personally benefiting from all the
new technologies, although that will certainly be amazing. I mean the
experience of watching a long-held set of ideals materialize in front of us
all at once. I think many will be literally moved to tears by it.
I think the Culture’s values are a winning strategy because they’re the sum
of a million small decisions that have clear moral force and that tend to
pull everyone together onto the same side. Basic human intuitions of
fairness, cooperation, curiosity, and autonomy are hard to argue with, and
are cumulative in a way that our more destructive impulses often aren’t. It
is easy to argue that children shouldn’t die of disease if we can prevent it,
and easy from there to argue that everyone’s children deserve that right
equally. From there it is not hard to argue that we should all band together
and apply our intellects to achieve this outcome. Few disagree that people
should be punished for attacking or hurting others unnecessarily, and
from there it’s not much of a leap to the idea that punishments should be
consistent and systematic across people. It is similarly intuitive that
people should have autonomy and responsibility over their own lives and
choices. These simple intuitions, if taken to their logical conclusion, lead
eventually to rule of law, democracy, and Enlightenment values. If not
inevitably, then at least as a statistical tendency, this is where humanity
was already headed. AI simply offers an opportunity to get us there more
quickly—to make the logic starker and the destination clearer.
Thanks to Kevin Esvelt, Parag Mallick, Stuart Ritchie, Matt Yglesias, Erik
Brynjolfsson, Jim McClave, Allan Dafoe, and many people at Anthropic for
reviewing drafts of this essay.
To the winners of the 2024 Nobel prize in Chemistry, for showing us all
the way.
Footnotes
1https://allpoetry.com/All-Watched-Over-By-Machines-Of-Loving-Grace
5This is roughly the current speed of AI systems – for example they can
read a page of text in a couple seconds and write a page of text in maybe
20 seconds, which is 10-100x the speed at which humans can do these
things. Over time larger models tend to make this slower but more
powerful chips tend to make it faster; to date the two effects have roughly
canceled out. ↩
6This might seem like a strawman position, but careful thinkers like Tyler
Cowen and Matt Yglesias have raised it as a serious concern (though I
don’t think they fully hold the view), and I don’t think it is crazy. ↩
7The closest economics work that I’m aware of to tackling this question is
work on “general purpose technologies” and “intangible investments” that
serve as complements to general purpose technologies. ↩
14I didn't want to clog up the text with speculation about what specific
future discoveries AI-enabled science could make, but here is a
brainstorm of some possibilities:
— Design of better computational tools like AlphaFold and AlphaProteo —
that is, a general AI system speeding up our ability to make specialized AI
computational biology tools.
— More efficient and selective CRISPR.
— More advanced cell therapies.
— Materials science and miniaturization breakthroughs leading to better
implanted devices.
— Better control over stem cells, cell differentiation, and de-
differentiation, and a resulting ability to regrow or reshape tissue.
— Better control over the immune system: turning it on selectively to
address cancer and infectious disease, and turning it off selectively to
address autoimmune diseases. ↩
15AI may of course also help with being smarter about choosing what
experiments to run: improving experimental design, learning more from a
first round of experiments so that the second round can narrow in on key
questions, and so on. ↩
20As an example, I’m told that an increase in productivity growth per year
of 1% or even 0.5% would be transformative in projections related to
these programs. If the ideas contemplated in this essay come to pass,
productivity gains could be much larger than this. ↩
21The media loves to portray high status psychopaths, but the average
psychopath is probably a person with poor economic prospects and poor
impulse control who ends up spending significant time in prison. ↩
22I think this is somewhat analogous to the fact that many, though likely
not all, of the results we’re learning from interpretability would continue to
be relevant even if some of the architectural details of our current artificial
neural nets, such as the attention mechanism, were changed or replaced
in some way. ↩
25For example, cell phones were initially a technology for the rich, but
quickly became very cheap with year-over-year improvements happening
so fast as to obviate any advantage of buying a “luxury” cell phone, and
today most people have phones of similar quality. ↩
26This is the title of a forthcoming paper from RAND, that lays out roughly
the strategy I describe. ↩
29I am breaking my own rule not to make this about science fiction, but
I’ve found it hard not to refer to it at least a bit. The truth is that science
fiction is one of our only sources of expansive thought experiments about
the future; I think it says something bad that it’s entangled so heavily with
a particular narrow subculture. ↩
Back to top
Privacy policy