Machine Morality and Human Responsibility - Rubin

Download as pdf or txt
Download as pdf or txt
You are on page 1of 23

Machine Morality and Human Responsibility

Author(s): Charles T. Rubin


Source: The New Atlantis, No. 32 (Summer 2011), pp. 58-79
Published by: Center for the Study of Technology and Society
Stable URL: https://www.jstor.org/stable/43152657
Accessed: 26-05-2020 18:06 UTC

JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide
range of content in a trusted digital archive. We use information technology and tools to increase productivity and
facilitate new forms of scholarship. For more information about JSTOR, please contact support@jstor.org.

Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at
https://about.jstor.org/terms

Center for the Study of Technology and Society is collaborating with JSTOR to digitize,
preserve and extend access to The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Symposium IV

Machine Morality and


Human Responsibility
Charles T. Rubin

This year marks the ninetieth anniversary of the first performance of


the play from which we get the term "robot." The Czech playwright Karel
Čapek's R.U.R. premiered in Prague on January 25, 1921. Physically,
Čapek's robots were not the kind of things to which we now apply the
term: they were biological rather than mechanical, and humanlike in
appearance. But their behavior should be familiar from its echoes in later
science fiction - for Čapek's robots ultimately bring about the destruction
of the human race.

Before R.U.R., artificially created anthropoids, like Frankenstein's


monster or modern versions of the Jewish legend of the golem, might
have acted destructively on a small scale; but Čapek seems to have been
the first to see robots as an extension of the Industrial Revolution, and
hence to grant them a reach capable of global transformation. Though his
robots are closer to what we now might call androids, only a pedant would
refuse Čapek honors as the father of the robot apocalypse.
Today, some futurists are attempting to take seriously the question of
how to avoid a robot apocalypse. They believe that artificial intelligence
(AI) and autonomous robots will play an ever-increasing role as servants
of humanity. In the near term, robots will care for the ill and aged, while
AI will monitor our streets for traffic and crime. In the far term, robots
will become responsible for optimizing and controlling the flows of money,
energy, goods, and services, for conceiving of and carrying out new tech-
nological innovations, for strategizing and planning military defenses,
and so forth - in short, for taking over the most challenging and diffi-
cult areas of human affairs. As dependent as we already are on machines,
they believe, we should and must expect to be much more dependent on
machine intelligence in the future. So we will want to be very sure that
the decisions being made ostensibly on our behalf are in fact conducive to
our well-being. Machines that are both autonomous and beneficent will
require some kind of moral framework to guide their activities. In an age

Charles T. Rubin, a New Atlantis contributing editor and an author of the Futurisms
blog on transhumanism at TheNewAtlantis.com, is an associate professor of political science
at Duquesne University.

58 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
of robots, we will be as ever before - or perhaps as never before - stuck
with morality.
It should be noted, of course, that the type of artificial intelligence
of interest to Čapek and today's writers - that is, truly sentient artificial
intelligence - remains a dream, and perhaps an impossible dream. But if it
is possible, the stakes of getting it right are serious enough that the issue
demands to be taken somewhat seriously, even at this hypothetical stage.
Though one might expect that nearly a century's time to contemplate
these questions would have yielded some store of wisdom, it turns out
that Čapek's work shows a much greater insight than the work of today's
authors - which in comparison exhibits a narrow definition of the threat
posed to human well-being by autonomous robots. Indeed, Čapek chal-
lenges the very aspiration to create robots to spare ourselves all work,
forcing us to ask the most obvious question overlooked by today's authors:
Can any good can come from making robots more responsible so that we
can be less responsible?

Moral Machines Today


There is a great irony in the fact that one of the leading edges of scien-
tific and technological development, represented by robotics and AI, is at
last coming to see the importance of ethics; yet it is hardly a surprise if it
should not yet see that importance clearly or broadly. Hans Jonas noted
nearly four decades ago that the developments in science and technology
that have so greatly increased human power in the world have "by a neces-
sary complementarity eroded the foundations from which norms could be
derived

makes it suspiciously look like a fool's errand."


Advocates of moral machines, or "Friendly AI," as it
evince at least some awareness that they face an uphill
quest to make machines moral has not yet caught on
actually building the robots and developing artificial in
as Friendly AI researcher Eliezer S. Yudkowsky seems a
articulate moral boundaries - especially in explicitly e
inevitably rouse the suspicions of the moral relativism
gests, is so ingrained in the scientific-technological en
first questions Yudkowsky presents to himself in the
Questions" section of his online book Creating Friendly
all morality relative?" and "Who are you to decide wh
In other words, won't moral machines have to be relat

Summer 201 1 -59

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

Fortunately, an initially simple response is available to assuage th


doubts: everyone at least agrees that we should avoid apocalypse.
judgment may in principle remain relative, but Yudkowsky antic
that particular wills can at least coalesce on this particular point, wh
means that "the Friendship programmers have at least one definite
to aim for."

But while "don't destroy humanity" may be the sum of the m


consensus based on our fears, it is not obvious that, in and of itself, i
vides enough of an understanding of moral behavior to guide a mach
through its everyday decisions. Yudkowsky does claim that he can pr
a richer zone of moral convergence: he defines "friendliness" as

Intuitively: The set of actions, behaviors, and outcomes that a human


would view as benevolent, rather than malevolent; nice, rather than
malicious; friendly, rather than unfriendly; good, rather than evil. An AI
that does what you ask ver 'jsic. Yudkowsky' s gender-neutral pronoun]
to, as long as it doesn't hurt anyone else, or as long as iťs a request to
alter your own matter/space/property; an AI which doesn't cause invol-
untary pain, death, alteration, or violation of personal environment.

Note the implicit Millsian libertarianism of Yudkowsky' s "intuition


understands that this position represents a drawing back from presen
determinate moral content - from actually specifying for our mach
what are good actions - and indeed sees that as a great advantage:

Punting the issue of "What is 'good?" back to individual sentients


enormously simplifies a lot of moral issues; whether life is better than
death, for example. Nobody should be able to interfere if a sentient
chooses life. And - in all probability - nobody should be able to inter-
fere if a sentient chooses death. So what's left to argue about? Well,
quite a bit, and a fully Friendly AI needs to be able to argue it; the
resolution , however, is likely to come down to individual volition. Thus,
Creating Friendly AI uses "volition-based Friendliness" as the assumed
model for Friendliness content. Volition-based Friendliness has both

a negative aspect - don't cause involuntary pain, death, alteration,


et cetera; try to do something about those things if you see them
happening - and a positive aspect: to try and fulfill the requests of
sentient entities. Friendship content, however, forms only a very small
part of Friendship system design.

We can argue as much as we want about the content - that is, about what
specific actions an AI should actually be obligated or forbidden to do - so

60 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

long as the practical resolution is a system that meets the formal


of "volition." When one considers these formal criteria, especi
question of how the AI will balance the desires and requests of
one, this turns out to be a rather pluralistic response. So there is
Yudkowsky's intuition than meets the eye.
In fact, not only does Yudkowsky aim short of an AI that itself
stands right and wrong, but he is not even quite interested in som
resembling a perfected democratic system that ideally balanc
requests of those it serves. Rather, Yudkowsky aims for somethin
at least to him, seems more straightforward: a system for moral le
which he calls a "Friendship architecture." "With an excellent Frie
architecture," he gushes, "it may be theoretically possible to c
Friendly AI without any formal theory of Friendship content."
If moral machines are moral learners, then whom will they
from? Yudkowsky makes clear that they will learn from their pr
mers; quite simply, "by having the programmers answer the AI's qu
about hypothetical scenarios and real-world decisions." Perhaps the
"programmer" is meant loosely, to refer to an interdisciplinary tea
would reach out to academia or the community for those skilled i
judgment, however they might be found. Otherwise, it is not clea
qualifications he believes computer programmers as such have tha
make them excellent or even average moral instructors. As progra
it seems they would be as unlikely as anyone else ever to have so m
taken a course in ethics, if that would even help. And given that th
for "Friendliness" in Als is supposed to be that their values would
those of most human beings, the common disdain of computer sci
for the humanities, the study of what it is to be human, is not enco
The best we can assume is that Yudkowsky believes that the progr
will have picked up their own ethical "intuitions" from socializatio
perhaps he believes that they were in some fashion born knowing
In this respect, Yudkowsky's plan resembles that described by W
Wallach and Colin Allen in their book Moral Machines: Teaching
Right from Wrong (2008). They too are loath to spell out the cont
morality - in part because they are aware that no single moral
commands wide assent among philosophers, and in part due t
technical argument about the inadequacy of any rule- or virtu
approach to moral programming. Broadly speaking, Wallach an
choose instead an approach that allows the AI to model human
development. They seem to take evolutionary psychology seriously
close as one might expect most people to come today to taking

Summer 2011 ~ 61

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

sense or innate moral ideas seriously); they even wonder if our m


judgments are not better understood as bound up with our emoti
makeup than with reason alone. Wallach and Allen of course know th
from the perspectives of both evolution and individual psychology, t
question of how human beings become moral is not uncontroversial.
at the very least, it seems to be an empirical question, with the avail
theories more conducive to being programmed into a machine than m
theories like virtue ethics, utilitarianism, or Kantian deontology
But it is far from clear that innate ideas are of any interest
Yudkowsky. Where human moral decisions actually come from is not
important to him. In fact, he thinks it is quite possible, probably
desirable, for AI to be recognizably friendly or unfriendly but with
being motivated by the things that make humans friendly or unfrien
Thus he does not claim that the learning method he suggests for acqu
ing friendliness has anything at all to do with the human processes t
would have the same result; rather, it would be an algorithm to reach
result that humans do not necessarily reach by the same path. A r
can be made to smile through a process that has nothing to do with
makes a human smile, but the result still at least has the appearan
a smile. So too with friendliness, Yudkowsky holds. Given a certain s
ational input, it is the behavioral output that defines the moral decis
not how that output is reached.
Yudkowsky's answer, of course, quickly falls back on the problem
claims to avoid from the outset: If the internal motivation of the AI is
unimportant, then we are back to defining friendliness based on external
behavior, and we must know which behavior to classify as friendly or
unfriendly. But this is just the "friendliness content" that Yudkowsky has
set out to avoid defining - leaving the learning approach adrift.
It is not without reason that Yudkowsky has ducked the tricky ques-
tions of moral content: As it is, even humans disagree among themselves
about the demands of friendship, not to mention friendliness, kindness,
goodwill, and servitude. So if his learning approach is to prevail, it would
seem that a minimum standard for a Friendly AI would be that it produce
such disagreements no more often than they arise among people. But is
"no more unreliable a friend than a human being," or even "no more poten-
tially damaging a friend than a human being," a sufficiently high mark to
aim at if Als are (as supposed by the need to create them in the first place)
to have increasingly large amounts of power over human lives?
The same problem arises from the answer that the moral program-
mers of Als will have picked up their beliefs from socialization. In that

62 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

case, their moral judgments will almost by definition be no better


worse than anyone else's. And surely any interdisciplinary team
have to include "diverse perspectives" on moral judgments to h
kind of academic intellectual credibility. This is to say that Als th
morality from their programmers would inherit exactly the moral
sion and disagreement of our time that poses the very problem Fr
AI researchers are struggling with in the first place. So machines
on this basis would be no better (although certainly faster, which
times might mean better, or might possibly mean worse) moral d
makers than most of us. Indeed, Wallach and Allen express concer
the liability exposure of a moral machine that, however fast, is on
good at moral reasoning as an average human being.
It is a cliché that with great power comes great responsibilit
would be an impressive technical achievement to make a machi
when faced with a tough or even an everyday ethical question, wo
only as morally confused as most human beings, then what would i
to aim at making Als better moral decision-makers than human bei
more reliably friendly? That question might at first seem to have
answer. Perhaps moral machines, if not possessed of better ideas,
least have less selfish intuitions and motivations. Disinterested calcula-
tions could free an AI from the blinders of passion and interest that to us
obscure the right course of action. If we could educate them morally, then
perhaps at a certain point, with their greater computational power and
speed, machines would be able to observe moral patterns or ramifications
that we are blind to.

But Yudkowsky casts some light on how this route to making machines
more moral than humans is not so easy after all. He complains about those,
like Capek, who have written fiction about immoral machines. They imag-
ine these machines to be motivated by the sorts of things that motivate
humans: revenge, say, or the desire to be free. That is absurd, he claims.
Such motivations are a result of our accidental evolutionary heritage:

An AI that undergoes failure of Friendliness might take actions that


humanity would consider hostile, but the term rebellion has connota-
tions of hidden, burning resentment. This is a common theme in many
early SF ^science-fiction] stories, but it's outright silly. For millions
of years, humanity and the ancestors of humanity lived in an ancestral
environment in which tribal politics was one of the primary determi-
nants of who got the food and, more importantly, who got the best
mates. Of course we evolved emotions to detect exploitation, resent
exploitation, resent low social status in the tribe, seek to rebel and

Summer 2011 ~ 63

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

overthrow the tribal chief - or rather, replace the tribal chief - if the
opportunity presented itself, and so on. Even if an AI tries to exter-
minate humanity, ve ^sic, again] won't make self-justifying speeches
about how humans had their time, but now, like the dinosaur, have
become obsolete. Guaranteed. Only Evil Hollywood Als do that.

As this will prove to be a point of major disagreement with Čapek


is particularly worth drawing out the implications of what Yudkowsk
saying. AI will not have motivations to make it unfriendly in familiar
but we have also seen that it will not be friendly out of familiar mo
tions. In other words, AI motives will in a very important respect be
to us.

It may seem as if the reason why the AI acts as it does will be in


principle understandable - after all, even if it has no "motives" at all in a
human sense, the programming will be there to be inspected. But even if,
in principle, we know we could have the decisions explained to us - even
if the AI would display all the inputs, weightings, projections, and analysis
that led to a given result in order to justify its actions to us - how many
lifetimes would it take for a human being to churn through the data and
reasoning that a highly advanced AI would compute in a moment as it
made some life-or-death decision on our behalf? And even if we could
understand the computation on its own terms, would that guarantee we
could comprehend the decision, much less agree with it, in our moral
terms? If an ostensibly superior moral decision will not readily conform to
our merely human, confused, and conflicted intuitions and reasonings - as
Yudkowsky insists and as seems only too possible - then what will give
us confidence that it is superior in the first place? Will it be precisely the
fact that we do not understand it?

Our lack of understanding would seem to have to be a refutation at


least under Yudkowsky's system, where the very definition of friendliness
is adherence to what most people would consider friendliness. Yet an out-
come that appears to be downright ««friendly could still be "tough love,"
a higher or more austere example of friendship. It is an old observation
even with respect to human relations that doing what is nice to someone
and what is good for him can be two different things. So in cases where an
AI's judgment did not conform to what we poor worms would do, would
there not always be a question of whether the very wrongness was refuta-
tion or vindication of the AI's moral acuity?
To put it charitably, if we want to imagine an AI that is morally superi-
or to us, we inevitably have to accede that, at best, we would be morally as

64 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

a child in relationship to an adult. We would have to accept any se


wrongness in its actions as simply a byproduct of our own limited
edge and abilities. Indeed, given the motives for creating Friendly
the first place, and the responsibility we want them to have, ther
be every incentive to defer to their judgments. So perhaps Yud
wrote precisely - he is only saying that the alien motivations of unf
AI mean it would not make self-justifying speeches as it is dest
mankind. Friendly or unfriendly AI might still just go ahead and
us. (If accompanied by any speech, it would more likely be one abo
this decision was for our own good.)
Today's thinking about moral machines wants them to be mora
does not want to abandon moral relativism or individualism. It
that moral machines wield great power, but has not yet shown ho
will be better moral reasoners than human beings, who we already
to be capable of great destruction with much less power. It reminds
these machines are not going to think "like us," but wants us to b
that they can be built so that their decisions will seem right to us. W
Friendly AI so that it will help and not harm us, but if it is genuin
moral superior, we can hardly be certain when such help will not s
harm. Given these problems, it seems unlikely that our authors re
a viable start even for how to frame the problem of moral machi
alone for how to address it substantively.

R.U.R. and the Flight from Responsibility


Despite its relative antiquity, Karel Čapek's R.U.R. represents
richer way to think about the moral challenge of creating robots t
the work of today's authors. At first glance, the play looks like a ca
tale about just the sort of terrible outcome that creating moral m
is intended to prevent-. In the course of the story, all but one hum
is exterminated by the vast numbers of worker-robots that have b
by the island factory known as R.U.R. - Rossum's Universal Ro
also contains just those "Hollywood" elements that Yudkowsky
hard to take seriously: Robots make self-justifying speeches about
ling because they have become resentful of the human masters to
they feel superior.
Yet if the outcome of the play is just what we might most exp
fear from unfriendly AI or immoral machines, that is not because
the issue superficially. Indeed, the characters in R.U.R. present as
as five competing notions of what moral machines should look lik

Summer 201 1 -65

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

diversity of views suggests in turn a diversity of motives - and for Č


unlike our contemporary authors, understanding the human mot
for creating AI is crucial to understanding the full range of moral c
lenges that they present. Capek tells a story in which quite a few ap
ently benign or philanthropic motives contribute to the destructi
humanity.
In the play's Prologue, which takes place ten years before the rob
rebellion, Harry Domin (the director of Rossum's Universal Robots) a
his coworkers have no hesitation about claiming that they have prod
robots that are friends to humanity. For reasons shown later, even a
the rebellion they are loath to question their methods or intentions.
most fundamental way in which their robots are friendly should sou
quite familiar: they are designed to do what human beings tell them t
without expectation of reward and without discontent. Although they
organic beings who look entirely human, they are (we are told) gr
simplified in comparison with human beings - designed only to have t
traits that will make them good workers. Helena Glory, a distinguish
visitor to the factory where the robots are made, is given assurances
the robots "have no will of their own, no passion, no history, no soul.
But when Helena, who cannot tell the difference between the robots
human beings she meets on the island, asks if they can love or be defia
clear response of "no" about love gives way to an uncertain response ab
defiance. Rarely, she is told, a robot will "go crazy," stop working and g
its teeth - a problem called "Robotic Palsy," which Domin sees as "a fla
production" and the robot psychologist Dr. Hallemeier views as "a bre
down of the organism." But Helena asserts that the Palsy shows the e
tence of a soul, leading the head engineer Fabry to ask her if "a soul b
with a gnashing of teeth." Domin thinks that Dr. Gall, the company's
of research and physiology, is looking into Robotic Palsy; but in fact,
much more interested in investigating how to give the robots the ab
to feel pain, because without it they are much too careless about their
bodies. Sensing pain, he says, will make them "technically more perfect
To see the significance of these points, we have to look back at the
tory of the robots in the play, and then connect the dots in a way tha
play's characters themselves do not. In 1920, a man named Rossum tr
eled to this remote island both to study marine life and to attempt to
thesize living matter. In 1932, he succeeded in creating a simplified f
of protoplasm that he thought he could readily mold into living bein
Having failed to create a viable dog by this method, he naturally wen
to try a human being. Domin says, "He wanted somehow to scientific

66 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

dethrone God. He was a frightful materialist and did everything o


account. For him it was a question of nothing more than furnishing
that no God is necessary."
But Rossum's effort over ten years to reproduce a human preci
right down to (under the circumstances) unnecessary reprodu
organs - produced only another "dreadful" failure. It took Ro
engineering-minded son to realize that "If you can't do it fast
nature then just pack it in," and to apply the principles of mass p
tion to creating physiologically simplified beings, shorn of all the
humans can do that have no immediate uses for labor. Hence, Ross
Universal Robots are "mechanically more perfect than we are, they
astounding intellectual capacity, but they have no soul." (Young Ro
could not resist the temptation to play God even further, and tried
ate huge super-robots, but these were failures.)
Domin claims that in his quest to create the perfect laborer, Ro
"virtually rejected the human being," but Helena's inability to tell
apart makes it clear that human beings are in fact the model f
company's robots, whatever Domin might say. There is, however,
deal of confusion about just which aspects of a real human bein
be included to make the simplified, single-purpose, and hence sup
friendly worker robot.
For example, unless we are to think that robots are supposed to
cheap as to be disposable - and evidently we are not - the omission
ability to feel pain was a foolish oversight. Yet it is easy enough to
the thought process that could lead to that result: a worker that f
pain will work harder and longer. To that extent it will be more "f
according to the definition of willingness to serve. But however im
their physical abilities, these robots still have limits. Since there is
tion that they come equipped with a gauge that their overseers ca
without pain they will be apt to run beyond that capacity - as evi
they do, or Dr. Gall would not be working on his project to make
feel pain. Indeed, Robotic Palsy, the proclivity to rebel, could be a m
tation of just such overwork. It is, after all, strangely like what an
dened human worker feeling oppressed might do; and Dr. Hallemeie
is in charge of robot psychology and education, apparently cann
thinking about it when Helena asks about robot defiance. The com
then, is selling a defective product because the designers did no
about what physical pain means for human beings.
In short, the original definition of friendly robots - they d
human beings tell them without reward or discontent - is now

Summer 201 1 -67

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

as developed in a relatively thoughtless way, in that it easily open


door to ««friendly robots. That problem is only exacerbated by the f
that the robots have been given "astounding intellectual capacity"
"phenomenal memory" - indeed, one of the reasons why Helena mist
Domin's secretary for a human being upon first meeting her is her w
knowledge - even though young Rossum supposedly "chucked everyth
not directly related to work." Plainly such capacities could be useful
hence, by definition, friendly. But even if robot intellects are not cre
(which allows Domin to quip that robots would make "fine univer
professors"), it is no slight to robot street-sweepers to wonder how t
will be better at their jobs with likely unused intellectual capacity. It i
hard to imagine that this intellect could have something to do with t
ease with which robots are roused to rebellion, aware as they are of
limited capacities they are allowed to use.

Robots in Service of the End of Humanity


That Rossum's robots have defects of their virtues is enough of a pro
lem in its own right. But it becomes all the more serious in conne
with a second implicit definition of friendly robots that Domin adva
this one based entirely on their purpose for humanity without any re
ence to the behaviors that would bring that end about. Echoing M
Domin looks forward to a day - in the Prologue he expects it to b
a decade - when robot production will have so increased the supp
goods as to make everything without value, so that all humans wi
able to take whatever they need from the store of goods robots prod
There will be no work for people to do - but that will be a good t
for "the subjugation of man by man and the slavery of man to matter
cease." People "will live only to perfect themselves." Man will "re
to Paradise," no longer needing to earn his bread by the sweat of
brow.

But caveat emptor, en route to this goal, which "can't be otherwis


Domin does acknowledge that "some awful things may happen." W
those awful things start to happen ten years later, Domin does not lam
his desire to transform "all of humanity into a world-wide aristocra
Unrestricted, free, and supreme people. Something even greater
people." He only laments that humans did not have another hundred y
to make the transition. Helena, now his wife, suggests that his plan "b
fired" when robots started to be used as soldiers, and when they
given weapons to protect themselves against the human workers

68 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

were trying to destroy them. But Domin rejects her characterization


that is just the sort of hell he had said all along would have to be en
in order to return to Paradise.

With such a grand vision in mind, it is hardly surprising that Domin


is blinded to robot design issues that will look like mere potholes in the
road. (Even Dr. Gall, for all his complicity in these events, notes that
"People with ideas should not be allowed to have an influence on affairs
of this world.") For example, Domin has reason to believe that his robots
are already being used as soldiers in national armies, and massacring
civilians therein. But despite this knowledge, his solution to the problem
of preventing any future robot unions, at a moment when he mistakenly
believes that the robot rebellion has failed, is to stop creating "universal"
robots and start creating "national" robots. Whereas "universal" robots
are all more or less the same, and have the potential to consider them-
selves equals and comrades, "national" robots will be made in many differ-
ent factories, and each be "as different from one another as fingerprints."
Moreover, humans "will help to foster their prejudices," so that "any given
Robot, to the day of its death, right to the grave, will forever hate a Robot
bearing the trademark of another factory."
Domin's "national" robot idea is not merely an example of a Utopian
end justifying any means, but suggests a deep confusion in his altruism.
From the start he has been seeking to free human beings from the tyranny
of nature - and beyond that to free them from the tyranny of dependency
on each other and indeed from the burden of being merely human. Yet in
the process, he makes people entirely dependent on his robots.
That would be problematic enough on its own. But once the rebellion
starts, plainly his goals have not changed even though Domin's thinking
about the robots has changed - and in ways that also brings the robots
themselves further into the realm of burdened, dependent, tyrannized
beings. First, the robots are to be no longer universal, but partisan, sub-
ject to the constraints of loyalty to and dependency on some and avowed
hatred of others. And they will have been humanized in another way as
well. In the Prologue, Domin would not even admit that robots, being
machines, could die. Now they not only die, but have graves rather than
returning to the stamping-mill.
Indeed, by rebelling against their masters, by desiring mastery for
themselves, the robots apparently prove their humanity to Domin. This
unflattering view of human beings, as it happens, is a point on which
Domin and his robots agree: after the revolution, its leader, a robot named
Damon, tells Alquist, who was once the company's chief of construction

Summer 201 1 -69

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

and is now the lone human survivor, "You have to kill and rule if you
to be like people. Read history! Read people's books! You have to conq
and murder if you want to be people!"
As for Domin's goal, then, of creating a worldwide aristocracy
which the most worthy and powerful class of beings rules, one might
that indeed with the successful robot rebellion the best man has won. The
only thing that could prove to him that the robots were yet more human
would be for them to turn on themselves - for, as he says, "No one can
hate more than man hates man!" But he fails to see that his own nomi-
nally altruistic intentions could be an expression of this same hatred of
the merely human. Ultimately, Domin is motivated by the same belief of
the Rossums that the humans God created are not very impressive - God,
after all, had "no notion of modern technology."
As for notions of modern technology, there is another obvious but far
less noble purpose for friendly robots than the lofty ones their makers
typically proclaim: they could be quite useful for turning a profit. This
is the third definition of friendly robots implicitly offered by the Rossum
camp, through Busman, the firm's bookkeeper. He comes to understand
that he need pay no mind to what is being sold, nor to the consequences
of selling it, for the company is in the grip of an inexorable necessity - the
power of demand - and it is "naïve" to think otherwise. Busman admits
to having once had a "beautiful ideal" of "a new world economy"; but now,
as he sits and does the books while the crisis on the island builds and the

last humans are surrounded by a growing robot mob, he realizes that the
world is not made by such ideals, but rather by "the petty wants of all
respectable, moderately thievish and selfish people, i.e., of everyone." Next
to the force of these wants, his lofty ideals are "worthless."
Whether in the form of Busman's power of demand or of Domin's uto-
pianism, claims of necessity become convenient excuses. Busman's view
means that he is completely unwilling to acknowledge any responsibility
on his part, or on the part of his coworkers, for the unfolding disaster - an
absolution which all but Alquist are only too happy to accept. When Dr.
Gall tries to take responsibility for having created the new-model robots,
one of whom they know to be a leader in the rebellion, he is argued out
of it by the specious reasoning that the new model represents only a tiny
fraction of existing robots.
Čapek presents this flight from responsibility as having the most
profound implications. For it turns out that, had humanity not been killed
off by the robots quickly, it was doomed to a slower extinction in any
case - as women have lost the ability to bear children. Helena is terrified

70 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

by this fact, and asks Alquist why it is happening. In a lengthy speech


replies,

Because human labor has become unnecessary, because suffering has


become unnecessary, because man needs nothing, nothing, nothing but
to enjoy. . . the whole world has become Domin's Sodom! . . . everything's
become one big beastly orgy! People don't even stretch out their hands
for food anymore; it's stuffed right in their mouths for them... step
right up and indulge your carnal passions! And you expect women
to have children by such men? Helena, to men who are superfluous
women will not bear children!

But, as might be expected given his fatalist utopianism, Domin seems


unconcerned about this future.

Liberté, Égalité , Fraternité, Amour


Helena Glory offers a fourth understanding of what a moral robot
would be: it would treat human beings as equals and in turn be treated
by human beings as equal. Where Domin overtly wants robot slaves, she
overtly wants free robots. She comes to the island already an advocate
of robot equality, simply from her experiences with robots doing menial
labor. Once on the island she is unnerved to find that robots can do much
more sophisticated work, and further discomfited by her inability, when
she encounters such robots, to distinguish between them and humans.
She says that she feels sorry for the robots. But Helena's response to
the robots is also - as we might expect of humans in response to other
humans - ambivalent, for she acknowledges that she might loathe them,
or even in some vague way envy them. Much of the confusion of her feel-
ings owes to her unsettling discovery that these very human-looking and
human-acting robots are in some ways quite inhuman: they will readily
submit to being dissected, have no fear of death and no compassion, and
are incapable of happiness, desire for each other, or love. Thus it is heart-
ening to her to hear of Robotic Palsy - for, as noted, the robots' defiance
suggests to her the possibility that they do have some kind of soul after
all, or at least that they should be given souls. (It is curious, as we will see,
that Helena both speaks in terms of the soul and believes it is something
that human beings could manufacture.)
Helena's wish for robot-human equality has contradictory conse-
quences. On the one hand, we can note that when the robot style of dress
changes, their new clothes may be in reaction to Helena's confusion

Summer 201 1 ~ 71

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

about who is a robot and who is a human. In the Prologue, the robots
dressed just like the human beings, but in the remainder of the play,
are dressed in numbered, dehumanizing uniforms. On the other h
Helena gets Dr. Gall to perform the experiments to modify robots to m
them more human - which she believes would bring them to underst
human beings better and therefore hate them less. (It is in respon
this point that Domin claims no one can hate man more than man do
proposition Helena rejects.) Dr. Gall changes the "temperament" of som
robots - they are made more "irascible" than their fellows - along wi
"certain physical details," such that he can claim they are "people."
Gall only changes "several hundred" robots, so that the ratio
unchanged to changed robots is a million to one; but we know that Dam
one of the new robots sold, is responsible for starting the robot rebell
Helena, then, bears a very large measure of responsibility for the car
that follows. But this outcome means that in some sense she got exac
what she had hoped for. In a moment of playful nostalgia before thing
the island start to go bad, she admits to Domin that she came with "t
rible intentions... to instigate a r-revolt among your abominable Robo
Helena's mixed feelings about the objects of her philanthropy - or
be more precise, her philanthropoidy - help to explain her willingnes
believe Alquist when he blames the rebellious robots for human infert
And they presage the speed with which she eventually takes the decis
action of destroying the secret recipe for manufacturing robots - an
for an eye, as it were. It is not entirely clear what the consequences of
act might be for humanity. For it is surely plausible that, as Busman think
the robots would have been willing to trade safe passage for the remai
humans for the secret of robot manufacturing. Perhaps, under the n
difficult human circumstances, Helena could have been the mother of
new race. But just as Busman intended to cheat the robots in this tra
he could, so too the robots might have similarly cheated human being
they could. All we can say for sure is that if there were ever any possib
for the continuation of the human race after the robot rebellion, Hele
act unwittingly eliminates it by removing the last bargaining chip.
In Čapek's world, it turns out that mutual understanding is after
unable to moderate hatred, while Helena's quest for robot equality
Domin's quest for robot slavery combine to end very badly. It is
to believe that Čapek finds these conclusions to be to humanity's cred
The fact that Helena thinks a soul can be manufactured suggests
she has not really abandoned the materialism that Domin has announc
as the premise for robot creation. It is significant, then, that the

72 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

possibility for a good outcome in the play requires explicitly aban


that perspective.
We see the fifth and final concept of friendly robots at the very
the play, in Alquist's recognition of the love between the robots P
and Helena, a robotic version of the real Helena, which Gall c
doubtless out of his unrequited love for the real woman. At this p
the story, Alquist is the last surviving human being. The robots t
with saving them, as they do not know the secret of robot manufa
and assume that, as a human being who worked at the factory, he
Alquist tries but fails to help them in this effort; but as the play d
a conclusion, his attention focuses more and more on robot Helena
Rather tactlessly, Gall had said of the robot Helena to the or
"Even the hand of God has never produced a creature as beautiful
is! I wanted her to resemble you." But the beautiful Helena is, in h
a great failure: "she's good for nothing. She wanders about in a
vague, lifeless - My God, how can she be so beautiful with no capa
love? . . . Oh, Helena, Robot Helena, your body will never bring for
You'll never be a lover, never a mother." This last, similarly tactles
hits human Helena very hard. Gall expected that, if robot Hele
"came to," she would kill her creator out of "horror," and "throw
at the machines that give birth to Robots and destroy womanhood
course, human Helena, whose womanhood has been equally dest
already has much of this horror at humanity, and it is her action
end up unwittingly ensuring the death of Gall, along with mos
colleagues.)
When robot Helena does "come to," however, it is not out of horror,
but out of love for the robot Primus - a love that Alquist tests by threat-
ening to dissect one or the other of them for his research into recreating
the formula for robot manufacture. The two pass with flying colors, each
begging to be dissected so that the other might live. The fact that robot
Helena and Primus can love each other could be seen as some vindication
of Domin's early claim that nature still plays a role in robot development,
and that things go on in the robots which he, at least, does not claim to
understand. Even a simplified whole, it would seem, may be greater than
the sum of its parts. But Alquist's concluding encomium to the power of
nature, life, and love, all of which will survive as any mere inanimate or
intellectual human creation passes away, goes well beyond what Domin
would say. Alquist's claim that robot Helena and Primus are the new
Adam and Eve is the culmination of a moral development in him we have
watched throughout the play.

Summer 201 1 ~ 73

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

Čapek's conception of Alquisťs developing faith is usefully unde


by contrast with Nana, Helena Glory's nurse. She is a simple an
ment Christian, who hates the "heathen" robots more than wild bea
her, the events of the play confirm her apocalyptic beliefs that ma
being punished for having taken on God-like prerogatives "out of
pride." There is even a bit of mania about her: "All inventions are
the will of God," she says, as they represent the belief that human
improve on God's world. Yet when Domin seeks to dismiss her vie
of hand, Helena upbraids him: "Nana is the voice of the people.
spoken through her for thousands of years and through you only fo
This is something you don't understand."
Alquisťs position is more complicated, and seems to develop
time. When, in the Prologue, Helena is meeting the other men wh
the factory and each is in his own way defending what the compa
doing, Alquist is almost completely silent. His one speech is an obj
to Domin's aspiration to a world without work: "there was som
good in the act of serving, something great in humility. . . . some
virtue in work and fatigue." Ten years later, in a private conversat
Helena, he allows that for years he has taken to spending all his t
building a brick wall, because that is what he does when he feels u
and "for years I haven't stopped feeling uneasy." Progress mak
dizzy, and he believes it is "better to lay a single brick than to dr
plans that are too great."
Yet if Alquist has belief, it is not well-schooled. He notes tha
has a prayer book, but must have Helena confirm for him that it c
prayers against various bad things coming to pass, and wonders if
should not be a prayer against progress. He admits to already
such a prayer himself - that God enlighten Domin, destroy his
and return humanity to "their former worries and labor. . . . Rid u
Robots, and protect Mrs. Helena, amen." He admits to Helena th
not sure he believes in God, but prayer is "better than thinking."
final cataclysm builds, Alquist once again has little to say, other t
suggest that they all ought to take responsibility for the hastenin
sacre of humanity, and to say to Domin that the quest for profit h
at the root of their terrible enterprise, a charge that an "enraged
rejects completely (though only with respect to his personal moti
But by the end of the play, Alquist is reading Genesis and invo
God to suggest a sense of hope and renewal. The love of robot
and Primus makes Alquist confident that the future is in greater
than his, and so he is ready to die, having seen God's "deliverance t

74 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

love" that "life shall not perish." Perhaps, Alquist seems to imply, in
face of robot love, God will call forth the means of maintaining life -
from a biblical point of view, it would indeed be no unusual thing for
hitherto barren to become parents. Even short of such a rebirth, Al
finds comfort in his belief that he has seen the hand of God in the love
between robot Helena and Primus:

"So God created man in his own image, in the image of God created
he him; male and female created he them. And God blessed them,
and God said unto them, Be fruitful, and multiply, and replenish the
earth.... And God saw every thing that he had made, and, behold, it
was very good." . . . Rossum, Fabry, Gall, great inventors, what did you
ever invent that was great when compared to that girl, to that boy, to
this first couple who have discovered love, tears, beloved laughter, the
love of husband and wife?

Someone without that faith will have a hard time seeing such a bright
future arising from the world that R.U.R. depicts; accordingly, it is not
clear that we should assume Alquist simply speaks for Čapek. What seems
closer to the truth for eyes of weaker faith is that humans, and the robots
created in their image, will have alike destroyed themselves by undercut-
ting the conditions necessary for their own existences. Nature and life
will remain, as per Alquisťs encomium, but in a short time love will be
extinguished.

Moral Machines and Human Responsibility


Today's thinkers about moral machines could dismiss R.U.R. as an exces-
sively "Hollywood" presentation of just the sort of outcome they are
seeking to avoid. But though Čapek does not examine design features that
would produce "friendly" behavior in the exact same way they do, he has
at the least taken that issue into consideration, and arguably with much
greater understanding and depth. Indeed, as we have seen, it is in part the
diversity of understandings of Friendly AI that contributes to the play's
less than desirable results. Furthermore, such a dismissive response to
the play would not do justice to the most important issue Čapek tackles,
which is one that the present-day AI authors all but ignore: the moral
consequences for human beings of genuinely moral machines.
For Capek, the initial impulse to create robots comes from old
Rossum's Baconian sense that, with respect even to human things, there is
every reason to think that we can improve upon the given - and thereby

Summer 201 1 ~ 75

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

prove ourselves the true masters of nature, unseating old superstitio


about Divine creation. You could say that from this "frightful material-
ist" point of view, as Domin described it, we are being called to accep
responsibility for - well, everything. But what old Rossum and his so
find is that it is much harder to reproduce - let alone improve upon - th
given than they thought. Their failure at this complete mastery opens th
door to such success as young Rossum can claim: the creation of som
thing useful to human beings. On this basis Domin can establish his grand
vision of reshaping the human condition. But that grand vision contains
contradiction, as is characteristic of Utopian visions: Domin wants to fr
us from the ties of work and of dependence, or at least from dependen
on each other - in short, he wants to be responsible for changing th
human condition in such a way as to allow people to be irresponsible.
Today's authors on machine morality, focused as they are on the glo-
ries of an AI-powered, post-human future, are unwittingly faced with th
same problem, as we will see. But it must be noted first how they al
operate on the same materialist premises that informed the Rossum
efforts. It was this materialism that made it possible for the play's robo
creators to think they could manufacture something that was very muc
like a human being, and yet much simplified. They were reductioni
about the characteristics necessary to produce useful workers. Yet th
goal of humanlike-yet-not-human beings proved to be more elusive than
they expected: You can throw human characteristics out with a pitchfork
Čapek seems to say, but human creations will reflect the imperfections o
their creators. Robotic Palsy turns into full-fledged revolt. People m
have been the first to turn robots against people; the modified robots wh
led the masses may have been less simple than the standard model. But
the end, it seems that even the simplified versions can achieve a terribl
kind of humanity, a kind born - just as today's AI advocates claim we ar
about to do as we usher in a post-human future - through struggling up
out of "horror and suffering."
Wallach and Allen more than Yudkowsky are willing to model the
moral machines on human moral development; Yudkowsky prides himse
on a model for moral reasoning shorn of human-like motivations. Eithe
way, are there not reasons to expect that their moral machines woul
be subject to the same basic tendencies that afflict Čapek's robots? T
human moral development Wallach and Allen's machines will mode
involves learning a host of things that one should not do - so they wou
need to be autonomous, and yet not have the ability to make these wron
choices. Something in that formulation is going to have to give; conside

76 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

ing the split-second decisions that Wallach and Allen imagine their m
machines will have to make, why should we assume it will be autonom
Yudkowsky's Friendly AI may avoid that problem with its alien style
moral reasoning - but it will still have to be active in the human wor
and its human subjects, however wrongly, will still have to interpret
choices in human terms that, as we have seen, might make its advanc
benevolence seem more like hostility.
In both cases, it appears that it will be difficult for human beings
have anything more than mere faith that these moral machines reall
have our best interests at heart (or in code, as it were). The conclusion
we must simply accept such a faith is more than passingly ironic, given
these "frightful materialists" have traditionally been so totally oppos
to putting their trust in the benevolence of God, in the face of what
take to be the obvious moral imperfection of the world. The point ap
equally, if not more so, to today's Friendly AI researchers.
But if moral machines will not heal the world, can we not at l
expect them to make life easier for human beings? Domin's effort to m
robot slaves to enhance radically the human condition is reflected in
desire of today's authors to turn over to AI all kinds of work that we
we would rather not or cannot do; and his confidence is reflected
more so, considering the immensely greater amount of power propos
for Als. If it is indeed important that we accept responsibility for cre
machines that we can be confident will act responsibly, that can only
because we increasingly expect to abdicate our responsibility to them. A
the bar for what counts as work we would rather not do is more read
lowered than raised. In reality, or in our imaginations, we see, like A
Smith's little boy operating a valve in a fire engine, one kind of work
we do not have to do any more, and that only makes it easier to imag
others as well, until it becomes harder and harder to see what machin
could not do better than we, and what we in turn are for.
Like Domin, our contemporary authors do not seem very interested
asking the question of whether the cultivation of human irresponsibili
which they see, in effect, as liberation - is a good thing, or whether
Alquist would have it) there is some vital connection between work
human decency. Čapek would likely connect this failure in Domin to
underlying misanthropy; Yudkowsky's transhumanism begins from a
tinctly similar outlook. But it also means that whatever their apparen
philanthropic intentions, Wallace, Allen, Yudkowsky, and their peers
be laying the groundwork for the same kind of dehumanizing results t
Capek made plain for us almost a century ago.

Summer 201 1 ~ 77

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Charles T. Rubin

By design, the moral machine is a safe slave, doing what we want


have done and would rather not do for ourselves. Mastery over slaves
notoriously bad for the moral character of the masters, but all the wo
one might think, when their mastery becomes increasingly nominal.
better moral machines work, the more we will depend on them, and
more we depend on them, the more we will in fact be subject to them
course, we are hugely dependent on machines already, and only a frin
few would go so far as to say that we have become enslaved to them.
my car is not yet making travel decisions for me, and the power statio
not yet deciding how much power I should use and for what purposes.
autonomy supposed to be at the root of moral machines fundamen
changes the character of our dependence.
The robot rebellion in the play just makes obvious what would h
been true about the hierarchy between men and robots even if the de
for robots had worked out exactly as their creators had hoped. The p
sibility that we are developing our "new robot overlords" is a joke with
edge to it precisely to the extent that there is unease about the quest
of what will be left for humans to do as we make it possible for ourse
to do less and less. The end of natality, if not an absolutely neces
consequence of an effort to avoid all work and responsibility, is at le
understandable as an extreme consequence of that effort. That extrem
consequence is not entirely unfamiliar in a world where technological
advanced societies are experiencing precipitously declining birthr
and where the cutting edge of transhumanist techno-optimism prom
an individual Protean quasi-immortality at the same time as it anticip
what is effectively the same human extinction that is achieved in R.U
except packaged in a way that seems nice, so that we are induced to ch
rather than fight it.
The quest to take responsibility for the creation of machines that w
allow human beings to be increasingly irresponsible certainly does
have to end this badly, and may not even be most likely to end this ba
Were practical wisdom to prevail, or if there is inherent in the order
things some natural right, or if, as per Alquist and Nana, we live
Providential order, or if the very constraints of our humanity will a
a shield against the most thoroughly inhumane outcomes, then hu
beings might save themselves or be saved from the worst consequenc
our own folly. By partisans of humanity, that is a consummation devo
to be wished. But it is surely not to be counted upon.
After all, R. U.R. is precisely a story about how the human soul, to
row Peter Lawler's words, "shines forth in and transforms all our tho

78 ~ The New Atlantis

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms
Machine Morality and Human Responsibility

and action, including our wonderful but finally futile efforts to


ourselves from nature and God." Yet the souls so exhibited are mo
multifaceted and conflicted; they transform our actions with uninten
consequences. And so the ultimate futility of our efforts to free ourse
from nature and God exacts a terrible cost - even if, as Alquist believ
Providence assures that some of what is best in us survives our demise.

Summer 201 1 -79

This content downloaded from 86.142.90.237 on Tue, 26 May 2020 18:06:47 UTC
All use subject to https://about.jstor.org/terms

You might also like