Building Moral Robots 2019
Building Moral Robots 2019
https://doi.org/10.1007/s11948-019-00084-5
ORIGINAL RESEARCH/SCHOLARSHIP
John‑Stewart Gordon1,2
Abstract
This paper examines the ethical pitfalls and challenges that non-ethicists, such as
researchers and programmers in the fields of computer science, artificial intelligence
and robotics, face when building moral machines. Whether ethics is “computable”
depends on how programmers understand ethics in the first place and on the ade-
quacy of their understanding of the ethical problems and methodological challenges
in these fields. Researchers and programmers face at least two types of problems
due to their general lack of ethical knowledge or expertise. The first type is so-called
rookie mistakes, which could be addressed by providing these people with the nec-
essary ethical knowledge. The second, more difficult methodological issue concerns
areas of peer disagreement in ethics, where no easy solutions are currently avail-
able. This paper examines several existing approaches to highlight the ethical pit-
falls and challenges involved. Familiarity with these and similar problems can help
programmers to avoid pitfalls and build better moral machines. The paper concludes
that ethical decisions regarding moral robots should be based on avoiding what is
immoral (i.e. prohibiting certain immoral actions) in combination with a pluralistic
ethical method of solving moral problems, rather than relying on a particular ethical
approach, so as to avoid a normative bias.
* John‑Stewart Gordon
[email protected]; [email protected]
http://www.johnstgordon.com
1
Department of Philosophy and Social Critique, Faculty of Political Science and Diplomacy,
Vytautas Magnus University, V. Putvinskio g. 23 (R 403), 44243 Kaunas, Lithuania
2
Research Cluster for Applied Ethics, Faculty of Law, Vytautas Magnus University, V. Putvinskio
g. 23 (R 403), 44243 Kaunas, Lithuania
13
Vol.:(0123456789)
J.-S. Gordon
Introduction
In recent years, the interdisciplinary field of machine ethics—that is, how to pro-
gram robots with ethical rules so that they become either implicit or explicit moral
agents (Moor 2006)—has become of utmost importance because of current and
anticipated technological developments in the fields of computer science, artificial
intelligence (AI) and robotics (Lin et al. 2014; Wallach and Allen 2010; Anderson
and Anderson 2011a; Gunkel and Bryson 2014). One of the most important issues
in machine ethics concerns the goal of making artificially intelligent robots “ethi-
cal”. Susan Anderson, a pioneer in the field, defines this goal as “to create a machine
that follows an ideal ethical principle or set of principles in guiding its behaviour;
in other words, it is guided by this principle, or these principles, in the decisions it
makes about possible courses of action it could take. We can say, more simply, that
this involves ‘adding an ethical dimension’ to the machine” (2011c: 22).
Based on the great technological advances in AI, the advent of fully autonomous,
human-like, intelligent robots capable of ethical reasoning and decision making
seems unavoidable. There exist various ethical programs with numerous approaches
that could also—sooner or later—be used for robots. This paper is not concerned
with the moral and legal status of such intelligent artificial beings. Rather, it focuses
on the ethical pitfalls and challenges that researchers and programmers in the fields
of computer science, AI and robotics face when they attempt to build such moral
machines. Whether ethics is “computable”1 depends in part on how those research-
ers understand ethics and attempt to implement that understanding in programs,
but also more generally on whether they have some expertise in ethics and moral
philosophy in the first place. Based on a review of current programs in the field, it
appears that at this time, researchers and programmers have neither a good enough
understanding nor sufficient ethical expertise to build moral machines that would be
comparable to human beings with respect to ethical reasoning and decision making.
In general, one can observe two main types of problems that non-ethicists commit
when they attempt to add an “ethical dimension” to their machines. The first type
could be called rookie mistakes; they involve a misunderstanding of moral issues
or starting from wrong ethical assumptions (such as overestimating the capacity of
a given ethical approach to solve complex moral cases). These problems show that
the researchers and programmers committing the errors are somewhat unaware of
the moral complexity of the issues with which they are grappling or the knowledge
already achieved in the field of ethics. The second category contains methodological
issues that pose a challenge even to ethical experts because they disagree on how to
solve those issues, as will be discussed below.
Building moral machines is quite conceivable and could be a valuable advance
for humanity. However, achieving this goal will require the builders to have suitable
ethical expertise. This paper highlights some of the most significant ethical pitfalls
1
In this paper, the concept of computability is used in a rather broad sense. It concerns the way in which
a machine is built to make ethical decisions that generally seek to resemble the moral reasoning and deci-
sion making of human beings.
13
Building Moral Robots: Ethical Pitfalls and Challenges
and challenges that stand in the way of building moral machines. It argues that ethi-
cal decisions of moral robots should rather be based on (1) avoiding the immoral, or
prohibiting certain immoral actions, and (2) a pluralistic ethical method for solving
moral problems, as opposed to reliance on a particular ethical approach.
The next section of this paper provides some brief preliminary remarks on
the meaning of moral agency in the context of AI. The third part examines four
recent ethical programs that are supposed to mirror human-like ethical reasoning
and decision making: (1) the casuistic approach (Guarini 2006), (2) the top-down
approach of the MoralDM (Dehghani et al. 2011), (3) the hybrid approach (Wallach
et al. 2010), and (4) the web-based approach (Rzepka and Araki 2005). Based on
this review, the fourth part explains why researchers and programmers have failed
to achieve ethical reasoning and decision making in their programs. Four issues
are presented—lack of ethical expertise, pluralism in ethics, dilemma cases, and
machine bias—that indicate that the arrival of moral machines is a long way off. The
paper then deals with two vital methodological issues—the problem of justification
and avoiding the immoral—that should be taken into account in the construction of
moral machines. The final part offers some conclusions.
Preliminary Remarks
According to Bekey (2014: 18), the notion of autonomy can be defined as “the
capacity to operate in the real-world environment without any form of external
control, once the machine is activated and at least in some areas of operation, for
extended periods of time”. That means that if we maximally increase the capacity
to operate in the real-world environment, we could create intelligent robots (IRs)
that are able to perform various tasks of utmost complexity, most probably even
unmatched by human beings. The way to increase this capacity is to perfect the use
of AI in robots, i.e. “making machines do things that would require intelligence if
done by men” (Marvin Minsky, cited by Copeland 1993: 1).
Technically speaking, IRs need proper algorithms to solve any problem or issue
based on the input perceived. It has been suggested, however, that IRs cannot
become moral agents because the algorithms are finite or deterministic by nature
and hence the robots lack free will, which is commonly understood as a prerequisite
for moral agency. If IRs must act according to a fixed set of rules, then they could
not be seen as moral agents (Johnson and Axinn 2014). This objection has already
been addressed at length (Gordon 2018), but two important points should be briefly
reviewed here. First, it is possible that IRs will be able to learn from their past expe-
riences in such a way as to re-program themselves and hence alter their algorithms
by changing their pre-fixed set of rules to adapt to new situations. Second, many
influential philosophers and neurobiologists working on the interdisciplinary and
complex issue of free will have forcefully questioned (1) whether human beings
have free will and (2) whether free will is a necessary precondition of moral agency.
To show that moral machines will never be able to make human-like autonomous
13
J.-S. Gordon
decisions, opponents must counter these two objections. Such a decisive refutation
has not been provided yet, and it is very unlikely that it will be provided any time
soon. Therefore, the free will objection—i.e. to deny IRs moral agency because
of the nature of their algorithms—is unconvincing, because it suggests that robots
should meet a supposedly necessary precondition for moral agency (free will) the
fulfilment of which even by human beings remains in dispute.
In his influential paper “The Nature, Importance, and Difficulty of Machine Eth-
ics” (2006), James H. Moor made an important distinction between various forms of
ethical agents—ethical-impact agents, implicit ethical agents, explicit ethical agents
and full ethical agents—which has been frequently quoted by other authors. Accord-
ing to Moor, an ethical-impact agent is a machine that helps to avoid an otherwise
immoral situation; as an example, he cites the development of robotic camel jockeys
to replace small boys who are kept as slaves to ride the camels in races. “Computing
technology often has important ethical impact. The young boys replaced by robotic
camel jockeys are freed from slavery” (Moor 2006: 19). An implicit ethical agent is
a machine programmed to act ethically or to avoid unethical behaviour by following
the programming of a designer who in turn follows ethical principles (for example,
ATMs, web banking software or autopilot). An explicit ethical agent, in contrast,
actually uses ethical principles to solve ethical issues. Finally, “A full ethical agent
can make explicit ethical judgments and generally is competent to reasonably justify
them” (Moor 2006: 20). Explicit ethical agents and full ethical agents both use ethi-
cal principles to make ethical decisions, but only the latter can provide a justification
for those decisions.
Most authors refer to only two of these four categories (implicit and explicit) and
thereby leave the impression that explicit ethical agents are the prime example of
moral agents. This is unfortunate, since it confines the discussion to “limited explicit
ethical agents” instead of full ethical agents (Moor 2006: 21).2 Moor himself states:
A full ethical agent can make explicit ethical judgements and generally is
competent to reasonably justify them. An average adult human is a full ethi-
cal agent. We typically regard humans as having consciousness, intentionality,
and free will. Can a machine be a full ethical agent? It’s here that the debate
about machine ethics becomes most heated. Many believe a bright line exists
between the senses of machine ethics discussed so far and a full ethical agent.
For them, a machine can’t cross this line. The bright line marks a crucial onto-
logical difference between humans and whatever machines might be in the
future. (Moor 2006: 20)
2
For example, Anderson and Anderson (2007: 15) observe, “Using Moor’s terminology, most of those
working on machine ethics would say that the ultimate goal is to create a machine that is an explicit
ethical agent.” This might be true, but we should be prepared for the advent of IRs who are full ethical
agents, regardless of how long it may take for such IRs to become a reality.
13
Building Moral Robots: Ethical Pitfalls and Challenges
Moor believes that it is not possible to answer the question of whether machines
will become full ethical agents by adhering to “philosophical argument or empirical
research” and therefore suggests focusing on “developing limited explicit agents”
even though they will “fall short of being full ethical agents”. They could, according
to Moor, still “help prevent unethical outcomes”, but one should not be “too opti-
mistic about our ability to develop” even such machines (Moor 2006: 21).
This perspective faces two problems. First, it is unclear whether Moor believes
that it is, in general, impossible to answer by philosophical or empirical means the
question of whether machines could become full ethical agents, or only that humans
are currently unable to do so. A more optimistic approach will assume only the latter
claim. Second, his cautionary statement that it might be impossible to develop even
machines that are explicit ethical agents is rather unconvincing. His reasons are as
follows: (1) even human beings have a limited understanding of the complex topic
of ethics, (2) humans’ lack of understanding of the phenomenon of learning, particu-
larly concerning the child machine envisioned by Turing,3 and (3) the computer’s
lack of common sense and world knowledge (Moor 2006: 21). The main reason why
one should reject these objections is that the trajectory of technological development
seems quite promising; moreover, many would argue that moral philosophy itself
has, despite the continued and undisputed existence of unresolved complex moral
issues, nonetheless made significant progress during the last centuries as well.
However, one should not overlook two vital points concerning the contested ques-
tion of moral progress. First, it seems easier to formulate moral rules that classify
certain actions as immoral than to prescribe correct moral behaviour for all people
(see Gordon 2016b). This is because there exist a plurality of lifestyles from which
different normative actions may emerge that can be seen as equally good or, at least,
not morally reprehensible, whereas we can identify particular actions which are gen-
erally acknowledged as immoral, such as rape, torture, theft and child molestation.
Second, there have been important, undeniable moral developments in the history
of humanity that can be counted as moral progress, such as the abolition of slavery
(even though many people are still kept as slaves around the world), the recogni-
tion of universal human rights (and their transition into international legislation), the
idea of a fair trial and that the punishment should always match the crime, gender
equality and the recognition of animal rights. The notion of moral progress does not
claim that all these points are fully accepted and acknowledged by all people, but
that humanity overall has made great moral progress on these issues during the last
few centuries.
Hence, there is no reason not to consider seriously the possibility of machines
that will become full ethical agents. Should IRs become more complex, more pow-
erful, highly autonomous and endowed with an unmatched artificial intelligence,
moral standards that match the great freedom of such machines will be essential to
avoid causing harm (see Picard 1997).
3
See Allen et al. (2000: 251–261) for the interesting idea of the so-called Moral Turing Test (MTT) that
robots would pass if they (1) give correct answers to moral dilemmas and (2) offer the appropriate justifi-
cation for their claims. This would amount to understanding ethics.
13
J.-S. Gordon
This section offers a brief overview of some recent attempts to show how intelligent
machines should actually make moral decisions.4 The different approaches can be
categorized as follows:
Whereas Guarini’s system faces problems with regard to the reclassification of cases
due to the lack of proper reflection and exact representation (Guarini admits that
casuistry alone is insufficient for machine ethics), Rzepka and Araki’s web-based
GENTA approach offers a kind of majority-rule ethics based on the average behav-
iour of human beings (including immoral actions). From a moral point of view, the
idea of relying on a majority-based decision procedure in ethics is not only mislead-
ing, but can never be considered an appropriate means of ethical reasoning and deci-
sion making. The MoralDM approach of Dehghani et al. is of utmost significance in
that it pays proper respect to the two main ethical theories (deontology and utilitari-
anism) and combines them in a way that is fruitful and quite promising. However,
their additional strategy of using empirical studies to mirror human moral decisions
by considering as correct only those decisions that align with the majority of human
subjects is misleading and seriously flawed. In this regard, Susan Anderson (2011a:
525) rightly claims that their system can be regarded only as a model for descrip-
tive and not for normative ethics. The same objection could be raised against the
hybrid model of moral reasoning developed by Wallach, Franklin and Allen, who
attempt to re-create human decision making by appealing to a complex combination
of top-down and bottom-up approaches. In addition, their somewhat idiosyncratic
4
It is not claimed that the approaches analysed in the table are the only important ones, but they are a
representative sample of the existing range of accounts that have been extensively discussed in recent
years.
13
Table 1 Machine ethics: four approaches
Bottom-up approach Top-down approach Mixed approach Other
Guarini (2006): The casuistic approach Dehghani et al. (2011): The MoralDM Wallach et al. (2010), Wallach and Allen Rzepka and Araki (2005): The web-based
(2010): The hybrid approach approach
The system (using a neural network) Their system combines two main ethical Their hybrid model of human cognition Their system, GENTA (General Belief
bases its ethical decisions on a learn- theories, utilitarianism and deontol- combines a so-called top-down (the- Retrieving Agent), makes ethical deci-
ing process in which it is presented ogy, in combination with analogi- ory-driven reasoning) and a bottom-up sions based on statistical analysis using
Building Moral Robots: Ethical Pitfalls and Challenges
with known correct answers to ethical cal reasoning. Utilitarian reasoning approach (shaped by evolution and the full spectrum of web-based knowl-
dilemmas. After that, the system is applies until “sacred values” are learning) that is both considered the edge with respect to human behaviour,
able to solve new ethical dilemmas on involved, at which point the system foundation for moral reasoning and opinions, common consequences,
its own operates in a deontological mode and decision making. The result thus far is exceptions, etc
becomes less sensitive to the utility LIDA, an artificial general intelligence
of actions and consequences. To software offering a comprehensive,
align the system with human moral conceptual and computational model
decisions, they evaluate the system that covers a large portion of human
against psychological studies of how cognition
the majority of human beings decided
particular cases
13
J.-S. Gordon
understanding of both approaches in moral philosophy does not really match how
moral philosophers understand and use them in normative ethics.5
The ethical pitfalls and challenges that undermine all four approaches—reliance
on casuistry alone (Guarini), the use of majority-based decision procedures in ethics
(MoralDM and GENTA), and the idea of combining two types of moral reasoning
and decision making that are commonly viewed as mutually exclusive (LIDA)—
could have been avoided through a more comprehensive understanding of ethics and
moral philosophy. Although debate continues over the ethical merit of casuistry (it
seems to be a matter of degree), relying on majority opinion as a basis for discerning
ethical truth goes beyond any reasonable and acceptable standards in ethics. This
objection is also substantiated by common sense. The only solution to such prob-
lems is to work closely with moral philosophers so as to avoid steps that the philoso-
phers will readily identify as rookie mistakes. There are other more complex ethical
issues that even moral philosophers cannot solve completely, but even in those cases
they can clearly determine the demarcation lines and describe the problem suffi-
ciently to understand the implications of a given issue.
The next section examines some of the main ethical pitfalls and challenges that
occur in efforts to build moral machines. Researchers and programmers should
become more aware of the deep methodological problems related to ethical reason-
ing and decision making that—without proper understanding—may hinder them in
their endeavours to construct intelligent moral machines.
The above-mentioned approaches that attempt to empower IRs to make sound moral
decisions are unconvincing for several reasons, which will be examined in this sec-
tion. This critique is not intended to imply that IRs will never be able to make moral
decisions and thereby become full ethical agents; rather, the analysis suggests that
ethical reasoning and decision making are, by virtue of their complex nature, more
difficult than AI researchers and programmers may assume at first glance (Anderson
2011b).6 In the long term, however, there is no principled reason why moral IRs
should not eventually become a reality (Anderson 2011b: 164).
5
The model presented by Wallach et al. is not necessarily inadequate with respect to how moral deci-
sion making works in an empirical sense, but their approach is descriptive rather than normative by
nature. Therefore, their empirical model does not solve the normative problem of how moral machines
should actually act. Descriptive ethics and normative ethics are two different things. The first tells us how
human beings make moral decisions; the latter is concerned with how we should act.
6
For a brief but good overview of the philosophical concerns, see Anderson (2011b: 162–167) and Gips
(2011: 244–253).
13
Building Moral Robots: Ethical Pitfalls and Challenges
Because of the lack of sufficient ethical expertise among non-ethicists who work in
the field of machine ethics, there is a great danger that AI researchers will start from
wrong ethical assumptions and build their ethical approach to machines on shaky
grounds. For example, Anderson and Anderson (2007), who are among the pioneers
in the field of machine ethics, developed three different applications: (1) a medi-
cal advisor system for dilemmas (MedEthEx), (2) a medication reminder system for
elderly people (EthEl), and (3) the instantiation of EthEl in a Nao robot, which they
believe is the “first example of a robot that follows an ethical principle in determin-
ing which actions it will take” (2011b: 480). However, they base their famous medi-
cal advisor system (MedEthEx) on some contested claims that undermine their oth-
erwise interesting account. They state, “Because there is more agreement between
ethicists working on biomedical ethics than in other areas, and because there are
fewer duties, we decided to begin to develop our prima facie duty approach to com-
puting ethics using Beauchamp and Childress’s principles of biomedical ethics”
(Anderson and Anderson 2007: 22).
Experts in bioethics acknowledge that their field is one of the most complex
and challenging areas in all of applied ethics. There is no common ground or unity
among bioethicists. Furthermore, the suggested bioethical approach of principlism,
even though it has been widely used for many decades by practitioners and bioethi-
cists alike, is complex and not easy to apply, in particular with respect to the meth-
ods of specification and balancing (Gordon 2011a, b). As a result, Anderson and
Anderson start from misleading assumptions and use a theory that is quite demand-
ing and, according to Clouser and Gert (1990), highly controversial. They thus jeop-
ardise their own project because they misjudge the ethical complexity present in the
field of bioethics and overestimate the capacity of principlism to function as a role
model for machine ethics.
In another work, Susan Anderson (2011c: 21–27) takes a somewhat different
approach. Rightly pointing out that ethics is too complex for everyone to adhere to
a simple absolute ethical theory, she suggests instead starting from particular cases
and eventually using “a large number of cases in a fashion that permits automated
analysis. From this information, general ethical principles may emerge” (25). How-
ever, her own approach of multiple prima facie duties, based on Ross (1930), is
eventually faced with the general problem of making reasonable decisions between
conflicting duties. Furthermore, she interestingly suggests starting by working
on machines that will serve as “ethical advisors” before creating “moral robots”,
because the latter category requires that “a difficult judgment be made about the sta-
tus of the machine itself” (26). Granted, one can thereby avoid the difficult questions
of the moral status and moral rights of intelligent machines, but the earlier one starts
to think about these issues, the better equipped one will be when the time comes.
Furthermore, the idea of using machines as moral advisors is also supported by, for
example, Giubilini and Savulescu (2018), McLaren (2011: 310–312), and Torrance
(2008). For example, Torrance argues that, if we build such machines, we will also
learn something about human morality, and that such machines will, in general, be
useful and instructive to human beings.
13
J.-S. Gordon
Pluralism in Ethics
Numerous ethical theories have been used to solve moral issues, among which the
most prominent are virtue ethics, deontology, utilitarianism, casuistry and contract
theory. There is currently no common ground among moral experts as to which ethi-
cal theory to use. Ethicists, like laypeople, choose among these varying approaches
when dealing with moral problems. This pluralism is equally visible in approaches
to machine ethics. The underlying idea that IRs will be better at making moral deci-
sions than human beings (Hall 2011: 522–5237; Dietrich 2011)8 faces at least two
problems. First, before one can claim that machines make morally better decisions,
one must have a proper moral approach that is, in fact, better than the other avail-
able theories.9 Second, the idea that many ethical issues could have equally good
but different decisions must be taken into account as well (for example, the same
moral problem could be adequately solved differently in different cultural settings).
Furthermore, when it comes to pluralism in ethics, it seems reasonable to avoid
extreme moral relativism and to accept a firm core of universal moral norms that all
human beings acknowledge (e.g. one must not commit murder or rape, insult other
people, or violate the human rights of others). Beyond this core of moral norms,
people should be free to act according to their particular moralities in their given
community, as long as those particular norms do not conflict with the core of uni-
versal norms or with the lingua franca of international human rights (Gordon 2015;
Macklin 1999). This view is further substantiated by Ruth Macklin’s pithy depiction
of the problem of moral relativism:
A long-standing debate surrounds the question whether ethics are relative to
time and place. One side argues that there is no obvious source of a universal
morality and that ethical rightness and wrongness are products of their cultural
and historical setting. Opponents claim that even if a universal set of ethical
norms has not yet been articulated or agreed upon, ethical relativism is a per-
nicious doctrine that must be rejected. The first group replies that the search
for universal ethical precepts is a quest for the Holy Grail. The second group
responds with the telling charge: If ethics were relative to time, place, and cul-
ture, then what the Nazis did was “right” for them, and there is no basis for
moral criticism by anyone outside the Nazi society. (Macklin 1999: 4)
7
Hall (2011: 523) claims, “In the long run, AIs will run everything, and should, because they can do a
better job than humans. Not only will their intellectual prowess exceed ours, but their moral judgment as
well.”
8
Dietrich argues that human beings, as biological entities, might have a genetic predisposition for selfish
behaviour as a survival mechanism with respect to competition with others. Machines lack this predis-
position and could be made in such a way that they both possess the good features of human beings and
follow ethical principles. They could then be seen as “Homo sapiens 2.0”—enhanced or improved human
beings.
9
One might also claim here that the ability of moral machines to perform “better” than humans in the
realm of ethics relates to the fact that they do not suffer from moral imperfections—e.g. prejudice, self-
ishness, partiality, weakness of will, and being carried away by emotions—in their ethical reasoning and
decision making.
13
Building Moral Robots: Ethical Pitfalls and Challenges
Dilemma Cases
10
See Pereira and Saptawijaya (2011) for a detailed description of how their system, ACORDA, uses
prospective logic programming to solve various dilemmas like the trolley case.
11
See, for example, the interesting programmable “moral machine” developed by Scalable Cooperation
at MIT Media Lab concerning the question “Between whom is the self-driving car deciding?” in a case
where the car’s brakes fail. It is a “platform for gathering a human perspective on moral decisions made
by machine intelligence, such as self-driving cars … where a driverless car must choose the lesser of
two evils, such as killing two passengers or five pedestrians. As an outside observer, you judge which
outcome you think is more acceptable” (www.moralmachine.mit.edu, accessed 2 July 2018). However,
the MIT trial does not involve cars working in concert, but only in isolation or in a way that involves
human–robot interaction. Self-driving cars may eventually be part of a larger system in which machines
will communicate with each other and thereby solve upcoming problems more fully than can occur when
each machine functions in isolation, particularly in the context of automated driving (Borenstein et al.
2017).
13
J.-S. Gordon
Machine Bias
Empirically speaking, one can regularly make an interesting but important observa-
tion, while sitting on ethics committees, with regard to the difference between the
commonly agreed-upon solution, on one hand, and the varying justifications given
by members on the other hand. Usually, the decisions made by ethics committees are
13
Building Moral Robots: Ethical Pitfalls and Challenges
quite uniform, yet the committee members often provide very different justifications
for their decisions (Bayertz 1991). This result suggests that different ethical theo-
ries or perspectives overlap in their view of many moral issues, even though they
apply completely different lines of argumentations. Of course, it could be objected
that ethics committees generally are seeking the best pragmatic decision, whereas
“normal” moral reasoning and decision making are concerned with the right moral
decision, independent of any external factors that might favour a more pragmatic
solution. The goal of ethics committees is hence twofold—namely, to find the best
available and most pragmatic solution in their decision making.
The key point here is that AI researchers must be aware, when programming
moral machines, of the nuance that different ethical theories often provide the same
decisions but for different reasons. To avoid bias in ethics, it would seem worthwhile
to create moral machines that examine moral problems from different ethical per-
spectives, using multiple ethical theories, and then—after having determined and
examined all possible issues—make a well-balanced moral decision. This type of
bias is different from the machine bias problem discussed above, where either (1)
improper data or (2) a defective algorithm guides the machine towards disadvan-
taging particular groups, such as an ethnic group or women. This theoretical bias
really concerns how ethical reasoning and decision making are done by moral phi-
losophers or laypeople. As noted above, the moral solution of some difficult cases,
such as the trolley case, may differ substantially depending on the ethical theory
used. For example, a neo-Kantian approach emphasizes the core values of rational-
ity and autonomy in its ethical reasoning and decision making, whereas a hedonistic,
utilitarian approach favours sentience, i.e. the ability to suffer and to feel pain. As a
result, the two ethical theories may arrive at different moral outcomes based on their
different core values. Therefore, some moral philosophers such as Brody (1988)
and Gordon (2016a) have suggested a more pluralistic approach based on several
key concepts in order to avoid a theoretical bias in decision making. This reasoning
suggests that, to be theoretically unbiased, machine ethics should use a method that
does justice to the complexity of the moral life.
Machine ethics should “focus on the similarities between the various ethical
approaches in order to gain a better understanding of what one should not do. In
other words, one should try to spell out in more detail not how moral agents should
act but, rather, what they should avoid doing” (Gordon 2016b: 290). This approach
has an analogy in the context of moral enhancement, where a similar question—
namely, which ethical approach should be used as a default position to make human
beings more moral—has been raised. One reasonable answer to this question is to
try to agree “on a binding list of things we want to avoid in our moral lives” (Gor-
don 2016b: 290) and to create machines that will act accordingly while, at the same
time, permitting them to make moral decisions within a framework that allows for
different but equally acceptable moral solutions. After all, that seems to be the best
that one can currently hope for in the area of machine ethics.
13
J.-S. Gordon
Conclusions
12
The idea that machines might come up with new or unexpressed ethical principles has been advanced
by Anderson and Anderson (2011b: 476–492, esp. 479).
13
Building Moral Robots: Ethical Pitfalls and Challenges
be severely undermined. How to decide what actions are morally correct is one of
the most difficult questions in our lives. Understanding the ethical pitfalls and chal-
lenges involved in these decisions is absolutely necessary to build intelligent, unbi-
ased moral machines.
Acknowledgements I would like to thank the anonymous reviewers for their valuable comments. This
research is funded by the European Social Fund according to the activity ‘Improvement of researchers’
qualification by implementing world-class R&D projects of Measure No. 09.3.3-LMT-K-712.
References
Abney, K. (2014). Robotics, ethical theory, and metaethics: A guide for the perplexed. In P. Lin, K.
Abney, & G. Bekey (Eds.), Robot ethics: The ethical and social implications of robotics (pp. 35–52).
Cambridge, MA: MIT Press.
Aletras, N., Tsarapatsanis, D., Preoţiuc-Pietro, D., & Lampos, V. (2016). Predicting judicial decisions of
the European Court of Human Rights: A natural language processing perspective. PeerJ Computer
Science, 2(e93), 1–19.
Allen, C., Varner, G., & Zinser, J. (2000). Prolegomena to any future artificial moral agent. Journal of
Experimental & Theoretical Artificial Intelligence, 12(3), 251–261.
Allen, C., Wallach, W., & Smit, I. (2011). Why machine ethics? In M. Anderson & S. Anderson (Eds.),
Machine ethics (pp. 51–61). Cambridge, MA: Cambridge University Press.
Anderson, S. L. (2011a). How machines might help us achieve breakthroughs in ethical theory and
inspire us to behave better. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 524–
5309). Cambridge, MA: Cambridge University Press.
Anderson, S. L. (2011b). Philosophical concerns with machine ethics. In M. Anderson & S. L. Anderson
(Eds.), Machine ethics (pp. 162–167). Cambridge, MA: Cambridge University Press.
Anderson, S. L. (2011c). Machine metaethics. In M. Anderson & S. L. Anderson (Eds.), Machine ethics
(pp. 21–27). Cambridge, MA: Cambridge University Press.
Anderson, M., & Anderson, S. L. (2007). Machine ethics: Creating an ethical intelligent agent. AI Maga-
zine, 28(4), 15–26.
Anderson, M., & Anderson, S. L. (2011a). Machine ethics. Cambridge, MA: Cambridge University Press.
Anderson, M., & Anderson, S. L. (2011b). A prima facie duty approach to machine ethics. In M. Ander-
son & S. L. Anderson (Eds.), Machine ethics (pp. 476–492). Cambridge, MA: Cambridge Univer-
sity Press.
Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016). Machine bias. ProPublica. https://www.propu
blica.org/article/machine-bias-risk-assessments-in-criminal-sentencing. Accessed 8 March 2018.
Bayertz, K. (1991). Praktische Philosophie als angewandte Ethik”. In K. Bayertz (Ed.), Praktische Phi-
losophie: Grundorientierungen angewandter Ethik (pp. 7–47). Hamburg: Rowohlt Taschenbuch
Verlag.
Bekey, G. A. (2014). Current trends in robotics: Technology and ethics. In P. Lin, K. Abney, & G. Bekey
(Eds.), Robot ethics: The ethical and social implications of robotics (pp. 17–34). Cambridge, MA:
MIT Press.
Borenstein, J., Herkert, J. R., & Miller, K. W. (2017). Self-driving cars and engineering ethics: The
need for a system level analysis. Science and Engineering Ethics. https://doi.org/10.1007/s1194
8-017-0006-0.
Bringsjord, S., & Taylor, J. (2014). The divine-command approach to robot ethics. In P. Lin, K. Abney, &
G. Bekey (Eds.), Robot ethics: The ethical and social implications of robotics (pp. 85–108). Cam-
bridge, MA: MIT Press.
Brody, B. (1988). Life and death decision making. Oxford: Oxford University Press.
Clouser, K. D., & Gert, B. (1990). A critique of principlism. Journal of Medicine and Philosophy, 15(2),
219–236.
Coeckelbergh, M. (2014). The moral standing of machines: Towards a relational and non-Cartesian moral
hermeneutics. Philosophy & Technology, 27(1), 61–77.
Copeland, J. (1993). Artificial intelligence: A philosophical introduction. Oxford: Wiley.
13
J.-S. Gordon
Dehghani, M., Forbus, K., Tomai, E., & Klenk, M. (2011). An integrated reasoning approach to moral
decision making. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 422–441). Cam-
bridge, MA: Cambridge University Press.
Dietrich, E. (2011). Homo sapiens 2.0: Building the better robots of our nature. In M. Anderson & S. L.
Anderson (Eds.), Machine ethics (pp. 531–538). Cambridge, MA: Cambridge University Press.
Floridi, L. (2011). On the morality of artificial agents. In M. Anderson & S. L. Anderson (Eds.), Machine
ethics (pp. 184–212). Cambridge: Cambridge University Press.
Floridi, L., & Sanders, J. W. (2004). On the morality of artificial agents. Minds and Machines, 14(3),
349–379.
Foot, P. (1967). The problem of abortion and the doctrine of double effect. Oxford Review, 5, 5–15.
Gips, J. (2011). Towards the ethical robot. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp.
244–253). Cambridge, MA: Cambridge University Press.
Giubilini, A., & Savulescu, J. (2018). The artificial moral advisor. The “Ideal Observer” meets artificial
intelligence. Philosophy & Technology, 31(2), 169–188.
Gordon, J.-S. (2011). Global ethics and principlism. Kennedy Institute of Ethics Journal, 21(3), 251–276.
Gordon, J.-S. (2015). Human rights and cultural identity. Baltic Journal of Law & Politics, 8(2), 112–135.
Gordon, J. S. (2016a). Ethics as a method. In J. Schildmann, J. S. Gordon, & J. Vollmann (Eds.), Clini-
cal ethics consultation: Theories and methods, implementation, evaluation (pp. 11–20). London:
Routledge.
Gordon, J.-S. (2016b). Should moral enhancement be compulsory? Ethical Perspectives, 23(2), 277–305.
Gordon, J.-S. (2018). What do we owe to intelligent robots? AI & Society. https://doi.org/10.1007/s0014
6-018-0844-6.
Gordon, J.-S., Rauprich, O., & Vollmann, J. (2011). Applying the four-principle approach. Bioethics,
25(6), 293–300.
Guarini, M. (2006). Particularism and the classification and reclassification of moral cases. IEEE Intel-
ligent Systems, 21(4), 22–28.
Gunkel, D. J., & Bryson, J. (2014). The machine as moral agent and patient. Philosophy & Technology,
27(1), 5–142.
Hall, J. S. (2011). Ethics for self-improving machines. In M. Anderson & S. L. Anderson (Eds.), Machine
ethics (pp. 512–523). Cambridge, MA: Cambridge University Press.
Hughes, J. (2014). Compassionate AI and selfless robots: A Buddhist approach. In P. Lin, K. Abney, & G.
Bekey (Eds.), Robot ethics: The ethical and social implications of robotics (pp. 69–84). Cambridge,
MA: MIT Press.
Johnson, A. M., & Axinn, S. (2014). Acting versus being moral: The limits of technological moral actors.
In Proceedings of the IEEE 2014 international symposium on ethics in engineering, science, and
technology (pp. 30(1)–30(4)). Piscataway, NJ: IEEE Press.
Kant, I., & Paton, H. J. (2009). Groundwork of the metaphysic of morals. New York: Harper Perennial
Modern Classics.
Katz, D. M., Bommarito, M. J., II, & Blackman, J. (2017). A general approach for predicting the behavior
of the Supreme Court of the United States. PLoS ONE, 12(4), 1–18.
Lin, P., Abney, K., & Bekey, G. A. (Eds.). (2014). Robot ethics: The ethical and social implications of
robotics. Intelligent robotics and autonomous agents. Cambridge, MA: MIT Press.
Macklin, R. (1999). Against relativism. Cultural diversity and the search for ethical universals in medi-
cine. New York: Oxford University Press.
McLaren, B. M. (2011). Computational models of ethical reasoning: Challenges, initial steps, and future
directions. In M. Anderson & S. L. Anderson (Eds.), Machine ethics (pp. 297–315). Cambridge,
MA: Cambridge University Press.
Moor, J. H. (2006). The nature, importance, and difficulty of machine ethics. Research Gate, 21(4),
18–21.
Nadeau, J. E. (2006). Only androids can be ethical. In K. Ford & C. Glymour (Eds.), Thinking about
android epistemology (pp. 241–248). Cambridge, MA: MIT Press.
Pereira, L. M., & Saptawijaya, A. (2011). Modeling morality with prospective logic. In M. Anderson & S.
L. Anderson (Eds.), Machine ethics (pp. 398–421). Cambridge, MA: Cambridge University Press.
Picard, R. (1997). Affective computing. Cambridge: The MIT Press.
Pontier, M. A., & Hoorn, J. F. (2012). Toward machines that behave ethically better than humans do. In
N. Miyake, B. Peebles, & R. P. Cooper (Eds.), Proceedings of the 34th international annual confer-
ence of the cognitive science society (pp. 2198–2203). CogSci’12.
13
Building Moral Robots: Ethical Pitfalls and Challenges
Powers, T. (2011). Prospectives for a Kantian machine. In M. Anderson & S. L. Anderson (Eds.),
Machine ethics (pp. 464–475). Cambridge, MA: Cambridge University Press.
Rodogno, R. (2016). Robots and the limits of morality. In M. Norskov (Ed.), Social robots: Boundaries,
potential, challenges. London: Routledge. http://pure.au.dk/portal/files/90856828/Robots_and_the_
Limits_of_Morality.pdf. Accessed 12 March 2016.
Ross, W. D. (1930). The right and the good. Oxford: Clarendon Press.
Rzepka, R., & Araki, K. (2005). What statistics could do for ethics? The idea of common sense pro-
cessing based safety valve. In AAAI fall symposium on machine ethics, Technical Report FS-05-06:
85-87.
Sullins, J. P. (2011). When is a robot a moral agent? In M. Anderson & S. L. Anderson (Eds.), Machine
ethics (pp. 151–161). Cambridge, MA: Cambridge University Press.
Thomson, J. J. (1976). Killing, letting die, and the trolley problem. The Monist, 59(2), 204–217.
Torrance, S. (2008). Ethics and consciousness in artificial agents. AI & Society, 22(4), 495–521.
Wallach, W., & Allen, C. (2010). Moral machines: Teaching robots right from wrong. Oxford: Oxford
University Press.
Wallach, W., Franklin, S., & Allen, C. (2010). A conceptual and computational model of moral decision
making in human and artificial agents. Topics in Cognitive Science, 2(3), 454–485.
Publisher’s Note Springer Nature remains neutral with regard to jurisdictional claims in published
maps and institutional affiliations.
13