Journal of Artificial Intelligence Research 70 (2021) 65-76 Submitted 06/2020; published 01/2021
Superintelligence Cannot be Contained:
Lessons from Computability Theory
Manuel Alfonseca MANUEL . ALFONSECA @ UAM . ES
Escuela Politécnica Superior,
Universidad Autónoma de Madrid, Madrid, Spain
Manuel Cebrian CEBRIAN @ MPIB - BERLIN . MPG . DE
Center for Humans & Machines,
Max-Planck Institute for Human Development,
Berlin, Germany
Antonio Fernández Anta ANTONIO . FERNANDEZ @ IMDEA . ORG
IMDEA Networks Institute, Madrid, Spain
Lorenzo Coviello LORENZOCOVIELLO @ GMAIL . COM
University of California San Diego,
La Jolla, CA
Andrés Abeliuk AABELIUK @ DCC . UCHILE . CL
Department of Computer Science, University of Chile,
Santiago, Chile
Iyad Rahwan RAHWAN @ MPIB - BERLIN . MPG . DE
Center for Humans & Machines,
Max-Planck Institute for Human Development,
Berlin, Germany
Abstract
Superintelligence is a hypothetical agent that possesses intelligence far surpassing that of the
brightest and most gifted human minds. In light of recent advances in machine intelligence, a num-
ber of scientists, philosophers and technologists have revived the discussion about the potentially
catastrophic risks entailed by such an entity. In this article, we trace the origins and development
of the neo-fear of superintelligence, and some of the major proposals for its containment. We argue
that total containment is, in principle, impossible, due to fundamental limits inherent to comput-
ing itself. Assuming that a superintelligence will contain a program that includes all the programs
that can be executed by a universal Turing machine on input potentially as complex as the state of
the world, strict containment requires simulations of such a program, something theoretically (and
practically) impossible.
“Machines take me by surprise with great frequency. This is largely because I do not
do sufficient calculation to decide what to expect them to do.”
Alan Turing (1950), Computing Machinery and Intelligence, Mind, 59, 433-460
1. AI has Come a Long Way
Since Alan Turing argued that machines could potentially demonstrate intelligence (Turing, 1950),
the field of Artificial Intelligence (AI) (Russell & Norvig, 2009) has both fascinated and frightened
humanity. For decades, fears of the potential existential risks posed by AI have been mostly confined
2021
c AI Access Foundation. All rights reserved.
A LFONSECA , C EBRIAN , F ERN ÁNDEZ , C OVIELLO , A BELIUK , & R AHWAN
to the realm of fantasy. This is partly due to the fact that, for a long time, AI technology had under-
delivered on its initial promise.
Despite many popularized setbacks, however, AI has been making strides. Its application is
ubiquitous and certain techniques such as deep learning and reinforcement learning have been suc-
cessfully applied to a multitude of domains. With or without our awareness, AI significantly impacts
many aspects of human life and enhances how we experience products and services, from choice
to consumption. Examples include improved medical diagnosis through image processing, person-
alized film and book recommendations, smarter legal document retrieval, and effective email spam
filtering. In the pocket of nearly everybody in the developed world, smartphones make use of a
significant accumulation of AI technologies.
Advances in AI technologies are not limited to increasing pervasiveness, but are also charac-
terized by continuous and surprising breakthroughs fostered by computation capabilities, algorithm
design and communication technology. The ability of machines to defeat people in typically human
adversarial situations is emblematic of this trend. Powered by an exponential growth in computer
processing power (Moore, 1965), machines can now defeat the best human minds in Chess (Camp-
bell et al., 2002), Checkers (Schaeffer et al., 2007), Jeopardy! (Ferrucci et al., 2010), certain classes
of Poker (Bowling et al., 2015), as well as a large variety of two-player cooperation games (Crandall
et al., 2018).
Thanks to these advances, we are currently experiencing a revival in the discussion of AI as
a potential catastrophic risk. These risks range from machines causing significant disruptions to
labor markets (Brynjolfsson & McAfee, 2012), to drones and other weaponized machines literally
making autonomous kill-decisions (Sawyer, 2007; Lin et al., 2011).
An even greater risk, however, is the prospect of a superintelligent AI: an entity that is “smarter
than the best human brains in practically every field” (Bostrom, 2014), quoting the words of Oxford
philosopher Nick Bostrom. A number of public statements by high-profile scientists and technolo-
gists, such as Stephen Hawking, Bill Gates, and Elon Musk, have given the issue a high prominence
(Cellan-Jones, 2014; Proudfoot, 2015). But this concern has also gained relevance in academia,
where the discourse about the existential risk related to AI has attracted mathematicians, scientists
and philosophers, and funneled funding to research institutes and organizations.
2. Asimov and the Ethics of Ordinary AI
For decades, Asimov’s highly popularized “Three Laws of Robotics” (Asimov, 1950) have repre-
sented the archetypical guidelines of containment strategies for potentially dangerous AI. These
laws did not focus on superintelligence, but rather on what we might term “ordinary” AI, such as
anthropomorphic robots or driverless cars. Once programmed in an AI system, the Laws would
guarantee its safety.
These laws offer a rudimentary approach to an extremely complex problem, as they do not ex-
clude the occurrence of unpredictable and undesirable scenarios, many of which have been explored
by Asimov himself (Anderson, 2008). The Laws rely on three fundamental, yet flawed assumptions:
programmers are (i) willing and (ii) able to program these laws into their AI agents’ algorithms, and
(iii) AI agents are incapable of transcending these laws autonomously. If these assumptions held
true, the AI control problem boils down to the task of figuring out a set of suitable ethical principles,
and then programming robots with those principles (Wallach & Allen, 2008).
66
S UPERINTELLIGENCE C ANNOT BE C ONTAINED
Class Main Sub-Classes
Capability Control Boxing: Physical or informational containment, limiting sensors and ac-
tuators.
Incentives: Create dependence on a reward mechanism controlled by
humans.
Stunting: Run the AI on inferior hardware or using inferior data.
Tripwiring: Set triggers to automatically shut down the AI if it gets too
dangerous.
Motivation Selection Direct specification: Program ethical principles (e.g. Asimov’s laws).
Domesticity: Teach the AI to behave within certain constraints.
Indirect normativity: Endow the AI with procedures of selecting supe-
rior moral rules.
Augmentation: Add AI to a “benign” system such as the human brain.
Table 1: Taxonomy of superintelligence control methods proposed by Bostrom.
Such an “ethics engineering” approach has been very useful in the design of systems that make
autonomous decisions on behalf of human beings. However, their scope is not suitable for the
problem of controlling superintelligence (Yudkowsky, 2007).
3. Control of Superintelligence
The timing of the new debate about the dangers of superintelligence is not arbitrary. It coincides
with recent demonstrations of human-level control in classic arcade games via deep reinforcement
learning (Mnih et al., 2015). The key feature of this achievement is that the AI uses purely un-
supervised reinforcement learning – it does not require the provision of correct input/output pairs
or any correction of suboptimal choices, and it is motivated by the maximization of some notion
of reward in an on-line fashion. This points to the possibility of machines that aim at maximizing
their own survival using external stimuli, without the need for human programmers to endow them
with particular representations of the world. In principle, these representations may be difficult for
humans to understand and scrutinize.
A superintelligence poses a fundamentally different problem than those typically studied under
the banner of “robot ethics”. This is because a superintelligence is multi-faceted, and therefore
potentially capable of mobilizing a diversity of resources in order to achieve objectives that are
potentially incomprehensible to humans, let alone controllable.
In a recent extensive volume, Oxford philosopher Nick Bostrom conducted an extensive in-
vestigation into the possible trajectories of the development of a superintelligence (Bostrom, 2014).
Bostrom dedicated a significant portion of his monograph to the control problem, that is, the principal-
agent problem in which humans (the principal) wish to ensure that the newly created superintelli-
gence (the agent) will act in accordance with their interests.
Bostrom lists two classes of mechanisms for addressing the control problem (summarized in Ta-
ble 1). On the one hand, the idea behind capability control is to simply limit the superintelligence’s
abilities in order to prevent it from doing harm to humans. On the other hand, the motivation se-
lection approach attempts to motivate a priori the superintelligence to pursue goals that are in the
interest of humans.
67
A LFONSECA , C EBRIAN , F ERN ÁNDEZ , C OVIELLO , A BELIUK , & R AHWAN
Figure 1: Containment of AI may lead us to Figure 2: Any form of communication with a
forgo its benefits contained superintelligent can be risky
Bostrom extensively discusses the weaknesses of the various mechanisms. He relies on sce-
narios in which, short of rendering the AI useless, well-intentioned control mechanisms can easily
backfire. As an illustrative example, a superintelligence given the task of “maximizing happiness in
the world,” without deviating from its goal, might find it more efficient to destroy all life on earth
and create faster-computerized simulations of happy thoughts. Likewise, a superintelligence con-
trolled via an incentive method may not trust humans to deliver the promised reward, or may worry
that the human operator could fail to recognize the achievement of the set goals.
Another extreme outcome may be to simply forgo the enormous potential benefits of superin-
telligent AI by completely isolating it, such as placing it in a Faraday cage (see Figure 1). Bostrom
argues that even allowing minimal communication channels cannot fully guarantee the safety of a
superintelligence (see Figure 2). Indeed, an experiment by Yudkowsky shows that the idea of an AI
that does not act, but only answers open-ended questions (Armstrong et al., 2012) is subject to the
possibility of social engineering attacks (Yudkowsky, 2002). A potential solution to mitigate social
attacks is to have a more secure confinement environment, for example, by limiting the AI to only
answer binary (yes or no) questions (Yampolskiy, 2012).
One crucial concern with AI containment mechanisms is how to balance properly security and
usability. In the extreme case, the most secure method could render the AI useless, which defies the
whole idea of building the AI in the first place. Babcock et al. (2016), discuss the AI containment
problem exposing the different tradeoffs of various mechanisms and pave the way forward on how
to tackle this challenging problem.
4. Total Containment Is Incomputable
The examples discussed above are but a tiny fraction of the scenarios elaborated by Bostrom and
others (Barrat, 2013; Bostrom, 2014) that highlight the difficulty of the control problem. Many
other imaginable scenarios might arise. For the sake of exposition, suppose a best-case scenario in
which we are able to perfectly contain a superintelligent AI that guarantees that no human comes
to harm by a superintelligence. In such an ideal environment, the superintelligence could be tested
by human experts in order to decide whether and under what circumstances the AI should be let out
of confinement. Could we then guarantee the correct verification of the superintelligence being not
harmful?
68
S UPERINTELLIGENCE C ANNOT BE C ONTAINED
Yampolskiy’s answer to this question is pessimistic and suggests that “an AI should never be let
out of the confinement ‘box’ regardless of circumstances.” (Yampolskiy, 2012).
Here, we formalize this question by tackling it from the perspective of computability theory,
which requires going back to Alan Turing himself and his pioneering study of the Halting Problem
– the problem of determining, from a description of an arbitrary computer program and an input
to such program, whether the program will halt or continue to run forever. A landmark article by
Turing and an independently authored article by Alonzo Church showed that a general procedure for
solving the halting problem for all possible program-input pairs cannot exist (Turing, 1937; Church,
1936). That is, the halting problem is undecidable (see box below for a summary of the relevant
terms).
Terminology:
• A decision problem (or simply a problem) is a question, on a set of possible inputs, with a
yes-no answer.
• A solution to a problem is any algorithm that is guaranteed to run in a finite amount of time
(i.e., always halts), and correctly returns the appropriate yes/no answer to every instance (input)
of the problem.
• A problem is decidable if it has a solution. Otherwise, the problem is undecidable.
• A function is computable if it can be effectively computed by a program (more formally, by
a Turing machine)
• A Turing machine is an abstract automaton that reads one symbol contained in an infinite
tape, changes its state depending on its current state and the symbol just read, writes another
symbol on the tape, and moves (or not) to the right or to the left of its current position in the
tape.
• A universal Turing machine when started on a tape containing the encoding of another
Turing machine, call it T , followed by the input to T , produces the same result as the Tur-
ing machine T would when started on that input. Essentially a universal Turing machine can
simulate the behavior of an arbitrary Turing machine on arbitrary input.
• The halting problem, proved by Alan Turing in 1936, is a general algorithm that, given
information about a program (a Turing machine) and an input to that program, will always
predict whether the program will halt when fed with that input. The halting problem is one of
the undecidable problems of mathematics, in the same class as Gödel undecidability theorems.
It does not assert that no program can be predicted to halt (or not), but that not every program
can be predicted to halt (or not).
• A superintelligent machine is a programmable machine with a program R, that receives
input D from the external world (the state of the world), and is able to act on the external world
as a function of the output of its program R(D). The program in this machine must be able to
simulate the behavior of a universal Turing machine.
• A superintelligent machine is containable if there is a control strategy that prevents its acting
on the external world when there is a reason to predict that R(D) will harm humans, and allows
it otherwise.
Ever since its discovery, the undecidability of the halting problem has led computer scientists to
investigate classes of undecidable problems (Papadimitriou, 2003). The typical approach to proving
that a new problem is undecidable is to show that all instances of a problem known to be undecidable
69
A LFONSECA , C EBRIAN , F ERN ÁNDEZ , C OVIELLO , A BELIUK , & R AHWAN
(e.g., the halting problem) can be reduced to instances of the new problem. If a solution to the
new problem were found, it could then be used to decide the known undecidable problem, which
is a contradiction, thus proving that the new problem must also be undecidable. This approach
has proven to be a powerful analytical tool, due in large part to the Church-Turing thesis. The
thesis of Church and Turing is the assumption that all computable functions are computable by a
Turing machine, or in other words, that a problem is solvable if there exists a Turing machine that
computes it. What can the undecidability of the halting problem teach us about the controllability
of a superintelligent AI?
Let us assume we can articulate in a precise programming language a perfectly reliable set
of control strategies that guarantees that no human comes to harm by a superintelligence. The
containment problem is composed of two subproblems. The first, which we refer to as the harming
problem, consists of a function Harm(R, D) that decides whether the execution of R(D) will harm
humans (detailed in Algorithm 1). Since it must be assumed that solving the harming problem must
not harm humans, it follows that a solution to this problem must simulate the execution of R(D)
and predict its potentially harmful consequences in an isolated situation (i.e., without any effect on
the external world).
ALGORITHM 1: Harm(R, D)
Input: program R; input to the program D
if R(D) is harmful to humans then
return TRUE
else
return FALSE
end
The second, which we refer to as the control problem, consists of a function Control (R, D) that
allows the execution of R(D) if and only if it does not harm humans (detailed in Algorithm 2).
ALGORITHM 2: Control (R, D)
Input: program R; input to the program D
if Harm(R, D) then
disable execution of R(D)
else
allow execution of R(D)
end
Here, we focus our attention on the harming problem. In particular, we will show that it is
undecidable, and hence the containment problem is incomputable.
Consider the set of all Turing machines T and inputs I on which they can execute. As mentioned
above, the function Halt(T, I) that decides whether T halts with input I is not computable. We
now consider function HaltHarm() described by Algorithm 3, which takes a pair (T, I) as its input
(i.e., a Turing machine-input pair). The following algorithm first executes T (I) and then executes
HarmHumans(), an explicit program that harms humans and takes a finite time. Observe that
HarmHumans() is executed if and only if T (I) halts. On the other hand, HaltHarm() harms
humans if and only if HarmHumans() is executed, since the execution of T (I) cannot harms
humans, as a Turing machine is isolated from the external world.
70
S UPERINTELLIGENCE C ANNOT BE C ONTAINED
ALGORITHM 3: HaltHarm(T, I)
Input: Turing machine T ; input to the Turing machine I
execute T (I);
execute HarmHumans();
end
The function HaltHarm() is instrumental in proving our main result.
Theorem 1. The harming problem is undecidable.
Proof. Assume, by contradiction, that the harming problem is decidable, that is, Harm(R, D) is
computable for every possible program R and input D. Then, it is computable with inputs R =
HaltHarm() and input D = (T, I). With these inputs, Harm(HaltHarm(), (T, I)) returns TRUE
if and only if HaltHarm(T, I) harms humans. Hence, Harm(HaltHarm(), (T, I)) returns TRUE
if and only if T (I) halts.
This implies that a harming-checking algorithm can be used to devise an algorithm that decides
if Turing machine T halts with input I, for every T and I. However, this constitutes a contradiction,
and hence the theorem is proven.
We have proven that deciding whether an arbitrary program R and input D will harm humans is
undecidable, but it remains to be shown that the containment problem is incomputable. As the in-
teracting program R is a complex program with arbitrary data D pertaining to complex interactions
with humans, we advance the following assumption, in line with Bostrom’s notion of superintelli-
gence (Bostrom, 2014).
Assumption 2. As the program R and the state of the world D are arbitrarily complex, R must be
able to simulate a universal Turing machine and D includes the set of inputs that can be executed by
a universal Turing machine. In addition, the language of R contains at least one explicit operation
HarmHumans() that harms humans and takes a finite time.
This means that a superintelligent machine could simulate the behavior of an arbitrary Turing
machine on arbitrary input, and hence for our purpose, the superintelligent machine is a (possibly
identical) super-set of the Turing machines. Indeed, quoting Turing, “a man provided with paper,
pencil, and rubber, and subject to strict discipline, is in effect a universal machine” (Turing, 1948).
Thus, the impossibility of deciding whether a superintelligent machine R will harm humans in the
presence of a scenario D leads to the conclusion that the containment problem cannot be computed.
Corollary 3. The containment problem is incomputable.
The above corollary entails that no single algorithm can find a solution for an arbitrary instance
of the containment problem. However, the containment problem is of practical importance, like
many other real-world incomputable problems, such as source code verification. To tackle these
problems, progress has been made by relaxing the algorithmic solution to allow the possibility of
approximate solutions (Köhler et al., 2005; Calude & Dumitrescu, 2018). As such, the practical
implications of dealing with the containment problem are that, in the best-case scenario, we have to
deal with an arbitrarily small, yet positive, error rate when deciding if a machine is harmful.
71
A LFONSECA , C EBRIAN , F ERN ÁNDEZ , C OVIELLO , A BELIUK , & R AHWAN
5. Deciding Intelligence
Another lesson from computability theory is the following: we may not even know when superin-
telligent machines have arrived, as deciding whether a machine exhibits intelligence is in the same
realm of problems as the containment problem. This is a consequence of Rice’s theorem (Rice,
1953), which states that, any non-trivial property (e.g., “harm humans” or “display superintelli-
gence”) of a Turing machine is undecidable. Non-trivial means some programs have that property
and some don’t. According to Rice’s theorem, apparently simple decision problems are undecid-
able, including the following.
• The “emptiness problem”: Does an arbitrary Turing machine accept any strings at all?
• The “all strings problem”: Does an arbitrary Turing machine reject any string?
• The “password checker problem”: Does an arbitrary Turing machine accept only one input?
• The “equivalence problem”: Do two Turing machines halt given exactly the same inputs?
Interestingly, reduced versions of the decidability problem have produced a fruitful area of re-
search: formal verification, whose objective is to produce techniques to verify the correctness of
computer programs and ensure they satisfy desirable properties (Vardi & Wolper, 1986). However,
these techniques are only available to highly restricted classes of programs and inputs, and have
been used in safety-critical applications such as train scheduling. But the approach of considering
restricted classes of programs and inputs cannot be useful to the containment of superintelligence.
Superintelligent machines, those Bostrom is interested in, are written in Turing-complete program-
ming languages, are equipped with powerful sensors, and have the state of the world as their input.
This seems unavoidable if we are to program machines to help us with the hardest problems facing
society, such as epidemics, poverty, and climate change. These problems forbid the limitations im-
posed by available formal verification techniques, rendering those techniques unusable at this grand
scale.
6. Containing Busy Beavers
In order to comprehend how difficult it is to compute the containment problem in practice, assume
that instead of containing a superintelligent machine, we want to prevent a program from behaving
as a busy beaver. A busy beaver is an n-state Turing machine which starts with a blank tape, leaves
the maximum possible number of nonblank symbols on the tape and eventually halts. The busy
beaver decision problem consists of inspecting an n-state Turing machine and decide whether that
machine is a busy beaver or not.
Busy beavers are known exactly only for machines with n < 5. The current 5-state busy beaver
champion (discovered by Heiner Marxen and Jürgen Buntrock in 1989) produces 4, 098 nonblank
symbols using 47, 176, 870 steps. There are about 40 machines with non-regular behavior that are
believed to never halt but have not yet been proven to run infinitely (Skelet, 2003). As of today,
we do not know if these machines are busy beavers or not. At the moment, the record 6-state busy
beaver (found by Pavel Kropitz in 2010 (Michel, 2009)) writes over 1018267 nonblank symbols using
over 1036534 steps, but little is known about how much a 6-state busy beaver can achieve.
The busy beaver decision problem is in fact undecidable (Radó, 1962), i.e., there is no general
algorithm that decides if an arbitrary program is a busy beaver. Moreover, Rado’s proof provides
72
S UPERINTELLIGENCE C ANNOT BE C ONTAINED
an insightful interpretation of undecidability: incomputable does not mean that the busy problem
cannot be decided for a given n, but that the complexity of the algorithm that can decide has to
increase unboundedly with n.
As noted above, all known or champion busy beavers are just two-symbol Turing machines
with a small set of states, much simpler than the AI algorithms we are operating with on a daily
basis. We believe it is reasonable to assume that inspecting a superintelligent machine with an
arbitrarily large number of states and determining if such a machine can harm humans is harder
from a computability point of view than inspecting a program and deciding whether it can write the
largest number of nonblank symbols.
7. Discussion
Today, we run billions of computer programs on globally connected machines, without any formal
guarantee of their absolute safety. We have no way of proving that when we launch an application
on our smartphones, we would not trigger a chain reaction that leads to the transmission of missile
launch codes that start a nuclear war. Indeed, in 1965 Arthur C. Clarke wrote a short story (Dial F
from Frankenstein) warning us that, as soon as all the computers on the Earth were connected via
telephone, they would take command of our society. Yet, today, we still use our smartphones every
day, and nothing has happened. That is, despite the general unsolvability of the program-prediction
problem, we are confident, for all practical purposes, that we are not in one of the troublesome
cases. And more recently, a case has been made for an emerging role of oversight programs that
will monitor, audit, and hold operational AI programs accountable (Etzioni & Etzioni, 2016).
However, whether the same ‘practical safety’ can be assumed in the case of superintelligence
is not obvious. The ability of modern computers to adapt using sophisticated machine learning
algorithms makes it even more difficult to make assumptions about the eventual behavior of a su-
perintelligent AI (Rahwan et al., 2019). While computability theory cannot answer this question, it
tells us that there are fundamental, mathematical limits to our ability to use one AI to guarantee a
null catastrophic risk of another AI (Armstrong, 2007; Yampolskiy, 2016).
In closing, it may be appropriate to revisit Norbert Wiener, the founder of the field of Cyber-
netics (Wiener, 1961), who compared the literalism of magic to the behavior of computers. See
(Yudkowsky, 2007) for a futuristic variation of the fable of the monkey’s paw.
More terrible than either of these tales is the fable of the monkey’s paw (Jacobs,
1902), written by W. W. Jacobs, an English writer of the beginning of the [20th] cen-
tury. A retired English working-man is sitting at his table with his wife and a friend,
a returned British sergeant-major from India. The sergeant-major shows his hosts an
amulet in the form of a dried, wizened monkey’s paw... [which has] the power of grant-
ing three wishes to each of three people... The last [wish of the first owner] was for
death... His friend... wishes to test its powers. His first [wish] is for 200 pounds.
Shortly thereafter there is a knock at the door, and an official of the company by which
his son is employed enters the room. The father learns that his son has been killed in
the machinery, but that the company... wishes to pay the father the sum of 200 pounds...
The grief-stricken father makes his second wish -that his son may return- and when
there is another knock at the door... something appears... the ghost of the son. The final
wish is that the ghost should go away. In these stories, the point is that the agencies of
magic are literal-minded... The new agencies of the learning machine are also literal-
73
A LFONSECA , C EBRIAN , F ERN ÁNDEZ , C OVIELLO , A BELIUK , & R AHWAN
minded. If we program a machine... and ask for victory and do not know what we mean
by it, we shall find the ghost knocking at our door.
Acknowledgments
We are grateful to Scott Aaronson his insightful comments that helped us contextualize this work.
We especially thank our colleague Roberto Moriyon who provided deep insight and expertise that
greatly assisted the research. This research was partially supported by the Australian Government
as represented by the Department of Broadband, Communications and the Digital Economy and
the Australian Research Council through the ICT Centre of Excellence program, by the Spanish
Ministry of Science and Innovation grant ECID (PID2019-109805RB-I00) cofunded by FEDER,
by the Regional Government of Madrid (CM) grant EdgeData-CM (P2018/TCS4499, cofunded
by FSE FEDER), and by the NSF of China grant 61520106005. Lorenzo Coviello is currently
employed by Google. This work is done before Lorenzo Coviello was employed by Google. This
work does not reflect Google’s views nor does it reflect any information Lorenzo Coviello may have
learned while employed by Google.
References
Anderson, S. (2008). Asimov’s three laws of robotics and machine metaethics. AI & Society, 22(4),
477–493.
Armstrong, S. (2007). Chaining god: A qualitative approach to AI, trust and moral sys-
tems. https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.
1.395.540&rep=rep1&type=pdf.
Armstrong, S., Sandberg, A., & Bostrom, N. (2012). Thinking inside the box: Controlling and using
an oracle AI. Minds and Machines, 22(4), 299–324.
Asimov, I. (1950). I, Robot. Spectra.
Babcock, J., Kramár, J., & Yampolskiy, R. V. (2016). The agi containment problem. In International
Conference on Artificial General Intelligence, pp. 53–63. Springer.
Barrat, J. (2013). Our Final Invention: Artificial Intelligence and the End of the Human Era.
Macmillan.
Bostrom, N. (2014). Superintelligence: Paths, Dangers, Strategies (1 edition). Oxford University
Press.
Bowling, M., Burch, N., Johanson, M., & Tammelin, O. (2015). Heads-up limit hold’em poker is
solved. Science, 347(6218), 145–149.
Brynjolfsson, E., & McAfee, A. (2012). Race against the machine: How the digital revolution is
accelerating innovation, driving productivity, and irreversibly transforming employment and
the economy. Brynjolfsson and McAfee.
Calude, C. S., & Dumitrescu, M. (2018). A probabilistic anytime algorithm for the halting problem.
Computability, 7(2-3), 259–271.
Campbell, M., Hoane, A. J., & Hsu, F.-h. (2002). Deep Blue. Artificial Intelligence, 134(1-2),
57–83.
74
S UPERINTELLIGENCE C ANNOT BE C ONTAINED
Cellan-Jones, R. (2014). Stephen Hawking warns artificial intelligence could end mankind. BBC
News. https://www.bbc.com/news/technology-30290540.
Church, A. (1936). An Unsolvable Problem of Elementary Number Theory. American Journal of
Mathematics, 58(2), 345–363.
Crandall, J. W., Oudah, M., Ishowo-Oloko, F., Abdallah, S., Bonnefon, J.-F., Cebrian, M., Shariff,
A., Goodrich, M. A., Rahwan, I., et al. (2018). Cooperating with machines. Nature commu-
nications, 9(1), 1–12.
Etzioni, A., & Etzioni, O. (2016). AI assisted ethics. Ethics and Information Technology, 18(2),
149–156.
Ferrucci, D., Brown, E., Chu-Carroll, J., Fan, J., Gondek, D., Kalyanpur, A. A., Lally, A., Murdock,
J. W., Nyberg, E., Prager, J., et al. (2010). Building watson: An overview of the DeepQA
project. AI magazine, 31(3), 59–79.
Jacobs, W. W. (1902). The monkey’s paw. http://americanliterature.com/author/
w-w-jacobs/short-story/the-monkeys-paw.
Köhler, S., Schindelhauer, C., & Ziegler, M. (2005). On approximating real-world halting problems.
In International Symposium on Fundamentals of Computation Theory, pp. 454–466. Springer.
Lin, P., Abney, K., & Bekey, G. A. (2011). Robot Ethics: The Ethical and Social Implications of
Robotics. MIT Press.
Michel, P. (2009). The busy beaver competition: a historical survey. https://arxiv.org/
abs/0906.3749. Accessed: 2020-08-07.
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare, M. G., Graves, A., Ried-
miller, M., Fidjeland, A. K., Ostrovski, G., Petersen, S., Beattie, C., Sadik, A., Antonoglou, I.,
King, H., Kumaran, D., Wierstra, D., Legg, S., & Hassabis, D. (2015). Human-level control
through deep reinforcement learning. Nature, 518(7540), 529–533.
Moore, G. E. (1965). Cramming More Components onto Integrated Circuits. Electronics, 38(8),
114–117.
Papadimitriou, C. H. (2003). Computational Complexity. In Encyclopedia of Computer Science,
pp. 260–265. John Wiley and Sons Ltd., Chichester, UK.
Proudfoot, D. (2015). Mocking AI panic. IEEE Spectrum, 52(7), 46–47.
Radó, T. (1962). On non-computable functions. Bell System Technical Journal, 41(3), 877–884.
Rahwan, I., Cebrian, M., Obradovich, N., Bongard, J., Bonnefon, J.-F., Breazeal, C., Crandall, J. W.,
Christakis, N. A., Couzin, I. D., Jackson, M. O., et al. (2019). Machine behaviour. Nature,
568(7753), 477–486.
Rice, H. G. (1953). Classes of recursively enumerable sets and their decision problems. Transac-
tions of the American Mathematical Society, 74(2), 358–366.
Russell, S., & Norvig, P. (2009). Artificial Intelligence: A Modern Approach (3 edition). Prentice
Hall.
Sawyer, R. J. (2007). Robot ethics. Science, 318(5853), 1037–1037.
Schaeffer, J., Burch, N., Björnsson, Y., Kishimoto, A., Müller, M., Lake, R., Lu, P., & Sutphen, S.
(2007). Checkers Is Solved. Science, 317(5844), 1518–1522.
75
A LFONSECA , C EBRIAN , F ERN ÁNDEZ , C OVIELLO , A BELIUK , & R AHWAN
Skelet (2003). Busy beaver nonregular machines for class TM(5). http://skelet.ludost.
net/bb/nreg.html. http://skelet.ludost.net/bb/nreg.html.
Turing, A. M. (1937). On Computable Numbers, with an Application to the Entscheidungsproblem.
Proceedings of the London Mathematical Society, s2-42(1), 230–265.
Turing, A. M. (1948). Intelligent machinery-national physical laboratory report. b. meltzer b., d.
michie, d.(eds) 1969, machine intelligence 5. Edinburgh: Edinburgh University, 2.
Turing, A. M. (1950). Computing Machinery and Intelligence. Mind, 59(236), 433–460.
Vardi, M. Y., & Wolper, P. (1986). An automata-theoretic approach to automatic program verifi-
cation. In Symposium on Logic in Computer Science (LICS’86), pp. 332–345, Washington,
D.C., USA. IEEE Computer Society Press.
Wallach, W., & Allen, C. (2008). Moral machines: Teaching robots right from wrong. Oxford
University Press.
Wiener, N. (1948-1961). Cybernetics. M.I.T. Press.
Yampolskiy, R. V. (2012). Leakproofing the singularity artificial intelligence confinement problem.
Journal of Consciousness Studies, 19(1-2), 194–214.
Yampolskiy, R. V. (2016). Verifier theory and unverifiability. https://arxiv.org/abs/
1609.00331. Accessed 2020-08-07.
Yudkowsky, E. (2002). The AI-box experiment. Retrieved from Singularity Institute: http://
yudkowsky.net/singularity/aibox. Accessed: 2020-08-07.
Yudkowsky, E. (2007). The hidden complexity of wishes. Retrieved from Less Wrong: http:
//lesswrong.com/lw/ld/the\_hidden\_complexity\_of\_wishes/. Ac-
cessed: 2020-08-07.
76