Quantum Computing in The Arts and Humanities An Introduction To Core Concepts, Theory and Applications
Quantum Computing in The Arts and Humanities An Introduction To Core Concepts, Theory and Applications
Quantum Computing
in the Arts and
Humanities
An Introduction to Core Concepts,
Theory and Applications
Quantum Computing in the Arts and Humanities
Eduardo Reck Miranda
Editor
Quantum Computing
in the Arts and Humanities
An Introduction to Core Concepts, Theory
and Applications
Editor
Prof. Eduardo Reck Miranda
Interdisciplinary Centre for Computer
Music Research (ICCMR)
University of Plymouth
Plymouth, UK
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
v
vi Preface
be able to do that a standard computer would not be capable of? Today’s high-
performance supercomputers are rather powerful. And they will continue to improve.
Consider the Fugaku supercomputer, developed by Fujitsu and RIKEN in Japan: it
performs circa 600 quadrillion operations per second. To put this in perspective, a
top of the range desktop computer does approximately 100 billion operations per
second. That is, Fugaku is six million times faster than a decent desktop computer.
All the same, quantum computing is here to stay. What started as a desire to build
a machine for physicists to study quantum mechanics1 has been evolving towards
the ambition to develop general-purpose quantum computing technology for a wide
range of applications, some of which we have not even dreamed of yet. And more,
I advocate that processing speed is not the only benefit that such machines will
bring to society. They will bring new philosophies, new approaches to design, new
algorithms, new creative artefacts, new economies and so on. In the long run, using
an actual quantum computer for mundane applications may as well become a matter
of choice or luxury rather than necessity. Who knows?
Much ongoing research into quantum computing worldwide focuses on improving
hardware and developing solutions for science and technology. Nevertheless, there
also is growing activity within areas where quantum computing may impact beyond
science and technology, including the arts, humanities and fringe interdisciplinary
initiatives. These are exciting because they might influence developments in the
quantum computing industry; e.g., create unforeseen new markets.
This book brings a collection of chapters by scholars championing research in
their respective areas (such as Philosophy, Linguistics and Music, amongst others)
informed by quantum computing. The authors offer thought-provoking discussions
about cognition, perception, language, music, games, visualization of the sub-atomic
world, brain-machine communication and more.
I am indebted to all contributing authors who enthusiastically supported this book
project. They made possible what seemed impossible. We have here what is probably
the first book ever on Quantum Computing in the Arts and Humanities.
References
Deutsch, D. (1985). Quantum theory, the church-turing principle and the universal quantum
computer. Proceedings of the Royal Society of London A, 400, 97–117.
Feynman, R. P. (1982). Simulating physics with computers. International Journal of Theoretical
Physics, 21, 467–88.
1 The impetus to develop a quantum computer stemmed from the fact that some simulations to
study quantum mechanics would allegedly require computing power that even state of the art
supercomputers might take years to run. The idea then is to harness quantum mechanics to build
quantum computers rather than simulate quantum mechanics on standard computers.
Preface vii
Holevo, A. S. (1973). Bounds for the quantity of information transmitted by a quantum commu-
nication channel, Problemy Peredachi Informatsii, 9, 3–11. English translation in Problems of
Information Transmission, 9, 177–83.
Ingarden, R. S. (1976). Quantum information theory. Reports on Mathematical Physics, 10, 43–72.
Poplavskii, R. P. (1975). Thermodynamical models of information processing (in Russian), Uspekhi
Fizicheskikh Nauk, 115, 465–501.
Contents
ix
From Digital Humanities to Quantum
Humanities: Potentials and Applications
Johanna Barzen
1 Introduction
As quantum computers are becoming real there are several application areas where
the use of quantum computers is expected to be especially promising (National
Academies of Sciences, Engineering, and Medicine, 2019): one is simulation,
focusing on molecular properties used in material science or in pharma industry, and
the other is machine learning e.g. classification, clustering or dimension reduction
making heavy use of optimization.
The main difference between classical computers and quantum computers is the
different information unit used, namely bits for classical computers and qubits for
quantum computers (Nielsen & Chuang, 2010). Qubits are not restricted to the two
states 0 and 1 but can be in an infinite number of different combinations of these
J. Barzen (B)
Institute of Architecture of Application Systems (IAAS), University of Stuttgart, Stuttgart,
Germany
e-mail: [email protected]
how they are used for restricted Boltzmann machines and autoencoders, and sketches
their realizations on quantum computers. As many quantum algorithms are hybrid,
meaning they consist of both, a quantum part and a classical part, Sect. 6 focuses on
variational hybrid quantum–classical algorithms. The main idea is outlined as well
as an application for clustering, namely a quantum version for solving the maximum
cut problem. Therefore, techniques as quantum approximate optimization algorithm
(QAOA) and variational quantum eigensolver (VQE) are discussed. To support access
to the described techniques, Sect. 7 focuses on our quantum humanities analysis tool
(QHAna) that is used for pattern detection. In addition to the substantive goal of
the use case from the MUSE project to extract costume patterns in films, QHAna
provides easy access for people without quantum computing background to evaluate
the benefits of using quantum algorithms and allows them to gain initial application
knowledge in the field of quantum humanities. Since both classical analysis algo-
rithms and their quantum counterparts are included, the comparison of their results
allows a deeper understanding within quantum humanities. Section 8 concludes the
chapter and gives an outlook on future work in quantum humanities.
How to read this chapter: As quantum humanities brings together very heteroge-
nous disciplines (e.g. the respective humanities discipline, physics, mathematics,
computer science) and each has its own approaches, terminology, concepts and
research culture, a fundamental challenge of quantum humanities is to find a common
language to communicate between these disciplines. Thus, the chapter starts with an
easy-to-understand introduction to the overall topic, the vision and potentials that
does not require any previous knowledge (Sects. 1–4). The underlying ideas and
concepts get more and more refined, up to mathematical definitions (Sects. 5–6). We
use mathematical definitions as they give little room for interpretation and therefore
seems appropriate as long as there is no common language yet. As the mathematical
definitions may be challenging to some readers they are accomplished by descriptions
and pictures giving a high-level view on the core concepts, so some readers may skip
over the mathematical formalisms. Please note that algorithmic details themselves
are not the focus. Instead, the underlying concepts of the (classical and quantum)
algorithms are provided that are relevant for our approach, as well as the tooling
provided by QHAna. QHAna than outlines the practical usage of the introduced
concepts in a use case centric manner (Sect. 7).
How beneficial the use of computers, their practical support as well as techniques
and methods from computer science can contribute to research in the humanities has
been proven by the establishment of the digital humanities. With quantum computers
becoming available, it is promising to extend the usage of classical computers as
done in the digital humanities by the use of quantum computers. As stressed in the
introduction, working with quantum computers, programing algorithms and letting
them run on quantum hardware is very different to working with classical computers.
4 J. Barzen
When speaking about digital humanities the term combines a broad variety of topics,
contents and contexts. This is reflected by the continuously growing amount of
literature on divergent approaches in this domain (Berry, 2012; Berry & Fagerjord,
2017; Brown, 2020; Jannidis et al., 2017; Terras et al., 2013). But there are three
core elements that can be identified when taking a closer look at approaches that
define digital humanities: (i) digital humanities is about bringing together humani-
ties research and information technology, (ii) it is a diverse field, and (iii) it is a still
emerging field (DHQ, 2021).
The first of those highlighted three core elements stresses the combination of
computer science and the humanities research. Digital humanities therefore are a
highly inter- and trans-disciplinary field operating between the traditional humanities
and the (in comparison) rather young computer science. The aim is to examine
existing and new questions from the humanities by using concepts, methods and
tools from computer science. Therefore, digital humanities is understood as a field
of methods that unites different topics, approaches and sub-disciplines and as such
fosters synergies between different scientific disciplines.
This is rather complex and hints to core element two: digital humanities are a
diverse field. ‘Humanities’ already encompasses a large variety of disciplines, such
as philosophy, linguistics, arts, cultural studies, archaeology, musicology and media
science, to name just a few. It gets even broader when taking into account that also
“computer science” covers quite different subject areas like databases, visualization,
or theoretical computer science. In the compilation of so many fields of interest,
digital humanities is a very heterogeneous area. It attempts to include the application
as well as the further development of tools and methods of computer science in
order to possibly obtain new and profound answers to research questions in the
humanities while performing a critical self-reflection of the methodological and
theoretical approaches.
Even though discussions on digital humanities are around for about 60 years and
core areas such as computational linguistics have established themselves through
their own courses of study, digital humanities is still a young field of research that
has only recently become an independent discipline (Rehbein & Sahle, 2013). This is
the third core element of many definitions of digital humanities: Digital humanities
is a young field whose own definition is constantly being developed (What is Digital
Humanities? 2021).
Nevertheless, a broad variety of research programs and projects, conferences
and competence centers, stress the benefits of using computers as tools as well as
From Digital Humanities to Quantum Humanities … 5
The description of the potentials outlined above are from an abstract point of view
and must be considered much more comprehensively in its complexity, taking into
account various factors from related issues of quantum computing as well as from
digital humanities.
From a hardware perspective, the potentials stated above apply to an “ideal”
quantum computer, which is not yet available. Today’s quantum computers are error-
prone and have only limited capabilities (Preskill, 2018; Leymann & Barzen, 2020a).
They provide only a small number of qubits (LaRose, 2019) allowing only a limited
set of input data to be represented within the quantum computer. Besides this, noise
affects calculations on quantum computers (Knill, 2007; Preskill, 2018): The states
of the qubits are only stable for a certain amount of time—an effect referred to
as decoherence—due to unintended interactions between the qubits and their envi-
ronment (Leymann et al., 2020; Nielsen & Chuang, 2002). In addition, the opera-
tions that manipulate the qubits are also error-prone. Nevertheless, so-called NISQ
(Noisy Intermediate-Scale Quantum) machines (Preskill, 2018) allow first applica-
tions which can be used for prototypical implementations addressing small real-world
problems. This can contribute to establish application knowledge—also in the field of
quantum humanities. Taking a closer look at the recently published quantum roadmap
of IBM (Gambetta, 2020) it emphasizes the importance of starting to build application
knowledge now: Here IBM outlines its roadmap to 1 million error-corrected qubits
within a decade, providing more than 1,000 qubits in 2023, continuously increasing
the number of qubits for the coming years.
Next to the issues related to the hardware of quantum computers, from the software
perspective too several questions need to be addressed when working with quantum
computers. Even though quantum computers are becoming commercially available
(e.g. IBM, D-Wave, Rigetti, Honeywell, IONQ), questions regarding usability and
accessibility must be considered for each vendor. Also, more concrete questions like
the following need to be taken into account: How to encode your data properly based
on the processing demands of the chosen quantum algorithm (Weigold et al., 2021)?
How to expand oracles that many algorithms contain to keep the influence of such
an expansion on the number of qubits and operations required small? How to cope
with readout errors? (Leymann & Barzen, 2020a).
It is also necessary—even if the quantum computer is superior to classical
computers in certain areas as stated above—to address several open questions from
the application side. In a first step those existing and new problems from the digital
humanities need to be determined that are suitable to be considered being solved
8 J. Barzen
In recent years, more and more research projects in the field of digital humanities
can be identified that are based on data and data analysis to provide new insights
into questions stated in the humanities. For example, when taking a closer look at the
books of abstracts of the DHd (Digital Humanites im deutschspachigen Raum (DHd,
2021)) conferences—as a well-established conference in the digital humanities—
over the last three years (Sahle, 2019; Schöch, 2020; Vogeler, 2018) the term data
(and its German translation, as well as all compound words containing the term)
occurs more than 5,600 times and hint to the significance of data in digital humanities
research. Also, in analogy, terms hinting to analysing this data via techniques from
machine learning (e.g. searching for: machine learning, artificial intelligence, un-
/supervised learning, clustering and classification) has about 1,000 counts, giving
an impression of the importance of this approach to analyse data gained during the
last years. Since quantum computers are expected to have substantial superiority
especially in the field of machine learning (Schuld & Petruccione, 2018), the use
case presented in the following is positioned in this area. Here, data from the digital
humanities project MUSE is analysed with the of help machine learning and quantum
machine learning techniques.
The project MUSE (2021) aims at identifying a pattern language for costumes in
films. It contains the method as well as a supporting toolchain to capture all relevant
information about costumes, to analyse them and to abstract the significant infor-
mation into costume patterns. Costume patterns capture those significant elements
of a costume (e.g. colour, material, way of wearing) that are used to communicate a
certain stereotype or a character trait, for example. When speaking of a “nerd” or a
“wild west outlaw” most recipients do have a rather clear idea of what these stereo-
types should look like to be recognized as “nerd” or “wild west outlaw” based on
From Digital Humanities to Quantum Humanities … 9
Step 3: Capturing detailed information about the costumes of the initial film corpus.
To support capturing of all relevant parameters, a repository was designed and imple-
mented: The MUSE-repository (an open-source implementation is available under
(MUSE GitHub, 2021)) is based on the ontology described in Sect. 3.1 and assists to
collect all information about the films of the corpus, the characters and the costumes.
Currently (March 2021), the data set contains 5,231 costumes described from 58
films. These costumes contain 28,798 base elements and 63,711 primitives, 159,360
selections of colours and 180,110 selections of materials.
Step 5: Abstracting the results of the analysis step 4 into costume patterns. Patterns
in the tradition of Alexander et al. (1977) are documents that follow a pre-defined
format to capture knowledge about proven solutions in an abstract way to make this
knowledge easily accessible and reusable for other applications. Patterns are related
to each other and compose a pattern language based on these relations. As stated
above, costume patterns capture the proven solutions about how costume designers
address the problem to communicate a certain stereotype like a “wild west sheriff” in
terms of all the significant elements of the costume. This contains e.g. base elements,
primitives, their relations, colours, ways of wearing, material, if they proved to be
significant in the analyse step. As the costume patterns are part of a costume pattern
From Digital Humanities to Quantum Humanities … 11
4 Analysing Data
Analysing data has two major purposes: discovery and prediction. The collection of
techniques focusing primarily on the first purpose is called data mining, and the set
of techniques focusing primarily on the second purpose is called machine learning.
Despite their different primary purposes, data mining and machine learning have a
large overlap in terms of the methods they use: Optimization and statistics are core of
both approaches, and data mining is even using selective techniques from machine
learning. This is why both disciplines are often subsumed by the term data science.
Another difference is how the suitability or appropriateness of a data mining or a
machine learning solution is assessed: The appropriateness of a data mining solution
is assessed by its capability to discover previously unknown facts, while the appropri-
ateness of a machine learning solution is assessed by correctly reproducing already
known knowledge. Once a machine learning solution reproduced known knowledge
the solution is considered to correctly deliver previously unknown knowledge.
Despite these differences, the development of a solution based on either of both
disciplines follows the same overall procedure which is described in the next section.
In the past, we used data mining technologies to analyse data about costumes in
films (see Sect. 3.4, for more details; see Falkenthal et al., 2016a; Barzen, 2018).
The general procedure shown in Fig. 1 has been used for this analysis, but it is also
applicable in analysing data with machine learning techniques (Sect. 5).
Data cleansing (Skiena, 2017) encompasses activities like data acquisition, format
conversion and so on. Data preparation (Skiena, 2017) deals with a proper treatment
and encoding of data, and feature engineering. Algorithm selection (Kubat, 2017)
determines the family of algorithms (e.g. classification, clustering), selects a proper
family member, and choses hyperparameters of the selected algorithm. Result eval-
uation is often based on visualizing the results (Skiena, 2017) and the creation of
hypotheses, which are finally validated. Because these last steps are critical for the
success of a data analysis project they are depicted as separate steps in Fig. 1. If no
hypothesis can be built or successfully validated, other algorithms may be tried out to
finally succeed with at least one proper hypothesis. Successfully validated hypotheses
trigger further processing towards the final goal of the project, like finding patterns
(see Sect. 4.3).
Note that within a data analysis project, most time is typically spend in data
cleansing and data preparation (Skiena, 2017). Also, several algorithms, even several
algorithms of the same family of algorithms, are typically applied within an overall
data analysis project. And each algorithm applied involves performing several steps
of the process shown in Fig. 1. For example, in order to determine patterns of costumes
in films in MUSE, the mostly categorial data must be prepared, features must be engi-
neered, and the resulting data must be clustered (see the black shapes in Fig. 2). After
clusters have been determined, data about newly captured clothes can be classified
to identify which costume the clothes represents.
Each of these tasks can be achieved by a variety of algorithms (grey shapes below
black shapes). Data preparation, for example, involves one-hot encoding to turn
categorial data into binary vectors, or the set of data points representing clothes may
be turned into a distance space based on using Wu-Palmer similarity (see Sect. 4.2).
Feature engineering may be performed by training an autoencoder (see Sect. 5.5) or
multi-dimensional scaling after the transformation of the data into a distance space.
Data Feature
Clustering Classification
Preparation Engineering
Distance Space
Autoencoder MaxCut SVM
Computation
Multi-Dim
OPTICS
Scaling
Fig. 2 Data analysis pipeline for determining patterns and succeeding classification
From Digital Humanities to Quantum Humanities … 13
Clustering may be based on solving the maximum cut problem (see Sect. 6.2), training
a corresponding Boltzmann machine (see Sect. 5.4), or using the k-means algorithm
(Arthur & Vassilvitskii, 2007). Classification can be done by means of a Boltzmann
machine too, or by using a SVM.
Furthermore, each of the tasks depicted as black shapes in Fig. 2 is in fact
performed as a process chain (see the small process chains refining two of the
black shapes). For example, feature engineering requires to select a corresponding
algorithm, e.g. multi-dimensional scaling and setting proper hyperparameters (see
Sect. 7.3). The corresponding embedding needs the data in a certain format which
has to be prepared. After executing the algorithm, the resulting data representation
must be evaluated; if it is not appropriate, the hyperparameters must be adapted, and
the algorithms must be run again. If the results based on this algorithm is not appro-
priate at all, another algorithm (e.g. principal component analysis (PCA)) must be
tried out. Once the features are properly engineered, clustering is performed, which
means that the corresponding process chain is executed. Thus, a pipeline results that
consists of a sequence of several process chains.
Many data in the Humanities are categorical data, i.e. data with a finite set of
values, often strings, that cannot be used for calculations. For example, computing
the maximum of a set of textures or the average of jobs is meaningless. But most
machine learning algorithms presume numerical data or even metrical data. Thus,
categorial data must be transformed into metrical data to be able to be processed by
such algorithms.
Barzen et al. (2021) discussed a mechanism of how to turn tuples of categorical
data the domains of which are tree structured into an approximately equivalent set
of vectors in Rn , i.e. into a set of metrical data. Our mechanism is (i) based in the
Wu-Palmer similarity (Wu & Palmer, 1994), (ii) the fact that similarities can be
turned into distance measures, and (iii) that a finite set with a distance measure can
be embedded into an appropriate Rn by means of multidimensional scaling (MDS)
(Cox & Cox, 2001): see Barzen et al. (2021) for all the details.
Data elements the values of which are given by means of a taxonomy are such data
with a tree structured domain. In our application area of costumes in films (Barzen,
2018), most data types have domains defined by taxonomies. Thus, based on our
mechanism this data can be embedded into a vector space and, consequently, can be
processed by machine learning algorithms.
Figure 3 shows a small fraction of one of the taxonomies of our costume data set.
To compute the similarity of two nodes in a tree (e.g. a “swimming shorts" and a
“cycling shorts” in the figure) the lowest common ancestor of these two nodes must
be determined: in our example, this is the node “short pants”. Next, the length of the
paths (in terms of number of edges) from each of these nodes to their lowest common
ancestor must be determined: in our example, these lengths are L 1 = 2 and L 2 = 1.
14 J. Barzen
L1 = 2
L3 = 4
L2 = 1
Then, the length of the path from their lowest common ancestor to the root of the tree
must be determined: in our example, this length is L 3 = 4. Finally, the Wu-Palmer
similarity is defined as
2L 3 8 8
ω(swimming shorts, cycling shorts) = = = ≈ 0.72.
L 1 + L 2 + 2L 3 2+1+8 11
The similarity between two nodes is between 0 and 1, i.e. “swimming shorts” and
“cycling shorts” are quite similar. A first quick overview of the similarities between
the data elements of a data set can be given by a similarity matrix that presents the
similarities of pairs of data elements and may render these values by means of a
colour code (see Sect. 7.1 for an example).
The method to determine the similarity ω between two values of a tree structured
domain (part 1 in Fig. 4) can be extended to determine the similarity of two sets of
values of the same tree structured domain: in a nutshell, the similarities ω of all pairs is
determined whose first component is from the first set and whose second component
is from the second set; a certain mean value of these similarities is computed, resulting
in a similarity measure σ between two sets (part 2 of Fig. 4). Based on the similarity
measure σ the similarity of tuples of the same dimension can be determined whose
components are sets of values of the same tree structured domain (part 3 in Fig. 4): the
similarities between each components of the tuples is determined (e.g. the similarity
of the colours of two tuples representing costume A and costume B, the similarity
of their materials and so on); the mean value of these similarities is computed which
results in a similarity measure μ between two tuples. See Barzen et al. (2021) for
the details.
From Digital Humanities to Quantum Humanities … 15
σ
σ
The similarity ω(i, j) between two nodes i and j (and the similarity σ (i, j) of
two sets or the similarity μ(i, j) between two tuples i and j, respectively) can be
transformed into the dissimilarity or distance δ(i, j) between i and j, respectively, by
means of
δ(i, j) = ω(i, i) + ω( j, j) − 2ω(i, j)
“n” and “p” are hyperparameters that must be chosen based on former experience.
Embeddings can be computed by multidimensional scaling (MDS) (Cox & Cox,
2001).
This way, a set of data points M with components whose domains are defined by
taxonomies can be transformed into a set of data points in Rn such that the distances
in M are approximately the distances in Rn . After this transformation, the data points
can be further processed by most machine learning algorithms. Our costume data set
is transformed in this manner (see Sect. 7.1) and is, thus, available to be analysed by
several machine learning algorithms.
The main purpose of our project MUSE is the creation of pattern languages of
costumes in films. Pattern languages can be built in several ways. Often, they are
built from experience: practitioners recognize that a particular problem occurred
repeatedly in the past perhaps in various contexts and that the solutions applied
to solve the problem are very similar. Thus, these solutions can be abstracted to
identify their underlying principles such that these principles can be applied to solve
that particular problem even in new contexts. This way, a new pattern has been
identified. Similarly, relations between patterns are established based on experience.
Those relations between patterns become links with an associated semantics.
What results is a pattern language: A pattern language is a collection of patterns
within a particular domain and their relations. From an abstract point of view, a
pattern language is a weighted, directed graph. The nodes of this graph are the
patterns, the edges are the links between the patterns, and the weights of the edges
are the semantics of the links.
In contrast to derive patterns from experience, our project MUSE strives towards
deriving patterns by means of analysing data. The method we developed in MUSE
is generally applicable, and its use has already been initially verified also in the
domain of music (Barzen et al., 2016) and is envisioned by other aspects of films (see
Falkenthal et al., 2016b). Applying the MUSE method in other areas of the humanities
successfully seems to be possible. After having used data mining techniques in our
method first, we are now exploiting machine learning techniques and the initial results
we get seem to be more promising than using mining techniques.
To identify a pattern language based on data analysis the process sketched in Fig. 1
in Sect. 4.1 must be extended (see Fig. 6). If the hypothesis validation step resulted
Pattern
Hypothesis Pattern Pattern Pattern
Candidate
Validation Verification Relations Language
Identification
in a verified hypothesis, this hypothesis must be turned into a pattern document. That
is the various sections of the pattern document corresponding to the hypothesis must
be authored. Thus, a pattern candidate is created: the resulting document is “only”
a candidate but not yet a pattern because its general applicability still needs to be
verified. Such a verification is performed in the succeeding pattern verification step.
Reiners (2014) proposes a system based on which a community e.g. can discuss
the candidate, exchange experiences with it, and can jointly decide to promote the
candidate to a pattern. The new pattern must be related to the other patterns of the
domain in the pattern relations step. This way a pattern language results. Note that
a pattern language is a “living” entity, i.e. over the time patterns will be added to
the pattern language. Typically, a first version of a pattern language will be made
available or published when “major” problems of the underlying domain can be
solved with it.
An artificial neural network (or neural network for short) is a mathematical model that
mimics the biological neural networks of brains of animals. Such a neural network
is represented as a directed graph the nodes of which abstract neurons and the edges
abstract synapses of a brain. In this section, neurons and neural networks are defined,
perceptrons and their use for classification are sketched, and restricted Boltzmann
machines, as well as autoencoders are described.
5.1 Neurons
The first model of an artificial neuron roots back to McCulloch and Pitts (1943),
nowadays called a McCulloch-Pitts neuron. Such a neuron accepts binary data as
input and produces a binary output if a predefined threshold is exceeded. This model
was extended and called perceptron (see Sect. 5.3) in Rosenblatt (1958). Today a
neuron is abstracted as a mathematical function (Zurada, 1992): it receives multiple
values as input and produces a single value as output. What characterizes functions
that represent neurons is the way in which the output of a neuron is computed, i.e.
the ingredients of such functions and how they interact.
Definition Let w1 , . . . , wn ∈ R be real numbers called weights; the function
n
: Rn → R, x → wi xi
i=1
n
ν : Rn → R, x → α wi xi .
i=1
Figure 7 shows a graphical representation of an artificial neuron: the left side of
the figure represents the details of the definition before, the right side of the figure
depicts a more condensed version of this graphical representation in which the two
nodes representing Π and α are combined into a single node denoted with Π|α. The
input values x1 , . . . , xn of the neuron ν are connected via directed edges with the
neuron’s propagation function Π; the weight wi of the input x i is shown as a number
associated with the corresponding edge. The purpose of the propagation function Π
is to aggregate the input of the neuron considering the corresponding weights of the
input values. The purpose of the activation function α is (i) to decide whether its
input suffice to pass on a non-zero output and (ii) what the output value will be.
An important variant defines a neuron as a map ν : {1} × Rn ⊆ Rn+1 → R, with
n
ν(1, x1 , . . . , xn ) = α w0 + wi xi .
i=1
Thus, a neuron has one input x 0 of constant value “1” (see Fig. 8). The corre-
sponding weight w0 is called bias, denoted as “b”. The bias influences the firing of
the neuron: a negative bias results in the neuron firing less frequently, a positive bias
results in the neuron firing more frequently.
One problem with realizing neurons on a quantum computer is that quantum algo-
rithms are unitary transformation, thus, linear, but activation functions of neurons,
are typically non-linear. Cao et al. (2017) and Wan et al. (2017) present quantum
neurons with different activation functions. A special neuron (a so-called percep-
tron—see Sect. 5.3) on a quantum computer of the IBM Quantum Experience family
has been described in Tacchino et al. (2018). This implementation proves the expo-
nential advantage of using quantum computers in terms of requiring only N qubits
in order to represent input bit-strings of 2N bits. But in order to prepare this input an
exponential number of 1-qubit gates and CNOTs are needed, in general (Leymann &
Barzen, 2020a).
Neural networks are directed graphs the nodes of which are neurons (Kubat, 2017;
Skiena, 2017). The edges of the graph specify which neuron passes its output as
input to which other neuron. The neurons are partitioned into disjoint sets L1 ,…, Ln
(so-called layers) such that neurons of layer Li are connected with neurons of layer
Li+1 . The neurons of layer L1 get their input from the outside, i.e. not from other
neurons; this layer is called the input layer of the neural net. Neurons of layer Ln
pass their output to the outside, i.e. not to other neurons; this layer is called the output
layer of the neural network. All other layers are referred to as hidden layers. A neural
network with at least one hidden layer is sometimes called a deep neural network.
Definition A neural network is a tuple (N, E, L , X, Y) with the following
properties:
• G = (N, E) is a directed graph,
• L = {L1 ,…, Ln } ⊆ P(N) is a partition of the set of nodes, i.e. ∀i: Li = ∅, Li ∩ Lj
= ∅ for i = j, and ∪ Li = N; Li is called layer, L1 input layer, Ln output layer,
and Li (1 < i < n) hidden
layer,
• each node νi = w ji , i , αi ∈ N is a neuron with a set of weights {wji }, a
propagation function Π i , and an activation function αi ,
• X = {x1 ,…, xm } is the set of input values,
• Y = {y1 ,…, yk } is the set of output values,
• the set of edges E ⊆ (N × N) ∪ (X × N) ∪ (N × Y) connects two neurons, or an
input value and a neuron, or a neuron and an output value,
• for ν ∈ Li and (ν, ν’) ∈ E ∩ (N × N) it is ν’ ∈ Li+1 ,
• for ν ∈ Li , x ∈ X, and (x, ν) ∈ E it is i = 1,
• for ν ∈ Li , y ∈ Y, and (ν, y) ∈ E it is i = n, and y is the output of ν,
• for (νj , νi ) ∈ E ∩ (N × N) the output of νj is passed to νi where it is processed by
νi ’s propagation function Π i weighted by wji ,
• for ν ∈ L1 and (x, ν) ∈ E, ν processes x by its propagation function Π weighted
by wji .
20 J. Barzen
Effectively, a neural network that consumes m input values and that produces k
output values is a function Rm → Rk . The computational power of neural networks
stems from the fact that any continuous function on a compact set K ⊇ Rm → Rk
can be approximated with arbitrary precision by a neural network (see Kidger and
Lyons (2020) for details).
Figure 9 shows the principle structure of a neural network according to our defi-
nition before. Note that there are some variants of our definition: sometimes, edges
between neurons with the same layer are supported, sometimes edges are allowed
between neurons of any layer, sometimes a neuron may be connected to itself, and
so an. However, such variations are for use in special cases.
The structure of a neural network, i.e. its set of layers and its edges, is referred
to as its topology. Defining the topology of a neural network, specifying the activa-
tion functions and propagation functions of its neurons is called modelling a neural
network. Note that most often the propagation function from our definition in Sect. 5.1
is used, but activation functions change across problem domains; even different
activation functions for different neurons of a given neural network may be defined.
Training a neural network consists of choosing appropriate weights and biases. For
this purpose, the neural network
is considered as a function F : Rm → Rk . Further-
more, a set of input vectors xj ∈ R is given together with the corresponding
m
known correct output vectors y j ∈ Rk ; note that the set of pairs (x j , y j ) is
called training data. Ideally, the neural network will output yj for each input xj ; but
the neural network is not (yet) realizing the ideal function, i.e. it is just approximating
it: it will output Y j = F(x j ). Thus, the neural network will produce an error |y j −Y j |
in computing yj . The goal of training the neural network is to minimize these errors
considering all training data by adapting the weights and biases.
For this purpose, the so-called loss function
2 2
L(w, b) = yj − Yj = y j − F(x j )
j j
is minimized (as usual, the square of errors is used instead of their modulus). Here,
“w” denotes all weights of the neural network and “b” all its biases. Training now
means to choose w and b such that L(w, b) is minimized. Lots of mathematical proce-
dures are known to minimize L, e.g. gradient-based methods (e.g. stochastic gradient
descent) or derivative-free methods (e.g. Nelder-Mead) can be chosen (Nocedal &
Wright, 2006). Note that gradient-based methods require differentiability of the loss
function which in turn requires differentiability of F which in turn requires differ-
entiability of the activation functions of the neurons composing the neural network
implementing F (see Sect. 5.1). After this training phase, the neural network is ready
for use. It is said that the neural network has learned. The use of training data, i.e. input
values with corresponding known output values, is referred to as supervised learning:
the processing of the input data by the neural network is supervised by comparing
its output with the known given results associated with the input. Learning without
supervision is referred to as unsupervised learning: this kind of learning is discussed
in Sect. 5.5 in the context of autoencoders.
An implementation of quantum neural network that requires a single qubit for
each neuron (plus additional ancillae) has been proposed by Cao et al. (2017). The
training of this quantum neural network is performed in a hybrid quantum–classical
manner, i.e. the optimization is executed by classical software using the Nelder-
Mead algorithm. In contrast to training classical neural networks where the individual
training data is processed sequentially, the training of this quantum neural network
can be done based on a superposition of the input/output pairs of the training data—
something impossible for classical neural networks. A quantum neural network has
been realized on a near-term quantum processor as described by (Farhi & Neven,
2018); their neural network has been successfully used as a binary classifier on the
MNIST data set to distinguish two different handwritten digits. A set of requirements
on implementations of quantum algorithms that represent “meaningful” quantum
neural networks has been posed in Schuld et al. (2014).
5.3 Perceptrons
A perceptron can decide whether a given data point represented as input is left or
right of a given straight line in the plane (or in higher dimensions, right or left of a
hyperplane). Thus, a perceptron can be used as a binary classifier of linear separable
22 J. Barzen
H H
x
data sets: the left side of Fig. 10 shows two kinds of data points (grey points and
black points) in the plane. They are linear separable because a straight line H can be
found that splits these two data sets such that on one side of H only the black data
points are located, on the other side only grey data points reside.
Any straight line H (or hyperplane in higher dimensions) can be defined by means
of a vector w that is orthogonal to H, and whose length w = θ is the distance of H
from the origin (so-called Hesse normal form—see the right side of the figure). For
any point x, the scalar product w, x is the length of the projection of x onto w, and
w, x · x itself is the projection of x onto the straight line defined by w. Thus, a point
x is on H if and only if w, x − θ = 0. Furthermore, x is “on the right side” of H if
the length w, x of the projection of x onto w is greater than θ (i.e. w, x − θ > 0),
and x is “on the left side” of H if the length w, x of the projection of x onto w is
less than θ (i.e. w, x − θ < 0). Consequently, the value of w, x − θ determines
whether a point x is right or left of the hyperplane H defined by w, i.e. it serves as
is a binary classifier: a positive value classifies x as a black point, a negative value
classifies x as a gray point.
Definition A perceptron is a neuron ν with Heaviside-function θ as activation
function, where θ := (w1 , . . . , wn ) for the weights w1 , . . . , wn of the neuron.
The vector w = (w1 , . . . , wn ) of the weights of the perceptron ν determines a
hyperplane H (with w = θ ). When a point x = (x1 , . . . , xn ) is passed as input to
the propagation function Π of ν, the scalar product of w and x results:
n
(x) = wi xi = w, x.
i=1
With
From Digital Humanities to Quantum Humanities … 23
n
ν(x) = θ wi xi ∈ {0, 1},
i=1
1
m
e(τ ) = |dk − yk (τ )|,
m k=1
i.e. e(τ ) < γ , where γ is a predetermined error threshold. If the error is below this
threshold the algorithms stops.
• Otherwise, a new bias and new weights are determined: chose an arbitrary data
point t j (i.e. 1 ≤ j ≤ m) and compute
24 J. Barzen
wi (τ + 1) = wi (τ ) + r · c j − y j (τ ) · xi j ,
H1 H2
H
H3
Fig. 11 Computing separating hyperplanes via perceptrons and support vector machines, and using
a computed hyperplane as classifier for new data
From Digital Humanities to Quantum Humanities … 25
black point. Furthermore, support vector machines can even be used to classify non-
linear separable data sets (Bennett & Campbell, 2000) by embedding the data into a
high-dimensional vector space (which already hints that quantum implementations
of support vector machines are advantageous because the state spaces of quantum
computers are extremely huge).
Wiebe et al. (2016) proposed two quantum perceptrons: the first one can be trained
with a quadratic speedup in size of the training data, and the second one quadratically
reduces the training time required to achieve a given precision when compared with
the algorithm presented in this section. A quantum perceptron has been suggested in
Tacchino et al. (2018) that requires exponentially less storage resources for training;
it has been implemented and validated on a near-term quantum computer of the IBM
Q family of systems.
x2
y1
x3 .
.
. .
.
yk
.
xm
Input Visible Hidden Output
Values Layer {vi} Layer {hj} Values
26 J. Barzen
Fig. 13 Using restricted Boltzmann machines for classification and feature learning
From Digital Humanities to Quantum Humanities … 27
the input and the output are the features learned. Typically, these features are used
by another algorithm for further processing, e.g. by another restricted Boltzmann
machine (so-called “stacked” restricted Boltzmann machines).
A special kind of quantum restricted Boltzmann machines that can approximate
any quantum circuit has been specified in Wu et al. (2020). Also, an implementation of
such kind of a quantum restricted Boltzmann machine on a NISQ device of the IBM Q
family has been provided. Quantum Boltzmann machines (i.e. a general Boltzmann
machine which supports also connections between neurons of the same layer) as
well as quantum restricted Boltzmann machines have been investigated by Amin
et al. (2018) with the result that their quantum Boltzmann machine outperforms a
classical Boltzmann machine on a specific learning task. A circuit realizing a quantum
restricted Boltzmann machine is given in Zhang and Zhou (2015). No implementation
on a real quantum computer has been evaluated but a simulation: it turned out that
the quantum restricted Boltzmann machine performed a classification task faster and
with higher precision than a classical machine. An implementation of a restricted
Boltzmann machine to be realized on a D-Wave quantum annealer is described in
Denil and Freitas (2011); obstructions for a real implementation are discussed.
5.5 Autoencoders
The encoder transforms its input into a code (also called a representation of the
input). If the input is a data point in Rm and the code is a data point in Rk , the encoder
is a map ϕ : Rm → Rk . Because the output layer must have the same number of
neurons as the input layer, the output is again a data point in Rm , i.e. the decoder is a
map ψ : Rk → Rm . By definition, it is ψ ◦ϕ ≈ id. That is, for any input x, the output
is ψ ◦ ϕ(x) ≈ id(x) = x; thus, the difference (or error) between x and ψ ◦ ϕ(x)
square of the error (x − ψ ◦ ϕ(x)) is
2
should be as small as possible (as usual kthe
taken). For a set of training data T = x , the sum of these squared errors should
be minimal. Now, ϕ and ψ perform their mapping based on their weights wij and wi j
as well as their biases bi and bi , respectively. Consequently, the sum of the squared
errors is a function L (“loss function”) of these weights and biases:
2
L(w, b) = xk − ψ ◦ ϕ xk .
k
data points intact. While some other algorithms that might be used to process data
after dimension reduction require similar distances between the high-dimensional
data points and the low-dimensional data points, some other do not consider these
distances. Thus, the next algorithm used to process the data influences the decision
whether an embedding has to be used or whether an encoder suffice.
Based on the work of Lamata et al. (2019), a quantum autoencoder requiring
for each neuron a separate qubit has been worked out in Bondarenko and Feld-
mann (2020). Each neuron has a unitary operation associated with it that operates
on the neuron and all neurons of the preceding layer. The number of such operations
needed to apply the autoencoder grows exponentially with the maximum number of
neurons per layer. Training of their autoencoder has been simulated in MATLAB. An
implementation of a (small) autoencoder on a Rigetti Quantum Computer has been
described in Ding et al. (2019); the corresponding circuits are provided. This autoen-
coder was successfully used for dimension reduction. Similar has been achieved in
Pepper et al. (2019) based on a photonic quantum computer. Romero et al. (2017)
described a variational quantum–classical way to implement a quantum autoencoder.
Figure 15 depicts (for illustration only, not for reading) our implementation of the
quantum part of the algorithm based on the simulator of IBM Q und Qiskit (2021).
The input data is prepared in four qubits that are transformed by the encoder. Once
the encoder finished its processing, the first two qubits are set to |0 while the last
two qubits are kept in the state the encoder produced. Effectively, the code of the
autoencoder consists of these two last qubits, i.e. a dimension reduction by 50% is
achieved. Next, the decoder tries to reconstruct the original input from the code. The
decoding result is measured and analysed by a classical part (not shown in the figure);
based on this analysis another iteration may be required. Once the autoencoder is
successfully trained, the encoder can be used for dimension reduction.
Some quantum algorithms consist of both, a pure quantum part and a pure classical
part (e.g. Shor’s algorithm for factorization): its quantum part is performed on a
quantum computer, and its classical part is performed on a classical computer. That
is, such an algorithm inherently requires classical pre- or post-processing to achieve
its goal, it would not succeed without its classical part. Such a split algorithm is
referred to as a hybrid quantum–classical algorithm.
Instead of such algorithms that are inherently hybrid, algorithms might be
designed to be hybrid from the outset to limit their processing on a quantum computer.
For example, this is enforced by today’s near-term quantum devices that restrict the
amount of processing that can be performed with sufficient precision. Such an algo-
rithm requires additional classical processing to compensate for the restricted amount
of work performed by the quantum computer: the algorithm is not inherently hybrid
quantum–classical but it is so by design.
Two main problem domains are addressed by such kind of hybrid algorithms by
design: problems that at their core can be solved based on the Raleigh-Ritz principle
(Yuan et al., 2019), and problems that can be reduced to combinatorial optimiza-
tion problems. Both kind of algorithms make use of a quantum part that consists
of parameterized quantum circuits that prepare a state on a quantum computer and
measure it, and a classical part that optimizes the corresponding parameters in itera-
tions in dependence of the measured results (see Fig. 16). The parameterized unitary
operator U ( p1 , ..., pk ) in Fig. 16 is called an “ansatz”: its goal is to prepare states that
deliver measurement results m that are good candidates for arguments optimizing a
given function F. Thus, the value F(m) is dependent on the parameters p1 , ..., pk . By
The maximum cut problem is to partition the node set of a graph in two sets such
that the number of edges between nodes of the different sets is maximal. In Fig. 17,
part 1, a graph with four edges {A,B,C,D} is shown where each node is connected
to each other node. Part 2 and part 3 of the figure depict two different maximum
cuts: the members of one node set are coloured black while the members of the other
node set are left white. Dashed edges connect nodes of the same set, i.e. they do not
contribute to the cut, while solid lines connect nodes of the two different set, i.e. they
contribute to the cut. The number of edges of each cut is four.
Part 4 of the figure adds a weight to each edge, e.g. the weight of an edge indicates
the distance between the nodes connected by the edge; such a distance, for example,
might be determined by means of a distance measure (see Sect. 4.2). The weighted
maximum cut problem strives towards finding a partition of the node set such that
the sum of the weights of the edges between nodes of different sets is maximal. Note
that by setting each weight to “1”, the maximum cut problem is seen to be a special
case of the weighted maximum cut problem. Part (4) shows a weighted maximum
cut: the node set is partitioned into {A,B,C} and {D}, and the sum of the weights of
the corresponding edges is 20. If the weights are distances, the weighted maximum
cut is nicely interpreted as determining clusters: the two node sets consist of nodes
that are close to each other within each set, but nodes of different sets are far apart.
For example, D has distance 6, 9, 5 to A, B, C, respectively, but A, B, C are quite
close to each other (distances 1, 1, 2). Thus, an algorithm for determining a weighted
maximum cut can be used for clustering.
Let G = (N, E) be a graph with n nodes (i.e. card(N) = n), and let S, T be a cut of
N (i.e. S ∩ T = ∅, S ∪ T = N, S = ∅, and T = ∅). The function z : N → {−1, +1}
indicates membership to S or T, i.e. z(u) = +1 ⇐⇒ u ∈ S and z(u) = −1 ⇐⇒
u ∈ T (elements of S are labelled +1, elements of T are labelled −1—or “black”
and “white” in the figure). For brevity, zu : = z(u).
For i, j ∈ S, it is zi = zj = + 1, thus zi · zj = 1; and for i, j ∈ T, it is zi = zj −1,
thus zi · zj = 1. Consequently, it is i, j ∈ S ∨ i, j ∈ T ⇔ zi · zj = 1, i.e. (i, j) is not a
cut edge. Similarly, (i ∈ S ∧ j ∈ T) ∨ (j ∈ S ∧ i ∈ T) ⇔ zi zj = − 1, i.e. (i, j) is a cut
edge. Reformulation results in:
• (i, j) ∈ E is not a cut edge ⇔ 1 − zi zj = 0 ⇔ ½ · (1 − zi zj ) = 0
• (i, j) ∈ E is a cut edge ⇔ 1 − zi zj = 2 ⇔ ½ · (1 − zi zj ) = 1
Thus, a maximum cut maximizes the following cost function:
1
C(z) = 1 − zi z j .
2 (i, j)∈E
This is because each cut edge contributes a “1” to the sum, and a non-cut edge
contributes a “0”: the more cut edges exist, the higher is C(z). Such kind of a formula,
i.e. a formula with parameters with values from {−1, +1} is called an Ising formula,
which is important for solving quantum problems on so-called quantum annealers.
In this chapter it is assumed that algorithms are formulated in the gate model. For
this purpose, parameters from {0, 1} are better suited, i.e. the above cost formula has
been transformed. To achieve this, a function x : N → {0, 1} indicates membership
to S or T, i.e. x(u) = 1 ⇐⇒ u ∈ S and x(u) = 0 ⇐⇒ u ∈ T . For brevity, x u : =
x(u). This implies:
• (i, j) ∈ E is not a cut edge ⇔ i, j ∈ S ∨ i, j ∈ T ⇔ xi · (1 − xj ) + xj · (1 − xi ) = 0
From Digital Humanities to Quantum Humanities … 33
1
C(x) = xi · (1 − x j ) + x j · (1 − xi ) .
2 (i, j)∈E
m
C(z) = Cα ,
α=1
6.3 QAOA
The concept of the quantum approximate optimization algorithm (QAOA) has been
introduced by Farhi et al. (2014). It is used to approximately solve combinatorial
optimization problems on near-term quantum devices and has been first applied to the
maximum cut problem in Farhi et al. (2014). Crooks (2018) provided a QAOA-based
implementation of a maximum cut algorithm on a near-term quantum computer of
Rigetti and showed that this implementation is faster than the well-known classical
Goemans–Williamson algorithm. The results can even be improved following the
warm starting idea for QAOA (Egger et al., 2020).
The basic idea is as follows: The cost function C of a combinatorial optimiza-
tion problem induces a map C : Hn → Hn by interpreting z ∈ {0, 1}n as the
computational basis vector |z ∈ Hn and defining
m
C|z := Cα (z) · |z = f (z) · |z .
α=1
Thus, the matrix of C in the computational basis is a diagonal matrix, i.e. the map
C is Hermitian and can be used to measure the states of a quantum computer (C is an
observable). Let z ∈ {0, 1}n such that f (z ) = max{f (z)} = C max . The expectation
value Cψ of measuring C in any state |ψ = x z |z is Cψ = ψ|Cψ =
34 J. Barzen
x z |z|x z f (z)|z = |x z |2 f (z) ≤ |x z |2 f (z ) = f (z )|x z |2 = f (z ) =
Cmax , i.e. the expectation value is less than or equal Cmax . If |ψ is close to |z , Cψ
will be close to Cmax . This means that with
|ψ = x z |z = x z |z + x z z ,
z∈{0,1}n z =z
Let G = (N, E) be a graph with n nodes (i.e. card(N) = n), and let S, T be a partition
of N. Such a partition can be described by a bit vector x ∈ {0, 1}n as follows:
xi = 1 ⇐⇒ i ∈ S and xi = 0 ⇐⇒ i ∈ T . There are 2n different bit vectors, i.e.
partitions. For each partition the number wd of edges between nodes of different sets
of the partition as well as the number ws of edges between nodes of the same set of the
partition is determined. Next, x ∈ {0, 1}n is considered as the binary representation
of a natural number x ∈ N, and wx := 21 (ws − wd ) is associated with this number x.
Finally, the matrix M ∈ C2 ×2 is defined where m i j = 0 ⇐⇒ i = j and m ii = wi ,
n n
i.e. the i-th row of M has as diagonal element the number wi corresponding to the
partition i ∈ {0, 1}n , and all other elements of the row are zero.
M is a diagonal matrix and, thus, Hermitian, i.e. all eigenvalues are real numbers.
Each vector of the computational basis is an eigenvector of M. It can be proven that
each vector of the computational basis that is an eigenvector of the lowest eigenvalue
defines a maximum cut. Thus, determining the lowest eigenvalue of the matrix M
and one of its eigenvectors means to determine a maximum cut of the corresponding
graph. Note that an analogue construction can be made for arbitrary weighted graphs.
Using eigenvalues to solve the maximum cut problem is based on Poljak and Rendl
(1995).
6.5 VQE
ψ|O|ψ|
λmin ≤ for any |ψ = 0.
ψ, ψ
With ψ|O|ψ = O|ψ for each state |ψ (i.e. ψ = 1), the expectation value
of O provides an upper bound of the lowest eigenvalue of O: λmin ≤ O|ψ . Thus, if
a state |ψ can be found that minimizes the expectation value of O, O|ψ is close to
the lowest eigenvalue of O. (Note that O|ψ = λmin holds for an eigenvector |ψ
of λmin ).
In order to find such a ψ̂ , series of states |ψ( p1 , ..., pk ) are iteratively
constructed that depend on parameters p1 , ..., pk ; in Fig. 16 the unitary operator
U ( p1 , ..., pk ) prepares this state. The expectation value O|ψ( p1 ,..., pk ) is measured
for each parameterized state, and the measurement result is subject to a clas-
sical optimization algorithm that determines new parameters p1 , ..., pk that reduce
the expectation value further; in Fig. 16 the function F(m) is this expectation
value. These iterations approximate |ψ , i.e. the final set of parameters determine
O|ψ( p1 ,..., pk ) ≈ λmin and |ψ( p1 , ..., pk ) is an approximation of the corresponding
eigenvector.
An efficient measurement of the expectation values is done as follows: Each
Hermitian operator O can be written as a linear combination of “simpler” Hermitian
operators Oi , i.e. O = xi Oi . Here, “simpler” means that Oi is a combination of
Pauli operators. The expectation value Oi v of such a combination can be efficiently
measured in a quantum computer (Hamamura & Imamichi, 2019), and it is Ov =
xi Oi v , i.e. the expectation value of O can be easily computed classically based
on the measured Oi v —i.e. in Fig. 16 the function F computes xi Oi . For all the
details see Peruzzo et al. (2014).
As described in Sects. 5 and 6 there are several machine learning techniques for which
first implementations are available on quantum computers. Some can be used to just
gain initial experiences, others already show improvements in terms of speedup or
precision. Bringing together the aspects and concepts outlined in the above sections,
QHAna (2021) was developed. QHAna is our quantum humanities data analysis
36 J. Barzen
tool that aims for several goals: (i) There is the content-related goal of supporting
the identification of patterns by data analysis. A feasibility study is provided by the
use case (Sect. 3) that focuses on analysing costume data to improve the under-
standing of vestimentary communication in films. (ii) By performing this analysis
by classical and hybrid quantum–classical algorithms a comparison of both methods
is supported, allowing to assess the benefits of using quantum algorithms. (iii) This
comparison supports the goal of improving the understanding of potential benefits
quantum computing may provide for digital humanities research (see Sect. 2.2). (iv)
Additionally, QHAna allows the integration of heterogeneous tools from different
quantum computing vendors in a single analysis pipeline. (v) Thus, one of the overar-
ching goals of QHAna is to provide easy access to people without quantum computing
background to gain first application knowledge in the domain of quantum humanities.
In the following QHAna will be introduced by this core objectives and functions.
The primary goal of QHAna is to support analysing data when aiming at the iden-
tification of patterns based on the data analysis pipeline defined in Sect. 4.1. By
providing easy access to the analysis methods and techniques outlined above, QHAna
is designed to e.g., improve the understanding of vestimentary communication in
films by supporting the identification of costume patterns in the MUSE dataset (see
Sect. 3.3). Note that QHAna is data independent and therefore, not limited to the
data of our use case. Data from other application areas like from the MUSE4Music
project (Barzen et al., 2016) is planned to be analysed.
Based on initial question like (i) “which base elements are used to communicate
specific stereotypes?”, (ii) “do certain age impression and attributes like colour or
condition communicate certain character traits within one genre?”, (iii) “and if this
is true, how to group new costumes to the suitable category when being captured?”
several analysis steps need to be performed. The upper part of Fig. 18 gives an
impression of the main menu that guides through the pipeline for analysing data. To
approach an answer to these sample questions, for example, question (ii) is addressed
by clustering algorithms, focusing on finding those costumes that achieve the same
effect, while the question (iii) is addressed by classification algorithms. For both, the
first step is about preparing the data depending on the requirements of the algorithms.
As this is the basis for all the other following steps, it will be described in more detail
how QHAna supports this step.
As Fig. 18 depicts, the tab “Data Preparation” allows to prepare the data depending
on the requirements of specific analysis algorithms. As described in Sect. 4.2, several
algorithms require numerical data and the MUSE data is mostly categorical data.
Therefore, the subtabs “Distance Space Calculation” and “One-Hot Encoding” as
shown in Fig. 18 provide different options to transform categorical data into numer-
ical ones. For our example we chose the “Distance Space Calculations” that supports
the approach described in Sect. 4.2: Here, a dropdown menu allows to choose the
From Digital Humanities to Quantum Humanities … 37
domain of the distance space. In our MUSE use case, the “Costume” data set is
chosen. Figure 18 depicts how to specify a view for analysing a defined subset of
the attributes of the MUSE data set. For this purpose, the attributes of interest of
our example question are chosen, e.g., “Dominant Color”, “Dominant Condition”,
“Dominant Character Trait”, “Dominant Age Impression”, and “Genre” are selected.
Furthermore, the user can define which element comparer (that allow to compare two
elements within an attribute category) and attribute comparer (that allow to compare
sets of elements within an attribute), or which aggregator type (that define how data
points are aggregated) and transformer type (that describe the function used to trans-
form similarity measures into distance measures), are most suitable for the use case,
as well as how empty attributes are to be treated (e.g. if a costume has no value for
the attribute category “Color” the missing element should be “ignored”). The table
at the top provides an overview of all selections made.
The tap “Documentation” (right of “Data Preparation”) allows to retrieve general
information about the currently processed data and its structure (e.g. attribute
lists, taxonomies), explanations about the components already implemented in
QHAna (e.g. different cluster or classification algorithms), references to software
dependencies as well as all relevant related papers used to implement the tool.
After computing the distance space, QHAna supports calculating corresponding
distance matrices (see Sect. 4.2). Here, the user can choose how many costumes
should be analysed and whether these costumes should be selected randomly from
the whole data set or custom specific. Figure 19 gives an example of such a distance
38 J. Barzen
Fig. 19 Screenshot of a
distance matrix
matrix of 25 costumes. Note that the visualization of the distance may become very
complex with an increasing number of input data and the resulting visualization may
become difficult to comprehend.
Distances between pairs of data elements (e.g. costumes) are represented by a
colour code: The distance value of two costumes becomes a coefficient of the distance
matrix and this value is turned into a colour: dark blue corresponds to no distance at
all, while dark red indicates maximum distance. Each costume is represented by a
number (0–24 in Fig. 19). As the distance matrix depicts, some of the costumes are
very close to each other based on the selected attributes and others are very far away
(i.e. they are not really similar). For example, costume 5 is very close to costume
9 and costume 15, while it has hardly any similarity with (i.e. it is far away from)
costume 11 and 22. For a more detailed comparison of these costumes the “Entity
Table” subtab of the “Overview” tab (see Fig. 20) allows to compare the costumes of
interest in detail by their attributes chosen. Figure 20 depicts costumes 5, 9, 11, 15
and 22 stressing that 5, 9 and 15 have the same attribute values, while costume 11 and
15 differ in their attribute values. As most costumes have more than just 5 attributes,
changing the input parameters in the costume distance space definition may have a
deep impact on the similarity measures of these costumes. Also, experts sometimes
need to verify the concrete costumes manually. For this purpose, the entity table
provides for each costume links to the MUSE-repository (the film, the character, the
concrete costume) allowing further investigation of the costumes with all the detailed
information, in particular it provides the visual representations of the costumes from
the MUSE-repository.
After the data is prepared the next step allows to—if needed—perform “Feature
Engineering” to reduce the number of relevant attributes. The tab “Feature Engi-
neering” provides several techniques for dimension reduction and feature extraction
such as autoencoders, embeddings, or PCA. This provides the basis to use clustering
From Digital Humanities to Quantum Humanities … 39
Entity Table
[5,9,11,15,22] Table Entities of List
['High school
15 Link Link Link ['Light Blue'] ['Ironed'] ['Active', 'Strong'] ['40s'] comedy',
'Love movie']
['Good', 'Passive', ['High school
22 Link Link Link ['Dark Brown'] ['Worn Out'] ['Youth']
Weak'] comedy']
techniques that allow to group those costumes together that have the same effect,
e.g. communicating a Sheriff by certain base elements like a Sheriff star, Cowboy
boots and so on (Barzen, 2018). As there are new costumes captured on a daily
basis and they need to be classified in terms of being mapped to the costume pattern
they contribute to, running classification algorithms is very promising for the MUSE
use case. All those steps supported by QHAna aim at improving the understanding
of vestimentary communication in films. How feature engineering, clustering, and
classification are supported is unfolded in what follows.
To get the optimal results when approaching an answer to a stated question the
different analysis steps are performed based on different algorithms, often in parallel
by different classical implementations as well as by different quantum–classical
implementations. This allows comparing the different results of classicals machine
learning algorithms amongst each other, comparing different results of quantum
machine learning algorithms amongst each other, and comparing results from classi-
cals machine learning algorithms with results from quantum machine learning algo-
rithms. This enables finally selecting the most suitable algorithm for the problem
at hand as well as improving the application of an algorithms when comparing the
results and optimizing iteratively the hyperparameters. Please note that due to the
limitations of today’s quantum computers only small sets of data can be processed
when using quantum machine learning algorithms.
40 J. Barzen
Fig. 21 Distance matrix as result of a hybrid autoencoder (left side) and computed via the Wu and
Palmer similarity (right side)
From Digital Humanities to Quantum Humanities … 41
results have been achieved by using these two different techniques: The statistical
approach (autoencoder) on the one hand and the approach using taxonomies (Wu-
Palmer similarities) on the other hand, identified the same costumes as being close
to each other. This hints to a potential costume pattern. Also, it provides first insights
on comparing a classical and a quantum-based approach to compute distance values.
Classical and Quantum-based Approaches to Clustering
Implementations of both, classical clustering algorithms and first quantum–clas-
sical clustering algorithms are available under the tab “Clustering” in QHAna.
Currently, this includes an implementation of the OPTICS cluster algorithm (Ankerst
et al., 1999), four different implementations of the maximum cut (a.k.a. maxcut)
algorithm (see Sect. 6.2) and four different implementations of the k-means algo-
rithm (Arthur & Vassilvitskii, 2007; Khan et al., 2019). The implementations of
the maximum cut range from a classical naive implementation (“classicNaiveMax-
Cut”), to two approximative implementations (“sdpMaxCut” based on semidefinite
programming (Boumal, 2015) and “bmMaxCut” using a Bureir-Monteiro solver
(Boumal et al., 2016)), to a quantum–classical implementation (“qaoaMaxCut”) that
is based on QAOA (see Sect. 6.3) and implemented in Qiskit (see Max Cut, 2021).
Figure 22 allows a comparison of the results of using different maximum cut
implementations. Input are the same 10 costumes as in Fig. 21, using the default
values of the hyperparameters per algorithm provided by QHAna. Diagram 1 of
the figure shows the costumes and their distances (upper left corner) in an MDS-
based embedding. The subtab “Embedding” of “Feature Engineering” supports to
use MDS as one implemented approach to map the distance measures of the chosen
costumes with their chosen attributes to a metrical feature space. As a result of the
embedding via MDS the data is mapped into Rn in such a way that the original
distances of the data elements are nearly the distances in the feature space (see
Sect. 4.2). The other three diagrams present clustering results achieved by different
maximum cut implementations namely “qaoaMaxCut” (Diagram 2), “sdpMaxCut”
(Diagram 3), and “bmMaxCut” (Diagram 4). It can be seen that the two approximative
implementations (Diagrams 3 and 4) have identical results: The first cluster (red
circles) contains the costumes 0–4, while the second cluster (blue circles) contains
the costumes 5–9. The result of the quantum–classical implementation (part 2) is
rather different: cluster one (red circles) contains the costumes 3, 4, 7, and 9 while
the second cluster (blue circles) contains the costumes 0, 1, 2, 5, 6, and 8. In providing
easy access to several implementation the comparison of such results is supported
by QHAna. In addition, the results can be manually verified using the entity table of
QHAna (see Fig. 20) as outlined in Sect. 7.2. The entity table allows the evaluation
of the attributes of the costumes that are part of the clusters and provides all the
details of the costumes by linking them to the MUSE-repository. This can be used
to improve the understanding of the benefits quantum computing can have, e.g. for
quantum humanities research.
42 J. Barzen
Fig. 22 Values of 10 costumes embedded via MDS (1) and their clustering results of different
maximum cut implementations (2–4)
In a fist attempt when taking a closer look at the comparison of Fig. 22 one might
decide for the classical implemented clustering for further analysis as it seems more
precise. But QHAna also allows to improve the results by adjusting the hyperpa-
rameters of the algorithms. For the results of Fig. 22 this would include to poten-
tially improve the results of the quantum–classical implementation of the algo-
rithm as shown in Fig. 22 to equal the results of the classical implementation.
Thus, QHAna lists the selectable parameters (a.k.a. hyperparameters) specific to
the chosen algorithms, together with descriptions that support selecting the right
parameters. Figure 23 gives an impression of the hyperparameters to be selected for
the quantum–classical implementation of a maximum cut algorithm including the
user token required to run the algorithm on the IBM Q systems.
Figure 24 outlines the influence of the maximum number of iterations performed
by the algorithm, while all other parameters keep the default settings of QHAna.
As can be seen, increasing the number of iterations seems to improve the results:
Diagram 1 is performed with only one iteration, while the number of maximum
From Digital Humanities to Quantum Humanities … 43
Fig. 24 Clustering results of the quantum–classical maximum cut implementations when changing
the number of maximum iterations performed
Fig. 25 Classification results of the classically implemented SVM (left) and of the QKE based
implemented SVM (right)
From Digital Humanities to Quantum Humanities … 45
results of the classically implemented SVM (Diagram 1) and of the QKE based SVM
(Diagram 2). Input for this example are 30 costumes and their distance values based
on the attributes stereotype and role relevance embedded by MDS, and classified as
“positive” or “negative” based on another training set for the classifier. As can be
seen in Fig. 25 these classes are not linear separable. Thus, finding the right classifier
is a challenging task. Figure 25 depicts that (by using the default parameters QHAna
provides) the result of the QKE variant (Diagram 2) is much more precise (accuracy
of 94% versus 89%) than the result of the classical variant (Diagram 1).
The prototype of QHAna has a modular structure. Thus, it can easily be extended by
further algorithm implementations. Various implementations are already available
(see Sect. 7.1–7.3). Because QHAna can also serve REST APIs the effective inte-
gration of components accessible via such APIs is straightforward; this way, further
implementations of techniques described in the theoretical sections of this contribu-
tion into QHAna is achieved: This includes—but is not restricted to—extending the
feature engineering step by quantum principal component analysis, for example, or
add quantum restricted Boltzmann machines, using more clustering algorithms and
integrating more classification algorithms. Also, algorithms used in other domains
(like HHL, VQE and so on) can be integrated into QHAna.
Another significant advantage of QHAna is that it provides comfortable access
to different backends to run the algorithms on. Figure 23 shows, for example, that
there are different simulators and IBM quantum computing backends integrated
and selectable already. Currently we are integrating more backends like PennyLane
(PennyLane, 2021) and TensorFlow.
Working with quantum computers still requires advanced knowledge in physics and
mathematics and is therefore a great challenge for research in the field of digital
humanities. As there are promising potentials in using the quantum computer in this
area (see Sect. 2.2), it is of interest to provide a simple, straightforward access to
this new technology without having to deal with all the mathematical, physical, and
algorithmic details. This is what QHAna aims for. As Fig. 23 depicts, even very
complex quantum–classical algorithms are provided in an abstract manner allowing
to adjust their hyperparameters, but default values of these parameters are set. Thus,
these algorithms and their benefits can be applied by someone not familiar with qubits,
QPU connectivity, quantum circuits, and so on. In Sects. 5 and 6 we provided the
core concepts to enable an understanding of the underlying ideas of the (classical and
quantum) algorithms used in QHAna, but no details about classical algorithms nor
46 J. Barzen
quantum algorithms are given. These algorithms and their implementations evolve
rapidly and potentially change quite frequently. Their details are not the focus of
QHAna. Instead, providing access to a complex new technology to learn about and
participate in the advantages given for further research is at the heart of QHAna.
domain can be seen (Leymann, 2019; Weigold et al., 2020, 2021). It is of particular
relevance to provide application knowledge in a field where the mathematical and
physical basics required to use this technology cannot be taken for granted, but fields,
which are essential for the critical reflection of digital methods, such as the digital
humanities. Especially, if a quantum computer is not only to be used as a tool like for
quantum machine learning but for thinking about totally new questions that have not
yet been tackled at all, perhaps not even identified or considered, knowledge about
the corresponding concepts, methods, and application potentials is key. To identify
and explore these possible new application areas is an exciting task and needs to
be started by building knowledge on how quantum computers can be applied to
problems stated in the humanities.
Acknowledgements I am very grateful to Frank Lemann for discussing several subjects of this
chapter. Also, I would like to thank Felix Truger, Philipp Wundrack, Marcel Messer, Daniel Fink
and Fabian Bühler for their valuable input and implementing several aspects of our use case.
This work was partially funded by the BMWi project PlanQK (01MK20005N) and the Terra
Incognita project Quantum Humanities funded by the University of Stuttgart.
References
Aharonov, D., van Dam, W., Kempe, J., Landau, Z., Lloyd, S., & Regev, O. (2008). Adiabatic
quantum computation is equivalent to standard quantum computation. SIAM Review, 50(4), 755–
787.
Alexander, C., Ishikawa, S., Silverstein, M., Jacobson, M., Fiksdahl-King, I., & Angel, S. (1977).
A pattern language: Towns, buildings, constructions. Oxford University Press.
Amin, M. A., Andriyash, E., Rolfe, J., Kulchytskyy., B., & Melko, R. (2018). Quantum Boltzmann
machine. Physical Review X, 8, 021050.
Ankerst, M., Breunig, M. M., Kriegel, H., & Sander, J. (1999). OPTICS: ordering points to identify
the clustering structure. In ACM SIGMOD international conference on Management of data.
ACM Press.
Anschuetz, E. R., Olson, J. P., Aspuru-Guzik, A., & Cao, Y. (2018). Variational quantum factoring.
arXiv:1808.08927
Arthur, D., & Vassilvitskii, S. (2007). “k-means++: The advantages of careful seeding”. In Proceed-
ings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms (pp. 1027–1035).
Society for Industrial and Applied Mathematics.
Barzen, J. (2013). Taxonomien kostümrelevanter Parameter: Annäherung an eine Ontologisierung
der Domäne des Filmkostüms. Technical Report No. 2013/04, Universität Stuttgart. https://www.
iaas.uni-stuttgart.de/publications/TR-2013-04_Taxonomien-_kostuemrelevanter_Parameter.pdf
Barzen, J. (2018). Wenn Kosüme sprechen—Musterforschung in den Digital Humanities am
Beispiel vestimentärer Kommunikation im Film. Dissertation, University of Cologne. https://
kups.ub.uni-koeln.de/9134/
Barzen, J., Breitenbücher, U., Eusterbrock, L., Falkenthal, M., Hentschel, F., & Leymann, F. (2016).
The vision for MUSE4Music. Applying the MUSE method in musicology. In: Computer science—
Research and development. Advancements of service computing: Proceedings of summer SoC
2016 (Vol. 32, pp. 3–4).
Barzen, J., Falkenthal, M., & Leymann, F. (2018). Wenn Kostüme sprechen könnten: MUSE—Ein
musterbasierter Ansatz an die vestimentäre Kommunikation im Film. In P. Bockwinkel, B. Nickel,
G. Viehhauser (Eds.), Digital humanities. Perspektiven der Praxis, Frank & Timme (pp. 223–241).
48 J. Barzen
Barzen, J., & Leymann, F. (2014). Costume languages as pattern languages. In Proceedings of
Pursuit of Pattern Languages for Societal Change—Preparatory Workshop.
Barzen, J., & Leymann, F. (2019). Quantum humanities: A vision for quantum computing in digital
humanities. In SICS software-intensive cyber-physical systems (pp. 1–6). https://doi.org/10.1007/
s00450-019-00419-4
Barzen, J., & Leymann, F. (2020). Quantum humanities: A first use case for quantum-ML in media
science. In ISAAI’19 Proceedings—Artificial Intelligence, Digitale Welt (Vol. 4(1), pp. 102–103).
Barzen, J., Leymann, F., Falkenthal, M., Vietz, D., Weder, B., & Wild, K. (2021). Relevance of
near-term quantum computing in the cloud: A humanities perspective. In Cloud Computing and
Services Science, Communications in Computer and Information Science. Springer (to appear).
Bennett, K. P., & Campbell, C. (2000). Support vector machines: Hype or Hallelujah? ACM
SIGKDD, 2(Issue 2).
Berry, D. M. (2012). Understanding digital humanities. Palgrave Macmillan.
Berry, D. M., & Fagerjord, A. (2017). Digital humanities: Knowledge and critique in a digital age.
Polity Press.
Bhaskar, M. K., Hadfield, S., Papageorgiou, A., & Petras, I. (2015). Quantum algorithms and circuits
for scientific computing. arXiv:1511.08253
Biamonte, J. (2020). Lectures on quantum tensor networks. arXiv:1912.10049
Bondarenko, D., & Feldmann, P. (2020). Quantum autoencoders to denoise quantum data. Physical
Review Letters, 124, 130502.
Boumal, N. (2015). A Riemannian low-rank method for optimization over semidefinite matrices
with block-diagonal constraints. arXiv:1506.00575
Boumal, N., Voroninski, V., & Bandeira, A. S. (2016). The non-convex Burer-Monteiro approach
works on smooth semidefinite programs. In Advances in Neural Information Processing Systems
29 (NIPS 2016).
Bravo-Prieto, C., LaRose, R., Cerezo, M., Subasi, Y., Cincio, L., & Coles, P. J. (2020). Variational
quantum linear solver. arXiv:1909.05820
Brown, K. (Ed.) (2020). The Routledge companion to digital humanities and art history. Routledge.
Burges, Ch. J. C. (1998). A tutorial on support vector machines for pattern recognition. Data Mining
and Knowledge Discovery, 2, 121–167.
Cao, Y., Guerreschi, G. G., & Aspuru-Guzik, A. (2017). Quantum neuron: an elementary building
block for machine learning on quantum computers. arXiv:1711.11240
Chen, G., & Srihari, S. H. (2015). Restricted Boltzmann machine for classification with hierarchical
correlated prior. arXiv:1406.3407v2
Cox, T. F., & Cox, M. A. A. (2001). Multidimensional scaling. Chapman and Hall.
Crooks, G. E. (2018). Performance of the quantum approximate optimization algorithm on the
maximum cut problem. arXiv:1811.08419v1
da Silva, I. N., Spatti, D. H., Flauzino, R. A., Liboni, L. H. B., & dos Reis Alves, S. F. (2017).
Artificial neural networks: A practical course. Springer International Publishing.
Denil, M., & de Freitas, N. (2011). Toward the implementation of a quantum RBM. Oxford University
Research Archive.
Deutsch, D. (1985). Quantum theory, the Church-Turing principle, and the universal quantum Turing
machine. Proceedings of the Royal Society of London, A400, 97117.
DHd. (2021). Digital Humanites im deutschspachingen Raum. https://dig-hum.de/
DHQ. (2021). Digital humanities quarterly. http://www.digitalhumanities.org/dhq/.
Dickel, C. (2018). A cloud quantum computer business plan. In: QuTech. http://blog.qutech.nl/
index.php/2018/07/18/a-cloud-quantum-computer-business-plan/
Ding, Y., Lamata, L., Sanz, M., Chen, X., & Solano, E. (2019). Experimental implementation of a
quantum autoencoder via quantum adders. Advanced Quantum Technologies, 1800065, 2019.
Dunjko, V., Taylor, J. M., & Briegel H. J. (2016). Quantum-enhanced machine learning. arXiv:
1610.08251
Egger, D. J., Marecek, J. & Woerner, S. (2020). Warm-starting quantum optimization. arXiv:2009.
10095
From Digital Humanities to Quantum Humanities … 49
Falkenthal, M., Barzen, J., Breitenbücher, U., Brügmann, S., Joos, D., Leymann, F., & Wurster, M.
(2016a). Pattern research in the digital humanities: How data mining techniques support the iden-
tification of costume patterns. In Computer Science—Research and Development. Advancements
of Service Computing: Proceedings of SummerSoC 2016 (Vol. 32, pp. 3–4).
Falkenthal, M., Barzen, J., Breitenbücher, U., Fehling, C., Leymann, F., Hadjakos, A., Hentschel,
F., & Schulze, H. (2016b). Leveraging pattern applications via pattern refinement. In Proceed-
ings of the International Conference on Pursuit of Pattern Languages for Societal Change
(PURPLSOC).
Falkenthal, M., Barzen, J., Dörner, S., Elkind, V., Fauser, J., Leymann, F, & Strehl, T. (2015).
Datenanalyse in den Digital Humanities—Eine Annäherung an Kostümmuster mittels OLAP
Cubes. In Datenbanksysteme für Business, Technologie und Web (BTW) (Vol. 16). Fachtagung
des GI-Fachbereichs “Datenbanken und Informationssysteme” (DBIS).
Farhi, E., Goldstone, J., & Gutmann, S. (2014). A quantum approximate optimization algorithm.
In MIT-CTP/4610.
Farhi, E., & Neven, H. (2018). Classification with quantum neural networks on near term processors.
arXiv:1802.06002v2
Fehling, C., Barzen, J., Falkenthal, M., & Leymann, F. (2014). PatternPedia—Collaborative Pattern
identification and authoring. In Proceedings of Pursuit of Pattern Languages for Societal
Change—Preparatory Workshop.
Gambetta, J. (2020). IBM’s roadmap for scaling quantum technology. In IBM Research Blog.
Retrieved September 15, 2020, from https://www.ibm.com/blogs/research/2020/09/ibm-qua
ntum-roadmap/
Grover, L. K. (1996). A fast quantum mechanical algorithm for database search. In Proceedings of
the twenty-eighth annual ACM symposium on Theory of computing (pp. 212–219).
Hamamura, I., & Imamichi, T. (2019). Efficient evaluation of quantum observables using entangled
measurements. arXiv:1909.09119v2
Haven, E., & Khrennikov, A. (2013). A Quantum social science. Cambridge University Press.
Havenstein, C., Thomas, D., & Chandrasekaran, S. (2018). Comparisons of performance between
quantum and classical machine learning. SMU Data Science Review, 1(4), Article 11.
Havlicek, V., Córcoles, A. D., Temme, K., Harrow, A. W., Kandala, A., Chow, J. M., & Gambetta,
J. M. (2018). Supervised learning with quantum enhanced feature spaces. arXiv:1804.11326
Higgott, O., Wang, D., & Brierley, S. (2019). Variational quantum computation of excited states.
Quantum, 3, 156.
Hinton, G. E., (2012). A practical guide to training restricted Boltzmann machines. In G. Montavon,
G. B. Orr, K. R. Müller (Eds.), Neural networks: Tricks of the trade. Lecture Notes in Computer
Science (Vol. 7700). Springer.
Hinton, G. E., & Zemel, R. S. (1994). Autoencoders, minimum description length and Helmholtz
free energy. In Advances in neural information processing systems (Vol. 6, pp. 3–10).
Jannidis, F., Kohle, H., & Rehbein, M. (Eds.) (2017). Digital humanities: Eine Einführung. J.B.
Metzler.
Jozsa, R. (2006). An introduction to measurement based quantum computation. NATO Science
Series, III: Computer and Systems Sciences. Quantum Information Processing-From Theory to
Experiment, 199, 137–158.
Kandala, A., Mezzacapo, A., Temme, K., Takita, M., Brink, M., Chow, J. M., & Gambetta, J. M.
(2017). Hardware-efficient variational quantum Eigensolver for small molecules and quantum
magnets. arXiv:1704.05018
Karalekas, P. J., Tezak, N. A., Peterson, E. C., Ryan, C. A., da Silva, M. P., & Smith, R. S. (2020). A
quantum-classical cloud platform optimized for variational hybrid algorithms. Quantum Science
and Technology, 5, 024003.
Khan, S. U., Awan, A. J., & Vall-Llosera, G. (2019). K-means clustering on noisy intermediate
scale quantum computers. arXiv:1909.12183
50 J. Barzen
Kidger, P., & Lyons, T. (2020). Universal approximation with deep narrow networks. In Machine
learning research (Vol. 125, pp. 1–22). Proc. 33rd Annual Conference on Learning Theory (COLT
2020).
Knill, E. (2007). Quantum computing with very noisy devices. arXiv:quant-ph/0410199 (preprint)
Kubat, M. (2017). An introduction to machine learning. Springer International Publishing AG.
Lamata, L., Alvarez-Rodriguez, U., Martın-Guerrero, J. D., Sanz, M., & Solano, E. (2019). Quantum
autoencoders via quantum adders with genetic algorithms. Quantum Science and Technology, 4,
014007.
Larochelle, H., Mandel, M., Pascanu, R., & Bengio, Y. (2012). Learning algorithms for the
classification restricted Boltzmann machine. Journal of Machine Learning Research, 13,
643–669.
LaRose, R. (2019). Overview and comparison of gate level quantum software platforms. Quantum,
3, 130. arXiv:1807.02500
Leymann, F. (2019). Towards a pattern language for quantum algorithms. In First international
workshop (QTOP 2019).
Leymann, F., & Barzen, J. (2020a). The bitter truth about gate-based quantum algorithms in the
NISQ era. Quantum Science and Technology, 5, 044007.
Leymann, F., & Barzen, J. (2020b). Pattern atlas. arXiv:2006.05120
Leymann, F., Barzen, J., Falkenthal, M., Vietz, D., Weder, B., & Wild, K. (2020). Quantum in the
cloud: Application potentials and research opportunities. In Proceedings of the 10 th International
Conference on Cloud Computing and Services Science (pp. 9–24). SciTePress.
Lubasch, M., Joo, J., Moinier, P., Kiffner, M., & Jaksch, D. (2020). Variational quantum algorithms
for nonlinear problems. Physical Review A, 101, 010301.
McClean, J. R., et al. (2017). OpenFermion: The electronic structure package for quantum
computers. arXiv:1710.07629
McClean, J. R., Romero, J., Babbush, R., & Aspuru-Guzik, A. (2016). The theory of variational
hybrid quantum-classical algorithms. New Journal of Physics, 18, 023023.
McCulloch, W. S., & Pitts, W. (1943). A logical calculus of the ideas immanent in nervous activity.
Bulletin of Mathematical Biophysics, 5, 115–133.
Michielsen, K., Nocon, M., Willsch, D., Jin, F., Lippert, T., & De Raedt, H. (2017). Benchmarking
gate-based quantum computers. Computer Physics Communications, 220, 44–55.
MUSE. (2021). Muster Suchen und Erkennen. https://www.iaas.uni-stuttgart.de/en/projects/muse/
MUSE GitHub. (2021). https://github.com/Muster-Suchen-und-Erkennen/muse-docker
National Academies of Sciences, Engineering, and Medicine. (2019). Quantum computing: progress
and prospects. The National Academies Press. https://doi.org/10.17226/25196
Nielsen, M., & Chuang, I. L. (2010). Quantum computation and quantum information. Cambridge
University Press.
Nocedal, J., & Wright, S. J. (2006). Numerical optimization. Springer Science + Business Media,
LLC.
Novikoff, A. (1962). On convergence proofs for perceptrons. In Proc. Mathematical Theory of
Automata, Vol. 12, pp. 615–622.
Otterbach, J. S., et al. (2017). Unsupervised machine learning on a hybrid quantum computer. arXiv:
1712.05771v1
Palmer, M. (2020). Quantum computers will create better versions of Alexa and Siri. In: Sifted.
Retrieved December 10, 2020, from https://sifted.eu/articles/quantum-computers-nlp/
PennyLane. (2021). https://pennylane.ai/
Pepper, A., Tischler, N., & Pryde. G. J. (2019). Experimental realization of a quantum autoencoder:
The compression of qutrits via machine learning. Physical Review Letters, 122, 060501.
Peruzzo, A., McClean, J., Shadbolt, P., Yung, M.-H., Zhou, X.-Q., Love, P. J., AspuruGuzik,
A., & O’Brien, J. L. (2014). A variational eigenvalue solver on a quantum processor. Nature
Communications, 5, 4213.
Poljak, S., & Rendl, F. (1995). Solving the max-cut problem using eigenvalues. Discrete Applied
Mathematics, 62, 249–278.
From Digital Humanities to Quantum Humanities … 51
Preskill, J. (2018). Quantum computing in the NISQ era and beyond. Quantum, 2, 79.
PyTorch. (2021). https://pytorch.org/
QH. (2021). Quantum humanities. https://www.iaas.uni-stuttgart.de/forschung/projekte/quantum-
humanities/
QHAna. (2021). Quantum humanities data analysis tool. https://github.com/UST-QuAntiL/qhana
Qiskit. (2021). https://qiskit.org/
Qiskit Max Cut. (2021). https://qiskit.org/documentation/locale/de_DE/tutorials/optimization/6_e
xamples_max_cut_and_tsp.html
Raz, R., & Tal, A. (2018). Oracle separation of BQP and PH. Iin: Electronic colloquium on
computational complexity, Report No. 107.
Rehbein, M., & Sahle, P. (2013). Digital Humanities lehren und lernen. Modelle, Strategien,
Erwartungen. In: H. Neuroth, N. Lossau, A. Rapp (Eds.), Evolution der Informationsinfrastruktur
Kooperation zwischen Bibliothek und Wissenschaft. Verlag Werner Hülsbusch.
Reiners, R. (2014). An evolving pattern library for collaborative project documentation. Disserta-
tion, University of Aachen.
Rieffel, E., & Polak, W. (2011). Quantum computing: A gentle introduction. The MIT Press.
Romero, J., Olson, J. P., & Aspuru-Guzik, A. (2017). Quantum autoencoders for efficient
compression of quantum data. Quantum Science Technology, 2, 045001.
Rønnow, T., Wang, Z., Job, J., Boixo, S., Isakov, S. V., Wecker, D., Martinis, J. M., Lidar, D. A., &
Troyer, M. (2014). Defining and detecting quantum speedup. Science, 345(6195).
Rosenblatt, F. (1958). The perceptron: A probabilistic model for information storage and
organization. Psychological Review, 65(6).
Sahle, P. (Ed.) (2019). DHd 2019. Digital humanities: Multimedial & multimodal. Konferenzab-
stracts. Mainz/Frankfurt a M.
Schöch, C. (Ed.) (2020). DHd 2020. Spielräume Digital Humanities zwischen Modellierung und
Interpretation. Konferenzabstracts. https://doi.org/10.5281/zenodo.3666690
Schuld, M., & Petruccione, F. (2018). Supervised learning with quantum computers. Springer.
Schuld, M., Sinayskiy, I., & Petruccione, F. (2014). The quest for a Quantum Neural Network.
arXiv:1408.7005v1
Schumm, D., Barzen, J., Leymann, F., & Ellrich, L. (2012). A pattern language for costumes in films.
In Proceedings of the 17th European Conference on Pattern Languages of Programs (EuroPLoP
2012).
Shaydulin, R., Safro, I., & Larson, J. (2019). Multistart methods for quantum approximate
optimization. arXiv:1905.08768
Shor, P. W. (1995). Polynomial-time algorithms for prime factorization and discrete logarithms on
a quantum computer. arXiv:quant-ph/9508027
Sierra-Sosa, D., Arcila-Moreno, J., Garcia-Zapirain, B., Castillo-Olea, C., & Elmaghraby, A. (2020).
Dementia prediction applying variational quantum classifier. arXiv:2007.08653
Skiena, S. S. (2017). The data science design manual. Springer International Publishing AG.
Smolensky, P. (1986). Information processing in dynamical systems: Foundations of harmony
theory. In D. E. Rumelhart & J. L. McLelland (Eds.), Parallel distributed processing: explorations
in the microstructure of cognition, Volume 1: Foundations (pp. 194–281). MIT Press.
Tacchino, F., Macchiavello, C., Gerace, D., & Bajoni, D. (2018). An artificial neuron implemented
on an actual quantum processor. arXiv:1811.02266v1
TensorFlow. (2021). https://www.tensorflow.org/
TensorFlow Quantum. (2021). https://www.tensorflow.org/quantum
Terras, M., Nyhan, J., & Vanhoutte, E. (2013). Defining digital humanities. A reader.
Tomczak, J. M. (2016). Learning informative features from restricted Boltzmann machines. Neural
Processing Letters, 2016(44), 735–750.
Vogeler, G. (Ed.) (2018). DHd 2018. Kritik der digitalen Vernunft. Konferenzabstracts.
Wan, K. H., Dahlsten, O., Kristjánsson, H., Gardner, R., & Kim, M. S. (2017). Quantum
generalisation of feedforward neural networks. npj Quantum Information, 3, 36.
52 J. Barzen
Weigold, M., Barzen, J., Breitenbücher, U., Falkenthal, F., Leymann, F., & Wild, K. (2020). Pattern
views: Concept and tooling for interconnected pattern languages. In S. Dustdar (Ed.), Service-
oriented computing. SummerSOC 2020. Communications in computer and information science
(Vol. 1310). Springer, Cham. https://doi.org/10.1007/978-3-030-64846-6_6
Weigold, M., Barzen, J., Leymann, F., & Salm, M. (2021). Data encoding patterns for quantum algo-
rithms. In The Hillside Group (Ed.), Proceedings of the 27th Conference on Pattern Languages
of Programs (PLoP’20) (to appear).
What is Digital Humanities? (2021). http://whatisdigitalhumanities.com/
Wiebe, N., Kapoor, A., & Svore, K. M. (2016). Quantum perceptron models. arXiv:1602.04799
Wu, Y., Wei, C., Qin, S., Wen, Q., & Gao, F. (2020). Quantum restricted Boltzmann machine
universal for quantum computation. arXiv:2005.11970v3
Wu, Z., & Palmer, M. (1994). Verb semantics and lexical selection. In ACL’94 Proceedings of the
32nd annual meeting on Association for Computational Linguistics.
Yuan, X., Endo, S., Zhao, Q., Li, Y., & Benjamin, S. (2019). Theory of variational quantum
simulation. Quantum, 3, 191.
Zhang, P., Li, S., & Zhou, Y. (2015). An algorithm of quantum restricted Boltzmann machine
network based on quantum gates and its application. Shock and Vibration Volume. Article ID
756969.
Zheng, X., Wu, Z., Meng, H., Li, W., & Cai, L. (2013). Feature learning with Gaussian restricted
Boltzmann machine for robust speech recognition. arXiv:1309.6176
Zurada, J. M. (1992). Introduction to artificial neural systems. West Publishing Company.
(All links have been last followed March 11, 2021)
Quantum Computing and Cognitive
Simulation
Martha Lewis
Abstract Cognitive Science is the study of the mind and how it relates to human
behaviour. As a field, it is highly interdisciplinary, bringing together theories from
Psychology, Neuroscience, Biology, and others. The field of cognitive science oper-
ates at a number of levels. On the one hand, some aspects of cognitive science look
at the behaviour of neurons, whereas other aspects look to explain human behaviour
at a more abstract level, seeking to explain human decision making or reasoning.
A key area of research in cognitive science is how to formalise human behaviours
around judgements of similarity, categorization, and decision making. In the field
of Physics, Quantum Mechanics has fundamentally altered our understanding of the
way in which particles behave. Quantum mechanics has a number of unintuitive phe-
nomena, some of which can be used model unusual aspects of human behaviour. The
application of quantum theory to model human behaviour is wide-ranging. In this
chapter we will look at three main areas in which it has been applied. One key area
is in how similarity judgements can be modelled. There are a number of phenomena
around similarity judgements that are not well modelled using a view of concepts
that does not take the state of the observer into account. These include the asymmetry
of similarity judgements and the fact that similarity can change depending on other
exemplars that are present. Another key area is in judgement and decision-making.
Again, puzzling phenomena have been observed regarding the fact that judgements
do not follow classical probability theory, known as the conjunction fallacy, can
depend crucially on the order in which questions are presented or on the amount of
knowledge about the world. Finally, approaches to modelling cognitive phenomena
at the neural level will also be discussed within a quantum-theoretic framework.
M. Lewis (B)
Department of Engineering Mathematics, University of Bristol, Bristol, UK
e-mail: [email protected]
1 Introduction
Cognitive science is the study of the mind and how it relates to human behaviour.
As a field, it is highly interdisciplinary, bringing together theories from Psychol-
ogy, Neuroscience, Biology, and others. The field of cognitive science operates at
a number of levels. On the one hand, some aspects of cognitive science look at the
behaviour of neurons, whereas other aspects look to explain human behaviour at a
more abstract level, seeking to explain human decision-making or reasoning. There
is a long-standing history of vector space models in cognitive science. Theories of
categorization such as those developed by Ashby and Gott (1988), Nosofsky (1986),
Rosch (1975) utilize notions of distance and similarity that can readily be incor-
porated in vector space models of meaning. Hampton (1987), Smith and Osherson
(1984), Tversky (1977) encode meanings as feature vectors, and models of high-level
cognitive reasoning have been implemented within vector symbolic architectures
(Gayler, 2003; Plate, 1995; Smolensky, 1990).
Another field in which vector space models play an important role is Physics,
and especially quantum theory. Though seemingly unrelated to language, intriguing
connections have recently been uncovered. The link between Physics and natural lan-
guage semantics that vector space models provide has been successfully exploited,
providing novel solutions and a fresh perspective for a number of problems related
to cognitive science, such as modelling logical aspects in vector spaces (Widdows
& Peters, 2003). Methods from quantum logic have also been applied to cognitive
processes related to the human mental lexicon, such as word association (Bruza
et al., 2009), decision-making (Pothos & Busemeyer, 2013), and human probability
judgements (Busemeyer et al., 2011). Furthermore, the categorical model of Coecke
et al. (2010), inspired by Quantum Mechanics, has provided a convincing account
of compositionality in vector space models and an extensible framework for linguis-
tically motivated research on sentential semantics. More recently, the link between
Physics and text meaning was made more concrete by a number of proposals that aim
at replacing the traditional notion of a word vector with that of a density matrix—a
concept borrowed from Quantum Mechanics which can be seen as a probability dis-
tribution over vectors (Bankova et al., 2019; Piedeleu et al., 2015; Sadrzadeh et al.,
2018).
A key area of research in cognitive science is how to formalize human behaviours
around judgements of similarity, categorization, and decision-making. However, for-
malizing concepts within a vector space, and relying on distance within vector space
to give a notion of similarity, can lead to some puzzling features, at least if we
assume that distance and similarity behave classically. For example, it has been
shown (Tversky, 1977) that judgements of similarity are not always symmetric, as
you would expect if they are just based on distance in a vector space.
Quantum Computing and Cognitive Simulation 55
In the field of Physics, Quantum Mechanics has fundamentally altered our under-
standing of the way in which particles behave. Quantum mechanics has a number
of unintuitive phenomena, some of which can be used to model unusual aspects
of human behaviour. This is not to say that a description of the brain and mind is
provided in terms of quantum phenomena at a small scale, although this sort of mod-
elling has been proposed (Hameroff, 2014; Penrose, 1990). Instead, the idea is that
the formalism arising from quantum theory provides the right sort of mathematical
tools to model these puzzling aspects of human behaviour.
The application of quantum theory to model human behaviour is wide-ranging. In
this chapter, we will look at three main areas in which it has been applied. One key
area is how similarity judgements can be modelled. There are a number of phenom-
ena around similarity judgements that are not well modelled using a view of concepts
that does not take the state of the observer into account. These include the asymmetry
of similarity judgements and the fact that similarity can change depending on other
exemplars that are present (Tversky, 1977). These phenomena have been addressed
in Pothos et al. (2013). Another key area is in judgement and decision-making. Again,
puzzling phenomena have been observed regarding the fact that judgements do not
follow classical probability theory, known as the conjunction fallacy (Tversky &
Kahneman, 1983), and can depend crucially on the order in which questions are
presented (Moore, 2002) or on the amount of knowledge about the world (Tversky
& Shafir, 1992). These phenomena are modelled within a quantum framework in
Busemeyer et al. (2011), Franco (2009), Khrennikov and Haven (2009), Pothos and
Busemeyer (2009), Wang and Busemeyer (2013). Related to this area is the phe-
nomenon of contextuality. This can be summarized as the idea that there may be
sets of random variables (in Psychology, results of an experiment) that have pairwise
joint distributions but for which no joint distribution across the whole set of random
variables can be found. This is one of the key aspects of quantum theory. Amazingly,
the same theory was also developed in Psychology, and has been formalized in Dzha-
farov and Kujala (2014, 2016). Finally, quantum theory has been applied to describe
categorization and concept combination. Again, the ways in which humans use con-
cepts has been shown not to be well modelled by classical views of combination like
fuzzy set theory, where problems known as over- and underextension are observed
(Hampton, 1987, 1988a, 1988b, 1997; Smith & Osherson, 1984). Approaches to
answer these phenomena have been proposed in Aerts (2009), Aerts and Gabora
(2005a, 2005b), Aerts et al. (2015), Sozzo (2014, 2015). The problem of modelling
how concepts compose can also be addressed via a linguistic route. The problem of
modelling concept composition by means of the grammatical structure in language
has been addressed in Coecke et al. (2010), and applications of this to the problem
of overextension was proposed in Coecke and Lewis (2015).
Approaches to modelling cognitive phenomena at the neural level have also been
considered within a quantum-theoretic framework. Fuss and Navarro (2013) show
that a quantum random walk approach to modelling choice tasks better simulates
human reaction times. Consideration of neuronal activation within a quantum frame-
work is also discussed in Pothos and Truebloodm (2015). The question of modelling
how concepts can ‘bind’ together (for example, the combination of an adjective and
56 M. Lewis
a noun) was addressed at a vector-based level in Smolensky (1990) and has been
investigated within a cognitive science context in Martin and Doumas (2020). The
compositional distributional framework of Coecke et al. (2010) has the potential to
model these aspects of neuronal activity well, and combining this theory with the
tensor product binding of Smolensky (1990) is an area for future research.
In the remainder of the chapter, we cover the following. We provide a short section
on the mathematical notation we will use. In Sect. 2, we will summarize the cognitive
phenomena that have been described by quantum theory. In Sect. 3, we describe
how quantum theory has been used to address each of these phenomena, as well as
discussing quantum-theoretic approaches to modelling neuronal-level phenomena. In
Sect. 4, we discuss the implications for using quantum computing to model cognitive
phenomena and Artificial Intelligence more generally.
We assume that the reader has a general understanding of linear algebra. We use
Dirac’s bra-ket notation to represent vectors, their duals, and inner products.
• A ket |v is a column vector in a Hilbert space H. We will always consider the
space to be finite-dimensional, so H = Cn .
• A bra v| is the vector dual to |v. It is the conjugate transpose of |v and can
be thought of as a row vector whose elements are the complex conjugates of the
elements in |v.
• The inner product of two vectors |v, |w is represented by v|w.
• If a matrix is represented by M, its multiplication with a vector |v is given by
M|v and its multiplication with a bra v| is given by v|M. Often, these will be
combined to give a scalar v|M|v.
• The
√ absolute value of a complex number α = a + ib is represented as |α| =
a 2 + b2 . √
• The Euclidean norm of a vector is written as ||v|| = v|v.
2 Cognitive Phenomena
d(x, x) = 0 (1)
d(x, y) = d(y, x) (symmetry) (2)
d(x, z) ≤ d(x, y) + d(y, z) (triangle inequality) (3)
Tversky (1977) show that these assumptions do not necessarily hold. In a series of
experiments, Tversky showed that when one object is considered more prominent
than another, the less prominent object is considered more similar to the more promi-
nent object and vice versa. Firstly, pairs of countries are assessed (by Israeli college
students) for which country is most prominent. The pairs of countries are, for exam-
ple, China and Vietnam,1 the USA and Mexico, or Belgium and Luxembourg. A
separate group of students were then asked to judge which phrase they preferred to
use when describing the similarity of the two countries out of “country a is similar to
country b” and “country b is similar to country a”. Across all pairs, the majority of
students chose the ordering in which the more prominent country was given second,
for example “Vietnam is similar to China” rather than “China is similar to Vietnam”.
This kind of effect was seen across a range of topics and modalities, for example,
when judging the similarity of shapes, letters, and sequences of sounds.
2.2 Diagnosticity
Tversky (1977) further show that judgements of similarity can be altered by the
presence of comparison objects. According to a geometric model of similarity, the
similarity of two objects should not be affected by the presence of other objects.
1 In fact, the countries assessed are Red China and North Vietnam.
58 M. Lewis
However, this is shown not to hold for human similarity judgements in the following
manner. Pairs of quadruples of countries were designed that differ in only one entry,
i.e. there are pairs of sets {a, b, c, p} and {a, b, c, q}. Participants are asked to say
which of b, c and q or p country a is most similar to. For example, one such pair is the
sets of countries {Austria, Sweden, Poland, Hungary} and {Austria, Sweden, Norway,
Hungary}. Participants see only one quadruple and are asked to judge which country
Austria is most similar to. Tversky (1977) find that in the set {Austria, Sweden,
Poland, Hungary}, most participants judge Austria to be most similar to Sweden, but
in the quadruple {Austria, Sweden, Norway, Hungary} participants judge Austria to
be most similar to Hungary.
The description is followed by eight statements about the person’s career or per-
sonality:
• Linda is a teacher in elementary school.
• Linda works in a bookstore and takes Yoga classes.
• Linda is active in the feminist movement (F).
• Linda is a psychiatric social worker.
• Linda is a member of the League of Women Voters.
• Linda is a bank teller (T ).
• Linda is an insurance salesperson.
• Linda is a bank teller and is active in the feminist movement (T &F).
Participants are asked to rank the eight statements associated with each description
by how probable each statement is. A large majority (between 85 and 90%, depending
on the participant pool) rank the statements in the order T &F > T , that is, they
judge that the probability that Linda is both a feminist and a bank teller is higher than
the probability that Linda is a bank teller. This phenomenon still holds (to a lesser
Quantum Computing and Cognitive Simulation 59
extent) in other contexts where the participants are asked to place a bet on the two
statements T and T &F, in a domain of medical experts (where the story that is told
is about symptoms and the task is to diagnose the patient), and in a range of other
situations designed to reduce the incidence of the conjunction fallacy. Experiments
where fallacy does not hold to such a great extent are when the participant pool is
with a range of statistically sophisticated participants, and when the experiment is
phrased in terms of numbers of people.
In Moore (2002), the effect of order on answers to questions is addressed. If you are
asked “Do you try to eat healthily?” and “Do you like McDonalds?”, your answer
to each question may change depending on which question is asked first. There are
a number of ways in which the ordering could affect the answers. Moore (2002)
identifies four types of such effect, termed contrast, consistency, additive, and sub-
tractive. The consistency effect runs as follows. A group of participants is asked
a pair of questions. For a concrete example, consider the questions “Do you think
Bill Clinton is trustworthy” and “Do you think Al Gore is trustworthy”. For context,
these questions were asked in 1997. If the question is asked first, it is considered
to be asked in a non-comparative context, meaning that there is nothing to imme-
diately compare the question to. If the question is asked second, it is considered to
be asked in a comparative context, because it can be considered in comparison to
the first. The answers to the questions were distributed as in Table 1. The effect of
the question ordering is that the answer to the question in the comparative context is
altered to make it more consistent with the answer in the non-comparative context.
So, if a participant is first asked whether Gore is honest and trustworthy, and they
answer positively, then they are more likely to answer positively to the question of
whether Clinton is honest and trustworthy. On the other hand, if a participant is first
asked whether Clinton is honest and trustworthy, and answers negatively, then they
are more likely to answer negatively to the question of whether Gore is honest and
trustworthy. The effect is to make the answers to the two questions more consistent.
Table 1 Example of the consistency effect. ∗ indicates significant difference. Figures from Moore
(2002)
Per cent saying yes
Context Clinton Gore Gap
First 50 68 +18∗
Second 57 60 +3
Difference +7∗ −8∗ −15∗
60 M. Lewis
Table 2 Example of the contrast effect. ∗ indicates significant difference. Figures from Moore
(2002)
Per cent saying yes
Context Gingrich Dole Gap
First 41 60 +19∗
Second 33 64 +31∗
Difference −8∗ +4∗ +12∗
Table 3 Example of the additive effect. ∗ indicates significant difference. Figures from Moore
(2002)
Per cent saying ‘All’ or ‘Many’
Context White Black Gap
First 41 46 +5
Second 53 56 +3
Difference +12∗ +10∗ −2
The contrast effect occurs when people alter their answers to make the answers
contrast with each other. For example, in a Gallup poll in 1995 when asked the
question of whether the label ‘honest and trustworthy’ applies to either Bob Dole or
Newt Gingrich, people’s answers in the comparative context altered to increase the
gap between them, emphasizing the differences between the two. Figures are shown
in Table 2.
Under the additive effect, considering either question first has the same effect. In
a series of racial hostility polls in 1996, people were asked the questions “Do you
think that only a few white people dislike blacks, many white people dislike blacks,
or almost all white people dislike blacks?” and “Do you think that only a few black
people dislike whites, many black people dislike whites, or almost all black people
dislike whites?”. In both cases, the percentage responding ‘All’ or ‘Many’ increased
in the comparative context; see Table 3.
Lastly, in the subtractive context considering either question first again has the
same effect, but in this case the effect is to decrease the number of positive answers.
The example given in Moore (2002) is the question of whether baseball players Pete
Rose and, separately, Shoeless Joe Jackson should be eligible for admission to the
Baseball Hall of Fame. In each case, the players are ineligible, and the participants
are told why. In each case, the number of people answering favourably to the question
of whether they should be allowed decreased in the comparative context (Table 4).
These four examples show clearly the different kinds of effects that question order
can have on responses, and that the context in which a question is asked is crucial
for its response. These effects are addressed in Wang and Busemeyer (2013), Wang
et al. (2013).
Quantum Computing and Cognitive Simulation 61
Table 4 Example of the subtractive effect. ∗ indicates significant difference. Figures from Moore
(2002)
Per cent saying ‘Favourable’
Context Rose Jackson Gap
First 64 45 +19∗
Second 52 33 +19∗
Difference −12∗ −12∗ 0
31% of respondents said that they would buy the vacation package, 7% said that they
would not buy the vacation package, and 60% said that they would reserve the low
price for the small fee-paying to wait until the uncertainty was resolved.
However, when the participants were asked to imagine they knew the outcome of
the exam, a majority said that they would buy the package:
Imagine that you have just taken a tough qualifying examination. It is the end of the semester,
you feel tired and run-down, and you find out that you [passed the exam/failed the exam.
You will have to take it again in a couple of months-after the Christmas holidays]. You now
have an opportunity to buy a very attractive 5-day Christmas vacation package to Hawaii at
an exceptionally low price. The special offer expires tomorrow. Would you
In the context where the participant had passed the exam, 54% of respondents said
that they would buy the vacation package, 16% of respondents said that they would
not buy the vacation package, and 30% of students said that they would pay the $5
waiting fee. Very similarly, in the case where the participant had failed the exam, 57%
of respondents said that they would buy the vacation package, 12% of respondents
said that they would not buy the vacation package, and 31% of students said that they
would pay the $5 waiting fee. Essentially, in the context where the outcome of the
exam is known, a majority of participants decide to book the holiday—whether or
not it is pass or fail. According to the sure thing principle, those participants should
decide to book the holiday in that case where the outcome of the exam is not known.
But, as can be seen, this does not happen.
A similar experiment involves gambling:
Imagine that you have just played a game of chance that gave you a 50% chance to win $200
and a 50% chance to lose $100. The coin was tossed and you have [won $200/lost $100].
You are now offered a second identical gamble
50% chance to win $200 and 50% chance to lose $100. Would you:
In the context that the participants had won the first gamble, 69% accepted and 31%
rejected the second. In the context that participants had lost the first gamble, the
split was 59% accept, 41% reject. Either way, a majority of participants accepted the
second gamble.
However, in a context where they did not know the outcome of the first gamble,
a majority of the same set of participants rejected the second gamble. The split was
36% accept, 64% reject.
The explanation given in Tversky and Shafir (1992) is that the presence of uncer-
tainty makes it more difficult for people to focus on the implications of each outcome.
A further example is given in a prisoner’s dilemma situation. In the prisoner’s
dilemma, a hypothetical prisoner has the choice either to cooperate with the other
prisoner, i.e. keep quiet about their activities, or defect, i.e. tell the authorities what
they know. The possible outcomes of cooperating or defecting are presented in the
following payoff matrix (Table 5).
In an experimental situation, if the participant is told that the other has cooperated,
the majority of participants choose to defect. Also, if the participant is told that the
other has defected, the majority of participants choose to defect. However, if the
Quantum Computing and Cognitive Simulation 63
participant is not told what the choice of the other is, the majority of participants
choose to cooperate.
2.6 Categorization
One of the most important things that humans can do is to categorize their experi-
ences. Is this fruit good to eat, or poisonous? Is that large striped animal dangerous?
Categorization can be thought of in terms of similarity to other experiences, and to
help us do this we use concepts. There have been a number of theories of concepts.
The classical view of concepts dates back to Plato and Aristotle. Under the classical
view of concepts, a concept is viewed as a set of necessary and sufficient conditions,
formulated as a list of properties. Such a view of concepts is inadequate. Firstly,
for many natural concepts it is impossible to give a set of necessary and sufficient
conditions (Wittgenstein, 1953). Instead, instances of a category can be thought of
as sharing a set of ‘family resemblances’. For example: what makes the images in
Fig. 1 pictures of chairs, and the images in Fig. 2 not pictures of chairs? It is difficult
to decide on a set of necessary and sufficient conditions that rules the kneeling chair
to be a chair, but rules out a bar stool.2,3,4,5
Furthermore, the classical view does not take into account the graded and context-
dependent nature of concepts. Consider the much simpler concept tall. Whether or
not to describe something as tall will firstly be dependent on the set of things that
form its context, for example, 4-year-old children, or Dutch women, or giraffes, or
mountains. Secondly, even if we have fixed a context, it is not a crisp question of
2 https://commons.wikimedia.org/wiki/File:Philipp_Rumpf_Studienblatt_eines_Stuhls.jpg.
3 https://commons.wikimedia.org/wiki/File:Sleekform_Kneeling_Chair.jpg.
4 https://commons.wikimedia.org/wiki/File:Shaker_no._7_rocking_chair_Rocking_Chair,_1878
%E2%80%931910_(CH_18460985).jpg.
5 https://commons.wikimedia.org/wiki/File:LibertyChair.jpg.
64 M. Lewis
Fig. 2 Non-chairs
whether we can divide some people as tall. In Fig. 3, how are we to decide what is
the necessary and sufficient condition for being tall? Certainly, anyone to the right of
Jeremy Clarkson is tall, and anyone to the left of Nikita Khrushchev is not tall, but
should we state the condition to be 180 cm? 184 cm? If we say 184 cm and someone
is 183.5 cm, are they really not tall?6,7,8,9
These sorts of considerations led to the formulation of prototype theory (Rosch,
1975). In prototype theory, concepts are assumed to have a prototype, and members
of the concept are judged as being more or less similar to the prototype. For example,
in the concept bird, a robin is judged as a very prototypical bird, whereas an ostrich
or a penguin would be judged as much less typical. This allows the consideration
6 https://commons.wikimedia.org/wiki/File:Bar_Stools_(49907001456).jpg.
7 https://commons.wikimedia.org/wiki/File:52.8_Inch_Sofa_Couch_Loveseat,_Grey.jpg.
8 https://commons.wikimedia.org/wiki/File:TabouretAFDB.jpg.
9 https://commons.wikimedia.org/wiki/File:Wooden_Rocking_Horse_with_Weels.jpg.
Quantum Computing and Cognitive Simulation 65
of concepts in terms of family resemblances, and accounts for the graded nature of
concepts.
Further, the exemplar theory of concepts views a category as based on specific
instances, or exemplars, that are stored in memory. Categorization judgements are
based on judgements of similarity to the exemplars, which are easily called to mind.
Exemplar theory and prototype theory are both backed by experimental evidence
(Nosofsky, 1986; Rosch, 1975). Key characteristics of these theories are that they can
account for the graded nature of concepts, and in both cases, we represent concept
instances within some kind of feature space, equipped with a distance metric.
Another key feature of human concept use is that we are able to combine concepts
to create new ones. This can be done in very simple ways, for example, combining
the concept blue and the concept book, we obtain a blue book. How can we formalize
this? One approach to formalizing concept combination used fuzzy set theory. Under
this view, a concept is a fuzzy set, that is, a set whose members have membership
values in the interval [0, 1] rather than in the set {0, 1}. We write
μ A : X → [0, 1] (4)
and use the notation μ A (x) to describe the membership of the item x in the concept
A. In fuzzy set theory, we can form the conjunction or disjunction of two sets as
follows:
Under these rules, combinations quickly become problematic. The classic exam-
ple is that of a pet fish (Smith & Osherson, 1984). Consider a fish like a goldfish. A
goldfish is not a particularly typical fish, nor a very typical pet, so we might assign
low memberships to goldfish in each of these concepts:
However, it is a very typical pet fish, i.e. we would assign something like
2.7 Negation
Suspect Tall Not tall Tall and not tall Neither tall nor not tall
True False True False True False True False
5 4 1.3 98.7 94.7 3.9 14.5 76.3 27.6 65.8
5 7 5.3 93.4 78.9 17.1 21.1 65.8 31.6 57.9
5 11 46.1 44.7 25 67.1 44.7 40.8 53.9 42.1
6 2 80.3 10.5 9.2 82.9 28.9 56.6 36.9 55.3
6 6 98.7 1.3 0 100 5.3 81.6 6.6 89.5
There is a strong correlation between the values for the borderline statements ‘tall
and not tall’ and ‘neither tall nor not tall’. Participants were willing to agree to both
of these statements. In the first statement ‘tall and not tall’, a participant would seem
to ascribe the value true both to ‘tall’ and to ‘not tall’, whereas in the second, the
participant would also judge that he is neither.
The problem of negation is also considered by Hampton, in a setting combined
with conjunction (Hampton, 1997). In this setting, Hampton considers the typicality
of items to conjunctions of concepts. For example, we can consider the typicality
of parrot to the conjunction Pets which are also birds and Pets which are not birds.
Of course, we would consider the typicality of parrot in the first to be high and in
the second to be low. However, if concepts are being combined in line with log-
ical reasoning, then the sum of the typicality in Pets which are also birds and in
Pets which are not birds should not exceed the typicality of parrot in Pets. This is
another form of overextension, as we have seen in previous descriptions.
Quantum Computing and Cognitive Simulation 67
3 Quantum Models
The problems for modelling cognitive phenomena outlined in Sect. 2 above have all
been tackled using methods from quantum theory. Below, in Sect. 3.1 we give an
overview of the key ideas before explaining the models in detail. After this overview,
we discuss approaches to modelling similarity, diagnosticity, and the conjunction
fallacy. These approaches have been developed over a research programme laid out
in Busemeyer et al. (2011), Franco (2009), Pothos et al. (2013) and other papers,
and crucially depend on the notion of interference. Following this, we consider the
sure thing principle and question order models. Again, these models depend on the
notion of interference. We go on to discuss models of concept combination and
negation. These approaches use the quantum phenomenon of entanglement as well
as interference to describe how humans combine concepts. Linked to these is the
notion of contextuality. Finally, we talk about quantum models that are applied at a
neuronal level.
The following gives an overview of some of the key quantum concepts used in
approaches such as those in Busemeyer et al. (2011), Pothos et al. (2013). Within
applications of quantum theory to Psychology, questions are considered from the
point of view of an agent. We consider an agent as having a certain state |ψ, and as
questions are asked of the agent, or the agent is exposed to a stimulus, the state of
the agent changes. Within quantum theory, a question about a system is termed an
observable: a self-adjoint linear operator. The eigenvectors {|ai }i of an observable
A form an orthonormal basis of the Hilbert space, and the eigenvalues {ai } associ-
ated with the {|ai }i form the values that the observable can take, i.e. the answers
to the questions. To give a simple example: a yes/no question such as “is the sky
blue?” might be encoded in a two-dimensional vector space. An observable B on the
space will have two orthonormal eigenvectors |blue and |not blue. We can set the
eigenvalues associated with |blue and |not blue to be 1 and 0, respectively.
Now, we represent an agent’s belief about the state of the sky as a vector |s.
Given that the agent is ambivalent about what the weather is (possibly clear and
blue, possibly overcast and grey), the probability that we receive the answer ‘yes!’
to the question “Is the sky blue” is given by the following calculation:
This quantity is the square of the projection onto the blue axis, illustrated in Fig.
4.
This notion can be extended. Firstly, there may be more than one basis vector
corresponding to a given answer, so that that answer corresponds to a subspace of
68 M. Lewis
Fig. 4 A two-dimensional
space with orthonormal basis blue
|blue, |not blue. The state
|s0 represents belief in a
| s0⟩
mostly blue sky with a ⟨blue | s0⟩
couple of clouds, and the
state |s1 represents belief in
a mostly grey sky. The
projection onto the blue axis
is high for |s0 and low for
|s1 , whereas the projection
onto the not blue axis is low | s1⟩
for |s0 and higher for |s1 .
⟨blue | s1⟩
not blue
⟨not blue | s1⟩
⟨not blue | s0⟩
the vector space of dimension greater than 1. For example, we could have a three-
dimensional vector space with one dimension |blue corresponding to blue sky, one
dimension |cloudy corresponding to cloudy sky, and one dimension |night corre-
sponding to nighttime, and suppose we define an observable D to ask the question
“is it daytime?” D has eigenvectors |blue and |cloudy with eigenvalue 1 (corre-
sponding to yes) and eigenvector |night with eigenvalue 0 (corresponding to no).
Then,
illustrated in Fig. 5.
More generally again, the current state of affairs may be uncertain. Quantum
theory models this as a mixture of states
ρ= si |si si | (13)
i
where i si = 1.
Then the probability of the answer ‘yes!’ to the question “is it daytime?” is given
by
T r (Dρ) (14)
Quantum Computing and Cognitive Simulation 69
blue
where T r stands for matrix trace. If we are not at all uncertain about the state, then
ρ = |ss| and this is equivalent to the formula (12).
A key difference between classical probability theory and quantum probability
theory is the treatment of conjunctions of events. In classical probability theory, the
probability P(AB) conjunction of two events A and B is equal to the probability
P(B A) of the two events in the opposite order. In quantum probability, this does not
hold, due to the fact that in general the matrices corresponding to the observables Â
and B̂ will not commute.
This means that conjunction in quantum probability needs to be defined in sequen-
tial terms. Specifically, suppose we want to calculate the probability of an agent in
state |s judging A and then B. We have
where |s A is the state of knowledge on the agent after judging that A holds, i.e.
A|s
|s A = (18)
||A|s||
i.e. the agent projects first to subspace A, and then to subspace B. In general, this is
not equal to ||AB|s||2 , as it would be in classical probability.
In Pothos et al. (2013), the model described above is used to account for the asym-
metry of similarity judgements. Recall that Tversky (1977) showed that when asked
to consider the similarity of one entity, such as a country, to another, the judgements
of similarity are higher if the entity that is more salient is the one being compared
to. So, for example, more participants agreed with the statement “Korea is similar
to China” than “China is similar to Korea” (the participants were all Israeli college
students).
Pothos et al. (2013) explicate this as follows. The entities being judged, for exam-
ple, China and Korea are modelled as subspaces of a given vector space. Each of
these subspaces has an associated projector PChina , PKorea . The state of belief of the
agent is modelled as a quantum state |ψ. The probability that the agent’s belief state
is consistent with the concept China or the concept Korea is given by the square of
the projection onto the relevant subspace: ||PChina |ψ||2 = ψ|PChina ψ. The initial
belief state | psi is set up so that ψ|PChina ψ = ψ|PKorea ψ.
Then, the similarity of Korea to China is calculated but first projecting | psi onto
the Korea subspace, and then projecting the resulting vector onto the China subspace:
This quantity calculates the probability that the original state |ψ is consistent with
Korea, multiplied by the probability that this new state is consistent with China.
The agent projects their belief state onto Korea, after which their new belief state is
|ψ
normalized, so that |ψKorea = ||PPKorea
Korea |ψ||
. This new state is projected onto China, and
is consistent with the concept China with probability ||PChina |ψKorea . So, we have
that
||PChina |ψKorea ||2 · ||PKorea |ψ||2 ||PChina PKorea |ψ||2 (22)
Now, as long as the projectors PChina and PKorea do not commute, we have that
Sim(Korea, China) = Sim(China, Korea). Moreover, Pothos et al. (2013) show that
the similarity of Korea to China can be made to be higher than the similarity of China
to Korea by modelling the more salient entity, in this case China, as a larger (higher
dimensionality) subspace.
Quantum Computing and Cognitive Simulation 71
3.3 Diagnosticity
Pothos et al. (2013) also show that the phenomenon of diagnosticity (Sect. 2.2) can be
modelled. In this model, each country is modelled as a single-dimensional subspace.
The diagnosticity effect is produced by the order in which the projectors for the
countries are applied to the agent’s state vector |ψ.
Recall from Sect. 2.2 that the diagnosticity effect says that similarity judgements
can be affected by the context they are made in. An example from Tversky (1977)
is the following. Take the sets of countries {Austria, Sweden, Poland, Hungary} and
{Austria, Sweden, Norway, Hungary}. Participants see only one quadruple and are
asked to judge which country Austria is most similar to. In the quadruple {Austria,
Sweden, Poland, Hungary}, Austria is judged most similar to Sweden by a majority
of participants. However, in {Austria, Sweden, Norway, Hungary}, Austria is judged
most similar to Hungary. Given that both are available in each set of countries, why
does the similarity change?
Pothos et al. (2013) set each country to be represented as a ray. The similarity
of Austria to any of the other countries in the set is modelled by saying that the
agent’s belief state |ψ first projects onto the subspaces corresponding to the other
countries. There are two orders in which this can happen, since there are two other
countries. Therefore, these orders are averaged. Then, depending on the placing of
the rays in the Hilbert space, the required effect can be produced. Pothos et al. (2013)
show that with some reasonable assumptions about the similarities between the pairs
of countries (i.e. we assume that Sweden is similar to Norway, and that Hungary is
similar to Poland), the effect can be reliably reproduced over a number of simulations.
Franco (2009) provide an account of the conjunction fallacy using the notion of
interference effect. The model works as follows. Recall that in the conjunction fallacy
(Sect. 2.3), the problem is that the conjunction of two events (that Linda is a bank
teller and a feminist) is judged more probable than the single event that Linda is a
feminist. According to classical probability, this cannot hold.
Busemeyer et al. (2011), Franco (2009) model this using the quantum formalism
described in Sect. 3.1. They consider the two questions A: “is Linda a feminist?”
and B: “is Linda a bank teller?”. The question A is represented by an observable
on a two-dimensional Hilbert space with eigenvectors |a0 and |a1 which have the
eigenvalues a0 = 0 (Linda is not a feminist) and a1 = 1 (Linda is a feminist), respec-
tively. Question B is similarly represented. Furthermore, Franco (2009) represent the
probability P(b j |ai ) as the probability that the answer to question B is b j , given that
the answer to A has already been determined to be ai . If this latter condition is the
case, then within this two-dimensional setting this means that the current state we are
in is |ai itself, and therefore P(b j |ai ) = |b j |ai |2 , following Eq. (10). Notice that
72 M. Lewis
this means that P(b j |ai ) = |b j |ai |2 = |ai |b j |2 = P(ai |b j ). Now as in Eq. (20),
the probability of the answer that Linda is a feminist and a bank teller, P(a1 -and-b1 )
is modelled as P(a1 )P(b1 |a1 ).
Franco (2009) then lays out a specific relation between the eigenvectors of  and
the eigenvectors of B̂, as follows.
The two observables  and B̂ both operate on the same two-dimensional Hilbert
space, and the eigenvectors of  can be transformed into the eigenvectors of B̂ via
a unitary transformation
√ √
e−iξ√ P(b1 |a1 ) e−iφ √1−P(b1 |a1 )
U=
−eiφ 1−P(b1 |a1 ) eiξ P(b1 |a1 )
Visually, we have a pair of orthonormal bases as given in Fig. 6, where |bi = U |ai .
Now, given this representation, the state of knowledge of the agent is represented
as a vector:
|s = P(a0 )|a0 + eiφa P(a1 )|a1
which gives the correct probabilities for the agent to judge whether or not Linda is a
feminist.
Expressing this vector in the basis of B̂, we obtain
Quantum Computing and Cognitive Simulation 73
| a1⟩ feminist
P( feminist and then bank teller)
P(bank teller) | s⟩
Fig. 7 Starting in the state |s, the probability that Linda is judged to be a bank teller is less than
that the probability that Linda is judged first to be a feminist and then to be a bank teller
|s = P(a0 )U T |b0 + P(a1 )U T |b1 (23)
= P(a0 ) P(b1 |a1 )eiξ − P(a1 ) P(b1 |a0 )eiφa −φ |b0 (24)
+ P(a0 ) P(b1 |a0 )eiφ + P(a1 ) P(b1 |a1 )eiφa −ξ |b1 (25)
Now, calculating P(b1 ), i.e. the probability that the agent judges that Linda is a bank
teller, we get
P(b1 ) = | P(a0 ) P(b1 |a0 )eiφ + P(a1 ) P(b1 |a1 )eiφa −ξ |2 (26)
= P(a0 )P(b1 |a0 ) + P(a1 )P(b1 |a1 ) + 2 P(a0 )P(a1 )P(b1 |a0 )P(b1 |a1 )cos(φ + ξ + φa )
(27)
The last term in this sum is called the interference term I (s, A). If P(a0 )P(b1 |a0 ) +
I (s, A) is negative, then we obtain that P(b1 ) < P(a1 and then b1 ), i.e. that the prob-
ability that Linda is a bank teller is judged less likely than the probability that Linda
is a feminist and a bank teller. This can be understood visually in Fig. 7.
This calculation is expressed succinctly in Busemeyer et al. (2011) as follows.
Consider that the event of answering ‘yes’ to “is Linda a feminist?” is represented by
the projector PF , and that the event of answering ‘yes’ to “is Linda a bank teller?” is
given by PB . The projectors corresponding to the answer ‘no’ to each question are
given by P¬F = 1 − PF and P¬B = 1 − PB , respectively.
Now, as in Eq. (20), given a state |s we have
We can compare this value with the value P(B) = ||PB |s||2 as follows:
P(bank teller) < P(feminist and bank teller) < P(feminist or bank teller) < P(feminist)
Busemeyer et al. (2011) show that their probability model can also account for
this phenomenon. They consider the probability of answering ‘no’ to the question
“is Linda neither a bank teller nor a feminist?”. This is modelled as the probability
1 − ||P¬F P¬B ||2 . Notice that here, the order of the projectors is such that the state is
first projected onto P¬B —this is because this subspace is more probable. Consider
||P¬F |s||2 = ||P¬F P¬B ||2 + ||P¬F PB ||2 + s¬F B |s¬F¬B + s¬F¬B |s¬F B (33)
Again, if the interference term s¬F B |s¬F¬B + s¬F¬B |s¬F B is sufficiently negative,
we have ||P¬F PB ||2 + s¬F B |s¬F¬B + s¬F¬B |s¬F B < 0 and then ||P¬F |s||2 <
||P¬F P¬B ||2 , as needed for the disjunction fallacy.
Whilst this model covers a number of kinds of probability judgement error, one phe-
nomenon that is not covered is the phenomenon of double overextension. This occurs
when the probability of an item belonging to a conjunction of concepts is judged
to be higher than the probability of the item belonging to either of the constituent
concepts. In the Linda story, this would mean that the probability of Linda being
judged to be a feminist and a bank teller would not only be judged higher than the
Quantum Computing and Cognitive Simulation 75
probability of her being a bank teller, but also higher than her being judged to be a
feminist.
A quick look at the formalism shows that this cannot hold, at least as long as we
assume the order of conjunction is fixed.
As long as we hold the order fixed, we cannot have that both P(F) and P(B) are
lower than P(F and then B).
However, this kind of double overextension has been empirically observed, as
described in Sect. 2.6. In Sect. 3.8, we will describe an approach in which the problem
of double overextension can be modelled.
The ‘sure thing’ principle states that if an agent would take a specific course of action
C in all possible states of the world, then they should take action C when they do
not know what the state of the world is. However, as described in Sect. 2.5, humans
usually do not behave like this. This is seen in a prisoner’s dilemma situation, the
situation where we are choosing whether to book a holiday, or a situation where we
are choosing whether to make a bet, as well as many more. Pothos and Busemeyer
(2009) show that this phenomenon can be modelled using the same kind of quantum
similarity model described in Sect. 3.2. Other approaches include that in Khrennikov
and Haven (2009).
The model is applied to a prisoner’s dilemma situation. The situation can be
modelled as the tensor product of one space modelling our opponents’ actions, and
one space modelling our actions. Each space is two-dimensional, with one dimension
corresponding to the ‘cooperate’ (C) action, and the other corresponding to the
‘defect’ (D) action. The agent’s state at the start of the game is therefore represented
as the state
1
|ψ0 = |D D + |DC + |C D + |CC (36)
2
If the opponent is known to defect, then the state changes to
|ψ D = α D D |D D + α DC |DC (37)
where α2D D + α2DC = 1. On the other hand, if the opponent is known to cooperate,
the state changes to
|ψC = αC D |C D + αCC |CC (38)
76 M. Lewis
Now, if we know that our opponent defected, then our probability of defection is
where |α| indicates the absolute value of α. However, if we know that our opponent
cooperated, then our probability of defection is
We model not knowing whether the opponent has defected or not as the superposition
|ψ = √12 (|ψ D + |ψC ) When we are in this state of ignorance, we have
1 ∗
ψ|M|ψ = (α D D| + α∗DC DC| + αC∗ D C D| + αCC
∗
CC|)(|D DD D| + |C DC D|)
2 DD
(44)
(α D D |D D + α DC |DC + αC D |C D + αCC |CC) (45)
1 2
= (α D D + α2DC ) + Re(α D D D D|α DC DC) (46)
2
Again, this last term is an interference term, and may be negative. This means that
we can obtain the situation that the probability of defecting is smaller in the state of
ignorance than it is in the case that we know what the action of our opponent is.
p(Ax and then By) = ||PAx |s||2 ||PBy |s Ax ||2 = ||PBy PAx |s||2 (47)
An order effect for a given question occurs when the probability of giving an
answer to a question in the comparative context differs from the probability of giv-
ing that answer in the non-comparative context. In the Clinton/Gore setting, we
have that the total probability T PC y of answering ‘yes’ to the question “is Clinton
trustworthy?” in the comparative context is given by
However, as we saw in Sect. 3.4, the probability of answering ‘yes’ in the non-
comparative context is given by
This last quantity p(C y and then Gn) + p(Cn and then Gy) − p(Gy and then Cn)
− p(Gn and then C y) is termed q.
More generally, take the quantity p AB = p(Ay and then Bn) + p(An and then
By) to be the probability that the answers to the two questions are different in the AB
order, and take p B A = p(By and then An) + p(Bn and then Ay) in the B A order.
Then the quantum model predicts that
q = p AB − p B A = 0 (61)
Wang et al. (2013) test this theory on the data from Moore (2002) (described in
Sect. 2.4) using the sample frequencies to compute the values p AB and p B A , and then
using a z-test to compute whether q is significantly different from 0.
The conclusions are that the quantum question order model fits the data extremely
well. The three datasets exemplifying the consistency, contrast, and additive effects
have q-values that are not at all significantly different from 0. The fourth dataset
(Rose-Jackson) was not well modelled by the QQ model. However, the authors
argue that there is a violation of the assumptions in this model. Specifically, in this
case, additional information is supplied about each of the characters, and this entails
that the value θ is not the same in both cases. Hence, the cancellations worked out
in Eq. (56) do not hold in this case.
A crucial part of the argument in this paper is also that other probabilistic models
cannot account for both the question order effects seen and the QQ equality. In the
case of a Bayesian probability model, the assumption is that the event Ay is a set
contained in a sample space, and that By is another set contained in the same sample
space. The probability of the conjunction is then defined by the intersection of the
two events, which is commutative. In order to introduce order effects, two new events
must be introduced: an event O1: “A was asked before B”, and an event O2: “B was
asked before A”. However, in general, introducing these two events does not predict
the QQ equality to obtain.
Markov models can incorporate question order effects, since they can encode the
memory of previous events into the state representation. However, the authors argue
that in general, QQ equality is not upheld.
Quantum Computing and Cognitive Simulation 79
3.7 Contextuality
However, within quantum theory this inequality is violated. Consider the quantum
state |ψ = |01+|10
√
2
, and suppose Alice can perform measurements A = Z 1 or A =
X 1 on the first qubit and Bob can perform measurements B = −Z√2 −X
2
2
or B = Z 2√−X
2
2
1 1 1 1
AB = √ , A B = √ , A B = √ , AB = − √ (65)
2 2 2 2
√
meaning that E(AB + A B + A B − AB ) = 2 2, and violating the inequality in
(64).
The violation of this inequality has been demonstrated experimentally, and shows
that either there is no set of definite values of the random variables A, A , B, and
B that exist independent of our observation, or that the act of Alice’s measurement
affects the act of Bob’s measurement.
The notion of contextuality also arises within cognitive science. Contextual-
ity is crucially different from context-dependence. The phenomenon of context-
dependence arises when the answer to a question depends on the context that the
question is asked. This kind of context-dependence is called direct influences (Dzha-
farov & Kujala, 2016). However, another kind of context-dependence that is analo-
gous to quantum contextuality may be present. In Basieva et al. (2019), the following
example is given. Suppose we have the three questions
1. Do you like chocolate?
2. Are you afraid of pain?
3. Do you see your dentist regularly?
Imagine that you are asked a pair of these questions. It may well be that the context in
which you are asked a given question directly influences your answer to that question.
So if you are asked (1) and (2) together, you may respond positively to (1), but if you
are asked (1) and (3) together, the thought of going to the dentist may incline you to
answer negatively to (1). This is the notion of direct influence. Basieva et al. (2019),
following Dzhafarov and Kujala (2016), use the following notation: a response to
j
question qi asked in context c j is a random variable denoted Ri , and described by
j Y es N o response
Ri : j j (66)
pi 1 − pi probability
These random variables can be laid out as follows (Dzhafarov & Kujala, 2016):
R11 R21 c1
R22 R32 c2
(67)
R13 R33 c3
q1 q2 q3 system R3
This says that in context 1 (c1 ), questions 1 and 2 (q1 and q2 ) are asked, and so on.
Now, we might assume that all direct influences have somehow been eliminated,
so that you have the same probability of answering yes or no to a given question no
matter what the context. We then have that
Quantum Computing and Cognitive Simulation 81
However, we can also imagine a setup in which, if you answer yes to one question
in the context, you always answer no to the other question. Representing yes and no
by 1 and −1, we then have
This is saying that the pairs of random variables R11 and R13 , R21 and R22 , R32 and
R33 cannot be identical. This kind of context-dependence is not reducible to direct
influences.
Within quantum theory, contextuality applies only to systems in which there is
no notion of direct influences. Measurements of a property in different conditions
are described by the same random variable. The theory developed in Dzhafarov
and Kujala (2016), called contextuality-by-default or CbD, can also take account
of the presence of direct influences, and is therefore an extension of the quantum-
theoretic notion. CbD is able to quantify how much the distributions of responses
to different questions differ in different contexts, and how much the identities of the
random variables differ in different contexts. By comparing these two quantities, it
is possible to show how much contextuality is present over and above the effect of
direct influences.
In previous work examining the contextuality of human judgements, the impact
of direct influences was not always considered (Aerts et al., 2013; Dzhafarov &
Kujala, 2014; Khrennikov et al., 2014), and so it was not clear the extent to which
true contextuality was actually present in human decision-making. For example, in
Bruza et al. (2012, 2015), whilst the effect of direct influences was taken into account,
the presence of true contextuality was only minimally detected. However, in Basieva
et al. (2019), Cervantes and Dzhafarov (2019), Dzhafarov (2019), experiments have
been carried out to show that this phenomenon is indeed present in human judgement.
82 M. Lewis
μ : × M × → [0, 1] (75)
which is a probability function that describes how a state p under the influence of
context e changes into state q.
which gives the weight of a property a for the concept in the state p. So, for example,
after considering the concept ‘pet’ in the context “the pet is chewing a bone”, the
weight of the property ‘is furry’ might increase.
ν( p, a) = p|Pa | p (78)
In Aerts and Gabora (2005a), a Hilbert space representation of the theory is given.
Concepts are represented in a Hilbert space in which the basis dimensions of the space
represent basic states of the concept. The basic states of a concept are defined via
basic contexts of a concept, which roughly means that the basic concept contexts are
the most specific. For example, we might have the contexts a1 = “the pet is chewing
a bone”, and a2 = “the pet is chewing a bone in the garden”, and we would say that
a2 is more specific than a1 . The set of atomic contexts is denoted by X , and these are
considered to be rank-1 projectors. The basic states are then the states corresponding
to these rank-1 projectors. The ground state p̂ of a concept is given by the following
Quantum Computing and Cognitive Simulation 83
so that the ground state has an equal probability of changing to any basic state under
one of the basic contexts.
General contexts are represented by projectors that are not merely of rank 1. So
suppose that we have 100 basic contexts, and 25 of these basic contexts are more
specific than ebone = the pet is chewing a bone, i.e. they are more detailed contexts
in which the pet is chewing a bone.11 Then we would build a projector corresponding
to the context ebone by summing together the projectors for the basic contexts in the
set E bone = {u ∈ X |u is more specific than ebone }:
Pbone = |uu| (80)
u∈E bone
The new state under the influence of this context is calculated in the standard way,
i.e. in the same way that the knowledge state of the agent is evolved in Sect. 3.1,
Eq. (18):
Pbone | p̂
| pbone = (81)
||Pbone | p̂||
In our example with 100 basic contexts and 25 of these basic contexts are more
specific than e1 , we have that
1
Pbone | p̂ = |uu|| p̂ = √ |u (82)
u∈E bone u∈E bone
100
1 25
||Pbone | p̂||2 = p̂|Pbone | p̂ = | p̂|u|2 = = (83)
u∈E bone u∈E bone
100 100
so that
1 1
| pbone = √ |u = |u (84)
u∈E bone 25 u∈E bone
5
Recall that in the pet fish problem, the problem is that the exemplar ‘goldfish’ is
viewed as more typical to the ‘pet fish’, conceived as the conjunction of two concepts
‘pet’ and ‘fish’ than it is to either of the constituent concepts. To describe the notion
of the typicality of an exemplar to a concept, Aerts and Gabora (2005a) use the
function μ described in Eq. (75). Aerts and Gabora (2005a) explicate the notion
11 Note that there might be basic contexts that are unrelated to “the pet is chewing a bone”, for
example, “the pet is chewing a cracker in a cage and scratching its wing”—although this might be
a basic context, it is not related to the context “the pet is chewing a bone”.
84 M. Lewis
Table 6 Numbers of basic states assigned to each exemplar in the concepts pet and fish
Pet Fish
Dog 50 0
Cat 35 0
Goldfish 10 10
Shark 5 20
Tuna 0 20
Earlier, we said that there are 25 out of 100 basic contexts in which the pet is
chewing a bone. Suppose there are 10 contexts where the pet is a goldfish, but only
1 context where the pet is chewing a bone and is a goldfish. Then we will have
1
μ( pgf , egf , pbone ) = pbone |Pgf | pbone = v|uu|w (86)
u∈E gf v∈E bone w∈E bone
25
1 1
= = (87)
u∈E bone ∩E gf
25 25
So the weight of the exemplar goldfish in the context the pet is chewing a bone is
1
25
.This quantity is what Aerts et al. (2015) use to formalize the notion of typicality
of an exemplar to a concept.
To examine the pet fish situation, we will first set up an example situation. In Aerts
and Gabora (2005a), these figures were generated from a psychological experiment
involving asking humans for the typicality figures. In this chapter, we simplify the
figures for explanatory purposes.
Suppose that we have a pool of 100 basic states for pet and a pool of 50 basic
states for fish. The basic states are distributed amongst the following exemplars as in
Table 6.
For each concept, we outline how many basic contexts are included in each of the
x
contexts epet = the x is a pet, efish
x
= the x is a fish, efurry
x
= the x is furry, and eswim
x
=
the x is swimming around in a glass tank in the house.
Quantum Computing and Cognitive Simulation 85
Table 7 Number of basic contexts assigned to each context for each exemplar in the concept pet
pet pet pet pet
epet efish efurry eswim
Dog 50 0 45 0
Cat 35 0 30 0
Goldfish 10 10 0 9
Shark 5 5 0 5
Tuna 0 0 0 0
Table 8 Numbers of basic contexts assigned to each context for each exemplar in the concept fish
fish fish fish fish
epet efish efurry eswim
Dog 0 0 0 0
Cat 0 0 0 0
Goldfish 10 10 0 9
Shark 5 20 0 5
Tuna 0 20 0 0
μ(goldfish, goldfish, pet) = pet|Pgoldfish | pet = pet| |uu|| pet (88)
u∈E goldfish
1 10
= v|uu|w = (89)
100 u∈E v∈E pet w∈E pet
100
goldfish
20
A similar calculation gives us that the typicality of goldfish in the concept fish is 100 .
pet fish pet fish
The shared contexts of pet and fish are efish , epet , eswim , and eswim . Now, to form com-
binations of concepts, such as pet fish, the concept is represented in the tensor product
space of the two concepts pet and fish. The basic contexts of the concept pet fish are
denoted as E pet fish which is the intersection of the set E pet of the basic contexts of pet,
and the set E fish of the basic contexts of fish. The assumption is made for the exam-
pet fish pet pet
ple that E pet fish = E fish E pet , where E fish = {u ∈ E pet |u is more specific than efish },
i.e. it is the set of basic contexts of the concept pet that are more specific than
the pet is a fish. Note that in Tables 7 and 8, the cardinality of that set is 15.
The concept pet fish is modelled as
1
|pet fish = √ |u ⊗ |u (90)
u∈E pet fish
15
86 M. Lewis
pet
G = Pgoldfish ⊗ 1fish = |uu| ⊗ I (91)
pet
u∈E goldfish
1 10
= v|uu|vv|w = (94)
15 pet
15
v∈E pet fish u∈E goldfish w∈E pet fish
We can see that the typicality of goldfish in pet fish at 23 is much higher than the
1 2
typicality of goldfish in either pet ( 10 ) or fish ( 10 ).
Moreover, although we have only applied the context goldfish to the pet subsystem,
the fish subsystem has changed as well:
1 1
= √ |uu|w ⊗ |w = √ |u ⊗ |u (96)
10 pet pet fish 10 pet
u∈E goldfish w∈E u∈E goldfish
If we constructed a similar context saying that the fish is a goldfish, the concept would
collapse to the same state. In the pet fish concept, saying that the pet is a goldfish has
the same effect as saying that the fish is a goldfish—and this is the desired outcome.
In contrast, if the concept pet fish was modelled without entanglement, for example,
as
Then
G|pet fish = |uu| ⊗ I |v ⊗ |w (99)
pet pet fish
u∈E goldfish v∈E fish w∈E pet
= |u ⊗ |w (100)
pet fish
u∈E goldfish w∈E pet
where although the pet subsystem has changed to a goldfish, the fish subsystem has
not. If we constructed a similar context saying that the fish was a goldfish, we would
not obtain the same state. We would instead have a state consisting of a generic
pet, and a goldfish. As such, the entanglement is essential to correctly model this
problem.
Aerts and Gabora (2005b) go on to describe that this approach could be extended
to model the composition of concepts in more complicated structures, up to the
sentence level.
Whilst the approach given in Aerts and Gabora (2005b) is able to model the
phenomenon of overextension, it suffers from the problem that the space in which
concepts must be modelled gets longer dependent on the phrase or sentence length.
In Coecke et al. (2010), an approach to modelling language was proposed that com-
bines the very successful distributional models of meaning with the compositional
approaches from formal semantics. In brief, this approach models words in differing
vector spaces based on their grammatical type. Nouns are modelled in a vector space
N for nouns, and sentences are modelled in a vector space S for sentences. Adjectives
are modelled as linear maps on the noun space, i.e. matrices in N ⊗ N , and transitive
verbs are bilinear maps from two copies of the noun space to the sentence space, or
tensors in N ⊗ S ⊗ N . The benefit of this approach is that arbitrarily long sentences
can all be mapped into the same vector space S. Whilst this approach was originally
developed with an eye to natural language processing, it has also been used to tackle
problems in cognitive science.
In Coecke and Lewis (2015), the compositional distributional model is used to
examine the pet fish problem, and more generally the problem of overextension.
The first approach to tackling the pet fish problem is to view the word pet as an
adjective, rather than a noun. Then the parameters of the matrix can be tuned so
that the similarity of the vector goldfish to pet fish is greater than the similarity to
either pet or fish. The sentence can further be expanded out to look at the conjunc-
tions pet which is a fish and fish which is a pet, and this format is used to model the
importance of attributes in the conjunction of concepts. In Hampton (1987), data is
collected on the importance of attributes in concepts and their combination. Con-
cepts are considered in pairs that are related to some degree, for example, ‘Pets’ and
‘Birds’. Six pairs are considered in total. Participants are asked to generate attributes
for each concept and for their conjunctions, where conjunction in this case is ren-
dered as “Pets which are also Birds”, or “Birds which are also Pets”. For example,
attributes such as ‘lives in the house’, ‘is an animal’, and ‘has two legs’ are generated
for ‘Pets’, ‘Birds’. For each pair of concepts and their conjunction, attributes that
88 M. Lewis
had been generated by at least 3 out of the 10 participants were collated. Participants
were then asked to rate the importance of each attribute to each concept and to each
conjunction. Importance ratings were made on a 7-point verbal scale ranging from
“Necessarily true of all examples of the concept” to “Necessarily false of all exam-
ples of the concept”. Numerical ratings were subsequently imposed ranging from 4
to −2, respectively.
The question then arises of how the importance of attributes in the conjunction
of the concepts is related to the importance of attributes in the constituent concepts.
Key phenomena are that conjunction is not commutative, that inheritance failure can
occur (i.e. an attribute that is important in one of the concepts is not transferred to the
conjunction), that attribute emergence can occur, where an attribute that is important
in neither of the conjuncts becomes important in the conjunction, and further, that
necessity and impossibility are preserved. In order to model this data, Hampton uses
a multilinear regression.
In Coecke and Lewis (2015), the phrase A which is a B is modelled as
1
μA or B (X ) = (A| + B|)M X (|A + |B)
2
1
= (A|M X |A + B|M X |B + Re(A|M X |B)
2
1
= (μA (X ) + μB (X )) + Re(A|M X |B)
2
where this last term is the interference term, as in Sect. 3.4.
On the other hand, the disjunction of the two concepts can be modelled in two
copies of H, as |A ⊗ |B. The operator M X ⊗ M X is a projection operator over the
space that asks whether X is a member of A and whether X is a member of B. To
determine the membership of X in A or B, we therefore calculate
and the conjunction in one copy of H is modelled as √12 (|A + |B) again. This
means that in the single space, conjunction and disjunction are the same.
The single space and the double space are combined in a Fock space H ⊗ H ⊕ H.
The concept overall is represented as
neiφ
ψ(A, B) = meiθ |A ⊗ |B + √ (|A + |B)
2
where m 2 + n 2 = 1. Then,
μA and B (X ) =
neiφ neiφ
meiθ A| ⊗ B| + √ (A| + B|) M X ⊗ M X ⊕ M X meiθ |A ⊗ |B + √ (|A + |B)
2 2
1
= m 2 μ A (X )μ B (X ) + n 2 ( (μ A (X ) + μ B (X )) + Re(A|M X |B))
2
1
μA or B (X ) = m 2 (μ A (X ) + μ B (X ) − μ A (X )μ B (X )) + n 2 ( (μ A (X ) + μ B (X )) + Re(A|M X |B))
2
90 M. Lewis
A|I − M X |A = 1 − μ A (X )
This system is able to successfully model data from Hampton (1997) and from Sozzo
(2015).
A key part of human concept use is how we express these concepts in language.
However, when concepts are expressed in language, the language can have one word
which refers to the same concept. A well-used example is the word bank. If I say
“I’m going to the bank to deposit some money”, then bank has a different meaning
than if I say “I’m going to the bank to fish”. In Bruza et al. (2012), an experiment
is performed that looks into pairs of such ambiguous words and their combinations.
An example is apple chip. This could mean a part of a Macintosh computer, or a
slice of dried fruit. If I interpret apple in the sense computer, then it is very unlikely
that I should interpret chip as a slice, or part, or the computer. Similarly, if I interpret
apple in the sense fruit, then it is very unlikely that I will interpret chip in the sense
of a computer chip.
A quantum modelling of this phenomenon would seem very appropriate: if we
view each individual concept as modelled in a Hilbert space H, and I disambiguate
the meanings so that I have 4 vectors |applecomputer , |applefruit , |chipcomputer , and
|chipslice , then we can model the combined concept in H ⊗ H as
Quantum Computing and Cognitive Simulation 91
Given a state |ψ that the agent is in, the interpretation of the combination will
collapse to either the computer sense or the fruit sense.
However, another form of combination is as follows. We could model
apple as |apple = √12 |applecomputer + |applefruit , and |chi p = √12 |chipcomputer +
|chipslice
In this case, a participant could interpret the combination apple chip in any of the
four possible combinations.
This kind of phenomenon is investigated in Bruza et al. (2015) In these experi-
ments, the participant is first primed for one of the senses of the combination—for
example, the word banana is shown to them before they are asked to make a decision
on the interpretation of the compound. The participants are then asked to describe
the interpretation they have made. These interpretations are used to build a set of
probabilities for the interpretations of the compounds, and these probabilities are
analysed to determine whether the compound is separable or entangled. There are
a number of factors which need to be accounted for in this, which are discussed in
Dzhafarov and Kujala (2016) and which we discussed in Sect. 3.7.
The results are that in general, there is no clear evidence for the entanglement of
these concepts in the human mental lexicon. However, in a number of cases, there
is some evidence of phenomena occurring which are not completely explained by a
classical probabilistic model.
Another approach to modelling the ambiguity of words is to model words as
density matrices (Meyer & Lewis, 2020; Piedeleu et al., 2015). This was proposed
as an extension of the compositional distributional model of Coecke et al. (2010). As
mentioned in Sect. 3.1, a density matrix is a probability distribution over a number
of quantum states:
ρ= αi |vi vi | (104)
i
where ∀i, αi > 0 and i αi = 1. The idea of using density matrices to model ambi-
guity runs as follows. Each sense of a word is modelled as a vector, and then these
senses are mixed together following Eq. (104) above. Crucially, the same composi-
tional framework as in the standard vector-based model is available. This means that
as words are used in context, their meaning can be disambiguated, in the same way
that humans are able to. In Meyer and Lewis (2020), a neural approach to building
density matrices for words is proposed and tested on a set of sentence similarity
92 M. Lewis
In the foregoing sections, the level of modelling was at a high, cognitive level. There
is no attempt to form any match between these models and actual processes in the
brain. However, there has been a range of research into using quantum theory to
model cognitive processes at a level that is closer to a neuronal level.
Fuss and Navarro (2013) describe a quantum model of two-alternative forced
choice (2AFC) experiments. In this sort of experiment, the participant is asked to
make a choice about a stimulus. For example, there may be two shapes on a computer
screen and the participant is asked which one is a square. The participant must choose
one or the other, and does not have to choose to say neither. Information about the
nature of processing is inferred from reaction times. For example, the average reaction
time to the image in Fig. 8a may be faster than the average reaction time to the image
in Fig. 8b.
A standard way of modelling reaction times in this sort of experiment is to say that
a pair of neurons, or groups of neurons, is responsible for accumulating evidence for
one choice or the other, and that when the evidence hits a threshold, the participant
reacts. The two accumulators are in a race to get to the threshold. So in the examples in
Fig. 8, evidence is being accumulated by one accumulator for left and one for right.
In Fig. 8a, evidence for right would only accumulate slowly, and so the reaction
left would happen more quickly, whereas in Fig. 8b, more of the evidence would
be added to the right accumulator, and so overall the threshold would take longer to
reach. This is usually modelled as a random walk with drift. An extension to a simple
random walk model is to consider that multiple random walks could be happening in
parallel. However, having multiple random walks happening in parallel can still only
produce a model where evidence is accumulated. On the other hand, a model that
explicitly models neurons and the connections between them (connectionist models)
can account for inhibitory connections, i.e. that evidence for one choice can inhibit
the accumulation of evidence for the other choice.
In Fuss and Navarro (2013), a parallel random walk model using quantum the-
ory is proposed. This allows for the kinds of inhibition we just mentioned to be
modelled, by means of interference. This kind of model can produce qualitatively
different probability distributions over the search space, allowing for more efficient
decision-making to be carried out. The authors show that this kind of model can fit
2AFC experimental data more closely than a simple standard random walk with drift
(although there are more complex models which may do better).
More recently, Khrennikov et al. (2018) have begun to formulate a description
of neuronal substrates within quantum terms. This does not claim that neuronal and
cognitive activities are being modelled at the quantum level (although models to
explain cognitive activity through actual quantum modelling has been proposed in
Hameroff (2014), Penrose (1990)). Rather, the idea is that quantum theory can be
used to represent the information and activity present in neuronal assemblies.
s = v1 ⊗ r0 + v2 ⊗ r1
s = v1 ⊗ r0 + (v3 ⊗ r0 + v4 ⊗ r1 ) ⊗ r1
= v1 ⊗ r0 + v3 ⊗ (r0 ⊗ r1 ) + v4 ⊗ (r1 ⊗ r1 )
≡ v1 ⊗ r0 + v3 ⊗ r01 + v4 ⊗ r11
A notable feature of this representation is that the vector space in which concepts
live must be arbitrarily large, depending on the size of the structure to be represented.
Symbols at depth d in a binary tree are realized by S(d) , the F R d -dimensional vector
space formed from vectors of the form f ⊗ ri ⊗ r j ⊗ · · · ⊗ rk with d role vectors,
where F is the dimension of the filler vectors f and R is the dimension of the individual
role vectors ri . A vector space containing all vectors in S(d) for all d is
Vectors s(i) are embedded into this space, meaning that the normal operation of vector
addition can be used to combine sentence components.
Once a set of concepts (e.g. Jen, loves, and Junpa) have been combined into
one vector representation, it is also possible to pull the representation apart, and
recombine it. This is done by means of an unbinding mechanism. Suppose we form
the sentence representation
we can ask “who is the subject of the sentence” by unbinding |subj from the
sentence. If role vectors are all orthogonal, then the unbinding is simply done by
taking the inner product of |S with the role vector: S|subj = |J en.
Tensor product networks have recently been used extensively to examine the
structure of recurrent neural networks (McCoy et al., 2019), and also as a potential
substrate for quantum language processing (Wiebe et al., 2019). Moreover, very
similar structures have been used in the neural-based computing framework Nengo
(Bekolay et al., 2014). These latter structures are different in that they use circular
convolution (Plate, 1995) rather than tensor product to bind roles and fillers. The
benefit of using circular convolution rather than the tensor product is that all vectors
are mapped into one shared space. An approximate unbinding mechanism can be
employed by using circular correlation with a query vector.
Tensor product representations have also been used in Pothos and Truebloodm
(2015) to model how the similarity of objects can depend on where the similarity
is to be found. For example, given two birds with a spotted pattern, if the spotted
pattern is found on the wings of one bird and the back of the other, they will not be
judged as similar as if they both have the spotted pattern on their wings. The use of
role vectors in the tensor product representations of Smolensky (1990), Smolensky
and Legendre (2006) is leveraged to account for this phenomenon.
The compositional distributional model Coecke et al. (2010), mentioned in Sect.
3.8, has the property that all representations are mapped into one shared space, and
shares some structures similar to the ICS model. In Al-Mehairi et al. (2017), it
is shown that the ICS model can be mapped onto the compositional distributional
model, to form a categorical compositional model of cognition.
A further mapping from ICS to the compositional distributional model can be
considered. One of the most useful aspects of the compositional distributional model
is the way that functional words such as adjectives and verbs modify the nouns
they are applied to, either by mapping them to another vector in the noun space,
or by mapping them into a sentence space. Within the ICS model, this does not
happen in the same way. Consider the following problem, as investigated in Martin
and Doumas (2020). Under the assumption that the concepts we use are expressed
by the language we use, the way that we compose concepts can be examined by
looking at the properties of words and phrases. Within the field of neural modelling,
a key question is how neural representations of two concepts can be composed into
a single combined representation. Neural representations of concepts are commonly
modelled as (normalized) vectors, and similarity between concept representations
is calculated by the inner product. Two common proposals for how to combine
neural representations are via firstly tensor product binding or holographic reduced
representations, as described above, or secondly vector addition (Hummel, 2011).
The problem with tensor product binding is that it does not transform the argument
of a relation in the necessary way. Consider vectors |fluffy, |cat, and |dog. If the
combinations fluffy dog and fluffy cat are represented by the tensor product of the
two component vectors:
96 M. Lewis
then the similarity between fluffy cat and fluffy dog is just the similarity between cat
and dog, since
On the other hand, vector addition has been shown to be successful in modelling
concept combination Hummel (2011), Martin and Doumas (2020), and has a clear
neural correlate, termed synchronous binding.12,13,14,15
However, it is arguably not able to express gradations in meaning that are needed
in terms like red car versus red wine. For example, given |red, |wine, and |car,
suppose that wine and car are not very similar, say wine|car = 0.1, and they are
each fairly similar to red, say red|car = red|wine = 0.5. Then, adding vectors
and normalizing:
|red + |car
red car = √
3
|red + |wine
red wine = √
3
and
12 https://commons.wikimedia.org/wiki/File:Wayfield%27s_Young_Argos,
_the_Labrador_Retriever.jpg. Attribution: Andrew Skolnick, en:User:Askolnick, CC BY-SA
3.0. http://creativecommons.org/licenses/by-sa/3.0/, via Wikimedia Commons.
13 https://commons.wikimedia.org/wiki/File:Cat_November_2010-1a.jpg. Attribution: Alvesgas-
that fluffy cats can be moved closer to fluffy dogs, i.e. the problem in Eq. (105) is
not encountered. On the other hand, the application of an adjective like red does not
force red car to become similar to red wine. How? Essentially, the use of a matrix
gives enough ‘space’ for different uses of words to be represented.
Another line of work looks at using vector space methods to model structure in
concepts. A key characteristic of the concepts we use is that they are hierarchical.
Besides categorizing individual entities, we want to talk about types of entities—for
example, we want to simply say that tigers are dangerous, rather than “that tiger
is dangerous. And that tiger is dangerous. And that one. And...”. Within a quantum
model, we can start doing this by considering concepts as subspaces of a vector space,
and using the lattice structure of the subspaces, or equivalently their projectors, to
reason with. This was originally proposed by Birkhoff and Von Neumann (1936),
and was applied in an information retrieval setting by Widdows and Peters (2003).
In that work, the authors describe a means of applying negation and conjunction in
document queries, based on quantum logic.
The density matrix variant of the compositional distributional model of Coecke
et al. (2010) forms an extension of the quantum logic formed by the lattice of projec-
tors. Words are modelled as positive operators, and hierarchy in concepts is modelled
via the Löwner order. In Bankova et al. (2019), Sadrzadeh et al. (2018), it is shown
that the Löwner order is compatible with the compositional model used. In Lewis
(2019), a concrete modelling of words as positive operators is given which is shown
to model human judgements of entailment very well. Further work into modelling
negation is underway (Lewis, 2020), and work on modelling other forms of logical
structure such as generalized quantifiers has also been undertaken by Hedges and
Sadrzadeh (2019), Dostál et al. (2021).
4 Concluding Remarks
as the similarity of Korea to China. However, in the model of the conjunction fallacy,
this is interpreted as the conjunction of A and then B, as it also is in the question
order model.
The conjunction and negation model proposed by Aerts and Gabora (2005b) also
models concepts as subspaces of a vector space. Here, the subspaces have very high
dimensionality and are based on very specific contexts that objects are seen in. The
notion of entanglement plays a key part in modelling the combination of concepts
such as pet fish. Aerts (2009), Aerts et al. (2015), Sozzo (2015) model concepts as
states within a vector space, and again use the notion of entanglement. In this latter
work, the vector spaces used are smaller—for example, around eight-dimensional.
A drawback of these compositional methods is that in all cases, the size of the
underlying space grows as more concepts are added to the combination. This has at
least two drawbacks: for more complex concepts, spaces may grow unfeasible large,
and furthermore it means that combinations of unequal numbers of concepts are not
easy to compare.
Various papers have examined whether a form of contextuality is present in human
concept use. In some cases, a claim of contextuality is erroneous, as argued by
Dzhafarov and Kujala (2016). On the other hand, Basieva et al. (2019), Dzhafarov
(2019) have shown that experiments can be designed in which human decision-
making does show contextuality. If human decision-making does indeed display true
contextuality, then the use of quantum computing to model these forms of decisions
may provide a very useful substrate. However, in many cases there is no evidence of
contextuality over and above the effect of direct influences, a more straightforward
form of context-dependence.
Vector-based models of meaning and compositionality have been extremely suc-
cessful in modelling language (Coeckeetal., 2010; Kartsaklis & Sadrzadeh, 2016;
McCoy et al., 2019; Sadrzadeh et al., 2018; Smolensky, 1990). These models use the
same fundamental structures as quantum theory, and extensions of the compositional
distributional model have incorporated crucially quantum phenomena such as den-
sity matrices Bankova et al. (2019), Meyer and Lewis (2020), Piedeleu et al. (2015),
Sadrzadeh et al. (2018). The models proposed in Smolensky (1990) can be argued
to have relevance for realistic neuronal modelling (Bekolay et al., 2014). Moreover,
theory from Smolensky (1990) has been developed for potential implementation on
quantum computers in Wiebe et al. (2019), and the model from Coecke et al. (2010)
has actually been implemented on a modern quantum computer (Lorenz et al., 2021).
This implementation takes advantage of the fact that the fundamental structures in
this model are the same as those used in quantum theory. In Lorenz et al. (2021),
experiments on sentence classification are run, one to classify sentences as belonging
to one of two topics, and a second to classify noun phrases as containing either a
subject relative clause or an object relative clause. The general methodology of the
experiments is to first generate a linguistic parse of a given sentence, and to express
the sentence in terms of a DisCoCat diagram (as in Coecke et al. (2010)). The diagram
is rewritten to eliminate complex aspects of the diagram—namely by straightening
wires in the diagram in a certain way. Following this step, it is then possible to build a
quantum circuit whose structure matches the grammatical structure of the sentence.
100 M. Lewis
A quantum compiler translates the circuit into instructions for a physical quantum
computer, which runs the circuit a certain number of times, and the relative frequency
of given answers can be recorded.
So what is the potential for cognitive science within the realm of quantum computing?
On the one hand, we can simply say that human cognitive faculties are not computers,
whether classical or quantum. However, this is unnecessarily negative. There are a
wide range of cognitive processes that are well modelled using quantum theory.
Models of similarity, judgement, and decision-making moreover seem to be well
modelled even in fairly low-dimensional spaces.
At the same time, models that incorporate compositionality at a more complex
level, using the grammar of natural language, have been implemented on quantum
computers, and the potential for the benefits of this are large. The successes in Lorenz
et al. (2021) have the potential to be extended to the cognitive domain. Drawing
together some of the theory from the similarity and decision-making models to
form a unified whole with models of language is an important way forward for this
endeavour. One way to progress this would be to consider a form of compositionality
that can be applied at the cognitive, rather than the language level, like a language
of thought (Fodor, 1975).
Fodor’s language of thought hypothesis is summarized (by himself (Fodor &
Pylyshyn, 1988)) as the idea that cognition is reliant on mental representations, and
that those mental representations have a combinatorial syntax and semantics. This
argument is based on the fact that human cognitive abilities have three key properties:
• Productivity: We are able to entertain a large number of thoughts and to generate
these on the fly, as combinations of representations that we already have.
• Systematicity: This is the property that if, for example, “Junpa loves Jen” can be
understood, then “Jen loves Junpa” should also be understood, whether or not it
is true.
• Compositionality: This is the property that the meaning of complexes of repre-
sentations should be understandable from the meanings of the individual repre-
sentations and how they are combined.
Whilst natural language arguably has these properties, and furthermore is intrinsically
related to our thoughts, we do not have a ‘grammar’ that we can apply to thoughts to
combine them in the same way that grammar exists for natural languages. Developing
such a model together with the right sort of representations and notions of similarity
and decision-making has the potential to form a very powerful model of cognition.
Utilizing the compositional approach proposed in Coecke et al. (2010) may provide
a way forward for a form of cognitive architecture that can be implemented on a
quantum computer.
Quantum Computing and Cognitive Simulation 101
References
Gayler, R. W. (2003, July). Vector symbolic architectures answer Jackendoff’s challenges for cog-
nitive neuroscience. In Joint International Conference on Cognitive Science (p. 6). https://arxiv.
org/abs/cs/0412059
Hameroff, S. (2014, January). Quantum walks in brain microtubules-a biomolecular basis for quan-
tum cognition? Topics in Cognitive Science, 6(1), 91–97. ISSN 17568757. https://doi.org/10.
1111/tops.12068
Hampton, J. A. (1987, January). Inheritance of attributes in natural concept conjunctions. Memory
& Cognition, 15(1), 55–71. ISSN 1532-5946. https://doi.org/10.3758/BF03197712
Hampton, J. A. (1988a, November). Disjunction of natural concepts. Memory & Cognition, 16(6),
579–591. ISSN 1532-5946. https://doi.org/10.3758/BF03197059
Hampton, J. A. (1988b). Overextension of conjunctive concepts: Evidence for a unitary model of
concept typicality and class inclusion. Journal of Experimental Psychology: Learning, Memory,
and Cognition, 14(1), 12–32. ISSN 1939-1285(Electronic),0278-7393(Print). https://doi.org/10.
1037/0278-7393.14.1.12
Hampton, J. A. (1997, November). Conceptual combination: Conjunction and negation of natural
concepts. Memory & Cognition, 25(6), 888–909. ISSN 0090-502X, 1532-5946. https://doi.org/
10.3758/BF03211333. http://link.springer.com/10.3758/BF03211333
Hedges, J., & Sadrzadeh, M. (2019, June). A generalised quantifier theory of natural lan-
guage in categorical compositional distributional semantics with bialgebras. Mathemat-
ical Structures in Computer Science, 29(6), 783–809. ISSN 0960-1295, 1469-8072.
https://doi.org/10.1017/S0960129518000324. https://www.cambridge.org/core/journals/
mathematical-structures-in-computer-science/article/abs/generalised-quantifier-theory-of-
natural-language-in-categorical-compositional-distributional-semantics-with-bialgebras/
9738AE3A40B9A51AE2134D13E4FD3C5B
Hummel, J. E. (2011, June). Getting symbols out of a neural architecture. Connection Science,
23(2), 109–118. ISSN 0954-0091. https://doi.org/10.1080/09540091.2011.569880
Kartsaklis, D., & Sadrzadeh, M. (2016, December). Distributional inclusion hypothesis for tensor-
based composition. In Proceedings of COLING 2016, the 26th International Conference on
Computational Linguistics: Technical Papers, Osaka, Japan (pp. 2849–2860). The COLING
2016 Organizing Committee. https://www.aclweb.org/anthology/C16-1268
Khrennikov, A., Basieva, I., Dzhafarov, E. N., & Busemeyer, J. R. (2014, October) Quantum models
for psychological measurements: An unsolved problem. PLoS ONE, 9(10), e110909. ISSN 1932-
6203. https://doi.org/10.1371/journal.pone.0110909
Khrennikov, A., Basieva, I., Pothos, E. M., & Yamato, I. (2018, November). Quantum probabil-
ity in decision making from quantum information representation of neuronal states. Scientific
Reports, 8(1), 16225. ISSN 2045-2322. https://doi.org/10.1038/s41598-018-34531-3. https://
www.nature.com/articles/s41598-018-34531-3
Khrennikov, A. Yu., & Haven, E. (2009, October). Quantum mechanics and violations of the sure-
thing principle: The use of probability interference and other concepts. Journal of Mathematical
Psychology, 53(5), 378–388. ISSN 0022-2496. https://doi.org/10.1016/j.jmp.2009.01.007. http://
www.sciencedirect.com/science/article/pii/S0022249609000182
Lewis, M. (2019, September). Compositional hyponymy with positive operators. In Proceedings
of the International Conference on Recent Advances in Natural Language Processing (RANLP
2019), Varna, Bulgaria (pp. 638–647). INCOMA Ltd. https://doi.org/10.26615/978-954-452-
056-4_075. https://www.aclweb.org/anthology/R19-1075
Lewis, M. (2020, September). Towards logical negation for compositional distributional semantics.
IfCoLog Journal of Applied Logics, 7(5), 771–794.
Lorenz, R., Pearson, A., Meichanetzidis, K., Kartsaklis, D., & Coecke, B. (2021, February). QNLP in
practice: Running compositional models of meaning on a quantum computer. arXiv:2102.12846
[quant-ph]. http://arxiv.org/abs/2102.12846
Martin, A. E., & Doumas, L. A. A. (2020, February). Tensors and compositionality in
neural systems. Philosophical Transactions of the Royal Society B: Biological Sciences,
104 M. Lewis
Smolensky, P. (1990, November). Tensor product variable binding and the representation
of symbolic structures in connectionist systems. Artificial Intelligence, 46(1–2), 159–216.
ISSN 00043702. https://doi.org/10.1016/0004-3702(90)90007-M. https://linkinghub.elsevier.
com/retrieve/pii/000437029090007M
Smolensky, P., & Legendre, G. (2006, July). The harmonic mind. The MIT Press. https://mitpress.
mit.edu/books/harmonic-mind-2-vol-set
Sozzo, S. (2014, June). A quantum probability explanation in Fock space for borderline contradic-
tions. arXiv:1311.6050 [physics, physics:quant-ph]. http://arxiv.org/abs/1311.6050
Sozzo, S. (2015, June). Conjunction and negation of natural concepts: A quantum-theoretic model-
ing. Journal of Mathematical Psychology, 66, 83–102. ISSN 0022-2496. https://doi.org/10.1016/
j.jmp.2015.01.005. https://www.sciencedirect.com/science/article/pii/S0022249615000073
Tversky, A. (1977). Features of similarity. Psychological Review, 84(4), 327–352. ISSN 1939-1471
(Electronic), 0033-295X (Print). https://doi.org/10.1037/0033-295X.84.4.327
Tversky, A., & Kahneman, D. (1983). Extensional versus intuitive reasoning: The conjunction
fallacy in probability judgment. Psychological Review, 90(4), 293–315. ISSN 1939-1471 (Elec-
tronic), 0033-295X (Print). https://doi.org/10.1037/0033-295X.90.4.293
Tversky, A., & Shafir, E. (1992, September). The disjunction effect in choice under uncertainty.
Psychological Science, 3(5), 305–310. ISSN 0956-7976. https://doi.org/10.1111/j.1467-9280.
1992.tb00678.x
Wang, Z., & Busemeyer, J. R. (2013, September). A quantum question order model supported by
empirical tests of an a priori and precise prediction. Topics in Cognitive Science, n/a–n/a. ISSN
17568757. https://doi.org/10.1111/tops.12040
Wang, Z., Busemeyer, J. R., Atmanspacher, H., & Pothos, E. M. (2013, September). The potential
of using quantum theory to build models of cognition. Topics in Cognitive Science, n/a–n/a. ISSN
17568757. https://doi.org/10.1111/tops.12043
Widdows, D., & Peters, S. (2003). Word vectors and quantum logic experiments with negation and
disjunction (p. 14). Stanford University.
Wiebe, N., Bocharov, A., Smolensky, P., Troyer, M., & Svore, K. M. (2019, February). Quantum
language processing. arXiv:1902.05162 [quant-ph]. http://arxiv.org/abs/1902.05162
Wittgenstein, L. (1953). Philosophical investigations. Philosophische Untersuchungen. Macmillan.
The Philosophy of Quantum Computing
Michael E. Cuffaro
My work on this chapter benefited significantly from my interactions with students and other
audience members during and after a series of lectures I gave at the University of Urbino’s twenty-
third international summer school in philosophy of physics, held online in June 2020, in the midst
of the first wave of the COVID-19 pandemic, as well as a further lecture I gave for Michel Janssen’s
“The Age of Entanglement” honors seminar at the University of Minnesota in December 2020, as
the second wave of the pandemic began in earnest. Thanks to Ari Duwell, Eduardo Reck Miranda,
Philip Papayannopoulos, and Lev Vaidman for comments on a previous draft of this chapter. I
am also grateful for informal discussions, over the years, with Guido Bacciagaluppi, Jim Baggot,
Michel Janssen, Christoph Lehner, Lev Vaidman, and David Wallace; my presentation of the Everett
interpretation in Sect. 3, in particular, is significantly informed by what I take myself to have learned
from these discussions, though I hold only myself responsible for any mistakes or misunderstandings
in my presentation of the Everettian view. Section 2, on “Fundamental concepts” is heavily informed
by my recent work on related topics with Stephan Hartmann, Michael Janas, Michel Janssen,
and Markus Müller, as well as by informal correspondence with Jeffrey Bub and (the late) Bill
Demopoulos; though here again, I take sole responsibility for any mistakes. Finally, I gratefully
acknowledge the generous financial support of the Alexander von Humboldt Stiftung.
M. E. Cuffaro (B)
Munich Center for Mathematical Philosophy, Ludwig-Maximilians-Universität München,
Munich, Germany
e-mail: [email protected]
1 Introduction
From the philosopher’s perspective, the interest in quantum computation stems pri-
marily from the way that it combines fundamental concepts from two distinct sci-
ences: physics (especially quantum mechanics) and computer science, each long a
subject of philosophical speculation and analysis in its own right. Quantum comput-
ing combines both of these more traditional areas of inquiry into one wholly new
(if not quite independent) science. There are philosophical questions that arise from
this merger, and philosophical lessons to be learned. Over the course of this chapter,
we will be discussing what I take to be some of the most important.1
We begin, in Sect. 2, by introducing the fundamental concepts from physics and
computation that will be essential for framing the further philosophical discussion
that will follow. Section 2.1 and Sect. 2.2 introduce concepts from classical mechanics
and the classical theory of computation, respectively. In Sect. 2.1, we discuss the
concept of the state of a physical system as it is given in classical mechanics. We
emphasize in particular the way that we are invited, in classical mechanics, to think of
the state of a physical system as a compact description of the properties possessed by
it. These properties determine, in advance, the answers to the experimental questions
that we can pose of a system. And considering these questions and their answers leads
one to an alternative representation of a system’s state that is useful for representing
computation, a subject we then take up in more detail in Sect. 2.2.
In Sect. 2.2, we introduce the concept of a model of computation (or computational
architecture), the concept of the cost of carrying out a computation under a given
computational model, and explain what it means to solve a problem just as easily
under one computational model as under another. After discussing some of the more
important computational complexity classes, we then introduce two theses that can
be used to ground the model-independence of computational cost: the so-called
universality of Turing efficiency thesis (sometimes also referred to as the “strong”,
or “physical”, or “extended” Church-Turing thesis), and the invariance thesis. These
theses are both called into question by the existence of quantum computers. And
since some have taken the absolute model-independence guaranteed by them to
be foundational for the science of computational complexity theory, the question
arises of what to make of those foundations in light of the existence of quantum
computation. We will discuss this question, in an abstract way, in Sect. 2.3, where
I will argue that although the universality thesis must indeed be given up, this is
not a great loss. The invariance thesis, by contrast, remains, but in a different form,
transformed from a metaphysical to a methodological principle.
1Space does not permit me to exhaustively survey all of the philosophical issues brought up by
quantum computing. The interested reader can find a summary of other important issues in Hagar
and Cuffaro (2019).
The Philosophy of Quantum Computing 109
In Sect. 2.4, we turn to the physics behind quantum computation and begin by
introducing some of the more important concepts of quantum theory. Most impor-
tantly, we introduce the concept of a quantum state. We emphasize that the way that
a quantum state determines what the answers to experimental questions will be is
fundamentally different than the way that they are determined by a classical state.
We then turn to quantum computation proper in Sect. 3, where we review the basic
concepts of quantum computing and consider what we can say regarding the physical
explanation of the power of quantum computers. The many-worlds explanation of
quantum computing—the idea that quantum computers outperform classical com-
puters by running their computations in exponentially many physical universes—is
then introduced. We note two major problems that arise for this interpretation. The
first arises from the so-called preferred basis problem. This problem is a challenge
(that is arguably surmountable) for the many-worlds view in the more general context
of quantum mechanics. But we will see that it is especially problematic in the context
of quantum computers. The second major problem arises from the fact that there are
many different models of quantum computation, but the many-worlds explanation
of quantum computing only seems motivated by one of them.
In Sect. 4, we consider the role of quantum entanglement in enabling quantum
computers to outperform classical computers. We summarize an unsuccessful argu-
ment to the conclusion that quantum entanglement is insufficient to enable this quan-
tum “speedup” in Sect. 5, noting in Sect. 6 that reflecting on what it means to provide
a classical computer simulation of a quantum phenomenon should convince us to
reach the opposite conclusion. We continue the discussion of classically simulating
quantum computers in Sect. 7, and reflect on general aspects of the computational
perspective that the study of quantum computing provides on physical theory. We
note that reflecting on quantum computing emphasizes that there are important dif-
ferences in the methodological presuppositions that lie at the basis of physics and
computer science, respectively, and that conflating these can lead to confusion. We
also note the emphasis that studying quantum computation places on the fact that
quantum mechanics and classical mechanics are each alternative universal languages
for describing physical systems, and that the difference between quantum mechan-
ics and classical mechanics lies, fundamentally, in the differences in the expressive
power of these languages.2 We reflect on this in Sect. 8.
2By “quantum mechanics” I mean the fundamental theoretical framework shared in common by
every specific quantum-mechanical theory (quantum field theories, for instance) of a particular class
of systems; see Aaronson (2013b, pp. 110–111), Janas et al. (2022, Chap. 1 and §6.3), Nielsen and
Chuang (2000, p. 2), and Wallace (2019).
110 M. E. Cuffaro
2 Fundamental Concepts
Classical mechanics (in its various forms3 ) was, prior to quantum mechanics, our
fundamental theoretical framework for describing the dynamics of physical systems
(i.e., how physical systems change over time). What is meant by a physical system
here is just one of the concrete objects that a particular physical theory describes. All
such systems are described as evolving through time in accordance with the dynam-
ical constraints that a theoretical framework applies universally to every physical
system. In classical mechanics (and the same is true, as we will see later, in quantum
mechanics), a physical system can be anything from a single particle to (in princi-
ple) the entire physical universe. Mathematically, though, what the dynamical laws
of a theory actually govern are the relations between the possible state descriptions
of systems. By the state description (or state specification) of a system is meant a
description of the particular physical situation that it happens to be in at a given
moment in time. For instance, at a given moment in time, a classical-mechanical sys-
tem will have a particular kinetic energy, it will be accelerating (or not) in a particular
direction, and so on.
It turns out that we do not have to explicitly specify each and every one of a sys-
tem’s dynamical properties to exactly and exhaustively specify the system’s dynam-
ics. In classical mechanics, given the dynamical laws, it is enough to specify values
for a system’s position and momentum. The value of any other dynamical prop-
erty can then be calculated by relating these with each other and with the values of
non-dynamical properties of the system such as, for instance, its charge or its mass.
Figure 1 illustrates the two different strategies described in the previous paragraph
for specifying the state of a physical system, for instance, a system composed of a
particle attached to the end of a spring constrained to move in one direction (see
Hughes, 1989, p. 73). On the left, we explicitly specify values (v1 , v2 , . . . ) for each
and every one of the physical system’s dynamical parameters. On the right, the values
of momentum, p, and position, q, are specified, and all other dynamical quantities are
calculated on their basis. In particular, the total energy, H , of the system is defined as
the sum of the kinetic energy, T , of the particle, and the potential energy, V , stored in
the spring, which are in turn defined in terms of p and q, respectively. Note that m, the
system’s mass, is a non-dynamical parameter that does not change over the history
of the system, and k is the force (the first derivative of momentum) per unit distance
required for the spring to be displaced from its equilibrium point (i.e., the point at
which it is neither stretched nor compressed). Other forces (e.g., gravitational forces)
are neglected but in principle can be included (for further discussion, see Hughes,
1989, §3.2.5).
3 These include Newtonian, Lagrangian, Hamiltonian, relativistic, and classical statistical mechan-
ics. For a recent comparison and philosophical discussion of Lagrangian and Hamiltonian mechan-
ics, see Curiel (2014).
The Philosophy of Quantum Computing 111
Fig. 1 Two different strategies for specifying the state of a physical system, for instance, a system
composed of a particle attached to the end of a spring constrained to move in one direction (see
Hughes, 1989, p. 73). On the left, we explicitly specify values (v1 , v2 , . . . ) for each and every one of
the physical system’s dynamical properties. On the right, the values of momentum, p, and position,
q, are specified, and all other dynamical quantities are calculated on their basis
The upshot is that specifying the position and momentum of a system provides
us with enough information, in classical mechanics, to completely characterize all
of the other dynamical properties of that system. Accordingly, the dynamical state,
ωt , of a system at some particular time t is given in classical mechanics by
ωt = (qt , pt ), (2.1)
where qt and pt are vectors (in three dimensions) representing the system’s position
and momentum, respectively, at t. Further, we can infer from the state of a system at
t, and the classical-mechanical laws of motion, exactly what the state of the system
will be in the next moment and at every other time both in the system’s past and in
its future (Hughes, 1989, §2.6).
Classical-mechanical states have another feature. Imagine all of the possible
(experimental) yes-or-no questions one might want to ask about the dynamical
properties of a particular system at a particular time, questions like: Is the value
of the dynamical property A within the range ? Completely specifying a sys-
tem’s dynamical state, i.e., specifying its momentum, p, and its position, q, yields a
simultaneous answer to all such questions irrespective of whether any question has
actually been asked. Indeed, this is what actually justifies our having considered the
values of p, q, and the other quantities derived from them to be (observable) prop-
erties of the system in the first place, properties possessed by the system whether or
not, and however, we enquire concerning them (Janas et al., 2022, §6.3). The same
goes for a system composed of many parts; for, after all, any observable property
A1 of a subsystem S1 of some larger system S is also an observable property of S.
Thus, given the state specification for a particular system we can construct a sort of
“truth table” which could be used, in principle, to answer any conceivable yes-or-no
question concerning the system and its subsystems (see Fig. 2).
In Fig. 2, I wrote Y and N (for “yes” and “no”), but I could have equally well used T
(for true) and F (for false), or alternately the binary digits 0 and 1. Using binary digits,
i.e., “bits”, is convenient because they are useful for representing numbers, which can
112 M. E. Cuffaro
Fig. 2 A “truth table” for a classical system composed of two subsystems whose states, respectively,
are (p1 , q1 ) and (p2 , q2 ). Various combinations of values for the state parameters pi and qi are given
on the left of the double-vertical line, with the superscript j in v j indicating the jth combination.
Relative to a given combination of values, the answers to experimental questions concerning the
values of derived observable quantities A, B, . . . on the right of the double-vertical line can be
determined
Fig. 3 A physical system consisting of four light switches. For each switch, we can ask: “Is switch
S turned on?” Using 0 to represent yes and 1 to represent no, the state of the overall system relative
to this question can be represented using the bit-string 0110 (which, in base-ten, is the number 6).
This is just one of the 24 = 16 possible states that a system like this one can be in. More generally,
for a system made up of n two-dimensional subsystems, the number of possible states for the system
as a whole is 2n
Fig. 4 Some of the logico-mathematical operations, or “logic gates”, that can be used to manipulate
bits. OR and NOT together constitute a universal set of gates, in the sense that any more complicated
logico-mathematical operation, such as the “if-then” gate at the right, can be constructed using
combinations of OR and NOT. There are other universal sets besides {OR, NOT}, for instance:
{AND, NOT}, {NOR}, and {NAND}, where NAND is the “not-and” gate and NOR is the “not-or”
gate
When the word “computer” is used in popular discourse today, generally what is
meant is a physical system like the one depicted in Fig. 3—not, of course, literally
a collection of light switches, but a physical system composed of subsystems that,
like the switches, can reliably be interpreted as being either on or off, and that can be
organized into registers, random access memory, hard drives, and other peripheral
systems such as keyboards, “mice”, touchpads, and so on, that can be manipulated
to reliably realize various logical operations like the ones depicted in Fig. 4. The
laptop I am writing this chapter with is an example of such a device. We turn such
devices on when we need them, and turn them off or “put them to sleep” when we
do not. But the word “computer” did not always signify a kind of machine. Prior
to the 1930s, “computation” generally meant an activity carried out by a human
being. Even after Alan Turing’s introduction of what later came to be known as
the “Turing machine” in 1936—an abstract device that is itself modeled on human
computation—“computers” were generally understood to be human beings until well
into the 1960s.6
As for the Turing machine: this, as mentioned, is also not a concrete physical
device (although it could, in principle, be instantiated by one) but an abstract math-
ematical model. Turing’s primary reason for introducing it had been to address the
so-called Entscheidungsproblem (which is the German word for ‘decision problem’),
an important but unsettled question that had arisen in the foundations of number
theory. The Entscheidungsproblem concerned the existence of an effective proce-
6The role of human computers in the United States of America’s space program, for instance, has
been documented in Shetterly (2016).
114 M. E. Cuffaro
dure for deciding whether an arbitrarily given expression in first-order logic can be
proven from the axioms of first-order logic. The answer, as Turing and, indepen-
dently, Alonzo Church were able to demonstrate, is that there is no such procedure.
Turing’s own proof relied on a penetrating philosophical analysis of the notion of
effective (human) computation, and a corresponding argument that one could design
an automatic machine (now called a Turing machine; see Fig. 5) to carry out the
essential activities of a human computer.
Turing (1936–7, pp. 249–51) argued that, for a computer to carry out a computa-
tion, it is essential that she have access to a notebook from which she can read and
onto which she can write various symbols in the course of her work. These symbols
need to be distinguishable by her from one another, on the basis of which Turing
argued that the alphabet from which she chooses her symbols must be a finite alpha-
bet. At any given moment during a computation, a computer may find herself in one
of a finite number (again, because she must be able to distinguish them) of states of
mind relevant to the task at hand which summarize her memory of the actions she
has performed up until that point along with her awareness of what she must now
do (pp. 253–4). The actions that are available to her are characterized by a finite set
of elementary operations, such as “read the next symbol” from the notebook, “write
symbol a” to the notebook, and so on. In an automatic machine, a control unit is
constructed to encode the machine’s “state of mind” (i.e., a logical representation
of its current state, its current input, and its transition function), which in general
changes after every operation of the read-write head. The latter moves back and
forth along a one-dimensional tape (the machine’s “notebook”), from which it reads
and onto which it writes various symbols from a finite alphabet, in conformity with
a particular finite set of exactly specifiable rules. Turing was able to show that no
automatic machine of this kind can be used to solve the Entscheidungsproblem.
Through the work of Turing, Church, and others, the modern science of computa-
tion, and in particular computability theory—the science of which problems can and
which cannot be solved by a computer, i.e., by anything or any person that can carry
out (but is restricted to) the essential activities associated with human computation—
was born.7
Irrespective of whether it is a machine or a human being that is doing the comput-
ing, the question of how much it actually costs to compute a solution to a computable
problem is also a very important one. In particular, we generally would like to know
how much time and how much space are required (though time is usually regarded
as the more important measure). The question of the cost of a given computation had
become especially important as more and more problems came to be carried out by
machines in the 1950s and 1960s. Even the earliest digital computers of the 1930s
and 1940s performed their computations far faster than human beings could possibly
perform them, and with improvements in design and in technology, machines could
be made to run ever faster. Through the need to use these machines’ resources effi-
7 For more on the Entscheidungsproblem and the early history of computer science, see Copeland
(2020), Davis (2000), Dawson Jr. (2007), Lupacchini (2018).
The Philosophy of Quantum Computing 115
Fig. 5 A version of what is now called a ‘Turing machine’, with a single bi-directional tape, a
control unit encoding a particular set of instructions or program, and a read-write head. Such a
machine is an idealized representation of the components involved in human computation. In a
universal Turing machine, the computer’s program is read from the tape at the beginning of the
computation. For a good introduction to the Turing machine and other related topics, see Martin
(1997)
ciently, a sub-field of computer science began to develop whose concern was with
how efficiently a given problem can be solved and on which types of machine.
In the context of the Turing machine model, time and space are quantified in
terms of the number of computational ‘steps’ (i.e., the number of basic operations)
and the number of squares of tape, respectively, that are needed to carry out a given
computation. Although it is in principle possible (assuming we have enough tape)
to instantiate a Turing machine physically,8 generally it makes more sense, from
a practical perspective, to design a mechanical computer in accordance with some
other model of computation. Particularly important is the von Neumann architecture,
a type of stored-program computer,9 that involves a central processing unit (CPU),
an instruction register and program counter, internal memory used to store data
(including the program being run), external memory used to store less frequently
needed data, and finally an input and an output mechanism. A von Neumann machine
can in principle compute all of the same problems as a universal Turing machine (see
Fig. 5), but a time step for a von Neumann machine is not concretely interpreted in the
8 A Turing machine’s tape does not need to be actually infinite in length. What is required is only that
the tape be indefinitely long, so that, for a given (finite) computation, the machine can be supplied
with enough tape to carry it out. To put it another way: What constitutes ‘enough tape’ to carry out
a computation is not part of the general definition of a Turing machine. It is, rather, assumed, for
the purposes of that definition, that enough tape to carry out a given computation can be supplied.
That said, certain variations on the Turing machine model restrict the ways in which tape can be
read by the control unit in various ways. For instance, some variants employ separate tape(s) for
the machine to write “rough work” on in addition to an output tape, some variants only allow the
read-write head to move in one direction along the tape, and so on.
9 Although the von Neumann architecture, or ‘von Neumann machine’ is only one of a number of
various types of stored-program computer, the terms have today (inappropriately, from a historical
point of view) come to be understood synonymously (Copeland, 2017).
116 M. E. Cuffaro
same way as it is for a Turing machine, just as ‘space’ for a von Neumann machine
is not measured in squares of tape.
There are many other universal models of computation besides von Neumann
machines and Turing machines. Nevertheless, at least when it comes to reasonable
classical computational models, it is possible to talk about the computational cost
of solving a particular problem (like, for instance, the problem of finding the prime
factors of a given integer), in a model-independent way; i.e., without having to
specify which machine model that cost is defined with respect to. There are two
things involved in explicating what it means for the computational cost of solving a
given problem to be model-independent in this sense. The first, mentioned just above
is the idea of a reasonable machine architecture.
Roughly, a reasonable (universal) machine architecture or model is one that it
would be possible to physically instantiate. Turing machines like the one described
in Fig. 5 and von Neumann machines are examples of reasonable machine models. A
computational model that employs an unbounded amount of parallel processing, by
contrast, is an example of an unreasonable model, or what van Emde Boas (1990, p. 5)
calls a model of the “second machine class”. There are legitimate theoretical reasons
for studying the complexity-theoretic properties of unreasonable models. But no
finite physical system could possibly instantiate an unreasonable architecture. There
are thus good reasons to want to distinguish the reasonable from the unreasonable
ones (see also, Dean, 2016).
The second thing we need to do, in order to explicate what it means for the compu-
tational cost of solving a given problem to be model-independent, is to make sense
of what it means to solve a given problem just as easily with one computational
architecture as with some other one. But before we can make sense of solving some-
thing just as easily, we need to first try and make sense of what it means to solve
something easily under one computational model. Consider, to begin with, a human
computer working alone. For a human computer, the difference between a problem
that could in principle require up to a thousand steps to solve, and one that could in
principle require up to a million steps, is very great indeed. With some justification,
this person might call the former problem easy to solve, and the latter problem hard.
Any criterion based on a fixed number of computational steps, however, would not
be very useful in the context of machines, at any rate certainly not any criterion based
on numbers this low. Even by the 1950s, a mechanical computer could carry out a
computation far faster than a human being. As of 2020, a typical laptop computer
can perform hundreds of thousands of Millions of Instructions per Second (MIPS)
(Wikipedia contributors, 2020), and computers are becoming more and more per-
formant every year. In the context of machine computation, it is far more useful to
ask how a given solution to a computational problem scales; i.e., how the number
of resources required to carry out a given solution to that problem grows with the
size of the input to the problem. Take, for instance, the problem to sort a given set of
numbers. All else equal, the more numbers there are to sort, the longer it will take.
How much longer? The fastest sorting algorithms (one of which is MergeSort)
will take on the order of n log n steps to sort n numbers in the worst case (Mehlhorn
& Sanders, 2008).
The Philosophy of Quantum Computing 117
10 “On the order of” is a technical term, usually symbolized in “big-oh notation” as O(T (n)). An
algorithm is O(T (n)) for some function T (n) if for every sufficiently large n, its actual running
time t (n) ≤ c · T (n) for some constant c. For instance, an algorithm that never takes more than 5n 3
steps is O(n 3 ).
11 A famous example of a problem for which only exponential-time solutions are known is the
principle, to some extent, in a private letter he wrote to John von Neumann in 1956. For further
discussion, see Cuffaro (2018b).
13 This correspondence is not perfect, but the usefulness of the polynomial principle is such that we
steps, where n is the number of bits used to represent the input, and k and l are finite constants. The
total number of steps needed to carry out this program will be n k+l . If k and l are finite constants
then so is k + l. In other words, n k+l is still a polynomial.
118 M. E. Cuffaro
Fig. 6 Inserting any number of calls to polynomial-time subroutines within the body of a
polynomial-time algorithm results in a polynomial-time algorithm
number of extra time steps as compared to M2 .15 Another way of putting this is that
anything that M2 can do is efficiently simulable by M1 .
Any problem that is solvable in polynomial time by a Turing machine is said to
belong to the class PTIME (for “polynomial time”), though it is usually referred
to simply as P. P is, of course, not the only complexity class. If a given problem is
such that, for a given solution to it, there is a Turing machine that will verify that the
solution is indeed a solution, then the problem is said to belong to the class NP, for
“nondeterministic polynomial time”. The reason for the name is that, equivalently,
this class of problems can be defined as those that can be solved in polynomial
time on a nondeterministic Turing machine. A nondeterministic Turing machine, or
choice machine, is such that at a given step of a given computation, an imagined
external operator of the machine can choose to have it transition in one way rather
than another. This is unlike a standard (deterministic) Turing machine for which
every computational step is completely determined in advance given a particular
input. Choice machines are interesting from a theoretical point of view, but given
that they involve an external operator they are not really automatic machines in the
sense that is relevant to our discussion here.
A variant of the choice machine that is automatic is a probabilistic Turing machine.
Such a machine’s choices are not made by an external operator but by the machine
itself, which we can imagine as equipped with the equivalent of a fair coin that it can
flip a number of times to decide whether to transition one way or another whenever
such a choice is open to it.16 The class of problems solvable in polynomial time
on a probabilistic Turing machine is called BPP (for “bounded error probabilistic
polynomial”). Note that what it means to be able to solve a given problem is not the
same for a probabilistic Turing machine as it is for a deterministic Turing machine.
In particular, a given “solution” output by a probabilistic Turing machine is allowed
to be wrong. We only demand that it be right with high enough probability so that,
15 Note that it makes sense to talk about solving a given problem just as easily on M1 as on M2
even when the problem under consideration is actually intractable for both. For instance, if some
problem requires 2n steps to solve on M1 and 2n + n 3 steps to solve on M2 then it is no harder,
from the point of view of the polynomial principle, to solve it on M2 than on M1 .
16 For more on probabilistic and nondeterministic Turing machines and how they compare to their
if we re-run the machine for the given input on the order of a polynomial number of
further times, our confidence in the majority answer will approach certainty.
There are very many more complexity classes besides these ones (see Aaronson,
2016). But we will stop here as we now have the basic ingredients with which to
state a couple of related theses that can be used to explicate what it means for the
cost of solving a given problem to be model-independent. As we will see a little
later, there are good reasons to prefer the second thesis, but for now, we simply
introduce both. First, the universality of Turing efficiency thesis (sometimes also
called the “strong”, “physical”, or “extended” Church-Turing thesis17 ) asserts that
any problem that can be efficiently solved on any reasonable machine model M (von
Neumann architecture, Harvard architecture, or whatever) can be efficiently solved
on a probabilistic Turing machine, or more formally
PolyM = BPP. (2.2)
In other words, the thesis implies that the set of problems solvable in polynomial
time does not grow beyond BPP if we allow ourselves to vary the underlying model.
Assuming the thesis is true, we do not need to worry about what model an algorithm is
implemented on when discussing the computational complexity of various problems;
we can simply focus on the abstract probabilistic Turing machine model and carry
out our analyses in relation to it.18
The second thesis we will introduce is called the invariance thesis (van Emde
Boas, 1990, p. 5),19 which asserts that given any two reasonable machine models
Mi and M j , Mi can efficiently simulate M j ; i.e., Mi can solve any problem just as
easily (in the sense explained above) as M j can. More formally
poly
∀i, j Mi ∼ M j . (2.3)
Note that the invariance thesis implies the universality thesis, but not vice versa.
17 See Timpson (2013, Chap. 6) for discussion of a different, more general thesis, that is only
indirectly relevant to computational complexity theory. For a discussion of how these theses relate,
see Cuffaro (2018b, §11.4).
18 We could have also expressed the thesis in terms of P rather than BPP. Although it was thought,
for many years, that there are more problems efficiently solvable on a probabilistic Turing machine
than on a standard Turing machine, a number of recent results have pointed in the opposite direction
(e.g., Agrawal et al., 2004), and it is now generally believed that classical probabilistic computation
does not offer any performance advantage over classical deterministic computation (Arora Barak,
2009, Chap. 20). In other words, it is now widely believed that P = BPP, or that it is just as easy
to solve a given problem on a deterministic Turing machine as it is on a probabilistic one. We have
nevertheless stated the universality thesis in terms of BPP because this will prove convenient when
it comes time to discuss the differences between classical and quantum computation. A quantum
computer is, from one point of view, just another kind of probabilistic computer (that calculates
probabilities differently), and it has the same success criterion as a classical probabilistic computer,
i.e., we only demand that a given “solution” be correct with “high enough” probability.
19 See also: Goldreich (2008, p. 33), who names it differently.
120 M. E. Cuffaro
Neither the universality of Turing efficiency thesis, nor the invariance thesis, is a
mathematical theorem. These statements can be true or false, and for a long time
they were thought to be true, for none of the reasonable (classical) universal models
of computation that had been developed since the time of Turing were found to
be more efficient than the Turing model by more than a polynomial factor, and all
of them had been shown to be able to simulate one another efficiently (van Emde
Boas, 1990). Over the last three decades, however, evidence has been mounting
against universality and invariance, primarily as a result of the advent of quantum
computing (Aaronson, 2013b, Chaps. 10 and 15).
Quantum mechanics, as we will see in the next section, is an irreducibly probabilis-
tic theory. Thus, a quantum computer is a type of probabilistic machine. Analogously
to the way we defined BPP as the class of problems solvable (in the probabilistic
sense) in polynomial time on a probabilistic Turing machine, we can also define the
complexity class BQP (for “bounded error quantum polynomial”) as the class of
problems solvable (in the same sense) in polynomial time on a quantum computer.
It is easy for a quantum computer to simulate a probabilistic Turing machine. Thus
20 For instance, the quantum Turing model (Deutsch, 1985), the quantum circuit model (Deutsch,
1989), the cluster-state model (Briegel et al., 2009), the adiabatic model (Farhi et al., 2000), and so
on.
21 See, for instance, Aharonov et al. (2007), Nishimura and Ozawa (2009), Raussendorf and Briegel
(2002).
122 M. E. Cuffaro
We saw earlier that in classical mechanics, assigning a particular state, ω = (q, p), to
a system fixes the answer to every conceivable yes-or-no experimental question that
we can ask about it in advance, irrespective of whether we actually ask any questions
or not. And we saw how to arrange the possible questions we can ask about a system,
and their possible answers, into a truth-table-like structure like the one given in Fig.
2. Note that specifying the values of the answers to the questions on the right-hand
side of the double-vertical line in Fig. 2 is another way of representing the state of
a system like the one depicted in Fig. 3. We saw how one can manipulate such a
system in ways that can be represented abstractly as logico-mathematical operations
on binary digits, where the binary digits are abstract representations of the properties
of the system’s (two-level) subsystems.
The subject matter of quantum mechanics, just as it is for classical mechanics,
is physical systems, where these can be anything from a single particle to the entire
physical universe. In other words, quantum mechanics is not just a theory of the
small. It is, rather, a new universal language for describing physical systems,22 and it
describes our experience with those systems better and more accurately than classical
mechanics describes it. In quantum mechanics, just as in classical mechanics, a
system can be, at a given moment in time, in any one of a number of possible
physical states. These states are represented by state vectors (or “wave functions”).
The state vector for a two-dimensional quantum system or “qubit”, for instance, is
given in general by
When the complex numbers α and β are both non-zero, a qubit is said to be in
a superposition of the basis vectors |0 and |1. Just as an ordinary vector in the
Fig. 7 An ordinary vector in the Cartesian plane. It can be decomposed into x and y coordinates,
as on the left, or into the alternative coordinates x and y as on the right. Similarly, a state vector,
in quantum mechanics, can be decomposed into more than one Hilbert space basis. Two common
bases (but not the only two) with which to decompose the state vector for a qubit are the {|0, |1}
basis (known as the computational basis) and the {|+, |−} basis
Pr = |α|2
M α|0 + β|1 −−−−→ |0 (2.7)
Pr = |β|2
M α|0 + β|1 −−−−→ |1 (2.8)
and unitarily:
when these subsystems are correlated with one another (as the state of one side of a
coin is correlated with the state of the other side, for instance), as long as we include
facts about those correlations in our description of the overall system, then its overall
state can be factored into the individual states of its subsystems (Bell, 2004 [1981]).
This is true even if the subsystems happen to be far apart, for instance, if I print out
two copies of a document and mail each of them to distant parts of the globe.
In quantum mechanics, in contrast, a system composed of multiple subsystems
can sometimes be in a state like the following:
1
√ |0 A |1 B − |1 A |0 B . (2.11)
2
This state, describing a system composed of two subsystems A and B, is not a product
of the individual states of its subsystems, for there is no way to factorize it into a state
of the form: |ψ A |φ B (the reader is invited to try). It is sometimes said that in an
entangled quantum system the whole is somehow prior to the parts (Howard, 1989),
or that the systems are nonlocally correlated (Maudlin, 2011), for if Alice measures
her qubit in the computational basis and receives the result |0 A , then a measurement
by Bob on his qubit (which in principle he may have taken to the opposite side of
the universe) is thereby instantaneously determined to be |1 B , although there is no
way for Bob to take advantage of this instantaneous determination for the purpose
of predicting the outcome of a measurement on his subsystem, short of waiting for
Alice to send him a classical signal (a text message or a phone call, for instance).25
Note that the correlation between the results of computational (i.e., {|0, |1}) basis
measurements on an entangled system in this state do not, in themselves, represent
as much of a departure as one might think from what one can describe using classical
language. For with respect to this specific correlation, Bell (2004 [1964]) showed that
it can be seen as arising from a classically describable system that has been prepared
in a particular way. It is only when we consider other experiments on such a system
that we see a departure from the possibilities inherent in classical description. In
particular, because
1
√ |+ A |− B − |− A |+ B . (2.13)
2
Indeed, we obtain the same form for the state of the system regardless of which basis
we choose to express it in:
25This restriction on the use that can be made of nonlocal correlations in quantum mechanics is
called “no signaling”. For discussion, see Bub (2016), Cuffaro (2020).
126 M. E. Cuffaro
1
√ |b1 A |b2 B − |b2 A |b1 B . (2.14)
2
Equations (2.11), (2.13), and (2.14) all yield the same probabilities for the results of
experiments.
If we collect a number of pairs of subsystems prepared in this entangled state, and
then ask of each pair: “Is this pair in the state |0|1 (as opposed to |1|0)?,” then
we will find that the answers “yes” and “no” will be obtained with equal probability.
Conditional upon this question, we can imagine these answers as arising from a prior
classical probability distribution over the properties of the subsystem pairs, half of
which we imagine to be in the state |0|1, and half of which we imagine to be in
the state |1|0. This prior probability distribution is incompatible, however, with
the prior classical probability distribution that we might imagine to be responsible
for the answers yielded from repeatedly asking the question: “Is this pair in the
state |+|− (as opposed to |−|+)?” In the context of the latter question, we can
imagine the answers as arising from a prior classical probability distribution over
the properties of the subsystem pairs, half of which we imagine to be in the state
|+|−, and half of which we imagine to be in the state |−|+. Asking a question
of an imagined ensemble of systems, half of which are in the state |0|1, and half of
which are in the state |1|0, will yield different statistics, however, than the ones that
would be yielded by asking the same question of an imagined ensemble of systems,
half of which are in the state |+|−, and half of which are in the state |−|+.
These probability distributions are, in this sense, incompatible. And yet all of these
statistics are predicted by one and the same quantum state, the one expressed in the
unitarily equivalent forms given in Eqs. (2.11–2.14). In quantum mechanics, unlike
in classical mechanics, the questions we ask or do not ask actually matter for the
ways that we can conceive of the underlying properties of a physical system, even
when the state of that system is fully specified (Janas et al., 2022, §6.5).
The are problems that are hard for a classical computer, but that a quantum computer
can solve easily (where “easy” and “hard” are meant in the computational sense
defined above in Sect. 2.2).26 The most famous example of such a problem is that of
factoring large integers into primes. The best-known classical algorithm for factoring
is the number field sieve (Lenstra et al., 1990), which takes on the order of 2(log N )
1/3
steps to factor a given integer N .27 No one has yet proven that this is the best that a
26 There are also problems for which a quantum computer, despite being unable to solve them easily,
can nevertheless solve them significantly more easily than a classical computer can. An example is
the problem to search an unstructured database, for which a quantum (“Grover’s”) algorithm can
reduce the number of steps required by a quadratic factor over any known classical algorithm. See:
Bennett et al. (1997), Grover (1996), and for further discussion see Cuffaro (2018b, p. 269).
27 For the meaning of ‘on the order of’ see fn. 10.
The Philosophy of Quantum Computing 127
classical computer can do. All the same, encryption algorithms such as RSA (named
for the paper by Rivest, Shamir, and Adleman in which it was introduced), that are
today widely used for secure transactions on the internet, rely on the assumption that
it is; i.e., that one cannot factor in polynomial time. In 1994, Peter Shor discovered
a quantum algorithm—now known as Shor’s algorithm—that can factor integers in
on the order of log N steps for a given integer N —an exponential “speedup” over
the number field sieve.
What explains this phenomenon? What is it, exactly, that allows quantum com-
puters to compute certain problems more efficiently than classical computers can?
Surprising as it may seem, there is still no consensus regarding the answer, neither
among the researchers working in the field, nor among the philosophers comment-
ing on this work. By contrast, in the popular literature on the topic, one answer
has tended to dominate all the others. This is the idea that quantum computers are
actually parallel devices that perform exponentially many classical computations
simultaneously, with each computation taking place in a parallel physical universe
(or “world”), different in certain ways from our own but just as real as it is.
Perhaps the strongest proponent of this view, which we will call the many-worlds
explanation of the power of quantum computers, is one of the fathers of the field,
David Deutsch, discoverer of the first quantum algorithm and of the universal quan-
tum Turing machine (Deutsch, 1985). The many-worlds explanation of quantum
computing is, for Deutsch, not just a speculative conjecture. For Deutsch, many-
worlds are the only plausible explanation for how quantum computers work. As he
puts it in his 1997 book, The Fabric of Reality: “[t]o those who still cling to a single-
universe world-view, I issue this challenge: Explain how Shor’s algorithm works”
(1997, p. 217).
Before we discuss what those who, like Deutsch, defend this view of quantum
computation are really saying, let us review briefly what the view amounts to in
more general terms in the context of quantum mechanics, for what the many-worlds
explanation of quantum computing amounts to is an application of this more gen-
eral philosophical interpretation of quantum mechanics to the case of quantum-
mechanical computers.28 Thus far, our examples of quantum-mechanical systems
have been of qubits, either considered singly or in combination with other qubits.
But quantum mechanics is not restricted, of course, to describing qubits. Quantum
mechanics is, rather, a universal language for describing any physical system, be it a
single particle or (in principle) the entire physical universe. From the point of view
of quantum mechanics, you and I, the laptop I am writing this chapter with, the table
28 The interpretation of quantum mechanics that we will be discussing in this section is one of
a number of related interpretations of quantum mechanics that are collectively referred to as the
“Everett interpretation”. These include, but are not limited to Hugh Everett III’s original formu-
lation (Everett III, 1956), the “Berlin Everettianism” of Lehner (1997), Lev Vaidman’s version of
Everett (Vaidman, 1998), so-called “many minds” variants (Albert & Loewer, 1988), and finally the
“many-worlds” variants that are the direct inspiration for the many-worlds explanation of quantum
computing. Belonging to the last group are DeWitt’s (1973 [1971]) view, as well as the “Oxford
Everett” interpretation (Deutsch, 1997; Saunders, 1995; Wallace, 2003, 2012) with which we will
be mostly concerned here.
128 M. E. Cuffaro
2
n
−1 2
n
−1
|xn |0 → |xn | f (x), (3.1)
x=0 x=0
Note that, in Eq. (3.2), the n input qubits are together in a superposition of all of
the possible computational basis states for a system of n qubits. This is obscured
somewhat, both by the summation notation as well as by the shorthand, |xn , being
used to represent each superposition term. To clarify this, recall that for a two-
dimensional system, i.e., a single qubit, the computational basis states are, as we
pointed out earlier (see Eq. (2.6) and Fig. 7), the states |0 and |1. As for a system
composed of two qubits, it has four computational basis states, namely, |0|0, |0|1,
|1|0, and |1|1. In general, for a system of n qubits there are 2n basis states.
|0|0 . . . |0|0,
|0|0 . . . |0|1,
|0|0 . . . |1|0,
...
|1|1 . . . |1|1, (3.3)
each of which can be thought of as representing a binary number. For short, we can
represent these in decimal as {|0, |1, |2, |3, . . . |2n−1 }. A superposition of all
of these basis states is then given by
2
n
−1
|0 + |1 + |2 + |3 + . . . + |2 n−1
= |xn , (3.4)
x=0
the computer literally, then the fact that only one of them can ever be accessed in
a particular run should give us some pause (cf. Mermin, 2007, p. 38). That said, in
any given run, there will be an, in general, non-zero probability of obtaining any one
of them, and this, perhaps, speaks in favor of viewing them all as somehow literally
there. As for the goal (which we should not lose sight of) of actually solving the
problem at hand, what is just as important as achieving the form of the state of the
computer on the right-hand side of Eq. (3.1) is the next step in a computation like
this one, which requires that we manipulate the system cleverly enough so that one
of the desired solutions is found with higher probability than the other, undesirable,
solutions (Pitowsky, 2002, §3).
The many-worlds explanation of a quantum computational process enjoins us to
take an evolution like the one given in Eq. (3.1) ontologically seriously. It affirms,
in other words, that things are indeed as they seem: The computer most definitely
is performing many simultaneous function evaluations in parallel when it is in a
state like the one on the right-hand side of Eq. (3.1). Moreover, the many-worlds
explanation directly answers the question of where this parallel computation occurs,
namely, in distinct physical universes. For this reason, it is also, arguably, one of the
more intuitive of the physical explanations of quantum speedup. And it is certainly
thought-provoking. It is no wonder, then, that it is the one most often mentioned in
the popular literature on quantum computation. And it has advocates in the more
serious literature as well.29
So far we have highlighted one potential problem for the many-worlds explana-
tion: Despite the appearance of parallel processing in an equation like Eq. (3.1), only
one of these function evaluations is ever accessible on a given run of the computer.
This is not to say that no story can be told from the point of view of the many-worlds
explanation about why this is so, but at the very least this is enough to cast doubt on
the claim that an evolution like the one given in Eq. (3.1) constitutes, all by itself,
evidence for the many-worlds explanation of the power of quantum computers. We
will return to this issue later. But for now, we need to consider a somewhat deeper
problem faced by the advocate of the many-worlds explanation. Recall that decom-
posing a qubit in the state |ψ into the computational basis states |0 and |1 (see
Eq. (2.6) and Fig. 7) is only one way to express this state. The same state can also be
expressed in other bases besides the computational basis. This means that a system
that is in the state
1 1
√ |0 + √ |1, (3.5)
2 2
from the point of view of the computational basis, is simply in the state:
|+ (3.6)
29In addition to Deutsch’s 1997 book, see Deutsch (2010), and see also Vaidman (2018 [2002],
§7) and Wallace (2012, Chap. 10). The strongest and most in-depth defence of the many-worlds
explanation of quantum computing that I am aware of is the one given by Hewitt-Horsman (2009).
The Philosophy of Quantum Computing 131
from the point of view of the {|+, |−} basis (see Eq. (2.12)). But now it is no longer
clear just how many universes we should take this qubit to exist in. If we decompose
its state in the computational basis, then it would seem that it exists in two worlds,
and if we decompose it into the {|+, |−} basis then it would seem that it only exists
in one. Which is it? If we decide that one of these decompositions is to be preferred
to the other, then the challenge for the advocate of the many-worlds explanation is to
give a compelling reason why, for prima facie there does not appear to be any reason
for preferring one of these decompositions over the other.
This is known as the preferred basis problem, and it is a problem for the many-
worlds view more generally, i.e., not just in the context of quantum computing. In
this more general context, advocates of the many-worlds picture (see, for instance,
Wallace, 2003) attempt to solve the preferred basis problem by appealing to the
dynamical process of decoherence. In quantum mechanics, as we saw (see Eq. (2.6)
and Fig. 7), the state of a system at a particular moment in time is described by
a state vector that evolves linearly and unitarily. However, the evolution thereby
described is the evolution of a system that is completely isolated from its external
environment, which is an idealization; in reality, it is actually never possible to
completely isolate a system from its environment,30 unless, perhaps, we take our
system of interest to be the universe in its entirety.31 But barring the universe as a
whole, when a system interacts with an external environment—for example, with
our measurement apparatus as we ask the system an experimental question—then
the terms in the superposition describing its state begin to decohere (Zurek, 2003
[1991]) and come to achieve a kind of independence from one another (although they
never decohere completely). On the many-worlds picture, we are to think of such
(approximately) decoherent terms as existing in independently evolving physical
universes. And in each of these independently evolving physical universes, there is,
in addition, a different version of ourselves, all of whom receive a different answer
to the experimental question that was asked. And with every additional question, the
universe as we know it is branching, spawning exponentially more and more versions
of the system, and more and more versions of ourselves along with it, and so on and
on and on.
Fundamentally, however, decoherence is an approximate phenomenon, and some
small amount of residual interference between worlds always remains despite it.
Nevertheless, decoherence tells us that, when the environment and system happen
to be correlated in a particular way, then a particular basis will emerge with respect
to which we can effectively describe the superposition terms expressed in the state
of the system as evolving independently of one another. As Wallace (2003, p. 90)
puts it: “the basic idea is that dynamical processes cause a preferred basis to emerge
rather than having to be specified a priori”. In this way, superposition terms that for
all practical purposes evolve stably and independently over time with respect to the
decoherence basis can be identified with different copies of measurement pointers,
30 At the very least, the gravitational effects of other distant systems will not be able to be neglected.
31 Some philosophers have questioned whether we should think of even the universe as a whole as
a closed system (see, for instance, Cuffaro & Hartmann, 2021).
132 M. E. Cuffaro
cats, experimenters, and whatever else is theoretically useful for us to include in our
ontology: “the existence of a pattern as a real thing depends on the usefulness—in
particular, the explanatory power and predictive reliability—of theories which admit
that pattern in their ontology” (Wallace, 2003, p. 93). Whatever else one may think
of the many-worlds view, decoherence, at least, does provide a principled way to
identify worlds in the wave-function.
For the advocate of the many-worlds explanation of quantum computation, how-
ever, there is still a problem. Appealing to decoherence may solve the preferred basis
problem for the purposes of describing the world of our everyday experience,32 but
the inner workings of a quantum computer are not part of that everyday experience.
The problem is not just that qubits are too small to see. The problem is that the super-
positions characteristic of quantum algorithms are coherent superpositions (Nielsen
Chuang, 2000, p. 278; see also Aaronson, 2013c; Cuffaro, 2012). Thus, the terms
in the wave-function of a quantum computer do not seem to meet the criterion for
world-identification advocated for by the many-worlds view.
Now, to be fair, a similar thing can be said with regard to our everyday experience.
So-called decoherent superpositions are (as I mentioned) not really decoherent, after
all. But they are decoherent enough, and for long enough, that it is useful (or so says
the many-worlds advocate) to think of the terms in such a superposition as inde-
pendent, and the worlds that they describe as ontologically real. Likewise, although
the superposition on the right-hand side of Eq. (3.1) is actually a coherent superpo-
sition, it may nevertheless be useful to think of the terms in that superposition as
independent, at least for the short time that the quantum computer is in that state
(Hewitt-Horsman, 2009, p. 876). The problem, however, is that it is the very fact
that they are coherent that allows us to “cleverly” extract desirable solutions from
these superpositions with higher probability than undesirable solutions (Bub, 2010;
Duwell, 2007).
Even if we grant that there is some heuristic value in describing a quantum com-
puter in a superposition state as evaluating functions in exponentially many parallel
worlds (I do not doubt that this was of some heuristic use to Deutsch, for instance,
even if I question whether it is necessary to think of such superpositions in this way),
it does not follow that this is enough to licence granting ontological status to those
worlds. Wallace (2003, p. 93) mentions (as we saw) explanatory power and predictive
reliability, for instance, and discusses the way that these and other ideas are applied
in contemporary physics to support the many-worlds view outside of the context of
quantum computing. It is not at all clear that these criteria are met in the context of
quantum computing, however, and even Wallace admits that they, for the most part,
are not: “There is no particular reason to assume that all or even most interesting
quantum algorithms operate by any sort of ‘quantum parallelism”’ (Wallace, 2010,
p. 70, n. 17). Wallace goes on: “Shor’s algorithm, at least, does seem to operate in
32 The preferred basis problem is not the only challenge that needs to be met by an advocate of the
Everett interpretation of quantum mechanics. Another issue that has been much discussed in recent
literature is the problem of how to account for probabilities on the Everettian view. For more on
this issue, see Adlam (2014), Dawid and Thébault (2015), Greaves and Myrvold (2010), Vaidman
(1998, 2012), and Wallace (2007).
The Philosophy of Quantum Computing 133
this way” (ibid.), but he does not describe how. Yet there are very plausible accounts
of how Shor’s algorithm works that do not appeal to massive parallelism at all (see
Bub, 2010). Far from it, on Jeffrey Bub’s account of Shor’s algorithm, the quan-
tum algorithm is more efficient than known classical algorithms because it performs
fewer, not more, computations (see also Bub, 2008).
The final reason that I will mention for being skeptical of the many-worlds expla-
nation of quantum computing is that it only really seems to be useful in the context
of one particular model of quantum computing. This is the so-called quantum circuit
model, the model for which many of the first quantum algorithms were designed
(Deutsch, 1989). This model is useful for abstract theoretical purposes, as well as
for pedagogical purposes, as it borrows many familiar ideas from the classical cir-
cuit model of computation (see Fig. 4). In the quantum circuit model, similarly to
the classical circuit model, logical circuits are constructed out of various “quantum
logic gates”. These instantiate unitary transformations of one or more qubits that are
prepared beforehand in computational basis states (typically qubits begin a compu-
tation in the state |0). The unitary gates transform the qubits’ states into various
superpositions of computational basis states, and at the end of the computation a
measurement is performed, again in the computational basis, on (some of) the qubits
and the results are read out.
Figure 8a depicts a number of important one- and two-qubit quantum gates. The
X gate implements a qubit-flip operation; i.e., it takes |0 → |1 and vice versa.
The Y gate takes |0 → i|1 and |1 → −i|0. The Z gate takes |0 → |0 and
|1 → −|1. The R gate takes |0 → |0 and |1 → i|1. The H (or Hadamard)
√ √
gate takes |0 → (|0 + |1)/ 2 and |1 → (|0 − |1)/ 2. The S gate takes |0 → |0 and
|1 → eiπ/4 |1. At the extreme right is the two-qubit CNOT (or controlled-not) gate. It
leaves the topmost qubit unchanged. The bottom qubit is then assigned the output of
taking the exclusive-or of both, i.e., this gate takes |0|0 → |0|0, |0|1 → |0|1,
|1|0 → |1|1, and |1|1 → |1|0. The X , Y , Z , R, H , and CNOT gates together
form the Clifford group of gates, which we will have more to say about later. If we
add the S gate to the Clifford group, they together form a universal set of gates,
i.e., any quantum circuit implementing any series of unitary transformations can be
simulated to arbitrary accuracy using combinations of these seven gates.
Figure 8b depicts a quantum circuit diagram for Deutsch’s Algorithm, which
determines whether a given function f on one bit is constant ( f (0) = f (1)) or
balanced ( f (0)
= f (1)): Two qubits are prepared in the product state |0|0 and are
each sent through an X -gate and a Hadamard gate, after which they are together input
to the two-qubit entangling unitary gate U f . The first qubit is then sent through a
further Hadamard gate and finally measured to yield the answer (see Deutsch, 1989).
Just as there are various models of universal classical computation (for instance,
the various versions of the Turing machine, as well as the von Neumann architecture,
and so on, that I mentioned above), there are various models of universal quantum
computation. One computational model that presents a particularly difficult problem
for those who would advocate for the many-worlds explanation is the cluster-state
model of quantum computing, also known as measurement-based and one-way quan-
tum computing (Raussendorf & Briegel, 2002; Raussendorf et al., 2003; Nielsen,
134 M. E. Cuffaro
Fig. 8 a A number of important one- and two-qubit quantum gates. The X , Y , Z , R, H , and CNOT
gates together form the Clifford group. If we add the S gate to the Clifford group, they together form
a universal set. b A quantum circuit diagram for Deutsch’s Algorithm, which determines whether
a given function f on one bit is constant ( f (0) = f (1)) or balanced ( f (0)
= f (1))
2006).33 In the cluster-state model, the computer begins, not in a product state but in
a highly entangled state, and measurements are performed not only at the end of the
computation but throughout it. These measurements are adaptive, in the sense that
each measurement is performed in a different basis, which depends on the random
outcomes of whatever previous measurements have been performed.
Why is this a problem for the advocate of the many-worlds explanation of quantum
computing? Because the fact that measurements are adaptive means, on the one hand,
that there is no principled way to select a preferred basis a priori in the context of a
given computation (Cuffaro, 2012, §4). Whichever basis we choose, few qubits will
actually be measured in that basis in the course of the computation. On the other
hand, there is no sense in which we can say that a preferred basis “emerges” from
the computational process. There is, therefore, no way to identify the worlds that the
computation as a whole is supposed to be happening in.
As I alluded to above, both the cluster-state model and the circuit model are
universal models of quantum computing. Thus, anything that the circuit model can
do can be done (and, as it happens, just as efficiently) in the cluster-state model and
vice versa. Perhaps, then, it could be argued that understanding the “true nature” of
algorithms in the cluster-state model requires that we first translate them into the
language of the circuit model, though I can think of no reason why one should think
so other than a desire to hold on to the many-worlds idea come what may. The proper
response to anyone who would put forward such an argument is that a large part
of what motivates those who adhere to the many-worlds explanation of quantum
computing in the first place is that it is useful for algorithm analysis and design to
33For introductions to cluster-state quantum computing aimed at philosophers, see Cuffaro (2012,
§4) and Duwell (2018, §4.5).
The Philosophy of Quantum Computing 135
believe that a quantum computer is carrying out its computations in parallel worlds.
This does not seem to be so for the cluster-state model. On the contrary, dogmatically
holding on to the view that many-worlds are, at root, physically responsible for
the speedup evinced in the cluster-state model is at best useless, for it is of no
help in designing algorithms for the cluster-state model. At worst, dogmatically
holding on to the many-worlds idea could prove positively detrimental if it prevents
us from exploiting the power of the cluster-state model or discovering other quantum
computational models in the future.
The many-worlds explanation of the power of quantum computers is not the only
explanation that has been proposed by philosophers. I have already mentioned the
upshot of Bub’s analysis of Shor’s algorithm in the previous section. Bub’s view,
more generally (2006, 2010), is that the key to understanding the power of quan-
tum computers lies in understanding the way that they exploit the novel logico-
probabilistic structure of quantum mechanics (Pitowsky, 1989). We saw above that
in quantum mechanics, the (classical) probability distribution over the answers to a
particular experimental question—the one that we infer, conditional upon our asking
that question—cannot be embedded into a global prior probability distribution over
all of the answers to all of the questions we might want to ask. But although these
individual probability distributions do not logically fit together as they do for a classi-
cal system, there are logical relations between them nonetheless, that are compactly
described by the quantum state.
For instance, as a general rule (which admits exceptions) one cannot use a quan-
tum system to compute a logical disjunction by computing the individual values
of each disjunct. This is simply because in the general case both disjuncts will not
be globally defined. However, quantum mechanics’ logical structure provides other
ways to compute a logical disjunction, and these other ways (which are not available
to a classical computer) are exploited by a quantum computer to compute certain
problems more efficiently than a classical computer can compute them.
Another view is Duwell’s (2018, 2021), who in contrast, agrees with the proponent
of the many-worlds explanation in identifying quantum parallelism as at least part of
the source of the power of quantum computers. Duwell, however, resists the tempta-
tion to make the a metaphysical inference from parallelism to many computational
worlds. For Duwell, the source of the power of a quantum computer lies in the way
that it can efficiently correlate multiple values of a function and use these correlations
to efficiently extract global information about the function. The disagreement over
whether a quantum computer performs more, or fewer, computations than a classical
computer is one that Duwell views as arising from conflicting intuitions about how
to appropriately describe the quantum systems that perform computational tasks.
In these and other candidate explanations of the power of quantum computers that
one encounters in the philosophical literature on the topic, the fact that quantum-
136 M. E. Cuffaro
mechanical systems exhibit entanglement (see Sect. 2.4) invariably plays an impor-
tant role. For Bub entanglement is absolutely central, as entanglement is a (direct)
manifestation of the fact that the logical structures of quantum and classical mechan-
ics differ. As for Duwell, he takes his quantum parallelism thesis to be completely
compatible with the idea that entanglement plays a central role in quantum comput-
ing, even if his explanation emphasizes the correlations between the values of the
function being evaluated by a system rather than the logical structure of its under-
lying state space (Duwell, 2018, p. 101). The many-worlds advocate, as well, views
quantum entanglement to be indispensable in the analysis of the power of quantum
computing (Hewitt-Horsman, 2009, p. 889), even if for the many-worlds advocate it
does not, by itself, suffice as a philosophical explanation for it.
The debate over the interpretation of the phenomenon of quantum entanglement
has historically been one of the central debates in the controversy over quantum the-
ory’s conceptual foundations. First emphasized by Albert Einstein, Boris Podolsky,
and Nathan Rosen in their 1935 criticism of the orthodox interpretation of quan-
tum mechanics, Erwin Schrödinger called it “the characteristic trait of quantum
mechanics, the one that enforces its entire departure from classical lines of thought”
(Schrödinger, 1935b, p. 555, emphasis in the original). It was only with the work of
Bell (2004 [1964], 2004 [1966]), however, that its significance for our understanding
of the break that quantum mechanics makes with classical physics was first made
fully clear.
Consider Louise and Robert, two friends who live in the 15th and 11th arrondisse-
ments, on the left bank and the right bank of the Seine, respectively, in the city of
Paris. Every morning, Louise and Robert each receive a letter in their respective
mailboxes that consists of a single sheet of paper inscribed with either a large bass
clef:
After awhile, Louise determines that it is equally likely, on any given day, that she
will receive a bass clef as it is that she will receive a treble clef. After awhile, Robert
determines the same. But when they compare their notes, they find that whenever
Robert receives a bass clef, so does Louise. Similarly, every time he receives a treble
clef, she does too. In other words, there is a one-in-two chance, on any given day, that
they both receive a bass clef, and a one-in-two chance that they both receive a treble
clef. No other combinations are ever observed. Thus, their outcomes are correlated,
The Philosophy of Quantum Computing 137
such that the probability distribution over the possible combinations of outcomes is
given by34 :
(4.1)
What explains this correlation? Well in this case it turns out that Louise and Robert
both play in a jazz ensemble. The band’s leader, Carsten, lives in the center of the city
(on the Île Saint-Louis). Every afternoon he flips a fair coin. Depending on whether
the coin lands heads or tails, he either writes a large bass clef or a large treble clef
on a sheet of paper, photocopies it, and sends one copy each to Robert and Louise. If
Robert receives a treble clef from Carsten, then he will know to take his tenor horn
to the jazz club that night. Otherwise, he will take his trombone. As for Linda, if she
receives a treble clef, she will know to bring her soprano clarinet. Otherwise, she
will bring her bassoon.
The result of Carsten’s coin flip is called the common cause of Louise’s and
Robert’s outcomes, and the story we tell about how Carsten’s coin flip determines
Louise’s and Robert’s outcomes is called a causal model for the correlations that they
see (see Pearl, 2009). If Louise and Robert do not know how Carsten is determining
what to send them, then they will wonder about the hidden-variable that explains
their correlation. In this case, the result of Carsten’s coin flip is actually a local
hidden-variable, since the process by which the outcome of the coin flip determines
what is written on the letters is confined to the small localized region in the vicinity
of the desk in Carsten’s central office. He flips the coin, takes note of the outcome
(the local hidden-variable), and writes the corresponding symbol.
Instead of simply flipping a coin while sitting at his desk, we can imagine a
more complicated, spatially distributed, process by which Carsten determines what to
write. For instance, Carsten might begin by flipping his coin, and then, corresponding
to heads or tails he might telephone Tilde or Bjarne, who live in the northern and
southern suburbs of the city, respectively, and ask whichever of them he calls to roll a
six-sided die at exactly seven-o’clock in the evening, and to afterward call him back
at either 7:13 PM, if it is Bjarne, or 7:18 PM, if it is Tilde, to tell him the outcome.
Then, if the result of the die roll is one, three, or six, Carsten will write a treble clef on
the sheet of paper before photocopying it and sending copies to Robert and Louise,
while if it is two, four, or five, he will write a bass clef. All of these actions together
constitute, just like the coin flip in our simpler scenario, a locally causal model to
explain Louise’s and Robert’s correlation. Why do we call it local even though it is
spatially distributed? Because the physical processes (the coin flips, die rolls, and
telephone calls) by which Carsten determines what to write propagate locally in the
physical sense, i.e., at a finite speed that is less than the speed of light.
Bell (2004 [1964]) showed, first, that there are certain probabilistic constraints—
what are now called Bell inequalities—on the statistics arising from measurements on
any locally correlated system. He then showed that these constraints are violated by
statistics that are predicted to arise from certain experiments on a system composed
of two qubits in the quantum-mechanical state35
1
√ |0 A |1 B − |1 A |0 B , (4.2)
2
where the A and B subsystems can be as far apart as one likes. The proof of this
violation is known as Bell’s theorem and it, and its variants, have since been experi-
mentally confirmed many times over (Genovese, 2016).
The predicted violation only occurs for certain measurements. If we measure
both the A and B qubits in the computational basis (see Eq. (2.6) and Fig. 7), then
the predicted statistics will actually be compatible with the constraints imposed by
local hidden-variable theories, as Bell himself showed (2004 [1964], p. 16). But
as we rotate the measurement basis (see Fig. 7) that we use for B away from the
measurement basis that we use for A, quantum mechanics predicts that we will see
a violation, that it will reach a maximum at a certain point, and then decrease again
as the measurement basis for B begins to line up again with the measurement basis
for A.
Another name for the computational basis is the Z -basis. We call it the Z -basis
because the two computational basis vectors |0 and |1 correspond to the two possible
outcomes of a Z -basis measurement on a qubit, where “measuring a qubit in the Z -
basis” means sending it through a Z -gate (see Fig. 8) and then recording its state
(Janas et al., 2022, §6.5). Similarly, the basis vectors |+ and |− are the two possible
outcomes of an X -basis measurement on a qubit, and |y + and |y − are the two
possible outcomes of a Y -basis measurement. The X , Y , and Z gates, together with
the trivial I gate that leaves a qubit’s state unchanged, are known as the Pauli gates.
I will have more to say about this family of gates later. For now, I want to point out
that, for a system of qubits in the state given by Eq. (4.2), as long as both the A and
B qubit are measured in one of the Pauli bases, the predicted statistics arising from
those measurements will not violate the constraints that Bell’s inequality imposes
on local hidden-variable theories. In other words, if all that we have access to are
measurement devices that can measure a qubit in one of the Pauli bases, then there
will be no way to disprove some local hidden-variable story a skeptic might cook up to
explain the observed statistics. To experimentally disprove such a story, we will need
to have measurement devices that can measure in other measurement bases besides
these ones, the ones for which Bell showed that a violation of the Bell inequalities
will occur.
35 This state is identical to the one given in Eq. (2.11) but we repeat it here for convenience.
The Philosophy of Quantum Computing 139
Bell’s theorem should convince us that no local hidden-variable theory can repro-
duce the correlations arising from such measurements (i.e., from measurements in
bases other than X , Y , and Z ), but this does not bar a skeptic from positing a nonlocal
hidden-variable theory to recover them instead. An example of such a theory is one
in which the outcome of a measurement on A depends on the measurement basis
used to measure B. But since A and B could in principle be far apart in space, the
causal influence from B to A in such a theory will have to be propagated faster than
the speed of light. This is a hard pill to swallow, but one might be inclined to swallow
it anyway in order to avoid the other alternatives.36
Coming back to our discussion of quantum computers, recall that when we dis-
cussed the many-worlds explanation of quantum computing, we noted that the com-
puter’s state as given on the right-hand side of Eq. (3.1) is entangled. The question
arises, then, as to what role is played by entanglement more generally in enabling
quantum computers to outperform their classical competitors. There are two ways to
think of this question. We might ask, on the one hand, whether realizing an entangled
state is necessary to enable a quantum speedup. On the other hand, we might ask
whether it is sufficient to enable it.
There has been some debate surrounding the first question. On the one hand, Jozsa
and Linden (2003) have proven that realizing an entangled state is necessary to enable
speedup if a quantum computer can be assumed to be in a so-called “pure state”, i.e.,
in a state that represents a maximally specific description of the system from the point
of view of the quantum-mechanical formalism. This is the case when we represent
a system with a state vector as we have been doing up until now (and will continue
to do). On the other hand, it has been argued (Biham et al., 2004) that quantum
computers that are in “mixed states”—i.e., states which describe the computer in
a less than maximally specific way, either because we are ignorant of the actual
state of the computer, or because the computer is coupled to its environment—are in
some cases capable of achieving a quantum speedup over their classical competitors
without ever being entangled. There is insufficient space to discuss this (largely
technical) debate here, but in Cuffaro (2013), I argue that the purported counter-
examples to what I there call the “necessity of entanglement thesis” do not actually
demonstrate what they purport to show, but instead clarify the necessary role that
entanglement does play in quantum computation.
From the philosopher’s point of view, the more interesting question in relation to
the role of entanglement in quantum computing is the question regarding sufficiency,
for as we will see, reflecting on this question sheds light on the tension between
physical and computational ways of thinking that is at the heart of the science of
quantum computing, the tension that is the primary source of the insight this new
science brings into both physics and computation.
The main reason for being skeptical of the idea that entanglement suffices to enable
quantum speedup is the Gottesman–Knill theorem (Gottesman, 1999). This theorem
states that any quantum algorithm that employs (exclusively) some combination of
the following operations (which together form what we will call the Gottesman–
Knill set) is efficiently simulable by a classical computer: (i) The Clifford group of
gates (see Fig. 8), i.e., the X , Y , Z , R, H , and CNOT gates; (ii) Clifford group gates
conditioned on the values of classical bits (indicating, e.g., the results of previous
measurements); (iii) state preparation of a qubit in the computational basis (as one
typically does for each qubit prior to the beginning of a computation); (iv) measure-
ments in one of the Pauli bases (as one does at the end of a computation) (Nielsen &
Chuang, 2000, §10.5.4).
The reason the theorem represents a challenge, to the view that entanglement
suffices to enable a quantum computer to outperform a classical computer, is that by
using only the operations given in (i)–(iv), which according to the theorem are all
efficiently classically simulable, it is possible to generate an entangled state. Figure 9
depicts a quantum circuit to realize an entangled state using only operations from the
Gottesman–Knill set. It begins with the preparation of two qubits in the computational
basis state |0, then subjects them each to an X -gate, as a result of which they will
both be in the state |1. The first qubit is then run through a Hadamard gate which
√
transforms it into the superposition state (|0 − |1)/ 2, and following that both qubits
are run through a CNOT gate. The resulting state of the computer is an entangled
state.
It would be wrong to conclude (cf. Jozsa & Linden, 2003, pp. 2029–30), on the
basis of this, that entanglement is not sufficient to enable a quantum computer to out-
perform a classical computer. Why? Well, let us reflect on what the Gottesman–Knill
theorem is saying. The Gottesman–Knill theorem shows that there is a certain set of
quantum operations that can be efficiently simulated on a classical computer. Let us
then consider what we mean when we say that we have “efficiently classically sim-
ulated” something. We can say, to start with, that in a classical computer simulation
(efficient or otherwise), whatever computer is doing the simulating will be such that
Fig. 9 A quantum circuit (above), written out explicitly in terms of the unitary transformations
required to realize it (below), involving only operations from the Gottesman–Knill set. At the end
of this sequence of operations, the computer will be in an entangled state
The Philosophy of Quantum Computing 141
37For a review of some of the uses envisioned for quantum computers in music, see Miranda (2021).
38This is actually the attitude (and for similar reasons) that many theorists take toward certain claims
by private corporations to have built a working quantum computer (see, e.g., Aaronson (2013a),
Shin et al. (2014)).
142 M. E. Cuffaro
systems composed of three or more parties as well (see Cuffaro, 2017, §A.2, which
summarizes a technique first described in Tessier, 2004).39
The upshot of the Gottesman–Knill theorem is that a certain number of quantum
operations are efficiently classically simulable despite being capable of generating
entangled states. And we have just seen that to say that some phenomenon is “effi-
ciently classically simulable” is to say that it can be efficiently redescribed in a
locally causal way. Since any sequence of Gottesman–Knill operations is equivalent
to a measurement in one of the Pauli bases, X , Y , Z , on a system whose state is
given by one of the basis vectors of a Pauli basis, the Gottesman–Knill theorem is
essentially telling us that there are some statistics associated with measurements on
systems in entangled states that admit of a locally causal description.
We do not really need the Gottesman–Knill theorem to tell us this, for Bell’s
and related inequalities amount to general constraints that any locally causal model
of combined measurement statistics must satisfy, and we already know that in
some cases, i.e., for measurements in one of the Pauli bases, the predictions of
a locally causal model are compatible with quantum-mechanical predictions. It is
therefore misleading to conclude, on the basis of the Gottesman–Knill theorem, that
entanglement is not sufficient to enable quantum computational speedup. What the
Gottesman–Knill theorem shows us is that one also needs to use this entanglement
to its full potential, in particular, by not restricting a quantum computer to use only
a small proportion of the operations of which it is capable (i.e., by not restricting it
to the Gottesman–Knill set).
In the context of more general discussions of quantum mechanics, there are those
who claim that entanglement is the one and only distinguishing feature of the theory
(Schrödinger, 1935b). This is a controversial claim, but it is not proved false by the
fact that Bell (2004 [1964]) showed that some operations on some entangled states
can be redescribed in a locally causal way. Likewise, pointing out essentially the
same thing in the context of quantum computation is not an objection to the claim
that quantum entanglement constitutes a sufficient physical resource for realizing
speedup over classical computation. Indeed, the Gottesman–Knill theorem serves to
highlight the role that is actually played by entanglement in a quantum computer and
to clarify why and in what sense it is sufficient to preclude the computer’s evolution
from plausibly being classically simulated (for more on this, see Cuffaro, 2017).
Above we mentioned that techniques have been developed for providing classical
computer simulations, i.e., locally causal descriptions, to recover the statistics aris-
ing from Pauli-basis measurements on entangled systems whose subsystems are in
superpositions of eigenstates of Pauli observables, for any number of parties. The
39 The n-party case, for n ≥ 3, introduces subtleties which we will discuss in the next section.
The Philosophy of Quantum Computing 143
classical computer simulations of three or more party systems are especially inter-
esting in that they necessarily involve a small amount of communication. For three
parties, A, B, and C, one bit must be communicated between B and A in order to
make the protocol work. For n parties, n − 2 bits are required, i.e., the number of
bits required scales linearly with n (Tessier, 2004). Note that this counts as being
“easy” in the complexity-theoretic sense we discussed in Sect. 2.2.
From a certain point of view, however, it may seem quite odd to call a classi-
cal computer simulation of quantum correlations, that involves a certain amount of
communication between parties, a locally causal model of those correlations. Recall
that, earlier, when we introduced Bell’s theorem, we claimed that communication
between distant parties in the context of a Bell experiment constitutes a nonlocal
influence, i.e., one that would have to be communicated at a speed exceeding that of
light.
We can start to resolve the tension here if we consider, first, just what a locally
causal redescription of quantum measurement statistics is required to do. If a pair
of qubits, A and B, are prepared in an entangled state and then spatially separated,
quantum mechanics tells us that when, for instance, a Z -basis measurement is per-
formed on B, A’s state changes “instantaneously”, so that a Z -basis measurement on
A will yield an outcome that is correlated with the result of the experiment on B in a
particular way. Our challenge, however, is not to reproduce these apparent dynamics.
Our challenge is to come up with an alternative description of the statistics that we
actually observe in our experiments.
In order to assert that one has actually observed the results of a combined exper-
iment on two or more spatially separated subsystems, it must be assumed that the
results associated with the individual experiments on those subsystems have actually
been combined in some way. That is, we must somehow gather together the results
registered locally at the sites of the experiments on A and B. If Alice (a bassist)
is assigned to measure A, for instance, and Bob (who plays the piano) to measure
B, then once they record their individual results, they need to communicate them,
either to one another, or perhaps to Carsten who is sitting in his office on the Île
Saint-Louis, where he is in the process of showing Quadeisha how his quantum
computer is but one part of a network of spatially distributed quantum computers
that are together performing a computation that depends on the prior generation of
a system in an entangled state. While Alice and Bob are making their way toward
Carsten and Quadeisha with their results, however, there is time for Alice and Bob to,
surreptitiously, exchange further information with one another by local means (Bob
might call Alice on his cell phone, for instance, as they are making their way toward
Carsten’s office) in order to coordinate the outcomes of their individual measure-
ments and “correct” them if necessary, unbeknownst to Quadeisha. And since this
further information is propagated locally, i.e., at a speed no greater than that of light,
it can be considered as part of the common cause of the actually observed (combined)
measurement event, and thus as part of a classical, locally causal, description of that
measurement event (see Cuffaro, 2017, §4, Fig. 1).
Now, given a particular combined calculational outcome at a time t, one can
try to explain it in one of two ways: (i) Carsten’s way, i.e., as the result of a bona
144 M. E. Cuffaro
40Quadeisha’s loophole is actually conceptually similar to the “collapse locality” loophole at the
heart of Adrian Kent’s causal quantum theory (Kent, 2005). For discussion see Cuffaro (2017, pp.
104–106). For a more general discussion of the methodology of no-go theorems, see Dardashti
(2021).
The Philosophy of Quantum Computing 145
41 Note that I am taking ‘tractable’ here in a relative sense. That is, the resources required by a
classical computer to reproduce a particular effect should differ tractably from those required by a
quantum computer. Or in other words: it should not be essentially harder for the classical system
to produce the same effect as the quantum system.
42 The general subject of classical simulations of quantum systems is an important and burgeoning
area of modern physics (see, for example, Borges et al., 2010; Lee & Thomas, 2002).
146 M. E. Cuffaro
‘plausible’ in each case. In the practical context we assume that a system has been
built by a rational agent to achieve a particular purpose. We do not presuppose this
in the theoretical context.43 Clearly, certain assumptions that would be regarded as
plausible in the former context will not be in the latter and it is important to be on
guard against conflating them. As we saw in the previous section, practical investi-
gators attempting to isolate and/or quantify the computational resources provided by
physical systems may be in danger of conceptual confusion if they are not cognisant
of the differences between the practical and theoretical contexts. More positively, it
must not be assumed by practical investigators that every ‘no-go result’, formulated
in the theoretical context, will constrain what they can accomplish in the practical
context (Cuffaro, 2018a). More generally it is important to think seriously about
what one means by a ‘plausible locally causal model’ regardless of the context of
investigation.
Calling a spade a spade gives us, moreover, insight into how the practical context
can illuminate the investigations of the traditional philosopher of physics in the con-
text of quantum mechanics. The silly story we told in the previous section about how
Carsten, Alice, and Bob manage to reproduce, by communicating just a few classical
bits, all of the correlational phenomena associated with Pauli-basis measurements on
an n-party quantum system in an entangled state, besides representing just another
silly classical story, emphasizes perhaps the most fundamental difference that we can
point to between classical and quantum mechanics. Classical and quantum mechanics
are, fundamentally, alternative universal languages for describing physical systems
(Janas et al., 2022, §6.3). And by investigating the computational power inherent
in those languages we gain insight into the respective logical structures of classical
and quantum theory, into the logical structure of quantum mechanics that enables
the efficient representation of correlational phenomena by quantum-mechanical sys-
tems, and the logical structure of classical mechanics which precludes this (Janas et
al., 2022, §4.2).
8 Concluding Remarks
This is a lesson with which the artist will be familiar.44 From time to time, in
literature, music, and in other forms of art, new methods of writing and composing
emerge that allow one to express the subtleties and nuances of experience more
43 This statement is not meant to express any sort of theological opinion. It is merely a statement
about how science operates, at least in this century.
44 A similar point is made in Janas et al. (2022, Sect. 6.3, note 22).
The Philosophy of Quantum Computing 147
easily than before. E. M. Forster (1942, p. 28) once said, about Virginia Woolf, that
“She pushed the light of the English language a little further against the darkness.”
In music, the (re-)introduction, into the Western musical tradition, of modality and
chromaticism through Jazz, Blues, and other forms of popular music, and also through
art music (see Vincent, 1951), has enabled modern composers in the Western tradition
to explore musical landscapes and express aspects of our emotional and intellectual
experience that composers limited to the major and minor scales, so dominant during
Western music’s classical period, cannot easily capture. In his musings about quantum
mechanics, Bohr himself was wont to appeal to an analogy with poetry. Werner
Heisenberg recalls, in his memoirs, a conversation he had with Bohr in which the
latter stated that:
We must be clear that, when it comes to atoms, language can be used only as poetry. The poet,
too, is not nearly so concerned with describing facts as with creating images and establishing
mental connections” (Heisenberg, 1971, p. 41).
These and like analogies were not lost on actually practising poets and other artists
of the same period (see Mairhofer, 2021). The subject of quantum mechanics has
even been found to lend itself very naturally to the format of the comic book (see Bub
& Bub, 2018, and the review by Cuffaro & Doyle, 2021, especially §4). As for the
science of quantum computing: The science of quantum computing takes advantage,
as we have seen throughout the course of this chapter, of the increased expressive
power of the quantum-mechanical language. It shows us that using the new language
of quantum mechanics, we can interact with what we take to be quantum-mechanical
systems in ways that cannot be easily replicated classically.
In the course of this chapter we have reflected on the fundamentals of classical and
quantum physics and computation, on the resources used and on the explanation of the
power of quantum computers, and finally on the broader insights that can be gleaned
from the science of quantum computing both for physics and for computation. The
philosophical issues addressed in this chapter are not the only issues brought up by
the science of quantum computing. But I hope to have convinced the reader, with
this sample of some of the more central ones, of quantum computing’s potential for
illuminating the world that we are all participants in.
References
Aaronson, S. (2013a). D-Wave: Truth finally starts to emerge. Posted: June 05, 2013. Retrieved
August 11, 2014, from
www.scottaaronson.com/blog/?p=1400
Aaronson, S. (2013b). Quantum computing since Democritus. Cambridge University Press.
Aaronson, S. (2013c). Why philosophers should care about computational complexity. In B. J.
Copeland, C. J. Posy, & O. Shagrir (Eds.), Computability: Turing, Gödel, Church, and beyond
(pp. 261–327). MIT Press.
Aaronson, S. (2016). Complexity zoo. complexityzoo.uwaterloo.ca/Complexity_Zoo.
Adlam, E. (2014). The problem of confirmation in the Everett interpretation. Studies in History and
Philosophy of Modern Physics, 47, 21–32.
148 M. E. Cuffaro
Agrawal, M., Kayal, N., & Saxena, N. (2004). PRIMES is in P. Annals of Mathematics, 160, 781–
793.
Aharonov, D., van Dam, W., Kempe, J., Landau, Z., Lloyd, S., & Regev, O. (2007). Adiabatic quan-
tum computation is equivalent to standard quantum computation. SIAM Journal on Computing,
37, 166–194.
Albert, D., & Loewer, B. (1988). Interpreting the many worlds interpretation. Synthese, 77, 195–213.
Arora, S., & Barak, B. (2009). Computational complexity: A modern approach. Cambridge Uni-
versity Press.
Bell, J. S. (2004 [1964]). On the Einstein-Podolsky-Rosen paradox. In Speakable and unspeakable
in quantum mechanics (pp. 14–21). Cambridge University Press.
Bell, J. S. (2004 [1966]). On the problem of hidden variables in quantum mechanics. In Speakable
and unspeakable in quantum mechanics (pp. 1–13). Cambridge University Press.
Bell, J. S. (2004 [1981]). Bertlmann’s socks and the nature of reality. In Speakable and unspeakable
in quantum mechanics (pp. 139–158). Cambridge University Press.
Bennett, C. H., Bernstein, E., Brassard, G., & Vazirani, U. (1997). Strengths and weaknesses of
quantum computing. SIAM Journal on Computing, 26, 1510–1523.
Bernstein, E., & Vazirani, U. (1997). Quantum complexity theory. SIAM Journal on Computing,
26, 1411–1473.
Biham, E., Brassard, G., Kenigsberg, D., & Mor, T. (2004). Quantum computing without entangle-
ment. Theoretical Computer Science, 320, 15–33.
Bohr, N. (1948). On the notions of causality and complementarity. Dialectica, 2, 312–319.
Boole, G. (1847). The mathematical analysis of logic. Philosophical Library. Reprinted: Thoemmes
Press (1998).
Borges, C. V. S., Hor-Meyll, M., Huguenin, J. A. O., & Khoury, A. Z. (2010). Bell-like inequality
for the spin-orbit separability of a laser beam. Physical Review A, 82, 033833.
Brassard, G., Cleve, R., & Tapp, A. (1999). Cost of exactly simulating quantum entanglement with
classical communication. Physical Review Letters, 83, 1874–1877.
Briegel, H. J., Browne, D. E., Dür, W., Raussendorf, R., & den Nest, M. V. (2009). Measurement-
based quantum computation. Nature Physics, 5, 19–26.
Bub, J. (2006). Quantum computation from a quantum logical perspective.
arXiv:quant-ph/0605243v2
Bub, J. (2008). Quantum computation and pseudotelepathic games. Philosophy of Science, 75,
458–472.
Bub, J. (2010). Quantum computation: Where does the speed-up come from? In A. Bokulich & G.
Jaeger (Eds.), Philosophy of quantum information and entanglement (pp. 231–246). Cambridge
University Press.
Bub, J. (2016). Bananaworld, quantum mechanics for primates. Oxford University Press.
Bub, J., & Pitowsky, I. (2010). Two dogmas about quantum mechanics. In S. Saunders, J. Barrett,
A. Kent, & D. Wallace (Eds.), Many worlds? Everett, quantum theory, and reality (pp. 433–459).
Oxford University Press.
Bub, T., & Bub, J. (2018). Totally random: Why nobody understands quantum mechanics. Princeton
University Press.
Church, A. (1936). An unsolvable problem of elementary number theory. American Journal of
Mathematics, 58, 345–363.
Cobham, A. (1965). The intrinsic computational difficulty of functions. In Y. Bar-Hillel (Ed.),
Logic, methodology and philosophy of science: Proceedings of the 1964 international congress
(pp. 24–30). North-Holland.
Cook, W. J. (2012). In pursuit of the traveling salesman: Mathematics at the limits of computation.
Princeton University Press.
Copeland, J. B. (2017). The modern history of computing. In E. N. Zalta (Ed.), The Stanford
encyclopedia of philosophy (Winter 2017 ed.). Metaphysics Research Lab, Stanford University.
Copeland, J. B. (2020). The Church-Turing thesis. In E. N. Zalta (Ed.), The Stanford encyclopedia
of philosophy (Summer 2020 ed.). Metaphysics Research Lab, Stanford University.
The Philosophy of Quantum Computing 149
Cuffaro, M. E. (2012). Many worlds, the cluster-state quantum computer, and the problem of the
preferred basis. Studies in History and Philosophy of Modern Physics, 43, 35–42.
Cuffaro, M. E. (2013). On the necessity of entanglement for the explanation of quantum speedup.
arXiv:1112.1347v5
Cuffaro, M. E. (2017). On the significance of the Gottesman-Knill theorem. The British Journal for
the Philosophy of Science, 68, 91–121.
Cuffaro, M. E. (2018). Reconsidering no-go-theorems from a practical perspective. The British
Journal for the Philosophy of Science, 69, 633–655.
Cuffaro, M. E. (2018). Universality, invariance, and the foundations of computational complexity
in the light of the quantum computer. In S. O. Hansson (Ed.), Technology and mathematics:
Philosophical and historical investigations. Springer.
Cuffaro, M. E. (2020). Information causality, the Tsirelson bound, and the ‘being-thus’ of things.
Studies in History and Philosophy of Modern Physics, 72, 266–277.
Cuffaro, M. E., & Doyle, E. P. (2021). Essay review of Tanya and Jeffrey Bub’s Totally random:
Why nobody understands quantum mechanics: A serious comic on entanglement. Foundations of
Physics, 51, 28:1–28:16.
Cuffaro, M. E., & Hartmann, S. (2021). The open systems view. arXiv:2112.11095v1
Curiel, E. (2014). Classical mechanics is Lagrangian; it is not Hamiltonian. The British Journal for
Philosophy of Science, 65, 269–321.
Dardashti, R. (2021). No-go theorems: What are they good for? Studies in History and Philosophy
of Science, 86, 47–55.
Davis, M. (2000). The universal computer: The road from Leibniz to Turing. W. W. Norton and
Company.
Dawid, R., & Thébault, K. P. Y. (2015). Many worlds: Decoherent or incoherent? Synthese, 192,
1559–1580.
Dawson, J. W., Jr. (2007). Classical logic’s coming of age. In D. Jacquette (Ed.), Philosophy of
logic (pp. 497–522). Elsevier.
Dean, W. (2016). Squeezing feasibility. In A. Beckmann, L. Bienvenu, & N. Jonoska (Eds.) Pursuit
of the Universal: Proceedings of the 12th Conference on Computability in Europe (pp. 78–88).
Springer International Publishing.
Deutsch, D. (1985). Quantum theory, the Church-Turing principle and the universal quantum com-
puter. Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences,
400, 97–117.
Deutsch, D. (1989). Quantum computational networks. Proceedings of the Royal Society of London.
Series A. Mathematical and Physical Sciences, 425, 73–90.
Deutsch, D. (1997). The fabric of reality. Penguin.
Deutsch, D. (2010). Apart from universes. In S. Saunders, J. Barrett, A. Kent, & D. Wallace (Eds.),
Many worlds? Everett, quantum theory, and reality (pp. 542–552). Oxford University Press.
DeWitt, B. S. (1973 [1971]). The many-universes interpretation of quantum mechanics. In DeWitt
& Graham (1973, pp. 167–218).
Duwell, A. (2007). The many-worlds interpretation and quantum computation. Philosophy of Sci-
ence, 74, 1007–1018.
Duwell, A. (2018). How to make orthogonal positions parallel: Revisiting the quantum parallelism
thesis. In M. E. Cuffaro & S. C. Fletcher (Eds.), Physical perspectives on computation, compu-
tational perspectives on physics (pp. 83–102). Cambridge University Press.
Duwell, A. (2021). Computation and physics. Cambridge University Press.
Edmonds, J. (1965). Paths, trees, and flowers. Canadian Journal of Mathematics, 17, 449–467.
Einstein, A., Podolsky, B., & Rosen, N. (1935). Can quantum-mechanical description of physical
reality be considered complete? Physical Review, 47, 777–780.
Everett III, H. (1956). The theory of the universal wave function. In DeWitt & Graham (1973, pp.
3–140).
Farhi, E., Goldstone, J., Gutmann, S., & Sipser, M. (2000). Quantum computation by adiabatic
evolution. Tech. Rep. MIT-CTP-2936, MIT. arXiv:quant-ph/0001106
150 M. E. Cuffaro
Andrei Khrennikov
A. Khrennikov (B)
International Center for Mathematical Modeling in Physics and Cognitive Sciences, Linnaeus
University, 351 95 Växjö, Sweden
e-mail: [email protected]
1 Introduction
‘Information’ is the new paradigm reflecting the main character of modern human
society as an information society. The tremendous development of information tech-
nologies of the last 20 years has dramatically changed our lifestyle through the World
Wide Web of the Internet and the mobile-connectivity web that have led to the cre-
ation of virtual social networks. Recent information technologies caused the digital
transformation of human communications and, consequently, of the whole society.
Unavoidably, challenges of the new information culture become the focus of inten-
sive studies and a meeting point for researchers not only from physics, information
engineering, and artificial intelligence, but also from biology, psychology, decision-
making, cognitive, social and political sciences, economics and finance. The aim of
this note to review briefly the recent applications of the quantum information and
probability theories outside of physics. This line of research is known as quantum-like
modeling (Asano et al., 2011, 2012, 2014, 2015a, 2015b, 2017a, 2017b; Basieva &
Khrennikov, 2015; Bagarello & Oliveri, 2013; Busemeyer & Bruza, 2012; Bagarello
et al., 2019; 2018; Basieva et al., 2017; Busemeyer et al., 2006; Boyer-Kassem et
al., 2015;Busemeyer et al., 2014; Dzhafarov & Kujala, 2012; Dzhafarov et al., 2015;
Haven, 2005; Haven & Khrennikov, 2013; Haven et al., 2017; Khrennikov, 1999,
2003, 2004a, 2004b, 2006, 2010, 2014, 2015a, 2015b, 2016; Khrennikov et al., 2014;
Khrennikova, 2016, 2017; Khrennikov & Basieva, 2014b,2014a; Khrennikov et al.,
2018; Ozawa & Khrennikov, 2020a, 2020b; Pothos & Busemeyer, 2009; Surov et al.,
2019; Wang & Busemeyer, 2013, 2014; White et al., 2014). This is not consideration
of genuine quantum physical processes in cognitive systems, but modeling behavior
of generally macroscopic cognitive or AI agents processing information on the basis
of quantum laws.
We want to present in more detail consequences of such information processing
for rationality. In classical decision-making, rational agents are mathematically mod-
eled as probabilistic information processors using Bayesian update of probabilities:
rational = Bayesian. Quantum state update is generally non-Bayesian (Ozawa &
Khrennikov, 2020b). We define quantum rationality as decision-making that is based
on quantum state update. Quantum and classical rational agents behave differently.
For instance, a quantum(-like) agent can violate the Savage Sure Thing Principle
(Savage, 1954) (see Busemeyer & Bruza, 2012; Haven & Khrennikov, 2013; Haven
et al., 2017) and the Aumann theorem (Aumann, 1976) on impossibility of agree-
ing to disagree (see Haven et al., 2017; Khrennikov & Basieva, 2015b; Khrennikov,
2014b).
In quantum-like modeling, the brain is treated as a black box in which informa-
tion processing cannot be described by classical probability (Kolmogorov, 1933)
(CP) (cf. with genuine quantum physical models of brain’s functioning (Arndt et
al., 2009; Bernroider, 2017; Bernroider & Summhammer, 2012; Hameroff, 1994;
Penrose, 1989; Umezawa, 1993; Vitiello, 1995, 2001)). And there is a plenty of
such nonclassical statistical data—in cognitive psychology, game theory, decision-
making, social science, economics, finances, and politics. In decision theory, such
Quantum-Like Cognition and Rationality: Biological and Artificial Intelligence Systems 155
data was coupled to probability fallacies and irrational behavior of agents. We pro-
pose to apply the most well-developed nonclassical theory of probability, namely,
based on the mathematical formalism of quantum theory.
One may think that the appeal to quantum probability (Khrennikov, 2016) (QP)
and information to model decision-making by humans is too exotic. However, we
recall that as early as the 1970s, Tversky (one of the most cited psychologists of all
time) and Kahneman (Nobel prize in economics in 2002, for prospect theory, which
he co-developed with Tversky) have been demonstrating cases where CP-prescription
and human behavior persistently diverge (Kahneman, 2003; Kahneman & Tversky,
1972, 1979, 1984; Kahneman & Thaler, 2006). Today, we are at the theoretical
cross-roads, with huge divisions across conflicting, entrenched theoretical positions.
Should scientists continue use CP as the basis for descriptive and normative
predictions in decision making and accept continuous generation of paradoxes?
Should we abandon probability theory completely and instead pursue explana-
tions based on heuristics, as Tversky and Kahneman proposed?
The use of the probabilistic and statistical methods is really the cornerstone of the
modern scientific methodology. Thus, although the heuristic approach to decision-
making cannot be discarded completely, it seems more natural to search novel proba-
bilistic models for decision-making. Our suggestion is to use QP and more generally
quantum information, instead of heuristics of Tversky and Kahneman.
We stress that quantum-like modeling does not appeal to genuine quantum physi-
cal processes in biosystems. Quantum-like information processing can be performed
by macroscopic systems, as cells, animals, or humans. Even artificial intelligence
need not be based on quantum physical processors, as, e.g., quantum computers.
Quantum-like modeling opens the door to creation of AI-systems processing infor-
mation by respecting the laws of quantum theory, but equipped with classical phys-
ical processors. Some step in this direction was done within the recent studies on
quantum(-like) information retrieval (see, e.g., Aerts et al., 2019; Melucci, 2015;
van Rijsbergen, 2004): algorithms based on the complex Hilbert state representation
of information, but driven on classical computers, demonstrate superiority compar-
ing with the traditional (“classical”) algorithms. Of course, creation of successfully
working genuine quantum computers and simulators would give the possibility for
creation of the genuine quantum AI-systems. However, since behavior of quantum-
like and genuine quantum AI-systems is based on the same formalism and method-
ology, the theory of quantum-like cognition, rationality, and artificial intelligence
would be useful even for quantum physical AI-systems.
For reader’s convenience, this paper contains basics of CP and QP. We present
them in two steps. Section 2 is the informal comparative introduction to CP versus
QP; then Sects. 7, 9 present briefly, but mathematically rigorously the CP and QP
formalisms (the later is presented as a part of the quantum formalism: quantum states
and observables, the Born rule for probability of outcome of a measurement, the
projection postulate and quantum state update). However, these formal presentations
of CP and QP are used only in the last part of this work devoted to comparison of
classical and quantum versions of the Aumann theorem. The preceding part devoted
to classical vs. quantum rationality is written at the heuristic level.
156 A. Khrennikov
In fact, powerful integration theory that is needed for calculation of averages demands
σ-additivity:
P(∪ j O j ) = P(O j ), (1)
j
where O j ∩ Oi = ∅, i = j.
QP is the calculus of complex amplitudes or in the abstract formalism complex
vectors. Thus, instead of operations on probability measures one operates with vec-
tors. We can say that QP is a vector model of probabilistic reasoning. Each complex
amplitude ψ gives the probability by Born’s rule: Probability is obtained as the
square of the absolute value of the complex amplitude.
p = |ψ|2 (2)
(for the Hilbert space formalization, see Sect. 9, Formula (17). By operating with
complex probability amplitudes, instead of the direct operation with probabilities,
one can violate the basic laws of CP.
In CP, the law of total probability (LTP) is derived by using additivity of probability
and the Bayes formula, the definition of conditional probability,
P(O2 ∩ O1 )
P(O2 |O1 ) = , P(O1 ) > 0. (3)
P(O1 )
+2 cos θα1 α2 P(A = α1 )P(B = β|A = α1 )P(A = α2 )P(B = β|a = α2 )
α1 <α2
(5)
If the interference term is positive, then the QP-calculus would generate a probability
that is larger than its CP-counterpart given by the classical LTP (2). In particular, this
probability amplification is the basis of the quantum computing supremacy.
Therefore, violation of LTP and, hence, of STP (and classical rationality) is generated
either by violation of additivity of probability or the Bayes law for conditional proba-
bility or by the combination of these factors. Generally, this leads to the impossibility
to use in decision-making Bayesian inference. Quantum(-like) agents proceed with
more general inference machinery based on the quantum state update.
Hence, classical rationality is Bayesian inference rationality and quantum ratio-
nality is non-Bayesian inference rationality.1 In the light of above considerations,
one can ask:
Are quantum agents irrational?
As was discussed, by using QP it is possible to violate LTP and hence STP.
Therefore, generally quantum-like agents are (classically) irrational. However, we
can question the classical probabilistic approach to mathematical formalization of
decision making and, consequently, the corresponding notion of rationality. We define
quantum(-like) rationality as respecting the quantum calculus of probabilities and
the quantum formula for interference of probabilities, LTP with the interference term
(5).
1Of course, non-Bayesian probability updates are not reduced to quantum, given by state transfor-
mations in the complex Hilbert space. One may expect that human decision-making violates not
only classical, but even quantum rationality.
Quantum-Like Cognition and Rationality: Biological and Artificial Intelligence Systems 159
We point out that recently humans started to use widely the quantum decision-making
technique, especially in social and political decision-making. In particular, this can
explain the recent trend in increasing dysfunctional disagreement in, e.g., political
debates and generally in social life of the modern society. This disagreement cannot
be explained by insufficient information supply. There is a lot of information. The
problem is in opposite—in information overload (Khrennikov, 2020). The informa-
tion flows generated by mass-media and internet are so powerful, that by making a
decision on the concrete complex problem humans are not able to construct the joint
algebra for all possible events related to the problem under consideration. Different
issues involved in a problem are treated in different algebras. Moreover, these alge-
bras can differ essentially depending on agent’s social network and his information
environment (or it would better to say “a part of the information environment that is
selected by this agent”).
In the situation when an agent is overloaded with variety of information and
he or she should make rapidly the decision on a complex problem, the quantum
information processing (without attempting to unify all received information within
a single Boolean algebra) demonstrates its superiority over classical one, at least
from the viewpoint of minimization of computational resources and speed up of
decision-making. Thus, in the modern society quantum rationality beats classical
rationality, with “just one” casualty—inconsistency of some decisions. However,
this is inconsistency w.r.t. classical Boolean logic. If all agents behave quantumly,
then such inconsistencies become invisible.
The above discussion is applicable not only to biological systems, as say humans,
but also to AI-systems. The latter when operating with powerful information flows
would also prefer to use the quantum logic and QP. For each problem, a quantum(-
like) AI-system, say robot, selects a proper Boolean algebra, but at the same time
it keeps the possibility to use other Boolean algebras, corresponding to selection
of different orthonormal bases in the Hilbert state space. However, transition from
one Boolean algebra to another realized with a unitary operator U in the state space
demands a lot of computational resources, because state space has big dimension;
for n-qubit state space, it is 2n . In situation of information overload and temporal
constraints, an agent (biological or AI) has no possibility to perform such transition.
Moreover, to get more or less complete image of the situation, an agent has to make
transitions to a variety of Boolean algebras corresponding to incompatible variables
which are represented by noncommuting Hermitian operators.
160 A. Khrennikov
One of the consequences of information overload is that information loses its con-
tent. A human has no possibility to analyze deeply the content of communications
delivered by mass-media and social networks. People process information without
even attempting to construct an extended Boolean algebra of events. They oper-
ate with labels such as say COVID-19, vaccination, pandemy without trying to go
deeper beyond this labels. Contentless information behaves as a bosonic quantum
field which is similar to the quantum electromagnetic field. Interaction of humans
with such quantum information field can generate a variety of quantum-like behav-
ioral effects. One of them is social lasing, stimulated amplification of social actions
(SASA) (Khrennikov, 2015a, 2016, 2018; Khrennikov et al., 2018, 2019; Tsarev
et al., 2019). In social laser theory, humans play the role of atoms, social atoms
(s-atoms). Interaction of the information field composed of indistinguishable (up
to some parameters, as say social energy) excitations with gain medium composed
of s-atoms generate the cascade type process of emission of social actions. SASA
describes well, e.g., color revolutions and other types of mass protests (see Khren-
nikov, 2020 for detailed presentation).
Aumann’s approach (Aumann, 1976) to common knowledge and his theorem that
rational agents cannot agree on disagree play the crucial role in theory of rationality.
This theorem implies that rational agents acting under very natural conditions would
never agreeing on disagree. Hence, it couples rationality with consistency of actions
of decision makers. The main puzzle raised by Aumann’s theorem is that people
often “agree on disagree”.
How can we explain this contradiction between the statement which was mathe-
matically proved in the well-established framework of decision theory and the real
behavior of humans?
The typical solution is that one of the two basic assumptions of the Aumann
theorem is violated in the real processes of decision-making. We recall the Aumann’s
assumptions:
1. Common knowledge.
2. Common priors.
However, in modern information society these assumptions are very natural. Society
is homogeneous and the majority of people have common priors. Information is
openly distributed via mass-media and internet. Of course, there are attempts to
use insider information. But, such attempts, e.g., at the financial market, are subject
Quantum-Like Cognition and Rationality: Biological and Artificial Intelligence Systems 161
The notion of common knowledge plays the crucial role in various problems of coordi-
nation of actions and approaching conventions—in philosophy, economics (including
accounting and capital market research), game theory, statistics, computer science,
artificial intelligence. This notion is not simple and we shall devote this section to
the informal discussion, irrelevant to the use of CP or QP. For illustration, we shall
present the examples of the crucial impact of common knowledge in the problems
of social coordination. The first one is based on author’s personal experience and the
second one is commonly used in literature on common knowledge.
162 A. Khrennikov
2 Comment: earthquakes nearby Tokyo happen often; in some periods things in room shake practi-
cally everyday. Amplitudes vary day to day; of course each time I estimated their strength. But this
is difficult to do subjectively....
Quantum-Like Cognition and Rationality: Biological and Artificial Intelligence Systems 163
eyes. For island’s inhabitants, the number k is unknown. At the beginning nobody
knows the color of her/his eyes. There is the very strict rule:
If a person finds that she/he has blue eyes, that person must move from the island
before sunrise.
At this island everybody knows eye colors of others. But, it is forbidden to discuss
eye colors and there are no mirrors or similar devices. Once a stranger comes to the
island and announced to all the people:
“At least one of you has blue eyes”.
Thus it became common knowledge. The problem: what is the eventual outcome
of this public announcement?
Consider first the simplest case, k = 1. Some person will recognize that she/he
alone has blue eyes (by seeing only green eyes in the others) and leave at the first
sunrise. Let now k = 2. At the first sunrise nobody leaves the island. Then two people
having blue eyes by seeing only one person with blue eyes, and that no one left on
the first sunrise understand that k > 1. They leave at the second sunrise. And so on
by using the inductive argument. The paradox is that if k > 1 then the stranger told
to people at this island what they already have known: there are blue-eyed people.
However, without stranger’s announcement this fact was not common knowledge.
Its becoming common knowledge had dramatic consequences for inhabitants of the
island.
For k = 2, it is first-order knowledge. Each person having blue eyes knows that
there is someone with blue eyes, but she/he does not know that the other blue-eyed
person has this same knowledge. For k = 3, it is second order knowledge. After 2
days, each person having blue eyes knows that a second blue-eyed person knows that
a third person has blue eyes, but no one knows that there is a third blue-eyed person
with that knowledge, until the third day arrives. And so on...
(, F, P),
We remind that a σ-algebra is a system of sets which contains and empty set,
it is closed with respect to the operations of countable union and intersection and to
the operation of taking the complement of a set. For example, the collection of all
subsets of is a σ-algebra. This σ-algebra is used in the case of finite or countable
sets:
= {ω1 , ..., ωn , ...}. (6)
However, for “continuous sets”, e.g., = [a, b] ⊂ R, the collection of all possi-
ble subsets is too large to have applications. Typically it is impossible to describe a
σ-algebra in the direct terms. To define a σ-algebra, one starts with a simple system
of subsets of and then consider the σ-algebra which is generated from this sim-
ple system with the aid of aforementioned operations. In particular, one of the most
important for applications σ-algebras, the so-called σBorel-algebra, is constructed
in this way by staring with the system consisting of all open and closed intervals of
the real line. In a metric space (in particular, in a Hilbert space), the Borel σ-algebra
is constructed by starting with the system of all open and closed balls.
Finally, we remark that in American literature the term σ-field is typically used,
instead of σ-algebra.
The probability is defined as a measure, i.e., a map from F to nonnegative real
numbers which is σ-additive:
P(∪ j A j ) = P(A j ), (7)
j
In the case of a discrete probability space, see (6), the probability measures have
the form
P(A) = p j , p j = P({ω j }).
ω j ∈A
In fact, any finite measure μ, i.e, μ() < ∞, can be transformed into the probability
measure by normalization:
μ(A)
P(A) = , A ∈ F. (9)
μ()
In the same way we define the real (and complex) vector-valued random variables,
ξ : → Rn and ξ : → Cn .
Let ξ1 , ..., ξk be real-valued random variables. Their join probability distribution
Pξ1 ,...,ξk is defined as the probability distribution of the vector-valued random vari-
able ξ = (ξ1 , ..., ξk ). To determine this probability measure, it is sufficient to define
probabilities
where j , j = 1, ..., k, are intervals (open, closed, half-open) of the real line.
We remark once again that LTP (5) is a theorem within Kolmogorov probability
theory (Kolmogorov, 1933). We also recall that LTP plays the basic role in Bayesian
probability inference.
Each agent creates its information representation for possible states of the world
based on its own possibilities to perform measurements, “to ask questions to the
world.” Mathematically these representations are given by partitions of : P (i) =
(P j(i) ), where, for each agent i,
Thus an agent cannot get to know the state of the world ω precisely; she/he can
only get to know to which element of its information partition P j(i) ≡ P (i) (ω) this ω
belongs. In this set-theoretic model of knowledge, by definition the agent i knows
166 A. Khrennikov
an event E in the state of the world ω if the element of his information partition
containing this ω is contained in E :
In logical terms, this can be written as P (i) (ω) ⇒ E, the event P (i) (ω) implies the
event E; we also remark that {ω} ⇒ P (i) (ω).
P j(1)
1
∩ ... ∩ P j(N
N
)
→ p j1 ... jN ≡ p(P j(1)
1
∩ ... ∩ P j(N
N
)
). (12)
Thus by agreeing on the prior the agents have to agree on numerous conjunctive
probabilities.
O = ∪m P j(i)
m
, (13)
In theory of common knowledge, the basic role is played by the set of all states
of the world for which E is common knowledge; it is denoted by the symbol κE.
As was shown by Aumann (1976), this set of states of the world belongs to P̃ and,
hence, for each i, it can be represented (in the case κE = ∅) in the form (see (13)):
κE = ∪m P j(i)
m
. (14)
Let E be an event. For a state of the world ω, each agent i updates the common prior
p(E) on the basis of the observation the element P (i) (ω) of its information partition.
(For this agent, it means that the state of the world ω is contained in P (i) (ω).) This
update is given by the conditional probability
We remark that the conditional probability qi (ω) is defined to be the same for all
states of the world ω in a given element of partition. Thus, in fact,
qi (ω) ≡ qik ,
are common knowledge and prior probabilities are the same, then necessarily q1 =
q2 —simply because
where Cq1 q2 is the event (15): “the first agent by updating the prior probability of the
event E assigns the value q1 and the second agent the value q2 .”
168 A. Khrennikov
9.1 States
and according to the projection postulate the post-measurement state is obtained via
the state transformation:
Ê A (x)ρ Ê A (x)
ρ → ρx = . (18)
T r Ê A (x)ρ Ê A (x)
For reader’s convenience, we present these formulas for a pure initial state ψ ∈ H.
The Born’s rule has the form:
The map x → I A (x) given by (21) is the simplest (but very important) example
of quantum instrument (Davies, 1976; Davies & Lewis, 1970; Ozawa, 1984, 1997;
Yuen, 1987).
Following von Neumann (1955) and Birkhoff and von Neumann (1936) we represent
events, propositions, as orthogonal projectors in complex Hilbert space H.
For an orthogonal projector P, we set H P = P(H ), its image, and vice versa, for
subspace L of H, the corresponding orthogonal projector is denoted by the symbol
PL .
The set of orthogonal projectors is a lattice with the order structure: P ≤ Q
iff H P ⊂ HQ or equivalently, for any ∈ H, |P ≤ |Q. This lattice is
known as quantum logic. Thus in classical Boolean logic events are represented by
sets and in quantum logic events are represented by orthogonal projectors.
We recall that the lattice of projectors is endowed with operations “and” (∧),
conjunction, and “or” (∨), disjunction. For two projectors P1 , P2 , the projector
R = P1 ∧ P2 is defined as the projector onto the subspace H R = H P1 ∩ H P2 and
the projector S = P1 ∨ P2 is defined as the projector onto the subspace H R defined
as the minimal linear subspace containing the set-theoretic union H P1 ∪ H P2 of sub-
spaces H P1 , H P2 : this is the space of all linear combinations of vectors belonging
these subspaces. The operation of negation is defined as the orthogonal complement:
In the language of subspaces the operation “and”, conjunction, coincides with the
usual set-theoretic intersection, but the operations “or”, disjunction, and “not”, nega-
tion, are nontrivial deformations of the corresponding set-theoretic operations. It is
natural to expect that such deformations can induce deviations from classical Boolean
logic.
Consider the following simple example. Let H be two-dimensional
√ Hilbert space
with the orthonormal basis (e1 , e2 ) and let v = (e1 + e2 )/ 2. Then Pv ∧ Pe1 = 0 and
Pv ∧ Pe2 = 0, but Pv ∧ (Pe1 ∨ Pe2 ) = Pv . Hence, for quantum events, in general the
distributivity law is violated:
P ∧ (P1 ∨ P2 ) = (P ∧ P1 ) ∨ (P ∧ P2 ) (22)
As can be seen from our example, even mutual orthogonality of the events P1 and
P2 does not help to save the Boolean laws.
Thus quantum logic relaxes some constraints set by classical Boolean logic, in
particular, the distributivity between the operations of conjunction and disjunction.
170 A. Khrennikov
In our quantum-like model, the “states of the world” are given by pure states. Thus
the unit sphere S1 (H ) in a complex Hilbert space H represents (up to phase factors)
all possible states of the world.
3 Although in quantum physics the magnitudes of these numbers play an important role, in quantum
information theory the eigenvalues are merely formal labels encoding information which can be
extracted from a state with the aid of an observable. In the case of dichotomous answers, we can
simply use zero to encode “no” and one to encode “yes”.
Quantum-Like Cognition and Rationality: Biological and Artificial Intelligence Systems 171
p (Pk(i)
0
) = TrP Pk(i)0 = 1 and, for k = k0 , p (Pk(i) ) = TrP Pk(i) = 0.
Then
Q (i)
= min{P ∈ P̃
(i)
: P ≤ P}, (29)
4We state again that in the classical probability model the states of the world are encoded by points
of . Take one fixed state ω. Since information representation of each agent is a partition of , for
each i there exists an element of partition, say P j(i) , containing this ω. For this state of the world,
the ith agent should definitely get the answer a (i) (i)
j corresponding the element P j . Thus any agent
is able to resolve uncertainty at least for her/his information representation (although she/he is not
able to completely resolve uncertainty about the state of the world). In the quantum case, an agent
is not able to resolve uncertainty even at the level of her/his information representation. And the
prior probability is updated in this uncertainty context.
172 A. Khrennikov
The ith agent picture of the world can be represented by the density operator
(mixed quantum state)
ρ(i) = Pk(i) P Pk(i) = p (Pk(i) )ρ(i)
k ,
k k
Pk(i) P Pk(i)
where ρ(i)
k = . Since each ρ(i)
k is a pure state, the ith picture of the world
TrPk(i) P Pk(i)
is given by the mixture of pure states, corresponding to “cells" Pk(i) .
Definition 1 For the -state of the world and the event E, the ith agent knows E if
Q (i)
≤ E. (30)
It is evident that if, for the state of the world , the ith agent knows E, then
∈ HE . In general the latter does not imply that E is known (for the state ).5
However, if ∈ E = P j(i) , then this event is known for i. The same is valid for any
event of the form E = P j(i)1
∨ ... ∨ P j(i)
k
(= P j(i)
1
+ ... + P j(i)
k
); if ∈ HE , then such
E is known for i.
We remark that the straightforward analog of the classical definition, see (11),
would be based on condition P j(i) ≤ E for
P ≤ P j(i) , (31)
instead of more general condition (30). However, it would trivialize the class of
possible states of the world, because condition (31) is very restrictive.
We shall use the standard definition of common knowledge, see ACN, but based
on the quantum representation of knowing an event, see Definition 1. As in the
classical case, we have that “Where something is common knowledge, everybody
knows it.”
We recall that in the classical case, for each event E, there is considered the set
of all states of the world for which E is common knowledge. It is denoted by the
symbol κE.
5For example, the state space H is four dimensional with the orthonormal basis (e1 , e2 , e3 , e4 ), the
projectors P1 and P2 project H onto the subspaces with the bases (e1 , e2 ) and (e3 , e4 ), respectively.
Here (P1 , P2 ) is information representation of√an agent. Let E be the projector onto the subspace
with the basis (e1 , e4 ) and let = (e1 + e4 )/ 2. Then Q = I, the unit operator. Hence, E is not
known for this agent, although it belongs to H E .
Quantum-Like Cognition and Rationality: Biological and Artificial Intelligence Systems 173
P̃ = ∩i P̃ (i) .
We remark that (by definition) a projector P ∈ P̃ if and only if, for each i = 1, ..., N ,
it can be represented in the form
P= P j(i)
m
. (32)
m
Examples illustrating how this common knowledge structure works can be found
in Haven et al. (2017).
Now, before to formulate the quantum version of Aumann’s theorem, we recall that
while in the classical Aumann scheme the update of the prior probability distribution
on the basis of information representations of agents plays the crucial role. The
quantum analog of the Aumann scheme is based on the quantum procedure of the
state update as the result of measurement, in the simplest case this is the projection
postulate based update.
The quantum common prior assumption is formulated naturally: both agents assign
to possible states of the world the same quantum probability distribution given by
the density operator ρ, a priori state. The agents do not know exactly the real state
of the world which is always a pure state and in general a possible state of the world
appears for them as a mixed quantum state ρ.
If the amplitude of right-hand side of (33), the interference term between updates
of the probability of the event E by two different agents, is small, we can say that the
agents named i and s practically agree. The interference term can be considered as
a measure of “agreement on disagree” between the agents.
174 A. Khrennikov
11 Concluding Discussion
References
Aerts, D., Khrennikov, A., Melucci, M., & Toni, B. (Eds.). (2019). Quantum-like models for infor-
mation retrieval and decision-making. STEAM-H: Science, technology, engineering, agriculture,
mathematics & health. Springer.
Arndt, M., Juffmann, T., & Vedral, V. (2009). Quantum physics meets biology. HFSP Journal,3(6),
386–400. https://doi.org/10.2976/1.3244985
Asano, M., Basieva, I., Khrennikov, A., & Yamato, I. (2017a). A model of differentiation in quantum
bioinformatics. Progress in Biophysics and Molecular Biology,130, Part A, 88–98.
Asano, M., Basieva, I., Khrennikov, A., Ohya, M., Tanaka, Y., & Yamato, I. (2012). Quantum-
like model for the adaptive dynamics of the genetic regulation of E. coli’s metabolism of glu-
cose/lactose. Systems and Synthetic Biology, 6, 1–7.
Asano, M., Basieva, I., Khrennikov, A., Ohya, M., Tanaka, Y., & Yamato, I. (2015b). Quantum
information biology: From information interpretation of quantum mechanics to applications in
molecular biology and cognitive psychology. Foundations of Physics, 45(10), 1362–1378.
Asano, M., Khrennikov, A., Ohya, M., Tanaka, Y., & Yamato, I. (2014). Violation of contextual
generalization of the Leggett-Garg inequality for recognition of ambiguous figures. Physica
Scripta,T163, 014006.
Asano, M., Khrennikov, A., Ohya, M., Tanaka, Y., & Yamato, I. (2015a). Quantum adaptivity in
biology: From genetics to cognition. Springer.
Asano, M., Basieva, I., Khrennikov, A., Ohya, M., & Tanaka, Y. (2017b). A quantum-like model of
selection behavior. Journal of Mathematical Psychology, 78, 2–12.
Asano, M., Ohya, M., Tanaka, Y., Basieva, I., & Khrennikov, A. (2011). Quantum-like model of
brain’s functioning: Decision making from decoherence. Journal of Theoretical Biology, 281(1),
56–64.
Aumann, R. J. (1976). Agreeing on disagree. The Annals of Statistics,4, 1236–1239.
Bagarello, F. (2019). Quantum concepts in the social, ecological and biological sciences. Cambridge
University Press.
Bagarello, F., Basieva, I., Pothos, E. M., & Khrennikov, A. (2018). Quantum like modeling of deci-
sion making: Quantifying uncertainty with the aid of Heisenberg-Robertson inequality. Journal
of Mathematical Psychology, 84, 49–56.
Bagarello, F., & Oliveri, F. (2013). A phenomenological operator description of interactions between
populations with applications to migration. Mathematical Models and Methods in Applied Sci-
ences, 23(03), 471–492.
Basieva, I., & Khrennikov, A. (2015). On the possibility to combine the order effect with sequential
reproducibility for quantum measurements. Foundations of Physics, 45(10), 1379–1393.
Basieva, I., Pothos, E., Trueblood, J., Khrennikov, A., & Busemeyer, J. (2017). Quantum probability
updating from zero prior (by-passing Cromwell’s rule). Journal of Mathematical Psychology, 77,
58–69.
Bernroider, G. (2017). Neuroecology: Modeling neural systems and environments, from the quan-
tum to the classical level and the question of consciousness. Journal of Advanced Neuroscience
Research, 4, 1–9.
Bernroider, G., & Summhammer, J. (2012). Can quantum entanglement between ion transition
states effect action potential initiation? Cognitive Computation, 4, 29–37.
Birkhoff, J., & von Neumann, J. (1936). The logic of quantum mechanics. Annals of Mathematics,
37(4), 823–843.
Boyer-Kassem, T., Duchene, S., & Guerci, E. (2015). Quantum-like models cannot account for the
conjunction fallacy. Theory and Decision, 10, 1–32.
Busemeyer, J., & Bruza, P. (2012). Quantum models of cognition and decision. Cambridge Univer-
sity Press.
Busemeyer, J. R., Wang, Z., Khrennikov, A., & Basieva, I. (2014). Applying quantum principles to
psychology. Physica Scripta,T163, 014007.
176 A. Khrennikov
Busemeyer, J. R., Wang, Z., & Townsend, J. T. (2006). Quantum dynamics of human decision
making. Journal of Mathematical Psychology, 50, 220–241.
Davies, E. B. (1976). Quantum theory of open systems. Academic Press.
Davies, E. B., & Lewis, J. T. (1970). An operational approach to quantum probability. Communi-
cations in Mathematical Physics, 17, 239–260.
Dzhafarov, E. N., & Kujala, J. V. (2012). Selectivity in probabilistic causality: Where psychology
runs into quantum physics. Journal of Mathematical Psychology, 56, 54–63.
Dzhafarov, E. N., Zhang, R., & Kujala, J. V. (2015). Is there contextuality in behavioral and social
systems? Philosophical Transactions of the Royal Society A, 374, 20150099.
Friedell, M. (1969). On the structure of shared awareness. Behavioral Science, 14, 28–39.
Hameroff, S. (1994). Quantum coherence in microtubules. A neural basis for emergent conscious-
ness? Journal of Consciousness Studies, 1, 91–118.
Haven, E., & Khrennikov, A. (2013). Quantum social science. Cambridge University Press.
Haven, E., Khrennikov, A., & Robinson, T. R. (2017). Quantum Methods in Social Science: A First
Course. WSP.
Haven, E. (2005). Pilot-wave theory and financial option pricing. International Journal of Theoret-
ical Physics, 44(11), 1957–1962.
Kahneman, D., & Thaler., R. (2006). Utility maximization and experienced utility. Journal of
Economic Perspectives,20, 221–232.
Kahneman, D. (2003). Maps of bounded rationality: Psychology for behavioral economics. Amer-
ican Economic Review, 93(5), 1449–1475.
Kahneman, D., & Tversky, A. (1972). Subjective probability: A judgment of representativeness.
Cognitive Psychology, 3(3), 430–454.
Kahneman, D., & Tversky, A. (1979). Prospect theory: An analysis of decision under risk. Econo-
metrica, 47, 263–291.
Kahneman, D., & Tversky, A. (1984). Choices, values and frames. American Psychologist, 39(4),
341–350.
Khrennikov, A. (2004b). Information dynamics in cognitive, psychological, social, and anomalous
phenomena. Ser.: Fundamental Theories of Physics.
Khrennikov, A. (2010). Ubiquitous quantum structure: From psychology to finances. Springer.
Khrennikov, A. (2015a). Quantum-like model of unconscious-conscious dynamics. Frontiers in
Psychology,6, Art. N 997.
Khrennikov, A. (2016). Probability and randomness: Quantum versus classical. Imperial College
Press.
Khrennikov, A. (2020). Social laser. Jenny Stanford Publ.
Khrennikov, A., & Basieva, I. (2014a). Quantum model for psychological measurements: From the
projection postulate to interference of mental observables represented as positive operator valued
measures. NeuroQuantology,12, 324–336.
Khrennikov, A., Alodjants, A. Trofimova., A., & Tsarev, D. (2018). On interpretational questions
for quantum-like modeling of social lasing. Entropy,20(12), 921.
Khrennikov, A., Basieva, I., Dzhafarov, E.N., Busemeyer, J. R. (2014). Quantum models for psy-
chological measurements: An unsolved problem. PLOS ONE, 9, Art. e110909.
Khrennikov, A., Basieva, I., Pothos, E. M., & Yamato, I. (2018). Quantum probability in decision
making from quantum information representation of neuronal states. Scientific Reports, 8. Article
number: 16225.
Khrennikov, A. (1999). Classical and quantum mechanics on information spaces with applications
to cognitive, psychological, social and anomalous phenomena. Foundations of Physics, 29, 1065–
1098.
Khrennikov, A. (2003). Quantum-like formalism for cognitive measurements. Biosystems, 70, 211–
233.
Khrennikov, A. (2004a). On quantum-like probabilistic structure of mental information. Open Sys-
tems and Information Dynamics, 11(3), 267–275.
Khrennikov, A. (2006). Quantum-like brain: Interference of minds. BioSystems, 84, 225–241.
Quantum-Like Cognition and Rationality: Biological and Artificial Intelligence Systems 177
Wang, Z., & Busemeyer, J. R. (2013). A quantum question order model supported by empirical
tests of an a priori and precise prediction. Topics in Cognitive Science, 5, 689–710.
Wang, Z., Solloway, T., Shiffrin, R. M., & Busemeyer, J. R. (2014). Context effects produced by
question orders reveal quantum nature of human judgments. PNAS, 111, 9431–9436.
White, L. C., Pothos, E. M., & Busemeyer, J. R. (2014). Sometimes it does hurt to ask: The
constructive role of articulating impressions. Cognition, 133(1), 48–64.
Yuen, H. P. (1987). Characterization and realization of general quantum measurements. In M.
Namiki et al. (Ed.), Proceedings of the 2nd International Symposium of Foundations of Quantum
Mechanics (pp. 360–363).
Quantum Music, Quantum Arts
and Their Perception
Abstract The expression of human art, and supposedly sentient art in general, is
modulated by the available rendition, receiving and communication techniques. The
components or instruments of these techniques ultimately exhibit a physical, in par-
ticular, quantum layer, which in turn translates into physical and technological capac-
ities to comprehend and utilize what is possible in our universe. In this sense, we can
apply a sort of Church-Turing thesis to art, or at least to its rendition.
V. Putz (B)
Pädagogische Hochschule Wien, Grenzackerstraße 18, 1100 Vienna, Austria
e-mail: [email protected]
K. Svozil
Institute for Theoretical Physics, Vienna University of Technology, Wiedner Hauptstraße
8-10/136, 1040 Vienna, Austria
e-mail: [email protected]
It thus comes of no surprise that the evolution of quantum physics brought about
the quest for the quantum arts; and in particular, for quantum music Putz and Svozil
(2017) and quantum fine arts, especially quantum visual art. Indeed, every aspect of
human life can be re-evaluated and reframed in terms of the quantum paradigm.
In our (not so humble) opinion there are two immediate issues: One issue is the
tendency to re-introduce irrational “magic”, a sort of “quantum hocus pocus” Svozil
(2016) that brings it close to the esoteric, and fosters a kind of pseudo-religion
allegedly justified by the most advanced contemporary physics.
Another, converse, issue is the temptation to argue that, just like in quantum
computing (Mermin, 2007, Section 1.1), “any art is quantum” as the “underlying
physical layer” of any (classical) artistic expression is governed by the laws of quan-
tum mechanics. However, we emphasize upfront that we have to resist this temptation
towards a premature flattening and folding of the quantum phenomena into classi-
cal molds. Rather we consider quantum arts, and, in particular, quantum music, as
operations exploiting certain very special transformations of physical internal states,
subject to very carefully controlled conditions.
So what exactly are these very special transformations that characterize quantum
art? In this regard, we can proceed almost in parallel to the development of quantum
computation Fortnow (2003); Nielsen and Chuang (2010); Mermin (2007), and point
out some central assets or capacities:
(i) parallelization through coherent superposition (aka simultaneous linear com-
bination) of classically mutually exclusive tones or signals that are acoustic,
optic, touch, taste, or otherwise sensory;
(ii) entanglement not merely by classical correlation Peres (1978) but by relational
encoding Schrödinger (1935); Brukner and Zeilinger (1999); Zeilinger (1999);
Brukner et al. (2002) of multi-partite states such that any classical information is
“scrambled” into relational, joint multi-partite properties while at the same time
losing value definiteness about the single constituents of such multi-partite states
—this can be seen as a sort of zero-sum game, a tradeoff between individual
and collective properties;
(iii) complementarity associated with value (in)definiteness of certain tones or sig-
nals that is acoustic, optic, touch, taste, or otherwise: if one such observable is
definite, another is not, and vice versa;
(iv) contextuality is an “enhanced” form of complementarity and value indefinite-
ness that can be defined in various ways Dzhafarov et al. (2017); Abram-
sky (2018); Grangier (2002); Aufféves and Grangier (2018, 2020); Grang-
ier (2020); Budroni et al. (2021), in particular, emphasizing homomorphic,
structure-preserving nonembeddability into classical schemes Specker (1960);
Kochen and Specker (1967); Svozil (2021)
Those criteria or observables constitute significant signatures of quantum behav-
ior. The transformations and processing of classical-to-quantum states or quantum
states exhibiting these features can be considered musical, optical, or other instru-
ments or “transmitters” for the creation of quantum art. Similarly, assorted trans-
Quantum Music, Quantum Arts and Their Perception 181
Fig. 1 (Color online) Temporal succession of quantum tones |c , |d , . . ., |b in the C major
scale forming the octave basis B
Pure quantum musical states could be represented as unit vectors |ψ ∈ C7 which
are linear combinations of the basis B; that is,
One of the mind-boggling quantum field theoretic features of a “bundling” within sin-
gle modes is the possibility of the simultaneous “co-existence” of classically exclud-
ing musical states, such as a 50:50 quantum g in the
C major scale obtained by sending
1 1
|0g through the Hadamard gate H = √2 1
, resulting in √12 |0g − |1g , and
1 −1
depicted in Fig. 2 by a 50 white 50 black; that is, gray, tone (though without the
relative “−” phase).
This novel form of musical expression might contribute to novel musical experi-
ences; in particular, if any such coherent superposition can be perceived by the audi-
ence in full quantum uniformity. This would require the cognition of the recipient
to experience quantum coherent superpositions—a capacity that is highly specula-
tive. It has been mentioned earlier that any such capacity is related to Schrödinger’s
cat Schrödinger (1935) and quantum jellyfish Schrödinger (1995) metaphors, as
well as to nestings of the Wigner’s friend type von Neumann (1996); Everett (1957);
Wigner (1995); Everett (2012).
In the following, we shall assume that quantum music is “reduced” to the continuous
infinity of its classical forms. Then, if a classical auditorium listens to the quan-
tum musical state |ψ in Eq. 1, the individual classical listeners may perceive |ψ
Fig. 2 Representation of a 50:50 quantum tone |g = √1 |0g − |1g in gray (without indicating
2
phase factors)
184 V. Putz and K. Svozil
very differently; that is, they will hear only a single one of the different tones with
probabilities of |αc |2 , |αd |2 , . . ., and |αb |2 , respectively.
Indeed, suppose that classical recipients (aka “listeners”) modeled by classical
measurement devices acting as information-theoretic receivers are assumed. Then
any perception (aka “listening” or reception) of a quantum musical state that is
in a coherent superposition—with some coefficients 0 < |αi | < 1—because of the
supposedly irreducably stochastic Zeilinger (2005) quantum-to-classical transla-
tion Svozil (2004) represents an “irreducible” Peres (1980); Scully and Drühl (1982);
Greenberger and YaSin (1989); Scully et al. (1991); Zajonc et al. (1991); Kwiat et al.
(1992); Pfau et al. (1994); Chapman et al. (1995); Herzog et al. (1995) stochastic
measurement. This can never render a unique classical listening experience, as the
probability to hear the tone i is |αi |2 . Therefore, partitions of the audience will hear
different manifestations of the quantum musical composition made up of all varieties
of successions of tones. These experiences multiply and diverge as more tones are
perceived.
For the sake of a demonstration, let us try a two-note quantum composition. We
start with a pure quantum mechanical state in the two-dimensional subspace spanned
by |c and |g , specified by
4 3 1 4
|ψ1 = |c + |g = . (3)
5 5 5 3
This means for the quantum melody of both quantum tones |ψ1 and |ψ2 in
succession—for the score, see Fig. 3—that in repeated measurements, in 0.642 =
40.96% of all cases c − g is heard, in 0.362 = 12.96% of all cases g − c, in
0.64 · 0.36 = 23.04% of all cases c − c or g − g, respectively.
1
| ± = √ (|0e |1a ± |1e |0a ) ,
2
(5)
1
|± = √ (|0e |0a ± |1e |1a ) ,
2
A necessary and sufficient condition for entanglement among the quantized notes
e and a is that the coefficients α1 , α2 , α3 , α4 of their general composite state
|ga = α1 |0e |0a + α2 |0e |1a + α3 |1e |0a + α4 |1e |1a obey α1 α4 = α2 α3
(Mermin, 2007, Sec. 1.5). This is clearly satisfied by Eqs. (5). Figure 4 depicts the
entangled musical Bell states.
− +
Fig. 5 (Color online) Quantum musical entangled states for bundled octaves |ea and |ea in
Fig. 6 Temporal succession of complementary tones (a) for binary occupancy |φa = αa |0a +
βa |1a , with |αa |2 + |βa |2 = 1 with increasing |αa | (decreasing occupancy), (b) in the bundled
octave model, separated by bars
Quantum Music, Quantum Arts and Their Perception 187
Fig. 7 Temporal succession of tones |c , |d , . . ., |b in an octave in the C major scale with
dicreasing mean occupancy
Just as for the performing arts such as music one could contemplate the quantum
options and varieties for the visual arts. Suffice it to say that the notion of “color”
experience can be extended to the full quantum optical varieties that result from
the electromagnetic field quantization, as already mentioned earlier. Incidentally,
Schrödinger published a series of papers on classical color perception Schrödinger
(1924); Schrödinger and Niall (2017) until around 1925. Yet to our best knowledge he
never considered the particular quantum aspects of human color and light perception.
Human rod cells respond to individual photons Hecht et al. (1942); Westheimer
(2016). Moreover, recent reports suggest that humans might be capable of “being
aware” of the detection of a single-photon incident on the cornea with a probability
significantly above chance Tinsley et al. (2016). It thus may be suspected that this
area of perception presents the most promising pathway into truly quantum percep-
tion. Speculations how this issue may be transferred to the perception of sound are
compelling.
Let us state up front that quantum visual art, and, in particular, quantum paral-
lelism, is not about additive color mixing, but it is about the simultaneous existence
of different, classically mutually exclusive “colors”, or visual impressions in general.
Quantum visual arts use the same central assets or capacities (i)–(iv) mentioned ear-
lier in Sect. 1. It can be developed very much in parallel to quantum music but requires
the creation of an entirely new nomenclature. The perception of quantum visual art is
subject to the same assumptions about the cognitive capacities to comprehend these
artifacts fully quantum mechanically or classically. This will be shortly discussed in
the following section.
Suppose for a moment that humans are capable to sense, receive and perceive quan-
tum signals not only classically but in a fully quantum mechanical way. Thereby, they
would, for instance, be capable of simultaneously “holding” different classically dis-
tinct tones at once—not just by interference but by parallel co-existence. This would
result in a transgression of classical art forms, and in entirely novel forms of art.
The existence of such possibilities depends on the neurophysiology of the human,
or, more generally, sentient, perception apparatus. Presently the question as to
whether or not this is feasible is open; the answer to it is unknown.
In the case that merely classical perceptions are feasible, we would end up with
a sort of Church-Turing thesis for music. In particular, quantum music would not be
able to “go beyond” classical music for a single observer, as only classical renditions
could be perceived. Of course, as we mentioned earlier, quantum music might “sound
differently for different observers”. To this end, we might conceptualize a kind of
Quantum Music, Quantum Arts and Their Perception 189
universal musical instrument that is capable of rendering all possible classical notes.
Pianos and organs might be “for all practical purposes good” approximations to such
a universal device.
Quantum music and quantum arts, just like quantum computing Deutsch (1985),
or computations starting and ending in real numbers but using imaginary numbers
as intermediaries Musil (1906), might be a sort of bridge crossing elegantly a gap
between two classical domains of perception. And yet they could be so much more
if only the quantum could be “heard” or “sensed”.
8 Summary
We have contemplated the various extensions of music, and arts in general, to the
quantum domain. Thereby we have located particular capacities which are genuine
properties. These involve parallelization through coherent superposition (aka simul-
taneous linear combination), entanglement, complementarity and contextuality. We
have reviewed the nomenclature introduced previously Putz and Svozil (2017) and
considered particular instances of quantum music. Then we have briefly discussed
quantum visual arts.
The perception of quantum arts depends on the capacity of the audience to either
perceive quantum physical states as such, or reduce them to classical signals. In the
first case, this might give rise to entirely novel artistic experiences. We believe that
these are important issues that deserve further attention, also for sentient perception
in general and human neurophysiology, in particular.
Acknowledgements This research was funded in whole, or in part, by the Austrian Science Fund
(FWF), Project No. I 4579-N. For the purpose of open access, the author has applied a CC BY
public copyright licence to any Author Accepted Manuscript version arising from this submission.
References
This, H. (2005). Modelling dishes and exploring culinary ‘precisions’: the two issues of molecular
gastronomy. British Journal of Nutrition, 93, S139.
Putz, V., & Svozil, K. (2017). Quantum music. Soft Computing, 21, 1467. arXiv:1503.09045.
Svozil, K. (2016). Quantum hocus-pocus, Ethics in Science and Environmental Politics (ESEP).
16, 25. arXiv:1605.08569.
Mermin, D. N. (2007). Quantum Computer Science. Cambridge: Cambridge University Press.
Fortnow, L. (2003). One complexity theorist’s view of quantum computing. Theoretical Computer
Science, 292, 597.
Nielsen, M. A., & Chuang, I. L. (2010). Quantum Computation and Quantum Information. Cam-
bridge: Cambridge University Press. 10th Anniversary Edition.
Peres, A. (1978). Unperformed experiments have no results. American Journal of Physics, 46, 745.
Schrödinger, E. (1935). Die gegenwärtige Situation in der Quantenmechanik. Naturwissenschaften,
23, 807.
190 V. Putz and K. Svozil
Brukner, Č, & Zeilinger, A. (1999). Operationally invariant information in quantum measurements.
Physical Review Letters, 83, 3354. quant-ph/0005084.
Zeilinger, A. (1999). A foundational principle for quantum mechanics. Foundations of Physics, 29,
631.
Brukner, Č., Zukowski, M., & Zeilinger, A. (2002). The essence of entanglement.
arXiv:quant-ph/0106119, translated to Chinese by Qiang Zhang and Yond-de Zhang, New
Advances in Physics (Journal of the Chinese Physical Society).
Dzhafarov, E. N., Cervantes, V. H., & Kujala, J. V. (2017). Contextuality in canonical systems of
random variables. Philosophical Transactions of the Royal Society A: Mathematical, Physical
and Engineering Sciences, 375, 20160389. arXiv:1703.01252 .
Abramsky, S. (2018). Contextuality: At the borders of paradox. In E. Landry (Ed.) Categories for the
Working Philosopher (pp. 262–285). Oxford University Press, Oxford, UK. arXiv:2011.04899.
Grangier, P. (2002). Contextual objectivity: a realistic interpretation of quantum mechanics. Euro-
pean Journal of Physics, 23, 331. arXiv:quant-ph/0012122.
Aufféves, A., Grangier, P. (2018). Extracontextuality and extravalence in quantum mechanics. Philo-
sophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences,
376, 20170311. arXiv:1801.01398.
Auffèves, A., & Grangier, P. (2020). Deriving born’s rule from an inference to the best explanation.
Foundations of Physics, 50, 1781. arXiv:1910.13738.
Grangier, P. (2020). Completing the quantum formalism in a contextually objective framework.
arXiv:2003.03121.
Budroni, C., Cabello, A., Gühne, O., & Kleinmann, M. J. (2021). Quantum contextuality: Åke
Larsson. arXiv:2102.13036 [quant-ph].
Specker, E. (1960). Die Logik nicht gleichzeitig entscheidbarer Aussagen. Dialectica, 14, 239.
arXiv:1103.4537 .
Kochen, S., Specker, E. P. (1967). The problem of hidden variables in quantum mechanics. Journal
of Mathematics and Mechanics (now Indiana University Mathematics Journal), 17, 59.
Svozil, K. (2021). Varieties of contextuality emphasizing (non)embeddability. arXiv:2103.06110.
Shannon, C. E. (1949). Bell System Technical Journal 27, 379 (1948), reprinted in C. E. Shannon and
W. Weaver: The Mathematical Theory of Communication, University of Illinois Press, Urbana,
Illinois.
Reck, M., Zeilinger, A., Bernstein, H. J., & Bertani, P. (1994). Experimental realization of any
discrete unitary operator. Physical Review Letters, 73, 58.
Zukowski, M., Zeilinger, A., & Horne, M. A. (1997). Realizable higher-dimensional two-particle
entanglements via multiport beam splitters. Physical Review A, 55, 2564.
Glauber, R. J. (1986). Amplifiers, attenuators, and schrödinger’s cat. Annals of the New York
Academy of Sciences, 480, 336.
Schrödinger, E. (1995). The Interpretation of Quantum Mechanics. Dublin Seminars (1949-1955)
and Other Unpublished Essays. Woodbridge, Connecticut: Ox Bow Press.
von Neumann, J. (1932, 1996). Mathematische Grundlagen der Quantenmechanik, (2nd ed.). Berlin,
Heidelberg: Springer, English translation in [54].
Everett, H., III. (1957). Relative State formulation of quantum mechanics. Reviews of Modern
Physics, 29, 454.
Wigner, E. P. (1961, 1962, 1995) Remarks on the mind-body question. In I. J. Good (Ed.), The
Scientist Speculates (pp. 284–302). London, New York, and Berlin: Heinemann, Basic Books,
and Springer-Verlag.
Everett, H., III. (2012). The Everett Interpretation of Quantum Mechanics: Collected Works 1955–
1980 with Commentary, edited by J. A. Barrett & P. Byrne. Princeton, NJ: Princeton University
Press.
Schwinger, J. (1960). Unitary operators bases. Proceedings of the National Academy of Sciences
(PNAS), 46, 570.
Zeilinger, A. (2005). The message of the quantum. Nature, 438, 743.
Quantum Music, Quantum Arts and Their Perception 191
Svozil, K. (2004). Quantum information via state partitions and the context translation principle.
Journal of Modern Optics, 51, 811. arXiv:quant-ph/0308110.
Peres, A. (1980). Can we undo quantum measurements? Physical Review D, 22, 879.
Scully, M. O., & Drühl, K. (1982). Quantum eraser: A proposed photon correlation experiment
concerning observation and “delayed choice” in quantum mechanics. Physical Review A, 25,
2208.
Greenberger, D. M., & YaSin, A. (1989). “Haunted” measurements in quantum theory. Foundation
of Physics, 19, 679.
Scully, M. O., Englert, B.-G., & Walther, H. (1991). Quantum optical tests of complementarity.
Nature, 351, 111.
Zajonc, A. G., Wang, L. J., Zou, X. Y., & Mandel, L. (1991). Quantum eraser. Nature, 353, 507.
Kwiat, P. G., Steinberg, A. M., & Chiao, R. Y. (1992). Observation of a “quantum eraser”: a revival
of coherence in a two-photon interference experiment. Physical Review A, 45, 7729.
Pfau, T., Spälter, S., Kurtsiefer, C., Ekstrom, C. R., & Mlynek, J. (1994). Loss of spatial coherence
by a single spontaneous emission. Physical Review Letters, 73, 1223.
Chapman, M. S., Hammond, T. D., Lenef, A., Schmiedmayer, J., Rubenstein, R. A., Smith, E., &
Pritchard, D. E. (1995). Photon scattering from atoms in an atom interferometer: Coherence lost
and regained. Physical Review Letters, 75, 3783.
Herzog, T. J., Kwiat, P. G., Weinfurter, H., & Zeilinger, A. (1995). Complementarity and the quantum
eraser. Physical Review Letters, 75, 3034.
Pauli, W. (1933). Die allgemeinen Prinzipien der Wellenmechanik. In H. Geiger & K. Scheel (Ed.),
Handbuch der Physik (Vol. 24, p. 126). Berlin: Springer.
Glauber, R. J. (1969). Quantum theory of coherence. In S. M. Kay & A. Maitland (Ed.), Quan-
tum Optics: Proceedings of the Scottish Universities’ Summer School in Physics 1969. London:
Academic Press.
Glauber, R. J. (2007). Amplifiers, attenuators and Schrödingers cat. In Quantum Theory of Optical
Coherence (pp. 537–576). Wiley-VCH Verlag GmbH & Co. KGaA.
Weinberg, S. (1977). The search for unity: Notes for a history of quantum field theory. Daedalus,
106, 17.
Schrödinger, E. (1924). Über den Ursprung der Empfindlichkeitskurven des Auges. Die Naturwis-
senschaften, 12, 925.
Schrödinger, E., Niall, K. K. (2017). Erwin Schrödinger’s Color Theory. Springer International
Publishing.
Hecht, S., Shlaer, S., & Pirenne, M. H. (1942). Energy, quanta, and vision. Journal of General
Physiology, 25, 819.
Westheimer, G. (2016). History of physiological optics in the twentieth century, in Handbook of
Visual Optics: Fundamentals and Eye Optics, Volume One (Chap. 1, pp. 1–10). CRC Press, Taylor
& Francis Group.
Tinsley, J. N., Molodtsov, M. I., Prevedel, R., Wartmann, D., Espigulé-Pons, J., Lauwers, M., &
Vaziri, A. (2016). Direct detection of a single photon by humans. Nature Communications, 7.
https://doi.org/10.1038/ncomms12172.
Deutsch, D. (1985). Quantum theory, the Church-Turing principle and the universal quantum com-
puter. Proceedings of the Royal Society of London. Series A, Mathematical and Physical Sciences
(1934–1990), 400, 97.
Musil, R. (1906). Die Verwirrungen des Zöglings Törleß. Wien und Leipzig: Wiener Verlag, project
Gutenberg ebook # 3471.
von Neumann, J. (1955). Mathematical Foundations of Quantum Mechanics. Princeton, NJ: Prince-
ton University Press, German original in [27].
Quanta in Sound, the Sound of Quanta:
A Voice-Informed Quantum Theoretical
Perspective on Sound
Abstract Humans have a privileged, embodied way to explore the world of sounds,
through vocal imitation. The Quantum Vocal Theory of Sounds (QVTS) starts from
the assumption that any sound can be expressed and described as the evolution of a
superposition of vocal states, i.e., phonation, turbulence, and supraglottal myoelastic
vibrations. The postulates of quantum mechanics, with the notions of observable,
measurement, and time evolution of state, provide a model that can be used for
sound processing, in both directions of analysis and synthesis. QVTS can give a
quantum-theoretic explanation to some auditory streaming phenomena, eventually
leading to practical solutions of relevant sound-processing problems, or it can be
creatively exploited to manipulate superpositions of sonic elements. Perhaps more
importantly, QVTS may be a fertile ground to host a dialogue between physicists,
computer scientists, musicians, and sound designers, possibly giving us unheard
manifestations of human creativity.
voi ce
1 Sound ↔ Quanta
Sometimes, when kids imitate sounds around them, they are blamed for producing
weird noises. However, they are unknowingly using their own voice as a probe to
investigate the world of sounds, and thus they are probably performing some exper-
iments. Some of the these kids will become sound designers, other ones composers,
other sound engineers and physicists; some other ones, will blame future kids, and
the cycle repeats.
M. Mannone
European Centre for Living Technology (ECLT), Ca’ Foscari University of Venice, Venice, Italy
M. Mannone (B) · D. Rocchesso
Department of Mathematics and Computer Sciences, University of Palermo, Palermo, Italy
e-mail: [email protected]; [email protected]
D. Rocchesso
e-mail: [email protected]
rable, because we can distinguish the components along different axes. In the case
of QVTS, the space is related with the vocal primitives, and it is separable as well,
as we will see in the next section.
A quantum state, that is a unit-length vector in the state space, can be seen as
a superposition of values with some probabilities. An eigeinstate is a characteristic
state of some operator. After the measurement process, the probability wave collapses
to a certain value: it is the eigenvalue of a certain operator, and the system is in an
eigenstate of that operator.
Probability is a key concept in quantum mechanics. According to the principle of
uncertainty, we cannot know, let’s say, the position and the momentum of a particle
with the same precision. The more is the information we have on position, the less
we know about momentum, and vice versa.
Let us say more on the idea of quantum measurement. Consider a Cartesian frame-
work with axes x, y, and z—a tridimensional space with three mutually orthogonal
axes. We can perform measurements along each of the axes. Quantum measure
implies a change in the measured entity. If the measurement along the direction x
can give a positive result, in all subsequent measurements along the same direction
we will have a positive value. A measure along x would zero out the y and z compo-
nents, while leaving only the x component with value 1. If, before measurement, there
is a given probability to get a specific outcome, after the effective measurement of
that outcome, the probability to get the same value in each subsequent measurement
along the same direction is 100%. In fact, in quantum mechanics, the measurement of
a state implies the destruction of part of the initial information, and thus the process is
called destructive measure. A quantum state is a superposition of eigenstates, which
are reduced to a single state after the measurement. Such state collapse happens in the
context including both the system and the measuring entity, through the interaction
of the two (Rovelli 2021). Intuitively, it’s like observing and taking a picture of a
person, and blocking him or her as the represented image along that specific shooting
direction. (Be careful the next time you’ll take pictures).
Dennis Gabor first exploited the paradigm of quantum theory to investigate
sound (Gabor 1947), instead of doing the usual vice versa, with sound and strings used
as metaphors to understand quantum waves. Gabor proposed the concept of quantum
of sound, as a unit-area cell in the time-frequency plane, which could be called phon,
from the Greek fwn». On the other hand, we start from a vocal description of sound,
to define the phon as the set of vocal primitive operators.
In a recent article (Rocchesso and Mannone 2020), we have proposed the basics for
a Quantum Vocal Theory of Sound (QVTS). Here, we summarize its main ideas, and
then we propose some hints for future developments.
First of all, let us define the phon formalism, where the word phon indicates the
quantum of sound, expressed in the state space of vocal primitives. With the phon
198 M. Mannone and D. Rocchesso
formalism, we can define vocal states, and extend the quantum bit (qubit) language
to the human voice. Some quantum-mechanical concepts, such as state preparation
and measurement, can be extended to the domain of voice as well.
Consider a space with three independent directions: x, y, and z. In the QVTS, the
three axes of this “phonetic space” have a vocal meaning:
• z is the phonation, giving waveforms of different pitches;
• x is the turbulence, giving noises of different brightnesses;
• y is the myoelasticity, giving pulsations at different tempos (thought of as slow
pulse trains).
Such three-dimensional space is sketched in Fig. 1 where, at the intersections between
each of the axes and the unit (called Bloch) sphere, we can find two mutually orthog-
onal vectors, each depicted as a tiny sketchy spectrogram.
Given a sound recording of human voice, if we measure phonation using a specific
computational tool (such as SMS-tools Bonada et al. 2011), it is possible to separate
such component from the rest, and all subsequent measurements of phonation will be
giving the same result. If we measure a primitive component first, and then another
one, the result is generally dependent on the order of the two operations: A fact that
is known as non-commutativity. Figure 5 shows a couple of example spectrograms
illustrating the difference.
A vocal state can be described as a superposition of phonation, turbulence, and
myoelasticity with certain probabilities. We can thus define a phon operator σ as a
3-vector operator, providing information on the x, y, and z components through its
specific directions in the 3d phonetic space. Each component of σ is represented by
a linear operator, so we have σx , σ y , and σz .
Quanta in Sound, the Sound of Quanta 199
A measurement along the z axis is performed through the operator σz . The eigenvec-
tors (or eigenstates) of σz are |u and |d, corresponding to pitch-up phonation and
pitch-down phonation, with eigenvalues λu = +1 and λd = −1, respectively:
The eigenstates |u and |d are orthogonal, i.e., u|d = 0, and they can be represented
as column vectors
1 0
|u = , |d = . (2)
0 1
Applying a measurement along the z direction to a generic phon state |ψ corre-
sponds to pre-multiply it by one of the measurement operators (or projectors)
10
Mu = |u u| =
00
200 M. Mannone and D. Rocchesso
or
00
Md = |d d| = ,
01
and to normalize the resulting vector to have length one. Such operators satisfy the
completeness relation Mu + Md = I , summing up to the unit operator.
A generic phon state |ψ can be expressed as
where the coefficients are complex numbers, αu = u|ψ, and αd = d|ψ. Being
the system in state |ψ, the probability to measure pitch-up is
pu = ψ| Mu† Mu |ψ = ψ|u u|u u|ψ = ψ|u u|ψ = αu∗ αu (5)
where the sum within brackets is called the observable of the measurement.
In quantum computing terminology, the vectors2 (2) give the computational basis
of a qubit vector space. The operator (3) corresponds to a Z gate, which acts as a
phase flip on the second state of the computational basis.
The eigenstates of the operator σx are |r and |l, corresponding to turbulent primitive
sounds having different spectral distributions, one with the rightmost (or highest-
frequency) centroid and the other with the lowest-frequency centroid. Their respec-
tive eigenvalues are λr = +1 and λl = −1, such that
σx |r = |r , σx |l = − |l .
2 In quantum computing, the vectors of the computational basis are normally called |0 and |1.
Quanta in Sound, the Sound of Quanta 201
If the phon is prepared |r (turbulent) and then the measurement apparatus is set
to measure σz , there will be equal probabilities of getting pitch-up or pitch-down
phonation as an outcome. This measurement property is satisfied if |r is defined as
1 1
|r = √ |u + √ |d . (7)
2 2
A similar definition is given for |l, such that the two eigenstates of turbulence are
orthogonal (r |l = 0):
1 1
|l = √ |u − √ |d . (8)
2 2
√1 √1
|r = 2 , |l = 2 . (10)
√1
2
− √12
Applying a measurement along the x direction to a generic phon state |ψ corre-
sponds to pre-multiply it by one of the measurement operators
1 11
Mr = |r r | =
2 11
or
1 1 −1
Ml = |l l| = ,
2 −1 1
and to normalize the resulting vector to have length one. Such operators satisfy the
completeness relation Mr + Ml = I .
In quantum computing, the operator (9) corresponds to a X gate, which is the
equivalent of the NOT gate in classical logic circuits, as it flips the states of the
computational basis. The vectors (7) and (8) form the Hadamard basis, often denoted
with the symbols {|+ , |−}.
Preparation in one of the states of the Hadamard basis {|r , |l}, followed by
measurement along the z axis, results in an operation that is equivalent to coin
flipping, +1 or −1 being obtained with equal probability.
202 M. Mannone and D. Rocchesso
The eigenstates of the operator σ y are | f and |s, corresponding to slow myoelastic
pulsations, one faster and one slower,3 with eigenvalues λu = +1 and λd = −1, such
that
σ y | f = | f σ y |s = − |s .
If the phon is prepared | f (pulsating) and then the measurement apparatus is set to
measure σz , there will be equal probabilities for |u or |d phonation as an outcome.
This measurement property is satisfied if
1 i
| f = √ |u + √ |d , (11)
2 2
1 i
|s = √ |u − √ |d . (12)
2 2
√1 √1
|f = 2 , |s = 2 .
√i
2
− √i 2
Applying a measurement along the y direction to a generic phon state |ψ corre-
sponds to pre-multiply it by one of the measurement operators
1 1 −i
Mf = | f f | =
2 i 1
or
1 1 i
Ms = |s s| = ,
2 −i 1
and to normalize the resulting vector to have length one. Such operators satisfy the
completeness relation M f + Ms = I .
3In describing the spin eigenstates, the symbols |i and |o are often used, to denote the in–out
direction.
Quanta in Sound, the Sound of Quanta 203
The matrices (3), (9), and (13) are called the Pauli matrices. In quantum computing,
these are all useful one-qubit gates.
The Harmonic plus Noise model (Bonada et al. 2011) is well suited to describe
measurement and preparation in the phonation-turbulance planar section of the 3d
phonetic space. An arbitrary direction in such plane is described by the operator
cos θ sin θ
σn = , (15)
sin θ − cos θ
and the two are orthogonal. Suppose we prepare the phon to pitch-up |u. If we rotate
the measurement system along n, the probability to measure +1 is
Symmetrically, if we prepare the phon in state |λ1 and we measure along the z
axis, we get a pitch-up with probability cos2 θ/2 and a pitch-down with probability
sin2 θ/2.
More generally, the Sines plus Noise plus Transients model (Verma et al. 1997)
may be suitable to describe measurement and preparation in the whole 3d phonetic
space, where supraglottal myoelastic vibrations are made to correspond to transient
pulse trains. For example, consider the vocal fragment4 whose spectrogram is repre-
sented in Fig. 2. An extractor of pitch salience and an extractor of onsets5 have been
applied to highlight respectively the phonation (horizontal dotted line) and myoelas-
tic (vertical dotted lines) components in the spectrogram. In the z − y plane, there
would be a measurement orientation and a measurement operator that admit such
sound as an eigenstate.
4 It is one of the example vocal sounds considered in Rocchesso et al. (2016), and taken from New-
man (2004).
5 The feature extractors are found in the Essentia library (Bogdanov et al. 2013).
Quanta in Sound, the Sound of Quanta 205
the observer, who has some degree of epistemic uncertainty. Thus, the concept of
density matrix generalizes the concept of state superposition. The density operator
is defined as
ρ= p j |ψ j ψ j | , (21)
j
where p j indicates the probability for the j-state. The density operator for a pure
state is ρ = |ψψ|, and the trace of the associated density matrix is tr[ρ 2 ] = 1. For a
mixed state, tr[ρ 2 ] < 1. It can be shown that the density matrix (21) can be expressed
as a composition of Pauli matrices as in (14), with the addition of the identity matrix.
From such representation, pure states can be shown to lay on the surface of the Bloch
sphere, while mixed states stay inside the sphere, with the completely chaotic state
being found at the origin (Cariolaro 2015). A pure state can contain a superposition,
but such a composition is defined with certainty. A mixed state is a probabilistic
mixing. The mixed state is inseparable. The generalization introduced by mixed
states can represent the audio concept of mixing, thus coming useful in composition
of auditory scenes.
In the wonderland of quantum mechanics, it can happen that, the better we know
something, the lesser we know something else. In QVTS, the more precise our
knowledge of phonation, the less precise our measurement of turbulence. In quantum
mechanics, if we measure two observables L and M simultaneously in a single
experiment, the system is left in a simultaneous eigenvector of the observables only
if L and M commute, i.e., if their commutator [L, M] = LM − ML vanishes. When
the commutator is different from zero, we say that the two operators do not commute.
This
happens
with measurement operators along different axes. It is the case of
σx , σ y = 2iσz . As a consequence for QVTS, phonation and turbulence cannot be
simultaneously measured with certainty.
The uncertainty principle is one of the key ideas of quantum mechanics. It is
based on Cauchy–Schwarz inequality in complex vector spaces. According to the
uncertainty principle, the product of the two uncertainties is at least as large as half
the magnitude of the commutator:
1
LM ≥ |ψ| [L, M] |ψ| (22)
2
Kids learn that multiplying a times b, with a, b natural numbers, is the same as
multiplying b times a—and, since early age, they think that commutativity is always
verified. Reading a book and then going for a walk might be the same stuff as going
for a walk and then reading a book (maybe). Quantum mechanics does not work that
way, and the same for QVTS. If we record a singer, we take away vowels, and then we
take again away vowels, the result is the same—the recording is in an autostate of no-
vowels. If we take away vowels, and then we take away the noise, the result is different
from what we could hear if we do the opposite, that is, taking away the noise and
then the vowels. More precisely, the measurement operators oriented along different
axes do not commute. For example, let A be an audio segment. The measurement (by
extraction) of turbulence by the measurement operator turbulence-right Mr = |r r |
leads to Mr (A) = A . A successive measurement of phonation by the measurement
operator pitch-up Mu = |uu| gives Mu (A ) = A , thus Mu (A ) = Mu Mr (A) =
A . If we perform the measurements in the opposite order, with phonation first and
turbulence later, we obtain Mr Mu (A) = Mr (A∗ ) = A∗∗ . We expect that [Mr , Mu ] =
0, and thus, that A∗∗ = A . The diagram in Fig. 3 shows non-commutativity in the
style of category theory.
In bra-ket notation, this fact can be expressed as
Given that r |u is a scalar and u|r is its complex conjugate, and that |u r | is
generally non-Hermitian, we get
Fig. 3 A non-commutative Mu Mr
diagram representing the
non-commutativity of
measurements of phonation
A
Mr Mu A
(Mu ) and turbulence (Mr ) on A =
audio A Mu Mr A∗∗
A∗
Mr M u
Quanta in Sound, the Sound of Quanta 207
audio
sines+noise sines+noise
Harmonic Stochastic
sines+noise sines+noise
Stochastic Harmonic
Fig. 4 On the left, an audio segment is analyzed via the sines+noise model. Then, the noise part
is submitted to a new analysis. In this way, a measurement of phonation follows a measurement of
turbulence. On the right, the measurement of turbulence follows a measurement of phonation. This
can be described via projectors through equation (23), and diagrammatically in Fig. 3
1 11 10 1 10 11
[Mr , Mu ] = − =
2 11 00 2 00 11
(25)
1 10 1 11 1 0 −1 i
= − = = σ y = 0.
2 10 2 00 2 1 0 2
The variation of quantum states in time can be obtained through the application of
time evolution operators on them. Similarly, suitable time operators can make the
density matrix vary in time as well. Given a density operator ρ(t0 ) at time t0 , its time
variation is obtained applying a unitary operator U (t0 , t):
208 M. Mannone and D. Rocchesso
This is the most general definition: There are no assumptions on states (mixed or
pure), and the only assumptions on the operator U are that it is unitary, i.e., U† U = I,
with I the identity matrix, and that it depends only on t and t0 .
But actually there is more. The unitary operator U, evaluated at a tiny time incre-
ment
, is related to the Hamiltonian H, describing the energy of the system:
U(
) = I − i
H. (27)
For a closed and isolated system, H is time-independent, and the unitary operator
becomes U(t) = eiH(t−t0 ) . However, nature is more complex, things are not isolated,
and usually H is time-dependent, and the time evolution is given by an integral.
To complicate things even more, with a non-commutative Hamiltonian, an explicit
solution cannot be found. The problem can be circumvented by considering local
time segments where the Hamiltonian is locally commutative.
Quanta in Sound, the Sound of Quanta 209
Where do the restoring forces come from, in the sound domain? Broadly speaking,
they come from the local temporal sound production context. Similarly to the con-
cept of coarticulation in phonetics, the locally-defined Hamiltonian is determined by
neighboring sounds, extending their effects in the short-term future or past (Daniloff
and Hammarberg 1973). In practice, we can rely on an audio analysis system, such as
the Short-Time Fourier Transform (STFT), to extract and manipulate slowly-varying
features such as pitch salience or spectral energy to determine the components of the
Hamiltonian (28). Considered a slice of time and an audio signal, the initial phon
state can be made to evolve subject to a time-dependent yet commutative Hamiltonian
expressed as
H (t) = e−kt S, (31)
which in general will be a superposition (4) in the phon space. A measurement in the
computational (phonation) basis will make it collapse to |u or |d according to the
probabilities αu∗ αu or αd∗ αd , respectively. If there are two competing and concurrent
pitch lines, the Hamiltonian evolution followed by measurement may thus make
a pitch following process stay on one line or jump to the other one. In this way,
auditory streaming processes (Bregman 1994) and figure-ground segregation can be
mimicked.
In this section, we present some examples that show how the quantum formalism,
as assimilated by the QVTS, can be used together with classical signal processing,
for creative yet controllable analysis/synthesis tasks. Given the time-frequency
representation of an audio signal, as provided by the STFT, the elements of the S
matrix of the Hamiltonian (31) can be computed from decimated audio features.
For example, pitch salience can be extracted from time-frequency analysis (Salamon
and Gomez 2012), and used as the n z component. The exponential factor can be set
to g(m) = e−km , where m is the frame number within a segment of M frames. The
time evolution (32) can be computed by approximating the integral with a cumulative
sum. Starting from an initial state (e.g., |u), the phon goes through repeated cycles of
Hamiltonian evolution, measurement, and collapse. The decision to measure phona-
tion or turbulence can be based on the degree of pitchiness that the evolution within
a certain audio segment has reached. Since the observable σz has eigenvalues ±1 for
eigenvectors |u and |d, a measure of the degree of pitchiness can be given by the
distance of σz |ψ from |ψ . The degrees of noisiness and transientness can be
similarly determined using the observables σx and σ y , respectively.
When doing actual audio signal processing based on the QVTS, several degrees
of freedom are available to experiment with: The decimation factor or number M of
frames in a segment; The exponential damping factor k; The thresholds for switching
to a certain measurement direction in phon space; The decision to collapse or not—
this is a freedom we have if we are using a classical computer!
Consider the beginning of the Fugue from the Toccata and Fugue in D Minor, BWV
565, by Johann Sebastian Bach (Fig. 6). In this fragment, there is only one voice,
played by the left hand. However, the design of this sequence actually creates the
illusion of two voices: an upper line with an ostinato A, and a lower line with notes
G, F, E, D, C, D, E, F, and so on.
Quanta in Sound, the Sound of Quanta 211
Fig. 6 Toccata and Fugue in D minor BWV 565 by J. S. Bach: beginning of the Fugue
Fig. 7 The synthetic recording of the excerpt in Fig. 6 through Hamiltonian evolution. The upper
line with the repeated A is evident in the first part of the graph, while the second part contains a
fragment of the melody of the lower line
The score fragment of Fig. 6 was automatically rendered with piano samples at 100
bpm and analyzed via the STFT,6 with pitch salience and noise energy extracted via
the SMS-tools (Bonada et al. 2011). Setting the parameters frame decimation M =
10, exponential damping k = 0.1, threshold of pitchiness 0.9, collapse decimation
5, we obtain a phon evolution from pitch-up phonation represented by the green dots
of Fig. 7. The red and yellow lines represent the two most salient pitches, as features
extracted from the STFT. In the first part of the evolution, the green dots mainly
correspond to the ostinato of note A, while, in the second part, they mainly follow
the ascending scale fragment (A B C D...). Even without any noise added, transient
and noise components are inherent in piano samples (e.g., key noise) and, therefore,
the phon is subject to non-negligible forces and actually moves from the initial
|u state. Different runs will give different evolutions, as the collapse is governed
by probability amplitudes, but in general we observe that the Hamiltonian listener
follows the upper or the lower melodic line for some fractions of time. Interestingly,
the melodic line following is stable, or even more stable, if we add a strong white
noise to the signal, with an amplitude that is about one tenth of the signal. An example
evolution is depicted in Fig. 8, where the effect of the added noise is only apparent
after second 5, when effectively there is no signal. Figure 9 shows the phon evolution
when the fugue is drowned into noise. In this case the melodic contour following is
more easily disrupted. The zero-pitch green dots represent points where measurement
and collapse have been oriented to the x direction, for the effect of thresholding in
6 Sample rate 44100Hz, window size 2048, transform size 4096, hop size 1024.
212 M. Mannone and D. Rocchesso
Fig. 8 The synthetic recording of the excerpt in Fig. 6 through Hamiltonian evolution, with the
added noise (amplitude 0.1). The upper line with the repeated A is evident in the first part of the
graph, while the second part contains a fragment of the melody of the lower line
Fig. 9 The same fragment of Fig. 6 completely drowned into noise (amplitude 1.0)
pitchiness. In a resynthesis, these dots correspond to noise bursts, while the other
dots come from z-oriented measurements and produce pitched notes.
In repeated runs of Hamiltonian fugue following, we can see multiple melodic
lines emerging as the time evolution of some initial sound/vocal state. The collapse
of the phon to a state or another can be interpreted as the attention shifts from figure
to ground, or vice versa (Bregman 1994; Bigand et al. 2000).
The proposed example can be the starting point for a wider investigation in the
field of auditory bistability. Bistability is an intriguing topic in cognition. As a refer-
ence for quantum effects in cognition, especially regarding superposition and non-
classical probability, we may refer to Yearsley and Busemeyer (2016) for a theoretical
quantum-based approach to explain cognitive acts. The idea of bistability is also faced
in an article on mathematics and image/music (Mannone et al. 2020). It exploits the
Dirac formalism used in quantum mechanics to represent images as superpositions
of essential visual forms. There is a minimum number of forms which allows the
Quanta in Sound, the Sound of Quanta 213
recognizability of the form. With a little abuse of terminology, we can consider this
limiting, minimum value of simple forms as the limit of a Gestalt neighborhood, as
the discrete version of a topological neighborhood, having in the center the initial,
complete, not-approximated visual form. When an image is bistable, we can imagine
to have two neighborhoods as the two faces of a thin cylinder. We can see one face
or the other one; but we cannot see the two faces together. This is the core idea of
bistability. While classic examples of bistability are visual, also auditory illusions
can be constructed, e.g., with different auditory streamings (Byrne et al. 2019). These
cases might be analyzed with the help of QVTS, as we did with the beginning of the
Fugue from the Toccata and Fugue BWV 565.
Fig. 10 Tracing the two most salient pitches and noise energy for two crossing glides interrupted
by noise
Fig. 11 Tracking the phon state under Hamiltonian evolution from pitch-up
Quanta in Sound, the Sound of Quanta 215
The quantum concept of mixing, briefly described in Sect. 2.3, can be related to the
familiar audio concept of mixing. At the start of a Hamiltonian evolution, the initial
state may be mixed, i.e., known only as a probabilistic mixture. For example, at time
zero we may start from a mixture having 13 probability of |u and 23 probability of
|d. The density matrix would evolve in time according to equation (26).
When a pitch measurement is taken, the outcome is up or down according to
and the density matrix that results from collapsing upon measurement is given by
Mi ρ Mi
ρ (i) = . (34)
T r [ρ Mi ]
The density matrix can be made audible in various ways, thus sonifying the Hamil-
tonian evolution. For example, the completely chaotic mixed state, corresponding to
the half-identity matrix ρ = 21 I, can be made to sound as noise, and the pure states
can be made to represent distinct components of an audio mix.
Given the same audio scene of the two crossing glides interrupted by noise (Fig. 10),
we may follow the Hamiltonian evolution from an initial mixed state. We can choose
to make the pure states to sound like the upper or the lower of the most salient
pitches, and the completely mixed state to sound like noise. These three components
can be mixed for states with intermediate degrees of purity. If pu and pd are the
respective probabilities of |u and |d as encoded in the mixed state, the resulting
mixed sound can be composed by a noise having amplitude min ( pu , pd ), by the
upper pitch weighted by pu − min ( pu , pd ), and by the lower pitch weighted by
pd − min ( pu , pd ). Figure 12 shows an example of evolution from the mixed state
having probabilities 13 and 23 , with periodic measurements and collapses ruled by
equation (34).
216 M. Mannone and D. Rocchesso
Fig. 12 Amplitudes of components |u, |d, and noise resulting from a Hamiltonian evolution from
a mixed state
Fig. 13 Sound synthesis obtained from the density matrix evolution from a mixed state, using the
component amplitudes depicted in Fig. 12
The analyzed audio scene and the model parameters, including the computed
Hamiltonian, are the same as used in the evolution of pure states described in
Sect. 3.1.2. The amplitudes of the three components can be used as automated knobs
to control two oscillators and a noise generator, producing the sound of spectro-
gram Fig. 13, characterized by a prevailing upward tone with a downward bifurcation
and a noisy tail.
Quanta in Sound, the Sound of Quanta 217
As another example of mixed state evolution, we consider again the vocal sound
whose spectrogram is depicted in Fig. 2. It was chosen as an example of actual
superposition of phonation and slow myoelastic vibration. Despite the presence of
only one definite pitch, we can prepare the phon in an initial mixed state, having
1
3
probability of |u and 23 probability of |d, and compute a Hamiltonian evolution
based on potentials deduced from the time-frequency analysis, namely pitch salience,
noise component, and detected onsets. As in the example of Sect. 3.2.1, we chose
to assign phonation amplitudes equal to pu − min ( pu , pd ) and pd − min ( pu , pd )
to the components |u and |d, respectively, and turbulence amplitude min ( pu , pd )
to the noise component. In addition, here we extract a pulsating component as well,
corresponding to slow myoelastic vibration, whose amplitude is derived from the
probabilities p f and ps of fast or slow pulsation. For example, p f is derived from
T r [ρ | f f |], which is similar to equation (33).
Figure 14 shows the amplitude profiles that are extracted from the Hamiltonian
evolution, where we chose to measure phonation when min ( pu , pd ) > 0.5, otherwise
measuring along the slow myoelastic vibration axis. With non-physical freedom, we
collapsed the mixed state, along |u, |d, | f , or |s, using an equation similar to (34),
once every five measurements. The resulting sound, which can be considered as a
quantum-inspired audio effect, has the spectrogram depicted in Fig. 15, where the
most salient pitch and the onsets have been extracted again and superimposed.
Fig. 14 Amplitudes of components |u, |d, turbulence, and slow myoelastic vibration, resulting
from a Hamiltonian evolution from a mixed state, run on the vocalization of Fig. 2
218 M. Mannone and D. Rocchesso
Fig. 15 Sound synthesis obtained from the density matrix evolution from a mixed state, using the
component amplitudes depicted in Fig. 14, and all three components of phonation, turbulence, and
slow myoelastic vibration. The most salient pitch and onsets, as extracted from the synthetic sound,
are displayed as red dashed lines
In digital audio, talking about quantization does not mean referring to quantum theory.
Instead, quantization is meant to be the reduction of a continuous range of signal
amplitude values to a finite set of discrete values, with a cardinality that depends on the
number of bits dedicated to represent each discrete-time sample. Signal quantization
introduces a kind of noise, which tends to have a spectrotemporal structure that
somehow follows the signal, thus becoming audible as a distortion for low-amplitude
signals. A cure for quantization noise is dithering, i.e., adding some tiny broadband
noise to the audio signal itself, before quantization, thus making quantization noise
more spectrally uniform and perceptually tolerable (Pohlmann 1995). That injecting
dither noise to signals and systems can make human and machine processing more
robust is a fact that has been known for a long time, and widely applied in a variety of
fields, including audio and image processing. In quantum-inspired sound processing,
as illustrated in the example of Sect. 3.1, dithering can be used to control how erratic
leading-pitch attribution can be, in auditory scenes of competing sources.
As opposed to low-amplitude noise, that may actually make the pitch evolution
of a phon more stable, when a high-amplitude noise burst is encountered, it actually
acts as a bounce on the phon state, making it rotate by an angle θ . A sequence of
bursts, such as that of the example in Sect. 3.2, is much like a sequence of bounces
between billiard balls of highly different weights. Recently, the classically mechanic
behavior of balls, whose weights are in ratios of powers of 100, has been shown to be
Quanta in Sound, the Sound of Quanta 219
perfectly analogous to the kernel of the Grover algorithm for quantum search (Brown
2020), which is based on unitary reflections in the state space.
In the examples of Sects. 3.1 and 3.2, the quantum evolution is driven by potentials
that are derived from the same audio that is being processed. To turn these evolu-
tions into quantum algorithms we should freeze a reference audio segment, extract
the feature-based potentials from the time-frequency representation, and convert
the elementary unitary transformations into quantum gates, arranged along left-to-
right wires. Each stage of the quantum algorithm would represent a bounce or a
measurement in the phon space, as long as the operators are consistent with the
postulates of quantum mechanics. It should be noted that we have only been consid-
ering single-qubit (or single-phon) operators. The universe of multiple and entangled
phons remains to be explored.
In both quantum mechanics and sound signal processing, unitary operators and
unitary transformations have a central role. In fact, in physically-inspired sound syn-
thesis and digital audio effects, unitary matrix transformations are often found, as
scattering elements in feedback delay networks for artificial reverberation (Pulkki
et al. 2011). In these structures, if the feedback matrix A is chosen to be unitary,
an initial pulse bounces indefinitely, at each bounce scattering into a multiplicity of
other pulses. In the ball-within-the-box (BaBo) model (Rocchesso 1995), the matrix
A can be interpreted as a scattering ball that redistributes the energy from incoming
wavefronts into different directions, each corresponding to a planar wave loop, which
produces a harmonic series. Indeed, the matrix A does not have to be unitary for the
feedback structure to be lossless (Schlecht and Habets 2017). However, even staying
within the class of unitary matrices, A can be chosen for its scattering properties,
ranging from the identity matrix (no scattering at all) to maximally-diffusive struc-
tures (Rocchesso 1997; Schlecht 2020). A promising perspective for future quantum
sound processing, is to find realizable quantum operators for such matrices. In par-
ticular, the Hadamard operator and the Householder reflection are extensively used
in quantum algorithms, and these were proposed as reference matrices for feedback
delay networks with maximally-diffusive properties (Jot 1997). In the context of
QVTS, a Hadamard gate H converts a phon from |r to |u, and from |l to |d. If
followed by a measurement in the computational basis, it can be used to discriminate
between the two turbulent states. If inserted in a phon manipulation sequence, it
determines a switch in the vocal state of sound. Loops are not allowed in quantum
computing (Nielsen 2010), but by spatially unfolding feedback, a reverberator based
on feedback delay networks may be converted to a quantum algorithm with several
stages of unitary operators, acting as scattering elements on a multiplicity of phons.
As a non-negligible detail, banks of delay lines of different lengths, in the order of
tens of milliseconds, should be interposed between consecutive scattering sections.
Shor’s algorithm for factorization of a large integer N relies on an efficient way
of finding the periodicity (modulo N ) of the function a x , constructed from a ran-
domly chosen integer a that is smaller than N and coprime with it. To compute
the periodicity of a function, the Quantum Fourier Transform (QFT) operator is
used, which transforms a superposition
of N = 2n computational basis states on
n qubits, with coefficients x = x0 , x1 , . . . , x N −1 , into another superposition with
220 M. Mannone and D. Rocchesso
coefficients X = X 0 , X 1 , . . . , X N −1 = D F T (x), that is the Discrete Fourier
Transform of x. Using quantum parallelism, such DFT is implemented with O(n 2 )
quantum gates, while classically that would take O(n2n ) steps. Recently, a direct
transposition of the Fast Fourier Transform (FFT) into the form of a quantum circuit
has been proposed and called the Quantum FFT (QFFT) (Asaka et al. 2020). Instead
of the amplitude encoding used for the QFT, a basis encoding is −1
used, where a data
sequence x is expressed as a tensor product of vector spaces Nj=0 |x j . A potential
impact on audio signal processing would be that quantum parallelism would allow
to perform all frames of a N -bins STFT simultaneously, with O(N log2 N ) gates.
The aggregate features of an audio segment would then be encoded in the resulting
vector of qubits.
A respectable scientific theory helps find new results, confirms expectations, extends
the validity of known laws bringing them toward the realm of the unknown and
(formerly) inexplicable, and so on.
An exciting scientific theory leaves room for imagination and artistic creativity.
New ideas can arise from the interdisciplinary dialogue between people of differ-
ent fields. QVTS is intrinsically interdisciplinary, and we think it can enhance the
dialogue between worlds.
Interchanges between music and quantum mechanics constitute a relatively new
and flourishing research area. Our contribution to this field is the addition of the
human voice, and the use of vocal primitives as a probe to more generally investigate
the world of sounds. In Sect. 3 we have proposed some examples of a creative use of
QVTS, where the Hamiltonian evolution is the starting point for sound synthesis. In
this section, we suggest some further creative applications.
The density matrix can be exploited to improve source separation techniques. In
fact, the operation of partial trace on density matrix allows us to separate a system
from the environment, the reservoir. Choosing on which part of the whole system
we are making the operation of partial trace, we can interchangeably choose which
part we are neglecting. For example, given a polyphonic vocal recording, we can
establish that singer 1 is the system and singers 2, 3, and 4 are the environment (thus,
we can perform the partial trace on singers 2-3-4), or that singer 2 is the system and
singers 1, 3, and 4 are the environment, and so on. In fact, as a practical interest
in the domain of QVTS, we can think of a general recording, with multiple voices,
and interpret it as a statistic mixture of states. Voices might be organized as a solo
singing voice against a a background of several other voices of a choir—a quantum
choir. Therefore, QVTS may help analyze choral music. In addition, it can give us
hints also on how to create music. Creativity can precisely take off from mixtures of
states and vocal polyphony.
Because QVTS constitutes a bridge between sounds and quantum formalism, we
can play with the symmetries of particle processes and transform them to musical
Quanta in Sound, the Sound of Quanta 221
Fig. 16 Sketch for a music composition based on Bloch’s sphere for QVTS (by M.M.).
Starting from kids’ playing and moving toward futuristic scenarios of quantum choirs
and Quantum Inventions, in this chapter we presented the fundamental ideas of
the Quantum Vocal Theory of Sound (QVTS), along with some proposal of future
developments.
We aimed to discuss a supplemental formalism to describe sound (with the vocal
probe), rather than proving any “wrongness” or obsolescence of the classical formal-
ism, such as Fourier analysis, for sound. Our supplemental formalism is an alternative
one, it gives a new perspective, and it has the advantage of providing more infor-
mation, especially regarding Gestalt-related phenomena, as in the case of bistable
states.
QVTS is interdisciplinary in nature, as it provides a bridge between sound sources,
sound production, human perception, and the intuitive identification of sounds and
sound sources. In addition, the uncertainty (or fuzziness) that is proper to quantum
Quanta in Sound, the Sound of Quanta 223
References
Asaka, R., Sakai, K., & Yahagi, R. (2020). Quantum circuit for the fast Fourier transform. Quantum
Information Processing, 19, 277.
Aspect, A., Dalibard, J., & Roger, G. (1982). Experimental test of Bell’s inequalities using time-
varying analyzers. Physical Review Letters, 49(25), 1804.
Bell, J. S. (1964). On the Einstein Podolsky Rosen Paradox. Physics Physique Fizika, 1(3), 195–200.
https://doi.org/10.1103/PhysicsPhysiqueFizika.1.195.
Bigand, E., McAdams, S., & Forêt, S. (2000). Divided attention in music. International Journal of
Psychology, 35(6), 270–278.
Blutner, R., & Beim Graben, P. (2020). Gauge models of musical forces. Journal of Mathematics
and Music. https://doi.org/10.1080/17459737.2020.1716404.
Bogdanov, D., Wack, N., Gómez Gutiérrez, E., Gulati, S., Herrera Boyer, P., Mayor, O., Roma
Trepat, G., Salamon, J., Zapata González, J. R., & Serra, X. (2013). Essentia: An audio analysis
library for music information retrieval. In Proceedings of the 14th Conference of the International
Society for Music Information Retrieval (ISMIR) (pp. 493–498), Curitiba, Brazil
Bonada, J., Serra, X., Amatriain, X., & Loscos, A. (2011). Spectral processing. In U. Zölzer (Ed.),
DAFX: Digital Audio Effects (pp. 393–445). John Wiley & Sons Ltd.
Bregman, A. S. (1994). Auditory scene analysis: The perceptual organization of sound. Cambridge,
MA: MIT Press.
Brown, A. R. (2020) Playing Pool with |ψ: from Bouncing Billiards to Quantum Search. Quantum
4, 357, Verein zur Förderung des Open Access Publizierens in den Quantenwissenschaften.
Byrne, Á., Rinzel, J., & Rankin, J. (2019). Auditory streaming and bistability paradigm extended
to a dynamic environment. Hearing Research,383, 107807.
Cádiz, R., & Ramos, J. (2014). Sound synthesis of a gaussian quantum particle in an infinite square
well. Computer Music Journal, 38(4), 53–67.
Cariolaro, G. (2015). Quantum Communications. Cham: Springer.
Ciocca, V., & Bregman, A. S. (1987). Perceived continuity of gliding and steady-state tones through
interrupting noise. Perception & Psychophysics, 42(5), 476–484, Springer .
Costa Hamido, O., Cirillo, G. A., & Giusto, E. (2020). Quantum synth: a quantum-computing-based
synthesizer. In Proceedings of the 15th International Conference on Audio Mostly (pp. 265–268)
(2020)
Dalla Chiara, M. L., Giuntini, R., Leporini, R., Negri, E., & Sergioli, G. (2015). Quantum informa-
tion, cognition, and music. Frontiers in Psychology, 6, 1583.
224 M. Mannone and D. Rocchesso
Nielsen, M. A. & Chuang. I. L. (2010). Quantum Computation and Quantum Information. Cam-
bridge University Press.
Oliveros, P. (2002) Quantum listening: From practice to theory (to practise practice). In K. Siu Tong
& C. Sin-wai (Eds.), Culture and Humanity in the New Millennium: The Future of Human Values
(pp. 27–44). The Chinese University Press.
Palazzeschi, A. (1909). La fontana malata, from Poemi. Firenze, Cesare Blanc. https://www.
pensieriparole.it/poesie/poesie-d-autore/poesia-38384
Pohlmann, K. C. (1995). Principles of Digital Audio (3rd ed.). USA: McGraw-Hill Inc.
Pulkki, V., Lokki, T., & Rocchesso, D. (2011) Spatial Effects. DAFX: Digital Audio Effects (pp.
139–183). Wiley.
Roads, C. (2001). Microsound. Cambridge, MA: MIT Press.
Rocchesso, D. (1995). The ball within the box: A sound-processing metaphor. Computer Music
Journal, 4(19), 47–57. The MIT Press. http://www.jstor.org/stable/3680990
Rocchesso, D., Mauro, D. A., & Delle Monache, S. (2016). MiMic: The Microphone as a Pencil. In
Proceedings of the TEI ’16: Tenth International Conference on Tangible, Embedded, and Embod-
ied Interaction, in Eindhoven, Netherlands (pp. 357–364). New York, NY, USA: Association for
Computing Machinery, TEI ’16. https://doi.org/10.1145/2839462.2839467
Rocchesso, D., Mauro, D., & Drioli, C. (2016). Organizing a sonic space through vocal imitations.
Journal of the Audio Engineering Society, 64(7/8), 474–483. http://www.aes.org/e-lib/browse.
cfm?elib=18333
Rocchesso, D. (1997). Maximally diffusive yet efficient feedback delay networks for artificial rever-
beration. IEEE Signal Processing Letters, 4(9), 252–255. https://doi.org/10.1109/97.623041.
Rocchesso, D., & Mannone, M. (2020). A quantum vocal theory of sound. Quantum Information
Processing, 19, 292. https://doi.org/10.1007/s11128-020-02772-9.
Rovelli, C. (2021). Helgoland. Penguin Books.
Salamon, J., & Gomez, E. (2012). Melody extraction from polyphonic music signals using pitch
contour characteristics. IEEE Transactions on Audio, Speech, and Language Processing, 20(6),
1759–1770.
Schlecht, S. J. (2020). FDNTB: The feedback delay network toolbox. In Proceedings of the 23rd
International Conference on Digital Audio Effects DAFx-20.
Schlecht, S. J., & Habets, E. A. P. (2017). On lossless feedback delay networks. IEEE Transactions
on Signal Processing, 65(6), 1554–1564. https://doi.org/10.1109/TSP.2016.2637323.
Shamma S. (2013). Spectro-temporal receptive fields. In D. Jaeger , R. Jung (Eds.), Encyclopedia
of Computational Neuroscience. New York, NY: Springer. https://doi.org/10.1007/978-1-4614-
7320-6_437-1.
Spence, C. (2011) Crossmodal correspondences: A tutorial review. Attention, Perception, & Psy-
chophysics, 73, 971–995. https://doi.org/10.3758/s13414-010-0073-7
Spence, C., Senkowski, D., & Röder, B. (2009). Crossmodal processing. Experimental Brain
Research, 198(107), 107–111 (2009). https://link.springer.com/article/10.1007/s00221-009-
1973-4.
Verma, T. S., Levine, S. N., & Meng, T. H. (1997). Transient modeling synthesis: A flexible anal-
ysis/synthesis tool for transient signals. In Proceedings of the International Computer Music
Conference (pp. 48–51)
Vicario, G. B. (1963). La “dislocazione temporale” nella percezione di successioni di stimoli dis-
creti (The “time displacement” in the perception of sequences of discrete stimuli). Rivista di
Psicologia,57(1), 17–87.
Vicario, G. (1960). L’effetto tunnel acustico (The acoustic tunnel effect). Rivista di Psicologia, 54,
41–52.
Vochlea microphone. Retrieved from https://vochlea.com/.
von Helmholtz, H. (1870). Die Lehre von den Tonempfindungen als physiologische Grundlagefürdie
Theorie der Musik. Braunschweig: F. Vieweg und sohn.
Weimer, H. (2010). Listen to quantum computer music. Retrieved Jan. 20, 2021, from http://www.
quantenblog.net/physics/quantum-computer-music.
226 M. Mannone and D. Rocchesso
Yearsley, J. M., & Busemeyer, J. R. (2016). Foundations of probability theory in psychology and
beyond (special issue). Journal of Mathematical Psychology, 74, 99.
Youssry, A., El-Rafei, A., & Elramly, S. (2015). A quantum mechanics-based framework for image
processing and its application to image segmentation. Quantum Information Processing, 14(10),
3613–3638.
Quantum Cinema and Quantum
Computing
1 Introduction
This chapter on Quantum Cinema and Quantum computing engages with intuitively
developed images realized as digital dynamic 3D geometry. Some of these images,
which could be relevant in the context of quantum computing, are presented here.
Quantum Cinema—a digital vision (2010–2013) 1 mentored by the then most eminent
Austrian physicist Helmut Rauch, who first proved the 4π-value of the spin of a
neutron (Rauch et al., 1975). As a pioneer in the field of neutron interferometry, he
said,
There are strong indications, that we may understand the physical world as interference
pattern (Rauch, in a private conversation, 2012).
The Quantum Cinema project aimed to develop visual tools for the perception
of quantum phenomena. After all, computer animation is the perfect medium for
overcoming projective geometry, which is foundational for many scientific concepts.
We hold that only the invention of digital media makes the appropriate outlook of
higher mathematics possible, and animated 3D graphics are the new medium for a
holographic depiction model which was realized with the help of a group of artists
with trained imagination skills.
Quantum Cinema is based on the idea that geometry must not remain restricted
to the plane, 2-dimensional page—merely a shadow of rigid shapes projected onto a
plane as it is the case with projective geometry—but may create an abstract movie of
previously unimaginable quantum phenomena in order to gain cognitive and visual
access to the quantum realms.—Hence, the title Quantum Cinema.
The term Quantum Cinema was originally coined by the Austrian media theorist
Peter Weibel—who was so kind as to head this project—for his (then) futuristic
vision of a cinema that can be perceived without further external device directly in
the brain (Weibel, 2005). Nowadays, human enhancement projects like Neuralink,
brain–machine interfaces (BMIs) for full-bandwidth data streaming from a device
are under development (Musk, 2019).
Questions of imagination and visualization of quantum phenomena relevant in
the context of quantum computation shall be discussed here:
– How can we understand quantum phenomena such as the wave-particle duality
and the superposition of particles and the entanglement of two or more particles?
These are important features for the application of quantum algorithms since they
are produced by a discrete Fourier transform over the amplitudes of a particle’s wave
function. Fourier transform allows for any wave to be uniquely decomposable into
an infinite superposition of trigonometric waves of different frequencies (Fourier,
1822).
– How can we imagine phenomena on the foundations of nature which serve as the
most effective ingredients for quantum computing technology?
– How can we imagine the tiniest processes in nature, known as wave-particle
duality, that serve as bits in quantum computers?
– How can we imagine quantum fluctuations, the wavelike state of a particle that
produces a numeric output by means of a quantum computer?
1I am grateful to Quantum Cinema artists Nikola Tasic, Christian Magnes, Rudi Friemel, Kathrin
Stumreich and the mathematician and civil engineer Dr. Hans Katzgraber.
230 R. C.-Z. Quehenberger
Lady Lovelace is meanwhile widely recognized for the conception of the first
computer algorithm, published as a footnote to her translation of Italian military
engineer Luigi Menabrea article into English (Lovelace, 1842).
Although the further development of algebras remained widely abstract, in the
1830s, Augusta Ada King, the Countess of Lovelace, developed her imagination of
algebraic operations in higher dimensions.
Her interest in mathematics was fostered by her mother, Lady Annabella Byron,
dubbed “the Princess of Parallelograms” for her own fascination with mathematics.
She tried to keep her, as Lord Byron’s daughter, away from poetry to prevent her
from following in his footsteps. So Lady Lovelace was tutored by the mathematician
Augustus De Morgan, who was, with George Boole, one of the founders of formal
logic and developers of modern algebra. In Boolean algebra, True and False are
replaced by 0 and 1, which became the cornerstones of a binary decision-making
system to which the third indeterminate value, the state of superposition, was added
for quantum computation. In quantum logic, De Morgan’s Laws play an important
role in the projection-valued truth value assignment of observational propositions in
quantum mechanics.
By 1841, Ada Lovelace was developing a concept of “Poetical Science” in which
scientific logic would be driven by imagination. “Poetical Science” can also be under-
stood as a precursor of SciArt, a collaboration between the arts and sciences as it
is accomplished partly nowadays. For Lovelace, imagination is “the discovering
faculty, pre-eminently—science and religion were the same and imagination itself
was divine.
Imagination penetrates into the unseen worlds around us, the worlds of Science.
She saw mathematics metaphysically, as “the language of the unseen relations
between things” (Holmes, 2015).
Quantum Cinema and Quantum Computing 231
The physicist David Bohm spoke of an art form of science. The scientist is an artisan
who has to create sensitive instruments in order to discover oneness and totality—
like quantum computation technology, one might add—which aid perception and so
make possible both the testing of new ideas for their truth or falsity, and the disclosure
of new and unexpected kinds of facts (Bohm, 2004, 105).
On the other hand, the artist may create visual access to subscendent structures,
to what Plato called the world of ideas, that express the harmony and beauty in
nature—categories which remained more valid in the arts than in science.—The
beautiful formula as a guarantee for truth became obsolete.
Computer-generated art is based on mathematical algorithms which were previ-
ously considered as an art form.
Algorism is the art by which at present we use those Indian figures, which number two times
five. (Alexandre de Villedieu, poem Carmen de Algorismo, 1240)
This phrase alludes to the Indian origin of the numerals (Indorum ars numerandi)
or Hindu numerals. The term algorithm goes back to the name of Persian mathe-
matician Muhammad ibn Musa al-Khwārizmı̄ (780–850), who is arguably the most
important mathematician of the Middle Ages. Besides calculating the most exact
calendar of his time and his contributions to architecture and poetry, he developed
two distinct branches of mathematics, both of which owe their name to him: algebra
and algorithms (Brezina, 2006).
For the French, thirteenth century poet Alexander de Villa Dei mathematics was
part of the Liberal Arts and algorithms an art form. A century later, Leonardo da
Vinci pursued painting as a scientific method. For example, his advice on the mode
of studying anatomy was,
Study the science first, and then follow the practice which results from that science. (Leonardo
da Vinci [1490] 1877, 13)
232 R. C.-Z. Quehenberger
The Austrian quantum physicist Wolfgang Pauli confirmed that at the time of the
“re-opening” of the Platonic Academy in Florence (1438–39) sponsored by Cosimo
de’ Medici, Leonardo was the scientist while Marsilio Ficino, the “first academic”,
and Pico della Mirandola founder of esotericism, did not practice sober science, but
religious mysticism, which originated in a new cosmic feeling and expressed itself
in particular in the deification of space (Pauli, 1955, 8). In the fifteenth century, the
German philosopher and theologian Nicolas of Cusa proclaimed that space is infinite
and boundless, like the n-dimensional space in mathematics later on.
The invention of perspective and projective geometry centuries before by Renais-
sance artists, such as Paolo Uccello and Albrecht Dürer, only gained importance in
mathematics in the nineteenth century. Projective geometry was the basis for the
invention of cinema in the nineteenth century. The projection of an image onto a
screen by a beam of light led to new worlds’ arising at the point at infinity. Whereas
according to Gilles Deleuze narrative cinema produces “time crystals” by freezing
space and time into moving images taken from the real world (Deleuze, 1991, 132).
3D digital quantum cinema renders higher-dimensional virtual space configurations
composed of 3-dimensional moving spaces in time.
It is less known that Dürer also explored shapes of higher-dimensional spaces. He
designed Penrose type irregular five-fold patterns, one of them exactly matching the
Ammann-Beenker Tiling (Dürer, folio 142, in Ruppich, 1969, 73). The astronomer
Johannes Kepler also designed 4-dimensional polyhedra during his search for the
shape of the space with the potential for the formation of matter and the emergence
of motion (Lück, 2000). A closer look tells us that Kepler’s Pythagorean approach
towards the discovery of the three laws of planetary motion lead him in fact into
higher-dimensional space configurations (Quehenberger, 2022).
The history of arts and literature is full of comments on scientific shortcomings.
For instance, in the second half of the nineteenth century, during the hype about the
4th dimension, the poet Edwin Abbott Abbott made fun of limited mathematical
imaginations in his satirical novel about “Flatlanders” who cannot imagine a sphere
(Abbott, 1884). H. G. Wells invented a “Time Machine” (1895) for travelling on
the time axis that inspired relativity theory significantly, while during the cubist art
movement in the 1910s, Marcel Duchamp and Pablo Picasso interpreted aspects of
the 4th dimension by depicting the simultaneity of different perspectives in their
paintings. Inspired by conversations with Henri Poincaré, Duchamp also criticized
the “free choice of metrics” as applied in metric tensors with wooden rulers carved
in the arbitrary shape of dropped threads in his artwork (3 Standard Stoppages,
1913–14).
The famous “melting clocks” (The Persistence of Memory, 1931) paintings by
the surrealist Salvatore Dalí are a reaction to General relativity, wherein “time and
space lose the last remnant of physical reality.” (Einstein, 1916, 117). It appears
that the pseudo 4-dimensional Minkowski space led to a mental loss of a proper
imagination of 4-dimensional systems. Quantum Cinema models are trying to over-
come this deficit. In 2012, we produced the Quantum Fluxus Dada Duett as Art’s
Birthday present, a satirical dadaist sound poem on 100s of different currently
Quantum Cinema and Quantum Computing 233
used space concepts with Erwin Schrödinger’s sentence, “this is absurd” as refrain
(Quehenberger & Weibel, 2012).
We like to recall pataphysique as proclaimed by the French artist and writer Alfred
Jarry at the beginning of the twentieth century. He called for an ideology-free space
in science and announced a science beyond physics and metaphysics, a science of
imagination that underlies everything (Jarry, [1911] 1996). Only recently, SciArt has
emerged as the 3rd culture, opening up new collaborations between art and science.
them, we can design them and study them. If therefore, by example, Mechanics with more
than three dimensions must be condemned as devoid of any object, it is not the same with
the hypergeometry (Poincaré, 1895, 1).
Quantum Cinema research worked on the assumption that the aperiodic Penrose
tiling—a 2-dimensional slice of the 5- or 10-dimensional space (depending on its
fivefold or tenfold symmetries) which is used as models for quasicrystals (Shechtman
et al., 1984) can be adopted to quantum physics (Fig. 1).
The dual of the Penrose patterns is a cubic grid formed by parallels into five distinct
directions of space, named Ammann-bars after their explorer, Robert Ammann
(Grünbaum & Shephard, 1987, 571ff.). This grid is associated with the icosahedral
group, isomorphous to the group A5 (de Brujin, 1981). Penrose tilings are defined
in the 6-dimensional space through a strip and projection method. This gives rise
to the idea that is appropriate for the visualization for the quantum space which
is differing from the usual position and momentum-space. Accordingly, our first
intuitively designed quantum space looks like as shown in Fig. 2.
In quaternionic quantum mechanics, the spin–orbit interaction demands that the
(Euclidean) space–time dimension be 6-dimensional. Pauli correspondence is char-
acterized by the fact that the state space HP1 n S4 of a spin 1/2 particle system admits
Quantum Cinema and Quantum Computing 235
Fig. 1 The convex epitahedron E+ whose edges are partitions of the pentagrid; a polyhedron in
4-dimensional space (image: Quantum Cinema/Renate Quehenberger 2012© Bildrecht, Vienna
2021)
a natural embedding in R5 (Brody and Gräfe, 2011). Also, Louis de Broglie was
convinced of a 5-dimensional wave mechanics (de Broglie, 1927).
These facts seem to support our visualization method based on the Penrose tiling,
the 2-dimensional slice of the 5-dimensional space.
The British physicist Alain Mackay showed that a set of six vectors lies in parallel
to the fivefold rotation axis of the icosahedron (Mackay, 1981). This could possibly
be useful for the hexagonal grids used in quantum computation.
Since the beginnings of modern quantum theory (around 1926) there has been
a dispute about whether or which “images” of quantum events are, in principle,
possible.
We are tempted to regard it as the biggest irony in the history of quantum
mechanics, that the aim for “Anschaulichkeit” (German for “visual accessibility”)
as expressed by Werner Heisenberg in his first published article on the uncertainty
principle (Heisenberg, 1927), turned into the very opposite when the ramifications
of the principle were codified in the Copenhagen Interpretation, which states that the
236 R. C.-Z. Quehenberger
Fig. 2 The complex space model of a particle with straight lines, parallel into 5 directions
constructed from two interlocked icosahedra (image: Quantum Cinema/Renate Quehenberger,
2011© Bildrecht, Vienna 2021)
wave equation is not a real phenomenon but allows only statistical predictions.—How
would a quantum computer work then?
The Copenhagen Interpretation, which appears to be “orthodox” quantum theory
today (cf. Bohr, Heisenberg, and others), called for a ban on images: In principle, the
Heisenberg Uncertainty Principle prohibits thinking about processes and quantum
mechanical formalism in general. This was compared with the Byzantine iconoclasm
of the eight and the ninth centuries, when the “magical power” of images was opposed
and the divine principle was promoted as an obscured power of a non-imaginative
entity. In orthodox quantum theory, formalism itself has ascribed a “magic” (Feynman
et al.), whereby “images” generated in computer models (of causal, so-called “hid-
den” parameters) the result of their materialistic approach, threatening that magic by
demystification (Grössing, 2003).
Various “no-go theorems” were cited, which supposedly proved that such
“images” (which could, via “hidden parameters”, to a certain extent consti-
tute a “deeper” explanatory authority “behind” the observed quantum mechan-
ical processes) are impossible, although the “causal” interpretation after Louis de
Broglie—who first proposed matter waves (de Broglie, 1925), David Bohm and
others—refuted exactly this as viable counterexample.
Quantum Cinema and Quantum Computing 237
Perhaps one has to conceive the transition from one stationary state to the next in a similar
way to the transition from one image to another in films. The transition does not take place
suddenly. Rather, the one image becomes weaker and the other slowly emerges and becomes
stronger. For a while, both images permeate each other and one does not know which is
actually meant. Perhaps there is thus a state in which one does not know whether an atom is
in the upper or the lower state. (Heisenberg, 1989, 31)
The young Heisenberg illustrated his ideas about quantum indeterminacy to Albert
Einstein with a movie frame analogy: Maybe we have to imagine the transition from
one stationary state to the next in a similar way like the transition from one picture
to the next in some films. There are not enough discrete frames available in order to
achieve a clear image and the same happens with measurement. Only an ensemble
of particles allows the appropriate measurement outcome, just as you need a series
of images to make pin-sharp movie pictures (Heisenberg, 2011 [1979], 39).
Heisenberg’s analogy argument that we will never be able to see the still picture
between one and the next film frame in terms of mathematical “Anschaulichkeit”,
was the topic of our geometrical experiment applied to a geometrical model.
In the Quantum Cinema project, Heisenberg’s movie frame analogy was applied
to moving straight lines, partitions of the 6-dimensional space grid, which must be
considered to be in constant movement. The still picture shows all possible positions
of the points on the two circles merged in the white lines. Our geometrical digital
experiment comprised of possible positions of two points connected to two circles
in blue and red. The straight lines rotating at different speeds generated traces of
white lines that can be interpreted as probability distribution for the localization of
a particle, as shown in the still images (Fig. 3).
238 R. C.-Z. Quehenberger
Fig. 3 3D animated digital experiment: slowly moving centre points of circles create distorted
white lines (b) the traces of fast movement points creating a distorted bundle of white lines (image:
Quantum Cinema/R. Quehenberger 2012© Bildrecht, Vienna 2021)
We take two circles (a, b) with centre points A and B on a triangle. The two circles
perform rotations over these 3 points. The distance AB can be regarded as minimum
distance h. Thereby the points on the circle outline run into turbulence and the line is
no longer a straight line but generates different forms of loops on the circle line. Any
point and position of the point on this loop will be in a fast wavelike motion. One
cannot “measure” both the position and momentum of the point if the simulation is
played at full speed.
Any point and position of the point on this loop will be in fast motion so that
one cannot measure the position and momentum simultaneously. But Heisenberg’s
idea that the still image between one and the next picture cannot be seen in terms of
mathematical visual access turns into something different: in 3D digital geometry,
one can see all images in-between if the movie is run in very slow motion. The still
image shows all possible positions of the points on the two circles merged in the
white lines.
The Quantum Cinema 3D digital geometry examples on the uncertainty principle
were actually inspired by Johannes Kepler’s concept of quantum and fluxus with
the idea of “exire”—points in motion that create lines. He was relying on Proclus’
dynamic interpretation of Euclid’s “Elements” which was also the formation principle
of the hyper-Euclidean geometry we have supposedly only (re-)invented.
Alan Mackay also mentioned that unfortunately, Kepler lived too early to learn
that the five regular polyhedra are in fact (as we now believe!) eigenfunctions of
Schrödinger’s equation and thus do describe the symmetries of the planetary electrons
surrounding an atomic nucleus (Mackay, 1990, 3).
Based on our higher-dimensional geometry, we reached the conclusion that the
cause for the Heisenberg indeterminacy lies in the fact that higher-dimensional space
itself has to be regarded as being in constant motion. On a moving background of a
6-dimensional space grid, it is impossible to fix a moving particle and measure the
position and momentum of an object.
Quantum Cinema and Quantum Computing 239
The discussion on higher dimensions goes back to Ancient Greece, while evidence
for higher-dimensional space models can be found in ancient Egypt. Meanwhile,
Plato’s 5th element could be identified as unit cells of the 5-dimensional space
(Quehenberger, 2016).
Plato’s description of the composition of triangles that form the well-known
Platonic solids starts with the tetrahedron, the octahedron and the icosahedron, and
ends with the cube. He does not mention the dodecahedron, which does not fit into his
concept because it does not consist of triangles and was ascribed only to his scholar
Theaetetus, who also proved that there are only five regular convex polyhedra, the
Platonic solids, before Euclid. Plato himself added instead this obscure remark:
[...] since there is another fifth configuration the god used it for the delineation of the
universe.” (Plato, Timaios, 55c)
Echoing Pythagorean harmonies, Alan L. Mackay, who coined the term “general-
ized crystallography”, anticipated the existence of aperiodic fivefold crystals before
the discovery of quasicrystals, revealed:
What the Pythagoreans really discovered was group theory and of course this can be applied
to a vast range of circumstances. (Mackay, 1990, 3)
At this point, we have to consider that Plato was visualizing Pythagoras, and
both are evidently relying on ancient Egyptian mystery schools and mythology that
explains the origami of 5-dimensional space conveyed as sex and crime stories about
gods and goddesses (Quehenberger, 2013, 2019). Pythagoras understood the signif-
icance of making the connection between the sensory aspects of our world and the
realm of numbers with the goal of realizing the pre-established harmony of the
cosmos (Miller, 1996, 181).
240 R. C.-Z. Quehenberger
The mathematicians Shing-Tung Yau and Steve Nadis mention that geometriza-
tion, the idea that the physics we observe springs directly from geometry, stands at
the core of both approaches. Would they be surprised to read that Plato’s 5th element
could be composed as 26-dimensional space, as required by String Theory for one
particle to inhabit in order to achieve the unification?
On digesting this higher-dimensional information from Plato’s world of ideas,
we may well convince ourselves that the Pythagoreans also worked with this
higher-dimensional space concept that was lost in religious doctrines in the ensuing
centuries.
Higher dimensions were again considered at the turn of the seventeenth century
by Renée Descartes, who was the first to introduce the notion of higher dimensions
in relation to equations of higher order (Descartes [1637] 1902, 13). Descartes is
well-known for the x, y coordinate system, however, as early as 1610, he also stated
that a combination of systems of coordinate axes is a prerequisite for the depiction of
higher-dimensional spaces. He worked with the German weaver Johann Faulhaber,
Quantum Cinema and Quantum Computing 241
who also produced calculations for Johannes Kepler—on higher order simplices.
Faulhaber worked on cubic equations based on Girolamo Cardano, who had first
published Tartaglia’s solution a century earlier (1545) and defined formulas for sums
up to the 17th power, which is known as the Faulhaber Formula (Faulhaber, 1631).
Faulhaber’s work is still relevant for computation because his works influenced
Jakob Bernoulli, and Bernoulli numbers are the core ingredients of the first so-called
software program of Ada Lovelace that she developed for the Analytical Engine
designed by Charles Babbage.
The Analytical Engine weaves algebraic patterns like the Jaquard loom weaves flowers and
leaves. (Lovelace, in Padua, 2015, 88)
It is notable that punched cards were invented for weaving on a loom in 1804
by Jean Marie Jacquard and were once commonly used in data storing and data
processing applications or to directly control automated machinery even after the
development of electronic computers in the 1950s and until recently.
Erwin Schrödinger, who first proposed entanglement imagined the wave func-
tion—which he formulated for the first time—as the description of a phenomenon
located in higher-dimensional space. He regarded the purely statistical Copen-
hagen interpretation, which takes the wave function merely as probability, intro-
duced by Max Born after Heisenberg discovered quantum indeterminacy, as a
terrible misinterpretation of his theory. This development of quantum theory went
in a completely different direction than Schrödinger had wanted. Abstract statistics
doesn’t demand any kind of imagination. Even the Nobel laureate Gerard’t Hooft
admitted, speaking about higher dimensions, “I leave them out whenever I can” (‘t
Hooft in QC-Interview, Quehenberger, 2011).
Contributing to its complexity, the American physicist Richard Feynman was the
first to imagine a quantum computer that would be able to calculate the simulation
of complex quantum systems:
But the full description of quantum mechanics for a large system with R particles is given
by a function ψ(x1 , x2 , . . ., xR , t) which we call the amplitude to find the particles x1 , . . .,
xR , and therefore, because it has too many variables, it cannot be simulated with a normal
computer with a number of elements proportional to R or proportional to N. (Feynman,
1982)
Only recently, the discussion on higher dimensions gained new momentum. In the
operational approach to general probabilistic theories, one distinguishes two spaces,
the state space of the “elementary systems” in the quantum world and the physical
space in which “laboratory devices” are embedded. Each of those spaces has its own
dimensionality. This means there is a different minimal number of real parameters
(coordinates) needed to specify the state of the system and for a point within the
physical space (Dakic and Brukner, 2016).
6 Schrödinger’s “Entanglement”
The rather romantic term entanglement was coined by the Austrian physicist—
and notorious womanizer who lived with two wives—Erwin Schrödinger, whose
name is tied to the famous Schrödinger wave equation.
Schrödinger tried to understand the possible unification of matrix mechanics by
Heisenberg, Jordan algebra and his wave mechanics, since partial differential equa-
tions are equivalent to quadratic forms of infinitely many variables (Schrödinger in
a letter to Born, April 14, 1926, in Meyenn, 2010, 201).
Finally, Paul Dirac’s attempt to prove the equivalence of matrix mechanics and
wave mechanics by essential use of the delta function and his bra-ket formalism
based on hypercomplex numbers similar to William Rowan Hamilton’s quaternions
led to a deeper understanding of the foundations (Dirac, 1930). As artists, we are
trying to understand mathematical abstract formalism in a way that can be visualized
geometrically. We are less focussed on measurement outcomes than on ontological
questions—what is really happening on the subscendent level-1?
In the information interpretation of quantum mechanics, information is the most
fundamental basic entity (Svozil, 2000). Every quantized system associated with a
definite discrete amount of information can be traced back to Erwin Schrödinger’s
“catalogue of expectation values” mentioned in his seminal papers “On the Current
Situation of Quantum Mechanics” (Schrödinger, 1935). Here, he also introduced the
famous simultaneously dead and alive “Schrödinger’s cat” metaphor.
In this drastically cruel thought experiment about a quantum cat in a closed
box with cyanide, while hidden from the observer, the cat is dead and alive at the
same time. It is intended to serve a better understanding of the phenomenon of the
superposition of a quantum state.
Well, ’state’ is a term that everybody uses today, even the holy PAM, but it doesn’t make it
more rich in content (Schrödinger in a letter to Pauli, Dublin, beginning of July 1935, in
Meyenn, 1985, 406)
The “holy PAM” is Paul Adrien Maurice Dirac, the British physicist who first
defined the spin of a particle. Wolfgang Pauli famously said of him: “There is no
God and Dirac is his prophet.” (Pauli, in Meyenn, 2010, 406).
Schrödinger’s (1935) articles were immediate answers to Albert Einstein’s
question “Can quantum–mechanical description of physical reality be considered
complete?” wherein he assumed a local deterministic realism of quantum states
(Einstein et al., 1935).
Schrödinger also reacted to the forbidden determinism by introducing the term
“entanglement” for the undetermined state depending on an earlier random event
instead of Einstein’s “spooky action at a distance” of which Isaac Newton had already
stated:
I believe no man who has in philosophical matters a competent faculty of thinking, can ever
fall into it. (Newton in a letter to Richard Bentley in 1693, quoted by Price, 2013)
Quantum non-locality does not prove that “signals” travel “faster than light”.
Rather, it shows that at a deep level of reality the speed of light as a limiting factor is
irrelevant because phenomena are instantaneously connected regardless of distance.
Because the spin of a particle does not exist until a measurement is made, the act
of making the measurement and determining the axis of spin of particle 1 will also
determine the spin of particle 2 no matter how far apart it is from particle 1. Particle
2 will instantly respond to the state of particle 1, even if it is on the other side of the
universe.
At any instant measurement of particle 1 is performed, particle 2, which may be
thousands of miles away, will acquire a definite spin—“up” or “down” if a vertical
axis, “left” or “right” and a horizontal axis is chosen. How does particle 2 know
which axis was chosen? There is no time for it to receive that information by any
conventional signal (Capra, 1982, 85).
Wolfgang Pauli approached the entangled quantum system as follows:
One must differentiate between different layers of reality, an R that contains all the informa-
tion that can be obtained from measurements on 1 and one (deducible from R) r which only
contains the information which can be obtained by measurements on 1 alone.
A complete description would have to assign characteristics to the state of the particle
1 which must already contain all those properties of 1 which — according to possible
measurements on 2, which do not interfere with 1 — can be predicted with certainty. (Pauli
in a letter to Heisenberg, 1935, in Meyenn, 1985, 404)
The effect of the environment through decoherence becomes more accessible in the
artistic interpretation of a higher-dimensional framework based on the Penrose tiling
shown above:
Let us assume that for any interconnected system of two or more measurable
quantum objects of higher order, say qubits (no matter how far they are separated),
the “miracle” of a “minimal distance effect” occurs in a spatial framework that
exhibits principles of the aperiodic Penrose tiling. For this, Conway’s theorem holds
that for any circular region of diameter d, the distance from the perimeter of the
“home town to the perimeter” of the duplicate town is never more than d times half
of the cube of the golden ratio, or 2.11 times d.
If you walk in the right direction, you need not go more than that distance to find yourself
inside an exact copy of your home town. The theorem also applies to the universe in which
you live. Every large circular pattern (there is an infinity of different ones) can be reached
by walking a distance in some direction that is certainly less than about twice the diameter
of the pattern and more likely about the same distance as the diameter. On a Penrose pattern
you are always very close to a duplicate of home. (Gardner, 1989, 10)
what Plato called a “figure of marriage”. In his symposium, Plato let Socrates quote
his teacher Diotima, “Love is the medium”.
The strikingly nonclassical properties of entangled states lead to more and more
novel demonstrations of fundamental properties of quantum mechanical systems
(Greenberger-Horne-Zeilinger, 1993). The young field of quantum information
exploits such entangled quantum states for new types of information transmission
and information processing (Bennett, 1995).
In 1995, the Austrian physicist Anton Zeilinger and co-workers reported
the first experimental realization of quantum communication, verifying the
increased capacity of a quantum information channel by “quantum dense coding”
(Mattle, Zeilinger et al., 1996). The applied scheme was theoretically proposed
by Bennett and Wiesner and utilizes two entangled two-state systems (Bennett &
Wiesner, 1992).
The output of a quantum information processor is, by convention, measured in
the basis states with Boolean values |0 and |1 denoted as A and B, often referred
to as Alice and Bob. Any controlled unitary gate can be realized using at most two
CNOT gates and four single-qubit gates. This acts on the state of two qubits, known
as the control qubit, C, and the target qubit, T (Barnett, 2009, 146). √
Typically, four specific two-qubit states with the maximal value of 2/ 2 are desig-
nated as Bell states. These Bell pairs represent all the ways that can be chosen to have
correlations or anticorrelations between the |0 and |1 states of the computational
basis and the |+ and |− states of the X basis.
These four states form a complete basis for two qubits. So any state, entangled or
not, can be expressed as a superposition of Bell states. They are known as the four
maximally entangled two-qubit Bell states and form a maximally entangled basis
known as the Bell basis of the 4-dimensional Hilbert space for two qubits.
From the perspective of quantum information theory, any two-state quantum
system (represented by a0 |0 + a1 |1) can be used to provide a physical imple-
mentation of a qubit like the bit in classical information theory. Examples that have
been realized in the laboratory include the two orthogonal polarization states of a
photon, the orientation of a spin-half particle, a pair of electronic energy levels in an
atom, ion, or quantum dot, and a pair of paths in an interferometer. However, it is
convenient to treat a qubit as though it were a spin-half particle and to describe it by
Pauli operators (σ_x, σ_y, σ_z) (Barnett, 2009, 46).
246 R. C.-Z. Quehenberger
In the 1950s, the physicist Léon Brillouin, who applied information theory to physics,
introduced space cells on the fundamental level of reality. Before he made major
contributions to crystallography by finding the elementary, irreducible zone, the
first Brillouin zone reduced by all symmetries in the point group of the lattice.
He sketched a “miscellaneous” cell on the foundations of the quantum realm with
information content per element of data (Brioullin, 1952, 72). The term “miscel-
laneous” stands for a compound of elements in movement which also applies to
the 6-dimensional space lattice and inherent symplectic manifolds in movement.
We envisage a compound of dynamic elements (points and straight lines) forming
5-simplices, 3D representation of the Penrose Kites and Darts tiling (E±). All parts
of these unit cells of the 5-dimensional space, its points, edges and faces have to be
thought of as being in constant movement, forming different angles and composi-
tions. We imagine iterations of the size of the lattice forming simplices with parti-
tions of circle decoration (cf. Conway decoration of the Penrose kites and darts
tiling) forming wavelike structures in movement. Their traces would, according to
Plato, lead to the formation of matter. We imagine that a particle as expressed by
Schrödinger’s wave equation is the result of permutations of space cells according
to symmetry operations in accordance with group theory.
Our Platonic approach seems to meet symplectic Clifford algebra as interpreted by
Basil Hiley: The basis is not matter or fields in interaction in space–time, but a notion
of “structure process” from which the geometry of space–time and its relationship to
matter emerge together, providing a way that could underpin general relativity and
quantum theory (Hiley, 2016). All the quantum properties of Schrödinger and Pauli
particles can be described entirely from within a Clifford algebra taken over the reals,
there is no need to resort to any “wave function” (Hiley & Callaghan, 2010).
Such a space in motion would not permit the measurement of the momentum
and position of a particle, and provides a logical geometrical framework for the
Heisenberg uncertainty principle: Higher-dimensional space, as we know it from
visualizations of hypercubes, is in motion.
In this respect, the dynamic structure of the infinite 5-dimensional space may be
a visualization of the “pattern that connects” as coined in cybernetics by Gregory
Bateson. Only by means of 3D animated geometry can we visualize the idea of
dynamic building blocks for systems like “Jumbled Boxes”, as Gordon Pask drew
them (Clarke, 2009, 50).
Various Unified Theories can be visualized by means of this 3D representation and
its composite configurations, such as of the 26-dimensional space continuum arising
from inside the 4-dimensional dodecahedron, identified as Poincaré’s dodecahedral
space (Quehenberger, 2018).
However, we do not expect that this model will be accepted anytime soon since the
physics community remains agnostic about an underlying space. In Quantum Gravity
emergent space–time scenarios, space and time are denied at the fundamental level
of reality (Oriti, 2006). In our model “time” is represented in the 4th dimension of
Quantum Cinema and Quantum Computing 247
reality on the globe, while if we apply the same model to quaternionic quantum
mechanics we get quantum space–time crystals, as proposed by Nobel Laureate
Frank Wilczek (2012). This system, which is periodic in time and also in space, was
realized experimentally by confining ions in a ring-shaped trapping potential with a
static magnetic field soon thereafter.
Progress in quantum mechanics could have been achieved much faster if John von
Neumann’s Mathematical Foundations of Quantum Mechanics (1932) would have
been less determining. His proof that no hidden variables can exist held for decades
until it was finally refuted by John Bell’s with his non-locality theory on hidden
variables (Bell, 1966). Everybody continued to cite the von Neumann proof until
John Bell (Bell, 1964) rediscovered that von Neumann’s no hidden variables proof
was based on an assumption that “can only be described as silly […] so silly […]”
(Mermin, 1993).
But back in 1935, the German philosopher and mathematician Grete Herman
already disproved John von Neumann’s proof that no hidden variables can exist.
She pointed out a glaring deficiency in his argument, however, she seems to have
been entirely ignored. When the Einstein–Podolsky–Rosen paper was published,
Wolfgang Pauli asked his German colleague Werner Heisenberg to write a “pedagogic
response” concerning the determinism in quantum mechanics to the “old boys” (Pauli
in a letter to Heisenberg, Zurich, 15 June, 1935 in Meyenn, 1985, 412).
Heisenberg replied that Grete Hermann already had already written a treatise
which he found “overall reasonable (despite a philosophical rational tendency)”
(letter from Heisenberg to Pauli, Leipzig, 2 July 1935, Meyenn, 1985, 408).
Grete Hermann was the first physicist who asked for the ontologic grounds of
observables. With her philosophical request to find other characteristics which deter-
mine the path of a particle, we should not stop searching for underlying causes.
She called for an improvement of the quantum mechanical formalism (Herzenberg,
2008):
We need to find the causations of the measurement results and predictions — there is only
one sufficient reason to stop looking for it — that you know already the source. (Hermann,
1935)
Grete Hermann showed that causality was retained in the sense that after an
interaction, causes could be assigned for a particular effect. This allows for the
248 R. C.-Z. Quehenberger
possibility that physical processes may be strictly determined even though the exact
prediction is not possible (Lenzen, 1969).
Her work with Werner Heisenberg, Carl Friedrich von Weizsäcker and other
prominent physicists during the early 1930s led Hermann to the idea of a rela-
tional interpretation of quantum mechanics. This term refers to an interpretation of
quantum theory that eliminates the concepts of the absolute state of a system, which
was only recently developed more fully by Carlo Rovelli. Niels Bohr arrived consid-
erably later at the relational concept of quantum mechanical states by considering
the Einstein–Podolsky–Rosen thought experiment, and he became known for this
approach, while Grete Hermann’s work in introducing it has been all but forgotten
(Jammer, 1974, 208).
Hermann has studied in Göttingen under Emmy Noether, one of the iconic figures
of twentieth century mathematics, who revealed the very important general connec-
tion between symmetries and conservation laws in physics, known as Noether’s
theorem.
In her thesis, Grete Hermann demonstrated that Noether’s proof of the Lasker–
Noether theorem could be turned into polynomial ideals in Noetherian rings. This was
the first efficient algorithm for computing primary decompositions for polynomial
rings over a field of characteristic 0 (Hermann, 1926).
Hermann completed her work before the availability of computers, and even before
the idea of an effective procedure had been formalized. Algorithms that compute
primary decompositions are very useful nowadays for quantum algorithms.
Ironically, Bohm’s hidden variable theory and Bohmian mechanics were falsely
declared as deterministic and was even “verboten” (forbidden) to be taught in recent
decades (Basil Hiley in QC-interview, Quehenberger, 2012a). In its original form, the
mathematical formalism of the de Broglie-Bohm theory refers to hypothetical ontic
elements, such as the quantum potential comparable with John Bell’s "beables”. John
Bell was inspired by his teacher David Bohm, and his idea of an underlying reality
that we can imagine like a pattern in movement (Bohm, 2004, 92).
“Bell’s inequality” and separately the Kochen–Specker theorem by Simon Kochen
and Ernst Specker (1967) first demanded contextuality to be a feature of quantum
phenomenology was a correction of the history of quantum mechanics that opened
up a new field of research, namely quantum information—and Bell’s non-locality
theorem became the basic feature of quantum computation.
Also, the Kochen–Specker theorem demonstrates that it is not possible to repro-
duce the predictions of quantum theory in terms of a hidden variable model
where the hidden variables assign a value to every projector deterministically and
non-contextually (Kunjwal & Spekkens, 2015).
When John Bell started to work on the foundations of quantum mechanics, there
was hardly any interest in such topics. Even worse, working on foundations was
not considered to be a proper topic for a physicist, so the CERN employée wrote his
seminal article during his sabbatical leave, as his then-colleague Reinhold Bertlmann
tells: John Clauser, the first who had the courage to carry out an experiment on
Bell inequalities in the 1970s had to struggle enormously to get the resources for
doing the experiment. Only after the experiments of Alain Aspect in the 1980s the
community slowly began to realize that there was something essential to Bell’s s
theorem. In the 1990s, the field of applications of entangled states and of Bell’s
theorem opened up, followed by experiments on quantum teleportation, quantum
cryptography, long-distance quantum communication and the realization of some
of the basic entanglement-based concepts in quantum computation (Bertlmann &
Zeilinger, 2017, vi).
Aspect was following Bell’s example by using entangled photons instead of elec-
trons, which proves physical reality must be un-local and all results show the impossi-
bility of a Local Realistic interpretation of quantum mechanics (Aspect et al., 1981).
After that, the perfect anti-correlation of entangled pairs has been repeated many
times since the easiest method of producing entangled states was achieved by Anton
Zeilinger’s group after their entangled photon experiment (Kwiat et al., 1995). The
pairs of polarization-entangled photons were produced by type-II down-conversion
in a nonlinear beta-barium-borate crystal that was incidentally found in the desk of
their colleague, Ralph Höpfel. The angle of 0.72° in their experimental set-up seems
to indicate the 72° of the 5-dimensional space grid.
The second obstacle seems to be the lack of realism attributed to the quantum
world. How could we build technology on something the existence of which is
questioned?
The realist quantum approaches often distinguish between ψ-ontic, ontological
quantum theories and ψ-epistemic theories (Fraassen, 1972). “Ontic states” are
250 R. C.-Z. Quehenberger
what Terry Rudolph calls real states versus “the mist (quantum state) is just knowl-
edge/information about the real state” referred to as “psi-ontic” and “psi-epistemic”,
respectively (Rudolph, 2017). In other words, type-I ψ-ontic interpretations are
defined as those in which the probabilities of measurement outcomes are determined
by intrinsic properties of the observed system. The other, ψ-epistemic, theories view
the quantum state as merely representing knowledge of an underlying objective
reality in a sense somewhat analogous to that in which a state in classical statistical
mechanics assigns a probability distribution to points in phase space (Cabello, 2017).
Schrödinger was considering to assign a physical realism not to the wave function
ψ itself but to the quadratic function of ψ (Schrödinger in a letter to H. A. Lorentz,
June 6, 1926, in Meyenn, 2010, 254ff.).
Twenty-five years ago, David Bohm and Basil Hiley described the close resem-
blance between key properties of deterministic, hidden variable, pilot-wave theory
and emergence theory in the book The Undivided Universe (1993). Their theory
describes the emergent formation of ordered (i.e., negentropic) states in nonlinear,
self-organizing systems, such as deterministic chaos (Walleczek et al., 2019).
David Bohm’s philosophy of the undividable wholeness where there are “no
separate whirlpools interacting” was inspired by the Fourier Universe where the
subscendent space is a carpet-like pattern of frequencies. This enfolded [or implicate]
order contains the general totality of what is called the “explicate” or “unfolded” order
(Bohm, 1980, xv).
We believe that the fractal nature of the Penrose tilings and their 3D representations
where each triangle is decorated with a fraction of a circle, as suggested by John
H. Conway for the Kites and Darts tiling, is appropriate to visualize this pattern
of frequencies. Similarly, like the triangles with curves, the underlying reality as
conceptualized by String theorists consists of fractions of flat space with a string
attached, named Dirichlet branes (Polchinski, 1995). This idea seems to meet our
concept with the triangular faces of epitahedra with the 1-dimensional “string-like”
circular decoration. Joseph Polchinski also found that the product of the electric and
magnetic charges is a single Dirac unit, and that the quantum of charge takes the
value required by string duality.
If we assign some realism to the straight lines and attribute electric force to them,
(like Maxwell and Faraday’s lines of force) and magnetic force to circles, we derive
the image of the wavelike nature of reality suggested by Joseph Fourier and Bohm.
Usually, the observables are pictured as points on the surface of a sphere, the
Bloch sphere—it is easy to visualize, but in some respects, it is quite misleadingly
simple (Bengtsson & Zyczkowski, 2006, 215).
It is a sphere of unit radius, with each point on its surface corresponding to a
different pure state. A given observable consists of two antipodal points on the surface
of the Bloch ball representing a pair of mutually orthogonal states two orthogonal
pure states. The north and south poles correspond to the quantum states so that the
four possible results 00, 01, 10 and 11 correspond to measurements of the two qubits
as the Bell basis.
During the Quantum Cinema research project, we were trying to find a more
detailed 4-dimensional representation of such a quantum system.
There are examples of quantum states depicted as compositions of nested Platonic
solids: Partially transposed states occupy the remaining four corners of the unit cube,
which is filled with two tetrahedra in an antipodal position. Two states are regarded
as “equally entangled” if they differ only by a choice of basis in Hilbert spaces H1
and H2 . The separable subset in 2 × 2 dimensions is the intersection of two tetrahedra
and can be represented by an octahedron (Vollbrecht and Werner, 2000).
Another representation of probabilities of maximally entangled states by using
an octahedron inside a tetrahedron is quite similar to our intuitive “entanglement”
image above (Fig. 4), where two epitahedra form the dodecahedron as Boolean
intersection. We may assume that this emerging dodecahedral space can be compared
to the maximal Kús-Zyczkowski ball of absolutely separable states that lies within
the double pyramid. It touches the faces of the pyramids at the four points which
mark the closest separable states to the four Bell states. The entangled Weyl states are
located in the four corners of the tetrahedron outside the double pyramid, extending
from the separable states to the maximally entangled Bell states at the tip (Bertlmann-
Friis-Bulusu, 2016).
There are several indicators for fivefold symmetries in the quantum system. Most
famously Klyachko and co-workers assigned a 5-vector an orthogonality graph in the
form of a pentagram, and in this way derive a Kochen–Specker inequality for spin
1 systems where the set of all functions forms a Vorob’ev-Keller cone (Klyachko
et al., 2008). This 5-vector can be visualized by the 5 edges of the pyramidal cone of
the 3D representation of the Penrose Kites and Darts tiling E±. We assume that E±
is a golden 5-simplex that corresponds to the unique optimal configuration of C5 , of
which it is said that it looks like an umbrella (Bharti, Cabello et al., 2019).
We maintain that Mermin’s magic pentagram as well as several other contextu-
ality proofs can be drawn back to higher-dimensional spatial structures, namely the
two antipodal entangled simplices (Quehenberger, 2017). The here presented entan-
gled simplex representation (2E±) (Fig. 5) also seems to conform to the Penrose
Dodecahedron as contextuality proof, an image that goes back to Penrose’s idea of 5
252 R. C.-Z. Quehenberger
Fig. 4 “Entangled state” of two epitahedra with a dodecahedron in the centre; 3D animation of
rotating faces result in full circles of the Conway circle decoration of the Penrose tiles, (image:
Quantum Cinema/Renate Quehenberger, 2011© Bildrecht, Vienna 2021)
Fig. 5 3D Representation of
a three-partite quantum
system indicating spin states
in the framework of the
Poincaré sphere. The sphere
is drawn with three great
circles and labels for six
basic polarization H (linear
horizontal), V (linear
vertical), D (linear diagonal),
A (linear antidiagonal), R
(right-hand circular) and L
(left-hand circular) and
images of the polarization
vectors for each where the (i,
0) (image: Quantum
Cinema/Renate
Quehenberger, 2017©
Bildrecht, Vienna 2021)
Quantum Cinema and Quantum Computing 253
interlocked cubes (Zimba & Penrose, 1993). In our “entanglement figure”, the apexes
of these two irregular pyramids (E±) with dynamical edges from partitions of the
pentagrid may be assigned to the spin up ↑/right and the spin down ↓/left, the other
two measurements would then show 01, 10 (see Fig. 11).
We assume that it was possible to translate the magic Mermin pentagram from
a finite projective geometry description by Michel Planat et al. (2006) into a hyper-
Euclidean representation by means of the double epitahedron (2 E+). The deviated
Pauli operators are in projective geometry represented by three points of PG(1, 2),
their three “Jacobson” counterparts, and the four points whose both coordinates are
zero-divisors.
These points on Mermin’s magic pentagram correspond with the intersecting
points of the entangled simplex figure.
Our 3D representation of the magic pentagram seems to correspond with the three
different interlocked spaces associated with Stokes relations, summing up to zero.
The coordinate system of Stokes vectors with components S1 , S2 and S3 (usually
depicted as three vectors) appear in their spatial form. It appears that from these two
entangled simplices (2E + , or better an indeterminate E± flipping from E+ to E-?)
we derive the picture for a three-partite quantum system. Inside the Poincaré sphere,
it illustrates the three spaces of possible polarizations of electromagnetic waves and
their corpuscular behaviour. In this model, we may assign one direction of the spin in
one epitahedron (E + beige), right (R), and another direction to the other (E + grey),
left (L) (see Fig. 5), where the states |0 and |1 are the eigenstates (Quehenberger,
2017).
Entangled “state” of spin 1/2 particles can be assigned to the Pauli operators, a
set of three 2 × 2 complex matrices (usually denoted as σ_x, σ_y, and σ_z … by the
Greek letter sigma (σ).
Here they are represented by the node points of the edges of the three intersecting
space cells while partitions of the edges surrounding the dodecahedral centre perform
the Clifford algebra of Pauli spin matrices as permuting elements. This construc-
tion, applicable for any number of degrees of freedom, reveals an interesting cyclic
geometric structure that reminds us of a “quantum- merry-go-round” (cf. Uchida
et al., 2015).
It looks as if this 8-dimensional representation of qubits (if we count two inter-
locked 4-dimensional systems) inside the Poincaré sphere could lead to a more
detailed vision of quantum circuits. The advantage of the entangled E± figure, the
two antipodal, spinning golden simplices, for the “conventional” Platonic structure
is that it can be considered as a “real” partition of higher-dimensional space and the
rational principles of algebraic structures may be applied in a dynamic way for a
simulation of movement. Yet, it seems that until there is more discussion, abstract
formalism will continue to prevail:
I’ve always regarded the arrangement into a 5-pointed star of the ten operators in the eight-
dimensional proof of the BKS theorem as nothing more than a useful mnemonic for what
they are. All you have to remember is that the four going along one line of the star are XXX,
XYY, YXY, and YYX.
(N. David Mermin in an email to the author, 05/21/2015)
254 R. C.-Z. Quehenberger
Charles Babbage first realized the idea of gates in his design of the Analytical Engine
with barrels that might act on 50, 100 or more vertical systems of control levers with
cranks, pivots and sliders (Bromley, 1982). These gates now became quantum gates—
for example, the Hadamard gate, which is the one-qubit version of the quantum
Fourier transform. Although Babbage knew Charles Fourier socially, Babbage did
not seem to grasp the significance of his 1811 work on heat propagation, nor did he
seem to know of Joule’s endeavours with heat and mechanical energy. Both could
not have had any idea that their works would merge in a quantum computer driven
by Fourier amplitudes.
In quantum computation, it is sometimes useful to represent a quantum operation
or protocol as a quantum circuit that is reminiscent of the equivalent circuits in
electrical and electronic engineering.
Here, De Morgan’s Laws are applied as powerful rules of Boolean algebra and set
theory that relates the three basic set operations (union, intersection and complement)
to each other. They denote a pair of transformation rules that are both valid rules of
inference, providing an indispensable tool when simplifying digital circuits involving
and/or, and not gates in quantum computing (Mano, 1988).
If A and B are subsets of a universal set U, de Morgan’s laws state that.
(A ∪ B)’ = A’ ∩ B’.
(A ∩ B)’ = A’ ∪ B’.
where ∪ denotes the union (OR), ∩ denotes the intersection (AND) and A’ denotes
the set complement (NOT) of A in U, i.e., A’ = U\A (Bernal, 2005).
In order to create a quantum algorithm, the CNOT and Hadamard gates followed
by measurement in the computational basis constitute a Bell-state measurement. It
is carried out on the first two qubits and leaves the last in one of the states |ψ-, σz|ψ|,
σx|ψ| and σy|ψ. These four possible measurements of Bell states are associated with
the binary numbers 00, 01, 10 and 11 (Barnett, 2009, 152).
The unitary operators on a single qubit, X, Y, Z correspond to the measurement
of the spin along the x, y, and z axes, respectively. The X matrix is often known as
the quantum gate, by analogy to the classical gate, while the Hadamard gate, which
is also key to quantum computing, is the one-qubit version of the quantum Fourier
transform.
The X and Z Pauli matrices are also sometimes referred to as the bit flip and phase
flip matrices: the X matrix takes |0 to |1 and |1 to |0, earning the name bit flip;
and the Z matrix leaves |0 invariant, and takes |1 to –|1, with the extra factor of
−1 added known as a phase factor, so justifying the term phase flip. The Pauli gates
X, Y and Z can be obtained by appropriately setting the phase shifts α = π/2, β2 =
π, or α = π/4 and β4 = –π/2 (Hao et al., 2019).
The computation is performed step by step as follows:
Alice performs a Pauli X and Z gate to her half of the entangled pair, dependent
on the message she would like to send. She then sends her half of the entangled pair
Quantum Cinema and Quantum Computing 255
in a string of N possible symbols through a CNOT gate to Bob, who then measures
the two qubits in the Bell basis to infer the intended 2-bit classical message.
Given Alice’s measurements, Bob performs one of four operations on his half of
the EPR pair and recovers the original quantum state. If Alice performs a measure-
ment and obtains the result |00 then Bob’s system will be in the state |ψ (Nielsen and
Chuang, 2010, 27).
The secret key determines a second transformation, this time on the encrypted
message. This transformation is known as decryption, the inverse of encryption,
allowing Alice to recover the original message (Nielsen and Chuang, 2010, 641).
Quantum computing and lattices have had a close relationship ever since the work of
Regev in 2004. He showed that the unique shortest vector problem is equivalent to
bounded distance decoding the hidden subgroup problem on dihedral groups (Regev,
2004). The ordered lattices of all classical crystals are derived from three Platonic
solids: the tetrahedron, the cube and the octahedron. Most lattice theories in quantum
electrodynamics (QED) also rely on cubic lattices up to the 25-dimensional Leech
lattice, which was discussed in the rise of higher-dimensional String Theories.
The grid of the 5-dimensional space formed by the edges of 12 E± (named
epita-dodecahedron) is also actually a cubic lattice forming 24 subspaces and two
counter-movements, totaling up to 26 dimensions if we count all subspaces and
movements (Quehenberger, 2013, 2014a, 2014b, 2014c). The issue of its dimension-
ality was discussed with Gerard’t Hooft, who first counted the 12 polyhedra forming
the epita-dodecahedron and considered the epita-dodecahedron as 12-dimensional.
On realizing that they form another 12 partitions from their intersection in the centre,
he also saw 24 dimensions (in a personal conversation during the Time and Matter
conference, Venice, 2012).
The essential ideas in quantum computation require operations upon quantum stuff
as a user-defined circuit mapped onto space–time. In the space–time computer frame-
work, interactions are taking place in discrete time intervals defined by a discrete set of
space–time points, each labelling a setting-outcome pair. Any given quantum compu-
tation may be expressed as “tomographically complete” sets of quantum circuits.
The structure of setting-outcome pairs at discrete events is the computational lattice
(Hardy & Lewis, 2019, 1).
The problem with this model is, that “space–time points” are not sufficient to
describe a continuum.
Here we suggest using a 6-dimensional space–time framework for one particle
and the same six dimensions for the real world—if we are prepared to restructure the
epistemic philosophical weltbild. The latter includes time in the 4th dimension and
light as electromagnetic turbulence in the 5th dimension (Quehenberger, 2012b).
256 R. C.-Z. Quehenberger
For quantum computing, a procedure to construct new lattices also enables the
labelling of the signals belonging to particular signal constellations (Vieira et al.,
2006). Signal constellations are considered to consist of the barycentres of the regular
hyperbolic polygons with 4 g edges, denoted by F4g (Carvalho, 2001).
Hyperbolic polygons are partitions of some hyperbolic Poincaré half-plane models
imagined as the signal space, where the noise is modelled as a hyperbolic Gaussian
random variable.
Here we depict the hyperbolic space as hemispheres, defining the hypersphere
forming the 4-dimensional space in imaginary movement. These hemispheres may
also represent Boolean sets A and B as shown in Fig. 6.
The transformation matrices associated with the generators of 4g, with m =
1,..., 2 g and.
iπ
e 4g 0
C= −iπ (1)
0 e 4g
Fig. 6 Cubic Lattice of the five-dimensional space and the 6-dimensional space–time model with
Venn diagram of subsets AB visualizing de Morgans’s law (image: Quantum Cinema/Renate
Quehenberger 2021© Bildrecht, Vienna 2021)
Quantum Cinema and Quantum Computing 257
Formula (1) above poses a challenge to our imagination—how can we bring together
the term e iπ as an infinite logarithmic wavelike phenomenon with the space–time
grid?
Within the context of the design of digital communications systems in homo-
geneous spaces, in particular in hyperbolic spaces, it is necessary to establish a
systematic procedure for the construction of lattices (order of a quaternion algebra)
as the basic entity to the design of signal constellations (quotient of an order by a
proper ideal) to be employed in the transmission of digital information.
The computational lattice x = (lx, xt), composed of “wired” events is a tessel-
lation of octagons and squares. Octagons are convenient because they adjoin to form
“causal-diamond-like structures”. The two-qubit gates situated on the octagons form
a regular lattice, the gate lattice.
The choice of “causal-diamond-like structures” is the simplest construction of a
rhomb by drawing two circles with the same radius, where the centrepoint of one
circle lies on the circumference of the second. The same rhomb is forming the pattern
of the golden mean, known as rhombic Penrose tiling and the lattice for quasicrystals,
which we use for visualization of quantum phenomena.
By proposing the pentagrid as a model for a quantum lattice we also think of the
infinite 5-dimensional space with angles in 72° (Fig. 7). Its inverse number, as well
as its value behind the comma, seem to indicate a relationship to the experimental
value for the Bell parameter function that was experimentally determined to be Sexp
= 2.73± = 0.02 (Weihs et al., 1998)—named “awkward Irish angles” by John Bell
(Bertmann and Friis, 2008).
9 The Hypersphere S3
The hypersphere S3 serves as a model for the Universe as well as for many features in
quantum physics. Despite its importance, usually in literature, for the understanding
of the topology of space, the depiction of the 3-sphere (S3 ) remains to date more
or less a question mark. There is no model for the 4-dimensional space and the
hypersphere (S3 ) is declared as “unimaginable”.
For the hyper-Euclidean model of the hypersphere, which may also serve as a
representation of the earth’s complex system, we take an ordinary 3-dimensional
sphere (in mathematics denoted as S2 : a 2-sphere, because it takes only the surface
into account) and attach hemispheres on each point of its surface (Quehenberger,
Rubin et al., 2022).
So we arrive at a complex 3-dimensional representation of a sphere with a
4-dimensional “real” Euclidean “tangent space” consisting of interconnected, hemi-
spheres in imaginary movement. The model satisfies William Rowan Hamilton’s
mathematical foundations as played out in his Algebra as the Science of Pure Time
(Hamilton, 1837), which appears to have been widely neglected. In his Lectures
on Quaternions (1853), Hamilton noted that complex numbers (quaternions) √ are a
composite of real and imaginary numbers (z = x + iy ∈ C where i = −1) living
258 R. C.-Z. Quehenberger
Fig. 8 Construction of the hypersphere: Attaching hemispheres on the ball started with icosahedral
symmetry. More and more hemispheres are added until the centre of one is in the radius of its
neighbour; the hypersphere fully packed with hemispheres exhibits hexagonal hyperbolic cells on
the surface of S3 (image: Quantum Cinema/Renate Quehenberger 2012© Bildrecht, Vienna 2021)
Fig. 9 Construction Epita-dodecahedron, 3D Kites and Darts Penrose tilings with circle decoration
in motion exhibit the spherical shape of a 3-sphere. (image: Quantum Cinema/Renate Quehenberger
2012© Bildrecht, Vienna 2021)
Fourier transform, or better quantum Fourier transform, became the dominant tool
in quantum computing. Fourier transform may be used to solve the hidden subgroup
problem, a generalization of the phase estimation and order-finding problems that
has among its special cases an efficient quantum algorithm for the discrete logarithm
problem, another problem thought to be intractable on a classical computer (Nielson
and Chuang, 2011, 217).
In quantum computing the Fourier transform is the key to a general procedure
known as phase estimation—also described as black box operation—which is, in
turn, the key for many quantum algorithms.
In mathematics, the discrete Fourier transform over an arbitrary ring generalizes
the discrete Fourier transform of a function whose values are complex numbers, i.e.,
performing the transform in the integers mod N for some integer N (Schönhage &
Strassen, 1971). The algorithm uses recursive Fast Fourier transforms in rings with
Quantum Cinema and Quantum Computing 263
2n+1 elements, a specific type of number theoretic transform (Knuth, 1997, 305).
Quantum algorithms can be summarized as follows: First, reduce the question at hand
to something resembling a hidden subgroup problem, then extend the existing hidden
subgroup machinery to give a quantum algorithm for this particular variant. The key
step in solving a hidden subgroup problem is the ability to compute the quantum
Fourier transform over a group with exponentially many elements. So the relationship
between the Fourier transforms over different cyclic groups of an arbitrary fixed
superposition (Hales & Hallgren, 2000).
Suppose a unitary operator U has an eigenvector |u > with eigenvalue e2πiφ , where
the value of φ is unknown. The product representation allows the construction of an
efficient quantum circuit computing the Fourier transform and provides insight into
algorithms based upon the quantum Fourier transform (Nielson and Chuang, 2011,
217).
A classical computer requires exponentially more operations to compute the
Fourier transform than a quantum computer. The quantum circuit construction of
the quantum Fourier transform apparently requires logic gates of exponential preci-
sion in the number of qubits used. Fourier analysis involves expanding periodic
functions on the unit circle written in terms of a series of sines and cosines functions
which are referred to as harmonics. Considering higher-dimensional analogues of
the harmonics on the unit n-sphere, one arrives at the spherical harmonics—which
we imagine as logarithmic spirals around the unit sphere.
In our visualization of the 4π evolution of e (Fig. 10), the natural algorithm e is
depicted by means of logarithmic spirals emerging from the unit sphere and curling
around it. This construction is similar to the idea of fibered spaces; we coloured the
spiral on the left red, and spirals on the right blue. Uneven numbers of logarithmic
spirals mix, while even numbered spirals appear in bundles of the same colour. This
is certainly another topic that requires deeper investigation.
At least it provides a suggestion for the wave state of a quantum particle. It reminds
us of the measurement of the 4π value of a neutron by Helmut Rauch, which we
discussed with him in depth during the Quantum Cinema project.
This wave bundle figure is reminiscent of a charged-particle motion (factors of
4π/c) and Maxwell’s equations describing a curl in an electric field that implies a
change in the magnetic field equal to the current density, occurring at the speed of
light. The given distribution of electric current density J is derived from the magnetic
field B simply as (c/4πi)∇ × B (Vasyliunas, 2005). This reminds us of Schrödinger’s
dream of a visual interpretation of the matrix elements “as amplitudes of partial
oscillations of the electrical moment of the atom”, which he had to abandon (Meyenn,
2010, 204).
According to Heisenberg’s theory, the Schrödinger equation of a light quantum
has a value of the order of 1 (Heisenberg in a letter to Pauli, May 30, 1935, in Meyenn,
1985, 397).
However, we may imagine the quantum object, the ideal quantum Fourier trans-
form on n qubits, results if the controlled-Rk gates are performed to a precision
r = 1/p(n) for some polynomial p(n).
264 R. C.-Z. Quehenberger
The prime state made of n qubits that corresponds to the quantum superposition
of all prime numbers is less than 2n (we take n > 1 so that 2n is not a prime), where
each prime number can be expressed in binary form p = p0 20 + p1 21 + ... + pn−1 2
n−1
.
All the states in the sum are orthogonal and the normalization of the state is related
to the squared root of the number of primes less than 2n , namely π(2n ). In the case
√ n = 3,
of, for example,
|P3 = 1/ 4 (|2 + |3 + |5 + |7
= 1/2 (| ↑↓↑+ | ↑↓↓ + | ↓↑↓ + | ↓↓↓) where the qubits |0 and |1 are described
by the spin polarized states ↑aand ↓nof a spin 1/2 particle (Latorre & Sierra, 2013).
Applying the controlled-R2 gate produces the state of the qubit system.
At most n/2 swaps are required, and each swap can be accomplished using three
controlled-CNOT gates. Accordingly, this circuit provides a (n2 ) algorithm for
performing the quantum Fourier transform.
The ability to tune the control frequency of a qubit within the silicon chip by engi-
neering its atomic configuration is extremely important. The molecules can be created
with different resonance frequencies. This means that controlling the spin of one
qubit will not affect the spin of the neighbouring qubit, leading to fewer errors—an
essential requirement for the development of a full-scale quantum computer (Voisin,
2020).
A quantum computer can factor a number exponentially faster than the best known
classical algorithms. The most spectacular discovery in quantum computing to date
is that quantum computers can efficiently perform some tasks which are not feasible
on a classical computer.
For example, finding the prime factorization of an n-bit integer is thought to
require exp((n1/3 log2/3 n)) operations using the best classical algorithm known at
266 R. C.-Z. Quehenberger
the time of writing, the so-called number field sieve (Nielsen and Chuang, 2010,
216).
Prime numbers are central objects in Mathematics and Computer Science. They
appeared dramatically in Quantum Computation through Shor’s algorithm, which
converts the hard problem of factorization into a polynomial one using quantum
interference (Shor, 1997). The entanglement properties of the Prime state remain to
be explored in more detail.
The Prime state is highly entangled, and its entanglement measures encode number
theoretical functions, such as the distribution of twin primes.
Bernoulli numbers are characterized by the three most useful features, namely, a
recursive equation, an explicit formula and a generating function. They were first
applied to computer science in the first conception for the first published complex
computer program by Ada Lovelace in note G on the Analytical Engine, where she
describes an algorithm for generating Bernoulli numbers with Babbage’s machine
(Lovelace, 1842). Bernoulli factory algorithms sample the probability f (λ). There is
an analogy to the Bernoulli factory problem called the quantum Bernoulli factory,
with the same goal of simulating functions of unknown probabilities. For every odd
n > 1, Bn = 0. For every even n > 0, Bn is negative if n is divisible by 4 and positive
otherwise.
Bernoulli numbers can also be expressed in terms of the famous Riemann zeta
(ξe) function, an extension of Euler’s zeta (ξ) function.
The Swiss mathematician Leonard Euler made pioneering and influential discov-
eries in many other branches of mathematics, such as analytic number theory,
complex analysis, and infinitesimal calculus. He was invited to Russia by Empress
Catherine II thanks to his good relations with the Bernoulli brothers and succeeded
Daniel Bernoulli as professor of mathematics at the Academy of St. Petersburg in
1733. Euler discovered a formula relating ξ(2 k) to the Bernoulli numbers, yielding
results such as ξ(2) = π2 /6 and ξ(4) = π4 /90. The values of the zeta function are
expressed as powers of π and depend on these Bernoulli numbers.
Bernhard Riemann noted that his zeta function had trivial zeros at −2, −4, −6, …
and that all non-trivial zeros were symmetric about the line Re(s) = 1/2. He calculated
the first six non-trivial zeros of the function and observed that they were all on the
same straight line. In a report published in 1859, Riemann stated that this might very
well be a general fact (Riemann, 1859).
In 1900, David Hilbert listed proving or disproving the Riemann hypothesis as
one of the most important unsolved problems confronting modern mathematics.
Since then numerous attempts to solve this conjecture have been published. It
is central to understanding the overall distribution of the primes. The Hilbert-Pólya
Conjecture proposes a spectral interpretation of the non-trivial zeros of the Riemann
zeta function. Hilbert called the set of eigenvalues of an operator the “spectrum”, by
Quantum Cinema and Quantum Computing 267
analogy with the spectral lines produced by the radiation frequencies of atoms. Prime
numbers are also related to the harmonic mean as formulated by the Pythagoreans.
For example, in 1901, von Koch showed that the Riemann hypothesis was equiv-
alent to the logarithmic expression similar to the one Euler found (von Koch,
1901).
or(x) = Li (x) + 0 (x1/2 ln x)
Jacques Hadamard and de la Vallee Poussin proved the prime number theorem,
by showing that. √
ad(x) = Li (x) + 0 (xe -a ln x )
for some positive constant a, and they did this by bounding the real part of the zeros
in the critical strip away from 0 and 1. The error term is directly dependent on what
was known about the zero-free region within the critical strip.
The function f satisfies the equation f (1 − s) = f(s) for the complex argument
s ≡ σ + iτ. Or, as the British mathematician Godfrey Harold Hardy put it, ξ (1/2 +
it) and proved that the Riemann zeta function has infinitely many real zeros. Hardy
also showed that infinite many non-trivial zeros lie on the critical line Re(s) = 1/2
(Hardy, 1914).
The Austro-Hungarian mathematician Marcel Riesz showed in 1916 that
the Riemann hypothesis was equivalent to the claim that it is true for any e larger
than 1/4 if we had, x tending to infinity on the positive real axis (Riesz, 1916). Also,
all points where f vanishes are located on the critical line σ = 1/2 and all lines of
constant phase of f corresponding to ±π, ±2π, ±3π… merge with the critical line
(Schleich et al., 2018). This π-periodicity is visualized in Fig. 11, where after every
π -turn the prime numbers appear on one line with the most beautiful pattern of
“entangled” prime numbers on the y-axis visualized by small blue circles.
It seems to be a visual confirmation of Godfrey H. Hardy and John E. Littlewood’s
Twin Prime Conjectures, (Hardy & Littlewood, 1921). An analytical explanation of
these periodical structures obtained is an open problem worthy of investigation by
the experts in the analytical number theory (Machadoa & Luchkob, 2021).
In our visualization, the numbers 1, 2, 3, 4 are lying inside the radius of the sphere,
while number 5 is already the first number on the spiral arm crossing the x-axis above
the north pole of the sphere, representing the imaginary part of the system just like
the red and blue bundles of lines in Fig. 10. By placing all numbers consecutively on
the ±x and ±y-axis, we arrive at all even numbers on the x-axis and all odd numbers
on the y-axis. Furthermore, we place the imaginary numbers i on a logarithmic spiral
originating on the surface of the unit sphere at point 1/2 of a Cartesian coordinate
system, which is the half diameter of the unit sphere that represents the real numbers.
This implies that in our visualization, the so-called “critical line” with non-trivial
zeros is a geodesic on the real number space unit sphere and the imaginary numbers
on the logarithmic spiral form the 4-dimensional space around a central unit sphere
which can be identified as Hamiltonian/quaternionic/Lorentzian space–time.
For quantum computation, Jose I. Latorre and German Sierra showed how to
construct the quantum prime state, a quantum sequence state based on the sequence
of prime numbers with an error below the bound that allows for verification of
268 R. C.-Z. Quehenberger
Fig. 11 The prime numbers in a logarithmic 4πi solution, all prime numbers on the y-axis, produce
a series of twin prime numbers which are depicted by small blue circles, seemingly “entangled”
(image: R. Quehenberger 2012© Bildrecht, Vienna 2021)
the Riemann hypothesis. This algorithm can be further combined with the quantum
Fourier transform to yield an estimate of the prime counting function more efficiently
than any classical algorithm (Latorre & Sierra, 2013). They assume it is likely that
the quantum correlations emerging from the Prime state are profoundly related to
theorems in Number Theory. Their claim is simple: All non-trivial zeros have a real
part of 1/2.
With the help of computers, more than ten billion zeros of the zeta function lying
on the critical, clearly highlighting Riemann’s extraordinary intuition, have been
calculated to date (Bayer, 2018). However, calculations ad infinitum do not provide
sufficient proof (van de Lune et al., 1986).
Finally, we call for the reestablishment of the authority of geometry for proving
algebraic operations so that we could claim the million dollars—at least for the
Poincaré conjecture since, despite many claims to have done so, Henri Poincaré’s
question has never really been answered in the way we have shown here.
Quantum Cinema and Quantum Computing 269
References
Bohm, D. (1982b). The holographic paradigm and other paradoxes, exploring the leading edge of
science. In: K. Wilber (Ed.). Shambala.
Bohm, D. (2004). On creativity. In: L. Nichol (Ed.). Routledge.
Bohm, D., & Hiley, B. (1993). The undivided universe. Routledge.
Bernal, R. A. (2005). De Morgan’s laws revisited: To be AND/OR NOT to be, paper PO25,
PharmaSUG.
Bertlmann, R. A., & Friis, N. (2008). Theoretical physics T2 quantum mechanics course of lectures,
T2–script of Sommersemester 2008 by Reinhold A. Bertlmann.
Bertlmann, R., & Zeilinger, A. (Eds.). (2017). Quantum [Un]Speakables II. Springer International
Publishing Switzerland.
Brezina, C. (2006). Al-Khwarizmi: The inventor of algebra. The Rosen Publishing Group.
Brillouin, L. (1956). Science and information theory. Academic Press.
Brody, D. C., & Gräfe, E.-M. (2011). Six-dimensional space-time from quaternionic quantum
mechanics. Physical Review D, 84, 125016. arXiv:1105.3604
de Broglie, L. (1925). Recherches sur la Théorie des Quanta, Thesis 1924. Annales de Physique,
10e série, t. III.
de Broglie, L. (1927). L’univers à cinq dimensions et la mécanique ondulatoire. Journal De Physique
Et Le Radium, 8, 66–73.
de Broglie, L. (1943). Die Elementarteilchen, Individualität und Wechselwirkung, H Coverts; orig.
(1941). Continu et discontinu en physique moderne (Continuous and discontinuous in Modern
Physics), Paris: Albin Michel.
Bromley, A. G. (1982). Charles Babbage’s analytical engine, 1838. Annals, 4(3).
Brouwer, L. E. J. (1913). Über den natürlichen Dimensionsbegriff. Reine Und Angewandte
Mathematik, J, 42, 146–215.
Cabello, A. (2017). Interpretations of quantum theory: A map of madness. In: O. Lombardi, S.
Fortin, F. Holik, C. López (Eds.), What is quantum information? Cambridge University Press.
Cavalcante, R. G., Lazari, H., Lima, J. D., & Palazzo, R. Jr. (2005). A new approach to the design of
digital communication systems. In: A. Ashikhimin & A. Barg (Eds.), Discrete mathematics and
theoretical computer science (Vol. 68, pp. 145–177). DIMACS Series. American Mathematical
Society.
Carvalho, E. D. (2001). Construction and labeling of geometrically uniform signal constellations
in Euclidean and hyperbolic spaces. Doctoral Dissertation, FEEC-UNICAMP.
Capra, F. (1982). The turning point. Simon and Schuster.
Clarke, B. (2009). Heinz von Foerster’s Demons. In: B. Clarke, M. B. N. Hansen (Eds.), Emergence
and embodiment: New essays on second-order systems theory. Duke University Press.
Dakić, B., & Brukner, Č. (2016). The classical limit of a physical theory and the dimensionality of
space. Fundamental Theory Physics, 181, 249–282.
Deleuze, G. (1989). Part 2: Cinema 2, University of Minnesota Press. Here quoted in German from:
Transl. Klaus Englert, (1991) Das Zeit-Bild, Kino 2, Frankfurt/Main: Suhrkamp.
Descartes, R. ([1637] 1902). Discours de la Méthode. Adam et Tannery.
Dirac, P. A. M. (1930). The principles of quantum mechanics. Clarendon Press.
de Bruijn, N. G. (1981). Algebraic Theory of Penrose’s Non-periodic Tilings of the Plane I and II,
Proc Koninklijke Nederlandse Akademie van Wetenschapen, 84, 39–52 and 53–66.
Einstein, A. (1916). Die Grundlage der allgemeine Relativitätstheorie. Annalen Der Physik, 49,
769–822.
Einstein, A., Podolsky, B., & Rosen, N. (1935). Can quantum-mechanical description of physical
reality be considered complete? Physical Review, 47, 777–780.
Faulhaber, J. (1631). Academia Algebrae, darinnen di miraculo’sische, Inventiones zu den höchsten
Costen weiters continuirt u. profitiert werden, Augsburg: Johann Remelins.
Feynman, R. (1982). Simulating physics with computers. International Journal of Theoretical
Physics, 21, 6/7, 467–488.
Fourier, J. B. J. (1822). Theorie analytique de la chaleur. Firmin Didot.
Quantum Cinema and Quantum Computing 271
van Fraassen, B. (1972). A formal approach to the philosophy of science in paradigms and paradoxes:
The philosophical challenge of the quantum domain. In R. G. Colodny (Ed.), Paradigms and
paradoxes: The philosophical challenge of the quantum domain (pp. 303–366). University of
Pittsburgh Press.
Fuchs, L. (1880). Über die Verallgemeinerung des Kehrungsproblems. Journal De Crelle, 89, 151–
169.
Gardner, M. (1989). Penrose tiles to trapdoor ciphers. H. Freeman & Co.
Greenberger, D. M., Horne, M. A., & Zeilinger, A. (1993). Multiparticle interferometry and the
superposition. Physics Today, 46(8), 22.
Grünbaum, B., & Shephard, G. C. (1987). Tilings and patterns. W. H. Freeman and Company.
Grössing, G. (2003). Zum Bilderstreit in der Quantentheorie, http://phaidon.philo.at/~nulleins/arc
hiv/archiv.html
Hales, L., & Hallgren, S. (2000) An improved quantum Fourier transform algorithm and appli-
cations. In: Proceedings of the 41st Annual Symposium on Foundations of Computer Science
(pp. 45–46).
Hao, L., Wang, Z., Yang, J., et al. (2019). Anomalous magnetoresistance due to longitudinal spin
fluctuations in a Jeff = 1/2 Mott semiconductor. Nature Communications, 10, 5301. https://doi.
org/10.1038/s41467-019-13271-6
Hardy, G. H. (1914). Sur les zeros de la fonction zeta ζ(s). Comptes Rendus De L’académie Des
Sciences, 158, 1012–1014.
Hardy, G. H., & Littlewood, J. E. (1921). The zeros of Riemann’s Zeta-function on the critical line.
Mathematische Zeitschrift, 10(3–4), 283–317. https://doi.org/10.1007/bf01211614
Hardy, L., & Lewis, A. G. M. (2019) Quantum computation with machine-learning-controlled
quantum stuff. arXiv:1911.13282v1
Hermann, G. (1926). Die Frage der endlich vielen Schritte in der Theorie der Polynomideale.
Mathematische Annalen, 95, 736–788.
Hermann, G. (1935). Die naturphilosophischen Grundlagen der Quantenmechanik. Abhandlungen
Der Fries’schen Schule, 6(2), 69–152.
Heisenberg, W. (1927) Über den anschaulichen Inhalt der quantentheoretischen Kinematik und
Mechanik (On the Perceptual Content of Quantum Theoretical Kinematics and Mechanics),
Copenhagen, Institut for theoretical Physics at the Kopenhagen University, March 23 and
Zeitschrift für Physik 43, 172–198.
Heisenberg, W. (1989). Encounters with Einstein: And other essays on people, places, and particles.
Princeton University Press.
Heisenberg, W. (2011). Quantenphysik und Philosophie, Stuttgart: Philip Reclam jun. (reprint from
1979).
Herzenberg, C. L. (2008) Grete Hermann: An early contributor to quantum theory.
arxiv.org/pdf/0812.3986.
Hiley, B. J., & Callaghan, R. E. (2010). The Clifford algebra approach to quantum mechanics B:
The Dirac particle and its relation to the Bohm approach. arXiv:1011.4033v1
Hiley, B. J.(2016). Structure process, weak values and local momentum. Journal of Physics:
Conference Series, 701(1). Article id. 012010
Hopf, H. (1931). Über¯ die Abbildungen der dreidimensionalen Sphäre auf die Kugelfäche,
Mathematische, Annalen, 104(1), 637–665.
Holmes, R. (2015). Computer Science: Enchantress of abstraction. Nature, 525, 30–32.
Toole, B. A. (2010) Ada, the enchantress of numbers: Poetical science. Critical Connection.
Jammer, M. (1974). The philosophy of quantum mechanics: The interpretations of quantum
mechanics in historical perspective. Wiley.
Jevtic, S., & Rudolph, T. (2015). How Einstein and/or Schrödinger should have discovered Bell’s
theorem in 1936. Journal of the Optical Society of America B - Optical Physics, 32, A50-A55.
ISSN 0740-3224
Jarry, A. ([1911] 1996). Exploits and Opinions of Dr. Faustroll, Pataphysician, Boston: Exact
Change.
272 R. C.-Z. Quehenberger
Jones, W. (1794). On the philosophy of the Asiatics. Asiatic Researches, IV, 164.
Jones, W. (1794). On the philosophy of the Asiatics. Asiatic Researches, IV, 164.
Klein, F. (1872). Vergleichende Betrachtungen über neuere geometrische Forschungen, Programm
zum Eintritt in die philosophische Fakultät und den Senat der Friedrich-Alexander-Universität
zu Erlangen, Erlangen: Deichert.
Klein, F. (1892). Vorlesungen über Nicht-Euklidische Geometrie. In Klein, Felix (1928)
Vorlesungen über Nicht-Euklidische Geometrie, Rosemann, W. (Ed.) Berlin: Springer.
Klyachko, A., Can, M. A., Binicioğlu, S., & Shumovsky, A. S. (2008). Simple test for hidden
variables in spin-1 systems. Physical Review Letters, 101, 020403.
Knuth, D. E. (1997). The art of computer programming, volume 2: Seminumerical algorithms, 3rd
edn. Addison-Wesley Professional.
von Koch, H. (1901). Sur la distribution des nombres premiers. Acta Mathematica, 24, 159–182.
Kochen, S., & Specker, E. P. (1967). The problem of hidden variables in quantum mechanics.
Journal of Mathematics and Mechanics, 17(1967), 59–87.
Kunjwal, R., Spekkens, R. W. (2015). From the Kochen-Specker theorem to noncontextuality
iinequalities without assuming determinism, submitted to QPL. www.cs.ox.ac.uk/qpl2015/pre
proceedings/50.pdf
Kwiat, P. G., Mattle, K., Weinfurter, H., Zeilinger, A., Sergienko, A. V., & Shih, Y. H. (1995).
New high-intensity source of polarization-entangled photon pairs. Physical Review Letters, 75,
4337–4341.
Latorre, J. I., & Sierra, G. (2013) Quantum computation of prime number functions. Quantum
Information & Computation, 14(7). arXiv:1302.6245v3
Lenzen, V. F. (1969). Concepts and reality in quantum mechanics. Philosophy of Science, 16(4),
279–286.
Leonardo da Vinci. (1877). Treatise on painting (J. F. Rigaud, Trans.) George Bell & Sons.
Lovelace, A. (1842). Notes to “A Sketch of the Analytical Engine invented by Charles Babbage by LF
Menabrea”, translated with a detailed commentary Ada Augusta, Countess of Lovelace, Biblio-
thèque Universelle de Genève, No. 82, October 1842. In: R. Taylor, (Ed.), Scientific memoirs,
selections from transactions of foreign academies and learned societies and from foreign journals
(Vol. 3, Article XXIX). FSA (1843).
van de Lune, J., te Riele, H. J. J., & Winter, D. T. (1986). On the zeros of the Riemann zeta function
in the critical strip, iv. Mathematics of Computation, 46(174), 667–681.
Lück, R. (2000). Dürer-Kepler-Penrose, the development of pentagon tilings. Materials Science
and Engineering A, 294–296.
Machadoa, T., & Luchkob, Y. (2021). Multidimensional scaling and visualization of patterns in
distribution of nontrivial zeros of the zeta-function. Communications in Nonlinear Science and
Numerical Simulation, 102, 105924.
Mackay, A. L. (1990). Crystals and fivefold symmetry. In: I. Hargittai (Ed.), Quasicrystals, networks,
and molecules of fivefold symmetry. VCH.
Mackay, A. L. (1981). De nive Quinquangula — On the Pentagonal Snowflake, Kristallogafiya 26,
1981, 910–919. Soviet Physics Crystallography, 26(1981), 517–522.
Mandolesi, A. L. G. (2019). Quantum fractionalism: The Born rule as a consequence of the complex
Pythagorean theorem. 1905.08429. Physics Letters A, 384(28), 126725.
Mattle, K., Weinfurter, H., Kwiat, P. G., & Zeilinger, A. (1996). Dense Coding in Experimental
Quantum Communication, Physical Review Letters, 76, 4656.
Mano, M. M. (1988). Computer engineering: Hardware design. Prentice Hall.
Mermin, N. D. (1990). Simple unified form for the major no-hidden-variables theorems. Physics
Review Letters, 65(27), 3373–3376.
Mermin, N. D. (1993). Hidden variables and the two theorems of John Bell. Reviews of Modern
Physics, 65, 803–815.
von Meyenn, K. (1985) Wolfgang Pauli: Wissenschaftlicher Briefwechsel mit Bohr, Einstein,
Heisenberg u.a. Band II: 1930–1939. Springer.
Quantum Cinema and Quantum Computing 273
von Meyenn, K. (2010). Eine Entdeckung von ganz außerordentlicher Tragweite: Schrödingers
Briefwechsel zur Wellenmechanik und zum Katzenparadoxon. Springer.
Miller, A. I. (1996). Insights of genius, imagery and creativity in science and art. Springer.
Milnor, J. (2000). The Poincaré conjecture. https://www.claymath.org/sites/default/files/poincare.
pdf
Milnor, J. (2003). Towards the Poincaré conjecture and the classification of 3-manifolds. Notices
of the AMS, 50(50), 1226–1233.
Musk, E. (1990) Neuralink, An integrated brain-machine interface platform with thousands of
channels. bioRxiv: 703801.https://doi.org/10.1101/70380
von Neumann, J. (1932). Mathematische Grundlagen der Quantenmechanik. Julius Spinger.
Nielsen, M. A., & Chuang, I. L. (2010). Quantum computation and quantum information, 10th
Anniversary Edition, CUP. Cambridge University Press.
Oriti, D. (2006). Disappearance and emergence of space and time in quantum gravity. In: D. Rickles
(Ed.), Studies in history and philosophy of modern physics. Special issue “Quantum Gravity”.
arXiv:1302.2849.
Padua, S. (2015). The thrilling adventures of Lovelace and Babbage—The (Mostly) true story of
the first computer. Pantheon.
Pauli, W. (1955). Die Wissenschaft und das abendländische Denken, Wolfgang Pauli Nachlass
im CERN-ARCH-PMC-06–291, WB 4/IVA, letter [2707]. http://cds.cern.ch/record/96387/files/
CERN-ARCH-PMC-06-291.pdf?version=1.
Pauka, J., et al. (2021). A cryogenic CMOS chip for generating control signals for multiple qubits.
Nature Electronics, 4, 64–70. https://doi.org/10.1038/s41928-020-00528-y
Penrose, R. (1974). The role of aesthetics in pure and applied mathematical research, southend-on-
sea. Bulletin of Mathematical Analysis and Applications, 10, 266–271.
Plato. (2003). Timaios. Reclam.
Poincaré, H. (1882). Theorie des Groupes Fuchsiens. Acta Mathematica, 1, 1–63.
Poincaré, H. (1895). Analysis Situs, Journal de l’École Polytechnique, 11e Série—Premier Cahier,
Paris. (Transl. by J. Stillwell: Papers on Topology, Analysis Situs and Its Five Supplements, Henri
Poincaré, 2009).
Poincaré, H. (1892). Sur l’Analysis situs, note de Henri Poincare. Comptes Rendus Des Seances
De L’academie Des Sciences, 115, 633–636.
Poincaré, H. (1900). Second supplement to analysis situs. Proceedings of the London Mathematical
Society, 32, 277–308.
Poincaré, H. (1904). Cinquième complément à l’analysis situs, Rend. Circ. Mat.: Palermo, 18,
45–110.
Poincaré, H. (1912). Pourquoi l’espace á trois dimensions. Revue De Métaphysique Et De Morale,
4, 483–504.
Polchinski, J. (1995). Dirichlet-Branes and Ramond-Ramond charges. arxiv.org/abs/hep-
th/9510017.
Pool, R. (2020). A new kind of quantum. https://spie.org/news/photonics-focus/novdec-2020/a-
new-kind-of-quantum?SSO=1.
Price, H. (2013). Taming the quantum spooks. https://aeon.co/essays/can-retrocausality-solve-the-
puzzle-of-action-at-a-distance.
Pusey, M. F., Barrett, J., & Rudolph, T. (2012). On the reality of the quantum state. arXiv:1111.
3328
Quehenberger, R. (2011). QC expert talk with Gerard ‘t Hooft (Utrecht University) 12 November
2011, Vienna. http://vimeo.com/34081638
Quehenberger, R., Weibel, P., Rauch, H., Katzgraber, H., & Friemel, R. (2012). A new digital
3-D dynamic geometry for the visualization of complex number space. In: Proceedings of
International Conference on Computational Electronics and Nanotechnology (ICOCENT-2012)
Conference, March 2012. Jaipur (pp. 287–292).
Quehenberger, R. (2012a). Quantum cinema expert-talk with Basil Hiley and Helmut Rauch.
vimeo.com/50374353.
274 R. C.-Z. Quehenberger
Quehenberger, R., & Weibel, P. (2012). Recording of “Quantum Fluxus Dada Duett“ with voice
of Erwin Schrödinger (recording “Do electrons think?” by courtesy of Ruth Braunitzer) Live
stream with Zabelka, M. (violin) as ARTs Birthday present 2012, Ö1 Kunstradio. http://www.
kunstradio.at/2012A/22_01_12.html
Quehenberger, R. (2012b). A reflection on theories of light. In: A. Khrennikov, H. Atmanspacher,
A. Migdall, & S. Polyakov (Eds.), Quantum theory: Reconsideration of foundations 6. AIP
Conference Proceedings 1508 (pp. 459–463).
Quehenberger, R. (2013). Quantum information traced back to ancient Egyptian mysteries.
Technoetic Arts: A Journal of Speculative Research, 11(3), 319–334.
Quehenberger, R. (2014a). A newly found golden heptahedron named epitahedron. In: G. Darvas
(Ed.) Journal Symmetry: Culture and Science. Symmetrion no. 25, 177–192.
Quehenberger, R. (2014b). A new visualization of the homology sphere. ICM 2014. In: International
Congress of Mathematicians, Seoul (KR) Poster presentation, P5 #29.
Quehenberger, R. (2014c). Epitadodecahedron, film presented at IMAGINARY at ICM 2014 &
International Conference of Mathematicians, ICM Seoul 2014. www.imaginary.org/film/the-
epita-dodecahedron-visualizing-poincares-dodecahedral-space
Quehenberger, R. (2016). Description of the cover-images: The Epitahedron. In: G. Darvas (Ed.),
Symmetry: Culture and science (Vol. 27(1), pp. 56–58). Symmetrion.
Quehenberger, R. (2017) A Proposal for a Psi-ontological model based on 5-dimensional
geometry. In: QCQMB Workshop: Quantum Contextuality in Quantum Mechanics and
Beyond, Prague 2017, http://www.psych.purdue.edu/~ehtibar/workshop/files/Poster_prag_Q
CQMB%202017_RCZQ_A0_corr.png
Quehenberger, R. (2018). A quest for an epistemic reset in higher dimensional space, Entropy, Best
Poster Awards at Linnaeus Conference: Towards Ultimate Quantum Theory (UQT). https://res.
mdpi.com/data/1st-place-r.c.-z.-quehenberger.pdf
Quehenberger, R. (2019) Enfolding the 5th dimension: With origami to the 3D representation of the
Penrose kites & darts pattern, symmetry: art and science, 2019. In: 11th Congress and Exhibition
of SIS Special Theme: Tradition and Innovation in Symmetry, Katachi 形 Kanazawa, Japan,
November 25–30, 2019.
Quehenberger, R., Kern L., & Tirelli, D. dC. Rubin, S. (2022). GAIA 5.0—A five-dimensional
geometry for the 3D visualization of Earth’ climate complexity. In: T. Lanza (Ed.) S. Illing-
worth et al., (Eds.) Special issue: Five years of earth sciences and art at the EGU (2015–2019).
Geoscience Communication, Copernicus, 2020–2021.
Quehenberger, R. (2021). Interactive comment on GAIA 5.0—A five-dimensional geometry for the
3D visualization of Earth’ climate complexity. Geoscience Communication Discussion. https://
doi.org/10.5194/gc-2020-27-AC3
Quehenberger, R. (2022, forthcoming) Kepler & die 5. Dimension, (Transl.: Kepler & the 5th
Dimension, forthcoming) Aichbergiana, Handreiche zum Alltag, No. 27.
Riemann, G. F. B. (1859). Über die Anzahl der Primzahlen unter einer gegebenen Grösse.
Monatsberichte der Berliner Akademie, 671–680.
Rauch, H., Zeilinger, A., Badurek, G., Wilfing, A., Bauspiess, W., & Bonse, U. (1975). Verification
of coherent spinor rotation of fermions. Physics Letters A, 54(6), 425–427.
Regev, O. (2004). Quantum computation and lattice problems. SIAM Journal on Computing, 33(3),
738–760.
Riesz, M. (1916). Sur l’hypothèse de Riemann. Acta Mathematica, 40, 185–190.
Rudolph, T. (2017). Q is for quantum. Kindle Edition by Terry Rudolph.
Rupprich, H. (Ed.) (1969). Dürer. Schriftlicher Nachlass, Berlin: Deutscher Verlag für Kunst-
wissenschaft, 1956/1966/1969, 3rd vol. 1969.
Schleich, W. P., Bezděková, I., Kim, M. B., Abbott, P. C., Maier, H., Montgomery, H. L., &
Neuberger, J. W. (2018). Equivalent formulations of the Riemann hypothesis based on lines
of constant phase. Physica Scripta, 93(6), [065201]. https://doi.org/10.1088/1402-4896/aabca9
Quantum Cinema and Quantum Computing 275
Schopenhauer, A. (1819). Die Welt als Wille und Vorstellung (Engl.: The World as Will and Repre-
sentation), in: Arthur Schopenhauers sämtliche Werke in sechs Bänden (Ed. Eduard Grisebach),
Verlag von Philipp Reclam jun., 1892.
Schönhage, A., & Strassen, V. (1971). Schnelle Multiplikation grosser Zahlen. Computing, 7, 281–
292.
Schrödinger, E. (1935). Die gegenwärtige Situation in der Quantenmechanik, Naturwissenschaften,
23:48, 807–812, 823–828; 844–849.
Schrödinger in a letter to Born, 14 April 1926, in Meyenn, 2010, p. 201) “... sich irgendwie mit der
Atommechanik von Heisenberg, Jordan und mir vereinigen lassen. Ich kann mir natürlich unge-
fähr denken, wie das geht, denn partielle Differentialgleichungen sind ganz allgemein äquivalent
mit quadratischen Formen von unendlich vielen Variabeln; aber ich wüsste doch gern genauer,
wie es zusammenhängt und wäre Ihnen sehr dankbar, wenn...”
Shannon, C. E (1948). A Mathematical Theory of Communication, 27, 379–423 & 623–656.
Shechtman, D., Blech, I., Gratias, D., & Cahn, J. W. (1984). Metallic phase with long-range
orientaional order and no translational symmetry. Physical Review Letters, 53(20), 1951–1953.
Shor, P. W. (1997). Polynomial-time algorithms for prime factorization and discrete logarithms on
a quantum computer. SIAM Journal on Computing, 26(5), 1484–1509.
Soiguine, A. (2017). Quantum computing in geometric algebra formalism: Light Beam Guide
Implementation Future Technologies Conference (FTC) 29–30 November 2017, Vancouver,
Canada
Svozil, K. (2000). The information interpretation of quantum mechanics. arxiv.org/abs/quant-
ph/0006033v2.
Threlfall, W. & Seifert, H. (1931).Topologische Untersuchung der Diskontinuitätsbereiche endlicher
Bewegungsgruppen des dreidimensionalen sphärischen Raumes, Mathematische Annalen 104(1),
1–70.
Thurston, W. P. ([1980] 2002). The geometry and topology of three-manifolds, notes distributed
by Princeton University, Electronic version 1.1, March 2002. http://www.msri.org/publications/
books/gt3m/
Toole, B. A. (2010). Ada, the Enchantress of Numbers: Poetical Science, Sausalito CA: Critical
Connection.
Tydecks, W. (2020). Re-entry nach Spencer-Brown, presentation at formlabor, Berlin, 31 January
2020.
Uchida, G., Bertlmann, R. A., & Hiesmayr, B. C. (2015) Entangled entanglement: The geometry
of GHZ states. Physics Letters A, 379, 2698, arxiv.org/abs/1410.7145.
Vasyliunas, V. M. (2005). Relation between magnetic fields and electric currents in plasmas. Annales
Geophysicae, 23(7), 2589–2597.
Vieira, L. V., Palazzo, R., & Faria, M. B. (2006). On the arithmetic Fuchsian groups derived from
quaternion orders VI. In: International Telecommunications Symposium (ITS2006). https://doi.
org/10.1109/ITS.2006.4433342
Verhulst, F. (2012). Henri Poincaré: Impatient Genius, Springer US.
Voisin, B., et al. (2020). Valley interference and spin exchange at the atomic scale in silicon. Nature
Communications. https://doi.org/10.1038/s41467-020-19835-1
Walleczek, J., Grössing, G., Pylkkänen, P., & Hiley, B. (2019). Emergent quantum mechanics:
David Bohm centennial perspectives. Entropy, 21(2), 113. https://doi.org/10.3390/e21020113
Weibel, P. (2005). Time slot—Geschichte und Zukunft der apparativen Wahrnehmung vom
Phenakistiskop bis zum Quantenkino, Cologne: Walther König.
Weihs, G., Jennewein, T., Simon, C., Weinfurter, H., & Zeilinger, A. (1998). Violation of Bell’s
inequality under strict Einstein locality conditions. Physical Review Letters, 81, 5039–5043.
Wheeler, J. A. (1983) Law without law. In: J. A. Wheeler & W. H. Zurek (Eds.). Quantum theory
and measurement. Princeton University Press.
Wheeler, J. A. (1989). Information, physics, quantum: the search for links. In: Proceedings III
International Symposium on Foundations of Quantum Mechanics (pp. 354–368). Tokyo.
276 R. C.-Z. Quehenberger
Wilczek, F. (2012). Quantum time crystals. Physical Review Letters, 109(16): 160401. arXiv:1202.
2539
Wüthrich, C. (2014). Raiders of the lost spacetime. In: D. Lehmkuhl (Ed.), Towards a theory of
spacetime theories. Birkhäuser. arXiv:1405.5552v1
Yau, S.-T., & Nadis, S. (2010). The shape of inner space, string theory and the geometry of the
universe’s hidden dimensions. Basic Books.
Zeilinger, A. (1999). A foundational principle for quantum mechanics. Foundations of Physics, 29,
631–643.
Zeilinger, A., Bernstein, H. J., & Horne, M. A. (1994). Information transfer with two-state, two-
particle quantum systems. Journal of Modern Optics, 41, 2375–2381.
Zimba, J., & Penrose, R. (1993). On Bell non-locality without probabilities: More curious geometry.
Studies in History and Philosophy of Science. SHPSB524
How to Make Qubits Speak
Abstract This is a story about making quantum computers speak, and doing so
in a quantum-native, compositional and meaning-aware manner. Recently we did
question-answering with an actual quantum computer. We explain what we did, stress
that this was all done in terms of pictures, and provide many pointers to the related
literature. In fact, besides natural language, many other things can be implemented
in a quantum-native, compositional and meaning-aware manner, and we provide
the reader with some indications of that broader pictorial landscape, including our
account on the notion of compositionality. We also provide some guidance for the
actual execution, so that the reader can give it a go as well.
1 Networks of Words
What is a word?
In particular, what is the meaning of a word? This could be what you find in a
dictionary. Then again, before there were dictionaries people also used words, and
so did people that could not read. In fact, most words we know we didn’t look up
in a dictionary. We probably learned them by hearing them used a lot. This idea of
meaning is closer to how machines learn meanings of words today: they learn them
from the context in which they are used, an idea coined by Wittgenstein as “meaning
is use”, and expressed by Firth’s derived dictum (Firth, 1957):
“You shall know a word by the company it keeps”.
What is a sentence?
In particular, what is the meaning of a sentence? A sentence is not just a “bag of
words”,1 but rather, a kind of network in which words interact in a particular fashion.
In fact, in a very particular fashion, given that when we hear a sentence that we never
heard before, provided we do understand the words that occur in it, then we surely
also understand that sentence. That’s exactly why we don’t have dictionaries for
sentences—besides the fact a ‘sentence dictionary’ would take up an entire library,
if not more. There is an important academic question here:
• How do we deduce the meaning of a sentence given the meanings of its words?
We can also ask the converse question:
• Can we infer the meanings of words in sentences, from the meanings of sentences?
Both turn out to also be essential practical questions.
Some 10 years ago, in (Coecke et al., 2010), BC Mehrnoosh Sadrzadeh and Steve
Clark started to draw networks in order to address the first of these questions. These
networks look like this:
that
Alice likes (the) flowers Bob gives Claire
(1)
The usual technical term for these networks is “string diagram” (see e.g. Baez & Stay
2011). These string diagrams provided a straight-forward manner for how meanings
of words combine in order to produce the meaning of a sentence, even resulting in a
cover-heading feature in New Scientist.2
In order to better see how these string diagrams do so, let’s consider a simpler
example:
Alice hates Bob
(2)
The idea here is that the boxes represent the meanings of words and that the wires
are channels through which these meanings can be transmitted. So, in the example
above, the subject Alice and the object Bob are both sent to the verb hates, and
1 Here, ‘bag’ is a widely used technical term, referring to the fact that in many ‘natural language
processing’ applications grammatical structure has been entirely ignored for many years (Harris,
1954).
2 https://www.newscientist.com/article/mg20827903-200-quantum-links-let-computers-
understand-language/.
How to Make Qubits Speak 279
together they then make up the meaning of the sentence. This idea scales to much
bigger sentences like the one depicted above, and even to large text made up of
multiple sentences (Coecke, 2019).
This flow of words in sentences can be traced back to work originally started in the
1930s by Adjukiewicz (1935) and later by Bar-Hillel (1953), Chomsky (1956, 1957)
and Lambek (1958, 2008), among others. What they did was unifying grammatical
structures across different languages as a single mathematical structure. An overview
of these developments from the perspective of string diagrams is in Coecke (2013).
2 Language Is Quantum-Native
One particularly interesting aspect of this graphical framework for linguistics was
that the string diagrams were inherited from previous work that provided quantum
theory with a network-like language (Coecke, 2005). This work was accumulated
eventually in a 900-page book (Coecke & Kissinger, 2017).
In summary, a direct correspondence was established between, on the one hand,
the meanings of words and quantum states, and on the other hand, grammatical
structures and quantum measurements. This is illustrated in Fig. 1.
Obviously this led to the question: Can one make quantum computers handle
natural language? This was first proposed in a paper by Will Zeng and BC in 2016
(Zeng & Coecke, 2016), creating a new paradigm for natural language processing
(NLP) in a quantum computing context.
The idea of making quantum computers process natural language is just not only
incredibly cool but also a very natural thing to do for the reasons indicated above.
More researchers started to take an interest, and at some point Intel supported a basic
attempt to develop some of the ideas contained in (Zeng & Coecke, 2016) on their
quantum simulator (Intel News Release, 2019; O’Riordan et al., 2020).
who
Quantum measurement
Bell-tests
Fig. 1 Illustration of how word meanings can be interpreted as quantum states, and grammatical
structure in terms of quantum measurement
280 B. Coecke et al.
However, there were some significant challenges to the proposal. Most impor-
tantly, there weren’t any sufficiently capable quantum computers that could imple-
ment the NLP tasks proposed. Additionally, an assumption was made that one could
encode word meanings on the quantum computer using quantum random access
memory (QRAM) (Giovannetti et al., 2008), which to this day, and despite theoreti-
cal progress and experimental proposals, remains a distant possibility.
In the past year, we have been examining ways to use existing NISQ (= Noisy
Intermediate Scale Quantum) devices, in the first instance one of IBM’s quantum
devices, for NLP.
The string diagrams as depicted above can’t be interpreted directly by IBM’s
machine, which instead, needs something in the form of a ‘quantum circuit’. Natural
language when we map it to a ‘quantum circuit skeleton’ now looks like this (Coecke
et al., 2020):
Alice αA αp αB Bob
βA βB
α
β
hates
α + γ
β
γ
(3)
In this picture as well as in our experiments we used the ‘ZX-calculus’ (Coecke &
Duncan, 2011) for drawing quantum circuits, which is again part of the same string
diagram language of quantum theory—we give the translation of ZX-calculus circuit
components in standard circuit gates in Fig. 2.
How to Make Qubits Speak 281
α α
⎛ ⎞
1 0 0 0
1 1 0 1 0 ⎜0 1 0 0⎟
H◦ ◦H ⎜ ⎟ 1 0
0 0 eiα 0 eiα ⎝0 0 0 1⎠
0 0 1 0
Fig. 2 Translation of ZX-calculus circuit components in standard circuit gates. For a short tutorial
see (Coecke & Duncan, 2012), or for a more extensive ones see (Coecke & Kissinger, 2017; van de
Wetering, 2020)
... ...
α α
... ... (4)
...
α ...
... = = =
... α+β
β
...
...
(5)
So we are genuinely dealing with an entirely pictorial language, and in fact, recently it
was shown in (Hadzihasanovic, 2018) that all the equations that can be derived using
the usual quantum mechanical formalism, can also be derived using only pictures!
The key part of the passage from a sentence diagram representing grammatical
structure like (2) to a quantum circuit uses the ZX-calculus in a fundamental way.
For example, the fact that the CNOT-gate arises by sticking together two spiders as
follows:
(6)
is essential. Using this decomposition, the circuit arises as follows from (2):
282 B. Coecke et al.
Alice Bob
Alice *hates* Bob
= *hates*
(8)
From this (3) arises when parametrising the word meanings by gates. A detailed
discussion of the entire passage is in (Coecke et al., 2020).
In the form (3), natural language can be implemented on NISQ devices, and of
course, will still work well as these devices scale in terms of size and performance.
Crucially, our solution provides a way forward in the absence of QRAM. By employ-
ing quantum machine learning we do not directly encode the meanings of words, but
instead quantum gates—those in the circuit (3) carrying Greek letters—learn their
meanings directly from text (Meichanetzidis et al., 2020). By way of analogy with
classical machine learning, in quantum machine learning we can indeed use quantum
circuits instead of classical neural networks in order to learn patterns from data (Ma
et al., 2019; Benedetti et al., 2019). Interestingly, neural network architectures are the
state-of-the-art in classical NLP, but the majority of methods do not take advantage
of grammatical structures. In contrast, we saw that our approach to natural language
naturally accommodates both grammar and meaning.
Using our framework, once the meanings of words and phrases are encoded as
quantum gates, we are able to encode the meaning of grammatical sentences on
quantum hardware. Posing a question to the quantum computer, constructed by the
vocabulary and grammar the quantum computer has learned, it returns the answer—
as illustrated in Fig. 3.3
Naturally, we next turned our attention toward the design and execution of an
experiment that is non-trivial, not least of all since our design is predicated on the
program being scalable. This means that the dimension of the meaning space grows
significantly with the number of qubits available whilst the size of the circuits dictated
by the grammar does not grow too large with the size of the sentence.
3Earlier theoretical work by our team towards question-answering includes (Coecke et al., 2018;
Felice et al., 2019).
How to Make Qubits Speak 283
Fig. 4 A schematic presentation of the actual experiment on IBM quantum hardware, depicting
both the training and testing parts, with θ = {α A , β A , α B , β B , α, β, . . .}
284 B. Coecke et al.
subject(α A , β A ) = Alice
object(α B , β B ) = Bob
verb(α, β, . . .) = hates
The values are determined empirically by a text corpus and are then used to answer
questions about the corpus. In order to ensure that our experiment can be executed
effectively on near-term NISQ devices, but at the same time be complex enough to
be interesting, we chose a vocabulary of a few words, for example:
and generated not some, but all grammatical sentences from their combinations. From
these sentences, we created their corresponding parameterised circuits. Moreover,
we interpret the language diagrams such that the sentence space is one-dimensional,
i.e. just a number indicating the truth-value of the sentence:
0 ≤ sentence(α A , β A , α B , β B , α, β, . . .) ≤ 1
A value close to 1 represents “tr ue” and a value close to 0 indicates “ f alse”.
The labeled toy corpus would look like:
Cor pus = {(Alice loves Bob, f alse), (Bob is silly, tr ue), (Alice is rich, tr ue), ...}
in a training set T rain and a test set T est. Sentences in the training set T rain are
used to do supervised quantum machine learning in order to learn the parameters
that result in the correct measurement of the truth-labels. In this way, the parame-
ters for the circuits that prepare the meaning states for nouns {Alice, Bob}, verbs
{is, loves, hates}, and adjectives {rich, silly}, are learned.
The scheme for learning the parameters for words in sentences in the training set
is as follows. The circuit of a sentence in the training set is evaluated for the current
set of parameters on the quantum computer. By sampling measurement outcomes
we estimate:
|sentence(α A , β A , α B , β B , α, β, . . .)|2
This number is read by a classical computer that checks how well this matches the
desired truth label of the sentence. If there is a mismatch, our system updates the
parameters so that the quantum computer may evaluate an updated circuit. Iterating
this procedure until the parameters converge and all truth labels are reproduced for
the sentences in the training set.
How to Make Qubits Speak 285
After training, sentences in T est are used to estimate how well the truth labels
of new sentences, i.e. not in T rain, are inferred. These new sentences share the
same vocabulary as the sentences used for training, but they are grammatically and
semantically different.
Note that finding the optimal sequence of updates is, in general, a so-called ‘hard
optimization problem’, so it is important that our quantum meaning space is well
designed and allows the learning to be as tractable as possible. This design feature
is critical.
With this learning framework, we can now ask questions of the quantum com-
puter, as long as the questions are grammatical sentences expressed in terms of the
vocabulary and grammar used during training. We are pleased to add that in our
experiment, questions can, in fact, be posed as compositions of already learned sen-
tences. For example, we can use the relative pronoun who (Sadrzadeh et al., 2013)
(which we model in terms of CNOT gates within a quantum circuit) and ask:
who who
Bob is silly loves Alice is rich
(9)
is tr ue. This amounts to evaluating a bigger quantum circuit than the one that the
model has been trained on. However, because the model was trained on the same
grammar and vocabulary as used to express the question, we get the expected truth
label, false in this case.
One critical enabling feature of our experiment is making computers understand
the pictures that we have been drawing. For this we used GdF and AT’s DisCoPy
Python library (Felice et al., 2020). Another critical feature is effective and efficient
compiling and optimization. In order to run a circuit on a quantum device, it needed
to be compiled. Compiling entails morphing the circuit such that quantum operations
are expressed in terms of device-native operations, as well as accommodating for
the quantum processor’s restricted connectivity. For this task, we used Cambridge
Quantum Computing’s quantum software development platform, t|ket (Sivarajah
et al., 2020) which again crucially makes use of ZX-calculus.
286 B. Coecke et al.
5 QNLP
We’ve done it, it exists, so let’s give the kid a name: ‘quantum natural language pro-
cessing’, or QNLP in short. A much more detailed account on QNLP as well as the
data for a larger-scale experiment can be found in (Coecke et al., 2020; Meichanet-
zidis et al., 2020; Lorenz et al., 2021). With the successful execution of QNLP on
quantum hardware, and the great promise for the future, Cambridge Quantum Com-
puting has now resourced an Oxford based team dedicated to QNLP, and structure-
driven quantum AI more generally, evidently including each of us. Let’s single out
a few special features of QNLP.
Firstly, as already indicated above, QNLP is quantum-native. What we mean by
that is that the model of natural language that we employ is a quantum model (Clark
et al., 2014). Not just because we think that having a quantum model is particularly
cool, but because it was effectively the case that the most economic manner for
bringing language meaning and grammatical structure together was provided by the
‘categorical quantum mechanics’ framework (Abramsky & Coecke, 2004; Coecke,
2005; Coecke & Kissinger, 2017).4 The immediate consequence of this is that when
we try to ‘simulate’ QNLP on a classical computer, it would be exponentially expen-
sive, just as it is the case for simulation quantum systems on a classical computer.
Hence, QNLP truly loves being on quantum hardware.
Secondly, it is fair to say that it is meaning-aware, at the very least, it is definitely
more meaning-aware than the current deep learning based NLP implementations.
Indeed, the language pictures clearly indicate how the meanings flow in a meaningful
way. We say a bit more about this below in Sect. 9.
Thirdly, as already indicated in (Zeng & Coecke, 2016), there are many more
advantages to using quantum computers, most notably, we get quantum speed-up for
a large variety of NLP tasks, like question-answering. This space of opportunities is
in fact still largely unexploited, and the mission of our QNLP team is to do just that,
in as many possible manners as we can.
6 Beyond Language
We are now well on the way to make qubits speak. But can we make them do other
things as well? We are pretty sure we can, a first thing could be to make qubits
represent all of the inputs that we get through our fleshy physical embodiment,
like taste, smell, vision, hearing. The technical term for the kinds of spaces that
represent these sensory modes is ‘conceptual spaces’. They were introduced by Peter
Gärdenfors (Gärdenfors, 2000, 2014). There already are existing frameworks that
represent the input of these senses in a manner that exactly matches our language
4Categorical quantum mechanics is really just a fancy name for doing quantum mechanics in terms
of pictures.
How to Make Qubits Speak 287
diagrams, and hence allow for these sensory inputs to interact with each other (Bolt
et al., 2018).
These conceptual spaces are convex, and fit well with some other work within
the framework of our natural language diagrams, namely on the use of the quantum
mechanical concept of density matrices, which also form convex spaces (Ludwig,
1985; Barrett, 2007). This additional convex structure allows one to bring many more
linguistic features into play, like word ambiguity. For example, the word queen is
clearly ambiguous, as it can be a royal, a rock band, a bee, a chess piece, a playing card,
etc. In order to represent that ambiguity, we simply add all the different meanings in
order to form a density matrix (Piedeleu et al., 2015):
ρqueen = |queen-royalqueen-royal|
+ |queen-bandqueen-band|
+ |queen-beequeen-bee|
+ |queen-chessqueen-chess|
+ |queen-cardqueen-card|
+ ...
ρlion = |00|
ρtiger = |++|
ρcheeta = |−−|
ρbig cat = |00| + |11|
ρmammal = |00| + |11| + |22|
ρvertebrate = |00| + |11| + |22| + |33|
In (Coecke & Meichanetzidis, 2020; Lewis, 2019a, b; De las Cuevas et al., 2020;
Meyer & Lewis, 2020) this use of density matrices in language is further elaborated
upon.
Sticking with the theme of big cats, what’s the difference between a tiger and a lion?
In particular, why does a tiger have stripes, while a lion doesn’t (Fig. 5).
A more traditional scientist will tell you that these days we understand all of that
perfectly well. Before, if we dissected these animals, we found exactly the same
288 B. Coecke et al.
organs. If we do some smaller scale animal chopping, we end up with cells that
again have exactly the same structure. However, digging even further, and hitting the
molecular level, we encounter DNA, and then we truly understand the difference.
Really? Some obscure humongously large code that is impossible to grasp by any
human ‘explains’ the difference? It explains as much the difference between these
two animals as the codes of two computer programs written in terms of 0’s and 1’s
explain the difference between accounting software and the operating system of your
smart phone.
That said, most of the exact sciences have adopted this perspective of understand-
ing things by breaking them down to smaller things, like elementary particles in
particle physics, and the elements of a set in set-theory. In contrast, in the arts and
humanities things really get their meanings by putting them in a context. For example,
in the social sciences properties are mostly thought of in terms of behaviours, rather
than by trying to build a map of the inside of someone’s brain, and most modern art
has very little meaning without a context.
We think that we can all agree that for our tiger and our lion, what explains the
difference is really the hunting process which each of them needs to do in order to
survive, and their respective coats provide the best camouflage for doing so, given they
hunt in very different environments. While of course these differences are encoded in
the genes, it is the process of evolution that makes our tiger and our lion look as they
do. The role of DNA in explaining the stripes of tigers is similar to the role of our
parameters α A , β A , α B , . . . in computing the meaning of sentences. Indeed, learning
the optimal parameters is an evolution-like process that depends on the environment
of the words, i.e. on the training data that we feed the machine.
So the two views explained above are really about either focussing on components
vs. context, and the latter is very much in line with Wittgenstein’s “meaning is use”.
We are saying all of this because the language diagrams that we saw above are the
mathematical incarnation of this way of thinking, for example:
How to Make Qubits Speak 289
(10)
8 Compositionality
Above we saw how we can turn language diagrams into circuits. A typical thing
about circuits is, of course, that you can compose different circuits together for
forming bigger circuits. This is the principle of “compositionality”, and in fact,
string diagrams themselves are a result of this way of thinking. We can now, once
we moved to circuits, compose several sentences together in order to produce larger
text:
290 B. Coecke et al.
(11)
brother
hangs
gang
shoots
shoots
(12)
(13)
Maybe there is something deeper here. Maybe the origin of language should be
looked for in the visual world, rather than in some symbolic abstract realm. The
passage from these visual circuits to the more complex grammatical structure may
actually boil down to forcing something that wants to live in two (or more) dimensions
on a line, and force additional bureaucracy (or, bureau-crazy?) like the ordering of
words, which ends up being different for different languages. Why forced into one-
dimension? Since it seems to be hard for us humans to communicate verbally in any
other way than in terms of one-dimensional strings of sounds.
In fact, the very success of string diagrams is that it allows things to live in two
(or more) dimensions. Here’s a demonstration of this fact. Consider the operations of
parallel composition ⊗ (a.k.a. ‘while’) and sequential composition ◦ (a.k.a. ‘after’).
Typically, in textbooks, this structure is represented by one-dimensional strings
of algebraic symbols, and then, equations are needed to express their interaction
(a.k.a. bifunctoriality (Mac Lane, 1998)):
f1 f2
g1 ⊗ g2 ◦ f1 ⊗ f2 = g1 g2 ◦ f1 f2 =
g1 g2
⎛ ⎞ ⎛ ⎞
⎜ f1 ⎟ ⎜ f2 ⎟ f1 f2
⎜ ⎟ ⎜ ⎟
g1 ◦ f1 ⊗ g2 ◦ f2 =⎜ ⎟⊗⎜ ⎟ =
⎝ g ⎠ ⎝ g ⎠ g1 g2
1 2
That is, by using the two-dimensional format of diagrams, the symbolic equation
is always satisfied, so it is not needed anymore. In other words, the equation was a
piece of bureaucracy due to the one dimensional symbolic representation.
Let’s end with a bit of philosophy, which puts our language diagrams within a broader
context of different approaches that one may take towards machine learning. We’ll
again do this by means of a metaphor. Assume one wants to design a home. One way
to do so would be to train a machine using many different images of homes, possibly
with some indications of appreciation. Then, let the machine be creative and design
a home based on all of that input.
And it comes up with an amazing design. But what dit it really do? It kind of
carved everything out of one big block of stuff, and there will be forks and plates on
the table, chairs at the table, pictures on the wall, since that’s were they were in all of
the training images. However, the machine is unaware of what the use is of each of
these objects, and in particular that there is a very special relationship between the
plate and the fork, and also with the chair at the table (and less so with the picture
on the wall), and even more so, that there is a person who happens to be not in the
picture, but who wants to exploit this special relationship with respect to the food
they are cooking in the kitchen.
On the other hand, our language diagrams, when applied to a broader context, can
capture these relationships, in contrast to the black-box approach that is taken within
machine learning:
How to Make Qubits Speak 293
is at
is on
is on
is in
is in
(14)
First, we need some data to play with. For this we have two options: (1) we generate
an artificial dataset by designing the grammar ourselves, this is the simplest option
and the one we have implemented in (Coecke et al., 2020) or (2) we take some real-
world text data (e.g. every sentence from Alice in Wonderland) and use a parser to
construct the diagrams automatically for us. In any case, we begin our experiment
with a collection of diagrams for sentences, with some annotation telling us whether
each sentence is true or false (or whether or not it comes from Alice in Wonderland).
Next, we need to choose an ansatz for each type of word, i.e. an appropriate circuit
shape. Instantaneous quantum polynomials (IQP) (Shepherd & Bremner, 2009) are
a reasonable choice, since they are shallow circuits but are still believed to be hard
to simulate for classical computers. Once we’ve chosen the circuit shape for each
type, this defines a landscape, the total number of parameters defining our model:
the sum of the parameters defining each word. For each point in this landscape, i.e.
for each combination of parameters, we can run the circuits corresponding to each
294 B. Coecke et al.
sentence in our dataset and check how far away we are from the correct label. The
sum of the distance between our model’s prediction and the correct label is called
the loss function, this is the objective that we want to minimise.
Once we have defined the landscape and the loss, we can feed them as input
to the optimisation algorithm of our choice. The easiest choice, and the one we
implemented in (Coecke et al., 2020), is to pick a black box optimiser such as
simultaneous perturbation stochastic approximation (SPSA) (Spall, 1992). A more
advanced solution is to go fully diagrammatic and compute the gradient of our loss
directly from the shape of the diagrams for sentences (Toumi et al., 2021). When
the process has converged to some (locally) optimal parameters, tadam! We have
successfully trained our quantum natural language model. We may now evaluate
it on some testing data, typically a subset of the sentences that we withhold from
the training set. Note that crucially, even though our model has never seen these
sentences during training, hopefully it has already learnt the meanings for the words,
and it knows how to combine these meanings in new ways.
Thus, a typical QNLP experiment may be summarised as follows: (1) we draw
the diagrams for each sentences, (2) we pick a circuit shape for each word type,
(3) we define a loss function, (4) we find the optimal parameters and (5) we test
the model’s prediction. These five steps fit in a few lines of Python using the tools
from the DisCoPy library, the result can then be presented in the form of a Jupyter
notebook, see the documentation for plenty of examples:
https://discopy.readthedocs.io/en/main/notebooks.qnlp.html
Also, here is a more detailed tutorial as well:
https://discopy.readthedocs.io/en/main/notebooks/qnlp-tutorial.html
11 Outlook
So what’s next? Well, as we speak the editor of this volume is himself getting to play,
and obviously, we will also embark on a musical adventure. Evidently, we will also
further develop QNLP bringing in more conceptual depth, for the specific case of
language as well as towards broader AI features. Experimentally, we keep pushing
along as hardware improves.
All together this is just the beginning of an adventure that, we hope, on the one
hand, will create the next generation of AI, and on the other hand, will further the
quest towards better understanding the human mind and its interaction with the world
around us.
How to Make Qubits Speak 295
References
Abramsky, S. (1996). Retracing some paths in process algebra. In CONCUR’96: concurrency theory,
vol. 1119 of Lecture notes in computer science (pp. 1–17). Springer.
Abramsky, S., & Coecke, B. (2004). A categorical semantics of quantum protocols. In Proceed-
ings of the 19th annual IEEE symposium on logic in computer science (LICS) (pp. 415–425).
arXiv:quant-ph/0402130.
Adjukiewicz, K. (1935) Die syntaktische Konnexität. Studia Philosophica.
Baez, J. C., & Fong, B. (2015). A compositional framework for passive linear networks.
arXiv:1504.05625
Baez, J. C., & Stay, M. (2011). Physics, topology, logic and computation: a Rosetta stone. In
B. Coecke (ed.), New structures for physics, lecture notes in physics (pp. 95–172). Springer.
Bankova, D., Coecke, B., Lewis, M., & Marsden, D. (2019). Graded hyponymy for compositional
distributional semantics. Journal of Language Modelling, 6(2), 225–260.
Bar-Hillel, Y. (1953). A quasiarithmetical notation for syntactic description. Language, 29, 47–58.
Barrett, J. (2007). Information processing in generalized probabilistic theories. Physical Review A,
75, 032304.
Benedetti, M., Lloyd, E., Sack, S., & Fiorentini, M. (2019). Parameterized quantum circuits as
machine learning models. Quantum Science and Technology, 4, 043001.
Bolt, J., Coecke, B., Genovese, F., Lewis, M., Marsden, D., & Piedeleu, R. (2018). Interacting
conceptual spaces I: Grammatical composition of concepts. In M. Kaipainen, A. Hautamäki,
P. Gärdenfors, & F. Zenker (eds.), Concepts and their applications, synthese library, studies in
epistemology, logic, methodology, and philosophy of science. Springer.
Bonchi, F., Sobocinski, P., & Zanasi, F. (2014). A categorical semantics of signal flow graphs. In
CONCUR’14: concurrency theory. Lecture notes in computer science (vol. 8704, pp. 435–450).
Springer.
Chomsky, N. (1956). Tree models for the description of language. I.R.E. Transactions on Information
Theory, IT, 2, 113–124.
Chomsky, N. (1957). Syntactic Structures. Mouton.
Clark, S., Coecke, B., Grefenstette, E., Pulman, S., & Sadrzadeh, M. (2014). A quantum teleportation
inspired algorithm produces sentence meaning from word meaning and grammatical structure.
Malaysian Journal of Mathematical Sciences, 8, 15–25. arXiv:1305.0556.
Coecke, B. (2005). Kindergarten quantum mechanics. In A. Khrennikov (ed.), Quantum theory:
Reconsiderations of the foundations III (pp. 81–98). AIP Press. arXiv:quant-ph/0510032
Coecke, B. (2013). An alternative Gospel of structure: Order, composition, processes. In C. Heunen,
M. Sadrzadeh, & E. Grefenstette (eds.), Quantum physics and linguistics. A compositional, dia-
grammatic discourse (pp. 1 – 22). Oxford University Press. arXiv:1307.4038
Coecke, B. (2017). From quantum foundations via natural language meaning to a theory of every-
thing. In S. B. Cooper & M. I. Soskova (eds.), The incomputable: Journeys beyond the turing
barrier, theory and applications of computability (pp. 63–80). Springer International Publishing.
arXiv:1602.07618
Coecke, B. (2019). The mathematics of text structure. arXiv:1904.03478
Coecke, B., De Felice, G., Marsden, D., & Toumi, A. (2018). Towards compositional distributional
discourse analysis. In M. Lewis, B. Coecke, J. Hedges, D. Kartsaklis, & D. Marsden (eds.),
Proceedings of the 2018 workshop on compositional approaches in physics, NLP, and social
sciences. Electronic proceedings in theoretical computer science (vol. 283, pp. 1–12)
Coecke, B., de Felice, G., Meichanetzidis, K., & Toumi, A. (2020). Foundations for near-term
quantum natural language processing. arXiv preprint arXiv:2012.03755
Coecke, B., & Duncan, R. (2011). Interacting quantum observables: Categorical algebra and dia-
grammatics. New Journal of Physics, 13, 043016. arXiv:quant-ph/09064725
Coecke, B., & Duncan, R. (2012). Tutorial: Graphical calculus for quantum circuits. In International
workshop on reversible computation (pp. 1–13). Springer
296 B. Coecke et al.
Coecke, B., & Kissinger, A. (2017). Picturing quantum processes. A first course in quantum theory
and diagrammatic reasoning. Cambridge University Press.
Coecke, B., & Meichanetzidis, K. (2020). Meaning updating of density matrices. arXiv:2001.00862
Coecke, B., & Paquette, É. O. (2011). Categories for the practicing physicist. In B. Coecke (ed.),
New structures for physics, lecture notes in physics (pp. 167–271.) Springer. arXiv:0905.3010
Coecke, B., Sadrzadeh, M., & Clark, S. (2010). Mathematical foundations for a compositional
distributional model of meaning. In J. van Benthem, M. Moortgat, & W. Buszkowski (eds.), A
festschrift for Jim Lambek. Linguistic analysis (vol. 36, pp. 345–384). arxiv:1003.4394
Coecke, B., & Wang, V. (2020). De-humanising grammar for improved machine use. Draft paper.
Cowtan, A., Dilkes, S., Duncan, R., Simmons, W., & Sivarajah, S. (2019). Phase gadget synthesis
for shallow circuits. arXiv preprint arXiv:1906.01734
de Beaudrap, N., Bian, X., & Wang, Q. (2020). Fast and effective techniques for t-count reduc-
tion via spider nest identities. In 15th conference on the theory of quantum computation, com-
munication and cryptography (TQC 2020). Schloss Dagstuhl-Leibniz-Zentrum für Informatik.
arXiv:2004.05164
de Felice, G., Di Lavore, E., Román, M., Toumi, A. (2020). Functorial language games for question
answering. arXiv preprint arXiv:2005.09439
de Felice, G., Meichanetzidis, K., & Toumi, A. (2019). Functorial question answering. arXiv preprint
arXiv:1905.07408
de Felice, G., Toumi, A., & Coecke, B. (2020). DisCoPy: Monoidal categories in python. Accepted
for ACT 2020. arXiv preprint arXiv:2005.02975
De las Cuevas, G., Klingler, A., Lewis, M., & Netzer, T. (2020). Cats climb entails mammals move:
preserving hyponymy in compositional distributional semantics. arXiv:2005.14134 [quant-ph]
Duncan, R., Kissinger, A., Pedrix, S., & van de Wetering, J. (2019). Graph-theoretic simplification
of quantum circuits with the zx-calculus. arXiv preprint arXiv:1902.03178
Firth, J. R. (1957). A synopsis of linguistic theory, 1930–1955. Studies in Linguistic Analysis, 1–32.
Gärdenfors, P. (2000). Conceptual spaces: The geometry of thought. MIT Press.
Gärdenfors, P. (2014). The geometry of meaning: Semantics based on conceptual spaces. MIT Press.
Ghani, N., Hedges, J., Winschel, V., & Zahn, P. (2018). Compositional game theory. In Proceedings
of the 33rd annual ACM/IEEE symposium on logic in computer science, (pp. 472–481). ACM.
Giovannetti, V., Lloyd, S., & Maccone, L. (2008). Quantum random access memory. Physical review
letters, 100(16), 160501.
Hadzihasanovic, A., Ng, K. F., & Wang, Q. (2018). Two complete axiomatisations of pure-state
qubit quantum computing. In Proceedings of the 33rd annual ACM/IEEE symposium on logic in
computer science (pp. 502–511). ACM.
Harris, Z. S. (1954). Distributional structure. Word, 10(2–3), 146–162.
Kissinger, A., & van de Wetering, J. (2019). Pyzx: Large scale automated diagrammatic reasoning.
arXiv preprint arXiv:1904.04735
Lambek, J. (1958). The mathematics of sentence structure. American Mathematics Monthly, 65.
Lambek, J. (2008). From word to sentence. Milan: Polimetrica.
Lewis, M. (2019a). Compositional hyponymy with positive operators (pp. 638–647). Varna, Bul-
garia: In Proceedings of Recent Advances in Natural Language Processing.
Lewis, M. (2019b). Towards negation in discocat. In: Proceedings of SemSpace.
Lorenz, R., Pearson, A., Meichanetzidis, K., Kartsalkis, D., & Coecke, B. (2021). Qnlp in
practice: Running compositional models of meaning on a quantum computer. arXiv preprint
arXiv:2102.12846.
Ludwig, G. (1985). An axiomatic basis of quantum mechanics (vol. 1). Berlin Heidelberg: Derivation
of Hilbert Space. Springer-Verlag.
Ma, Y., Tresp, V., Zhao, L., & Wang, Y. (2019). Variational quantum circuit model for knowledge
graph embedding. Advanced Quantum Technologies, 1800078.
Mac Lane, S. (1998). Categories for the working mathematician. Springer.
How to Make Qubits Speak 297
Meichanetzidis, K., Gogioso, S., de Felice, G., Toumi, A., Chiappori, N., & Coecke, B. (2020).
Quantum natural language processing on near-term quantum computers. Accepted for QPL 2020.
arXiv preprint arXiv:2005.04147
Meichanetzidis, K., Toumi, A., de Felice, G., & Coecke, B. (2020). Grammar-aware question-
answering on quantum computers. arXiv:2012.03756
Meyer, F., & Lewis, M. (2020). Modelling lexical ambiguity with density matrices. In Proceed-
ings of the 24th Conference on Computational Natural Language Learning (pp. 276–290).
arXiv:2010.05670 [cs.CL]
O’Riordan, L. J., Doyle, M., Baruffa, F., & Kannan, V. (2020). A hybrid classical-quantum workflow
for natural language processing. arXiv preprint arXiv:2004.06800
Pavlovic, D. (2013). Monoidal computer I: Basic computability by string diagrams. Information
and computation, 226, 94–116.
Piedeleu, R., Kartsaklis, D., Coecke, B., & Sadrzadeh, M. (2015). Open system categorical quantum
semantics in natural language processing. In 6th conference on algebra and coalgebra in computer
science (CALCO 2015). Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik.
Intel News Release. (2019). Intel to support the Irish Centre for High End Computing on new col-
laborative quantum computing project, March 6, 2019. https://newsroom.intel.ie/news-releases/
intel-to-support-the-irish-centre-for-high-end-computing-on-new-collaborative-quantum-
computing-project/
Sadrzadeh, M., Clark, S., & Coecke, B. (2013). The Frobenius anatomy of word meanings I: Subject
and object relative pronouns. Journal of Logic and Computation, 23, 1293–1317. arXiv:1404.5278
Selinger, P. (2011). A survey of graphical languages for monoidal categories. In B. Coecke (ed.),
New structures for physics, Lecture notes in physics (pp. 275–337). Springer. arXiv:0908.3347
Shepherd, D., & Bremner, M. J. (2009). Temporally unstructured quantum computation. Proceed-
ings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 465(2105), 1413–
1439.
Signorelli, C. M., Wang, Q., & Khan, I. (2020). A compositional model of consciousness based on
consciousness-only. arXiv preprint arXiv:2007.16138
Sivarajah, S., Dilkes, S., Cowtan, A., Simmons, W., Edgington, A., & Duncan, R. (2020). t|ket>:
A retargetable compiler for NISQ devices. arXiv preprint arXiv:2003.10611
Spall, J. C., et al. (1992). Multivariate stochastic approximation using a simultaneous perturbation
gradient approximation. IEEE Transactions on Automatic Control, 37(3), 332–341.
Toumi, A., Yeung, R., & de Felice, G. (2021). Diagrammatic differentiation for quantum machine
learning. arXiv preprint arXiv:2103.07960
Tull, S., & Kleiner, J. (2020). Integrated information in process theories. arXiv preprint
arXiv:2002.07654
van de Wetering, J. (2020). ZX-calculus for the working quantum computer scientist.
arXiv:2012.13966
Zeng, W., & Coecke, B. (2016). Quantum algorithms for compositional natural language processing.
Electronic Proceedings in Theoretical Computer Science, 221. arXiv:1608.01406
A Quantum Computing Approach
to Harnessing the Logic of the Mind
for Brain–Computer Interfacing
1 Introduction
The human brain is allegedly the most complex object known to mankind: it has circa
one hundred billion neurones forming a network of quadrillions of connections. The
amount of information that circulates through this network is immense.
There is a school of thought, called Dualism, which considers the mind and the
brain as separate entities (Rozemond, 1988). What is more, it has been suggested
that minds would not even need brains to exist. Although the separation between
mind and brain enjoys some currency in philosophical circles, it is generally agreed
nowadays that the mind results from the functioning of the brain.
The scientific community does not have yet a clear understanding of how brain
activity gives rise to the mind. Even though the behaviour of individual neurones is
fairly well understood nowadays, the way in which they cooperate in ensembles of
E. R. Miranda (B)
Interdisciplinary Centre for Computer Music Research (ICCMR), University of Plymouth,
Plymouth, UK
e-mail: [email protected]
millions has been difficult to unpack. Yet, this is of paramount importance to fathom
how the brain creates the mind.
The advent of increasingly sophisticated brain scanning technology has been
enabling a plethora of research activity to comprehend the neuronal correlates of
mental activities (Kraft et al., 2008; Vartanian et al., 2013). This comprehension is
paramount for philosophy, psychology, medicine and engineering. Indeed, emerging
technology that enables users to control systems with their mind banks on such
developments.
A Brain–Computer Interface, or BCI, is a piece of equipment that enables users to
control systems with their mind. It reads signals from the brain and harnesses them
for communicating with a computer; for example, to control a mechanical device,
such as a robotic arm, a musical instrument or a wheelchair (Fig. 1).
This chapter proposes a novel approach to study and harness neuronal correlates
of mental activity for the development of BCI systems. It introduces the notion of a
logic of the mind, where neurophysiological data are encoded as logical expressions
representing mental activity. These representations then become instructions for BCI
control. Ultimately, the logic of the mind is tool for studying the neuronal correlates
of mental activities.
Effective logical expressions are likely to be extensive and complex, involving
dozens of variables. Large expressions require considerable computational power to
be processed. This is problematic for BCI applications because brain data need to be
processed rapidly in order to execute sequences of commands on the fly.
Quantum computers hold much promise in terms of processing speed for a number
of problems that are intractable by classical computers, including those involving
evaluation of logical expressions. Hence, the rational for using quantum computers
to process the logic of the mind.
Quantum computers are fundamentally different from the typical computer as
we know it. The speed-up prophecy depends, amongst other things, on algorithms
that cleverly exploit fundamental properties of quantum physics, such as superpo-
sition, entanglement and interference. As it will be demonstrated below, the proposed
quantum algorithm to process logic of the mind expressions takes advantage of those
properties.
As a proof-of-concept, the chapter presents a system that reads the electroen-
cephalogram of a person at given lapses of time and builds logic expressions encoding
brain activity. The system generates respective quantum circuits and submits them to
a quantum computer. The computer then checks the satisfiability of the expressions.
Those expressions that are satisfied trigger commands to control a robotic arm.
As an additional example, we introduce an application of a BCI to make music. In
this case, the system converts the satisfiability results into sounds. The BCI becomes
a musical instrument controlled by the mind of the player.
The paper begins with a brief introduction to BCI and the electroencephalogram
(EEG) which is the neurophysiological signal that is normally used in BCI systems,
including the one presented in this chapter. Then, it briefly discusses how the EEG
corresponds to mental states, and presents the logic of the mind concept. Next, there
is a short introduction to quantum computing. Here, we focus on the basics deemed
necessary to understand the proposed quantum algorithm that deals with logic of
the mind expressions. Then, the paper shows how those expressions are processed
and describes the proof-of-concept BCI systems. The limitations of current quantum
computing hardware technology are also discussed.
2 Brain–Computer Interfacing
By and large, BCI research is concerned with the development of assistive technology
for persons with severe motor impairment. However, BCI technology is also being
developed for other applications, such as computer games (Hasan & Gan, 2012),
302 E. R. Miranda
biometrics (Palaniappan, 2008) and cursor control (Wilson & Palaniappan, 2011).
And this author has developed BCI for making music (Miranda, 2006; Miranda et al.,
2011).
There are different types of brain signals and respective sensors to read them.
The type of signal that is most commonly used for BCI technology is the
electroencephalogram.
Neurones communicate with one another through electrical impulses. Neuronal elec-
trical activity can be recorded with electrodes placed on the scalp (Fig. 2). This
recording is called the electroencephalogram, or EEG. The EEG conveys the overall
activity of millions of neurones in the brain in terms of electric current. This is
measured as the voltage difference between two or more electrodes, one of which is
taken as a reference. See (Marcuse et al., 2015) for more details.
It is also possible to record electrical brain activity with electrodes surgically
implanted under the skull, on the surface of the cortex or deep inside the brain. This
is often called electrocorticography (ECoG) or intracranial EEG (iEEG). Whereas
implanted electrodes provide a far better signal to work with than surface ones, brain
implants are not yet routinely used for BCI systems for health and safety reasons.
Other technologies for brain scanning include functional Magnetic Resonance
1 https://www.gtec.at/.
304 E. R. Miranda
Power spectrum analysis is a popular method to harness the EEG. This method
breaks the EEG signal into different frequency bands and reveals the distribution
of power between them. Power spectrum analysis is useful because it can reveal
patterns of brain activity, and a computer can be programmed to recognize and
translate them into commands for a system. Although this chapter focuses on power
spectrum analysis, it is worth noting that there are other EEG analysis methods for
detecting mental activity as well; e.g. Hjorth analysis (Oh et al., 2014).
Typically, users must learn how to voluntarily produce specific patterns of EEG
signals in order to be able to control something with a BCI. A fair amount of research
is underway to develop lexicons of detectable EEG patterns, understand what they
mean and develop methods to train users to produce them.
The EEG conveys information about mental activity (Anderson & Sijercic, 1996;
Petsche & Etlinger, 1998). In medicine, the EEG is an important tool for diagnosis
of brain disorders.
There has been an increasing amount of research aimed at the identification of
EEG correlates of all sorts of mental activities (Hinterberger et al., 2014; Jeunet
et al., 2020; Nicolas-Alonso & Gomez-Gil, 2012; So et al., 2017; Yelamanchili,
2018). For instance, Giannitrapani (1985) identified EEG patterns associated with
abstract intellectual tasks, such as doing arithmetic operations mentally. Guenter and
Brumberg (2011) discovered neural correlates for the production of speech, which
can be detected in the EEG. And more recently, Daly et al. (2018) detected EEG
signatures correlated with emotions elicited while subjects were listening to music.
It is generally known that the distribution of power in the spectrum of the EEG can
indicate different states of mind. For example, a spectrum with salient low-frequency
components is often associated with a state of drowsiness, whereas a spectrum with
salient high-frequencies is associated with a state of alertness. Giannitrapani (1985)
linked the prominence of low-frequency components with a passive state of mind,
as opposed to an active state, which is characterized by high-frequency spectral
components.
Research exploring mental correlates of the EEG normally considers spectral
components up to 40 Hz (Kropotov, 2008). There are four recognized spectral
frequency bands, also known as EEG rhythms, each of which is associated with
specific mental states (Table 1).
The exact boundaries of the bands listed in Table 1 and their respective mental
states vary from one author to another.
The EEG that gives raise to mental states is highly dynamic. Figure 4 shows a
snapshot of EEG mapped onto 2D and 3D topographic representations using the
OpenVibe software (Renard et al., 2010). The line cursor on the ‘Signal display’
pane shows the precise moment of the snapshot. At this moment, prominent beta
rhythms are detected at the lateral sides of the brain.
A Quantum Computing Approach to Harnessing the Logic … 305
Table 1 Typical EEG rhythms and associated mental states. Frequency bands are in Hertz (Hz)
Bands Rhythms Mental states
f <4 Delta Sleep; can indicate of cerebral anomaly
4≤ f <8 Theta Drowsiness; hypnotic state; can indicate cerebral anomaly
8 ≤ f < 15 Alpha Relaxed, meditative, unfocused, almost drowsy state of mind
15 ≤ f < 40 Beta High arousal, active thinking, consciously focusing state of mind
Fig. 4 Snapshot of EEG activity mapped onto 2D and 3D topographical representations (The
electrodes’ labelling arrangement here is slightly different from the convention introduced in Fig. 2.)
It is common knowledge that different regions of the brain have distinct roles;
e.g. the visual cortex processes images, the auditory cortex processes sound, the
motor cortex controls our limbs and so on (Squire et al., 2008). Assorted regions
of the brain cooperate to perform mental tasks. Moreover, spectral amplitudes of
scalp EEG are constantly changing over the skull surface. Therefore, states of mind
cannot be accurately inferred simply by looking at a snapshot of the averaged EEG
from the whole set of electrodes at once. Rather, one needs to look at the time-based
interrelationships between spectral components of the EEG recorded simultaneously
at different locations on the scalp.
306 E. R. Miranda
Petsche and Etlinger (1998) advocate that different bandwidths of the EEG spectrum
are distinct windows to a dynamic landscape of electrical brain activity. It is suggested
that mental states are correlated with interrelations between these windows. We
follow this up by proposing a method to represent those interrelationships using
logical expressions. Hence, the notion of an EEG-based logic of the mind.
Consider a system, which extracts information from the EEG of a number of
electrodes at given lapses of time; e.g. periods lasting for 500 ms. For the sake of
clarity, let us say that the system extracts beta and alpha rhythms from the EEG signals
(Table 1). With this information, the system tracks the behaviour of these rhythms
over the scalp as time progresses. For this example, it tracks the electrodes’ locations
where beta and alpha rhythms displayed most power. This is shown in Fig. 5: beta
rhythms (denoted by the red electrodes) were detected prominently by electrode Fp2
at time t 0 . Next, at time t 1 they were detected prominently by electrode Fz and then
by electrode T4 at time t 2 . And alpha rhythms (denoted by the blue electrodes) were
detected prominently by electrodes T5, T3 and C3, respectively.
Next, at each time step t n , the EEG information is encoded as a logic expression.
The variables represent the activity of the respective electrodes.
In order to keep the example simple, let us consider only the following subset of
electrodes: {Fp2, Fz, T3, C3, T4, T5}. In this case, the electrodes that detect the most
prominent EEG signals are represented in the expression as ‘True’. All the others are
‘False’. Thus, the expressions for the time lapses in Fig. 5 could be written follows2 :
2 See (Smith, 2020) for an introduction to formal logic. The symbol ¬ is the negation operator (i.e.
NOT), and ∧ stands for the logical conjunction operator (i.e. AND). Variables corresponding to a
True electrode are in bold for clarity. These expressions could be stated in reduced form, but we
leave them expanded here for the sake of intelligibility.
A Quantum Computing Approach to Harnessing the Logic … 307
The first term, inside the parenthesis on the left side of the conjunction operator,
corresponds to the beta rhythms and the second term to the alpha ones.
As a matter of fact, the expressions above could have been written with other
logical operators. For instance, t 0 could have been written with the logical disjunction
operator3 between the beta and alpha terms:
The forms of those logical expressions depend on what they are meant to repre-
sent. They might, for example, depend on the correlated mental states which they are
supposed to stand for. For instance, it is generally agreed that increased beta activity in
the pre-frontal cortex corresponds to ‘focused attention in problem-solving’ (Ligeza
et al., 2015). The pre-frontal cortex is covered by electrodes Fp1 and Fp2, and to some
extent F7, F3, Fz, F4 and F8. Thus, the expressions for t 0 and t 1 above would corre-
spond to (or ‘satisfy’ in logic parlance) this mental state. Conversely, the expression
for t 2 would not.
The problem of establishing which values for the variables of a logical expression
can render the whole expression True, or satisfiable, is known as the Boolean satis-
fiability problem. As an illustration, consider this simple expression, with variables
A and B: (¬A ∧ B). If A = False and B = True, then this expression is True. Hence,
these values for A and B render the expression satisfiable. But if, say, B = False then
the whole expression would be False.
The ability to represent EEG correlates in terms of logical expressions is useful
for implementing BCI systems because they can be programmed to activate actions
associated to specific expressions. As the EEG is acquired and analysed, a system
would check when the information satisfies given logical expressions. Then, it would
perform the respective actions for those expressions that return True. For example:
If at t n , β(a ∧ ¬(b ∧ c ∧ d ∧ e ∧ f )) ∨ α(¬(a ∧ b ∧ c ∧ d ∧ e) ∧ f )) = True.
then { move robot arm to the left by 180°};
This section introduces the basics of quantum computing and logic operations. It
focuses on the basics deemed necessary to contextualize and understand how to
program quantum computers for solving logical expressions. For more detailed expla-
nations of quantum computing, the reader is referred to (Bernhardt, 2019; Grum-
bling & Horowitz, 2019; Johnston et al., 2019; Mermin, 2007; Rieffel & Polak,
2011).
Classical computers manipulate information represented in terms of binary digits,
each of which can value 1 or 0. They work with microprocessors made up of billions
of tiny switches activated by electric signals. Values equal to 1 and 0 reflect the on
and off states of the switches, respectively.
A Quantum Computing Approach to Harnessing the Logic … 309
The state vector | is a superposition of vectors |0 and |1 in a two-dimensional
complex space, referred to as Hilbert space, with amplitudes α and β. Consider the
squared values of α and β as probability values representing the likelihood of the
310 E. R. Miranda
Fig. 7 The X gate rotates a qubit’s state vector (pointing upwards on the figure on the left side) by
180° around the x-axis (pointing downwards on the figure on the right side)
Fig. 8 The Hadamard gate puts the qubit into a superposition state halfway two opposing poles
Fig. 9 A quantum algorithm depicted as a circuit of quantum gates. The squares with dials represent
measurements, which are saved on classical registers represented at the bottom line
gate is an entangling gate. Depending on the input control state, it can put 2 qubits
in entanglement, which is the second fundamental property of quantum computing
mentioned at the introduction. Entangled qubits can no longer be thought of as
independent units. They become one quantum entity described by a state vector of
its own right.
The CX gate applies an X gate on a qubit only if the state of another qubit is |1.
Thus, the CX gate establishes a dependency of the state of 1 qubit with the value of
another. The schematic representation of the CX gate is shown in Fig. 10. In fact,
any quantum gate can be built in controlled form. And entanglement can take place
between more than 2 qubits.
The Bloch sphere is useful for visualizing what happens with a single qubit, but it
is not suitable for multiple qubits, in particular when they are entangled. Entangled
qubits can no longer be thought of as independent units. They become one quantum
entity described by a state vector of its own right. Hence, from now on, we will have
to use mathematics to represent quantum systems.
The notation used above to represent quantum states (|, |0, |1), is called Dirac
notation. It provides an abbreviated way to represent vectors. For instance, |0 and
|1 represent the following vectors, respectively:
Mathematically, quantum gates are expressed as matrices. For instance, the X gate
is represented as
01
X=
10
Therefore, quantum gate operations are expressed as matrix operations. Thus, the
application of an X gate to |0 is the multiplication of a matrix (gate) by a vector
(qubit state), which looks like this:
01 1 0
X (|0) = × = = |1
10 0 1
Table 2 shows the resulting quantum states of CX gate operations, where the
second qubit flips only if the first qubit is |1. Note that in quantum computing,
qubit strings are usually enumerated from the right end of the string to the left: e.g.
|q2 ⊗ |q1 ⊗ |q0 . This is the norm adopted in this chapter from now on. Thus, the
‘first qubit’ is the rightmost one.
Another useful controlled gate is the multiple controlled form of the X gate, also
known as the Toffoli gate (Aharonov, 2003). An example of a 3-qubit Toffoli gate
(also known as CCX or CCNOT) in shown in Fig. 11.
Table 3 shows resulting quantum states of the 3-qubit Toffoli gate portrayed in
Fig. 11.
In digital logic, one can build any logic operator and entire Boolean expressions with
just one basic NAND operator (Akerkar and Akerkar, 2004). Likewise, quantum
A Quantum Computing Approach to Harnessing the Logic … 315
logic operators and expressions can be built using only the X gate and its controlled
forms. A few examples of such quantum logic operators are shown in Fig. 12.
In addition to the X gate, another useful basic gate with which we can build
Boolean expressions is the Rz(ϑ) gate. This gate rotates the state vector of a qubit
around the z-axis of the Bloch sphere by a given angle ϑ. In cases where the angle
ϑ is equal to 180° then the Rz(ϑ) gate is often called as the Z gate (Fig. 13).
Rotations around the z-axis represent changes in the phase of a qubit. For a qubit
in uniform superposition with an equal probability to be measured |0 or |1, the state
vector will point to the equator line. The Z gate thus reverses the phase of the qubit,
while maintaining its superposition.
Boolean gates built using controlled Z gates encode the results of operations in
the phases of the qubits. This allows for the representation of multiple outcomes in
superposition, which is something that cannot be done using digital bits. Hence, the
beauty and processing power of quantum computing.
Fig. 12 Quantum logic operators built with X gates. It is assumed that qubit c is initialized to |0
316 E. R. Miranda
Fig. 13 The Z gate rotates a qubit’s state vector by 180° around the z-axis. The opposite ends for
this state vector are notated as |+ and |−
A controlled Z gate acts only when both, the control qubit(s) and the target qubit are
pointing to |1. Therefore, the logic circuit will change the phases only of those qubits
that satisfy the operations they represent. These changes are referred to as ‘spin-
marking’ the possible outcomes. Examples of logic operators built with controlled
Z gates are shown in Fig. 14.
Notice that one cannot simply link sequences logic statements using a combination
of controlled X and Z gates in a circuit. Their ‘inputs’ and ‘outputs’ are incompatible
because they operate on different computational basis.
Bear in mind that rotations on the z-axis do not affect the amplitudes of the state
vector on the standard computational basis (z-axis of the Bloch sphere). And if
one measures a phase logic operator on the standard basis, phase information is
In order to explain how to build logical expressions from EEG information auto-
matically, let us begin by defining an example of a scheme for representing EEG
information as logic expressions with only three clauses and three logical variables
(A, B, and C) as follows:
Each clause has two terms of the form (T er m 1 ∨ T er m 2 ). Below are examples
of logic expressions in the proposed format:
4 In the case of controlled Z gates, the black dot and the boxed circuit representation have the same
effect; sometimes only a vertical line of black dots are used.
318 E. R. Miranda
Next, consider that the clauses represent EEG information captured by electrodes
positioned at specific places on the head (Fig. 3), as shown in Table 4 and illustrated
in Fig. 16.
In this example, we are dealing with a BCI system that banks on EEG beta rhythms
in order to activate commands to control a hypothetical system.
The terms of the clauses are stipulated as follows: for each clause, the system
selects the two electrodes with the two highest EEG amplitudes. As an example, let
us pretend that at a certain moment in time the electrodes Fp1 and O1 registered
the highest amplitudes. Then, the system analyses the spectrum of the EEG from
these two electrodes and extracts the most prominent frequency component in the
spectrum, for each electrode. Let us say that the strongest component for Fp1 was
33.18 Hz and for O1 was 23.61 Hz. These correspond to terms A and C of Clause1 ;
that is, A = 33.18 Hz and C = 23.61 Hz.
Next, the system checks if the frequencies conveyed by the respective terms are
beta rhythms. If a frequency is equal to, or higher than, 15 Hz, then the respective
term is True. Otherwise, it is False. In this case A = True and C = True. Therefore,
these two terms form the clause (A ∨ C). Had Fp1 been equal to, say, 10.36 Hz, then
this term would have been (¬A ∨ C) instead.
For this example, the EEG analysis for the nine electrodes resulted in the following
expression: (A∨B)∧(¬B∨¬C)∧(A∨C). The next step is to verify if this expression
is satisfiable. In other words, the system checks if those EEG values would render
this expression True. Let us examine how the system generates a quantum circuit to
check the satisfiability of logical expressions.
The system uses X and Z gates, and their multiple-controlled forms, to build logical
expressions. The one introduced above are formed by three clauses linked by conjunc-
tion operators. And each of the three clauses are formed by terms linked by disjunction
operators.
The three disjunction clauses are built with X gates. And then, they are linked
with conjunction operators implemented with a controlled Z gate. The circuit requires
only 6 qubits: three of them (q0, q1 and q2) represent the logical variables A, B and
C. The other three (q3, q4 and q5) are ancillary qubits, which are used to represent
the results of the disjunction clauses. The full circuit is shown in Fig. 17 and the
respective Quil5 code in Code 1.
To begin with, all qubits are initialized to |0 and the ones representing the logical
variables are put in uniform superposition with the H gate.
The disjunction clauses are specified, one after the other, as shown in dashed boxes
at the top of Fig. 17. The results from each clause are held in ancillary qubits: N1 in q3,
N2 in q4 and N3 in q5. Then, the tripartite conjunction operation is implemented with
a 3-qubit controlled Z gate applied to the ancillary qubits: they will be spin-marked
by the phase-logic AND operation.
Next, the three disjunction clauses need to be uncomputed in order to return
the ancillary qubits to their initial states. Uncomputation is achieved by running the
respective logic sub-circuits back to front; the ‘Inverted clauses’ in the middle section
of Fig. 17. Then, amplitude amplification is applied to the logic variables to reveal
the spin-marked qubits in terms of amplitudes. Finally, the qubits representing the
logic values are measured to yield the result.
A Quantum Computing Approach to Harnessing the Logic … 321
Fig. 18 Result yielded by running the circuit example for 5,000 shots. Binary numbers were
converted to decimals for plotting on the horizontal axis. Vertical axis are the times each of the
results were observed
The ancillary qubits are not part of the result. Therefore, they are not measured
in the end. However, they can interfere with the measurements because they are
entangled with the qubits that represent the logical variables. Hence, the ancillary
qubits must be uncomputed to make them return to their unentangled state.6
In this example, the circuit spin-marks the quantum states |001, |011 and |101.
Thus, each time the algorithm runs, it will output either 001, 011 and 101 with much
higher probability than any of the other possible values. The plot in Fig. 18 shows the
times that each of the possible outcomes were observed after running and measuring
the circuit for 5,000 times, or 5,000 ‘shots’ in quantum computing terminology, on
a Rigetti’s Quantum Virtual Machine (QVM).7 See Appendix 1 for more examples.
What does this result mean? Bearing in mind that in quantum computing qubit
strings are enumerated from the right end of the string to the left (i.e. |CBA), the
expression (A ∨ B) ∧ (¬B ∨ ¬C) ∧ (A ∨ C) is satisfied when:
Case 1, |001: A = True, B = False and C = False.
Case 2, |011: A = True, B = True and C = False.
Case 3, |101: A = True, B = False and C = True.
In other words, a BCI command associated with this expression is triggered when
significant beta rhythms are detected as follows:
(a) by electrodes Fp1 and Fp2 (Case 1)
6 Refer to (Johnston et al., 2019) for a didactic discussion on uncomputing ancillary qubits and
implications to measurement.
7 The Rigetti Quantum Virtual Machine is an implementation of the Quantum Abstract Machine
described in (Smith et al., 2017). Noise simulation models to account for the effect of quantum
hardware decoherence were not used in the examples discussed in this chapter.
322 E. R. Miranda
In the example presented in Sect. 6, the system spin-marked the quantum states |001,
|011 and |101. As shown in Fig. 18, it outputted 001, 011 and 101 significantly
more times than any of the other possible values. But there are five outliers, whose
probability of being observed were nearly 12% of the spin-marked ones. These
outliers take place due to the very nature of the amplitude amplification algorithm
(Johnston et al., 2019).
The probabilities of the outliers appearing could have been squashed further by
repeating the circuit between the initial Hadamard gates and the measurement section
a few times. What happens here is that interference can amplify the amplitudes (or
‘probabilities’) of spin-marked qubits and decrease the amplitudes of all others.
For the example above, repeating the circuit three times squashes the outliers to
negligible levels, as shown in Fig. 19. However, one must exercise caution here.
The number of repetitions needs to be specified carefully. In this particular case, if
the circuit is repeated only twice, then the probabilities of the outliers would have
increased instead. Moreover, a more important caveat is that such repetitions increase
circuit’s depth (i.e. the overall length of the circuit) considerably. On quantum
hardware, this escalates significantly the likelihood of errors due to decoherence.
The problem of decoherence poses severe limitations to the quantity of successive
gates that can be used in a circuit for a real quantum processor. The higher the number
of gates sequenced one after the other, the higher the likelihood of decoherence to
occur.
A Quantum Computing Approach to Harnessing the Logic … 323
Fig. 19 Results yielded by running three copies of the circuit in Fig. 17 for 5,000 shots on a Rigetti’s
QVM. Binary numbers were converted to decimals for plotting on the horizontal axis. Vertical axis
are the times each of the results were observed
At the time of writing, quantum processors struggle to maintain coherence for more
than a dozen successive gates involving superposition and entanglement. Currently,
coherence is assessed in terms of a few microseconds rather than seconds. However,
research is progressing fast to improve this (Cho, 2020). In addition to improving
hardware technology, there is much research activity to develop efficient quantum
error correction methods.
In order to illustrate how challenging the problem of decoherence is, let us
examine what happened when we ran Code 1 on the state of the art Rigetti’s Aspen-8
quantum chip, with no added noise correction algorithm. Aspen-8 affords three-
fold connectivity (i.e. supports 3-qubit gates) and coherence times of approximately
20 µs.
Figure 20 plots the outcomes after running the code for 2,048 shots on the quantum
chip. The results do not match the ones obtained with the QVM simulator; they are
all over the place. This is because the circuit is indeed too deep for this chip.
In truth, even the circuit for the first logic clause (A ∨ B) is problematic. The Quil
code to examine the satisfiability of this simple clause is shown in Code 2. Table 5
shows the outcomes that satisfy this clause. Thus, one can predict that each time we
run Code 2, it would output either 000, 101, 110, 111. Indeed, Fig. 21 shows the times
that each of the possible outcomes were observed after running the circuit for 2,048
shots on Rigetti’s QVM. They match the prediction perfectly. Yet, the outcomes from
running exactly the same code on Aspen-8 are not accurate (Fig. 22). Why is this so?
324 E. R. Miranda
Fig. 20 Results from running the circuit shown in Fig. 17 for 2,048 shots on a Rigetti’s Aspen-
8 quantum chip. Binary numbers were converted to decimals for plotting on the horizontal axis.
Vertical axis are the times each of the results were observed
Fig. 21 Results from running the circuit shown in Code 2 for 2,048 shots on Rigetti’s Quantum
Virtual Machine (QVM). Binary numbers were converted to decimals for plotting on the horizontal
axis. Vertical axis are the times each of the results were observed
Fig. 22 Results from running the circuit shown in Code 2 for 2048 shots for 2048 trials on Rigetti’s
Aspen-8 quantum chip. Binary numbers were converted to decimals for plotting on the horizontal
axis. Vertical axis are the times each of the results were observed
326 E. R. Miranda
The expressions in Sect. 6 are limited to three logic variables. One may not even need
a computer to check their satisfiability, even less so a quantum computer. Should the
clauses have entailed a unique variable for each electrode, then the task of checking
the expressions’ satisfiability would have been harder. But still, not hard enough to
justify the need for a quantum computer.
However, if the expressions encompassed the whole set of 20 electrodes shown in
Fig. 3, each of which corresponding to a unique logic variable, then the satisfiability
problem would become considerably harder. In this case a quantum computer could
well be advantageous.
8 XY(ϑ) produces a parameterized iSWAP gate, which is not priority for discussion on this chapter.
For more details, see (Abrams et al., 2019).
9 A transpiler is a program that translates a piece of code into another at the same level of abstraction.
It is different from a compiler whose output is in a lower level of abstraction than the input.
A Quantum Computing Approach to Harnessing the Logic … 327
This section presents a practical application of the logic of the mind: a BCI sound
synthesiser.
We developed a method to render sounds with the results from checking the
satisfiability of the kinds of logical expressions introduced in Sects. 5 and 6 (Fig. 23).
The system generates sounds using an additive synthesiser. The additive synthesis
technique is informed by the theory of Fast Fourier Transform, or FFT (Muller, 2015).
It is based on the notion that sounds can be characterized as a sum of sine waves.
Additive synthesis works by deploying a number of sine wave sound generators
(e.g. digital oscillators) to produce partials, which are added up to produce the final
328 E. R. Miranda
result. A sine wave is characterised by a frequency value (i.e. speed of the cycle) and
an amplitude value (i.e. strength of the signal).
Our synthesiser comprises eight digital oscillators, each of which requires a
frequency value in Hertz (Hz) and an amplitude, whose value varies between 0
(silence) and 1 (loudest). The outputs are summed and a Hanning function is applied
to the result to give a smooth fade in and fade out bell shape to the sound.
In Fig. 24, individual partials are represented on the left-hand side of the figure,
where a bar on the ‘freq’ axis (frequency domain) has a certain magnitude on the
‘amp’ axis (amplitude domain). The spectrum of the resulting sound is represented
on the right-hand side; it contains eight partials. A schematic representation of the
resulting sound, depicting its bell-like shape is also shown.
For every lapse of time the system checks the satisfiability of the respective logic
expression, as discussed earlier, and uses the results to activate the oscillators of the
synthesiser. Figure 25 delineates how the qubit measurements plotted in Fig. 18 are
associated with oscillators. Notice that there are as many oscillators as the number
of different quantum states that the respective circuit can return: in this case, 23 = 8.
That is, each possible output is associated with a different partial of the resulting
sound. In additive synthesis, changes to the frequencies and amplitudes of the oscil-
lators modify the timbre of the resulting sound. A schematic representation of this
is given in Fig. 26.
Each oscillator of the synthesizer is assigned a frequency; for instance: osc 1 =
55.0 Hz, osc 2 = 164.81, osc 3 = 329.63 Hz and so on. This is fully customisable, and
the system can be set to change these frequencies algorithmically. The amplitudes are
normalized proportionally to the number of times the respective quantum state was
observed after a number of pre-specified shots. Thus, in Fig. 25, the first oscillator,
whose amplitude is controlled by the number of times the state |000 was observed,
A Quantum Computing Approach to Harnessing the Logic … 329
will produce a partial that is much quieter than the second oscillator, whose amplitude
is given by number of times that state |001 was observed.
The duration of the time lapses is also customisable. However, it should not be
shorter than the window of time that is required to read and analyse EEG data.
Our experiments suggest that at least 1 s is needed to capture sufficient EEG data
to associate with a mental state. But in order to perceive differences clearly in the
timbre of the sounds, it is suggested to let it play for no less than 5 s. Of course, time
can vary and can be defined algorithmically should one wish to do so.
As a sound is being played, the system processes the EEG for the next time lapse,
builds and runs the circuit, and synthesizes a new sound immediately after the current
one. And the cycle continues for as long as required.
330 E. R. Miranda
Fig. 26 Schematic representation of a sequence of five sounds and their respective partials shown
above the time lapse line
A Quantum Computing Approach to Harnessing the Logic … 331
9 Concluding Remarks
This chapter presented an approach to interfacing the brain and a quantum computer.
Central to this approach is a method to encode brain activity as logic expressions
representing states of mind. These expressions are associated with commands to
control systems and/or machines with brain signals.
In addition to building brain–computer interfaces, the chapter put forward the
notion of a logic of the mind as a means to study neural correlates of the mind
and brain functioning, with potential benefits to medicine; e.g. for diagnosing brain
disorder. The EEG is an important physiological signal for diagnostics (Tatum IV,
2008). It is used to detect traces of epilepsy, dementia, cancer, inflammation, sleep
disorders and more. For instance, Clarke et al. (2013) reported that persistent excess
of beta rhythms at the frontal regions of the brain can indicate hyperactivity disorder.
It would be straightforward to encode this as a logic of the mind expression.
The EEG is currently the best signal for BCI systems. However, it is not necessarily
the best for research into understanding brain functioning and mental correlates. For
instance, functional magnetic resonance imaging (fMRI) measures the minuscule
variations in blood flow that occurs with brain activity. It is used to determine which
parts of the brain are more active than others when handling particular tasks. It has
higher spatial resolution than EEG has, and it can detect information deep inside the
brain.
The logic of the mind requires the processing of large Boolean expressions, which
are computationally very demanding. Even more so if we were to use fMRI scanning.
Today’s average classical computer would take hours, if not days, to evaluate a
handful of expressions. Hence, the rationale for using quantum computers.
At the time of writing, quantum computing hardware is of limited capacity for
realistic applications. But once fully developed, quantum computers promise to be
considerably faster than their classical counterparts to run certain types of algorithms,
such as the one introduced in this chapter. In the meantime, simulations on virtual
quantum machines enable research and development of quantum algorithms, which
would eventually run optimally when robust quantum hardware is available.
By way of a practical demonstration, we presented a BCI system that renders the
logic of the mind into sounds. We introduced a method to sonify quantum measure-
ments obtained from evaluating logic expressions. In this case, the states of mind
do not activate control commands. Instead, they are mapped directly onto sound.
Effectively, the system can be thought of as a novel musical instrument.
Incidentally, the sound synthesis technique introduced above is a contribution to
the field quantum computing on its own right, in the sense that it provides a method
to represent the wavefunction of a quantum system auditorily.
Acknowledgements The author would like to thank Mathew Wilson and James Hefford of the
Department of Computer Science at the University of Oxford for reviewing the quantum computing
content of this chapter. Also, many thanks to Palaniappan Ramaswamy of the School of Computing at
the University of Kent for scrutinizing the BCI content. Their meticulous comments and suggestions
contributed significantly to the clarity and rigour of this work.
332 E. R. Miranda
Below are four examples of results taken from a run of the BCI system presented in
this chapter, using Rigetti’s QVM.
In relation to the BCI musical instrument, the different results produced variations
in the timbre of the sounds. Recordings of the respective sounds, and other examples,
are available online at SoundClick: https://soundclick.com/LogicoftheMind.
Example 1
This is an example where the logic expression is unsatisfiable (Table 6 and Fig. 27).
Fig. 27 Results from running the circuit for Example 1 for 1,000 times. Binary numbers were
converted to decimals for plotting on the horizontal axis. Vertical axis are the times each of the
results were observed
334 E. R. Miranda
Example 2
Fig. 28 Results from running the circuit for Example 2 for 1,000 times. Binary numbers were
converted to decimals for plotting on the horizontal axis. Vertical axis are the times each of the
results were observed
A Quantum Computing Approach to Harnessing the Logic … 335
Example 3
Fig. 29 Results from running the circuit for Example 3 for 1,000 times. Binary numbers were
converted to decimals for plotting on the horizontal axis. Vertical axis are the times each of the
results were observed
A Quantum Computing Approach to Harnessing the Logic … 337
Example 4
Fig. 30 Results from running the circuit for Example 4 for 1,000 times. Binary numbers were
converted to decimals for plotting on the horizontal axis. Vertical axis are the times each of the
results were observed
338 E. R. Miranda
Figure 31 shows the results from running the excerpt corresponding to the logic clause
(A ∨ B) on an IBM quantum computer simulator, for 2,048 shots. The Quil code
shown in Code 2, in Sect. 7.2, was translated into the IBM’s OpenQASM language,
as shown in Code 8.
Not surprisingly, the results shown in Fig. 31 are identical to those obtained with
Rigetti’s QMV, as shown in Fig. 21.
The outcomes from running the exact OpenQASM code on IBM’s Santiago
processor (ibmq_santiago v1.1.1) is shown Fig. 32. Although there were errors,
in overall the results are comparable to those produced by the simulator. IBM’s
Santiago processor proved to be more resilient than Rigetti’s Aspen-8 in this case.
A Quantum Computing Approach to Harnessing the Logic … 339
Fig. 31 Results from running the circuit for the logic clause (A ∨ B) for 2,048 trials on IBM’s
simulator
However, it should be noted the programming for both machines were kept simple,
with no error correction and no a priori hardware calibration. Aspen-8 might have
performed just as well with appropriate calibration and noise mitigation procedures.
A detailed technical discussion on the idiosyncrasies of these machines, calibration,
error mitigation strategies and so on, is beyond the scope of this chapter.
Fig. 32 Results from running the circuit for the logic clause (A ∨ B) for 2,048 trials on IBM’s
Santiago processor
340 E. R. Miranda
In order to assess how well IBM’s hardware would fare with a deeper circuit,
we added a third logic variable and a phase-logic AND operation to implement the
expression ((A ∨ B) ∧ C), as shown in Fig. 33.
Considering that qubit strings are enumerated from the right end of the string to
the left (i.e. |CBA), the expression ((A ∨ B) ∧ C is satisfied when:
Case 1, |101: A = True, B = False and C = True.
Case 2, |110: A = False, B = True and C = True.
Case 3, |111: A = True, B = True and C = True.
Figure 34 shows the outcomes from running the circuit on the IBM quantum
computer simulator, for 2,048 shots. Not surprisingly, the measurements are fairly
accurate. Yet, the outcomes from running exactly the same code on IBM’s Santiago
processor (ibmq_santiago v1.1.1) are not accurate (Fig. 35). The transpiled code
comprised 75 gates; the circuit is too deep for this processor.
Fig. 34 Results from running the circuit in Fig. 33 for 2,048 trials on IBM’s simulator
A Quantum Computing Approach to Harnessing the Logic … 341
Fig. 35 Results from running the circuit in Fig. 33 for 2048 trials on IBM’s Santiago processor
References
Abrams, D. M., Didier, N., Johnson, B. R., da Silva, M. P., & Ryan, C. A. (2019). Implementation
of the XY interaction family with calibration of a single pulse. Retrieved October 24, 2020, from
https://arxiv.org/abs/1912.04424v1
Aharonov, D. (2003). A simple proof that Toffoli and Hadamard are quantum universal. https://
arxiv.org/abs/quant-ph/0301040
Akerkar, R., & Akerkar, R. (2004). Discrete mathematics. Person Education India.
Anderson, C., & Sijercic, Z. (1996). Classification of EEG signals from four subjects during
five mental tasks. In Solving Engineering Problems with Neural Networks: Proceedings of the
Conference on Engineering Applications in Neural Networks (EANN’96), London, UK.
Ball, P. (2014). Questioning quantum speed. Physics World, January: 38–41.
Bernhardt, C. (2019). Quantum computing for everyone. The MIT Press.
Berta, R., Bellotti, F., De Gloria, A., Pranantha, D., & Schatten, C. (2013). Electroencephalo-
gram and physiological signal analysis for assessing flow in games. IEEE Transactions on
Computational Intelligence and AI in Games, 5(2), 164–175.
Cho, A. (2020). IBM promises 1000-qubit quantum computer—A milestone—By 2023.
ScienceMag.org 15 Sep 2020. Retrieved October 30, 2020, from https://www.sciencemag.org/
news/2020/09/ibm-promises-1000-qubit-quantum-computer-milestone-2023
Clarke, A. R., Barry, R. J., Dupuy, F. E., McCarthy, R., Selikowitz, M., & Johnstone, S. J. (2013).
Excess beta activity in the EEG of children with attention-deficit/hyperactivity disorder: A
disorder of arousal? International Journal of Psychophysiology, 89(3), 314–319.
Daly, I., Williams, D., Malik, A., Weaver, J., Kirke, A., Hwang, F., Miranda, E. R., & Nasuto, S.
J. (2018). Personalised, multi-modal, affective state detection for hybrid brain-computer music
interfacing. IEEE Transactions on Affective Computing, 11(1), 111–124.
Giannitrapani, D. (1985). The electrophysiology of intellectual functions. Karger.
Grover, L. K. (1996). A fast quantum mechanical algorithm for database search. In Proceedings of
28th Annual ACM Symposium on the Theory of Computing. https://arxiv.org/abs/quant-ph/960
5043
342 E. R. Miranda
Grumbling, E., & Horowitz, M. (Eds.). (2019). Quantum computing: Progress and prospects.
National Academies Press. https://doi.org/10.17226/25196
Guenter, F. H.„ & Brumberg, J. S. (2011). Brain-machine interfaces for real-time speech synthesis.
In 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology
Society, Boston, MA, pp. 5360–5363.
Hasan, B. A. S., & Gan, J. Q. (2012). Hangman BCI: An unsupervised adaptive self-paced brain-
computer interface for playing games. Computers in Biology and Medicine, 42(5), 598–606.
Hinterberger, T., Zlabinger, M., & Blasser, K. (2014). Neurophysiological correlates of various
mental perspectives. Frontiers in Human Neuroscience, 8(637).
Jeunet, C., Tonin, L., Albert, L., Chavarriaga, R., Bideau, B., Argelaguet, F., Millan, J. del. R.,
Lecuyer, A., & Kulpa, R. (2020). Uncovering EEG correlates of covert attention in soccer
goalkeepers: Towards innovative sport training procedures. Scientific Reports, 10(1705).
Johnston, E. R., Harrigan, N., & Gimeno-Segovia, M. (2019). Programming quantum computers.
O’Reilly Media Inc.
Kraft, E., Gulyás, B., & Pöppel, E. (2008). Neural correlates of thinking. Spriger Science & Business
Media.
Kropotov, J. (2008). Quantitative EEG, event-related potentials and neurotherapy. Academic Press.
Ligeza, T., Wyczesany, M., Tymorek, A. D., & Kaminski, M. (2015). Interactions between the
prefrontal cortex and attentional systems during volitional affective regulation: An effective
connectivity reappraisal study. Brain Topography, 29, 253–261.
Marcuse, L. V., Fields, M. C., & Yoo, J. J. (2015). Rowan’s primer of EEG (2nd ed.). Elsevier.
McFarland, D. J., & Wolpaw, J. R. (2017). EEG-based brain-computer interfaces. Current Opinion
in Biomedical Engineering, 4, 194–200.
Mermin, N. D. (2007). Quantum computer science: An introduction. Cambridge University Press.
Miranda, E. R. (2006). Brain-computer music interface for composition and performance.
International Journal on Disability and Human Development, 5(2), 119–125.
Miranda, E. R., Magee, W. L., Wilson, J. J., & Palaniappan, R. (2011). Brain-computer music
interfacing (BCMI): From basic research to the real world of special needs. Music and Medicine,
3(3), 134–140.
Muller, M. (2015). Fundamentals of music processing: Audio, analysis, algorithms, applications.
Springer.
Nicolas-Alonso, L. F., & Gomez-Gil, J. (2012). Brain computer interfaces, a review. Sensors, 12(2),
1211–1279.
Oh, S.-H., Lee, Y.-R., & Kim, H.-N. (2014). A novel EEG feature extraction method using Hjorth
parameter. International Journal of Electronics and Electrical Engineering, 2(2), 106–110.
https://doi.org/10.12720/ijeee.2.2.106-110
Palaniappan, R. (2008). Two-stage biometric authentication method using thought activity brain
waves. International Journal of Neural Systems, 18(1), 59–66.
Petsche, H., & Etlinger, S. C. (1998). EEG and thinking: Power and coherence analysis of cognitive
processes. Austrian Academy of Sciences.
Renard, Y., Lotte, F., Gibert, G., Congedo, M., Maby, E., Delannoy, V., Bertrand, O., & Lecuyer,
A. (2010). OpenViBE: An open-source software platform to design, test, and use brain-computer
interfaces in real and virtual environments. Presence, 19(1), 35–53.
Rieffel, E., & Polak, W. (2011). Quantum computing: A gentle introduction. The MIT Press.
Rozemond, M. (1988). Descartes’s dualism (p. 1998). Harvard University Press.
Smith, P. (2020). An introduction to formal logic. Logic matters, 2nd edn. Retrieved October 10,
2020, from https://www.logicmatters.net/ifl/
Smith, R., Curtis, M. J., & Zeng, W. (2017). A practical quantum instruction set architecture.
Retrieved October 13, 2020, from https://arxiv.org/abs/1608.03355v2
So, W. K. Y., Wong, S. W. H., Mak, J. N., & Chan, R. H. M. (2017). An evaluation of mental
workload with frontal EEG. PLoS One, 12(4), e0174949.
Squire, L., Berg, D., Bloom, F., du Lac, S., Ghosh, A., & Spitzer, N. (Eds.). (2008). Fundamental
neuroscience (3rd ed.). Academic Press.
A Quantum Computing Approach to Harnessing the Logic … 343
Vartanian, O., Bristol, A. S., & Kaufman, J. C. (Eds.). (2013). Neuroscience of creativity. The MIT
Press.
Wilson, J. J., & Palaniappan, R. (2011). Analogue mouse pointer control via an online steady state
visual evoked potential (SSVEP) brain–computer interface. Journal of Neural Engineering, 8(2),
025026.
Yelamanchili, T. (2018). Neural correlates of flow, boredom, and anxiety in gaming: An elec-
troencephalogram study. Missouri University of Science and Technology, Masters Theses 7812.
Retrieved October 10, 2020, from https://scholarsmine.mst.edu/masters_theses/7812
The History of Games for (Quantum)
Computers
James R. Wootton
Abstract Computer games are not just one application of computers, they are a
multitude. A wide variety of computational tasks are combined, all running as fast
as possible, to deliver the best possible experience to the player. It is reasonable to
expect that, somewhere in this maelstrom of computation, we can find something
that quantum computers will excel at. If the quest to combine games and quantum
computers was a game itself, it would be one of open-world exploration. Though
there may be no well-defined, concrete goals to achieve, there is something that
will guide us: this game is a sequel. By looking at how we combined games with
computers the first time, we can get some idea of what we can expect in this quantum
successor.
1 Introduction
Computer games are not just one application of computers, they are a multitude. A
wide variety of computational tasks are combined, all running as fast as possible,
to deliver the best possible experience to the player. It is reasonable to expect that,
somewhere in this maelstrom of computation, we can find something that quantum
computers will excel at.
With this we begin the adventure! If the quest to combine games and quantum
computers was a game itself, it would be one of open-world exploration. Though
there may be no well-defined, concrete goals to achieve, there is something that
will guide us: this game is a sequel! By looking at how we combined games with
computers the first time, we can get some idea of what we can expect in this quantum
successor.
J. R. Wootton (B)
IBM Quantum, IBM Research – Zurich, Zurich, Switzerland
e-mail: [email protected]
Since the early history of computing, people have been interested in how it relates to
games. This has come in multiple forms: Can computers implement games? Can they
play games, and perhaps even beat humans? Could they offer experiences that would
be impossible without them? We now know that the answer to all these questions is
‘yes’ since the last few decades have conclusively shown what computers can do for
games.
Back in the 1950s, however, the combination of games and computers got off to a
slow start. Early examples were less about what computers could do for games and
more about what games could do for computers.
One of the first examples was Bertie the Brain exhibited at the Canadian National
Exhibition in 1950. It used vacuum tubes and light bulbs and played Tic-tac-toe. Not
because it offered a better play experience than using pen and paper, but because
it helped to show off the vacuum tubes. The fact that we are writing about it now,
over 70 years later, shows that it certainly had some impact. It showed us the first
possible reason for making games with computers: to showcase new technology.
Unfortunately, in this case, however, the specific technology it was showcasing soon
became obsolete.
Then, there was Nimrod in 1951. Again, vacuum tubes and light bulbs, but this time
the game was Nim. It was shown at the Festival of Britain with a stated aim, among
other things, to ‘…illustrate the algorithm and programming principles involved’.
This gave us a second reason to make games with computers: for education. Program-
ming would have seemed a strange and arcane art in those days. How better to make
it relatable than to show it playing a game?
Next was a research project in 1952 at Cambridge University, now commonly
referred to as OXO. The idea was to study human–computer interaction, done using
another implementation of tic-tac-toe (or ‘noughts and crosses’ as it is known in
Cambridge). It again used vacuum tubes, but this time had cathode ray displays: the
first graphical upgrade of the games industry. Nevertheless, the driving principle was
research, giving us another reason to make games with computers.
Another project to note was that done at IBM Research in the 1950s: making
artificial intelligence (AI) to play Checkers and beat a human opponent (Samuel,
1959). This would be something that IBM would repeat decades later, using Deep
Blue to beat Gary Kasparov at chess, and then Watson to beat champions at Jeopardy.
DeepMind has also been prominent in this area, using AlphaGo to beat a champion
at Go. All of which are more examples of using games to showcase technology.
There were more games in the 1950s, but the ones here show what the decade was
all about. Whether it was to showcase the technology, to educate about the technology
or to research the technology, games were used to help computers rather than the
other way around. An exception was Tennis for Two, a game made in 1958 for fun;
and it signalled what was coming.
The History of Games for (Quantum) Computers 347
In 1962, a team at MIT decided to put their new PDP-1 computer to good use. They
embarked on a project that would:
• Test out the new device (and later, new installations).
• Showcase its capabilities.
• Be fun!
They created a game. But it wasn’t just an existing game, with a computer-based
implementation tacked on as a gimmick. It was something new: space battles without
the need to build your own spaceship. It was called Spacewar.
This was the first clear example of computers doing something for games. They
were used to make something new, and something fun. Functions made for trigono-
metric calculations were repurposed to let the player experience flying around a
star.
It was not the only game this decade to be set in space. There was also the original
Lunar Lander as well as the unimaginatively named Space Travel. More down to
Earth was an economics simulator, The Sumerian Game, which was an early ancestor
of games like SimCity.
In each case, the computer was used to do exactly what computers were made
for: crunching numbers and performing simulations. But rather than being used for
payroll calculations or planning a real trip to the moon, the computers were used to
provide unique play experiences.
Not many people know about Spacewar! or Space Travel. Any familiarity with Lunar
Lander is usually due to one of its later incarnations, rather than the 1960s text-based
original. This was not due to any secrecy on the part of the developers: they usually
freely shared the source code. It was more because few had the room for a room-sized
computer, never mind the money.
The 1970s is when it became possible to bring the hardware to the masses. The first
arcade game was Computer Space, conceived of as a coin-operated Spacewar! that
ran on hardware dedicated only to the game. The first generation of home consoles
soon followed. With these breakthroughs, computer games became something you
could sell to the masses, either in an arcade or at home. The games industry had truly
begun.
Something else also started in the 1970s, when Paul Benioff began to think
about computers that operate according to the principles of quantum mechanics
(Benioff, 1980). It has taken a long time since then to fully formulate the idea,
explore applications and develop the required hardware. But we are finally close to
fruition.
348 J. R. Wootton
By 2010 we already had many examples of algorithms that could run on quantum
computers, to solve problems that would be intractable for even the biggest supercom-
puter; see here for a few examples (Jordan, 2021). However, the hardware required
to actually run these algorithms was far beyond what was available. Instead of the
many thousands of high-quality qubits that are needed, just having two qubits was
pretty impressive in 2010. This hardware was also not easy to use, being hidden away
in labs around the world and programmed primarily through the manual labour of
research students.
The state of affairs did not seem much better in 2015 when this author, a
theoretical physicist working on quantum error correction, proposed a quantum
computing experiment for 5-qubits (Wootton, 2016). Proposing an experiment was
easy. Convincing a research group with the required hardware to actually run it would
be much harder. And running it would be harder still.
This all changed in 2016 when IBM released their Quantum Experience: A web
interface allowing direct access to a 5-qubit device (Quantum Composer and Lab).
On one fateful day, just before lunchtime, this author suddenly realized that this
provided exactly what was required for his proposed experiment. By the time he left
the work, the experiment had been run. A revolution had begun!
Since then, a great many others like this author have used direct and simple access
to quantum hardware as part of their scientific research. But that is beyond the scope
of this chapter. With access to new computational hardware, there are also other
things we can do. It took us until early 2017 to realize that this included making
games.
Most probably the first-ever game using a quantum computer is a simple version of
Battleships (Wootton, 2017). The display was just ASCII in a terminal. The gameplay
was no better than you could do with pen and paper. Rather than being made to be
played, it was made to be the basis of a blog post.
The idea was to give an example of quantum programming that wasn’t a complex
algorithm, but instead something more tangible. The states of a qubit were used to
encode whether a ship had been sunk or not. Quantum gates were used to imple-
ment attacks. Without realizing it at the time, the author repeated the history of
Nimrod, mentioned above. Just as normal programming seemed like an arcane art
then, quantum computing can seem so now. So the game was designed to ‘illustrate
the algorithm and programming principles involved’. It was a quantum game made
for the purpose of education.
The next major quantum game was Quantum Awesomeness (Wootton, 2018).
This was based on the fact that there are many aspects of a quantum processor that
determine how useful it is: number of qubits, which pairs of qubits we are allowed to
apply a two-qubit manipulation to, and what imperfections are present. Thus, a puzzle
game was designed to help people get an idea of how different devices compare.
The History of Games for (Quantum) Computers 349
The game was designed so that the puzzles were based on the specifications of
the device used to run it. The more qubits you had, and the more complex the layout
of possible two-qubit gates, the better the puzzles would be. The more imperfections
there were, the less it would seem to make any sense. Therefore, to compare devices
and see which was best, one only needed to see which was the most fun. It is not as
informative as looking at benchmarking data, but it is a lot more accessible to the
public.
An example can be seen in Fig. 1. The coloured circles represent qubits on the
device. The connections between them show us which pairs of qubits may be directly
manipulated through two-qubit gates. And hence the device is itself the puzzle board.
The principle behind the puzzle is simple: a random pairing of the qubits is chosen,
and the player has to work out what the pairing is. This is implemented by running a
quantum program whose details are defined by the chosen pairing. From the results,
a number is calculated for each qubit. This number should be exactly equal for the
two qubits of each pair, and so the player’s job should be a simple one: they just need
to find the pairs of equal numbers.
For an example such as the one above, finding this pairing is indeed a very easy job;
it is the pairs labelled H, I, J, K, L, M G and V if the reader is wondering. However,
these results come from a simulation of a quantum computer, done without any of the
imperfections inherent in real quantum hardware. An example from real hardware
(with a different pairing) can be seen in Fig. 2.
In the case shown in Fig. 2, the numbers that should be equal are not as equal
as they should be. We see that some pairs of qubits are behaving fairly well. Others
qubits, like the one whose number is 14, seem to be giving us total nonsense. This,
therefore, allows us to get an idea of how well the qubits work on this device, and
also allows us to compare qubits between devices.
This game was designed so that the player didn’t need to have direct access
to the devices. Instead, puzzles could be made once and played later. This is also
demonstrated in the image above, where you can still play the game despite the
device it ran on having been decommissioned some time ago. This was an example
of making quantum games for the purpose of research, repeating the history of the
1950s.
The next game to use a real quantum device was Q|Cards, made in 2019 by
a team working at Quantum Wheel, the Fifth Game Jam held in Helsinki, Finland
(Cards). This is a card game in which each player has a qubit and each card represents
a quantum gate. The quantum computer is used at the very end to judge who played
their gates best.
It is a fun game to play, but it also helps to teach people about quantum gates.
Instead of getting them to sit through a lecture or read through a textbook, playing
the game helps us explain the basics of quantum computing in a friendly atmosphere.
Another quantum game made for education.
The motivation behind these games was exactly the same as in the 1950s: education
and research. They also served the other big purpose of 1950s games, to showcase
the new technology, since their very existence helped to get the word out in the form
of blog posts, tweets and more.
The short summaries of the above-mentioned games are intended simply to give
a flavour of the quantum games of the 2010s. We will look at one of these examples
in detail later in this chapter.
If the history of computer games began in the 1950s, then the 1940s was its prehistory.
One relic of this era was Turochamp, a chess AI developed by Alan Turing and David
Champernowne (Copeland, 2004). It had one big problem: it was too complex for
the computer hardware of the time. It was run in 1948, but with a human manually
simulating the processes. This too has a parallel in recent experiments with quantum
games. It is easier to simulate simple quantum programs on a normal computer than
to run them on real quantum hardware. This allows us greater flexibility to explore
what qubits can do, within the familiar and fun context of hacking together simple
games.
Many games have been made in this way; for instance, QPong (a quantum version
of Pong) (QPong), QiskitBlocks (building quantum programs in a Minecraft-like
world) (QiskitBlocks), to cite but two.
One example of our own was not so much a game, but a method for procedural
generation. It used just a single qubit (the smallest element of quantum computation)
to create a landscape to walk around. The generated terrain was far simpler than can
easily be created using non-quantum methods, but that was beside the point. The
project was intended to provide a hands-on means by which people could learn all
the intricacies of single-qubit manipulations, by creating their own terrain generator.
This project will be explained in detail in Sect. 9.
The History of Games for (Quantum) Computers 351
What can quantum computers actually do for games? Like everything, they bring
some benefits but also some design constraints. For one thing, the need to cool the
hardware to near absolute zero means that you’ll never get one inside a home console.
They’ll live on the cloud, and so probably aren’t something that you want to send jobs
to every frame. For any game that might use a quantum computer, the vast majority
of it will still run on a normal computer. We just need to find a small corner that
quantum computers can excel at.
Though speeding up computation is the big promise of quantum computers, the
aim is mostly to take problems that could never otherwise be tackled within a human
lifetime and instead solve them within a reasonable timescale. Tasks that can already
be done very quickly with conventional hardware, will continue to be done with
conventional hardware, with no hope of a quantum speedup.
The most reasonable points in a game to insert quantum results are those where
responses are not expected in less than a second. This includes tasks such as setting
up the levels, puzzles or worlds that the player will experience. These jobs can be
done during a loading screen, or even during the game design process, ensuring that
the quantum computer has time to do its thing. The future of quantum computation
in games, therefore, looks likely to be dominated by procedural generation.
As with any attempt to find applications for quantum computers, we also need to
think about the different eras that the technology will go through. Our main aim is
a future era, where all imperfections are removed by error correction, where you’ll
have more qubits than you’ll ever need, and where you’ll be able to do any two-qubit
gates you want. This is the era of fault-tolerant, scalable quantum computing. With
this hardware, we will be able to implement the many algorithms developed over the
last few decades. By looking at the kinds of problems that they can solve, we can
begin to speculate on how they will be used in procedural generation.
Certain possibilities immediately jump out. Quantum computation has been
shown to offer more efficient computational complexity for constraint satisfiability
problems, which could be useful in searching the space of different forms of content
for those that satisfy desired properties. Also, since many problems in procedural
generation can be expressed in the abstract language of nodes connected by a network,
quantum speed-ups for the analysis of such networks could offer additional tools.
These are motivating examples, but they are admittedly not defined very concretely.
This is because quantum computation is a field that has mostly concerned itself with
asymptotic properties of a computation’s runtime: When thinking about how this
time grows with the size of a problem, we know much about what shape the graph
would be but little about the exact values. Games, on the other hand, are an applica-
tion that depends greatly on the exact speed at which calculations can be performed.
A factor of 10 in speed can be the difference between something being useful or
useless in a game.
352 J. R. Wootton
As discussed in the last section, one reason we might want to combine games with
quantum computers is for education. The most obvious method is to make a game
such that players will learn about quantum computing through play. But people also
enjoy making games, allowing us to also explore a very different possibility: learning
about quantum computing by using it to make a game.
This was the idea behind the pioneering Battleships with partial NOT gates
(Wootton, 2017). It was not made to offer a good player experience but instead
was made to form the basis of a blog post. It was used as a simple and relatable
example of quantum computers doing something, which would be less imposing
than the quantum algorithms of textbooks. It was the Nimrod of quantum games, as
mentioned earlier, built to ‘…illustrate the algorithm and programming principles
involved’.
The game itself is a variant of the abovementioned Battleships. The central idea is
that there are ships that get attacked, and after a given number of hits they will sink.
Everything else, such as inputs, outputs and the handling of turns, is just bookkeeping.
Like any quantum algorithm, any quantum game will require a hybrid approach.
Conventional digital computers will be used for some tasks, and quantum computers
will be used for others. In this case we will use quantum computation for the central
game mechanic. Thus, it will keep track of how damaged each ship is and implement
attacks. Before we do this, let us start with an explanation of what a bit is. Even
The History of Games for (Quantum) Computers 353
though they have become a ubiquitous concept in our increasingly digital world,
they are not something that people spend much time thinking about.
A bit is simply the smallest possible chunk of information. It can take one of two
possible values, which we typically call 0 and 1. We can use bits to store binary
values. For example, if we are recording whether a ship has been destroyed or not,
we could use 1 for it being destroyed and 0 for it being fully intact. To read out the
information, we simply look at the value that was recorded.
The qubits at the heart of quantum computing are the quantum version of a bit.
They too can only give outputs of 0 or 1. So we will use exactly the encoding described
above, but with a qubit instead of a bit: The qubit state is certain to output 0 to encode
the state of a ship that is perfectly undamaged, and that for 1 for a ship that has been
destroyed.
A program for qubits takes the form of a so-called ‘quantum circuit’. These are
like the Boolean circuits with which conventional digital computing can be expressed
when considering very low levels of abstraction. We will use the Qiskit framework
(Qiskit) created by IBM Quantum and built in Python, to write these circuits. In this
Battleships game, each player has five ships, so we begin by creating a circuit for
five qubits:
from qiskit import QuantumCircuit
qc = QuantumCircuit(5)
On its own, we cannot do much with this circuit. We cannot even extract an output
of 0 or 1 from the qubits. For that we would need a circuit which does not just have
a qubit but has a bit on which an output can be written. So, let us modify our circuit
to not only have 5 qubits but also 5 bits to serve as their outputs.
qc = QuantumCircuit(5,5)
To actually extract these outputs, we need to use a so-called measure gate, which
specifies exactly when the process of extracting the output should be performed. For
convenience we will keep this in a separate circuit, defined for 5 qubits (and bits) as
follows.
m_z = QuantumCircuit(5,5)
for j in range(5):
m_z.measure(j,j)
Note that qubits and bits are indexed from 0 in Qiskit. So, the five qubits and
bits are referred to as 0, 1, 2, 3 and 4. The measure(j,j) command is telling the
circuit to extract an output from the jth qubit and write the answer to the jth bit.
After creating a circuit, the final step is to run it. For complex circuits we have
no choice but to use actual quantum hardware for this. For 5 qubits, however, it is
possible to simulate using a conventional digital computer:
from qiskit import Aer
backend = Aer.get_backend(’aer_simulator’)
job = backend.run(qc.compose(m_z), shots=10, memory=True)
result = job.result()
354 J. R. Wootton
outputs = result.get_memory()
Let us unpack what is happening here. Firstly, the circuit we ran was specifically
the combination of qc and m_z: our original five qubits with all the measurements
placed at the end. This combined circuit is created in Qiskit with the command
qc.compose(m_z). The circuit was run on a so-called ‘backend’. In this case the
simulator known as the ‘Aer simulator’ was used. Using a real quantum device is as
easy as changing the backend object to one that refers to a real quantum device so
that the ‘run‘ function sends the job to that device over the cloud.
The process is repeated for ten shots, so we can see if there is any randomness in
the results. The output of each process will be a 5-bit string, given that our circuit
has five output bits. Since we used shots = 10 to ask for ten runs of the same
process, the results are returned as a list of the ten outputs. Specifically, we will get
the following list:
[‘00000’, ‘00000’, ‘00000’, ‘00000’, ‘00000’, ‘00000’, ‘00000’,
‘00000’, ‘00000’, ‘00000’]
Each output is ’00,000’. This means that each of the five qubits output a 0 in
each case. This is due to the way qubits are initialized: they always start out in the
state that is certain to output a 0. Our five ships, therefore, start off fully undamaged.
To implement an attack on a ship we need to apply an operation to its qubits, known
as a gate. The simplest example would be a completely fatal attack, flipping the ship
directly from undamaged to destroyed. The corresponding qubit must therefore be
flipped from the state certain to output 0 to the one certain to output 1. This effect can
be achieved by two quantum gates, known as X and Y gates. Both can be regarded
as quantum equivalents of the NOT gate for standard bits, which flips between the 0
and 1 states. In the following, we will focus on the X gate to implement the attack.
So, let us get on with it, and launch our most destructive torpedo at the first ship,
which is the one corresponding to the qubit labelled 0:
qc.x(0)
This adds the new operation to the circuit qc: an X gate on qubit 0. If we run the
combined circuit now, we will get the result ’00,001’. Here the result for the first
ship appears at the first bit on the right. It is a 1, denoting that the first ship has been
destroyed. Therefore, our X gate had the desired effect.
Now let us load up a new torpedo and take aim for the second ship. But let us
suppose that this ship is made of sturdier stuff, so one torpedo will not be enough to
sink it. Instead, it will only half destroy its target. In order to implement this attack,
we would like to do something like ‘half of an X gate’. This means we need to find
a gate which, when applied twice, leads to an end result that is exactly the same as
an X gate.
If we were using a conventional digital computer, with a conventional bit repre-
senting each ship and a NOT gate to implement attacks, this would not be possible.
Either you flip the bit value, or you do nothing. There is no in-between. More bits
The History of Games for (Quantum) Computers 355
and different types of gate would be required to implement the possibility of a half-
damaged ship. Of course, using more conventional bits is exactly the best course of
action in this case, but it is not what our game is about. We will doggedly stick with
just the one bit for each ship, and happily pay the price of making it quantum.
With quantum bits we can access so-called ‘superposition states’, which are not
quite 0 and not quite 1. That is, when we measure them, the results will appear random,
making it easy to think that they are simply the result of an overly engineered coin
flip. However, that is not the case. These states are just as definite and well-defined
as the ones whose outputs are certain. We could define measurements for which they
would give certain outputs instead of random ones, as we will see in the next section.
But when we ask the question of whether they are 0 or 1, they do not fit neatly into
either box. They cannot use a third option to declare that they are neither: then they
would no longer be bits since bits are only allowed two options by definition. Thus,
they simply answer randomly. With that in mind, let us implement the following
attack on the second ship (the one whose qubit is labelled 1):
from math import pi
qc.rx(pi/2, 1)
Instead of just an X gate, this is an RX gate. This gate requires two arguments: as
before, we specify which qubit it acts on, but we also supply the value of pi/2. If we
run the combined circuit now, we will find that the result is randomly either ’00001’
or ’00011’. In both cases, the ships that we have not attacked remain undamaged
(represented by their 0 s) and the first ship remains destroyed (represented by its 1).
The second ship, however, which is represented by the second qubit on the right, is
randomly either 0 or 1. By taking multiple samples we can determine that these two
possibilities occur with equal probability.
In order to see that this is not just a fancy coin flip, let us implement the attack
again with another qc.rx(pi/2, 1). The entire quantum circuit qc is now:
qc = QuantumCircuit(5) # attack on the first ship
qc.x(0) # attack on the second ship
qc.rx(pi/2, 1)
qc.rx(pi/2, 1)
Running the combined circuit now yields the result ’00011’ with certainty.
The first RX gate took its qubit from the state certain to output 0 to a superposition
state between 0 and 1. The second RX gate completed the journey, taking the qubit
to the state certain to output 1. Together they had exactly the same effect as a single
X gate.
Now let us attack the next ship (the one whose qubit is labelled 2). This is even
sturdier than the last and will require three torpedoes to sink. The required effect can
be implemented with rx(pi/3, 2). Applying this gate once will give random
results, but with a bias towards 0. For two applications the bias will instead favour
1. For three we get 1 with certainty, and the ship is destroyed.
Similarly, we can use pi/4 for a gate that must be repeated four times to form an X
gate, and so on. This argument can be thought of as representing an angle, as we will
356 J. R. Wootton
see in the next section. The angle pi is required for an X gate (and so rx(pi, 0)
would be another way to express the X gate with which we attacked the first ship).
For multiple sequential applications of RX gates on the same qubit, these angles
effectively add together.
This is basically all there is to the quantum machinery behind Battleships with
partial NOT gates. It does not show us how or why one might want to build qubits, or
why the ability to do partial forms of the gates used in conventional digital computers
will confer any computational advantage. The aim that it does achieve, however, is
to show them in action. And to show that quantum computers can do a relatable job
without the need for an abstract algorithmic approach.
Now that the qubit is no longer completely unfamiliar, let us get to know it better.
This can again be done in the context of using it to implement something that can be
used in a game.
As in the last section, we will be considering the concept of quantum circuits. We will
also repeat the trick of having a circuit called qc, which contains all the manipulations
we want to do before extracting an output:
qc = QuantumCircuit(1,1)
Next, we will use an additional circuit to extract the output, via the process known
as ‘measurement’. The standard form of measurement, as we used in the previous
section, is implemented simply using the measure gate in Qiskit:
m_z = QuantumCircuit(1, 1)
m_z = m_z.measure(0, 0)
As we might expect from the previous section, the result is 0 with certainty:
[‘0’, ‘0’, ‘0’, ‘0’, ‘0’, ‘0’, ‘0’, ‘0’, ‘0’, ‘0’]
This is the behaviour we can expect from a freshly initialized qubit whenever it is
measured using m_z. However, this is not the only way to measure. Another method
is the so-called ‘x measurement’. This is done with the following circuit, m_x:
m_x = QuantumCircuit(1, 1)
m_x.h(0)
m_x.measure(0, 0)
As you might have noticed, this requires an extra element compared to m_z. This
is because, at the time of writing, Qiskit doesn’t allow us to ask for the x measurement
directly, and so we need to use a little hack.
By running the combination of qc and m_x we will get a random mixture of 0s
and 1s as our results. To see the probability for each possible output, we need to
extract many results. For this it is more convenient to get the results in a less verbose
form, which simply lists each type of result along with the number of samples for
which it occurred. The syntax for this is as follows, for 1000 samples:
backend.run(qc.compose(m_x), shots=1000).result().get_counts()
358 J. R. Wootton
This gets the so-called ‘counts’ dictionary. Here’s an example of the kind of thing
you’ll see:
{‘1’: 499, ‘0’: 501}
Here the two possible results, 0 and 1, both occur roughly 500 times out of the 1000
total samples. From this we can conclude that the results are completely random,
giving 0 or 1 with 50/50 probability.
There are actually an infinite number of ways to extract an output from a qubit. But
we can create a complete description by using just three. The x and z measurements
are two of them. The final one is known as the ‘y measurement’:
m_y = QuantumCircuit(1, 1)
m_y.rx(pi/2, 0)
m_y.measure(0, 0)
By running the combination of qc and m_y we will again get a random mixture
of 0 s and 1 s as our results. With more samples, we can confirm that it is again
50/50. With statistics on the results of these three types of measurement, we can
fully characterize a single qubit state.
Here we characterized the state that a qubit is in when initialized at the beginning
of a circuit. The result was that such a qubit is certain to give an output of 0 for m_z,
but gives random results for the other two.
Now we can investigate the properties of other possible qubit states. We will do
this by applying all of the possible single-qubit operations to a qubit.
Let us examine the effects of the quantum operator known as the X gate:
qc = QuantumCircuit(1)
qc.x(0)
Since we are now well acquainted with all the important ways to get an output,
we can run them all for this new qc and see what happens. We will find that the x
and y measurements still give random results, but the z measurement is now certain
to output a 1. The X gate simply flips the bit value that is output for m_z. It serves
as the quantum form of a standard NOT logic operator, as discussed above.
If we add another X gate to the circuit, we will see the flip effect again: the output
for m_z will go back to being 0.
qc = QuantumCircuit(1)
qc.h(0)
If we run this with our three types of measurement, we will find that it is now
m_y and m_z that give random results, and m_x for which the outcome is certain
to be 0. This is an effect of what we might call the ‘conservation of certainty’ in
quantum systems. The H gate has unlocked the ability to store a definite bit value in
the results of m_x. However, it also takes away the ability to do so with m_z. This
prevents us from storing 2-bit values in the same qubit, and so preserves its identity
as truly being a type of bits.
A similar effect can be done using the RX gate that we saw earlier. This requires
an additional parameter: an angle expressed in radians. For example, with the angle
–pi/2:
qc = QuantumCircuit(1)
qc.rx(–pi/2, q)
The effect of this is to make it m_y that gives the output 0 with certainty, and the
other two that are random.
For both H and RX gates with this specific angle, the qubit was certain to output
0 for one of the measurements. With other gates we can make these certain to be 1
instead. The simplest way is to add an x gate at the beginning of each circuit.
With the gates we have so far, we are able to set the output of a qubit to be 0
or 1, and also to change which of the measurements see randomness and which see
certainty. These represent the most basic kinds of operations, known as the Clifford
gates.
The simplest way to move beyond the Clifford gates is to revisit the RX gate, and
simply use an angle that is not a multiple of pi/2. Specifically, let us use the following
code:
qc = QuantumCircuit(1).
qc.rx(–pi/4,q).
The results for this will not be as clear cut as we’ve seen before. Here is an example
of what you might see for 1000 samples for each type of measurement:
The results from m_y are similarly random but with a bias towards 1. The limited
certainty of the qubit has therefore been shared between these two possible types of
output. Both are mostly, but not completely certain of what output to give.
To fully understand what is happening here, we need a way of visualizing the
results. Specifically, we will plot the probability for the outcomes 0 and 1 for each of
the three types of measurement on a 3D plot. Since this is well known to be a useful
thing to do, it has a name: the Bloch sphere.
Let us first use this visualization on an empty circuit: one that outputs a 0 with
certainty for a z measurement, but random results for the others:
qc = QuantumCircuit(1)
The state of the qubit is represented as a point in 3D space. The x, y and z axes are
used to show the probabilities for x, y and z measurements, respectively. If the output
is certain to be 0 for each measurement, the qubit state is depicted on one extreme
along the corresponding axis. If it is certain to be 1, it is depicted on the other side.
For a completely random result, it is in the middle. In this case the certainty of a 0 for
the z measurement puts the point at the very top of the image. For the x and y axes,
the point lies in the middle. The state corresponds to the point labelled |0 (Fig. 3).
Next, let us plot the state after applying the X gate to the qubit, for which the z
measurement is certain to output 1:
qc = QuantumCircuit(1)
qc.x(0)
Here the point is at the bottom of the Bloch sphere (Fig. 4). The positions of
these outcomes are completely opposed to each other. The state in this case has been
labelled |1.
We find similarities for the states whose output is certain for the x measurement.
Let us examine again the state where the H gate is used to ensure that the qubit is
certain to output 0 for the x measurement:
qc = QuantumCircuit(1)
qc.h(0)
Although the visualization does not include a specific label for this state, it is
usually referred to as the |+ state (Fig. 5).
Let us now construct a circuit that is certain to output 1 for an x measurement:
qc = QuantumCircuit(1)
qc.x(0)
qc.h(0)
As we should expect, the corresponding point lies on the opposite side to the one
above. It is a state known as |– (Fig. 6).
The states corresponding to the certainty of a 0 or 1 outcome for y measurement
similarly correspond to the opposite points on the y-axis. This then gives us six
extreme points on the surface of a sphere. Each represents an outcome that is certain
for one of the three types of measurement, but completely random for the other two.
With other gates, we can explore the entire surface of the sphere. The RX gate for
a given angle corresponds to the rotation of the state around the x-axis by that angle.
There are similarly ry and rz gates that rotate around their respective axes.
For example, here is the –pi/4 rotation with RX gate that we considered earlier,
with the point corresponding to the state highlighted as a blue point. From its place-
ment, it is easy to see why it is biased towards 0 for a z measurement and 1 for a y
measurement: it lies exactly between those points (Fig. 7).
The fact that the qubit state appears bound to the surface of a sphere is a conse-
quence of the conservation of certainty. It is an expression of exactly how the trade-off
for certainty between the different measurements works.
Now we have seen all that needs to be known about a single qubit. All single-qubit
manipulations can be constructed from the gates discussed above. Everything else,
like special names for specific choices of angle and axis, is just administration.
The History of Games for (Quantum) Computers 363
Using the properties of the single qubit, we will now devise an algorithm for terrain
generation. This is based upon a proof-of-principle game created for the Ludum Dare
45 game jam (Wootton, 2019).
This algorithm for procedural generation is based around a single function:
get_height(i, j).
The function above will simply tell us the height of the terrain at any given point,
specified by the coordinates i and j. Note that here we use i and j as coordinates,
rather than the usual x and y because this section has already been using X and Y as
gates, and x and y as types of measurements.
The height in this function is calculated using a single-qubit quantum circuit.
Specifically, gates are applied that depend upon the coordinates of the point. The
choice of gates presented here was chosen quite arbitrarily. The reader is encouraged
to play around with alternatives.
One aspect of the choice of gates that was done very deliberately was to ensure that
the circuit created for any given point (i, j) is quite similar to that for its neighbours,
to ensure smooth terrain. For a specific example of this, let us consider a circuit that
consists of an rx and a ry rotation. For these, we shall slowly change the angle for
the RX gate as we move from point to point along the horizontal axis of the map,
and similarly change the angle for ry as we move along the vertical axis of the map.
Note that this is an arbitrary choice, and does not reflect any deep meaning. There
is no connection between the axes on the map and those of the Bloch sphere. It was
chosen simply because it seemed like a nice choice. Again, the reader is encouraged
to play around with alternatives.
After these gates, a z measurement is performed and the probability of getting the
output 1 is calculated. This probability could itself be used as the height. However,
it was found that more appealing terrain results when using p as the height. Again,
364 J. R. Wootton
this is a rather arbitrary choice and the reader can try different approaches. Below is
the resulting function:
def get_height(i, j):
qc = QuantumCircuit(1, 1)
# perform rotations, whose angles depend on i and j
qc.x(0)
# low frequency rotations to create island shape
qc.rx((1/32)*i*pi, 0)
qc.ry((1/32)*j*pi, 0)
# perform a z measurement
qc.measure(0, 0)
# determine the probability of a 1
counts = backend.run(
qc,shots=1000
).result().get_counts()
try:
p = counts[’1’]/1000
except:
p=0
# return p2 as the height
return p**2
In order to see what kind of terrain this generates, we can run it for a set of points
and plot the output. Figure 8 plots the results for a 30 × 30 set of points around
the origin. A terrain colour map is used, which colours low values as blue, and then
green, up to high values as brown and then white.
Though this is unmistakably some terrain, it clearly isn’t very interesting or real-
istic. In order to make it nicer, we can add some more gates. As one example, we
could break up the single peak by adding gates for which the angles of rotation make
larger changes as we move from one point to its neighbours, as follows (Fig. 9):
The example shown in Fig. 9 still is not very realistic. In order to create something
better, one should experiment with various new combinations of gates. We could also
try adding in a randomly generated seed. Note that this would not mean using the
randomness of the quantum computer: our calculation and use of a probability for the
height is an effort to avoid this randomness and make something deterministic from
it instead. Rather it is a random choice of parameters in the circuits we run in order to
366 J. R. Wootton
make each map unique. There are many ways that the process could be seeded, so we
again need to make an arbitrary choice. Since this method for procedural generation
was developed for a game jam, the method for seed generation was inspired by the
theme of the jam. In this case, the theme of Ludum Dare 45 was ‘Start with nothing’.
For this reason, the game starts with no seed, and hence no terrain. As the player
explores, their (presumably random) path was used to generate the world as they
explore it.
Another way of improving this method is to not just run the circuit with a z
measurement, but with x and y measurements too. This additional information could
then also be used to determine the characteristics of the terrain at any given point.
There are many further ideas that can be explored, but the limitations of this
method must also be remembered. Manipulating a single qubit is, as we saw from
the Bloch sphere, much like rotating a ball. As such, we won’t get anything uniquely
quantum that could not be done using standard tools for 3D rotations.
10 Conclusion
References
Benioff, P. (1980). The computer as a physical system: A microscopic quantum mechanical Hamil-
tonian model of computers as represented by Turing machines. Journal of Statistical Physics.,
22(5), 563–591.
Q Cards. https://zhamul.itch.io/qcards
Copeland, B. J. (2004). The essential Turing (pp. 563–564). Oxford University Press
Jordan, S. (2021) Quantum algorithm zoo. https://quantumalgorithmzoo.org
Qiskit. https://qiskit.org
QiskitBlocks. https://github.com/JavaFXpert/QiskitBlocks
The History of Games for (Quantum) Computers 367
QPong. https://github.com/HuangJunye/QPong
IBM Quantum Composer and Lab. https://quantum-computing.ibm.com
Samuel, A. L. (1959). Some studies in machine learning using the game of checkers. IBM Journal
of Research and Development, 3(3), 210–229.
Wootton, J. R. (2016). Demonstrating non-Abelian braiding of surface code defects in a five qubit
experiment. Quantum Science and Technology, 2(1).
Wootton, J. R. (2017). How to program a quantum computer. Retrieved December 03, 2020, from
https://medium.com/qiskit/how-to-program-a-quantum-computer-982a9329ed02
Wootton, J. R. (2018). Benchmarking of quantum processors with random circuits. https://arxiv.
org/abs/1806.02736
Wootton, J. R. (2019). Genesis. https://ldjam.com/events/ludum-dare/45/genesis-1
Wootton, J. R. (2020a). A quantum procedure for map generation, 2020 IEEE Conference on Games
(CoG), 73-80.
Wootton, J. R. (2020b). Procedural generation using quantum computation, International Confer-
ence on the Foundations of Digital Games, 98.