Download Quantum Measurement Theory and its Applications 1st Edition Kurt Jacobs ebook All Chapters PDF
Download Quantum Measurement Theory and its Applications 1st Edition Kurt Jacobs ebook All Chapters PDF
Download Quantum Measurement Theory and its Applications 1st Edition Kurt Jacobs ebook All Chapters PDF
https://ebookgate.com/product/handbook-of-implicit-social-cognition-
measurement-theory-and-applications-1st-edition-bertram-gawronski/
ebookgate.com
https://ebookgate.com/product/quantum-reality-theory-and-
philosophy-1st-edition-allday/
ebookgate.com
Quantum Information Theory and the Foundations of Quantum
Mechanics 1st Edition Christopher G. Timpson
https://ebookgate.com/product/quantum-information-theory-and-the-
foundations-of-quantum-mechanics-1st-edition-christopher-g-timpson/
ebookgate.com
https://ebookgate.com/product/group-theory-and-quantum-mechanics-
michael-tinkham/
ebookgate.com
https://ebookgate.com/product/game-theory-and-its-applications-in-the-
social-and-biological-sciences-2nd-edition-andrew-m-colman/
ebookgate.com
https://ebookgate.com/product/quantum-groups-and-lie-theory-1st-
edition-andrew-pressley/
ebookgate.com
https://ebookgate.com/product/quantum-mechanics-concepts-and-
applications-1st-edition-nouredine-zettili/
ebookgate.com
QUANTUM MEASUREMENT THEORY
AND ITS APPLICATIONS
KURT JACOBS
University of Massachusetts at Boston
University Printing House, Cambridge CB2 8BS, United Kingdom
www.cambridge.org
Information on this title: www.cambridge.org/9781107025486
c Kurt Jacobs 2014
This publication is in copyright. Subject to statutory exception
and to the provisions of relevant collective licensing agreements,
no reproduction of any part may take place without the written
permission of Cambridge University Press.
First published 2014
Printed in the United Kingdom by Clays, St Ives plc
A catalogue record for this publication is available from the British Library
Library of Congress Cataloguing in Publication data
Jacobs, Kurt (Kurt Aaron), author.
Quantum measurement theory and its applications / Kurt Jacobs, University of Massachusetts at Boston.
pages cm
Includes bibliographical references and index.
ISBN 978-1-107-02548-6 (hardback)
1. Quantum measure theory. I. Title.
QC174.17.M4J33 2014
530.801–dc23 2014011297
ISBN 978-1-107-02548-6 Hardback
Cambridge University Press has no responsibility for the persistence or accuracy of
URLs for external or third-party internet websites referred to in this publication,
and does not guarantee that any content on such websites is, or will remain,
accurate or appropriate.
To my mother, Sandra Jacobs, for many things.
Not least for the Nelson Southern Link Decision, a great triumph unsung.
Contents
Preface page xi
1 Quantum measurement theory 1
1.1 Introduction and overview 1
1.2 Classical measurement theory 4
1.2.1 Understanding Bayes’ theorem 6
1.2.2 Multiple measurements and Gaussian distributions 9
1.2.3 Prior states-of-knowledge and invariance 11
1.3 Quantum measurement theory 15
1.3.1 The measurement postulate 15
1.3.2 Quantum states-of-knowledge: density matrices 15
1.3.3 Quantum measurements 20
1.4 Understanding quantum measurements 28
1.4.1 Relationship to classical measurements 28
1.4.2 Measurements of observables and resolving power 30
1.4.3 A measurement of position 31
1.4.4 The polar decomposition: bare measurements and feedback 34
1.5 Describing measurements within unitary evolution 37
1.6 Inefficient measurements 39
1.7 Measurements on ensembles of states 40
2 Useful concepts from information theory 48
2.1 Quantifying information 48
2.1.1 The entropy 48
2.1.2 The mutual information 53
2.2 Quantifying uncertainty about a quantum system 55
2.2.1 The von Neumann entropy 55
2.2.2 Majorization and density matrices 58
2.2.3 Ensembles corresponding to a density matrix 61
2.3 Quantum measurements and information 63
2.3.1 Information-theoretic properties 64
2.3.2 Quantifying disturbance 72
vii
viii Contents
I would like to thank here a number of people to whom I am indebted in one way or
another. To begin, there are five people from whose insights I especially benefited in my
formative years in physics. In order of appearance: Sze M. Tan, for teaching me classi-
cal measurement theory, and introducing me to information theory and thermodynamics;
Howard M. Wiseman, for teaching me about quantum measurement theory; Salman Habib,
for teaching me about open systems and classical chaos; Tanmoy Bhattacharya, for enlight-
enment on a great variety of topics, and especially for the insight that measurement is
driven by diffusion gradients; Gerard Jungman, for mathematical and physical insights,
and for introducing me to many beautiful curiosities.
I am very grateful to a number of people who helped directly to make this book what it
is: Os Vy, Luciano Silvestri, Benjamin Cruikshank, Alexandre Zagoskin, Gelo Tabia, Justin
Finn, Josh Combes, Tauno Palomaki, Andreas Nunnenkamp, and Sai Vinjanampathy who
read various chapters and provided valuable suggestions that improved the book. Xiaot-
ing Wang who derived Eqs. (G.43)–(G.48). Jason Ralph who enlightened me on some
superconductor facts that were strangely difficult to extract from the literature. Jason also
helped me with the brief history of superconductivity and quantum superconducting cir-
cuits in Chapter 7. Justin Guttermuth who saved our asses when we had a house to move
into, rooms to paint, a new baby, and I had this book to finish. My colleagues in the
UMass Boston physics department for their support — especially Maxim Olchanyi, Bala
Sundaram, Vanja Dunjco, and Steve Arnason. And last but not least, my wonderful wife
Jacqueline, who helped me with the figures and the cover, and put up with the long hours
this book required.
I apologize in advance for any errors and inadvertent omissions in this book. I would
be most grateful to be notified of any errors that you may find. I will include corrections
to all errors, as they are found, in an errata file on my website. I am very grateful to a
number of readers who sent me corrections for my previous book, Stochastic Processes
for Physicists, all of which have been made in the current printing. I was also able to
acknowledge these readers in the current printing, which due to their efforts now appears
to be largely error-free.
While I have endeavored to cite a fairly comprehensive and representative set of research
papers on the topics I have covered in this text, it is likely that I have omitted some that
xi
xii Preface
deserve to be included. If you discover that your important paper on topic X has been
missed, please send me the reference and I will be glad to correct this omission in any
further edition.
Finally, it is a pleasure to acknowledge an ARO MURI grant, W911NF-11-1-0268, that
was led by Daniel Lidar and administered by Harry Chang. This grant provided partial
support for a number of research projects during the writing of this book, and from which
it greatly benefited.
1
Quantum measurement theory
1
2 Quantum measurement theory
There is another important connection between quantum measurement theory and ther-
modynamics: thermal baths that induce thermalization and damping also carry away a
continuous stream of information from the system with which they interact. Because of
this, thermal baths can mediate continuous measurements, and there is a useful overlap
between the descriptions of the two. Further, the continuous measurements induced by a
system’s environment are sufficient to induce the quantum-to-classical transition, in which
the trajectories of classical dynamics emerge as a result of quantum mechanics.
It is now possible to construct and manipulate individual quantum systems in the lab-
oratory in an increasingly wide range of physical settings. Because of this, we complete
our coverage of measurement theory by applying it to the measurement and control of
a variety of concrete mesoscopic systems. We discuss nano-electromechanical systems,
in which superconducting circuits are coupled to nano-mechanical resonators, and opto-
mechanical systems in which superconducting circuit elements are replace by the modes
of optical resonators. In Chapter 7 we introduce these systems and show how to deter-
mine their Hamiltonians and the interactions between them. In this chapter we also explain
how the continuous quantum measurements introduced in Chapter 3 are realized in these
systems, and rephrase them in the language usually used by experimentalists, that of
amplifiers. In Chapter 8 we consider a number of examples in which the above sys-
tems are controlled, including realizations of the feedback control techniques described in
Chapter 5.
Despite its importance, quantum measurement theory has had an uneasy relationship
with physicists since the inception of quantum mechanics. For a long time the tendency
was to ignore its predictions as far as measurements on individual systems were concerned,
and consider only the predictions involved with averages of ensembles of systems. But as
experimental technology increased, and experimentalists were increasingly able not only to
prepare and measure individual quantum systems in well-defined states with few quanta but
to make repeated measurements on the same system, physicists were increasingly forced
to use the full predictions of measurement theory to explain observations.
The uneasiness associated with measurement theory stems from a philosophical issue,
and this has disturbed physicists enough that many still seem uncomfortable with using the
language of measurement in quantum mechanics, and assigning the knowledge gained to
an observer. In fact, as will be explained in due course, all processes that involve quantum
measurements can be described without them, with the possible exception of a single mea-
surement relegated to the end of the process. In using this second description the language
changes, but the two descriptions are completely equivalent. Thus the use of the language
of measurement theory is entirely justified. What is more, analysis that makes explicit use
of measurement theory is often vastly more convenient than its unitary equivalent, pro-
viding a strong practical reason for using it. This should not be taken to mean that the
philosophical problem has gone away – far from it. In fact, while the focus of this book is
applications, we feel that no education in quantum measurement theory is quite complete
without an understanding of “the measurement problem,” and because of this we include a
discussion of it at the end of Chapter 4.
1.1 Introduction and overview 3
We do not cover all the applications of quantum measurement theory, most notably those
to do with quantum information theory. The latter is now itself a large subfield of physics,
and there are a number of textbooks devoted to it. While our goal is to focus on applications
that are not usually covered in books on quantum information, given the fundamental con-
nection between measurement and information the two subjects are intertwined. A number
of concepts from information theory are therefore invaluable in broader applications of
measurement, and so we discuss information theory in Chapter 2. We do not cover “foun-
dational” questions in quantum theory, with the sole exception of the so-called “quantum
measurement problem.” Another topic within measurement theory that we do not discuss
is that termed “weak values,” not to be confused with weak measurements which we treat
in detail.
Quantum measurement theory is a weird and wonderful subject. I hope I have whetted
your appetite, and that you find some topics in this book as stimulating and fascinating as
I have.
metrology; Sections 7.1 through 7.6, and Sections 8.1 and 8.3 on controlling mesoscopic
systems.
Here P(y) is the probability distribution for y irrespective of the value of x (also called the
marginal distribution for y), and is given by
∞
P(y) = P(x, y) dx. (1.2)
−∞
Now P(x) is our state-of-knowledge of x prior to making the measurement, and it is there-
fore the probability density for x irrespective of the value of y. It is therefore also the
marginal probability for x, and thus given by
∞
P(x) = P(x, y) dy. (1.3)
−∞
While the relationships in Eq. (1.1) are fairly intuitive, they are explained further in, for
example, references [288] and [592].
6 Quantum measurement theory
Rearranging Eq. (1.1) we obtain the famous relationship known as Bayes’ theorem,
being
P(y|x)P(x)
P(x|y) = . (1.4)
P(y)
Upon examining Eq. (1.4) we will see that it tells us exactly how to change our state-of-
knowledge when we obtain the measurement result y. First note that since P(x|y) must be
normalized (that is, its integral over x must be unity), the value of P(y) on the bottom line
is completely determined by this normalization. We can therefore write Bayes’ theorem as
∞
P(y|x)P(x)
P(x|y) = , where N = P(y|x)P(x) dx = P(y). (1.5)
N −∞
We see that on the right-hand side (RHS) we have our state-of-knowledge of x before we
obtain the data y, and on the left-hand side (LHS) the probability for x given that value of
y. The LHS is therefore our state-of-knowledge after obtaining the value of y. In Eq. (1.5)
P(x) is called the prior probability, and P(x|y) the posterior probability. So Bayes’ theorem
tells us that to obtain our new state-of-knowledge once we have made our measurement:
we simply multiply our current state-of-knowledge by the likelihood function P(y|x), and
normalize the result. Note that the prior is simply the marginal (overall) distribution of
x. The relationship given by Eq. (1.5), Bayes’ theorem, is the fundamental theorem of
classical measurement theory.
P(1|x)P(x) P(1|x)P(x)
P(x|1) = )P(x )
= . (1.6)
x P(1|x P(1|0)P(0) + P(1|1)P(1)
case we would expect the measurement to tell us a lot about x. Conversely, if α is close to
1/2, then the reverse is true. For the sake of concreteness let us choose 0. 5 < α < 1. With
this choice, if x = 0 then it is more likely that we will get the result y = 0, and if x = 1 it is
more likely that we will get y = 1.
Now assume that we initially know nothing about x, so that our prior state of knowledge
is P(0) = 0. 5 = P(1). What happens when we make the measurement and get the result
y = 1? Since our prior is flat, by which we mean that it does not change with x, it cancels
on the top and bottom lines of Bayes’ theorem, telling us that our posterior is simply the
likelihood function, normalized if necessary:
P(y|x) P(y|x)
P(x|y) =
= . (1.7)
x P(y|x ) P(y|0) + P(y|1)
Our posterior is thus P(0|1) = 1 − α and P(1|1) = α. So it is now more likely that x = 1
than x = 0. This is indeed intuitively reasonable. The likelihood function tells us that if x
were to be 0 then it would be less likely that we would get the result 1, so it is reasonable
that since we obtained the result 1, it is more likely that x = 1 than that x = 0.
The above discussion shows that if the prior is uniform, Bayes’ theorem tells us that the
values of x that are more likely are those for which the result we have obtained is the more
likely outcome.
Now let us examine why we need to include the prior, in addition to the likelihood func-
tion, when calculating our posterior. This is clearest when the prior is strongly weighted
toward one value of x. Consider a situation in which the prior is P(0) = 0. 999 and
P(1) = 0. 001. This means that in the absence of any further data, on average x will only be
equal to 1 one time in a thousand cases. We now consider a slightly different two-outcome
measurement from the one above, as this will make the logic simple. The likelihood func-
tion for the new measurement is given in Table 1.2, and in words is as follows: if x = 1
then y is always equal to 1. If x = 0, then y = 0 with probability 0.999 and y = 1 only one
time in a thousand. This means that, if the prior is flat, upon obtaining the result y = 1 the
value of x would be equal to 1 approximately nine hundred and ninety-nine times out of a
8 Quantum measurement theory
thousand:
P(1|1) 0. 999
P(x = 1|y = 1) = = . (1.8)
P(1|0) + P(1|1) 1. 001
So what is the case when the prior is highly weighted toward x = 0, as described above?
Well, in this case x is equal to 1 only one time in a thousand. Now, if we get the result
y = 1 there are two possibilities. Either x = 1, which happens one time in one thousand, or
x = 0 and we got the result y = 1 anyway, which also happens approximately one time in a
thousand. (The precise figure is the frequency of x = 0 multiplied by the frequency of the
result y = 1 given x = 0, which is 0. 999 × 0. 001 ≈ 1/1001.) Thus the result y = 1 happens
approximately one time in 500, and half of these are due to x = 0, and half due to x = 1. So
when we obtain the result y = 1, there is only a 50% chance that x = 1. This is, of course,
exactly what Bayes’ theorem tells us; by multiplying the likelihood function that weights
x = 1 very highly, by the prior that weights x = 0 very highly, we obtain an approximately
flat posterior.
The example we have just considered applies directly to a real and very important sit-
uation: testing for the HIV virus. Each test is pretty reliable, giving a false positive only
about one time in a thousand. On that basis alone one might think that when a result comes
back positive, there is little reason to perform a follow-up test to confirm it. But this is very
wrong. Since very few patients have HIV, false positives come up just as frequently as real
positives. Thus, whenever a positive test result comes back, it is essential to do a follow-
up test to check that it is not a false positive. Bayesian inference, and thus measurement
theory, is therefore crucial in real-world problems.
To complete our discussion of Bayes’ theorem it is worth noting that our state-of-
knowledge does not necessarily become more certain when we make a measurement. To
take the example of the HIV test above, before we obtain the result of the test we are almost
certain that the patient is HIV negative, since the vast majority of patients are. However,
upon obtaining a positive test result, there is an approximately fifty-fifty chance that the
patient is HIV positive. Thus, after obtaining the measurement result, we are less certain
of the HIV status of the patient. Even in view of this, all classical measurements do have
the property that, upon making the measurement, we become more certain on average of
1.2 Classical measurement theory 9
the value of the measured quantity (where the average is taken over all the possible mea-
surement results). We will be able to make this statement precise once we have described
how to quantify the concept of information in Chapter 2.
P(y1 , . . . , yN |x)P(x)
P(x|y1, . . . , yN ) = , (1.9)
N
∞
with N = −∞ P(y1 , . . . , yN |x)P(x) dx. The fact that all the measurement results are
independent means that
where Nn = P(yn ). So we see that each time we make another independent measurement we
update our state-of-knowledge by multiplying it by the likelihood function and normalizing
the result.
P(yA |x)Pprior(x)
PA (x) = , (1.12)
NA
P(yB |x)Pprior (x)
PB (x) = , (1.13)
NB
10 Quantum measurement theory
where yA and yB are the vectors of data obtained by the respective observers. Since we
intend the data the observers have obtained to represent all the information they each have
about x, Pprior(x) is the prior that describes having no initial knowledge about x. The problem
of determining such a prior can be surprisingly difficult, and we discuss it further below. If
we assume for now that we know what Pprior is, and we choose the measure for integration
over x so that Pprior(x) = 1 (that is, we absorb Pprior into the measure), then an observer who
has access to the data of both A and B has the state-of-knowledge
So all we have to do to pool the knowledge of two (or more) observers is to multiply their
states-of-knowledge together, and normalize the result.
1
e−(y−x) /(2V) .
2
P(y|x) = √ (1.15)
2πV
This is the likelihood function for the measurement. If we have absolutely no knowledge
of x before the measurement (never the case in reality, of course), then we can set P(x) = 1.
Our knowledge of x after making the measurement is then simply the likelihood function,
normalized so that it is a valid probability distribution over x for each value of y. In this
case the normalization is already correct, and so
1
e−(x−y) /(2V) .
2
P(x|y) = √ (1.16)
2πV
This tells us that the value of x is a Gaussian centered at y, with variance V, and thus that the
most likely value of x is y, and the expectation value of x is also equal to y. It is customary
to quote the error on a√measured quantity as twice the standard deviation. Thus we write
the value of x as y ± 2 V.
1.2 Classical measurement theory 11
The central limit theorem states that so long as the variances of all the convolved distribu-
tions are finite, then as one convolves more distributions together, the closer the result is to
a Gaussian.
To obtain a similar result for multiplying a number of probability distributions together,
we take the Fourier transforms of these probability distributions. Multiplying two functions
together is the same as convolving their Fourier transforms. That is, the Fourier transform
of the product of two functions is the convolution of their Fourier transforms. So the central
limit theorem tells us that if we convolve the Fourier transforms of a set of distributions
together, then the result is close to a Gaussian. Since the inverse Fourier transform of
this approximate Gaussian is the product of all the probability distributions, and since
the Fourier transform of a Gaussian is also Gaussian, the product of the distributions is
approximately Gaussian.
So when we make multiple measurements to learn about some quantity, or when our
knowledge about a quantity is derived from many different sources, then our state-of-
knowledge will be approximately Gaussian. Just as Gaussian errors are common in nature,
Gaussian states-of-knowledge are also common.
Adding noise and making measurements can be regarded as opposite processes. The
more noise (random variables) we add to a quantity, the more uncertain it becomes. The
more measurements we make on a quantity, the more certain (on average) it becomes. It is
interesting to note that the mathematical operations that describe each of these processes,
convolution for adding noise, and multiplication for making measurements, are dual to
each other from the point of view of the Fourier transform.
then Bayes’ theorem is strictly correct and open to no interpretation. However, for many sit-
uations it is not clear what the prior should be, since the observer usually knows something
before making measurements, but it is not always possible to quantify this knowledge. A
sensible solution to this problem is to assume that the observer has no initial knowledge,
as this prevents any biasing of the final state-of-knowledge with unwarranted prejudice.
The problem is that it is not always easy to determine the prior that quantifies “no knowl-
edge.” Here we will discuss a useful method for obtaining such priors (and give references
for further methods), but before we do this it is worth noting that in many situations the
choice of prior is, in fact, unimportant. This is the case when we can obtain enough data
that the likelihood function is much sharper (has much less uncertainty) than any “reason-
able” prior. In this case the prior has very little influence on the final expectation value
of the measured quantity, and is therefore irrelevant. In this case we can ignore the prior
completely and set the posterior equal to the (correctly normalized) likelihood function for
the set of measurements.
The question of what it means to know nothing is actually quite subtle. In the case of
measuring the position of an object in space, it seems quite obvious that the prior should
be constant and extend from −∞ to ∞. This prior is not normalizable, but this does not
matter, since for any measurement that provides a reasonable amount of data, the likelihood
function, and thus the posterior, will be normalizable. On the contrary, the question of what
it means to know nothing about a positive quantity, such as the volume of water in a lake,
is not at all obvious. As will become clear, a prior that is constant on the interval (0, ∞) is
not the correct choice.
The problem of what it means to know nothing about a positive quantity is solved by
using the idea that the prior should be invariant under a transformation. The key is to
realize that if one knows nothing about the value of a positive quantity, λ, then multiplying
λ by any positive number should not change one’s state-of-knowledge. That is, if one is
completely ignorant of λ, then one must be ignorant of the overall scale of λ. There is only
one state-of-knowledge that is invariant under a scaling transformation, and that is
where a is any dimensionless constant. This prior is also not normalizable, but this is fine
so long as the likelihood function is normalizable. The set of all scaling transformations,
which is the set of all positive real numbers, forms what is called a transitive group. It is
a group because it satisfies the properties of a group (e.g., the consecutive application of
two scaling transformations applied consecutively is another scaling transformation), and
the term “transitive” means that we can transform any value of our positive quantity to any
other value by using one of these transformations.
The powerful technique of identifying transformations, in particular transformations that
are transitive groups, over which our prior state-of-knowledge must be invariant, was devel-
oped by Jeffreys [312, 314, 313] and Jaynes [522, 311]. Returning to the problem of the
prior for the location of an object, we see that the flat prior is the one that is invariant under
1.2 Classical measurement theory 13
all translations, being the transitive group for that space. Transitive groups for a given
quantity must map all admissible values of the quantity to admissible values. That is why
the translation transformation is not appropriate for quantities that can only be positive.
An excellent example of the use of invariance for determining priors is given by Jaynes’
solution to “Bertrand’s problem,” and can be found in [311].
Using the notion of invariance under a transitive group it is simple to determine priors
for quantities that have such groups. The simplest case is a “six-sided” die. If the die
is perfectly symmetric, then our prior should be invariant under all permutations of the
six faces of the die. The result is a prior that assigns equal probability to all the faces.
A less trivial example is a prior over the possible states of an N-dimensional quantum
system. In this case the set of possibilities is an N-dimensional complex vector space,
and the transitive group is that of all the unitary transformations on this vector space. The
zero-knowledge prior is the one that is invariant under all unitaries.
Instead of specifying the prior in terms of a set of coordinates that parametrizes the
space of possibilities, one usually specifies the measure over which to integrate in terms
of these coordinates, and chooses this measure so that it is invariant under the required
transformations. With this choice for the measure, the invariant prior is merely equal to
unity. For unitary transformations the invariant measure is called the Haar measure, and is
given in Appendix G. The definition of a measure, using a minimum of technical jargon,
can be found in reference [288].
The method of transitive groups fails when the set of possibilities has no such group. The
interval [0, 1] is one such example, and applies when one is inferring a probability. More
sophisticated methods have been devised to address these situations. We will not discuss
these methods here, but the reader can find further details in [52, 479, 363].
What is wrong with the above analysis is the hidden assumption we have made about
what it means to initially “know nothing” about the amount of money in the envelopes.
That is, we have arbitrarily chosen a prior, and this causes problems with our reasoning.
Now let us consider the problem with the prior made explicit. Let us call the envelopes A
and B, and say that envelope A contains x dollars, and envelope B contains 2x dollars. To
perform any reasoning about probabilities we must assign a probability distribution to x
to start with, so let us call this distribution Px (x). Now we pick one envelope at random.
Call the envelope we pick envelope 1, the amount of money we find in it y, and denote
by z the unknown amount of money in envelope 2. Since the probability that envelope 1
is envelope A is independent of which envelope is which, Bayes’ theorem tells us that the
relative probabilities that envelope 2 has amounts z = y/2 and z = 2y are determined by the
prior alone. Thus
Px (y/2) Px (2y)
Prob(z=y/2) = , Prob(z=2y) = . (1.19)
Px (y/2) + Px (2y) Px (y/2) + Px(2y)
If we chose to keep envelope 1 then we will have y dollars. The expectation value of the
amount of money we will have if we choose to keep envelope 2 instead is
(y/2)Px(y/2) + 2yPx(2y)
z = . (1.20)
Px (y/2) + Px(2y)
We will want to keep envelope 2 if z > y.
Now that we have made the prior explicit, we can see that our initial reasoning about
the problem corresponds to choosing the prior to be flat, from which it follows that the
probability that z = 2y is equal to the probability that z = y/2. But our discussion in the
previous section makes us suspicious of this prior. If x were merely an unknown positive
quantity it would have Px (x) ∝ 1/x. The present case is a little different, but we might expect
a somewhat similar prior. Let us first consider a prior in which the expectation value of x
is finite, as this allows us to see how opening envelope 1 provides us with information. If
you put in an explicit function for Px (x) such that x is finite, you will find the following.
If y is large enough then we are sufficiently sure that envelope 2 is the one with less money,
and we keep envelope 1. If y is small enough then it is sufficiently likely that envelope 2 is
the one with more money, and so we keep it instead. For every Px with finite x there is a
threshold value yc such that when y > yc we keep envelope 1, and if not we keep envelope 2.
Now let us return to the problem of what it might mean to know nothing about x prior
to opening an envelope. Let us denote the “zero-knowledge” distribution that we seek by
P0 (x). From our discussion in the previous section, P0 (x) should have no scale, since a
scale, such as a given value for x , means that we know something about x. Because
P0 (x) must be scale-free, when we open an envelope it cannot provide a number yc that
determines which envelope we should keep. The only remaining option is that when we
open an envelope we learn nothing about which envelope to keep. This in turn requires
that no matter what value we find for y, it will always be true that z = y, giving us
no preference for either envelope. You can determine for yourself from Eq. (1.20) that the
1.3 Quantum measurement theory 15
√
only prior distribution satisfying this requirement is P0 (x) = 1/ x. While this is not strictly
scale-invariant, it is scale-free as desired. The lesson is that you cannot just pick any prior
and assume that it captures what it means to know nothing.
for some complex coefficients cn , where n |cn |2 = 1. In analyzing quantum measure-
ments we will often assume that the system is finite-dimensional. This simplifies the
analysis while losing very little in the way of generality: since all systems ultimately have
bounded energy, any real system can always be approximated arbitrarily well using a finite
number of states.
The basis of quantum measurement theory is the following postulate: We can choose
any basis, and look to see which one of these basis states the system is in. When we do
so, we will find the system to be in one of these basis states, even though it may have
been in any state |ψ before the measurement. Which basis state we find is random. If
the system is initially in the state |ψ then the probability that we will find state |n is
given by |cn |2 . A measurement like this, for which the result is one of a set of basis states,
is called a von Neumann measurement. Before we use this basic measurement postulate
to derive quantum measurement theory (which has, necessarily, a very similar structure
to classical measurement theory), we need to know how to describe states-of-knowledge
about quantum systems.
where X is the operator corresponding to that observable. So while the expectation value
of an observable is quadratic in the vector |ψ , it is linear in the operator (matrix) |ψ ψ|.
Expectation values are also linear in classical probabilities. If our state-of-knowledge is a
probability distribution over the M states {|φm }, where the probability of the system being
in the state labeled by m is pm , then the expectation value of X is
X = pm φm |X|φm = pm Tr[X|φm φm |] = Tr [Xρ] , (1.23)
m m
where
ρ≡ pm |φm φm |. (1.24)
m
So we see that the matrix ρ is sufficient to calculate the expectation value of any operator.
This is precisely because expectation values are a linear function of each of the matrices
|φm φm |, and thus all we have to do to include classical probabilities is to weight each of
the |φm φm | by its classical probability.
We will see below that ρ is also sufficient to calculate the results of any measurement
performed on the system. Note that it is only the results of measurements on quantum sys-
tems that determine events in the macroscopic world. This is because it is measurements
that determine a set of mutually exclusive outcomes, and since the macroscopic world
behaves classically, it is only sets of mutually exclusive possibilities that appear in it. (See
also the discussions in Sections 1.5, 4.5, and 4.4.) We can conclude that questions about the
future behavior of a quantum system are ultimately questions about the results of measure-
ments. Thus ρ is sufficient to fully characterize the future behavior of a quantum system,
and this is why it is a sufficient description of one’s state-of-knowledge for many purposes.
In the absence of measurements the evolution of ρ is very simple. This evolution is
given by evolving each of its component states. Since the evolution of a quantum state |ψ
is given by applying to it a unitary operator, U(t), the evolution of ρ is
ρ(t) = pm |φm (t) φm (t)| = pm U(t)|φm (0) φm (0)|U †(t)
m m
= U(t)ρ(0)U (t).†
(1.25)
1.3 Quantum measurement theory 17
The density matrix, ρ, is therefore a simple and compact way to describe our state-of-
knowledge of a quantum system. The term density matrix comes from its role as a quantum
equivalent of a probability density for classical systems, and also because its diagonal
elements constitute a probability distribution. Specifically, if we express our states in the
basis {|j }, so that
|φm = cjm |j , (1.26)
j
Since |cjm |2 is the (conditional) probability of finding the system in the state |j given
that it is initially in the state |φm , ρjj is the total probability of finding the system in the
state |j .
If the density matrix consists of only a single state, so that ρ = |ψ ψ|, then it is
described as being pure. If it consists of a sum over more than one state, then it is described
as being mixed, and the system is said to be in a statistical mixture, or simply a mixture, of
states. As an example, if the system is in the pure state |ψ = j cj |j , then
ρ = |ψ ψ| = cj c∗k |j k| (1.29)
jk
and the system is said to be in a superposition of the basis states |j . If the system is in the
state
ρ= pj |j j|, (1.30)
j
then it is said to be in a mixture of the basis states |j . In the latter case, we can think of the
system as really being in one of the basis states, and the density matrix merely describes
the fact that we do not know which one (although some may be more likely than others).
On the other hand, if the system is in a superposition of two states, then we cannot describe
the system as really being in one state or the other, and we discuss this further in the next
section.
The density matrix is said to be completely mixed if it is equal to I/N, where N is the
dimension of the system. If this is the case, then each of its eigenstates are equally likely.
Further, every state in the Hilbert space (the complex vector space) is equally likely, since
18 Quantum measurement theory
ψ|(I/N)|ψ = 1/N for every state |ψ . In this case we have no information about the
system.
It is not difficult to show that the density matrix is Hermitian (ρ = ρ † ), and has unit trace
(Tr[ρ] = 1). Since the density matrix is Hermitian it has a complete set of eigenvectors
(eigenstates), and can always be diagonalized (that is, written in the basis of its eigen-
states so that it is diagonal). If the eigenvalues are λj , j = 1, . . . , N, and the corresponding
eigenvectors are |λj , the density matrix is
ρ= λj |λj λj |. (1.31)
j
Since Psup (x) is not equal to P(x), and P(x) is a necessary result of the electron really being
in state |ψ1 or |ψ2 , a superposition cannot correspond to the electron really being in
either of these two states. For want of better terminology, one usually refers to a system
that is in a superposition of two states as being in both states at once, since this seems more
reasonable than saying that it is in neither.
1 Note that this is also what happens to water waves if they encounter a barrier with two small openings. While
Schrödinger’s wave-equation is not the same as that for water waves, all wave-equations have this feature in
common.
20 Quantum measurement theory
Theorem 1 If the state of a quantum system is ρ, then the most likely pure state is the
eigenstate of ρ with the largest eigenvalue. This eigenvalue is the probability that the
system will be found in that state.
Proof We need to find the state |ψ that maximizes ψ|ρ|ψ . Let us denote the eigenbasis
of ρ as {|λn }. Writing the density matrix in terms of its eigenbasis (Eq. 1.31), and writing
|ψ = n cn |λn , we have
ψ|ρ|ψ = |cn |2 λn . (1.36)
n
Since n |cn |2 = 1, the above expression is the average of the eigenvalues over the prob-
ability distribution given by pn = |cn |2 . Thus to maximize it we must place all of the
probability on the largest eigenvalue. If we denote this eigenvalue by λj , then this means
that |cn |2 = δnj , and therefore |ψ = |λj .
A density matrix consists of a set of states, {|ψn }, in which each state has an associated
probability, pn . Such a set of states and probabilities is called an ensemble. In fact, there
is no reason why every state in an ensemble must be a pure state – a set of states {ρn }
with associated probabilities {pn } is also an ensemble, and generates a density matrix ρ =
n pn ρn . If an ensemble contains only pure states, then it is called a pure-state ensemble.
We will often write an ensemble for the set of states {ρn }, with associated probabilities
{pn }, as {ρn , pn }.
Since every pure-state ensemble corresponds to some density matrix, a natural question
to ask is, for a given density matrix, can one determine all the possible pure-state ensembles
that correspond to it? It turns out that not only is the answer yes, but the collection of
ensembles can be characterized very simply. We leave these facts to Chapter 2, however,
because one of the characterizations uses the concept of majorization. This is a very simple
concept which captures the intuitive notion of uncertainty, and it will be defined in the next
chapter.
Pn ≡ |n n|. (1.37)
Now, if we make a von Neumann measurement in the basis {|n }, then after the measure-
ment we will find the system to be in one of these basis states, say |m , with probability
pm = m|ρ|m . To reduce an arbitrary initial density matrix, ρ, to the final density matrix
1.3 Quantum measurement theory 21
|m m|, we must sandwich ρ between two copies of the projector Pm , and normalize the
result. That is, the final state is given by
Pm ρPm
ρ̃m = |m m| = . (1.38)
Tr [Pm ρPm ]
We include the tilde on top of ρm to indicate that it is a state that results from a measure-
ment. This convention will help to make our expressions clearer later on. Note that the
expression on the bottom line is actually the probability pm :
pm = m|ρ|m = Tr [Pm ρPm ] = Tr (Pm )2 ρ = Tr [Pm ρ] , (1.39)
where we have used the cyclic property of the trace: Tr[ABC] = Tr[CAB] = Tr[BCA]. The
reason why we chose to write the expression for the final state using the projectors Pn is
that this form will make it clear later on that von Neumann measurements are a special
case of the more general measurements that will be derived below.
Note that a von Neumann measurement induces a nonlinear change in the density. The
operation of applying the projector Pn is linear, since it is matrix multiplication, but the
operation of rescaling the density matrix so that the final state is normalized is not linear.
Recall that an operation Q( · ) on a vector (a set of numbers) is linear if Q(αv + βu) =
αQ(v) +βQ(u), for all real numbers α, β and vectors v and u. In our case the density matrix
is the “vector,” and it is not hard to see that the above linearity condition is not satisfied by
von Neumann measurements. This nonlinearity is a significant departure from the linear
evolution of Schrödinger’s equation. In classical measurement theory measurements also
change the observer’s state-of-knowledge in a nonlinear way, but they do not affect the
dynamics of the system itself. In Section 1.4.3 we will see that quantum measurements do
cause dynamical changes. Because of this they cause the dynamics of quantum systems to
be nonlinear.
Von Neumann measurements are certainly not the only kinds of measurements one can
make on a quantum system. To gain an understanding of quantum measurements we need
to examine different kinds of measurements, and what kind of information they extract,
and how they affect the systems being measured. We could start by considering a number
of physical examples, and follow this by deriving a general formalism that describes them
all. We have chosen instead to do it the other way around, as we feel that this is quicker.
But instead of following this path, if you prefer, you can skip the derivation and go straight
to Theorem 2 (p. 24), which presents the concise mathematical description of (almost)
all measurements. From Theorem 2 you can then go to Section 1.4 which explains, via
some examples, how this description relates to classical measurements and the classical
likelihood function, and why quantum measurements induce a dynamical change in the
system being measured.
22 Quantum measurement theory
ρcomb = |0 0| ⊗ ρ, (1.40)
where ρ is the initial state of the target. We will always write the state of the probe on the
left-hand side of the tensor product, and that of the target on the right. To understand the
following it is essential to be familiar with how a composite quantum system, consisting of
two subsystems, is described by the tensor product of the spaces of these two subsystems.
If you are not familiar with how to combine two quantum systems using the tensor product,
denoted by ‘⊗’, the details are presented in Appendix A.
where |sk are a set of basis states for the target, and unk,n k are the matrix elements of U.
For each pair of probe states |n and |n , there is a sub-block of U that acts in the space of
the target. We can alternatively write U in terms of these sub-blocks as
U= |n n | ⊗ Ann , (1.42)
nn
1.3 Quantum measurement theory 23
The relationship between the matrices Ann and U is most clear when they are written as
matrices, rather than in Dirac notation. This relationship is
⎛ ⎞
A00 A01 ··· A0M
⎜ A10 A11 ··· A1M ⎟
⎜ ⎟
U=⎜ .. .. .. .. ⎟ (1.44)
⎝ . . . . ⎠
AM0 AM1 ··· AMM
Recall that the system has dimension N and so each matrix Ann is N-dimensional. In what
follows we will use the notation An as shorthand for the matrices An0 . These matrices
constitute the first column of the sub-blocks of U. Let us now denote the M × M sub-
blocks of the matrix U † U by Bnn . Because U is unitary, U † U = I. This means that each of
the sub-blocks Bnn (those on the diagonal of U † U) must be equal to the identity. Further,
the sub-block B00 is the product of the first column of the sub-blocks of U, multiplied by
the first row of the sub-blocks of U † . This gives us immediately the relationship
I = B00 = A†n An . (1.45)
n
The final important fact about the structure of U is that, apart from the restriction given
by Eq. (1.45), we can choose the sub-blocks An ≡ A0n to be any set of operators. This
is because Eq. (1.45) is alone sufficient to ensure that there exists a unitary U with the
operators An as its first column of sub-blocks. To see why this is true, note that the operators
An , being the first column of sub-blocks, constitute the first N columns of U, where N is
the dimension of the target. To show that Eq. (1.45) ensures that U is unitary, we need
merely show that Eq. (1.45) implies that these first N columns are mutually orthonormal.
So long as this is the case, we can always choose the remaining columns of U so that all
the columns of U form an orthonormal basis, making U unitary. This task is not especially
difficult, and we leave it as an exercise.
We have placed a caron ( ˇ ) over the final state, σ , to indicate that it is not necessarily
normalized, due to the fact that we have projected the system onto a subspace. Writing U
24 Quantum measurement theory
in terms of its sub-blocks, we immediately obtain a simple form for the final state:
σ̌ = |n n| ⊗ An ρA†n . (1.47)
An ρA†n
ρ̃n = , (1.48)
Tr[A†n An ρ]
We now have a complete description of what happens to a quantum system under a gen-
eral measurement process. Further, we know that every set of operators {An } that satisfies
Eq. (1.45) describes a measurement that can be realized by using an interaction with a
probe system. For emphasis we state this result as a theorem.
An ρA†n
ρ̃n = , (1.50)
pn
pn = Tr[A†n An ρ]. (1.51)
We refer to the operators {An } as the measurement operators for the measurement.
In fact, the form given for quantum measurements in the above theorem, in which
each outcome is associated with a single measurement operator, does not cover all pos-
sible measurements. But the additions required to describe every possible measurement
1.3 Quantum measurement theory 25
are purely classical. We need merely include classical uncertainty about the result of
the measurement, or about what measurement was performed. Because of this all the
purely quantum features of any measurement are captured by a set of operators {An }, with
†
n An An = I. We show how to add additional classical uncertainty to quantum measure-
ments in Section 1.6 below. Measurements that have no additional classical uncertainty are
referred to as efficient measurements, and we explain the reason for this terminology in
Section 1.6.
Theorem 2 provides a simple mathematical description of quantum measurements, but
we don’t yet know the physical meaning of the operators {An }. That is, we don’t know how
the form of a particular operator An is related to the physical properties of the measurement.
We turn to this question in Section 1.4, but before we do, we discuss a few useful concepts
and topics.
It is useful to define a “partial inner product” that is closely related to the partial projector
P. We define the inner product n|A |ψ as
n|A |ψ = n|A αmk |m A |k B ≡ αmk n|m A |k B , (1.53)
mk mk
which is a state of system B alone. The partial inner product takes the inner product with
the states of one system, and leaves the states of the other alone. One way to define this
inner product is to say that when we have an expression of the form a|A |ψ , where a|A is
a state of one subsystem, and |ψ is a joint state of both systems, then a|A is shorthand for
the unnormalized vector k a|A k|B .
To make clearer the relationship between the projector Pn and the partial inner product,
denote the state of system B that is produced by the partial inner product as |ψn B =
n|A |ψ . If we apply Pn = |n A n|A ⊗ IB to |ψ then this contains within it the application
of the partial inner product:
With this definition of the partial inner product, if we have a joint operator X of the two
subsystems, then n|A X|n A is a sub-block of the matrix X that operates only on system B.
If you are not familiar with this fact then you can show it as an exercise.
We also note that the partial trace, described in detail in Appendix A, can be written
in terms of the partial inner product. If the joint density matrix is ρ, the partial trace over
system A is ρb = n n|A ρ|n A .
Discarding a system
Let us say we have a quantum system S that is initially in a state ρ. It then interacts with
a second system B over which we have no control, and to which we have no access. What
is our state-of-knowledge of S following the interaction? One way to answer this question
is to realize that another observer could measure B, just as we measured the probe in the
scenario above. Since we do not know the outcome of this measurement, our state-of-
knowledge is given by averaging over all the outcomes of the measurement on B. It is
therefore
ρ̃ = pn ρ̃n = An ρA†n = n|U(|0 0| ⊗ ρ)U †|n
n n n
†
≡ TrB [U(|0 0| ⊗ ρ)U ], (1.55)
where we have used the partial-inner-product notation, |n denotes a basis state of the
probe, |0 0| is the initial state of the probe, and TrB denotes the partial trace over system
B, which is defined and discussed in detail in Appendix A.
So our state-of-knowledge after the interaction is given by the partial trace over the sec-
ond system. In fact, the partial trace is completely independent of the measurement that
the observer makes on the second system. Of course, this is crucial, because we cannot
know what measurement the observer will make, so our state-of-knowledge must be inde-
pendent of this measurement. The invariance of the partial trace under operations on the
traced system is proved in Appendix A.
evolution of the joint system is positive. To obtain the map for the first system we start
with the map for the joint system, and take the trace over the second system. The subtlety
is that not all positive linear maps for the first system are derivable from positive linear
maps for the joint system. Only the subset of positive maps that can be derived in this way
are valid evolutions. These maps are referred to as being completely positive. We note that
the definition of a “quantum operation” is any completely positive linear map that may or
may not preserve the trace of the density matrix.
Every realizable quantum evolution, L(ρ), can be written in the form
L(ρ) = An ρA†n , (1.56)
n
where n A†n An = I. This follows immediately from our analysis of a quantum measure-
ment above. The evolution of every quantum system is ultimately the result of a unitary
evolution on a larger system of which the system is a subsystem, and the evolution for the
subsystem is given by tracing out the rest of the larger system. This trace can be written
as the average over the results of a von Neumann measurement on the rest of the larger
system, and this results in the form in Eq. (1.56), being the average over the results of a
measurement on the system.
The form given in Eq. (1.56) is called an “operator-sum” representation for a quantum
evolution. The result that any linear map is completely positive if and only if it can be
written in the operator-sum representation was first proved by Choi [124]. Another useful
fact that is not obvious is that the evolution of an N-dimensional quantum system can
always be written using no more than N 2 operators {An }. This was first shown by Hellwig
and Kraus [247, 345], and the measurement operators {An } are often referred to as Kraus
operators. Schumacher proved this result in an elegant and more direct way by defining a
linear map using the partial inner product, and we give his proof in Appendix H.
Prob(n ∈ M) = pn = Tr[A†n An ρ] = Tr A†n An ρ . (1.57)
n∈M n∈M n∈M
28 Quantum measurement theory
Since the operators A†n An are all positive, the operator in the parentheses is also positive.
This is the positive operator associated with the set M, and is thus the positive operator of
the “positive-operator valued measure.”
Note that in this case the system is not in a superposition of the basis states. It actually is
in one of the basis states; we merely do not know which one. Recall that we refer to this
situation by saying that the system is in a statistical mixture (or simply a mixture) of the
states {|n }. Now, if we make a measurement on the system purely to obtain information
about the basis state it is in, then we might expect this to be described perfectly well
by a classical measurement. This is because the basis states are distinguishable, just like
classical states, and our state-of-knowledge is a probability distribution over these states,
just like a classical state of knowledge.
With the above thoughts in mind, consider a quantum measurement in which all the
measurement operators Aj are diagonal in the same basis as the density matrix. Note that
we can think of the diagonal elements of the measurement operators, and of ρ, simply as
functions of the index n. That is, we can write
ρ= P(n)|n n|, (1.59)
n
Aj = A(j, n)|n n|, (1.60)
n
where P(n) and A(j, n) are the functions in question. The final (or posterior) state-of-
knowledge that results from our measurement is
Aj ρA†j nA
2 (j, n)P(n)|n n|
ρ̃j = = . (1.61)
Tr[A†j Aj ρ] N
1.4 Understanding quantum measurements 29
This is precisely the same form as the update rule for classical measurement theory. All we
have to do is write the posterior state ρ̃j as
ρ̃j = P(n|j)|n n|, (1.62)
n
P(j|n)P(n)
P(n|j) = . (1.63)
N
So, at least in this case, quantum measurements are merely classical measurements, and the
diagonal of the measurement operator Aj corresponds to the square-root of the likelihood
function evaluated for the jth outcome.
Two operators A and B are diagonal in the same basis if and only if they commute with
one another. In view of the above results, we will refer to quantum measurements for which
If κ = 1 (or κ = 0) then both measurement operators are projectors, and after the mea-
surement we are therefore certain about the state of the system. If κ = 1/2 then both
operators are proportional to the identity. They therefore have no action on the system, and
we learn nothing from the measurement. For κ ∈ (1/2, 1) the measurement is an “incom-
plete” measurement. It changes the probabilities p0 and p1 , but does not reduce either to
zero.
30 Quantum measurement theory
As discussed in the previous section, since the measurement operators commute with
the density matrix, their action on the diagonal elements of the density matrix is simply
that of Bayesian inference. The measurement operators we have chosen above correspond
precisely to the simple two-outcome measurement that we discussed in Section 1.2.1, with
the identification α = κ. So, for example, if κ > 1/2 and we get the outcome 0, then the
diagonal elements of ρ are re-weighted so as to make the state |0 more likely than it was
before. Finite-strength measurements thus provide some information about which state the
system is in, but not complete information.
We will see in Section 1.4.4 that we can, in fact, understand the action of all quantum
measurements, to a large extent, in terms of Bayesian inference.
Aα = f (α)e−k(X−α) ,
2
(1.68)
and where f (α) is the scaling required for each operator Aα so that α Aα = I. The larger
the parameter k, the higher the resolving power of the measurement.
“We could see all the men jumping overboard,” he wrote. “She
continued heeling over, and it was simply agonising to watch the
wretched men struggling out of the ports over the ship’s bottom in
masses. All this, of course, happened in less time than it takes to
write. You could see the poor men who, in their hurry to jump over,
jumped on to the screw being cut to pieces as it revolved. She
heeled right over, the water rushing in through her funnels. A great
explosion of steam rose; she turned right over, and you could see all
the men eagerly endeavouring to crawl over her bottom, when, with
a plunge, she went down bows first. We could see her stern rise
right out of the water and plunge down, the screws still revolving. It
was simply a dreadful sight. We could not realise it. Personally, I was
away in my boat, pulling as hard as we could to the scene of the
disaster.... After pulling up and down for two hours, we reorganised
the fleet, leaving two ships on the scene of the disaster; and,
making for Tripoli, anchored for the night. No one can realise the
dreadful nature of the accident.
“However, dropping the Victoria for a minute, we must turn to the
Camperdown. She appeared to be in a very bad way. Her bow was
sinking gradually, and I must say at the time I thought it quite on
the cards that she might be lost also; but, thanks to the indomitable
way in which the crew worked, they managed to check the inrush by
means of the collision mat and water-tight doors. All last night,
however, they were working hard to keep her afloat.
“You can imagine our feelings—the flagship sunk with nearly all
hands, the other flagship anchored in a sinking condition. We have a
lot of the survivors of the Victoria on board, but their accounts vary
greatly.... Anyhow, what is quite certain is that the admiral did not
realise the gravity of his situation, or else he would have abandoned
the ship at once, instead of trying to save her. The discipline was
magnificent. Not until the order was given did a single man jump
overboard.
“The last thing that was seen was the admiral refusing to try to save
himself, whilst his coxswain was entreating him to go. Another
instance of pluck was exhibited by the boatswain of signals, who
was making a general semaphore until the water washed him away.
Unfortunately the poor chap was drowned. Many of the survivors are
in a dreadful state of mental prostration. Most people say that
Admiral Markham should have refused to obey the signal, but I think
that Admiral Tryon infused so much awe in most of the captains of
the fleet that few would have disobeyed him. However, he stuck to
his ship to the last, and went down in her.”
Thus was the Victoria lost; less than a quarter of an hour after being
struck she was lying at the bottom of the Mediterranean, Admiral
Tryon and 400 gallant seamen going with her.
At the court-martial Captain Bourke was absolved of all blame for
the loss of the ship, the finding being that the disaster was entirely
due to Admiral Tryon’s order to turn the two lines sixteen points
inward when they were only six cables apart.
INCIDENTS IN THE SLAVE TRADE
Stories of the Traffic in Human Merchandise
WE shall not here deal with the history and abolition of slavery,
because every schoolboy knows all about that, and will doubtless be
glad to have something more exciting. And of excitement there is
abundance in the annals of slavery. The trade was always attended
by risks, even before the days when it was illegal to ship slaves, for
there was ever the danger of the negroes breaking loose and
running amok on the ship; or, what was perhaps worse, the holds of
the slavers were often little less than death-holes, with fever and
cholera rampant. Altogether, it was a game with big profits—and
mighty big risks, as the following story will show:
It was back in 1769 that the slaver Delight (Captain Millroy) was the
scene of an uprising of negroes, which resulted in a rousing fight
and fatal effects to a good many aboard.
About three o’clock one Sunday morning Surgeon Boulton and the
men with him in the aft-cabin were awakened by a chorus of
screams and shrieks overhead, a rushing of feet, a pandemonium of
noise which told that something serious was afoot. Boulton slipped
out of his bunk and dashed towards the captain’s cabin, half
guessing what was taking place. He reached the cabin, and,
entering, shook Millroy fiercely to awaken him. He had barely
succeeded in rousing the captain when a billet of wood came
hurtling through the air and caught him on the shoulder, and a
cutlass pierced his neck. Turning, Boulton saw that a couple of
negroes had, all unseen and unheard, crept below, intent on putting
the captain hors de combat while he was asleep; and, finding the
surgeon interfering with their plot, they attacked him in quick time.
Millroy, now properly aroused, joined forces with Boulton, who forgot
his own danger in the thought of what was happening above, and
the pair chased the negroes on to the deck, Boulton carrying a pistol
and the captain a cutlass.
When they reached deck they found themselves in a very inferno.
Hundreds of negroes were swarming all over the place, some armed
with wooden spars, others with cutlasses; and with these weapons
they were hard at it taking vengeance on their captors. The herd of
savages flung themselves upon the seamen, cutting off legs and
arms, mutilating bodies dreadfully, their yells making the air ring.
Boulton and the captain, realising that it was a case for prompt and
vigorous action, hurled themselves into the heaving fight with a will.
Down went one negro, killed by Millroy’s cutlass; then another; while
Boulton did all he could. But the “all” of these two men was but
little, and presently Millroy fell to the deck, overpowered by
numbers, and literally hacked to pieces. Boulton, more fortunate,
escaped injury, and made a dash for the rigging, up which he
scrambled till he came to the maintop, where he discovered the cook
and a boy had already taken refuge.
Perched on their lofty platform, the three looked down upon deck,
watching as though fascinated the drama being enacted before their
eyes, seeing the now maddened negroes wreaking vengeance on
the men who were bearing them from freedom to slavery. The
bloodlust was upon them, and they searched the ship to take their
fill.
Suddenly the watchers saw two men come up from below and make
a rush across the deck to the rigging. Like lightning the negroes
dashed after them, and one man was brought to deck by a dozen
billets flung at him, and his body was cut to pieces. The second
man, more fortunate, managed to reach the rigging, and clambered
up like a monkey.
The negroes, having satisfied themselves that they had accounted
for all the crew with the exception of those in the maintop, whom
they decided to deal with presently, ransacked the ship, seeking
arms; and meanwhile Boulton, knowing that safety depended upon
weapons, went on a tour of exploration. He wormed his way into the
foretop to see what might be there, and luckily found a knife, with
which he set out to return to the maintop. On the way the negroes
saw him, and began to pelt him with billets of wood, all of which
missed, however; so that Boulton reached his comrades safely. The
one dread in the minds of the four survivors was that the negroes
would find the arms-chest, in which case it seemed to them
hopeless to expect to escape. While the slaves remained armed only
with wooden spars and cutlasses, Boulton did not feel particularly
anxious, knowing that he and his companions would be able to
tackle any who dared to ascend the rigging to try and get them
down. One thing that kept him hopeful was the fact that another
slaver, the Apollo, was almost within hailing distance, and the
Delight, unsteered and sails untrimmed, was rapidly drifting towards
her, which would make the men on the Apollo aware that something
had happened. But Boulton’s luck was out. The negroes found the
arms-chest, and, breaking it open, armed themselves with muskets,
and set to work in earnest to put the survivors out of action.
Shot after shot sang by the maintop, and one of the men there,
fearing that he would be killed if he stayed, and might be saved if he
trusted himself to the mercy of the negroes, like a madman
descended to the deck. Barely had his foot touched it when a negro
fell upon him with an axe and split his head in two; and a dozen
pairs of hands seized him and pitched him overboard to the sharks
which were following the ship, their appetites whetted by the feasts
already given them by the negroes.
While this was going on, other slaves were still shooting away at the
maintop, fruitlessly; and Boulton was calling madly on the Apollo,
now not far away. Presently the captain of the other vessel, realising
what was afoot, gave the word, and a broadside hurtled across the
deck of the Delight, in the hope of frightening the slaves. They
seemed to take little notice of this, however, and Boulton began to
fear that all was over, especially as the negroes, seeing that they
could not hit the men in the maintop, ceased fire, and a giant black,
cutlass in one hand and a pistol in the other, sprang into the rigging,
bent, apparently, on storming the position. Boulton waited calmly.
He had no weapons but his knife and a quart bottle; but he felt that
he was in a good position to meet an attack. Presently the negro’s
head appeared above the platform, and then—whack! The bottle fell
upon it with a sickening thud, the black lost his hold, and went
hurtling into the sea.
Meanwhile, the Apollo was firing at the Delight, and the latter was
returning the fire as well as it could, the negroes evidently knowing
that to give in was to court disaster, and to lose what they had stood
in a fair way to gaining. For four hours they fought the Apollo, and
at the same time kept up their fusillade on the maintop.
Then came the end. Not because the negroes were not able to keep
up any longer, but because a shot from the Apollo fell into a barrel of
gunpowder and exploded it, with the result that the Delight took fire,
and the slaves could not cope with the flames and their enemy at
the same time. The revolt fizzled out as quickly as it had arisen.
While the negroes rushed about seeking to put out the fire, Boulton,
taking his life in his hands, descended to the deck, at the same time
that a boat set out from the Apollo with a crew to tackle the flames
and the negroes, who, filled with consternation, now stood quietly
by watching the fire-fighters. They were absolutely cowed; they had
made their bid for freedom, and had failed, and they knew it. They
allowed themselves to be driven below and secured. The result of
their revolt was that nine of the crew of the Delight were butchered,
one man on the Apollo was killed, and eighteen of the negroes
found death instead of liberty—perhaps death to them was better
than freedom; certainly better than the lot of those poor human
cattle they left behind them.
Such incidents as this were of frequent occurrence, and the recital of
one must suffice.
After the Abolition Act had been passed, severe measures were
brought into operation, giving the Navy a wide scope—so wide that,
even although a vessel had no slaves on board, yet, if the naval
officers had reason to suspect that slaving was her business, they
could apprehend her. Special ships were fitted out and commissioned
to deal with the traffic in the South Atlantic, both off Central America
and the West Coast of Africa. So effective were the measures taken
that the slavers resorted to all manner of disguises to turn suspicion
away from their vessels, which had hitherto been of a distinctive
kind—long, rakish craft with tall spars, the whole effect being one of
beauty, and the idea being speed. The traders changed all this by
having ships more after the fashion of the ordinary merchant vessel,
so that the hunters had a more difficult task in front of them. But
they worked energetically, and swept the seas month after month,
on the look-out for the human cattle-ships, and, as all the world
knows, succeeded in clearing them from the seas.
The subjoined account from the Sierra Leone Watchman for
November 15, 1846, gives a striking picture of the conditions against
which the Navy were doing such good work.
The vessel referred to is the Brazilian brigantine Paqueta de Rio,
captured off Sherbro:
“The 547 human beings—besides the crew and passengers (as they
styled themselves), twenty-eight in number—were stowed in a
vessel of 74 tons. The slaves were all stowed together, perfectly
naked, with nothing on which to rest but the surfaces of the water-
casks. These were made level by filling in billets of wood, and
formed the slave-deck. The slaves who were confined in the hold—it
being utterly impossible for the whole of them to remain on deck at
one time—were in a profuse perspiration, and panting like so many
hounds for water. The smell on board was dreadful. I was informed
that, on the officers of the Cygnet boarding the slaver, the greater
part of the slaves were chained together with pieces of chain, which
were passed through iron collars round their necks; iron shackles
were also secured round their legs and arms. After the officers had
boarded, and the slaves were made to understand they were free,
their acclamations were long and loud. They set to work, and, with
the billets of wood which had hitherto formed their bed, knocked off
each other’s shackles, and threw most of them overboard. There
were several left, which were shown to me. We will leave it to the
imagination of your readers what must have been the feelings of
these poor people when they found they were again free—free
through the energy and activity of a British cruiser. On examining the
poor creatures, who were principally of the Kosso nation, I found
they belonged to, and were shipped to, different individuals; they
were branded like sheep. Letters were burnt in the skin two inches
in length. Many of them, from the recent period it had been done,
were in a state of ulceration. Both males and females were marked
as follows: On the right breast ‘J’; on the left arm, ‘P’; over women’s
right and left breasts, ‘S’ and ‘A’; under the left shoulder, ‘P’; right
breast, ‘R’ and ‘RJ’; on the right and left breasts, ‘SS’; and on the
right and left shoulder, ‘SS.’ This is the same vessel that cleared out
from here about three weeks previous to her capture for Rio de
Janeiro. The slaves were all embarked from the slave factories at
Gallinas, under the notorious Don Luiz, and the vessel under way in
five hours; and had there been the slightest breeze she would have
escaped. Among the slaves there were two men belonging to Sierra
Leone—a man named Peter, once employed by Mr. Elliott, the pilot.
He stated that he had been employed by a Mr. Smith, a Popohman,
to go to Sherbro to purchase palm-oil, and that whilst pursuing that
object he was seized and sold by a Sherbro chief named Sherry.”
A RACE TO SUCCOUR
An Incident of the United States Revenue
Service
THE records of the revenue men of the United States teem with
heroic deeds done in the execution of their duty. The present story is
typical of the thrilling determination of men who will not be beaten,
and incidentally shows a healthy rivalry between the revenue men
and the lifeboatmen.
On January 11, 1891, the three-masted schooner Ada Barker
encountered a terrific storm which played shuttlecock with her, and
after a fierce conflict pitched her on to the Junk of Pork, the
euphonious name of a large rock near outer Green Island, off the
coast of Maine. The Junk of Pork rises a sheer fifty feet out of the
water, and all round it are reefs and boulders, a literal death-trap to
any unfortunate vessel that should get caught there. The Ada
Barker, after having her sails torn to shreds and her rigging
hopelessly entangled, began to ship water, and though her men
worked hard and long at the pumps, they could not save her; then
she was bowled on to the outer reef at night; the bottom dropped
out of her, and she heeled over. To the men on board it seemed that
the end of all things had come, and they gave themselves up for
lost.
“Though her men worked hard and long at the pumps, they
could not save her”
Our website is not just a platform for buying books, but a bridge
connecting readers to the timeless values of culture and wisdom. With
an elegant, user-friendly interface and an intelligent search system,
we are committed to providing a quick and convenient shopping
experience. Additionally, our special promotions and home delivery
services ensure that you save time and fully enjoy the joy of reading.
ebookgate.com