Markov Chains and Mixing Times Second Edition David A. Levin All Chapter Instant Download
Markov Chains and Mixing Times Second Edition David A. Levin All Chapter Instant Download
Markov Chains and Mixing Times Second Edition David A. Levin All Chapter Instant Download
com
https://textbookfull.com/product/markov-
chains-and-mixing-times-second-edition-david-
a-levin/
textbookfull
More products digital (pdf, epub, mobi) instant
download maybe you interests ...
https://textbookfull.com/product/understanding-markov-chains-
examples-and-applications-2nd-edition-nicolas-privault/
https://textbookfull.com/product/foundations-of-average-cost-
nonhomogeneous-controlled-markov-chains-xi-ren-cao/
https://textbookfull.com/product/mixing-music-1st-edition-russ-
hepworth-sawyer/
https://textbookfull.com/product/english-as-a-global-language-
second-edition-david-crystal/
Understanding Your Migraines: A Guide for Patients and
Families 1st Edition Morris Levin
https://textbookfull.com/product/understanding-your-migraines-a-
guide-for-patients-and-families-1st-edition-morris-levin/
https://textbookfull.com/product/gmo-food-a-reference-handbook-
second-edition-david-e-newton/
https://textbookfull.com/product/the-mixing-engineer-s-
handbook-3rd-edition-bobby-owsinski/
https://textbookfull.com/product/the-mixing-engineer-s-
handbook-4th-edition-bobby-owsinski/
https://textbookfull.com/product/wide-range-and-multi-frequency-
antennas-levin/
10.1090/mbk/107
Markov Chains
and Mixing Times
Second Edition
David A. Levin
University of Oregon
Yuval Peres
Microsoft Research
With contributions by
Elizabeth L. Wilmer
A M E R I C A N M AT H E M AT I C A L S O C I E T Y
Providence, Rhode Island
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
2010 Mathematics Subject Classification. Primary 60J10, 60J27, 60B15, 60C05, 65C05,
60K35, 68W20, 68U20, 82C22.
FRONT COVER: The figure on the bottom left of the front cover, courtesy of
David B. Wilson, is a uniformly random lozenge tiling of a hexagon (see Section
25.2). The figure on the bottom center, also from David B. Wilson, is a random
sample of an Ising model at its critical temperature (see Sections 3.3.5 and 25.2)
with mixed boundary conditions. The figure on the bottom right, courtesy of Eyal
Lubetzky, is a portion of an expander graph (see Section 13.6).
Copying and reprinting. Individual readers of this publication, and nonprofit libraries
acting for them, are permitted to make fair use of the material, such as to copy select pages for
use in teaching or research. Permission is granted to quote brief passages from this publication in
reviews, provided the customary acknowledgment of the source is given.
Republication, systematic copying, or multiple reproduction of any material in this publication
is permitted only under license from the American Mathematical Society. Permissions to reuse
portions of AMS publication content are handled by Copyright Clearance Center’s RightsLink
service. For more information, please visit: http://www.ams.org/rightslink.
Send requests for translation rights and licensed reprints to [email protected].
Excluded from these provisions is material for which the author holds copyright. In such cases,
requests for permission to reuse or reprint material should be addressed directly to the author(s).
Copyright ownership is indicated on the copyright page, or on the lower right-hand corner of the
first page of each article within proceedings volumes.
Second edition c 2017 by the authors. All rights reserved.
First edition
c 2009 by the authors. All rights reserved.
Printed in the United States of America.
∞ The paper used in this book is acid-free and falls within the guidelines
established to ensure permanence and durability.
Visit the AMS home page at http://www.ams.org/
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
Contents
Preface ix
Preface to the Second Edition ix
Preface to the First Edition ix
Overview xi
For the Reader xii
For the Instructor xiii
For the Expert xiv
Acknowledgements xvi
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
iv CONTENTS
Chapter 5. Coupling 60
5.1. Definition 60
5.2. Bounding Total Variation Distance 61
5.3. Examples 62
5.4. Grand Couplings 69
Exercises 73
Notes 73
Chapter 6. Strong Stationary Times 75
6.1. Top-to-Random Shuffle 75
6.2. Markov Chains with Filtrations 76
6.3. Stationary Times 77
6.4. Strong Stationary Times and Bounding Distance 78
6.5. Examples 81
6.6. Stationary Times and Cesàro Mixing Time 83
6.7. Optimal Strong Stationary Times* 84
Exercises 85
Notes 86
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
CONTENTS v
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
vi CONTENTS
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
CONTENTS vii
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
viii CONTENTS
Bibliography 425
Index 441
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
Preface
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
x PREFACE
the stationary distribution is prescribed; the number of steps required to reach this
target is called the mixing time of the chain. Now, the goal is to understand how
the mixing time grows as the size of the state space increases.
The modern theory of Markov chain mixing is the result of the convergence, in
the 1980s and 1990s, of several threads. (We mention only a few names here; see
the chapter Notes for references.)
For statistical physicists Markov chains become useful in Monte Carlo simu-
lation, especially for models on finite grids. The mixing time can determine the
running time for simulation. However, Markov chains are used not only for sim-
ulation and sampling purposes, but also as models of dynamical processes. Deep
connections were found between rapid mixing and spatial properties of spin systems,
e.g., by Dobrushin, Shlosman, Stroock, Zegarlinski, Martinelli, and Olivieri.
In theoretical computer science, Markov chains play a key role in sampling and
approximate counting algorithms. Often the goal was to prove that the mixing
time is polynomial in the logarithm of the state space size. (In this book, we are
generally interested in more precise asymptotics.)
At the same time, mathematicians including Aldous and Diaconis were inten-
sively studying card shuffling and other random walks on groups. Both spectral
methods and probabilistic techniques, such as coupling, played important roles.
Alon and Milman, Jerrum and Sinclair, and Lawler and Sokal elucidated the con-
nection between eigenvalues and expansion properties. Ingenious constructions of
“expander” graphs (on which random walks mix especially fast) were found using
probability, representation theory, and number theory.
In the 1990s there was substantial interaction between these communities, as
computer scientists studied spin systems and as ideas from physics were used for
sampling combinatorial structures. Using the geometry of the underlying graph to
find (or exclude) bottlenecks played a key role in many results.
There are many methods for determining the asymptotics of convergence to
stationarity as a function of the state space size and geometry. We hope to present
these exciting developments in an accessible way.
We will only give a taste of the applications to computer science and statistical
physics; our focus will be on the common underlying mathematics. The prerequi-
sites are all at the undergraduate level. We will draw primarily on probability and
linear algebra, but we will also use the theory of groups and tools from analysis
when appropriate.
Why should mathematicians study Markov chain convergence? First of all, it is
a lively and central part of modern probability theory. But there are ties to several
other mathematical areas as well. The behavior of the random walk on a graph
reveals features of the graph’s geometry. Many phenomena that can be observed in
the setting of finite graphs also occur in differential geometry. Indeed, the two fields
enjoy active cross-fertilization, with ideas in each playing useful roles in the other.
Reversible finite Markov chains can be viewed as resistor networks; the resulting
discrete potential theory has strong connections with classical potential theory. It
is amusing to interpret random walks on the symmetric group as card shuffles—and
real shuffles have inspired some extremely serious mathematics—but these chains
are closely tied to core areas in algebraic combinatorics and representation theory.
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
OVERVIEW xi
In the spring of 2005, mixing times of finite Markov chains were a major theme
of the multidisciplinary research program Probability, Algorithms, and Statistical
Physics, held at the Mathematical Sciences Research Institute. We began work on
this book there.
Overview
We have divided the book into two parts.
In Part I, the focus is on techniques, and the examples are illustrative and
accessible. Chapter 1 defines Markov chains and develops the conditions necessary
for the existence of a unique stationary distribution. Chapters 2 and 3 both cover
examples. In Chapter 2, they are either classical or useful—and generally both;
we include accounts of several chains, such as the gambler’s ruin and the coupon
collector, that come up throughout probability. In Chapter 3, we discuss Glauber
dynamics and the Metropolis algorithm in the context of “spin systems.” These
chains are important in statistical mechanics and theoretical computer science.
Chapter 4 proves that, under mild conditions, Markov chains do, in fact, con-
verge to their stationary distributions and defines total variation distance and
mixing time, the key tools for quantifying that convergence. The techniques of
Chapters 5, 6, and 7, on coupling, strong stationary times, and methods for lower
bounding distance from stationarity, respectively, are central to the area.
In Chapter 8, we pause to examine card shuffling chains. Random walks on the
symmetric group are an important mathematical area in their own right, but we
hope that readers will appreciate a rich class of examples appearing at this stage
in the exposition.
Chapter 9 describes the relationship between random walks on graphs and
electrical networks, while Chapters 10 and 11 discuss hitting times and cover times.
Chapter 12 introduces eigenvalue techniques and discusses the role of the re-
laxation time (the reciprocal of the spectral gap) in the mixing of the chain.
In Part II, we cover more sophisticated techniques and present several detailed
case studies of particular families of chains. Much of this material appears here for
the first time in textbook form.
Chapter 13 covers advanced spectral techniques, including comparison of Dirich-
let forms and Wilson’s method for lower bounding mixing.
Chapters 14 and 15 cover some of the most important families of “large” chains
studied in computer science and statistical mechanics and some of the most impor-
tant methods used in their analysis. Chapter 14 introduces the path coupling
method, which is useful in both sampling and approximate counting. Chapter 15
looks at the Ising model on several different graphs, both above and below the
critical temperature.
Chapter 16 revisits shuffling, looking at two examples—one with an application
to genomics—whose analysis requires the spectral techniques of Chapter 13.
Chapter 17 begins with a brief introduction to martingales and then presents
some applications of the evolving sets process.
Chapter 18 considers the cutoff phenomenon. For many families of chains where
we can prove sharp upper and lower bounds on mixing time, the distance from
stationarity drops from near 1 to near 0 over an interval asymptotically smaller
than the mixing time. Understanding why cutoff is so common for families of
interest is a central question.
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
xii PREFACE
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
FOR THE INSTRUCTOR xiii
hope the reader will realize that this omission of floor or ceiling brackets (and the
details of analyzing the resulting perturbations) is in her or his best interest as
much as it is in ours.
Chapters 1 through 7, shown in gray, form the core material, but there are
several ways to proceed afterwards. Chapter 8 on shuffling gives an early rich
application but is not required for the rest of Part I. A course with a probabilistic
focus might cover Chapters 9, 10, and 11. To emphasize spectral methods and
combinatorics, cover Chapters 8 and 12 and perhaps continue on to Chapters 13
and 16.
While our primary focus is on chains with finite state spaces run in discrete time,
continuous-time and countable-state-space chains are both discussed—in Chapters
20 and 21, respectively.
We have also included Appendix B, an introduction to simulation methods, to
help motivate the study of Markov chains for students with more applied interests.
A course leaning towards theoretical computer science and/or statistical mechan-
ics might start with Appendix B, cover the core material, and then move on to
Chapters 14, 15, and 25.
Of course, depending on the interests of the instructor and the ambitions and
abilities of the students, any of the material can be taught! Below we include
a full diagram of dependencies of chapters. Its tangled nature results from the
interconnectedness of the area: a given technique can be applied in many situations,
while a particular problem may require several techniques for full analysis.
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
xiv PREFACE
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
FOR THE EXPERT xv
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
Acknowledgements
The authors thank the Mathematical Sciences Research Institute, the National
Science Foundation VIGRE grant to the Department of Statistics at the University
of California, Berkeley, and National Science Foundation grants DMS-0244479 and
DMS-0104073 for support. We also thank Hugo Rossi for suggesting we embark on
this project. Thanks to Blair Ahlquist, Tonci Antunovic, Elisa Celis, Paul Cuff,
Jian Ding, Ori Gurel-Gurevich, Tom Hayes, Itamar Landau, Yun Long, Karola
Mészáros, Shobhana Murali, Weiyang Ning, Tomoyuki Shirai, Walter Sun, Sith-
parran Vanniasegaram, and Ariel Yadin for corrections to an earlier version and
making valuable suggestions. Yelena Shvets made the illustration in Section 6.5.4.
The simulations of the Ising model in Chapter 15 are due to Raissa D’Souza. We
thank László Lovász for useful discussions. We are indebted to Alistair Sinclair for
his work co-organizing the M.S.R.I. program Probability, Algorithms, and Statisti-
cal Physics in 2005, where work on this book began. We thank Robert Calhoun
for technical assistance.
Finally, we are greatly indebted to David Aldous and Persi Diaconis, who initi-
ated the modern point of view on finite Markov chains and taught us much of what
we know about the subject.
xvi
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
Part I: Basic Methods and Examples
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
10.1090/mbk/107/01
CHAPTER 1
Example 1.1. A certain frog lives in a pond with two lily pads, east and west.
A long time ago, he found two coins at the bottom of the pond and brought one
up to each lily pad. Every morning, the frog decides whether to jump by tossing
the current lily pad’s coin. If the coin lands heads up, the frog jumps to the other
lily pad. If the coin lands tails up, he remains where he is.
Let X = {e, w}, and let (X0 , X1 , . . . ) be the sequence of lily pads occupied
by the frog on Sunday, Monday, . . .. Given the source of the coins, we should not
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
1.1. MARKOV CHAINS 3
assume that they are fair! Say the coin on the east pad has probability p of landing
heads up, while the coin on the west pad has probability q of landing heads up.
The frog’s rules for jumping imply that if we set
P (e, e) P (e, w) 1−p p
P = = , (1.2)
P (w, e) P (w, w) q 1−q
then (X0 , X1 , . . . ) is a Markov chain with transition matrix P . Note that the first
row of P is the conditional distribution of Xt+1 given that Xt = e, while the second
row is the conditional distribution of Xt+1 given that Xt = w.
Assume that the frog spends Sunday on the east pad. When he awakens Mon-
day, he has probability p of moving to the west pad and probability 1 − p of staying
on the east pad. That is,
P{X1 = e | X0 = e} = 1 − p, P{X1 = w | X0 = e} = p. (1.3)
What happens Tuesday? By considering the two possibilities for X1 , we see that
P{X2 = e | X0 = e} = (1 − p)(1 − p) + pq (1.4)
and
1 1 1
0 10 20 0 10 20 0 10 20
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
4 1. INTRODUCTION TO FINITE MARKOV CHAINS
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
1.2. RANDOM MAPPING REPRESENTATION 5
That is, the x-th entry of P f tells us the expected value of the function f at
tomorrow’s state, given that we are at state x today.
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
6 1. INTRODUCTION TO FINITE MARKOV CHAINS
Rather than writing down the transition matrix in (1.11), this chain can be
specified simply in words: at each step, a coin is tossed. If the coin lands heads up,
the walk moves one step clockwise. If the coin lands tails up, the walk moves one
step counterclockwise.
More precisely, suppose that Z is a random variable which is equally likely to
take on the values −1 and +1. If the current state of the chain is j ∈ Zn , then the
next state is j + Z mod n. For any k ∈ Zn ,
P{(j + Z) mod n = k} = P (j, k).
In other words, the distribution of (j + Z) mod n equals P (j, ·).
A random mapping representation of a transition matrix P on state space
X is a function f : X ×Λ → X , along with a Λ-valued random variable Z, satisfying
P{f (x, Z) = y} = P (x, y).
The reader should check that if Z1 , Z2 , . . . is a sequence of independent random
variables, each having the same distribution as Z, and the random variable X0
has distribution μ and is independent of (Zt )t≥1 , then the sequence (X0 , X1 , . . . )
defined by
Xn = f (Xn−1 , Zn ) for n ≥ 1
is a Markov chain with transition matrix P and initial distribution μ.
For the example of the simple random walk on the cycle, setting Λ = {1, −1},
each Zi uniform on Λ, and f (x, z) = x + z mod n yields a random mapping repre-
sentation.
Proposition 1.5. Every transition matrix on a finite state space has a random
mapping representation.
Proof. Let P be the transition matrix of a Markov chain with state space
X = {x1 , . . . , xn }. Take Λ = [0, 1]; our auxiliary random variables Z, Z1 , Z2 , . . .
will be uniformly chosen in this interval. Set Fj,k = ki=1 P (xj , xi ) and define
f (xj , z) := xk when Fj,k−1 < z ≤ Fj,k .
We have
P{f (xj , Z) = xk } = P{Fj,k−1 < Z ≤ Fj,k } = P (xj , xk ).
Note that, unlike transition matrices, random mapping representations are far
from unique. For instance, replacing the function f (x, z) in the proof of Proposition
1.5 with f (x, 1 − z) yields a different representation of the same transition matrix.
Random mapping representations are crucial for simulating large chains. They
can also be the most convenient way to describe a chain. We will often give rules for
how a chain proceeds from state to state, using some extra randomness to determine
where to go next; such discussions are implicit random mapping representations.
Finally, random mapping representations provide a way to coordinate two (or more)
chain trajectories, as we can simply use the same sequence of auxiliary random
variables to determine updates. This technique will be exploited in Chapter 5, on
coupling Markov chain trajectories, and elsewhere.
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
1.3. IRREDUCIBILITY AND APERIODICITY 7
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
8 1. INTRODUCTION TO FINITE MARKOV CHAINS
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
1.5. STATIONARY DISTRIBUTIONS 9
2 4
1 3 5
Figure 1.4. An example of a graph with vertex set {1, 2, 3, 4, 5}
and 6 edges.
Example 1.12. Consider simple random walk on a graph G = (V, E). For any
vertex y ∈ V ,
deg(x)
deg(x)P (x, y) = = deg(y). (1.16)
x∼y
deg(x)
x∈V
To get a probability, we simply normalize by y∈V deg(y) = 2|E| (a fact the reader
should check). We conclude that the probability measure
deg(y)
π(y) = for all y ∈ X ,
2|E|
which is proportional to the degrees, is always a stationary distribution for the
walk. For the graph in Figure 1.4,
2 3 4 2 1
π = 12 , 12 , 12 , 12 , 12 .
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
10 1. INTRODUCTION TO FINITE MARKOV CHAINS
If G has the property that every vertex has the same degree d, we call G d-regular .
In this case, 2|E| = d|V | and the uniform distribution π(y) = 1/|V | for every y ∈ V
is stationary.
A central goal of this chapter and of Chapter 4 is to prove a general yet precise
version of the statement that “finite Markov chains converge to their stationary
distributions.” Before we can analyze the time required to be close to stationar-
ity, we must be sure that it is finite! In this section we show that, under mild
restrictions, stationary distributions exist and are unique. Our strategy of building
a candidate distribution, then verifying that it has the necessary properties, may
seem cumbersome. However, the tools we construct here will be applied in many
other places. In Section 4.3, we will show that irreducible and aperiodic chains do,
in fact, converge to their stationary distributions in a precise sense.
1.5.2. Hitting and first return times. Throughout this section, we assume
that the Markov chain (X0 , X1 , . . . ) under discussion has finite state space X and
transition matrix P . For x ∈ X , define the hitting time for x to be
τx := min{t ≥ 0 : Xt = x},
the first time at which the chain visits state x. For situations where only a visit to
x at a positive time will do, we also define
Lemma 1.13. For any states x and y of an irreducible chain, Ex (τy+ ) < ∞.
Proof. The definition of irreducibility implies that there exist an integer r > 0
and a real ε > 0 with the following property: for any states z, w ∈ X , there exists a
j ≤ r with P j (z, w) > ε. Thus, for any value of Xt , the probability of hitting state
y at a time between t and t + r is at least ε. Hence for k > 0 we have
Since Px {τy+ > t} is a decreasing function of t, (1.18) suffices to bound all terms of
the corresponding expression for Ex (τy+ ):
Ex (τy+ ) = Px {τy+ > t} ≤ rPx {τy+ > kr} ≤ r (1 − ε)k < ∞.
t≥0 k≥0 k≥0
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
1.5. STATIONARY DISTRIBUTIONS 11
Remark 1.15. Recall that Lemma 1.13 shows that if P is irreducible, then
Ez (τz+ ) < ∞. We will show in Section 1.7 that the assumptions of (i) and (ii) are
always equivalent (Corollary 1.27) and there always exists z satisfying both.
Proof. For any state y, we have π̃(y) ≤ Ez τz+ . Hence Lemma 1.13 ensures
that π̃(y) < ∞ for all y ∈ X . We check that π̃ is stationary, starting from the
definition:
∞
π̃(x)P (x, y) = Pz {Xt = x, τz+ > t}P (x, y). (1.20)
x∈X x∈X t=0
Reversing the order of summation in (1.20) and using the identity (1.21) shows that
∞
π̃(x)P (x, y) = Pz {Xt+1 = y, τz+ ≥ t + 1}
x∈X t=0
∞
= Pz {Xt = y, τz+ ≥ t}. (1.22)
t=1
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
12 1. INTRODUCTION TO FINITE MARKOV CHAINS
The expression in (1.22) is very similar to (1.19), so we are almost done. In fact,
∞
Pz {Xt = y, τz+ ≥ t}
t=1
∞
= π̃(y) − Pz {X0 = y, τz+ > 0} + Pz {Xt = y, τz+ = t}
t=1
= π̃(y) − Pz {X0 = y} + Pz {Xτz+ = y} (1.23)
= π̃(y). (1.24)
The equality (1.24) follows by considering two cases:
y = z: Since X0 = z and Xτz+ = z, the last two terms of (1.23) are both 1, and
they cancel each other out.
y = z: Here both terms of (1.23) are 0.
Therefore, combining (1.22) with (1.24) shows that π̃ = π̃P .
Finally, to get a probability measure, we normalize by x π̃(x) = Ez (τz+ ):
π̃(x)
π(x) = satisfies π = πP. (1.25)
Ez (τz+ )
The computation at the heart of the proof of Proposition 1.14 can be gen-
eralized; see Lemma 10.5. Informally speaking, a stopping time τ for (Xt ) is a
{0, 1, . . . , } ∪ {∞}-valued random variable such that, for each t, the event {τ = t}
is determined by X0 , . . . , Xt . (Stopping times are defined precisely in Section 6.2.)
If a stopping time τ replaces τz+ in the definition (1.19) of π̃, then the proof that
π̃ satisfies π̃ = π̃P works, provided that τ satisfies both Pz {τ < ∞} = 1 and
Pz {Xτ = z} = 1.
1.5.4. Uniqueness of the stationary distribution. Earlier in this chapter
we pointed out the difference between multiplying a row vector by P on the right
and a column vector by P on the left: the former advances a distribution by one
step of the chain, while the latter gives the expectation of a function on states, one
step of the chain later. We call distributions invariant under right multiplication by
P stationary . What about functions that are invariant under left multiplication?
Call a function h : X → R harmonic at x if
h(x) = P (x, y)h(y). (1.26)
y∈X
a contradiction. It follows that h(z) = M for all states z such that P (x0 , z) > 0.
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
1.6. REVERSIBILITY AND TIME REVERSALS 13
Remark 1.18. Another proof of Corollary 1.17 follows from the Convergence
Theorem (Theorem 4.9, proved below). Another simple direct proof is suggested in
Exercise 1.11.
Proposition 1.19. If P is an irreducible transition matrix and π is the unique
probability distribution solving π = πP , then for all states z,
1
π(z) = . (1.28)
Ez τz+
Proof. Let π̃z (y) equal π̃(y) as defined in (1.19), and write πz (y) = π̃z (y)/Ez τz+ .
Proposition 1.14 implies that πz is a stationary distribution, so πz = π. Therefore,
π̃z (z) 1
π(z) = πz (z) = + = .
Ez τz Ez τz+
since P is stochastic.
Checking detailed balance is often the simplest way to verify that a particular
distribution is stationary. Furthermore, when (1.29) holds,
π(x0 )P (x0 , x1 ) · · · P (xn−1 , xn ) = π(xn )P (xn , xn−1 ) · · · P (x1 , x0 ). (1.30)
We can rewrite (1.30) in the following suggestive form:
Pπ {X0 = x0 , . . . , Xn = xn } = Pπ {X0 = xn , X1 = xn−1 , . . . , Xn = x0 }. (1.31)
Copyright no copyright American Mathematical Society. Duplication prohibited. Please report unauthorized use to [email protected].
Thank You! Your purchase supports the AMS' mission, programs, and services for the mathematical community.
Another random document with
no related content on Scribd:
period. A severe penance was imposed on the insolent offender; but
one night he found means to evade the vigilance of his guardians,
escaped from the convent, crossed the intervening country, and after
some days of joyous gipsying and vagabond wanderings, he arrived
at Palermo. Some knowledge of the principles of chemistry and
medicine was about the total of the advantages he had derived from
the discipline of conventual life. His uncle began to despair of him,
but advice and remonstrances were alike lost upon the young
reprobate, who derided them all, and employing a certain portion of
his time in the cultivation of a natural taste for drawing, he otherwise
abandoned himself to unbridled excesses. He associated with
rascals and ne’er-do-wells; his drunkenness, gambling, and general
libertinage, led him into perpetual brawling; and he was frequently in
the hands of the police, whom he is said to have taken special
pleasure in resisting, frequently delivering by force the prisoners
whom they had arrested. He has been also accused of forging
tickets of admission to the theatres, and selling them with
characteristic effrontery. One of his uncles coaxed him back for a
time into his house, and was rewarded by the robbery of a
considerable quantity of money and some valuable effects. He
became an intermediary in the amorous intercourse of a female
cousin with one of his friends. He carried billets-doux to and fro
between them, and made the entire transaction personally profitable
by extorting money from his friend, persuading him that the fair
cousin had a partiality for presents, including both money and
jewellery, and, of course, appropriating the funds which were
entrusted to him. Graver crimes were soon laid to his charge. There
was a certain dissolute Marquis Maurigi in Palermo who coveted an
inheritance which had been willed to a pious establishment, and
knowing Balsamo, to him were his projects confided, and an
expedient was presently forthcoming. Joseph had a relative who was
a notary, and by frequenting his office he found means to forge a will,
bearing every mark of authenticity, in favour of the Marquis, who
made good his claim to the estate, and no doubt liberally
recompensed the skill and pains of his confederate. The falsification
was discovered many years after, but the guilty parties were both of
them far away. It was also rumoured that Balsamo was a party to the
assassination of a wealthy canon, but the matter is exceedingly
doubtful. He was many times arrested on various charges, but
eluded justice, either by the absence of direct proof against him, or
by the credit of his relations, and the exertions of reputable persons
of Palermo, who took interest in his family. It will scarcely be credited
that at this period Balsamo was only fourteen years of age. Naturally
endowed with artistic aptitudes, he soon began to give lessons in
drawing, and seems to have been many times on a fair way to
reformation. His skill in arms is also acknowledged, but, conscious of
his superiority, his street brawls frequently ended in duels; his
impetuosity even prompted him to take up the gauntlet for his
companions, and he scorned danger.
The most notorious of his youthful exploits, and that which caused
him to commence his life-long wanderings, was the adventure of the
concealed treasure, which has been variously related.
An avaricious goldsmith, named Marano, resided at Palermo. He
was a weak, superstitious man—a believer in magic, says M. Louis
Figuier—and he was much attracted by the mystery which, even at
this period, is declared by Figuier to have surrounded the life and
escapades of Balsamo, who already posed as an initiate of the
occult sciences. Joseph was now seventeen years of age, of
handsome mien and haughty carriage, speaking little, but holding his
hearers spell-bound by the magnetic fascination of his glance. He
had been seen evoking spirits; he was believed to converse with
angels, and to obtain by their agency an insight into the most
interesting secrets. He had, in fact, radically changed; the common
rogue was developing into the transcendental impostor. Marano lent
an attentive ear to the stories concerning him, and burned with
anxiety to behold “the friend of the celestial spirits.” The first
interview took place in the lodging of Balsamo; the goldsmith fell on
his knees before him, and Balsamo, after receiving his homage,
raised him condescendingly from the ground, and demanded in a
solemn manner why he had come to him.
“Thanks to your daily communion with spirits, you will easily know,”
answered Marano, “and you should have no difficulty in assisting me
to recover the money which I have wasted among false alchemists,
or even to procure me more.”
“I can perform this service for you, provided you believe,” said
Balsamo, with composure.
“Provided I believe!” cried the goldsmith; “I believe, indeed.”
An appointment was made for the next day in a meadow beyond
the town, and the interview ended without another word.
This version of the story is more romantic than probable, and we
owe it to the vivacity of a Frenchman’s imagination, which is never
more brilliant than when employed in the perversion or
embellishment of history. According to the more sober Aventures de
Cagliostro, Marano had for some time been acquainted with the
youthful charlatan, who sought him one day at his own residence,
and said to him: “You are aware of my communications with the
supernal spirits; you are aware of the illimitable potency of the
incantations to which I devote myself. Listen! In an olive field, at no
great distance from Palermo, there is a buried treasure according to
my certain knowledge, and by the help of a ceremonial evocation I
can discover the precise spot where the spade of the seeker should
be driven in. The operation, however, requires some expensive
preliminaries; sixty ounces of gold are absolutely needed. Will you
place them at my disposal?”
Marano declaimed against the preposterous extravagance of the
demand, maintaining that the herbs and drugs utilised in alchemical
experiments were exceedingly moderate in their price.
“’Tis well,” said Balsamo, coldly. “The matter is soon settled; I shall
enjoy the vast treasure alone. A blessing when shared is but half a
blessing for those who participate in it.”
On the morrow, however, Marano sought out the enchanter,
having been agonised by the gold fever the whole night.
“I am furnished with the sum you require,” he said. “But I pray you
to bargain a little with the spirits, and endeavour to beat them down.”
“Do you take them for sordid speculators?” cried the magician,
indignant. “The devil is no Jew, though he abode full long in Judea.
He is a magnificent seigneur, living generously in every country of
the world. Treat him with respect, he returns a hundredfold. I shall
find elsewhere the sixty ounces of gold, and can afford to dispense
with your assistance.”
“It is here,” said Marano, drawing quickly a leather bag from his
pocket, and the arrangements were soon made.
At moonlight they repaired to the olive field, where Balsamo had
secretly made preparations for the approaching evocation. The
incantatory preliminaries were sufficiently protracted, and Marano
panted with terror under the influence of the magical charms, till it
seemed to him that the very earth shivered beneath his feet and
phantoms issued from the ground. Marano fell prostrate on his face,
an action apparently foreseen, for there and then the wretched
goldsmith was belaboured unmercifully with sticks by the infernal
spirits, who left him at length for dead, taking flight in the company of
the enchanter, and fortified by the possession of the sixty ounces of
gold. On the morrow, the goldsmith, fortunately discovered by
muleteers, was carried disconsolately home, and forthwith
denounced Balsamo to the law. The adventure spread everywhere,
but the magician had sailed for Messina.
These are the facts of the case, but the mendacious chronicle of
Louis Figuier, alchemical critic and universal manufacturer of light
scientific literature, offers us a far more ornate and attractive version.
There the adept and his miserable dupe repair to a place appointed
at six o’clock in the morning, Balsamo in dignified silence motioning
the goldsmith to follow him, and proceeding with a pre-occupied
aspect along the road to the chapel of Saint Rosalia for the space of
a whole hour. They stopped at length in the middle of a wild
meadow, and in front of a grotto, before which Balsamo extended his
hand, and solemnly declared that a treasure was buried within it
which he himself was forbidden to touch, which was guarded by
devils of hell, which devils might, however, be bound for a brief
period by the angels who commonly responded to his potent magical
call.
“It only remains to be ascertained,” he remarked in conclusion,
“whether you will scrupulously fulfil the conditions which must be
imposed on you. At that price, the treasure may be yours.”
The credulous goldsmith impetuously implored him to name them.
“They cannot be learned from my lips,” said Balsamo loftily. “On
your knees, in the first place!”
He himself had already assumed the posture of adoration. Marano
hastened to imitate him, and immediately a clear, harmonious voice
in the celestial altitude pronounced the following words—words, says
the Frenchman, more delicious in the ears of the covetous miser
than all the symphonies of aërial choirs.
“Sixty ounces of pearls, sixty ounces of rubies, sixty ounces of
diamonds, in a coffer of enchased gold, weighing one hundred and
twenty ounces. The infernal genii who protect this treasure will place
it in the hands of the worthy man whom our friend has brought, if he
be fifty years of age, if he be no Christian—if—if—if—” and a series
of conditions followed which Marano perfectly united in his own
penurious person, even to the last, which was thus formulated:
—“And if he deposit at the entrance of the grotto, before setting foot
therein, sixty ounces of gold to propitiate the guardians.”
“You have heard,” said Balsamo, who, already on his feet, began
to retrace his steps, completely ignoring the utter stupefaction of his
companion.
“Sixty ounces of gold!” ejaculated the miser with a dismal groan,
and torn by the internal conflict of avarice and cupidity; but Balsamo
heeded the exclamation as little as the groan, and regained the town
in silence.
When they were on the point of separating, Marano appeared to
have resolved.
“Grant me one instant!” he cried in a piteous voice. “Sixty ounces
of gold? Is that the irrevocable condition?”
“Undoubtedly,” said Balsamo, carelessly.
“Alas! alas! And at what hour to-morrow?”
“At six o’clock in the morning and, mark, at the same spot.”
“I will be there.”
This was the parting speech of the goldsmith, and, as it were, the
last gasp of his conquered avarice. On the morrow, punctual to the
appointed time, they met as before, Balsamo with his habitual
coolness, Marano with his gold. They arrived in due course at the
grotto, where the angels, consulted as on the previous day, returned
the same oracles. Balsamo assumed ignorance of what would take
place. With a terrific struggle, Marano deposited his gold and
prepared to cross the threshold. He took one step forward, then
started back, inquired if there were no danger in penetrating into the
depths of the cavern, was assured of safety if the gold had been
faithfully weighed, entered with more confidence, and again
returned, these manœuvres being repeated several times, under the
eyes of the adept, whose expression indicated the most uninterested
indifference. At length, Marano took courage and proceeded so far
that a return was impossible, for three black, muscular devils started
out from the shadows and barred his path, giving vent to the most
alarming growls. They seized him, forced him to whirl round and
round for a long time, and then while the unhappy creature vainly
invoked the assistance of Balsamo, they proceeded to cudgel him
lustily till he dropped overwhelmed to the ground, when a clear voice
bade him remain absolutely silent and motionless, for he would be
instantaneously despatched if he stirred either hand or foot. The
wretched man did not dare to disobey, but after a long swoon the
complete stillness encouraged him to raise his head; he dragged
himself as best he could to the mouth of the terrible grotto, looked
round him, and found that the adept, the demons, and the gold had
alike vanished.