0% found this document useful (0 votes)
95 views49 pages

Percolation and Random Walks On Graphs - Perla Sousi

This document discusses percolation and random walks on graphs. It begins by defining bond percolation and coupling all percolation processes. There is a phase transition parameter pc such that an infinite cluster exists almost surely if p > pc and does not exist if p < pc. For dimensions d ≥ 2, it is shown that pc is between 0 and 1. The document also discusses self-avoiding walks, connective constants, and open conjectures regarding percolation and self-avoiding walks on lattices. It then briefly introduces random walks on graphs and electrical networks.

Uploaded by

skb
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
95 views49 pages

Percolation and Random Walks On Graphs - Perla Sousi

This document discusses percolation and random walks on graphs. It begins by defining bond percolation and coupling all percolation processes. There is a phase transition parameter pc such that an infinite cluster exists almost surely if p > pc and does not exist if p < pc. For dimensions d ≥ 2, it is shown that pc is between 0 and 1. The document also discusses self-avoiding walks, connective constants, and open conjectures regarding percolation and self-avoiding walks on lattices. It then briefly introduces random walks on graphs and electrical networks.

Uploaded by

skb
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 49

Percolation and Random walks on graphs

Perla Sousi∗

May 14, 2018

Contents

1 Percolation 2
1.1 Definition of the model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Coupling of percolation processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.3 Phase transition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3.1 Self-avoiding walks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.3.2 Existence and uniqueness of the infinite cluster . . . . . . . . . . . . . . . . . 5
1.4 Correlation inequalities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.5 Russo’s formula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.6 Subcritical phase . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
1.7 Supercritical phase in Z2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
1.8 Russo Seymour Welsh theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
1.9 Power law inequalities at the critical point . . . . . . . . . . . . . . . . . . . . . . . . 24
1.10 Grimmett Marstrand theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
1.11 Conformal invariance of crossing probabilities p = pc . . . . . . . . . . . . . . . . . . 28

2 Random walks on graphs 30


2.1 Electrical networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
2.2 Effective resistance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
2.3 Transience vs recurrence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
2.4 Spanning trees . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
2.5 Wilson’s algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
2.6 Uniform spanning trees and forests . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46

University of Cambridge

1
1 Percolation

These notes are largely based on Grimmett’s books Percolation [3] and Probability on Graphs [4].

1.1 Definition of the model

Let G = (V, E) be a graph with set of vertices V and set of edges E and let p ∈ [0, 1]. There are two
models of percolation, bond and site percolation. We will mainly be interested in bond percolation.
To define it consider the probability space Ω = {0, 1}E and endow it with the σ-algebra generated
by the cylinder sets, i.e. sets of the form

{ω(e) = xe , ∀ e ∈ A},

where A is a finite set and xe ∈ {0, 1} for all e. The associated probability measure is going to
be the product measure Pp (·), which means that each edge is 1 (in this case we call it open) with
probability p and 0 (closed) with probability 1 − p and the states of different edges are independent.
We will denote the state of the system by a vector ηp ∈ {0, 1}E .
For this course we will focus on the case where G = Ld , i.e. the d-dimensional Euclidean lattice.
We write x ↔ y if there is an open path of edges from x to y. We write

C(x) = {y : y ↔ x}

to be the cluster of x. We also write x ↔ ∞ if |C(x)| = ∞. By translation invariance, it is clear


that |C(x)| has the same distribution as |C(0)| for all x. We can now ask the following questions:

• Is there p ∈ (0, 1) such that Pp (|C(0)| = ∞) > 0?

• Is the probability Pp (|C(0)| = ∞) monotone in p?

• In the case where Pp (|C(0)| = ∞) > 0, how many infinite clusters are there?

To answer the questions above and more, we will introduce some machinery. The first tool that we
will use a lot to study percolation is a coupling of all percolation processes.

Definition 1.1. Coupling of two probability measures µ and ν, defined on possibly different
probability spaces, is a random variable (X, Y ) defined on a single probability space so that the
marginal distribution of X is µ and the marginal of Y is ν.

We will see the importance of coupling a lot of times during this course. From the definition it’s
not clear why this is such an important notion. If you have taken a Markov chains course before,
you must have seen the use of coupling when you proved the convergence to stationarity.

1.2 Coupling of percolation processes

We are going to couple all percolation processes on Zd for different values of p. To emphasise the
dependence on p we write ηp (e) to denote the state of the edge e when percolation probability is p.

2
Let (U (e))e∈Ld be i.i.d. uniform random variables on [0, 1]. We set
(
1 if U (e) ≤ p
ηp (e) =
0 otherwise.

Then ηp has the law of bond percolation with parameter p.


Using this coupling we now immediately see that if p ≤ q, then ηp (e) ≤ ηq (e), and hence this proves

Lemma 1.2. The probability θ(p) = Pp (|C(0)| = ∞) is increasing as a function of p.

1.3 Phase transition

We now define
pc (d) = sup{p ∈ [0, 1] : θ(p) = 0}

and using the monotonicity of the function θ(p) we get that if p > pc , then θ(p) > 0, while for
p < pc we have θ(p) = 0.
It is known that θ(p) is a C ∞ function on (pc , 1], but it is not known whether θ(p) is continuous
at pc for d = 3.

Conjecture 1.3. The function θ(p) is continuous at pc for d = 3.

We next ask: is pc ∈ (0, 1)? We will answer this in the next theorem.
When d = 1, it is obvious that for all p < 1 we have Pp (|C(0)| = ∞) = 0, since with probability
one we will encounter closed edges. Therefore, for d = 1 we have pc = 1. The picture is different
for higher dimensions.

Theorem 1.4. For all d ≥ 2 we have pc (d) ∈ (0, 1).

Proof that pc (d) > 0. Let Σn denote the number of open self-avoiding paths of length n starting
from 0. We clearly have

Pp (|C(0)| = ∞) = Pp (∀ n : Σn ≥ 1) = lim Pp (Σn ≥ 1) ≤ lim Ep [Σn ] ,


n→∞ n→∞

where for the last inequality we used Markov’s inequality. We now need to upper bound Ep [Σn ].
Let σn denote the number of self-avoiding paths starting from 0 of length n. Then

Ep [Σn ] ≤ σn · pn ,

so it suffices to upper bound σn . A crude upper bound is as follows: for the first step we have 2d
possible choices. For later steps, there are at most 2d − 1 choices. Therefore

2d
Ep [Σn ] ≤ σn · pn ≤ (2d) · (2d − 1)n−1 pn = · ((2d − 1)p)n .
2d − 1
1
If (2d − 1)p < 1, then limn→∞ Ep [Σn ] = 0. Therefore, this shows that pc (d) ≥ 2d−1 .

3
1.3.1 Self-avoiding walks

In the proof above we used the number of self-avoiding walks (paths that do not contain circuits)
and used a very crude bound on their number. Let’s see what else we can say about them.

Lemma 1.5. Let σn be the number of self-avoiding paths of length n. Then for all m, n we have

σn+m ≤ σn · σm .

Proof. The proof of the result above follows from the observation that a self-avoiding path of
length n + m can be written uniquely as a concatenation of a self-avoiding path of length n started
from 0 and a translation of another self-avoiding path of length m.

Corollary 1.6. There is a constant λ so that

log σn
lim = λ.
n→∞ n

Proof. From the previous lemma, the sequence log σn is subadditive, i.e.

log σn+m ≤ log σn + log σm .

Exercise 1 completes the proof.

Remark 1.7. From the above corollary we get that the number of self-avoiding paths of length n
grows like σn = enλ(1+o(1)) . We now define κ = eλ to be the connective constant of the lattice.

There are many open questions in this area. For instance, what is κ for
p the √
Euclidean lattice? It
has been established recently by Duminil-Copin and Smirnov that κ = 2 + 2 for the hexagonal
lattice.

Conjecture 1.8. The number of self-avoiding walks satisfies


 11
n
n 32 κ d=2



nγ κn

d=3
σn ≈


 (log n)1/4 κn d=4

κn

d ≥ 5.

This conjecture was verified for d ≥ 5 in the work of Hara and Slade using lace expansion.
For lower d, in particular d = 2, the bounds are still very far from the truth.

Theorem 1.9 (Hammersley and Welsh). For all d the number of self avoiding walks σn satisfies
√ 
σn ≤ exp cd n κnd ,

where cd is a positive constant.

Tom Hutchcroft proved a small improvement last month:

4
Theorem 1.10 (Hutchcroft). For all d we have

σn ≤ exp(o( n))κn .

The study of the self-avoiding walk is a central topic in probability theory. Another big open
question is to understand the scaling limit of a self-avoiding walk of length n started from the origin
sampled uniformly at random. It is believed that this is characterised by an SLE(8/3) process, but
there is no proof. Last summer, Ewain Gwynne and Jason Miller proved this statement for self-
avoiding walks on random surfaces.

1.3.2 Existence and uniqueness of the infinite cluster

Definition 1.11. Let G be a planar graph, i.e. a graph that can be embedded on the plane in such
a way that no edges cross.
The dual of a planar graph G (called primal) is obtained by placing a vertex in each face of G and
connecting two vertices by an edge if their faces share a boundary edge.
Remark 1.12. The lattice Z2 satisfies the duality property, which means that Z2 is isomorphic
to its dual lattice.

Proof of Theorem 1.4. (pc (d) < 1) First of all we note that it is enough to prove that pc (2) < 1,
since for all d we can embed Zd into Zd+1 and hence, if there exists an infinite cluster in Zd , this
would imply the same statement for Zd+1 .
The reason why we choose to prove for d = 2 is that the 2-dimensional lattice satisfies the duality
property, Remark 1.12.
The following argument is usually referred to as Peierls argument. (Peierls used it to prove phase
transition in the two-dimensional Ising model.) Consider the dual of Z2 and perform bond perco-
lation, by declaring an edge open if the edge that it crosses from Z2 is open and vice versa. It is
clear from a picture, that |C(0)| < ∞ if and only if there exists a closed dual circuit that contains 0.
Let Dn denote the number of closed dual circuits of length n that surround the origin. Then we
obtain

X
Pp (|C(0)| < ∞) = Pp (∃ n : Dn ≥ 1) ≤ Ep [Dn ] .
n=4

Exercise 3 shows that the number of dual circuits of length n that surround 0 is at most n4n . So
we obtain
Ep [Dn ] ≤ n4n (1 − p)n .

5
So by taking p sufficiently close to 1, say p > 1 − δ, we can make Pp (|C(0)| < ∞) < 1 − ε, which
means that pc (d) ≤ 1 − δ.

Lemma 1.13. Let A∞ be the event that there exists an infinite cluster. Then we have the following
dichotomy:
(a) If θ(p) = 0, then Pp (A∞ ) = 0.
(b) If θ(p) > 0, then Pp (A∞ ) = 1.

Proof. (a) Suppose that θ(p) = 0. By the union bound we have


X X
Pp (A∞ ) = Pp (∃ x : |C(x)| = ∞) ≤ Pp (|C(x)| = ∞) = θ(p) = 0.
x∈Zd x∈Zd

(b) If θ(p) > 0, in order to deduce that Pp (A∞ ) = 1, we will use an ergodicity argument.
Recall Kolmogorov’s 0-1 law: let (Xi ) be i.i.d. random variables, Fn = σ(Xk : k ≥ n) and
F∞ = ∩n≥0 Fn . Then F∞ is trivial, i.e. every A ∈ F∞ satisfies P(A) ∈ {0, 1}.
Order the edges of Zd and let ω(e1 ), ω(e2 ), . . . be the states of the ordered edges. Then they
are i.i.d. and A∞ ∈ F∞ , since changing the states of finitely many edges does not change the
occurrence of A∞ . So Pp (A∞ ) ∈ {0, 1}. Since θ(p) > 0 and Pp (A∞ ) ≥ θ(p) > 0, this means that
Pp (A∞ ) = 1.

Once we have established that Pp (A∞ ) satisfies a 0 − 1 law, we ask the following: in the case where
Pp (A∞ ) = 1, how many infinite clusters are there?
Theorem 1.14. Let N be the number of infinite clusters. For all p > pc we have that

Pp (N = 1) = 1.

Remark 1.15. To prove the above theorem, first we would like to apply an ergodicity argument
as in the previous proof. Notice that here we cannot use Kolmogorov’s 0 − 1 law. Indeed, the event
{N = k} is not a tail event, since changing just one edge could change its occurrence. Only for
k = 0 or k = ∞ this is the case. However, N is a translation invariant random variable, and hence
by Exercise 4 it is almost surely a constant.

Theorem 1.14 was first proved by Aizenman, Kesten and Newman. Here we present the argument
by Burton and Keane and we follow the exposition of Grimmett [3] and Nolin [7].

Proof of Theorem 1.14. First by Exercise 4 we have that N is almost surely a constant. Thus
there exists k ∈ {0, 1, . . .} ∪ {∞} such that Pp (N = k) = 1. We know from Lemma 1.13 that k ≥ 1.
Therefore we will exclude the case where k ≥ 2 and k = ∞.
We first prove that if we assume that k < ∞, then k = 1.
Let B(n) = [−n, n]d ∩ Zd be the d-dimensional box of side length 2n + 1 and let ∂B(n) denote its
boundary. Then as n → ∞ we get

Pp (N = k, all k ∞-clusters intersect ∂B(n)) → 1.

6
Figure 1: Example of a trifurcation

So there exists n sufficiently large such that

1
Pp (all ∞-clusters intersect ∂B(n)) ≥ .
2
Now we note that the event above is independent of the states of the edges in the box B(n). So we
obtain

Pp (N = 1) ≥ Pp (all ∞-clusters intersect ∂B(n), all edges of B(n) are open)


1 E(B(n))
= Pp (all ∞-clusters intersect ∂B(n)) Pp (all edges of B(n) are open) ≥ ·p ,
2
where E(B(n)) is the number of edges in the box B(n). We thus showed that with positive
probability N = 1, but since N is almost surely a constant, this implies that under the assumption
that k < ∞ we get that k = 1.
To finish the proof we now have to exclude the possibility k = ∞. Suppose that Pp (N = ∞) = 1.
Let S(n) be the diamond
S(n) = {x ∈ Zd : kxk1 ≤ n},

where kxk1 = di=1 |xi | for all x = (x1 , . . . , xd ). The assumption that N = ∞ almost surely now
P

gives that as n → ∞
 
d
Pp at least three ∞-clusters in Z \ S(n) intersect ∂S(n) → 1.

We note again that the event above is independent of the states of the edges in S(n). We now
take n sufficiently large so that
  1
1 2 3
Pp ∃ C∞ , C∞ , C∞ ⊆ Zd \ S(n) intersecting ∂S(n) ≥ . (1.1)
2

We next introduce the notion of a trifurcation. A vertex x is called a trifurcation if the following
three conditions are satisfied:
(i) x belongs to an ∞ open cluster C∞ ,
(ii) there exist exactly three open edges adjacent to x and
(iii) the set C∞ \ {x} contains exactly three infinite clusters and no finite ones.

7
By translation invariance it is clear that Pp (x is a trifurcation) = Pp (0 is a trifurcation) for all x.
Our next goal now is to prove that

Pp (0 is a trifurcation) ≥ c > 0,

for some constant c. Let x1 , x2 , x3 ∈ ∂S(n) be three distinct points. Then there exist three disjoint
paths (γi )i≤3 in S(n) joining each xi to 0. Check! (This is why we choose to work with S(n) instead
of B(n).) Let J(x1 , x2 , x3 ) be the event that all the edges on these three paths are open and all
other edges of S(n) are closed.
We can now lower bound the probability that 0 is a trifurcation. To this end we have
 
1 2 3
Pp (0 is a trifurcation) ≥ Pp ∃ C∞ , C∞ , C∞ ⊆ Zd \ S(n) intersecting ∂S(n), J(x1 , x2 , x3 ) ,

where {xi } = C∞ i ∩ ∂S(n) for i ≤ 3. Since the points x can be chosen in a deterministic way from
i
the configuration outside S(n) and using the fact that the two events above are independent, we
obtain
 
1 2 3
Pp ∃ C∞ , C∞ , C∞ ⊆ Zd \ S(n) intersecting ∂S(n), J(x1 , x2 , x3 )
 
1 2 3
= Pp J(x1 , x2 , x3 ) ∃ C∞ , C∞ , C∞ ⊆ Zd \ S(n) intersecting ∂S(n)


1 2 3
 1
× Pp ∃ C∞ , C∞ , C∞ ⊆ Zd \ S(n) intersecting ∂S(n) ≥ min(p, (1 − p))E(S(n)) · ,
2
where we used (1.1) and again E(S(n)) stands for the number of edges in S(n), which is finite.
This proves that
Pp (0 is a trifurcation) ≥ c > 0.

We will now arrive at a contradiction. First of all using the above we get
 
X
Ep  1(x is a trifurcation) ≥ c · |B(n)| = c · (2n + 1)d . (1.2)
x∈B(n)

We now claim that almost surely


X
1(x is a trifurcation) ≤ |∂B(n)|. (1.3)
x∈B(n)

Combining this with (1.2) gives the contradiction, since |∂B(n)| grows as nd−1 . So it only remains
to prove (1.3).
Let x1 be a trifurcation. Then by definition there exist three disjoint self avoiding paths from x1
to ∂B(n). Let x2 be another trifurcation. Then there will exist again three disjoint self avoiding
paths to ∂B(n). These paths could intersect the paths of x1 , but they cannot create a cycle,
since this would violate the definition of a trifurcation. Continuing in this way, we explore all
trifurcations, and this gives rise to a forest. The trifurcations are vertices of degree 3 in the forest
and the boundary vertices are at the intersection with the boundary of B(n), see Figure 1.3.2. It

8
Figure 2: Forest of trifurcations in the box and leaves on the boundary

is easy to show that for any finite tree we have

# {vertices of degree 3} ≤ # leaves.

Therefore, this shows that the number of trifurcations is upper bounded by the number of leaves,
which is clearly upper bounded by the size of ∂B(n) and this concludes the proof.

1.4 Correlation inequalities

The goal of this section is to state and prove some correlation inequalities for product measure
(they are actually more general). The tools that we will develop will be used a lot in the proofs of
important results in percolation theory.
Same setup as before, let G = (V, E) be a graph and Ω = {0, 1}E be endowed with the σ-algebra F
generated by the cylinder sets and the probability measure is the product measure Pp (·) with
p ∈ [0, 1].

Definition 1.16. For two configurations ω and ω 0 we say that ω 0 is larger than ω and write ω 0 ≥ ω
if ω 0 (e) ≥ ω(e) for all e ∈ E. This defines a partial order on the set Ω.
A random variable X is called increasing if whenever ω ≤ ω 0 , then X(ω) ≤ X(ω 0 ). It is called
decreasing if −X is increasing. An event A is called increasing if its indicator is increasing.

Example 1.17. For instance the event {|C(0)| = ∞} is increasing.

Using the coupling of all percolation processes we can immediately prove

Theorem 1.18. If N is an increasing random variables and p1 ≤ p2 , then

Ep1 [N ] ≤ Ep2 [N ] .

Similarly if A is an increasing event, then

Pp1 (A) ≤ Pp2 (A) .

9
The first inequality that we will prove says that two increasing events are positively correlated.
In other words, conditioning on an increasing event, increases the probability of the occurrence of
another increasing event. For instance, suppose we condition on {x ↔ y}. Then this increases the
probability of the event {u ↔ v}.

Theorem 1.19 (FKG inequality). Let X and Y be two increasing variables on (Ω, F) such that
Ep X 2 < ∞ and Ep Y 2 < ∞. Then
   

Ep [XY ] ≥ Ep [X] Ep [Y ] .

In particular, if A and B are two increasing events, then

Pp (A ∩ B) ≥ Pp (A) Pp (B) .

Remark 1.20. FKG stands for Fortuin, Kasteleyn and Ginibre. Another way of writing the above
inequality is that for A and B increasing events one has

Pp (A | B) ≥ Pp (A) ,

i.e. knowledge of B increases the probability of A.

Example 1.21. Let G be a graph and for every vertex x we define

pc (x) = sup{p ∈ [0, 1] : Pp (|C(x)| = ∞) = 0}.

Then as a consequence of the FKG inequality we get that pc (x) = pc (y) for all x, y. So this is saying
that for bond (same argument for site) percolation the choice of x is irrelevant in the definition of
the critical probability.

Proof of the FKG inequality. We start by proving the FKG inequality for random variables X
and Y that only depend on a finite set of edges e1 , . . . , en . Then we extend to the general case.
We will prove the first claim by induction on n. First suppose that X and Y depend only on the
state of the edge e1 , which we call ω(e1 ). Using the increasing property of X and Y we obtain that
for all ω1 , ω2 ∈ {0, 1}

(X(ω1 ) − X(ω2 ))(Y (ω1 ) − Y (ω2 )) ≥ 0.

So this immediately gives now


X
0≤ (X(ω1 ) − X(ω2 ))(Y (ω1 ) − Y (ω2 ))Pp (ω(e1 ) = ω1 ) Pp (ω(e1 ) = ω2 )
ω1 ,ω2

= 2(Ep [XY ] − Ep [X] Ep [Y ]).

Suppose now that the claim holds for all n < k. Let X and Y be increasing functions of ω(e1 ), . . . , ω(ek ).
Using properties of conditional expectation we obtain

Ep [XY ] = Ep [Ep [XY | ω(e1 ), . . . , ω(ek−1 )]] .

10
We now note that after conditioning on ω(e1 ), . . . , ω(ek−1 ) the variables X and Y become increasing
functions of ω(ek ) only, and hence

Ep [XY | ω(e1 ), . . . , ω(ek−1 )] ≥ Ep [X | ω(e1 ), . . . , ω(ek−1 )] Ep [Y | ω(e1 ), . . . , ω(ek−1 )]

Applying the induction step to Ep [X | ω(e1 ), . . . , ω(ek−1 )] and Ep [Y | ω(e1 ), . . . , ω(ek−1 )] we get

Ep [Ep [X | ω(e1 ), . . . , ω(ek−1 )] Ep [Y | ω(e1 ), . . . , ω(ek−1 )]] ≥ Ep [X] Ep [Y ] .

Next we prove it for X and Y that depend on countably many edges and have finite second moments.
Let e1 , . . . be an ordering of the edges of G and for each n define

Xn = Ep [X | ω(e1 ), . . . , ω(en )] and Yn = Ep [Y | ω(e1 ), . . . , ω(en )] .

So now Xn and Yn are increasing functions of the states of the edges e1 , . . . , en , and hence we can
apply what we proved earlier to get

Ep [Xn Yn ] ≥ Ep [Xn ] Ep [Yn ] . (1.4)

Since (Xn ) and (Yn ) are both martingales bounded in L2 , by the L2 martingale convergence theorem
we deduce as n → ∞
L2 L2
Xn → X and Yn → Y.

In particular, we have convergence in L1 , and hence taking limits in (1.4) we conclude

Ep [XY ] ≥ Ep [X] Ep [Y ] .

This completes the proof.

Remark 1.22. The FKG inequality gives us a lower bound on the probability of the intersection of
two increasing events A and B. Often one also needs an upper bound on this probability. However,
such a bound cannot always be useful for any increasing events A and B. It turns out that we can
obtain a useful upper bound if instead we look at the event A ◦ B, which means that A and B
occur disjointly, i.e. the set of edges that determines whether or not A occurs is disjoint from the
one corresponding to B.

Before giving the formal definition of disjoint occurrence let us start with an example.

Example 1.23. Perform bond percolation on Ld and let x, y, w, z be 4 distinct points in Zd . Then
we define {x ↔ y} ◦ {w ↔ z} to be the event that there exist two disjoint paths one joining x to y
and the other one joining w to z. Then if we condition on {w ↔ z}, it is intuitive that this should
decrease the probability of {x ↔ y}. This is exactly what the BK inequality says.

Definition 1.24. For every ω ∈ Ω = {0, 1}E and a subset S ⊆ E we write

[ω]S = {ω 0 ∈ Ω : ω 0 (e) = ω(e), ∀ e ∈ S}.

Let A and B be two events depending on a finite set of edges F . Then the disjoint occurrence of A

11
and B is defined via

A ◦ B = ω ∈ Ω : ∃ S ⊆ F with [ω]S ⊆ A and [ω]F \S ⊆ B .

Theorem 1.25 (BK inequality). Let F be a finite set and Ω = {0, 1}F . For all increasing events A
and B on Ω we have
Pp (A ◦ B) ≤ Pp (A) Pp (B) .

Proof. This theorem was proved by van den Berg and Kesten (see [3]). Here we present an easier
proof given by Bollobás and Leader (see for instance [1]).
The proof follows by induction on the size n of the set F . For n = 0, this trivially holds. Suppose
that it works for all sets of size n−1, we will show it also works for size n. For every set D ⊆ {0, 1}F
we define for i = 0, 1
Di = {(ω1 , . . . , ωn−1 ) : (ω1 , . . . , ωn−1 , i) ∈ D}.

By the independence of different coordinates we obtain

Pp (D) = pPp (D1 ) + (1 − p)Pp (D0 ) . (1.5)

Let A, B ⊆ {0, 1}F and C = A ◦ B. Then it is easy to check that

C0 = A0 ◦ B0 and C1 = (A1 ◦ B0 ) ∪ (A0 ◦ B1 ).

Also since A is increasing, we get A0 ⊆ A1 and both A0 and A1 are increasing events. Similarly
for B. Therefore,
C0 ⊆ (A0 ◦ B1 ) ∩ (A1 ◦ B0 ) and C1 ⊆ A1 ◦ B1 .

Using this and the induction hypothesis we get that

Pp (C0 ) = Pp (A0 ◦ B0 ) ≤ Pp (A0 ) Pp (B0 )


Pp (C1 ) ≤ Pp (A1 ◦ B1 ) ≤ Pp (A1 ) Pp (B1 )
Pp (C0 ) + Pp (C1 ) ≤ Pp ((A0 ◦ B1 ) ∩ (A1 ◦ B0 )) + Pp ((A1 ◦ B0 ) ∪ (A0 ◦ B1 ))
= Pp (A0 ◦ B1 ) + Pp (A1 ◦ B0 ) ≤ Pp (A0 ) Pp (B1 ) + Pp (A1 ) Pp (B0 ) .

Multiplying the first inequality by (1 − p)2 , the second p2 and the third one by p(1 − p). Adding
them all up we obtain

(1 − p)Pp (C0 ) + pPp (C1 ) ≤ (pPp (A1 ) + (1 − p)Pp (A0 )) (pPp (B1 ) + (1 − p)Pp (B0 )) .

From (1.5) we finally get

Pp (C) ≤ Pp (A) Pp (B)

and this concludes the proof.

After the BK inequality was proved, it was conjectured that it should hold more generally, not only
for increasing events A and B. This was later proved by Reimer. Despite being more general (it

12
actually includes both FKG and BK inequality) it has not found many applications, i.e. instances
where the other inequalities cannot be used.

Theorem 1.26 (Reimer’s inequality). Let F be a finite set and Ω = {0, 1}F . For all A and B we
have
Pp (A ◦ B) ≤ Pp (A) Pp (B) .

We end this section by presenting an application of the BK inequality. Let χ(p) = Ep [|C(0)|] be
the mean cluster size. Recall Bn = [−n, n]d ∩ Zd .

Theorem 1.27. Suppose that χ(p) < ∞. Then there exists a positive constant c so that for all
n ≥ 1 we have
Pp (0 ↔ ∂Bn ) ≤ e−cn .

This theorem was proved by Hammersley using a different method, but the use of the BK inequality
makes the proof much shorter. In a later section we will prove the same statement under the weaker
assumption that p < pc . The reason we chose to present the proof of the above result is because it
is a very nice application of the BK inequality and also to illustrate a method which is commonly
used.

Proof of Theorem 1.27. Let Xn = x∈Bn 1(0 ↔ x) be the number of vertices on ∂Bn that are
P

connected to 0 via an open path of edges. Then



X ∞ X
X X
Ep [Xn ] = Pp (0 ↔ x) = Pp (0 ↔ x) = χ(p).
n=0 n=0 x∈∂Bn x∈Zd

If there exists an open path of edges from 0 to ∂Bm+k , then there must exist a vertex x on ∂Bm
which is connected to 0 and ∂Bm+k with two disjoint paths. Therefore, using the BK inequality
this gives for all m, k ≥ 1
X
Pp (0 ↔ ∂Bm+k ) ≤ Pp (0 ↔ x) Pp (x ↔ ∂Bm+k )
x∈∂Bm
X
≤ Pp (0 ↔ x) Pp (0 ↔ ∂Bk ) = Ep [Xm ] Pp (0 ↔ ∂Bk ) ,
x∈∂Bm

where in the last step we used translation invariance. Since we assumed that χ(p) < ∞, this means
that Ep [Xn ] → 0 as n → ∞, and hence there exists m large so that

Ep [Xm ] < δ < 1.

Now for any n ∈ N with n ≥ m we can find q, r ≥ 0 with r ∈ [0, m − 1] such that n = qm + r. Then
we finally obtain

Pp (0 ↔ ∂Bn ) ≤ Pp (0 ↔ ∂Bqm ) ≤ δ q ≤ δ −1+n/m = e−cn

for a suitable constant c > 0. By adjusting the value of c if necessary to take care of values of
n < m, this concludes the proof.

13
1.5 Russo’s formula

We saw in Lemma 1.2 that using the coupling of all percolation processes, the probability θ(p) is an
increasing function of p. But, do we know the rate at which it changes when we increase p? This
is the goal of the current section, to understand the derivative of Pp (A) for all increasing events A
as a function of p.
Before stating Russo’s theorem we define what we mean by a pivotal edge.
Definition 1.28. Let A be an event and ω a percolation configuration. An edge e is called pivotal
for (A, ω) if 1(ω ∈ A) 6= 1(ω 0 ∈ A), where ω 0 (f ) = ω(f ) for all edges f 6= e and ω 0 (e) 6= ω(e). In
other words, e is pivotal for A if the occurrence of A depends crucially on whether the edge e is
present or not.
The event {e is pivotal for A} is equal to {ω : e is pivotal for (A, ω)}.
Example 1.29. Let A be the event that 0 is in an infinite cluster. Then an edge e is pivotal for A
if the removal of e leads to a finite component containing the origin and one endpoint of e and an
infinite component containing the other endpoint of e.
Theorem 1.30 (Russo’s formula). Let A be an increasing event that depends only on the states of
a finite number of edges. Then
d
Pp (A) = Ep [N (A)] ,
dp
where N (A) is the number of pivotal edges for A.

Proof. First of all we argue that the function in question is indeed differentiable. Suppose that A
depends on the states of m edges, e1 , . . . , em . We define the function

f (p1 , . . . , pm ) = Pp (A) ,

where p = (p1 , . . . , pm ) and pi is the probability that the i-th edge is open. Then clearly f is a
polynomial in pi ’s, and hence it is differentiable.
We now turn to establish the formula for the derivative. Recall the coupling of all percolation
processes. Let (X(e))e∈Ld be i.i.d. uniform random variables in [0, 1]. For a vector of probabilities
p = (p(e) : e ∈ Ld ) we write ηp (e) = 1 if X(e) ≤ p(e) and ηp (e) = 0 otherwise. We write Pp (·)
for the probability measure where edge e is open with probability p(e) and closed with 1 − p(e)
independently over different edges. So we have

Pp (A) = P(ηp ∈ A) .

Fix an edge f and let p0 be such that p0 (e) = p(e) for all e 6= f and p0 (f ) = p(f ) + δ with δ > 0.
Then we have
 
Pp0 (A) − Pp (A) = P ηp0 ∈ A − P(ηp ∈ A) = P ηp ∈
/ A, ηp0 ∈ A ,

where the last equality follows from the increasing property of A. We now claim that

P ηp ∈
/ A, ηp0 ∈ A = δ · Pp (f is pivotal for A) . (1.6)

14
Since p0 agrees with p everywhere except for the edge f , on the event {ηp ∈ / A, ηp0 ∈ A} we have
ηp (f ) = 0 while ηp0 (f ) = 1. Hence this means that f is pivotal for the event A. We also notice that
the event that f is pivotal for A is independent of the state of the edge f . Therefore we get

P ηp ∈
/ A, ηp0 ∈ A = Pp (f is pivotal for A, p < X(f ) ≤ p + δ)
= P(p < X(f ) ≤ p + δ) Pp (f is pivotal for A) = δ · Pp (f is pivotal for A) ,

which is exactly (1.6). So we can now conclude

∂ Pp0 (A) − Pp (A)


Pp (A) = lim = Pp (f is pivotal for A) .
∂p(f ) δ↓0 δ

Up to here we have not used the assumption that A depends only on finitely many edges. We only
used the fact that A is increasing. Since A depends on finitely many edges, say e1 , . . . , em , then as
we said above Pp (A) is a polynomial of p(e1 ), . . . , p(em ), and hence by the chain rule we obtain
m m
d X ∂ X
Pp (A) = Pp (A) |p=(p,...,p) = Pp (ei is pivotal for A) = Ep [N (A)]
dp ∂p(ei )
i=1 i=1

and this completes the proof.

Remark 1.31. We note that if A is an increasing event depending on an infinite number of edges,
then
Pp+δ (A) − Pp (A)
lim inf ≥ Ep [N (A)] .
δ↓0 δ
Indeed, let Bn = [−n, n]d ∩ Zd and define pn via
(
p if e ∈
/ Bn
pn (e) =
p+δ if e ∈ Bn .

Since A is an increasing event we now get

Pp+δ (A) − Pp (A) Pp (A) − Pp (A)


≥ n .
δ δ
If q and p only differ at the value of only one edge e by δ and q ≤ p, then by (1.6) we obtain

Pp (A) − Pq (A) = δ · Pq (e is pivotal for A) . (1.7)

Also for any event B if p and q differ in a finite number m of edges, then

|Pq (B) − Pp (B) | = |P(ηq ∈ B) − P(ηp ∈ B)|


(1.8)
≤ P(∃ edge e : ηp (e) 6= ηq (e)) ≤ m · P(X(e) ∈ [p(e) ∧ q(e), p(e) ∨ q(e)]) .

So by taking successive differences and using (1.7) and (1.8) this gives

Ppn (A) − Pp (A) X


lim inf ≥ Pp (e is pivotal for A) .
δ↓0 δ
e∈Bn

15
Taking the limit as n → ∞ completes the proof.
Corollary 1.32. Let A be an increasing event depending on the states of m edges and p ≤ q be
in (0, 1]. Then  m
q
Pq (A) ≤ Pp (A) .
p

Proof. As already mentioned in the proof of Russo’s formula, the event that the edge e is a pivotal
edge is independent of its state, i.e.

Pp (e is open and pivotal for A) = p · Pp (e is pivotal for A) .

Moreover, since A is increasing, we have that

{ω : ω(e) = 1 and 1(ω ∈ A) 6= 1(ω 0 ∈ A)} = {ω : e is pivotal for (A, ω) and A occurs},

where ω 0 agrees with ω everywhere except for the edge e. Using these two equalities above we thus
get from Russo’s formula

d 1 X
Pp (A) = · Pp (e is open and pivotal for A)
dp p e
1 X 1 X
= · Pp (e is pivotal for A, A) = · Pp (e e is pivotal for A | A) Pp (A)
p e p e
1
= · Ep [N (A) | A] Pp (A) .
p

Dividing through by Pp (A) and integrating from p to q we obtain


Z q 
1
Pq (A) = Pp (A) exp · Es [N (A) | A] ds .
p s

Using the trivial bound


X
Ep [N (A) | A] = Pp (e is pivotal for A) ≤ m
e

proves the corollary.

1.6 Subcritical phase

In this section we focus on p < pc . In this case we know that there is no infinite cluster almost surely.
However, one can ask what is the size of the cluster of 0. How do probabilities like Pp (|C(0)| ≥ n)
decay in n?
We write Bn = [−n, n]d ∩ Zd .
Theorem 1.33. Let d ≥ 2. Then the following are true
(a) If p < pc , then there exists a constant c so that for all n ≥ 1 we have

Pp (0 ↔ ∂Bn ) ≤ e−cn .

16
(b) If p > pc , then
p − pc
θ(p) = Pp (0 ↔ ∞) ≥ .
p(1 − pc )
This theorem was first proved by Aizenman and Barsky in a more general framework of long range
percolation. A different proof was discovered by Menshikov at the same time using the geometry of
the pivotal edges. Here we will present a new proof of Duminil-Copin and Tassion from last year,
which is much shorter than the previous ones. We follow closely [2].

Proof. The main idea of the proof is to introduce a different critical probability that is more
amenable to calculations. First, for all finite sets S containing 0 we define
 
S
X
ϕp (S) = p Pp 0 ←→ x ,
(x,y)∈∂S

S
where we write x ←→ y if x is connected to y via an open path of edges all lying in the set S and
∂S = {(x, y) ∈ E : x ∈ S and y ∈
/ S}. We now define

pec = sup{p ∈ [0, 1] : ∃ a finite set S 3 0 with ϕp (S) < 1}.

Since for any finite set S the function ϕp (S) is continuous in p, it follows that the set of p’s for
which there exists a finite set S 3 0 with ϕp (S) < 1 is an open subset of [0, 1]. Therefore, pec > 0.
We now notice that it suffices to prove the statement of the theorem with pc replaced by pec . Indeed,
from part (a) we would get that pec ≤ pc , since for all p < pec by letting n → ∞ would give θ(p) = 0.
Therefore, pec < 1, and hence from part (b) we would get that pc ≤ pec .
So we now turn to prove (a) and (b) with pc replaced by pec .
(a) Let p < pec . Then by the definition there exists a finite set S with 0 ∈ S and such that ϕp (S) < 1.
Take L large so that S ⊆ BL−1 . Let k ≥ 1 and assume that the event {0 ←→ ∂BkL } holds. Let

S
C = {x ∈ S : 0 ←→ x}.

Since S ∩ ∂BLk = ∅, on the event {0 ←→ ∂BkL } there must exist an edge (x, y) ∈ ∂S such that
S Cc
0 ←→ x, the edge (x, y) is open and y ←→ ∂BkL . So we obtain

Ac
 
S
X X
Pp (0 ←→ ∂BkL ) ≤ Pp 0 ←→ x, (x, y) is open, C = A, y ←→ ∂BkL .
(x,y)∈∂S A⊆S

S Ac
We now notice that the events {0 ←→ x, C = A}, {(x, y) is open} and {y ←→ ∂BkL } appearing
above are independent, since they depend on disjoint sets of edges. (This is a trivial application of
the BK inequality.) Therefore we deduce

Ac
   
S
X X
Pp (0 ←→ ∂BkL ) ≤ p Pp 0 ←→ x, C = A Pp y ←→ ∂BkL .
(x,y)∈∂S A⊆S

Since y ∈ BL , it follows that


Ac
  
Pp y ←→ ∂BkL ≤ Pp 0 ←→ ∂B(k−1)L ,

17
and hence we conclude
 
S
X  
Pp (0 ←→ ∂BkL ) ≤ p Pp 0 ←→ x Pp 0 ←→ ∂B(k−1)L = ϕp (S)Pp 0 ←→ ∂B(k−1)L .
(x,y)∈∂S

Iterating the above inequality gives

Pp (0 ←→ ∂BkL ) ≤ (ϕp (S))k−1 ,

and hence this establishes the exponential decay.


(b) To prove the second part it suffices to prove the following differential inequality for all p ∈ (0, 1)
and all n ≥ 1

d 1
Pp (0 ←→ ∂Bn ) ≥ · inf ϕp (S) · (1 − Pp (0 ←→ ∂Bn )). (1.9)
dp p(1 − p) S⊆Bn
0∈S

Indeed, once we show this inequality, then integrating from pec to p > pec and then taking the limit
as n → ∞ proves the desired inequality. So we turn to prove (1.9). By Russo’s formula since
{0 ←→ ∂Bn } is increasing and only depends on a finite set of edges, we get

d X
Pp (0 ←→ ∂Bn ) = Pp (e is pivotal for {0 ←→ ∂Bn })
dp
e∈Bn
1 X
= Pp (e is pivotal, 0 ←→
6 ∂Bn ) ,
1−p
e∈Bn

since the events {e is pivotal} and {e is closed} are independent and also

{e is pivotal and closed} = {e is pivotal, 0 ←→


6 ∂Bn }.

We next define the set


S = {x ∈ Bn : x ←→
6 ∂Bn }.

On the event {0 ←→
6 ∂Bn }, the set S is a subset of Bn containing 0. So we get

d 1 X X
Pp (0 ←→ ∂Bn ) = Pp (e is pivotal, S = A) .
dp 1−p
e∈Bn A⊆Bn
0∈A

We now observe that on the event {S = A}, the pivotal edges for the event {0 ↔ ∂Bn } are the
A
edges (x, y) ∈ ∂A such that 0 ←→ x. Therefore we obtain

d 1 X X 
A

Pp (0 ←→ ∂Bn ) = Pp 0 ←→ x, S = A .
dp 1−p
A⊆Bn (x,y)∈∂A
0∈A

Now the two events appearing above are independent, since in order to determine whether S = A,

18
we only need to look at what happens outside of A. Hence this now gives

d 1 X X 
A

Pp (0 ←→ ∂Bn ) = Pp 0 ←→ x Pp (S = A)
dp 1−p
A⊆Bn (x,y)∈∂A
0∈A
1 X 1
= ϕp (A)Pp (S = A) ≥ inf ϕp (S) Pp (0 ←→
6 ∂Bn )
p(1 − p) p(1 − p) S⊆Bn
A⊆Bn 0∈S
0∈A

and this concludes the proof.

Remark 1.34. Note that the assumption of Theorem 1.33 is that p < pc and not that θ(p) = 0.
We will see later that for d = 2, at the critical probability 1/2 we do not have exponential decay of
this probability.

Remark 1.35. Note that the probability Pp (0 ↔ ∂Bn ) is at least pn , and hence we cannot hope
for a faster convergence than exponential as in Theorem 1.33. As a consequence of Theorem 1.33
we get that when p < pc , then
χ(p) = Ep [|C(0)|] < ∞.

Theorem 1.33 also gives us an upper bound on the tail probabilities of the distribution of |C(0)|.
In particular, we get for p < pc
 
Pp (|C(0)| ≥ n) ≤ Pp (0 ↔ ∂Bn1/d ) ≤ exp −cn1/d .

However, this is not the best possible result, as the term n1/d in the exponential can be replaced
by n (see for instance [3, Theorem 6.75]). So if we restrict bond percolation on a large box Bn , i.e.
by taking n large enough, then the largest connected cluster will have size of order log n.

1.7 Supercritical phase in Z2

Conjecture 1.36. For all d ≥ 2 we have θ(pc (d)) = 0.

This has only been proven for d = 2 and d ≥ 11. We now present the proof for d = 2.

Theorem 1.37. For bond percolation on Z2 the critical probability pc = 1/2 and θ(1/2) = 0.

Proof. Different parts of this theorem were proved by Harris, Kesten, Russo, Seymour and Welsh.
Here we follow the proof given in Grimmett [3].
We will first prove that θ(1/2) = 0, which will then imply that pc ≥ 1/2. Next we will show that
pc ≤ 1/2. Both parts rely crucially on duality, which is only present in two dimensions.
Suppose that θ(1/2) > 0. Let Sn = [0, n]2 . Then there exists n sufficiently large so that
 4
1
P(∂Sn ↔ ∞) > 1 − . (1.10)
8

19
Let A` , At , Ar , Ab be the events that there exists an open path joining the left, respectively top,
right and bottom, side of the box Sn to ∞. By symmetry they all have the same probability. Also
clearly they are increasing events, and hence by the FKG inequality we obtain
4 4
P1/2 (∂Sn ←→
6 ∞) = P(Ac` ∩ Act ∩ Acr ∩ Acb ) ≥ P1/2 (Ac` ) = 1 − P1/2 (A` ) .

Therefore, taking n large enough so that (1.10) holds we get that


 4
4 1
1 − P1/2 (A` ) ≤ ,
8

which of course is equivalent to saying

7
P1/2 (A` ) ≥ .
8

Recall the definition of the dual lattice. Consider the dual box Tn = [0, n − 1]2 + (1/2, 1/2) and let
D` , Dt , Dr , Db be the events that the left (resp, top, right, bottom) side of Tn is connected to ∞
via a closed path of edges of the dual lattice.
Let G = A` ∩ Ar ∩ Dt ∩ Db . Then by symmetry P1/2 (Di ) = P1/2 (Ai ) for all i ∈ {`, r, t, b}, and hence

1
P1/2 (Gc ) ≤ P(Ac` ) + P(Acr ) + P(Dtc ) + P(Dbc ) ≤ ,
2
or equivalently P1/2 (G) ≥ 1/2. Using the uniqueness of the infinite open cluster above pc implies
that there must exist a connection between the two infinite open clusters. However, this in turn
implies that there exist two infinite closed dual paths, since they have to be separated by the open
path of the primal. But this has probability 0, by the uniqueness theorem again. We thus reached
a contradiction, so θ(1/2) = 0.
We now turn to prove that pc ≤ 1/2. Suppose for contradiction that pc > 1/2. Then by the
exponential decay in the subcritical regime, Theorem 1.33, there exists a positive constant c such
that for all n ≥ 1

P1/2 (0 ↔ ∂Bn ) ≤ e−cn . (1.11)

Let An be the event that the left side of the box Cn = [0, n + 1] × [0, n] is joined to the right side
along an open path.
Consider next the dual box Dn = {x + (1/2, 1/2) : 0 ≤ x1 ≤ n, −1 ≤ x2 ≤ n} and let Bn be the
event that there exists a closed path of dual edges joining a vertex on the top of the box to the
bottom. It is not hard to see that An ∩ Bn = ∅ and also that An and Bn partition the whole space.
Indeed, if both An and Bn occur, then there must exist an open edge which is crossed by a closed
edge of the dual, which is impossible. Also suppose that An does not occur, then there must exist
a closed path from top to bottom blocking all connections of the left side to the right side of Cn .
Therefore,
Pp (An ) + Pp (Bn ) = 1

20
and since the dual lattice is isomorphic to Z2 we also get Pp (An ) = P1−p (Bn ). Therefore,

P1/2 (An ) = 1/2. (1.12)

However, using the exponential decay below criticality, we obtain for a positive constant c

P1/2 (0 ↔ {(n, k) : k ∈ Z}) ≤ e−cn .

So letting Ln = {(n, k) : k ∈ Z}, we deduce


n
X
P1/2 (An ) ≤ P1/2 ((0, k) ↔ Ln ) ≤ (n + 1)e−cn ,
k=0

but this gives a contradiction to the fact that P1/2 (An ) = 1/2 and thus completes the proof.

1.8 Russo Seymour Welsh theorem

We saw in the proof of pc (2) = 1/2 the use of open crossings of boxes. We will now see the most
fundamental result of RSW which deals with crossings of rectangles.
Let us first fix some notation. First of all we write

B(k`, `) = [−`, (2k − 1)`] × [−`, `]

and B(`, `) = B(`) = [−`, `]2 . A left-right crossing of a rectangle B is an open path joining the left
side of the rectangle to the right side, but it is not allowed to use any of the boundary edges. (This
is simply for a technical reason.) We denote by LR(k`, `) the event that there exists a left to right
crossing of B(k`, `) and we write LR(`) for a crossing of B(`). We also denote by

A(`) = B(3`) \ B(`)

and we write O(`) for the event that there is an open circuit in A(`) containing 0 in its interior.

Theorem 1.38 (RSW). Suppose that Pp (LR(`)) = α. Then


 √ 4 12
Pp (O(`)) ≥ α 1 − 1 − α .

Remark 1.39. Note that the importance of the above theorem is not on the particular lower
bound that it gives, but rather that if we have a positive probability of a left to right crossing of
the box, then there is a positive probability of having an open circuit in the annulus. We will see
some applications of this to inequalities for θ(p) in the next section.

The proof of RSW is split into two unequal lemmas.

Lemma 1.40. Suppose that Pp (LR(`)) = α. Then


  
3 3
Pp LR `, ` ≥ 1− 1−α .
2

21
Figure 3: Crossings intersecting

Figure 4: Open circuit in the annulus

Lemma 1.41. We have that


   2
3
Pp (LR(2`, `)) ≥ Pp (LR(`)) Pp LR `, ` (1.13)
2
Pp (LR(3`, `)) ≥ Pp (LR(`)) (Pp (LR (2`, `)))2 (1.14)
Pp (O(`)) ≥ (Pp (LR(3`, `)))4 (1.15)

Proof of Theorem 1.38. The proof of RSW follows immediately by combining the two lemmas
above.

Proof of Lemma 1.41. This is the argument of Russo. We start by proving (1.13).
Let LR1 be the event that there exists a left to right crossing of the rectangle [0, 3`] × [−`, `], LR2
be the event that there exists a left to right crossing of [`, 4`] × [−`, `] and TB1 the event that there
exists a top to bottom crossing of [`, 3`] × [−`, `]. If all these events occur, then it is clear that there
exists a left to right crossing of the rectangle [0, 4`] × [−`, `] (see Figure 1.8), and hence we obtain

Pp (LR(2`, `)) ≥ Pp (LR1 ∩ LR2 ∩ TB1 ) ≥ Pp (LR1 ) Pp (LR2 ) Pp (TB1 ) ,

where the last inequality follows from the FKG inequality. So this proves (1.13). Inequality (1.14)
follows similarly.
For the last inequality we express the event A(`) as the union of four rectangles and we ask for left
to right crossings of each of those (see Figure 1.8). Then by the FKG inequality again we get the
result.

22
Sketch of proof of Lemma 1.40. First of all one can define a partial order on left to right
crossings of the box B(`). We say π1 ≤ π2 if π1 is contained in the closed bounded region of B(`)
below π2 . It can be shown that if a set of crossings is non-empty, then a lowest left to right crossing
exists and we call it Π. We call A the set of left to right crossings of B(`).
For π ∈ A we denote by (0, yπ ) the last vertex where the path and the vertical axis intersect. For
π ∈ A we write πr for the part of π that connects (0, yπ ) to the right side of B(`). We also write πr0
for the reflection of πr along the line {(`, k) : k ∈ Z}. We write A− for the set of paths π with
yπ ≤ 0 and A+ for those with yπ ≥ 0.
For a path π ∈ A we write Vπ for the event that all the edges of π are open. We also write Mπ for
the event that there exists an open path from the top of B(`)0 = [0, 2`] × [−`, `] to πr ∪ πr0 and Mπ−
for the event that this path ends on πr .
Let L− be the event that there exists an open path in A− , resp. for L+ . Let Lπ be the event that
π is the lowest left to right open crossing of B(`).
Finally we write N for the event that there exists a left to right crossing of B(`)0 and N − for the
event that this path starts from the negative part of the vertical axis.
It is clear from the picture that the event ∪π∈A− Vπ ∩ Mπ− ∩ N + implies the existence of an open
left to right crossing of B(3/2`, `). Hence we get

Pp (LR(3/2`, `)) ≥ Pp N + ∩ ∪π∈A− (Vπ ∩ Mπ− ) = Pp N + ∩ G .


 

We now observe that both events, N + and G, are increasing, and hence we can use the FKG
inequality to obtain
Pp (LR(3/2`, `)) ≥ Pp N + Pp (G) .


By the square root trick we get


q √
Pp N + ≥ 1 − 1 − Pp (N + ∪ N − ) = 1 − 1 − α,


since Pp (N + ∪ N − ) = Pp (LR(`)) = α by assumption.


It remains to lower bound the probability of G. For this we have
X
Pp (G) ≥ Pp ∪π∈A− (Lπ ∩ Mπ− ) = Pp Mπ− Lπ Pp (Lπ ) ,
 

π∈A−

since the events appearing in the union are disjoint. We now claim that for all π ∈ A we have

Pp Mπ− Lπ ≥ 1 − 1 − α.

(1.16)

Once we establish this, the rest of the proof follows easily. Indeed, substituting this into the bound
for Pp (G) we get
X √
Pp (G) ≥ Pp (Lπ ) (1 − 1 − α).
π∈A−

23
Pp (Lπ ) = Pp (L− ) and by the square root trick again we deduce
P
But π∈A−

+
 q √
Pp L ≥ 1 − 1 − Pp (L+ ∪ L− ) = 1 − 1 − α.

So it remains to prove (1.16).


The idea is to say that the event that π is the lowest left to right crossing is not giving us more
information than saying that all the edges of π are open. Indeed, what happens below π does not
affect the occurrence of the event Mπ . Therefore

Pp Mπ− Lπ “ = ”Pp Mπ− Vπ .


 

And now we can use the FKG inequality again since all events are increasing to get
q q √
Pp Mπ− Vπ ≥ Pp Mπ− ≥ 1 − 1 − Pp Mπ+ ∪ Mπ− = 1 − 1 − Pp (Mπ ) ≥ 1 − 1 − α,
  

since whenever there is a top to bottom crossing of B(`)0 the event Mπ happens. This now concludes
the proof.

1.9 Power law inequalities at the critical point

Theorem 1.42. There exist finite positive constants α1 , A1 , α2 , A2 , α3 , A3 , α4 , A4 so that for all
n ≥ 1 we have
1
√ ≤P1/2 (0 ↔ ∂Bn ) ≤ A1 · n−α1 (1.17)
2 n
1
√ ≤P1/2 (|C(0)| ≥ n) ≤ A2 · n−α2 (1.18)
2 n
E[|C(0)|α3 ] < ∞. (1.19)

Moreover, for all p > 1/2 we have

1 α4
 
θ(p) ≤ A4 · p − (1.20)
2

Remark 1.43. Recall that in Theorem 1.33, part (b) we proved that when p > 1/2, then
 
1
θ(p) ≥ c p − ,
2

where c is a positive constant.

The following easy application of RSW will be a crucial ingredient of the proof.

Lemma 1.44. Let O(`) be the event that there exists an open circuit in the annulus A(`) =
B(3`) \ B(`) containing 0 in its interior. Then there exists a positive constant ζ such that for
all ` ≥ 1 we have
P1/2 (O(`)) ≥ ζ.

24
Proof. Let An be the event that the left side of [0, n + 1] × [0, n] is joined to the right side along
an open path.
Then in (1.12) in the proof of Theorem 1.37 we showed using duality that

1
P1/2 (An ) = .
2
It is obvious that this left to right crossing also contains a left to right crossing of the box [0, n]×[0, n].
Therefore, we showed that P1/2 (LR(`)) ≥ 1/2 for all ` ≥ 1.
Hence, from Theorem 1.38 (RSW) we obtain that for all ` ≥ 1

1 √ 48
 
−24
P1/2 (O(`)) ≥ 2 1− 3
2

and this completes the proof.

Proof of Theorem 1.42. The left hand side of (1.17) is left as an exercise. We turn now to the
right hand side.
We work on the dual lattice again and write B(k)d = B(k) + (1/2, 1/2) for the dual box, A(`)d
for the dual annulus A(`)d = B(3`)d \ B(`)d . Let O(`)d be the event that there exists a closed
dual circuit in A(`)d containing the dual origin (1/2, 1/2) in its interior. Duality and Lemma 1.44
immediately give
P1/2 (O(`)d ) ≥ ζ, ∀ ` ≥ 1.

For all k ≥ 0 we now get


 
P1/2 0 ↔ ∂B(3k + 1) ≤ P1/2 (O(3r )d does not occur ∀ r < k) .

But notice that the annuli A(3r )d for different values of r are disjoint, and hence the events ap-
pearing on the right hand side above are independent. Therefore, we obtain
 
P1/2 0 ↔ ∂B(3k + 1) ≤ (1 − ζ)k .

For every n ≥ 1 we can find k such that 3k + 1 ≤ n < 3k+1 + 1. We thus deduce
 
P1/2 (0 ↔ ∂B(n)) ≤ P1/2 0 ↔ ∂B(3k + 1) ≤ (1 − ζ)k ≤ A1 · n−α1 ,

where A1 and α1 are positive constants. This proves (1.17).


The second inequality follows directly from (1.17). For (1.19) we have

X
γ
E[|C0 | ] = P1/2 (|C0 |γ ≥ n) < ∞
n=0

by choosing a suitable γ as a function of α2 .


We now prove (1.20). First of all we note that for all n ≥ 1

θ(p) ≤ Pp (0 ↔ ∂Bn ) .

25
Take now p > 1/2. Since there are at most 18n2 edges in the box Bn , from Corollary 1.32 we get
 18n2
1
P1/2 (0 ↔ ∂Bn ) ≥ · Pp (0 ↔ ∂Bn ) .
2p

This now gives us


2 2
θ(p) ≤ (2p)18n · P1/2 (0 ↔ ∂Bn ) ≤ (2p)18n · A1 · n−α1 ,

where for the last inequality we used (1.17). Taking now n = [(log(2p))−1/2 ] and noting that
n ∼ (2p − 1)−1/2 as p ↓ 1/2 yields

1 α4
 
θ(p) ≤ A1 · e18 · n−α1 ≤ A4 · p −
2

and this concludes the proof.

1.10 Grimmett Marstrand theorem

For any connected graph G we write pc (G) for the critical probability of bond percolation on G.
We now define for each k ≥ 1 the slab of thickness k by

Sk = Z2 × {0, 1, 2, . . . , k}d−2 .

We write pc (Sk ) for the critical probability of Sk . Since Sk ⊆ Sk+1 ⊆ Zd , we immediately get

pc (Sk ) ≥ pc (Sk+1 ) ≥ pc (Zd ).

Therefore, the limit limk→∞ pc (Sk ) exists and we call it pslab


c which satisfies pslab
c ≥ pc .
It was a big open question for many years to prove that pslab
c = pc .
This was finally proved by Grimmett and Marstrand and it is one of the most important theorems
in the study of the supercritical regime for d ≥ 3. For d = 2, the supercritical regime corresponds
to subcritical for the dual process, but in higher dimensions the picture is different.
A lot of theorems can be proved under the assumption that p > pslab
c .

Theorem 1.45 (Grimmett and Marstrand (1990)).


(a) Let d ≥ 2 and let F be an infinite connected subset of Zd with pc (F ) < 1. For each η > 0 there
exists an integer k such that
pc (2kF + Bk ) ≤ pc + η.

(b) If d ≥ 3, then pslab


c = pc .

Proof that (a)⇒ (b) Suppose that d ≥ 3. Taking F = Z2 × {0}d−2 we have that

2kF + Bk = Z2 × ([−k, k]d−2 ∩ Zd−2 ),

26
which is a translate of S2k . Therefore

pc (S2k ) = pc (2kF + Bk ) → pc as k → ∞,

which then gives that pc = pslab


c .

An application of Theorem 1.45 is the following theorem on the decay of the probability that 0 is
connected to distance n and belongs to a finite component when p > pc . For d = 2 this can be
proved using duality, but for d ≥ 3 it can be established for all p > pslab
c .

Theorem 1.46. Let d ≥ 3 and p > pc . There exists a positive constant c so that for all n ≥ 1 we
have
Pp (0 ↔ ∂Bn , |C(0)| < ∞) ≤ e−cn .

27
1.11 Conformal invariance of crossing probabilities p = pc

We follow [4].
In this section we consider site percolation on the triangular lattice T. Each vertex is painted black
or white with probability p and 1 − p respectively independently over different vertices.
It may be proved similarly to Z2 that pc (T) = 1/2.
Let us introduce some notation and definitions.
Let D ⊆ C be an open simply connected domain in R2 . We assume that its boundary ∂D is a
Jordan curve. Let a, b, c be distinct points of ∂D taken in anticlockwise order around ∂D. The
Riemann mapping theorem guarantees the existence of a conformal map ϕ from D to the interior
of the equilateral triangle T of C with vertices A = 0, B = 1 and C = eπi/3 . Such a ϕ can be
extended to the boundary ∂D in such a way that it becomes a homeomorphism from D ∪ ∂D to the
closed triangle T . There exists a unique conformal map ϕ that maps a, b, c to A, B, C respectively.
If x ∈ ∂D lying on the arc bc, then under the conformal map X = ϕ(x) lies on the arc BC of T .
We now rescale the triangular lattice T to have mesh size δ and we are interested in the probability
Pδ (ac ↔ bx in D) of having an open path joining ac to bx in the domain D. The RSW estimates
also hold for the triangular lattice, and hence the probability above can be shown to be uniformly
bounded away from 0 and 1 as δ → 0. So it seems reasonable that this probability should converge
as δ → 0. Cardy’s formula tells us the values of this limit.
Theorem 1.47 (Cardy’s formula). We have

Pδ (ac ↔ bx in D) → |BX| as δ → 0.

Cardy stated the limit of the probability above in terms of a hypergeometric function of a certain
cross-ratio. To derive it he used arguments from conformal field theory. Lennart Carlesson recog-
nised the hypergeometric function in terms of the conformal map from a rectangle to a triangle
and conjectured the simple form of the theorem above. The theorem was then proved by Smirnov
in 2001.
Theorem 1.48 (Smirnov, 2001). Cardy’s formula holds for the triangular lattice.
If two domains (Ω, a, b, c, d) and (Ω0 , a0 , b0 , c0 , d0 ) are conformally equivalent (i.e. there exists a
conformal map from one to the other one), then

P(ac ↔ bd in Ω) = P a0 c0 ↔ b0 d0 in Ω0 .


We now introduce the exploration process – dynamical way of discovering the configuration in a
discrete domain Ω, i.e. made of small hexagons. We use the following convenient way of representing
site percolation on T: we colour each face of the dual hexagonal lattice H in the same colour as the
corresponding site in T. Hence we can think of site percolation on T as a random colouring of the
faces of H.
Colour the positive half line black and the negative one white. Then one can define the exploration
process by always keeping a black hexagon on the right and a white one on the left. If we rescale
the mesh size of the hexagonal lattice, what is the scaling limit of this exploration curve?

28
It turns out that the limit is an SLE(6) curve. The proof of this uses the locality property of
percolation: what you have not explored so far is independent of what has been explored. The
same property holds for SLE(6). In particular both processes have the same hitting probabilities.
For the exploration process, one can get these hitting probabilities using Cardy’s formula.
Let us take a continuous curve γt in the upper half plane H. Let Kt be such that H \ Kt is the
unbounded component of H \ γ[0, t]. Then there is a conformal map gt : H \ Kt → H satisfying

d 2
gt (z) = ,
dt gt (z) − a(t)

where a(t) is a real-valued function of t and it is exactly the point where the tip of the curve is
mapped via gt , i.e. gt (γt ) = a(t). Also we have g0 (z) = z.
In order for this curve to describe the scaling limit of many discrete structures, the function a must
have the following properties:
(1) a must be continuous
(2) it must have i.i.d. increments
(3) a must be symmetric, i.e. a has the same law as −a.
There is only one type of function with all these properties, namely any multiple of a standard

Brownian motion. So taking a(t) = κBt gives us the SLE(κ) curve.
More on SLE next term!

29
2 Random walks on graphs

In this section we are following closely [5, Chapter 9].


Let G = (V, E) be a finite connected graph with set of vertices V and set of edges E. We endow it
with non-negative numbers (c(e))e∈E that we call conductances. We write c(x, y) = c({x, y}) and
clearly c(x, y) = c(y, x). The reciprocal r(e) = 1/c(e) is called the resistance of the edge e.
We now consider the Markov chain on the nodes of G with transition matrix
c(x, y)
P (x, y) = ,
c(x)
P
where c(x) = y:y∼x c(x, y). This process is called the weighted random walk on G with edge
weights (c(e))e .
P
This process is reversible with respect to π(x) = c(x)/cG , where cG = x∈V c(x), since

c(x) c(x, y) c(x, y)


π(x)P (x, y) = · = = π(y)P (y, x)
cG c(x) cG

and π is stationary for P , i.e. πP = π.


When c(e) = 1 for all edges e, we call the Markov chain with transition matrix P a simple random
walk on G. In this case (
1
if y ∼ x,
P (x, y) = d(x)
0 otherwise.
It is not hard to show that every reversible Markov chain is a weighted random walk on a graph.
Indeed, suppose that P is a transition matrix which is reversible with respect to the stationary
distribution π. Then we declare {x, y} an edge if P (x, y) > 0. Reversibility implies that P (y, x) > 0
if P (x, y) > 0, and hence this is well-defined. We define conductances on the edges by setting
c(x, y) = π(x)P (x, y). Again by reversibility this is symmetric and with this choice of weights we
get π(x) = c(x). The study of reversible Markov chains is thus equivalent to the study of random
walks on weighted graphs.
Let now P be a transition matrix which is irreducible with state space Ω. (We do not assume that
P is reversible.)
A function h : Ω → R is called harmonic for P at the vertex x if
X
h(x) = P (x, y)h(y).
y∈Ω

For a subset B ⊆ Ω we define the hitting time of B by

τB = min{t ≥ 0 : Xt ∈ B}.

Proposition 2.1. Let X be an irreducible Markov chain with transition matrix P and let B ⊆ Ω.
Let f : B → R be a function defined on B. Then the function h(x) = Ex [f (XτB )] is the unique
extension h : Ω → R of f such that h(x) = f (x) for all x ∈ B and h is harmonic for P at all
x ∈ Ω \ B.

30
Proof. Clearly h(x) = f (x) for all x ∈ B. By conditioning on the first step of the Markov chain
we get for x ∈
/B
X X X
h(x) = P (x, y)Ex [f (XτB ) | X1 = y] = P (x, y)Ey [f (XτB )] = P (x, y)h(y),
y∈Ω y y

where for the second equality we used the Markov property.


We now turn to show uniqueness. Let h0 be another function satisfying the same conditions as h.
Then the function g = h − h0 is harmonic on Ω \ B and g = 0 on B. We first show that g ≤ 0.
Assume the contrary and let x ∈
/ B belong to the set
 
A = x : g(x) = max g(y) .
y∈Ω\B

Suppose that P (x, y) > 0. If g(y) < g(x), then harmonicity of g on Ω \ B implies that
X X
g(x) = g(z)P (x, z) = g(y)P (x, y) + P (x, z)g(z) < max g(y),
y∈Ω\B
z∈Ω z6=y

which is clearly a contradiction. Hence it follows that g(y) = maxz g(z), which means that y ∈ A.
By irreducibility we continue in the same way and we finally get that g(x) = 0, since we eventually
get to the boundary B. Hence this proves that max g = 0. Similarly, we can prove that min g = 0.
So we deduce that g = 0.

2.1 Electrical networks

The goal of this section is to explain the connection between random walks and electrical networks.
Again here G = (V, E) will be a finite connected graph with conductances (c(e))e on the edges. We
distinguish two vertices a and b that will be the source and the sink respectively.
Definition 2.2. A flow θ on G is a function defined on oriented edges ~e = (x, y) of E satisfying

θ(x, y) = −θ(y, x).


P
The divergence of the flow θ is defined to be div θ(x) = y∼x θ(x, y).

We note that by the antisymmetric property of θ we get


X X
div θ(x) = (θ(x, y) + θ(y, x)) = 0.
x (x,y)∈E

Definition 2.3. A flow θ from a to b is a flow such that

• div θ(x) = 0 ∀ x ∈
/ {a, b} (“flow in equals flow out” – Kirchoff’s node law)

• div θ(a) ≥ 0.

The strength of a flow θ from a to b is defined to be kθk := div θ(a). A unit flow is a flow with
kθk = 1.

31
We note that div θ(b) = −divθ(a).
A voltage W is a harmonic function on V \ {a, b}. By Proposition 2.1 a voltage always exists and
is uniquely determined by its boundary values W (a) and W (b).
Given a voltage W on G we define the current flow I on oriented edges via

W (x) − W (y)
I(x, y) = = c(x, y)(W (x) − W (y)).
r(x, y)

Exercise: Check it is a flow and then prove that the unit current flow is unique.
By the definition we immediately see that the current flow satisfies Ohm’s law:

r(x, y)I(x, y) = W (x) − W (y).

The current flow also satisfies the cycle law: if the oriented edges ~e1 , . . . , ~en form an oriented cycle,
then
X n
r(~ei )I(~ei ) = 0.
i=1
Proposition 2.4. Let I be a current and θ a flow from a to z satisfying the cycle law for any
cycle. If kθk = kIk, then θ = I.

Proof. Consider the function f = θ − I. Then since θ and I are flows with the same strength, it
follows that f satisfies Kirchoff’s node law at all nodes and the cycle law. Suppose that θ 6= I. Then
P
without loss of generality, there must exist an edge ~e1 such that f (~e1 ) > 0. Since y∼x f (x, y) = 0,
we get that there must exist an edge ~e2 to which ~e1 leads such that f (~e2 ) > 0. Continuing in
this way we get a sequence of oriented edges with positive value of f . Since the graph is finite, at
some point this sequence of edges should revisit a point. This then violates the cycle law. Hence
θ = I.

2.2 Effective resistance

Let W0 be the voltage with W0 (a) = 1 and W0 (z) = 0. By the uniqueness of harmonic functions,
we obtain that any other voltage W is given by

W (x) = (W (a) − W (z))W0 (x) + W (z).

We call I0 the current flow associated with W0 . Note that by the definition of the current flow, its
strength is given by
X W (a) − W (x)
kIk = = (W (a) − W (z)) kI0 k .
x∼a
r(a, x)

We thus see that the ratio


W (a) − W (z)
kIk
is independent of W . We define this to be the effective resistance

W (a) − W (z)
Reff (a, z) :=
kIk

32
and the reciprocal is called the effective conductance, Ceff (a, z).
Why is it called effective resistance? Suppose that we wanted to replace the whole network by
single edge joining a and z with resistance Reff (a, z). Then if we apply the same voltage at a and
z in both networks, then the same amount of current would flow through.
We are now ready to state the connection between random walks and electrical networks.
We define τx+ = min{t ≥ 1 : Xt = x}.

Proposition 2.5. Let X be a reversible chain on a finite state space. For any x, a, z ∈ Ω we have

1
Pa τz < τa+ =

.
c(a)Reff (a, z)

Proof. The function f (x) = Px (τz < τa ) is a harmonic function on Ω \ {a, z} and f (a) = 0 and
f (z) = 1. So from Proposition 2.1 we get that f has to be equal to the function

W (a) − W (x)
h(x) = ,
W (a) − W (z)

where W is a voltage, since they are both harmonic with the same boundary values. Therefore, we
obtain
 X X c(a, x) W (a) − W (x)
Pa τz < τa+ = P (a, x)Px (τz < τa ) = · .
x x∼a
c(a) W (a) − W (z)

By the definition of the current flow, the above sum is equal to


P
x∼a I(a, x) 1
=
c(a)(W (a) − W (z)) c(a)Reff (a, z)

and this proves the proposition.

Definition 2.6. The Green’s function for a random walk stopped at a stopping time τ is defined
to be "∞ #
X
Gτ (a, x) := Ea 1(Xt = x, τ > t) .
t=0

Lemma 2.7. Let X be a reversible Markov chain. Then for all a, z we have

Gτz (a, a) = c(a)Reff (a, z).

Proof. The number of visits to a before the first hitting time of z has a geometric distribution
with parameter Pa (τz < τa+ ). The statement follows from Proposition 2.5.

There are some ways of simplifying a network without changing quantities of interest.
Conductances in parallel add: let e1 and e2 be edges sharing the same endvertices. Then we
can replace both edges by a single edge of conductance equal to the sum of the conductances. Then
the same current will flow through and the same voltage difference will be applied. To see it, check
Kirchoff’s and Ohm’s laws with I(~e) = I(~e1 ) + I(~e2 ).

33
Resistances in series add: if v ∈ V \ {a, z} is a node of degree 2 with neighbours v1 and v2 ,
we could replace the edges (v, v1 ) and (v, v2 ) by a single edge (v1 , v2 ) of resistance r(v1 , v2 ) =
r(v, v1 ) + r(v, v2 ). To see it, check Kirchoff’s and Ohm’s laws with I(v1 , v2 ) = I(v1 , v) = I(v, v2 )
and the same as before everywhere else.
Gluing: If two vertices have the same voltage, we can glue them to a single vertex, while keeping
all existing edges. Since current never flows between vertices with the same voltage, potentials and
currents remain unchanged.
Example 2.8. Let a and b be two vertices on a finite connected tree T . Then the effective resistance
Reff (a, b) is equal to the distance on the tree between a and b.
Definition 2.9. Let θ be a flow on a finite connected graph G. We define its energy by
X
E(θ) = (θ(e))2 r(e),
e

where the sum is taken over all unoriented edges e. (Note that since θ is antisymmetric, we did not
need to take direction on e.)

The following theorem gives an equivalent definition of effective resistance as the minimal energy
over all flows of unit strength from a to z.
Theorem 2.10 (Thomson’s principle). Let G be a finite connected graph with edge conductances
(c(e))e . For all a and z we have

Reff (a, z) = inf{E(θ) : θ is a unit flow from a to z}.

The unique minimiser above is the unit current flow from a to z.

Proof. We follow [4].


Let i be the unit current flow from a to z associated to the potential ϕ.
We start by showing that
Reff (a, z) = E(i).

Using that i is a flow from a to z and Ohm’s law we have

1X 1X
E(i) = i(u, v)2 r(u, v) = i(u, v)(ϕ(u) − ϕ(v)) = ϕ(a) − ϕ(z) = Reff (a, z).
2 u,v 2 u,v
u∼v u∼v

Let j be another flow from a to z of unit strength. The goal is to show that E(j) ≥ E(i).
We define k = j − i. Then this is a flow of 0 strength. So we now get
X X
E(j) = (j(e))2 r(e) = (i(e) + k(e))2 r(e)
e e
X X X
= (i(e))2 r(e) + (k(e))2 r(e) + 2 k(e)i(e)r(e)
e e e
X
= E(i) + E(k) + 2 k(e)i(e)r(e).
e

34
We now show that X
k(e)i(e)r(e) = 0.
e

Since i is the unit current flow associated with ϕ, for e = (x, y) it satisfies

ϕ(x) − ϕ(y)
i(x, y) = .
r(x, y)

Substituting this above we obtain


X 1 XX 1 XX 1 XX
k(e)i(e)r(e) = · (ϕ(x) − ϕ(y))k(x, y) = · ϕ(x)k(x, y) + · ϕ(x)k(x, y),
e
2 x y∼x 2 x y∼x 2 x y∼x

where for the last equality we used the antisymmetric property of k. Since k is a flow of 0 strength,
we get that both these sums are equal to 0. Therefore this proves that

E(j) ≥ E(i)

with equality if and only if E(k) = 0 which is equivalent to k = 0.

Theorem 2.11 (Rayleigh monotonicity principle). The effective resistance is a monotone in-
creasing function as a function of the component resistances, i.e. if (r(e))e and (r0 (e))e satisfy
r(e) ≤ r0 (e) for all e, then
Reff (a, z; r) ≤ Reff (a, z; r0 ).

Proof. Let i and i0 be the unit current flows associated to the resistances r(e) and r0 (e) respectively.
Then by Thomson’s principle we get
X X
Reff (a, z; r) = (i(e))2 r(e) ≤ (i0 (e))2 r(e),
e e

where the inequality follows, since the energy is minimised by the unit current flow i. Using the
assumption on the resistances we now obtain
X X
(i0 (e))2 r(e) ≤ (i0 (e))2 r0 (e) = Reff (a, z; r0 )
e e

and this concludes the proof.

Corollary 2.12. Let G be a finite connected graph. Suppose we add an edge which is not adjacent
to a. Then this increases the escape probability Pa (τz < τa+ ).

Proof. Recall from Proposition 2.5 that

1
Pa τz < τa+ =

.
c(a)Reff (a, z)

Adding an edge means that we decrease the resistance of the edge from ∞ to a finite number.
Hence from Rayleigh’s monotonicity principle we get that the effective resistance will decrease.

35
Corollary 2.13. The operation of gluing vertices together cannot increase the effective resistance.

We now present a nice technique due to Nash and Williams to obtain lower bounds on effective
resistances.

Definition 2.14. We call a set of edges Π an edge–cutset separating a from z if every path from a
to z uses an edge of Π.

Proposition 2.15 (Nash-Williams inequality). If (Πk ) are disjoint edge-cutsets which separate a
from z, then
 −1
X X
Reff (a, z) ≥  c(e) .
k e∈Πk

Proof. By Thomson’s principle it suffices to prove that for any flow θ from a to z of unit strength
we have  −1
X X X
(θ(e))2 r(e) ≥  c(e) .
e k e∈Πk

Since the sets Πk are disjoint, we get


X XX
(θ(e))2 r(e) ≥ (θ(e))2 r(e).
e k e∈Πk

By the Cauchy-Schwarz inequality we now get


     2  2
X X Xp p X
 c(e) ·  r(e)(θ(e))2  ≥  c(e) r(e)|θ(e)| =  |θ(e)| .
e∈Πk e∈Πk e∈Πk e∈Πk

But since the sets Πk are cutsets separating a from z and θ has unit strength, this last sum is at
least 1. Rearranging completes the proof.

Proposition 2.16. Let a = (1, 1) and z = (n, n) be the opposite corners of the box Bn = [1, n]2 ∩Z2 .
Then
log(n − 1)
Reff (a, z) ≥ .
2
Remark 2.17. The effective resistance between a and z is also upper bounded by log n. To
prove it one defines a flow and shows that its energy is at most log n. For more details see [5,
Proposition 9.5].

Proof of Proposition 2.16. Let Πk = {(v, u) ∈ Bn : kvk∞ = k and kuk∞ = k − 1}. Then (Πk )
are disjoint edge-cutsets separating a from z. By counting the number of edges we get |Πk | = 2(k−1)
and since c(e) = 1 for all edges e we get
n
X 1 log(n − 1)
Reff (a, z) ≥ ≥
2(k − 1) 2
k=2

and this completes the proof.

36
Lemma 2.18. Let X be an irreducible Markov chain on a finite state space. Let τ be a stopping
time satisfying E[τ ] < ∞ and Pa (Xτ = a) = 1 for some a in the state space. Then for all x we
have
Gτ (a, x) = Ea [τ ] · π(x).

Proposition 2.19 (Commute time identity). Let X be a reversible Markov chain on a finite state
space. Then for all a, b we have

Ea [τa,b ] = Ea [τb ] + Eb [τa ] = c(G) · Reff (a, z),


P
where c(G) = 2 e c(e).

Proof. The stopping time τa,b which is the first time the walk comes back to a after having visited b

satisfies Pa Xτa,b = a = 1. Also note that we only visit a before time τb , i.e.

Gτa,b (a, a) = Gτb (a, a) = c(a)Reff (a, b),

where the last equality follows from Proposition 2.5. We can now apply the previous lemma to
finish the proof.

Remark 2.20. The previous identity immediately gives us that the effective resistance satisfies
the triangle inequality. Hence the effective resistance defines a metric on the graph G (the other
two properties are trivially satisfied).

2.3 Transience vs recurrence

So far we have been focusing on finite state spaces. In this section we will see how we can use the
electrical network point of view to determine transience and recurrence properties of graphs.
Let G = (V, E) be a countable graph and let 0 be a distinguished point. Let Gk = (Vk , Ek ) be an
exhaustion of G by finite graphs, i.e. Vn ⊆ Vn+1 and En contains all edges of E with endpoints
in Vn for all n, 0 ∈ Vn for all n and ∪n Vn = V .
For every n we construct a graph G∗n by gluing all the points of V \ Vn into a single point zn and
P
setting c(x, zn ) = z∈V \Vn c(x, z) and define

Reff (0, ∞) = lim Reff (0, zn ; G∗n ).


n→∞

Check by Rayleigh’s monotonicity principle that this limit does exist and is independent of the
choice of exhaustion. We thus have
1 1
P0 τ0+ = ∞ = lim Pa τzn < τ0+ = lim
 
= . (2.1)
n→∞ n→∞ c(0)Reff (0, zn ; G∗n ) c(0)Reff (0, ∞)

We can define a flow from 0 to ∞ on an infinite graph as an antisymmetric function on the edges
with divθ(x) = 0 for all x 6= 0.

Proposition 2.21. Let G be a countable connected weighted graph with conductances (c(e))e and
let 0 be a distinguished vertex.

37
(a) A random walk on G is recurrent if and only if Reff (0, ∞) = ∞.
(b) A random walk on G is transient if and only if there exists a unit flow i from 0 to ∞ of finite
energy E(i) = e (i(e))2 r(e) < ∞.
P

Proof. (a) The first part follows directly from (2.1). If the walk is recurrent, then

P0 τ0+ = ∞ = 0,


and hence Reff (0, ∞) = ∞ and vice versa.


(b) For the second part let Gn be an exhaustion of G by finite graphs. Then by definition

Reff (0, ∞) = lim Reff (0, zn ).


n→∞

Let in be the unit current flow from 0 to zn on the graph Gn and let vn be the corresponding
voltage. Then by Thomson’s principle we get

Reff (0, zn ) = E(in ).

Suppose now that there exists a unit flow θ from 0 to ∞ of finite energy. We call θn the restriction
of θ to the graph Gn . Then θn is a unit flow from a to zn in Gn . Applying Thomson’s principle we
obtain
E(in ) ≤ E(θn ) ≤ E(θ) < ∞.

Therefore we get
Reff (0, ∞) = lim E(in ) < ∞,
n→∞

which implies that the walk is transient from the first part.
Suppose now that the walk is transient. Then Reff (0, ∞) < ∞ from the first part. We want to
construct a unit flow from 0 to ∞ of finite energy. Since Reff (0, ∞) < ∞ we get that limn→∞ E(in ) <
∞, so there exists M > 0 such that E(in ) ≤ M for all n.
We now start a random walk from 0 and call Yn (x) the number of visits to x up until it hits zn .
We also call Y (x) the total number of visits to x. It is clear that Yn (x) ↑ Y (x) as n → ∞, and
hence by monotone convergence we get

lim E0 [Yn (x)] ↑ E0 [Y (x)] .


n→∞

Since the walk is assumed to be transient, we have E0 [Y (x)] < ∞. It is not hard to check that the
function Gτzn (0, x)/c(x) is a harmonic function with value 0 at zn , and it is equal to vn (x). (Use
reversibility and same proof as in Proposition 2.1). So we get

lim c(x)vn (x) = c(x)v(x)


n→∞

for some function v which is finite. Therefore, we can define

i(x, y) = c(x, y)(v(x) − v(y)) = lim c(x, y)(vn (x) − vn (y)).


n→∞

38
Since E(in ) ≤ M for all n using dominated convergence one can show that i is a unit flow from 0
to ∞ of finite energy (Check!).

Corollary 2.22. If G ⊆ G0 and G0 is recurrent, then G is also recurrent. If G is transient, then


G0 is also transient.

Corollary 2.23. Simple random walk is recurrent on Z2 and transient on Zd for all d ≥ 3.

Proof. For d = 2 we construct a new graph in which for each k we identify all vertices at distance k
from 0. By the series/parallel law we see that

n−1
X 1
Reff (0, ∂Λn ) ≥ .
8i − 4
i=1

Therefore, we get that


Reff (0, n) ≥ c log n → ∞ as n → ∞.

For d = 3 we are going to construct a flow of finite energy. To each directed edge of Z3 we attach
an orthogonal unit square intersecting e at its midpoint me . We now define the absolute value of
θ(~e) to be the area of the radial projection of this square onto the sphere of radius 1/4 centred at
the origin. We take θ(~e) with positive sign if ~e points in the same direction as the radial vector
from 0 to me and negative otherwise. By considering the projections of all faces of the unit cube
centred at a lattice point, we can see that θ satisfies Kirchoff’s node law at all vertices except for 0
(Check!). Hence θ is a flow from 0 to ∞ in Z3 . Its energy is given by
 c 2
2
X
E(θ) ≤ c1 n2 · < ∞,
n
n2

and hence this proves transience.


An alternative proof goes via embedding a binary tree with resistance between edges from level
n − 1 to n equal to ρn for a suitable ρ > 0. Then the effective resistance of this tree is given by
∞  
X ρ n
Reff (0, ∞) = .
2
n=1

Taking ρ < 2 makes it finite.


We now want to embed this tree in Z3 in such a way that a vertex at distance n − 1 and a neighbour
at distance n are separated by a path of length ρn .
The surface of a ball of radius k in R3 is of order k 2 , so in order to be able to accommodate this
tree in Z3 we need
(ρn )2 ≥ 2n ,

which then gives ρ > 2. This then gives that the effective resistance from 0 to ∞ in Z3 is bounded
by the effective resistance of the tree, and hence it is finite.
This idea has been used to show that random walk on the infinite component of supercritical
percolation cluster is transient in dimensions d ≥ 3 (see Grimmett, Kesten and Zhang (1993)).

39
2.4 Spanning trees

Definition 2.24. Let G be a finite connected graph. A spanning tree of G is a subgraph that is a
tree (no cycles) and which contains all the vertices of G.

Let G be a finite connected graph and let T be the set of spanning trees of G. We pick T uniformly
at random from T . We call T a uniform spanning tree (UST).
We will prove that T has the property of negative association, i.e.

Theorem 2.25. Let G = (V, E) be a finite graph. Let f, g ∈ E with f 6= g. Let T be a UST. Then

P(f ∈ T | g ∈ T ) ≤ P(f ∈ T ) .

In order to prove this theorem we are first going to establish a connection between spanning trees
and electrical networks.
Let N (s, a, b, t) be the set of spanning trees of G with the property that the unique path from s
to t passes along the edge (a, b) in the direction from a to b. We write N (s, a, b, t) = |N (s, a, b, t)|.
Let N be the total number of spanning trees of G. We then have the following theorem:

Theorem 2.26. The function

N (s, a, b, t) − N (s, b, a, t)
i(a, b) =
N
for all (a, b) ∈ E defines a unit flow from s to t satisfying Kirchoff ’s laws.

Remark 2.27. The above expression for i(a, b) is also equal to

i(a, b) = P(T ∈ N (s, a, b, t)) − P(T ∈ N (s, b, a, t)) .

Exactly the same proof as below would work if G is a weighted graph. In this case we would define
the weight of a tree to be Y
w(T ) = c(e)
e∈T

and we would set


X X
N∗ = w(T ) and N ∗ (s, a, b, t) = w(T ).
T ∈T T ∈N (s,a,b,t)

Then Theorem 2.26 would still be valid with


N ∗ (s, a, b, t) − N ∗ (s, b, a, t)
i∗ (a, b) =
N∗
for all edges (a, b). The negative association theorem would also be true in this case, i.e. when a
tree T is picked with probability proportional to its weight.

40
Proof of Theorem 2.26. It is obvious from definition that i is an antisymmetric function. We
next check that it satisfies Kirchoff’s node law, i.e. for all x ∈
/ {s, t} we have
X
i(a, x) = 0.
x∼a

We now count the contribution of each spanning tree T to the sum above. We now consider the
unique path from s to t in this spanning tree. If a is a vertex on this path, then there are two
edges on the path with endpoint a that contribute to the sum. The edge going into a and the one
going out of a. The first one will contribute −1/N and the second one 1/N . Now if a is not on
the path, then there is no contribution to the sum from T . Hence the overall contribution of T is
−1/N + 1/N = 0 and this proves Kirchoff’s node law.
We now check that it satisfies the cycle law. Let v1 , . . . , vn , vn+1 = v1 constitute a cycle C. We will
show that
n
X
i(vi , vi+1 ) = 0. (2.2)
i=1

To do this we will work with bushes instead of trees. We define an s/t bush to be a forest consisting
of exactly two trees Ts and Tt such that s ∈ Ts and t ∈ Tt . Let e = (a, b) be an edge. We
define B(s, a, b, t) as the set of s/t bushes with a ∈ Ts and b ∈ Tt .
We now claim that |B(s, a, b, t)| = N (s, a, b, t). Indeed, for every bush in B(s, a, b, t) by adding the
edge e we obtain a spanning tree of N (s, a, b, t). Also for every spanning tree T ∈ N (s, a, b, t) by
removing the edge e we obtain a bush in B(s, a, b, t).
So instead of counting the contribution of each spanning tree to the sum in (2.2) we look at bushes.
Let B be an s/t bush. Then B makes a contribution to i(a, b) of 1/N if B ∈ B(s, a, b, t), −1/N
if B ∈ B(s, b, a, t) and 0 otherwise.
So in total an s/t bush B contributes (F+ − F− )/N , where F+ is the number of pairs (vj , vj+1 ) so
that B ∈ B(s, vj , vj+1 , t) and similarly for F− .
But since C is a cycle, if there is a pair (vj , vj+1 ) in F+ , then there must be a pair (vi , vi+1 ) in F− .
(If not, this would violate the no cycle property of the tree.) Therefore we get F+ = F− and hence
the total contribution of B is 0.
Finally we need to check that i is a unit flow, i.e.
X
i(s, x) = 1.
x∼s

First we note that N (s, x, s, t) = 0 for all x ∼ s. Every spanning tree must contain a path from s
to t, and hence this gives that X
N (s, s, x, t) = N
x∼s

and concludes the proof.

Proof of Theorem 2.25. We consider G as a network with every edge having conductance 1. Let

41
e = (s, t) be an edge. Then from Theorem 2.26 we get that i is a unit current flow from s to t and

N (s, s, t, t)
i(s, t) = ,
N
where N (s, s, t, t) is the number of spanning trees that use the edge (s, t). Hence

N (s, s, t, t)
= P(e ∈ T ) .
N
Since the network has unit conductances, we get that

i(s, t) = ϕ(s) − ϕ(t),

where ϕ is the potential associated to the unit current i. Therefore the effective resistance between s
and t is given by
Reff (s, t) = i(s, t) = P(e ∈ T ) .

Let e and g be distinct edges of G. We write G.g for the graph obtained by gluing both endpoints of
g to a single vertex. In this way we obtain a one to one correspondence between spanning trees of G
containing g and spanning trees of G.g. Therefore, P(e ∈ T | g ∈ T ) is the proportion of spanning
trees of G.g containing e. So from the above

P(f ∈ T | g ∈ T ) = Reff (s, t; G.g).

Gluing the two endpoints of g decreases the effective resistance by Rayleigh’s principle, and hence

Reff (s, t; G.g) ≤ Reff (s, t; G),

which is exactly the statement of the theorem.

Definition 2.28. Let G be a finite connected graph. We write F for the set of forests of G (subsets
of G that do not contain cycles). Let F be a forest picked uniformly at random among all forests
in F. We refer to it as USF.

Conjecture 2.29. For f, g ∈ E with f 6= g the USF satisfies

P(f ∈ F | g ∈ F ) ≤ P(f ∈ F ) .

There is a computer aided proof (Grimmett and Winkler) which shows that for graphs on 8 or
fewer vertices this conjecture is true.

Theorem 2.30 (Foster’s theorem). Let G be a finite connected network on n vertices with conduc-
tances (c(e)) on the edges. Then X
c(e)Reff (e) = n − 1.
e∈E

P
Proof. Note that if T is a UST in G, then e∈E P(e ∈ T ) = n − 1. Using that P(e ∈ T ) =
c(e)Reff (e) concludes the proof.

42
2.5 Wilson’s algorithm

This section has been taken from Lyons and Peres [6].
In this section we will present a way to sample a uniform spanning tree of a given graph. This
algorithm was invented by David Wilson.
In order to describe the algorithm, we start by explaining the loop-erasure of a path. Let γ =
hx1 , . . . , xk i be a finite path in a graph which could be directed or undirected. We define its loop
erasure by erasing cycles in the chronological order in which they appear. More formally, let y0 = x0
and if xk = x0 , then we set m = 0 and stop here. Otherwise let i = max{j ≥ 1 : xj = x0 } and
define y1 = xi+1 . If y1 = xk , then we terminate, otherwise we continue in the same way as before.
Now in order to generate a spanning tree of a graph G = (V, E), we distinguish one vertex which
we call the root r and choose any ordering of V \ {r}. We now define a growing sequence of trees
by first setting T0 = {r} and then inductively: if Ti spans the whole graph, then we stop, otherwise
we pick the first vertex in our ordering which is not in Ti and start a simple random walk on the
graph until the first time that it hits Ti . Then we define the tree Ti+1 by appending to Ti the loop
erasure of the random walk path that we generated. This is called Wilson’s method.

Theorem 2.31 (Wilson’s algorithm). Wilson’s method yields a uniform spanning tree.

To prove the above miraculous statement we first prove some deterministic results about cycle
popping.
Here is another way of thinking about a random walk on the graph. We imagine an infinite stack
of cards (or instructions) lying under every vertex. On the i-th visit to the vertex we use the i-th
card, which tells us to which vertex we should jump at the next step. Every time we use a card,
we throw it away. The instructions on each card are i.i.d. and on vertex x the instructions contain
one of the neighbours of x with equal probability.
This is a way of generating a random walk (or more generally any Markov chain), but here we want
to sample a uniform spanning tree. So we put an empty stack of cards under the root r and we use
the other stacks as follows: at any time the top cards of the stacks give rise to a directed graph
whose edges are (x, y) where y is the instruction given in the top card underneath x. We call this
the visible graph at that time. If it happens that this contains no cycles, then we are done and
we have obtained a spanning tree rooted at r. If, however, there is a cycle, then we pop it, which
means that we remove the top card in the stacks under the vertices that lie on the cycle. Then we
continue with another cycle, until there is no cycle.
We claim that in this way we obtain a spanning tree with probability 1 and it has the uniform
distribution. Moreover, as we will show it has the same distribution as the spanning tree obtained
by running Wilson’s algorithm.
To prove both statements, we will keep track of the locations in the stack of the cards that are
popped, and we will call them colours. So we say that an edge (x, Sxi ) has colour i, where Sxi is the
i-th instruction at the vertex x. A coloured cycle is a cycle whose vertices are coloured (but they
do not need to have the same colour).
We now state and prove a deterministic statement about popping cycles.

43
Lemma 2.32. Given any stacks under the states, the order in which cycles are popped is irrelevant
in the sense that every order pops an infinite number of cycles or every order pops the same (finite
set of ) coloured cycles, thus leaving the same coloured spanning tree on top in the latter case.

Proof. We will show that if C is any coloured cycle that can be popped, that is, there is some
sequence C1 , C2 , . . . , Cn = C that may be popped in that order, but some coloured cycle C 0 6= C1
happens to be the first coloured cycle popped, then (1) C = C 0 , or else (2) C can still be popped from
the stacks after C 0 is popped. Once we show this, we are done, since if there are an infinite number
of coloured cycles that can be popped, the popping can never stop; whereas in the alternative case,
every coloured cycle that can be popped will be popped. Now if all the vertices of C 0 are disjoint
from those of C1 , C2 , . . . , Cn , then of course C can still be popped. Otherwise, let Ck be the first
cycle that has a vertex in common with C 0 . Now, all the edges of C 0 have colour 1. Consider any
vertex x in C 0 ∩ Ck . Since x ∈ / C1 ∪ . . . Ck−1 ,the edge in Ck leading out of x also has colour 1, so it
leads to the same vertex as it does in C 0 . We can repeat the argument for this successor vertex of
x, then for its successor, and so on, until we arrive at the conclusion that C 0 = Ck . Thus, C 0 = C
or we can pop C in the order C 0 , C1 , C2 , . . . , Ck1 , Ck+1 , . . . , Cn .

Proof of Theorem 2.31. Wilsons method (using loop-erased parts of a Markov chain) certainly
stops with probability 1 at a spanning tree. Using stacks to run the Markov chain and noting that
loop erasure in order of cycle creation is one way of popping cycles, we see that Wilsons method
pops all the cycles lying over a spanning tree. Because of Lemma 4.2, popping cycles in any other
manner also stops with probability 1 and with the same distribution. Furthermore, if we think
of the stacks as given in advance, then we see that all our choices inherent in Wilson’s method
have no effect whatsoever on the resulting spanning tree. Now to show that the distribution is the
desired one, think of a given set of stacks as defining a finite set O of coloured cycles lying over a
non-coloured spanning tree T . We do not need to keep track of the colours in the spanning tree,
since they are easily recovered from the colours in the cycles over it. Let X be the set of all pairs
(O, T ) that can arise from stacks corresponding to our given Markov chain. If (O, T ) ∈ X, then
also (O, T 0 ) ∈ X for any other spanning tree T 0 : indeed, anything at all can be in the stacks under
any finite set O of coloured cycles. That is, X = X1 × X2 , where X1 is a certain collection of sets of
coloured cycles and X2 is the set of all non-coloured spanning trees. Extend our definition of Ψ to
Q Q
coloured cycles C by Ψ(C) := e∈C p(e) and to sets O of coloured cycles by Ψ(O) := C∈O Ψ(C).
What is the chance of seeing a given set O of coloured cycles lying over a given spanning tree T ?
S
It is simply the probability of seeing all the arrows in O ∪ T in their respective places, which is
S
simply the product of p(e) for all e ∈ O ∪ T , in other words, Ψ(O)Ψ(T ). Letting P be the law of
(O, T ), we get P = µ1 × µ2 , where µi are probability measures proportional to Ψ on Xi . Therefore,
the set of coloured cycles seen is independent of the coloured spanning tree and the probability of
seeing the tree T is proportional to Ψ(T ). This shows that Wilson’s method does what we claimed
it does.

Corollary 2.33 (Cayley, 1889). The number of labeled unrooted trees with n vertices is nn−2 . Here,
a labeled tree is one whose vertices are labeled 1 through n.

We now mention another way of sampling uniform spanning trees due to Aldous and Broder. Their
algorithm proceeds as follows: choose a vertex r of G and start a simple random walk until the first

44
time that all vertices have been visited at least once. Then for every w ∈ V with w 6= r let (u, w)
be the directed edge that was traversed by the walk on its first visit to w. The edges obtained in
this way undirected constitute a uniform spanning tree.

45
2.6 Uniform spanning trees and forests

In the previous section we described an algorithm to generate a uniform spanning tree of a finite
connected graph. We will now see how to define a measure on spanning trees on infinite recurrent
graphs and uniform spanning forests on transient graphs.
We will do this by taking limits over exhaustions of the infinite graph by finite subgraphs.
Let us first look at the recurrent case. Let G be an infinite recurrent graph and let r be a dis-
tinguished vertex that will be the root of our spanning tree. We now start a random walk from
some other vertex of G and run it until it first hits r. Then we take its loop erasure. We now start
another walk from a point not already taken and start a random walk until the first time it hits
the existing tree and we apply loop erasure. We continue in the same way. In this way we obtain
a spanning tree of the infinite graph. The reason we call it a uniform spanning tree is because if
we restrict it to a big finite subgraph G0 of G, then the probabilities that a uniform spanning tree
of G0 contains a finite set of edges is close to the probability that a uniform spanning tree of G
contains the same set of edges.
In the case of a transient graph we start by defining F0 = ∅. Then inductively to define Fi+1 we
run a random walk from a point not already taken and we stop the random walk if it hits Fi or if
not, we let it run indefinitely. Then take its loop erasure, which is well defined since the walk is
transient, so it visits every state only finitely many times. Then we append the loop erasure to the
previous set and we continue in the same way. It is clear that what we get will contain no cycles.
This method of generating a uniform spanning forest was called by Benjamini, Lyons, Pemantle
and Peres (2001) Wilson’s method rooted at infinity.
Let (Gn ) be an exhaustion of an infinite graph G. Let µn be the uniform spanning tree measure
on Gn where we glue all the vertices of G \ Gn into a single vertex. The spanning forest we obtain in
the limit is also called wired uniform spanning forest as opposed to the free spanning forest where
there is no gluing taking place.

Theorem 2.34. The weak limit µ = limn→∞ µn exists and is supported on the set of spanning
forests of G. Moreover, the measure µ is the distribution of the spanning forest generated using
Wilson’s method rooted at infinity.

For a proof see [4, Theorem 2.11] and Lyons and Peres (Chapter 10).
We now turn to see in which cases the uniform spanning forest consists of a single tree. From
the construction it is clear that for a recurrent graph, this is always the case by construction. For
transient graphs the answer is in the next theorem.
Pemantle first answered this in 1991 and it was quite striking.

Proposition 2.35 (Pemantle (1991)). The uniform spanning forest is almost surely a tree if and
only if starting from every vertex a simple random walk and an independent loop-erased random
walk intersect infinitely many times with probability 1. Moreover, the probability that x and y are
in the same tree is equal to the probability that simple random walk started from x intersects an
independent loop-erased random walk started from y.

46
This proposition is obvious using Wilson’s method, but this method of generating uniform spanning
forests was not known to Pemantle.
In 2003 Lyons, Peres and Schramm showed that a random walk intersects an independent loop-
erased walk infinitely often if and only if the same is true for two independent random walks. So
then we get the following:
Theorem 2.36. Let G be an infinite transient graph. Then the uniform spanning forest is a single
tree if and only if two random walks started at any two vertices intersect with probability 1.

Let us now apply the previous theorem to spanning forests on Zd .


Theorem 2.37. The uniform spanning forest in Zd is a single tree almost surely for all d ≤ 4.

Proof for d ≥ 5. Let X and Y be two independent simple random walks on Zd starting from
X0 = x and Y0 = y.
We start by calculating the expected number of intersections
∞ X
X ∞
I= 1(Xt = Ys ).
t=0 s=0

For this we have


∞ X
X ∞ X ∞
X
Ex,y [I] = Px,y (Xt = Ys ) = Px−y (Xt+s = 0) = (t + 1) · Px−y (Xt = 0) .
t=0 s=0 t,s t=0

Since simple random walk only moves distance 1 at each time step, the sum above should start
from kx − yk. Then using that there exists a positive constant c so that for all x we have
c
Px (Xt ) ≤
td/2
we get that

X c
Ex,y [I] ≤ .
td/2−1
t=kx−yk

Now for kx − yk sufficiently large and d ≥ 5 we see that this last sum can be made smaller than ε.
Therefore, we get by Markov’s inequality that

Px,y (I > 0) ≤ Ex,y [I] < ε.

For every ε > 0 we can find x, y so that Ex,y [I] < ε. Therefore, from Wilson’s method rooted at
infinity we obtain
P(USF is connected) ≤ Px,y (X and Y intersect) ≤ ε.

Since this is true for all ε > 0, this gives that with probability 1 the uniform spanning forest consists
of a single tree.

Let us now prove that for simple random walk on Zd for all x and all t we have
c
Px (Xt = 0) ≤ .
td/2

47
It is easy to check by direct calculation and using Stirling’s formula that for a simple random
walk Y on Z we have
c
Px (Yt = 0) ≤ √ ,
t
where c is a positive constant. Now if we write Xt = (Xt1 , . . . , Xtd ) and denote by ti the number of
steps that the walk X i performed up to time t, then we get
   
t t c1
Px (Xt = 0) ≤ Px Xt = 0, ti > , ∀ i + P ∃i : ti < ≤ d/2 + e−c2 t .
2d 2d t

48
References

[1] Béla Bollobás and Oliver Riordan. Percolation. Cambridge University Press, New York, 2006.

[2] Hugo Duminil-Copin and Vincent Tassion. A new proof of the sharpness of the phase transition
for bernoulli percolation on Zd . Available at arXiv:1502.03051.

[3] Geoffrey Grimmett. Percolation, volume 321 of Grundlehren der Mathematischen Wis-
senschaften [Fundamental Principles of Mathematical Sciences]. Springer-Verlag, Berlin, second
edition, 1999.

[4] Geoffrey Grimmett. Probability on graphs, volume 1 of Institute of Mathematical Statistics


Textbooks. Cambridge University Press, Cambridge, 2010. Random processes on graphs and
lattices.

[5] David A. Levin, Yuval Peres, and Elizabeth L. Wilmer. Markov chains and mixing times.
American Mathematical Society, Providence, RI, 2009. With a chapter by James G. Propp and
David B. Wilson.

[6] Russell Lyons and Yuval Peres. Probability on Trees and Networks. Cambridge University Press,
New York, 2016. Available at http://pages.iu.edu/~rdlyons/.

[7] P. Nolin. Lecture Notes on Percolation Theory.

49

You might also like