Book
Book
2
Contents
I Basic Models 1
1 Random Graphs 3
1.1 Models and Relationships . . . . . . . . . . . . . . . . . . . . . . 3
1.2 Thresholds and Sharp Thresholds . . . . . . . . . . . . . . . . . . 9
1.3 Pseudo-Graphs . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
1.4 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
1.5 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2 Evolution 21
2.1 Sub-Critical Phase . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.2 Super-Critical Phase . . . . . . . . . . . . . . . . . . . . . . . . 34
2.3 Phase Transition . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
2.4 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
2.5 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
3 Vertex Degrees 51
3.1 Degrees of Sparse Random Graphs . . . . . . . . . . . . . . . . . 51
3.2 Degrees of Dense Random Graphs . . . . . . . . . . . . . . . . . 57
3.3 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
3.4 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
4 Connectivity 67
4.1 Connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
4.2 k-connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
4.3 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
4.4 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
5 Small Subgraphs 75
5.1 Thresholds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
5.2 Asymptotic Distributions . . . . . . . . . . . . . . . . . . . . . . 79
5.3 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
i
ii CONTENTS
5.4 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
6 Spanning Subgraphs 85
6.1 Perfect Matchings . . . . . . . . . . . . . . . . . . . . . . . . . . 85
6.2 Hamilton Cycles . . . . . . . . . . . . . . . . . . . . . . . . . . 93
6.3 Long Paths and Cycles in Sparse Random Graphs . . . . . . . . . 97
6.4 Greedy Matching Algorithm . . . . . . . . . . . . . . . . . . . . 99
6.5 Random Subgraphs of Graphs with Large Minimum Degree . . . 103
6.6 Spanning Subgraphs . . . . . . . . . . . . . . . . . . . . . . . . 106
6.7 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
6.8 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
9 Resilience 161
9.1 Perfect Matchings . . . . . . . . . . . . . . . . . . . . . . . . . . 161
9.2 Hamilton Cycles . . . . . . . . . . . . . . . . . . . . . . . . . . 162
9.3 The chromatic number . . . . . . . . . . . . . . . . . . . . . . . 174
9.4 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175
9.5 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175
13 Digraphs 267
13.1 Strong Connectivity . . . . . . . . . . . . . . . . . . . . . . . . . 267
13.2 Hamilton Cycles . . . . . . . . . . . . . . . . . . . . . . . . . . 275
13.3 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 278
13.4 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 279
14 Hypergraphs 281
14.1 Component Size . . . . . . . . . . . . . . . . . . . . . . . . . . . 281
14.2 Hamilton Cycles . . . . . . . . . . . . . . . . . . . . . . . . . . 286
14.3 Thresholds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 290
14.4 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 304
14.5 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 307
16 Mappings 347
16.1 Permutations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 347
16.2 Mappings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 350
16.3 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 357
16.4 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 358
17 k-out 361
17.1 Connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 361
17.2 Perfect Matchings . . . . . . . . . . . . . . . . . . . . . . . . . . 364
17.3 Hamilton Cycles . . . . . . . . . . . . . . . . . . . . . . . . . . 373
17.4 Nearest Neighbor Graphs . . . . . . . . . . . . . . . . . . . . . . 376
17.5 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 379
17.6 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 380
22 Inequalities 465
22.1 Binomial Coefficient Approximation . . . . . . . . . . . . . . . . 465
22.2 Balls in Boxes . . . . . . . . . . . . . . . . . . . . . . . . . . . . 466
22.3 FKG Inequality . . . . . . . . . . . . . . . . . . . . . . . . . . . 468
22.4 Sums of Independent Bounded Random Variables . . . . . . . . . 469
22.5 Sampling Without Replacement . . . . . . . . . . . . . . . . . . 475
22.6 Janson’s Inequality . . . . . . . . . . . . . . . . . . . . . . . . . 476
22.7 Martingales. Azuma-Hoeffding Bounds . . . . . . . . . . . . . . 479
22.8 Talagrand’s Inequality . . . . . . . . . . . . . . . . . . . . . . . . 485
22.9 Dominance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 488
25 Entropy 499
25.1 Basic Notions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 499
25.2 Shearer’s Lemma . . . . . . . . . . . . . . . . . . . . . . . . . . 502
26 Indices 559
Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 560
Main Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 567
vi CONTENTS
Preface
History
Random graphs were used by Erdős [285] to give a probabilistic construction of
a graph with large girth and large chromatic number. It was only later that Erdős
and Rényi began a systematic study of random graphs as objects of interest in their
own right. Early on they defined the random graph Gn,m and founded the subject.
Often neglected in this story is the contribution of Gilbert [382] who introduced
the model Gn,p , but clearly the credit for getting the subject off the ground goes to
Erdős and Rényi. Their seminal series of papers [286], [288], [289], [290] and in
particular [287], on the evolution of random graphs laid the groundwork for other
mathematicians to become involved in studying properties of random graphs.
In the early eighties the subject was beginning to blossom and it received a
boost from two sources. First was the publication of the landmark book of Béla
Bollobás [135] on random graphs. Around the same time, the Discrete Mathemat-
ics group in Adam Mickiewicz University began a series of conferences in 1983.
This series continues biennially to this day and is now a conference attracting
more and more participants.
The next important event in the subject was the start of the journal Random
Structures and Algorithms in 1990 followed by Combinatorics, Probability and
vii
viii CONTENTS
Computing a few years later. These journals provided a dedicated outlet for work
in the area and are flourishing today.
Acknowledgement
Several people have helped with the writing of this book and we would like to
acknowledge their help. First there are the students who have sat in on courses
based on early versions of this book and who helped to iron out the many typo’s
etc.
We would next like to thank the following people for reading parts of the book
before final submission: Andrew Beveridge, Deepak Bal, Malgosia Bednarska,
Patrick Bennett, Mindaugas Blozneliz, Antony Bonato, Boris Bukh, Fan Chung,
Amin Coja-Oghlan, Colin Cooper, Andrzej Dudek, Asaf Ferber, Nikolas Foun-
toulakis, Catherine Greenhill, Dan Hefetz, Paul Horn, Hsien–Kuei Hwang, Tal
Hershko, Jerzy Jaworski, Tony Johansson, Mihyun Kang, Michael Krivelevich,
Tomasz Łuczak, Colin McDiarmid, Andrew McDowell, Hosam Mahmoud, Mike
CONTENTS ix
Molloy, Tobias Müller, Rajko Nenadov, Wesley Pegden, Boris Pittel, Dan Poole,
Pawel Prałat, Oliver Riordan, Andrzej Ruciński, Katarzyna Rybarczyk, Wojtek
Samotij, Yilun Shang, Matas Šilekis, Greg Sorkin, Joel Spencer, Sam Spiro, Dud-
ley Stark, Angelika Steger, Prasad Tetali, Andrew Thomason, Linnus Wästlund,
Nick Wormald, Stephen Young.
Thanks also to Béla Bollobás for his advice on the structure of the book.
Conventions/Notation
Often in what follows, we will give an expression for a large positive integer. It
might not be obvious that the expression is actually an integer. In which case, the
reader can rest assured that he/she can round up or down and obtained any required
property. We avoid this rounding for convenience and for notational purposes.
In addition we list the following notation:
Mathematical Relations
• f (x) = O(g(x)): | f (x)| ≤ K|g(x)| for some constant K > 0 and all x ∈ R.
• A B: A/B → 0 as n → ∞.
• A B: A/B → ∞ as n → ∞.
• A . B or B & A if A ≤ (1 + o(1))B.
• [n]: This is {1, 2, . . . , n}. In general, if a < b are positive integers, then
[a, b] = {a, a + 1, . . . , b}.
Graph Notation
• G = (V, E): V = V (G) is the vertex set and E = E(G) is the edge set.
• N(S) = NG (S) = {w ∈
/ S : ∃v ∈ S such that {v, w} ∈ E} and dG (S) = |NG (S)|
for S ⊆ V (G).
• Gn,m : The family of all labeled graphs with vertex set V = [n] = {1, 2, . . . , n}
and exactly m edges.
• En,m = E(Gn,m ).
• Gn,p : A random graph on vertex set [n] where each possible edge occurs
independently with probability p.
• En,p = E(Gn,p ).
≥k : G
• Gδn,m n,m , conditioned on having minimum degree at least k.
• Gn,n,p : A random bipartite graph with vertex set consisting of two disjoint
copies of [n] where each of the n2 possible edges occurs independently with
probability p.
• Gn,d : The set of graphs with vertex set [n] and degree sequence
d = (d1 , d2 , . . . , dn ).
• Hn,p;k : A random k-uniform hypergraph on vertex set [n] where each of the
n
k possibles edge occurs independently with probability p.
CONTENTS xi
• ~Gk−out : A random digraph on vertex set [n] where each v ∈ [n] indepen-
dently chooses k random out-neighbors.
• Gk−out : The graph obtained from ~Gk−out by ignoring orientation and coa-
lescing multiple edges.
Probability
• N(0, 1): A random variable with the normal distribution, mean 0 and vari-
ance 1.
• Bin(n, p): A random variable with the binomial distribution with parameters
n, the number of trials and p, the probability of success.
Basic Models
1
Chapter 1
Random Graphs
Graph theory is a vast subject in which the goals are to relate various graph prop-
erties i.e. proving that Property A implies Property B for various properties A,B.
In some sense, the goals of Random Graph theory are to prove results of the form
“Property A almost always implies Property B”. In many cases Property A could
simply be “Graph G has m edges”. A more interesting example would be the fol-
lowing: Property A is “G is an r-regular graph, r ≥ 3” and Property B is “G is
r-connected”. This is proved in Chapter 11.
Before studying questions such as these, we will need to describe the basic
models of a random graph.
assign a probability
n−1
P(G) = 2 .
m
Equivalently, we start with an empty graph on the set [n], and insert m edges
n
in such a way that all possible (m2) choices are equally likely. We denote such a
random graph by Gn,m = ([n], En,m ) and call it a uniform random graph.
We now describe a similar model. Fix 0 ≤ p ≤ 1. Then for 0 ≤ m ≤ n2 , assign
to each graph G with vertex set [n] and m edges a probability
n
P(G) = pm (1 − p)(2)−m ,
3
4 CHAPTER 1. RANDOM GRAPHS
where 0 ≤ p ≤1. Equivalently, we start with an empty graph with vertex set [n]
and perform n2 Bernoulli experiments inserting edges independently with proba-
bility p. We call such a random graph, a binomial random graph and denote it by
Gn,p = ([n], En,p ). This was introduced by Gilbert [382]
As one may expect there is a close relationship between these two models of
random graphs. We start with a simple observation.
Lemma 1.1. A random graph Gn,p , given that its number of edges is m, is equally
n
likely to be one of the (2) graphs that have m edges.
m
{Gn,p = G0 } ⊆ {|En,p | = m}
we have
P(Gn,p = G0 , |En,p | = m)
P(Gn,p = G0 | |En,p | = m) =
P(|En,p | = m)
P(Gn,p = G0 )
=
P(|En,p | = m)
n
pm (1 − p)(2)−m
= n
(2) pm (1 − p)(n2)−m
m
n−1
= 2 .
m
Thus Gn,p conditioned on the event {Gn,p has m edges} is equal in distribu-
tion to Gn,m , the graph chosen uniformly at random from all graphs with m edges.
Obviously, the main difference between those two models of random graphs is that
in Gn,m we choose its number of edges, while in the case of Gn,p the number of
edges is the Binomial random variable with the parameters n2 and p. Intuitively,
for large n random graphs Gn,m and Gn,p should behave in a similar fashion when
the number of edges m in Gn,m equals or is “close” to the expected number of
edges of Gn,p , i.e., when
n2 p
n
m= p≈ , (1.1)
2 2
or, equivalently, when the edge probability in Gn,p
2m
p≈ . (1.2)
n2
1.1. MODELS AND RELATIONSHIPS 5
1 − p = (1 − p1 )(1 − p2 ), (1.3)
or, equivalently,
p = p1 + p2 − p1 p2 .
Thus an edge is not included in Gn,p if it is not included in either of Gn,p1 or Gn,p2 .
It follows that
Gn,p = Gn,p1 ∪ Gn,p2 ,
where the two graphs Gn,p1 , Gn,p2 are independent. So when we write
Gn,p1 ⊆ Gn,p ,
we mean that the two graphs are coupled so that Gn,p is obtained from Gn,p1 by
superimposing it with Gn,p2 and replacing eventual double edges by a single one.
We can also couple random graphs Gn,m1 and Gn,m2 where m2 ≥ m1 via
Gn,m2 = Gn,m1 ∪ H.
Here H is the random graph on vertex set [n] that has m = m2 − m1 edges chosen
uniformly at random from [n]
2 \ En,m1 .
Consider now a graph property P defined as a subset of the set of all labeled
n
graphs on vertex set [n], i.e., P ⊆ 2(2) . For example, all connected graphs (on n
vertices), graphs with a Hamiltonian cycle, graphs containing a given subgraph,
planar graphs, and graphs with a vertex of given degree form a specific “graph
property”.
We will state below two simple observations which show a general relation-
ship between Gn,m and Gn,p in the context of the probabilities of having a given
graph property P. The constant 10 in the next lemma is not best possible, but in
the context of the usage of the lemma, any constant will suffice.
Next recall that the number of edges |En,p | of a random graph Gn,p is a random
variable with the Binomial distribution with parameters n2 and p. Applying Stir-
ling’s Formula:
k
k √
k! = (1 + o(1)) 2πk, (1.5)
e
and putting N = n2 , we get, after substituting (1.5) for the factorials in Nm ,
N m n
P(|En,p | = m) = p (1 − p)(2)−m
m
√
N N 2πN pm (1 − p)N−m
= (1 + o(1)) p (1.6)
mm (N − m)N−m 2π m(N − m)
s
N
= (1 + o(1)) ,
2πm(N − m)
Hence
1
P(|En,p | = m) ≥ √ ,
10 m
so
P(Gn,m ∈ P) ≤ 10m1/2 P(Gn,p ∈ P).
1.1. MODELS AND RELATIONSHIPS 7
and
P(Gn,m ∈ P) ≤ P(Gn,m0 ∈ P), (1.8)
respectively.
For monotone increasing graph properties we can get a much better upper bound
on P(Gn,m ∈ P), in terms of P(Gn,p ∈ P), than that given by Lemma 1.2.
N
P(Gn,p ∈ P) = ∑ P(Gn,k ∈ P) P(|En,p| = k)
k=0
N
≥ ∑ P(Gn,k ∈ P) P(|En,p| = k)
k=m
The number of edges |En,p | in Gn,p has the Binomial distribution with parameters
N, p. Hence
N
P(Gn,p ∈ P) ≥ P(Gn,m ∈ P) ∑ P(|En,p| = k)
k=m
N
= P(Gn,m ∈ P) ∑ uk , (1.9)
k=m
where
N k
uk = p (1 − p)N−k .
k
Now, using Stirling’s formula,
N N pm (1 − p)N−m 1 + o(1)
um = (1 + o(1)) 1/2
= .
m
m (N − m) N−m (2πm) (2πm)1/2
after using Lemma 22.1(a),(b) to obtain the inequality. and our assumptions on
N, p to obtain the second.
It follows that for 0 ≤ t ≤ m1/2 ,
( )
t−1
1 + o(1) s s+1
um+t ≥ exp − ∑ − ≥
(2πm)1/2 s=0 N − m − s m
n 2 o
t
exp − 2m − o(1)
,
(2πm)1/2
other situations we can use a stronger and more widely applicable result. The
theorem below, which we state without proof, gives precise conditions for the
asymptotic equivalence of random graphs Gn,p and Gn,m . It is due to Łuczak
[555].
p
Theorem 1.4. Let 0 ≤ p0 ≤ 1, s(n) = n p(1 − p) → ∞, and ω(n) → ∞ arbitrarily
slowly as n → ∞.
(i) Suppose that P is a graph property such that P(Gn,m ∈ P) → p0 for all
n n
m∈ p − ω(n)s(n), p + ω(n)s(n) .
2 2
Then P(Gn,p ∈ P) → p0 as n → ∞,
as n → ∞.
P(Gn,p(ε) ∈ P) = ε.
Gn,1−(1−p)k ⊆ Gn,kp ,
/ P implies G1 , G2 , . . . , Gk ∈
and so Gn,kp ∈ / P. Hence
/ P) ≤ [P(Gn,p ∈
P(Gn,kp ∈ / P)]k .
1.2. THRESHOLDS AND SHARP THRESHOLDS 11
/ P) ≤ 2−ω = o(1).
P(Gn,ω p∗ ∈
lim P(En ) = 1.
n→∞
Thus the statement that says p∗ is a threshold for a property P in Gn,p is the same
as saying that Gn,p 6∈ P w.h.p. if p p∗ , while Gn,p ∈ P w.h.p. if p p∗ .
In many situations we can observe that for some monotone graph properties
more “subtle” thresholds hold. We call them “sharp thresholds”. More precisely,
Definition 1.8. A function m∗ = m∗ (n) is a sharp threshold for a monotone in-
creasing property P in the random graph Gn,m if for every ε > 0,
0 i f m/m∗ ≤ 1 − ε
lim P(Gn,m ∈ P) =
n→∞ 1 i f m/m∗ ≥ 1 + ε.
0 i f p/p∗ ≤ 1 − ε
lim P(Gn,p ∈ P) =
n→∞ 1 i f p/p∗ ≥ 1 + ε.
This simple graph property is clearly monotone increasing and we will show be-
low that p∗ = 1/n2 is a threshold for a random graph Gn,p of having at least one
edge (being non-empty).
Lemma 1.10. Let P be the property defined above, i.e., stating that Gn,p contains
at least one edge. Then
(
0 if p n−2
lim P(Gn,p ∈ P) =
n→∞ 1 if p n−2 .
Proof. Let X be a random variable counting edgesin Gn,p . Since X has the Bino-
mial distribution, then E X = 2 p, and Var X = n2 p(1 − p) = (1 − p) E X.
n
A standard way to show the first part of the threshold statement, i.e. that w.h.p.
a random graph Gn,p is empty when p = o(n−2 ), is a very simple consequence of
the Markov inequality, called the First Moment Method, see Lemma 21.2. It states
that if X is a non-negative integer valued random variable, then
Var X
P(X > 0) ≥ 1 − .
(E X)2
Consider the monotone decreasing graph property that a graph contains an isolated
vertex, i.e. a vertex of degree zero:
We will show that m∗ = 21 n log n is the sharp threshold function for the above
property P in Gn,m .
Lemma 1.11. Let P be the property that a graph on n vertices contains at least
one isolated vertex and let m = 21 n(log n + ω(n)). Then
(
1 if ω(n) → −∞
lim P(Gn,m ∈ P) =
n→∞ 0 if ω(n) → ∞.
Proof. To see that the second statement of Lemma 1.11 holds we use the First
Moment Method. Namely, let X0 = Xn,0 be the number of isolated vertices in the
random graph Gn,m . Then X0 can be represented as the sum of indicator random
variables
X0 = ∑ Iv ,
v∈V
where (
1 if v is an isolated vertex in Gn,m
Iv =
0 otherwise.
14 CHAPTER 1. RANDOM GRAPHS
So
(n−1
2 )
m
E X0 = ∑ E Iv = n n =
v∈V
() 2
m
m m−1
n−2 4i
n ∏ 1− =
n i=0 n(n − 1)(n − 2) − 2i(n − 2)
n−2 m (log n)2
n 1+O , (1.10)
n n
assuming that ω = o(log n).
(For the product we use 1 ≥ ∏m−1 m−1
i=0 (1 − xi ) ≥ 1 − ∑i=0 xi which is valid for all
0 ≤ x0 , x1 , . . . , xm−1 ≤ 1.)
Hence,
n−2 m
2m
E X0 ≤ n ≤ ne− n = e−ω ,
n
for m = 12 n(log n + ω(n)).
(1 + x ≤ ex is one of the basic inequalities stated in Lemma 22.1.)
So E X0 → 0 when ω(n) → ∞ as n → ∞ and the First Moment Method implies
that X0 = 0 w.h.p.
To show that Lemma 1.11 holds in the case when ω → −∞ we first observe
from (1.10) that in this case
n−2 m
E X0 = (1 − o(1))n
n
2m
≥ (1 − o(1))n exp −
n−2
−ω
≥ (1 − o(1))e → ∞, (1.11)
The second inequality in the above comes from Lemma 22.1(b), and we have once
again assumed that ω = o(log n) to justify the first equation.
We caution the reader that E X0 → ∞ does not prove that X0 > 0 w.h.p. In
Chapter 5 we will see an example of a random variable XH , where E XH → ∞ and
yet XH = 0 w.h.p.
We will now use a stronger version of the Second Moment Method (for its
proof see Section 21.1 of Chapter 21). It states that if X is a non-negative integer
valued random variable then
(E X)2 Var X
P(X > 0) ≥ 2
= 1− . (1.12)
EX EX2
Notice that
1.2. THRESHOLDS AND SHARP THRESHOLDS 15
!2
E X02 = E ∑ Iv = ∑ E(Iu Iv )
v∈V u,v∈V
= ∑ P(Iu = 1, Iv = 1)
u,v∈V
= ∑ P(Iu = 1, Iv = 1) + ∑ P(Iu = 1, Iv = 1)
u6=v u=v
n−2
( ) 2
= n(n − 1) mn + E X0
() 2
m
2m
n−2
≤ n2 + E X0
n
= (1 + o(1))(E X0 )2 + E X0 .
The last equation follows from (1.10).
Hence, by (1.12),
(E X0 )2
P(X0 > 0) ≥
E X02
(E X0 )2
=
1 + o(1))((E X0 )2 + E X0 )
1
=
(1 + o(1)) + (E X0 )−1
= 1 − o(1),
on using (1.11). Hence P(X0 > 0) → 1 when ω(n) → −∞ as n → ∞, and so we can
conclude that m = m(n) is the sharp threshold for the property that Gn,m contains
isolated vertices.
For this simple random variable, we worked with Gn,m . We will in general
work with the more congenial independent model Gn,p and translate the results to
Gn,m if so desired.
For another simple example of the use of the second moment method, we will
prove
Theorem 1.12. If m/n → ∞ then w.h.p. Gn,m contains at least one triangle.
Proof. Because having a triangle is a monotone increasing property we can prove
the result in Gn,p assuming that np → ∞.
Assume first that np = ω ≤ log n where ω = ω(n) → ∞ and let Z be the number
of triangles in Gn,p . Then
ω3
n 3
EZ = p ≥ (1 − o(1)) → ∞.
3 6
16 CHAPTER 1. RANDOM GRAPHS
We remind the reader that simply having E Z → ∞ is not sufficient to prove that
Z > 0 w.h.p.
Next let T1 , T2 , . . . , TM , M = n3 denote the triangles of Kn . Then
M
E Z2 = ∑ P(Ti , T j ∈ Gn,p )
i, j=1
M M
= ∑ P(Ti ∈ Gn,p ) ∑ P(T j ∈ Gn,p | Ti ∈ Gn,p ) (1.13)
i=1 j=1
M
= M P(T1 ∈ Gn,p ) ∑ P(T j ∈ Gn,p | T1 ∈ Gn,p ) (1.14)
j=1
M
= E Z × ∑ P(T j ∈ Gn,p | T1 ∈ Gn,p ).
j=1
This proves the theorem for p ≤ logn n . For larger p we can use (1.7).
We can in fact use the second moment method to show that if m/n → ∞ then
w.h.p. Gn,m contains a copy of a k-cycle Ck for any fixed k ≥ 3. See Theorem 5.3,
see also Exercise 1.4.7.
1.3. PSEUDO-GRAPHS 17
1.3 Pseudo-Graphs
We sometimes use one of the two the following models that are related to Gn,m
and have a little more independence. (We will use Model A in Section 7.3 and
Model B in Section 6.4).
Model A: We let x = (x1 , x2 , . . . , x2m ) be chosen uniformly at random from
[n]2m .
Model B: We let x = (x1 , x2 , . . . , x2m ) be chosen uniformly at random from
[n]m
2 .
(X)
The (multi-)graph Gn,m , X ∈ {A, B} has vertex set [n] and edge set Em =
{{x2i−1 , x2i } : 1 ≤ i ≤ m}. Basically, we are choosing edges with replacement. In
Model A we allow loops and in Model B we do not. We get simple graphs from
(X∗)
by removing loops and multiple edges to obtain graphs Gn,m with m∗ edges. It
is not difficult to see that for X ∈ {A, B} and conditional on the value of m∗ that
(X∗)
Gn,m is distributed as Gn,m∗ , see Exercise (1.4.11).
More importantly, we have that for G1 , G2 ∈ Gn,m ,
(X) (X) (X) (X)
P(Gn,m = G1 | Gn,m is simple) = P(Gn,m = G2 | Gn,m is simple), (1.15)
for X = A, B.
This is because for i = 1, 2,
Indeed, we can permute the edges in m! ways and permute the vertices within
(X)
edges in 2m ways without changing the underlying graph. This relies on Gn,m
being simple.
Secondly, if m = cn for a constant c > 0 then with N = n2 , and using Lemma
22.2,
N m!2m
(X)
P(Gn,m is simple) ≥ ≥
m n2m
Nm m2 m3 m!2m
(1 − o(1)) exp − −
m! 2N 6N 2 n2m
2 +c)
= (1 − o(1))e−(c . (1.16)
(X) (X)
P(Gn,m ∈ P) = P(Gn,m ∈ P | Gn,m is simple) ≤
18 CHAPTER 1. RANDOM GRAPHS
2 +c (X)
(1 + o(1))ec P(Gn,m ∈ P). (1.17)
Here we have used the inequality P(A | B) ≤ P(A)/ P(B) for events A, B.
(X)
We will use this model a couple of times and (1.17) shows that if P(Gn,m ∈
P) = o(1) then P(Gn,m ∈ P) = o(1), for m = O(n).
(A)
Model Gn,m was introduced independently by Bollobás and Frieze [146] and
by Chvátal [195].
1.4 Exercises
We point out here that in the following exercises, we have not asked for best pos-
sible results. These exercises are for practise. You will need to use the inequalities
from Section 22.1.
1.4.1 Suppose that p = d/n where d = o(n1/3 ). Show that w.h.p. Gn,p has no
copies of K4 .
1.4.2 Suppose that p = d/n where d > 1. Show that w.h.p. Gn,p contains an
induced path of length (log n)1/2 .
1.4.3 Suppose that p = d/n where d = O(1). Prove that w.h.p., in Gn,p , for all
S ⊆ [n], |S| ≤ n/ log n, we have e(S) ≤ 2|S|, where e(S) is the number of
edges contained in S.
1.4.4 Suppose that p = log n/n. Let a vertex of Gn,p be small if its degree is less
than log n/100. Show that w.h.p. there is no edge of Gn,p joining two small
vertices.
1.4.5 Suppose that p = d/n where d is constant. Prove that w.h.p., in Gn,p , no
vertex belongs to more than one triangle.
1.4.6 Suppose that p = d/n where
l d is constant.
m Prove that w.h.p. Gn,p contains
a vertex of degree exactly (log n)1/2 .
1.4.7 Suppose that k ≥ 3 is constant and that np → ∞. Show that w.h.p. Gn,p
contains a copy of the k-cycle, Ck .
1.4.8 Suppose that 0 < p < 1 is constant. Show that w.h.p. Gn,p has diameter
two.
1.4.9 Let f : [n] → [n] be chosen uniformly at random from all nn functions from
[n] → [n]. Let X = { j :6 ∃i s.t. f (i) = j}. Show that w.h.p. |X| ≈ e−1 n.
1.5. NOTES 19
1.5 Notes
Friedgut and Kalai [329] and Friedgut [330] and Bourgain [160] and Bourgain
and Kalai [159] provide much greater insight into the notion of sharp thresholds.
Friedgut [328] gives a survey of these aspects. For a graph property A let µ(p, A )
be the probability that the random graph Gn,p has property A . A threshold is
)
coarse if it is not sharp. We can identify coarse thresholds with p dµ(p,A dp < C for
some absolute constant 0 < C. The main insight into coarse thresholds is that to
exist, the occurrence of A can in the main be attributed to the existence of one
of a bounded number of small subgraphs. For example, Theorem 2.1 of [328]
states that there exists a function K(C, ε) such that the following holds. Let A be
a monotone property of graphs that is invariant under automorphism and assume
)
that p dµ(p,A
dp < C for some constant 0 < C. Then for every ε > 0 there exists a
finite list of graphs G1 , G2 , . . . , Gm all of which have no more than K(ε,C) edges,
such that if B is the family of graphs having one of these graphs as a subgraph
then µ(p, A ∆B) ≤ ε.
20 CHAPTER 1. RANDOM GRAPHS
Chapter 2
Evolution
Here begins our story of the typical growth of a random graph. All the results up
to Section 2.3 were first proved in a landmark paper by Erdős and Rényi [287].
The notion of the evolution of a random graph stems from a dynamic view of a
graph process: viz. a sequence of graphs:
G0 = ([n], 0),
/ G1 , G2 , . . . , Gm , . . . , GN = Kn .
21
22 CHAPTER 2. EVOLUTION
Theorem 2.2. If m n1/2 then Gm is the union of isolated vertices and edges
w.h.p.
Proof. Let p = m/N, m = n1/2 /ω and let X be the number of paths of length two
in the random graph Gn,p . By the First Moment Method,
n4
n 2
P(X > 0) ≤ E X = 3 p ≤ → 0,
3 2N 2 ω 2
2.1. SUB-CRITICAL PHASE 23
as n → ∞. Hence
Notice that the property that a graph contains a path of a given length two is
monotone increasing, so by Lemma 1.3,
Proof. Let p = Nm , m = ωn1/2 and X be the number of paths of length two in Gn,p .
Then
n 2
EX = 3 p ≈ 2ω 2 → ∞,
3
as n → ∞. This however does not imply that X > 0 w.h.p.! To show that X > 0
w.h.p. we will apply the Second Moment Method
Let P2 be the set of all paths of length two in the complete graph Kn , and let
X̂ be the number of isolated paths of length two in Gn,p i.e. paths that are also
components of Gn,p . We will show that w.h.p. Gn,p contains such an isolated
path. Now,
X̂ = ∑ IP⊆i Gn,p .
P∈P2
We always use IE to denote the indicator for an event E . The notation ⊆i indicates
that P is contained in Gn,p as a component (i.e. P is isolated). Having a path of
length two is a monotone increasing property. Therefore we can assume that m =
o(n) and so np = o(1) and the result for larger m will follow from monotonicity
and coupling. Then
n 2
E X̂ = 3 p (1 − p)3(n−3)+1
3
n3 4ω 2 n
≥ (1 − o(1)) (1 − 3np) → ∞,
2 n4
as n → ∞.
In order to compute the second moment of the random variable X̂ notice that,
∗
X̂ 2 = ∑ ∑ IP⊆i Gn,p IQ⊆i Gn,p = ∑P,Q∈P IP⊆i Gn,p IQ⊆i Gn,p ,
2
P∈P2 Q∈P2
24 CHAPTER 2. EVOLUTION
where the last sum is taken over P, Q ∈ P2 such that either P = Q or P and Q are
vertex disjoint. The simplification that provides the last summation is precisely
the reason that we introduce path-components (isolated paths). Now
( )
E X̂ 2 = ∑ ∑ P(Q ⊆i Gn,p| P ⊆i Gn,p) P(P ⊆i Gn,p ).
P Q
The expression inside the brackets is the same for all P and so
where P{1,2,3} denotes the path on vertex set [3] = {1, 2, 3} with middle vertex
2. By conditioning on the event P(1,2,3) ⊆i Gn,p , i.e, assuming that P(1,2,3) is a
component of Gn,p , we see that all of the nine edges between Q and P(1,2,3) must
be missing. Therefore
n 2
2 3(n−6)+1
≤ E X̂ 1 + (1 − p)−9 E X̂ .
E X̂ ≤ E X̂ 1 + 3 p (1 − p)
3
So, by the Second Moment Method (see Lemma 21.5),
(E X̂)2 (E X̂)2
P(X̂ > 0) ≥ ≥
E X̂ 1 + (1 − p)−9 E X̂
E X̂ 2
1
= →1
(1 − p)−9 + [E X̂]−1
as n → ∞, since p → 0 and E X̂ → ∞. Thus
P(Gn,p contains an isolated path of length two) → 1,
which implies that P(Gn,p contains a path of length two) → 1. As the property of
having a path of length two is monotone increasing it in turn implies that
P(Gm contains a path of length two) → 1
for m n1/2 and the theorem follows.
From Theorems 2.2 and 2.3 we obtain the following corollary.
Corollary 2.4. The function m∗ (n) = n1/2 is the threshold for the property that a
random graph Gm contains a path of length two, i.e.,
(
o(1) if m n1/2 .
P(Gm contains a path of length two) =
1 − o(1) if m n1/2 .
2.1. SUB-CRITICAL PHASE 25
As we keep adding edges, trees on more than three vertices start to appear.
Note that isolated vertices, edges and paths of length two are also trees on one,
two and three vertices, respectively. The next two theorems show how long we
have to “wait” until trees with a given number of vertices appear w.h.p.
k−2
Theorem 2.5. Fix k ≥ 3. If m n k−1 , then w.h.p. Gm contains no tree with k
vertices.
k−2
2 3
Proof. Let m = n k−1 /ω and then p = Nm ≈ ωnk/(k−1) ≤ ωnk/(k−1) . Let Xk denote the
number of trees with k vertices in Gn,p . Let T1 , T2 , . . . , TM be an enumeration of
the copies of k-vertex trees in Kn . Let
The probability that a tree T occurs in Gn,p is pe(T ) , where e(T ) is the number of
edges of T . So,
M
E Xk = ∑ P(At ) = M pk−1.
t=1
n k−2
since one can choose a set of k vertices in nk ways and then by
But M = k k
Cayley’s formula choose a tree on these vertices in kk−2 ways. Hence
n k−2 k−1
E Xk = k p . (2.1)
k
Noting also that (see Lemma 22.1(c))
n ne k
≤ ,
k k
we see that
ne k k−1
k−2 3
E Xk ≤ k
k ωnk/(k−1)
3k−1 ek
= → 0,
k2 ω k−1
as n → ∞, seeing as k is fixed.
Thus we see by the first moment method that,
Let us check what happens if the number of edges in Gm is much larger than
k−2
n .
k−1
k−2
Theorem 2.6. Fix k ≥ 3. If m n k−1 , then w.h.p. Gm contains a copy of every
fixed tree with k vertices.
k−2
Proof. Let p = Nm , m = ωn k−1 where ω = o(log n) and fix some tree T with k
vertices. Denote by X̂k the number of isolated copies of T (T -components) in
Gn,p . Let aut(H) denote the number of automorphisms of a graph H. Note that
there are k!/aut(T ) copies of T in the complete graph Kk . To see this choose a
copy of T with vertex set [k]. There are k! ways of mapping the vertices of T to
the vertices of Kk . Each map f induces a copy of T and two maps f1 , f2 induce
the same copy iff f2 f1−1 is an automorphism of T .
So,
n k! k
E X̂k = pk−1 (1 − p)k(n−k)+(2)−k+1 (2.2)
k aut(T )
(2ω)k−1
= (1 + o(1)) → ∞.
aut(T )
k
In (2.2) we have approximated nk ≤ nk! and used the fact that ω = o(log n) in
k
order to show that (1 − p)k(n−k)+(2)−k+1 = 1 − o(1).
Next let T be the set of copies of T in Kn and T[k] be a fixed copy of T on
vertices [k] of Kn . Then, arguing as in (2.3),
2
Notice that the (1 − p)−k factor comes from conditioning on the event
T[k] ⊆i Gn,p which forces the non-existence of fewer than k2 edges.
Hence, by the Second Moment Method,
(E X̂k )2
P(X̂k > 0) ≥ 2
→ 1.
E X̂k 1 + (1 − p)−k E X̂k
2.1. SUB-CRITICAL PHASE 27
In the next theorem we show that “on the threshold” for k vertex trees, i.e., if
k−2
m = cn k−1 , where c is a constant, c > 0, the number of tree components of a given
order asymptotically follows the Poisson distribution. This time we will formulate
both the result and its proof in terms of Gm .
k−2
Theorem 2.8. If m = cn k−1 , where c > 0, and T is a fixed tree with k ≥ 3 vertices,
then
P(Gm contains an isolated copy of tree T ) → 1 − e−λ ,
k−1
as n → ∞, where λ = (2c)
aut(T ) .
More precisely, the number of copies of T is asymptotically distributed as the
Poisson distribution with expectation λ .
Proof. Let T1 , T2 , . . . , TM be an enumeration of the copies of some k vertex tree T
in Kn .
Let
Ai = {Ti occurs as a component in Gm }.
Suppose J ⊆ [M] = {1, 2, . . . , M} with |J| = t, where t is fixed. Let AJ = j∈J A j .
T
and so
m2
n−kt
→ 0.
2
Then from Lemma 22.1(f),
m−(k−1)t
n−kt
N 1 − O ktn
2 = (1 + o(1))
m − (k − 1)t (m − (k − 1)t)!
N m−(k−1)t 1 − O mkt
n
= (1 + o(1))
(m − (k − 1)t)!
N m−(k−1)t
= (1 + o(1)) .
(m − (k − 1)t)!
Nm
N
= (1 + o(1)) ,
m m!
and so
m! m (k−1)t
P(AJ ) = (1 + o(1)) N −(k−1)t = (1 + o(1)) .
(m − (k − 1)t)! N
Theorem 2.9. If m = 21 cn, where 0 < c < 1 is a constant, then w.h.p. the order of
the largest component of a random graph Gm is O(log n).
The above theorem follows from the next three lemmas stated and proved in
terms of Gn,p with p = c/n, 0 < c < 1. In fact the first of those three lemmas
covers a little bit more than the case of p = c/n, 0 < c < 1.
Proof. Suppose that there is a pair of cycles that are in the same component.
If such a pair exists then there is minimal pair C1 ,C2 , i.e., either C1 and C2 are
connected by a path (or meet at a vertex) or they form a cycle with a diagonal path
(see Figure 2.1). Then in either case, C1 ∪C2 consists of a path P plus another two
distinct edges, one from each endpoint of P joining it to another vertex in P. The
number of such graphs on k labeled vertices can be bounded by k2 k!.
111
000 111
000 000
111
111
000
000
111 000
111 000
111
000
111
000
111 000
111
000
111 00
11 000
111 000
111
000
111 000
111 000
111 11
00 000
111 111
000
111
000 00
11 000
111
000
111
11
00 000
111 00
11 000
111
00
11
00
11 000
111
00
11 000
111 11
00
00
11
00
11
00
11
000
111 111
000
000
111 00
11 111
000
000
111 111
000
000
111 111
000
000
111 11
00
111
000 000
111 11
00 000
111 000
111 000
111 111
000 00
11
00
11
000
111
000
111 000
111 00
11 000
111 000
111 000
111 000
111 00
11
000
111 000
111 00
11 000
111 000
111 000
111 000
111
000
111
000
111
000
111
111
000 11
00
000
111 00
11
00
11
000
111 00
11 111
000
000
111
111
000
000
111 000
111 000
111
000
111 000
111
000
111
000
111 000
111 000
111 00
11
000
111 111
000 111
000 11
00
000
111
000
111 000
111
000
111 00
11
000
111 000
111 00
11
000
111
111
000
111
000
000
111 000
111
000
111 000
111 111
000
000
111
000
111 000
111 111
000 000
111
000
111 000
111 000
111
000
111
000
111 000
111
000
111 11
00
00
11
00
11
00
11
111
000
000
111
000
111
00
11
11
00 000
111
00
11 000
111 11
00
00
11 00
11
00
11
00
11
000
111 000
111
111
000
111
000
000
111 000
111
000
111 000
111 000
111
000
111
111
000 000
111 111
000 111
000
000
111
000
111 000
111 000
111
000
111
000
111 000
111 000
111
000
111 000
111 000
111
Let X be the number of subgraphs of the above kind (shown in Figure 2.1) in the
random graph Gn,p . By the first moment method (see Lemma 21.2),
n
n 2 k+1
P(X > 0) ≤ E X ≤ ∑ k k!p (2.3)
k=4 k
n
nk 2 ω k+1
1
≤ ∑ k k! k+1 1 − 1/3
k=4 k! n n
Z ∞ 2
x ωx
≤ exp − 1/3 dx
0 n n
2
= 3
ω
= o(1).
We remark for later use that if p = c/n, 0 < c < 1 then (2.3) implies
n
P(X > 0) ≤ ∑ k2ck+1n−1 = O(n−1). (2.4)
k=4
vertices in [k]. This bounds the number C(k, k) of connected graphs on [k] with k
edges. This is off by a factor O(k1/2 ) from the exact formula which is given below
for completeness:
k r
k (r − 1)! k−r−1 π k−1/2
C(k, k) = ∑ rk ≈ k . (2.6)
r=3 r 2 8
2.1. SUB-CRITICAL PHASE 31
The remaining factor, other than nk , in (2.5) is the probability that the k edges of
the unicyclic component exist and that there are no other edges on Gn,p incident
with the k chosen vertices.
Noting also that by Lemma 22.1(d),
nk k(k−1)
n
≤ e− 2n .
k k!
Assume next that c < 1 and then we get
nk − k(k−1) k+1 ck −ck+ ck(k−1) + ck
E Xk ≤ e 2n k e 2n 2n (2.7)
k! nk
ek k(k−1) k(k−1) c
≤ k e− 2n kk+1 ck e−ck+ 2n + 2 (2.8)
k
k c
≤ k ce1−c e 2 .
So,
n n k c
E ∑ Xk ≤ ∑k ce1−c e 2 = O(1), (2.9)
k=3 k=3
since ce1−c < 1 for c 6= 1. By the Markov inequality, if ω = ω(n) → ∞, (see
Lemma 21.1) !
n
1
P ∑ Xk ≥ ω = O → 0 as n → ∞,
k=3 ω
and the Lemma follows for c < 1. If c > 1 then we cannot deduce (22.19) from
2
(22.18). If however k = o(n) then this does not matter, since them ek /n = eo(k) .
Now we show in the proof of Theorem 2.14 below that when c > 1 there is w.h.p.
a unique component of size Ω(n) and all other components are of size O(log n).
This enables us to complete the proof of this lemma for c > 1.
After proving the first two lemmas one can easily see that the only remaining
candidate for the largest component of our random graph is an isolated tree.
Lemma 2.12. Let p = nc , where c 6= 1 is a constant, α = c − 1 − log c, and ω =
ω(n) → ∞, ω = o(log log n). Then
(i) w.h.p. there exists an isolated tree of order
1 5
k− = log n − log log n − ω,
α 2
Thus, by the Chebyshev inequality (see Lemma 21.3), we see that for any fixed
ε > 0,
1 2ck2
P (|Xk − E Xk | ≥ ε E Xk ) ≤ 2 + 2 = o(1). (2.13)
ε E Xk ε n
Thus w.h.p. Xk ≥ Aeαω/2 and this completes the proof of (i).
For (ii) we go back to the formula (2.10) and write, for some new constant
A > 0,
2An 1−bck k
≤ cbk e ,
cbk k5/2
k
where cbk = c 1 − 2n .
In the case c < 1 we have cbk e1−bck ≤ ce1−c and cbk ≈ c and so we can write
k
n
3An n ce1−c 3An ∞
X ≤
∑ k c ∑ k5/2
E ≤ 5/2 ∑ e−αk =
k=k+ k=k+ ck+ k=k+
3Ane−αk+ (3A + o(1))α 5/2 e−αω
= = = o(1). (2.14)
5/2
ck+ (1 − e−α ) c(1 − e−α )
If c > 1 then for k ≤ logn n we use cbk e1−bck = e−α−O(1/ log n) and for k > n
log n we use
ck ≥ c/2 and cbk e1−bck ≤ 1 and replace (2.14) by
n n/ log n n
3An 6An 1
∑ E Xk ≤ 5/2 ∑ e−(α+O(1/ log n))k + ∑ = o(1).
k=k+ ck+ k=k+ c k=n/ log n k5/2
Finally, applying Lemmas 2.11 and 2.12 we can prove the following useful
identity: Suppose that x = x(c) is given as
(
c c≤1
x = x(c) = −x −c
.
The solution in (0, 1) to xe = ce c>1
Note that xe−x increases continuously as x increases from 0 to 1 and then de-
creases. This justifies the existence and uniqueness of x.
Lemma 2.13. If c > 0, c 6= 1 is a constant, and x = x(c) is defined above, then
1 ∞ kk−1 k
∑ ce−c = 1.
x k=1 k!
Proof. Let p = nc . Assume first that c < 1 and let X be the total number of vertices
of Gn,p that lie in non-tree components. Let Xk be the number of tree-components
of order k. Then,
n
n= ∑ kXk + X.
k=1
So,
n
n= ∑ k E Xk + E X.
k=1
Now,
34 CHAPTER 2. EVOLUTION
n k+ kk−1 k
n = o(n) + ∑ ce−c
c k=1 k!
n ∞ kk−1 k
= o(n) + ∑ ce−c .
c k=1 k!
ce1−c+cβ1 < 1,
and let β0 = β0 (c) be large enough so that
β log n 1
1−c+o(1) 0
ce < 2.
n
If we choose β1 and β0 as above then it follows that w.h.p. there is no component
of order k ∈ [β0 log n, β1 n].
Our next task is to estimate the number of vertices on small components i.e. those
of size at most β0 log n.
We first estimate the total number of vertices on small tree components, i.e.,
on isolated trees of order at most β0 log n.
1
Assume first that 1 ≤ k ≤ k0 , where k0 = 2α log n, where α is from Lemma 2.12.
It follows from (2.11) that
!
k0
n k0 kk−1 k
E ∑ kXk ≈ ∑ ce−c
k=1 c k=1 k!
n ∞ kk−1 k
≈ ∑ ce−c ,
c k=1 k!
using kk−1 /k! < ek , and ce−c < e−1 for c 6= 1 to extend the summation from k0 to
infinity.
Putting ε = 1/ log n and using (2.13) we see that the probability that any
Xk , 1 ≤ k ≤ k0 , deviates from its mean by more than 1 ± ε is at most
k0
(log n)2 (log n)4
∑ 1/2−o(1) + O n
= o(1),
k=1 n
where the n1/2−o(1) term comes from putting ω ≈ k0 /2 in (2.12). (ω = o(log log n)
in Lemma 2.12. We can however use ω ≈ k0 /2 for k at most k+ /2.)
Thus, if x = x(c), 0 < x < 1 is the unique solution in (0, 1) of the equation xe−x =
ce−c , then w.h.p.,
k0
n ∞ kk−1 k
∑ kXk ≈ c ∑ k! xe−x
k=1 k=1
36 CHAPTER 2. EVOLUTION
nx
= ,
c
by Lemma 2.13.
Now consider k0 < k ≤ β0 log n.
!
β0 log n
n β0 log n 1−c+ck/n k
∑ kXk ≤ ce
c k=k∑
E
k=k0 +1 +1
0
= O n(ce1−c )k0
1/2+o(1)
=O n .
!
β0 log n β0 log n
n k−1 k c k c k(n−k)
E ∑ kYk ≤ ∑ k 1−
k=1 k=1 k 2 n n
β0 log n k
≤ ∑ k ce1−c+ck/n
k=1
= O(1).
Summarising, we have proved so far that w.h.p. there are approximately nx c ver-
tices on components of order k, where 1 ≤ k ≤ β0 log n and all the remaining giant
components are of size at least β1 n.
We complete the proof by showing the uniqueness of the giant component. Let
log n c1
c1 = c − and p1 = .
n n
Define p2 by
1 − p = (1 − p1 )(1 − p2 )
2.2. SUPER-CRITICAL PHASE 37
log n
and note that p2 ≥ n2
. Then, see Section 1.2,
If x1 e−x1 = c1 e−c1 , then x1 ≈ x and so, by our previous analysis, w.h.p., Gn,p1
has no components with number of vertices in the range [β0 log n, β1 n].
Suppose there are components C1 ,C2 , . . . ,Cl with |Ci | > β1 n. Here l ≤ 1/β1 .
Now we add edges of Gn,p2 to Gn,p1 . Then
l 2
(1 − p2 )(β1 n)
P ∃i, j : no Gn,p2 edge joins Ci with C j ≤
2
2
≤ l 2 e−β1 log n
= o(1).
So w.h.p.Gn,p has a unique component with more than β0 log n vertices and it has
≈ 1 − xc n vertices.
We now consider the number of edges in the giant C0 . Now we switch to
G = Gn,m . Suppose that the edges of G are e1 , e2 , . . . , em in random order. We
estimate the probability that e = em = {x, y} is an edge of the giant. Let G1 be the
graph induced by {e1 , e2 , . . . , em−1 }. G1 is distributed as Gn,m−1 and so we know
that w.h.p. G1 has a unique giant C1 and other components are of size O(log n).
So the probability that e is an edge of the giant is o(1) plus the probability that x
or y is a vertex of C1 . Thus,
x x
P e 6∈ C0 | |C1 | ≈ n 1 − = P e ∩C1 = 0/ | |C1 | ≈ n 1 −
c c
|C1 | |C1 | + 1 x 2
= 1− 1− ≈ . (2.16)
n n c
P(ei , e j ⊆ C0 ) =
o(1) + P(e j ∩C2 6= 0/ | ei ∩C2 6= 0)
/ P(ei ∩C2 6= 0)
/
= (1 + o(1)) P(ei ⊆ C0 ) P(e j ⊆ C0 ).
which has probability o(1). We should double this o(1) probability here to account
for switching the roles of i, j.
The Chebyshev inequality can now be used to show that the number of edges
is concentrated as claimed.
We will see later, see Theorem 2.18, that w.h.p. each of the small components
have at most one cycle.
From the above theorem and the results of previous sections we see that, when
m = cn/2 and c passes the critical value equal to 1, the typical structure of a
random graph changes from a scattered collection of small trees and unicyclic
components to a coagulated lump of components (the giant component) that dom-
inates the graph. This short period when the giant component emerges is called
the phase transition. We will look at this fascinating period of the evolution more
closely in Section 2.3.
We know that w.h.p. the giant
2
component of Gn,m , m = cn/2, c > 1 has ≈
1 − xc n vertices and ≈ 1 − xc2 cn
2 edges. So, if we look at the graph H induced
by the vertices outside the giant, then w.h.p. H has ≈ n1 = nx
c vertices and ≈ m1 =
xn1 /2 edges. Thus we should expect H to resemble Gn1 .m1 , which is sub-critical
since x < 1. This can be made precise, but the intuition is clear.
Now increase m further and look on the outside of the giant component. The
giant component subsequently consumes the small components not yet attached
to it. When m is such that m/n → ∞ then unicyclic components disappear and a
random graph Gm achieves the structure described in the next theorem.
Tree-components of order k die out in the reverse order they were born, i.e.,
larger trees are ”swallowed” by the giant earlier than smaller ones.
Cores
Given a positive integer k, the k-core of a graph G = (V, E) is the largest set S ⊆ V
such that the minimum degree δS in the vertex induced subgraph G[S] is at least
k. This is unique because if δS ≥ k and δT ≥ k then δS∪T ≥ k. Cores were first
discussed by Bollobás [134]. It was shown by Łuczak [559] that for k ≥ 3 either
there is no k-core in Gn,p or one of linear size, w.h.p. The precise size and first
occurrence of k-cores for k ≥ 3 was established in Pittel, Spencer and Wormald
[654]. The 2-core, C2 which is the set of vertices that lie on at least one cycle
2.2. SUPER-CRITICAL PHASE 39
behaves differently to the other cores, k ≥ 3. It grows gradually. We will need the
following result in Section 17.2.
Lemma 2.16. Suppose that c > 1 and that x < 1 is the solution to xe−x = ce−c .
Then w.h.p. the 2-core C2 of Gn,p , p = c/n has (1 − x) 1 − xc + o(1) n vertices
2
and 1 − xc + o(1) cn2 edges.
Proof. Fix v ∈ [n]. We estimate P(v ∈ C2 ). Let C1 denote the unique giant compo-
nent of G1 = Gn,p − v. Now G1 is distributed as Gn−1,p and so C1 exists w.h.p. To
be in C2 , either (i) v has two neighbors in C1 or (ii) v has two neighbors in some
other component. Now because all components other than C1 have size O(log n)
w.h.p., we see that
O(log n) c 2
P((ii)) = o(1) + n = o(1).
2 n
Now w.h.p. |C1 | ≈ 1 − xc n and it is independent of the edges incident with v and
so
P((i)) = 1 − P(0 or 1 neighbors in C1 ) =
c |C1 | c |C1 |−1 c
= o(1) + (1 + o(1)) E 1 − 1 − + |C1 | 1 − (2.17)
n n n
−c+x −c+x
= o(1) + 1 − (e + (c − x)e )
x
= o(1) + (1 − x) 1 − ,
c
where the last line follows from the fact that e−c+x = xc . Also, one has to be care-
|C |
ful when estimating something like E 1 − nc 1 . For this we note that Jensen’s
inequality implies that
c |C1 | c E |C1 |
E 1− ≥ 1− = e−c+x+o(1) .
n n
On the other hand, if ng = 1 − xc n,
c |C1 |
E 1− ≤
n
c |C1 |
E 1− |C1 | ≥ (1 − o(1))ng P (|C1 | ≥ (1 − o(1))ng )
n
+ P (|C1 | ≤ (1 − o(1))ng ) = e−c+x+o(1) .
It follows from (2.17) that E(|C2 |) ≈ (1 − x) 1 − xc n. To prove concentration
of |C2 |, we can use the Chebyshev inequality as we did in the proof of Theorem
2.14 to prove concentration for the number of edges in the giant.
40 CHAPTER 2. EVOLUTION
the transition from O(log n) through Θ(n2/3 ) to Ω(n) the “double jump”. They
did not study the regime m = n/2 + o(n). Bollobás [133] opened the detailed
study of this and Łuczak [557] continued this analysis. He established the precise
size of the “scaling window” by removing a logarithmic factor from Bollobás’s
estimates. The component structure of Gn,m for m = n/2 + o(n) is rather compli-
cated and the proofs are technically challenging. We will begin by stating several
results that give a an idea of the component structure in this range, referring the
reader elsewhere for proofs: Chapter 5 of Janson, Łuczak and Ruciński [449]; Al-
dous [16]; Bollobás [133]; Janson [436]; Janson, Knuth, Łuczak and Pittel [453];
Łuczak [557], [558], [562]; Łuczak, Pittel and Wierman [565]. We will finish
with a proof by Nachmias and Peres that when p = 1/n the largest component is
likely to have size of order n2/3 .
The first theorem is a refinement of Lemma 2.10.
Theorem 2.17. Let m = 2n − s, where s = s(n) ≥ 0.
(a) The probability that Gn,m contains a complex component is at most n2 /4s3 .
(b) If n2/3 s n then w.h.p. the largest component is a tree of size asymptotic
n2 s3
to 2s 2 log n .
The next theorem indicates when the phase in which we may have more than
one complex component “ends”, i.e., when a single giant component emerges.
Theorem 2.18. Let m = n2 + s, where s = s(n) ≥ 0. Then the probability that Gn,m
contains more than one complex component is at most 6n2/9 /s1/3 .
For larger s, the next theorem gives a precise estimate of the size of the largest
component for s n2/3 . For s > 0 we let s̄ > 0 be defined by
2s̄ 2s̄ 2s 2s
1− exp = 1+ exp − .
n n n n
n
Theorem 2.19. Let m = 2 + s where s n2/3 . Then with probability at least
1 − 7n2/9 /s1/3 ,
n2/3
L1 − 2(s + s̄)n ≤
n + 2s 5
where L1 is the size of the largest component in Gn,m . In addition, the largest
component is complex and all other components are either trees or unicyclic com-
ponents.
42 CHAPTER 2. EVOLUTION
4s2
3
s
s̄ = s − +O 2 .
3n n
The next theorem gives some information about `-components inside the scal-
ing window m = n/2 + O(n2/3 ). An `-component is one that has ` more edges
than vertices. So trees are (-1)-components.
Theorem 2.20. Let m = 2n +O(n2/3 ) and let r` denote the number of `-components
in Gn,m . For every 0 < δ < 1 there exists Cδ such that if n is sufficiently large,
then with probability at least 1 − δ , ∑`≥3 `r` ≤ Cδ and the number of vertices on
complex components is at most Cδ n2/3 .
One of the difficulties in analysing the phase transition stems from the need
to estimate C(k, `), which is the number of connected graphs with vertex set [k]
and ` edges. We need good estimates for use in first moment calculations. We
have seen the values for C(k, k − 1) (Cayley’s formula) and C(k, k), see (2.6).
For ` > 0, things become more tricky. Wright [758], [759], [760] showed that
Ck,k+` ≈ γ` kk+(3`−1)/2 for ` = o(k1/3 ) where the Wright coefficients γ` satisfy an
explicit recurrence and have been related to Brownian motion, see Aldous [16]
and Spencer [719]. In a breakthrough paper, Bender, Canfield and McKay [75]
gave an asymptotic formula valid for all k. Łuczak [556] in a beautiful argument
simplified a large part of their argument, see Exercise (4.3.6). Bollobás [135]
proved the useful simple estimate Ck,k+` ≤ c`−`/2 kk+(3`−1)/2 for some absolute
constant c > 0. It is difficult to prove tight statements about Gn,m in the phase
transition window without these estimates. Nevertheless, it is possible to see that
the largest component should be of size order n2/3 , using a nice argument from
Nachmias and Peres. They have published a stronger version of this argument in
[623].
Theorem 2.21. Let p = n1 and A be a large constant. Let Z be the size of the
largest component in Gn,p . Then
1 2/3
(i) P Z ≤ n = O(A−1 ),
A
(ii) P Z ≥ An2/3 = O(A−1 ).
Proof. We will prove part (i) of the theorem first. This is a standard application of
the first moment method, see for example Bollobás [135]. Let Xk be the number
2.3. PHASE TRANSITION 43
h i
1 2/3
of tree components of order k and let k ∈ An , An2/3 . Then, see also (2.10),
n k−2 k−1 k
E Xk = k p (1 − p)k(n−k)+(2)−k+1 .
k
But
k 2 /2
(1 − p)k(n−k)+(2)−k+1 ≈ (1 − p)kn−k
= exp{(kn − k2 /2) log(1 − p)}
kn − k2 /2
≈ exp − .
n
k3
n
E Xk ≈ √ exp − 2 . (2.18)
2π k5/2 6n
So if
An2/3
X= ∑ Xk ,
1 2/3
An
then
3
1 A e−x /6
Z
EX ≈ √ dx
2π x= A1 x5/2
4
= √ A3/2 + O(A1/2 ).
3 π
E Xk2 ≤ E Xk + (1 + o(1))(E Xk )2 ,
1
P(X > 0) ≥
(E X)−1 + 1 + o(1)
= 1 − O(A−1 ),
44 CHAPTER 2. EVOLUTION
To prove (ii) we first consider a breadth first search (BFS) starting from, say,
vertex x. We construct a sequence of sets S1 = {x}, S2 , . . ., where
[
Si+1 = {v 6∈ S j : ∃w ∈ Si such that (v, w) ∈ E(Gn,p )}.
j≤i
We have
|Si |
E(|Si+1 | |Si ) ≤ (n − |Si |) 1 − (1 − p)
≤ (n − |Si |)|Si |p
≤ |Si |.
So
E |Si+1 | ≤ E |Si | ≤ · · · ≤ E |S1 | = 1. (2.19)
We prove next that
4
πk = P(Sk 6= 0)
/ ≤ . (2.20)
k
This is clearly true for k ≤ 4 and we obtain (2.20) by induction from
n−1
n−1 i
πk+1 ≤ ∑ p (1 − p)n−1−i (1 − (1 − πk )i ). (2.21)
i=1 i
To explain the above inequality note that we can couple the construction of S1 , S2 , . . . , Sk
with a (branching) process where T1 = {1} and Tk+1 is obtained from Tk as fol-
lows: each Tk independently spawns Bin(n − 1, p) individuals. Note that |Tk |
stochastically dominates |Sk |. This is because in the BFS process, each w ∈ Sk
gives rise to at most Bin(n − 1, p) new vertices. Inequality (2.21) follows, because
Tk+1 6= 0/ implies that at least one of 1’s children give rise to descendants at level
k. Going back to (2.21) we get
where n o
X1 = x : |Cx | ≥ n2/3 and ρx ≤ n1/3 ,
n o
X2 = x : ρx > n1/3 .
So we have
ECmax ≤ 6n2/3
and part (ii) of the theorem follows from the Markov inequality (see Lemma 21.1).
2.4 Exercises
2.4.1 Prove Theorem 2.15.
2.4.2 Show that if p = ω/n where ω = ω(n) → ∞ then w.h.p. Gn,p contains no
unicyclic components. (A component is unicyclic if it contains exactly one
cycle i.e. is a tree plus one extra edge).
2.4.3 Prove Theorem 2.17.
2.4.4 Suppose that m = cn/2 where c > 1 is a constant. Let C1 denote the giant
component of Gn,m , assuming that it exists. Suppose that C1 has n0 ≤ n
vertices and m0 ≤ m edges. Let G1 , G2 be two connected graphs with n0
vertices from [n] and m0 edges. Show that
P(C1 = G1 ) = P(C1 = G2 ).
(I.e. C1 is a uniformly random connected graph with n0 vertices and m0
edges).
2.4.5 Suppose that Z is the length of the cycle in a randomly chosen connected
n
unicyclic graph on vertex set [n]. Show that, where N = 2 ,
nn−2 (N − n + 1)
EZ = .
C(n, n)
2.4.6 Suppose that c < 1. Show that w.h.p. the length of the longest path in Gn,p
log n
, p = nc is ≈ log 1/c .
2.4.7 Suppose that c 6= 1 is constant. Show that w.h.p. the number of edges in the
log n
largest component that is a path in Gn,p , p = nc is ≈ c−log c.
2.4.8 Let Gn,n,p denote the random bipartite graph derived from the complete bi-
partite graph Kn,n where each edge is included independently with probabil-
ity p. Show that if p = c/n where c > 1 is a constant then w.h.p. Gn,n,p has
a unique giant component of size ≈ 2G(c)n where G(c) is as in Theorem
2.14.
2.4. EXERCISES 47
2.4.9 Consider the bipartite random graph Gn,n,p=c/n , with constant c > 1. Define
0 < x < 1 to be the solution to xe−x = ce−c . Prove that w.h.p. the 2-core of
2
Gn,n,p=c/n has ≈ 2(1 − x) 1 − xc n vertices and ≈ c 1 − xc n edges.
2.4.11 Let m = 2n +s, where s = s(n) ≥ 0. Show that if s n2/3 then w.h.p. the ran-
dom graph Gn,m contains exactly one complex component. (A component
C is complex if it contains at least two distinct cycles. In terms of edges, C
is complex iff it contains at last |C| + 1 edges).
2.4.12 Let mk (n) = n(log n + (k − 1) log log n + ω)/(2k), where |ω| → ∞, |ω| =
o(log n). Show that
(
o(1) if ω → −∞
P(Gmk 6⊇ k-vertex-tree-component) = .
1 − o(1) if ω → ∞
2.4.13 Let k ≥ 3 be fixed and let p = nc . Show that if c is sufficiently large, then
w.h.p. the k-core of Gn,p is non-empty.
2.4.14 Let k ≥ 3 be fixed and let p = nc . Show that there exists θ = θ (c, k) > 0
such that w.h.p. all vertex sets S with |S| ≤ θ n contain fewer than k|S|/2
edges. Deduce that w.h.p. either the k-core of Gn,p is empty or it has size at
least θ n.
2.4.15 Suppose that p = nc where c > 1 is a constant. Show that w.h.p. the giant
component of Gn,p is non-planar. (Hint: Assume that c = 1 + ε where ε is
small. Remove a few vertices from the giant so that the girth is large. Now
use Euler’s formula).
2.4.16 Show that if ω = ω(n) → ∞ then w.h.p. Gn,p has at most ω complex com-
ponents.
2.4.17 Suppose that np → ∞ and 3 ≤ k = O(1). Show that Gn,p contains a k-cycle
w.h.p.
2.4.18 Suppose that p = c/n where c > 1 is constant and let β = β (c) be the
smallest root of the equation
1
cβ + (1 − β )ce−cβ = log c(1 − β )(β −1)/β .
2
48 CHAPTER 2. EVOLUTION
2.4.19 Show that if c 6= 1 and xe−x = ce−c where 0 < x < 1 then
(
1 ∞ kk−2 −c k 1 − 2c c < 1.
∑ (ce ) = x x
c 1− 2
c k=1 k! c > 1.
≥k
2.4.20 Let GδN,M denote a graph chosen uniformly at random from the set of graphs
with vertex set [N], M edges and minimum degree at least k. Let Ck denote
the k core of Gn,m (if it exists). Show that conditional on |Ck | = N and
≥k
|E(Ck )| = M that the graph induced by Ck is distributed as GδN,M .
2.5 Notes
Phase transition
The paper by Łuczak, Pittel and Wierman [565] contains a great deal of informa-
tion about the phase transition. In particular, [565] shows that if m = n/2 + λ n2/3
then the probability that Gn,m is planar tends to a limit p(λ ), where p(λ ) → 0 as
λ → ∞. The landmark paper by Janson, Knuth, Łuczak and Pittel [453] gives the
most detailed analysis to date of the events in the scaling window.
Outside of the critical window 2n ± O(n2/3 ) the size of the largest component
is asymptotically determined. Theorem 2.17 describes Gn,m before reaching the
window and on the other hand a unique “giant” component of size ≈ 4s begins to
emerge at around m = n2 + s, for s n2/3 . Ding, Kim, Lubetzky and Peres [256]
give a useful model for the structure of this giant.
Achlioptas processes
Dimitris Achlipotas proposed the following variation on the basic graph process.
Suppose that instead of adding a random edge ei to add to Gi−1 to create Gi ,
one is given a choice of two random edges ei , fi and one chooses one of them
to add. He asked whether it was possible to come up with a choice rule that
would delay the occurrence of some graph property P. As an initial challenge
he asked whether it was possible to delay the production of a giant component
beyond n/2. Bohman and Frieze [117] showed that this was possible by the use
2.5. NOTES 49
of a simple rule. Since that time this has grown into a large area of research. Kang,
Perkins and Spencer [481] have given a more detailed analysis of the “Bohman-
Frieze” process. Bohman and Kravitz [124] and in greater generality Spencer and
Wormald [721] analyse “bounded size algorithms” in respect of avoiding giant
components. Flaxman, Gamarnik and Sorkin [319] consider how to speed up the
occurrence of a giant component. Riordan and Warnke [675] discuss the speed of
transition at a critical point in an Achlioptas process.
The above papers concern component structure. Krivelevich, Loh and Su-
dakov [524] considered rules for avoiding specific subgraphs. Krivelevich, Lubet-
zky and Sudakov [525] discuss rules for speeding up Hamiltonicity.
Graph Minors
Fountoulakis, Kühn and Osthus [325] show that for every ε > 0 there exists Cε
> Cε and p = o(1) then w.h.p. Gn,p contains a complete minor of
such that if np
2
n p
size (1 ± ε) log np . This improves earlier results of Bollobás, Catlin and Erdős
[139] and Krivelevich and Sudakov [530]. Ajtai, Komlós and Szemerédi [10]
showed that if np ≥ 1 + ε and np = o(n1/2 ) then w.h.p. Gn,p contains a toplogical
clique of size almost as large as the maximum degree. If we know that Gn,p is non-
planar w.h.p. then it makes sense to determine its thickness. This is the minimum
number of planar graphs whose union is the whole graph. Cooper [208] showed
that the thickness of Gn,p is strongly related to its arboricity and is asymptotic to
np/2 for a large range of p.
50 CHAPTER 2. EVOLUTION
Chapter 3
Vertex Degrees
In this chapter we study some typical properties of the degree sequence of a ran-
dom graph. We begin by discussing the typical degrees in a sparse random graph
i.e. one with O(n) edges and prove some results on the asymptotic distribution
of degrees. Next we look at the typical values of the minimum and maximum
degrees in dense random graphs. We then describe a simple canonical labelling
algorithm for the graph isomorphism problem on a dense random graph.
E X0 = n(1 − p)n−1 ,
51
52 CHAPTER 3. VERTEX DEGREES
D
distribution. We write Xn → X to say that a random variable Xn converges in
distribution to a random variable X, as n → ∞.
The following theorem shows that the asymptotic distribution of X0 passes
through three phases: it starts in the Normal phase; next when isolated vertices
are close to “dying out”, it moves through a Poisson phase; it finally ends up at
the distribution concentrated at 0.
Theorem 3.1. Let X0 be the random variable counting isolated vertices in a ran-
dom graph Gn,p . Then, as n → ∞,
D
(i) X̃0 = (X0 − E X0 )/(Var X0 )1/2 → N(0, 1),
if n2 p → ∞ and np − log n → −∞,
D
(ii) X0 → Po(e−c ), if np − log n → c, c < ∞,
D
(iii) X0 → 0, if np − log n → ∞.
Proof. For the proof of (i) we refer the reader to Chapter 6 of Janson, Łuczak and
Ruciński [449] (or to [62] and [518]).
To prove (ii) one has to show that if p = p(n) is such that np − log n → c , then
e−ck −e−c
lim P(X0 = k) = e , (3.2)
n→∞ k!
for k = 0, 1, ... . Now,
X0 = ∑ Iv,
v∈V
where (
1 if v is an isolated vertex in Gn,p
Iv =
0 otherwise.
So
E X0 = ∑ E Iv = n(1 − p)n−1
v∈V
= n exp{(n − 1) log(1 − p)}
( )
∞
pk
= n exp −(n − 1) ∑
k=1 k
= n exp −(n − 1)p + O(np2 )
(log n)2
= n exp −(log n + c) + O
n
3.1. DEGREES OF SPARSE RANDOM GRAPHS 53
≈ e−c . (3.3)
The easiest way to show that (3.2) holds is to apply the Method of Moments
(see Chapter 21). Briefly, since the distribution of the random variable X0 is
uniquely determined by its moments, it is enough to show, that either the kth fac-
torial moment E X0 (X0 − 1) · · · (X0 − k + 1) of X0 , or its binomial moment E Xk0 ,
tend to the respective moments of the Poisson distribution, i.e., to either e−ck or
e−ck /k!. We choose the binomial moments, and so let
(n) X0
Bk = E ,
k
then, for every non-negative integer k,
(n)
Bk = ∑ P(Ivi1 = 1, Ivi2 = 1, . . . , Ivik = 1),
1≤i1 <i2 <···<ik ≤n
n k
= (1 − p)k(n−k)+(2) .
k
Hence
e−ck
(n)
lim Bk = ,
n→∞ k!
and part (ii) of the theorem follows by Theorem 21.11, with λ = e−c .
For part (iii), suppose that np = log n + ω where ω → ∞. We repeat the cal-
culation estimating E X0 and replace ≈ e−c in (3.3) by ≤ (1 + o(1))e−ω → 0 and
apply the first moment method.
From the above theorem we immediately see that if np − log n → c then
−c
lim P(X0 = 0) = e−e . (3.4)
n→∞
We next give a more general result describing the asymptotic distribution of
the number Xd = Xn,d , d ≥ 1 of vertices of any fixed degree d in a random graph.
Recall, that the degree of a vertex in Gn,p has the binomial distribution Bin(n−
1, p). Hence,
n−1 d
E Xd = n p (1 − p)n−1−d . (3.5)
d
Therefore, as n → ∞,
0 if p n−(d+1)/d ,
λ1 if p ≈ cn−(d+1)/d , c < ∞,
∞ if p n−(d+1)/d) but
E Xd → (3.6)
pn − log n − d log log n → −∞,
λ2 if pn − log n − d log log n → c, c < ∞,
0 if pn − log n − d log log n → ∞,
54 CHAPTER 3. VERTEX DEGREES
where
cd e−c
λ1 = and λ2 = . (3.7)
d! d!
The asymptotic behavior of the expectation of the random variable Xd suggests
possible asymptotic distributions for Xd , for a given edge probability p.
Proof. The proofs of statements (i) and (v) are straightforward applications of the
first moment method, while the proofs of (ii) and (iv) can be found in Chapter 3
of Bollobás [127] (see also Karoński and Ruciński [489] for estimates of the rate
of convergence). The proof of (iii) can be found in [62].
The next theorem shows the concentration of Xd around its expectation when
in Gn,p the edge probability p = c/n, i.e., when the average vertex degree is c.
Theorem 3.3. Let p = c/n where c is a constant. Let Xd denote the number of
vertices of degree d in Gn,p . Then, for d = O(1), w.h.p.
cd e−c
Xd ≈ n.
d!
Proof. Assume that vertices of Gn,p are labeled 1, 2, . . . , n. We first compute E Xd .
Thus,
E Xd = n P(deg(1) = d) =
n − 1 c d c n−1−d
=n 1−
d n n
3.1. DEGREES OF SPARSE RANDOM GRAPHS 55
2
nd
d c d c 1
=n 1+O exp −(n − 1 − d) +O 2
d! n n n n
d
c e−c
1
=n 1+O .
d! n
We now compute the second moment. For this we need to estimate
P(deg(1) = deg(2) = d)
c n−1−d 2
c n − 2 c d−1
= 1−
n d −1 n n
c n−2−d 2
c n − 2 c d
+ 1− 1−
n d n n
1
= P(deg(1) = d) P(deg(2) = d) 1 + O .
n
The first line here accounts for the case where {1, 2} is an edge and the second
line deals with the case where it is not.
Thus
Var Xd =
n n
=∑ ∑ [P(deg(i) = d, deg( j) = d) − P(deg(1) = d) P(deg(2) = d)]
i=1 j=1
n
1
≤ ∑ O + E Xd ≤ An,
i6= j=1 n
Theorem 3.4. Let ∆(Gn,p ) (δ (Gn,p )) denotes the maximum (minimum) degree of
vertices of Gn,p .
(i) If p = c/n for some constant c > 0 then w.h.p.
log n
∆(Gn,p ) ≈ .
log log n
56 CHAPTER 3. VERTEX DEGREES
log n 1
d log d ≥ · · (log log n − log log log n + o(1))
log log n 1 − 2λ
log n
= (1 + 2λ + O(λ 2 ))(log log n − log log log n + o(1))
log log n
log n
= (log log n + log log log n + o(1)). (3.9)
log log n
Plugging this into (3.8) shows that ∆(Gn,p ) ≤ d− w.h.p.
Now let d = d+ and let Xd be the number of vertices of degree d in Gn,p . Then
n − 1 c d c n−d−1
E(Xd ) = n 1−
d n n
= exp {log n − d log d + O(d)}
log n
= exp log n − (log log n − log log log n + o(1)) + O(d) (3.10)
log log n
→ ∞.
Here (3.10) is obtained by using −λ in place of λ in the argument for (3.9). Now,
for vertices v, w, by the same argument as in the proof of Theorem 3.3, we have
and the Chebyshev inequality implies that Xd > 0 w.h.p. This completes the proof
of (i).
Statement (ii) is an easy consequence of the Chernoff bounds, Corollary 22.7.
Let ε = ω −1/3 . Then
2 np/3 1/3 /3
P(∃v : |deg(v) − np| ≥ εnp) ≤ 2ne−ε = 2n−ω = o(n−1 ).
3.2. DEGREES OF DENSE RANDOM GRAPHS 57
Now
d r d
d n−1 q n−1
= 1+x =
(n − 1)p (n − 1)p
x2 q
r 3 r
q x pq
= exp x − + O 3/2 p+x
(n − 1)p 2(n − 1)p n n−1
r
pq 2
x q
3
x
= exp x + + O 3/2 ,
n − 1 2(n − 1) n
whereas
d 1− d
n − 1 − d 1− n−1
r
p n−1
= 1−x =
(n − 1)q (n − 1)q
x2 p
r 3 r
p x pq
= exp − x + + O 3/2 q−x
(n − 1)q 2(n − 1)q n n−1
58 CHAPTER 3. VERTEX DEGREES
x2 p
r 3
pq x
= exp −x + + O 3/2 ,
n − 1 2(n − 1) n
So
d 1− d
x2
3
d n−1 n−1−d n−1 x
= exp + O 3/2 ,
(n − 1)p (n − 1)q 2(n − 1) n
and lemma follows from (3.11).
The next lemma proves a strengthing of Theorem 3.5.
Lemma 3.7. Let ε = 1/10, and p be constant and q = 1 − p. If
p
d± = (n − 1)p + (1 ± ε) 2(n − 1)pq log n.
then w.h.p.
(i) ∆(Gn,p ) ≤ d+ ,
assuming that p
d ≤ dL = (n − 1)p + (log n)2 (n − 1)pq.
Also, if d > (n − 1)p then
Bd+1 (n − d − 1)p
= <1
Bd (d + 1)q
and so if d ≥ dL ,
E Xd ≤ E XdL ≤ n exp{−Ω(log n)4 }.
It follows that
∆(Gn,p ) ≤ dL w.h.p. (3.13)
Now if Yd = Xd + Xd+1 + · · · + XdL for d = d± then
!2
dL r
n 1 l − (n − 1)p
EYd ≈ ∑ exp − p
l=d 2π pq 2 (n − 1)pq
!2
∞ r
n 1 l − (n − 1)p
≈∑ exp − p (3.14)
l=d 2π pq 2 (n − 1)pq
!2
1 λ − (n − 1)p
r
n
Z ∞
≈ exp − p dλ .
2π pq λ =d 2 (n − 1)pq
and !2
d+ − (n − 1)p
r 1
n
exp − p = n−O(1) .
2π pq 2 (n − 1)pq
p
If d = (n − 1)p + x(n − 1)pq then, from (3.12) we have
!2
1 λ − (n − 1)p
r
n
Z ∞
EYd ≈ exp − p dλ
2π pq λ =d 2 (n − 1)pq
60 CHAPTER 3. VERTEX DEGREES
r
n p ∞ Z
2
= (n − 1)pq e−y /2 dy
2π pq y=x
n 1 −x2 /2
≈√ e
2π x
(
≤ n−2ε(1+ε) d = d+
. (3.15)
≥ n2ε(1−ε) d = d−
since
n−2
ˆ = d1 )
P(d(1) d1
= n−1(1
− p)−1
P(deg(1) = d1 ) d1
3.2. DEGREES OF DENSE RANDOM GRAPHS 61
= 1 + Õ(n−1/2 ).
So, with d = d−
1
P Yd ≤ EYd
2
E(Yd (Yd − 1)) + EYd − (EYd )2
≤
(EYd )2 /4
1
= Õ ε
n
= o(1).
Now
dL
∑ ∑ ˆ = d1 − 1) P(d(2)
P(d(1) ˆ = d2 − 1)
d1 =d− |d2 −d1 |≤10
dL
−1/2 ˆ = d1 − 1) 2 ,
≤ 21(1 + Õ(n )) ∑ P(d(1)
d1 =d−
dL
1
Z ∞
ˆ = d1 − 1) 2 ≈ 2
e−y dy
∑ P(d(1)
d1 =d− 2π pqn y=x
1 ∞Z
−z2 /2
=√ √ e dz
8π pqn z=x 2
1 1 2
≈√ √ n−2(1−ε) ,
8π pqn x 2
62 CHAPTER 3. VERTEX DEGREES
ˆ = d1 2 . Thus
We get a similar bound for ∑ddL1 =d− ∑|d2 −d1 |≤10 P(d(1)
2
P(¬(iii)) = o n2−1−2(1−ε)
= o(1).
XL+1 XL+2 · · · Xn
G∼
= H ⇔ vi → wi is an isomorphism.
Proof. Lemma 3.7 implies that Step 1 succeeds w.h.p. We must now show that
w.h.p. Xi 6= X j for all i 6= j > L. There is a slight problem because the edges from
vi , i > L to v j , j ≤ L are conditioned by the fact that the latter vertices are those of
highest degree.
Now fix i, j and let Ĝ = Gn,p \ {vi , v j }. It follows from Lemma 3.7 that if
i, j > L then w.h.p. the L largest degree vertices of Ĝ and Gn,p coincide. So, w.h.p.,
we can compute Xi , X j with respect to Ĝ to create X̂i , X̂ j , which are independent of
the edges incident with vi , v j . It follows that if i, j > L then X̂i = Xi and X̂ j = X j and
this avoids our conditioning problem. Denote by NĜ (v) the set of the neighbors
of vertex v in graph Ĝ. Then
P(Step 2 fails)
≤ o(1) + P(∃vi , v j : NĜ (vi ) ∩ {v1 , . . . , vL } = NĜ (v j ) ∩ {v1 , . . . , vL })
n
≤ o(1) + (p2 + q2 )L
2
= o(1).
Corollary 3.9. If 0 < p < 1 is constant then w.h.p. Gn,p has a unique automor-
phism, i.e. the identity automorphism.
The chromatic index χ 0 (G) of a graph G is the minimum number of colors that
can be used to color the edges of G so that if two edges share a vertex, then they
have a different color. Vizing’s theorem states that
Also, if there is a unique vertex of maximum degree, then χ 0 (G) = ∆(G). So,
it follows from Theorem 3.5 (ii) that, for constant p, w.h.p. we have χ 0 (Gn,p ) =
∆(Gn,p ).
64 CHAPTER 3. VERTEX DEGREES
3.3 Exercises
3.3.1 Suppose that m = dn/2 where d is constant. Prove that the number of ver-
k e−d
tices of degree k in Gn,m is asymptotically equal to d k! n for any fixed
positive integer k.
3.3.2 Suppose that c > 1 and that x < 1 is the solution to xe−x = ce−c . Show
that if c= O(1) isfixed then w.h.p. the giant component of Gn,p , p = nc has
k e−c k
≈ c k! 1 − xc n vertices of degree k ≥ 1.
3.3.6 Show that if 0 < p < 1 is constant then w.h.p. the minimum degree δ in
Gn,p satisfies
p p
|δ − (n − 1)q − 2(n − 1)pq log n| ≤ ε 2(n − 1)pq log n,
3.3.7 Use the canonical labelling of Theorem 3.8 to show that w.h.p. Gn,1/2 has
exactly one automprphism, the identity automorphism. (An automorphism
of a graph G = (V, E) is a map ϕ : V → V such that {x, y} ∈ E if and only if
{ϕ(x), ϕ(y)} ∈ E.)
3.4 Notes
For the more detailed account of the properties of the degree sequence of Gn,p the
reader is referred to Chapter 3 of Bollobás [135].
Erdős and Rényi [286] and [288] were first to study the asymptotic distri-
bution of the number Xd of vertices of degree d in relation with connectivity
of a random graph. Bollobás [131] continued those investigations and provided
detailed study of the distribution of Xd in Gn,p when 0 < lim inf np(n)/ log n ≤
lim sup np(n)/ log n < ∞. Palka [637] determined certain range of the edge prob-
ability p for which the number of vertices of a given degree of a random graph
3.4. NOTES 65
Gn,p has a Normal distribution. Barbour [59] and Karoński and Ruciński [489]
studied the distribution of Xd using the Stein–Chen approach. A complete answer
to the asymptotic Normality of Xd was given by Barbour, Karoński and Ruciński
[62] (see also Kordecki [518]). Janson [442] extended those results and showed
that random variables counting vertices of given degree are jointly normal, when
p ≈ c/n in Gn,p and m ≈ cn in Gn,m , where c is a constant.
Ivchenko [431] was the first to analyze the asymptotic behavior of the kth-
largest and kth smallest element of the degree sequence of Gn,p . In particular he
analysed the span between the minimum and the maximum degree of sparse Gn,p .
Similar results were obtained independently by Bollobás [129] (see also Palka
[638]). Bollobás [131] answered the question for what values of p(n), Gn,p w.h.p.
has a unique vertex of maximum degree (see Theorem 3.5).
Bollobás [126], for constant p, 0 < p < 1, i.e., when Gn,p is dense, gave √ an es-
timate of the probability that maximum degree does not exceed pn + O( n log n).
A more precise result was proved by Riordan and Selby [672] who showed that
for constant
p p, the probability that the maximum degree ofn Gn,p does not exceed
pn + b np(1 − p), where b is fixed, is equal to (c + o(1)) , for c = c(b) the root
of a certain equation. Surprisingly, c(0) = 0.6102... is greater than 1/2 and c(b)
is independent of p.
McKay and Wormald [591] proved that for a wide range of functions p =
p(n), the distribution of the degree sequence of Gn,p can be approximated by
{(X1 , . . . , Xn )| ∑ Xi is even}, where X1 , . . . , Xn are independent random variables
each having the Binomial distribution Bin(n − 1, p0 ), where p0 is itself a random
variable with a particular truncated normal distribution
66 CHAPTER 3. VERTEX DEGREES
Chapter 4
Connectivity
We first establish, rather precisely, the threshold for connectivity. We then view
this property in terms of the graph process and show that w.h.p. the random graph
becomes connected at precisely the time when the last isolated vertex joins the
giant component. This “hitting time” result is the pre-cursor to several similar
results. After this we deal with k-connectivity.
4.1 Connectivity
The first result of this chapter is from Erdős and Rényi [286].
and use Theorem 1.4 to translate to Gm and then use (1.7) and monotonicity for
cn → ±∞.
Let Xk = Xk,n be the number of components with k vertices in Gn,p and con-
sider the complement of the event that Gn,p is connected. Then
67
68 CHAPTER 4. CONNECTIVITY
n/2
[
= P (Gn,p has a component of order k) =
k=1
n/2
[
P {Xk > 0} .
k=1
n/2
P(X1 > 0) ≤ P(Gn,p is not connected ) ≤ P(X1 > 0) + ∑ P(Xk > 0).
k=2
Now
n/2 n/2 n/2 n/2
n k−2 k−1 k(n−k)
∑ P(Xk > 0) ≤ ∑ E Xk ≤ ∑ k
k p (1 − p) = ∑ uk .
k=2 k=2 k=2 k=2
So
n/2
e−c log n n/2 1+o(1)−k/2
∑ uk ≤ (1 + o(1)) n
+ ∑ n
k=2 k=10
= O no(1)−1 .
It follows that
P(Gn,p is connected ) = P(X1 = 0) + o(1).
4.1. CONNECTIVITY 69
But we already know (see Theorem 3.1) that for p = (log n + c)/n the number
of isolated vertices in Gn,p has an asymptotically Poisson distribution and there-
fore, as in (3.4)
−c
lim P(X1 = 0) = e−e ,
n→∞
and so the theorem follows.
It is possible to tweak the proof of Theorem 4.1 to give a more precise result
stating that a random graph becomes connected exactly at the moment when the
last isolated vertex disappears.
(i) Gm− consists of a giant connected component plus a set V1 of at most 2 log n
isolated vertices,
m− ≤ m∗1 ≤ m∗c ≤ m+ .
E X1 = n(1 − p− )n−1
≈ ne−np−
≈ log n.
Moreover
So,
Var X1 ≤ E X1 + 2(E X1 )2 p− ,
and
Having at least 2 log n isolated vertices is a monotone property and so w.h.p. Gm−
has less then 2 log n isolated vertices.
To show that the rest of Gm is a single connected component we let Xk , 2 ≤
k ≤ n/2 be the number of components with k vertices in G p− . Repeating the
calculations for p− from the proof of Theorem 4.1, we have
!
n/2
o(1)−1
E ∑ Xk = O n .
k=2
Let
E = {∃ component of order 2 ≤ k ≤ n/2}.
4.2. K-CONNECTIVITY 71
Then
√
P(Gm− ∈ E ) ≤ O( n) P(Gn,p− ∈ E )
= o(1),
4.2 k-connectivity
In this section we show that the threshold for the existence of vertices of degree k
is also the threshold for the k-connectivity of a random graph. Recall that a graph
G is k-connected if the removal of at most k − 1 vertices of G does not disconnect
it. In the light of the previous result it should be expected that a random graph
becomes k-connected as soon as the last vertex of degree k − 1 disappears. This
is true and follows from the results of Erdős and Rényi [288]. Here is a weaker
statement.
Proof. Let
log n+(k − 1) log log n + c
p= .
n
We will prove that, in Gn,p , with edge probability p above,
(i) the expected number of vertices of degree at most k − 2 is o(1),
e−c
(ii) the expected number of vertices of degree k − 1 is, approximately (k−1)! .
72 CHAPTER 4. CONNECTIVITY
We have
then
1
P ∃S, T, |S| < k, 2 ≤ |T | ≤ (n − |S|) : A (S, T ) = o(1).
2
This implies that if δ (Gn,p ) ≥ k then Gn,p is k-connected and Theorem 4.3 fol-
lows. |T | ≥ 2 because if T = {v} then v has degree less than k.
We can assume that S is minimal and then N(T ) = S and denote s = |S|, t =
|T |. T is connected, and so it contains a tree with t − 1 edges. Also each vertex of
S is incident with an edge from S to T and so there are at least s edges between S
and T . Thus, if p = (1 + o(1)) logn n then
P(∃S, T ) ≤ o(1)+
k−1 (n−s)/2
n n t−2 t−1 st s
∑ ∑ s t t p s p (1 − p)t(n−s−t)
s=1 t=2
k−1 (n−s)/2 s t
ne
≤ p−1 ∑ ∑ · (te) · p · et p ne · p · e−(n−t)p
s=1 t=2 s
k−1 (n−s)/2
≤ p−1 ∑ ∑ At Bs (4.1)
s=1 t=2
where
Now if 2 ≤ t ≤ log n then A = n−1+o(1) and B = O((log n)2 ). On the other hand,
if t > log n then we can use A ≤ n−1/3 and B ≤ n2 to see that the sum in (4.1) is
o(1).
4.3. EXERCISES 73
4.3 Exercises
4.3.1 Let m = m∗1 be as in Theorem 4.2 and let em = (u, v) where u has degree one.
Let 0 < c < 1 be a positive constant. Show that w.h.p. there is no triangle
containing vertex v.
4.3.2 Let m = m∗1 as in Theorem 4.2 and let em = (u, v) where u has degree one.
Let 0 < c < 1 be a positive constant. Show that w.h.p. the degree of v in Gm
is at least c log n.
4.3.3 Suppose that n log n m ≤ n3/2 and let d = 2m/n. Let Si (v) be the set of
vertices at distance i from vertex v. Show that w.h.p. |Si (v)| ≥ (d/2)i for all
v ∈ [n] and 1 ≤ i ≤ 32 log
log n
d.
4.3.4 Suppose that m n log n and let d = m/n. Using the previous question,
show that w.h.p. there are at least d/2 internally vertex disjoint paths of
length at most 34 log
log n
d between any pair of vertices in Gn,m .
4.3.5 Suppose that m n log n and let d = m/n. Suppose that we randomly color
(log n)2
the edges of Gn,m with q colors where q (log d)2
. Show that w.h.p. there
is a rainbow path between every pair of vertices. (A path is rainbow if each
of its edges has a different color).
4.3.6 Let Ck,k+` denote the number of connected graphs with vertex set [k] and
k + ` edges where ` → ∞ with k and ` = o(k). Use the inequality
n k n
Ck,k+` pk+` (1 − p)(2)−k−`+k(n−k) ≤
k k
4.3.7 Let Gn,n,p be the random bipartite graph with vertex bi-partition V = (A, B),
A = [1, n], B = [n + 1, 2n] in which each of the n2 possible edges appears
independently with probability p. Let p = log n+ω
n , where ω → ∞. Show
that w.h.p. Gn,n,p is connected.
74 CHAPTER 4. CONNECTIVITY
4.4 Notes
Disjoint paths
Being k-connected means that we can find disjoint paths between any two sets
of vertices A = {a1 , a2 , . . . , ak } and B = {b1 , b2 , . . . , bk }. In this statement there
is no control over the endpoints of the paths i.e. we cannot specify a path from
ai to bi for i = 1, 2, . . . , k. Specifying the endpoints leads to the notion of linked-
ness. Broder, Frieze, Suen and Upfal [169] proved that when we are above the
connectivity threshold, we can w.h.p. link any two k-sets by edge disjoint paths,
provided some natural restrictions apply. The result is optimal up to constants.
Broder, Frieze, Suen and Upfal [168] considered the case of vertex disjoint paths.
Frieze and Zhao [367] considered the edge disjoint path version in random regular
graphs.
Rainbow Connection
The rainbow connection rc(G) of a connected graph G is the minimum number
of colors needed to color the edges of G so that there is a rainbow path between
everyppair of vertices. Caro, Lev, Roditty, Tuza and Yuster [177] proved that
p = log n/n is the sharp threshold for the property rc(G) ≤ 2. This was sharp-
ened to a hitting time result by Heckel and Riordan [418]. He and Liang [417] fur-
ther studied the rainbow connection of random graphs. Specifically, they obtain a
threshold for the property rc(G) ≤ d where d is constant. Frieze and Tsourakakis
[366] studied the rainbow connection of G = G(n, p) at the connectivity threshold
p = log n+ω
n where ω → ∞ and ω = o(log n). They showed that w.h.p. rc(G) is
asymptotically equal to max {diam(G), Z1 (G)}, where Z1 is the number of ver-
tices of degree one.
Chapter 5
Small Subgraphs
Graph theory is replete with theorems stating conditions for the existence of a
subgraph H in a larger graph G. For example Turán’s theorem [738] states that a
2
graph with n vertices and more than 1 − 1r n2 edges must contain a copy of Kr+1 .
In this chapter we see instead how many random edges are required to have a
particular fixed size subgraph w.h.p. In addition, we will consider the distribution
of the number of copies.
5.1 Thresholds
In this section we will look for a threshold for the appearance of any fixed graph
H, with vH = |V (H)| vertices and eH = |E(H)| edges. The property that a random
graph contains H as a subgraph is clearly monotone increasing. It is also trans-
parent that ”denser” graphs appear in a random graph ”later” than ”sparser” ones.
More precisely, denote by
eH
d(H) = , (5.1)
vH
the density of a graph H. Notice that 2d(H) is the average vertex degree in H.
We begin with the analysis of the asymptotic behavior of the expected number of
copies of H in the random graph Gn,p .
75
76 CHAPTER 5. SMALL SUBGRAPHS
n
Proof. The complete graph on n vertices Kn contains vH aH distinct copies of H,
where aH is the number of copies of H in KvH . Thus
n
E XH = aH peH ,
vH
aH × aut(H) = vH !.
Theorem 5.2. Let H be a fixed graph with eH > 0. Suppose p = o n−1/d(H) .
Then w.h.p. Gn,p contains no copies of H.
Thus
P(XH > 0) ≤ E XH → 0 as n → ∞.
However, as we will see, this is not always enough for Gn,p to contain a copy of a
given graph H w.h.p. To see this, consider the graph H given in Figure 5.1 below.
5.1. THRESHOLDS 77
000
111
000
111
000
111
000
111
000
111
111111
000000
111111 0111111
000
111
1000000
000000
000000 1
111111 0111111
000000
111111
000000
000000
111111 0
1
0000000
111111
111111
000000 1
0111111
1000000
000000
111111
000000 1
111111 0
1000000
111111
000000
111111
000000
111111 0
0000000
111111
000000
111111 1
0111111
1000000
000000 1
111111
111111 0000000
111111
000000
111111
000000 0
1111111
000000
000
111
111111
000000 0111111
1
0
000000
111111
000000000
111
000
111
111111
000000 11
0111111
000000000
111
000
111
000
11111111111111111
00000000000000
0
1 000
111
000
111
1111111
00000001
000
111
1111111 1111111
0
0000000000
111
0000000
000
111
1111111
0000000
1111111
0
1
0000000
1111111
0
1
0000000000
111
00000000
1111111
0000000
1111111 0
0000000
1111111
1
0000000
1111111 0000000
1111111
1
0
0000000
1111111
1
0000000
1111111
00000001
1111111 0
0000000
1111111
0
1111111
0000000
0000000
1111111
1
1111111
0
0000000
1
1111111
0000000
000000010
1111111
0000000
1111111
0000000
1111111 0
0000000
1111111
1
0
1111111
0000000
0000000
1111111
1
0
1111111
0000000
1111111
0000000
1
1111111
0
0000000
1
000
111
000
111
000
111
000
111
000
111 00000000
11111111
00011111111
111 00000000
00000000
11111111 000
111
00000000
11111111 000
111
000
111
00000000 1111111111
11111111
00000000
11111111 000
111
0000000000
00000000
11111111
000
111 000
111
0000000000
1111111111
000
111
00000000
11111111 0000000000
1111111111
000
111
0000000000
1111111111
000
111
000
111
000
111
000
111
Here vH = 6 and eH = 8. Let p = n−5/7 . Now 1/d(H) = 6/8 > 5/7 and so
E XH ≈ cH n6−8×5/7 → ∞.
E XĤ ≤ n4−6×5/7 → 0,
A graph H is balanced if m(H) = d(H). It is strictly balanced if d(H) > d(K) for
all proper subgraphs K ⊂ H.
Now we are ready to determine the threshold for the existence of a copy of
H in Gn,p . Erdős and Rényi [287] proved this result for balanced graphs. The
threshold for any graph H was first found by Bollobás in [127] and an alternative,
deterministic argument to derive the threshold was presented in [488]. A simple
proof, given here, is due to Ruciński and Vince [686].
Observe that random variables Ii and I j are independent iff Hi and H j are edge
disjoint. In this case Cov(Ii , I j ) = 0 and such terms vanish from the above sum-
mation. Therefore we consider only pairs (Hi , H j ) with Hi ∩ H j = K , for some
graph K with eK > 0. So,
2vH −vK 2eH −eK 2eH
Var XH = O ∑ n p −p
K⊆H,e >0
K
= O n2vH p2eH ∑ n−vK p−eK .
K⊆H,eK >0
can be written as
E(XH )k = ∑ P(IHi1 = 1, IHi2 = 1, . . . , IHik = 1)
i1 ,i2 ,...,ik
= Dk + Dk ,
where the summation is taken over all k-element sequences of distinct indices i j
from {1, 2, . . . ,t}, while Dk and Dk denote the partial sums taken over all (ordered)
k tuples of copies of H which are, respectively, pairwise vertex disjoint (Dk ) and
not all pairwise vertex disjoint (Dk ). Now, observe that
Dk = ∑ P(IHi1 = 1) P(IHi2 = 1) · · · P(IHik = 1)
i1 ,i2 ,...,ik
n
= (aH peH )k
vH , vH , . . . , vH
≈ (E XH )k .
fF = fF 0 + fHik − fK = fF 0 − fK < 0,
which completes the induction and implies that d(F) > m(H).
Let CF be the number of sequences Hi1 , Hi2 , . . . , Hik of k distinct copies of H,
such that
k k
Hi j ∼
[ [
V Hi j = {1, 2, . . . , vF } and = F.
j=1 j=1
5.3 Exercises
5.3.1 Draw a graph which is : (a) balanced but not strictly balanced, (b) unbal-
anced.
5.3.2 Are the small graphs listed below, balanced or unbalanced: (a) a tree, (b) a
cycle, (c) a complete graph, (d) a regular graph, (d) the Petersen graph, (e)
82 CHAPTER 5. SMALL SUBGRAPHS
5.3.3 Determine (directly, not from the statement of Theorem 5.3) thresholds p̂
for Gn,p ⊇ G, for graphs listed in exercise (ii). Do the same for the thresh-
olds of G in Gn,m .
5.3.5 Let F be a graph obtained by taking a union of triangles such that not every
pair of them is vertex-disjoint, Show (by induction) that eF > vF .
fF = a vF + b eF ,
5.3.7 Determine (directly, using exercise (v)) when the random variable counting
the number of copies of a triangle in Gn,p has asymptotically the Poisson
distribution.
5.3.8 Let Xe be the number of isolated edges (edge-components) in Gn,p and let
Prove that
(
0 if p n−2 or ω(n) → ∞
P(Xe > 0) →
1 if p n−2 and ω(n) → ∞.
5.3.9 Determine when the random variable Xe defined in exercise (vii) has asymp-
totically the Poisson distribution.
5.4 Notes
Distributional Questions
In 1982 Barbour [59] adapted the Stein–Chen technique for obtaining estimates
of the rate of convergence to the Poisson and the normal distribution (see Section
21.3 or [60]) to random graphs. The method was next applied by Karoński and
Ruciński [489] to prove the convergence results for semi-induced graph properties
of random graphs.
Barbour, Karoński and Ruciński [62] used the original Stein’s method for nor-
mal approximation to prove a general central limit theorem for the wide class of
decomposable random variables. Their result is illustrated by a variety of appli-
cations to random graphs. For example, one can deduce from it the asymptotic
distribution of the number of k-vertex tree-components in Gn,p , as well as of the
number of vertices of fixed degree d in Gn,p (in fact, Theorem 3.2 is a direct
consequence of the last result).
Barbour, Janson, Karoński and Ruciński [61] studied the number Xk of maxi-
mal complete subgraphs (cliques) of a given fixed size k ≥ 2 in the random graph
Gn,p . They show that if the edge probability p = p(n) is such that the E Xk tends to
a finite constant λ as n → ∞, then Xk tends in distribution to the Poisson random
variable with the expectation λ . When its expectation tends to infinity, Xk con-
verges in distribution to a random variable which is normally distributed. Poisson
convergence was proved using Stein–Chen method, while for the proof of the nor-
mal part, different methods for different ranges of p were used such as the first
projection method or martingale limit theorem (for details of these methods see
Chapter 6 of Janson, Łuczak and Ruciński [449]).
Svante Janson in an a sequence of papers [432],[433], [434], [437] (see also
[450]) developed or accommodated various methods to establish asymptotic nor-
mality of various numerical random graph characteristics. In particular, in [433]
he established the normal convergence by higher semi-invariants of sums of de-
pendent random variables with direct applications to random graphs. In [434] he
proved a functional limit theorem for subgraph count statistics in random graphs
(see also [450]).
In 1997 Janson [432] answered the question posed by Paul Erdős: What is the
length Yn of the first cycle appearing in the random graph process Gm ? He proved
that
1
Z 1
2 /4 √
lim P(Yn = j) = t j−1 et/2+t 1 − t dt, for every j ≥ 3.
n→∞ 2 0
84 CHAPTER 5. SMALL SUBGRAPHS
Spanning Subgraphs
The previous chapter dealt with the existence of small subgraphs of a fixed size.
In this chapter we concern ourselves with the existence of large subgraphs, most
notably perfect matchings and Hamilton Cycles. The celebrated theorems of Hall
and Tutte give necessary and sufficient conditions for a bipartite and arbitrary
graph respectively to contain a perfect matching. Hall’s theorem in particular can
be used to establish that the threshold for having a perfect matching in a random
bipartite graph can be identified with that of having no isolated vertices.
For general graphs we view a perfect matching as half a Hamilton cycle and
prove thresholds for the existence of perfect matchings and Hamilton cycles in a
similar way.
Having dealt with perfect matchings and Hamilton cycles, we turn our atten-
tion to long paths in sparse random graphs, i.e. in those where we expect a linear
number of edges. We then analyse a simple greedy matching algorithm using
differential equations.
We then consider random subgraphs of some fixed graph G, as opposed to
random subgraphs of Kn . We give sufficient conditions for the existence of long
paths and cycles.
We finally consider the existence of arbitrary spanning subgraphs H where we
bound the maximum degree ∆(H).
85
86 CHAPTER 6. SPANNING SUBGRAPHS
random graph.
Bipartite Graphs
Let Gn,n,p be the random bipartite graph with vertex bi-partition V = (A, B), A =
[1, n], B = [n + 1, 2n] in which each of the n2 possible edges appears independently
with probability p. The following theorem was first proved by Erdős and Rényi
[289].
log n+ω
Theorem 6.1. Let ω = ω(n), c > 0 be a constant, and p = n . Then
0
if ω → −∞
−c
lim P(Gn,n,p has a perfect matching) = e−2e if ω → c
n→∞
1 if ω → ∞.
Moreover,
Proof. We will use Hall’s condition for the existence of a perfect matching in a
bipartite graph. It states that a bipartite graph contains a perfect matching if and
only if the following condition is satisfied:
(i) If |S| > |T |+1, we can remove |S|−|T |−1 vertices from |S| – contradiction.
(ii) Suppose ∃w ∈ T such that w has less than 2 neighbors in S. Remove w and
its (unique) neighbor in |S| – contradiction..
6.1. PERFECT MATCHINGS 87
It follows that
Here e(S : T ) denotes the number of edges between S and T , and e(S : T ) cab be
assumed to be at least 2k − 2, because of (b) above.
Suppose now that p = lognn+c for some constant c. Then let Y denote the
number of sets S and T not satisfying the conditions (6.2), (6.3). Then
n/2
n n k(k − 1) 2k−2
EY ≤ 2 ∑ p (1 − p)k(n−k)
k=2 k
k − 1 2k − 2
n/2
ne k−1 ke(log n + c) 2k−2 −npk(1−k/n)
ne k
≤2∑ e
k=2 k k−1 2n
!k
n/2
eO(1) nk/n (log n)2
≤ ∑n
k=2 n
n/2
= ∑ uk .
k=2
Case 1: 2 ≤ k ≤ n3/4 .
To prove the case for |ω| → ∞ we can use monotonicity and (1.7) and the fact that
−2c −2c
e−e → 0 if c → −∞ and e−e → 1 if c → ∞.
Non-Bipartite Graphs
We now consider Gn,p . We could try to replace Hall’s theorem by Tutte’s theorem.
A proof along these lines was given by Erdős and Rényi [290]. We can however
get away with a simpler approach based on simple expansion properties of Gn,p .
The proof here can be traced back to Bollobás and Frieze [146].
log n+cn
Theorem 6.2. Let ω = ω(n), c > 0 be a constant, and let p= n . Then
0
if cn → −∞
lim P(G has a perfect matching) = e−e−c if cn → c
n→∞ n,p
n even
1 if cn → ∞.
Moreover,
Proof. We will for convenience only consider the case where cn = ω → ∞ and
ω = o(log n). If cn → −∞ then there are isolated vertices, w.h.p. and our proof
can easily be modified to handle the case cn → c.
Our combinatorial tool that replaces Tutte’s theorem is the following: We say
that a matching M isolates a vertex v if no edge of M contains v.
For a graph G we let
Let G = (V, E) be a graph without a perfect matching i.e. µ(G) < b|V |/2c. Fix
v ∈ V and suppose that M is a maximum matching that isolates v. Let S0 (v, M) =
{u 6= v : M isolates u}. If u ∈ S0 (v, M) and e = {x, y} ∈ M and f = {u, x} ∈ E
then flipping e, f replaces M by M 0 = M + f − e. Here e is flipped-out. Note that
y ∈ S0 (v, M 0 ).
Now fix a maximum matching M that isolates v and let
S0 (v, M 0 )
[
A(v, M) =
M0
Lemma 6.3. Let G be a graph without a perfect matching and let M be a maximum
matching and v be a vertex isolated by M. Then |NG (A(v, M))| < |A(v, M)|.
Proof. Suppose that x ∈ NG (A(v, M)) and that f = {u, x} ∈ E where u ∈ A(v, M).
Now there exists y such that e = {x, y} ∈ M, else x ∈ S0 (M) ⊆ A(v, M). We claim
that y ∈ A(v, M) and this will prove the lemma. Since then, every neighbor of
A(v, M) is the neighbor via an edge of M.
Suppose that y ∈ / A(v, M). Let M 0 be a maximum matching that (i) isolates u
and (ii) is obtainable from M by a sequence of flips. Now e ∈ M 0 because if e has
been flipped out then either x or y is placed in A(v, M). But then we can do another
flip with M 0 , e and the edge f = {u, x}, placing y ∈ A(v, M), contradiction.
We now change notation and write A(v) in place of A(v, M), understanding that
there is some maximum matching that isolates v. Note that if u ∈ A(v) then there is
some maximum matching that isolates u and so A(u) is well-defined. Furthermore,
it always that case that if v is isolated by some maximum matching and u ∈ A(v)
then µ(G + {u, v}) = µ(G) + 1.
Now let
log n + θ log log n + ω
p=
n
where θ ≥ 0 is a fixed integer and ω → ∞ and ω = o(log log n).
We have introduced θ so that we can use some of the following results for the
Hamilton cycle problem.
We write
Gn,p = Gn,p1 ∪ Gn,p2 ,
where
log n + θ log log n + ω/2
p1 =
n
and
ω
1 − p = (1 − p1 )(1 − p2 ) so that p2 ≈ .
2n
Note that Theorem 4.3 implies:
We consider a process where we add the edges of Gn,p2 one at a time to Gn,p1 .
We want to argue that if the current graph does not have a perfect matching then
there is a good chance that adding such an edge {x, y} will increase the size of a
largest matching. This will happen if y ∈ A(x). If we know that w.h.p. every set S
for which |NGn,p1 (S)| < |S| satisfies |S| ≥ αn for some constant α > 0, then
αn
− i α2
P(y ∈ A(x)) ≥ 2 n ≥ , (6.6)
2
2
90 CHAPTER 6. SPANNING SUBGRAPHS
provided i = O(n).
This is because
the edges we add will be uniformly random and there will be
at least αn2 edges {x, y} where y ∈ A(x). Here given an initial x we can include
edges {x , y } where x ∈ A(x) and y0 ∈ A(x0 ). We have subtracted i to account for
0 0 0
Proof. Let a vertex of graph G1 = Gn,p1 be large if its degree is at least λ = log n
100 ,
and small otherwise. Denote by LARGE and SMALL, the set of large and small
vertices in G1 , respectively.
Claim 2. W.h.p. Gn,p1 does not have a 4-cycle containing a small vertex.
Proof.
n |S| log n
Claim 3. W.h.p. in Gn,p1 for every S ⊆ [n], |S| ≤ 2eM , e(S) < M .
Proof.
n |S| log n
P ∃|S| ≤ and e(S) ≥
2eM M
n/2eM s
n 2 s log n/M
≤ ∑ p1
s=log n/M
s s log n/M
!log n/M s
n/2eM 1+o(1)
ne Me s
≤ ∑
s=log n/M
s 2n
n/2eM s
s −1+log n/M 1+o(1) log n/M
≤ ∑ · (Me )
s=log n/M
n
= o(1).
n
Then if |T | ≤ (θ + 4)|S| we have |S ∪ T | ≤ (θ + 5)|S| ≤ 2eM and
|S ∪ T | 1 2 |S ∪ T | log n
e(S ∪ T ) ≥ − log n = .
θ +5 100 M M
|N(S)|
≥ |N(S1 )| + |N(S2 )| − |N(S1 ) ∩ S2 | − |N(S2 ) ∩ S1 | − |N(S1 ) ∩ N(S2 )|
≥ |N(S1 )| + |N(S2 )| − |S2 | − |N(S2 ) ∩ S1 | − |N(S1 ) ∩ N(S2 )|.
But Claim 1 and Claim 2 and minimum degree at least θ + 1 imply that
We now go back to the proof of Theorem 6.2 for the case c = ω → ∞. Let
the edges of Gn,p2 be { f1 , f2 , . . . , fs } in random order, where s ≈ ωn/4. Let G0 =
Gn,p1 and Gi = Gn,p1 + { f1 , f2 , . . . , fi } for i ≥ 1. It follows from Lemmas 6.3 and
6.4 that with µ(G) as in (6.4), and if µ(Gi ) < n/2 then, assuming Gn,p1 has the
1
expansion claimed in Lemma 6.4, with θ = 0 and α = 10eM ,
α2
P(µ(Gi+1 ) ≥ µ(Gi ) + 1 | f1 , f2 , . . . , fi ) ≥ , (6.8)
2
see (6.6).
It follows that
We have used the notion of dominance, see Section 22.9 in order to use the bino-
mial distribution in the above inequality.
6.2. HAMILTON CYCLES 93
Moreover,
Proof. We will first give a proof of the first statement under the assumption that
cn = ω → ∞ where ω = o(log log n). The proof of the second statement is post-
poned to Section 6.3. Under this assumption, we have δ (Gn,p ) ≥ 2 w.h.p., see
Theorem 4.3. The result for larger p follows by monotonicity.
We now set up the main tool, viz. Pósa’s Lemma. Let P be a path with end
points a, b, as in Figure 6.1. Suppose that b does not have a neighbor outside of P.
P
111
000
000
111
000
111
000
111
000
111
000
111
000
111
000
111
000
111
000
111
000
111
000
111
00
11
00
11
00
11
y00
11
00
11
00
11
000
111
000
111 000
111
000
111 000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
0000
1111
0000
1111 000
1111111
0000 000
111
000
111 0000
1111
000
111 0000
1111000
111
000
111 0000
1111
0000
1111000
111
000
1110000
1111
0000
1111000
111
0000
1111 111
0001111
1110000 111
1111 000 1111
111
0000 1111
0001111
0000111
000 0000
1111000
1110000
1111111
000
000
0000 111
1111 0000000
0000
1111
1111
111
0000
000
000 1111
111 0000
000
111 0000
0000
11110001111
111 0000111
1111 0000111
0001111
1111111000
000
111
0000
1111
0000
1111 0001111
111 000
111
0000 111
1111 000 0000
1111
0001111
111 000
111
0000111
1111000 0000
0000
1111000
000
1110000
0000
1111000
111
000
111
0000 111
1111 0000000
0000
1111 0000
1111
000
111
000 1111
111 0000
000
111 0000
0000
11110001111
111 0000111
1111 0000111
0001111
1111111000
000
111
0000
1111
1111 000
1110000
1111 000
111 0000
1111
000
111 0000
1111000
111 00000000000000
111
0000
0000
1111 0001111
111 111
000
0000 111
1111 000 0000
1111
0001111
111 111
000
0000111
1111000 1111
0000
0000
1111111
000
000
1111111
0000
0000
1111000
111
000
111
0000 111
1111 0000000
0000
1111 0000
1111
000
111
000 1111
111 0000
000
111 0000
0000
11110001111
111 0000111
1111 0000111
0001111
1111111000
000
111
0000
1111
000
111
0000
1111 000
111
000
111 000
111
0000 111
1111 000
111
000 0000
1111
000
111
000
111 000
111
0000111
1111 0000
000
111
000 11110000000
00
11 000
111
00
11
000
111
1111 000
1110000
1111
000
111
0000 111 000
111
111
000 0000
1111
000
111 0000
1111
000
111 0000
000
111 000
1110000
1111
0000111
00
11 000
00
11
000
111 000
000
111 000
1110000
1111
000
111 0000111
1111
000
111 000 1111
0000111
000
111
1111
00000001111
00
11 000
111
00
11
000
111 000
111 000
111 000
111 000
111 00
11 00
11
000
111
a 000
111 000
111 000
111
x000
111 00
11 00
11 b
Notice that the P0 below in Figure 6.2 is a path of the same length as P, ob-
tained by a rotation with vertex a as the fixed endpoint. To be precise, suppose
that P = (a, . . . , x, y, y0 , . . . , b0 , b) and {b, x} is an edge where x is an interior vertex
of P. The path P0 = (a, . . . , x, b, b0 , . . . , y0 , y) is said to be obtained from P by a
rotation.
Now let END = END(P) denote the set of vertices v such that there exists a
path Pv from a to v such that Pv is obtained from P by a sequence of rotations with
vertex a fixed as in Figure 6.3.
94 CHAPTER 6. SPANNING SUBGRAPHS
P’
111
000
000
111
000
111
000
111
000
111
000
111
000
111
000
111
000
111
111
000
000
111
000
111
y 00
11
00
11
00
11
00
11
00
11
00
11
000
111
000111
111 000
111
000
111 000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
1111
0000 0001111
0000 111
000 1111
111
0000
000 111
000 1111
0000111
0001111
0000
1111
0000 111
0000
1111 0000000
1111 111
000 1111
000
111 0000
000
111 111
000
111
1111
0001111
0000
111
0000111
0000000111
1111
0001111
0000
000
000
111
0000
1111 000
111
1110000
1111 000
111 0000
1111
000
111 000
111 0000
1111000
1110000
1111000
111
0000
1111 111
000
1111
0001111
0000 111
0000 000 1111
111
0000
000
1111
111
0000
000 000
111 0000
1111000
1110000
1111111
000
0000 111
1111
1111 0000000
1111 000 1111
111 0000
000
111 0001111
111 0000111
11111110000111
0001111
1111000
000
111
0000
1111
0000 0001111
111 111
000
0000 111
1111 000 0000
1111
000
111 111
000
111
000 0000
1111
0000000
111
0000000
1111
0000000
111
1111 000
111
0000 1110000 0000
1111
000
111
000 1111
111 0001111
111 0000111
1111 0000111
0001111
1111111000
0000
1111 000
111
000
1111
0000
1111
0000 000
111 111
0000
000
1111
111
0000
000 000
111 00000000000111
000
111
000
0000
1111
1111 111
0001111 000
111
0000 111
1111 1111
111
0000
000 000
111 1111
0000000
1110000
1111111
000
0000
000
111
1111 000
1110000
000
111
0000 111 000
000
111
111
000 0000
1111
000
111
000
111 111
000 1111
0000
000
111
111
000 1111111
0001111
0000
0000111
00
11
111
0000111
0001111000
00
11
000
111
0000
1111 000
1110000
1111
000
111 000
111
000
111 0000
1111
000
111
000
111 000
111
000
111 0000
1111 00
11
0000000
1111000
111
00
11
000
111 000
000
111 000
1111111
111
0000
000
000
111 000
111
0000
1111 00
11 111
000
00
11
000
111 000
111 000
111 000
111 000
111 00
11 00
11
000
111
a 000
111 000
111 000
111
x 000
111 00
11 00
11 b
Here the set END consists of all the white vertices on the path drawn below in
Figure 6.4.
Proof. Suppose to the contrary that x, y are the neighbors of v on P and that v, x, y 6∈
END and that v is adjacent to w ∈ END. Consider the path Pw . Let {r,t} be the
neighbors of v on Pw . Now {r,t} = {x, y} because if a rotation deleted {v, y} say
then v or y becomes an endpoint. But then after a further rotation from Pw we see
that x ∈ END or y ∈ END.
6.2. HAMILTON CYCLES 95
000
111 111
000 000
111 00
11 00
11
000
111
000
111 000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
000
111
000
111 000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
1111
0000
0000
1111 1111
0000 000
111
000
111 1111
111
0000
000 1111
0000000
111
000
1110000
1111
0000
1111000
111
000
1110000
1111
0000
1111111
000
1111
0000 0000 111
1111
0000
1111
1111
111
0000
0001111
000 1111
0000
000
111
1111
0000111
0000 1111
0001111111
00001111111
0001111111
000
0000111
000
1111
0000 0000
1111 111
000 0000
1111
000
111 0000
1111111
00000000000000000
111
0000
1111
0000
1111 1111 000
111
0000 111
000 1111
111
0000 1111
0001111 000
111
0000111
0000000
1111
0000
1111000
111
000
1110000
1111
0000
1111111
000
0000
1111 0000
1111
0000
1111
1111
111
0000
000
000 1111
111 0000
000
111 0000
0000
11110001111
1110000111
1111 0000111
0001111
1111111000
000
111
1111
0000
1111 0000
1111 111
000 0000
1111
000
111 0000
1111111
00000000000000000
111
0000
0000
1111
111
000
0000 111
1111 000 0000
1111
0001111
111 111
000
0000111
1111000
1111
0000
0000
1111
111
000
000
111
1111
0000
0000
1111000
111
0000
1111 1111
0000 1111
111
0000
000
000 1111
111 00000001111
1110000111
1111 0000111
0001111
1111111000
0000
1111 0000
1111 000
111
0000 111
1111
111
0000
000
0000
1111
000
111
1111
0000000
111
0000111
1111 00000000000111
000
000
111
000
111
1111
0000 0000
1111 000
111
000 0000
1111 000
111
000
111 0000
1111 000
111
0001111
0000 00
11
111
0001111
0000
000
111
1111
0000
000
111 000
111
111
000
000
1110000
1111 000
111
000
111 1111
000
111 000
111
0000111
0001111
000
111
0000
11110000000111
00
11
00001111111
00
11
000
000
111
000
111 000
111 000
111 000
111 00
11
000
111
a 000
111 000
111 000
111 00
11 b
111
000 000
111 000
111
r 000
111 t 00
11 00
11
000
111
000
111 000
111
000
111 000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
000
111
000111
111 000
111
000
111 000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
1111
0000 0001111
0000 111
000 1111
111
0000
000 1111
0000111
000 1111
0000111
0001111
0000
0000 111
1111
0000
1111 0000000
1111 000 1111
111
000
111 111
0000
000 1111
00000001111
111
000
111 0000111
1111
00000000000111
0001111
1111111
0000
000
111
000
0000
1111 000
1110000
1111
0001111
111 000
111
0000 111
1111 0000
1111
000
111
0000
1111 0000
1111
0001111
111 000
111
0000111
1111 0000
1111000
1110000
1111000
111
000
111
0000
1111 000
111
0000 111
1111 0000 000
111 0000
1111
000
111
000 1111 0000000 0000
1111
0001111
111 000
111
0000111
1111 0000
1111
0000111
0001111
1111111000
0000
1111 000
000
1110000
1111
0000
1111 000
111 0000
000
111
0000
1111
000
111 0000
1111
0000
1111000
111 00000000000000
111
000
111
0000
1111
1111 0001111
111 000
111
0000 111
1111 0000
1111
0001111
111 000
111
0000111
1111 0000
1111000
1110000
1111000
111
0000
1111 000
111
0000 0000000 000
111 0000
1111
000
111
000 0000 0000000
111 1111
0000
1111
0001111
111
000
111
0000111
1111
0000
0000111
1111
0001111000
1111
0000 1111111
0000
1111111 111
000 1111
111
000 1111
0000111
000 00000000000111
000
0000
1111 000
111
0000000 111
0000
1111 000 1111
1111111
0000
000
1111
111
0000
000
1111111
0000
0000000 0000
1111000
1110000
1111111
000
000
111
000
111
0000
1111
000
111
0000 111
1111 0000000
1111
000
111 000
111
000
111
000
111
000
111 0000
1111 000
111
000
111 0000
1111
000
111 000
111
000
111 0000111
1111
000
111
000
111 0000
1111000
1110000111
00
11
1111
00
11
0000000
1111
000
00
11
000
111
00
11
000
111 000
111
000
111 000
1110000
1111
000
111 0000 111
1111
000
111 000
0000
1111 00
11 000
111
00
11
000
111
000
111 000
111
000
111 000
111
000
111 000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
a v w
Corollary 6.7.
|N(END)| < 2|END|.
1
|END| ≥ αn where α = . (6.9)
12eM
We now consider the following algorithm that searches for a Hamilton cycle
in a connected graph G. The probability p1 is above the connectivity threshold
and so Gn,p1 is connected w.h.p. Our algorithm will proceed in stages. At the
beginning of Stage k we will have a path of length k in G and we will try to grow
it by one vertex in order to reach Stage k + 1. In Stage n − 1, our aim is simply
to create a Hamilton cycle, given a Hamilton path. We start the whole procedure
with an arbitrary path of G.
Algorithm Pósa:
(a) Let P be our path at the beginning of Stage k. Let its endpoints be x0 , y0 . If x0
or y0 have neighbors outside P then we can simply extend P to include one of
these neighbors and move to stage k + 1.
96 CHAPTER 6. SPANNING SUBGRAPHS
(b) Failing this, we do a sequence of rotations with x0 as the fixed vertex until one
of two things happens: (i) We produce a path Q with an endpoint y that has a
neighbor outside of Q. In this case we extend Q and proceed to stage k + 1.
(ii) No sequence of rotations leads to Case (i). In this case let END denote
the set of endpoints of the paths produced. If y ∈ END then Py denotes a path
with endpoints x0 , y that is obtained from P by a sequence of rotations.
(c) If we are in Case (bii) then for each y ∈ END we let END(y) denote the set
of vertices z such that there exists a longest path Qz from y to z such that Qz is
obtained from Py by a sequence of rotations with vertex y fixed. Repeating the
argument above in (b) for each y ∈ END, we either extend a path and begin
Stage k + 1 or we go to (d).
(d) Suppose now that we do not reach Stage k + 1 by an extension and that we
have constructed the sets END and END(y) for all y ∈ END. Suppose that
G contains an edge (y, z) where z ∈ END(y). Such an edge would imply
the existence of a cycle C = (z, Qy , z). If this is not a Hamilton cycle then
connectivity implies that there exist u ∈ C and v ∈ / C such that u, v are joined
by an edge. Let w be a neighbor of u on C and let P0 be the path obtained from
C by deleting the edge (u, w). This creates a path of length k + 1 viz. the path
w, P0 , v, and we can move to Stage k + 1.
α2
P(λ (Gi+1 ) ≥ λ (Gi ) + 1 | f1 , f2 , . . . , fi ) ≥ , (6.10)
2
Proof. We prove this theorem by analysing simple properties of Depth First Search
(DFS). This is a well known algorithm for exploring the vertices of a component
of a graph. We can describe the progress of this algorithm using three sets: U is
the set of unexplored vertices that have not yet been reached by the search. D is
the set of dead vertices. These have been fully explored and no longer take part in
the process. A = {a1 , a2 , . . . , ar } is the set of active vertices and they form a path
from a1 to ar . We start the algorithm by choosing a vertex v from which to start
the process. Then we let
We now describe how these sets change during one step of the algorithm.
Step (a) If there is an edge {ar , w} for some w ∈ U then we choose one such w
and extend the path defined by A to include w.
D ← D ∪ {ar }; A ← A \ {ar }; r ← r − 1.
98 CHAPTER 6. SPANNING SUBGRAPHS
Claim 5. Let 0 < α < 1 be a positive constant. If p = c/n and c > α2 log αe then
w.h.p. in Gn,p , every pair of disjoint sets S1 , S2 of size at least αn − 1 are joined
by at least one edge.
Proof. The probability that there exist sets S1 , S2 of size (at least) αn − 1 with no
joining edge is at most
2 !αn−1
n 2 e2+o(1)
(1 − p)(αn−1) ≤ e−cα = o(1).
αn − 1 α2
To complete the proof of the theorem, we apply the above lemma to the ver-
tices S1 , S2 on the two sub-paths P1 , P2 of length 3 log c
c n at each end of P. There
will w.h.p. be an edge joining S1 , S2 , creating the cycle of the claimed length.
Krivelevich and Sudakov [531] used DFS to give simple proofs of good bounds
on the size of the largest component in Gn,p for p = 1+ε n where ε is a small con-
stant. Exercises 6.7.19, 6.7.20 and 6.7.21 elaborate on their results.
Now, conditional on Gn,p1 having minimum degree at least two, the proof of
the statement of Lemma 6.4 goes through without change for θ = 1 i.e. S ⊆
n
[n], |S| ≤ 10000 implies |N(S)| ≥ 2|S|. We can then use use the extension-rotation
argument
that
we used to prove Theorem
6.5(c).
This time we only need to close
n log log n n
O log n cycles and we have Ω log log n edges. Thus (6.11) is replaced by
(G \ {x, y} is the graph obtained from G by deleting the vertices x, y and all
incident edges.)
(B)
We will study this algorithm in the context of the pseudo-graph model Gn,m of
Section 1.3 and apply (1.17) to bring the results back to Gn,m . We will argue next
that if at some stage G has ν vertices and µ edges then G is equally likely to be
any pseudo-graph with these parameters.
We will use the method of deferred decisions, a term coined in Knuth, Mot-
wani and Pittel [507]. In this scenario, we do not expose the edges of the pseudo-
graph until we actually need to. So, as a thought experiment, think that initially
there are m boxes, each containing a uniformly random pair of distinct integers
from [n]. Until the box is opened, the contents are unknown except for their dis-
tribution. Observe that opening box A and observing its contents tells us nothing
more about the contents of box B. This would not be the case if as in Gn,m we
insisted that no two boxes had the same contents.
Remark 6.9. A step of GREEDY involves choosing the first unopened box at
random to expose its contents x, y.
After this, the contents of the remaining boxes will of course remain uniformly
random over V (G)
2 . The algorithm will then ask for each box with x or y to be
opened. Other boxes will remain unopened and all we will learn is that their
contents do not contain x or y and so they are still uniform over the remaining
possible edges.
Lemma 6.10. Suppose that m = cn for some constant c > 0. Then w.h.p. the
(B)
maximum degree in Gn,m is at most log n.
Proof. The degree of a vertex is distributed as Bin(m, 2/n). So, if ∆ denotes the
(B)
maximum degree in Gn,m , then with ` = log n,
`
2ce `
m 2
P(∆ ≥ `) ≤ n ≤n = o(1).
` n `
Now let X(t) = (ν(t), µ(t)),t = 1, 2, . . . , denote the number of vertices and
edges in the graph at the start of the tth iterations of GREEDY. Also, let Gt =
(Vt , Et ) = G at this point and let Gt0 = (Vt , Et \ e) where e is a uniform random
(B)
edge of Et . Thus ν(1) = n, µ(1) = m and G1 = Gn,m . Now ν(t + 1) = ν(t) − 2
6.4. GREEDY MATCHING ALGORITHM 101
and so ν(t) = n − 2t. Let dt (·) denote degree in Gt0 and let θt (x, y) denote the
number of copies of the edge {x, y} in Gt , excluding e. Then we have
E(µ(t + 1) | Gt ) = µ(t) − (dt (x) + dt (y) − 1 + θt (x, y)).
Taking expectations over Gt we have
E(µ(t + 1)) = E(µ(t)) − E(dt (x)) − E(dt (y)) + 1 + E(θt (x, y)).
Now
ν(t)
dt (i)
E(dt (x) | Gt ) = ∑ 2µ(t) dt (i)
i=1
ν(t)
dt (i)
E(dt (y) | Gt ) = Ex ∑ dt (i)
i=1 2µ(t) − 1
i6=x
ν(t)
dt (x)2
dt (i)
=∑ dt (i) − E (6.12)
i=1 2µ(t) − 1 2µ(t) − 1
ν(t)
dt (i)2
1
=∑ +O .
i=1 2µ(t) n − 2t
We will see momentarity that E(dt (x)2 ) = O(1). In the model GBn,m ,
!
ν(t) µ(t)
2 µ(t)−k
2 2 µ(t) 2
E ∑ dt (i) = ν(t) ∑ k 1−
i=1 k=0 k ν(t)k ν(t)
2 2µ(t)
= 2µ(t) 1 − + .
ν(t) ν(t)
So,
4E(µ(t)) 1
E(µ(t + 1)) = E(µ(t)) − −1+O . (6.13)
n − 2t n − 2t
Here we use Remark 6.9 to argue that E θt (x, y)) = O(1/(n − 2t)).
This suggests that w.h.p. µ(t) ≈ nz(t/n) where z(0) = c and z(τ) is the solution
to the differential equation
dz 4z(τ)
=− − 1.
dτ 1 − 2τ
This is easy to solve and gives
1 1 − 2τ
z(τ) = c + (1 − 2τ)2 − .
2 2
c
The smallest root of z(τ) = 0 is τ = 2c+1 . This suggests the following theorem.
102 CHAPTER 6. SPANNING SUBGRAPHS
Theorem 6.11. W.h.p., running GREEDY on Gn,cn finds a matching of size c+o(1)
2c+1 n.
(B)
Proof. We will replace Gn,m by Gn,m and consider the random sequence µ(t),
t = 1, 2, . . .. The number of edges in the matching found by GREEDY equals one
less than the first value of t for which µ(t) = 0. We show that w.h.p. µ(t) > 0 if
and only if t ≤ c+o(1)
2c+1 n. We will use Theorem 23.1 of Chapter 23.
In our set up for the theorem we let
4x
f (τ, x) = − − 1.
1 − 2τ
1 c 1
D = (τ, x) : − < t < TD = ,0 < x < .
n 2c + 1 2
We let X(t) = µ(t) for the statement of the theorem. Then we have to check the
conditions:
(P1) |µ(t)| ≤ cn, ∀t < TD = TD n.
(P2) |µ(t + 1) − µ(t)| ≤ 2 log n, ∀t < TD .
(P3) | E(µ(t + 1) − µ(t)|Ht , E ) − f (t/n, X(t)/n)| ≤ An , ∀t < TD .
Here E = {∆ ≤ log n} and this is needed for (P2).
(P4) f (t, x) is continuous and satisfies a Lipschitz condition
| f (t, x) − f (t 0 , x0 )| ≤ Lk(t, x) − (t 0 , x0 )k∞ where L = 10(2c + 1)2 ,
for (t, x), (t 0 , x0 ) ∈ D ∩ {(t, x) : t ≥ 0}
4x
Here f (t, x) = −1 − 1−2t and we can justify L of P4 as follows:
| f (t, x) − f (t 0 , x0 )|
0
4x 4x
= −
1 − 2t 1 − 2t 0
0) 8x0 (t − t 0 ) 80t(x − x0 )
4(x − x
≤ + +
(1 − 2t)(1 − 2t 0 ) (1 − 2t)(1 − 2t 0 ) (1 − 2t)(1 − 2t 0 )
≤ 10(2c + 1)2 .
Now let β = n1/5 and λ = n−1/20 and σ = T D − 10λ and apply the theorem.
This shows that w.h.p. µ(t) = nz(t/n) + O(n19/20 ) for t ≤ σ n.
The result in Theorem 6.11 is taken from Dyer, Frieze and Pittel [280], where a
central limit theorem is proven for the size of the matching produced by GREEDY.
The use of differential equations to approximate the trajectory of a stochastic
process is quite natural and is often very useful. It is however not always best
practise to try and use an “off the shelf” theorem like Theorem 23.1 in order to
get a best result. It is hard to design a general theorem that can deal optimally
with terms that are o(n).
6.5. RANDOM SUBGRAPHS OF GRAPHS WITH LARGE MINIMUM DEGREE103
Proof. We will assume that G has n vertices. We let T denote the forest produced
by depth first search. We also let D,U, A be as in the proof of Theorem 6.8. Let
v be a vertex of the rooted forest T . There is a unique vertical path from v to the
root of its component. We write A (v) for the set of ancestors of v, i.e., vertices
(excluding v) on this path. We write D(v) for the set of descendants of v, again
excluding v. Thus w ∈ D(v) if and only if v ∈ A (w). The distance d(u, v) between
two vertices u and v on a common vertical path is just their graph distance along
this path. We write Ai (v) and Di (v) for the set of ancestors/descendants of v
at distance exactly i, and A≤i (v), D≤i (v) for those at distance at most i. By the
depth of a vertex we mean its distance from the root. The height of a vertex v
is max {i : Di (v) 6= 0}.
/ Let R denote the set of edges of G that are not tested for
inclusion in G p during the exploration.
Lemma 6.13. Every edge e of R joins two vertices on some vertical path in T .
Proof. Let e = {u, v} and suppose that u is placed in D before v. When u is placed
in D, v cannot be in U, else {u, v} would have been tested. Also, v cannot be in D
by our choice of u. Therefore at this time v ∈ A and there is a vertical path from v
to u.
Lemma 6.14. With high probability, at most 2n/p = o(kn) edges are tested during
the depth first search exploration.
Proof. Each time an edge is tested, the test succeeds (the edge is found to be
present) with probability p. The Chernoff bound implies that the probability that
104 CHAPTER 6. SPANNING SUBGRAPHS
more than 2n/p tests are made but fewer than n succeed is o(1). But every suc-
cessful test contributes an edge to the forest T , so w.h.p. at most n tests are suc-
cessful.
From now on let us fix an arbitrary (small) constant 0 < ε < 1/10. We call a
vertex v full if it is incident with at least (1 − ε)k edges in R.
Lemma 6.15. With high probability, all but o(n) vertices of Tk are full.
Proof. For each rich vertex v, let P(v) be a set of dεke descendants of v, obtained
by choosing vertices of D(v) one-by-one starting with those furthest from v. For
every w ∈ P(v) we have D(w) ⊆ P(v), so |D(w)| < εk, i.e., w is poor. Consider
the set S1 of ordered pairs (v, w) with v rich and w ∈ P(v). Each of the n − o(n)
rich vertices appears in at least εk pairs, so |S1 | ≥ (1 − o(1))εkn.
For any vertex w we have |A≤i (w)| ≤ i, since there is only one ancestor at
each distance, until we hit the root. Since (v, w) ∈ S1 implies that w is poor and
v ∈ A (w), and there are only o(n) poor vertices, at most o(Ckn) = o(kn) pairs
(v, w) ∈ S1 satisfy d(v, w) ≤ Ck. Thus S10 = {(v, w) ∈ S1 : d(v, w) > Ck} satisfies
|S10 | ≥ (1 − o(1))εkn. Since each vertex v is the first vertex of at most dεke ≈ εk
pairs in S1 ⊇ S10 , it follows that n − o(n) vertices v appear in pairs (v, w) ∈ S10 .
Since any such v has height at least Ck, the proof is complete.
Let us call a vertex v light if |D ≤(1−5ε)k (v)| ≤ (1 − 4ε)k, and heavy otherwise.
Let H denote the set of heavy vertices in T .
Lemma 6.17. Suppose that T = Tk contains o(n) poor vertices, and let X ⊆ V (T )
with |X| = o(n). Then, for k large enough, T contains a vertical path P of length
at least ε −2 k containing at most ε 2 k vertices in X ∪ H.
Proof. Let S2 be the set of pairs (u, v) where u is an ancestor of v and 0 < d(u, v) ≤
(1 − 5ε)k. Since a vertex has at most one ancestor at any given distance, we have
|S2 | ≤ (1 − 5ε)kn. On the other hand, by Lemma 6.16 all but o(n) vertices u are
at height at least k and so appear in at least (1 − 5ε)k pairs (u, v) ∈ S2 . It follows
that only o(n) vertices u are in more than (1 − 4ε)k such pairs, i.e., |H| = o(n).
6.5. RANDOM SUBGRAPHS OF GRAPHS WITH LARGE MINIMUM DEGREE105
for some vertex v. Then, since εkp → ∞, testing the relevant edges {u, v} one-by-
one, w.h.p we find one present in G p , forming, together with T , the required long
cycle. On the other hand, suppose that (6.14) fails for every v. Suppose that some
vertex v is full but poor. Since v has at most εk descendants, there are at least
(1 − 2ε)k pairs {u, v} ∈ R with u ∈ A (v). Since v has only one ancestor at each
distance, it follows that (6.14) holds for v, a contradiction.
We have shown that we can assume that no poor vertex is full. Hence there
are o(n) poor vertices, and we may apply Lemma 6.17, with X the set of vertices
that are not full. Let P be the path whose existence is guaranteed by the lemma,
and let Z be the set of vertices on P that are full and light, so |V (P) \ Z| ≤ ε 2 k. For
any v ∈ Z, since v is full, there are at least (1 − ε)k vertices u ∈ A (v) ∪ D(v) with
{u, v} ∈ R. Since (6.14) does not hold, at least (1 − 2ε)k of these vertices satisfy
d(u, v) ≤ (1−5ε)k. Since v is light, in turn at least 2εk of these u must be in A (v).
Recalling that a vertex has at most one ancestor at each distance, we find a set R(v)
of at least εk vertices u ∈ A (v) with {u, v} ∈ R and εk ≤ d(u, v) ≤ (1 − 5ε)k ≤ k.
It is now easy to find a (very) long cycle w.h.p. Recall that Z ⊆ V (P) with
|V (P) \ Z| ≤ ε 2 k. Thinking of P as oriented upwards towards the root, let v0 be the
lowest vertex in Z. Since |R(v0 )| ≥ εk and kp → ∞, w.h.p. there is an edge {u0 , v0 }
in G p with u0 ∈ R(v0 ). Let v1 be the first vertex below u0 along P with v1 ∈ Z. Note
that we go up at least εk steps from v0 to u0 and down at most 1+|V (P)\Z| ≤ 2ε 2 k
from u0 to v1 , so v1 is above v0 . Again w.h.p. there is an edge {u1 , v1 } in G p with
u1 ∈ R(v1 ), and so at least εk steps above v1 . Continue downwards from u1 to the
first v2 ∈ Z, and so on. Since ε −1 = O(1), w.h.p. we may continue in this way
106 CHAPTER 6. SPANNING SUBGRAPHS
within P as each upwards step has length at most k.) These chords combine with
P to give a cycle of length at least (1 − 2ε −1 × 2ε 2 )k = (1 − 4ε)k, as shown in
Figure 6.6.
11
00 11
00 11
00 11
00 11
00 11
00 11
00 11
00 11
00 11
00
00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11
v0 v1 u0 v2 u1 v3 u2 v4 u3 u4
Figure 6.6: The path P, with the root off to the right. Each chord {vi , ui } has
length at least εk (and at most k); from ui to vi+1 is at most 2ε 2 k steps back along
P. The chords and the thick part of P form a cycle.
10 logbn/(∆2 + 1)c
p∆ > , (6.15)
bn/(∆2 + 1)c
Proof. To prove this we first apply the Hajnal-Szemerédi Theorem to the square
H 2 of our graph H.
Recall that we square a graph if we add an edge between any two vertices of our
original graph which are at distance at most two. The Hajnal-Szemerédi Theorem
states that every graph with n vertices and maximum vertex degree at most d is
d + 1-colorable with all color classes of size bn/(d + 1)c or dn/(d + 1)e, i.e, the
(d + 1)-coloring is equitable.
Since the maximum degree of H 2 is at most ∆2 , there exists an equitable ∆2 + 1-
coloring of H 2 which induces a partition of the vertex set of H, say U = U(H),
6.6. SPANNING SUBGRAPHS 107
log n
Corollary 6.20. Let n = d 2 and p = ω(n)
n1/4
, where ω(n), d → ∞. Then w.h.p.
Gn,p contains a copy of the 2-dimensional lattice Ld .
6.7 Exercises
6.7.1 Consider the bipartite graph process Γm , m = 0, 1, 2, . . . , n2 where we add
the n2 edges in A × B in random order, one by one. Show that w.h.p. the
hitting time for Γm to have a perfect matching is identical with the hitting
time for minimum degree at least one.
6.7.3 Show that if p = log n+(k−1)nlog log n+ω where k = O(1) and ω → ∞ then w.h.p.
Gn,n,p contains a k-regular spanning subgraph.
6.7.4 Consider the random bipartite graph G with bi-partition A, B where |A| =
|B| = n. Each vertex a ∈ A independently chooses d2 log ne random neigh-
bors in B. Show that w.h.p. G contains a perfect matching.
6.7.5 Show that if p = log n+(k−1)nlog log n+ω where k = O(1) and ω → ∞ then w.h.p.
Gn,p contains bk/2c edge disjoint Hamilton cycles. If k is odd, show that
in addition there is an edge disjoint matching of size bn/2c. (Hint: Use
Lemma 6.4 to argue that after “peeling off” a few Hamilton cycles, we can
still use the arguments of Sections 6.1, 6.2).
6.7.6 Let m∗k denote the first time that Gm has minimum degree at least k. Show
that w.h.p. in the graph process (i) Gm∗1 contains a perfect matching and (ii)
Gm∗2 contains a Hamilton cycle.
6.7.7 Show that if p = log n+logn log n+ω where ω → ∞ then w.h.p.Gn,n,p contains
a Hamilton cycle. (Hint: Start with a 2-regular spanning subgraph from
(ii). Delete an edge from a cycle. Argue that rotations will always produce
paths beginning and ending at different sides of the partition. Proceed more
or less as in Section 6.2).
6.7. EXERCISES 109
6.7.8 Show that if p = log n+logn log n+ω where n is even and ω → ∞ then w.h.p. Gn,p
contains a pair of vertex disjoint n/2-cycles. (Hint: Randomly partition [n]
into two sets of size n/2. Then move some vertices between parts to make
the minimum degree at least two in both parts).
6.7.9 Show that if three divides n and np2 log n then w.h.p. Gn,p contains n/3
vertex disjoint triangles. (Hint: Randomly partition [n] into three sets A, B,C
of size n/3. Choose a perfect matching M between A and B and then match
C into M).
6.7.10 Let G = (X,Y, E) be an arbitrary bipartite graph where the bi-partition X,Y
satisfies |X| = |Y | = n. Suppose that G has minimum degree at least 3n/4.
Let p = K log
n
n
where K is a large constant. Show that w.h.p. G p contains a
perfect matching.
6.7.11 Let p = (1+ε) logn n for some fixed ε > 0. Prove that w.h.p. Gn,p is Hamilton
connected i.e. every pair of vertices are the endpoints of a Hamilton path.
6.7.12 Show that if p = (1+ε)n log n for ε > 0 constant, then w.h.p. Gn,p contains a
copy of a caterpillar on n vertices. The diagram below is the case n = 16.
6.7.13 Show that for any fixed ε > 0 there exists cε such that if c ≥ cε then Gn,p
2
contains a cycle of length (1 − ε)n with probability 1 − e−cε n/10 .
6.7.14 Let p = (1+ε) logn n for some fixed ε > 0. Prove that w.h.p. Gn,p is pancyclic
i.e. it contains a cycle of length k for every 3 ≤ k ≤ n.
(See Cooper and Frieze [213] and Cooper [207], [209]).
6.7.16 Let T be a tree on n vertices and maximum degree less than c1 log n. Sup-
pose that T has at least c2 n leaves. Show that there exists K = K(c1 , c2 )
such that if p ≥ K log
n
n
then Gn,p contains a copy of T w.h.p.
110 CHAPTER 6. SPANNING SUBGRAPHS
6.7.17 Let p = 1000n and G = Gn,p . Show that w.h.p. any red-blue coloring of the
n
edges of G contains a mono-chromatic path of length 1000 . (Hint: Apply the
argument of Section 6.3 to both the red and blue sub-graphs of G to show
that if there is no long monochromatic path then there is a pair of large sets
S, T such that no edge joins S, T .)
This question is taken from Dudek and Pralat [270]
6.7.18 Suppose that p = n−α for some constant α > 0. Show that if α > 13 then
w.h.p. Gn,p does not contain a maximal spanning planar subgraph i.e. a
planar subgraph with 3n − 6 edges. Show that if α < 31 then it contains one
w.h.p. (see Bollobás and Frieze [147]).
6.7.19 Show that the hitting time for the existence of k edge-disjoint spanning trees
coincides w.h.p. with the hitting time for minimum degree k, for k = O(1).
(See Palmer and Spencer [639]).
6.7.20 Let p = nc where c > 1 is constant. Consider the greedy algorithm for con-
structing a large independent set I: choose a random vertex v and put v into
I. Then delete v and all of its neighbors. Repeat until there are no vertices
left. Use the differential equation method (see Section 6.4) and show that
w.h.p. this algorithm chooses an independent set of size at least logc c n.
6.7.21 Consider the modified greedy matching algorithm where you first choose
a random vertex x and then choose a random edge {x, y} incident with x.
Show that m = cn, that w.h.p. it produces a matching
applied to Gn,m , with
−2c)
of size 1
2 + o(1) − log(2−e 4c n.
6.8 Notes
Hamilton cycles
Multiple Hamilton cycles
There are several results pertaining to the number of distinct Hamilton cycles
in Gn,m . Cooper and Frieze [212] showed that in the graph process Gm∗2 con-
tains (log n)n−o(n) distinct Hamilton cycles w.h.p. This number
was improved by
n
Glebov and Krivelevich [384] to n!pn eo(n) for Gn,p and loge n eo(n) at time m∗2 .
McDiarmid [584] showed that for Hamilton cycles, perfect matchings, spanning
trees the expected number was much higher. This comes from the fact that al-
though there is a small probability that m∗2 is of order n2 , most of the expectation
comes from here. (m∗k is defined in Exercise 6.7.5).
Bollobás and Frieze [146] (see Exercise 6.7.4) showed that in the graph pro-
cess, Gm∗k contains bk/2c edge disjoint Hamilton cycles plus another edge disjoint
matching of size bn/2c if n is odd. We call this property Ak . This was the case
k = O(1). The more difficult case of the occurrence of Ak at m∗k , where k → ∞
was verified in two papers, Krivelevich and Samotij [528] and Knox, Kühn and
Osthus [508].
Suppose that instead of taking enough edges to make the minimum degree in Gn,m
two very likely, we instead condition on having minimum degree at least two.
≥k denote G
Let Gδn,m n,m conditioned on having minimum degree at least k = O(1).
Bollobás, Fenner and Frieze [144] proved that if
n log n
m= + k log log n + ω(n)
2 k+1
≥k has A w.h.p.
then Gδn,m k
≥k has prop-
Bollobás, Cooper, Fenner and Frieze [141] prove that w.h.p. Gδn,cn
erty Ak−1 w.h.p. provided 3 ≤ k = O(1) and c ≥ (k + 1)3 . For k = 3, Frieze [342]
≥3 is Hamiltonian w.h.p. for c ≥ 10.
showed that Gδn,cn
δ ≥k
The k-core of a random graphs is distributed like Gν,µ for some (random)
ν, µ. Krivelevich, Lubetzky and Sudakov [527] prove that when a k-core first
appears, k ≥ 15, w.h.p. it has b(k − 3)/2c edge disjoint Hamilton cycles.
112 CHAPTER 6. SPANNING SUBGRAPHS
Long cycles
A sequence of improvements, Bollobás [130]; Bollobás, Fenner and Frieze [145]
to Theorem 6.8 in the sense of replacing O(log c/c) by something smaller led
finally to Frieze [335]. He showed that w.h.p. there is a cycle of length n(1 −
ce−c (1 + εc )) where εc → 0 with c. Up to the value of εc this is best possible.
Glebov, Naves and Sudakov [385] prove the following generalisation of (part
of) Theorem 6.5. They prove that if a graph G has minimum degree at least k and
k+ωk (1)
p ≥ log k+log log
k then w.h.p. G p has a cycle of length at least k + 1.
Spanning Subgraphs
Riordan [671] used a second moment calculation to prove the existence of a cer-
tain (sequence of) spanning subgraphs H = H (i) in Gn,p . Suppose that we denote
the number of vertices in a graph H by |H| and the number of edges by e(H).
Suppose that |H| = n. For k ∈ [n] we let eH (k) = max{e(F) : F ⊆ H, |F| = k} and
H (k)
γ = max3≤k≤n ek−2 . Riordan proved that if the following conditions hold, then
Gn,p contains a copy of H w.h.p.: (i) e(H) ≥ n, (ii) N p, (1 − p)n1/2 → ∞, (iii)
npγ /∆(H)4 → ∞.
This for example replaces the 12 in Corollary 6.19 by 14 .
Spanning trees
Gao, Pérez-Giménez and Sato [373] considered the existence of k edge disjoint
spanning trees in Gn,p . Using a characterisation
of Nash-Williams [626] they were
m
able to show that w.h.p. one can find min δ , n−1 edge disjoint spanning trees.
Here δ denotes the minimum degree and m denotes the number of edges.
When it comes to spanning trees of a fixed structure, Kahn conjectured that
the threshold for the existence of any fixed bounded degree tree T , in terms of
number of edges, is O(n log n). For example, a comb consists of a path P of length
n1/2 with each v ∈ P being one endpoint of a path Pv of the same length. The
paths Pv , Pw being vertex disjoint for v 6= w. Hefetz, Krivelevich and Szabó [420]
6.8. NOTES 113
proved this for a restricted class of trees i.e. those with a linear number of leaves
or with an induced path of length Ω(n). Kahn, Lubetzky and Wormald [472],
[473] verified the conjecture for combs. Montgomery [605], [606] sharpened the
result for combs, replacing m = Cn log n by m = (1 + ε)n log n and proved that
any tree can be found w.h.p. when m = O(∆n(log n)5 ), where ∆ is the maximum
degree of T . More recently, Montgomery [608] improved the upper bound on m
to the optimal, m = O(∆n(log n)).
Large Matchings
Karp and Sipser [494] analysed a greedy algorithm for finding a large matching
in the random graph Gn,p , p = c/n where c > 0 is a constant. It has a much better
performance than the algorithm described in Section 6.4. It follows from their
work that if µ(G) denotes the size of the largest matching in G then w.h.p.
µ(Gn,p ) γ ∗ + γ∗ + γ ∗ γ∗
≈ 1−
n 2c
where γ∗ is the smallest root of x = c exp {−ce−x } and γ ∗ = ce−γ∗ .
Later, Aronson, Frieze and Pittel [40] tightened their analysis. This led to
δ ≥2 . Frieze and Pittel
the consideration of the size of the largest matching in Gn,m=cn
[362] showed that w.h.p. this graph contains a matching of size n/2 − Z where
Z is a random variable with bounded expectation. Frieze [340] proved that in
the bipartite analogue of this problem, a perfect matching exists w.h.p. Building
on this work, Chebolu, Frieze and Melsted [182] showed how to find an exact
maximum sized matching in Gn,m , m = cn in O(n) expected time.
H-factors
By an H-factor of a graph G, we mean a collection of vertex disjoint copies of
a fixed graph H that together cover all the vertices of G. Some early results on
the existence of H-factors in random graphs are given in Alon and Yuster [33]
and Ruciński [687]. For the case of when H is a tree, Łuczak and Ruciński [567]
found the precise threshold. For general H, there is a recent breakthrough paper
of Johansson, Kahn and Vu [468] that gives the threshold for strictly balanced H
and good estimates in general. See Gerke and McDowell [372] for some further
results.
114 CHAPTER 6. SPANNING SUBGRAPHS
Chapter 7
Extreme Characteristics
This chapter is devoted to the extremes of certain graph parameters. We look first
at the diameter of random graphs i.e. the extreme value of the shortest distance
between a pair of vertices. Then we look at the size of the largest independent set
and the the related value of the chromatic number. We decribe an important recent
result on “interpolation” that proves certain limits exist. We end the chapter with
the likely values of the first and second eigenvalues of a random graph.
7.1 Diameter
In this section we will first discuss the threshold for Gn,p to have diameter d,
when d ≥ 2 is a constant. The diameter of a connected graph G is the maximum
over distinct vertices v, w of dist(v, w) where dist(v, w) is the minimum number
of edges in a path from v to w. The theorem below was proved independently by
Burtin [171], [172] and by Bollobás [128]. The proof we give is due to Spencer
[718].
Theorem 7.1. Let d ≥ 2 be a fixed positive integer. Suppose that c > 0 and
Then (
e−c/2 if k = d
lim P(diam(Gn,p ) = k) =
n→∞ 1 − e−c/2 if k = d + 1.
115
116 CHAPTER 7. EXTREME CHARACTERISTICS
111
000 000
111
Y 00
11 00
11
000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
000
111 000
111 00
11 00
11
000111
111
0000
1111 000
111
0001111
0000
1111 00
11
1111
0000000
111 00
11
0000
1111000
111
0000
1111 000
111
0000 111
1111 0000 0000
1111000
111
0000111
1111 0000
1111
0000111
0001111
1111111000
0000
1111 000
000
1110000
1111
0000
1111 0000
11110000000000
111
000
111
1111
0000 0001111
111
0000 111
1111 0000
1111 1111
0000111
000
0000111
1111 1111
0000
0001111000
111
0000111
0000 000
1111 000
1110000 0000111
1111 0000000
0001111
1111111
1111
0000
0000
1111 0001111
1110000
0000
1111 1111
0000
0000
1111000
000
1110000
0000
1111
000
111
000
111
0000 111
1111
0000
1111
000
000
1110000
1111
0000
1111 0000111
1111
0000
11110000000111
0001111
1111111
0000
000
000
111
0000
1111 000
111
000
1110000
1111
0000
1111 0000
1111000
1110000
1111000
111
000
111
000
111
0000
1111 000
111
000
1110000
1111 000
111
0000
1111 00
11
000
1110000
1111 00
11
000
111
000
111 000
111
0000 111
1111 000
000
111 000
111
000
111
0000
11110000000111
00
11
0000111
1111 1111
00
11
000
00
11
000
111
00
11
000
111
000
111 000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
v X w
7.1. DIAMETER 117
So
P(∃v, w : dist(v, w) > d + 1) = o(n−1 ).
We now consider the probability that d or d + 1 is the diameter. We will use
Janson’s inequality, see Section 22.6. More precisely, we will use the earlier in-
equality, Corollary 22.14, from Janson, Łuczak and Ruciński [448].
We will first use this to estimate the probability of the following event: Let
v 6= w ∈ [n] and let
Let
Z = ∑ Zx ,
x
where (
1 if Bv,x,w occurs
Zx =
0 otherwise.
Janson’s inequality allows us to estimate the probability that Z = 0, which is pre-
cisely the probability of Av,w .
Now
2
d n 1
µ = E Z = (n − 2)(n − 3) · · · (n − d)p = log 1+O .
c n
∆= ∑ P(Bx ∩ By )
x,y:x6=y
v,x,w and v,y,w
share an edge
118 CHAPTER 7. EXTREME CHARACTERISTICS
d−1
d 2(d−1)−t 2d−t
≤ ∑ n p , t is the number of shared edges,
t=1 t
!
d−1 2d−t
2(d−1)−t− d−1
d (2d−t)
=O ∑n (log n) d
t=1
!
d−1
=O ∑ n−t/d+o(1)
t=1
= o(1).
i=1
and re-define
Z = ∑ Zx ,
x
7.1. DIAMETER 119
where now (
1 if Bx occurs
Zx =
0 otherwise.
then w.h.p.
2n
|N≤k0 | ≤ ∑ ((1 + γ)np)k ≤ 2((1 + γ)np)k0 = 3 + o(1)
k≤k0
log n
and so the diameter of Gn,p is at least (1 − o(1)) log np .
We can assume that np = no(1) as larger p are dealt with in Theorem 7.1. Now
fix v, w ∈ [n] and let Ni be as in the previous paragraph. Now consider a Breadth
First Search (BFS) that constructs N1 , N2 , . . . , Nk1 where
3 log n
k1 = .
5 log np
It follows that if (7.5) holds then for k ≤ k1 we have
Observe now that the edges from Ni to [n] \ N≤i are unconditioned by the BFS up
to layer k and so for x ∈ [n] \ N≤k ,
The events x ∈ Nk+1 are independent and so |Nk+1 | stochastically dominates the bi-
nomial Bin(n − n3/4 , ρk ). Assume inductively that |Nk | ≥ (1 − γ)k (np)k for some
k ≥ 1. This is true w.h.p. for k = 1 by (7.5). Let Ak be the event that (7.6) holds.
It follows that
E(|Nk+1 | | Ak ) ≥ np|Nk |(1 − O(n−1/5 )).
It then follows from the Chernoff bounds (Theorem 22.6) that
2
γ
P(|Nk+1 | ≤ ((1 − γ)np)k+1
≤ exp − |Nk |np = o(n−anyconstant ).
4
There is a small point to be made about conditioning here. We can condition
on (7.5) holding and then argue that this only multiplies small probabilities by
1 + o(1) if we use P(A | B) ≤ P(A)/ P(B).
It follows that if
log n log n
k2 = ≈
2(log np + log(1 − γ) 2 log np
then w.h.p. we have
|Nk2 | ≥ n1/2 .
7.2. LARGEST INDEPENDENT SETS 121
Dense case
The following theorem was first proved by Matula [576].
1
Theorem 7.3. Suppose 0 < p < 1 is a constant and b = 1−p . Then w.h.p.
α(Gn,p ) ≈ 2 logb n.
Let
∆= ∑ P(Si , S j are independent in Gn,p ),
i, j
Si ∼S j
where S1 , S2 , . . . , S(n) are all the k-subsets of [n] and Si ∼ S j iff |Si ∩ S j | ≥ 2. By
k
Janson’s inequality, see Theorem 22.13,
(E Xk )2
P(Xk = 0) ≤ exp − .
2∆
Here we apply the inequality in the context of Xk being the number of k-cliques in
the complement of Gn,p . The set [N] will be the edges of the complete graph and
the sets Di will the edges of the k-cliques. Now
u j+1 k− j k− j
= (1 − p)− j
uj n − 2k + j + 1 j + 1
k (1 − p)− j
2
logb n
≤ 1+O .
n n( j + 1)
Therefore,
2 j−2
uj k 2(1 − p)−( j−2)( j+1)/2
≤ (1 + o(1))
u2 n j!
j−2
2k2 e
j+1
≤ (1 + o(1)) (1 − p)− 2 ≤ 1.
nj
So
(E Xk )2 1 n2 (1 − p)
≥ ≥ .
∆ ku2 k5
7.2. LARGEST INDEPENDENT SETS 123
Therefore
2 /(log n)5 )
P(Xk = 0) ≤ e−Ω(n . (7.7)
Matula used the Chebyshev inequality and so he was not able to prove an
exponential bound like (7.7). This will be important when we come to discuss the
chromatic number.
Sparse Case
We now consider the case where p = d/n and d is a large constant. Frieze [339]
proved
Theorem 7.4. Let ε > 0 be a fixed constant. Then for d ≥ d(ε) we have that
w.h.p.
2n
α(Gn,p )) − (log d − log log d − log 2 + 1) ≤ εn .
d d
Dani and Moore [239] have recently given an even sharper result.
In this section we will prove that if p = d/n and d is sufficiently large then
w.h.p.
α(Gn,p ) − 2 log d ε log d
≤
n n. (7.8)
d d
This will follow from the following. Let Xk be as defined in the previous section.
Let
(2 − ε/8) log d (2 + ε/8) log d
k0 = n and k1 = n.
d d
Then,
(log d)2
ε log d
P α(Gn,p ) − E(α(Gn,p )) ≥
n ≤ exp −Ω n . (7.9)
8d d2
(log d)2
P(α(Gn,p ) ≥ k1 ) = P(Xk1 > 0) ≤ exp −Ω n . (7.10)
d
( ! )
(log d)3/2
P(α(Gn,p ) ≥ k0 ) = P(Xk0 > 0) ≥ exp −O n . (7.11)
d2
Let us see how (7.8) follows from these two. Indeed, (7.9) and (7.11) imply that
ε log d
E(α(Gn,p )) ≥ k0 − n. (7.12)
8d
Furthermore (7.9) and (7.10) imply that
ε log d
E(α(Gn,p )) ≤ k1 + n. (7.13)
8d
124 CHAPTER 7. EXTREME CHARACTERISTICS
t2
P(|Z − E(Z)| ≥ t) ≤ exp − .
2n − 2
ε log d
Setting t = 8d n yields (7.9).
Proof of (7.10): The first moment method gives
(k1 ) (k1 −1)/2 !k1
n d 2 ne d
Pr(Xk1 > 0) ≤ 1− ≤ · 1−
k1 n k1 n
k1
(log d)2
de −(1+ε/5)
≤ ·d = exp −Ω n .
2 log d d
k0
= ∑ v j. (7.15)
j=0
7.3 Interpolation
The following theorem is taken from Bayati, Gamarnik and Tetali [67]. Note that
it is not implied by Theorem 7.4. This paper proves a number of other results of
a similar flavor for other parameters. It is an important paper in that it verifies
some very natural conjectures about some graph parameters, that have not been
susceptible to proof until now.
126 CHAPTER 7. EXTREME CHARACTERISTICS
E(α(Gn,bdnc ))
lim = H(d).
n→∞ n
(A)
Proof. For this proof we use the model Gn,m of Section 1.3. This is proper since
we we know that w.h.p.
(A) (A)
|α(Gn,m ) − α(Gn,m )| ≤ ||E(Gn,m )| − m| ≤ log n.
(A)
Thus the sequence un = E(α(Gn,bdnc )) satisfies the conditions of Lemma 7.6 be-
low and the proof of Theorem 7.5 follows.
Proof of (7.17): We begin by constructing a sequence of graphs interpolating
(A) (A) (A)
between Gn,bdnc and a disjoint union of Gn1 ,m1 and Gn2 ,m2 . Given n, n1 , n2 such
that n1 + n2 = n and any 0 ≤ r ≤ m = bdnc, let G(n, m, r) be the random (pseudo-
)graph on vertex set [n] obtained as follows. It contains precisely m edges. The
first r edges e1 , e2 , . . . , er are selected randomly from [n]2 . The remaining m −
r edges er+1 , . . . , em are generated as follows. For each j = r + 1, . . . , m, with
probability n j /n, e j is selected randomly from M1 = [n1 ]2 and with probability
n2 /n, e j is selected randomly from M2 =[n1 + 1, n]2 . Observe that when r = m we
(A)
have G(n, m, r) = G(A) (n, m) and when r = 0 it is the disjoint union of Gn1 ,m1 and
(A)
Gn2 ,m2 where m j = Bin(m, n j /n) for j = 1, 2. We will show next that
(A)
E(α(Gn,m )) = E(α(G(n, m, m))) ≥
(A) (A)
E(α(G(n, m, 0))) = E(α(Gn1 ,m1 )) + E(α(Gn2 ,m2 ))
7.4. CHROMATIC NUMBER 127
which is (7.17).
Proof of (7.19): Observe that G(n, m, r − 1) is obtained from
G(n, m, r) by deleting the random edge er and then adding an edge from M1 or M2 .
Let G0 be the graph obtained after deleting er , but before adding its replacement.
Remember that
G(n, m, r) = G0 + er .
We will show something stronger than (7.19) viz. that
E(α(G(n, m, r)) | G0 ) ≥ E(α(G(n, m, r − 1)) | G0 ) for r = 1, . . . , m. (7.20)
Now let O∗ ⊆ [n] be the set of vertices that belong to every largest independent set
in G0 . Then for er = (x, y), α(G0 + e) = α(G0 ) − 1 if x, y ∈ O∗ and α(G0 + e) =
/ O∗ or y ∈
α(G0 ) if x ∈ / O∗ . Because er is randomly chosen, we have
∗ 2
|O |
E(α(G0 + er ) | G0 ) − E(α(G0 )) = − .
n
By a similar argument
E(α(G(n, m, r − 1) | G0 ) − α(G0 )
n1 |O∗ ∩ M1 | 2 n2 |O∗ ∩ M2 | 2
=− −
n n1 n n2
n1 |O∗ ∩ M1 | n2 |O∗ ∩ M2 | 2
≤− +
n n1 n n2
∗ 2
|O |
=−
n
= E(α(G0 + er ) | G0 ) − E(α(G0 )),
completing the proof of (7.20).
The proof of the following lemma is left as an exercise.
Lemma 7.6. Given γ ∈ (0, 1), suppose that the non-negative sequence un , n ≥ 1
satisfies
un ≥ un1 + un2 − O(nγ )
for every n1 , n2 such that n1 + n2 = n. Then limn→∞ unn exists.
Dense Graphs
We will first describe the asymptotic behavior of the chromatic number of dense
random graphs. The following theorem is a major result, due to Bollobás [136].
The upper bound without the 2 in the denominator follows directly from Theorem
7.3. An intermediate result giving 3/2 instead of 2 was already proved by Matula
[577].
1
Theorem 7.7. Suppose 0 < p < 1 is a constant and b = 1−p . Then w.h.p.
n
χ(Gn,p ) ≈ .
2 logb n
So assume that every set of order at least ν contains an independent set of order
at least k0 . We repeatedly choose an independent set of order k0 among the set
of uncolored vertices. Give each vertex in this set a new color. Repeat until the
number of uncolored vertices is at most ν. Give each remaining uncolored vertex
its own color. The number of colors used is at most
n n
+ν ≈ .
k0 2 logb n
It should be noted that Bollobás did not have the Janson inequality available
to him and he had to make a clever choice of random variable for use with the
Azuma-Hoeffding inequality. His choice was the maximum size of a family of
edge independent independent sets. Łuczak [560] proved the corresponding result
to Theorem 7.7 in the case where np → 0.
7.4. CHROMATIC NUMBER 129
Concentration
Proof. Write
χ = Z(Y1 ,Y2 , . . . ,Yn ) (7.22)
where
Y j = {(i, j) ∈ E(Gn,p ) : i < j}.
Then
|Z(Y1 ,Y2 , . . . ,Yn ) − Z(Y1 ,Y2 , . . . , Ŷi , . . . ,Yn )| ≤ 1
and the theorem follows from the Azuma-Hoeffding inequality, see Section 22.7,
in particular Lemma 22.17.
Algorithm GREEDY
• A is the current set of vertices that might get color k in the current round.
begin
k ←− 0, A ←− [n], U ←− [n], Ck ←− 0.
/
while U 6= 0/ do
k ←− k + 1 A ←− U
while A 6= 0/
begin
130 CHAPTER 7. EXTREME CHARACTERISTICS
1
Theorem 7.9. Suppose 0 < p < 1 is a constant and b = 1−p . Then w.h.p. algo-
rithm GREEDY uses approximately n/ logb n colors to color the vertices of Gn,p .
Proof. At the start of an iteration the edges inside U are un-examined. Suppose
that
n
|U| ≥ ν = .
(logb n)2
We show that approximately logb n vertices get color k i.e. at the end of round k,
|Ck | ≈ logb n.
Each iteration chooses a maximal independent set from the remaining uncolored
vertices. Let k0 = logb n − 5 logb logb n. Then
So the probability that we fail to use at least k0 colors while |U| ≥ ν is at most
1 3
ne− 2 (logb ν) = o(1).
Sparse Graphs
We now consider the case of sparse random graphs. We first state an important
conjecture about the chromatic number.
Conjecture: Let k ≥ 3 be a fixed positive integer. Then there exists dk > 0
such that if ε is an arbitrary positive constant and p = dn then w.h.p. (i) χ(Gn,p ) ≤ k
for d ≤ dk − ε and (ii) χ(Gn,p ) ≥ k + 1 for d ≥ dk + ε.
In the absence of a proof of this conjecture, we present the following result due
to Łuczak [561]. It should be noted that Shamir and Spencer [708] had already
proved six point concentration.
132 CHAPTER 7. EXTREME CHARACTERISTICS
Theorem 7.10. If p < n−5/6−δ , δ > 0, then the chromatic number of Gn,p is w.h.p.
two point concentrated.
(a) Let 0 < δ < 1/10, 0 ≤ p < 1 and d = np. Then w.h.p. each subgraph H of
Gn,p on less than nd −3(1+2δ ) vertices has less than (3/2 − δ )|H| edges.
(b) Let 0 < δ < 1.0001 and let 0 ≤ p ≤ δ /n. Then w.h.p. each subgraph H of
Gn,p has less than 3|H|/2 edges.
The above lemma can be proved easily by the first moment method, see Ex-
ercise 7.6.6. Note also that Lemma 7.11 implies that each subgraph H satisfying
the conditions of the lemma has minimum degree less than three, and thus is 3-
colorable, due to the following simple observation (see Bollobás [137] Theorem
V.1)
Lemma 7.12. Let k = maxH⊆G δ (H), where the maximum is taken over all in-
duced subgraphs of G. Then χ(G) ≤ k + 1.
1
P (χ(Gn,p ) ≥ k) > . (7.23)
log log n
Then w.h.p. all but at most n1/2 log n vertices of Gn,p can be properly colored
using k colors.
Proof. Let Z be the maximum number of vertices in Gn,p that can be properly
colored with k colors. Write Z = Z(Y1 ,Y2 , . . . ,Yn ) as in (7.22). Then we have
2
1 t
P(Z = n) > and P(|Z − E Z)| ≥ t) ≤ 2 exp − . (7.24)
log log n 2n
7.5. EIGENVALUES 133
Putting t = 21 n1/2 log n into (7.24) shows that E Z ≥ n − t and the lemma follows
after applying the concentration inequality in (7.24) once again.
Now we are ready to present Łuczak’s ingenious argument to prove Theorem
7.10. Note first that when p is such that np → 0 as n → ∞, then by Theorem 2.1
Gn,p is a forest w.h.p. and so its chromatic number is either 1 or 2. Furthermore,
for 1/ log n < d < 1.0001 the random graph Gn,p w.h.p. contains at least one edge
and no subgraph with minimal degree larger than two (see Lemma 7.11), which
implies that χ(Gn,p ) is equal to 2 or 3 (see Lemma 7.12). Now let us assume that
the edge probability p is such that 1.0001 < d = np < n1/6−δ . Observe that in this
range of p the random graph Gn,p w.h.p. contains an odd cycle, so χ(Gn,p ) ≥ 3.
Let k be as in Lemma 7.13 and let U0 be a set of size at most u0 = n1/2 log n
such that [n] \U0 can be properly colored with k colors. Let us construct a nested
sequence of subsets of vertices U0 ⊆ U1 ⊆ . . . ⊆ Um of Gn,p , where we define
Ui+1 = Ui ∪ {v, w}, where v, w 6∈ Ui are connected by an edge and both v and w
have a neighbor in Ui . The construction stops at i = m if such a pair {v, w} does
not exist.
Notice that m can not exceed m0 = n1/2 log n, since if m > m0 then a subgraph of
Gn,p induced by vertices of Um0 would have
vertices and at least 3m0 ≥ (3/2 − δ )|Um0 | edges, contradicting the statement of
Lemma 7.11.
As a result, the construction produces a set Um in Gn,p , such that its size is smaller
than nd −3(1+2δ ) and, moreover, all neighbors N(Um ) of Um form an independent
set, thus “isolating” Um from the “outside world”.
Now, the coloring of the vertices of Gn,p is an easy task. Namely, by Lemma 7.13,
we can color the vertices of Gn,p outside the set Um ∪ N(Um ) with k colors. Then
we can color the vertices from N(Um ) with color k + 1, and finally, due to Lemmas
7.11 and 7.12, the subgraph induced by Um is 3-colorable and we can color Um
with any three of the first k colors.
7.5 Eigenvalues
Separation of first and remaining eigenvalues
The following theorem is a weaker version of a theorem of Füredi and Komlós
[370], which was itself a strengthening of a result of Juhász [471]. See also Coja–
Oghlan [198] and Vu [749]. In their papers, 2ω log n is replaced by 2 + o(1) and
this is best possible.
134 CHAPTER 7. EXTREME CHARACTERISTICS
where
kMk = max |Mx| = max {|λ1 (M)|, |λn (M)|} .
|x|=1
We first show that the lemma implies the theorem. Let e denote the all 1’s vector.
Suppose that |ξ | = 1 and ξ ⊥e. Then Jξ = 0 and
p
|Aξ | = |Mξ | ≤ kMk ≤ 2ω log n np(1 − p).
Now let |x| = 1 and let x = αu + β y where u = √1 e and y⊥e and |y| = 1. Then
n
= |pJu| − |Mu|
p
≥ np − 2ω log n np(1 − p),
|Aξ |
≤ max
06=ξ ⊥u |ξ |
|Mξ |
≤ max
06=ξ ⊥u |ξ |
p
≤ 2ω log n np(1 − p)
λn = min ξ T Aξ ≥ min ξ T Aξ − pξ T Jξ
|ξ |=1 |ξ |=1
p
= min −ξ T Mξ ≥ −kMk ≥ −2ω log n np(1 − p).
|ξ |=1
(i) E mi j = 0
(ii) Var mi j ≤ p(1 − p) = σ 2
(iii) mi j , mi0 j0 are independent, unless (i0 , j0 ) = ( j, i),
in which case they are identical.
We estimate
kM̂k ≤ Trace(M̂ k )1/k ,
where k = ω log n.
Now,
Recall that the i, jth entry of M̂ k is the sum over all products
mi,i1 mi1 ,i2 · · · mik−1 j .
Continuing, we therefore have
k
E kM̂kk ≤ ∑ En,k,ρ
ρ=2
where !
k−1
En,k,ρ = ∏ mi j i j+1 .
∑ E
i0 ,i1 ,...,ik−1 ∈[n] j=0
|{i0 ,i1 ,i2 ,...,ik−1 }|=ρ
if the walk W (i) contains an edge that is crossed exactly once, by condition (i).
On the other hand, |mi j | ≤ 1 and so by conditions (ii), (iii),
!
k−1
E ∏ mi j i j+1 ≤ σ 2(ρ−1)
j=0
where nρ bounds from above the the number of choices of ρ distinct vertices,
while kk bounds the number of walks of length k.
We have
1 1
2 k+1 2 k+1 1
E kM̂kk ≤ ∑ Rk,ρ σ 2(ρ−1) ≤ ∑ nρ kk σ 2(ρ−1) ≤ 2n 2 k+1 kk σ k .
ρ=2 ρ=2
Therefore,
E kM̂kk
1 k
1
k
P kM̂k ≥ 2kσ n = P kM̂k ≥ 2kσ n
2 2 ≤
1 k
2kσ n 2
1
!k k
2n 2 k+1 kk σ k (2n)1/k
1
≤ = = + o(1) = o(1).
1 k 2 2
2kσ n 2
p
It follows that w.h.p. kM̂k ≤ 2σ ω(log n)n1/2 ≤ 2ω log n np(1 − p) and com-
pletes the proof of Theorem 7.14.
Concentration of eigenvalues
We show here how one can use Talagrand’s inequality, Theorem 22.18, to show
that the eigenvalues of random matrices are highly concentrated around their me-
dian values. The result is from Alon, Krivelevich and Vu [30].
Theorem 7.16. Let A be an n × n random symmetric matrix with independent en-
tries ai, j = a j,i , 1 ≤ i ≤ j ≤ n with absolute value at most one. Let its eigenvalues
be λ1 (A) ≥ λ2 (A) ≥ · · · ≥ λn (A). Suppose that 1 ≤ s ≤ n. Let µs denote the me-
dian value of λs (A) i.e. µs = infµ {P(λs (A) ≤ µ) ≥ 1/2}. Then for any t ≥ 0 we
have
2 2
P(|λs (A) − µs | ≥ t) ≤ 4e−t /32s .
The same estimate holds for the probability that λn−s+1 (A) deviates from its me-
dian by more than t.
Fix B ∈ B and let v(1) , v(2) , . . . , v(s) be an orthonormal set of eigenvectors for
(k) (k) (k)
the s largest eigenvalues of B. Let v(k) = (v1 , v2 , . . . , vn ),
s
(k) 2
αi,i = ∑ (vi ) for 1 ≤ i ≤ n
k=1
and s s
s s
(k) (k) 2
αi, j = 2 ∑ (vi )2 ∑ (v j ) for 1 ≤ i < j ≤ n.
k=1 k=1
Lemma 7.17.
∑ αi,2 j ≤ 2s2 .
1≤i≤ j≤n
Proof.
!2 !
n s s s
(k) 2 (k) 2 (k) 2
∑ αi,2 j = ∑ ∑ (vi ) +4 ∑ ∑ (vi ) ∑ (v j )
1≤i≤ j≤n i=1 k=1 1≤i< j≤n k=1 k=1
!2 !2
n s s n
(k) (k) 2
≤2 ∑ ∑ (vi )2 =2 ∑ ∑ (vi ) = 2s2 ,
i=1 k=1 k=1 i=1
where we have used the fact that each v(k) is a unit vector.
∑ αi, j ≥ t/2.
1≤i≤ j≤n:ai, j 6=bi, j
Fix A ∈ A . Let u = ∑sk=1 ck v(k) be a unit vector in the span S of the vectors
v(k) , k = 1, 2, . . . , s which is orthogonal to the eigenvectors of the (s − 1) largest
eigenvalues of A. Recall that v(k) , k = 1, 2, . . . , s are eigenvectors of B. Then
∑sk=1 c2k = 1 and ut Au ≤ λs (A) ≤ M, whereas ut Bu ≥ minv∈S vt Bv = λs (B) ≥ M +
t. Recall that all entries of A and B are bounded in absolute value by 1, implying
that |bi, j − ai, j | ≤ 2 for all 1 ≤ i, j ≤ n. It follows that if X is the set of ordered
pairs (i, j) for which ai, j 6= bi, j then
!t
s s
t (k) (k)
t ≤ u (B − A)u = ∑ (bi, j − ai, j ) ∑ ck vi ∑ ck v j
(i, j)∈X k=1 k=1
s s
(k) (k)
≤ 2 ∑ ∑ ck vi ∑ ck v j
(i, j)∈X k=1
k=1
7.6. EXERCISES 139
s s ! s s !
s s 2 s s
2
(k) (k)
≤2 ∑ ∑ c2k ∑ vi ∑ c2k ∑ vj
(i, j)∈X k=1 k=1 k=1 k=1
=2 ∑ αi, j
(i, j)∈X
as claimed. (We obtained the third inequality by use of the Cauchy-Schwarz in-
equality).
By the above two lemmas, and by Theorem 22.18 for every M and every t > 0
2 /(32s2 )
P(λs (A) ≤ M) P(λs (B) ≥ M + t) ≤ e−t . (7.26)
7.6 Exercises
7.6.1 Let p = d/n where d is a positive constant. Let S be the set of vertices of
2 log n
degree at least 3 log log n . Show that w.h.p., S is an independent set.
7.6.2 Let p = d/n where d is a large positive constant. Use the first moment
method to show that w.h.p.
2n
α(Gn,p ) ≤ (log d − log log d − log 2 + 1 + ε)
d
for any positive constant ε.
7.6.3 Complete the proof of Theorem 7.4.
Let m = d/(log d)2 and partition [n] into n0 = mn sets S1 , S2 , . . . , Sn0 of size
m. Let β (G) be the maximum size of an independent set S that satisfies
|S ∩ Si | ≤ 1 for i = 1, 2, . . . , n0 . Use the proof idea of Theorem 7.4 to show
that w.h.p.
2n
β (Gn,p ) ≥ k−ε = (log d − log log d − log 2 + 1 − ε).
d
140 CHAPTER 7. EXTREME CHARACTERISTICS
7.6.8 A topological clique of size s is a graph obtained from the complete graph
Ks by subdividing edges. Let tc(G) denote the size of the largest topological
clique contained in a graph G. Prove that w.h.p. tc(Gn,1/2 ) = Θ(n1/2 ).
7.6.10 Show that if d > 2k log k for a positive integer k ≥ 2 then w.h.p. G(n, d/n) is
not k-colorable. (Hint:Consider the expected number of proper k-coloring’s).
7.6.11 Let p = K log n/n for some large constant K > 0. Show that w.h.p. the
diameter of Gn,p is Θ(log n/ log log n).
7.6.12 Suppose that 1 + ε ≤ np = o(log n), where ε > 0 is constant. Show that
given A > 0, there exists B = B(A) such that
log n
P diam(K) ≥ B ≤ n−A ,
log np
7.6.13 Let p = d/n for some constant d > 0. Let A be the adjacency matrix of
Gn,p . Show that w.h.p. λ1 (A) ≈ ∆1/2 where ∆ is the maximum degree in
Gn,p . (Hint: the maximum eigenvalue of the adjacency matrix of K1,m is
m1/2 ).
7.6.15 The set chromatic number χs (G) of a graph G = (V, E) is defined as follows:
Let C denote a set of colors. Color each v ∈ V with a color f (v) ∈ C.
Let Cv = { f (w) : {v, w} ∈ G}. The coloring is proper if Cv 6= Cw whenever
{v, w} ∈ E. χs is the minimum size of C in a proper coloring of G. Prove
that if 0 < p < 1 is constant then w.h.p. χs (Gn,p ) ≈ r log2 n where r = log 21/s
2
and s = min q2` + (1 − q` )2 : ` = 1, 2, . . . where q = 1 − p. (This question
7.7 Notes
Chromatic number
There has been a lot of progress in determining the chromatic number of sparse
random graphs. Alon and Krivelevich [27] extended the result in [561] to the
range p ≤ n−1/2−δ . A breakthrough came when Achlioptas and Naor [6] identi-
fied the two possible values for np = d where d = O(1): Let kd be the smallest
integer k such that d < 2k log k. Then w.h.p. χ(Gn,p ) ∈ {kd , kd + 1}. This im-
plies that dk , the (conjectured) threshold for a random graph to have chromatic
number at most k, satisfies dk ≥ 2k log k − 2 logk −2 + ok (1) where ok (1) → 0 as
k → ∞. Coja–Oghlan, Panagiotou and Steger [200] extended the result of [6]
to np ≤ n1/4−ε , although here the guaranteed range is three values. More re-
cently, Coja–Oghlan and Vilenchik [201] proved the following. Let dk,cond =
2k log k − log k − 2 log 2. Then w.h.p. dk ≥ dk,cond − ok (1). On the other hand
Coja–Oghlan [199] proved that dk ≤ dk,cond + (2 log 2 − 1) + ok (1).
It follows from Chapter 2 that the chromatic number of Gn,p , p ≤ 1/n is w.h.p.
at most 3. Achlioptas and Moore [4] proved that in fact χ(Gn,p ) ≤ 3 w.h.p. for
p ≤ 4.03/n. Now a graph G is s-colorable iff it has a homomorphism ϕ : G → Ks .
(A homomorphism from G to H is a mapping ϕ : V (G) → V (H) such that if
{u, v} ∈ E(G) then (ϕ(u), ϕ(v)) ∈ E(H)). It is therefore of interest in the con-
text of coloring, to consider homomorphisms from Gn,p to other graphs. Frieze
and Pegden [361] show that for any ` > 1 there is an ε > 0 such that with high
probability, Gn, 1+ε either has odd-girth < 2` + 1 or has a homomorphism to the
n
odd cycle C2`+1 . They also showed that w.h.p. there is no homomorphism from
Gn,p , p = 4/n to C5 . Previously, Hatami [413] has shown that w.h.p. there is no
homomorphism from a random cubic graph to C7 .
Alon and Sudakov [32] considered how many edges one must add to Gn,p in
order to significantly increase the chromatic number. They show that if n−1/3+δ ≤
p ≤ 1/2 for some fixed δ > 0 then w.h.p. for every set E of
2−12 ε 2 n2
(log (np))2
edges, the chromatic number of Gn,p ∪ E is still at most 2 (1+ε)n
log (np) .
b b
142 CHAPTER 7. EXTREME CHARACTERISTICS
Algorithmic questions
We have seen that the Greedy algorithm applied to Gn,p generally produces a
coloring that uses roughly twice the minimum number of colors needed. Note
also that the analysis of Theorem 7.9, when k = 1, implies that a simple greedy
algorithm for finding a large independent set produces one of roughly half the
maximum size. In spite of much effort neither of these two results have been sig-
nificantly improved. We mention some negative results. Jerrum [465] showed that
the Metropolis algorithm was unlikely to do very well in finding an independent
set that was significantly larger than GREEDY. Other earlier negative results in-
clude: Chvátal [194], who showed that for a significant set of densities, a large
class of algorithms will w.h.p. take exponential time to find the size of the largest
independent set and McDiarmid [581] who carried out a similar analysis for the
chromatic number.
Frieze, Mitsche, Pérez-Giménez and Pralat [359] study list coloring in an on-
line setting and show that for a wide range of p, one can asymptotically match the
best known constants of the off-line case. Moreover, if pn ≥ logω n, then they get
the same multiplicative factor of 2 + o(1).
let Ωk (G) be the set of proper k-coloring’s of the vertices of G. There has been a
good deal of work on the problem of efficiently choosing a (near) random member
of Ωk (G). For example, Vigoda [746] has described an algorithm that produces a
(near) random sample in polynomial time provided k > 11∆(G)/6. When it comes
to Gn,p , Dyer, Flaxman, Frieze and Vigoda [277] showed that if p = d/n, d = O(1)
then w.h.p. one can sample a random coloring if k = O(log log n) = o(∆). The
bound on k was reduced to k = O(d O(1) ) by Mossell and Sly [615] and then to
k = O(d) by Efthymiou [283].
Extremal Properties
8.1 Containers
Ramsey theory and the Turán problem constitute two of the most important areas
in extremal graph theory. For a fixed graph H we can ask how large should n be
so that in any r-coloring of the edges of Kn can we be sure of finding a monochro-
matic copy of H – a basic question in Ramsey theory. Or we can ask for the
maximum α > 0 such that we take an α proportion of the edges of Kn without
including a copy of H – a basic question related to the Turán problem. Both of
these questions have analogues where we replace Kn by Gn,p .
There have been recent breakthroughs in transferring extremal results to the
context of random graphs and hypergraphs. Conlon and Gowers [206], Schacht
[702], Balogh, Morris and Samotij [56] and Saxton and Thomason [700] have
proved general theorems enabling such transfers. One of the key ideas being
to bound the number of independent sets in carefully chosen hypergraphs. Our
presentation will use the framework of [700] where it could just as easily have
used [56]. The use of containers is a developing field and seems to have a growing
number of applications.
In this section, we present a special case of Theorem 2.3 of [700] that will
enable us to deal with Ramsey and Turán properties of random graphs. For a
graph H with e(H) ≥ 2 we let
e(H 0 ) − 1
m2 (H) = max . (8.1)
H 0 ⊆H,e(H 0 )>1 v(H 0 ) − 2
145
146 CHAPTER 8. EXTREMAL PROPERTIES
Next let
ex(n, H)
π(H) = lim n (8.2)
n→∞
2
Theorem 8.1. Let H be a graph with e(H) ≥ 2 and let ε be a positive constant.
For some constant h = h(H, ε) > 0 and n sufficiently large, there exists a collection
C of graphs on vertex set [n] such that the following holds. The graphs C are the
containers:
(a) For every H-free graph Γ there exists T ⊆ Γ ⊆ C(T ) ∈ C such that e(T ) ≤
hn2−1/m2 (H) . (C depends only on T and not on Γ.)
n
(b) C contains at most εnv(H) copies of H and e(C) ≤ (π(H) + ε) 2 for every
C ∈ C.
We prove Theorem 8.1 in Section 8.4. We have extracted just enough from
Saxton and Thomason [700] and [701] to give a complete proof. But first we give
a couple of examples of the use of this theorem.
Theorem 8.2. For any graph H with e(H) ≥ v(H) and r ≥ 2, there exist c0 , c1 > 0
such that (
o(1) p ≤ c0 n−1/m2 (H)
P(Gn,p → (H)er ) =
1 − o(1) p ≥ c1 n−1/m2 (H)
8.2. RAMSEY PROPERTIES 147
The density p0 = n−1/m2 (H) is the threshold for every edge of Gn,p to be con-
tained in a copy of H. When p ≤ cp0 for small c, the copies of H in Gn,p will
be spread out and the associated 0-statement is not so surprising. We will use
Theorem 8.1 to prove the 1-statement for p ≥ c1 p0 . The proof of the 0-statement
follows [628] and is given in Exercises 8.5.1 to 8.5.6.
We begin with a couple of lemmas:
Lemma 8.3. For every graph H and r ≥ 2 there exist constants α > 0 and n0 such
that for all n ≥ n0 every r-coloring of the edges of Kn contains at least αnv(H)
monochromatic copies of H.
Proof. From Ramsey’s theorem we know that there exists N = N(H, r) such that
every r-coloring of the edges of KN contains a monochromatic copy of H. Thus,
in any r-coloring of Kn , every N-subset of the vertices of Kn contains at least one
n−v(H)
monochromatic copy of H. As every copy of H is contained in at most N−v(H)
N-subsets, the theorem follows with α = 1/N v(H) .
From this we get
Corollary 8.4. For every graph H and every positive integer r there exist con-
stants n0 and δ , ε > 0 such that the following is true: If n ≥ n0 , then for any
E1 , E2 , . . . , Er ⊆ E(Kn ) such that for all 1 ≤ i ≤ r the set Ei contains at most εnv(H)
copies of H, we have
|E(Kn ) \ (E1 ∪ E2 ∪ · · · ∪ Er )| ≥ δ n2 .
Proof. Let α and n0 be as given in Lemma 8.3 for H and r + 1. Further, let
Er+1 = E(Kn ) \ (E1 ∪ E2 ∪ · · · ∪ Er ), and consider the coloring f : E(Kn ) → [r + 1]
given by f (e) = mini∈[r+1] {e ∈ Ei }. By Lemma 8.3 there exist at least αnv(H)
monochromatic copies of H under coloring f , and so by our assumption on the
sets Ei , 1 ≤ i ≤ r, Er+1 must contain at least αnv(H) copies. As every edge is
contained in at most e(H)nv(H)−2 copies and E1 ∪ E2 ∪ · · · Er contains at most
rεnv(H) copies of H, we see that Er+1 contains at least (α − rεe(H))nv(H) copies
(α−rεe(H))nv(H)
of H. It follows that |Er+1 | ≥ e(H)nv(H)−2
and so the corollary follows with
α−re(H)ε α
δ= e(H) . Here we take ε ≤ 2re(H) .
Proof. We can now proceed to the proof of the 1-statement of Theorem 8.2. If
Gn,p 6→ (H)er then there must exist a coloring f : E(Gn,p ) → [r] such that for all
1 ≤ i ≤ r the set Ei = f −1 (i) does not contain a copy of H. By Theorem 8.1 we
have that for every such Ei there exists Ti and a container Ci such that Ti ⊆ Ei ⊆ Ci .
The crucial observation is that Gn,p completely avoids E0 = E(Kn ) \ (C1 ∪ C2 ∪
· · · ∪Cr ), which by Corollary 8.4 and a choice of ε has size at least δ n2 .
148 CHAPTER 8. EXTREMAL PROPERTIES
Therefore, we can bound P(Gn,p 6→ (H)er ) by the probability that there ex-
ist T = {T1 , . . . , Tr } and C = {Ci = C(Ti ) : i = 1, 2, . . . , r} such that E0 is edge-
disjoint from Gn,p . Thus,
P((Gn,p 6→ (H)er ) ≤ ∑ P(Ti ⊆ Gn,p , 1 ≤ i ≤ r ∧ E(Gn,p ) ∩ E0 = 0).
/
Ti ,1≤i≤r
Note that the two events in the above probability are independent and can thus be
bounded by pa (1 − p)b where a = | i Ti | and b = δ n2 . The sum can be bounded
S
by first deciding on a ≤ rhn2−1/m2 (H) (h from Theorem 8.1) and then choosing a
n
edges ( (a2) choices) and then deciding for every edge in which Ti it appears (ra
choices). Thus,
rhn2−1/m2 (H) n
e −δ n2 p
P((Gn,p 6→ (H)r ) ≤ e ∑ 2 (rp)a
a=0 a
rhn2−1/m2 (H) 2 a
−δ n2 p en rp
≤e ∑ .
a=0 2a
dom graphs. Our proof is taken from [700], although Conlon and Gowers [206]
gave a proof for 2-balanced H and Schacht [702] gave a proof for general H.
8.3. TURÁN PROPERTIES 149
Theorem 8.5. Suppose that 0 < γ < 1 and H is not a matching. Then there exists
A > 0 such that if p ≥ An−1/m2 (H) and n is sufficiently large then the following
3 n
event occurs with probability at least 1 − e−γ (2) p/384 :
n
Every H-free subgraph of Gn,p has at most (π(H) + γ) p edges.
2
If H is a matching then m2 (H) = 1/2 and then the lower bound on p in the
theorem is O(n−2 ) we would not be claiming a high probability result.
To prove the theorem, we first prove the following lemma:
Lemma 8.6. Given 0 < η < 1 and h ≥ 1, there is a constant ϕ = ϕ(η, h) such
that the following holds: Let M be a set, |M| = N and let I ⊆ 2M . Let t ≥ 1,
ϕt/N ≤ p ≤ 1 and let ηN/2 ≤ d ≤ N. Suppose there exists C : 2M → 2M and
T ⊆ ≤t M
such that for each I ∈ I there exists TI ∈ T such that TI ⊆ I and
CI = C(TI ) ⊆ M, where |CI | ≤ d. Let X ⊆ M be a random subset where each
element is chosen independently with probability p. Then
2 d p/24
P(∃I ∈ I : |CI ∩ X| > (1 + η)pd and I ⊆ X) ≤ e−η . (8.3)
With this lemma in hand, we can complete the proof of Theorem 8.5.
Let I be the set of H-free graphs on vertex set [n]. We take M = [n]
2 and
X = E(Gn,p ) and N = n2 . For I ∈ I , let TI and h = h(H, ε) be given by Theorem
8.1. Each H-free graph I ∈ I is contained in CI and so if Gn,p contains an H-
free subgraph with (π(H) + γ)N p edges then there exists I such that |X ∩ CI | ≥
(π(H) + γ)N p. Our aim is to apply Lemma 8.6 with
γ γ
η = , d = π(H) + N, t = hn2−1/m2 (H) .
2 4
The conditions of Lemma 8.6 then hold after noting that d ≥ ηN/2 and that p ≥
An−1/m2 (H) ≥ ϕt/N if A is large enough. Note also that |CI | ≤ d. Now (1 +
η)d p ≤ (π(H) + γ)N p, and so the probability that the event in the statement of
the theorem fails to occur is bounded by
γ 3N p
−η 2 d p/24
e ≤ exp −
384
Theorem 8.7. Let H be an `-graph with e(H) ≥ 2 and let ε > 0. For some h > 0
and for every N ≥ h, there exists a collection C of `-graphs on vertex set [N] such
that
(a) for every H-free `-graph I on vertex set [N], there exists C ∈ C with I ⊆ C,
(c) moreover, for every I in (a), there exists T ⊆ I, e(T ) ≤ hN `−1/m(H) , such
that C = C(T ).
1
µ(S) = ∑ d(u).
nd u∈S
1 µ(S)nd
e(G[S]) ≤ ∑ d(v) = = µ(S)e(G). (8.4)
r v∈S r
We now state the main theorem. An independent set of an `-graph is a set I such
that e ∈ E(G) implies e 6⊆ I.
Theorem 8.9. Let r ∈ N. Let G be an r-graph with average degree d and vertex
set [n]. Suppose that we can choose 0 < c, τ < 1 such that
Then there is a function C : P[n] → P[n], such that, for every independent set
I ⊆ [n] there exists T ⊆ I with
(a) I ⊆ C(T ),
(b) µ(T ) ≤ τ,
(d) µ(C(T )) ≤ 1 − c.
Corollary 8.10. Let r ∈ N and let ε > 0. Let G be an r-graph of average degree d
on vertex set [n]. Suppose that we can choose 0 < c, τ < 1 such that (8.5) holds.
Then there is a function C : P[n] → P[n], such that, for every independent set
I ⊆ [n] there exists T ⊆ I with
(a) I ⊆ C(T ),
The algorithm
We now describe an algorithm which given independent set I, constructs the quan-
tities in Theorem 8.9 and Corollary 8.10. It runs in two modes, prune mode, builds
T ⊆ I and build mode, which constructs C ⊇ I.
The algorithm builds multigraphs Ps , s ∈ [r] and then we define the degree of
σ in the multigraph Ps to be
where we are counting edges with multiplicity in the multiset E(Ps ). (Naturally
we may write ds (v) instead of ds ({v}) if v ∈ [n].)
The algorithm uses a threshold function which makes use of a real number δ .
Lemma 8.13. Let G be an r-graph on vertex set [n] with average degree d. Let
Pr = E(G) and let Pr−1 , . . . , P1 be the multisets constructed during the algorithm,
either in build mode or in prune mode. Then
by Lemma 8.13 with U = [n]. Thus µ(Ts ) ≤ (τ/ζ )(1 + rδ ), and µ(T ) ≤ µ(T1 ) +
· · · + µ(Tr−1 ) ≤ (r − 1)(τ/ζ )(1 + rδ ).
Lemma 8.15. Let C be the set produced by the algorithm in build mode. Let
D = ([n] \ C) ∪ T ∪ B. Define es by the equation |Ps | = es τ r−s nd for 1 ≤ s ≤ r.
Then
Proof. The way the algorithm builds C means that T ∪B ⊆ C. Let C0 = C \(T ∪B),
so D = [n] \C0 . For v ∈ [n] let fs+1 (v) be the number of sets in Ps+1 for which v is
the first vertex in the vertex ordering. Then
|Ps+1 | = ∑ fs+1 (v) = ∑ fs+1 (v) + ∑ fs+1 (v) for 1 ≤ s < r. (8.7)
v∈[n] v∈C0 v∈D
By definition of |Fv,s |, of the fs+1 (v) sets in Ps+1 beginning with v, fs+1 (v) − |Fv,s |
of them contain some σ ∈ Γs . If v ∈ C0 then v ∈ / B and v ∈
/ T and so, since v ∈ C,
we have |Fv,s | < ζ τ r−s−1 d(v). Therefore, writing PΓ for the multiset of edges in
Ps+1 that contain some σ ∈ Γs , we have
Finally, making use of (8.7) and (8.8) together with Lemma 8.13, we have
The bound (8.9) for ∑σ ∈Γs ds+1 (σ ) now gives the result claimed.
Proof of Theorem 8.9. We begin by choosing the constant c. Let γ = 2r1 r2 and
25r 2
c = γ r . Let
√ G be as in the theorem and let τ be chosen so that (8.5) is satisfied.
Let ζ = 2rγ. For later use, we note c ≤ γ ≤ ζ /2r ≤ 2rζ < 1.
As might be expected, we prove the theorem by using the containers C and the
sets T supplied by the algorithm. However, the input parameters we supply to the
algorithm are not τ and ζ as just defined, but instead τ∗ = γτ and ζ .
We therefore remind the reader that the values of τ and ζ appearing in the
lemmas above are those values input to the algorithm. Hence in the present case,
where we are using inputs τ ∗ and ζ , the conclusions of the lemmas hold with τ ∗
in place of τ. Again, as highlighted earlier, the value of δ in the lemmas is that
supplied by Definition 8.11 with τ ∗ in place of τ. Explicitly, δ is (by definition)
minimal such that d(σ ) ≤ δ dτ ∗(|σ |−1) for all σ . Now τ was chosen to satisfy (8.5),
so we know that d(σ ) ≤ cdτ (|σ |−1) . Since h = γ r this implies we know, for all σ ,
that d(σ ) ≤ γ r dτ (|σ |−1) ≤ γdτ ∗(|σ |−1) , because γ ≤ 1 and |σ | ≤ r. Consequently,
by the minimality of δ , we have δ ≤ γ.
What remains is to verify the claims of the theorem. Condition (a) follows
from the general properties of the algorithm.
Now cτ r−1 = γτ∗r−1 ≤ (ζ /r)τ∗r−1 , and cτ r = τ∗r ≤ (r/ζ )τ∗r . So, by Lemma 8.14,
µ(T ) ≤ (rτ∗ /ζ )(1 + rδ ) ≤ 2rτ∗ /ζ = 2rγτ/ζ = ζ τ, easily establishing condi-
tion (b). Moreover T ∩ B = 0, / so |T |ζ d ≤ ∑v∈T d(v) = ndµ(T ) ≤ ndζ τ, giving
condition (c). To show that condition (d) holds, note that 2rδ ≤ 2rγ ≤ ζ , and so
by Lemma 8.15 we comfortably have es+1 ≤ r2s es + µ(D) + 2ζ for r − 1 ≥ s ≥ 1.
s+1
Dividing the bound for es+1 by rs+1 2( 2 ) and adding over s = 1, . . . , r − 1, we
obtain
er 1 1 1 1 1 2
r ≤ (µ(D) + 2ζ ) 2
+ 3 3 + 4 6 + · · · ≤ (µ(D) + 2ζ ) 2 .
rr 2(2) r r 2 r 2 r
r
Recall that er nd = |Pr | = e(G) = nd/r so er = 1/r. Hence µ(D)+2ζ ≥ r−r 2−(2) =
5γ 1/2 2r/2 ≥ 5ζ . So µ(D) ≥ 3ζ . By definition, D = [n] − (C − (T ∪ B)). Thus
µ(C) ≤ 1 − µ(D) + µ(T ) + µ(B). We showed previously that µ(T ) ≤ ζ τ ≤ ζ .
156 CHAPTER 8. EXTREMAL PROPERTIES
Proof of Corollary 8.10. Write c∗ for the constant c from Theorem 8.9. We prove
the corollary with c = ε`−r c∗ , where ` = d(log ε)/ log(1 − c∗ )e. Let G, I and τ
be as stated in the corollary. We shall apply Theorem 8.9 several times. Each
time we apply the theorem, we do so with with τ∗ = τ/` in place of τ, with the
same I, but with different graphs G, as follows (we leave it till later to check that
the necessary conditions always hold). Given I, apply the theorem to find T1 ⊆ I
and I ⊆ C1 = C(T1 ), where |T1 | ≤ τ∗ n and µ(C1 ) ≤ 1 − c∗ . It is easily shown
that e(G[C1 ]) ≤ µ(C1 )e(G) ≤ (1 − c∗ )e(G) see (8.4). Now I is independent in the
graph G[C1 ] so apply the theorem again, to the r-graph G[C1 ], to find T2 ⊆ I and a
container I ⊆ C2 . We have |T2 | ≤ τ∗ |C1 |, and e(G[C2 ]) ≤ (1 − c∗ )e(G[C1 ]) ≤ (1 −
c∗ )2 e(G). We note that, in the first application, the algorithm in build mode would
have constructed C1 from input T1 ∪ T2 , and would likewise have constructed C2
from input T1 ∪ T2 in the second application. Thus C2 is a function of T1 ∪ T2 .
We repeat this process k times until we obtain the desired container C = Ck with
e(G[C]) ≤ εe(G). Since e(G[C]) ≤ (1 − c∗ )k e(G) this occurs with k ≤ `. Put
T = T1 ∪ · · · ∪ Tk . Then C is a function of T ⊆ I.
We must check that the requirements of Theorem 8.9 are fulfilled at each ap-
plication. Observe that, if d j is the average degree of G[C j ] for j < k, then |C j |d j =
re(G[C j ]) > rεe(G) = εnd, and since |C j | ≤ n we have d j ≥ εd. The conditions of
|σ |−1
Corollary 8.10 mean that d(σ ) ≤ cdτ |σ |−1 = ε`−r c∗ dτ |σ |−1 < c∗ d j τ∗ ; since
the degree of σ in G[C j ] is at most d(σ ), this means that (8.5) is satisfied every
time Theorem 8.9 is applied.
Finally condition (c) of the theorem implies |T j | ≤ τ∗ |C j | ≤ τ∗ n = τn/`, and
so |T | ≤ kτn/` ≤ τn, giving condition (b) of the corollary and completing the
proof.
H-free graphs
In this section we prove Theorem 8.7. We will apply the container theorem given
by Corollary 8.10 to the following hypergraph, whose independent sets corre-
spond to H-free `-graphs on vertex set [N].
Definition 8.16. Let H be an `-graph. Let r = e(H). The r-graph GH has vertex
set V (GH ) = [N] V (GH )
` , where B = {v1 , ..., vr } ∈ r is an edge whenever B, con-
sidered as an `-graph with vertices in [N] and with r edges, is isomorphic to H. So
B ∈ Mr where M = [N]
` .
8.4. CONTAINERS AND THE PROOF OF THEOREM 8.1 157
All that remains before applying the container theorem to GH is to verify (8.5).
Lemma 8.17. Let H be an `-graph with r = e(H) ≥ 2 and let τ = 2`!v(H)!N −1/m(H) .
N
Let N be sufficiently large. Suppose that e(GH ) = αH v(H) where αH ≥ 1 depends
re(GH ) rαH N v(H)
only on H. The average degree d in GH satisfies d = = . Then,
v(GH ) (N` )
1 |σ |−1
d(σ ) ≤ dτ , holds for all σ , |σ | ≥ 2.
αr
Proof of Theorem 8.7. Let η = η(ε, H) be given by Proposition 8.18, and let β =
min{ε, η}. Recall that r = e(H). Apply Corollary 8.10 to GH with c = α1H r and
τ = 2`!v(H)!N −1/m(H) and with β playing the role of ε in the corollary. The
conditions of Corollary 8.10 are satisfied; denote by c̃ the constant c appearing in
the corollary. The collection of containers C satisfies the following.
• For every independent set I there exists some C ∈ C with I ⊆ C. This implies
condition (a) of the theorem,
158 CHAPTER 8. EXTREMAL PROPERTIES
• Finally, for every set I as above, there exists T ⊆ I such that C = C(T ),
|T | ≤ c̃τ N` . This implies condition (c).
8.5 Exercises
8.5.1 An edge e of G is H-open if it is contained in at most one copy of H and H-
closed otherwise. The H-core ĜH of G is obtained by repeatedly deleting
H-open edges. Show that G → (H)e2 implies that ĜH 0 → (H 0 )e2 for every
H 0 ⊆ H. (Thus one only needs to prove the 0-statement of Theorem 8.2
for strictly 2-balanced H. A graph H is strictly 2-balanced if H 0 = H is the
unique maximiser in (8.1)).
8.5.3 Show that there exists a sufficiently small c > 0 and a constant L = L(H, c)
such that if H is 2-balanced and p ≤ cn−1/m2 (H) then w.h.p. every inclusion
minimal H-closed subgraph of Gn,p has size at most L. (Try c = o(1) first
here).
8.5.4 Show that if e(G)/v(G) ≤ m2 (H) and m2 (H) > 1 then G 6→ (H)e2 .
8.5.5 Show that if H is 2-balanced and p = cn−1/m2 (H) then w.h.p. every subgraph
G of Gn,p with v(G) ≤ L = O(1) satisfies e(G)/v(G) ≤ m2 (H).
8.6 Notes
The largest triangle-free subgraph of a random graph
Babai, Simonovits and Spencer [45] proved that if p ≥ 1/2 then w.h.p. the largest
triangle-free subgraph of Gn,p is bipartite. They used Szemerédi’s regularity
8.6. NOTES 159
lemma in the proof. Using the sparse version of this lemma, Brightwell, Pana-
giotou and Steger [164] improved the lower bound on p to n−c for some (unspec-
ified) positive constant c. DeMarco and Kahn [244] improved the lower bound
to p ≥ Cn−1/2 (log n)1/2 , which is best possible up to the value of the constant C.
And in [245] they extended their result to Kr -free graphs.
Anti-Ramsey Property
Let H be a fixed graph. A copy of H in an edge colored graph G is said to be
rainbow colored if all of its edges have a different color. The study of rainbow
copies of H was initiated by Erdős, Simonovits and Sós [291]. An edge-coloring
of a graph G is said to be b-bounded if no color is used more than b times. A
graph is G said to have property A (b, H) if there is a rainbow copy of H in
every b-bounded coloring. Bohman, Frieze, Pikhurko and Smyth [121] studied
the threshold for Gn,p to have property A (b, H). For graphs H containing at
least one cycle they prove that there exists b0 such that if b ≥ b0 then there exist
c1 , c2 > 0 such that
(
0 p ≤ c1 n−1/m2 (H)
lim P(Gn,p ∈ A (b, H)) = . (8.10)
n→∞ 1 p ≥ c2 n−1/m2 (H)
A reviewer of this paper pointed out a simple proof of the 1-statement. Given a b-
bounded coloring of G, let the edges colored i be denoted ei,1 , ei,2 , . . . , ei,bi where
bi ≤ b for all i. Now consider the auxiliary coloring in which edge ei, j is colored
with j. At most b colors are used and so in the auxiliary coloring there will be a
monochromatic copy of H. The definition of the auxiliary coloring implies that
this copy of H is rainbow in the original coloring. So the 1-statement follows
directly from the results of Rödl and Ruciński [681], i.e. Theorem 8.2.
Nenadov, Person, Škorić and Steger [627] gave further threshold results on
both Ramsey and Anti-Ramsey theory of random graphs. In particular they proved
that in many cases b0 = 2 in (8.10).
160 CHAPTER 8. EXTREMAL PROPERTIES
I NPUT
an r-graph G on vertex set [n], with average degree d
parameters τ, ζ > 0
in prune mode a subset I ⊆ [n]
in build mode a subset T ⊆ [n]
O UTPUT
in prune mode a subset T ⊆ [n]
in build mode a subset C ⊆ [n]
I NITIALISATION
put B = {v ∈ [n] : d(v) < ζ d}
evaluate the thresholds θs (σ ), σ ∈ [n](≤s) , 1 ≤ i ≤ r
A: put Pr = E(G), Ps = 0,/ Γs = 0,/ s = 1, 2, . . . , r − 1
in prune mode put T = 0/
in build mode put C = [n]
for v = 1, 2, . . . , n do:
for s = 1, 2, . . . , r − 1 do:
let Fv,s = { f ∈ [v + 1, n](s) : {v} ∪ f ∈ Ps+1 , and 6 ∃ σ ∈ Γs , σ ⊆ f }
[here Fv,s is a multiset with multiplicities inherited from Ps+1 ]
if v ∈/ B, and |Fv,s | ≥ ζ τ r−s−1 d(v) for some s
in prune mode if v ∈ I, add v to T
in build mode if v ∈ / T , remove v from C
if v ∈ T then for s = 1, 2, . . . , r − 1 do:
add Fv,s to Ps
for each σ ∈ [v + 1, n](≤s) , if ds (σ ) ≥ θs (σ ), add σ to Γs
Resilience
Sudakov and Vu [729] introduced the idea of the local resilience of a monotone
increasing graph property P. Suppose we delete the edges of some graph H
on vertex set [n] from Gn,p . Suppose that p is above the threshold for Gn,p to
have the property. What can we say about the value ∆ so that w.h.p. the graph
G = ([n], E(Gn,p ) \ E(H)) has property P for all H with maximum degree at
most ∆? We will denote the maximum ∆ by ∆P
In this chapter we discuss the resilience of various properties. In Section 9.1
we discuss the resilience of having a perfect matching. In Section 9.2 we discuss
the resileince of having a Hamilton cycle. In Section 9.3 we discuss the resilience
of the chromatic number.
[n] into two subsets X,Y of sizes m + 1 and m − 1 respectively. Now delete all
edges inside X so that X becomes an independent set. Clearly, the remaining
graph contains no perfect matching. The Chernoff bounds, Corollary 22.7, imply
that we have deleted ≈ np/2 edges incident with each vertex.
The lower bound requires a little more work. Theorem 3.4 implies that w.h.p.
the minimum degree in G is at least (1 − o(1)) 12 + ε np. We randomly partition
161
162 CHAPTER 9. RESILIENCE
np
PM2 e(S, T ) ≤ 1 + ε3 4 |S| for all S ⊆ X, T ⊆ Y, |S| = |T | ≤ n/4.
Property PM1 follows immediately from the Chernoff bounds, Corollary 22.7,
1
and the fact that dG (v) & 2 + ε np log n.
Property PM2 is derived as follows:
Given, PM1, PM2, we see that if there exists S ⊆ X, |S| ≤ n/4 such that |NX (S)| ≤
|S| then for T = NX (S),
1 ε ε np
+ np|S| . e(S, T ) ≤ 1 + |S|,
4 2 3 4
contradiction. We finish the proof that Hall’s condition holds, i.e. deal with |S| >
n/4 just as we did for |S| > n/2 in Theorem 6.1.
Going even further, Montgomery [607] and Nenadov, Steger and Trujić [631]
have given tight hitting time versions. The proofs in these papers rely on the use of
Pósa rotations, as in Chapter 6. Some recent papers have introduced the use of the
absorbing method from extremal combinatorics to related problems. The method
was initiated by Rödl, Ruciński and Szemerédi [679]. Our purpose in this section
is to give an example of this important technique. Our exposition closely follows
the paper of Ferber, Nenadov, Noever, Peter and Trujić [308]. They consider the
resilience of Hamiltonicity in the context of random digraphs, but their proof can
be adapted and simplified when considering graphs. Their proof in turn utilises
ideas from Montgomery [607].
9.2. HAMILTON CYCLES 163
log10 n
Theorem 9.3. Suppose that p ≥ n . Then w.h.p. in Gn,p ,
1 1
− ε np ≤ ∆H ≤ + ε np
2 2
From our previous remarks, we can see that log10 n is not optimal. The proof
we give can be tightened, but probably not down to log n. The proof of Theorem
9.3 takes up the remainder of this section.
The proof of the upper bound is essentially the same as for Theorem 9.1. After
making X independent, there is no possibility of a Hamilton cycle.
The lower bound requires more work.
A pseudo-random condition
We say that a graph G = (V, E) with |V | = n is (n, α, p)-pseudo-random if
10 log2 n
Q2 eG (S) ≤ |S| log3 n for all S ⊆ V, |S| ≤ p .
log2 n
Q3 eG (S, T ) ≤ 1 + α4 |S||T |p for all disjoint S, T ⊆ V , |S|, |T | ≥ p .
Proof. Q1: This follows from the fact that w.h.p. every vertex of Gn,p has degree
(1 + o(1))np, see Theorem 3.4(ii).
Q2: We show that this is true w.h.p. in G and hence in G. Indeed,
10 log2 n
3
P ∃S : eG (S) ≥ |S| log n and |S| ≤ ≤
p
10p−1 log2 n s 10p−1 log2 n log3 n !s
n 2 3 ne sep
∑ 3 ps log n ≤ ∑ ≤
s=log n s s log n s=log n s 2 log3 n
10p−1 log2 n 3 !s
5e log n
ne
∑ = o(1).
s=log n s log n
164 CHAPTER 9. RESILIENCE
Q3: We show that this is true w.h.p. in G and hence in G. We first note that
the Chernoff bounds, Corollary 22.7, imply that
α 2
P eG (S, T ) ≥ 1 + |S||T |p ≤ e−α |S||T |p/50 .
4
So,
log2 n
P ∃S, T : |S|, |T | ≥ and eG (S, T ) ≥ (1 + α)|S||T |p ≤
p
2
!s 2
!t
n n
ne1−α t p/100 ne1−α sp/100
n n −α 2 st p/50
∑ 2 s t e ≤ ∑ 2 ≤
−1 −1
s t
s,t=p log n s,t=p log n
2 2
!s 2 2
!t
n
ne1−α log n/100 ne1−α log n/100
∑ 2 ≤
−1
s t
s,t=p log n
2 log2 n/100
!s 2
n 1−α
∑ ne = o(1).
−1 2 s
s=p log n
The proof now rests on two lemmas: the following quantities are fixed for the
remainder of the proof:
4 log3 n
` = 12 dlog ne + 3 and t = (9.2)
p
C1 |K| `t logt.
1
C2 For every v ∈ K ∪ L we have |dK (v)| & 2 + α p|K|.
With these two lemmas in hand, we can show that G is Hamiltonian. Let P∗
be as in Lemma 9.6 and let U = (V2 ∪V3 ∪V4 ∪V5 ) \V (P∗ ). If v ∈ U then
1
dU (v) ≥ dV5 (v) & + α |V5 |p
2
1 5 + 7α 1 3α
& +α np ≥ + |U|p. (9.3)
2 5 + 10α 2 4
|U| log5 n
j k j k
n
Next let k = n and s = log5 n
. Randomly choose disjoint sets
S1 , S2 , . . . , Sk ⊆ U of size s and let S = ki=1 Si and S0 = U \ S. It follows from (9.3)
S
that w.h.p.
1 α
dSi (v) ≥ + |Si |p for all i ∈ [k], v ∈ U. (9.4)
2 2
Claim 6. Assuming (9.4), we see that there is a perfect matching Mi between
Si , Si+1 for 1 ≤ i < k.
166 CHAPTER 9. RESILIENCE
which contradicts Q3. (If |Y | < p−1 log2 n then we can add arbitrary vertices from
Si+1 \Y to Y so that we can apply Q3.)
The case |X| > s/2 is dealt with just as we did for |S| > n/2 in Theorem 6.1. This
completes the proof of Claim 6.
End of proof of Claim 6
x1 Q1 y1
P1
x2 Q2 y2
P2
xs Qs ys
Pt
Pt 0 −1
xs+1 xt 0 = yt 0
Qt 0
xt P∗ yt
2 log2 n
D2 |NG (X,Y )| ≥ p .
1
log2 n
D3 |NG (S,Y )| ≥ 2 + α4 |Y | for all S ⊆ Y, |S| ≥ p .
1
Then there exists x ∈ X such that |NG` (x,Y )| ≥
2 + α8 |Y |
2 log2 n
Proof. We first show that there exists x ∈ X such that |NG`−1 (x,Y )| ≥ p . For
this we use the following claim:
2 log2 n
Claim 7. Let i < ` and A ⊆ X be such that |NGi (A,Y )| ≥ p . Then there exists
2
2 log n
A0 ⊆ A such that |A0 | ≤ d|A|/2e and |NGi+1 (A0 ,Y )| ≥ p .
We prove the claim below. Using D2 and the claim ` − 2 times, we obtain a set
2
X 0 ⊆ X such that |X 0 | ≤ |X|/2`−2 and |NG`−1 (X 0 ,Y )| ≥ 2 logp n . But ` − 2 ≥ log2 n
l 2 m
and so we have |X 0 | = 1. Let X 0 = {x} and M ⊆ NG (x,Y ) be of size logp n .
By definition, there is a path Pw of length ` − 1 from x to each w ∈ M. Let V ∗ =
( w∈M V (Pw )) \ {x}. Then D2 and D3 imply
S
` ∗ 1 α 1 α
|NG (x,Y )| ≥ |NG (M,Y \V )| ≥ + |Y | − `|M| ≥ + |Y |.
2 4 2 8
9.2. HAMILTON CYCLES 169
Proof of Claim 7
First note that if A1 , A2 is a partition of A with |A1 | = d|A|/2e then we have
2 log2 n
|NGi (A1 ,Y )| + |NGi (A2 ,Y )| ≥ |NGi (A,Y )| ≥ .
p
We can assume therefore that there exists A0 ⊆ A, |A0 | ≤ d|A|/2e such that |NGi (A0 ,Y )| ≥
log2 n
p .
log2 n
l m
Choose B ⊆ NGi (A0 ,Y ), |B| = 1
p . Then D3 implies that |NG (B,Y )| ≥ 2 + α4 |Y |.
Each v ∈ B is the endpoint of a path Pv of length i from a vertex in A0 . Let
V ∗ = v∈B V (Pv ). Then,
S
2 log2 n
1 α
|NGi+1 (A0 ,Y )| ≥ |NG (B,Y )| − |V ∗ | ≥ + |Y | − `|B| ≥ .
2 4 p
48t`
E1 |RA |, |RB | ≥ α .
1
+ α4 |RZ | for all S ⊆ RA ∪ RB ∪ ti=1 {ai , bi }
E2 For Z = A, B, |NG (S, RZ )| ≥
S
2
log2 n
such that |S| ≥ p .
Then there exists a set I ⊆ [t], |I| = bt/2c and internally disjoint paths Pi , i ∈ I such
that Pi connects ai to bi and V (Pi ) \ {ai , bi } ⊆ RA ∪ RB .
Proof. We prove this by induction. Assume that we have found s < bt/2c paths
Pi from ai to bi for i ∈ J ⊆ I, |J| = s. Then let
We verify Claim 8 below. Assume its truth for now. Let S = NGhA (ai , R0A ). Then
47t` log2 n
1 α 1 α
|S| ≥ + (RA − s`) ≥ + ≥ .
2 8 2 8 α p
Now from Claim 8 we have that |NGhB (bi , R0B )| ≥ 12 + α8 |R0B | and so
NGhA +1 (ai , R0B ) ∩ NGhB (bi , R0B ) 6= 0/ and there is a path as claimed.
It only remains to prove Claim 8. Assume inductively
hA 0 1 α
0 that we have found v1 , v2 , . . . , vk ∈
{ai : i ∈ K} such that |NG (vi , RA )| ≥ 2 + 16 |RA | for i ∈ [k]. The base case is
k = 0. We apply Lemma 9.8 with Y = R0A and X = {ai : i ∈ K} \ {v1 , v2 , . . . , vk }.
47` log n 3
We check that the lemma’s conditions are satisfied. |R0A | ≥ 48t`
α −t` ≥ αp and
0
so D1 is satisfied. On the other hand E2 implies that if S ⊆ RA ∪ {ai : i ∈ K} is of
log2 n
size at least p then
log2 n
1 α 1 α
|NG (S, R0A )| ≥ |NG (S, RA )| − t` ≥ + |RA | − t` ≥ + |RA | .
2 4 2 8 p
2
So, D3 is satisfied and also D2 if |X| ≥ logp n i.e. if k ≤ t/2, completing the
induction. So, we obtain IA ⊆ [t], |IA | = bt/2c + 1 such that
1 α
|N hA
(ai , R0A )| ≥ + |R0A | for i ∈ IA .
2 8
h 1 α
0 the existence of IB ⊆ [t], |IB | = bt/2c + 1 such that
A similar argument proves
0
|N (bi , RB )| ≥ 2 + 8 |RB | for i ∈ IB and the claim follows, since IA ∩ IB 6= 0/ and
B
|K|
m = dlog2 te + 1, si = 2t, i ∈ [2m], s2m+1 = s2m+2 = , k = 2m + 2.
4
9.2. HAMILTON CYCLES 171
Thus
!
t s−1
|K| t |K| |K|
|RA | ≥
4
−O ∑ ∑ 2 j · 2 j ` = 4 − O(`t logt) ≥ 5 `t logt,
i=1 j=1
and similarly for RB and so E2 holds. It now follows from Lemma 9.9 that there
are bt/2c indices i for which there is a path from x j to y j and these yield at least
dt/2s e indices Is and a path from ai to bi for i ∈ Is .
It remains only to prove Lemma 9.10.
Proof. Let VA (s) denote the endpoints of Qs in Ss . Because of (9.6), we can reduce
this to the following: given a bipartite graph Γ with bipartion X = x1 , x2 , . . . , xbt/2c ⊆
Ss \VA (s), B = Ss+1 = {y1 , y2 , . . . , y2t } and minimum degree at least 21 + α2 psi+1 ≥
4 log3 n and
such that Q2, Q3 hold, there exists
a partition of B into t pairs
zi,1 , zi,2 , i ∈ [t] such that both edges xi , zi,l , l = 1, 2 exist in Γ. (The reader
can check that after each round, there are t/2 vertices that are leaves of the current
active trees, and need two neighbors to grow the tree. We say that a tree is active
if its root is not the endpoint of a path in Qs .)
We need to verify the following condition:
2 log2 n
Case 1: |S| ≤ 3p .
2
Let T = NΓ (S, B). If (9.7) fails then |S ∪ T | ≤ 2 logp n and eG (S ∪ T ) > |S| log3 n,
violating Q2.
2n
Case 2: |S| > 2 log3p .
If (9.7) fails then
1 α α 2
+ si |S|p ≤ eG (S, T ) ≤ 2 1 + |S| p. (9.8)
2 2 4
The lower bound is from (9.6) and the upper bound in (9.8) is from Q3. Equation
(9.8) is a contradiction, because si = 2t ≥ 4|S|.
9.2. HAMILTON CYCLES 173
sx x tx
s1x tx1 s2x tx2 six txi s2k 2k
x tx
Figure 9.2: The absorber for k1 = 3. The cycle Cx is drawn with solid lines. The
dashed lines represent paths. The part inside the rectangle can be repeated to make
larger absorbers.
Let k, ` be integers and consider the graph Ax with 3 + 2k(` + 1) vertices con-
structed as follows:
Proof. We take
At this point we have paths Px , Px0 for x ∈ V1 . We finally construct P∗ , using Lemma
9.5 to connect the paths Px0 , x ∈ V1 . We let t = h − 1 where V1 = {x1 , x2 , . . . , xh }.
We take ai = txi and bi = sxi+1 for i ∈ [h − 1] and K = V4 .
It is easy to see that this construction has the desired property. Where necessary,
we can absorb x ∈ V1 by replacing Px0 by Px .
n
W.h.p. every set of t ≤ vertices, span fewer than
10 log2 n
2npt
edges of Gn,p . (9.9)
log2 n
Indeed,
n/10 log2 n
t
n 2
Pr(∃ a set negating (9.9)) ≤ ∑ 2 t 2npt/ log2 n p2npt/ log n
2
t=2np/ log n
n/10 log2 n ne t t 2 ep log2 n 2npt/ log2 n
≤ ∑ ≤
t 4npt
t=2np/ log2 n
!t
n/10 log2 n 2n
ne te 2np/ log
∑ · = o(1).
t 4n
t=2np/ log2 n
We let s = 20∆ log2 n and randomly partition [n] into s sets V1 ,V2 , . . . ,Vs of size
n/s. Let Y denote the number of edges of H that have endpoints in the same set
of the partition. Then E(Y ) ≤ |E(H)|
s ≤ ∆n
2s . It follows from the Markov inequality
1 n
that Pr Y ≥ s ≤ 2 and so there exists a partition for which Y ≤ ∆n
∆n
s = 20 log2 n .
9.4. EXERCISES 175
Furthermore, it follows from (7.21), that w.h.p. the subgraphs Gi of Gn,p induced
by each Vi have chromatic number ≈ 2 logn/s(n/s) .
b
Given V1 ,V2 , . . . ,Vs , we color G as follows: we color the edges of Gi with ≈
n/s n n
2 logb (n/s) colors, using different colors for each set and ≈ 2 logb (n/s) ≈ 2 logb (n) ≈
χ(Gn,p ) colors overall. We must of course deal with the at most Y edges that could
be improperly colored. Let W denote the endpoints of these edges. Then |W | ≤
n
10 log2 n
. It follows from (9.9) that we can write W = {w1 .w2 , . . . , wm } such that wi
2np
has at most log2 n
+ ∆ neighbors in {w1 , w2 , . . . , wi−1 } i.e. the coloring number of
2np
the subgraph of Gn,p induced by W is at most log2 n
+ ∆. It follows that we can
2np
re-color the Y badly colored edges using at most log2 n + ∆ + 1 = o(χ(Gn,p )) new
colors.
9.4 Exercises
9.4.1 Prove that if p ≥ (1+η)n log n for a postive constant η then ∆C ≥ 1
2 − ε np,
where C denotes connectivity. (See Haller and Trujić [407].)
9.4.2 Show that for every ε > 0 there exists cε > 0 such that the following is true
w.h.p. If c ≥ cε and p = c/n and we remove any set of at most (1 − ε)cn/2
edges from Gn,p , then the remaining graph contains a component of size at
least εn/2.
9.5 Notes
Sudakov and Vu [729] were the first to discuss local resilience in the context of
random graphs. Our examples are taken from this paper except that we have given
a proof of hamiltonicity that introduces the absorbing method.
Hamiltonicity
4
Sudakov and Vu proved local resilience for p ≥ logn n and ∆H = (1−o(1))np
2 . The
expression for ∆H is best posible, but the needed value for p has been lowered.
Frieze and Krivelevich [349] showed that there exist constants K, α such that
w.h.p. ∆H ≥ αnp for p ≥ K log n
n . Ben-Shimon, Krivelevich and Sudakov [79]
improved this to α ≥ 1−ε6 holds w.h.p. and then in [80] they obtained a result
on resilience for np − (log n + log log n) → ∞, but with K close to 13 . (Vertices
np
of degree less than 100 can lose all but two incident edges.) Lee and Sudakov
176 CHAPTER 9. RESILIENCE
[541] proved the sought after result that for every positive ε there exists C = C(ε)
such that w.h.p. ∆H ≥ (1−ε)np 2 holds for p ≥ C log n
n . Condon, Espuny Dı́az, Kim,
Kühn and Osthus [204] refined [541]. Let H be a graph with degree sequence
d1 ≥ d2 ≥ · · · ≥ dn where di ≤ (n − i)p − εnp for i < n/2. They say that G is
ε-Pósa-resilient if G − H is Hamiltonian for all such H. Given ε > 0 there is a
constant C = C(ε) such that if p ≥ C log n
n
then Gn,p is ε-Pósa-resilient w.h.p. The
result in [541] has now been improved to give a hitting time result, see Mont-
gomery [607] and Nenadov, Steger and Trujić [631]. The latter paper also proves
the optimal resilience of the 2-core when p = (1+ε) 3n
log n
.
Fischer, Škorić, Steger and Trujić [312] have shown that there exists C > 0 such
3n
that if p ≥ C nlog
1/2 then not only is there the square of a Hamilton cycle w.h.p., but
containing a square is resilient to the deletion of not too many triangles incident
with each vertex.
Krivelevich, Lee and Sudakov [522] proved that G = Gn,p , p n−1/2 remains
pancyclic w.h.p. if a subgraph H of maximum degree ( 12 − ε)np is deleted, i.e.
pancyclicity is locally resilient. The same is true for random regular graphs when
r n1/2 .
Hefetz, Steger and Sudakov [421] began the study of the resilience of Hamiltonic-
ity for random digraphs. They showed that if p log n1/2
n
then w.h.p. the Hamil-
1
tonicity of Dn,p is resilient to the deletion of up to ( 2 − o(1))np edges incident
8
with each vertex. The value of p was reduced to p logn n by Ferber, Nenadov,
Noever, Peter and Škorić [308]. Finally, Montgomery [609] proved that in the ran-
dom digraph process, at the hitting time for Hamiltonicity, the property is resilient
w.h.p.
Part II
177
Chapter 10
Inhomogeneous Graphs
Thus far we have concentrated on the properties of the random graphs Gn,m and
Gn,p . We first consider a generalisation of Gn,p where the probability of edge
(i, j) is pi j is not the same for all pairs i, j. We call this the generalized binomial
graph . Our main result on this model concerns the probability that it is connected.
For this model we concentrate on its degree sequence and the existence of a giant
component. After this we move onto a special case of this model, viz. the expected
degree model. Here pi j is proportional to wi w j for weights wi . In this model, we
prove results about the size of the largest components. We finally consider another
special case of the generalized binomial graph, viz. the Kronecker random graph.
Note that Qi is the probability that vertex i is isolated and λn is the expected
number of isolated vertices. Next let
Rik = min qi j1 · · · qi jk .
1≤ j1 < j2 <···< jk ≤n
179
180 CHAPTER 10. INHOMOGENEOUS GRAPHS
Suppose that the edge probabilities pi j are chosen in such a way that the following
conditions are simultaneously satisfied as n → ∞:
max Qi → 0, (10.1)
1≤i≤n
Theorem 10.1. Let X0 denote the number of isolated vertices in the random graph
Gn,P . If conditions (10.1) (10.2) and (10.3) hold, then
λ k −λ
lim P(X0 = k) = e
n→∞ k!
for k = 0, 1, . . ., i.e., the number of isolated vertices is asymptotically Poisson
distributed with mean λ .
Proof. Let (
1 with prob. pi j
Xi j =
0 with prob. qi j = 1 − pi j .
Denote by Xi , for i = 1, 2, . . . n, the indicator of the event that vertex i is isolated
in Gn,P . To show that X0 converges in distribution to the Poisson random variable
with mean λ one has to show (see Theorem 21.11) that for any natural number k
!
λk
E ∑ X X
i1 i2 · · · Xik → (10.4)
1≤i1 <i2 <...<ik ≤n k!
as n → ∞. But
k
E (Xi1 Xi2 · · · Xik ) = ∏ P Xir = 1|Xi1 = . . . = Xir−1 = 1 , (10.5)
r=1
10.1. GENERALIZED BINOMIAL GRAPH 181
1
∑ Qi1 · · · Qik = ∑ Qi1 · · · Qik ≥
1≤i1 <···<ik ≤n k! 1≤i
1 6=···6=ir ≤n
!
1 k n 2
Q · · · Q − ∑ Qi Qi1 · · · Qik−2
k! 1≤i1 ∑ ∑
i1 ik
,...,ik ≤n k! i=1 1≤i1 ,...,ik−2 ≤n
λnk λk λk
≥ − (max Qi )λnk−1 = n − (max Qi )λnk−1 → , (10.7)
k! i k! i k!
as n → ∞.
Now,
n n
Qi
∑ ≥ λn = ∑ Qi,
i=1 Rik i=1
k
n/2 1
and if lim sup ∑ni=1 RQiki n Qi
> λ then lim sup ∑k=1 k! ∑i=1 Rik > eλ − 1, which con-
n→∞ n→∞
tradicts (10.3). It follows that
n
Qi
lim
n→∞
∑ Rik = λ .
i=1
Therefore !k
n
1 Qi λk
∑ Qi1 · · · Qik ≤ k! ∑ Rik →
k!
.
1≤i1 <...<ik ≤n i=1
182 CHAPTER 10. INHOMOGENEOUS GRAPHS
as n → ∞.
Combining this with (10.7) gives us (10.4) and completes the proof of Theorem
10.1.
One can check that the conditions of the theorem are satisfied when
log n + xi j
pi j = ,
n
where xi j ’s are uniformly bounded by a constant.
The next theorem shows that under certain circumstances, the random graph
Gn,P behaves in a similar way to Gn,p at the connectivity threshold.
Theorem 10.2. If the conditions (10.1), (10.2) and (10.3) hold, then
Proof. To prove the this we will show that if (10.1), (10.2) and (10.3) are satisfied
then w.h.p. Gn,P consists of X0 + 1 connected components, i.e., Gn,P consists of
a single giant component plus components that are isolated vertices only. This,
together with Theorem 10.1, implies the conclusion of Theorem 10.2.
Let U ⊆ V be a subset of the vertex set V . We say that U is closed if Xi j = 0 for
every i and j, where i ∈ U and j ∈ V \ U. Furthermore, a closed set U is called
simple if either U or V \ U consists of isolated vertices only. Denote the number
of non-empty closed sets in Gn,P by Y1 and the number of non-empty simple sets
by Y . Clearly Y1 ≥ Y .
We will prove first that
λk
lim P(Y = 2k+1 − 1) = e−λ .
n→∞ k!
Observe that for any ` ≥ 0,
`
EY ≥ ∑ (2k+1 − 1) P(Y = 2k+1 − 1)
k=0
10.1. GENERALIZED BINOMIAL GRAPH 183
and hence
`
λ k e−λ
lim inf EY ≥
n→∞
∑ (2k+1 − 1) k!
.
k=0
So,
`
λ k e−λ
lim inf EY ≥ lim
n→∞
∑ (2k+1 − 1)
`→∞ k=0 k!
= 2eλ − 1
To prove (10.9) denote by Zk the number of closed sets of order k in Gn,P so that
Y1 = ∑nk=1 Zk . Note that
Zk = ∑ Zi1 ...ik ,
i1 <...<ik
Hence !k
n
Qi 1 Qi
E Zk ≤ ∑ ∏ ≤ ∑ Rik .
i1 <...<ik i∈Ik Rik k! i=1
n/2
lim sup ∑ E Zk ≤ eλ − 1.
n→∞ k=1
To complete the estimation of E Zk (and thus for EY1 ) consider the case when
k > n/2. For convenience let us switch k with n − k, i.e, consider E Zn−k , when
0 ≤ k < n/2. Notice that E Zn = 1 since V is closed. So for 1 ≤ k < n/2
E Zn−k = ∑ ∏ qi j .
i1 <...<ik i∈Ik , j6∈Ik
184 CHAPTER 10. INHOMOGENEOUS GRAPHS
i.e., asymptotically, the probability that there is a closed set that is not simple,
tends to zero as n → ∞. It is easy to check that X0 < n w.h.p. and therefore
Y = 2X0 +1 − 1 w.h.p. and so w.h.p. Y1 = 2X0 +1 − 1. If Gn,P has more than X0 + 1
connected components then the graph after removal of all isolated vertices would
contain at least one closed set, i.e., the number of closed sets would be at least
2X0 +1 . But the probability of such an event tends to zero and the theorem follows.
∑ pi j ≥ c log n,
i∈S, j∈V \S
Hence the expected number of isolated vertices of Hi does not exceed |U|e−1 .
Therefore, by the Markov inequality, it is at most 2|U|e−1 with probability at least
1/2. But in this case the number of connected components of Hi is at most
−1 1 −1 1 −1
2|U|e + (|U| − 2|U|e ) = +e |U| < 0.9|U|,
2 2
and so (10.12) follows. Observe that if Ck > 1 then the total number of successful
stages is strictly less than log n/ log 0.9 < 10 log n. However, by (10.12), the prob-
ability of this event is at most the probability that a Binomial random variable with
parameters k and 1/2 will attain a value at most 10 log n. It follows from (22.22)
that if k = c log n = (20 + t) log n then the probability that Ck > 1 (i.e., that G0p0
e
2
is disconnected) is at most n−t /4c . This completes the proof of (10.11) and the
theorem follows.
We assume that maxi w2i < W so that pi j ≤ 1. The resulting graph is denoted as
Gn,Pw . Note that putting wi = np for i ∈ [n] yields the random graph Gn,p .
Notice that loops are allowed here but we will ignore them in what follows.
Moreover, for vertex i ∈ V its expected degree is
wi w j
∑ = wi .
j W
Denote the average vertex weight by w (average expected vertex degree) i.e.,
W
w= ,
n
while, for any subset U of a vertex set V define the volume of U as
w(U) = ∑ wk .
k∈U
10.2. EXPECTED DEGREE MODEL 187
Chung and Lu in [187] and [189] proved the following results summarized in
the next theorem.
Theorem 10.4. The random graph Gn,Pw with a given expected degree sequence
has a unique giant component w.h.p. if the average expected degree is strictly
greater than one (i.e., w > 1). Moreover, if w > 1 then w.h.p. the giant component
has volume √
λ0W + O n(log n)3.5 ,
where λ0 is the unique nonzero root of the following equation
n n
∑ wie−wiλ = (1 − λ ) ∑ wi,
i=1 i=1
(1 + o(1)µ(w) log n,
where (
1/(w − 1 − log w) if 1 < w < 2,
µ(w) =
1/(1 + log w − log 4) if w > 4/e.
Here we will prove a weaker and restricted version of the above theorem. In
the current context, a giant component is one with volume Ω(W ).
Theorem 10.5. If the average expected degree w > 4, then a random graph Gn,Pw
w.h.p. has a unique giant component and its volume is at least
2
1− √ W
ew
while the second-largest component w.h.p. has the size at most
log n
(1 + o(1)) .
1 + log w − log 4
The proof is based on a key lemma given below, proved under stronger condi-
tions on w than in fact Theorem 10.5 requires.
4
Lemma 10.6. For any positive ε < 1 and w > e(1−ε) 2 w.h.p. every connected
component in the random graph Gn,Pw either has volume at least εW or has at
log n
most 1+log w−log 4+2 log(1−ε) vertices.
188 CHAPTER 10. INHOMOGENEOUS GRAPHS
where
1 1
=ρ := .
W nw
So, the probability that S induces a connected subgraph of our random graph can
be bounded from above by
∑ P(T ) = ∑ ∏ wi j wil ρ,
T T {vi j ,vi }∈E(T )
l
where the sum ranges over all S ⊆ V, |S| = k. Now, we focus our attention on k-
vertex components whose volume is at most εW . We call such components small
or ε-small. So, if Yk is the number of small components of size k in Gn,Pw then
The function x2k−2 e−x(1−ε) achieves its maximum at x = (2k − 2)/(1 − ε). There-
fore
2k − 2 2k−2 −(2k−2)
k−1
n ρ
f (k) ≤ e
k kk 1−ε
ne k ρ k−1 2k − 2 2k−2
≤ e−(2k−2)
k kk 1−ε
2k
(nρ)k
2
≤ 2
e−k
4ρ(k − 1) 1 − ε
k
1 4
=
4ρ(k − 1)2 ew(1 − ε)2
e−ak
= ,
4ρ(k − 1)2
where
a = 1 + log w − log 4 + 2 log(1 − ε) > 0
under the assumption of Lemma 10.6.
Let k0 = loga n . When k satisfies k0 < k < 2k0 we have
1 1
f (k) ≤ =o ,
4nρ(k − 1)2 log n
190 CHAPTER 10. INHOMOGENEOUS GRAPHS
2 log n
while, when a ≤ k ≤ n, we have
1 1
f (k) ≤ 2 =o .
4n ρ(k − 1)2 n log n
So, the probability that there exists an ε-small component of size exceeding k0 is
at most
log n 1 1
∑ f (k) ≤ a × o log n + n × o n log n = o(1).
k>k0
To prove Theorem 10.5 assume that for some fixed δ > 0 we have
4 2
w = 4+δ = 2
where ε = 1 − (10.16)
e (1 − ε) (ew)1/2
Hence
1/2 1/3 δ
W ≤n +2 1+ n1/2 .
8
This is a contradiction since for our choice of w
W = nw ≥ 4(1 + δ )n.
10.2. EXPECTED DEGREE MODEL 191
1 + δ8
wi w j ρ ≥ w2i0 ρ ≥ .
i0
So the asymptotic behavior of G can be approximated by a random graph Gn,p
with n = i0 and p > 1/i0 . So, w.h.p. G has a component of size Θ(i0 ) = Ω(n1/3 ).
Applying Lemma 10.6 with ε as in (10.16) we see that any component with size
log n has volume at least εW .
Finally, consider the volume of a giant component. Suppose first that there
exists a giant component of volume cW which is ε-small i.e. c ≤ ε. By Lemma
10.6, the size of the giant component is then at most 2log n
log 2 . Hence, there must be
at least one vertex with weight w greater than or equal to the average
2cW log 2
w≥ .
log n
But it implies that w2 W , which contradicts the general assumption that all
pi j < 1.
We now prove uniqueness in the same way that we proved the uniqueness of
the giant component in Gn,p . Choose η > 0 such that w(1 − η) > 4. Then define
w0i = (1 − η)wi and decompose
Gn,Pw = G1 ∪ G2
w0i w0j
where the edge probability in G1 is p0i j = (1−η)W and the edge probability in G2
wi w j ηw w
is p00i j where 1 − W= (1 − p0i, j )(1 − p00i j ). Simple algebra gives p00i j ≥ Wi j . It
follows from the previous analysis that G1 contains between one and 1/ε giant
components. Let C1 ,C2 be two such components. The probability that there is no
G2 edge between them is at most
ηwi w j ηw(C1 )w(C2 )
∏ 1 − W ≤ exp − ≤ e−ηW = o(1).
i∈C1 W
j∈C2
Notice that
∑ j w2j W
= ≥ w2
= w.
W n
Chung and Lu [187] proved the following.
Theorem 10.7. If the average expected square degree w2 < 1 then, with proba-
2
w w2 √
bility at least 1 − 2 2 , all components of Gn,Pw have volume at most C n.
C 1−w
Proof. Let
x = P(∃S : w(S) ≥ Cn1/2 and S is a component).
Randomly choose two vertices u and v from V , each with probability proportional
to its weight.
√ Then, for√each vertex, the probability that it is in a set S with
w(S) ≥ C n is at least C nρ. Hence the probability that both vertices are in the
same component is at least
√
x(C nρ)2 = C2 xnρ 2 . (10.18)
On the other hand, for any two fixed vertices, say u and v, the probability Pk (u, v)
of u and v being connected via a path of length k + 1 can be bounded from above
as follows
Recall that the probabilities of u and v being chosen from V are wu ρ and wv ρ,
respectively. so the probability that a random pair of vertices are in the same
component is at most
2
wu wv ρ w2 ρ
∑ wuρ wvρ = .
u,v 1 − w2 1 − w2
which implies
2
w w2
x≤ ,
C2 1 − w2
and Theorem 10.7 follows.
where 0 < α, β , γ < 1, and let P[k] be the kth Kronecker power of P. Here P[k] is
obtained from P[k−1] as in the diagram below:
[k−1]
β P[k−1]
[k] αP
P =
β P[k−1] γP[k−1]
β 2 β γ γβ γ 2
Note that P[k] is symmetric and has size 2k × 2k .
We define a Kronecker random graph as a copy of Gn,P[k] for some k ≥ 1 and
n = 2k . Thus each vertex is a binary string of length k, and between any two such
vertices (strings) u, v we put an edge independently with probability
or equivalently
puv = α i β j γ k−i− j ,
where i is the number of positions t such that ut = vt = 1, j is the number of t
where ut 6= vt and hence k − i − j is the number of t that ut = vt = 0. We observe
that when α = β = γ then Gn,P[k] becomes Gn,p with n = 2k and p = α k .
194 CHAPTER 10. INHOMOGENEOUS GRAPHS
Connectivity
We will first examine, following Mahdian and Xu [570], conditions under which
is Gn,P[k] connected w.h.p.
k (k)
j
P(0 is isolated) = ∏(1 − p0v ) ≥ ∏ 1 − β j γ k− j
v j=0
(
k k j k− j )
j β γ
≥ exp − ∑ = e−1/ζ .
j=0 1−ζ
Now when α = β = 1, γ = 0, the vertex with all 1’s has degree n − 1 with proba-
bility one and so Gn,P[k] will be connected w.h.p. in this case.
It remains to show that the condition β + γ > 1 is also sufficient. To show
that β + γ > 1 implies connectivity we will apply Theorem 10.3. Notice that
the expected degree of vertex 0, excluding its self-loop, given that β and γ are
constants independent of k and β + γ > 1, is
(β + γ)k − γ k ≥ 2c log n,
∑ p0u ≥ c log n.
u∈S
10.3. KRONECKER GRAPHS 195
Take any vertices u, v and note that puv ≥ pu0 because we have assumed that
α ≥ β ≥ γ. Therefore
To add to the picture of the structure of Gn,P[k] when β + γ > 1 we state (with-
out proof) the following result on the diameter of Gn,P[k] .
Giant Component
We now consider when Gn,P[k] has a giant component (see Horn and Radcliffe
[429]).
Theorem 10.10. Gn,P[k] has a giant component of order Θ(n) w.h.p., if and only
if (α + β )(β + γ) > 1.
Proof. We prove a weaker version of the Theorem 10.10, assuming that for α ≥
β ≥ γ as in [570]. For the proof of the more general case, see [429].
We will show first that the above condition is necessary. We prove that if
(α + β )(β + γ) ≤ 1,
(α + β )(β + γ) = 1 − ε, ε > 0.
First consider those vertices with weight (counted as the number of 1’s in their
label) less than k/2 + k2/3 and let Xu be the degree of a vertex u with weight l
where l = 0, . . . , k. It is easily observed that
E Xu = (α + β )l (β + γ)k−l . (10.19)
Hence,
l k−l
l k−l
E Xu = ∑ puv = ∑ ∑ α i β j+l−i γ k−l− j
v∈V i=0 j=0 i j
l k−l
l
=∑ α i β l−i ∑ β j γ k−l−l
i=0 i j=0
and (10.19) follows. So, if l < k/2 + k2/3 , then assuming that α ≥ β ≥ γ,
2/3 2/3
E Xu ≤(α + β )k/2+k (β + γ)k−(k/2+k )
k2/3
k/2 α + β
= ((α + β )(β + γ))
β +γ
k2/3
α +β
=(1 − ε)k/2
β +γ
=o(1). (10.20)
Suppose now that l ≥ k/2 + k2/3 and let Y be the number of 1’s in the label of
a randomly chosen vertex of Gn,P[k] . Since EY = k/2, the Chernoff bound (see
(22.26)) implies that
k 4/3 1/3
P Y ≥ +k 2/3
≤ e−k /(3k/2) ≤ e−k /2 = o(1).
2
Therefore, there are o(n) vertices with l ≥ k/2 + k2/3 . It then follows from (10.20)
that the expected number of non-isolated vertices in Gn,P[k] is o(n) and the Markov
inequality then implies that this number is o(n) w.h.p.
Next, when α + β = β + γ = 1, which implies that α = β = γ = 1/2, then
random graph Gn,P[k] is equivalent to Gn,p with p = 1/n and so by Theorem 2.21
it does not have a component of order n, w.h.p.
To prove that the condition (α + β )(β + γ) > 1 is sufficient we show that the
subgraph of Gn,P[k] induced by the vertices of H of weight l ≥ k/2 is connected
w.h.p. This will suffice as there are at least n/2 such vertices. Notice that for any
vertex u ∈ H its expected degree, by (10.19), is at least
For the given vertex u let l be the weight of u. For a vertex v let i(v) be the
number of bits where ur = vr = 1, r = 1, . . . , k, while j(v) stands for the number
of bits where ur = 0 and vr = 1. Consider the partition
V \ H = S1 ∪ S2 ∪ S3 ,
where
S1 = {v : i(v) ≥ l/2, j(v) < (k − l)/2},
S2 = {v : i(v) < l/2, j(v) ≥ (k − l)/2},
S3 = {v : i(v) < l/2, j(v) < (k − l)/2}.
Next, take a vertex v ∈ S1 and turn it into v0 by flipping the bits of v which
correspond to 0’s of u. Surely, i(v0 ) = i(v) and
Notice that the weight of v0 is at least k/2 and so v0 ∈ H. Notice also that α ≥ β ≥
γ implies that puv0 ≥ puv . Different vertices v ∈ S1 map to different v0 . Hence
The same bound (10.23) holds for S2 and S3 in place of S1 . To prove the same
relationship for S2 one has to flip the bits of v corresponding to 1’s in u, while for
S3 one has to flip all the bits of v. Adding up these bounds over the partition of
V \ H we get
∑ puv ≤ 3 ∑ puv
v∈V \H v∈H
∑ ∑ puv ≥ 10 log n.
u∈S v∈H\S
198 CHAPTER 10. INHOMOGENEOUS GRAPHS
Without loss of generality assume that vertex 1 ∈ S. Equation (10.24) implies that
for any vertex u ∈ H either
∑ puv ≥ c log n, (10.25)
v∈S
or
∑ puv ≥ c log n. (10.26)
v∈H\S
Otherwise, (10.25) is true for every vertex u ∈ H. Since at least one such vertex
is in H \ S, we have
∑ ∑ puv ≥ c log n,
u∈S v∈H\S
10.4 Exercises
10.4.1 Prove Theorem 10.3 (with c = 10) using the result of Karger and Stein
[482] that in any weighted graph on n vertices the number of r-minimal
cuts is O (2n)2r . (A cut (S,V \ S), S ⊆ V, in a weighted graph G is called
r-minimal if its weight, i.e., the sum of weights of the edges connecting S
with V \ S, is at most r times the weight of minimal weighted cut of G).
10.4.2 Suppose that the entries of an n × n symmetric matrix A are all non-
negative. Show that for any positive constants c1 , c2 , . . . , cn , the largest
eigenvalue λ (A) satisfies
!
1 n
λ (A) ≤ max ∑ c j ai, j .
1≤i≤n ci j=1
10.4.3 Let A be the adjacency matrix of Gn,Pw and for a fixed value of x let
(
wi wi > x
ci = .
x wi ≤ x
1
Let m = max {wi : i ∈ [n]}. Let Xi = ci ∑nj=1 c j ai, j . Show that
m 2
E Xi ≤ w2 + x and Var Xi ≤ w + x.
x
10.5. NOTES 199
10.4.4 Apply Theorem 22.11 with a suitable value of x to show that w.h.p.
λ (A) ≤ w2 + (6(m log n)1/2 (w2 + log n))1/2 + 3(m log n)1/2 .
10.4.5 Show that if w2 > m1/2 log n then w.h.p. λ (A) = (1 + o(1))w2 .
10.4.8 Fix d ∈ N and let Zd denote the number of vertices of degree d in the
Kronecker random graph Gn,P[k] . Show that
k
k (α + β )dw (β + γ)d(k−w)
EZd = (1 + o(1)) ∑ ×
w=0 w d!
× exp −(α + β )w (β + γ)k−w + o(1).
or
EZd = o(2k ).
10.5 Notes
General model of inhomogeneous random graph
The most general model of inhomogeneous random graph was introduced by Bol-
lobás, Janson and Riordan in their seminal paper [148]. They concentrate on the
study of the phase transition phenomenon of their random graphs, which includes
as special cases the models presented in this chapter as well as, among others,
Dubins’ model (see Kalikow and Weiss [477] and Durrett [272]), the mean-field
scale-free model (see Riordan [673]), the CHKNS model (see Callaway, Hopcroft,
Kleinberg, Newman and Strogatz [175]) and Turova’s model (see [739], [740] and
[741]).
200 CHAPTER 10. INHOMOGENEOUS GRAPHS
The graph Gn,m is chosen uniformly at random from the set of graphs with vertex
set [n] and m edges. It is of great interest to refine this model so that all the graphs
chosen have a fixed degree sequence d = (d1 , d2 , . . . , dn ). Of particular interest
is the case where d1 = d2 = · · · = dn = r, i.e., the graph chosen is a uniformly
random r-regular graph. It is not obvious how to do this and this is the subject of
the current chapter. We discuss the configuration model in the next section and
show its usefulness in (i) estimating the number of graphs with a given degree
sequence and (ii) showing that w.h.p. random d-regular graphs are connected
w.h.p., for 3 ≤ d = O(1).
We finish by showing in Section 11.5 how for large r, Gn,m can be embedded
in a random r-regular graph. This allows one to extend some results for Gn,m to
the regular case.
Gn,d = {simple graphs with vertex set [n] s.t. degree d(i) = di , i ∈ [n]}
203
204 CHAPTER 11. FIXED DEGREE SEQUENCE
11
00
00
11 11
00
00
11 11
00
00
11 11
00
00
11
00
11 00
11 00
11 00
11
11
00
00
11 11
00
00
11 11
00
00
11 11
00
00
11 11
00
00
11
00
11 00
11 00
11 00
11 00
11
00
11 00
11 00
11 00
11 00
11 00
11 00
11
11
00 11
00 11
00 11
00 11
00 11
00 11
00
00
11 00
11 00
11 00
11 00
11 00
11 00
11
11
00 11
00 11
00 11
00 11
00 11
00 11
00 11
00
00
11 00
11 00
11 00
11 00
11 00
11 00
11 00
11
00
11 00
11 00
11 00
11 00
11 00
11 00
11 00
11
1 2 3 4 5 6 7 8
00
11 00
11 00
11 00
11
11
00
00
11 11
00
00
11 11
00
00
11 11
00
00
11
00
11 00
11 00
11 00
11 00
11
11
00
00
11 11
00
00
11 11
00
00
11 11
00
00
11 11
00
00
11
11
00 11
00 11
00 11
00 11
00 11
00 11
00
00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11 00
11
00
11
00
11 00
11 00
11 00
11 00
11 00
11 00
11 11
00
11
00 11
00 11
00 11
00 11
00 11
00 11
00 00
11
00
11
00
11 00
11 00
11 00
11 00
11 00
11 00
11
1 2 3 4 5 6 7 8
2 11
00
11
00
00
11
3
11
00
00
11
00
11
111
00 4
00
11
00
11 11
00
00
11
00
11
00
11
11
00 11
00
00
11 00
11
8 00
11 5
00
11 11
00
00
11
11
00 00
11
00
11
7 6
Algorithm F-GENERATOR
begin
U ←− W, F ←− 0/
for t = 1, 2, . . . , m do
begin
Choose x arbitrarily from U;
Choose y randomly from U \ {x};
F ←− F ∪ {(x, y)};
U ←− U \ {(x, y)}
end
end
Observe that the following relationship between a simple graph G ∈ Gn,d and
the number of configurations F for which γ(F) = G.
Lemma 11.1. If G ∈ Gn,d , then
n
|γ −1 (G)| = ∏ di ! .
i=1
206 CHAPTER 11. FIXED DEGREE SEQUENCE
Proof. Arrange the edges of G in lexicographic order. Now go through the se-
quence of 2m symbols, replacing each i by a new member of Wi . We obtain all F
for which γ(F) = G.
The above lemma implies that we can use random configurations to “approxi-
mate” random graphs with a given degree sequence.
Corollary 11.2. If F is chosen uniformly at random from the set of all configura-
tions Ω and G1 , G2 ∈ Gn,d then
P(γ(F) = G1 ) = P(γ(F) = G2 ).
So instead of sampling from the family Gn,d and counting graphs with a given
property, we can choose a random F and accept γ(F) iff there are no loops or
multiple edges, i.e. iff γ(F) is a simple graph.
This is only a useful exercise if γ(F) is simple with sufficiently high probabil-
ity. We will assume for the remainder of this section that
We will prove later (see Lemma 11.7 and Corollary 11.8) that if F is chosen
uniformly (at random) from Ω,
where
∑ di (di − 1)
λ= .
2 ∑ di
Hence, (11.1) and (11.2) will tell us not only how large is Gn,d , (Theorem 11.5)
but also lead to the following conclusion.
Theorem 11.3. Suppose that ∆ ≤ nα , α < 1/7. For any (multi)graph property P
Pr(L = 0 | Dk ⊆ F) Pr(Dk ⊆ F)
=∑ .
Dk Pr(L = 0)
Now because k = O(1), we see that the calculations that give us (11.4) will give
us Pr(L = 0 | Dk ⊆ F) ≈ Pr(L = 0). So,
D
E L = 0 ≈ ∑ Pr(Dk ⊆ F)
k Dk
d
1 2 d2i ϕ(i)
2
=
2 S,T∑ ∑ ∏ (2m − O(1))
⊆[n] ϕ:S→T i∈S
2
|S|=|T |=k
S∩T =0/
!2k
n
∆8
1 di (di − 1)
= ∑ 4m +O
k! i=1 m
208 CHAPTER 11. FIXED DEGREE SEQUENCE
λ 2k
≈ .
k!
It follows from Theorem 21.11 that
2
Pr(D = 0 | L = 0) ≈ e−λ (11.5)
(2m)!!
|Gn,d | ≈ e−λ (λ +1) .
∏ni=1 di !
∆ k1 (d1 + · · · + dn )k1
≤ o(1) +
2m k1 !
k1
∆e
≤ o(1) +
k1
= o(1).
The o(1) term in (11.7) accounts for the probability of having a double loop.
(c)
2 n 2
di ∆
P(F contains two adjacent double edges) ≤ ∑ ≤
i=1 2 2m − 8
∆5 m
= o(1).
(2m − 8)2
(d)
3
di dj 1
P(F contains a triple edge) ≤ ∑ 6 ≤
1≤i< j≤n 3 3 2m − 6
210 CHAPTER 11. FIXED DEGREE SEQUENCE
∆5 m
= o(1).
(2m − 6)3
(e) Let k2 = ∆2 log n.
P(F has at least k2 double edges)
n xi
di ∆
≤ o(1) + ∑ ∏ 2 · 2m − 4k2 (11.8)
x1 +···+xn =k2 , i=1
xi =0,1
∆ 2 k2 n
≤ o(1) +
m ∑ ∏ dixi
x1 +···+xn =k2 , i=1
xi =0,1
k
∆2 2 (d1 + · · · + dn )k2
≤ o(1) +
m k2 !
2 k2
2∆ e
≤ o(1) +
k2
= o(1).
The o(1) term in (11.8) accounts for adjacent multiple edges and triple edges. The
∆/(2m − 4k2 ) term can be justified as follows: We have chosen two points x1 , x2
in Wa in d2i ways and this term bounds the probability that x2 chooses a partner
in the same cell as x1 .
(f)
P(∃ vertex v incident to a loop and a multiple edge
n 2
di 1 ∆
≤∑
i=1 2 2m − 1 2m − 5
∆4 m
≤
(2m − 1)(2m − 5)
= o(1).
(g) Let X denote the number of triangles in F. Then
n di 2
∆ ∆3 m
E(F) ≤ ∑ 2 ≤ ≤ ∆3 .
i=1 2m − 4 2m − 4
Now use the Markov inequality.
(h) The probability that there is a vertex adjacent to two loops is at most
!2 4
n
1 n ∆ M(∆M)2 ∆4
d
∑ i 2∑ i i d (d − 1) ≤ = o(1).
i=1 i=1 M1 − O(1) (M − O(1))4
11.1. CONFIGURATION MODEL 211
Let now Ωi, j be the set of all F ∈ Ω such that F has i loops; j double edges, at
most ∆3 log n triangles and no double loops or triple edges and no vertex incident
with two double edges or with a loop and a multiple edge.
|Ω0,0 |
= (1 + o(1))e−λ (λ +1) ,
|Ω|
where
M2
λ= .
2M1
k1 k2
λ i+2 j
(1 − o(1))|Ω| = (1 + o(1))|Ω0,0 | ∑ ∑
i=0 j=0 i! j!
Wa Wb Wa Wb
x1 x3 x1 x3
x2 x4 x2 x4
So,
|Ωi+2, j−1 | j
= ,
|Ωi, j | (i + 1)(i + 2)
which shows that the first statement of the Switching Lemma holds.
Wb
x3 x3
Wa
x1 x1
x2 x2
x4 x4
Wc
i M1 − 2∆2 − 2i ≤ η ≤ iM1 ,
(11.9)
choosing an edge that is within distance two of the loop. We also note here that
forward d-switches do not increase the number of triangles.
Now for F ∈ Ω0, j let dL (F) denote the number of F 0 ∈ Ωi−1,0 that can be
obtained from F by an `-switch. Similarly, for F 0 ∈ Ωi−1,0 let dR (F 0 ) denote the
number of F ∈ Ωi,0 that can be transformed into F 0 by an `-switch. Then,
∑ dL (F) = ∑ dR (F 0 ).
F∈Ωi,0 F 0 ∈Ωi−1,0
while
M2 3 2 2 3
− 3∆ log n − 2∆ log n(∆ + ∆ ) ≤ |Ωi−1,0 | ≤
2
M2
∑ dR (F 0 ) ≤ |Ωi−1,0 |.
0
F ∈Ωi−1,0
2
So 5
|Ωi−1,0 | 2iM1 ∆ log n
= 1+O .
|Ωi,0 | M2 M1
Since an r-regular, r-connected graph, with n even, has a perfect matching, the
above theorem immediately implies the following Corollary.
ne−10 r−1
r(k + l) (rk+l)/2
n n rk
P(∃K, L) ≤ ∑ ∑ rk+l
k=4 l=0 k l 2 rn
ne−10 r−1
−( 2r −1)k+ 2l ek+l rk
≤ ∑ ∑n k l
2 (k + l)(rk+l)/2
k=4 l=0 k l
P(∃K, L)
n n rl ϕ(rk + rl − a)ϕ(r(n − k − l) + a)
≤∑ (11.12)
k,l,a k l a ϕ(rn)
ne k ne l
≤ Cr ∑ ×
k,l,a k l
(rk + rl − a)rk+rl−a (r(n − k − l) + a)r(n−k−l)+a
(rn)rn
ne k ne l (rk)rk+rl−a (r(n − k))r(n−k−l)+a
≤ Cr0 ∑
k,l,a k l (rn)rn
ne k ne l k rk k r(n−k)
00
≤ Cr ∑ 1−
k,l,a k l n n
r−1 !k
00 k 1−r/2 r/k
≤ Cr ∑ e n
k,l,a n
= o(1).
to be paired up in ϕ(rk + rl − a) ways and then the remaining points can be paired
up in ϕ(r(n − k − l) + a) ways. We then multiply by the probability 1/ϕ(rn) of
the final pairing.
11.3. EXISTENCE OF A GIANT COMPONENT 217
(b) If Λ > ε then w.h.p. there is a unique giant component of linear size ≈ Θn
where Θ is defined as follows: let K = ∑Li=1 iλi and
L
2α i/2
f (α) = K − 2α − ∑ iλi 1 − . (11.13)
i=1 K
(c) In Case (b), the degree sequence of the graph obtained by deleting the giant
component satisfies the conditions of (a).
Proof. We consider the execution of F-GENERATOR. We keep a sequence of
partitions Ut , At , Et ,t = 1, 2, . . . , m of W . Initially U0 = W and A0 = E0 = 0.
/ The
(t +1)th iteration of F-GENERATOR is now executed as follows: it is designed so
that we construct γ(F) component by component. At is the set of points associated
with the partially exposed vertices of the current component. These are vertices
in the current component, not all of whose points have been paired. Ut is the set of
unpaired points associated with the entirely unexposed vertices that have not been
added to any component so far. Et is the set of paired points. Whenever possible,
we choose to make a pairing that involves the current component.
(i) If At = 0/ then choose x from Ut . Go to (iii).
We begin the exploration of a new component of γ(F).
(a) We fix a vertex v and estimate the size of the component containing v. We keep
track of the size of At for t = O(log n) steps. Observe that
Here M1 = ∑Li=1 iλi n as before. The explanation for (11.14) is that |A| increases
only in Step (vi) and there it increases by i − 2 with probability . Miλ1 −2t
in
. The two
points x, y are missing from At+1 and this explains the -2.
Let ε1 = ε/L and let
(
|At | + ε1t |A1 |, |A2 |, . . . , |At | > 0.
Yt =
0 Otherwise.
It follows from (11.14) that if t = O(log n) and Y1 ,Y2 , . . . ,Yt > 0 then
P(Aτ 6= 0,
/ 1 ≤ τ ≤ t) ≤ P(Yt = Z1 + Z2 + · · · + Zt > 0),
It follows that with probability 1 −O(n−2 ) that At will become empty after at most
16ε1−2 log n rounds. Thus for any fixed vertex v, with probability 1 − O(n−2 ) the
11.3. EXISTENCE OF A GIANT COMPONENT 219
component contain v has size at most 4ε1−2 log n. (We can expose the component
containing v through our choice of x in Step (i).) Thus the probability there is a
component of size greater than 16ε1−2 log n is O(n−1 ). This completes the proof
of (a).
(b)
If t ≤ δ n for a small positive constant δ then
P(At 6= 0)
/ ≥ P(Yt = Z1 + Z2 + · · · + Zt > 0),
We now deal with the special case where λ1 = 0. There are two cases. If in
addition we have λ2 = 1 then w.h.p. Gd is the union of O(log n) vertex disjoint
cycles, see Exercise 10.5.1. If λ1 = 0 and λ2 < 1 then the only solutions to f (α) =
0 are α = 0, K/2. For then 0 < α < K/2 implies
L
2α i/2 L
2α
∑ iλi 1 − K < ∑ iλi 1 −
K
= K − 2α.
i=2 i=2
This gives Θ = 1. Exercise 10.5.2 asks for a proof that w.h.p. in this case, Gn,d
consists a giant component plus a collection of small components that are cycles
of size O(log n).
Assume now then that λ1 > 0. We show that w.h.p. there are Ω(n) isolated
edges. This together with the rest of the proof implies that Ψ < K/2 and hence
that Θ < 1. Indeed, if Z denotes the number components that are isolated edges,
then
λ1 n 1 λ1 n 6
E(Z) = and E(Z(Z − 1)) =
2 2M1 − 1 4 (2M1 − 1)(2M1 − 3)
and so the Chebyshev inequality (21.3) implies that Z = Ω(n) w.h.p.
Now for i such that λi > 0, we let Xi,t denote the number of entirely unexposed
vertices of degree i. We focus on the number of unexposed vertices of a give
degree. Then,
iXi,t
E(Xi,t+1 − Xi,t ) = − . (11.16)
M1 − 2t − 1
This suggests that we employ the differential equation approach of Section 23 in
order to keep track of the Xi,t . We would expect the trajectory of (t/n, Xi,t /n) to
follow the solution to the differential equation
dx ix
=− (11.17)
dτ K − 2τ
x(0) = λi . Note that K = M/n.
The solution to (11.17) is
2τ i/2
x = λi 1 − . (11.18)
K
In what follows, we use the notation of Section 23, except that we replace λ0
by ξ0 = n−1/4 to avoid confusion with λi .
(P0) D = (τ, x) : 0 < τ < Θ−ε
2 , 2ξ 0 < x < 1 where ε is small and positive.
(P1) C0 = 1.
11.3. EXISTENCE OF A GIANT COMPONENT 221
(P2) β = L.
ix
(P3) f (τ, x) = − K−2τ and γ = 0.
(P4) The Lipschitz constant L1 = 2K/(K − 2Θ)2 . This needs justification and
follows from
x x0 K(x − x0 ) + 2τ(x − x0 ) + 2x(τ − τ 0 )
− = .
K − 2τ K − 2τ 0 (K − 2τ)(K − 2τ 0 )
1/4 )
Theorem 23.1 then implies that with probability 1 − O(n1/4 e−Ω(n ),
i/2
2t
Xi,t − niλi 1 − = O(n3/4 ), (11.19)
K
up a point where Xi,t = O(ξ0 n). (The o(n3/4 ) term for the number of vertices of
degree i is absorbed into the RHS of (11.19).)
Now because
L L
|At | = M1 − 2t − ∑ iXi,t = Kn − 2t − ∑ iXi,t ,
i=1 i=1
2Ψ i/2
λi0 = λi 1 − .
K
222 CHAPTER 11. FIXED DEGREE SEQUENCE
(The important thing here is that the number of vertices of degree i is asymptot-
ically proportional to λi0 .) Next choose ε1 > 0 sufficiently small and let tε1 =
max {t : |At | ≥ ε1 n}. There must exist ε2 < ε1 such that tε1 ≤ (Ψ − ε2 )n and
f 0 (Ψ − ε2 ) ≤ −ε1 , else f cannot reach zero. Recall that Ψ < K/2 here and then,
2Ψ − 2ε2 i/2
0 1 2
−ε1 ≥ f (Ψ − ε2 ) = −2 + ∑ i λi 1 − K
K − 2(Ψ − ε2 ) i≥1
2Ψ i/2
1 + O(ε2 ) 2
= −2 + ∑ i λi 1 − K
K − 2Ψ i≥1
i/2 !
2Ψ i/2
1 + O(ε2 ) 2Ψ
= −2 ∑ iλi 1 − + ∑ i2 λi 1 −
K − 2Ψ i≥1 K i≥1 K
i/2
1 + O(ε2 ) 2Ψ
= ∑ i(i − 2)λi 1 −
K − 2Ψ i≥1 K
1 + O(ε2 )
=
K − 2Ψ i≥1∑ i(i − 2)λi0. (11.21)
Let E denote the high probability event in Lemma 11.12. We will condition
on the occurence of E .
Now for v ∈ [n] let Sk (v) denote the set of vertices at distance k from v and let
S
S≤k (v) = j≤k S j (v). We note that
Furthermore, Lemma 11.12 implies that w.h.p. we have that for all v, w ∈ [n], 1 ≤
k ≤ `0 ,
This is because there can be at most one cycle in S≤`0 (v) and the sizes of the
relevant sets are reduced by having the cycle as close to v, w as possible.
Now consider k > `0 . Consider doing breadth first search from v or v, w ex-
posing the configuration pairing as we go. Let an edge be dispensable if exposing
it joins two vertices already known to be in S≤k . Lemma 11.12 implies that w.h.p.
there is at most one dispensable edge in S≤`0 .
Lemma 11.13. With probability 1 − o(n−2 ), (i) at most 20 of the first n2/5 exposed
edges are dispensable and (ii) at most n1/4 of the first n3/5 exposed edges are
dispensable.
Proof. The probability that the kth edge is dispensable is at most (k−1)r
rn−2k , indepen-
dent of the history of the process. Hence,
2/5 !20
n rn2/5
P(∃ 20 dispensable edges in first n2/5 ) ≤ = o(n−2 ).
20 rn − o(n)
3/5 !n1/4
n rn3/5
1/4 3/5
P(∃ n dispensable edges in first n ) ≤ 1/4 = o(n−2 ).
n rn − o(n)
l m l m
Now let `1 = logr−1 n2/5 and `2 = logr−1 n3/5 . Then we have that, con-
ditional on E , with probability 1 − o(n−2 ),
|Sk (w)| ≥ ((r − 2)(r + 1) − o(1))(r − 1)k−2 ≈ (r − 2)(r + 1)(r − 1)a−2 n4/7 .
|Sk (w) \ Sk (v)| ≥ (r − 2 − o(1))(r − 1)k−1 ≈ (r − 2)(r − 1)a−1 n4/7 .
Suppose now that we consider the execution of breadth first search up until we
have exposed Sk (v). Then in order to have dk (v) = dk (w), conditional on the
history of the search, there has to be an exact outcome for |Sk (w) \ Sk (v)|. Now
consider the pairings of the Wx , x ∈ Sk (w) \ Sk (v). Now at most n1/4 of these
pairings are with vertices in S≤k (v)∪S≤k (w). Condition on these. There must now
be s = Θ(n4/7 ) pairings between Wx , x ∈ Sk (w) \ Sk (v) and Wy , y ∈
/ Sk (v) ∪ Sk (w).
Furthermore, to have dk (v) = dk (w) these s pairings must involve exactly t of
the sets Wy , y ∈
/ Sk (v) ∪ Sk (w), where t is determined before the choice of these s
pairings. The following lemma will easily show that G is asymmetric w.h.p.
Lemma 11.14. Let R = m
S
i=1 Ri be a partitioning of an rm set R into m subsets of
size r. Suppose that S is a random s-subset of R, where m5/9 < s < m3/5 . Let XS
denote the number of sets Ri intersected by S. Then
c0 m1/2
max P(XS = j) ≤ ,
j s
for some constant c0 .
Proof. We may assume that s ≥ m1/2 . The probability that S has at least 3 ele-
ments in some set Ri is at most
m 3r rm−3
s−3 r3 s3 r3 m1/2
rm ≤ ≤ .
s
6m2 6s
But
P(XS = j) ≤ P max |S ∩ Ri | ≥ 3 + P XS = j and max |S ∩ Ri | ≤ 2 .
i i
c1 m1/2
Pj = P XS = j and max |S ∩ Ri | ≤ 2 ≤ , (11.25)
i s
11.4. GN,R IS ASYMMETRIC 225
m j 2 j−s r s− j
j s− j r 2
Pj = rm . (11.26)
s
Pj+1 (m − j)(s − j) 2r
= . (11.27)
Pj (2 j + 2 − s)(2 j + 1 − s) r − 1
2
j0 − s − (r − 1)s ≤ 1.
2rm
s
Furthermore, if j1 = j0 − m1/2 then
Pj+1 m1/2
≤ 1 + c3 for j1 ≤ j ≤ j0 ,
Pj s
!−( j0 − j1 ) ( !)
m1/2 m1/2 m
Pj ≥ Pj0 1 + c3 = Pj0 exp −( j0 − j1 ) c3 +O 2
s s s
≥ Pj0 e−2c3 .
This proves
Theorem 11.15. W.h.p. Gn,r has a unique trivial automorphism.
and m = b(1 − γ)nr/2c, then there is a joint distribution of G(n, m) and Gn,r such
that
P(Gn,m ⊂ Gn,r ) → 1.
Corollary 11.17. Let Q be an increasing property of graphs such that Gn,m satis-
fies Q w.h.p. for some m = m(n), n log n m n2 . Then Gn,r satisfies Q w.h.p.
for r = r(n) ≈ 2m
n .
Our approach to proving Theorem 11.16 is to represent Gn,m and Gn,r as the
outcomes of two graph processes which behave similarly enough to permit a good
coupling. For this let M = nr/2 and define
GM = (ε1 , . . . , εM )
11.5. GN,R VERSUS GN,P 227
to be an ordered random uniform graph on the vertex set [n], that is, Gn,M with a
random uniform ordering of edges. Similarly, let
Gr = (η1 , . . . , ηM )
be an ordered random r-regular graph on [n], that is, Gn,r with a random uniform
ordering of edges. Further, write GM (t) = (ε1 , . . . , εt ) and Gr (t) = (η1 , . . . , ηt ),
t = 0, . . . , M.
For every ordered graph G of size t and every edge e ∈ Kn \ G we have
1
Pr (εt+1 = e | GM (t) = G) = n .
2 −t
This is not true if we replace GM by Gr , except for the very first step t = 0.
However, it turns out that for most of time the conditional distribution of the next
edge in the process Gr (t) is approximately uniform, which is made precise in the
lemma below. For 0 < ε < 1, and t = 0, . . . , M consider the inequalities
1−ε
Pr (ηt+1 = e | Gr (t)) ≥ n for every e ∈ Kn \ Gr (t), (11.28)
2 −t
Our aim is to couple GM and Gr up to the time Tε . For this we will define a
graph process G0r := (ηt0 ),t = 1, . . . , M such that the conditional distribution of
(ηt0 ) coincides with that of (ηt ) and w.h.p. (ηt0 ) shares many edges with GM .
228 CHAPTER 11. FIXED DEGREE SEQUENCE
Suppose that Gr = G0r (t) and GM = GM (t) have been exposed and for every
e∈
/ Gr the inequality
1−ε
pt+1 (e|Gr ) ≥ n (11.31)
2 −t
holds (we have such a situation, in particular, if t ≤ Tε ). Generate a Bernoulli
(1 − ε) random variable ξt+1 independently of everything that has been revealed
so far; expose the edge εt+1 . Moreover, generate a random edge ζt+1 ∈ Kn \ Gr
according to the distribution
1−ε
pt+1 (e|Gr ) −
(n2)−t
P(ζt+1 = e|G0r (t) = Gr , GM (t) = GM ) := ≥ 0,
ε
where the inequality holds because of the assumption (11.31). Observe also that
Note that
ξt+1 = 1 ⇒ εt+1 ∈ G0r (t + 1). (11.32)
We keep generating ξt ’s even after the stopping time has passed, that is, for t > Tε ,
0
whereas ηt+1 is then sampled according to probabilities (11.30), without coupling.
Note that ξt ’s are i.i.d. and independent of GM . We check that
0
P(ηt+1 = e | G0r (t) = Gr , GM (t) = GM )
= P(εt+1 = e) P(ξt+1 = 1) + P(ζt+1 = e) P(ξt+1 = 0)
pt+1 (e|Gr ) − 1−ε
1−ε n
(2)−t
= n + ε
2 −t
ε
= pt+1 (e|Gr )
for all admissible Gr , GM , i.e., such that P (Gr (t) = Gr , GM (t) = GM ) > 0, and
for all e 6∈ Gr .
Further, define a set of edges which are potentially shared by GM and Gr :
S := {εi : ξi = 1 , 1 ≤ i ≤ (1 − ε)M} .
11.5. GN,R VERSUS GN,P 229
Note that
b(1−ε)Mc
|S| = ∑ ξi
i=1
is distributed as Bin(b(1 − ε)Mc, 1 − ε).
Since (ξi ) and (εi ) are independent, conditioning on |S| ≥ m, the first m edges
in the set S comprise a graph which is distributed as Gn,m . Moreover, if Tε ≥
(1 − ε)M, then by (11.32) we have S ⊂ Gr , therefore
We have E |S| ≥ (1 − 2ε)M. Recall that ε = γ/3 and therefore m = b(1 − γ)Mc =
b(1 − 3ε)Mc. Applying the Chernoff bounds and our assumption on ε, we get
2 m)
P (|S| < m) ≤ e−Ω(γ = o(1).
In particular w.h.p.
x2
P (|X − tr/M| ≥ x) ≤ 2 exp − .
2 (τr + x/3)
230 CHAPTER 11. FIXED DEGREE SEQUENCE
√
Let x = 6 τr log n. From (11.29), assuming C0 ≥ 1, we get
r r
x log n log n
=6 ≤6 ≤ 6ε,
τr τr εr
and so x ≤ 6τr. Using this, we obtain
1 36τr log n
P (|X − tr/M| ≥ x) ≤ exp − = n−6 .
2 2(τr + 2τr)
or, equivalently, ε ≥ 144 log n/r, which is implied by (11.29) with C0 ≥ 144.
Given an ordered graph G = (e1 , . . . , et ), we say that an ordered r-regular graph
H is an extension of G if the first t edges of H are equal to G. Let GG (n, r) be the
family of extensions of G and GG = GG (n, r) be a graph chosen uniformly at
random from GG (n, r).
Further, for a graph H ∈ GG (n, r) and u, v ∈ [n] let
Note that degH|G (u, v) is not in general symmetric in u and v, but for G = 0/ coin-
cides with the usual co-degree in a graph H.
The next fact is used in the proof of Lemma 11.21 only.
Lemma 11.20. Let graph G with t ≤ (1 − ε)M edges be such that GG (n, r) is
nonempty. For each e ∈
/ G we have
4r
P (e ∈ GG ) ≤ . (11.35)
εn
Moreover, if l ≥ l0 := 4r2 /(εn), then for every u, v ∈ [n] we have
P degGG |G (u, v) > l ≤ 2−(l−l0 ) . (11.36)
11.5. GN,R VERSUS GN,P 231
Let us define an auxiliary bipartite graph B between Ge∈ and Ge∈/ in which H ∈ Ge∈
is connected to H 0 ∈ Ge∈/ whenever H 0 can be obtained from H by the following
switching operation. Fix an ordered edge {w, x} in H \ G which is disjoint from
e = {u, v} and such that there are no edges between {u, v} and {w, x} and replace
the edges {u, v} and {w, x} by {u, w} and {v, x} to obtain H 0 . Writing f (H) for
the number of graphs H 0 ∈ Ge∈/ which can be obtained from H by a switching,
and b(H 0 ) for the number of graphs H ∈ Ge∈ such that H 0 can be obtained H by a
switching, we get that
We have b(H 0 ) ≤ degH 0 (u) degH 0 (v) ≤ r2 . On the other hand, recalling that t ≤
(1 − ε)M, for every H ∈ Ge∈ we get
2r2
2 εM
f (H) ≥ M − t − 2r ≥ εM 1 − ≥ ,
εM 2
2r2 4r n 1/3 4 1
≤ 0 ≤ 0≤ .
εM C n r C 2
Therefore (11.37) implies that
|Ge∈ | 2r2 4r
P (e ∈ GG ) ≤ ≤ = ,
|Ge∈/ | εM εn
u0 w0 u0 w0
u u
w w
v v
Figure 11.6: Switching between G (l) and G (l − 1): Before and after.
3r2
2
f (H) ≥ 2l(M − t − 3r ) ≥ 2lεM 1 − ≥ lεM,
εM
3r2 6r 6 r 2/3 1
= ≤ 0 ≤ ,
εM εn C n 2
and the number of ways to apply a backward switching is b(H) ≤ r3 . So,
(i) e00 ∈
/H
0 00 0 00
(ii) max degH|G∪e0 (u , u ), degH|G∪e0 (v , v ) ≤ l0 + log2 n,
Pr M0 ∈ 0
/ Gnice | M0 ∈ G 0 = P GG∪e0 (n, r) 6∈ Gnice
0
4r ε
≤ + 2 × 2− log2 n ≤ . (11.41)
εn 4
We have
Pr M00 ∈ G 00 | M0 ∈ Gnice
0
Pr M0 ∈ Gnice
0
| M0 ∈ G 0 =
0 ) P(M0 ∈ G 0 , M0 ∈ G 0 )
P(M00 ∈ G 00 , M0 ∈ Gnice nice
· ≤
P(M0 ∈ Gnice0 ) P(M0 ∈ G 0 )
P (M00 ∈ G 00 )
. (11.42)
P (M0 ∈ G 0 )
To complete the proof of the claim, it suffices to show that
ε
Pr M00 ∈ G 00 | M0 ∈ Gnice
0
≥ 1− , (11.43)
4
since plugging (11.41) and (11.43) into (11.42) will complete the proof of the
statement.
To prove (11.43), fix H ∈ Gnice 0 and condition on M0 = H. A loop can only be
created in M when u is incident to u0 in H \ (G ∪ e0 ) and the randomly chosen
00 00
edge is {u0 , u00 }, or, provided v0 6= v00 , when v00 is incident to v0 in H \ (G ∪ e0 ) and
we randomly choose {v0 , v00 }. Therefore, recalling that ∆G ≤ (1 − ε/2)r, we get
1 1
Pr M00 has a loop | M0 = H ≤
+
degH\(G∪e0 ) (u00 ) degH\(G∪e0 ) (v00 )
4 ε
≤ ≤ , (11.44)
εr 8
where the second term is present only if e0 ∩ e00 = 0,
/ and the last inequality is
implied by (11.29).
11.5. GN,R VERSUS GN,P 235
A multiple edge can be created in three ways: (i) by choosing, among the
edges incident to u00 , an edge {u00 , w} ∈ H \ (G ∪ e0 ) such that {u0 , w} ∈ H; (ii) sim-
ilarly for v00 (if v0 6= v00 ); (iii) choosing both edges {u00 , v0 } and {v00 , u0 } (provided
they exist in H \ (G ∪ e0 )). Therefore, by (ii) and assumption ∆G ≤ (1 − ε/2)r,
and we can choose arbitrarily large C0 . (Again, in case when |e0 ∩ e00 | = 1, the
R-H-S of (11.45) reduces to only the first summand.)
Combining (11.44) and (11.45), we have shown (11.43).
Proof of Lemma 11.18. In view of Lemma 11.19 it suffices to show that
1−ε
Pr (ηt+1 = e | Gr (t) = G) ≥ n , e∈
/ G.
2 −t
By (11.38) we have
and similarly for the family G 00 . This yields, after a few cancellations, that
τ −δ 2 2δ 2
r r
log n log n ε
≥ 1− ≥ 1 − 24 ≥ 1 − 24 ≥ 1− ,
τ +δ τ τr εr 2
P (M00 ∈ G 00 ) ε
≥ 1− ,
P (M ∈ G )
0 0 2
11.6 Exercises
11.6.1 Show that w.h.p. a random 2-regular graph on n vertices consists of O(log n)
vertex disjoint cycles.
11.6.4 Let x = (x1 , x2 , . . . , x2m ) be chosen uniformly at random from [n]2m . Let Gx
be the multigraph with vertex set [n] and edges (x2i−1 , x2i ), i = 1, 2, . . . , m.
Let dx (i) be the number of times that i appears in x.
Show that conditional on dx (i) = di , i ∈ [n], Gx has the same distribution
as the multigraph γ(F) of Section 11.1.
λ i e−λ
P(Z = i) = for i ≥ k,
i! fk (λ )
11.6.9 Show that if graph G = G1 ∪G2 then its rainbow connection satisfies rc(G)
≤ rc(G1 ) + rc(G2 ) + |E(G1 ) ∩ E(G2 )|. Using the contiguity of Gn,r to the
union of r independent matchings, (see Chapter 20), show that rc(Gn,r ) =
O(logr n) for r ≥ 6.
11.7 Notes
Giant Components and Cores
Hatami and Molloy [414] discuss the size of the largest component in the scaling
window for a random graph with a fixed degree sequence.
Cooper [210] and Janson and Luczak [446] discuss the sizes of the cores of
random graphs with a given degree sequence.
Hamilton cycles
Robinson and Wormald [677], [680] showed that random r-regular graphs are
Hamiltonian for 3 ≤ r = O(1). In doing this, they introduced the important new
method of small subgraph conditioning. It is a refinement on the Chebyshev in-
equality. Somewhat later Cooper, Frieze and Reed [235] and Krivelevich, Su-
dakov, Vu Wormald [532] removed the restriction r = O(1). Frieze, Jerrum, Mol-
loy, Robinson and Wormald [346] gave a polynomial time algorithm that w.h.p.
finds a Hamilton cycle in a random regular graph. Cooper, Frieze and Krivelevich
[229] considered the existence of Hamilton cycles in Gn,d for certain classes of
degree sequence.
Chromatic number
r
Frieze and Łuczak [355] proved that w.h.p. χ(Gn,r ) = (1 + or (1)) 2 log r for r =
O(1). Here or (1) → 0 as r → ∞. Achlioptas and Moore [5] determined the chro-
matic number of a random r-regular graph to within three values, w.h.p. Kemkes,
Pérez-Giménez and Wormald [498] reduced the range to two values. Shi and
11.7. NOTES 239
Wormald [713], [714] consider the chromatic number of Gn,r for small r. In par-
ticular they show that w.h.p. χ(Gn,4 ) = 3. Frieze, Krivelevich and Smyth [350]
gave estimates for the chromatic number of Gn,d for certain classes of degree se-
quence.
Eigenvalues
The largest eigenvalue of the adjacency matrix of Gn,r is always r. Kahn and
Szemerédi [475] showed that w.h.p. the second eigenvalue is of order O(r1/2 ).
Friedman [331] proved that w.h.p. the second eigenvalue is at most 2(r − 1)1/2 +
o(1). Broder, Frieze, Suen and Upfal [169] considered Gn,d where C−1 d ≤ di ≤
Cd for some constant C > 0 and d ≤ n1/10 . They show that w.h.p. the second
eigenvalue of the adjacency matrix is O(d 1/2 ).
Rainbow Connection
Dudek, Frieze and Tsourakakis [268] studied the rainbow connection of random
regular graphs. They showed that if 4 ≤ r = O(1) then rc(Gn,r ) = O(log n). This is
best possible up to constants, since rc(Gn,r ) ≥ diam(Gn,r ) = Ω(log n). Kamčev,
Krivelevich and Sudakov [478] gave a simpler proof when r ≥ 5, with a better
hidden constant.
240 CHAPTER 11. FIXED DEGREE SEQUENCE
Chapter 12
Intersection Graphs
241
242 CHAPTER 12. INTERSECTION GRAPHS
set bipartition
Equivalence
One of the first interesting problems to be considered is the question as to when
the random graphs G(n, m, p) and Gn,p have asymptotically the same properties.
Intuitively, it should be the case when the edges of G(n, m, p) occur “almost inde-
pendently”, i.e., when there are no vertices of degree greater than two in M in the
generator Gn,m,p of G(n, m, p). Then each of its edges is induced by a vertex of
degree two in M, “almost” independently
of other edges. One can show that this
1/3
happens w.h.p. when p = o 1/(nm ) , which in turn implies that both random
12.1. BINOMIAL RANDOM INTERSECTION GRAPHS 243
graphs are asymptotically equivalent for all graph properties P. Recall that a
graph property P is defined as a subset of the family of all labeled graphs on ver-
n
tex set [n], i.e., P ⊆ 2(2) . The following equivalence result is due to Rybarczyk
[692] and Fill, Scheinerman and Singer-Cohen [311].
Theorem 12.1. Let 0 ≤ a ≤ 1, P be any graph property, p = o 1/(nm1/3 ) and
Then
P(Gn, p̂ ∈ P) → a
if and only if
P(G(n, m, p) ∈ P) → a
as n → ∞.
Proof. Let X and Y be random variables taking values in a common finite (or
countable) set S. Consider the probability measures L (X) and L (Y ) on S whose
values at A ⊆ S are P(X ∈ A) and P(Y ∈ A). Define the total variation distance
between L (X) and L (Y ) as
which is equivalent to
1
dTV (L (X), L (Y )) = ∑ | P(X = s) − P(Y = s)|.
2 s∈S
Notice (see Fact 4 of [311]) that if there exists a probability space on which ran-
dom variables X 0 and Y 0 are both defined, with L (X) = L (X 0 ) and L (Y ) =
L (Y 0 ), then
dTV (L (X), L (Y )) ≤ P(X 0 6= Y 0 ). (12.2)
Furthermore (see Fact 3 in [311]) if there exist random variables Z and Z 0 such
that L (X|Z = z) = L (Y |Z 0 = z), for all z, then
We will need one more observation. Suppose that a random variable X has distri-
bution the Bin(n, p), while a random variable Y has the Poisson distribution, and
E X = EY . Then
dTV (X,Y ) = O(p). (12.4)
244 CHAPTER 12. INTERSECTION GRAPHS
Lemma 12.2. If a random variable Z has the Poisson distribution with expec-
tation λ then Ni (Z), i = 1, 2, . . . , r, are independent and identically Poisson dis-
tributed random variables, with expectation λ γ. Moreover the random variable
X(Z) has the distribution Bin(r, 1 − e−λ γ ).
Let us consider the following special case of the scheme defined above, as-
suming that r = n2 and γ = 1/ n2 . Here each coupon represents a distinct edge
of Kn .
Lemma 12.3. Suppose p = o(1/n) and let a random variable Z be the
n 2
Bin m, 2 p (1 − p)n−2 distributed,
while a random variable Y be the
n −mp 2 (1−p)n−2
Bin 2 , 1 − e distributed. Then
Proof. Let Z 0 be a Poisson random variable with the same expectation as Z, i.e.,
0 n 2
EZ = m p (1 − p)n−2 .
2
12.1. BINOMIAL RANDOM INTERSECTION GRAPHS 245
dTV (L (Y ), L (X(Z)))
= dTV (L (X(Z 0 )), L (X(Z))) ≤ 2dTV (L (Z 0 ), L (Z))
n 2 n−2
= O n2 p2 = o(1).
≤O p (1 − p)
2
Now define a random intersection graph G2 (n, m, p) as follows. Its vertex set
is V = {1, 2, . . . , n}, while e = {i, j} is an edge in G2 (n, m, p) iff in a (generator)
bipartite random graph Gn,m,p , there is a vertex w ∈ M of degree two such that
both i and j are connected by an edge with w.
To complete the proof of our theorem, notice that,
for p = o(1/(nm1/3 ).
Hence it remains to show that
and
P(Gn, p̂ = G) = P(Gn, p̂ = G0 ).
Equation (12.6) now follows from Lemma 12.3. The theorem follows immedi-
ately.
For monotone properties (see Chapter 1) the relationship between the classical
binomial random graph and the respective intersection graph is more precise and
was established by Rybarczyk [692].
Small subgraphs
Let H be any fixed graph. A clique cover C is a collection of subsets of vertex
set V (H) such that, each induces a complete subgraph (clique) of H, and for every
edge {u, v} ∈ E(H), there exists C ∈ C , such that u, v ∈ C. Hence, the cliques
induced by sets from C exactly cover the edges of H. A clique cover is allowed
to have more than one copy of a given set. We say that C is reducible if for
some C ∈ C , the edges of H induced by C are contained in the union of the
edges induced by C \C, otherwise C is irreducible. Note that if C ∈ C and C is
irreducible, then |C| ≥ 2.
In this section, |C | stands for the number of cliques in C , while ∑ C denotes
the sum of clique sizes in C , and we put ∑ C = 0 if C = 0. /
Let C = {C1 ,C2 , . . . ,Ck } be a clique cover of H. For S ⊆ V (H) define the
following two restricted clique covers
Finally, let
τ(H) = min max {τ1 , τ2 },
C S⊆V (H)
where the minimum is taken over all clique covers C of H. We can in this calcu-
lation restrict our attention to irreducible covers.
Karoński, Scheinerman and Singer-Cohen [490] proved the following theorem.
As an illustration, we will use this theorem to show the threshold for complete
graphs in G(n, m, p), when m = nα , for different ranges of α > 0.
Proof. There are many possibilities for clique covers to generate a copy of a com-
plete graph Kh in G(n, m, p). However in the case of Kh only two play a dominat-
ing role. Indeed, we will show that for α ≤ α0 , α0 = 2h/(h − 1) the clique cover
C = {V (Kh )} composed of one set containing all h vertices of Kh only matters,
while for α ≥ α0 the clique cover C = 2 , consisting of h2 pairs of endpoints
Kh
Finally, when |S| = 1, then τ1 = (nm)−1 , while τ2 = 0, both smaller than n−1 m−1/h ,
and so equation (12.7) follows.
For the edge-clique cover {E} we have a similar expression, viz.
max{τ1 (Kh , {E}, S), τ2 (Kh , {E}, S)} = τ1 (Kh , {E},V ). (12.8)
S⊆V
Let S ⊂ V , with s = |S| ≤ h − 1, and consider restricted clique covers with cliques
of size at most two, and exactly two.
For τ1 , the clique cover restricted to S is the edge-clique cover of Ks , plus a
1-clique for each of the h − s external edges for each vertex of Ks , so
τ1 (Kh , {E}, S)
−1
= ns/[s(s−1)+s(h−s)] m[s(s−1)/2+s(h−s)]/[s(s−1)+s(h−s)]
−1
= n1/(h−1) m[h−(s+1)/2]/(h−1)
−1
≤ n1/(h−1) mh/(2(h−1))
−1
< n1/(h−1) m1/2 ,
Thus,
τ1 (Kh , C ,V ) ≥ min {τ1 (Kh , {V },V ), τ1 (Kh , {E},V )} . (12.9)
12.1. BINOMIAL RANDOM INTERSECTION GRAPHS 249
where h − 1 counts all other vertices aside from u since they must appear in some
clique with u.
For any v ∈ V (Kh ) we have
∑ C + |{i : Ci 3 v}| − (h − 1) ≥ ∑ C + r − (h − 1)
≥ (h − 1) + r + 2(|C | − r) + r − (h − 1)
= 2|C |.
Summing the above inequality over all v ∈ V (Kh ),
h ∑ C + ∑ C − h(h − 1) ≥ 2h|C |,
and dividing both sides by 2h ∑ C , we finally get
|C | h+1 h−1
≤ − .
∑C 2h 2∑C
Now, using the above bound,
h |C | 2h
xC (α0 ) = +
∑C ∑C h−1
250 CHAPTER 12. INTERSECTION GRAPHS
h h+1 h−1 2h
≤ + −
∑C 2h 2∑C h−1
2
= 1+
h−1
= xV (α0 ).
Now, since xC (α) ≤ xV (α) at both α = 0 and α = α0 , and both functions are
linear, xC (α) ≤ xV (α) throughout the interval (0, α0 ).
Since xE (α0 ) = xV (α0 ) we also have xC (α0 ) ≤ xE (α0 ). The slope of xC (α)
|C |
is ∑ C , and by the assumption that C consists of cliques of size at least 2, this is
at most 1/2. But the slope of xE (α) is exactly 1/2. Thus for all α ≥ α0 , xC (α) ≤
xE (α). Hence the bounds given by formula (12.9) hold.
One can show (see [690]) that for any irreducible clique-cover C that is not
{V } nor {E},
max{τ1 (Kh , C , S), τ2 (Kh , C , S)} ≥ τ1 (Kh , C ,V ).
S
Hence, by (12.9),
max{τ1 (Kh , C , S), τ2 (Kh , C , S)} ≥ min{τ1 (Kh , {V },V ), τ1 (Kh , {E},V )}.
S
λn = EXn ≈ ch /h!
and
dTV (L (Xn ), Po(λn )) = O n−α/h ;
12.2. RANDOM GEOMETRIC GRAPHS 251
2h −(h+1)/(h−1)
(ii) If α = h−1 , p ≈ cn then
h h(h−1)
λn = EXn ≈ c + c /h!
and
dTV (L (Xn ), Po(λn )) = O n−2/(h−1) ;
2h −1/(h−1) m−1/2
(iii) If α > h−1 , p ≈ cn then
and α(h−1)
2
h− 2 − h−1 −1
dTV (L (Xn ), Po(λn )) = O n +n .
Connectivity
The threshold (in terms of r) for connectivity was shown to be identical with that
for minimum degree one, by Gupta and Kumar [403]. This was extended to k-
connectivity by Penrose [645]. We do not aim for tremendous accuracy. The
simple proof of connectivity was provided to us by Tobias Müller [617].
252 CHAPTER 12. INTERSECTION GRAPHS
q
log n
Theorem 12.8. Let ε > 0 be arbitrarily small and let r0 = r0 (n) = πn . Then
w.h.p.
The factor (1 − πr2 )n−1 bounds the probability that none of X2 , X3 , . . . , Xn lie in
B(X1 , r), given that B(X1 , r) ⊆ D. It is exact for points far enough from the bound-
ary of D.
Now
(1 − ε) log n n
2 n−1
(1 − πr ) ≥ 1− = nε−1+o(1) .
n
So if I is the set of isolated vertices then E(|I|) ≥ nε−1+o(1) → ∞. Now
n−2
πr2
P(X1 ∈ I | X2 ∈ I) ≤ 1 − ≤ (1 + o(1)) P(X1 ∈ I).
1 − πr2
πr2
The expression 1 − 1−πr2 is the probability that a random point does not lie in
B(X1 , r), given that it does not lie in B(X2 , r), and that |X2 − X1 | ≥ 2r. Equation
(12.10) now follows from the Chebyshev inequality (21.3).
Now consider (12.11). Let η ε be a sufficiently small constant and divide
D into `20 sub-squares of side length ηr, where `0 = 1/ηr. We refer to these sub-
squares as cells. We can assume that η is chosen so that `0 is an integer. We
say that a cell is good if contains at least i0 = η 3 log n members of X and bad
otherwise. We next let K = 100/η 2 and consider the number of bad cells in a
K × K square block of cells.
Lemma 12.9. Let B be a K × K square block of cells. The following hold w.h.p.:
(a) If B is further than 100r from the closest boundary edge of D then B contains
at most k0 = (1 − ε/10)π/η 2 bad cells.
(b) If B is within distance 100r of exactly one boundary edge of D then B contains
at most k0 /2 bad cells.
12.2. RANDOM GEOMETRIC GRAPHS 253
Proof. (a) There are less than `20 < n such blocks. Furthermore, the probability
that a fixed block contains k0 or more bad cells is at most
!k0
2 i0
K n
k0 ∑ i (η 2r2)i(1 − η 2r2)n−i
i=0
2 k0 i0 !k0
K e ne 2 2
≤ 2 (η 2 r2 )i0 e−η r (n−i0 ) . (12.12)
k0 i0
2
!(1−ε/10)π/η 2
2K 2 en−η (1+ε/2)/π
≤ n−1−ε/3 . (12.14)
(1 − ε/10)π/η 2
Part (a) follows after inflating the RHS of (12.14) by n to account for the number
of choices of block.
(b) Replacing k0 by k0 /2 replaces the LHS of (12.14) by
2
!(1−ε/10)π/2η 2
4K 2 en−η (1+ε/2)/π
≤ n−1/2−ε/6 . (12.15)
(1 − ε/10)π/2η 2
Observe now that the number of choices of block is O(`0 ) = o(n1/2 ) and then Part
(b) follows after inflating the RHS of (12.15) by o(n1/2 ) to account for the number
of choices of block.
(c) Equation (12.13) bounds the probability that a single cell is bad. The number
of cells in question in this case is O(1) and (c) follows.
We now do a simple geometric computation in order to place a lower bound on
the number of cells within a ball B(X, r).
254 CHAPTER 12. INTERSECTION GRAPHS
√
Lemma 12.10. A half-disk of radius r1 = r(1 − η 2) with diameter part of the
grid of cells contains at least (1 − 2η 1/2 )π/2η 2 cells.
Proof. We place the half-disk in a 2r1 × r1 rectangle. Then we partition the rect-
angle into ζ1 = r1 /rη rows of 2ζ1 cells. The circumference of the circle will cut
the ith row at a point which is r1 (1j− i2 η 2 )1/2 from thek centre of the row. Thus
the ith row will contain at least 2 r1 (1 − i2 η 2 )1/2 /rη complete cells. So the
half-disk contains at least
2r1 1/η
Z 1/η−1
2r1
∑ ((1 − i2 η 2 )1/2 − η) ≥ ((1 − x2 η 2 )1/2 − η)dx
rη i=1 rη x=1
Z arcsin(1−η)
2r1
= 2 (cos2 (θ ) − η cos(θ ))dθ
rη θ =arcsin(η)
arcsin(1−η)
2r1 θ sin(2θ )
≥ 2 − −η .
rη 2 4 θ =arcsin(η)
Now
π
arcsin(1 − η) ≥ − 2η 1/2 and arcsin(η) ≤ 2η.
2
So the number of cells is at least
2r1 π 1/2
− η − η .
rη 2 4
Now let Γ be the graph whose vertex set consists of the good cells and where cells
c1 , c2 are adjacent iff their centres are within distance r1 . Note that if c1 , c2 are
adjacent in Γ then any point in X ∩ c1 is adjacent in GX ,r to any point in X ∩ c2 .
It follows from (12.16) that all we need to do now is show that Γ is connected.
It follows from Lemma 12.9 that at most π/η 2 rows of a K ×K block contain a bad
cell. Thus more than 95% of the rows and of the columns of such a block are free
of bad cells. Call such a row or column good. The cells in a good row or column
of some K × K block form part of the same component of Γ. Two neighboring
blocks must have two touching good rows or columns so the cells in a good row or
column of some block form part of a single component of Γ. Any other component
C must be in a block bounded by good rows and columns. But the existence of
such a component means that it is surrounded by bad cells and then by Lemma
12.2. RANDOM GEOMETRIC GRAPHS 255
12.10 that there is a block B with at least (1 − 3η 1/2 )π/η 2 bad cells if it is far
from the boundary and at least half of this if it is close to the boundary. But this
contradicts Lemma 12.9. To see this, consider a cell in C whose center c has the
largest second component i.e. is highest in C. Now consider the half disk H of
radius r1 that is centered at c. We can assume
√ (i) H is contained entirely in B and
(ii) at least (1 − 2η 1/2 )π/2η 2 − (1 − η 2)/η ≥ (1 − 3η 1/2 )π/2η 2 cells in H are
bad. Property (i) arises because cells above c whose centers are at distance at most
r1 are all bad and for (ii) we have discounted any bad cells on the diameter through
c that might be in C. This provides half the claimed bad cells. We obtain the rest
by considering a lowest cell of C. Near the boundary, we only need to consider
one half disk with diameter parallel to the closest boundary. Finally observe that
there are no bad cells close to a corner.
Hamiltonicity
The first inroads on the Hamilton cycle problem were made by Diaz, Mitsche and
Pérez-Giménez [252]. Best possible results were later given by Balogh, Bollobás,
Krivelevich, Müller and Walters [51] and by Müller, Pérez-Giménez and
Wormald [618]. As one might expect Hamiltonicity has a threshold at r close to
r0 . We now have enough to prove the result from [252].
We start with a simple lemma, taken from [51].
Proof. Consider a spanning tree T of γ that minimises the sum of the lengths of
the edges joining the centres of the cells. Then T does not have any vertex of
degree greater than 6. This is because, if centre v were to have degree at least
7, then there are two neighboring centres u, w of v such that the angle between
the line segments [v, u] and [v, w] is strictly less than 60 degrees. We can assume
without loss of generality that [v, u] is shorter than [v, w]. Note that if we remove
the edge {v, w} and add the edge {u, w} then we obtain another spanning tree but
with strictly smaller total edge-length, a contradiction. Hence T has maximum
degree at most 6.
Proof. We begin with the tree T promised by Lemma 12.11. Let c be a good cell.
We partition the points of X ∩ c into 2d roughly equal size sets P1 , P2 , . . . , P2d
256 CHAPTER 12. INTERSECTION GRAPHS
Chromatic number
We look at the chromatic number of GX ,r in a limited range. Suppose that nπr2 =
log n
ωr where ωr → ∞, ωr = O(log n). We are below the threshold for connectivity
here. We will show that w.h.p.
where will use cl to denote the size of the largest clique. This is a special case of
a result of McDiarmid [585].
We first bound the maximum degree.
Lemma 12.13.
log n
∆(GX ,r ) ≈ w.h.p.
log ωr
Proof. Let Zk denote the number of vertices of degree k and let Z≥k denote the
number of vertices of degree at least k. Let k0 = log n
ωd where ωd → ∞ and ωd =
o(ωr ). Then
log n log n
n 2 k0 neωd log n ωd eωd ωd
E(Z≥k0 ) ≤ n (πr ) ≤ n =n .
k0 nωr log n ωr
So,
log n
log(E(Z≥k0 )) ≤ (ωd + 1 + log ωd − log ωr ) . (12.17)
ωd
12.2. RANDOM GEOMETRIC GRAPHS 257
−1/2
Now let ε0 = ωr . Then if
ωd + log ωd + 1 ≤ (1 − ε0 ) log ωr
then (12.17) implies that E(Zk ) → 0. This verifies the upper bound on ∆ claimed
in the lemma.
Now let k1 = log n
b where ωd is the solution to
ω
b
d
bd + log ω
ω bd + 1 = (1 + ε0 ) log ωr .
Next let M denote the set of vertices that are at distance greater than r from any
edge of D. Let Mk be the set of vertices of degree k in M. If Zbk = |Mk | then
n − 1
E(Zbk1 ) ≥ n P(X1 ∈ M) × (πr2 )k1 (1 − πr2 )n−1−k1 .
k1
E(Zbk1 ) ≥
k1
n (n − 1)e 2 /(1−πr2 )
(1 − 4r) 1/2
(πr2 )k1 e−nπr
3k1 k1
log n
n1−1/ωr
eω
bd ω
bd
≥ (1 − o(1)) 1/2
.
3k1 ωr
So,
log(E(Zbk1 )) ≥
log n ω
bd
− o(1) − O(log log n) + ωbd + 1 + log ωbd − log ωr −
ωbd ωr
!
ε0 log n log ωr log n
=Ω =Ω 1/2
→ ∞.
ωd
b ωr
An application of the Chebyshev inequality finishes the proof of the lemma. In-
deed,
Now cl(GX ,r ) ≤ ∆(GX ,r ) + 1 and so we now lower bound cl(GX ,r ) w.h.p. But
this is easy. It follows from Lemma 12.13 that w.h.p. there is a vertex X j with at
log n
least (1 − o(1)) log(4ω r)
vertices in its r/2 ball B(X j , r/2). But such a ball provides
log n
a clique of size (1 − o(1)) log(4ω r)
. We have therefore proved
log n
Theorem 12.14. Suppose that nπr2 = ωr where ωr → ∞, ωr = O(log n). Then
w.h.p.
log n
χ(GX ,r ) ≈ ∆(GX ,r ) ≈ cl(GX ,r ) ≈ .
log ωr
We now consider larger r.
Theorem 12.15. Suppose that nπr2 = ωr log n where ωr → ∞, ωr = o(n/ log n).
Then w.h.p. √
ωr 3 log n
χ(GX ,r ) ≈ .
2π
Proof. First consider the triangular lattice in the plane. This
√ is the set of points
T = {m1 a + m2 b : m1 , m2 ∈ Z} where a = (0, 1), b = (1/2, 3/2), see Figure 12.1.
For the lower bound we use a classic result on packing disks in the plane.
Lemma 12.17. Let An = [0, n]2 and C be a√collection of disjoint disks of unit area
that touch An . Then |C | ≤ (1 + o(1))πn2 / 12.
Proof. Thue’s theorem states that the densest packing of disjoint same
√ size disks0
in the plane is the hexagonal packing which has density λ = π/ 12. Let C
denote the disks that are contained entirely in An . Then we have
πn2
|C 0 | ≥ |C | − O(n) and |C 0 | ≤ √ .
12
The first inequality comes from the fact that if C ∈ C \ C 0 then it is contained in a
perimeter of width O(1) surrounding An .
260 CHAPTER 12. INTERSECTION GRAPHS
12.3 Exercises
√
12.3.1 Show that if p = ω(n)/ (n m), and ω(n) → ∞, then G(n, m, p) has w.h.p.
at least one edge.
12.3.7 Let X denotes the number of isolated vertices in the binomial random in-
tersection graph G(n, m, p), where m = nα , α > 0. Show that if
(
(log n + ϕ(n))/m when α ≤ 1
p= p
(log n + ϕ(n))/(nm) when α > 1,
12.3.8 Find the variance of the random variable X counting isolated vertices in
G(n, m, p).
12.3.9 Let Y be a random variable which counts vertices of degree greater than
one in G(n, m, p), with m = nα and α > 1. Show that for p2 m2 n log n
12.3.14 Given X and an integer k we define the k-nearest neighbor graph Gk−NN,X
as follows: We add an edge between x and y of X iff y is one of x’s k near-
est neighbors, in Euclidean distance or vice-versa. Show that if k ≥ C log n
for a sufficiently large C then Gk−NN,X is connected w.h.p.
12.4 Notes
Binomial Random Intersection Graphs
For G(n, m, p) with m = nα , α constant, Rybarczyk and Stark [691] provided a
condition, called strictly α-balanced for the Poisson convergence for the number
of induced copies of a fixed subgraph, thus complementing the results of Theorem
12.5 and generalising Theorem 12.7. (Thresholds for small subgraphs in a related
model of random intersection digraph are studied by Kurauskas [538]).
Rybarczyk [693] introduced a coupling method to find thresholds for many prop-
erties of the binomial random intersection graph. The method is used to establish
sharp threshold functions for k-connectivity, the existence of a perfect matching
and the existence of a Hamilton cycle.
Stark [722] determined the distribution of the degree of a typical vertex of G(n, m, p),
m = nα and showed that it changes sharply between α < 1, α = 1 and α > 1.
Behrisch [70] studied the evolution of the order of the largest component in G(n, m, p),
m = nα when α 6= 1. He showed that when α > 1 the random graph G(n, m, p)
behaves like Gn,p in that a giant component of size order n appears w.h.p. when
262 CHAPTER 12. INTERSECTION GRAPHS
the expected vertex degree exceeds one. This is not the case when α < 1. There is
a jump in the order of size of the largest component, but not to one of linear size.
Further study of the component structure of G(n, m, p) for α = 1 is due to Lageras
and Lindholm in [540].
Behrisch, Taraz and Ueckerdt [71] study the evolution of the chromatic number
of a random intersection graph and showed that, in a certain range of parameters,
these random graphs can be colored optimally with high probability using various
greedy algorithms.
threshold for the property that each vertex of Gs (n, m, r) has degree at least k is
the same as that for Gs (n, m, r) being k-connected (for related results see [768]).
rackas in [100]), sharp threshold functions for connectivity, matchings and Hamil-
tonian cycles (by Rybarczyk in [693]) as well as the size of the largest component
(by Bradonjić, Elsässer, Friedrich, Sauerwald and Stauffer in [162]).
To learn more about different models of random intersection graphs and about
other results we refer the reader to recent review papers [101] and [102].
w.h.p. Friedrich, Sauerwald and Stauffer [333] extended this to higher dimen-
sions.
A recent interesting development can be described as Random Hyperbolic Graphs.
These are related to the graphs of Section 12.2 and are posed as models of real
world networks. Here points are randomly embedded into hyperbolic, as op-
posed to Euclidean space. See for example Bode, Fountoulakis and Müller [110],
[111]; Candellero and Fountoulakis [176]; Chen, Fang, Hu and Mahoney [185];
Friedrich and Krohmer [332]; Krioukov, Papadopolous, Kitsak, Vahdat and Boguñá
[520]; Fountoulakis [322]; Gugelmann, Panagiotou and Peter [402]; Papadopolous,
Krioukov, Boguñá and Vahdat [642]. One version of this model is described in
[322]. The models are a little complicated to describe and we refer the reader to
the above references.
266 CHAPTER 12. INTERSECTION GRAPHS
Chapter 13
Digraphs
267
268 CHAPTER 13. DIGRAPHS
Theorem 13.2. Let p = c/n, where c is a constant, c > 1, and let x be defined by
x < 1 and xe−x = ce−c . Then w.h.p. Dn,p contains a unique strong component of
2
size ≈ 1 − xc n. All other strong components are of logarithmic size.
Lemma 13.3. There exist constants α, β , dependent only on c, such that w.h.p.
6 ∃ v such that |D± (v)| ∈ [α log n, β n].
Proof. If there is a v such that |D+ (v)| = s then Dn,p contains a tree T of size s,
rooted at v such that
Now ce1−c < 1 for c 6= 1 and so there exists β such that when s ≤ β n we can
bound ce1−c+s/n by some constant γ < 1 (γ depends only on c). In which case
n s 4
γ ≤ n−3 for log n ≤ s ≤ β n.
cs log 1/γ
Fix a vertex v ∈ [n] and consider a directed breadth first search from v. Let S0+ =
S0+ (v) = {v} and given S0+ , S1+ = S1+ (v), . . . , Sk+ = s+
k (v) ⊆ [n] let Tk+ = Tk+ (v) =
Sk +
i=1 Si and let
+
= w 6∈ Tk+ : ∃x ∈ Tk+ such that (x, w) ∈ E(Dn,p ) .
Sk+1
We similarly define S0− = S0− (v), S1− = S1− (v), . . . , Sk− = Sk− , Tk− (v) ⊆ [n] with re-
spect to a directed breadth first search into v.
Not surprisingly, we can show that the subgraph Γk induced by Tk+ is close in dis-
tribution to the tree defined by the first k + 1 levels of a Galton-Watson branching
process with Po(c) as the distribution of the number of offspring from a single
parent. See Chapter 24 for some salient facts about such a process. Here Po(c) is
the Poisson random variable with mean c i.e.
ck e−c
P(Po(c) = k) = for k = 0, 1, 2, . . . , .
k!
Lemma 13.4. If Sˆ0 , Sˆ1 , . . . , Sˆk and T̂k are defined with respect to the Galton-
Watson branching process and if k ≤ k0 = (log n)3 and s0 , s1 , . . . , sk ≤ (log n)4
then
+
1
P |Si | = si , 0 ≤ i ≤ k = 1 + O 1−o(1) P |Ŝi | = si , 0 ≤ i ≤ k .
n
Proof. We use the fact that if Po(a), Po(b) are independent then
Po(a) + Po(b) has the same distribution as Po(a + b). It follows that
k
(csi−1 )si e−csi−1
P |Ŝi | = si , 0 ≤ i ≤ k = ∏ .
i=1 si !
270 CHAPTER 13. DIGRAPHS
+
Furthermore, putting ti−1 = s0 + s1 + . . . + si−1 we have for v ∈
/ Ti−1 ,
(log n)7
+ si−1
P(v ∈ Si ) = 1 − (1 − p) = si−1 p 1 + O . (13.1)
n
P |Si+ | = si , 0 ≤ i ≤ k =
(13.2)
k si
(log n)7
n − ti−1 si−1 c
=∏ 1+O
i=1 si n n
n−ti−1 −si
(log n)7
si−1 c
× 1− 1+O
n n
Here we use the fact that given si−1 ,ti−1 , the distribution of |Si+ | is the binomial
with n − ti−1 trials and probability of success given in (13.1). The lemma follows
by simple estimations.
Proof.
(a) P |Si+ | ≥ s log n||Si−1
+
| = s ≤ P (Bin(sn, c/n) ≥ s log n)
sn c s log n
≤
s log n n
s log n
snec
≤
sn log n
ec log n
≤
log n
≤ n−10 .
The proof of (b) is similar.
Lemma 13.6.
x
P(F ) = 1 − + o(1).
c
Proof. Applying Lemma 13.4 we see that
be the probability generating function of Po(c). Then Theorem 24.1 implies that
ρ = P(Eˆ ) is the smallest non-negative solution to G(ρ) = ρ. Thus
ρ = ecρ−c .
ξ
Substituting ρ = c we see that
ξ ξ
P(Eˆ ) = where = eξ −c , (13.5)
c c
and so ξ = x.
The lemma will follow from (13.4) and (13.5) and P(Fˆ |¬Eˆ ) = 1 and
P(Fˆ ∩ Eˆ ) = o(1).
(cs)k e−cs
∑ ≥ 1 − e−αs
k!
k≥ c+1
2 s
272 CHAPTER 13. DIGRAPHS
Given a population size between (log n)2 and (log n)3 at level i0 , let si denote the
population size at level i0 + i log n. Then Lemma 13.7 and the Chernoff bounds
imply that
1 2 1 2 2
P si+1 ≤ εsi (log n) ≤ exp − ε si (log n) .
2 8
It follows that
i !
1
P(Eˆ | Fˆ ) ≤ P ∃i : si ≤ ε(log n)2 2
s0 s0 ≥ (log n)
2
(
i )
∞
1 2 1 2 2
≤ ∑ exp − ε ε(log n) (log n) = o(1).
i=1 8 2
Lemma 13.8.
x
P |D− (v)| ≥ (log n)2 | |D+ (v)| ≥ (log n)2 = 1 − + o(1).
c
Proof. Expose S0+ , S1+ , . . . , Sk+ until either Sk+ = 0/ or we see that |Tk+ | ∈ [(log n)2 ,
(log n)3 ]. Now let S denote the set of edges/vertices defined by
S0+ , S1+ , . . . , Sk+ .
Let C be the event that there are no edges from Tl− to Sk+ where Tl− is the set of
vertices we reach through our BFS into v, up to the point where we first realise
13.1. STRONG CONNECTIVITY 273
that D− (v) < (log n)2 (because Si− = 0/ and |Ti− | ≤ (log n)2 ) or we realise that
D− (v) ≥ (log n)2 . Then
(log n)4
1
P(¬C ) = O = 1−o(1)
n n
and, as in (13.2),
P |Si− | = si , 0 ≤ i ≤ k | C =
k 0 si
(log n)7
n − ti−1 si−1 c
=∏ 1+O
i=1 si n n
n0 −ti−1 −si
(log n)7
si−1 c
× 1− 1+O
n n
where n0 = n − |Tk+ |.
Given this we can prove a conditional version of Lemma 13.4 and continue as
before.
Recall that any v 6∈ S is in a strong component of size ≤ α log n and so the second
part of the theorem will also be done.
274 CHAPTER 13. DIGRAPHS
P ∃ v, w ∈ S : w 6∈ D+ (v) = o(1).
(13.9)
In which case, we know that w.h.p. there is a path from each v ∈ S to every other
vertex w 6= v in S.
To prove (13.9) we expose S0+ , S1+ , . . . , Sk+ until we find that
|Tk+ (v)| ≥ n1/2 log n. At the same time we expose S0− , S1− , . . . , Sl− until we find that
|T − (w)| ≥ n1/2 log n. If w 6∈ D+ (v) then this experiment will have tried at least
l 2
n1/2 log n times to find an edge from D+ (v) to D− (w) and failed every time.
The probability of this is at most
c n(log n)2
1− = o(n−2 ).
n
This completes the proof of Theorem 13.2.
Given this, one only has to show that if ω 6→ −∞ then w.h.p. there does not exist
a set S such that (i) 2 ≤ |S| ≤ n/2 and (ii) E(S : S̄) = 0/ or E(S̄ : S) = 0/ and (iii)
S induces a connected component in the graph obtained by ignoring orientation.
But, here with s = |S|,
n/2
n s−2
P(∃ S) ≤ 2 ∑ s (2p)s−1 (1 − p)s(n−s)
s=2 s
13.2. HAMILTON CYCLES 275
Theorem 13.10.
(b) not (a) and there exists a Hamilton cycle if at least one of →
−
ei , ←
e−i is present,
or
(a) and (c) give the same conditional probability of Hamiltonicity in Γi , Γi−1 . In
Γi−1 (b) happens with probability p. In Γi we consider two cases (i) exactly one of
→
−
ei , ←
e−i yields Hamiltonicity and in this case the conditional probability is p and (ii)
either of →−ei , ←
e−i yields Hamiltonicity and in this case the conditional probability is
1 − (1 − p)2 > p.
Note that we will never require that both → −
ei , ←
e−i occur.
Theorem 13.10 was subsequently improved by Frieze [337], who proved the equiv-
alent of Theorem 6.5.
log n+cn
Theorem 13.11. Let p = n . Then
0
if cn → −∞
−c
lim P(Dn,p has a Hamilton cycle) = e−2e if cn → c
n→∞
1 if cn → ∞.
distinct families P obtained from this partitioning in this manner. Since this
occurs w.h.p. we conclude (applying the Markov inequality to the number of
partitions for which the bound fails) that this bound holds for (1 − o(1))-fraction
of such partitions. Since there are (n−s)!` such partitions, one can find at least
(m!)
(n − s)!
(1 − o(1)) `
(1 − o(1))n−s (m!)` pn−s
(m!)
= (1 − o(1))n−s (n − s)!pn−s = (1 − o(1))n n!pn
We show next how to close a given family of paths into a Hamilton cycle. For
each such family P, let A := A(P) denote the collection of all pairs (sP ,tP )
where sP is a starting point and tP is the endpoint of a path P ∈ P, and define
an auxiliary directed graph D(A) as follows. The vertex set of D(A) is V (A) =
S ∪ {zP = (sP ,tP ) : zP ∈ A}. Edges of D(A) are determined as follows: if u, v ∈ S
and (u, v) ∈ E(D) then (u, v) is an edge of D(A). The in-neighbors (out-neighbors)
of vertices zP in S are the in-neighbors of sP in D (out-neighbors of tP ). Lastly,
(zP , zQ ) is an edge of D(A) if (tP , sQ ) is an edge D.
Clearly D(A) is distributed as Ds+m,p , and that a Hamilton cycle in D(A) corre-
sponds to a Hamilton cycle in D after adding the corresponding paths between
each sP and tP . Now distinct families P 6= P 0 yield distinct Hamilton cycles (to
see this, just delete the vertices of S from the Hamilton cycle, to recover the paths).
Using Theorem 13.11 we see that for p = ω (log n/(s + m)) = ω (log(s + m)/(s + m)),
the probability that D(A) does not have a Hamilton cycle is o(1). Therefore,
using the Markov inequality we see that for almost all of the families P, the
corresponding auxiliary graph D(A) is indeed Hamiltonian and we have at least
(1 − o(1))n n!pn distinct Hamilton cycles, as desired.
278 CHAPTER 13. DIGRAPHS
13.3 Exercises
13.3.1 Let p = log n+(k−1)nlog log n+ω for a constant k = 1, 2, . . .. Show that w.h.p.
Dnp is k-strongly connected.
Show that if p = ω(log n/n) then with probability 1 − n−ω(1) , Gn,n,p con-
tains (1 − o(1))np edge disjoint perfect matchings.
13.3.3 Show that if p = ω((log n)2 /n) then w.h.p. Gn,p contains eo(n) n!pn distinct
Hamilton cycles.
13.3.5 Let T be a random tournament. Show that w.h.p. the size of the largest
acyclic sub-tournament is asymptotic to 2 log2 n. (A tournament is acyclic
if it contains no directed cycles).
13.3.6 Suppose that 0 < p < 1 is constant. Show that w.h.p. the size of the
largest acyclic tournament contained in Dnp is asymptotic to 2 logb n where
b = 1/p.
13.3.7 Let mas(D) denote the number of vertices in the largest acyclic subgraph
of a digraph D. Suppose that 0 < p < 1 is constant. Show that w.h.p.
mas(Dn,p ) ≤ 4log
log n 1
q where q = 1−p .
13.3.8 Consider the random digraph Dn obtained from Gn,1/2 by orienting edge
(i, j) from i to j when i < j. This can be viewed as a partial order on [n]
and is called a Random Graph Order. Show that w.h.p. Dn contains a path
of length at least 0.51n. (In terms of partial orders, this bounds the height
of the order).
13.3.9 Show that if np ≥ log10 n then w.h.p. Dn,p is 12 ±o(1) resilient, i.e. 21 − ε np ≤
∆H ≤ 12 + ε np. (Hint: just tweak the proof of Theorem 9.3 so that the
13.3.10 Let O represent an orientation of the edges of a Hamilton cycle. Show that
13.4 Notes
Packing
The paper of Frieze [337] was in terms of the hitting time for a digraph process Dt .
It proves that the first time that the δ + (Gt ), δ − (Gt ) ≥ k is w.h.p. the time when Gt
has k edge disjoint Hamilton cycles. The paper of Ferber, Kronenberg and Long
[305] shows that if p = ω((log n)4 /n) then w.h.p. Dn,p contains (1 −o(1))np edge
disjoint Hamilton cycles.
Long Cycles
The papers by Hefetz, Steger and Sudakov [422] and by Ferber, Nenadov, Noever,
Peter and Škorić [308] study the local resilience of having a Hamilton cycle. In
8
particular, [308] proves that if p (lognn) then w.h.p. one can delete any subgraph
H of Dn,p with maximum degree at most ( 21 − ε)np and still leave a Hamiltonian
subgraph.
Krivelevich, Lubetzky and Sudakov [526] proved that w.h.p. the random digraph
Dn,p , p = c/n contains a directed cycle of length (1 − (1 + εc )e−c )n where εc → 0
as c → ∞.
Cooper, Frieze and Molloy [231] showed that a random regular digraph with in-
degree = outdegree = r is Hamiltonian w.h.p. iff r ≥ 3.
Connectivity
Cooper and Frieze [220] studied the size of the largest strong component in a
random digraph with a given degree sequence. The strong connectivity of an
inhomogeneous random digraph was studied by Bloznelis, Götze and Jaworski in
[104].
280 CHAPTER 13. DIGRAPHS
Chapter 14
Hypergraphs
281
282 CHAPTER 14. HYPERGRAPHS
A hypertree of size 5.
((k − 1)`)!
Nk (`) = ((k − 1)` + 1)`−1 .
`! ((k − 1)!)`
b !`+b !t (n−s)
k−1
n s s `−1 ((k − 1)`)! c c
t 1 − n−1 (14.1)
s a k `!((k − 1)!)` n−1
k−1 k−1
!`+b
ne s sa+kb k−2
` (k − 1)k−1 ` eo(1) c(k − 1)!
≤ t `−1 k−2+o(1) e−(c+o(1))t
s k!b e (k − 1)! nk−1
14.1. COMPONENT SIZE 283
!`
skb−1 cb na+1−b(k−1) ea+1−c e1−(k−1)c+o(1) c(k − 1)k−1 `k−2t
≤
kbt sk−1
!`
skb−1 cb na+1−b(k−1) ea+1−c eo(1) e1−(k−1)c c(k − 1)k `k−1
≤
kbt ((k − 1)`)k−1
skb−1 cb na+1−b(k−1) ea+1−c o(1) 1−(k−1)c `
= e ce (k − 1) (14.2)
kbt
eo(1) ce1−(k−1)c (k − 1) ≤ 1 − εc .
na+1−b(k−1)+O(b log log n/ log n) e−εc ` ≤ n1−b+O(b log log n/ log n) e−εc ` . (14.3)
For the second expression, we used the fact that a ≤ b(k − 2). The second expres-
sion in (14.3) tends to zero if b > 1 and so we can assume that b ≤ 1. The first
expression in (14.3) then tends to zero if a + 1 < b(k − 1) and this verifies Part
(a). Because a + 1 − b(k − 1) ≤ 1, we see that Part (b) of the lemma follows with
A = 2/εc . Part (c) follows from the Markov inequality.
Lemma 14.4. W.h.p. there are no components in the range [A log n, log4 n] and all
but logk+1 n of the small components (of size at most A log n) are hypertrees.
Proof. Now suppose that S is a small component of size s which is not a hypertree
and let C be the set of vertices of a maximal hypertree with t = (k −1)`+1 vertices
and ` edges that is a subgraph of S. (Maximal in the sense that it is not contained
in any other hypertree of S.) Lemma 14.3 implies that w.h.p. there is at most
one edge e in S that is not part of C but is incident with at least two vertices of
C. Furthermore, Lemma 14.3(c) implies that the number of sets with b = 1 is
O(logk+1 n) w.h.p.
1
Lemma 14.5. If c < k−1 then w.h.p. the largest component has size O(log n).
Proof. Fix an edge e and do a Breadth First Search (BFS) on the edges starting
with e. We start with L1 = e and let Lt denote the number of vertices at depth t in
the search i.e the neighbors of Lt−1 that are not in Lt−1 . Then |Lt+1 | is dominated
n
by (k − 1)Bin |Lt | k−1 , p . So,
n
E(|Lt+1 | | Lt ) ≤ (k − 1)|Lt | p ≤ θ |Lt |
k−1
284 CHAPTER 14. HYPERGRAPHS
2 log n
where θ = (k − 1)c < 1. It follows that if t0 = log 1/θ then
/ ≤ nkθ t0 = o(1).
Pr(∃e : Lt0 6= 0)
So, w.h.p. there are at most t0 levels. Furthermore, if |Lt | ever reaches log2 n then
the Chernoff bounds imply that w.h.p. |Lt+1 | ≤ |Lt |. This implies that the maxi-
mum size of a component is O(log3 n) and hence, by Lemma 14.4, the maximum
size is O(log n).
1
Lemma 14.6. If c > k−1 then w.h.p. there is a unique giant component of size
Ω(n) and all other components are of size O(log n).
Now
n − o(n)
E(Yt ) = (k − 1)|Lt | p = (1 − o(1))c|Lt | = θ |Lt | where θ > 1.
k−1
n−o(n)
The Chernoff bounds applied to Bin |Lt | k−1 , p imply that
(θ − 1)2 |Lt |
1+θ
P Yt ≤ |Lt | ≤ exp − .
2 3(k − 1)
1+θ t
P |Lt | ≥
2
( ` )!
t (θ − 1)2 1+θ t log8 n
2
≥ ∏ 1 − exp − −O (14.4)
`=1 3(k − 1) n
14.1. COMPONENT SIZE 285
≥ γ,
where γ > 0 is a positive1+θconstant. This lower bound of γ follows from the fact
`
2
(θ −1) ( 2 )
that ∑∞`=1 exp − 3(k−1) converges, see Apostol [39], Theorem 8.55. Note
t
that if t = 2 log1+θ
log n
then 1+θ = log2 n. It follows that the probability we fail to
log 2
2
grow a component of size log2 n after s attempts is at most (1 − γ)s . Choosing
s = log21/(1−γ)
log n
we see that after exploring O(log3 n) vertices, we find a component
of size at least log2 n, with probability 1 − n−(1−o(1)) .
We show next that with (conditional) probability 1 − O(n−(1−o(1)) ) the component
of size at least log2 n will in fact have at least n0 = n(k−1)/k log n vertices. We
handle Xt ,Yt exactly as above. Going back to (14.4), if we run BFS for another
O(log n) steps then, starting with |Lt0 | ≈ log2 n, we have
1+θ t 2
P |Lt0 +t | ≥ log n
2
(θ − 1)2 log2 n
t 9
log n
≥ ∏ 1 − exp − −O
`=1 3(k − 1) n
= 1 − n−(1−o(1)) . (14.5)
It follows that w.h.p. Hn,p;k only contains components of size O(log n) and Ω(no ).
For this we use the fact that we only need to apply (14.5) less than n/n0 times.
We now prove that there is a unique giant component. This is a simple sprin-
kling argument. Suppose that we let (1 − p) = (1 − p1 )(1 − p2 ) where p2 = ωn1k−1
for some ω → ∞ slowly. Then we know from Lemma 14.4 that there is a gap in
component sizes for H(n, p1 , k). Now add in the second round of edges with prob-
ability p2 . If C1 ,C2 are distinct components of size at least n0 then the probability
there is no C1 : C2 edge added is at most
( n0 )
n0
k−1 n0
∑k−1
n0 ∑
(1 − p2 ) i=1 ( i )( ) ≤ exp −
k−i
i=1 i k−i
≤
ωn k−1
( ) ( )
2k−1 nk0 2k−1 logk n
exp − k−1 = exp − .
ωn ω
So, w.h.p. all components of size at least n0 are merged into one component.
We now look at the size of this giant component. The fact that almost all small
1
components are hypertrees when c < k−1 yields the following lemma. The proof
286 CHAPTER 14. HYPERGRAPHS
follows that of Lemma 2.13 and is left as Exercise 13.4.5. For c > 0 we define
(
1
c c ≤ k−1
x = x(c) = 1 .
1
to xe−(k−1)x = ce−(k−1)c c > k−1
The solution in 0, k−1
Lemma 14.7. Suppose that c > 0 is constant and c(k − 1) 6= 1 and x is as defined
above. Then
∞
c` ((k − 1)` + 1)` e−c((k−1)`+1) x 1/(k−1)
∑ = .
`=0 `!((k − 1)` + 1) c
0
cn → −∞.
−c
lim P(Hn,p;k is connected) = e−e cn → c.
n→∞
1 cn → ∞.
Dudek, Frieze, Loh and Speiss [265] and in Ferber [303]. Also, see Section 14.3.
The following theorem is from Dudek and Frieze [264]. Furthermore, we assume
that k ≥ 3.
Theorem 14.10.
(i) If p ≤ (1 − ε)e/n, then w.h.p. Hn,p;k is not Hamiltonian.
(iii) For all fixed ε > 0, if k ≥ 4 and p ≥ (1 + ε)e/n, then w.h.p. Hn,p;k is Hamil-
tonian.
Proof. We will prove parts (i) and (ii) and leave the proof of (iii) as an exercise,
with a hint.
Let ([n], E ) be a k-uniform hypergraph. A permutation π of [n] is Hamilton cycle
inducing if
(We use the convention π(n + r) = π(r) for r > 0.) Let the term hamperm refer to
such a permutation.
Let X be the random variable that counts the number of hamperms π for Hn,p;k .
Every Hamilton cycle induces at least one hamperm and so we can concentrate on
estimating P(X > 0).
Now
E(X) = n!pn .
This is because π induces a Hamilton cycle if and only if a certain n edges are all
in Hn,p;k .
For part (i) we use Stirling’s formula to argue that
√ np n √
E(X) ≤ 3 n ≤ 3 n(1 − ε)n = o(1).
e
This verifies part (i).
We see that np n
E(X) ≥ →∞ (14.6)
e
in parts (ii) and (iii).
Fix a hamperm π. Let H(π) = (Eπ (1), Eπ (2), . . . , Eπ (n)) be the Hamilton cy-
cle induced by π. Then let N(b, a) be the number of permutations π 0 such that
|E(H(π)) ∩ E(H(π 0 ))| = b and E(H(π)) ∩ E(H(π 0 )) consists of a edge disjoint
paths. Here a path is a maximal sub-sequence F1 , F2 , . . . , Fq of the edges of H(π)
288 CHAPTER 14. HYPERGRAPHS
Sq
such that Fi ∩ Fi+1 6= 0/ for 1 ≤ i < q. The set j=1 Fj may contain other edges of
H(π). Observe that N(b, a) does not depend on π.
Note that
n b
E(X 2 ) n!N(0, 0)p2n n!N(b, a)p2n−b
= + ∑ ∑ E(X)2 .
E(X)2 E(X)2 b=1 a=1
Since trivially, N(0, 0) ≤ n!, we obtain,
n b
E(X 2 ) n!N(b, a)p2n−b
≤ 1 + ∑ ∑ E(X)2 . (14.7)
E(X)2 b=1 a=1
We show that
n b n b
n!N(b, a)p2n−b N(b, a)pn−b
∑ ∑ E(X)2 = ∑ ∑ E(X) = o(1). (14.8)
b=1 a=1 b=1 a=1
E(X 2 )
P(X = 0) ≤ − 1 = o(1),
E(X)2
as required.
It remains to show (14.8). First we find an upper bound on N(b, a). Choose a
vertices vi , 1 ≤ i ≤ a, on π. We have at most
na (14.9)
choices. Let
b1 + b2 + · · · + ba = b,
where bi ≥ 1 is an integer for every 1 ≤ i ≤ a. Note that this equation has exactly
b−1
< 2b (14.10)
a−1
Thus, by the above considerations we can find a edge disjoint paths in H(π) with
the total of b edges in at most
na (2k)b (14.11)
many ways.
Let P1 , P2 , . . . , Pa be any collection of the above a paths. Now we count the number
of permutations π 0 containing these paths.
First we choose for every Pi a sequence of vertices inducing this path in π 0 . We
see the vertices in each edge of Pi in at most k! orderings. Crudely, every such
sequence can be chosen in at most (k!)bi ways. Thus, we have
a
∏(k!)bi = (k!)b (14.12)
i=1
(n − b − a(k − 1))!
Since p n n
E(X) = n!pn = (2 + o(1))πn pn ,
e
we get
N(b, a)pn−b e b+a(k−1)
< (1 + o(1))n2a (2k!k)b p−b .
E(X) n
Finally, since a ≤ b we estimate eb+a(k−1) ≤ ekb , and consequently,
b
N(b, a)pn−b 2k!kek
1 + o(1)
< . (14.15)
E(X) np na(k−3)
Proof of (ii):
Here k = 3 and np ≥ ω. Hence, we obtain in (14.15)
b
N(b, a)pn−b 2k!kek
≤ (1 + o(1)) .
E(X) ω
Thus,
n b n b
N(b, a)pn−b 2k!kek
∑ ∑ E(X) < (1 + o(1)) ∑ b ω = o(1). (14.16)
b=1 a=1 b=1
14.3 Thresholds
In this section we describe a breakthrough result of Frankston, Kahn, Narayanan
and Park [327]. It will give us good estimates of the thresholds for various struc-
tures.
14.3. THRESHOLDS 291
has a perfect matching w.h.p. This was first proved in a breakthrough paper by
Johansson, Kahn and Vu [468].
Loose Hamilton Cycles: We consider the case r = 1 of Section 14.2. In this case
[n]
X = r and the edges of H correspond to the loose Hamilton cycles of the com-
plete k-uniform hypergraph Hn,r on vertex set [n]. We see from Exercise 14.4.1
that we need at least Ω(n log n) random edges to have a loose Hamilton cycle
w.h.p. Now there are n/(k −1) edges in a loose cycle and so we take r = n/(k −1).
The number of loose Hamilton cycles in Hn,r is given by k−1 n!
2n · (k−2)!n/(k−1) , Exer-
cise 14.4.11, and for a set S = {e1 , e2 , . . . , es } ⊆ X we have
Arguing as for the Shamir problem, we see that there exists K > 0 such if m =
Kn log n then Hn,m;k has a loose hamilton cycle w.h.p. This being the result of
[341], [263] and [265].
Powers of Hamilton cycles: The kth power of a Hamilton cycle in a graph G=
(V, E) is a permutation x1 , x2 , . . . , xn of the vertices V such that xi , xi+ j is an
edge of G for all i ∈ [n], j ∈ [k]. Kühn and Osthus [537] studied the existence of
kth powers in Gn,p . They showed that for k ≥ 3 one could use Riordan’s Theorem
[671] to show that if npk → ∞ then Gn,p contains the kth power of a Hamilton
cycle w.h.p. This is tight as the first moment method shows that if npk → 0 then
w.h.p. there are no kth powers. The problem is more difficult for k = 2 and then
after a series of papers, Nenadov and Škorić [629], Fischer, Škorić, Steger and
2n
Trujić [312], Montgomery [610] we have an upper bound of p log n1/2
. Theorem
3/2
14.11 reduces the bound to O(n log n) in Gn,m , as we will now show.
We take X = [n] 2 and the edges of H correspond to the squares of Hamilton
We can therefore take κ = e−1 n1/2 , and then (14.18) yields the claimed upper
bound of O(n3/2 log n) on the threshold for the existence of the square of a Hamil-
ton cycle in Gn,m .
14.3. THRESHOLDS 293
we have
|E(H ) ∩ hSi|
= P(S ⊆ π(T )) ≤ κ −|S| . (14.21)
|E(H )|
χ(S,W ) = ψ(S ∪W ) \W ;
and say that the pair (S,W ) is bad if |χ(S,W )| > r0 and good otherwise. (It is
important to remember that ψ(S ∪ W ) is determined by S ∪ W and not by S. It
would be incorrect to take χ(S,W ) = S \W .)
294 CHAPTER 14. HYPERGRAPHS
The idea now is to choose a small random set W and argue that w.h.p. there
exists H ∈ H such that |H \W | is significantly smaller than |H|. We then repeat
the argument with respect to the hypergraph H \ W . In this way, we build up a
member of H bit by bit. After O(log r) iterations we can prove the existence of a
final small piece by using the second moment method.
Lemma 14.12. For H as above, and W chosen uniformly from np X
,
|{(S,W ) : (S,W ) is bad and |S| = s}| ≤ (γr)−1 N|H |C−r/3 . (14.23)
(Note that γr = r − r0 bounds the number of s for which the set in question can be
nonempty, whence the factor (γr)−1 .)
For Z ⊇ S ∈ Hs , we say that (S, Z) is pathological if there is T ⊆ S with t := |T | >
r0 and
|{S0 ∈ Hs : S0 ∈ [T, Z]}| > Cr/2 |H |κ −t ps−t . (14.24)
(Note that one would expect to bound the typical size of {S0 ∈ Hs : S0 ∈ [T, Z]} by
something of the order |H |κ −t ps−t .)
14.3. THRESHOLDS 295
choices for Z = W ∪ S.
Step 2. Given Z, let S0 = ψ(Z). Choose T := S ∩ S0 , for which there are at most
0
2|S | ≤ 2r possibilities, and set t = |T | > r0 . (If t ≤ r0 then (S,W ) cannot be bad, as
χ(S,W ) = S0 \W ⊆ T .)
Step 3. Since we are only interested in nonpathological choices, and we choose
S \ S0 by choosing S ∈ [T, Z], the number of possibilities for S is now at most
Cr/2 |H |κ −t ps−t .
2r+s N|H |Cr/2 (pκ)−t ≤ N|H |(4C1/2 )rC−t < N|H |[4C−(1/2−γ) ]r . (14.26)
(Here the left hand side counts the members of Hs in Z whose intersection with S
is precisely U. Of course, the existence of U as in (14.27) follows from (14.24).)
The number of possibilities for this choice is at most 2s−t .
296 CHAPTER 14. HYPERGRAPHS
Step 4. Choose Z \ S, the number of choices for which is less than N(2/C1/2 )r .
To see this, write Φ for the R.H.S. of (14.27). Noting that Z \ S must belong to
X\S X\S X\S
np ∪ np−1 ∪ · · · ∪ np−s , we consider, for Y drawn uniformly from this set,
(since n − s ≥ n/2). The Markov Inequality then bounds the probability in (14.28)
by θ /Φ, and this bounds the number of possibilities for Z \S by N(θ /Φ) (cf. (14.25)),
which is easily seen to be less than N(2/C1/2 )r .
Step 5. Complete the specification of (S,W ) by choosing S ∩ W , which can be
done in at most 2s ways.
Combining (and slightly simplifying), we find that the number of pathological
possibilities is at most
|H |N(16/C1/2 )r . (14.29)
Finally, the sum of the bounds in (14.26) and (14.29) is less than the
(γr)−1 N|H |C−r/3 of (14.23).
Small uniformities
Very small set sizes are handled by a simple use of the Janson inequality, Theorem
22.13.
Lemma 14.13. For an r-bounded, κ-spread G on Y , and α ∈ (0, 1),
!−1
r
r
P(Yα 6∈ hG i) ≤ exp − 2 ∑ (ακ)−t . (14.30)
t=1 t
14.3. THRESHOLDS 297
Proof. Denote the members of G by Si and set ζi = 1{Yα ⊇Si } . We add r − |Si | new
elements to each Si to create an r-uniform hypergraph G 0 on a set of vertices Y 0 .
Note that Yα0 ∈ hG 0 i implies Yα ∈ hG i. This is for the purposes of the proof only
and for the rest of the proof of this lemma, we let G = G 0 .
µ := ∑ E(ζi ) = α r |G |
and
∆= ∑ E(ζi ζ j ) =
i, j:Si ∩S j 6=0/
r r
∑α ∑ ∑ ∑ r
α r−t
≤ µ max ∑ α −t ∑ ∑ αr
t=1 T ⊆G H∩G=T G∈G T ⊆G H∩G=T
G∈G t=1
|T |=t |T |=t
( )
r r
≤ µ ∑ α −t max max ∑ α r
t=1 G∈G t T ⊆G H⊇T
|T |=t
r r r
−t r α |G | 2 −t r
≤µ ∑α = µ ∑ (ακ) . (14.31)
t=1 t κt t=1 t
where we use the fact that any binomial ξ with E[ξ ] ∈ Z satisfies P(ξ ≤ E[ξ ]) ≥
1/2.
298 CHAPTER 14. HYPERGRAPHS
C0 γ −1 log r C0 γ −1 log r
n
|X` | = n − `np ≥ n 1 − ≥ n 1 − −1 = .
κ 2γ C0 log r 2
For i ∈ [`] call Wi successful if |Hi | ≥ (1 − δ )|Hi−1 |, call W`+1 successful if it lies
in hH` i, and say a sequence of Wi ’s is successful if each of its entries is. We show
that h p i
P(W1 . . .W`+1 is successful) = 1 − exp −Ω( log r) . (14.33)
Now W1 . . .Wi−1 successful implies that |Hi−1 | > (1 − δ )` |H | > |H |/2. So for
I ⊆ Xi−1 we have
|H | |Hi−1 |
|Hi−1 ∩ hIi| ≤ |H ∩ hIi| ≤ ≤ .
(2κ)|I| κ |I|
We therefore have the spread condition (14.17) for Hi−1 . For i ∈ [`], according to
Lemma 14.12 (and the Markov Inequality),
Thus
`
P(W1 . . .W` is successful) > 1 − δ −1 ∑ C−ri−1 /3
i=1
n p o
> 1 − exp − log r/4 (14.34)
√
(using r` = log r).
Finally, if W1 . . .W` is successful, then Corollary 14.14 applied with G = H` ,
Y = X` , α = nq/|Y | ≥ q, r = r` , and W = W`+1 gives
n p o
P(W`+1 6∈ hH` i) ≤ 2 exp − log r/4 , (14.35)
Sparse Hypergraphs
The above proof can be adapted to prove the following result related to the Shamir
problem.
300 CHAPTER 14. HYPERGRAPHS
C
Theorem 14.15. Suppose that p = n where C is a constant and C > C0 . Then
(k−1 )
w.h.p. Hn,p;k contains a matching of size at least
n −Cγ/C0
1−e .
k
Proof. Let now ` = bC/C0 c and let δ = 1/(2`) as before. The first inequality in
(14.34) continues to hold with r1 , r2 , . . . , r`−1 ≥ r` = Ω(n). It follows that w.h.p.
there is a matching of size
n n
− r` ≥ 1 − (1 − γ)C/C0 .
k k
ft,S = |H |−1 | {J ∈ H : |J ∩ S| = t} |.
Theorem 14.17. Suppose that H is κ-spread and r-uniform and that there exist
constants α, K0 such that for all S ∈ H ,
t
K0
ft,S ≤ for 1 ≤ t ≤ αr. (14.36)
κ
Then
Cε |X|
∀ε > 0, ∃Cε such that m ≥ implies that
κ
Pr(Xm contains a copy of e ∈ H ) ≥ 1 − ε. (14.37)
Proof. The strategy here is similar to that used to prove Theorem 14.11. The
main difference is that we replace the O(log r) rounds by a single round where we
obtain almost all of an edge of H .
14.3. THRESHOLDS 301
Now
k
Var(Z) ≤ p2k
1 ∑ ∑ p−t
1 . (14.40)
t=1 A,B∈R
|A∩B|=t
k t k t
K0 K0
Var(Z) ≤ |R| |H |p2k
1 ∑ p−t
1 ≤ 2|R|2 p2k
1 ∑ ≤
t=1 κ t=1 κ p1
k t
K0
2|R|2 p2k
1 ∑ (using κ p1 ≥ κm/N ≥ C)
t=1 C
k t
2 K0 4K0
≤ 2 E(Z) ∑ ≤ E(Z)2 ,
t=1 C C
Var(Z) 4K0
Pr(Z = 0) ≤ ≤ .
E(Z)2 C
Proof of Claims
Proof of Claim 10 Fix t. We bound the number of bad (S,W )’s with |W ∩ S| =
E
t. In which case, |W \ S| = m − t and |W ∪ S| = m − t + r. Call Y ⊆ m−t+r
pathological if
−k/3 N −r N
| {S ⊆ Y : (S,Y \ S) is bad} | > C |H | .
m−t m−t +r
Pathological contributions:
Claim 11. For a fixed S ∈ H and random W \ S from E\S
m−t and t ≥ k and large
enough C,
N−r
−2k/3 m−t
E (| {H 3 J ⊆ W ∪ S : |J ∩ S| = t|}) ≤ C |H | N .
(14.42)
m−t+r
14.3. THRESHOLDS 303
for arbitrary J ∈ H , |J ∩ S| = t.
It is therefore enough to show that
m−t N
r−t m−t+r 1
ft,S N−r N−r ≤ . (14.45)
r−t m−t
C2k/3
Observe that
m−t r−t N r
r−t m−t m−t+r N −r
N−r
≤ and N−r
≤ .
r−t
N −r m−t
m−t
304 CHAPTER 14. HYPERGRAPHS
14.4 Exercises
14.4.1 Show that if m = cn log n then (i) ck < 1 implies that Hn,m;k has isolated
vertices w.h.p. and (ii) if ck > 1 then Hn,m;k is connected w.h.p.
(rn)!
rn/k n
e−(k−1)(r−1)/2 .
(k!) r! (rn/k)!
d nk−1 p
χ1 (Hn,p;k ) ≈ where d = .
2 log d (k − 2)!
14.4.13 Let U1 ,U2 , . . . ,Uk denote k disjoint sets of size n. Let H Pn,m,k denote the
set of k-partite, k-uniform hypergraphs with vertex set V = U1 ∪U2 ∪ · · · ∪
Uk and m edges. Here each edge contains exactly one vertex from each
Ui , 1 ≤ i ≤ k. The random hypergraph HPn,m,k is sampled uniformly from
H Pn,m,k . Prove the k-partite analogue of Shamir’s problem viz. there
exists a constant K > 0 such that if m ≥ Kn log n then
14.4.14 Find the threshold, up to a log n factor for the existence of the following
structures in Gn,p : replace 4 by n in an appropriate way.
14.5 Notes
Components and cores
If H = (V, E) is a k-uniform hypergraph and 1 ≤ j ≤ k −1 then two sets J1 , J2 ∈ Vj
The notion of a core extends simply to hypergraphs and the sizes of cores in ran-
dom hypergraphs has been considered by Molloy [602]. The r-core is the largest
sub-hypergraph with minimum degree r. Molloy proved the existence of a con-
stant ck,r such that if c < cr,k then w.h.p. Hn,cn;k has no r-core and that if c > cr,k
then w.h.p. Hn,cn;k has a r-core. The efficiency of the peeling algorithm for find-
ing a core has been considered by Jiang, Mitzenmacher and Thaler [466]. They
show that w.h.p. the number of rounds in the peeling algorithm is asymptotically
log log n
log(k−1)(r−1) if c < cr,k and Ω(log n) if c > cr,k . Gao and Molloy [372] show that
for |c − cr,k | ≤ n−δ , 0 < δ < 1/2, the number of rounds grows like Θ̃(nδ /2 ). In
this discussion, (r, k) 6= (2, 2).
Chromatic number
Krivelevich and Sudakov [529] studied the chromatic number of the random k-
uniform hypergraph Hn,p;k . For 1 ≤ γ ≤ k − 1 we say that a set of vertices S is
γ-independent in a hypergraph H if |S ∩ e| ≤ γ. The γ-chromatic number of a
hypergraph H = (V, E) is the minimum number of sets in a partition of V into
γ-independent sets. They show that if d (γ) = γ k−1
γ
n−1
k−1 p is sufficiently large
d (γ)
then w.h.p. (γ+1) log d (γ)
is a good estimate of the γ-chromatic number of Hn,p;k .
308 CHAPTER 14. HYPERGRAPHS
Dyer, Frieze and Greenhill [278] extended the results of [6] to hypergraphs. Let
uk,` = `k−1 log `. They show that if uk,`−1 < c < uk,` then w.h.p. the (weak (γ =
k − 1)) chromatic number of Hn,cn;k is either k or k + 1.
Achlioptas, Kim, Krivelevich and Tetali [3] studied the 2-colorability of H =
Hn,p;k . Let m = nk p be the expected number of edges in H. They show that
if m = c2k n and c > log2 2 then w.h.p. H is not 2-colorable. They also show that if
c is a small enough constant then w.h.p. H is 2-colorable.
Orientability
Gao and Wormald [374], Fountoulakis, Khosla and Panagiotou [324] and
Lelarge [542] discuss the orientability of random hypergraphs. Suppose that 0 <
` < k. To `-orient an edge e of a k-uniform hypergraph H = (V, E), we assign
positive signs to ` of its vertices and k − ` negative signs to the rest. An (`, r)-
orientation of H consists of an `-orientation of each of its edges so that each
vertex receives at most r positive signs due to incident edges. This notion has
uses in load balancing. The papers establish a threshold for the existence of an
(`, r)-orientation. Describing it it is somewhat complex and we refer the reader to
the papers themselves.
VC-dimension
Ycart and Ratsaby [764] discuss the VC-dimension of H = Hn,p;k . Let p = cn−α
for constants c, α. They give the likely VC-dimension of H for various values of
α. For example if h ∈ [k] and α = k − h(h−1)
h+1 then the VC-dimension is h or h − 1
w.h.p.
Erdős-Ko-Rado
The famous Erdős-Ko-Rado theorem states that if n > 2k then the maximum size
n−1
of a family of mutually intersecting k-subsets of [n] is k−1 and this is achieved
by all the subsets that contain the element 1. Such collections will be called stars.
Balogh, Bohman and Mubayi [49] considered this problem in relation to the ran-
dom hypergraph Hn,p;k . They consider for what values of k, p is it true that maxi-
mum size intersecting family of edges is w.h.p. a star. More recently Hamm and
Kahn [408], [409] have answered some of these questions. For many ranges of
k, p the answer is as yet unknown.
Bohman, Cooper, Frieze, Martin and Ruszinko [115] and Bohman,
Frieze, Martin, Ruszinko and Smyth [116] studied the k-uniform hypergraph H
obtained by adding random k-sets one by one, only adding a set if it intersects all
14.5. NOTES 309
previous sets. They prove that w.h.p. H is a star for k = o(n1/3 ) and were able to
analyse the structure of H for k = o(n5/12 ).
Other models
311
Chapter 15
Trees
The properties of various kinds of trees are one of the main objects of study in
graph theory mainly due to their wide range of application in various areas of
science. Here we concentrate our attention on the “average” properties of two
important classes of trees: labeled and recursive. The first class plays an important
role in both the sub-critical and super-critical phase of the evolution of random
graphs. On the other hand random recursive trees serve as an example of the very
popular random preferential attachment models. In particular we will point out,
an often overlooked fact, that the first demonstration of a power law for the degree
distribution in the preferential attachment model was shown in a special class of
inhomogeneous random recursive trees.
The families of random trees, whose properties are analyzed in this chapter, fall
into two major categories according to the order of their heights: they are either
of square root (labeled trees) or logarithmic (recursive trees) height. While most
of square-root-trees appear in probability context, most log-trees are encountered
in algorithmic applications.
313
314 CHAPTER 15. TREES
d1 + d2 + · · · + dn = 2(n − 1),
Theorem 15.1. Denote by ∆ = ∆(Tn ) the maximum degree of a random tree. Then
w.h.p.
log n
∆(Tn ) ≈ .
log log n
The classical approach to the study of the properties of labeled trees chosen at
random from the family of all labeled trees was purely combinatorial, i.e., via
counting trees with certain properties. In this way, Rényi and Szekeres [667],
using complex analysis, found the height of a random labeled tree on n vertices
(see also Stepanov [725], while for a general probabilistic context of their result,
see a survey paper by Biane, Pitman and Yor [93]).
Assume that a tree with vertex set V = [n] is rooted at vertex 1. Then there is a
unique path connecting the root with any other vertex of the tree. The height of
a tree is the length of the longest path from the root to any pendant vertex of the
tree. Pendant vertices are the vertices of degree one.
15.1. LABELED TREES 315
P(Y1 = r) = pr for r = 1, 2, . . . , N.
Then
M
P(ZM = N) = P(Y1 +Y2 + . . . +YN = N − M).
N
Now, instead of a random tree Tn chosen from the family of all labeled trees Tn
on n vertices, consider a tree chosen at random from the family of all (n + 1)n−1
trees on n + 1 vertices, with the root labeled 0 and all other vertices labeled from
1 to n. In such a random tree, with a natural orientation of the edges from the root
to pendant vertices, denote by Vt the set of vertices at distance t from the root 0.
Let the number of outgoing edges from a given vertex be called its out-degree and
+
Xr,t be the number of vertices of out-degree r in Vt . For our branching process,
choose the probabilities pr , for r = 0, 1, . . ., as equal to
λ r −λ
pr = e ,
r!
i.e., assume that the number of offspring has the Poisson distribution with mean
λ > 0. Note that λ is arbitrary here.
316 CHAPTER 15. TREES
Let Zr,t be the number of particles in the tth generation of the process, having
exactly r offspring. Next let X = [mr,t ], r,t = 0, 1, . . . , n be a matrix of non-negative
integers. Let st = ∑nr=0 mr,t and suppose that the matrix X satisfies the following
conditions:
(i) s0 = 1,
(iii) s0 + s1 + . . . + sn = n + 1.
Then, as proved by Kolchin [514], the following relationship holds between the
out-degrees of vertices in a random rooted tree and the number of offspring in the
Poisson process starting with a single particle.
Theorem 15.4.
+
P([Xr,t ] = X) = P([Zr,t ] = X|Z = n + 1).
1
P(Z1 = n + 1) = P(Y1 +Y2 + . . . +Yn+1 = n)
n+1
n+1 ri
1 λ −λ
= ∑ ∏ e
n + 1 r1 +...+rn+1 =n i=1 ri !
(n + 1)n λ n e−λ (n+1)
= .
(n + 1)!
Therefore
P([Zr,t ] = X|Z = n + 1) =
n st m0,t mn,t
∏t=0 m0,t ,...,mn,t p0 . . . pn
=
P(Z = n + 1)
mr,t
n st ! n λ r −λ
(n + 1)! ∏t=0 m0,t ! m1,t ! ... mn,t ! ∏r=0 r! e
=
(n + 1)n λ n e−λ (n+1)
(n + 1)! s1 ! s2 ! . . . sn ! n n 1
= n ∏ ∏ mr,t . (15.3)
(n + 1) t=0 r=0 mr,t ! (r!)
15.1. LABELED TREES 317
+
On the other hand, one can construct all rooted trees such that [Xr,t ] = X in the
following manner. We first layout an unlabelled tree in the plane. We choose
a single point (0, 0) for the root and then points St = {(i,t) : i = 1, 2, . . . , st } for
t = 1, 2, . . . , n. Then for each t, r we choose mr,t points of St that will be joined to
r points in St+1 . Then, for t = 0, 1, . . . , n − 1 we add edges. Note that Sn , if non-
empty, has a single point corresponding to a leaf. We go through St in increasing
order of the first component. Suppose that we have reached (i,t) and this has been
assigned out-degree r. Then we join (i,t) to the first r vertices of St+1 that have
not yet been joined by an edge to a point in St . Having put in these edges, we
Sn
assign labels 1, 2, . . . , n to t=1 St . The number of ways of doing this is
n
st !
∏ ∏n × n!.
t=1 r=1 mr,t !
+
The factor n! is an over count. As a set of edges, each tree with [Xr,t ] = X appears
n n m
exactly ∏t=0 ∏r=0 (r!) times, due to permutations of the trees below each ver-
r,t
tex. Summarising, the total number of tree with out-degrees given by the matrix
X is
n n
1
n! s1 ! s2 ! . . . sn ! ∏ ∏ mr,t ,
t=0 r=0 mr,t ! (r!)
which, after division by the total number of labeled trees on n + 1 vertices, i.e.,
by (n + 1)n−1 , results in an identical formula to that given for the random matrix
[X+r,t ] in the case of [Zr,t ], see (15.3). To complete the proof one has to notice that
for those matrices X which do not satisfy conditions (i) to (iii) both probabilities
in question are equal to zero.
Hence, roughly speaking, a random rooted labeled tree on n vertices has asymptot-
ically the same shape as a branching process with Poisson, parameter one in terms
of family sizes. Grimmett [399] uses this probabilistic representation to deduce
the asymptotic distribution of the distance from the root to the nearest pendant
vertex in a random labeled tree Tn , n ≥ 2. Denote this random variable by d(Tn ).
Theorem 15.5. As n → ∞,
( )
k−1
P(d(Tn ) ≥ k) → exp ∑ αi ,
i=1
Proof. Let k be a positive integer and consider the sub-tree of Tn induced by the
vertices at distance at most k from the root. Within any level (strata) of Tn , order
the vertices in increasing lexicographic order, and then delete all labels, excluding
that of the root. Denote the resulting tree by Tnk .
Now consider the following branching process constructed recursively according
to the following rules:
(i) Start with one particle (the unique member of generation zero).
(ii) For k ≥ 0, the (k + 1)th generation Ak+1 is the union of the families of de-
scendants of the kth generation together with one additional member which
is allocated at random to one of these families, each of the |Ak | families
having equal probability of being chosen for this allocation. As in Theorem
15.4, all family sizes are independent of each other and the past, and are
Poisson distributed with mean one.
Lemma 15.6. As n → ∞ the numerical characteristics of Tnk have the same dis-
tribution as the corresponding characteristics of the tree defined by the first k
generations of the branching process described above.
Proof. For a proof of Lemma 15.6, see the proof Theorem 3 of [399].
Let Yk be the size of the kth generation of our branching process and let Nk be the
number of members of the kth generation with no offspring. Let i = (i1 , i2 , . . . , ik )
be a sequence of positive integers, and let
A j = {N j = 0} and B j = {Y j = i j } for j = 1, 2, . . . , k.
P(d(Tn ) ≥ k) → P(A1 ∩ A2 ∩ . . . ∩ Ak ) .
Now,
k
P(A1 ∩ A2 ∩ . . . ∩ Ak ) = ∑ ∏ P(A j |A1 ∩ . . . ∩ A j−1 ∩ B1 ∩ . . . B j )
i j=1
× P(B j |A1 ∩ . . . ∩ A j−1 ∩ B1 ∩ . . . B j−1 ),
k i j −1
= ∑ ∏ 1 − e−1 C j (i j ), (15.4)
i j=1
Y j = 1 + Z + R1 + . . . + Ri j−1 −1 ,
where Z has the Poisson distribution and the Ri are independent random variables
with Poisson distribution conditioned on being non-zero. Hence
x i j−1 −1
x−1 e − 1
D j (x) = xe .
e−1
Now,
∞
−1 ik −1 Dk (1 − e−1 )
∑ (1 − e ) Ck (ik ) =
1 − e−1
.
ik =1
P(A1 ∩ A2 ∩ . . . ∩ Ak ) =
!ik−1 −1
k−1
i −1 eβ1 − 1
∑ ∏ β1 j C j (i j )eβ1 −1 , (15.5)
(i1 ,...,ik−1 ) j=1
e−1
P(A1 ∩ A2 ∩ . . . ∩ Ak ) =
!ik−2 −1
k−2
i −1 eβ2 − 1
∑ ∏ β1 j C j (i j )eβ1 +β2 −2 ,
(i1 ,...,ik−2 ) j=1
e−1
and αi = βi − 1. One can easily check that βi remains positive and decreases
monotonically as i → ∞, and so αi → −1.
320 CHAPTER 15. TREES
Another consequence of Lemma 15.3 is that, for a given N, one can associated
with the sequence Y1 ,Y2 , . . . ,YN , a generalized occupancy scheme of distributing
n particles into N cells (see [514]). In such scheme, the joint distribution of the
number of particles in each cell (ν1 , ν2 , . . . , νN ) is given, for r = 1, 2, . . . , N by
N !
P(νr = kr ) = P Yr = kr ∑ Yr = n . (15.6)
r=1
n +
Now, denote by Xr+ = ∑t=0 Xr,t the number of vertices of out-degree r in a random
n
tree on n + 1 vertices, rooted at a vertex labeled 0. Denote by Z (r) = ∑t=0 Zr,t , the
number of particles with exactly r offspring in the Poisson process µ(t). Then by
Theorem 15.4,
P(Xr+ = kr , r = 0, 1, . . . , n) = P(Z (r) = kr , r = 0, 1, . . . , n|Z1 = n + 1).
Hence by equation (15.1), the fact that we can choose λ = 1 in the process µ(t)
and (15.6), the joint distribution of out-degrees of a random tree coincides with the
joint distribution of the number of cells containing the given number of particles
in the classical model of distributing n particles into n + 1 cells, where each choice
of a cell by a particle is equally likely.
The above relationship, allows us to determine the asymptotic behavior of the
expectation of the number Xr of vertices of degree r in a random labeled tree Tn .
Corollary 15.7.
n
E Xr ≈ .
(r − 1)! e
Let Tn be a random recursive tree and let D+ n,i be the out-degree of the vertex with
label i, i.e the number of “children” of vertex i. We start with the exact probability
distribution of these random variables.
(i − 1)! n−i k
+
P(Dn,i = r) =
(n − 1)! k=r∑ r (i − 1)k−r |s(n − i, k)| (15.7)
Therefore
n−2 z
Φn,i (z) = Φn−1,i (z) + Φn−1,i (z).
n−1 n−1
With the boundary condition,
(i − 1)!
= (z + i − 1)(z + i) . . . (z + n − 2). (15.10)
(n − 1)!
Recall the definition of Stirling numbers of the first kind s(n, k). For non-negative
integers n and k
n
z(z − 1) . . . (z − n + 1) = ∑ s(n, k)zk .
k=1
Hence
(i − 1)! n−i
Φn,i (z) = ∑ |s(n − i, k)|(z + i − 1)k
(n − 1)! k=1
(i − 1)! n−i k k r
= ∑ ∑ r z (i − 1)k−r |s(n − i, k)|
(n − 1)! k=1 r=0
!
n−i
(i − 1)! n−i k
=∑
(n − 1)! ∑ r (i − 1)k−r |s(n − i, k)| zr .
r=0 k=r
n 1
E Ln = + . (15.11)
2 n−1
For a positive integer n, let ζn (s) = ∑nk=1 k−s be the incomplete Riemann zeta
function, and let Hn = ζ (1) = ∑nk=1 k−1 be the nth harmonic number, and let δn,k
denote the Kronecker function 1n=k .
while
Var Dn,i = Hn−1 − Hi−1 − ζn−1 (2) + ζi−1 (2).
15.2. RECURSIVE TREES 323
Proof. Let N j be the label of that vertex among vertices 1, 2, . . . j − 1 which is the
parent of vertex j. Then for j ≥ 1 and 1 ≤ i < j
n
Dn,i = ∑ δN j ,i . (15.12)
j=i+1
From the above theorem it follows that Var Dn,i ≤ E Dn,i . Moreover, for fixed i
and n large, E Dn,i ≈ log n, while for i growing with n the expectation E Dn,i ≈
log n − log i. The following theorem, see Kuba and Panholzer [535], shows a
standard limit behavior of the distribution of Dn,i .
Theorem 15.11. Let r ≥ 1 be fixed and let Xn,r be the number of vertices of degree
r in a random recursive tree Tn . Then, w.h.p.
Xn,r ≈ n/2r ,
and
Xn,r − n/2r d
√ → Yr ,
n
as n → ∞, where Yr has the N(0, σr2 ) distribution.
In place of proving the above theorem we will give a simple proof of its imme-
diate implication, i.e., the asymptotic behavior of the expectation of the random
variable Xn,r . The proof of asymptotic normality of suitably normalized Xn,r is
due to Janson and can be found in [441]. (In fact, in [441] a stronger statement
is proved, namely, that, asymptotically, for all r ≥ 1, random variables Xn,r are
jointly Normally distributed.)
Corollary 15.12. Let r ≥ 1 be fixed. Then
E Xn,r ≈ n/2r .
Proof. Let us introduce a random variable Yn,r counting the number of vertices of
degree at least r in Tn . Obviously,
Moreover, using a similar argument to that given for formula (15.7), we see that
for 2 ≤ r ≤ n,
n−2 1
E[Yn,r |Tn−1 ] = Yn−1,r + Yn−1,r−1 (15.15)
n−1 n−1
Notice, that the boundary condition for the recursive formula (15.15) is, trivially
given by
EYn,1 = n.
We will show, that EYn,r /n → 2−r+1 which, by (15.14), will imply the theorem.
Set
an,r := n2−r+1 − EYn,r . (15.16)
EYn,1 = n implies that an,1 = 0. We see from (15.11) that the expected number of
leaves in a random recursive tree on n vertices is given by
n 1
E Xn,1 = + .
2 n−1
15.2. RECURSIVE TREES 325
n−2 1
an,r = an−1,r + an−1,r−1 . (15.18)
n−1 n−1
Inductively assume that (15.17) holds for some n ≥ 3. Now, by (15.18), we get
n−2 1
an,r > an−1,r−1 + an−1,r−1 = an−1,r−1 .
n−1 n−1
Finally, notice that
2
an,n−1 = n22−n − ,
(n − 1)!
since there are only two recursive trees with n vertices and a vertex of degree
n − 1. So, we conclude that a(n, r) → 0 as n → ∞, for every r, and our theorem
follows.
Finally, consider the maximum degree ∆n = ∆n (Tn ) of a random recursive tree Tn .
It is easy to see that for large n, its expected value should exceed log n, since it
is as large as the expected degree of the vertex 1, which by Theorem 15.9 equals
Hn−1 ≈ log n. Szymański [731] proved that the upper bound is O(log2 n) (see Goh
and Schmutz [391] for a strengthening of his result). Finally, Devroye and Lu (see
[251]) have shown that in fact ∆n ≈ log2 n . This is somewhat surprising. While
each vertex in [1, n1−o(1) ] only has a small chance of having such a degree, there
are enough of these vertices to guarantee one w.h.p..
∆n ≈ log2 n.
The next theorem was originally proved by Devroye [246] and Pittel [652]. Both
proofs were based on an analysis of certain branching processes. The proof below
is related to [246].
Theorem 15.14. Let h(Tn ) be the height of a random recursive tree Tn . Then
w.h.p.
h(Tn ) ≈ e log n.
326 CHAPTER 15. TREES
Proof.
Upper Bound: For the upper bound we simply estimate the number ν1 of vertices
at height h1 = (1+ε)e log n where ε = o(1) but is sufficiently large so that claimed
inequalities are valid. Each vertex at this height can be associated with a path
i0 = 1, i1 , . . . , ih of length h in Tn . So, if S = {i1 , . . . , ih } refers to such a path, then
!h1
n
1 1 1
E ν1 = ∑ ∏ i − 1 ≤ h1! ∑ i ≤
|S|=h1 i∈S i=1
h1
(1 + log n)e
= o(1), (15.19)
h1
assuming that h1 ε →∞.
Explanation: If S = i1 = 1, i2 , . . . , ih1 then the term ∏hj=1
1
1/i j is the probability
that i j chooses i j−1 in the construction of Tn .
Lower Bound: The proof of the lower bound is more involved. We consider a
different model of tree construction and relate it to Tn . We consider a Yule process.
We run the process for a specific time t and construct a tree Y (t). We begin by
creating a single particle x1 at time 0 this will be the root of a tree Y (t). New
particles are generated at various times τ1 = 0, τ2 , . . . ,. Then at time τk there will
be k particles Xk = {x1 , x2 , . . . , xk } and we will have Y (t) = Y (τk ) for τk ≤ t < τk+1 .
After xk has been added to Y (τk ), each x ∈ Xk is associated with an exponential
random variable Ex with mean one1 . If zk is the particle in Xk that minimizes
Ex , x ∈ Xk then a new particle xk+1 is generated at time τk+1 = τk + Ezk and an
edge {zk , xk+1 } is added to Y (τk ) to create Y (τk+1 ). After this we independently
generate new random variables Ex , x ∈ Xk+1 .
Suppose then that we focus attention on Y (y; s,t), the sub-tree rooted at y contain-
ing all descendants of y that are generated after time s and before time t.
We observe three things:
(T1) The tree Y (τn ) has the same distribution as Tn . This is because each particle
in Xk is equally likely to be zk .
1 An exponential random variable Z with mean λ is characterised by P(Z ≥ x) = e−x/λ .
15.2. RECURSIVE TREES 327
(T2) If s < t and y ∈ Y (s) then Y (y; s,t) is distributed as Y (t − s). This follows
from Remark 15.15, because when zk ∈ / Y (y; s,t) it does not affect any of the
the variables Ex , x ∈ Y (y; s,t).
(T3) If x, y ∈ Y (s) then Y (x; s,t) and Y (y; s,t) are independent. This also follows
from Remark 15.15 for the same reasons as in (T2).
It is not difficult to prove (see Exercise (vii) or Feller [301]) that if Pn (t) is the
probability there are exactly n particles at time t then
Next let
t1 = (1 − ε) log n.
Then it follows from (15.20) that if ν(t) is the number of particles in our Yule
process at time t then
n−1
−t1 −t1 k−1 1
P(ν(t1 ) ≥ n) ≤ ∑ e (1 − e ) = 1 − 1−ε = o(1). (15.21)
k≥n n
We will show that w.h.p. the tree Tν(t1 ) has height at least
h0 = (1 − ε)et1
Suppose now that ν(s) ≥ ν0 and that the vertices of T1;0,s are
1/2
x1 , x2 , . . . , xν(s) . Let σ = ν0 and consider the sub-trees
A j , j = 1, 2, . . . , τ of T1;0,t1 rooted at x j , j = 1, 2, . . . , ν(s). We will show that
1
P(Tx1 ;s,t1 has height at least (1 − ε)3 e log n) ≥ . (15.23)
2σ log σ
Assuming that (15.23) holds, since the trees A1 , A2 , . . . , Aτ are independent, by T3,
we have
To prove all this we will associate a Galton-Watson branching process with each
of x1 , x2 , . . . , xτ . Consider for example x = x1 and let τ0 = log σ . The vertex x
will be the root of a branching process Π, which we now define. We will consider
the construction of Y (x; s,t) at times τi = s + iτ0 for i = 1, 2, . . . , i0 = (t1 − s)/τ0 .
The children of x in Π are the vertices at depth at least (1 − ε)eτ0 in Y (x; s, τ1 ). In
general, the particles in generation i will correspond to particles at depth at least
(1 − ε)eτ0 in the tree Y (ξ ; τi−1 , τi ) where ξ is a particle of Y (x; s,t) included in
generation i − 1 of Π.
If the process Π does not ultimately become extinct then generation i0 corresponds
to vertices in Y (t) that are at depth
1 m 1 1
D(h, m) = ∑ i−1 ∑ ∏
h i=2 j−1
S∈([2,m]\{i}
h−1 )
j∈S\{i}
m m
1 1 1 1 1 1
= ∑ ∏ j−1 ∑ ≥ ∑ ∏ j−1 ∑
h j∈S 16=k∈S
k−1 h j∈S k=h+1 k
S∈([2,m]
h−1 )
/ S∈([2,m]
h−1 )
15.2. RECURSIVE TREES 329
log m − log h − 1
≥ D(h − 1, m). (15.28)
h
Equation (15.27) follows by induction since D(1, m) ≥ log m.
Explanation of (15.28): We choose a path of length h by first choosing a vertex
i and then choosing S ⊆ [2, m] \ {i}. We divide by h because each h-set arises h
1
times in this way. Each choice will contribute ∏ j∈S∪{i} j−1 . We change the order
1 1
of summation i, S and then lower bound ∑m m
/ k−1 by ∑k=h+1 k .
16=k∈S
1 11
00
00
11 1 11
00
00
11 1 11
00
00
11
00
11
00
11 00
11
2 11
00
00
11
00
11
00
11
3 11
00
00
11 2 11
00
00
11 11
00
00
113 3 11
00
00
11
00
11
11
00
002
11
00
11
00
11 00
11
00
11 1 11
00 1 11
00
00
11
00
11
1 11
00 00
11
00
11
00
11
2 11
00
00
11
00
11
2 11
00
00
11
00
11
11
00
003
11
00
11
3 11
00
00
11
00
11
11
00
002
11
00
11
00
11
3 11
00
00
11
Assume now, as before that all the edges of a tree are directed toward the leaves,
and denote the out-degree of a vertex v by d + (v). Then the total number of exten-
sions of an plane-oriented recursive tree on n vertices is equal to
∑ (d +(v) + 1) = 2n − 1.
v∈V
So a new vertex can choose one those those 2n − 1 places to join the tree and
create a tree on n + 1 vertices. If we assume that this choice in each step is made
uniformly at random then a tree constructed this way is called a random plane-
oriented recursive tree. Notice that the probability that the vertex labeled n + 1
+ (v)+1
is attached to vertex v is equal to d 2n−1 i.e., it is proportional to the degree of
v. Such random trees, called plane-oriented because of the above geometric in-
terpretation, were introduced by Szymański [730] under the name of non-uniform
recursive trees. Earlier, Prodinger and Urbanek [664] described plane-oriented re-
cursive trees combinatorially, as labeled ordered (or plane) trees with the property
that labels along any path down from the root are increasing. Such trees are also
known in the literature as heap-ordered trees (see Chen and Ni [184], Prodinger
[663], Morris, Panholzer and Prodinger [614]) or, more recently, as scale-free
trees. So, random plane-oriented recursive trees are the simplest example of ran-
dom preferential attachment graphs.
Denote by an the number of plane-oriented recursive trees on n vertices. This
number, for n ≥ 2 satisfies an obvious recurrence relation
This is also the number of Stirling permutations, introduced by Gessel and Stanley
[381], i.e. the number of permutations of the multiset
{1, 1, 2, 2, 3, 3, . . . , n, n}, with the additional property that, for each value of 1 ≤
i ≤ n, the values lying between the two copies of i are greater than i.
There is a one-to-one correspondence between such permutations and plane-oriented
recursive trees, given by Koganov [509] and, independently, by Janson [443]. To
see this relationship consider a plane-oriented recursive tree on n + 1 vertices la-
belled 0, 1, 2, . . . , n, where the vertex with label 0 is the root of the tree and is
connected to the vertex labeled 1 only, and the edges of the tree are oriented in
the direction from the root. Now, perform a depth first search of the tree in which
we start from the root. Next we go to the leftmost child of the root, explore that
branch recursively, go to the next child in order etc., until we stop at the root. No-
tice that every edge in such a walk is traversed twice. If every edge of the tree gets
a label equal to the label of its end-vertex furthest from the root, then the depth
first search encodes each tree by a string of length 2n, where each label 1, 2, . . . , n
appears twice. So the unique code of each tree is a unique permutation of the
multiset {1, 1, 2, 2, 3, 3, . . . , n, n} with additional property described above. Note
also that the insertion of a pair (n + 1, n + 1) into one of the 2n − 1 gaps between
labels of the permutation of this multiset, corresponds to the insertion of the vertex
labeled n + 1 into a plane-oriented recursive tree on n vertices.
Let us start with exact formulas for probability distribution of the out-degree D+ n,i
of a vertex with label i, i = 1, 2, . . . , n in a random plane-oriented recursive tree.
Kuba and Panholzer [535] proved the following theorem.
Theorem 15.16. For i = 1, 2, . . . , n and r = 1, 2, . . . , n − 1,
r
+ r Γ(n − 3/2)Γ(i − 1/2)
P(Dn,i = r) = ∑ (−1)k ,
k=0 k Γ(i − 1 − k/2)Γ(n − 1/2)
R ∞ z−1 −t
where Γ(z) = 0 t e dt is the Gamma function. Moreover,
2i−2 n−i
i−1 4
E(D+
n,i ) = 2n−2
−1 (15.29)
n−1
For simplicity, we show below that the formula (15.29) holds for i = 1, i.e., the
expected value of the out-degree of the root of a random plane-oriented recursive
tree, and investigate its behavior as n → ∞. It is then interesting to compare the
15.3. INHOMOGENEOUS RECURSIVE TREES 333
latter with the asymptotic behavior of the degree of the root of a random recursive
tree. Recall that for large n this is roughly log n (see Theorem 15.10).
The result below was proved by Mahmoud, Smythe and Szymański [574].
Corollary 15.17. For n ≥ 2 the expected value of the degree of the root of a
random plane-oriented recursive tree is
4n−1
E(D+
n,1 ) = 2n−2
− 1,
n−1
and, √
E(D+
n,1 ) ≈ πn.
Proof. Denote by
n
4n 2i (2n)!!
un = 2n = ∏ 2i − 1 = (2n − 1)!! .
n i=1
P(D+
n+1,1 = r) =
r+1 r
1− P(D+
n,1 = r) + P(D+
n,1 = r − 1).
2n − 1 2n − 1
Hence
E(D+n+1,1 )
n
2n − r − 2 + r +
= ∑r P(Dn,1 = r) + P(Dn,1 = r − 1) =
r=1 2n − 1 2n − 1
!
n−1 n−1
1
2n − 1 ∑ r(2n − r − 2) P(D+n,1 = r) + ∑ (r + 1)2 P(D+n,1 = r)
r=1 r=1
n
1
= ∑ (2nr + 1) P(D+n,1 = r).
2n − 1 r=1
334 CHAPTER 15. TREES
The next theorem, due to Kuba and Panholzer [535], summarizes the asymptotic
behavior of the suitably normalized random variable D+
n,i .
(i) i = 1, then
d 2 /2
n−1/2 D+
n,1 → D1 , with density f D1 (x) = (x/2)e
−x
,
√
i.e., is asymptotically Rayleigh distributed with parameter σ = 2,
d
(ii) i ≥ 2, then n−1/2 D+
n,i → Di , with density
2i − 3
Z ∞
2 /4
fDi (x) = 2i−1 (t − x)2i−4 e−t dt.
2 (i − 2)! x
Let i = i(n) → ∞ as n → ∞. If
We now turn our attention to the number of vertices of a given out-degree. The
next theorem shows a characteristic feature of random
graphs built by preferential attachment rule where every new vertex prefers to
attach to a vertex with high degree (rich get richer rule). The proportion of vertices
with degree r in such a random graph with n vertices grows like n/rα , for some
15.3. INHOMOGENEOUS RECURSIVE TREES 335
constant α > 0, i.e., its distribution obeys a so called power law. The next result
was proved by Szymański [730] (see also [574] and [732]) and it indicates such a
behavior for the degrees of the vertices of a random plane-oriented recursive tree,
where α = 3.
Theorem 15.19. Let r be fixed and denote by Xn,r+ the number of vertices of out-
+ + r+1 + r +
E(Xn+1,r |Tn ) = Xn,r − Xn,r + Xn,r−1 + 1r=0 , (15.30)
2n − 1 2n − 1
which gives
+ 2n − r − 2 + r +
E Xn+1,r = E Xn,r + E Xn,r−1 + 1r=0 (15.31)
2n − 1 2n − 1
+
for r ≥ 1, (Xn,−1 = 0).
We will show that the difference
de f + 4n
an,r = E Xn,r − .
(r + 1)(r + 2)(r + 3)
is asymptotically negligible with respect to the leading term in the statement of
the theorem. Substitute an,r in the equation (15.31) to get that for r ≥ 1,
2n − r − 2 r 1
an+1,r = an,r + an,r−1 − . (15.32)
2n − 1 2n − 1 2n − 1
2
We want to show that |an,r | ≤ max{r,1} , for all n ≥ 1, r ≥ 0. Note that this is true
for all n and r = 0, 1, since from (15.31) it follows (inductively) that for n ≥ 2
+ 2n − 1 1
E Xn,0 = and so an,0 = − .
3 3
For n ≥ 2,
+ n 1 3 1 3
E Xn,1 = − + and so an,1 = − + .
6 12 4(2n − 3) 12 4(2n − 3)
We proceed by induction on r. By definition
4r
ar,r = − ,
(r + 1)(r + 2)(r + 3)
336 CHAPTER 15. TREES
and so,
2
|ar,r | < .
r
We then see from (15.32) that for and r ≥ 2 and n ≥ r that
2n − r − 2 2 r 2 1
|an+1,r | ≤ · + · − .
2n − 1 r 2n − 1 r − 1 2n − 1
r2
2 2 r
= − r+1− −
r (2n − 1)r r−1 2
2
≤ ,
r
which completes the induction and the proof of the theorem.
In fact much more can be proved.
|Yt+1 −Yt | ≤ 2.
15.3. INHOMOGENEOUS RECURSIVE TREES 337
For a proof of this, see the proof of Theorem 18.3. Applying the
Hoeffding- Azuma inequality (see Theorem 22.16) we get, for any fixed r,
+ +
| ≥ n log n) ≤ e−(1/8) log n = o(1).
p
P(|Xn,r − E Xn,r
+
√
But Theorem 15.19 shows that for any fixed r, E Xn,r n log n and (15.33)
follows.
Similarly, as for uniform random recursive trees, Pittel [652] established the asymp-
totic behavior of the height of a random plane-oriented recursive tree.
Theorem 15.21. Let h∗n be the height of a random plane-oriented recursive tree.
Then w.h.p.
log n
h∗n ≈ ,
2γ
where γ is the unique solution of the equation
γeγ+1 = 1,
1
i.e., γ = 0.27846.., so 2γ = 1.79556....
Suppose that Tn has been constructed for some n ≥ 1. Given Tn we add an edge
connecting one of its vertices with a new vertex labeled n + 1 and thus forming a
tree Tn+1 . A vertex vn ∈ {0, 1, 2, . . . , n} is chosen to be a neighbor of the incoming
vertex with probability
Obviously the model with such probability distribution is only a small generalisa-
tion of plane-oriented random recursive trees and we obtain the latter when we put
β = 0 in (15.35). Inhomogeneous random recursive trees of this type are known
in the literature as either scale free random trees or Barabási-Albert random trees.
For obvious reasons, we will call such graphs generalized random plane-oriented
recursive trees.
Let us focus the attention on the asymptotic behavior of the maximum degree of
such random trees. We start with some useful notation and observations.
Let Xn, j denote the weight of vertex j in a generalized plane-oriented random
recursive tree, with initial values X1,0 = X j, j = 1 + β for j > 0. Let
Γ n + β β+2
cn,k = , n ≥ 1, k ≥ 0,
β +k
Γ n + β +2
Lemma 15.22. Let Fn be the σ -field generated by the first n steps. If n ≥
max{1, j}, then Xn, j;k , Fn is a martingale.
Proof. Because Xn+1, j − Xn, j ∈ {0, 1}, we see that
Xn+1, j + k − 1
k
Xn, j + k − 1 Xn, j + k − 1 Xn+1, j − Xn, j
= +
k k−1 1
k(Xn+1, j − Xn, j )
Xn, j + k − 1
= 1+ .
k Xn, j
Hence, noting that
Xn, j
P(Xn+1, j − Xn, j = 1|Fn ) = ,
wn
and applying (15.37)
cn+1,k k
E(Xn+1, j;k |Fn ) = Xn, j;k 1+ = Xn, j;k ,
cn,k wn
we arrive at the lemma.
Thus, the random variable Xn, j;k , as a non-negative martingale, is
bounded in L1 and it almost surely converges to X jk /k!, where X j is the limit of
Xn, j;1 . Since Xn, j;k ≤ cXn, j;2k , where the constant c does not depend on n, it is also
bounded in L2 , which implies that it converges in L1 . Therefore we can determine
all moments of the random variable X j . Namely, for j ≥ 1,
X jk
β +k
= lim E Xn, j;k = X j, j;k = c j,k . (15.38)
k! n→∞ k
Let ∆n be the maximum degree in a generalized random plane-oriented recursive
tree Tn and let, for j ≤ n,
Note that since Xn,i is the weight of vertex i, i.e., its degree plus β , we find that
∆n,n = cn,1 (∆n + β ). Define
Now we are ready to prove the following result, due to Móri [613].
340 CHAPTER 15. TREES
Theorem 15.23.
−1/(β +2)
P lim n ∆n = ξ = 1.
n→∞
The limiting random variable ξ is almost surely finite and positive and it has
an absolutely continuous distribution. The convergence also holds in L p , for all
p, 1 ≤ p < ∞.
Proof. In the proof we skip the part dealing with the positivity of ξ and the abso-
lute continuity of its distribution.
By Lemma 15.22, ∆n,n is the maximum of martingales, therefore
(∆n,n |F ) is a non-negative sub-martingale, and so
n ∞
k k k β +k ∞
E ∆n,n ≤ ∑ E Xn, j;1 ≤ ∑ E X j = k! ∑ c j,k < ∞,
j=0 j=0 k j=0
Take the limit as n → ∞ of both sides of the above inequality. Applying (15.39)
and (15.38), we get
∞ ∞
−1/(β +2)
k
k β +k
E lim n ∆n − ξ j ≤ ∑ E ξi = k! ∑ c j,k .
n→∞
i= j+1 k i= j+1
15.4 Exercises
(i) Use the Prüfer code to show that there is one-to-one correspondence be-
tween the family of all labeled trees with vertex set [n] and the family of all
ordered sequences of length n − 2 consisting of elements of [n].
(iii) Let ∆ be the maximum degree of a random labeled tree on n vertices. Use
(15.1) to show that for every ε > 0, P(∆ > (1 + ε) log n/ log log n) tends to
0 as n → ∞.
(iv) Let ∆ be defined as in the previous exercise and let t(n, k) be the number
of labeled trees on n vertices
with maximum degreen at most k. Knowing
1 1
that t(n, k) < (n − 2)! 1 + 1 + 2! + . . . + (k−1)! , show that for every ε >
0, P(∆ < (1 − ε) log n/ log log n) tends to 0 as n → ∞.
(vi) Let Ln denote the number of leaves of a random recursive tree with n ver-
tices. Show that E Ln = n/2 and Var Ln = n/12.
(viii) Show that Φn,i (z) given in Theorem 15.8 is the probability generating func-
tion of the convolution of n−i independent Bernoulli random variables with
success probabilities equal to 1/(i + k − 1) for k = 1, 2, . . . , n − i.
(ix) Let Ln∗ denotes the number of leaves of a random plane-oriented recursive
tree with n vertices. Show that
2n − 1 2n(n − 2)
E Ln∗ = and Var Ln∗ = .
3 9(2n − 3)
15.5 Notes
Labeled trees
The literature on random labeled trees and their generalizations is very extensive.
For a comprehensive list of publications in this broad area we refer the reader to
a recent book of Drmota [261], to a chapter of Bollobás’s book [135] on random
graphs, as well as to the book by Kolchin [516]. For a review of some classical
results, including the most important contributions, forming the foundation of the
research on random trees, mainly due to Meir and Moon (see, for example : [594],
[595]and [597]), one may also consult a survey by Karoński [483].
342 CHAPTER 15. TREES
Recursive trees
Recursive trees have been introduced as probability models for system genera-
tion (Na and Rapoport [621]), spread of infection (Meir and Moon [596]), pyra-
mid schemes (Gastwirth [375]) and stemma construction in philology (Najock
and Heyde [625]). Most likely, the first place that such trees were introduced
in the literature, is the paper by Tapia and Myers [735], presented there under
the name “concave node-weighted trees”. Systematic studies of random recur-
sive trees were initiated by Meir and Moon ([596] and [611]) who investigated
distances between vertices as well as the process of cutting down such random
trees. Observe that there is a bijection between families of recursive trees and
binary search trees, and this has opened many interesting directions of research,
as shown in a survey by Mahmoud and Smythe [573] and the book by Mahmoud
[571].
Early papers on random recursive trees (see, for example, [621], [375] and [260])
were focused on the distribution of the degree of a given vertex and of the number
of vertices of a given degree. Later, these studies were extended to the distribution
of the number of vertices at each level, which is referred to as the profile. Recall,
that in a rooted tree, a level (strata) consists of all those vertices that are at the
same distance from the root.
The profile of a random recursive tree is analysed in many papers. For example,
Drmota and Hwang [262] derive asymptotic approximations to the correlation
coefficients of two level sizes in random recursive trees and binary search trees.
These coefficients undergo sharp sign-changes when one level is fixed and the
other is varying. They also propose a new means of deriving an asymptotic esti-
mate for the expected width, which is the number of nodes at the most abundant
level.
Devroye and Hwang [249] propose a new, direct, correlation-free approach based
on central moments of profiles to the asymptotics of width in a class of random
trees of logarithmic height. This class includes random recursive trees.
Fuchs, Hwang, Neininger [369] prove convergence in distribution for the profile,
normalized by its mean, of random recursive trees when the limit ratio α of the
level and the logarithm of tree size lies in [0, e). Convergence of all moments is
shown to hold only for α ∈ (0, 1) (with only convergence of finite moments when
α ∈ (1, e)).
van der Hofstadt, Hooghiemstra and Van Mieghem [426] study the covariance
structure of the number of nodes k and l steps away from the root in random
recursive trees and give an analytic expression valid for all k, l and tree sizes n.
For an arbitrary positive integer i ≤ in ≤ n − 1, a function of n, Su, Liu and Feng
[728] demonstrate the distance between nodes i and n in random recursive trees
Tn , is asymptotically normal as n → ∞ by using the classical limit theory method.
15.5. NOTES 343
Holmgren and Janson [428] proved limit theorems for the sums of functions of
sub-trees of binary search trees and random recursive trees. In particular, they
give new simple proofs of the fact that the number of fringe trees of size k = kn in
a binary search tree and the random recursive tree (of total size n) asymptotically
has a Poisson distribution
√ if k → ∞, and that the distribution is asymptotically
normal for k = o( n). Recall that a fringe tree is a sub-tree consisting of some
vertex of a tree and all its descendants (see Aldous [15]). For other results on that
topic see Devroye and Janson [250].
Feng, Mahmoud and Panholzer [302] study the variety of sub-trees lying on the
fringe of recursive trees and binary search trees by analysing the distributional
behavior of Xn,k , which counts the number of sub-trees of size k in a random tree
of size n, with k = k(n). Using analytic methods, they characterise for both tree
families the phase change behavior of Xn,k .
One should also notice interesting applications of random recursive trees. For
example, Mehrabian [593] presents a new technique for proving logarithmic up-
per bounds for diameters of evolving random graph models, which is based on
defining a coupling between random graphs and variants of random recursive
trees. Goldschmidt and Martin [392] describe a representation of the Bolthausen-
Sznitman coalescent in terms of the cutting of random recursive trees.
Bergeron, Flajolet, Salvy [84] have defined and studied a wide class of random
increasing trees. A a tree with vertices labeled {1, 2, . . . , n} is increasing if the
sequence of labels along any branch starting at the root is increasing. Obviously,
recursive trees and binary search trees (as well as the general class of inhomoge-
neous trees, including plane-oriented trees) are increasing. Such a general model,
which has been intensively studied, yields many important results for random trees
discussed in this chapter. Here we will restrict ourselves to pointing out just a few
papers dealing with random increasing trees authored by Dobrow and Smythe
[259], Kuba and Panholzer [535] and Panholzer and Prodinger [640], as well as
with their generalisations, i.e., random increasing k-trees, published by Zhang,
Rong, and Comellas [767], Panholzer and Seitz [641] and Darrasse, Hwang and
Soria [240],
gated the exact and limiting distributions of the size and the number of leaves in
the branches of the tree (see [443] for a follow up). Lu and Feng [552] considered
the strong convergence of the number of vertices of given degree as well as of the
degree of a fixed vertex (see also [573]). In Janson’s [441] paper, the distribution
of vertex degrees in random recursive trees and random plane recursive trees are
shown to be asymptotically normal. Brightwell and Luczak [163] investigate the
number Dn,k of vertices of each degree k at each time n, focusing particularly on
the case where k = k(n) is a growing function of n. They show that Dn,k is con-
centrated around its mean, which is approximately 4n/k3 , for all k ≤ (n log n)−1/3 ,
which is best possible up to a logarithmic factor.
Hwang [430] derives several limit results for the profile of random plane-oriented
recursive trees. These include the limit distribution of the normalized profile,
asymptotic bimodality of the variance, asymptotic approximation to the expected
width and the correlation coefficients of two level sizes.
Fuchs [368] outlines how to derive limit theorems for the number of sub-trees of
size k on the fringe of random plane-oriented recursive trees.
Finally, Janson, Kuba and Panholzer [445] consider generalized Stirling permuta-
tions and relate them with certain families of generalized plane recursive trees.
trees that does not use branching random walk. For further generalisations of
random recursive trees see Mahmoud [572].
346 CHAPTER 15. TREES
Chapter 16
Mappings
In the evolution of the random graph Gn,p , during its sub-critical phase, tree com-
ponents and components with exactly one cycle, i.e. graphs with the same number
of vertices and edges, are w.h.p. the only elements of its structure. Similarly, they
are the only graphs outside the giant component after the phase transition, until
the random graph becomes connected w.h.p. In the previous chapter we studied
the properties of random trees. Now we focus our attention on random mappings
of a finite set into itself. Such mappings can be represented as digraphs with the
same number of vertices and edges. So the study of their “average” properties
may help us to better understand the typical structure of classical random graphs.
We start the chapter with a short look at the basic properties of random permu-
tations (one-to-one mappings) and then continue to the general theory of random
mappings.
16.1 Permutations
Let f be chosen uniformly at random from the set of all n! permutations on the set
[n], i.e., from the set of all one-to-one functions [n] → [n]. In this section we will
concentrate our attention on the properties of a functional digraph representing a
random permutation.
Let D f be the functional digraph ([n], (i, f (i))). The digraph D f consists of vertex
disjoint cycles of any length 1, 2, . . . , n. Loops represent fixed points, see Figure
16.1.
Let Xn,t be the number of cycles of length t, t = 1, 2, . . . , n in the digraph D f . Thus
Xn,1 counts the number of fixed points of a random permutation. One can easily
check that
1 bn/tc (−1)i e−1/t
P(Xn,t = k) = k ∑ i → k as n → ∞, (16.1)
k!t i=0 t i! t k!
347
348 CHAPTER 16. MAPPINGS
x 1 2 3 4 5 6 7 8 9 10
f(x) 3 10 1 4 2 6 8 9 7 5
10
11
00
711
00
8
00
11
11
00
00
11
2 11
00 00
11 11
00 00
11
00
11 00
11 00
11 00
11
11
00
00
11
00
11
4
00
11 11
00
00
11
11
00
00
11 00
11
5 9
11
00 11
00 11
00
00
11
00
11 00
11 00
11
1 11
00 00
11
3 6
|s(n, k)|
P(Xn = k) = ,
n!
where the s(n, k) are Stirling numbers of the first kind, i.e., numbers satisfying the
following relation:
n
x(x − 1) · · · (x − n + 1) = ∑ s(n, k)xk .
k=0
Moreover,
n n
1 1
E Xn = Hn = ∑ , Var Xn = Hn − ∑ .
j=1 j j=1 j2
x(x + 1) · · · (x + n − 1)
Gn (x) = ,
n!
and the first part of the theorem follows. Note that
x+n−1 Γ(x + n)
Gn (x) = = ,
n Γ(x)Γ(n + 1)
The results for the expectation and variance of Xn can be obtained by calculating
the first two derivatives of Gn (x) and evaluating them at x = 1 in a standard way but
350 CHAPTER 16. MAPPINGS
one can also show them using only the fact that the cycles of functional digraphs
must be disjoint. Notice, for example, that
Similarly one can derive the second factorial moment of Xn counting ordered pairs
of cycles (see Exercises 16.3.2 and 16.3.3) which implies the formula for the vari-
ance.
Goncharov [395] proved a Central Limit Theorem for the number Xn of cycles.
Theorem 16.2.
Zx
Xn − log n 2
lim P √ ≤x = e−t /2 dt,
n→∞ log n −∞
Theorem 16.3.
E Ln 1 −y
Z ∞ Z ∞
lim = exp −x − e dy dx = 0.62432965....
n→∞ n 0 x y
16.2 Mappings
Let f be chosen uniformly at random from the set of all nn mappings from [n] →
[n]. Let D f be the functional digraph ([n], (i, f (i))) and let G f be the graph ob-
tained from D f by ignoring orientation. In general, D f has unicyclic components
only, where each component consists of a directed cycle C with trees rooted at
vertices of C, see the Figure 16.2.
Therefore the study of functional digraphs is based on results for permutations of
the set of cyclical vertices (these lying on cycles) and results for forests consisting
of trees rooted at these cyclical vertices (we allow also trivial one vertex trees).
For example, to show our first result on the connectivity of G f we will need the
following enumerative result for the forests.
16.2. MAPPINGS 351
x 1 2 3 4 5 6 7 8 9 10
f(x) 3 10 5 4 2 5 8 9 7 5
10
11
00
711
00
8
00
11
11
00
00
11
2 11
00 00
11 11
00 00
11
00
11 00
11 00
11 00
11
11
00
00
11
00
11
4
00
11 11
00
00
11
1111
000000
11 00
11
000011
1111
0000
1111
00
00
11
00
11 5 9
0000
1111 00
11
0000 11
1111
0000
1111
00
00
11
000
111
0000
1111 00
11
00
11
000
111
000
111 00
11
00
11
0000
1111 3 00
11
0000
1111
0000
1111
0000
1111
00
11 6
0000
1111
0000
1111
0000
1111
00
11
00
11
00
11
1
Lemma 16.4. Let T (n, k) denote the number of forests with vertex set [n], consist-
ing of k trees rooted at the vertices 1, 2, . . . , k. Then,
T (n, k) = knn−k−1 .
Theorem 16.5.
1 n (n)k
r
π
P(G f is connected ) = ∑ k ≈ .
n k=1 n 2n
Proof. If G f is connected then there is a cycle with k vertices say such that after
removing the cycle we have a forest consisting of k trees rooted at the vertices of
352 CHAPTER 16. MAPPINGS
If k ≥ n3/5 , then
k(k − 1) 1 1/6
uk ≤ exp − ≤ exp − n ,
2n 3
1 k−1
n −k j uk
E Zk = (k − 1)! n = ∏ 1 − = .
k k j=0 n k
If Z = Z1 + Z2 + · · · + Zn , then
n
uk 1 −x2 /2n 1 ∞ 1 −s/2n 1
Z ∞ Z
EZ = ∑ ≈ e dx = e ds ≈ log n.
k=1 k x=1 x 2 s=1 s 2
16.2. MAPPINGS 353
1 R ∞ 1 −s/2n
(To estimate the integral
R n/ω
2 s=1 s e
R ωn
we break I1+ I2 + I3 where I1 =
it into
ωn
· · · ds ≈ log n, ω =
log n, I2 = s=n/ω · · · ds ≤ log n/ω = o(log n) and I3 =
Rs=1
∞
s=ωn · · · ds = o(1).)
Moreover the expected number of vertices of cycles in a random mapping is equal
to
!
n n
r
πn
E ∑ kZk = ∑ uk ≈ .
k=1 k=1 2
P fˆ(i) = j = p j , j = 1, 2, . . . , n,
(16.2)
and
p1 + p2 + . . . + pn = 1.
This model was introduced (in a slightly more general form) independently by
Burtin [174] and Ross [683]. We will first prove a generalisation of Theorem
16.5.
Theorem 16.6.
P(G fˆ is connected ) =
!
= ∑ p2i 1 + ∑ p j + ∑ ∑ p j pk + ∑ ∑ ∑ p j pk pl + · · · .
i j6=i j6=i k6=i, j j6=i k6=i, j l6=i, j,k
To prove this theorem we use the powerful “Burtin–Ross Lemma”. The short
and elegant proof of this lemma given here is due to Jaworski [456] (His general
approach can be applied to study other characteristics of a random mappings, not
only their connectedness).
= ∑ pk ,
k∈U
Proof (of Theorem 16.6). Notice that G fˆ is connected if and only if there is a
subset U ⊆ [n] such that U spans a single cycle while there is no cycle on [n] \U.
Moreover, the events “U ⊆ [n] spans a cycle” and “there is no cycle on [n] \U” are
independent. Hence, by Lemma 16.7,
Pr(G fˆ is connected) =
= ∑ P(U ⊂ [n] spans a cycle) P(there is no cycle on [n] \U)
06/ =U⊆[n]
= ∑ (|U| − 1)! ∏ p j ∑ pk (16.3)
06/ =U⊂[n] j∈U k∈U
!
= ∑ p2i 1 + ∑ p j + ∑ ∑ p j pk + ∑ ∑ ∑ p j pk pl + · · · .
i j6=i j6=i k6=i, j j6=i k6=i, j l6=i, j,k
Using the same reasoning as in the above proof, one can show the following result
due to Jaworski [456].
Theorem 16.8. Let X be the number of components in G fˆ and Y be the number
of its cyclic vertices (vertices belonging to a cycle). Then for k = 1, 2, . . . , n,
where s(·, ·) is the Stirling number of the first kind. On the other hand,
P(Y = k) = k! ∑ ∏ p j − (k + 1)! ∑ ∏ p j .
U⊂[n] j∈U U⊂[n] j∈U
|U|=k |U|=k+1
Lemma 16.9 (Burtin-Ross Lemma - the second version). Let ĝ : [n] → [n] ∪ {0}
be a random mapping from the set [n] to the set [n] ∪ {0}, where, independently
for all i ∈ [n],
P (ĝ(i) = j) = q j , j = 0, 1, 2, . . . , n,
and
q0 + q1 + q2 + . . . + qn = 1.
Let Dĝ be the random directed graph on the vertex set [n] ∪ {0}, generated by the
mapping ĝ and let Gĝ denote its underlying simple graph. Then
P(Gĝ is connected ) = q0 .
356 CHAPTER 16. MAPPINGS
Notice that the event that Gĝ is connected is equivalent to the event that Dĝ is a
(directed) tree, rooted at vertex {0}, i.e., there are no cycles in Gĝ [[n]].
We will use this result and Lemma 16.9 to prove the next theorem (for more
general results, see [457]).
Theorem 16.10. Let D fˆ be the functional digraph of a mapping fˆ defined in
(16.2) and let ZR be the number of predecessors of a set R ⊂ [n], |R| = r, r ≥ 1, of
vertices of D fˆ , i.e.,
Proof. The distribution of ZR follows immediately from the next observation and
the application of Lemma 16.9. Denote by A the event that there is a forest
spanned on the set W = U ∪ R, where U ⊂ [n] \ R, composed of r (directed) trees
rooted at vertices of R. Then
where B is the event that all edges that begin in U end in W , while C denotes the
event that all edges that begin in [n] \W end in [n] \W . Now notice that
Furthermore,
P(A |B ∩C) = P(Gĝ is connected ),
where ĝ : U → U ∪ {0}, where {0} stands for the set R collapsed to a single vertex,
is such that for all u ∈ U independently,
pj ΣR
q j = P(ĝ(u) = j) = , for j ∈ U, while q0 = .
ΣW ΣW
We will finish this section by stating the central limit theorem for the number
of components of G f , where f is a uniform random mapping f : [n] → [n] (see
Stepanov [726]). It is an analogous result to Theorem 16.2 for random permuta-
tions.
Theorem 16.11.
1
Xn −
Z x
2 log n 2 /2
lim P q ≤ x = e−t dt,
n→∞ 1 −∞
2 log n
16.3 Exercises
16.3.1 Prove directly that if Xn,t is the number of cycles of length t in a random
permutation then E Xn,t = 1/t.
16.3.2 Find the expectation and the variance of the number Xn of cycles in a ran-
dom permutation using fact that the rth derivative of the gamma function
dr
equals (dx)r Γ(x) = 0 (logt)r t x−1 e−t dt,
R∞
16.3.4 Find the probability distribution for the length of a typical cycle in a ran-
dom permutation, i.e., the cycle that contains a given vertex (say vertex 1).
Determine the expectation and variance of this characteristic.
16.3.6 Show that the length of the cycle containing item 1 in a random permuta-
tion is uniformly distributed in [n].
16.3.7 Show that if X denotes the number of cycles in a random permutation of [n]
then P(X ≥ t) ≤ P(Bin(t, 1/2) ≤ dlog2 ne). Deduce that for every constant
K > 0, there exists a constant L > 0, such that P(X ≥ K log n) ≤ n−L .
358 CHAPTER 16. MAPPINGS
16.3.8 Now let X denote the number of cycles in the digraph D f of a random
mapping f : [n] → [n]. Show that for every constant K > 0, there exists a
constant L > 0, such that P(X ≥ K log n) ≤ n−L .
16.3.9 Determine the expectation and variance of the number of components in a
functional digraph D fˆ of a generalized random mapping fˆ (see Theorem
16.8)
16.3.10 Find the expectation and variance of the number of cyclic vertices in a
functional digraph D fˆ of a generalized random mapping fˆ (see Theorem
16.8).
16.3.11 Prove Theorem 16.8.
16.3.12 Show that Lemmas 16.7 and 16.9 are equivalent.
16.3.13 Prove the Burtin-Ross Lemma for a bipartite random mapping, i.e. a map-
ping with bipartition ([n], [m]), where each vertex i ∈ [n] chooses its unique
image in [m] independently with probability 1/m, and, similarly, each ver-
tex j ∈ [m] selects its image in [n] with probability 1/n.
16.3.14 Consider an evolutionary model of a random mapping (see [458],[459]),
i.e., a mapping fˆq [n] → [n], such that for i, j ∈ [n], P( fˆq (i) = j) = q if i = j
while, P( fˆq (i) = j) = (1 − q)/(n − 1) if i 6= j, where 0 ≤ q ≤ 1. Find the
probability that fˆq is connected.
16.3.15 Show that there is one-to-one correspondence between the family of nn
mappings f : [n] → [n] and the family of all doubly-rooted trees on the
vertex set [n] (Joyal bijection)
16.4 Notes
Permutations
Systematic studies of the properties of random permutations of n objects were ini-
tiated by Goncharov in [394] and [395]. Golomb [393] showed that the expected
length of the longest cycle of D f , divided by n is monotone decreasing and gave
a numerical value for the limit, while Shepp and Lloyd in [712] found the closed
form for this limit (see Theorem 16.3). They also gave the corresponding result
for kth moment of the rth longest cycle, for k, r = 1, 2, . . . and showed the limiting
distribution for the length of the rth longest cycle.
Kingman [503] and, independently, Vershik and Schmidt [745], proved that for a
random permutation of n objects, as n → ∞, the process giving the proportion of
16.4. NOTES 359
elements in the longest cycle, the second longest cycle, and so on, converges in
distribution to the Poisson-Dirichlet process with parameter 1 (for further results
in this direction see Arratia, Barbour and Tavaré [41]). Arratia and Tavaré [42]
provide explicit bounds on the total variation distance between the process which
counts the sizes of cycles in a random permutations and a process of independent
Poisson random variables.
For other results, not necessarily of a “graphical” nature, such as, for example, the
order of a random permutation, the number of derangements, or the number of
monotone sub-sequences, we refer the reader to the respective sections of books
by Feller [301], Bollobás [136] and Sachkov [697] or, in the case of monotone
sub-sequences, to a recent monograph by Romik [682].
Mappings
Uniform random mappings were introduced in the mid 1950’s by Rubin and Sit-
graves [684], Katz [497] and by Folkert [320]. More recently, much attention
has been focused on their usefulness as a model for epidemic processes, see for
example the papers of Gertsbakh [380], Ball, Mollison and Scalia-Tomba [55],
Berg [82], Mutafchiev [620], Pittel [649] and Jaworski [459]. The component
structure of a random functional digraph D f has been studied by Aldous [13].
He has shown, that the joint distribution of the normalized order statistics for the
component sizes of D f converges to the Poisson-Dirichlet distribution with pa-
rameter 1/2. For more results on uniform random mappings we refer the reader
to Kolchin’s monograph [515], or a chapter of Bollobás’ [136].
The general model of a random mapping fˆ, introduced by Burtin [174] and Ross
[683], has been intensively studied by many authors. The crucial Burtin-Ross
Lemma (see Lemmas: 16.7 and 16.9) has many alternative proofs (see [38]) but
the most useful seems to be the one used in this chapter, due to Jaworski [456].
His approach can also be applied to derive the distribution of many other char-
acteristics of a random digraph D f , as well as it can be used to prove generalisa-
tions of the Burtin-Ross Lemma for models of random mappings with independent
choices of images. (For an extensive review of results in that direction see [457]).
Aldous, Miermont, Pitman ([18],[19]) study the asymptotic structure of D fˆ using
an ingenious coding of the random mapping fˆ as a stochastic process on the in-
terval [0, 1] (see also the related work of Pitman [648], exploring the relationship
between random mappings and random forests).
Hansen and Jaworski (see [410], [411]) introduce a random mapping f D : [n] → [n]
with an in-degree sequence, which is a collection of exchangeable random vari-
ables (D1 , D2 , . . . , Dn ). In particular, they study predecessors and successors of a
given set of vertices, and apply their results to random mappings with preferential
360 CHAPTER 16. MAPPINGS
k-out
Several interesting graph properties require that the minimum degree of a graph
be at least a certain amount. E.g. having a Hamilton cycle requires that the mini-
mum degree is at least two. In Chapter 6 we saw that Gn,m being Hamiltonian and
having minimum degree at least two happen at the same time w.h.p. One is there-
fore interested in models of a random graph which guarantee a certain minimum
degree. We have already seen d-regular graphs in Chapter 11. In this chapter we
consider another simple and quite natural model Gk−out that generalises random
mappings. It seems to have first appeared in print as Problem 38 of “The Scottish
Book” [578]. We discuss the connectivity of this model and then matchings and
Hamilton cycles. We also consider a related model of “Nearest Neighbor Graphs”.
17.1 Connectivity
For an integer k, 1 ≤ k ≤ n − 1, let ~Gk−out be a random digraph on vertex set V =
{1, 2, . . . , n} with arcs (directed edges) generated independently for each v ∈ V
by a random choice of k distinct arcs (v, w), where w ∈ V \ {v}, so that each of
n−1
the k possible sets of arcs is equally likely to be chosen. Let Gk−out be the
random graph(multigraph) obtained from ~Gk−out by ignoring the orientation of its
arcs, but retaining all edges.
Note that ~G1−out is a functional digraph of a random mapping f : [n] → [n], with
a restriction that loops (fixed points) are not allowed. So for k = 1 the following
result holds.
Theorem 17.1.
lim P(G1−out is connected ) = 0.
n→∞
361
362 CHAPTER 17. K-OUT
The situation changes when each vertex is allowed to choose more than one neigh-
bor. Denote by κ(G) and λ (G) the vertex and edge connectivity of a graph G re-
spectively, i.e., the minimum number of vertices (respectively edges) the deletion
of which disconnects G. Let δ (G) be the minimum degree of G. The well known
Whitney’s Theorem states that, for any graph G,
In the next theorem we show that for random k − out graphs these parameters are
equal w.h.p. It is taken from Fenner and Frieze [309]. The Scottish Book [578]
contains a proof that Gk−out is connected for k ≥ 2.
Proof. In the light of Whitney’s Theorem, to prove our theorem we have to show
that the following two statements hold:
and
lim P(δ (Gk−out ) ≤ k) = 1. (17.2)
n→∞
Then, w.h.p.
k ≤ κ ≤ λ ≤ δ ≤ k,
and the theorem follows.
To prove statement (17.1) consider the deletion of r vertices from the random
graph Gk−out , where 1 ≤ r ≤ k − 1. If Gk−out can be disconnected by deleting r
vertices, then there exists a partition (R, S, T ) of the vertex set V , with | R |= r,
| S |= s and | T |= t = n − r − s, with k − r + 1 ≤ s ≤ n − k − 1, such that Gk−out
has no edge joining a vertex in S with a vertex in T . The probability of such an
event, for an arbitrary partition given above, is equal to
r+s−1 s n−s−1 n−r−s
! !
r + s sk n − s (n−r−s)k
k k
n−1 n−1
≤
k k
n n
Thus
b(n−r)/2c sk (n−r−s)k
n! r+s n−s
P(κ(Gk−out ) ≤ r) ≤ ∑
s=k−r+1 s!r!(n − r − s)! n n
17.1. CONNECTIVITY 363
where 1/2
n 2c
αs = α(s, n, r) ≤ c ≤ ,
s(n − r − s) s1/2
for some absolute constant c > 0.
Thus
b(n−r)/2c s (n−r−s)
1 r+s n−s
P(κ(Gk−out ) ≤ r) ≤ 2c ∑ us
s=k−r+1 s1/2 s n−r−s
where
us = (r + s)(k−1)s (n − s)(k−1)(n−r−s) nn−k(n−r) .
Now, s n−r−s
r+s n−s
≤ e2r ,
s n−r−s
and
(r + s)s (n − s)n−r−s
decreases monotonically, with increasing s, for s ≤ (n − r)/2. Furthermore, if
s ≤ n/4 then the decrease is by a factor of at least 2.
Therefore
!
n/4
2 n
P(κ(Gk−out ) ≤ r) ≤ 2ce2r ∑ 2−(k−1)(s−k+r−1) + n1/2 · 4 uk−r+1
s=k−r+1
where
a = (k + 1)(k−1)(k−r+1) .
It follows that
P(δ (Gk−out ) = k) → 1 as n → ∞.
n−2 n−1
! n−1
k
P(Ev ) = k
n−1
= 1− → e−k .
k
n − 1
Let Z denote the number of vertices of degree k in Gk−out . Then we have shown
that E(Z) ≈ ne−k . Now the random variable Z is determined by kn independent
random choices. Changing one of these choices can change the value of Z by at
most one. Applying the Azuma-Hoeffding concentration inequality – see Section
22.7, in particular Lemma 22.17 we see that for any t > 0
2t 2
P(Z ≤ E(Z) − t) ≤ exp − .
kn
Putting t = ne−k /2 we see that Z > 0 w.h.p. and the theorem follows.
Theorem 17.3.
(
0 if k = 1
lim P(Gk−out
n→∞
has a perfect matching) =
n even 1 if k ≥ 2.
We will only prove a weakening of the above result to where k ≥ 15. We follow
the ideas of Section 6.1. So, we begin by examining the expansion properties of
G = Ga−out , a ≥ 3.
1 1/(a−2)
Lemma 17.4. W.h.p. |NG (S)| ≥ |S| for all S ⊆ [n], |S| ≤ κa n where κa = 12 30
.
17.2. PERFECT MATCHINGS 365
Proof. The probability there exists a set S with insufficient expansion is at most
κa n as κa n as
n n 2s ne 2s 2s
∑ s s−1 n ≤ ∑ s n
s=3 s=3
κa n a−2 s
s 2 a
=∑ e 2 = o(1). (17.3)
s=3 n
Putting a = 8 gives b = 7 and a proof that G15−out , n even, has a perfect matching
w.h.p.
Bipartite graphs
We now consider the related problem of the existence of a perfect matching in a
random k-out bipartite graph.
Let U = {u1 , u2 , . . . , un },V = {v1 , v2 , . . . , vn } and let each vertex from U choose
independently and without repetition, k neighbors in V , and let each vertex from V
choose independently and without repetition k neighbors in U. Denote by ~Bk−out
the digraphs generated by the above procedure and let Bk−out be its underlying
simple bipartite graph.
366 CHAPTER 17. K-OUT
Theorem 17.6.
(
0 if k = 1
lim P(Bk−out has a perfect matching) =
n→∞ 1 if k ≥ 2.
We will give two different proofs. The first one - existential- of a combinatorial
nature is due to Walkup [750]. The second one - constructive- of an algorithmic
nature, is due to Karp, Rinnooy-Kan and Vohra [493]. We start with the combina-
torial approach.
Existence proof
Let X denote the number of perfect matchings in Bk−out . Then
The above bound follows from the following observations. There are n! ways
of pairing the vertices of U with the vertices of V . For each such pairing there
are 2n ways to assign directions for the connecting edges, and then each possible
matching has probability (k/n)n of appearing in Bk−out .
So, by Stirling’s formula,
Hence, for k = 2,
r 2r n − r 2(n−r)
Pr ≤ . (17.5)
n n
Then we use Stirling’s formula to show,
2
n n r n
=
r r−1 n−r+1 r
2(n−r)
r n n 2r n
≤ . (17.6)
n − r + 1 r(n − r) r n−r
To estimate Qr we have to consider condition (ii) which a minimal bad pair has to
satisfy. This implies that a vertex v ∈ S = N(R) is chosen by at least one vertex
from R (denote this event by Av ), or it chooses both its neighbors in R (denote
this event by Bv ). Then the events Av , v ∈ S are negatively correlated (see Section
22.2) and the events Bv , v ∈ S are independent of other events in this collection.
Let S = {v1 , v2 , . . . , vr−1 }. Then we can write
!
r−1
\
Qr ≤ P (Avi ∪ Bvi )
i=1
368 CHAPTER 17. K-OUT
i−1 !
r−1 \
= ∏ P Avi ∪ Bvi (Av j ∪ Bv j )
i=1 j=1
r−1
≤ ∏ P (Avi ∪ Bvi )
i=1
r−1
= 1 − P(Acv1 ) P(Bcv1 )
!!r−1
r
r − 2 2r
≤ 1− 1 − n2
r−1 2
≤ η r−1 (17.7)
for some absolute constant 0 < η < 1 when r ≤ (n + 1)/2.
Going back to (17.4), and using (17.5), (17.6), (17.7)
(n+1)/2 (n+1)/2
η r−1 n
∑ E Xr ≤ 2 ∑ = o(1).
r=2 r=2 (n − r)(n − r + 1)
Hence ∑r E Xr → 0 as n → ∞, which means that w.h.p. there are no bad pairs,
implying that Bk−out has a perfect matching w.h.p.
Frieze and Melsted [358] considered the related question. Suppose that M, N are
disjoint sets of size m, n and that each v ∈ M chooses d ≥ 3 neighbors in N. Sup-
pose that we condition on each vertex in N being chosen at least twice. They
show that w.h.p. there is a matching of size equal to min {m, n}. Fountoulakis and
Panagiotou [323] proved a slightly weaker result, in the same vein.
Algorithmic Proof
We will now give a rather elegant algorithmic proof of Theorem 17.6. It is due to
Karp, Rinnooy-Kan and Vohra [493]. We do this for two reasons. First, because
it is a lovely proof and second this proof is the basis of the proof that 2-in,2-out
is Hamiltonian in [217]. In particular, this latter example shows that constructive
proofs can sometimes be used to achieve results not obtainable through existence
proofs alone.
Start with the random digraph ~B2−out and consider two multigraphs, GU and GV
with labeled vertices and edges, generated by ~B2−out on the sets of the bipartition
(U,V ) in the following way. The vertex set of the graph GU is U and two vertices,
u and u0 , are connected by an edge, labeled v, if a vertex v ∈ V chooses u and u0
as its two neighbors in U. Similarly, the graph GV has vertex set V and we put an
edge labeled u between two vertices v and v0 , if a vertex u ∈ U chooses v and v0
as its two neighbors in V . Hence graphs GU and GV are random multigraphs with
exactly n labeled vertices and n labeled edges.
17.2. PERFECT MATCHINGS 369
• Step 2. Add the edge {x, y} , x, y ∈ V that has label u to the graph HV .
• Step 3. Let Cx ,Cy be the components of HV just before the edge labeled
u is added. Let C = Cx ∪ Cy . If all vertices in C are checked, go to Step
6. Otherwise, select an unchecked vertex v in C. If possible, select an
unchecked vertex v for which the edge labeled v in HU belongs to CORE.
We next argue that Algorithm PAIR, when it finishes at Step 5, does indeed pro-
duce a perfect matching in B2−out . There are two simple invariants of this process
that explain this:
(I1) The number of marked vertices plus the number of edges in HU is equal to n.
For I1, we observe that each round marks one vertex and deletes one edge of HU .
Similarly, for I2, we observe that each round checks one vertex and adds one edge
to HV .
Lemma 17.7. Up until (possible) failure in Step 6, the components of HV are
either trees with a unique unchecked vertex or are unicyclic components with all
vertices checked. Also, failure in Step 6 means that PAIR tries to add an edge to a
unicyclic component.
370 CHAPTER 17. K-OUT
Proof. This is true initially, as initially HV has no edges and all vertices are
unchecked. Assume this to be the case when we add an edge {x, y} to HV . If
Cx 6= Cy are both trees then we will have a choice of two unchecked vertices in
C = Cx ∪ Cy and C will be a tree. After checking one vertex, our claim will still
hold. The other possibilities are that Cx is a tree and Cy is unicyclic. In this case
there is one unchecked vertex and this will be checked and C will be unicyclic. The
other possibility is that C = Cx = Cy is a tree. Again there is only one unchecked
vertex and adding {x, y} will make C unicyclic.
Lemma 17.8. If HU consists of trees and unicyclic components then all the trees
in HU contain a marked vertex.
Proof. Suppose that HU contains k trees with marked vertices and ` trees with no
marked vertices and that the rest of the components are unicyclic. It follows that
HU contains n − k − ` edges and then (I1) implies that ` = 0.
Lemma 17.9. If the algorithm stops in Step 5, then we can extract a perfect match-
ing from HU , HV .
Proof. Suppose that we arrive at Step 5 after k rounds. Suppose that there are k
trees with a marked vertex. Let the component sizes in HU be n1 , n2 , . . . , nk for the
trees and m1 , m2 , . . . , m` for the remaining components. Then,
n1 + n2 + · · · + nk + m1 + m2 + · · · + m` = |V (HU )| = n.
|E(HU )| = n − k,
from I1 and so
(n1 − 1) + (n2 − 1) + · · · + (nk − 1)+
(≥ m1 ) + (≥ m2 ) + (≥ m` ) = n − k.
It follows that the components of HU that are not trees with a marked vertex have
as many edges as vertices and so are unicyclic.
We now show, given that HU , HV only contain trees and unicyclic components,
that we can extract a perfect matching. The edges of HU define a matching of
B2−out of size n − k. Consider a tree T component with marked vertex ρ. Orient
the edges of T away from ρ. Now consider an edge {x, y} of T , oriented from x
to y. Suppose that this edge has label z ∈ V . We add the edge {y, z} to M1 . These
edges are disjoint: z appears as the label of exactly one edge and y is the head of
exactly one oriented edge.
For the unicyclic components, we orient the unique cycle
C = (u1 , u2 , . . . , us ) arbitrarily in one of two ways. We then consider the trees
attached to each of the ui and orient them away from the ui . An oriented edge
{x, y} with label z yields a matching edge {y, z} as before.
17.2. PERFECT MATCHINGS 371
The remaining k edges needed for a perfect matching come from HV . We extract a
set of k matching edges out of HV in the same way we extracted n − k edges from
HU . We only need to check that these k edges are disjoint from those chosen from
HU . Let {y, z} be such an edge, obtained from the edge {x, y} of HV , which has
label z. z is marked in HU and so is the root of a tree and does not appear in any
matching edge of M1 . y is a checked vertex and so the edge labelled y has been
deleted from HU and this prevents y appearing in a matching edge of M1 .
Lemma 17.10. W.h.p. Algorithm PAIR cannot reach Step 6 in fewer than 0.49n
iterations.
Proof. It follows from Lemma 2.10 that w.h.p. after ≤ 0.499n rounds, HV only
contains trees and unicyclic components. The lemma now follows from Lemma
17.7.
To complete our analysis, it only remains to show
Lemma 17.11. W.h.p., at most 0.49n rounds are needed to make HU the union of
trees and unicyclic components.
(log n)2
(0.64)k
n k−2 0.49n
E Z ≤ o(1) + ∑ k (k − 1)! k−1 ×
k=1 k k−1 n
2
!.49n−(k−1)
k(n − k)
× 1− n (17.8)
2
(log n)2 k−2
k
≤ (1 + o(1))n ∑ (0.64)k e−0.98k
k=1 k!
372 CHAPTER 17. K-OUT
≤ (1 + o(1))n×
" k #
(0.64θ )2 (0.64θ )3 2(0.64θ )4 (0.64)e.02
∞
0.64θ + + + +∑
2 2 3 k=5 2k5/2
where θ = e−0.98
1
≤ (1 + o(1))n 0.279 +
2 × 55/2 (1 − (0.64)e.02 )
≤ (1 + o(1))n [0.279 + 0.026]
≤ (0.305)n.
Theorem 17.12. W.h.p. the algorithm PAIR finds a perfect matching in the ran-
dom graph B2−out in at most .49n steps.
One can ask whether one can w.h.p. secure a perfect matching in a bipartite ran-
dom graph having more edges then B1−out , but less than B2−out . To see that it is
possible, consider the following two-round procedure. In the first round assume
that each vertex from the set U chooses exactly one neighbor in V and, likewise,
every vertex from the set V chooses exactly one neighbor in U. In the next round,
only those vertices from U and V which have not been selected in the first round
get a second chance to make yet another random selection. It is easy to see that,
17.3. HAMILTON CYCLES 373
for large n, such a second chance is, on the average, given to approximately n/e
vertices on each side. I.e, that the average out-degree of vertices in U and V
is approximately 1 + 1/e. Therefore the underlying simple graph is denoted as
B(1+1/e)−out , and Karoński and Pittel [486] proved that the following result holds.
To see that this result is best possible note that one can show that w.h.p. the random
graph G2−out contains a vertex adjacent to three vertices of degree two, which pre-
vents the existence of a Hamiltonian Cycle. The proof that G3−out w.h.p. contains
a Hamiltonian Cycle is long and complicated, we will therefore prove the weaker
result given below which has a straightforward proof, using the ideas of Section
6.2. It is taken from Frieze and Łuczak [353].
Theorem 17.15.
degenerate 2-cycles consisting of two copies of the same edge and there may be a
path in the case n is odd.
We use rotations as in Section 6.2. Lemma 17.16 enables us to argue that we only
need to add random edges trying to find x, y where y ∈ END(x), at most O(log n)
times. We show next that H1 = G1 ∪ G2 has sufficient expansion.
17.3. HAMILTON CYCLES 375
Lemma 17.17. W.h.p. S ⊆ [n], |S| ≤ n/1000 implies that |NH1 (S)| ≥ 2|S|.
Proof. Let X be the number of vertex sets that violate the claim. Then,
!2 k
n/1000 3k
n n 2
EX ≤ ∑ n−1
k=1 k 2k 2
n/1000 3 3 k
e n 81k4
≤ ∑
k=1 4k3 n4
n/1000 k
81e3 k
= ∑
k=1 4n
= o(1).
If n is even then we begin our search for a Hamilton cycle by choosing a cycle
of H1 and removing an edge. This will give us our current path P. If n is odd
we use the path P joining the two vertices of degree one in M1 ∪ M2 . We can
ignore the case where the isolated vertex is the same in M1 and M2 because this
only happens with probability 1/n. We run Algorithm Pósa of Section 6.2 and
observe the following: At each point of the algorithm we will have a path P plus a
collection of vertex disjoint cycles spanning the vertices not in P. This is because
in Step (d) the edge {u, v} will join two cycles, one will be the newly closed cycle
and the other will be a cycle of M. It follows that w.h.p. we will only need to
execute Step (d) at most 3 log n times.
We now estimate the probability that we reach the start of Step (d) and fail to
close a cycle. Let the edges of G0 be {e1 , e2 , . . . , en } where ei is the edge chosen
by vertex i. Suppose that at the beginning of Step (d) we have identified END. We
can go through the vertices of END until we find x ∈ END such that ex = {x, y}
where y ∈ END(x). Because G0 and H1 are independent, we see by Lemma 17.17
that we can assume P(y ∈ END(x)) ≥ 1/1000. Here we use the fact that adding
edges to H1 will not decrease the size of neighborhoods. It follows that with
probability 1 − o(1/n) we will examine fewer than (log n)2 edges of G0 before we
succeed in closing a cycle.
Now we tryclosing cycles O(log n) times and w.h.p. each time we look at O((log n)2 )
edges of G0 . So, if we only examine an edge of G0 once, we will w.h.p. still al-
ways have n/1000 − O((log n)3 ) edges to try. The probability we fail to find a
Hamilton cycle this way, given that H1 has sufficient expansion, can therefore be
bounded by P(Bin(n/1000 − O((log n)3 ), 1/1000) ≤ 3 log n) = o(1).
376 CHAPTER 17. K-OUT
Theorem 17.18.
0
if k = 1,
lim P(Gk−nearest is connected ) = γ if k = 2,
n→∞
1 if k ≥ 3,
where 0.996636 ≤ γ.
The paper [656] contains an explicit formula for γ.
Consider the related problem of the existence of a perfect matching in the bipartite
k-th nearest neighbor graph Bk−nearest . For convenience, to simplify computations,
we will assume here that the Ci, j are iid exponential random variables with rate
1/n. Coppersmith and Sorkin [237] showed that the expected size of the largest
matching in B1−nearest (which itself is a forest) is w.h.p. asymptotic to
−1
−e−1 −e−e
2−e −e n ≈ 0.807n.
17.4. NEAREST NEIGHBOR GRAPHS 377
E( f 2 (a,Y1 ))
1 2 2 −ay 2
Z ∞
−ay −ay
≈ 1 − e − aye − a y e (1 − a)e−(1−a)y dy
0 2
Z ∞
= (1 − a) (e−(1−a)y − 2ey − 2aye−y − a2 y2 e−y + e−(1+a)y
0
1
+ 2aye−(1+a)y + 2a2 y2 e−(1+a)y + a3 y3 e−(1+a)y + a4 y4 e−(1+a)y )dy.
4
Now using
i!
Z ∞
yi e−cy dy = ,
0 ci+1
we obtain
2 1 1
E( f (a,Y1 )) = (1 − a) − 2 − 2a − 2a2 +
1−a 1+a
2a 4a 2 6a3 6a4
+ + + +
(1 + a)2 (1 + a)3 (1 + a)4 (1 + a)5
2a6 (10 + 5a + a2 )
= .
(1 + a)5
Letting
g(a) = E( f 2 (a,Y1 ))a/2 ,
we have
n
n n n g(a)g(1 − a)
Pk ≤ 2 (g(a)g(1 − a)) ≈ 2 2a = 2h(a)n .
k k−1 a (1 − a)2(1−a)
Numerical examination of the function h(a) shows that it is bounded below 1 for
a in the interval [δ , 0.5], which implies that the expected number of bad pairs is
exponentially small for any k > δ n, with k ≤ n/2. Taking δ < ε < 1/(2e2 ), we
conclude that, w.h.p., there are no bad pairs in B3−nearest , and so we arrive at the
theorem.
17.5 Exercises
17.5.1 Let p = log n+(m−1)n log log n+ω where ω → ∞. Show that w.h.p. it is possible
to orient the edges of Gn,p to obtain a digraph D such that the minimum
out-degree δ + (D) ≥ m.
380 CHAPTER 17. K-OUT
17.5.3 Show that w.h.p. the diameter of Gk−out is asymptotically equal to log2k n
for k ≥ 2.
17.5.5 Let G be a graph with n vertices and minimum degree ( 12 + ε)n for some
fixed ε > 0. Let H = ki=1 Hi where (i) k ≥ 2 and (ii) H1 , H2 , . . . , Hk are
S
17.5.6 Show that w.h.p. Gk−out contains k edge disjoint spanning trees. (Hint:
Use the Nash-Williams condition [626] – see Frieze and Łuczak [354]).
17.6 Notes
k-out process
Jaworski and Łuczak [460] studied the following process that generates Gk−out
along the way. Starting with the empty graph, a vertex v is chosen uniformly at
random from the set of vertices of minimum out-degree. We then add the arc
(v, w) where w is chosen uniformly at random from the set of vertices that are
not out-neighbors of v. After kn steps the digraph in question is precisely ~Gk−out .
Ignoring orientation, we denote the graph obtained after m steps by U(n, m). The
17.6. NOTES 381
paper [460] studied the structure of U(n, m) for n ≤ m ≤ 2m. These graphs sit
between random mappings and G2−out .
In this result, one can in a natural way allow k, ` ∈ [1, 2]. Hamiltonicity was dis-
cussed in [217] where it was shown that w.h.p. D2−in,2−out is Hamiltonian.
The random digraph Dn,p as well as ~Gk−out are special cases of a random digraph
where each vertex, independently of others, first chooses its out-degree d accord-
ing to some probability distribution and then the set of its images - uniformly from
all d-element subsets of the vertex set. If d is chosen according to the binomial
distribution then it is Dn,p while if d equals k with probability 1, then it is ~Gk−out .
Basic properties of the model (monotone properties, k-connectivity), were studied
in Jaworski and Smit [462] and in Jaworski and Palka [461] .
382 CHAPTER 17. K-OUT
There has recently been an increased interest in the networks that we see around
us in our every day lives. Most prominent are the Internet or the World Wide
Web or social networks like Facebook and Linked In. The networks are con-
structed by some random process. At least we do not properly understand their
construction. It is natural to model such networks by random graphs. When first
studying so-called “real world networks”, it was observed that often the degree
sequence exhibits a tail that decays polynomially, as opposed to classical random
graphs, whose tails decay exponentially. See, for example, Faloutsos, Faloutsos
and Faloutsos [299]. This has led to the development of other models of random
graphs such as the ones described below.
deg (w, Gt )
P(yi = w) = .
2mt
383
384 CHAPTER 18. REAL WORLD NETWORKS
This model was considered by Barabási and Albert [58]. This was followed by a
rigorous analysis of a marginally different model in Bollobás, Riordan, Spencer
and Tusnády [155].
E(Dk (t + 1)|Gt ) =
(k − 1)Dk−1 (t) kDk (t)
Dk (t) + m − + 1k=m + ε(k,t). (18.1)
2mt 2mt
Explanation of (18.1): The total degree of Gt is 2mt and so
(k−1)Dk−1 (t)
2mt is the probability that yi is a vertex of degree k − 1, creating a new
k (t)
vertex of degree k. Similarly, kD2mt is the probability that yi is a vertex of degree
k, destroying a vertex of degree k. At this point t + 1 has degree m and this
accounts for the term 1k=m . The term ε(k,t) is an error term that accounts for the
possibility that yi = y j for some i 6= j.
Thus
m k
ε(k,t) = O = Õ(t −1/2 ). (18.2)
2 mt
Taking expectations over Gt , we obtain
(k − 1)D̄k−1 (t) kD̄k (t)
D̄k (t + 1) = D̄k (t) + 1k=m + m − + ε(k,t). (18.3)
2mt 2mt
Under the assumption D̄k (t) ≈ dkt (justified below) we are led to consider the
recurrence
(k−1)dk−1 −kdk
1k=m +
2 if k ≥ m,
dk = (18.4)
0 if k < m,
or
k−1 2·1k=m
k+2 dk−1 + k+2
if k ≥ m,
dk =
0 if k < m.
Therefore
2
dm =
m+2
18.1. PREFERENTIAL ATTACHMENT GRAPH 385
k
l −1 2m(m + 1)
dk = dm ∏ = . (18.5)
l=m+1 l + 2 k(k + 1)(k + 2)
So for large k, under our assumption D̄k (t) ≈ dkt, we see that
2m(m + 1)
D̄k (t) ≈ t.
k3
We now show that the assumption D̄k (t) ≈ dkt can be justified. Note that the
Proof. Let
∆k (t) = D̄k (t) − dkt.
Then, replacing D̄k (t) by ∆k (t) + dkt in (18.3) and using (18.2) and (18.4) we get
k−1 k
∆k (t + 1) = ∆k−1 (t) + 1 − ∆k (t) + Õ(t −1/2 ). (18.6)
2t 2t
where (logt)β is the hidden power of logarithm in Õ(t −1/2 ) of (18.6) and A is an
unspecified constant.
This is trivially true for k < m also for small t if we make A large enough. So,
replacing Õ(t −1/2 ) in (18.6) by the more explicit αt −1/2 (logt)β we get
∆k (t + 1) ≤
k − 1 k
∆k (t) + αt −1/2 (logt)β
≤ ∆k−1 (t) + 1 −
2t 2t
k − 1 1/2 k
≤ β
At (logt) + 1 − At 1/2 (logt)β + αt −1/2 (logt)β
2t 2t
≤ (logt)β (At 1/2 + αt −1/2 ).
1 1/2
1/2 1/2 1
(t + 1) =t 1+ ≥ t 1/2 + 1/2 ,
t 3t
386 CHAPTER 18. REAL WORLD NETWORKS
and so
β 1/2 1 α
∆k (t + 1) ≤ (log(t + 1)) A (t + 1) − 1/2 + 1/2
3t t
≤ A (log(t + 1))β (t + 1)1/2 ,
Maximum Degree
Fix s ≤ t and let Xl be the degree of vertex s in Gl for s ≤ l ≤ t. We prove the
following high probability upper bound on the degree of vertex s.
Lemma 18.2.
Here our only choice will be λs . We show below that we can find a suitable value
for this, but first observe that if we manage this then
E eλs Xt ≤ E eλs+1 Xt−1 · · · ≤ E eλt Xs ≤ 1 + o(1).
Now
1 + mεt
λ j+1 ≤ 1 + λ j,
2j
implies that
( )
t t 1/2
1 + mεt 1 + mεt m t
λt = λs ∏ 1 + ≤ λs exp ∑ ≤e λs .
j=s 2j j=s 2j s
Theorem 18.3.
u2
P(|Dk (t) − D̄k (t)| ≥ u) ≤ 2 exp − . (18.7)
8mt
388 CHAPTER 18. REAL WORLD NETWORKS
000
111
111
000
000
111
000
111
111
000
000
111
v 00
11
11
00
00
11
00
11
11
00
00
11
v
000
111
000
111 000
111
000
111 00
11
00
11 00
11
00
11
000
111 00
11
111
000
000
111 11
00
00
11
000
111
000
111 00
11
00
11
111
000
000
111 11
00
00
11
000
111
000
111 00
11
00
11
000
111 00
11
111
000
000
111 111
000
000
111
000
111
000
111 000
111
000
111
000
111 000
111
111
000
000
111 11
00
00
11
000
111
000
111 00
11
00
11
000
111 00
11
G G
~ from G.
Figure 18.1: Constructing G
Proof. Let Y1 ,Y2 , . . . ,Ymt be the sequence of edge choices made in the construction
of Gt , and for Y1 ,Y2 , . . . ,Yi let
Zi = Zi (Y1 ,Y2 , . . . ,Yi ) = E(Dk (t) | Y1 ,Y2 , . . . ,Yi ). (18.8)
We will prove next that |Zi − Zi−1 | ≤ 4 and then (18.7) follows directly from the
Azuma-Hoeffding inequality, see Section 22.7. Fix Y1 ,Y2 , . . . ,Yi and Ŷi 6= Yi . We
define a map (measure preserving projection) ϕ of
Y1 ,Y2 , . . . ,Yi−1 ,Yi ,Yi+1 , . . . ,Ymt
to
Y1 ,Y2 , . . . ,Yi−1 , Ŷi , Ŷi+1 , . . . , Ŷmt
such that
|Zi (Y1 ,Y2 , . . . ,Yi ) − Zi (Y1 ,Y2 , . . . , Ŷi )| ≤ 4. (18.9)
In the preferential attachment model we can view vertex choices in the graph G as
random choices of arcs in a digraph G, ~ which is obtained by replacing every edge
of G by a directed 2-cycle (see Figure 18.1).
Indeed, if we choose a random arc and choose its head then v will be chosen with
probability proportional to the number of arcs with v as head i.e. its degree. Hence
Y1 ,Y2 , . . . can be viewed as a sequence of arc choices. Let
Yi = (x, y) where x > y
Ŷi = (x̂, ŷ) where x̂ > ŷ.
Note that x = x̂ if i mod m 6= 1.
Now suppose j > i and Y j = (u, v) arises from choosing (w, v). Then we define
(
Yj (w, v) 6= Yi
ϕ(Y j ) = (18.10)
(w, ŷ) (w, v) = Yi
18.1. PREFERENTIAL ATTACHMENT GRAPH 389
This map is measure preserving since each sequence ϕ(Y1 ,Y2 , . . . ,Yt ) occurs with
probability ∏tm −1
j=i+1 j . Only x, x̂, y, ŷ change degree under the map ϕ so Dk (t)
changes by at most four.
We will now study the degrees of early vertices.
and
di+1 (t) + · · · + dl (t) + ( j − l)dl+1 (t) ≺ di+1 (t) + · · · + ( j − l + 1)dl (t). (18.12)
This is trivial for j = l as the LHS is then the same as the RHS. Also, if true for
l = i then
di+1 (t) + · · · + d j (t) ≺ ( j − i)di+1 (t) ≺ ( j − i)di (t)
where the second inequality follows from Lemma 18.4 with j = i + 1.
Putting Z = di+1 (t), . . . , dl−1 (t) we see that (18.12) is implied by
after subtracting ( j − l)dl+1 (t). But the latter follows from Lemma 18.4.
Lemma 18.6. Fix 1 ≤ s = O(1) and let ω = log2 t and let Ds (t) = ∑s+ω
i=s+1 ds (t).
1/2
Then w.h.p. Ds (t) ≈ 2m(ωt) .
Proof.
We have from (18.11) that
s+ω t 1/2
E(Ds (t)) ≈ m ∑ ≈ 2m(ωt)1/2 .
i=s+1 i
Going back to the proof of Theorem 18.3 we consider the map ϕ as defined in
(18.10). Unfortunately, (18.9) does not hold here. But we can replace 4 by 10 logt,
most of the time. So we let Y1 ,Y2 , . . . ,Ymt be as in Theorem 18.3. Then let ψi
denote the number of times that (w, ν) = Yi in equation (18.10). Now ψ j is the sum
of mt − j independent Bernouilli random variables and E(ψi ) ≤ ∑mt j=i+1 1/m j ≤
m log mt. It follows from Hoeffding’s inequality that Pr(ψi ≥ 10 logt) ≤ t −10 .
−1
Given this, we define a new random variable dbs (t) and let D b s (t) = ∑ω dbs+ j (t).
j=1
Here dbs+ j (t) = ds+ j (t) for j = 1, 2, . . . , ω unless there exists i such that ψi ≥
10 logt. If there is an i such that ψi ≥ 10 logt then assuming that i is the first
such we let D b s (t) = Zi (Y1 ,Y2 , . . . ,Yi ) where Zi is as defined in (18.8), with Dk (t)
replaced by D b s (t). In summary we have
So,
b s (t)) − E(Ds (t)| ≤ t −9 .
| E(D
And finally,
|Zi − Zi−1 | ≤ 20 logt.
18.2. SPATIAL PREFERENTIAL ATTACHMENT 391
u2
Pr(|Ds (t) − E(Ds (t))| ≥ u) ≤ exp −
b b . (18.14)
800mt log2 t
A1 deg− (v;t) + A2
|S(v,t)| = min ,1 . (18.16)
t
392 CHAPTER 18. REAL WORLD NETWORKS
In [8] a stronger result is proved which indicates that fraction Ni,n /n follows a
power law. It is shown that for i = 0, 1, . . . , i f , where i f = (n/ log8 n) pA1 /(4pA1 +2) ,
w.h.p.
Ni,n = (1 + o(1))ci n.
Since, for some constant c,
ci = (1 + o(1))ci−(1+1/pA1 ) ,
it shows that for large i the expected proportion Ni,n /n follows a power law with
exponent 1 + pA1 1 , and concentration for all values of i up to i f .
To prove Theorem 18.8 we need the following result of Chung and Lu (see [191],
Lemma 3.1) on real sequeances
Lemma 18.9. If {αt }, {βt }and{γt } are real sequences satisfying the relation
βt
αt+1 = 1 − αt + γt .
t
Furthermore, suppose limt→∞ βt = β > 0 and limt→∞ γt = γ. Then limt→∞ αt t exists
and
αt γ
lim =
t→∞ t 1+β
18.2. SPATIAL PREFERENTIAL ATTACHMENT 393
we get that
E(N0,t ) = c0 + o(t),
where c0 as in (18.18).
For i > 0 Lemma 18.9 can be inductively applied with
A1 (i − 1) + A2
αt = E(Ni,t ), βt = p(A1 i + A2 ) and γt = E(Ni−1,t ) ,
t
to show that
E(Ni,t ) = ci + o(t),
where
A1 (i − 1) + A2
ci = pci−1 .
1 + p(A1 i + A2 )
One can easily verify that the expressions for c0 , and ci , i ≥ 1, given in (18.18)
and (18.19), satisfy the respective recurrence relations derived above.
Knowing the expected in-degree of a node, given its age, can be used to analyze
geometric properties of the SPA graph Gt . Let us note also that the result below
for i 1 was proved in [454] and extended to all i ≥ 1 in [232]. As before, let vi
be the node added at time i.
A2 t pA1 A2
E(deg− (vi ,t)) = (1 + o(1)) − ,
A1 i A1
(18.22)
394 CHAPTER 18. REAL WORLD NETWORKS
eA2 t pA1 A2
E(deg− (vi ,t)) ≤ − ,
A1 i A1
(18.23)
pA1 −1 −pA1
E(|S(vi ,t)|) ≤ (1 + o(1))eA2t i .
A2
X(vi ,t) = deg− (vi ,t) + . (18.24)
A1
We then have,
pA1 X(vi ,t) pA1 X(vi ,t)
E(X(vi ,t + 1) | X(vi ,t)) = (X(vi ,t) + 1) + X(vi ,t) 1 −
t t
pA1
= X(vi ,t) 1 + .
t
Since all nodes start with in-degree zero, X(vi , i) = AA21 . Note that, for 0 < x < 1,
log(1 + x) = x − O(x2 ). If i 1, one can use this to get
!
t−1
A2 t−1
pA1 A2 pA1
E(X(vi ,t)) = 1+ = (1 + o(1)) exp ∑ ,
A1 ∏j=i j A1 j=i j
Therefore, when i 1,
A2 t A2 t pA1
E(X(vi ,t)) = (1 + o(1)) exp pA1 log = (1 + o(1)) ,
A1 i A1 i
and (18.22) follows from (18.24) and (18.16). Moreover, for any i ≥ 1
A2 t eA t pA1
2
E(X(vi ,t)) ≤ exp pA1 log + 1/i ≤ ,
A1 i A1 i
and (18.23) follows from (18.24) and (18.16) as before, which completes the
proof.
Directed diameter
Consider the graph Gt produced by the SPA model. For a given pair of vertices
vi , v j ∈ Vt (1 ≤ i < j ≤ t), let l(vi , v j ) denote the length of the shortest directed
path from v j to vi if such a path exists, and let l(vi , v j ) = 0 otherwise. The directed
diameter of a graph Gt is defined as
Theorem 18.11. Consider the SPA model. There exists an absolute constant c1 >
0 such that w.h.p.
D(Gt ) ≤ c1 logt.
Proof. Let C = 18 max(A2 , 1). We prove that with probability 1 − o(t −2 ) we have
that for any 1 ≤ i < j ≤ t, Gt does not contain a directed (vi , v j )-path of length
exceeding k∗ = C logt. As there are at most t 2 pairs vi , v j , Theorem 18.11 will
follow.
In order to simplify the notation, we use v to denote the vertex added at step v ≤ t.
Let vPu be a directed (v, u)-path of length given by vPu = (v,tk−1 ,tk−2 , . . . ,t1 , u),
let t0 = u,tk = v.
k
A1 deg− (ti−1 ,ti ) + A2
Pr(vPu exists) = ∏ p .
i=1 ti
We first consider the case where u tends to infinity together with t. It follows from
Theorem 18.10 that
ti pA1 A2
− A2
E(deg (ti−1 ,ti )) = (1 + o(1)) − .
A1 ti−1 A1
Thus
k
1
k
A1 E(deg− (ti−1 ,ti )) + A2
E(N(v, u, k)) = ∑ p ∏ ti
u<t1 <···<tk−1 <v i=1
ti pA1
k
k 1 k
= ∑ (1 + o(1)) (A2 p) ∏
u<t1 <···<tk−1 <v i=1 ti ti−1
v pA1 1 k−1
1
= (1 + o(1))k (A2 p)k ∑ ∏ .
u v u<t1 <···<tk−1 <v i=1 ti
However
!k−1
k−1
1 1 1
∑ ∏ ti ≤ (k − 1)! ∑ s
u<t1 <···<tk−1 <v i=1 u<s<v
1
≤ (log v/u + 1/u)k−1
(k − 1)!
e(log v/u + 1/u) k−1
≤ .
k−1
The result follows for u tending to infinity. In the case where u is a constant, it
follows from Theorem 18.10 that a multiplicative correction of e can be used in
E(deg− (ti−1 ,ti )), leading to replacing e by e2 in (18.25) and then 6 in (18.26) by
2, giving a bound of O(2−18 logt ) = o(t −4 ) as before. This finishes the proof of the
theorem.
18.3. PREFERENTIAL ATTACHMENT WITH DELETION 397
E(Dk (t + 1)|Gt ) =
(k − 1)Dk−1 (t) kDk (t)
Dk (t) + m − + 1k=m + ε(k,t). (18.27)
2m(1 − α)t 2m(1 − α)t
Here the error ε(k,t) has also to absorb the possibility that we delete and edge
incident with t + 1. This is O(1/t) and so is negligible. Given this, we follow the
subsequent analysis and obtain
Theorem 18.12. For any sufficiently small constant δ there exists a sufficiently
large constant m = m(δ ) and a constant θ = θ (δ , m) such that w.h.p. Gn has a
“giant” connected component with size at least θ n.
The proof of this is quite complicated, but it does illustrate some new ideas overf
and above what we have seen so far in this book.
In the theorem above, the constants are phrased to indicate the suspected relation-
ship, although we do not attempt to optimize them. Our unoptimized calculations
work for δ ≤ 1/50 and m ≥ δ −2 × 108 and θ = 1/30.
The proof of Theorem 18.12 is based on an idea developed by Bollobas and Rior-
dan in [154]. There they couple the graph Gn with G(n, p), the Bernoulli random
graph, which has vertex set [n] and each pair of vertices appears as an edge inde-
pendently with probability p. We couple a carefully chosen induced subgraph of
Gn with G(n0 , p).
To describe the induced subgraph in our coupling, we now make a few definitions.
We say that a vertex v of Gt is good if it was created after time t/2 and the number
of its original edges that remain undeleted exceeds m/6. By original edges of v,
we mean the m edges that were created when v was added. Let Γt denote the set
of good vertices of Gt and γt = |Γt |. We say that a vertex of Gt is bad if it is not
good. Notice that once a vertex becomes bad it remains bad for the rest of the
process. On the other hand, a vertex that was good at time t1 can become bad at a
later time t2 , simply because it was created at a time before t2 /2.
Let
m
p=
1500n
and let ∼ denote “has the same distribution as”.
Theorem 18.13. For any sufficiently small constant δ there exists a sufficiently
large constant m = m(δ ) such that we can couple the construction of Gn and a
random graph Hn , with vertex set Γn , such that Hn ∼ G (γn , p) and w.h.p. |E(Hn ) \
2 7
E(Gn )| ≤ 10−3 e−δ m/10 mn.
In Section 18.3 we prove Theorem 18.13. In Section 18.3 we prove a lower bound
on the number of good vertices, a key ingredient for the proof of Theorem 18.12,
given in section 18.3.
Lemma 18.15. W.h.p., for all t with n/2 < t ≤ n we have γt ≥ t/10.
With these lemmas, the proof of Theorem 18.12 is only a few lines:
Let G = Gn and H = G(γn , p) be the graphs constructed in Theorem 18.13. Let
G0 = G ∩ H. Then E(H) \ E(G0 ) = E(H) \ E(G) and so w.h.p. |E(H) \ E(G0 )| ≤
2 7
10−3 e−δ m/10 mn. By Lemma 18.15,|G0 | = γn ≥ n/10 w.h.p. Since m is large
2 7
enough, p = m/1500n > 10/γn and 10−3 e−δ m/10 mn < n/1000 ≤ γn /100. Then,
by Lemma 18.14, w.h.p. G0 (and therefore G) has a component of size at least
|G0 |/3 ≥ n/30.
2 ε m/8 2 m/960
≤ e−δ 0 = e−δ .
degG[k] (v) m 1
Pr(ti = v) = ≥ =
2ηk 12mk 12k
Let ⊥ be a new symbol. For each i = 1, . . . , r we choose si ∈ Γk ∪{⊥} such that for
1
each v ∈ Γk we have Pr(si = v) = 12k . We couple the selection of the si ’s with the
selection of the ti ’s such that if si 6=⊥ then si = ti . Let S = {si : i = 1, . . . , r} \ {⊥}
and X = |S|. Let Y ∼ Bin (γk , p). If r ≥ mε0 (1 − δ /2) then
200m2
γk m 1 γk
E(X) ≥ r − ≥ (1 − δ /2)ε0 m −
12k 2 γk 12n n2
≥ (1 + δ )γk p = (1 + δ ) E(Y ).
Since E(X) ≥ (1 + δ ) E(Y ), the probability that (1 + δ /2)Y > X is at most the
probability that X or Y deviates from its mean by a factor of δ /5. And, since
k m m
E(X) ≥ E(Y ) = γk p ≥ ≥ .
10 1500n 30000
2 7
By Chernoff’s bound, Pr(Y ≥ (1 + δ /5) E[Y ]) is at most e−δ m/10 .
It follows from Azuma’s inequality that for any u > 0, Pr(|X − E(X)| > u) ≤
2
e−u /(2r) . This is because X is determined by r independent trials and changing
the outcome of a single trial can only change X by at most 1. Putting u = δ E(X)/5
we get
2 2
Pr(X ≤ (1 − δ /5) E(X)) ≤ e−δ r/50 ≤ e−δ m/12000 .
We say we have a type 2 failure if Y > X, so we have a type 2 failure with proba-
2 7
bility at most 2e−δ m/10 . In which case we generate H[k + 1] by joining xk+1 to
each vertex in H[k] independently with probability p.
Conditioning on X, the si ’s form a subset S of Γk of size X chosen uniformly at
random from all of these subsets. We choose S1 uniformly at random between all
the subsets of Γk of size Y , coupling the selection of S1 to the selection of S such
that S1 ⊆ S when Y ≤ X. Now, to generate H[k + 1], we join xk+1 to every vertex
in S1 (deterministically).
After the adversary deletes a (possible empty) set of vertices in G[k], we delete all
the vertices H[k] that don’t belong to Γk+1 , possibly including xb(k+1)/2c , simply
because of its age.
For k ≥ n/2 this process yields an H[k] with vertex set Γk and identically dis-
tributed with G(γk , p), so we have H[n] ∼ G (γn , p).
18.3. PREFERENTIAL ATTACHMENT WITH DELETION 401
and
∗ (1 + δ )ρmn
Pr M >
3000
n ! n !
(1 + δ )ρmn n δ n δ
≤ Pr M ∗ >
∑ ζk ≤ ρ 1 + + Pr ∑ ζk > ρ 1 +
3000
k=n/2
2 3 k=n/2
2 3
δ 2 n2
2 n n
n δ δ
≤ Pr(Bin ρ 1+ , p > 1+ ρ p) + Pr Bin , ρ > ρ 1 +
2 3 3 2 2 2 3
δ 2 n2 ρ 1 + δ3 p
nρδ 2
≤ exp − + exp −
54 54
δ 2 nmρ 1 + δ3
nρδ 2
= exp −
+ exp −
90000 54
−5 2 δ
≤ exp −10 δ 1 + nmρ
3
402 CHAPTER 18. REAL WORLD NETWORKS
In the coupling, after time t0 and before the first failure, an edge incident with
xs+1 and destined for deletion in P is matched with an edge incident with xs+1
and destined for deletion in P ? . So, until the first failure, Ts is bounded by Ts? , the
?
corresponding total degree
2 of V s ∩ D in Gs . In Lemma 18.17 below, we prove that
Pr(Ts? > sm/2) = O e−δ mn/6 and therefore the probability of having a failure
2 mn/6
2
is O ne −δ =O e −δ mn/7 .
To repeat, if there is no failure and if ei is deleted in P before time t we have
two possibilities: xs+1 is deleted or yi is deleted. In either case, xs+1 or y?i will
be deleted by time t in P ? and therefore e?i will be deleted, and Equation (18.30)
follows.
We will show that mt 2
?
Pr zt (P ) ≥ ≤ O(e−δ mn ), (18.31)
4
and then Inequality (18.29) follows from Equation (18.30).
To prove Inequality (18.31) let s be such that t/2 ≤ s ≤ t and xs 6∈ D. We want
to upper bound the probability in the process P ? that an edge created at time s
chooses its end point in D. For i = 1, . . . , m,
Ts?
Pr(y?i ∈ D | Ts? ) = .
2ms
2 mn
By Lemma 18.17 (below), we have Pr(Ts? ≥ mt/2) ≤ O(e−δ ) so
1
Pr(y∗i ∈ D) ≤ + o(1).
4
1
Therefore zt (P ? ) is stochastically dominated by Bin mt
2 , 4 + o(1) . Inequality
(18.31) now follows from Chernoff’s bound. This completes the proof of Lemma
18.16.
404 CHAPTER 18. REAL WORLD NETWORKS
Lemma 18.17. Let A ⊂ {x1 , . . . , xt }, with |A| ≤ δ n. Let t ≥ 1000δ n and let Gt be
a graph generated by preferential attachment (i.e. the process P, but without an
adversary). Let TA denote the total degree of the vertices in A. Then
2
Pr(∃A : TA ≥ mt/2) = O e−δ mn .
Proof. Let A0 = {x1 , . . . , xδ n } be the set of the oldest δ n vertices. We can couple
the construction of Gt with Gt0 , another graph generated by preferential attach-
ment, such that TA0 ≥ TA . Therefore Pr(TA ≥ mt) ≤ Pr(TA0 ≥ mt), and we can
assume A = A0 .
Now we consider the process P in δ −1 rounds, Each round consisting of δ n steps.
Let Ti be the total degree of A at the end of the ith round. Notice that T1 = 2δ mn
and T2 ≤ 3δ mn. For i ≥ 2, fix s with iδ n < s ≤ (i + 1)δ n. Then the probability
i +δ mn
that xs chooses a vertex in A is at most T2iδ . So given Ti , the difference Ti+1 − Ti
mn
Ti +δ mn
is stochastically dominated by Yi ∼ Bin δ mn, 2iδ mn .
Therefore, for i ≥ 2,
Pr(Ti+1 ≥ 3i2/3 δ mn) ≤ Pr(Ti+1 ≥ 3i2/3 δ mn | Ti ≤ 3(i − 1)2/3 δ mn)
+ Pr(Ti ≥ 3(i − 1)2/3 δ mn)
≤ Pr(Ti+1 ≥ 3i2/3 δ mn | Ti = 3(i − 1)2/3 δ mn)
+ Pr(Ti ≥ 3(i − 1)2/3 δ mn).
Now, if t ≥ t0 then j t k
k= ≥ 103 (18.32)
δn
and so
3(k − 1)2/3 δ mn ≤ tm/2.
Thus
2
Pr(Tt ≥ tm/2) ≤ e−2δ mn .
We inflate the above by δnn to get the bound in the lemma.
Proof of Lemma 18.14 If after deleting n/100 edges the maximum component
size is at most n/3 then Gn,c/n contains a set S of size n/3 ≤ s ≤ n/2 such that there
are at most n/100 edges joining S to V \ S. The expected number of edges across
9
this cut is s(n − s)c/n so when 1 − ε = 200c we have n/100 ≤ (1 − ε)s(n − s)c/n
and by applying the union bound and Chernoff’s bound we have
n/2
n −ε 2 s(n−s)c/(2n)
Pr(∃S) ≤ ∑ e
s=n/3
s
n/2 ne s
−ε 2 (n−s)c/(2n)
≤ ∑ e
s=n/3
s
= o(1).
Theorem 18.18. If r ≤ m and ω = log2 t and p ≥ ωt −1/2 then w.h.p. all vertices
in Gt get infected.
Proof. Given Theorem 18.7 we can assume that ds (t) ≥ mt 1/2 /ω 1/2 for 1 ≤ s ≤ m.
In which case, the probability that vertex s ≤ m is not infected in round 1 is at most
m−1 m−1
mt 1/2 /ω 1/2 i
1/2 1/2 1/2
∑ p (1 − p)mt /ω −i ≤ ∑ ω i/2 e−(1−o(1))mω = o(1).
i=1 i i=1
So, w.h.p. 1, 2, . . . , m are infected in round 1. After this we use induction and the
fact that every vertex i > s has m neighbors j < i.
θ = 2((1 − α)µ p + α µq ).
αγ µq
c = β µp + ,
1−α
(1 − α)(1 − β )µ p α(1 − γ)µq
d = + ,
θ θ
αδ
e = ,
1−α
α(1 − δ )
f = .
θ
We note that
c + e = a − 1 and b = d + f . (18.33)
Now define the sequence (d0 , d1 , . . . , dk , . . .) by d0 = 0, and for k ≥ 1
k−1
dk (a + bk) = (1 − α)pk + (c + d(k − 1))dk−1 + ∑ (e + f (k − j))q j dk− j . (18.34)
j=1
Theorem 18.19. There exists a constant M > 0 such that almost surely for all
t, k ≥ 1
|Dk (t) − tdk | ≤ Mt 1/2 logt.
This will be proved in Section 18.5.
It is shown in (18.35), that the number of vertices ν(t) at step t is w.h.p. asymp-
totic to (1 − α)t. It follows that the proportion of vertices of degree k is w.h.p.
asymptotic to
dk
d¯k = .
1−α
The next theorem summarises what is known about the sequence (dk ) defined by
(18.34).
Theorem 18.20. There exist constants C1 ,C2 ,C3 ,C4 > 0 such that
(i) C1 k−ζ ≤ dk ≤ C2 min{k−1 , k−ζ / j1 } where ζ = (1 + d + f µq )/(d + f ).
Here (18.38), (18.39), (18.40) are (respectively) the main terms of the change in
the expected number of vertices of degree k due to the effect on: terminal vertices
410 CHAPTER 18. REAL WORLD NETWORKS
in NEW, the initial vertex in OLD and the terminal vertices in OLD . Rearranging
the right hand side, we find:
from (18.33). So
C2 C2 b C2 (a − 1) C2
dk − ≤ + −
k a + bk (k − j1 )(a + bk) k
C2 (a − 1) C2 a
= −
(k − j1 )(a + bk) k(a + bk)
≤ 0,
for k ≥ j1 a.
We can now prove Theorem 18.19, which is restated here for convenience.
Theorem 18.22. There exists a constant M > 0 such that almost surely for t, k ≥ 1,
Proof. Let ∆k (t) = Dk (t) − tdk . It follows from (18.34) and (18.41) that
a + bk − 1
∆k (t + 1) = ∆k (t) 1 − + O(t −1/2 logt)
t
!
j1
1
+ (c + d(k − 1))∆k−1 (t) + ∑ (e + f (k − j))q j ∆k− j (t) . (18.43)
t j=1
Let L denote the hidden constant in O(t −1/2 logt). We can adjust M to deal with
small values of t, so we assume that t is sufficiently large. Let k0 (t) = t+1−b
a . If
t max{ j0 , j1 } C2
k > k0 (t) then we observe that (i) Dk (t) ≤ k (t) = O(1) and (ii) tdk ≤ t k (t) =
0 0
O(1) follows from Lemma 18.21, and so (18.42) holds trivially.
Assume inductively that ∆κ (τ) ≤ Mτ 1/2 log τ for κ + τ ≤ k + t and that k ≤ k0 (t).
Then (18.43) and k ≤ k0 implies that for M large,
logt
|∆k (t + 1)| ≤ L 1/2
+ Mt 1/2 logt ×
t !!
j1
1
1+ c + dk + ∑ (e + f k)q j − (a + bk − 1)
t j=1
logt
=L + Mt 1/2 logt
t 1/2
≤ M(t + 1)1/2 log(t + 1)
and thus
y
1− min{dk−1 , . . . , dk− j1 } ≤ dk ≤
a + bk
y
1− max{dk−1 , . . . , dk− j1 }. (18.45)
a + bk
It follows that
k
y
d j0 ∏ 1− ≤ dk ≤
j= j0 +1 a+bj
b(k− j0 )/ j1 c
y
max{d1 , d2 , . . . , d j0 } ∏ 1− . (18.46)
s=0 a + b(k − s j1 )
k
y
= d j0 ∏ 1− .
j= j0 +1 a+bj
The upper bound in (18.46) is proved as follows: Let di1 = max{dk−1 , . . . , dk− j1 },
and in general, let dit+1 = max{dit −1 , . . . , dit − j1 }. Using (18.45) we see there is a
sequence k − 1 ≥ i1 > i2 > · · · > i p > j0 ≥ i p+1 such that |it − it−1 | ≤ j1 for all t,
and p ≥ b(k − j0 )/ j1 c. Thus
p
y
dk ≤ di p+1 ∏ 1 − ,
t=0 a + bit
= C1 k−y/b .
This establishes the lower bound of the lemma. The upper bound follows simi-
larly, from the upper bound in (18.46).
The case j1 = 1
We prove Theorem 18.20(ii). When q1 = 1, p j = 0, j > j0 = Θ(1), the general
value of dk , k > j0 can be found directly, by iterating the recurrence (18.34). Thus
1
dk = (dk−1 ((a − 1) + b(k − 1)))
a + bk
1+b
= dk−1 1 −
a + bk
k
1+b
= d j0 ∏ 1− .
j= j0 +1 a + jb
dk ≈ C6 (a + bk)−x
414 CHAPTER 18. REAL WORLD NETWORKS
where
1 2
x = 1+ = 1+ .
b α(1 − δ ) + (1 − α)(1 − β ) + α(1 − γ)
The case f = 0
We prove Theorem 18.20(iii). The case ( f = 0) arises in two ways. Firstly if
α = 0 so that a new vertex is added at each step. Secondly, if α 6= 0 but δ = 1 so
that the initial vertex of an OLD choice is sampled u.a.r.
Observe that b = d now, see (18.33).
We first prove that for a sufficiently large absolute constant A2 > 0 and for all
sufficiently large k, that
dk 1+d ξ (k)
= 1− + 2 (18.47)
dk−1 a + dk k
where |ξ (k)| ≤ A2 .
We first re-write (18.34) as
j1 k−1
dk c + d(k − 1) eq j dt−1
= +∑ ∏ . (18.48)
dk−1 a + dk j=1 a + dk t=k− j+1 dt
where |ξ ∗ ( j, k)| ≤ A3 for some constant A3 > 0. (We use the fact that j1 is constant
here.)
Substituting (18.49) into (18.48) gives
]
18.6. SMALL WORLD 415
Watts-Strogatz Model
Milgram’s experiment suggests that large real-world networks although being
globally sparse, in terms of the number edges, have their nodes/vertices connected
by relatively short short paths. In addition, such networks are locally dense, i.e.
vertices lying in a small neighborhood of a given vertex are connected by many
edges. This observation is called the ”small world” phenomenon and it has gener-
ated many attempts, both theoretical and experimental to build and study appro-
priate models of small world networks. Unfortunately, for many reasons, the clas-
sical Erdős-Rényi- Gilbert random graph Gn,p is missing many important char-
acteristics of such networks. The first attempt to build more realistic model was
introduced by Watts and Strogatz in 1998 in Nature (see [753]).
The Watts-Strogatz model starts with a kth power of a n-vertex cycle, denoted
here as Cnk . To construct it fix n and k, n ≥ k ≥ 1 and take the vertex set as
V = [n] = {1, 2, . . . , n} and edge-set E = {{i, j} : i + 1 ≤ j ≤ j + k}, where the
additions are taken modulo n.
In particular, Cn1 = Cn is a cycle on n vertices. For an example of a square Cn2 of
Cn see Figure 18.6 below.
Notice, that for n > 2k graph Cnk is 2k-regular and has nk edges. Now choose each
of nk edges of Cnk , one by one, and independently with small probability p decide
to ”rewire” it or leave it unchanged. The procedure goes as follows. We start, say,
at vertex labeled 1, and move clockwise k times around the cycle. At the ith pasage
of the cycle, at each visted vertex, we take the edge connecting it to its neighbour
at distance i to the right and decide, with probability p, if its other endpoint should
be replaced by a uniformly random vertex of the cycle. Not however allowing the
creation a double edges. Notice that after this k round procedure is completed
the number of edges of the Watts-Strogatz random graph is kn, i.e., the same as
in ”starting” graph Cnk . To study properties the original Watts-Strogatz model on
a formal mathematical ground has proved rather difficult. Therefore Newman
and Watts (see [632]) proposed a modified version, where instead of rewiring the
edges of Cnk each of n2 − nk edges not in Cnk is added independently probability p.
In fact this modification, when k = 1 was introduced earlier by Ball, Mollison and
Scalia-Tomba in [55] as the great circle epidemic model. For a rigorous results
on typical distances in such random graph see the seminal papers of Barbour and
Reinert [64] and [65].
Much earlier Bollobás and Chung in [140] took a similar approach to introducing
”shortcuts” in Cn . Namely, let Cn be a cycle with n vertices labeled clockwise
1, 2, . . . , n, so that vertex i is adjacent to vertex i + 1 for 1 ≤ i ≤ n − 1. Consider the
graph Gn obtained by adding a randomly chosen perfect matching to Cn . (We will
assume that n is even. For odd n one can add a random near prefect matching.)
Note that the graphs generated by this procedure are 3-regular (see Figure 18.3
below).
It is easy to see that a cycle Cn itself has diameter n/2. Bollobás and Chung
proved that the diameter drops dramatically after adding to Cn such system of
random ”shortcuts”.
Figure 18.3: C8 ∪ M
18.6. SMALL WORLD 417
where dist(u, v) = distGn (u, v) denotes the length of a shortest path between u and
v in Gn .
Now define the following process for generating sets Si (u) and S≤i (u) in Gn , Start
with a fixed vertex u and ”uncover” the chord (edge of M) incident to vertex u.
This determines set S1 (u). Then we add the neighbours of S1 (u) one by one to
determine S2 (u) and proceed to determine Si (u).
A chord incident to a vertex in Si (u) is called ”inessential at level i” if the other
vertex in Si (u) is within distance 3 log2 n in Cn of the vertices determined so far.
Notice that |S≤i (u) ≤ 3 · 2i and so
18 · 2i+1 log2 n
P(a chord is inessential at level i | S≤i−1 (u)) ≤ . (18.50)
n
Denote by A the event that for every vertex u at most one of the chords chosen in
S≤i (u) is inessential and suppose that i ≤ 51 log2 n. Then
P(A c ) = P(∃u : at least two of the chords chosen in S≤i (u) are inessential)
2
3 · 2i+1 18 · 2i+1 log2 n
≤n = O n−1/5 (log n)2 .
2 n
For a fixed vertex u, consider those vertices v in Si (u) for which there is a unique
path from u to v of length i, say u = u0 , u1 , . . . , ui−1 , ui = v, such that
(i) if ui−1 is adjacent to v on the cycle Cn then S≤i (u) contains no vertex on
Cn within distance 3 log2 n on the opposite side to v (denote the set of such
vertices v by Ci (u)),
(ii) if {ui−1 , v} is a chord then S≤i (u) \ {v} contains no vertex within distance,
3 log2 n both to the left and to the right of v (denote the set of such verices
by Di (u)).
418 CHAPTER 18. REAL WORLD NETWORKS
Obviously,
Ci (u) ∪ Di (u) ⊆ Si (u).
Notice that if the event A holds then, for i ≤ 51 log2 n,
For v ∈ Ci (u) a new neighbor of v in Cn is a potential element of Ci+1 (u) and a new
neighbor, which is the end-vertex of the chord from v, is a potential element of
Di+1 (u). Also if v ∈ Di (u), then the two neigbors of v in Cn are potential elements
of Ci+1 (u). Here ”potential” means that the vertices in question become elements
of Ci+1 (u) and Di+1 (u) unless the corresponding edge is inessential.
Assuming that the events A and B both hold and 51 log2 n ≤ i ≤ 53 log2 n, then
while for i ≤ 15 log2 n the bounds given in (18.51) hold. Hence for all 3 ≤ i ≤
3
5 log2 n we have
Hence,
Kleinberg’s Model
The model can be generalized significantly, but to be specific we consider the
following. We start with the n × n grid G0 which has vertex set [n]2 and where
(i, j) is adjacent to (i0 , j0 ) iff d((i, j), (i0 , j0 )) = 1 where d((i, j), (k, `)) = |i − k| +
| j − `|. In addition, each vertex u = (i, j) will choose another random neighbor
ϕ(u) where
d(u, v)−2
P(ϕ(u) = v = (k, `)) =
Du
where
Dx = ∑ d(x, y)−2 .
y6=x
The random neighbors model “long range contacts”. Let the grid G0 plus the extra
random edges be denoted by G.
It is not difficult to show that w.h.p. these random contacts reduce the diameter
of G to order log n. This however, would not explain Milgram’s success. Instead,
Kleinberg proposed the following decentralized algorithm A for finding a path
from an initial vertex u0 = (i0 , j0 ) to a target vertex uτ = (iτ , jτ ): when at u move
to the neighbor closest in distance to uτ .
420 CHAPTER 18. REAL WORLD NETWORKS
Theorem 18.25. Algorithm A finds a path from initial to target vertex of order
O((log n)2 ), in expectation.
Proof. Note that each step of A finds a node closer to the target than the current
node and so the algorithm must terminate with a path.
Observe next that for any vertex x of G we have
2n−2 2n−2
−2
Dx ≤ ∑ 4j× j =4 ∑ j−1 ≤ 4 log(3n).
j=1 j=1
2j
|B j | ≥ 1 + ∑ i > 22 j−1 .
i=1
Note that by the triangle inequality, each member of B j is within distance 2 j+1 +
2 j < 22 j+2 of u.
Let X j ≤ 2 j+1 be the time spent in Phase j. Assume first that
log2 log2 n ≤ j ≤ log2 n. Phase j will end if the long range contact of the current
vertex lies in B j . The probability of this is at least
22 j−1 1
2 j+4
= .
4 log(3n)2 128 log(3n)
We can reveal the long range contacts as the algorithm progresses. In this way,
the long range contact of the current vertex will be independent of the previous
contacts of the path. Thus
∞ ∞ i
1
E X j = ∑ P(X j ≥ i) ≤ ∑ 1 − < 128 log(3n).
i=1 i=1 128 log(3n)
Now if 0 ≤ j ≤ log2 log2 n then X j ≤ 2 j+1 ≤ 2 log2 n. Thus the expected length of
the path found by A is at most 128 log(3n) × log2 n.
In the same paper, Kleinberg showed that replacing d(u, v)−2 by
d(u, v)−r for r 6= 2 led to non-polylogarithmic path length.
18.7. EXERCISES 421
18.7 Exercises
18.7.1 Show that w.h.p. the Preferential Attachment Graph of Section 18.1 has
diameter O(log n). (Hint: Using the idea that vertex t chooses a random
edge of the current graph, observe that half of these edges appeared at time
t/2 or less).
18.7.2 For the next few questions we modify the Preferential Attachment Graph
of Section 18.1 in the following way: First let m = 1 and preferentially
generate a sequence of graphs Γ1 , Γ2 , . . . , Γmn . Then if the edges of Γmn are
(ui , vi ), i = 1, 2, . . . , mn let the edges of Gn be (udi/me , vdi/me ), i = 1, 2, . . . , mn.
Show that (18.1) continues to hold.
18.7.3 Show that Gn of the previous question can also be generated in the follow-
ing way:
18.7.4 Show that the edges of the graph in the previous question can be generated
as follows:
18.7.6 Let L be a large constant and let ω = ω(n) → ∞ arbitrarily slowly. Then
let E be the event that
k
ϒk ≈ k for ∈ [ω, n] or k = mn + 1.
m
Show that
18.8 Notes
There are by now a vast number of papers on different models of “Real World
Networks”. We point out a few additional results in the area. The books by
Durrett [273] and Bollobás, Kozma and Miklós [152] cover the area. See also van
der Hofstadt [424].
[158] discuss “local algorithms” for finding a specific vertex or the largest degree
vertex. Frieze and Pegden [361] describe an algorithm for the same problem, but
with reduced storage requirements.
Geometric models
Some real world graphs have a geometric constraint. Flaxman, Frieze and Vera
[316], [317] considered a geometric version of the preferential attachment model.
Here the vertices X1 , X2 , . . . , Xn are randomly chosen points on the unit sphere
in R3 . Xi+1 chooses m neighbors and these vertices are chosen with probability
P(deg, dist) dependent on (i) their current degree and (ii) their distance from Xi+1 .
van den Esker [296] added fitness to the models in [316] and [317]. Jordan [469]
considered more general spaces than R3 . Jordan and Wade [470] considered the
case m = 1 and a variety of definitions of P that enable one to interpolate between
the preferential attachment graph and the on-line nearest neighbor graph.
The SPA model was introduced by Aiello, Bonato, Cooper, Janssen and Pralat
[8]. Here the vertices are points in the unit hyper-cube D in Rm , equipped with
a toroidal metric. At time t each vertex v has a domain of attraction S(v,t) of
−
volume A1 deg t(v,t)+A2 . Then at time t we generate a uniform random point Xt+1
as a new vertex. If the new point lies in the domain S(v,t) then we join Xt+1 to
v by an edge directed to v, with probability p. The paper [8] deals mainly with
the degree distribution. The papers by Jannsen, Pralat and Wilson [454], [455]
show that for graphs formed according to the SPA model it is possible to infer the
metric distance between vertices from the link structure of the graph. The paper
Cooper, Frieze and Pralat [232] shows that w.h.p. the directed diameter at time t
c1 logt
lies between log logt and c2 logt.
Random Apollonian networks were introduced by Zhou, Yan and Wang [769].
Here we build a random triangulation by inserting a vertex into a randomly chosen
face. Frieze and Tsourakakis [365] studied their degree sequence and eigenvalue
structure. Ebrahimzadeh, Farczadi, Gao, Mehrabian, Sato, Wormald and Zung
[281] studied their diameter and length of the longest path. Cooper and Frieze
[227] gave an improved longest path estimate and this was further improved by
Collevecchio, Mehrabian and Wormald [203].
Weighted Graphs
There are many cases in which we put weights Xe , e ∈ E on the edges of a graph or
digraph and ask for the minimum or maximum weight object. The optimisation
questions that arise from this are the backbone of Combinatorial Optimisation.
When the Xe are random variables we can ask for properties of the optimum value,
which will be a random variable. In this chapter we consider three of the most
basic optimisation problems viz. minimum weight spanning trees; shortest paths
and minimum weight matchings in bipartite graphs.
Theorem 19.1.
∞
1
lim E Ln = ζ (3) =
n→∞
∑ k3 = 1.202 · · ·
k=1
Proof. Suppose that T = T ({Xe }) is the MST, unique with probability one. We
use the identity
Z 1
a= 1{x≤a} dx.
0
Therefore
Ln = ∑ Xe
e∈T
425
426 CHAPTER 19. WEIGHTED GRAPHS
Z 1
= ∑ 1{p≤Xe } d p
e∈T p=0
Z 1
= ∑ 1{p≤Xe}d p
p=0 e∈T
Z 1
= |{e ∈ T : Xe ≥ p}|d p
p=0
Z 1
= (κ(G p ) − 1)d p,
p=0
6 log n
p≥ ⇒ E κ(G p ) = 1 + o(1).
n
Indeed, 1 ≤ E κ(G p ) and
Write
(log n)2 (log n)2
κ(G p ) = ∑ Ak + ∑ Bk +C,
k=1 k=1
where Ak stands for the number of components which are k vertex trees, Bk is the
number of k vertex components which are not trees and, finally, C denotes the
number of components on at least (log n)2 vertices. Then, for 1 ≤ k ≤ (log n)2 and
p ≤ p0 ,
n k−2 k−1 k
E Ak = k p (1 − p)k(n−k)+(2)−k+1
k
kk−2 k−1
= (1 + o(1))nk p (1 − p)kn .
k!
n k−2 k k
E Bk ≤ k p (1 − p)k(n−k)
k 2
≤ (1 + o(1))(npe1−np )k
≤ 1 + o(1).
n
C≤ .
(log n)2
Hence
6 log n (log n)2
6 log n
Z
n
∑ E Bk d p ≤ (log n)2 (1 + o(1)) = o(1),
p=0 k=1 n
and
6 log n
6 log n n
Z
n
Cd p ≤ = o(1).
p=0 n (log n)2
So
(log n)2 k−2 Z 6 log n
kk n
E Ln = o(1) + (1 + o(1)) ∑ n pk−1 (1 − p)kn d p.
k=1 k! p=0
But
(log n)2 k−2 Z 1
kk
∑ n pk−1 (1 − p)kn d p
k=1 k! p= 6 log
n
n
428 CHAPTER 19. WEIGHTED GRAPHS
= o(1).
Therefore
(log n)2
kk−2
Z 1
k
E Ln = o(1) + (1 + o(1)) ∑ n pk−1 (1 − p)kn d p
k=1 k! p=0
(log n)2
kk−2 (k − 1)!(kn))!
= o(1) + (1 + o(1)) ∑ nk
k=1 k! (k(n + 1))!
(log n)2 k
1
= o(1) + (1 + o(1)) ∑ nk kk−3 ∏
k=1 i=1 kn + i
(log n)2
1
= o(1) + (1 + o(1)) ∑
k=1 k3
∞
1
= o(1) + (1 + o(1)) ∑ 3
.
k=1 k
One can obtain the same result if the uniform [0, 1] random variable is replaced by
any random non-negative random variable with distribution F having a derivative
equal to one at the origin, e.g. an exponential variable with mean one, see Steele
[723].
(iii)
maxi, j Xi j
P − 3 ≥ ε → 0.
log n/n
Suppose that we want to find shortest paths from a vertex s to all other vertices in a
digraph with non-negative arc-lengths. Recall Dijkstra’s algorithm. After several
iterations there is a rooted tree T such that if v is a vertex of T then the tree path
from s to v is a shortest path. Let d(v) be its length. For x ∈ / T let d(x) be the
minimum length of a path P that goes from s to v to x where v ∈ T and the sub-
path of P that goes to v is the tree path from s to v. If d(y) = min {d(x) : x ∈ / T}
then d(y) is the length of a shortest path from s to y and y can be added to the tree.
Suppose that vertices are added to the tree in the order v1 , v2 , . . . , vn and that Y j =
dist(v1 , v j ) for j = 1, 2, . . . , n. It follows from property P1 that
1
where Ek is exponential with mean k(n−k) and is independent of Yk .
This is because Xvi ,v j is distributed as an independent exponential X conditioned
on X ≥ Yk −Yi . Hence
n−1
1
EYn = ∑ k(n − k)
k=1
n−1
1 1 1
= ∑ +
n k=1 k n − k
2 n−1 1
= ∑k
n k=1
2 log n
= + O(n−1 ).
n
Also, from the independence of Ek ,Yk ,
n−1
VarYn = ∑ Var Ek
k=1
430 CHAPTER 19. WEIGHTED GRAPHS
n−1 2
1
=∑
k=1 k(n − k)
n/2 2
1
≤2∑
k=1 k(n − k)
8 n/2 1
≤ ∑ k2
n2 k=1
= O(n−2 )
1 n i−1 1
E X1,2 = ∑ ∑
n − 1 i=2 k=1 k(n − k)
1 n−1 n − k
= ∑ k(n − k)
n − 1 k=1
1 n−1 1
= ∑k
n − 1 k=1
log n
= + O(n−1 ).
n
For the variance of X1,2 we have
where
1
δi ∈ {0, 1}; δ2 + δ3 + · · · + δn = 1; P(δi = 1) = .
n−1
n
Var X1,2 = ∑ Var(δiYi ) + ∑ Cov(δiYi , δ jY j )
i=2 i6= j
n
≤ ∑ Var(δiYi ).
i=2
So
n
Var X1,2 ≤ ∑ Var(δiYi )
i=2
n 2
1 i−1
1
≤∑ ∑
i=2 n − 1 k=1 k(n − k)
= O(n−2 ).
We can now use the Chebyshev inequality.
We turn now to proving (iii). We beginnwith a lower bound. o Let
(1−ε) log n
Yi = min Xi, j : i 6= j ∈ [n] . Let A = i : Yi ≥ n . Then we have that for
i ∈ [n],
(1 − ε) log n
Pr(i ∈ A) = exp −(n − 1) = n−1+ε+o(1) . (19.2)
n
An application of the Chebyshev inequality shows that |A| ≈ nε+o(1) w.h.p. Now
the expected number of paths from a1 ∈ A to a2 ∈ A of length at most (3−2ε)
n
log n
can be bounded by
log2 n
n2ε+o(1) × n2 × n−3ε+o(1) × = n−ε+o(1) . (19.3)
n2
Explanation for (19.3): The first factor n2ε+o(1) is the expected number of pairs
of vertices a1 , a2 ∈ A. The second factor is a bound on the number of choices
b1 , b2 for the neighbors of a1 , a2 on the path. The third factor F3 is a bound on the
expected number of paths of length at most α log n
n
from b1 to b2 , α = 1 − 3ε. This
factor comes from
`+1
` α log n 1
F3 ≤ ∑ n .
`≥0 n (` + 1)!
Here ` is the number of internal vertices on the path. There will be at most n`
choices for the sequence of vertices on the path. We then use the fact that the
exponential mean one random variable stochastically dominates the uniform [0, 1]
random variable U. The final two factors are the probability that the sum of ` + 1
independent copies of U sum to at most α log n
n . Continuing we have
The final factor in (19.3) is a bound on the probability that Xa1 b1 +Xa2 b2 ≤ (2+ε)n log n .
For this we use the fact that Xai bi , i = 1, 2 is distributed as (1−ε)n log n + Ei where
432 CHAPTER 19. WEIGHTED GRAPHS
a perfect matching in Kn,n by Cn . Aldous [14], [17] proved that limn→∞ ECn =
1
ζ (2) = ∑∞k=1 k2 . The following theorem was conjectured by Parisi [643]. It was
proved independently by Linusson and Wästlund [551] and Nair, Prabhakar and
Sharma [624]. The proof given here is from Wästlund [752].
Theorem 19.3.
n
1 1 1 1 1
ECn = ∑ k2 = 1 + 4 + 9 + 16 + · · · + n2 (19.4)
k=1
From the above theorem we immediately get the following corollary, first proved
by Aldous [17].
Corollary 19.4.
1 π2
∞
lim ECn = ζ (2) = ∑ 2 = = 1.6449 · · ·
n→∞
k=1 k 6
Pr(b∗ ∈ B∗r )
P(n, r) = lim . (19.5)
λ →0 λ
Its importance lies in the following lemma:
434 CHAPTER 19. WEIGHTED GRAPHS
Lemma 19.5.
P(n, r)
E(C(n, r) −C(n, r − 1)) = . (19.6)
r
Proof. Choose i randomly from [r] and let Bbi ⊆ Br be the B-vertices in the mini-
mum weight matching of (Ar \ {ai }) into B∗ . Let X = C(n, r) and let Y = C(n, r −
1). Let wi be the weight of the edge (ai , b∗ ), and let Ii denote the indicator variable
for the event that the minimum weight of an Ar matching that contains this edge
is smaller than the minimum weight of an Ar matching that does not use b∗ . We
can see that Ii is the indicator variable for the event {Yi + wi < X}, where Yi is
the minimum weight of a matching from Ar \ {ai } to B. Indeed, if (ai , b∗ ) ∈ Mr∗
then wi < X −Yi . Conversely, if wi < X −Yi and no other edge from b∗ has weight
smaller than X −Yi , then (ai , b∗ ) ∈ Mr∗ , and when λ → 0, the probability that there
are two distinct edges from b∗ of weight smaller than X −Yi is of order O(λ 2 ). In-
deed, let F denote the existence of two distinct edges from b∗ of weight smaller
than X and let Fi, j denote the event that (ai , b∗ ) and a j , b∗ ) both have weight
smaller than X.
Then,
We now proceed to estimate P(n, r). Fix r and assume that b∗ ∈ / B∗r−1 . Suppose
∗ ∗
that Mr is obtained from Mr−1 by finding an augmenting path P = (ar , . . . , aσ , bτ )
from ar to B \ Br−1 of minimum additional weight. We condition on (i) σ , (ii) the
lengths of all edges other than (aσ , b j ), b j ∈ B \ Br−1 and
(iii) min w(aσ , b j ) : b j ∈ B \ Br−1 . With this conditioning Mr−1 = Mr−1 ∗ will
0
be fixed and so will P = (ar , . . . , aσ ). We can now use the following fact: Let
X1 , X2 , . . . , XM be independent exponential random variables of rates λ1 , λ2 , . . . , λM .
19.3. MINIMUM WEIGHT ASSIGNMENT 435
Proof.
−1 ∗ n n−1 n−r+1
lim λ Pr(b ∈ B∗r ) =
lim λ −1
1− · ···
λ →0 λ →0 n+λ n−1+λ n−r+1+λ
−1 −1 !
λ λ
= lim λ −1 1 − 1 + ··· 1+
λ →0 n n−r+1
−1 1 1 1 2
= lim λ + +···+ λ + O(λ )
λ →0 n n−1 n−r+1
1 1 1
= + +···+ . (19.9)
n n−1 n−r+1
It follows that
E(Cn+1 −Cn ) =
1 n+1 n
1 r
1 1 1
= ∑ n−i+2 + ∑ r ∑ n−i+2 − n−i+1
n + 1 i=1 r=1 i=1
n+1 n
1 1 1 1 1
= +∑ +∑ − ,
(n + 1)2 i=2 (n + 1)(n − i + 2) r=1 r n+1 n−r+1
1
= . (19.10)
(n + 1)2
19.4 Exercises
19.4.1 Suppose that the edges of the complete bipartite graph Kn,n are given in-
(b)
dependent uniform [0, 1] edge weights. Show that if Ln is the length of
the minimum spanning tree, then
(b)
lim E Ln = 2ζ (3).
n→∞
19.4.2 Let G = Kαn,β n be the complete unbalanced bipartite graph with biparti-
tion sizes αn, β n. Suppose that the edges of G are given independent uni-
(b)
form [0, 1] edge weights. Show that if Ln is the length of the minimum
spanning tree, then
i −1 i −1
(b) 1 (i1 + i2 − 1)! γ i1 i12 i21
lim E Ln =γ+ + ∑ ,
n→∞ γ i1 ≥1,i2 ≥1 i1 !i2 ! (i1 + γi2 )i1 +i2
where γ = α/β .
19.4.4 Suppose that the edges of Kn are given independent uniform [0, 1] edge
weights. Let Zk denote the minimum total edge cost of the union of k
edge-disjoint spanning trees. Show that limk→∞ Zk /k2 = 1.
19.4.5 Suppose that the edges of Gn,p where 0 < p ≤ 1 is a constant, are given ex-
ponentially distributed weights with rate 1. Show that if Xi j is the shortest
distance from i to j then
(b)
max j Xi j 2
P − ≥ ε → 0.
log n/n p
19.4. EXERCISES 437
∑np=1 xip = 1 i = 1, 2, . . . , n
xip = 0/1.
Suppose now that the ai jpq are independent uniform [0, 1] random vari-
ables. Show that w.h.p. Zmin ≈ Zmax where Zmin (resp. Zmax ) denotes the
minimum (resp. maximum) value of Z, subject to the assignment con-
straints.
19.4.7 The 0/1 knapsack problem is to
Maximise
Z = ∑ni=1 ai xi
Sub ject to
∑ni=1 bi xi ≤ L
xi = 0/1 for i = 1, 2, . . . , n.
Suppose that the (ai , bi ) are chosen independently and uniformly from
[0, 1]2 and that L = αn. Show that w.h.p. the maximum value of Z, Zmax ,
satisfies 1/2
2
α n
α ≤ 14 .
2
Zmax ≈ (8α−8α −1)n 1 ≤ α ≤ 1 .
2 4 2
n 1
2 α≥2
19.4.8 Suppose that X1 , X2 , . . . , Xn are points chosen independently and uniformly
at random from [0, 1]2 . Let Zn denote the total Euclidean length of the
shortest tour (Hamilton cycle) through each point. Show that there exist
constants c1 , c2 such that c1 n1/2 ≤ Zn ≤ c2 n1/2 w.h.p.
19.4.9 Prove equation (19.11) below. See the preceding paragraph to explain the
3-dimensional assignment problem. The implied lower bound is easy.
For the upper bound let X = V1 ×V2 ×V3 and let
H = {(it , jt , kt ),t = 1, 2, . . . , n : V1 = {i1 , . . . , in } etc.}. Let p = C/n2 for
large C and let W0 ,W1 , . . . ,Wi , . . . be as in Claim 9 of Section 14.3. Show
that w.h.p. each Wi contributes O(1/n) × (1 − γ)i to the weight of the final
assignment.
438 CHAPTER 19. WEIGHTED GRAPHS
19.5 Notes
Shortest paths
There have been some strengthenings and generalisations of Theorem 19.2. For
example, Bhamidi and van der Hofstad [90] have found the (random) second-
order term in (i), i.e., convergence in distribution with the correct norming. They
have also studied the number of edges in the shortest path.
Spanning trees
Beveridge, Frieze and McDiarmid [89] considered the length of the minimum
spanning tree in regular graphs other than complete graphs. For graphs G of
large degree r they proved that the length MST (G) of an n-vertex randomly edge
weighted graph G satisfies MST (G) = nr (ζ (3)+or (1)) w.h.p., provided some mild
expansion condition holds. For r regular graphs of large girth g they proved that
if
∞
r 1
cr = 2 ∑ ,
(r − 1) k=1 k(k + ρ)(k + 2ρ)
3n
then w.h.p. |MST (G) − cr n| ≤ 2g .
Frieze, Ruszinko and Thoma [363] replaced expansion in [89] by connectivity and
in addition proved that MST (G) ≤ nr (ζ (3) + 1 + or (1)) for any r-regular graph.
Cooper, Frieze, Ince, Janson and Spencer [228] show that Theorem 19.1 can be
improved to yield E Ln = ζ (3) + cn1 + c2n+o(1)
4/3 for explicit constants c1 , c2 .
Bollobás, Gamarnik, Riordan and Sudakov [149] considered the
Steiner Tree problem on Kn with independent random edge weights, Xe , e ∈ E(Kn ).
Here they assume that the Xe have the same distribution X ≥ 0 where P(X ≤ x) =
x + o(x) as x → 0. The main result is that if one fixes k = o(n) vertices then w.h.p.
the minimum length W of a sub-tree of Kn that includes these k points satisfies
W ≈ k−1 n
n log k .
Angel, Flaxman and Wilson [37] considered the minimum length of a spanning
tree of Kn that has a fixed root and bounded depth k. The edges weights Xe are
independent exponential mean one. They prove that if k ≥ log2 log n + ω(1) then
w.h.p. the minimum length tends to ζ (3) as in the unbounded case. On the other
hand, if k ≤ log2 log n − ω(1) then w.h.p. the weight is doubly exponential in
log2 log n − k. They also considered bounded depth Steiner trees.
19.5. NOTES 439
Using Talagrand’s inequality, McDiarmid [583] proved that for any real t > 0 we
have P(|Ln − ζ (3)| ≥ t) ≤ e−δ1 n where δ1 = δ2 (t). Flaxman [314] proved that
P(|Ln − ζ (3)| ≤ ε) ≥ e−δ2 n where δ1 = δ2 (ε).
Assignment problem
Walkup [751] proved if the weights of edges are independent uniform [0, 1] then
ECn ≤ 3 (see (19.3)) and later Karp [491] proved that ECn ≤ 2. Dyer, Frieze and
McDiarmid [279] adapted Karp’s proof to something more general: Let Z be the
optimum value to the linear program:
n
Minimise ∑ c j x j , subject to x ∈ P = {x ∈ Rn : Ax = b, x ≥ 0} ,
j=1
(The upper bound uses the result of [468], see Section 14.3).
Frieze and Sorkin [364] give an O(n3 ) algorithm that w.h.p. finds a solution of
1
value n1−o(1) .
In another version of the 3-dimensional assignment problem we ask for a min-
imum weight collection of hyper-edges such that each pair of vertices v, w ∈ V
from different sets in the partition appear in exactly one edge. The optimal total
weight Z of this collection satisfies
(The upper bound uses the result of [279] to greedily solve a sequence of restricted
assignment problems).
440 CHAPTER 19. WEIGHTED GRAPHS
Chapter 20
There are several topics that we have not been able to cover and that might be of
interest to the reader. For these topics, we provide some short synopses and some
references that the reader may find useful.
Contiguity
Suppose that we have two sequences of probability models on graphs G1,n , G2,n on
the set of graphs with vertex set [n]. We say that the two sequences are contiguous
if for any sequence of events An we have
This for example, is useful for us, if we want to see what happens w.h.p. in
the model G1,n , but find it easier to work with G2,n . In this context, Gn,p and
Gn,m=n2 p/2 are almost contiguous.
Interest in this notion in random graphs was stimulated by the results of Robin-
son and Wormald [677], [680] that random r-regular graphs, r ≥ 3, r = O(1) are
Hamiltonian. As a result, we find that other non-uniform models of random regu-
lar graphs are contiguous to Gn,r e.g. the union rMn of r random perfect matchings
when n is even. (There is an implicit conditioning on rMn being simple here). The
most general result in this line is given by Wormald [761], improving on earlier
results of Janson [439] and Molloy, Robalewska, Robinson and Wormald [604]
and Kim and Wormald [502]. Suppose that r = 2 j + ∑r−1 i=1 iki , with all terms non-
negative. Then Gn,r is contiguous to the sum jHn + ∑r−1 i=1 ki Gn,i , where n is re-
stricted to even integers if ki 6= 0 for any odd i. Here jHn is the union of j edge
disjoint Hamilton cycles etc.
Chapter 8 of [449] is devoted to this subject.
441
442 CHAPTER 20. BRIEF NOTES ON UNCOVERED TOPICS
Games
Positional games can be considered to be a generalisation of the game of “Noughts
and Crosses” or “Tic-Tac-Toe”. There are two players A (Maker) and B (Breaker)
and in the context for this section, the board will be a graph G. Each player in
turn chooses an edge and at the end of the game, the winner is determined by the
partition of the edges claimed by the players. As a typical example, in the con-
nectivity game, player A is trying to ensure that the edges she collects contain a
spanning tree of G and player B is trying to prevent this. See Chvátal and Erdős
443
[196] for one of the earliest papers on the subject and books by Beck [68] and
Hefetz, Krivelevich, Stojaković and Szabó [419]. Most of the analyses have con-
sidered G = Kn and to make the problem interesting [196] introduced the notion
of bias. Thus in the connectivity game, player B is allowed to collect b edges
for each edge of A. Now the question becomes what is the largest value of b for
which A has a winning strategy. There is a striking though somewhat mysterious
connection between the optimal values of b for various games and thresholds for
associated properties in random graphs. For example in the connectivity game, the
threshold bias b ≈ logn n i.e. player A collects about 12 n log n edges, see Gebauer
and Szabó [376]. Another example is the biased H-game where Maker wins if
she can create a copy of some fixed graph H with at least two adjacent edges. The
1/m (H)
optimal threshold bias b for this game is of order Θ n 2 , Bednarska and
Łuczak [69]. For sufficiently small constant c > 0, if b ≤ cn 1/m 2 (H) , then Maker
Kierstead and and Zhu [66]. Bohman, Frieze and Sudakov [122] studied χg for
dense random graphs and proved that for such graphs, χg is within a constant
factor of the chromatic number. Keusch and Steger [499] proved that this factor is
asymptotically equal to two. Frieze, Haber and Lavrov [345] extended the results
of [122] to sparse random graphs.
Graph Searching
A collection of cops are placed on the vertices of a graph by player C and then
a robber is placed on a vertex by player R. The players take turns. C can move
all cops to a neighboring vertex and R can move the robber. The cop number of
a graph is them minimum number of cops needed so that C can win. The basic
rule being that if there is a cop occupying the same vertex as the robber, then C
wins. Łuczak and Pralat [566] proved a remarkable “zigzag” theorem giving the
cop number of a random graph. This number being nα where α = α(p) follows
a saw-toothed curve. Pralat and Wormald [659] proved that the cop number of
the random regular graph Gn,r is O(n1/2 ). It is worth noting that Meyniel has
conjectured O(n1/2 ) as a bound on the cop number of any connected n-vertex
graph. There are many variations on this game and the reader is referred to the
monograph by Bonato and Pralat [157].
Graph Cleaning
Initially, every edge and vertex of a graph G is dirty, and a fixed number of brushes
start on a set of vertices. At each time-step, a vertex v and all its incident edges
that are dirty may be cleaned if there are at least as many brushes on v as there
are incident dirty edges. When a vertex is cleaned, every incident dirty edge is
traversed (that is, cleaned) by one and only one brush, and brushes cannot traverse
a clean edge. The brush number b(G) is the minimum number of brushes needed
1−e−2d d
to clean G. Pralat [660], [661] proved that w.h.p. b(Gn,p ) ≈ 4 n for p = n
−2d
where d < 1 and w.h.p. b(Gn,p ) ≤ (1+o(1)) d + 1 − 1−e2d n
4 for d > 1. For the
random d-regular
graph Gn,d , Alon, Pralat and Wormald [25] proved that w.h.p.
3/2
b(Gn,d ) ≥ dn
4 1 − d21/2 .
445
Acquaintance Time
Let G = (V, E) be a finite connected graph. We start the process by placing one
agent on each vertex of G. Every pair of agents sharing an edge are declared to
be acquainted, and remain so throughout the process. In each round of the pro-
cess, we choose some matching M in G. The matching M need not be maximal;
perhaps it is a single edge. For each edge of M, we swap the agents occupying
its endpoints, which may cause more agents to become acquainted. We may view
the process as a graph searching game with one player, where the player’s strategy
consists of a sequence of matchings which allow all agents to become acquainted.
Some strategies may be better than others, which leads to a graph optimisation pa-
rameter. The acquaintance time of G, denoted by A (G), is the minimum number
of rounds required for all agents to become acquainted with one another. The pa-
rameter A (G) was 2introduced by Benjamini, Shinkar and Tsur [73], who showed
n log log n
that A (G) = O log n for an n vertex graph. The log log n factor was re-
moved by Kinnersley, Mitsche and Pralat [504]. The paper [504] also showed
that w.h.p. A (Gn,p ) = O p log n
for (1+ε)n log n ≤ p ≤ 1 − ε. The lower bound
was relaxed to np − log n → ∞ in Dudek and Pralat [271]. A lower bound,
here
Ω logp n for Gn,p and p ≥ n−1/2+ε was proved in [504].
H-free process
In an early attempt to estimate the Ramsey number R(3,t), Erdős, Suen and Win-
kler [294] considered the following process for generating a triangle free graph.
Let e1 , e2 , . . . , eN , N = n2 be a random ordering of the complete graph Kn . Let
P be a graph property e.g. being triangle free. We generate a sequence of ran-
dom graphs Γ0 , Γ1 , . . . , ΓN where Γi+1 = Γi + ei+1 if adding ei+1 does not destroy
P, otherwise Γi+1 = Γi . In this way we can generate a random graph that is
guaranteed to have property P.
For P is “bipartite” they show in [294] that ΓN has expected size greater than
(n2 − n)/4. When P is “triangle free” they show that w.h.p. that ΓN has size
Ω(n3/2 ) w.h.p. Bollobás and Riordan [153] studied the general H-free process.
More recently, Bohman [114] showed in the case of the triangle free process,
that w.h.p. ΓN has size Θ(n3/2 (logn)1/2). This provides an alternative proof to
t2
that of Kim [500] that R(3,t) = Ω logt . He made use of a careful use of the
differential equations method, see Chapter 23. Bohman and Keevash [123] and
Fiz Pontiveros, Griffiths and Morris [313] have improved this result and shown
1 3/2
that w.h.p. ΓN has size asymptotically equal to 2√ 2
n (log n)1/2 . They also show
that the independence number of ΓN is bounded by (1 + o(1))(2n log n)1/2 . This
446 CHAPTER 20. BRIEF NOTES ON UNCOVERED TOPICS
Planarity
We have said very little about random planar graphs. This is partially because
there is no simple way of generating a random planar graph. The study begins
with the seminal work of Tutte [742], [743] on counting planar maps. The number
of rooted maps on surfaces was found by Bender and Canfield [76]. The size of
447
the largest components were studied by Banderier, Flajolet, Schaeffer and Soria
[57].
When it comes to random labeled planar graphs, McDiarmid, Steger and Welsh
[588] showed that if pl(n) denotes the number of labeled planar graphs with n
vertices, then (pl(n)/n!)1/n tends to a limit γ as n → ∞. Osthus, Prömel and Taraz
[636] found an upper bound for γ, Bender, Gao and Wormald [77] found a lower
bound for γ. Finally, Giménez and Noy [383] proved that pl(n) ≈ cn−7/2 γ n n! for
explicit values of c, γ.
Next let pl(n, m) denote the number of labelled planar graphs with n vertices and
m edges. Gerke, Schlatter, Steger and Taraz [379] proved that if 0 ≤ a ≤ 3 then
(pl(n, an)/n!)1/n tends to a limit γa as n → ∞. Giménez and Noy [383] showed
that if 1 < a < 3 then pl(n, an) ≈ ca n−4 γan n!. Kang and Łuczak [480] proved the
existence of two critical ranges for the sizes of complex components.
Frieze and Shamir [167] showed that if c is sufficiently large then w.h.p. one can
in polynomial time find a Hamilton cycle H in Γ. While H may not necessarily
be C, this rules out a simple use of Hamilton cycles for a signature scheme.
Random Lifts
For a graph K, an n-lift G of K has vertex set V (K) × [n] where for each vertex
v ∈ V (K), {v} × [n] is called the fiber above v and will be denoted by Πv . The edge
set of a an n-lift G consists of a perfect matching between fibers Πu and Πw for
each edge {u, w} ∈ E(K). The set of n-lifts will be denoted Λn (K). In a random
n-lift, the matchings between fibers are chosen independently and uniformly at
random.
Lifts of graphs were introduced by Amit and Linial in [34] where they proved that
if K is a connected, simple graph with minimum degree δ ≥ 3, and G is a random
n-lift of K then G is δ (G)-connected w.h.p., where the asymptotics are for n → ∞.
They continued the study of random lifts in [35] where they proved expansion
properties of lifts. Together with Matoušek, they gave bounds on the indepen-
dence number and chromatic number of random lifts in [36]. Linial and Rozen-
man [550] give a tight analysis for when a random n-lift has a perfect matching.
Greenhill, Janson and Ruciński [398] consider the number of perfect matchings
in a random lift.
Łuczak, Witkowski and Witkowski [569] proved that a random lift of H is Hamil-
tonian w.h.p. if H has minimum degree at least 5 and contains two disjoint Hamil-
tonian cycles whose union is not a bipartite graph. Chebolu and Frieze [181] con-
sidered a directed version of lifts and showed that a random lift of the complete
digraph K ~ h is Hamiltonian w.h.p. provided h is sufficiently large.
Mixing Time
Generally speaking, the probability that a random walk is at a particular vertex
tends to a steady state probability deg(v)
2m . The mixing time is the time taken for
the distribution k-step distribution to get to within variation distance 1/4, say, of
the steady state. Above the threshold for connectivity, the mixing time of Gn,p is
certainly O(log n) w.h.p. For sparser graphs, the accent has been on finding the
mixing time for a random walk on the giant component. Fountoulakis and Reed
[326] and Benjamini, Kozma and Wormald [72] show that w.h.p. the mixing time
of a random walk on the giant component of Gn,p , p = c/n, c > 1 is O((log n)2 ).
Nachmias and Peres [622] showed that the mixing time of the largest component
of Gn,p , p = 1/n is in [εn, (1 − ε)n] with probability 1 − p(ε) where p(ε) → 0 as
ε → 0. Ding, Lubetzky and Peres [257] show that mixing time for the emerging
giant at p = (1 + ε)/n where λ = ε 3 n → ∞ is of order (n/λ )(log λ )2 . For random
regular graphs, the mixing time is O(log n) and Lubetzky and Sly [553] proved
that the mixing time exhibits a cut-off phenomenon i.e. the variation distance goes
from near one to near zero very rapidly.
450 CHAPTER 20. BRIEF NOTES ON UNCOVERED TOPICS
Cover Time
The covertime CG of a graph G is the maximum over starting vertex of the ex-
pected time for a random walk to visit every vertex of G. For G = Gn,p with
p = c log
n
n
where c > 1, Jonasson [467] showed that w.h.p. CG = Θ(n log n).
c
Cooper and Frieze [222] proved that CG ≈ A(c)n log n where A(c) = c log c−1 .
Then in [221] they showed that the cover time of a random r-regular graph is
w.h.p. asymptotic to r−1r−2 n log n, for r ≥ 3. Then in a series of papers they es-
tablished the asymptotic cover time for preferential attachment graphs [222]; the
giant component of Gn,p , p = c/n, where c > 1 is constant [223]; random geo-
metric graphs of dimension d ≥ 3, [224]; random directed graphs [225]; random
graphs with a fixed degree sequence [2], [230]; random hypergraphs [234]. The
asymptotic covertime of random geometric graphs for d = 2 is still unknown.
Avin and Ercal [43] prove that w.h.p. it is Θ(n log n). The paper [226] deals with
the structure of the subgraph Ht induced by the un-visited vertices in a random
walk on a random graph after t steps. It gives tight results on a phase transition
i.e. a point where H breaks up into small components. Cerny and Teixeira [178]
refined the result of [226] near the phase transition.
Stable Matching
In the stable matching problem we have a complete bipartite graph on vertex sets
A, B where A = {a1 , a2 , . . . , an } , B = {b1 , b2 , . . . , bn }. If we think of A as a set
of women and B as a set of men, then we refer to this as the stable marriage
problem. Each a ∈ A has a total ordering pa of B and each b ∈ B has a total
ordering pb of B. The problem is to find a perfect matching (ai , bi ), i = 1, 2, . . . , n
such that there does not exist a pair i, j such that b j > bi in the order pai and
ai > b j in the order pb j . The existence of i, j leads to an unstable matching.
Gale and Shapley [371] proved that there is always a stable matching and gave
an algorithm for finding one. We focus on the case where pa , pb are uniformly
random for all a ∈ A, b ∈ B. Wilson [757] showed that the expected number of
proposals in a sequential version of the Gale-Shapley algorithm is asymptotically
equal to n log n. Knuth, Motwani and Pittel [507] studied the likely number of
stable husbands for an element of A ∪ B. I.e. they show that w.h.p. there are
constants c < C such that for a fixed a ∈ A there are between c log n and C log n
choices b ∈ B such that a and b are matched together in some stable matching. The
question of how many distinct stable matchings there are likely to be was raised
in Pittel [651] who showed that w.h.p. there are at least n1/2−o(1) . More recently,
Lennon and Pittel [543] show that there are at least n log n with probability at
least 0.45. Thus the precise growth rate of the number of stable matchings is not
451
clear at the moment. Pittel, Shepp and Veklerov [655] considered the number
Zn,m of a ∈ A that have exactly m choices of stable husband. They show that
E(Z ) 1
limn→∞ (log n)n,mm+1 = (m−1)! .
Universal graphs
A graph G is universal for a class of graphs H if G contains a copy of every
H ∈ H . In particular, let H (n, d) denote the set of graphs with vertex set [n]
and maximum degree at most d. One question that has concerned researchers, is
to find the threshold for Gn,p being universal for H (n, d). A counting argument
shows that any H (n, d) universal graph has Ω(n2−2/d ) edges. For random graphs
this can be
nimproved
to Ω(n2−2/(d+1) (log n)O(1) ). This is because to contain the
union of d+1 disjoint copies of Kd+1 , all but at most d vertices must lie in a copy
of Kd+1 . This problem was first considered in Alon, Capalbo, Kohayakawa, Rödl,
Ruciński and Szemerédi [24]. Currently the best upper bound on the value of p
needed to make Gn,m H (n, d) universal is O(n2−1/d (log n)1/d ) in Dellamonica,
Kohayakawa, Rödl, and Ruciński [242]. Ferber, Nenadov and Peter [307] prove
that if p ∆8 n−1/2 log n then Gn,p is universal for the set of trees with maximum
degree ∆.
452 CHAPTER 20. BRIEF NOTES ON UNCOVERED TOPICS
Part IV
453
Chapter 21
Moments
455
456 CHAPTER 21. MOMENTS
Proof.
Var X EX2
P(X = 0) ≤ = −1
(E X)2 (E X)2
Var X
P(X = 0) ≤ P(|X − E X| ≥ E X) ≤
(E X)2
Var X (E X)2
P(X = 0) ≤ = 1 − .
EX2 EX2
Proof. Notice that
X = X · I{X≥1} .
Then, by the Cauchy-Schwarz inequality,
2
(E X)2 = E(X · I{X≥1} ) ≤ E I{X≥1}
2
E X 2 = P(X ≥ 1) E X 2 .
21.1. FIRST AND SECOND MOMENT METHOD 457
The bound in Lemma 21.5 is stronger than the bound in Lemma 21.4, since E X 2 ≥
(E X)2 . However, for many applications, these bounds are equally useful since the
Second Moment Method can be applied if
Var X
→ 0, (21.1)
(E X)2
or, equivalently,
EX2
→ 1, (21.2)
(E X)2
as n → ∞. In fact if (21.1) holds, then much more than P(X > 0) → 1 is true. Note
that
X 2
2
Var X X X
= Var =E − E
(E X)2 EX EX EX
2
X
= E −1
EX
Hence
2
X Var X
E −1 → 0 if → 0.
EX (E X)2
It simply means that
X L2
−→ 1. (21.3)
EX
In particular, it implies (as does the Chebyshev inequality) that
X P
−→ 1, (21.4)
EX
i.e., for every ε > 0,
So, we can only apply the Second Moment Method, if the random variable X has
its distribution asymptotically concentrated at a single value (X can be approxi-
mated by the non-random value E X, as stated at (21.3), (21.4) and (21.5)).
We complete this section with another lower bound on the probability P(Xn ≥ 1),
when Xn is a sum of (asymptotically) negatively correlated indicators. Notice that
in this case we do not need to compute the second moment of Xn .
458 CHAPTER 21. MOMENTS
(E Xn )2
P(Xn ≥ 1) ≥ .
E Xn2
Now
n n
E Xn2 = ∑ ∑ E(IiI j )
i=1 j=1
≤ E Xn + (1 + εn ) ∑ E Ii E I j
i6= j
!
n 2 n
= E Xn + (1 + εn ) ∑ E Ii − ∑ (E Ii )2
i=1 i=1
≤ E Xn + (1 + εn )(E Xn )2 .
The next result, which can be deduced from Theorem 21.7, provides a tool to
prove asymptotic Normality.
Corollary 21.8. Let X1 , X2 , . . . , Xn , . . . be a sequence of random variables with
finite moments and let a1 , a2 , . . . , an , . . . be a sequence of positive numbers, such
that
(
(2m)! k
m a + o(akn ), when k = 2m, m ≥ 1,
E(Xn − E Xn )k = 2 m!k n
o(an ), when k = 2m − 1, m ≥ 2,
as n → ∞. Then
Xn − E Xn D Xn − E Xn D
→ Z, and X̃n = √ → Z,
an Var Xn
where Z is a random variable with the standard Normal distribution N(0, 1).
A similar result for convergence to the Poisson distribution can also be deduced
from Theorem 21.7. Instead, we will show how to derive it directly from the
Inclusion-Exclusion Principle.
The following lemma sometimes simplifies the proof of some probabilistic in-
equalities: A boolean function f of events A1 , A2 , . . . , An⊆ Ω is a random variable
where f (A1 , A2 , . . . , An ) = S∈S ( i∈S Ai ) ∩ i6∈S Aci for some collection Si
S T T
Proof. Write ! !!
[ \ \
fi = Ai ∩ Aci ,
S∈Si i∈S i6∈S
for some real βS . If (21.6) holds, then βS ≥ 0 for every S, since we can choose
Ai = Ω if i ∈ S, and Ai = 0/ for i 6∈ S.
For J ⊆ [r] let AJ = i∈J Ai , and let S = | j : A j occurs | denote the number of
T
More precisely,
Lemma 21.10.
s
≤ (−1)k− j k Bk
s − j even.
∑ j
k= j
s
k− j k B
P(E j ) ≥ ∑ (−1) j k s − j odd
k= j
s
k− j k B
∑ (−1)
= s = r.
j k
k= j
21.2. CONVERGENCE OF MOMENTS 461
Proof. It follows from Lemma 21.9 that we only need to check the truth of the
statement for
P(Ai ) = 1 1 ≤ i ≤ `,
P(Ai ) = 0 ` < i ≤ r.
where 0 ≤ ` ≤ r is arbitrary.
Now (
1 if j = `,
P(S = j) =
0 6 `,
if j =
and
`
Bk = .
k
So,
s s
k− j k k− j k `
∑ (−1) j
Bk = ∑ (−1)
j k
k= j k= j
s
` k− j ` − j
= ∑ (−1) k − j . (21.7)
j k= j
If ` < j then P(E j ) = 0 and the sum in (21.7) reduces to zero. If ` = j then
P(E j ) = 1 and the sum in (21.7) reduces to one. Thus in this case, the sum is exact
for all s. Assume then that r ≥ ` > j. Then P(E j ) = 0 and
s s− j
k− j `− j t `− j s− j ` − j − 1
∑ (−1) k− j
= ∑ (−1)
t
= (−1)
s− j
.
k= j t=0
`− j−1
This explains the alternating signs of the theorem. Finally, observe that r− j
= 0, as required.
Now we are ready to state the main tool for proving convergence to the Poisson
distribution.
Theorem 21.11. Let Sn = ∑i≥1 Ii be a sequence of random variables, n ≥ 1 and
(n)
let Bk = E Skn . Suppose that there exists λ ≥ 0, such that for every fixed k ≥ 1,
(n) λk
lim Bk = .
n→∞ k!
Then, for every j ≥ 0,
λj
lim P(Sn = j) = e−λ ,
n→∞ j!
462 CHAPTER 21. MOMENTS
So, as n grows to ∞,
j+2l+1
k− j k (n)
∑ (−1) B ≤ lim inf P(Sn = j)
k= j j k n→∞
j+2l
k− j k (n)
≤ lim sup P(Sn = j) ≤ ∑ (−1) B .
n→∞ k= j j k
But,
j+m
λj m
k t
k− j k λ tλ λ j −λ
∑ (−1) = ∑ (−1) → e ,
k= j j k! j! t=0 t! j!
as m → ∞.
Notice that the falling factorial
counts number of ordered k-tuples of events with Ii = 1. Hence the binomial mo-
ments of Sn can be replaced in Theorem 21.11 by the factorial moments, defined
as
E(Sn )k = E[Sn (Sn − 1) · · · (Sn − k + 1)],
and one has to check whether, for every k ≥ 1,
lim E(Sn )k = λ k .
n→∞
the computations are often easier and fewer moment assumptions are required.
Moreover, it frequently leads to conditions for convergence weaker than those
obtainable by the method of moments.
Consider a sequence of random variables (Xn )∞ ∞
n=1 and let (λn )n=1 be a sequence
of positive integers, and let Po (λ ) denote, as before, the Poisson distribution with
expectation λ . We say that Xn is Poisson convergent if the total variation distance
between the distribution L (Xn ) of Xn and Po(λn ), λn = E Xn , distribution, tends
to zero as n tends to infinity. So, we ask for
λk
dTV (L (Xn ), Po(λn )) = sup P(Xn ∈ A) − ∑ n e−λn → 0, (21.8)
A⊆Z+ k∈A k!
Theorem 21.12. Let X = ∑a∈Γ Ia where the Ia are indicator random variables
with a dependency graph L. Then, with πa = E Ia and
λ = E X = ∑a∈Γ πa ,
where ∑ab∈E(L) means summing over all ordered pairs (a, b), such that {a, b} ∈
E(L).
464 CHAPTER 21. MOMENTS
Finally, let us briefly mention, that the original Stein method investigates the con-
vergence to the normal distribution in the following metric
Z Z
dS (L (Xn ), N(0, 1)) = sup ||h||−1 h(x)dFn (x) − h(x)dΦ(x), (21.9)
h
where the supremum is taken over all bounded test functions h with bounded
derivative, ||h|| = sup |h(x)| + sup |h0 (x)|.
Here Fn is the distribution function of Xn , while Φ denotes the distribution function
of the standard normal distribution. So, if dS (L (Xn ), N(0, 1)) → 0 as n → ∞, then
X̃n converges in distribution to N(0, 1) distributed random variable.
Barbour, Karoński and Ruciński [62] obtained an effective upper bound on dS (L (Xn ), N(0, 1))
if S belongs to a general class of decomposable random variables. This bound in-
volves the first three moments of S only.
For a detailed and comprehensive account of the Stein–Chen method the reader
is referred to the book by Barbour, Holst and Janson [60], or to Chapter 6 of the
book by Janson, Łuczak and Ruciński [449], where other interesting approaches to
study asymptotic distributions of random graph characteristics are also discussed.
For some applications of the Stein–Chen method in random graphs, one can look
at a survey by Karoński [484].
Chapter 22
Inequalities
(b)
1 − x ≥ e−x/(1−x) , 0 ≤ x < 1.
(c)
n ne k
≤ , ∀n, k.
k k
(d)
nk k k−1
n
≤ 1− , ∀n, k.
k k! 2n
(e)
nk nk
k(k − 1) n
1− ≤ ≤ e−k(k−1)/(2n) , ∀n, k.
k! 2n k k!
(f)
nk
n
≈ , i f k2 = o(n).
k k!
(g) If a ≥ b then
n−a
t − b b n − t − a + b a−b t b n − t a−b
t−b
≤ n ≤ .
n−b n−a t
n n−b
465
466 CHAPTER 22. INEQUALITIES
Proof. (g)
n−a
t−b (n − a)!t!(n − t)!
n =
t
n!(t − b)!(n − t − a + b)!
t(t − 1) · · · (t − b + 1) (n − t)(n − t − 1) · · · (n − t − a + b + 1)
= ×
n(n − 1) · · · (n − b + 1) (n − b)(n − b − 1) · · · (n − a + 1)
t b n − t a−b
≤ × .
n n−b
Proof.
nk k−1
n i
= ∏ 1−
k k! i=0 n
( )
k−1
nk
i
= exp ∑ log 1 −
k! i=0 n
( 4 )
k k−1 2
n i i k
= exp − ∑ + 2 +O 3
k! i=0 n 2n n
nk k2 k3
= (1 + o(1)) exp − − 2 .
k! 2n 6n
Theorem 22.3. Let S, T be disjoint subsets of [M] and let s,t be non-negative
integers. Then
P(WS ≤ s | WT ≤ t) ≤ P(WS ≤ s). (22.1)
P(WS ≥ s | WT ≤ t) ≥ P(WS ≥ s). (22.2)
P(WS ≥ s | WT ≥ t) ≤ P(WS ≥ s). (22.3)
P(WS ≤ s | WT ≥ t) ≥ P(WS ≤ s). (22.4)
Proof. Equation (22.2) follows immediately from (22.1). Also, equation (22.4)
follows immediately from (22.3). The proof of (22.3) is very similar to that of
(22.1) and so we will only prove (22.1).
Let
πi = P(WS ≤ s | WT = i).
Given WT = i, we are looking at throwing N − i balls into M − 1 boxes. It is clear
therefore that πi is monotone increasing in i. Now, let qi = P(WT = i). Then,
N
P(WS ≤ s) = ∑ πi qi .
i=0
t
qi
P(WS ≤ s | WT ≤ t) = ∑ πi .
i=0 q0 + · · · + qt
So, (22.1) reduces to
t N
(q0 + · · · + qN ) ∑ πi qi ≤ (q0 + · · · + qt ) ∑ πi qi ,
i=0 i=0
or
t N
(qt+1 + · · · + qN ) ∑ πi qi ≤ (q0 + · · · + qt ) ∑ πi qi ,
i=0 i=t+1
or
N t t N
∑ ∑ qiq j πi ≤ ∑ ∑ qi q j πi .
j=t+1 i=0 j=0 i=t+1
The result now follows from the monotonicity of πi .
The following is an immediate corollary:
Corollary 22.4. Let S1 , S2 , . . . , Sk be disjoint subsets of [M] and let s1 , s2 , . . . , sk be
non-negative integers. Then
!
k
\ k
P {WSi ≤ si } ≤ ∏ P({WSi ≤ si }).
i=1 i=1
!
\k k
P {WSi ≥ si } ≤ ∏ P({WSi ≥ si }).
i=1 i=1
468 CHAPTER 22. INEQUALITIES
a graph with vertex set [n]. Then A will be a set of graphs i.e. a graph property.
Suppose that f is the indicator function for A . Then f is monotone increasing, if
whenever G ∈ A and e ∈ / E(G) we have G + e ∈ A i.e. adding an edge does not
destroy the property. We will say that the set/property is monotone increasing. For
example if H is the set of Hamiltonian graphs then H is monotone increasing.
If P is the set of planar graphs then P is monotone decreasing. In other words a
property is monotone increasing iff its indicator function is monotone increasing.
Suppose next that we turn CN into a probability space by choosing some p1 , p2 ,
. . . , pN ∈ [0, 1] and then for x = (x1 , x2 , . . . , xN ) ∈ CN letting
P(x) = ∏ pj ∏ (1 − p j ). (22.5)
j:x j =1 j:x j =0
The following is a special case of the FKG inequality, Harris [412] and Fortuin,
Kasteleyn and Ginibre [321]:
E( f g | xN = 0) ≥ E( f | xN = 0) E(g | xN = 0) = 0
E( f g | xN = 1) ≥ E( f | xN = 1) E(g | xN = 1) ≥ 0
E( f g) = E( f g | xN = 0)(1 − pN ) + E( f g | xN = 1)pN
≥ E( f | xN = 1) E(g | xN = 1)pN . (22.6)
Furthermore,
E( f ) E(g) =
(E( f | xN = 0)(1 − pN ) + E( f | xN = 1)pN )×
(E(g | xN = 0)(1 − pN ) + E(g | xN = 1)pN )
= E( f | xN = 1) E(g | xN = 1)p2N . (22.7)
The result follows by comparing (22.6) and (22.7) and using the fact that E( f |
xN = 1), E(g | xN = 1) ≥ 0 and 0 ≤ pN ≤ 1.
In terms of monotone increasing sets A , B and the same probability (22.5) we
can express the FKG inequality as
Combining (22.9) and (22.10) one can obtain a bound for P(|S − µ| ≥ t).
470 CHAPTER 22. INEQUALITIES
and for λ ≤ 0
n
P(Sn ≤ µ − t) ≤ e−λ (µ−t) ∏ E(eλ Xi ). (22.12)
i=1
Note that E(eλ Xi ) in (22.11) and (22.12), likewise E(eλ S ) in (22.9) and (22.10)
are the moment generating functions of the Xi ’s and S, respectively. So finding
bounds boils down to the estimation of these functions.
Now the convexity of ex and 0 ≤ Xi ≤ 1 implies that
eλ Xi ≤ 1 − Xi + Xi eλ .
E(eλ Xi ) ≤ 1 − µi + µi eλ .
The second inequality follows from the fact that the geometric mean is at most
the arithmetic mean i.e. (x1 x2 · · · xn )1/n ≤ (x1 + x2 + · · · + xn )/n for non-negative
x1 , x2 , . . . , xn . This in turn follows from Jensen’s inequlaity and the concavity of
log x.
The right hand side of (22.13) attains its minimum, as a function of λ , at
(µ + t)(n − µ)
eλ = . (22.14)
(n − µ − t)µ
Hence, by (22.13) and(22.14), assuming that µ + t < n,
µ µ+t n − µ n−µ−t
P(Sn ≥ µ + t) ≤ , (22.15)
µ +t n− µ −t
while for t > n − µ this probability is zero.
22.4. SUMS OF INDEPENDENT BOUNDED RANDOM VARIABLES 471
Now let
ϕ(x) = (1 + x) log(1 + x) − x, x ≥ −1,
and let ϕ(x) = ∞ for x < −1. Now, for 0 ≤ t < n − µ, we can rewrite the bound
(22.15) as
−t
t
P(Sn ≥ µ + t) ≤ exp −µϕ − (n − µ)ϕ . (22.16)
µ n−µ
Since ϕ(x) ≥ 0 for every x, we get
x2
ϕ(x) ≥ . (22.20)
2(1 + x/3)
To see this observe that for |x| ≤ 1 we have
x2 ∞
1 1
ϕ(x) − = ∑ (−1)k − xk .
2(1 + x/3) k=2 k(k − 1) 2 · 3k−2
1
Equation (22.20) for |x| ≤ 1 follows from k(k−1) − 2·31k−2 ≥ 0 for k ≥ 2. We leave
it as an exercise to check that (22.20) remains true for x > 1.
Taking this into account we arrive at the following theorem, see Hoeffding [423].
t2
P(Sn ≥ µ + t) ≤ exp − (22.21)
2(µ + t/3)
and for t ≤ µ,
t2
P(Sn ≤ µ − t) ≤ exp − . (22.22)
2(µ − t/3)
472 CHAPTER 22. INEQUALITIES
Putting t = ε µ, for 0 < ε < 1, one can immediately obtain the following bounds.
µε 2
eε µ
P(Sn ≥ (1 + ε)µ) ≤ ≤ exp − , (22.23)
(1 + ε)1+ε 3
while
µε 2
P(Sn ≤ (1 − ε)µ) ≤ exp − (22.24)
2
Proof. The formula (22.24) follows directly from (22.22) and (22.23) follows
from (22.16).
One can “tailor” Chernoff bounds with respect to specific needs. For example,
for small ratios t/µ, the exponent in (22.21) is close to t 2 /2µ, and the following
bound holds.
Corollary 22.8.
2
t3
t
P(Sn ≥ µ + t) ≤ exp − + (22.25)
2µ 6µ 2
2
t
≤ exp − for t ≤ µ. (22.26)
3µ
(µ + t/3)−1 ≥ (µ − t/3)/µ 2 .
2t 2
P(Sn ≥ µ + t) ≤ exp − 2 . (22.28)
c1 + c22 + · · · + c2n
2t 2
P(Sn ≤ µ − t) ≤ exp − 2 . (22.29)
c1 + c22 + · · · + c2n
and so
λ Xi −λ µi µi µi λ ci
E(e )≤e 1− + e
ci ci
= e−θi pi 1 − pi + pi eθi , (22.30)
Now αβ
(α+β )2
≤ 1/4 and so f 00 (θi ) ≤ 1/4 anf therefore
1 λ 2 c2i
f (θi ) ≤ f (0) + f 0 (0)θi + θi2 = .
8 8
474 CHAPTER 22. INEQUALITIES
4
We obtain (22.28) by puttng λ = and (22.29) is proved in a similar manner.
∑ni=1 c2i
t2
P(Sn ≥ t) ≤ exp − (22.31)
2(σ 2 + t/3)
and
t2
P(Sn ≤ −t) ≤ exp − . (22.32)
2(σ 2 + t/3)
Proof. The strategy is once again to bound the moment generating function. Let
∞
λ r−2 E Xir ∞
λ r−2 σi2 eλ − 1 − λ
Fi = ∑ ≤∑ = .
r=2 r!σi2 r=2 r!σi
2 λ2
λ Xi λ r E Xir
∞
E(e ) = 1+ ∑
r=2 r!
= 1 + λ 2 σi2 Fi
2σ 2F
≤ eλ i i
n o
λ 2
≤ exp (e − λ − 1)σi .
So,
n n o
−λt λ 2
P(Sn ≥ t) ≤ e ∏ exp (e − λ − 1)σi
i=1
σ 2 (eλ −λ −1)−λt
=e
n t o
= exp −σ 2 ϕ .
σ2
22.5. SAMPLING WITHOUT REPLACEMENT 475
after assigning t
λ = log 1 + 2 .
σ
To obtain (22.31) we use (22.20). To obtain (22.32) we apply (22.31) to Yi =
−Xi , i = 1, 2, . . . , n.
E f (Wn ) ≤ E f (Sn ).
Proof. We write, where (A)n denotes the set of sequences of n distinct members
of A and (N)n = N(N − 1) · · · (N − n + 1) = |(A)n |,
1
E f (Sn ) = ∑n f (y1 + · · · + yn) =
N n y∈A
1
∑ g(x1, x2, . . . , xn) = E g(X), (22.33)
(N)n x∈(A)
n
g(x) ≥ f (x1 + · · · + xn ).
It follows that
E g(X) ≥ E f (Wn )
and the Lemma follows from (22.33).
As a consequence we have that (i) VarWn ≤ Var Sn and (ii) E eλWn ≤ E eλ Sn for any
λ ∈ R.
Thus all the inequalities developed in Section 22.4 can a fortiori be applied to
Wn in place of Sn . Of particular importance in this context, is the hypergeometric
distribution: Here we are given a set of S ⊆ [N], |S| = m and we choose a random
set X of size k from [N]. Let Z = |X ∩ S|. Then
m N−m
t k−t
P(Z = t) = N
, for 0 ≤ t ≤ k.
k
∆= ∑ E(Ii I j ) = µ + ∆ (22.34)
{i, j}:i∼ j
22.6. JANSON’S INEQUALITY 477
where
∆= ∑ E(Ii I j ). (22.35)
{i, j}:i∼ j
i6= j
Then by the FKG inequality (applied to the random set R and conditioned on
Ii = 1) we get, setting pi = E(Ii ) = ∏s∈Di qs ,
From (22.38) and (22.39), applying Jensen’s inequality to get (22.40) and remem-
bering that µ = ESn = ∑ni=1 pi , we get
ψ 0 (λ )
− (log ψ(λ ))0 = −
ψ(λ )
478 CHAPTER 22. INEQUALITIES
n
≥ ∑ pi E(e−λYi Ii = 1)
i=1
n
pi
≥µ∑ exp − E(λYi Ii = 1)
i=1 µ
( )
1 n
≥ µ exp − ∑ pi E(λYi Ii = 1) (22.40)
µ i=1
( )
λ n
= µ exp − ∑ E(Yi Ii )
µ i=1
= µe−λ ∆/µ .
So
−(log ψ(λ ))0 ≥ µe−λ ∆/µ (22.41)
which implies that
µ2
Z λ
− log ψ(λ ) ≥ µe−z∆/µ dz = (1 − e−λ ∆/µ ). (22.42)
0 ∆
Hence by (22.42) and (22.37)
µ2
log P(Sn ≤ µ − t) ≤ − (1 − e−λ ∆/µ ) + λ (µ − t), (22.43)
∆
which is minimized by choosing λ = − log(1 −t/µ)µ/∆. It yields the first bound
in (22.36), while the final bound in (22.36) follows from the fact that ϕ(x) ≥ x2 /2
for x ≤ 0.
The following Corollary is very useful:
Corollary 22.14 (Janson, Łuczak, Ruciński Inequality).
P(Sn = 0) ≤ e−µ+∆ .
2
n o
Proof. We put t = µ into (22.36) giving P(Sn = 0) ≤ exp − ϕ(−1)µ
∆ . Now note
µ2 µ2
that ϕ(−1) = 1 and ∆ ≥ µ+∆ ≥ µ − ∆.
Note that if D a trivial partition, i.e., D = D0 = {Ω} then P(A|D0 ) = P(A), while,
in general,
P(A) = E P(A|D). (22.44)
Suppose that X is a discrete random variable taking values
{x1 , x2 , . . . , xl } and write X as
l
X= ∑ x j IA j , (22.45)
j=1
Hence,
E(X|D)(ω
1 ) is the expected value of X conditional on the event
ω ∈ Di(ω1 ) .
Suppose that D and D 0 are two partitions of Ω. We say that D 0 is finer than D if
A (D) ⊆ A (D 0 ) and denote this as D ≺ D 0 .
If D is a partition of Ω and Y is a discrete random variable defined on Ω, then
Y is D-measurable if DY ≺ D, i.e., if the partition D is finer than the partition
induced by Y . It simply means that Y takes constant values yi on the atoms Di
of D, so Y can be written as Y = ∑m i=1 yi IDi , where some yi may be equal. Note
that a random variable Y is D0 -measurable if Y has a degenerate distribution, i.e.,
it takes a constant value on all ω ∈ Ω. Also, trivially, the random variable Y is
DY -measurable.
480 CHAPTER 22. INEQUALITIES
Indeed, if ω ∈ Ω then
E(E(X | D 0 ) | D)(ω) =
P(ω 00 ) 0
P(ω )
= ∑ ∑ X(ω 00 )
ω 0 ∈Di(ω) ω 00 ∈D0i(ω 0 )
P(D0i(ω 0 ) ) P(Di(ω) )
P(ω 0 )
= ∑ X(ω 00 ) P(ω 00 ) ∑0
00
ω ∈Di(w) 0
ω ∈Di(ω 00 )
P(D0i(ω 0 ) ) P(Di(ω) )
P(ω 0 )
= ∑ X(ω 00 ) P(ω 00 ) ∑
ω 00 ∈Di(w) ω 0 ∈D0i(ω 00 )
P(D0i(ω 00 ) ) P(Di(ω) )
P(ω 00 )
00
= ∑ X(ω )
ω 00 ∈Di(w)
P(Di(ω) )
= E(X | D)(ω).
Note that despite all the algebra, (22.47) just boils down to saying that the properly
weighted average of averages is just the average.
Finally, suppose a partition D of Ω is induced by a sequence of random variables
{Y1 ,Y2 , . . . ,Yn }. We denote such partition as DY1 ,Y2 ,...,Yn . Then the atoms of this
partition are defined as
where the yi range over all possible values of the Yi ’s. DY1 ,Y2 ,...,Yn is then the coars-
est partition such that Y1 ,Y2 , . . . ,Yn are all constant over the atoms of the partition.
For convenience, we simply write
E(X|Y1 ,Y2 , . . . ,Yn ), instead of E(X|DY1 ,Y2 ,...,Yn ).
D2 ≺ . . . ≺ Dn = D ∗ if
Xk is Dk -measurable
and
(a) E(Xk+1 | Dk ) = Xk k = 0, 1, . . . , n − 1.
(b) E(Xk+1 | Dk ) ≤ Xk k = 0, 1, . . . , n − 1.
(c) E(Xk+1 | Dk ) ≥ Xk k = 0, 1, . . . , n − 1.
We next give upper bounds for both the lower and upper tails of the probability
distributions of certain classes of martingales.
Theorem 22.16 (Azuma-Hoeffding bound). Let {Xk }n0 be a sequence of random
variables such that |Xk − Xk−1 | ≤ ck , k = 1, . . . , n and X0 is constant.
(a) If {Xk }n0 is a super-martingale then for all t > 0 we have
t2
P(Xn ≥ X0 + t) ≤ exp − n 2 .
2 ∑i=1 ci
t2
P(Xn ≤ X0 − t) ≤ exp − n 2 .
2 ∑i=1 ci
t2
P(|Xn − X0 | ≥ t) ≤ 2 exp − n 2 .
2 ∑i=1 ci
Proof. We only need to prove (a), since (b), (c) will then follow easily, since {Xk }n0
is a sub-martingale iff −{Xk }n0 is a super-martingale and {Xk }n0 is a martingale iff
it is a super-martingale and a sub-martingale.
22.7. MARTINGALES. AZUMA-HOEFFDING BOUNDS 483
Yk = Xk − Xk−1 , k = 1, . . . , n.
Then
n
∑ Yk = Xn − X0,
k=1
and
E(Yk+1 | Y0 ,Y1 , . . . ,Yk ) ≤ 0. (22.48)
Let λ > 0. Then
( ) !
n
P (Xn − X0 ≥ t) = P exp λ ∑ Yi ≥ eλt
i=1
( )!
n
≤ e−λt E exp λ ∑ Yi ,
i=1
The expectation in the middle term is over Y0 ,Y1 , . . . ,Yn−1 and the last inequality
follows by induction on n.
484 CHAPTER 22. INEQUALITIES
= ck .
22.8. TALAGRAND’S INEQUALITY 485
dα (A, x) = inf ∑ αi .
y∈A
i:yi 6=xi
Then we define
ρ(A, x) = sup dα (A, x),
|α|=1
At = {x ∈ Ω : ρ(A, x) ≤ t} .
Theorem 22.18.
2 /4
P(At )(1 − P(At )) ≤ e−t .
Lemma 22.20.
ρ(A, x) = min |v|.
v∈V (A,x)
Here |v| denotes the Euclidean norm of v. We leave the proof of this lemma as a
simple exercise in convex analysis.
We now give the proof of Lemma 22.19.
Proof. We use induction on the dimension n. For n = 1, ρ(A, x) = 1x∈A
/ so that
1 2 1
Z
exp ρ (A, x) = P(A) + (1 − P(A))e1/4 ≤
Ω 4 P(A)
If s ∈ V (B, x) and t ∈ V (Aω , x) then (s, 1) and (t, 0) are both in V (A, (x, ω)) and
hence for any λ ∈ [0, 1],
Then,
2 /4 1 1 1 (1−λ )2 /4 −λ
e(1−λ ) · = e r
P(Aω )λ P(B)1−λ P(B)
where x = P(A)/ P(B) ≤ 1. But x(2 − x) ≤ 1, completing the induction and hence
the theorem.
We call h : Ω → R Lipschitz if |h(x) − h(y)| ≤ 1 whenever x, y differ in at most
one coordinate.
Definition 22.21. Let f : N → N. h is f -certifiable if whenever h(x) ≥ s then there
exists I ⊆ [n] with |I| ≤ f (s) so that if y ∈ Ω agrees with x on coordinates I then
h(y) ≥ s.
agree with y on I and p agree with z outside I. By the certification h(y0 ) ≥ b. Now
y0 , z differ in at most t f (b) coordinates and so, by Lipschitz,
but then z ∈
/ A, a contradiction. So, P(X ≥ b) ≤ 1 − P(At ) and from Theorem
22.18,
2
P(X < b − t f (b)) P(X ≥ b) ≤ e−t /4 .
p
2
(b) Suppose that b − t f (b) ≥ m, then P(X ≥ b) ≤ 2e−t /4 .
p
22.9 Dominance
We say that a random variable X stochastically dominates a random variable Y if
There are many cases when we want to use our inequalities to bound the upper
tail of some random variable Y and (i) Y does not satisfy the necessary conditions
to apply the relevant inequality, but (ii) Y is dominated by some random variable
X that does. Clearly, we can use X as a surrogate for Y .
The following case arises quite often. Suppose that Y = Y1 +Y2 + · · · +Yn where
Y1 ,Y2 , . . . ,Yn are not independent, but instead we have that for all t in the range
[Ai , Bi ] of Yi ,
P(Yi ≥ t | Y1 ,Y2 , . . . ,Yi−1 ) ≤ Φ(t)
where Φ(t) decreases monotonically from 1 to 0 in [Ai , Bi ].
Let Xi be a random variable taking values in the same range as Yi and such that
P(Xi ≥ t) = Φ(t). Let X = X1 + · · · + Xn where X1 , X2 , . . . , Xn are independent of
each other and Y1 ,Y2 , . . . ,Yn . Then we have
22.9. DOMINANCE 489
nλ 3
λ = o(1) and α = 1.
β3
491
492 CHAPTER 23. DIFFERENTIAL EQUATIONS METHOD
where |ρ| ≤ (2L + 1)λ , since |θk | ≤ λ (by (P3)) and |ψk | ≤ Lβn k (by (P4)).
Now, given Ht , let
(
X(t + k) − X(t) − k f nt , X(t)
n − (2L + 1)kλ E
Zk = .
emptyset ¬E
493
Then
E(Zk+1 − Zk |Z0 , Z1 , . . . , Zk ) ≤ 0,
i.e., Z0 , Z1 , . . . , Zω is a super-martingale.
Also
t X(t)
|Zk+1 − Zk | ≤ β + f
, + (2L + 1)λ ≤ K0 β ,
n n
where K0 = O(1), since f nt , X(t) n = O(1) by continuity and boundedness of D.
So, using Theorem 22.16 we see that conditional on Ht , E ,
√
P X(t + ω) − X(t) − ω f (t/n, X(t)/n) ≥ (2L + 1)ωλ + K0 β 2αω
2K02 β 2 αω
≤ exp − 2 2
= e−α . (23.4)
2ωK0 β
Similarly,
√
P X(t + ω) − X(t) − ω f (t/n, X(t)/n) ≤ −(2L + 1)ωλ − K0 β 2αω
≤ e−α . (23.5)
Thus
√
P |X(t + ω) − X(t) − ω f (t/n, X(t)/n)| ≥ (2L + 1)ωλ + K0 β 2αω
≤ 2e−α .
√
We have that ωλ and β 2αω are both Θ(nλ 2 /β ) giving
√ nλ 2
(2L + 1)ωλ + K0 β 2αω ≤ K1 .
β
Now let ki = iω for i = 0, 1, . . . , i0 = bσ n/ωc. We will show by induction that
where j+1 !
Lω nλ
Bj = B 1+ −1 (23.7)
n L
and where B is another constant.
X(0)
The induction begins with z(0) = n and B0 = 0. Note that
Beσ L λ
Bi0 ≤ n = O(λ n). (23.8)
L
494 CHAPTER 23. DIFFERENTIAL EQUATIONS METHOD
Now write
z(ki+1 /n)
A3
A4
slope z0 (ki /n)
z(ki /n)
slope f (ki .n, X(ki )/n)
A1
X(ki+1 )/n
|A1 | ≤ Bi .
nλ 2
|A2 | ≤ K1 ,
β
Now
ω 0
z(ki+1 /n) − z(ki /n) =z (ki /n + ω̂/n)
n
for some 0 ≤ ω̂ ≤ ω and so (P4) implies that
0 0 ω2 nλ 2
|A3 | = ω|z (ki /n + ω/n) − z (ki /n + ω̂/n)| ≤ L ≤ 2L 2 .
n β
Finally, (P4) gives
ωL|A1 | ωL
|A4 | ≤ ≤ Bi .
n n
Thus for some B > 0,
Branching Processes
pk = P(X = k), k = 0, 1, 2, . . . .
Let
∞
G(z) = ∑ pk zk
k=0
be the probability generating function (p.g..f.) of X. Let µ = E X. Let
!
[
η =P {Zn = 0} (24.1)
n≥0
Proof. If Gn (z) is the p.g.f. of Zn , then Gn (z) = G(Gn−1 (z)). This follows from
the fact that Zn is the sum of Zn−1 independent copies of G. Let ηn = P(Zn = 0).
Then
ηn = Gn (0) = G(Gn−1 (0)) = G(ηn−1 ).
It follows from (24.1) that ηn % η. Let ψ be any other non-negative solution to
G(s) = s. We have
η1 = G(0) ≤ G(ψ) = ψ.
497
498 CHAPTER 24. BRANCHING PROCESSES
Entropy
We have a choice for the base of the logarithm here. We use the natural logarithm,
for use in Chapter 14.
Note that if X is chosen uniformly from RX , i.e. P(X = x) = 1/|RX | for all x ∈ RX
then then
log |RX |
h(X) = ∑ = log |RX |.
x∈RX |RX |
where Xy is the random variable with P(Xy = x) = P(X = x | Y = y). Here p(y) =
P(Y = y). The summation is over y such that p(y) > 0. We will use notation like
this from now on, without comment.
Chain Rule:
499
500 CHAPTER 25. ENTROPY
Lemma 25.1.
m
h(X1 , X2 , . . . , Xm ) = ∑ h(Xi | X1 , X2 , . . . , Xi−1 ). (25.2)
i=1
Now,
p(x1 , x2 )
h(X2 | X1 ) = − ∑ p(x1 , x2 ) log
x1 ,x2 p(x1 )
=− ∑ p(x1 , x2 ) log p(x1 , x2 ) + ∑ p(x1 , x2 ) log p(x1 )
x1 ,x2 x1 ,x2
= h(X1 , X2 ) + ∑ p(x1 ) log p(x1 )
x1
= h(X1 , X2 )−h(X1 ).
Inequalities:
Entropy is a measure of uncertainty and so we should not be surprised to learn that
h(X | Y ) ≤ h(X) for all random variables X,Y – here conditioning on Y represents
providing information. Our goal is to prove this and a little more.
Let p, q be probability measures on the finite set X. We define the
Kullback-Liebler distance
p(x)
D(p||q) = ∑ p(x) log q(x)
x∈A
= log 1
= 0.
Inequality (25.3) follows from Jensen’s inequality and the fact that log is a concave
function. Because log is strictly concave, will have equality in (25.3) iff p = q.
It follows from this that
h(X) ≤ log |RX |. (25.4)
Indeed, let u denote the uniform distribution over RX i.e. u(x) = 1/|RX |. Then
h(X | Y, Z) ≤ h(X | Y ).
h(X | Z) ≤ h(X).
Proof.
h(X | Y ) − h(X | Y, Z)
p(x, y) p(x, y, z)
= − ∑ p(x, y) log + ∑ p(x, y, z) log
x,y p(y) x,y,z p(y, z)
p(x, y) p(x, y, z)
= − ∑ p(x, y, z) log + ∑ p(x, y, z) log
x,y,z p(y) x,y,z p(y, z)
p(x, y, z)p(y)
= ∑ p(x, y, z) log p(x, y)p(y, z)
x,y,z
= D(px,y,z ||p(x, y)p(y, z)/p(y))
≥ 0.
Working through the above proof we see that h(X) = h(X | Z) iff p(x, z) = p(x)p(z)
for all x, z, i.e. iff X, Z are independent.
502 CHAPTER 25. ENTROPY
and
h(XAi ) = ∑ h(X j | X`, ` ∈ Ai, ` < j). (25.6)
j∈Ai
Here we obtain (25.8) from (25.7) by applying Lemma 25.3. We obtain (25.9)
from (25.8) and the fact that each j ∈ B appears in at least k Ai ’s. We then obtain
(25.10) by using (25.5).
Bibliography
[2] M. Abdullah, C.Cooper and A.M. Frieze, Cover time of a random graph
with given degree sequence, Discrete Mathematics 312 (2012) 3146-3163.
[4] D. Achlioptas and C. Moore, Almost all Graphs with Average Degree 4 are
3-colorable, Journal of Computer and System Sciences 67 (2003) 441-471.
[6] D. Achlioptas and A. Naor, The two possible values of the chromatic num-
ber of a random graph, Annals of Mathematics 162 (2005) 1335-1351.
[11] M. Ajtai, J. Komlós and E. Szemerédi, The longest path in a random graph,
Combinatorica 1 (1981) 1-12.
503
504 BIBLIOGRAPHY
[16] D. Aldous, Brownian excursions, critical random graphs and the multiplica-
tive coalescent, Annals of Probability 25 (1997) 812-854.
[17] D. Aldous, The ζ (2) limit in the random assignment problem, Random
Structures and Algorithms 4 (2001) 381-418.
[26] N. Alon and Z. Füredi, Spanning subgraphs of random graphs, Graphs and
Combinatorics 8 (1992) 91-94.
[31] N. Alon and J. Spencer, The probabilistic method, Third Edition, John Wi-
ley and Sons (2008).
[34] A. Amit and N. Linial, Random Graph Coverings I: General Theory and
Graph Connectivity, Combinatorica 22 (2002), 1-18.
[35] A. Amit and N. Linial, Random Lifts of Graphs II: Edge Expansion, Com-
binatorics Probability and Computing 15 (2006) 317-332.
[36] A. Amit, N. Linial and J. Matoušek, Random Lifts of Graphs III: Inde-
pendence and Chromatic Number, Random Structures and Algorithms 20
(2002) 1-22.
[40] J. Aronson, A.M. Frieze and B. Pittel, Maximum matchings in sparse ran-
dom graphs: Karp-Sipser revisited, Random Structures and Algorithms 12
(1998) 111-178.
[41] R. Arratia, A.D. Barbour and S. Tavaré, A tale of three couplings: Poisson-
Dirichlet and GEM approximations for random permutations, Combina-
torics, Probability and Computing 15 (2006) 31-62.
[42] R. Arratia and S. Tavaré, The cycle structure of random permutations, The
Annals of Probability 20 (1992) 1567-1591.
[43] C. Avin and G. Ercal, On the Cover Time of Random Geometric Graphs,
Automata, Languages and Programming, Lecture Notes in Computer Sci-
ence 3580 (2005) 677-689.
[44] L.Babai, P. Erdős, S.M. Selkow, Random graph isomorphism, SIAM Jour-
nal on Computing 9 (1980) 628-635.
[46] A. Backhaus and T.F. Móri, Local degree distribution in scale free random
graphs, Electronic Journal of Probability 16 (2011) 1465-1488.
[47] D. Bal, P. Bennett, A. Dudek and A.M. Frieze, The t-tone chromatic num-
ber of random graphs, see arxiv.org.
[48] D. Bal, P. Bennett, A.M. Frieze and P. Pralat, Power of k choices and rain-
bow spanning trees in random graphs, see arxiv.org.
[50] D. Bal and A.M. Frieze, Rainbow Matchings and Hamilton Cycles in Ran-
dom Graphs, see arxiv.org.
[59] A.D Barbour, Poisson convergence and random graph, Math. Proc. Camb.
Phil. Soc. 92 (1982) 349-359.
[60] A.D. Barbour, L. Holst and S. Janson, Poisson Approximation, 1992, Ox-
ford University Press, Oxford.
[62] A.D. Barbour, M. Karoński and A. Ruciński, A central limit theorem for de-
composable random variables with applications to random graphs, J. Com-
bin. Theory B 47 (1989) 125-145.
[63] A.D. Barbour and G. Reinert, The shortest distance in random multi-type
intersection graphs, Random Structures Algorithms 39 (2011) 179-209.
[64] A.D. Barbour and G. Reinert, Small worlds, Random Structures and Algo-
rithms 19(2001) 5474.
[69] M. Bednarska and T. Łuczak, Biased positional games for which random
strategies are nearly optimal, Combinatorica 20 (2000) 477-488.
[72] I. Benjamini, G. Kozma and N. Wormald, The mixing time of the gi-
ant component of a random graph, Random Structures and Algorithms 45
(2014) 383-407.
[74] E.A. Bender and E.R. Canfield, The asymptotic number of labelled graphs
with given degree sequences, Journal of Combinatorial Theory A 24 (1978)
296-307.
[75] E.A. Bender, E.R. Canfield and B.D. McKay, The asymptotic number of
labelled connected graphs, Random Structures and Algorithms 1 (1990)
127-169.
[76] E.A. Bender and E.R. Canfield, The number of rooted maps on an ori-
entable surface. Journal of Combinatorial Theory B 53 (1991) 293-299.
[78] P. Bennett and T. Bohman, A note on the random greedy independent set
algorithm, see arxiv.org.
[94] S.R. Blackbourn and S. Gerke, Connectivity of the uniform random inter-
section graph, Discrete Mathematics 309 (2009) 5130-5140.
[110] M. Bode, N. Fountoulakis and T. Müller, The probability that the hyper-
bolic random graph is connected, see arxiv.org.
[117] T. Bohman and A.M. Frieze, Avoiding a giant component, Random Struc-
tures and Algorithms (2001) 75-85.
[118] T. Bohman and A.M. Frieze, Hamilton cycles in 3-out, Random Structures
and Algorithms 35 (2009) 393-417.
[119] T. Bohman, A.M. Frieze and E. Lubetzky, A note on the random greedy
triangle-packing algorithm, Journal of Combinatorics 1 (2010) 477-488.
512 BIBLIOGRAPHY
[120] T. Bohman, A.M. Frieze and E. Lubetzky, Random triangle removal, see
arxiv.org.
[122] T. Bohman, A.M. Frieze and B.Sudakov, The game chromatic number of
random graphs, Random Structures and Algorithms 32 (2008) 223-235.
[135] B. Bollobás, Random Graphs, First Edition, Academic Press, London 1985,
Second Edition, Cambridge University Press, 2001.
[139] B. Bollobás, P. Catlin and P. Erdős, Hadwigers conjecture is true for almost
every graph, European Journal of Combinatorics 1 (1980) 195-199.
[140] B. Bollobás and F. Chung, The diameter of a cycle plus a random matching,
SIAM Journal on Discrete Mathematics 1 (1988) 328-333.
[141] B. Bollobás, C. Cooper, T.I. Fenner and A.M. Frieze, On Hamilton cycles
in sparse random graphs with minimum degree at least k, Journal of Graph
Theory 34 (2000) 42-59.
[143] B. Bollobás, T.I. Fenner and A.M. Frieze, An algorithm for finding hamil-
ton paths and cycles in random graphs, Combinatorica 7 (1987) 327-341.
[144] B. Bollobás, T.I. Fenner and A.M. Frieze, Hamilton cycles in random
graphs with minimal degree at least k, in A tribute to Paul Erdos, edited
by A.Baker, B.Bollobás and A.Hajnal (1990) 59-96.
[145] B. Bollobás, T.I. Fenner and A.M. Frieze, Long cycles in sparse ran-
dom graphs, In Graph Theory and Combinatorics, Proceedings, Cam-
bridge Combinatorial Conference in honour of Paul Erdős (B. Bollobás
Ed.) (1984) 59-64.
[154] B. Bollobás and O. Riordan, The diameter of a scale free random graph,
Combinatorica 24 (2004) 5-34.
[163] G.R. Brightwell and M.J. Luczak Vertices of high degree in the preferential
attachment tree, Elect. J. Probab. 17 (2012) 43pp.
[167] A. Broder, A.M. Frieze and E. Shamir, Finding hidden Hamilton cycles,
Random Structures and Algorithms 5 (1994) 395-410.
[168] A. Broder, A.M. Frieze, S. Suen and E. Upfal, An efficient algorithm for
the vertex-disjoint paths problem in random graphs, Proceedings of SODA
’96, 261-268.
[174] Y.D. Burtin, On a simple formula for a random mappings and its applica-
tions, J. Applied Probability 17 (1980) 403-414.
[175] D.S. Callaway, J.E. Hopcroft, J.M. Kleinberg, M.E.J. Newman and S.H.
Strogatz, Are randomly grown graphs really random? Phys.Rev. E 64
(2001) 041902.
[178] J. Cerny and A. Teixeira, Critical window for the vacant set left by random
walk on random regular graphs, see arxiv.org.
[181] P. Chebolu and A.M. Frieze, Hamilton cycles in random lifts of complete
directed graphs, SIAM Journal on Discrete Mathematics 22 (2008) 520-
540.
[183] L.H.Y. Chen, Poisson approximation for dependent trials, Annals of Prob-
ability 3 (1975) 534-545.
[184] W-C. Chen and W-C. Ni, Heap-ordered trees, 2-partitions and continued
fractions, European Journal of Combinatorics 15 (1994) 513-517.
[187] F. Chung and L. Lu, Connected components in random graphs with given
expected degree sequence Annals of Combinatorics, 6 (2002) 125-145.
[188] F. Chung and L. Lu. The average distances in random graphs with given
expected degree, Internet Math. 1 (2003) 91-113.
[189] F. Chung and L. Lu, The volume of the giant component of a random graph
with given expected degrees, SIAM Journal on Discrete Mathematics 20
(2006) 395-411.
[190] F. Chung and L. Lu, Coupling on-line and off-line analyses for random
power law graphs, Internet Mathematics 1 (2004) 409-461.
BIBLIOGRAPHY 517
[191] F. Chung and L. Lu. Complex Graphs and Networks, American Mathemat-
ical Society, Providence, RI, 2006.
[192] F. Chung, L. Lu and V. Vu, Eigenvalues of random power law graphs, An-
nals of Combinatorics 7 (2003) 21-33.
[193] F. Chung, L. Lu and V. Vu, The spectra of random graphs with expected de-
grees, Proceedings of the National Academy of Sciences 100 (2003) 6313-
6318.
[195] V. Chvátal, Almost all graphs with 1.44n edges are 3-colorable, Random
Structures and Algorithms 2 (1991) 11-28.
[196] V. Chvátal and P. Erdős, Biased positional games, Annals of Discrete Math-
ematics 2 (1978) 221-228.
[202] N. A. Cook and A. Dembo, Large deviations of subgraph counts for sparse
Erdő s–Rényi graphs, see arxiv.org.
[204] P. Condon, A. Espuny Dáz, J. Kim, D. Kühn and D. Osthus, Resilient de-
gree sequences with respect to Hamilton cycles and matchings in random
graphs, see arxiv.org.
518 BIBLIOGRAPHY
[205] O. Cooley, M. Kang and C. Koch, The size of the giant component in ran-
dom hypergraphs, see arxiv.org.
[210] C. Cooper, The size of the cores of a random graph with a given degree
sequence, Random Structures and Algorithms (2004) 353-375.
[212] C. Cooper and A.M. Frieze, On the number of hamilton cycles in a random
graph, Journal of Graph Theory 13 (1989) 719-735.
[213] C. Cooper and A.M. Frieze, Pancyclic random graphs, in Random Graphs,
(Ed: Karonski M, Javorski J, and Rucinski A.) Wiley (1990).
[214] C. Cooper and A.M. Frieze, The limiting probability that a-in, b-out is
strongly connected, Journal of Combinatorial Theory B 48 (1990) 117-
134.
[216] C. Cooper and A.M. Frieze, On the connectivity of random k-th nearest
neighbour graphs, Combinatorics, Probability and Computing 4 (196) 343-
362.
[217] C. Cooper and A.M. Frieze, Hamilton cycles in random graphs and directed
graphs, Random Structures and Algorithms 16 (2000) 369-401.
[219] C. Cooper and A.M. Frieze, On a general model of web graphs Random
Structures and Algorithms 22 (2003) 311-335.
[220] C. Cooper and A.M. Frieze, The Size of the Largest Strongly Connected
Component of a Random Digraph with a Given Degree Sequence, Combi-
natorics, Probability and Computing 13 (2004) 319-337.
[221] C. Cooper and A.M. Frieze, The cover time of random regular graphs,
SIAM Journal on Discrete Mathematics 18 (2005) 728-740.
[222] C. Cooper and A.M. Frieze, The cover time of the preferential attachment
graph, Journal of Combinatorial Theory B 97 (2007) 269-290.
[223] C. Cooper and A.M. Frieze, The cover time of the giant component of a
random graph, Random Structures and Algorithms 32 (2008) 401-439.
[224] C. Cooper and A.M. Frieze, The cover time of random geometric graphs,
Random Structures and Algorithms 38 (2011) 324-349.
[225] C. Cooper and A.M. Frieze, Stationary distribution and cover time of ran-
dom walks on random digraphs, Journal of Combinatorial Theory B 102
(2012) 329-362.
[227] C. Cooper and A.M. Frieze, Long paths in random Apollonian networks,
see arxiv.org.
[228] C. Cooper, A.M. Frieze, N. Ince, S. Janson and J. Spencer, On the length
of a random minimum spanning tree, see arxiv.org.
[230] C.Cooper, A.M. Frieze and E. Lubetzky, Cover time of a random graph
with given degree sequence II: Allowing vertices of degree two, Random
Structures and Algorithms 45 (2014) 627-674.
[231] C.Cooper, A.M. Frieze and M. Molloy, Hamilton cycles in random regular
digraphs, Combinatorics, Probability and Computing 3 (1994) 39-50.
520 BIBLIOGRAPHY
[232] C. Cooper, A,M, Frieze and P. Pralat, Some typical properties of the Spatial
Preferred Attachment model, Internet Mathematics 10 (2014), 27-47.
[234] C.Cooper, A.M. Frieze and T. Radzik, The cover time of random walks on
random uniform hypergraphs, Proceeedings of SIROCCO 2011, 210-221.
[235] C. Cooper, A.M. Frieze and B. Reed, Random regular graphs of non-
constant degree: connectivity and Hamilton cycles, Combinatorics, Proba-
bility and Computing 11 (2002) 249-262.
[236] C. Cooper, A.M. Frieze and J. Vera, Random deletions in a scale free ran-
dom graph process, Internet Mathematics 1 (2004) 463-483.
[239] V. Dani and C. Moore, Independent sets in random graphs from the
weighted second moment method, Proceedings of RANDOM 2011 (2011)
472-482.
[240] A. Darrasse, H-K. Hwang and M. Soria, Shape measures of random in-
creasing k-trees, manuscript (2013).
[241] M. Deijfen and W. Kets, Random intersection graphs with tunable degree
distribution and clustering, Probab. Engrg. Inform. Sci. 23 (2009) 661-674.
[243] B. DeMarco and J. Kahn, Tight upper tail bounds for cliques, Random
Structures and Algorithms 41 (2012)
[244] B. DeMarco and J. Kahn, Mantel’s theorem for random graphs, see
arxiv.org.
[245] B. DeMarco and J. Kahn, Turán’s theorem for random graphs, see arxiv.org.
BIBLIOGRAPHY 521
[249] L. Devroye and H-K. Hwang, Width and mode of the profile for some
random trees of logarithmic height, The Annals of Applied Probability 16
(2006) 886-918.
[250] L. Devroye and S. Janson, Protected nodes and fringe subtrees in some
random trees, Electronic Communications in Probability 19 (2014) 1-10.
[251] L. Devroy and J. Lu, The strong convergence of maximal degrees in uni-
form random recursive trees and dags, Random Structure and Algorithms 7
(1995) 1-14.
[255] J. Dı́az, M.D. Penrose, J. Petit and M.J. Serna, Approximating layout prob-
lems on random geometric graphs, Journal of Algorithms 39 (2001) 78-116.
[256] J. Ding, J.H. Kim, E. Lubetzky and Y. Peres, Anatomy of a young giant
component in the random graph, Random Structures and Algorithms 39
(2011) 139-178.
[259] R.P. Dobrow and R.T. Smythe, Poisson approximations for functionals of
random trees, Random Structures and Algorithms 9 (1996) 79-92.
[260] M. Dondajewski and J. Szymański, On a generalized random recursive tree,
unpublished manuscript.
[261] M. Drmota, Random Trees, Springer, Vienna, 2009.
[262] M. Drmota and H-K. Hwang, Profiles of random trees: Correlation and
width of random recursive trees and binary search trees, Adv. in Appl.
Probab. 37 (2005) 321-341.
[263] A. Dudek and A.M. Frieze, Loose Hamilton Cycles in Random k-Uniform
Hypergraphs, Electronic Journal of Combinatorics (2011) P48.
[264] A. Dudek and A.M. Frieze, Tight Hamilton Cycles in Random Uniform
Hypergraphs, Random structures and Algorithms 42 (2012) 374-385.
[265] A. Dudek, A.M. Frieze, P. Loh and S. Speiss Optimal divisibility conditions
for loose Hamilton cycles in random hypergraphs, Electronic Journal of
Combinatorics 19 (2012).
[266] A. Dudek, A.M. Frieze, A. Ruciński and M. Šileikis, Loose Hamilton Cy-
cles in Regular Hypergraphs, Combinatorics, Probability and Computing
24 (2015) 179-194.
[267] A. Dudek, A.M. Frieze, A. Ruciński and M. Šilekis, Embedding the Erdős-
Rényi Hypergraph into the Random Regular Hypergraph and Hamiltonic-
ity, see arxiv.org.
[268] A. Dudek, A.M. Frieze and C. Tsourakakis, Rainbow connection of random
regular graphs, see arxiv.org.
[269] A. Dudek, D. Mitsche and P. Pralat, The set chromatic number of random
graphs, see arxiv.org.
[270] A. Dudek and P. Pralat, An alternative proof of the linearity of the size-
Ramsey number of paths, to appear in Combinatorics, Probability and
Computing.
[271] A. Dudek, P. Pralat, Acquaintance time of random graphs near connectivity
threshold, see arxiv.org.
[272] R. Durrett, Rigoous result for CHKNS random graph model, In: Proc. Dis-
crete Random Walks (Paris 2003), Discrete Mathematics and Theoretical
Computer Science AC (2003) 95-104.
BIBLIOGRAPHY 523
[284] G.P. Egorychev, A solution of the Van der Waerden’s permanent problem,
Preprint IFSO-L3 M Academy of Sciences SSSR, Krasnoyarks (1980).
[286] P. Erdős and A. Rényi, On random graphs I, Publ. Math. Debrecen 6 (1959)
290-297.
524 BIBLIOGRAPHY
[287] P. Erdős and A. Rényi, On the evolution of random graphs, Publ. Math.
Inst. Hungar. Acad. Sci. 5 (1960) 17-61.
[289] P. Erdős and A. Rényi, On random matrices, Publ. Math. Inst. Hungar.
Acad. Sci. 8 (1964) 455-461.
[293] P. Erdős and J. Spencer, Evolution of the n-cube, Computers & Mathemat-
ics with Applications 5 (1979) 33-39.
[294] P. Erdős, S. Suen and P. Winkler, On the size of a random maximal graph,
Random Structures and Algorithms 6 (1995) 309-318.
[295] L. Eschenauer and V.D. Gligor, A key managment scheme for distributed
sensor networks, In: Proceedings of the 9th ACM Conference on Computer
and Communication Security (2002) 41-47.
[296] H. van den Esker, A geometric preferential attachment model with fitness,
see arxiv.org.
[298] D.I. Falikman, The proof of the Van der Waerden’s conjecture regarding to
doubly stochastic matrices, Mat. Zametki 29 (1981).
[302] Q. Feng, H.M. Mahmoud and A. Panholzer, Phase changes in subtree va-
rieties in random recursive and binary search trees, SIAM Journal Discrete
Math. 22 (2008) 160-184.
[309] T.I. Fenner and A.M. Frieze, On the connectivity of random m-orientable
graphs and digraphs, Combinatorica 2 (1982) 347-359.
[311] J.A. Fill, E.R. Scheinerman and K.B. Singer-Cohen, Random intersection
graphs when m = ω(n): An equivalence theorem relating the evolution the
evolution of the G(n, m, p) and G(n, p) models, Random Structures and
Algorithms 16 (2000) 156-176.
[313] G. Fiz Pontiveros, S. Griffiths and R. Morris, The triangle-free process and
R(3, k), see arxiv.org.
[314] A. Flaxman, The Lower Tail of the Random Minimum Spanning Tree, The
Electronic Journal of Combinatorics 14 (2007) N3.
526 BIBLIOGRAPHY
[315] A. Flaxman, A.M. Frieze and T. Fenner, High degree vertices and eigen-
values in the preferential attachment graph, Internet Mathematics 2 (2005)
1-20.
[319] A. Flaxman, D. Gamarnik and G.B. Sorkin, Embracing the giant compo-
nent, Random Structures and Algorithms 27 (2005) 277-289.
[323] N. Fountalakis and K. Panagiotou, Sharp load thresholds for cuckoo hash-
ing, Random Structures and Algorithms 41 (2012) 306-333.
[325] N. Fountoulakis, D. Kühn and D. Osthus, The order of the largest complete
minor in a random graph, Random Structures and Algorithms 33 (2008)
127-141.
[329] E. Friedgut and G. Kalai, Every Monotone Graph Property Has a Sharp
Threshold, Proceedings of the American Mathematical Society 124 (1996)
2993-3002.
[330] E. Friedgut, Sharp Thresholds of Graph Proprties, and the k-sat Problem,
Journal of the American Mathematical Society 12 (1999) 1017-1054.
[334] A.M. Frieze, On the value of a random minimum spanning tree problem,
Discrete Applied Mathematics 10 (1985) 47-56.
[335] A.M. Frieze, On large matchings and cycles in sparse random graphs, Dis-
crete Mathematics 59 (1986) 243-256.
[337] A.M. Frieze, An algorithm for finding hamilton cycles in random digraphs,
Journal of Algorithms 9 (1988) 181-204.
[340] A.M. Frieze, Perfect matchings in random bipartite graphs with minimal
degree at least 2, Random Structures and Algorithms 26 (2005) 319-358.
[344] A.M. Frieze and S. Haber, An almost linear time algorithm for finding
Hamilton cycles in sparse random graphs with minimum degree at least
three, to appear in Random Structures and Algorithms.
[345] A.M. Frieze, S. Haber and M. M. Lavrov, On the game chromatic number
of sparse random graphs, SIAM Journal of Discrete Mathematics 27 (2013)
768-790.
[347] A.M. Frieze and T. Johansson, On random k-out sub-graphs of large graphs,
see arxiv.org.
[348] A.M. Frieze and R. Kannan, A new approach to the planted clique prob-
lem, Proceedings of Foundations of Software Technology and Theoretical
Computer Science, Bangalore, India (2008).
[349] A.M. Frieze and M. Krivelevich, On two Hamilton cycle problems in ran-
dom graphs, Israel Journal of Mathematics 166 (2008) 221-234.
[351] A.M. Frieze, S. Haber and M. Lavrov, On the game chromatic number of
sparse random graphs, SIAM Journal of Discrete Mathematics 27 (2013)
768-790.
[352] A.M. Frieze and P. Loh, Rainbow Hamilton cycles in random graphs, Ran-
dom Structures and Algorithms 44 (2014) 328-354.
[353] A.M. Frieze and T. Łuczak, Hamiltonian cycles in a class of random graphs:
one step further, in em Proceedings of Random Graphs ’87, Edited by
M.Karonski, J.Jaworski and A.Rucinski, John Wiley and Sons (1990) 53-
59.
BIBLIOGRAPHY 529
[354] A.M. Frieze and T. Łuczak, Edge disjoint trees in random graphs, Periodica
Mathematica Hungarica 21 (1990) 28-30.
[355] A.M. Frieze and T. Łuczak, On the independence and chromatic numbers
of random regular graphs, Journal of Combinatorial Theory 54 (1992) 123-
132.
[356] A.M. Frieze and C.J.H. McDiarmid, On random minimum length spanning
trees, Combinatorica 9 (1989) 363 - 374.
[357] A.M. Frieze and B.D. McKay, Multicoloured trees in random graphs, Ran-
dom Structures and Algorithms 5 (1994) 45-56.
[360] A.M. Frieze and W. Pegden, Between 2- and 3-colorability, see arxiv.org.
[361] A.M. Frieze and W. Pegden, Looking for vertex number one, see arxiv.org.
[362] A.M. Frieze and B. Pittel, Perfect matchings in random graphs with pre-
scribed minimal degree, Trends in Mathematics, Birkhauser Verlag, Basel
(2004) 95-132.
[364] A.M. Frieze and G. Sorkin, Efficient algorithms for three-dimensional ax-
ial and planar random assignment problems, Random Structures and Algo-
rithms 46 (2015) 160-196.
[365] A.M. Frieze and C. Tsourakakis, High Degree Vertices, Eigenvalues and
Diameter of Random Apollonian Networks, WAW2012 (2012).
[369] M. Fuchs, H-K. Hwang, R. Neininger, Profiles of random trees: Limit the-
orems for random recursive trees and binary search trees,Algorithmica 46
(2006) 367-407.
[371] D. Gale and L. S. Shapley, College Admissions and the Stability of Mar-
riage, American Mathematical Monthly 69 (1962) 9-14.
[372] P. Gao and M. Molloy, The stripping process can be slow, see arxiv.org.
[376] H. Gebauer and T. Szabó, Asymptotic random graph intuition for the biased
connectivity game, Random Structures and Algorithms 35 (2009) 431-443.
[379] S. Gerke, D. Schlatter, A. Steger and A. Taraz, The random planar graphs
process, Random Structures and Algorithms 32 (2008) 236-261.
[383] O. Giménez and M. Noy, Asymptotic enumeration and limit laws of planar
graphs, Journal of the American Mathematical Society 22 (2009) 309-329.
[385] R. Glebov, H. Naves and B. Sudakov, The threshold probability for cycles,
see arxiv.org.
[386] Y.V. Glebskii, D.I. Kogan, M.I. Liagonkii and V.A. Talanov, Range and
degree of realizability of formulas in the restricted predicate calculus, Cy-
bernetics 5 (1969) 142-154.
[391] W. Goh, E. Schmutz, Limit distribution for the maximum degree of a ran-
dom recursive tree, Journal of Computing and Applied Mathematics 142
(2002) 61-82.
[392] Ch. Goldschmidt and J.B. Martin, Random recursive trees and the
Bolthausen-Sznitman coalescent, Electronic Journal of Probability 10
(2005) 718-745.
[394] V.L. Goncharov, Sur la distribution des cycles dans les permutations C.R.
Acad. Sci USSR (N.S.) 35, (1942) 267-269.
532 BIBLIOGRAPHY
[395] V.L. Goncharov, Some facts in combinatorics, Izv. Acad. Nauk SSSR, Ser.
Mat. 8 (1944) 3-48 (in Russian). [See also: On the field of combinatory
analysis, Trans. Amer. Math. Soc. 19 (1962) 1-46.]
[399] G.R. Grimmett, Random labelled trees and their branching networks, Jour-
nal of the Australian Mathematical Society 30 (1980) 229-237.
[403] P. Gupta and P.R. Kumar, Critical Power for Asymptotic Connectivity in
Wireless Networks, in Stochastic Analysis, Control, Optimization and Ap-
plications: A Volume in Honor of W. H. Fleming, W. M. McEneany, G. Yin,
and Q. Zhang, Eds. Boston, MA: Birkhauser, (1998) 547-566.
[406] S. Haber and M. Krivelevich, The logic of random regular graphs, Journal
of Combinatorics 1 (2010) 389-440.
[409] A. Hamm and J. Kahn, On Erdős-Ko-Rado for random hypergraphs II, see
arxiv.org.
[410] J.C. Hansen and J. Jaworski, Random mappings with exchangeable in-
degrees, Random Structures and Algorithms 33 (2008) 105-126.
[411] J.C. Hansen and J. Jaworski, Predecessors and successors in random map-
pings with exchangeable in-degrees, J. Applied Probability 50 (2013) 721-
740.
[412] T. Harris, A lower bound for the critical probability in a certain percolation,
Proceedings of the Cambridge Philosophical Society 56 (1960) 13-20.
[413] H. Hatami, Random cubic graphs are not homomorphic to the cycle of size
7, Journal of Combinatorial Theory B 93 (2005) 319-325.
[414] H. Hatami and M. Molloy, The scaling window for a random graph with a
given degree sequence, Random Structures and Algorithms 41 (2012) 99-
123.
[418] A. Heckel and O. Riordan, The hitting time of rainbow connection number
two, Electronic Journal on Combinatorics 19 (2012).
[420] D. Hefetz, M. Krivelevich and T. Szabo, Sharp threshold for the appear-
ance of certain spanning trees in random graphs, Random Structures and
Algorithms 41 (2012) 391-412.
[421] D. Hefetz, A. Steger and B. Sudakov, Random directed graphs are robustly
Hamiltonian, Random Structures and Algorithms 49 (2016) 345-362.
534 BIBLIOGRAPHY
[422] D. Hefetz, A. Steger and B. Sudakov, Random directed graphs are robustly
hamiltonian, see arxiv.org.
[423] W. Hoeffding, Probability inequalities for sums of bounded random vari-
ables, Journal of the American Statistical Association 58 (1963) 13-30.
[424] R. van der Hofstad, Random graphs and complex networks: Volume 1, see
arxiv.org.
[425] R. van der Hofstad, Critical behaviour in inhomogeneous random graphs,
Random Structures and Algorithms 42 (2013) 480-508.
[426] R. van der Hofstad, G. Hooghiemstra and P. Van Mieghem, On the covari-
ance of the level sizes in random recursive trees, Random Structures and
Algorithms 20 (2002) 519-539.
[427] R. van der Hofstad, S. Kliem and J. van Leeuwaarden, Cluster tails for
critical power-law inhomogeneous random graphs, see arxiv.org
[428] C. Holmgren and S. Janson, Limit laws for functions of fringe trees for
binary search trees and recursive trees, see arxiv.org
[429] P. Horn and M. Radcliffe, Giant components in Kronecker graphs, Random
Structures and Algorithms 40 (2012) 385-397.
[430] H-K. Hwang, Profiles of random trees: plane-oriented recursive trees, Ran-
dom Structures and Algorithms 30 (2007) 380-413.
[431] G.I. Ivchenko, On the asymptotic behaviour of the degrees of vertices in a
random graph, Theory of Probability and Applications 18 (1973) 188-195.
[432] S. Janson, Poisson convergence and Poisson processes with applications to
random graphs, Stochastic Process. Appl. 26 (1987) 1-30.
[433] S. Janson, Normal convergence by higher semi-invariants with applications
to random graphs, Ann. Probab. 16 (1988) 305-312.
[434] S. Janson, A functional limit theorem for random graphs with applications
to subgraph count statistics Random Structures and Algorithms 1 (1990)
15-37.
[435] S. Janson, Poisson approximation for large deviations, Random Structures
and Algorithms 1 (1990) 221-230.
[436] S. Janson, Multicyclic components in a random graph process, Random
Structures and Algorithms 4 (1993) 71-84.
BIBLIOGRAPHY 535
[437] S. Janson, The numbers of spanning trees, Hamilton cycles and perfect
matchings in a random graph, Combin. Probab. Comput. 3 (1994) 97-126.
[438] S. Janson, The minimal spanning tree in a complete graph and a functional
limit theorem for trees in a random graph, Random Structures Algorithms
7 (1995) 337-355.
[440] S. Janson, One, two and three times log n/n for paths in a complete graph
with random weights, Combinatorics, Probability and Computing 8 (1999)
347-361.
[446] S. Janson and M.J. Luczak, A simple solution to the k-core problem, Ran-
dom Structures Algorithms 30 (2007) 50-62.
[449] S. Janson, T. Łuczak and A. Ruciński, Random Graphs, John Wiley and
Sons, New York, 2000.
[453] S. Janson, D.E. Knuth, T. Łuczak and B.G. Pittel, The birth of the giant
component, Random Structures and Algorithms 4 (1993) 233-358.
[462] J. Jaworski and I.H. Smit, On a random digraph, Annals of Discrete Math.
33 (1987) 111-127.
BIBLIOGRAPHY 537
[464] J. Jaworski and D. Stark, The vertex degree distribution of passive random
intersection graph models, Combin. Prob Comput. 17 (2008) 549-558.
[465] M.R. Jerrum, Large Cliques Elude the Metropolis Process, Random Struc-
tures and Algorithms 3 (1992) 347-359.
[467] J. Jonasson, On the cover time of random walks on random graphs, Com-
binatorics, Probability and Computing 7 (1998), 265-279.
[470] J. Jordan and A. Wade, Phase transitions for random geometric preferential
attachment graphs, Advances in Applied Probability 47 (2015).
[472] J. Kahn, E. Lubetzky and N. Wormald, The threshold for combs in random
graphs, see arxiv.org.
[473] J. Kahn, E. Lubetzky and N. Wormald, Cycle factors and renewal theory,
see arxiv.org.
[474] J. Kahn, B. Narayanan and J. Park, The threshold for the square of a Hamil-
ton cycle, see arxiv.org.
[477] S. Kalikow and B. Weiss, When are random graphs connected? Israel J.
Math. 62 (1988) 257-268.
[480] M. Kang and T. Łuczak, Two critical periods in the evolution of random
planar graphs, Transactions of the American Mathematical Society 364
(2012) 4239-4265.
[482] D. Karger and C. Stein, A new approach to the minimum cut problem,
Journal of the ACM 43 (1996) 601-640.
[491] R.M. Karp, An upper bound on the expected cost of an optimal assignment,
Discrete Algorithms and Complexity: Proceedings of the Japan-US Joint
Seminar (D. Johnson et al., eds.), Academic Press, New York, 1987, 1-4.
[492] R.M. Karp, The transitive closure of a random digraph, Random Structures
and Algorithms 1 (1990) 73-93.
[493] R.M. Karp, A. Rinnooy-Kan and R.V. Vohra, Average Case Analysis of
a Heuristic for the Assignment Problem, Mathematics of Operations Re-
search 19 (1994) 513-522.
[494] R.M. Karp and M. Sipser, Maximum matchings in sparse random graphs,
Proceedings of the 22nd IEEE Symposium on the Foundations of Computer
Science (1981) 364-375.
[495] Zs. Katona, Width of scale-free tree, J. of Appl. Prob. 42 (2005) 839-850.
[499] R. Keusch and A. Steger, The game chromatic number of dense random
graphs, Electronic Journal of Combinatorics 21 (2014).
[500] J-H. Kim, The Ramsey number R(3,t) has order of magnitude t 2 / logt,
Random Structures and Algorithms (1995) 173-207.
[501] J-H. Kim and V. Vu, Sandwiching random graphs: universality between
random graph models, Advances in Mathematics 188 (2004) 444-469.
[502] J.H. Kim and N.C. Wormald, Random matchings which induce Hamilton
cycles, and hamiltonian decompositions of random regular graphs, Journal
of Combinatorial Theory B 81 (2001) 20-44.
[503] J.F.C. Kingman, The population structure associated with the Ewens sam-
pling formula, Theoretical Population Biology 11 (1977) 274-283.
540 BIBLIOGRAPHY
[505] M. Kiwi and D. Mitsche, A bound for the diameter of random hyperbolic
graphs, see arxiv.org.
[512] J. Komlós and E. Szemerédi, Limit distributions for the existence of Hamil-
ton circuits in a random graph, Discrete Mathematics 43 (1983) 55-63.
[513] V.F. Kolchin, A problem of the allocation of particles in cells and cycles of
random permutations, Theory Probab. Appl. 16 (1971) 74-90.
[514] V.F. Kolchin, Branching processes, random trees, and a generalized scheme
of arrangements of particles, Mathematical Notes 21 (1977) 386-394.
[515] V.F. Kolchin, Random Mappings, Optimization Software, Inc., New York
(1986).
[519] I.N. Kovalenko, Theory of random graphs, Cybernetics and Systems Anal-
ysis 7 (1971) 575-579.
[523] M. Krivelevich, C. Lee and B. Sudakov, Long paths and cycles in random
subgraphs of graphs with large minimum degree, see arxiv.org.
[535] M. Kuba and A. Panholzer, On the degree distribution of the nodes in in-
creasing trees, J. Combinatorial Theory (A) 114 (2007) 597-618.
[537] D. Kühn and D. Osthus, On Pósa’s conjecture for random graphs, SIAM
Journal on Discrete Mathematics 26 (2012) 1440-1457.
[540] A.N. Lageras and M. Lindholm, A note on the component structure in ran-
dom intersection graphs with tunable clustering, The Electronic Journal of
Combinatorics 15 (2008) #N10.
[541] C. Lee and B. Sudakov, Dirac’s theorem for random graphs, Random Struc-
tures and Algorithms 41 (2012) 293-305.
[543] C. Lennon and B. Pittel, On a likely number of solutions for a stable mar-
riage problem, Combinatorics, Probability and Computing (2009) 371-421.
[548] K. Lin and G. Reinert, Joint vertex degrees in the inhomogeneous random
graph model G (n, {pi j }), Adv. in Appl. Prob. 44 (2012) 139-165.
[552] J. Lu, and Q. Feng, (1998). Strong consistency of the number of vertices of
given degrees in nonuniform random recursive trees, Yokohama Math. J 45
(1998) 61-69.
[553] E. Lubetzky and A. Sly, Cutoff phenomena for random walks on random
regular graphs, Duke Mathematical Journal 153 (2010) 475-510.
[554] E. Lubetzky and Y. Zhao, On the variational problem for upper tails in
sparse random graphs, see arxiv.org.
[555] T. Łuczak, On the equivalence of two basic models of random graphs, Pro-
ceedings of Random Graphs’87, (Michal Karonski, Jerzy Jaworski, Andrzej
Rucinski, eds.), Wiley, Chichester (1990) 151-158.
[557] T. Łuczak, Component behaviour near the critical point, Random Structures
and Algorithms 1 (1990) 287-310.
[558] T. Łuczak, Cycles in a random graph near the critical point, Random Struc-
tures and Algorithms 2 (1991) 421-440.
[559] T. Łuczak, Size and connectivity of the k-core of a random graph, Discrete
Mathematics 91 (1991) 61-68.
[560] T. Łuczak, The chromatic number of random graphs, Combinatorica 11
(1991) 45-54.
[561] T. Łuczak, A note on the sharp concentration of the chromatic number of
random graphs, Combinatorica 11 (1991) 295-297.
[562] T. Łuczak, The phase transition in a random graph, Combinatorics, Paul
Erdős is Eighty, Vol. 2, eds. D. Miklós, V.T. Sós, and T. Szőnyi, Bolyai Soc.
Math. Stud. 2, J. Bolyai Math. Soc., Budapest (1996) 399-422.
[563] T. Łuczak, Random trees and random graphs, Random Structures and Al-
gorithms 13 (1998) 485-500.
[564] T. Łuczak, On triangle free random graphs, Random Structures and Algo-
rithms 16 (2000) 260-276.
[565] T. Łuczak, B. Pittel and J. C. Wierman, The structure of a random graph at
the point of the phase transition, Transactions of the American Mathemati-
cal Society 341 (1994) 721-748
[566] T. Łuczak and P. Pralat, Chasing robbers on random graphs: zigzag theo-
rem, Random Structures and Algorithms 37 (2010) 516-524.
[567] T. Łuczak and A. Ruciński, Tree-matchings in graph processes SIAM Jour-
nal on Discrete Mathematics 4 (1991) 107-120.
[568] T. Łuczak, A. Ruciński and B. Voigt, Ramsey properties of random graphs,
Journal of Combinatorial Theory B 56 (1992) 55-68.
[569] T. Łuczak, Ł. Witkowski and M. Witkowski, Hamilton Cycles in Random
Lifts of Graphs, see arxiv.org.
[570] M. Mahdian and Y. Xu. Stochastic Kronecker graphs, Random Structures
and Algorithms 38 (2011) 453-466.
[571] H. Mahmoud, Evolution of Random Search Trees, John Wiley & Sons, Inc.,
New York, 1992.
BIBLIOGRAPHY 545
[587] C. McDiarmid and B. Reed, Colouring proximity graphs in the plane, Dis-
crete Mathematics 199 (1999) 123-137.
[589] B.D. McKay, Asymptotics for symmetric 0-1 matrices with prescribed row
sums, Ars Combinatoria 19 (1985) 15-25.
[590] B.D. McKay and N.C. Wormald, Asymptotic enumeration by degree se-
quence of graphs with degrees o(n1/2 ), Combinatorica 11 (1991) 369-382.
[591] B.D. McKay and N.C. Wormald, The degree sequence of a random graph
I, The models, Random Structures and Algorithms 11 (1997) 97-117.
[594] A. Meir and J.W. Moon, The distance between points in random trees, J.
Combinatorial Theory 8 (1970) 99-103.
[595] A. Meir and J.W. Moon, Cutting down random trees, J. Austral. Math. Soc.
11 (1970) 313-324.
[596] A. Meir and J.W. Moon, Cutting down recursive trees, Mathematical Bio-
sciences 21 (1974) 173-181.
[597] A. Meir and J.W. Moon, On the altitude of nodes in random trees, Canadian
J. Math. 30 (1978) 997-1015.
[598] S.Micali and V.Vazirani, An O(|V |1/2 |E|) algorithm for finding maximum
matching in general graphs, Proceedings of the 21st IEEE Symposium on
Foundations of Computing (1980) 17-27.
[600] S. Milgram, The Small World Problem, Psychology Today 2 (1967) 60-67.
[618] T. Müller, X. Pérez and N. Wormald, Disjoint Hamilton cycles in the ran-
dom geometric graph, Journal of Graph Theory 68 (2011) 299-322.
[622] A. Nachmias and Y. Peres, Critical random graphs: diameter and mixing
time, Annals of Probability 36 (2008) 1267-1286.
[623] A. Nachmias and Y. Peres, The critical random graph, with martingales,
Israel Journal of Mathematics 176 (2010) 29-41.
[625] D. Najock and C.C. Heyde (1982). On the number of terminal vertices in
certain random trees with an application to stemma construction in philol-
ogy, Journal of Applied Probability 19 (1982) 675-680.
[628] R. Nenadov and A. Steger, A Short proof of the Random Ramsey Theorem,
to appear in Combinatorics, Probability and Computing.
[630] R. Nenadov, A. Steger and Stojacović, On the threshold for the Maker-
Breaker H-game, see arxiv.org.
[632] M.E.J. Newman and D.J. Watts, Scaling and percolation in the small-world
network model,Phys. Rev. E 60 (1999) 73327342.
[633] S. Nicoletseas, C. Raptopoulos and P.G. Spirakis, The existence and effi-
cient construction of large independent sets in general random intersection
graph, In: ICALP 2004, LNCS 3142, Springer (2004) 1029-1040.
[636] D. Osthus, H. J. Prömel, and A. Taraz, On random planar graphs, the num-
ber of planar graphs and their triangulations, Journal of Combinatorial The-
ory B 88 (2003) 119-134.
[639] E. Palmer and J. Spencer, Hitting time for k edge disjoint spanning trees in
a random graph, Period Math. Hungar. 91 (1995) 151-156.
[640] A. Panholzer and H. Prodinger, The level of nodes in increasing trees re-
visited, Random Structures and Algorithms 31 (2007) 203-226.
[644] E. Peköz, A. Röllin and N. Ross, Degree asymptotics with rates for pref-
erential attachment random graphs, The Annals of Applied Probability 23
(2013) 1188-1218.
550 BIBLIOGRAPHY
[647] N. Peterson and B. Pittel, Distance between two random k-out graphs, with
and without preferential attachment, see arxiv.org.
[648] J. Pitman, Random mappings, forest, and subsets associated with the Abel-
Cayley-Hurwitz multinomial expansions, Seminar Lotharingien de Combi-
natoire 46 (2001) 46.
[650] B.Pittel, On tree census and the giant component in sparse random graphs,
Random Structures and Algorithms 1 (1990), 311-342.
[652] B. Pittel, Note on the heights of random recursive trees and random m-ary
search trees, Random Structures and Algorithms 5 (1994) 337-347.
[656] B. Pittel and R. Weishar, The random bipartite nearest neighbor graphs,
Random Structures and Algorithms 15 (1999) 279-310.
[661] P. Pralat, Cleaning random d-regular graphs with Brooms, Graphs and
Combinatorics 27 (2011) 567-584.
[663] H. Prodinger, Depth and path length of heap ordered trees, International
Journal of Foundations of Computer Science 7 (1996) 293-299.
[668] M. Radcliffe and S.J. Young, Connectivity and giant component of stochas-
tic Kronecker graphs, see arxiv.org.
[670] A. Rényi, On connected graphs I, Publ. Math. Inst. Hungar. Acad. Sci. 4
(1959) 385-387.
[672] O. Riordan and A. Selby, The maximum degree of a random graph, Com-
binatorics, Probability and Computing 9 (2000) 549-572.
[673] O. Riordan, The small giant component in scale-free random graph, Com-
bin. Prob. Comput. 14 (2005) 897-938.
[674] O. Riordan, Long cycles in random subgraphs of graphs with large mini-
mum degree, see arxiv.org.
552 BIBLIOGRAPHY
[675] O. Riordan and L. Warnke, Achlioptas process phase transitions are con-
tinuous, Annals of Applied Probability 22 (2012) 1450-1464.
[676] O. Riordan and N. Wormald, The diameter of sparse random graphs, Com-
binatorics, Probability and Computing 19 (2010) 835-926.
[677] R.W. Robinson and N.C.Wormald, Almost all cubic graphs are Hamilto-
nian, Random Structures and Algorithms 3 (1992) 117-126.
[678] V. Rödl and A. Ruciński, Lower bounds on probability thresholds for Ram-
sey properties, in Combinatorics, Paul Erdős is Eighty, Vol.1, Bolyai Soc.
Math. Studies (1993), 317-346.
[680] R.W. Robinson and N.C.Wormald, Almost all regular graphs are Hamilto-
nian, Random Structures and Algorithms 5 (1994) 363-374.
[681] V. Rödl and A. Ruciński, Threshold functions for Ramsey properties Jour-
nal of the American Mathematical Society 8 (1995) 917-942.
[685] A. Ruciński, When a small subgraphs of a random graph are normally dis-
tributed, Probability Theory and Related Fields 78 (1988) 1-10.
[686] A. Ruciński and A. Vince, Strongly balanced graphs and random graphs,
Journal of Graph Theory 10 (1986) 251-264.
[688] A. Ruciński and N. Wormald, Random graph processes with degree restric-
tions, Combinatorics, Probability and Computing 1 (1992) 169-180.
[689] A. Rudas, B,Toth and B. Való, Random trees and general branching pro-
cesses, Random Structures and Algorithms 31 (2007) 186-202.
BIBLIOGRAPHY 553
[692] K. Rybarczyk, Equivalence of the random intersection graph and G(n, p),
Random Structures and Algorithms 38 (2011) 205-234.
[693] K. Rybarczyk, Sharp threshold functions for random intersection graphs via
a coupling method, The Electronic Journal of Combinatorics 18(1) (2011)
#P36.
[698] A.A. Saposhenko, Metric properties of almost all boolean functions, (in
Russian), Diskretny Analiz 10 (1967) 91-119.
[699] A.A. Saposhenko, Geometric structure of almost all boolean functions, (in
Russian), Problemy Kibernetiki 30 (1975) 227-261.
[701] D. Saxton and A. Thomason, Online containers for hypergraphs with ap-
plications to linear equations, see arxiv.org.
[702] M. Schacht, Extremal results for random discrete structures, see arxiv.org.
[711] S. Shelah and J. Spencer, Zero-One laws for sparse random graphs, Journal
of the American Mathematical Society 1 (1988) 97-115.
[712] L.A. Shepp and S.P. Lloyd, Ordered cycle length in a random permutation,
Trans. Amer. Math. Soc. 121 (1966) 340-357.
[720] J. Spencer, The Strange Logic of Random Graphs, Springer, Berlin 2001.
BIBLIOGRAPHY 555
[724] C. Stein, A bound for the error in the normal approximation to the distribu-
tion of a sum of dependent random variables, In:Proceedings of the Sixth
Berkeley Symposium on Mathematical Statistics and Probability , 1970,
Univ. of California Press, Berkeley, Vol. II (1972) 583-602.
[728] Ch. Su, J. Liu and Q. Feng, A note on the distance in random recursive
trees, Statistics & Probability Letters 76 (2006) 1748-1755
[729] B. Sudakov and V. Vu, Local resilience of graphs, Random Structures and
Algorithms 33 (2008) 409-433.
[731] J. Szymański, On the maximum degree and the height of random recursive
tree, In:Random Graphs’87, Wiley, New York (1990) 313-324.
[735] M.A. Tapia and B.R. Myers, Generation of concave node-weighted trees,
IEEE Trans. Circuits Systems 14 (1967) 229-230.
[736] S-H. Teng and F.F. Yao, k-nearest-neighbor clustering and percolation the-
ory, Algorithmica 49 (2007) 192-211.
[737] A.Thomason, A simple linear expected time algorithm for Hamilton cycles,
Discrete Mathematics 75 (1989) 373-379.
[739] T.S. Turova, Phase transition in dynamical random graphs, J. Stat. Phys.
123 (2006) 1007-1032.
[740] T.S. Turova, Diffusion approximation for the componentsin critical inho-
mogeneous random graphsof rank 1, Random Structures Algorithms 43
(2013) 486-539.
[741] T.S. Turova, Asymptotics for the size of the largest component scaled to
”log n” in inhomogeneous random graphs, Arkiv fur Matematik 51 (2013)
371-403.
[744] W.F. de la Vega, Long paths in random graphs, Studia Sci. Math. Hungar.
14 (1979) 335-340.
[748] V. Vu. On some degree conditions which guarantee the upper bound of
chromatic (choice) number of random graphs, Journal of Graph Theory 31
(1999) 201-226.
BIBLIOGRAPHY 557
[752] J. Wástlund, An easy proof of the ζ (2) limit in the random assignment
problem, Electronic Communications in Probability 14 (2009) 261-269.
[753] D.J. Watts and S.H. Strogatz, Collective dynamics of ”small-world” net-
works, Nature 393 (1998) 440442.
[754] K. Weber, Random graphs - a survey, Rostock Math. Kolloq. (1982) 83-98.
[756] E.P. Wigner, On the distribution of the roots of certain symmetric matrices,
Annals of Mathematics 67 (1958) 325-327.
[758] E.M. Wright, The number of connected sparsely edged graphs, Journal of
graph Theory 1 (1977) 317-330.
[759] E.M. Wright, The number of connected sparsely edged graphs II, Journal
of graph Theory 2 (1978) 299-305.
[760] E.M. Wright, The number of connected sparsely edged graphs III, Journal
of graph Theory 4 (1980) 393-407.
[761] N.C. Wormald, Models of random regular graphs, In J.D. Lamb and D.A.
Preece, editors, Surveys in Combinatorics, 1999, volume 267 of London
Mathematical Society Lecture Note Series, Cambridge University Press
(1999) 239-298.
[762] N. C. Wormald, The differential equation method for random graph pro-
cesses and greedy algorithms, Lectures on approximation and randomized
algorithms (M. Karoński and H.J. Prömel, eds.), Advanced Topics in Math-
ematics, Polish Scientific Publishers PWN, Warsaw, 1999, pp. 73-155.
558 BIBLIOGRAPHY
[765] S.J. Young, Random Dot Product Graphs: A Flexible Model of Complex
Networks, ProQuest (2008)
[766] S.J. Young and E.R. Scheinerman, Random dot product graph models for
social networks, In: Algorithms and Models for the Web-Graph, LNCS
4863, Springer (2007) 138-149.
[769] T. Zhou, G. Yan and B.H. Wang, Maximal planar networks with large clus-
tering coefficient and power-law degree distribution, Physics Reviews E 71
(2005) 046141.
Chapter 26
Indices
559
560 CHAPTER 26. INDICES
Author Index
Chvátal, V., 18, 138, 403 Faloutsos, C., 185, 194, 367
Coja–Oghlan, A., 129, 137, 409 Faloutsos, M., 367
Collevecchio, A., 386 Faloutsos, P., 367
Conlon, D., 140, 143 Fang, W., 255
Cooley, O., 294 Farczadi, L., 386
Cooper, C., 48, 106, 107, 229, 268, Feldman, V., 408
295, 296, 365, 373, 375, 385, Fenner, T., 107, 108, 345, 385
386, 399, 403, 411 Ferber, A, 156
Cover, T., 455 Ferber, A., 265, 267, 268, 275, 293,
403, 404, 412
Damarackas, J., 253 Fernholz, D., 139
Dani, V., 119 Fill, J., 232
de la Vega, W., 94 Fiz Pontiveros, G., 406
Deijfen, M., 193, 253 Flajolet, P., 408
Dellamonica, D., 412
Flaxman, A., 47, 138, 385, 400
DeMarco, B., 154
Fortuin, C., 428
Devroye, L., 192
Foulds, L., 410
Diaz, J., 245, 254
Fountoulakis, N., 47, 255, 295, 351,
Diepetro, R., 252
411
Ding, J., 47, 411
Fraiman, N., 192
Dudek, A., 106, 136, 217, 230, 275,
Frankl, F., 457
404, 406
Friedgut, E., 19
Durrett, R., 191, 228, 385, 418
Friedman, J., 230
Dyer, M., 99, 138, 294, 400, 410
Friedrich, T., 254, 255
Ebrahimzadeh, E., 386 Frieze, A., 18, 47, 70, 71, 85, 99,
Edmonds, J., 96 106–109, 119, 137, 138, 154,
Edmonson-Jones, M., 254 217, 229, 230, 264, 267, 268,
Efthymiou, C., 139 275, 294–296, 345, 347, 351,
Egorychev, G., 265 356, 364, 365, 373, 375, 385–
Elsässer, R., 254 387, 399–401, 403–405, 407–
England, J., 455 411
Ercal, G., 411 Frieze, A.,, 71
Erdős, P., vi, 3, 20, 40, 48, 60, 62,
64, 68, 74, 79, 80, 83, 85, Gale, D., 412
94, 125, 154, 295, 403, 406, Gamarnik, D., 47, 121, 400
407, 410 Gao, P., 109, 294, 295, 386
Eschenauer, L., 252 Gao, Z., 408
Gebauer, H., 404
Füredi, Z., 103, 129 Gerke, S., 110, 143, 252, 408
Fagin, R., 407 Gharamani, Z., 194
Falikman, D., 265 Gilbert, E., vi, 3
562 CHAPTER 26. INDICES
Knox, F., 107 Lubetzky, E., 47, 107, 109, 268, 407,
Knuth, D., 40, 47, 97, 412 411
Koch, C., 193, 194, 294 Lucier, B., 385
Kogan, D., 407 Luczak, M., 229
Kohayakawa, Y., 143, 278, 410, 412
Kolda, T,, 194 Müller, T., 241, 245, 254, 255, 404
Komlós, J., 48, 89, 94, 129, 410 Mahdian, M., 186
Kordecki, W., 62 Mahoney, M., 255
Kovalenko, I., 171, 172 Makai, T., 193, 194
Kozma, G., 385, 411 Makowski, A., 252
Kravitz, D., 47 Mancini, L., 252
Kreuter, B., 143 Marchetti-Spaccemela, A., 254
Krioukov, D., 255 Marczewski, E., 231
Krishnamachari, B., 254 Martin, A., 455
Krivelevich, M., 47, 48, 94, 99, 107– Martin, R., 295
109, 133, 137, 138, 229, 230, Martin-Löf, A., 193
245, 268, 294, 403, 404, 408 Matoušek, J., 409
Krivelevich. M., 95 Matula, D., 117, 123
Krohmer, A., 255 McColm, G., 254
Kronenberg, G., 265, 267, 403 McDiarmid, C., 107, 125, 138, 246,
Kucera, L., 408 248, 254, 264, 293, 399, 400,
Kumar, P., 241 408
Kurauskas, V., 251, 253 McDowell, A., 110
McKay, B., 41, 63, 200, 403
Lageras, A., 251, 253 Mehrabian, A., 386
Lavrov, M., 405 Mei, A., 252
Lee, C., 94, 99, 108, 156 Melsted, P., 109, 351
Leighton, T., 409 Meshulam, R., 410
Lelarge, M., 295 Micali, S., 96
Lennon, C., 412 Mihail, M., 385
Leskovec, J., 185, 194 Miklós, D., 385
Lev, A., 71 Milgram, S., 381
Liagonkii, M., 407 Mitsche, D., 136, 245, 406
Liang, H., 71 Mitzenmacher, M., 294
Lin, K., 192 Molloy, M., 212, 229, 268, 294, 296,
Lindholm, M., 251, 253 402
Linial, N., 409, 410 Montgomery, R., 109, 156
Linusson, S., 393 Moore, C., 119, 137, 229
Loh, P., 47, 275, 403 Morris, R., 140, 406
Long, E., 265, 267 Mossell, E., 139
Lu, L., 178, 184, 385 Motwani, R., 97, 412
564 CHAPTER 26. INDICES
Mousset, F., 403 Pralat, P., 106, 136, 385, 386, 403,
Mubayi, D., 295, 407 405, 406
Preciado, V, 254
Nachmias, A., 41, 411
Nair, C., 393 Rödl, V., 141, 143, 154, 156, 412
Naor, A., 137, 404 Röllin, A., 385
Nash-Williams, C., 109 Rényi, A., vi, 3, 20, 40, 62, 64, 68,
Naves, H., 108 74, 79, 83, 85, 94
Nenadov, R., 141, 154, 156, 268, 403, Radakrishnan, J., 457
412 Radcliffe, J., 187, 193
Newman, M., 191 Rademacher, L., 365
Nicoletseas, R., 252, 253 Radhakrishnan, J., 252
Noever, A., 156, 268 Rai, S., 254
Norros, I., 193 Ramachandran, V., 139
Noy, M., 408 Raptopoulos, C., 252, 253
Osthus, D., 47, 107, 408 Ratsaby, J., 295
Reed, B., 212, 229, 248, 296, 411
Pérez-Giménez, X., 109, 230, 245 Reinert, G., 192, 253
Pósa, L., 89 Reittu, H., 193
Palka, Z., 62, 63, 365 Reyzin, L., 408
Panagiotou, K., 137, 153, 255, 295, Rinnooy-Kan, A, 349, 351
351 Riordan, O., 47, 63, 71, 99, 108, 139,
Panconesi, A., 252 191–193, 367, 385, 400, 406
Papadimitriou, C., 385 Robalewska, H., 402
Papadopolous, F., 255 Robinson, R., 229, 402
Parisi, G., 393 Roditty, Y., 71
Pegden, W., 137, 385 Ross, N., 385
Peköz, R., 385 Rozenman, E., 409
Penrose, M., 241, 254 Ruciński, A., vii, 50, 52, 62, 74, 76,
Peres, Y., 41, 47, 411 78–80, 110, 141, 154, 156,
Perkins, W., 47 217, 407, 409, 412, 423, 424,
Person, Y., 154, 278 436, 453
Peter, U., 156, 255, 268, 403, 412 Ruszinko, M., 295, 399
Petit, J., 254 Rybarczyk, K., 232, 236, 240, 251–
Picolleli, M., 407 253
Pikhurko, O., 154
Pinar, A., 194 Söderberg, B., 191
Pittel, B., 37, 40, 47, 97, 99, 109, 412 Sós, V., 154
Poole, S., 194, 278 Samotij, W., 107, 140
Prömel, H., 143, 408 Saposhenko, A., 410
Prabhakar, B., 393 Sarkar, A., 364
AUTHOR INDEX 565
Sato, C., 109, 386 Steger, A., 137, 141, 143, 153, 154,
Sauerwald, T., 254 405, 408
Saxton, D., 140, 141 Stein, C., 190, 422
Schacht, M., 140, 143 Stojaković, M., 403, 404
Schaeffer, G., 408 Strogatz, S., 191
Scheinerman, E., 193, 231, 232, 237 Sudakov, B., 47, 48, 94, 95, 99, 107,
Schickinger, T., 143 108, 137, 138, 155, 156, 167,
Schlatter, D., 408 229, 230, 268, 294, 400, 405,
Schmidt, J., 279 408
Scott, A., 135 Suen, S., 70, 230, 406
Selby, A., 63 Sullivan, B., 194
Selkow, S., 60 Szabó, T., 109, 403, 404
Sen, S., 193 Szemerédi, E., 48, 89, 94, 156, 230,
Serna, J., 254 410, 412
Seshadhri, C., 194
Shamir, E., 279, 409 Talagrand, M., 445
Shannon, C., 455 Talanov, V., 407
Shapley, L., 412 Taraz, A., 251, 408
Sharma, M., 393 Teixeira, A., 411
Shearer, J., 457 Teng, S., 364
Shelah, S., 108, 407 Tetali, P., 121, 294
Shepp, L., 412 Thaler, J., 294
Shi, L., 230 Thoma, L., 399
Shikhelman, C., 403 Thomas, J., 455
Shinkar, I., 406 Thomason, A., 10, 108, 140, 141
Simonovits, M., 153, 154 Trujić, M, 156
Singer-Cohen, K., 231, 232, 237 Tsourakakis, C., 71, 230, 386
Sipser, M., 109, 409, 453 Tsur, G., 406
Sivasubramanian, S., 270 Turan, P., 72
Sly, A., 139, 411 Turova, T., 191
Smit, I., 365 Tusnády, G., 367
Smyth, C., 154, 230, 295 Tutte, W., 408
Soria, M., 408 Tuza, Z., 71
Sorkin, G., 47, 401 Ueckerdt, M., 251
Speiss, S., 275 Upfal, E., 70, 230
Spencer, J., 37, 41, 47, 106, 111, 153,
367, 399, 407, 410, 445 Vahdat, A., 255
Spirakis, P., 252, 253 van den Esker, H., 385
Stark, D., 240, 251 van der Hofstad, R., 193, 399
Stauffer, A., 254, 255 van Leeuwaarden, J., 193
Steger, A, 156 Vazirani, V., 96
566 CHAPTER 26. INDICES
Main Index
VC-Dimension, 295