Gradprob Notes3
Gradprob Notes3
1 Modes of convergence
Let (Ω, F, P) be a probability space. We will encounter various modes of conver-
gence for sequences of RVs on (Ω, F, P).
DEF 3.1 (Modes of convergence) Let {Xn }n be a sequence of (not necessarily
independent) RVs and let X be a RV. Then we have the following definitions.
• Convergence in probability: ∀ε > 0, P[|Xn −X| > ε] → 0 (as n → +∞);
which we denote by Xn →P X.
1
Lecture 3: Modes of convergence 2
Var[X]
P[|X − E[X]| > a] ≤ .
a2
Proof: Apply Markov’s inequality to Z = (X − E[X])2 .
An immediate application of Chebyshev’s inequality is the following.
THM 3.4 Let (Sn )n be a sequence of RVs with µn = E[Sn ] and σn2 = Var[Sn ]. If
σn2 /b2n → 0, then
Sn − µn
→P 0.
bn
DEF 3.6 (Infinitely often, eventually) Let (An )n be a sequence of events. Then
we define
Note that
1An i.o. = lim sup 1An .
n
Similarly,
[ +∞
\
An eventually (ev.) ≡ {ω : ω is in An for all large n} ≡ lim inf An ≡ An .
n
m n=m
Note that
1An ev. = limninf 1An .
Also we have (An ev.)c = (Acn i.o.).
Lecture 3: Modes of convergence 3
then
P[An i.o.] = 0.
as N → +∞. So P[∪+∞
n=M An ] = 1 and further
P ∩M ∪+∞
n=M A n = 1,
by monotonicity.
and
P
THM 3.16 (Weak law of large numbers) Let (Xn )n be IID and Sn = k≤n Xk .
A necessary and sufficient condition for the existence of constants (µn )n such that
Sn
− µn →P 0,
n
is
n P[|X1 | > n] → 0.
In that case, the choice
µn = E[X1 1|X1 |≤n ],
works.
Lecture 3: Modes of convergence 6
2 Further applications...
2.1 ...of Chebyshev’s inequality
Chebyshev’s inequality and Theorem 3.4 can be used to derive limit laws in some
cases where sequences are not necessarily IID. We give several important examples
from [D].
EX 3.17 (Occupancy problem) Suppose we throw r balls into n bins indepen-
dently uniformly at random. Let Nn be the number of empty boxes. If Ai is the
event that the i-th bin is empty, we have
1 r
P[Ai ] = 1 − ,
n
so that Nn = k≤n 1Ak (not independent) and
P
1 r
E[Nn ] = n 1 − .
n
In particular, if r/n → ρ we have
E[Nn ]
→ e−ρ .
n
Because there is no independence, the variance calculation is trickier. Note that
!2
n
1Am =
X X
E[Nn2 ] = E P[Am ∩ Am0 ],
m=1 1≤m,m0 ≤n
and
Var[Nn ] = E[Nn2 ] − (E[Nn ])2
X
= [P[Am ∩ Am0 ] − P[Am ]P[Am0 ]]
1≤m,m0 ≤n
and moments
1
E[N ] = ,
p
and
1−p 1
Var[N ] = ≤ 2 .
p2 p
Hence
n n
k − 1 −1
X X 1
E[Tn ] = 1− =n ∼ n log n,
n m
k=1 m=1
and
n n
k − 1 −2
X X 1
Var[Tn ] ≤ 1− = n2 ≤ Cn2 ,
n m2
k=1 m=1
Tn − n nm=1 m−1
P
→P 0,
n log n
or
Tn
→P 1.
n log n
indicator that the k-th element in this construction precedes the closure of a cycle.
E.g., we have X9,3 = X9,7 = X9,9 = 1. The construction above implies that the
Xn,k ’s are independent and
1
P[Xn,j = 1] = .
n−j+1
P is because only one of the remaining elements closes the cycle. Letting Sn =
That
k≤n Xn,k be the number of cycles in π we have
n
X 1
E[Sn ] = ∼ log n,
n−j+1
j=1
and
n
X n
X n
X
2
Var[Sn ] = Var[Xn,j ] ≤ E[Xn,j ]= E[Xn,j ] = E[Sn ].
j=1 j=1 j=1
Ln = max `m .
1≤m≤n
Note that P[`n = k] = (1/2)k+1 for all n, k. (The +1 in the exponent is for the
first −1.) We will prove
Ln
→ 1, a.s.
log2 n
For the lower bound, it suffices to divide the sequence into disjoint blocks to
use independence. Take blocks of size [(1 − ε) log2 n] + 1 so that a block is all-1
with probability at least
nε
n/ log2 n
−(1−ε)
P[Ln ≤ (1 − ε) log2 n] ≤ 1 − n /2 ≤ exp − ,
log2 n
which is summable. By (BC1),
Ln
lim inf ≥ 1 − ε, a.s.
n log2 n
The upper bound follows from (BC1). Indeed note that, for any ε > 0,
k+1
X 1
P[`n ≥ (1 + ε) log2 n] = ≤ n−(1+ε) ,
2
k≥(1+ε) log2 n
so that
P[`n ≥ (1 + ε) log2 n i.o.] = 0,
Hence, there is Nε (random) such that `n ≤ (1 + ε) log2 n for all n ≥ Nε and note
that the `n ’s with n < Nε are finite a.s. as they have a finite expectation. Therefore
Ln
lim sup ≤ 1 + ε, a.s.
n log2 n
Since ε is arbitrary, we get the upper bound.
1
Pn
Pnm=1 Am → 1, a.s.
m=1 P[Am ]
Var[Sn ] 1
P[|Sn − E[Sn ]| > δE[Sn ]] ≤ ≤ 2 → 0,
δ 2 E[Sn ]2 δ E[Sn ]
Lecture 3: Modes of convergence 10
by assumption. In particular,
Sn
→P 1.
E[Sn ]
We use a standard trick to obtain almost sure convergence. The idea is to take
subsequences, use (BC1), and sandwich the original sequence.)
1. Take
nk = inf{n : E[Sn ] ≥ k 2 },
and let Tk = Snk . Since E[Xn ] ≤ 1 we have in particular k 2 ≤ E[Tk ] ≤
k 2 + 1. Using Chebyshev again,
1
P[|Tk − E[Tk ]| > δE[Tk ]] ≤ ,
δ2 k2
which is summable so that, using (BC1) and the fact that δ is arbitrary,
Tk
→ 1, a.s.
E[Tk ]
We will see this argument again when we prove the strong law of large num-
bers.
Rn
→ 1, a.s.
log n
P[Am1 ∩ Am2 ] 1
= P[Am1 ] = .
P[Am2 ] m1
We have proved that the Ak ’s are pairwise independent and that P[Ak ] = 1/k.
Now use the fact that
n
X 1
∼ log n,
i
i=1
References
[Dur10] Rick Durrett. Probability: theory and examples. Cambridge Series in
Statistical and Probabilistic Mathematics. Cambridge University Press,
Cambridge, 2010.