Prob Notes
Prob Notes
Greg Lawler
Last Updated: March 21, 2016
Overview
This is an introduction to the mathematical foundations of probability theory. It is intended as a supplement
or follow-up to a graduate course in real analysis. The first two sections assume the knowledge of measure
spaces, measurable functions, Lebesgue integral, and notions of convergence of functions; the third assumes
Fubini’s Theorem; the fifth assumes knowledge of Fourier transform of nice (Schwartz) functions on R; and
Section 6 uses the Radon-Nikodym Theorem.
The mathematical foundations of probability theory are exactly the same as those of Lebesgue integration.
However, probability adds much intuition and leads to different developments of the area. These notes are
only intended to be a brief introduction — this might be considered what every graduate student should
know about the theory of probability.
Probability uses some different terminology than that of Lebesgue integration in R. These notes will
introduce the terminology and will also relate these ideas to those that would be encountered in an ele-
mentary (by which we will mean pre-measure theory) course in probability or statistics. Graduate students
encountering probabilty for the first time might want to also read an undergraduate book in probability.
1 Probability spaces
Definition A probability space is a measure space with total measure one. The standard notation is (Ω, F, P)
where:
• Ω is a set (sometimes called a sample space in elementary probability). Elements of Ω are denoted ω
and are sometimes called outcomes.
• F is a σ-algebra (or σ-field, we will use these terms synonymously) of subsets of Ω. Sets in F are called
events.
• P is a function from F to [0, 1] with P(Ω) = 1 and such that if E1 , E2 , . . . ∈ F are disjoint,
∞
[ X∞
P Ej = P[Ej ].
j=1 j=1
A discrete probability space is a probability space such that Ω is finite or countably infinite. In this case
we usually choose F to be all the subsetsP of Ω (this can be written F = 2Ω ), and the probability measure P
is given by a function p : Ω → [0, 1] with ω∈Ω p(ω) = 1. Then,
X
P(E) = p(ω).
ω∈E
We will consider another important example here, the probability space associated to an infinite number
of flips of a coin. Let
Ω = {ω = (ω1 , ω2 , . . .) : ωj = 0 or 1}.
1
We think of 0 as “tails” and 1 as “heads”. For each positive integer n, let
Ωn = {(ω1 , ω2 , . . . , ωn ) : ωj = 0 or 1}.
Each Ωn is a finite set with 2n elements. We can consider Ωn as a probability space with σ-algebra 2Ωn and
probability Pn induced by
pn (ω) = 2−n , ω ∈ Ωn .
Let Fn be the σ-algebra on Ω consisting of all events that depend only on the first n flips. More formally,
we define Fn to be the collection of all subsets A of Ω such that there is an E ∈ 2Ωn with
F1 ⊂ F2 ⊂ F3 ⊂ · · · .
Recall that an algebra of subsets of Ω is a collection of subsets containing Ω, closed under complementation,
and closed under finite unions. To show closure under finite unions, it suffices to show that if E1 , E2 ∈ F0 ,
then E1 ∪ E2 ∈ F0 .
E = {(1, 1, 1, . . .)}.
Proposition 1.2. The function P is a (countably additive) measure on F 0 , i.e., it satisfies P[∅] = 0, and if
E1 , E2 , . . . ∈ F 0 are disjoint with ∪∞ 0
n=1 En ∈ F , then
"∞ # ∞
[ X
P En = P(En ).
n=1 n=1
2
Proof. P[∅] = 0 is immediate. Also it is easy to see that P is finitely additive, i.e., if E1 , E2 , . . . , En ∈ F 0 are
disjoint then
n
[ Xn
P Ej =
P(Ej ).
j=1 j=1
(To see this, note that there must be an N such that E1 , . . . , En ∈ FN and then we can use the additivity
of PN .)
Showing countable subadditivity is harder. In fact, the following stronger fact holds: suppose E1 , E2 , . . . ∈
F 0 are disjoint and E = ∪∞ 0
n=1 En ∈ F . Then there is an N such that Ej = ∅ for j > N . Once we establish
this, countable additivity follows from finite additivity.
To establish this, we consider Ω as the topological space
{0, 1} × {0, 1} × · · ·
with the product topology where we have given each {0, 1} the discrete topology (all four subsets are open).
The product topology is the smallest topology such that all the sets in F 0 are open. Note also that all sets
in F 0 are closed since they are complements of sets in F 0 . It follows from Tychonoff’s Theorem that Ω is
a compact topological space under this topology. Suppose E1 , E2 , . . . are as above with E = ∪∞ 0
n=1 En ∈ F .
Then E1 , E2 , . . . is an open cover of the closed (and hence compact) set E. Therefore there is a finite
subcover, E1 , E2 , . . . , EN . Since the E1 , E2 , . . . are disjoint, this must imply that Ej = ∅ for j > N .
Let F be the smallest σ-algebra containing F 0 . Then the Carathéodory Extension Theorem tells us that
P can be extended uniquely to a complete measure space (P, F̄, P) where F ⊂ F̄.
The astute reader will note that the construction we just did is exactly the same as the construction of
Lebesgue measure on [0, 1]. Here we denote a real number x ∈ [0, 1] by its dyadic expansion
∞
X ωj
x = .ω1 ω2 ω3 · · · = .
j=1
2j
(There is a slight nuisance with the fact that .011111 · · · = .100000 · · · , but this can be handled.) The
σ-algebra F above corresponds to the Borel subsets of [0, 1] and the completion F corresponds to the
Lebesgue measurable sets. If the most complicated probability space we were interested were the space
above, then we could just use Lebesgue measure on [0, 1]. In fact, for almost all important applications
of probability, one could choose the measure space to be [0, 1] with Lebesgue measure (see Exercise 3).
However, this choice is not always the most convenient or natural.
Continuing on the last remark, one generally does not care what probability space one is working on.
What one observes are “random variables” which are discussed in the next section.
3
2 Random Variables and Expectation
Definition A random variable X is a measurable function from a probability space (Ω, F, P) to the reals1 ,
i.e., it is a function
X : Ω −→ (−∞, ∞)
such that for every Borel set B,
X −1 (B) = {X ∈ B} ∈ F.
Here we use the shorthand notation
{X ∈ B} = {ω ∈ Ω : X(ω) ∈ B}.
If X is a random variable, then for every Borel subset B of R, X −1 (B) ∈ F. We can define a function
on Borel sets by
µX (B) = P{X ∈ B} = P[X −1 (B)].
This function is in fact a measure, and (R, B, µX ) is a probability space. The measure µX is called the
distribution of the random variable. If µX gives measure one to a countable set of reals, then X is called
a discrete random variable. If µX gives zero measure to every singleton set, and hence to every countable
set, X is called a continuous random variable. Every random variable can be written as a sum of a discrete
random variable and a continuous random variable. All random variables defined on a discrete probability
space are discrete.
The distribution µX is often given in terms of the distribution function2 defined by
FX (x) = P{X ≤ x} = µX (−∞, x].
Note that F = FX satisfies the following:
• limx→−∞ F (x) = 0.
• limx→∞ F (x) = 1.
• F is a nondecreasing function.
• F is right continuous, i.e., for every x,
F (x+) := lim F (x + ) = F (x).
↓0
Conversely, any F satisfying the conditions above is the distribution function of a random variable. The
distribution can be obtained from the distribution function by setting
µX (−∞, x] = FX (x),
and extending uniquely to the Borel sets.
For some continuous random variables X, there is a function f = fX : R → [0, ∞) such that
Z b
P{a ≤ X ≤ b} = f (x) dx.
a
Such a function, if it exists, is called the density3 of the random variable. If the density exists, then
Z x
F (x) = f (t) dt.
−∞
1 or,
more generally, to any topological space
2 often
called cumulative distribution function (cdf) in elementary courses
3 More precisely, it is the density or Radon-Nikoydm derivative with respect to Lebesgue measure. In elementary courses,
4
If f is continuous at t, then the fundamental theorem of calculus implies that
f (x) = F 0 (x).
A density f satisfies Z ∞
f (x) dx = 1.
−∞
Conversely, any nonnegative function that integrates to one is the density of a random variable.
(The corresponding function in analysis is often called the characteristic function and denoted χE . Proba-
bilists never use the term characteristic function for the indicator function because the term characteristic
function has another meaning. The term indicator function has no ambiguity.)
Example Let (Ω, F, P) be the probability space for infinite tossings of a coin as in the previous section. Let
1, if nth flip heads,
Xn (ω1 , ω2 , . . .) = ωn = (2)
0, if nth flip tails.
Example Let µ be any probability measure on (R, B). Consider the trivial random variable
X = x,
defined on the probability space (R, B, µ). Then X is a random variable and µX = µ. Hence every probability
measure on R is the distribution of a random variable.
Example A random variable X has a normal distribution with mean µ and variance σ 2 if it has density
1 2
/2σ 2
f (x) = √ e−(x−µ) , −∞ < x < ∞.
2πσ 2
If µ = 0 and σ 2 = 1, X is said to have a standard normal distribution. The distribution function of the
standard normal is often denoted Φ,
Z x
1 2
Φ(x) = √ e−t /2 dt.
−∞ 2π
Example If X is a random variable and
g : (R, B) → R
is a Borel measurable function, then Y = g(X) is also a random variable.
Example Recall that the Cantor function is a continuous function F : [0, 1] → [0, 1] with F (0) = 0, F (1) = 1
and such that F 0 (x) = 0 for all x ∈ [0, 1] \ A where A denotes the “middle thirds” Cantor set. Extend F to R
by setting F (x) = 0 for x ≤ 0 and F (x) = 1 for x ≥ 1. Then F is a distribution function. A random variable
with this distribution function is continuous, since F is continuous. However, such a random variable has no
density.
5
The term random variable is a little misleading but it standard. It is perhaps easier to think of a random
variable as a “random number”. For example, in the case of coin-flips we get an infinite sequence of
random numbers corresponding to the results of the flips.
where the integral is the Lebesgue integral. If X is a random variable with E(|X|) < ∞, then we also define
the expectation by Z
E(X) = X dP.
If X takes on positive and negative values, and E(|X|) = ∞, the expectation is not defined.
Other terms used for expectation are expected value and mean (the term “expectation value” can be
found in the physics literature but it is not good English and sometimes refers to integrals with respect to
measures that are not probability measures). The letter µ is often used for mean. If X is a discrete random
variable taking on the values a1 , a2 , a3 , . . . we have the standard formula from elementary courses:
∞
X
E(X) = aj P{X = aj }, (4)
j=1
Then,
1
Xn − ≤ X ≤ Xn .
n
Hence
1
E[Xn ] − ≤ E[X] ≤ E[Xn ].
n
But,
X∞
1 k−1
E Xn − = P(Ak,n ),
n n
k=1
6
∞
X k
E[Xn ] = P(Ak,n ),
n
k=1
and
k−1 k−1 k k−1 k
Z
k
µX [ , )≤ x dµX ≤ µX [ , ).
n n n [ k−1 k
n ,n)
n n n
By summing we get Z
1
E Xn − ≤ x dµX ≤ E[Xn ].
n [0,∞)
By letting n go to infinity we get the result. The general case can be done by writing X = X + − X − . We
omit the details.
In particular, the expectation of a random variable depends only on its distribution and not on the
probability space on which it is defined. If X has a density f , then the measure µX is the same as f (x) dx
so we can write (as in elementary courses)
Z ∞
E[X] = x f (x) dx,
−∞
Since expectation is defined using the Lebesgue integral, all results about the Lebesgue integral (e.g.,
linearity, Monotone Convergence Theorem, Fatou’s Lemma, Dominated Convergence Theorem) also hold for
expectation.
Exercise 2.2. Use Hölder’s inequality to show that if X is a random variable and q ≥ 1, then E[|X|q ] ≥
(E[X])q . For which X does equality hold?
Example Consider the coin flipping random variables (2) and (3). Clearly
1 1 1
E[Xn ] = 1 · +0· = .
2 2 2
Hence
n
E[Sn ] = E[X1 + · · · + Xn ] = E[X1 ] + · · · + E[Xn ] = .
2
Also,
n
E[X1 + X1 + · · · + X1 ] = E[X1 ] + E[X1 ] + · · · + E[X1 ] = .
2
Note that in the above example, X1 + · · · + Xn and X1 + · · · + X1 do not have the same distribution,
yet the linearity rule for expectation show that they have the same expectation. A very powerful (although
trivial!) tool in probability is the linearity of the expectation even for random variables that are “correlated”.
If X1 , X2 , . . . are nonnegative we can even take countable sums,
"∞ # ∞
X X
E Xn = E[Xn ].
n=1 n=1
This can be derived from the rule for finite sums and the monotone convergence theorem by considering the
increasing sequence of random variables
X n
Yn = Xj .
j=1
7
Lemma 2.3. Suppose X is a random variable with distribution µX , and g is a Borel measurable function.
Then, Z
E[g(X)] = g(x) dµX .
R
Proof. Assume first that g is a nonnegative function. Then there exists an increasing sequence of nonnegative
simple functions with gn approaching g. Note that gn (X) is then a sequence of nonnegative simple random
variables approaching g(X). For the simple functions gn it is immediate from the definition that
Z
E[gn (X)] = gn (x) dµX .
R
The general case can be done by writing g = g + − g − and g(X) = g + (X) − g − (X).
If X has a density f , then we get the following formula found in elementary courses
Z ∞
E[g(X)] = g(x) f (x) dx.
−∞
Note that
The variance is often denoted σ 2 = σ 2 (X) and the square root of the variance, σ, is called the standard
deviation
There is a slightly different use of the term standard deviation in statistics; see any textbook on statistics
for a discussion.
Var(X)
P{|X − E(X)| ≥ a} ≤ .
a2
8
Proof. Let Xa = a1{|X|≥a} . Then Xa ≤ |X| and hence
This gives Markov’s inequality. For Chebyshev’s inequality, we apply Markov’s inequality to the random
variable [X − E(X)]2 to get
E([X − E(X)]2 )
P{[X − E(X)]2 ≥ a2 } ≤ .
a2
Exercise 2.5 (Generalized Chebyshev Inequality). Let f : [0, ∞) → [0, ∞) be a nondecreasing Borel function
and let X be a nonnegative random variable. Then for all a > 0,
E[f (X)]
P{X ≥ a} ≤ .
f (a)
3 Independence
Throughout this section we will assume that there is a probability space (Ω, A, P). All σ-algebras will be
sub-σ-algebras of A, and all random variables will be defined on this space.
Definition
• A collection of events {Aα } is called pairwise independent if for each distinct Aα1 , Aα2 ,
Clearly independence implies pairwise independence but the converse is false. An easy example can be
seen by rolling two dice and letting A1 = { sum of rolls is 7 }, A2 = { first roll is 1 }, A3 = { second roll is
6 }. Then
1
P(A1 ) = P(A2 ) = P(A3 ) = ,
6
1
P(A1 ∩ A2 ) = P(A1 ∩ A3 ) = P(A2 ∩ A3 ) = ,
36
but P(A1 ∩ A2 ∩ A3 ) = 1/36 6= P(A1 )P(A2 )P(A3 ).
9
In fact,
F(X) = {X −1 (B) : B Borel }.
(Clearly F(X) contains the right hand side, and we can check that the right hand side is a σ-algebra.) If
{Xα } is a collection of random variables, then F({Xα }) is the smallest σ-algebra F such that each Xα is
F-measurable.
Probabilists think of a σ-algebra as “information”. Roughly speaking, we have the information in the
σ-algebra F if for every event E ∈ F we know whether or not the event E has occured. A random
variable X is F-measurable if we can determine the value of X by knowing whether or not E has occured
for every F-measurable event E. For example suppose we roll two dice and X is the sum of the two dice.
To determine the value of X it suffices to know which of the following events occured:
Then F 0 is an algebra.
Exercise 3.2. Suppose X is a random variable and g is a Borel measurable function. Let Y = g(X). Then
F(Y ) ⊂ F(X). The inclusion can be a strict inclusion.
The following is a very useful way to check independence.
Lemma 3.3. Suppose F 0 and G 0 are two algebras of events that are independent, i.e., P(A∩B) = P(A) P(B)
for A ∈ F 0 , B ∈ G 0 . Then F := σ(F 0 ) and G := σ(G 0 ) are independent σ-algebras.
Proof. Let B ∈ G 0 and define the measure
Note that µB , µ̃B are finite measures and they agree on F 0 . Hence (using Cartheodory extension) they must
agree on F. Hence P(A ∩ B) = P(A) P(B) for A ∈ F, B ∈ G 0 . Now take A ∈ F and consider the measures
If X1 , . . . , Xn are random variables, we can consider them as a random vector (X1 , . . . , Xn ) and hence
as a random variable taking values in Rn . The joint distribution of the random variables is the measure
µ = µX1 ,...,Xn on Borel sets in Rn given by
Each random variable X1 , . . . , Xn also has a distribution on R, µXi ; these are called the marginal distribu-
tions. The joint distribution function F = FX1 ,...,Xn is defined by
10
Definition Random variables X1 , X2 , . . . , Xn are said to be independent if any of these (equivalent) condi-
tions hold:
The random variables X1 , . . . , Xn have a joint density (joint probability density function) f (x1 , . . . , xn )
if for all Borel sets B ∈ Rn ,
Z
P{(X1 , . . . , Xn ) ∈ B} = f (x1 , . . . , xn ) dx1 dx2 · · · dxn .
B
•
f (x1 , . . . , xn ) = fX1 (x1 ) · · · fXn (xn ),
where fXj (xj ) denotes the density of Xj .
An infinite collection of random variables is called independent if each finite subcollection is independent.
The following follows immediately from the definition and Exercise 3.2.
Proposition 3.4. If X1 , . . . , Xn are independent random variables and g1 , . . . , gn are Borel measurable
functions, then g1 (X1 ), . . . , gn (Xn ) are independent.
Proposition 3.5. If X, Y are independent random variables with E|X| < ∞, E|Y | < ∞, then E[|XY |] < ∞
and
E[XY ] = E[X] E[Y ]. (5)
Proof. First consider simple random variables
m
X n
X
S= ai 1Ai , T = bj 1Bj ,
i=1 j=1
where S is F(X) measurable and T is F(Y ) measurable. Then each Ai is independent of each Bj and (5) can
be derived immediately using P(Ai ∩ Bj ) = P(Ai ) P(Bj ). Suppose X, Y are nonnegative random variables.
Then there exists simple random variables Sn , measurable with respect to F(X), that increase to X and
simple random variables Tn , measurable with respect to F(Y ), increasing to Y . Note that Sn Tn increases
to XY . Hence by the monotone convergence theorem,
11
The multiplication rule for expectation can also be viewed as an application of Fubini’s Theorem. Let
µ, ν denote the distributions of X and Y . If they are independent, then the distribution of (X, Y ) is µ × ν.
Hence Z Z Z Z
E[XY ] = xy d(µ × ν) = xy dµ(x) dν(y) = E[X]y dν(y) = E[X] E[Y ].
R2
Random variables X, Y that satisfy E[XY ] = E[X]E[Y ] are called orthogonal. The last proposition shows
that independent integrable random variables are orthogonal.
Exercise 3.6. Give an example of orthogonal random variables that are not independent.
Proposition 3.7. Suppose X1 , . . . , Xn are random variables with finite variance. If X1 , . . . , Xn are pairwise
orthogonal, then
Var[X1 + · · · + Xn ] = Var[X1 ] + · · · + Var[Xn ].
Proof. Without loss of generality we may assume that all the random variables have zero mean (otherwise,
we subtract the mean without changing any of the variances). Then
Xn Xn
Var Xj = E ( Xj )2
j=1 j=1
n
X X
= E[Xj2 ] + E[Xi Xj ]
j=1 i6=j
n
X X
= Var(Xj ) + E[Xi Xj ].
j=1 i6=j
The last proposition is really just a generalization of the Pythagorean Theorem. A natural framework
for discussing random variables with zero mean and finite variance is the Hilbert space L2 with the inner
product hX, Y i = E[XY ]. In this case, X, Y are orthogonal iff hX, Y i = 0.
Exercise 3.8. Let (Ω, F, P) be the unit interval with Lebesgue measure on the Borel subsets. Follow this
outline to show that one can find independent random variables X1 , X2 , X3 , . . . defined on (Ω, F, P), each
normal mean zero, variance 1. Let use write x ∈ [0, 1] in its dyadic expansion
as in Section 1.
(i) Show that the random variable X(x) = x is a uniform random variable on [0, 1], i.e., has density
f (x) = 1[0,1] .
(ii) Use a diagonalization process to define U1 , U2 , . . ., independent random variables each with a uniform
distribution on [0, 1].
(iii) Show that Xj = Φ−1 (Uj ) has a normal distribution with mean zero, variance one. Here Φ is the
normal distribution function.
12
4 Sums of independent random variables
Definition
• A sequence of random variables Yn converges to Y almost surely (a.s.) or with probability one (w.p.1)
if there is an event E with P (E) = 1 and such that for ω ∈ E, Yn (ω) → Y (ω).
In other words, “in probability” is the same as “in measure” and “almost surely” is the analogue of “almost
everywhere”. The term almost surely is not very accurate (something that happens with probability one
happens surely!) so many people prefer to say “with probability one”.
Exercise 4.1. Suppose Y1 , Y2 , . . . is a sequence of random variables with E[Yn ] → µ and Var[Yn ] → 0. Show
that Yn → µ in probability. Give an example to show that it is not necessarily true that Yn → µ w.p.1.
Much of the focus of classical probability is on sums of independent random variables. Let (Ω, A, P) be
a probability space and assume that
X1 , X2 , . . .
are independent random variables on this probability space. We say that the random variables are identically
distributed if they all have the same distribution, and we write i.i.d. for “independent and identically
distributed.”
If X1 , X2 , . . . are independent (or, more generally, orthogonal) random variables each with mean µ and
variance σ 2 , then
X1 + · · · + Xn 1
E = [E(X1 ) + · · · + E(Xn )] = µ,
n n
σ2
X1 + · · · + Xn 1
Var = 2 [Var(X1 ) + · · · + Var(Xn )] = .
n n n
The equality for the expectations does not require the random variables to be orthogonal, but the expression
for the variance requires it. The mean of the weighted average is the√same as the mean of one of the random
variables, but the standard deviation of the weighted average is σ/ n which tends to zero.
Theorem 4.2 (Weak Law of Large Numbers). If X1 , X2 , . . . are independent random variables such that
E[Xn ] = µ and Var[Xn ] ≤ σ 2 for each n, then
X1 + · · · + Xn
−→ µ
n
in probability.
Proof. We have
σ2
X1 + · · · + Xn X1 + · · · + Xn
E = µ, Var ≤ → 0.
n n n
We now use Exercise 4.1.
13
The Strong Law of Large Numbers (SLLN) is a statement about almost sure convergence of the weighted
sums. We will prove a version of the strong law here. We start with an easy, but very useful lemma. Recall
that the limsup of events is defined by
∞ [
\ ∞
lim sup An = Am .
n=1 m=n
Probabilists often write {An i.o.} for lim sup An ; here “i.o.” stands for “infinitely often”.
The conclusion in (ii) is not necessarily true for dependent events. For example, if A is an event of
probability 1/2, and A1 = A2 = · · · = A, then lim sup An = A which has probability 1/2.
Proof. P
(i) If P(An ) < ∞,
∞ ∞
!
[ X
P(lim sup An ) = lim P An ≤ lim P(Am ) = 0.
n→∞ n→∞
m=n m=n
P
(ii) Assume P(An ) = ∞ and A1 , A2 , . . . are independent. We will show that
"∞ ∞ #
[ \
c c
P[(lim sup An ) ] = P Am = 0.
n=1 m=n
But by independence,
M
! M
( M
)
\ Y X
P Acm = [1 − P(Am )] ≤ exp − P(Am ) → 0.
m=n m=n n=m
14
Theorem 4.4 (Strong Law of Large Numbers). Let X1 , X2 , . . . be independent random variables each with
mean µ. Suppose there exists an M < ∞ such that E[Xn4 ] ≤ M for each n (this holds, for example, when
the sequence is i.i.d. and E[X14 ] < ∞). Then w.p.1,
X1 + X2 + · · · + Xn
−→ µ.
n
Proof. Without loss of generality we may assume that µ = 0 for otherwise we can consider the random
variables X1 − µ, X2 − µ, . . .. Our first goal is to show that
E[(X1 + X2 + · · · + Xn )4 ] ≤ 3M n2 . (6)
When we expand (X1 + X2 + · · · + Xn )4 and then use the sum rule for expectation we get n4 terms. We get
n terms of the form E[Xi4 ] and 3n(n − 1) terms of the form E[Xi2 Xj2 ] with i 6= j. There are also terms of the
form E[Xi Xj3 ], E[Xi Xj Xk2 ] and E[Xi Xj Xk Xl ] for distinct i, j, k, l; all of these expectations are zero by (7) so
we will not bother to count the exact number. We know that E[Xi4 ] ≤ M . Also (see Exercise 2.2), if i 6= j,
1/2
E[Xi2 Xj2 ] ≤ E[Xi2 ] E[Xj2 ] ≤ E[Xi4 ] E[Xj4 ] ≤ M.
Since the sum consists of at most 3n2 terms each bounded by M , we get (6).
Let An be the event
X1 + · · · + Xn
≥ 1 7/8
An =
n n1/8 = {|X1 + X2 + · · · + Xn | ≥ n }.
E[(X1 + · · · + Xn )4 ] 3M
P(An ) ≤ ≤ 3/2 .
(n7/8 )4 n
P
In particular, P(An ) < ∞. Hence by the Borel-Cantelli Lemma,
P(lim sup An ) = 0.
Suppose for some ω, (X1 (ω) + · · · + Xn (ω))/n does not converge to zero. Then there is an > 0 such that
for infinitely many values of n. This implies that ω ∈ lim sup An . Hence the set of ω at which we do not
have convergence has probability zero.
The strong law of large numbers holds in greater generality than under the conditions given here. In
fact, if X1 , X2 , . . . is any i.i.d. sequence of random variables with E(X1 ) = µ, then the strong law holds
X1 + · · · + Xn
→ µ, w.p.1.
n
We will not prove this here. However, as the next exercise shows, the strong law does not hold for all
independent sequences of random variables with the same mean.
15
Exercise 4.5. Let X1 , X2 , . . . be independent random variables. Suppose that
P{|Xn | ≥ n i.o.} = 0
F1 ⊂ F2 ⊂ · · · , G1 ⊃ G2 ⊃ · · · .
Let F be the σ-algebra generated by X1 , X2 , . . ., i.e., the smallest σ-algebra containing the algebra
∞
[
F 0 := Fn .
n=1
We write 4 for symmetric difference, A4B = (A \ B) ∪ (B \ A). The next lemma says roughly that F0 is
“dense” in F.
Lemma 4.7. Suppose F 0 is an algebra of events and F = σ(F 0 ). If A ∈ F, then for every > 0 there is
an A0 ∈ F 0 with
P(A4A0 ) < .
16
Proof. Let B be the collection of all events A such that for every > 0 there is an A0 ∈ F 0 such that
P(A4A0 ) < . Trivially, F 0 ⊂ B. We will show that B is a σ-algebra and this will imply that F ⊂ B.
Obviously, Ω ∈ B.
Suppose A ∈ B and let > 0. Find A0 ∈ F 0 such that P(A4A0 ) < . Then Ac0 ∈ F 0 and
Hence Ac ∈ B.
Suppose A1 , A2 , . . . ∈ B and let A = ∪Aj . Let > 0, and find n such that
n
[
P Aj ≥ P[A] − .
j=1
2
Proof. Let A ∈ T and let > 0. Find a set A0 ∈ F 0 with P(A4A0 ) < . Then A0 ∈ Fn for some n. Since
T ⊂ Gn and Fn is independent of Gn , A and A0 are independent. Therefore P(A ∩ A0 ) = P(A)P(A0 ). Note
that P(A ∩ A0 ) ≥ P(A) − P(A4A0 ) ≥ P(A) − . Letting → 0, we get P(A) = P(A)P(A). This implies
P(A) = 0 or P(A) = 1.
As an example, suppose X1 , X2 , . . . are independent random variables. Then the probability of the event
X1 (ω) + · · · + Xn (ω)
ω : lim =0
n→∞ n
is zero or one.
We will call g a Schwartz function it is C ∞ and all of its derivatives decay at ±∞ faster than every polynomial.
If g is Schwartz, then ĝ is Schwartz. (Roughly speaking, the Fourier transform sends smooth functions to
bounded functions and bounded functions to smooth functions. Very smooth and very bounded functions
are sent to very smooth and very bounded functions.) The inversion formula
Z ∞
1
g(x) = eixy ĝ(y) dy (8)
2π −∞
17
holds for Schwartz functions. A straightforward computation shows that
√ 2
−x2 /2 =
e\ 2πe−y /2 .
We also need
Z ∞ Z ∞ Z ∞
−ixy
f (x) ĝ(x) dx = f (x) g(y) e dy
−∞ −∞ −∞
Z ∞ Z ∞
= g(y) f (x) e−ixy dx dy
−∞ −∞
Z∞
= fˆ(y) g(y) dy.
−∞
This is valid provided that the interchange of integrals can be justified; in particular, it holds for Schwartz
f, g. The characteristic function is a version of the Fourier transform.
• φ is a continuous function of t. To see this note that the collection of random variables Ys = eisX are
dominated by the random variable 1 which has finite expectation. Hence by the dominated convergence
theorem, h i
lim φ(s) = lim E[eisX ] = E lim eisX = φ(t).
s→t s→t s→t
• The function MX (t) = etX is often called the moment generating function of X. Unlike the character-
istic function, the moment generating function does not always exist for all values of t. When it does
exist, however, we can use the formal relation φ(t) = M (it).
• If a random variable X has a density f then
Z ∞
φX (t) = eitx f (x) dx = fˆ(−t).
−∞
• If two random variables have the same characteristic function, then they have the same distribution.
This fact is not obvious — this will follow from Proposition 5.6 which gives an inversion formula for
the characteristic function similar to (8).
• If X1 , . . . , Xn are independent random variables, then
18
• If X is a normal random variable, mean zero, variance 1, then by completing the square in the expo-
nential one can compute Z ∞
1 2 2
φ(t) = eixt √ e−x /2 dx = e−t /2 .
−∞ 2π
If Y is normal mean µ, variance σ 2 , then Y has the same distribution as σX + µ, hence
2 2
φY (t) = φσX+µ (t) = eiµt φX (σt) = eiµt e−σ t /2
.
Hence, Z ∞
0
φ (t) = ixeitx dµ(x),
−∞
and, in particular,
φ0 (0) = i E[X].
19
Formally, (9) can be derived by writing
∞ j j ∞ j
X (iX) t X i E(X j ) j
E[eiXt ] = E = t ,
j=0
j! j=0
j!
differentiating both sides, and evaluating at t = 0. Proposition 5.3 tells us that the existence of E[|X|k ]
allows one to justify this rigorously up to the kth derivative.
Proposition 5.4. Let X1 , X2 , . . . be independent, identically distributed random variables with mean µ and
variance σ 2 and characteristic function φ. Let
X1 + · · · + Xn − nµ
Zn = √ ,
σ2 n
and let φn be the characteristic function of Zn . Then for each t,
2
lim φn (t) = e−t /2
.
n→∞
2
Recall that e−t /2
is the characteristic function of a normal mean zero, variance 1 random variable.
Proof. Without loss of generality we will assume µ = 0, σ 2 = 1 for otherwise we can consider Yj = (Xj −µ)/σ.
By Proposition 5.3, we have
1 t2
φ(t) = 1 + φ0 (0) t + φ00 (0) t2 + t t2 = 1 − + t t2 ,
2 2
where t → 0 as t → 0. Note that n
t
φn (t) = φ √ ,
n
and hence for fixed t,
n n
t2
t δn
lim φn (t) = lim φ √ = lim 1 − + ,
n→∞ n→∞ n n→∞ 2n n
where δn → 0 as n → ∞. Once n is sufficiently large that |δn − (t2 /2)| < n we can take logarithms and
expand in a Taylor series,
t2 t2
δn ρn
log 1 − + =− + , n→∞
2n n 2n n
where ρn → 0. (Since φ can take on complex values, we need to take a complex logarithm with log 1 = 0,
but there is no problem provided that |δn − (t2 /2)| < n.) Therefore
t2
lim log φn (t) = − .
n→∞ 2
Theorem 5.5 (Central Limit Theorem). Let X1 , X2 , . . . be independent, identically distributed random
variables with mean µ and finite variance. If −∞ < a < b < ∞, then
Z b
X1 + X2 + · · · + Xn − nµ 1 2
lim P a ≤ √ ≤b = √ e−x /2 dx.
n→∞ σ n a 2π
20
Proof. Let φn be the characteristic function of n−1/2 σ −1 (X1 + · · · + Xn − nµ). We have already shown that
2
for each t, φn (t) → e−t /2 . It suffices to show that if µn is any sequence of distributions such that their
2
characteristic functions φn converge pointwise to e−t /2 then for every a < b,
where Z x
1 2
Φ(x) = √ e−y /2 dy.
−∞ 2π
∞
Fix a, b and let > 0. We can find a C function g = ga,b, such that
g(x) = 1, a ≤ x ≤ b,
g(x) = 0, x∈
/ [a − , b + ].
We will show that Z ∞ Z ∞
1 2
lim g(x) dµn (x) = g(x) √ e−x /2 dx. (10)
n→∞ −∞ −∞ 2π
Since for any distribution µ, Z ∞
µ[a, b] ≤ g(x) dµ(x) ≤ µ[a − , b + ],
−∞
Since |φn ĝ| ≤ |ĝ| and ĝ is L1 , we can use the dominated convergence theorem to conclude
Z ∞ Z ∞
1 2
lim g(x) dµn (x) = e−y /2 ĝ(y) dy.
n→∞ −∞ 2π −∞
√ 2
−x2 /2 =
Recalling that e\ 2πe−y /2 , and that
Z Z
fˆg = f ĝ,
we get (10).
The next proposition establishes the uniqueness of characteristic functions by giving an inversion formula.
Note that in order to specify a distribution µ, it suffices to give µ[a, b] at all a, b with µ{a, b} = 0.
21
Proposition 5.6. Let X be a random variable with distribution µ, distribution function F , and characteristic
function φ. Then for every a < b such that F is continuous at a and b,
T
e−iya − e−iyb
Z
1
µ[a, b] = F (b) − F (a) = lim φ(y) dy. (11)
T →∞ 2π −T iy
but this integrand is not necessarily integrable on (−∞, ∞). We can write (12) if we interpret this as
the improper Riemann integral and not as the Lebesgue integral on (−∞, ∞).
and hence the integral on the right-hand side of (11) is well defined. The bound (13) allows us to use Fubini’s
Theorem:
Z T −iya Z T −iya Z ∞
− e−iyb − e−iyb
1 e 1 e iyx
φ(y) dy = e dµ(x) dy
2π −T iy 2π −T iy −∞
"
Z ∞ Z T iy(x−a) #
1 e − eiy(x−b)
= dy dµ(x)
2π −∞ −T iy
22
We do not need to worry about what happens at x = a or x = b since µ gives zero measure to these points.
Since the function Z T iy(x−a)
e − eiy(x−b)
gT (a, b, x) = dy
−T iy
is uniformly bounded in T, a, b, x, we can use dominated convergence theorem to conclude that
Z ∞ "Z T iy(x−a) #
− eiy(x−b)
Z
e
lim dy dµ(x) = 2π 1(a,b) dµ(x) = 2πµ[a, b].
T →∞ −∞ −T iy
Therefore,
T
e−iya − e−iyb
Z
1
lim φ(y) dy = µ[a, b].
T →∞ 2π −T iy
Exercise 5.7. A random variable X has a Poisson distribution with parameter λ > 0 if X takes on only
nonnegative integer values and
λk
P{X = k} = e−λ , k = 0, 1, 2, . . . .
k!
In this problem, suppose that X, Y are independent Poisson random variables on a probability space with
parameters λX , λY , respectively.
(i) Find E[X], Var[X], E[Y ], Var[Y ].
(ii) Show that X + Y has a Poisson distribution with parameter λX + λY by computing directly
P{X + Y = k}.
(iii) Find the characteristic function for X, Y and use this to find an alternative deriviation of (ii).
(iv) Suppose that for each integer n > λ we have independent random variables Z1,n , Z2,n , . . . , Zn,n with
λ
P{Zj,n = 1} = 1 − P{Zj,n = 0} = .
n
Find
lim P{Z1,n + Z2,n + · · · + Zn,n = k}.
n→∞
Exercise 5.8. A random variable X is infinitely divisible if for each positive integer n, we can find i.i.d. ran-
dom variables X1 , . . . , Xn such that X1 + · · · + Xn have the same distribution as X.
(i) Show that Poisson random variables and normal random variables are infinitely divisible.
(ii) Suppose that X has a uniform distribution on [0, 1]. Show that X is not infinitely divisible.
(iii) Find all infinitely divisible distributions that take on only a finite number of values.
6 Conditional Expectation
6.1 Definition and properties
Suppose X is a random variable with E[|X|] < ∞. We can think of E[X] as the best guess for the random
variable given no information about the outcome of the random experiment which produces the number X.
Conditional expectation deals with the case where one has some, but not all, information about an event.
Consider the example of flipping coins, i.e., the probability space
Ω = {(ω1 , ω2 , . . .) : ωj = 0 or 1}.
23
Let Fn be the σ-algebra of all events that depend on the first n flips of the coins. Intutitively we think of the
σ-algebra Fn as the “information” contained in viewing the first n flips. Let Xn (ω) = ωn be the indicator
function of the event “the nth flip is heads”. Let
Sn = X1 + · · · + Xn
be the total number of heads in the first n flips. Clearly E[Sn ] = n/2. Suppose we are interested in S3 but
we get to see the flip of the first coin. Then our “best guess” for S3 depends on the value of the first flip.
Using the notation of elementary probability courses we can write
E[S3 | X1 = 1] = 2, E[S3 | X1 = 0] = 1.
More formally we can write E[S3 | F1 ] for the random variable that is measurable with respect to the
σ-algebra F1 that takes on the value 2 on the event {X1 = 1} and 1 on the event {X1 = 0}.
Let (Ω, F, P) be a probability space and G ⊂ F another σ-algebra. Let X be an integrable random
variable.
Let us first consider the case when G is a finite σ-algebra. In this case there exist disjoint, nonempty
events A1 , A2 , . . . , Ak that generate G in the sense that G consists precisely of those sets obtained from unions
of the events A1 , . . . , Ak . If P(Aj ) > 0, we define E[X | G] on Aj to be the average of X on Aj ,
R
A
X dP E[X1Aj ]
E[X | G] = j = , on Aj . (14)
P(Aj ) P(Aj )
If P(Aj ) = 0, this definition does not make sense, and we just let E[X | G] take on an arbitrary value, say
cj , on Aj . Note that E[X | G] is a G-measurable random variable and that its definition is unique except
perhaps on a set of total probability zero.
Proposition 6.1. If G is a finite σ-algebra and E[X | G] is defined as in (14), then for every event A ∈ G,
Z Z
E[X | G] dP = X dP.
A A
l
R
X Aj
X dP
= P(Aj )
j=1
P(Aj )
l Z
X
= X dP
j=1 Aj
Z
= X dP.
A
24
Proof. The uniqueness is immediate since if Y, Ỹ are G-measurable random variables satisfying (15), then
Z
(Y − Ỹ ) dP = 0.
A
Note that ν << P . Hence by the Radon-Nikodym theorem there is a G-measurable random variable Y with
Z
ν(A) = Y dP.
A
Using the proposition, we can define the conditional expectation E[X | G] to be the random variable Y
in the proposition. It is characterized by the properties
• E[X | G] is G-measurable.
• For every A ∈ G, Z Z
E[X | G] dP = X dP. (16)
A A
Exercise 6.3. This exercise gives an alternative proof of the existence of E[X | G] using Hilbert spaces (but
not using Radon-Nikodym theorem). Let H denote the real Hilbert space of all square-integrable random
variables on (Ω, F, P).
1. Show that the space of G-measurable, square-integrable random varaibles is a closed subspace of H.
2. For square-integrable X, define E(X | G) to be the Hilbert space projection onto the space of G-
measurable random variables. Show that E(X | G) satisfies (16).
3. Show that E(X | G) exists for integrable X. (Hint: if X ≥ 0, we can approximate X from below by
simple random variables Xn . Define
25
Proof. Note that a E[X | G] + b E[Y | G] is G-measurable. Also, if A ∈ G,
Z Z Z
(a E[X | G] + b E[Y | G]) dP = a E[X | G] dP + b E[Y | G] dP P
A A A
Z Z
= a X dP + b Y dP
Z A A
= (aX + bY ) dP.
A
• If Y is G-measurable, then
E[XY | G] = Y E[X | G]. (18)
We will R case where X, Y ≥ 0. Note that this implies that E[X | G] ≥ 0 a.s.,
R first consider the
since A E[X | G] dP = A X dP ≥ 0 for every A ∈ G. Find simple G-measurable random variables
0 ≤ Y1 ≤ Y2 ≤ · · · such that Yn → Y . Then Yn X converges monotonically to Y X and Yn E[X | G]
converges monotonically to Y E[X | G]. If A ∈ G and
n
X
Z= cj 1Bj , Bj ∈ G,
j=1
26
is a G-measurable simple random variable,
Z n
X Z
Z E[X | G] dP = cj 1Bj E[X | G] dP
A j=1 A
Xn Z
= cj E[X | G] dP
j=1 A∩Bj
Xn Z
= cj X dP
j=1 A∩Bj
Z X n
= [ cj 1Bj ]X dP
A j=1
Z
= ZX dP.
A
Hence (19) holds for simple nonnegative Y and nonnegative X, and hence by the monotone convergence
theorem it holds for all nonnegative Y and nonnegative X. For general Y, X, we write Y = Y + −
Y − , X = X + − X − and use linearity of expectation and conditional expectation.
If X, Y are random variables and X is integrable we write E[X | Y ] for E[X | σ(Y )], Here σ(Y ) denotes
the σ-algebra generated by Y . Intuitively, E[X | Y ] is the best guess of X given the value of Y . Note that
E[X | Y ] can be written as φ(Y ) for some function φ, i.e., for each possible value of Y there is a value of
E[X | Y ]. Elementary texts often write this as E[X | Y = y] to indicate that for each value of y there is a
value of E[X | Y ]. Similarly, we can define E[X | Y1 , . . . Yn ].
Example Consider the coin-tossing random variables discussed earlier in the section. What is E[X1 | Sn ]?
Symmetry tells us that
E[X1 | Sn ] = E[X2 | Sn ] = · · · = E[Xn | Sn ].
Also, linearity gives
E[X1 + · · · + Xn | Sn ] = E[Sn | Sn ] = Sn .
Hence,
Sn
E[Xn | Sn ] = .
n
Note that E[X1 | Sn ] is not the same thing at E[X1 | X1 , . . . , Xn ] = E[X1 | Fn ] which would be just X1 .
The first conditioning is given only the number of heads on the first n flips while the second conditioning is
given all the outcomes of the n flips.
Example Suppose X, Y are two random variables with joint density function f (x, y). Let us take our
probability space to be R2 , let F be the Borel sets, and let P be the measure
Z
P(A) = f (x, y) dx dy.
A
Then the random variables X(x, y) = x and Y (x, y) = y have the joint density f (x, y). Then E[Y | X] is a
random variable with density
f (X, y)
f (y | X) = R ∞ .
−∞
f (X, z) dz
27
This formula is not valid when the denominator is zero. However, the set of x such that
Z ∞
f (x, z) dz = 0,
−∞
is a null set under the measure P and hence the density f (y | X) is well defined almost surely. This density
is often called the conditional density of Y given X.
6.2 Martingales
A martingale is a model of a fair game. Suppose we have a probability space (Ω, F, P) and an increasing
sequence of σ-algebras
F0 ⊂ F1 ⊂ · · ·
Such a sequence is called a filtration, and we think of Fn as representing the amount of knowledge we have
at time n. The inclusion of the σ-algebras imply that we do not lose any knowledge that we have gained.
Definition A martingale (with respect to the filtration {Fn }) is a sequence of integrable random variables
Mn such that Mn is Fn -measurable for each n and for all n < m,
E[Mm | Fn ] = Mn . (20)
E[Mn+1 | Fn ] = Mn .
This follows from successive applications of (17). If the filtration is not specified explicitly, then one assumes
that one uses the filtration generated by Mn , i.e., Fn is the σ-algebra generated by M1 , . . . , Mn .
Examples
We list some examples. We leave it as exercises to show that they are martingales.
Then Wn is a martingale.
28
• Suppose X1 , X2 , . . . are independent coin-flipping random variables, i.e., with distribution
1
P{Xj = 1} = P{Xj = −1} = .
2
A particular case of the last example is where B0 = 1 and Bn = 2n if X1 = X2 = · · · = X−1 = −1 and
Bn = 0 otherwise. In other words, one keeps doubling one’s bet until one wins. Note that
We want to prove a version of the optional stopping theorem which is a mathematical way of saying that
one cannot beat a fair game. The last example above shows that we need to be careful — if I play a fair
game and keep doubling my bet, I will eventually win and come out ahead (provided that I always have
enough money to put down the bet when I lose!)
Definition A stopping time with respect to a filtration {Fn } is a random variable taking values in {0, 1, 2 . . .}∪
{∞} such that for each n, the event
{T ≤ n} ∈ Fn .
In other words, we only need the information at time n to know whether we have stopped at time n.
Examples of stopping times are:
• Constant random variables are stopping times.
• If Mn is a martingale with respect to {Fn } and V ⊂ R is a Borel set, then
T = min{j : Mj ∈ V }
is a stopping time.
• If T1 , T2 are stopping times, then so are T1 ∧ T2 and T1 ∨ T2 .
• In particular, if T is a stopping time then so are Tn = T ∧ n. Note that T0 ≤ T1 ≤ . . . and T = lim Tn .
Proposition 6.4. If Mn is a martingale and T is a stopping time with respect to {Fn }, then Yn = MT ∧n
is a martingale with respect to {Fn }. In particular, E[Yn ] = E[Y0 ] = E[M0 ].
Proof. Note that
n−1
X
Yn = Mn 1{T ≥ n} + Mj 1{T = j}.
j=0
From this it is easy to see that Yn is Fn -measurable and E[|Yn |] < ∞. Also note that the event 1{T ≥ n + 1}
is the complement of the event 1{T ≤ n} and hence is Fn -measurable. Therefore, (18) implies
Examples
29
• Let X1 , X2 , . . . be independent, each with P{Xj = 1} = P{Xj = −1} = 1/2. Let Mn = 1+X1 +· · ·+Xn ,
N an integer greater than 1, and
T = min{j : Mn = 0 or Mn = N }.
T is the first time that a “random walker” starting at 1 reaches 0 or N . Then Mn∧T is a martingale.
In particular,
1 = E[M0 ] = E[Mn∧T ].
It is easy to check that w.p.1 T < ∞. Since Mn∧T is a sequence of bounded random variables
approaching MT , the dominated convergence theorem implies
E[MT ] = 1.
Note that
E[MT ] = N P{MT = N },
and hence we have computed a probability,
1
P{MT = N } = .
N
T = min{n : |Sn | = N }.
Indeed, we can justify this using the dominated convergence. It is not difficult to show (Exercise 6.5)
that E[T ] < ∞, and hence Mn∧T is dominated by the integrable random variable N 2 + T . This justifies
(22). Note that this implies that
E[T ] = N 2 .
Then WT ∧n is a martingale, and hence E[WT ∧n ] = E[W0 ] = 0. This could be checked easily using (21).
Note that P{T < ∞} = 1 and
lim WT ∧n = WT = 1.
T →∞
Hence,
1 = E[WT ] 6= lim E[WT ∧n ] = 0.
T →∞
In this case there is no integrable random variable Y such that |WT ∧N | ≤ Y for all N .
30
Exercise 6.5. Consider the second example above.
• Show that there exists c < ∞, ρ < 1 (depending on N ) such that for all n,
P{T > n} ≤ c ρn .
7 Brownian motion
Brownian motion is a model for random continuous motion. We will consider only the one-dimensional
version. Imagine that Bt = Bt (ω) denotes the position of a random particle at time t. For fixed t, Bt (ω) is a
random variable. For fixed ω, we have a function t 7→ Bt (ω). Hence, we can consider the Brownian motion
either as a collection of random varibles Bt parametrized by time or as a random function from [0, ∞) into
R.
Definition A collection of random variables Bt , t ∈ [0, ∞) defined on a probability space (Ω, F, P) is called
a Brownian motion if it satisfies the following.
(i) B0 = 0 w.p.1.
(ii) For each 0 ≤ s1 ≤ t1 ≤ s2 ≤ t2 ≤ · · · ≤ sn ≤ tn , the random variables
are independent.
(iii) For each 0 ≤ s ≤ t, the random variable Bt − Bs has the same distribution as Bt−s .
(iv) W.p.1., t 7→ Bt (ω) is a continuous function of t.
It is a nontrivial theorem, which we will not prove here, that states that if Bt is a processes satisfying
(i)–(iv) above, then the distribution of Bt − Bs must be normal. (This fact is closely related to but does not
immediately follow from the central limit theorem.) Note that
n h
X i
B1 = B j − B j−1 ,
n n
j=1
so a normal distribution is reasonable. However, one cannot conclude normality only from assumptions
(i)–(iii), see Exercise 5.8. Since we know Brownian motion must have normal increments (but we do not
want to prove it here!), we will modify our definition to include this fact.
Definition A collection of random variables Bt , t ∈ [0, ∞) defined on a probability space (Ω, F, P) is called
a Brownian motion with drift µ and variance parameter σ 2 , if it satisfies the following.
(i) B0 = 0 w.p.1.
(ii) For each 0 ≤ s1 ≤ t1 ≤ s2 ≤ t2 ≤ · · · ≤ sn ≤ tn , the random variables
are independent.
(iii) For each 0 ≤ s ≤ t, the random variable Bt − Bs has a normal distribution with mean µ(t − s) and
variance σ 2 (t − s).
(iv) W.p.1., t 7→ Bt (ω) is a continuous function of t.
Bt is called a standard Brownian motion if µ = 0, σ 2 = 1.
31
Exercise 7.1. Suppose Bt is a standard Brownian motion. Show that B̂t := σBt + µt is a Brownian motion
with drift µ and variance parameter σ 2 .
In this section, we will show that Brownian motion exists. We will take any probability space (Ω, F, P)
sufficiently large that there exist a countable sequence of standard normal random variables N1 , N2 , . . .
defined on the space. Exercise 3.8 shows that the unit interval with Lebesgue measure is large enough. For
ease, we will show how to construct a standard Brownian motion Bt , 0 ≤ t ≤ 1. Let
∞
[
Dn = {j/2n : j = 1, . . . , 2n }, D= Dn ,
n=0
and suppose that the random variables {Nt : t ∈ D} have been relabeled so they are indexed by the countable
set D. The basic strategy is as follows:
• Step 1. Use the random variables Nt , t ∈ D to define random variables Bt , t ∈ D that satisfy (i)–(iii),
restricted to s, t, sj , tj ∈ D.
• Step 2. Show that w.p.1 the function t 7→ Bt , t ∈ D is uniformly continuous.
• Step 3. On the event that t 7→ Bt , t ∈ D is uniformly continuous, define Bt , t ∈ [0, 1] by continuity.
Show this satisfies the defintion.
7.1 Step 1.
We will recursively define Bt for t ∈ Dn . In our construction we will see that at each stage the random
variables
∆(j, n) := Bj2−n − B(j−1)2−n , j = 1, . . . , 2n ,
are independent, normal, variance 2−n . We start with n = 0, and set B0 = 0, B1 = N1 . To do the next step,
we use the following lemma.
Lemma 7.2. Suppose X, Y are independent normal random variables with mean zero, variance σ 2 , and let
Z = X + Y, W = X − Y . Then Z, W are independent random variables with mean zero and variance 2σ 2 .
Proof. For ease, assume σ 2 = 1. If V ∈ R2 is a Borel set, let Ṽ = {(x, y) : (x + y, x − y) ∈ V }. Then
P{(Z, W ) ∈ V } = P{(X, Y ) ∈ Ṽ }
Z Z
1 −(x2 +y2 )/2 1 −(z2 +w2 )/4
= e dx dy = e dz dw.
Ṽ 2π V 4π
32
7.2 Step 2.
Let
Mn = sup{|Bs − Bt | : s, t ∈ D, |s − t| ≤ 2−n }.
Showing continuity is equivalent to showing that w.p.1, Mn → 0 as n → ∞, or, equivalently, for each > 0,
P {Mn > i.o.} = 0.
It is easier to work with a different quantity:
Z(j, n) = sup Bt − B(j−1)2−n : t ∈ D, (j − 1)2−n ≤ t ≤ j2−n ,
Zn = max{Z(j, n) : j = 1, . . . , 2n }.
Note that Mn ≤ 3 Zn . Hence, for any > 0,
n
2
X
P{Mn > 3} ≤ P{Zn > } ≤ P{Z(j, n) > } = 2n P{Z(1, n) > }.
j=1
√
Since Bt has the same distribution as t B1 , a little thought will show that the distribution of Z(1, n) is the
same as the distribution of
2−n/2 Z(1, 0) = 2−n/2 sup{|Bt | : t ∈ D} = 2−n/2 lim max{|Bt | : t ∈ Dm }.
m→∞
Since this holds for all m, Putting this all together, we have
P{Mn > 3} ≤ 2n P{Z(1, n) > 2−n/2 } ≤ 2n exp −2 2n−1 .
Therefore,
∞
X
P{Mn > 3} < ∞,
n=1
and by the Borel-Cantelli lemma,
P{Mn > 3 i.o.} = 0.
33
Exercise 7.3. Let α < 1/2 and
|Bt − Bs |
Y = sup : 0 ≤ s < t ≤ 1 .
(t − s)α
Show that w.p,.1., Y < ∞. Show that the result is not true for α = 1/2.
7.3 Step 3.
This step is straightforward and left as an exercise. One important fact that is used is that if X = lim Xn
and Xn is normal mean zero, variance σn2 and σn → σ, then X is normal mean zero, variance σ 2 .
8 Elementary probability
In this section we discuss some results that would usually be covered in an undergraduate calculus-based
course in probability.
34
n
The binomial coefficient k represents the number of ordered n-tuples of 0s and 1s that have exactly k 1s.
Clearly E[Xj ] = p and
2
Var[Xj ] = E[Xj2 ] − (E[Xj ]) = p − p2 = p (1 − p).
Hence, if Y has a binomial distribution with parameters n, p,
n
X n
X
E[Y ] = E[Xj ] = np, Var[Y ] = Var[Xj ] = np(1 − p).
j=1 j=1
λk
P{X = k} = e−λ , k = 0, 1, 2, . . .
k!
The Taylor series for the exponential shows that
∞
X
P{X = k} = 1.
k=0
The Poisson distribution arises as a limit of the binomial distribuiton as n → ∞ when the expected number
of successes stays fixed at λ. In other words, it is the limit of binomial random variables Yn = Yn,λ with
parameters n and λ/n. This limit is easily derived; if k is a nonnegative integer,
k n−k
n λ λ
lim P{Yn = k} = lim 1−
n→∞ n→∞ k n n
k n −k
n(n − 1) · · · (n − k + 1) λ λ λ
= lim 1− 1−
n→∞ k! n n n
λk e−λ
= .
k!
It is easy to check that
The expectation and variances could be derived from the corresponding quanitities for Yn ,
λ
E[Yn ] = λ, Var[Yn ] = λ 1 − .
n
X represents the number of Bernoulli trials needed until the first success assuming the probability of success
on each trial is p. (The event that the kth trial is the first success is the event that the first k − 1 trials are
failures and the last one is a success. From this, (23) follows immediately.) Note that
∞ ∞
X d X d 1−p 1
E[X] = k (1 − p)k−1 p = −p (1 − p)k = −p = .
dp dp p p
k=1 k=1
35
∞
X
E[X(X − 1)] = k (k − 1) (1 − p)k−1 p
k=1
d2 1 − p
2(1 − p)
= p(1 − p) = .
dp2 p p2
2 1 1−p
E[X 2 ] = − , Var[X] = E[X 2 ] − E[X]2 = .
p2 p p2
The geometric distributions can be seen to the only discrete distributions with support {1, 2, . . .} having the
loss of memory property: for all positive integers j, k,
The term geometric distribution is sometimes also used for the distribution of Z = X − 1,
P{Z = k} = (1 − p)k p, k = 0, 1, 2, . . .
We can think of Z as the number of failures before the first success in Bernoulli trials. Note that
1−p 1−p
E[Z] = E[X] − 1 = , Var[Z] = Var[X] = .
p p2
Note that Z b
F (b) = f (x) dx,
−∞
F 0 (x) = f (x),
36
The distribution function of Z is often denoted by Φ,
Z x
1 2
Φ(x) = √ e−t /2 dt.
0 2π
The normal distribution with mean µ and variance σ 2 , often denoted N (µ, σ 2 ) is the distribution of
X = σ Z + µ. It is easily checked that the density is
1 2
/2σ 2
√ e−(x−µ) .
2πσ 2
Indeed, if F denotes the distribution function for X,
x−µ x−µ
F (x) = P{σZ + µ ≤ x} = P Z ≤ =Φ .
σ σ
Hence,
1 x−µ 1 1 −(x−µ)2 /2σ2
F (x) = Φ0
0
= √ e .
σ σ σ 2π
An important property of normal distributions is the following.
• If X1 , . . . , Xn are independent random variables where Xj has a N (µj , σj2 ) distribution, then X1 +
. . . + Xn has a N (µ1 + · · · + µn , σ12 + · · · σn2 ) distribution.
or, equivalently, if
P{X ≥ t} = e−λt , t ≥ 0.
Note that Z ∞ Z λ
−λx 1 2
E[X] = λxe dx = , E[X ] =2
x2 e−λx dx = ,
0 λ 0 λ2
1
Var[X] = E[X 2 ] − E[X]2 =
.
λ2
The exponential distribution has the loss of memory property, i.e., if G(t) = P{X ≥ t}, then
It is not difficult to show that this property characterizes the exponential distribution. If G(t) = P{X ≥ t},
then the above equality translates to
G(s + t) = G(s) G(t),
and the only continuous functions on [0, ∞) satisfying the condition above with G(0) = 1 and G(∞) = 0 are
of the form G(t) = e−λt for some λ > 0.
37
8.2.3 Gamma distribution
A random variable X has a Gamma distribution with parameters λ > 0 and α > 0 if it has density
λα α−1 −λx
f (x) = x e , x > 0.
Γ(α)
Here Z ∞
Γ(α) = xα−1 e−x dx
0
is the usual Gamma function.
Example
• More generally, if X1 , . . . , Xn are independent random variables, each Gamma with parameters λ and
αj , then X1 + · · · + Xn has a Gamma distribution with parameters λ and α1 + · · · + αn .
• If Z has a standard normal distribution, then Z 2 has a Gamma distribution with parameters λ =
1/2, α = 1/2. To see this, let F denote the distribution of Z 2 . Then,
√
√
Z x
1 2
2
F (x) = P{Z ≤ x} = P{|Z| ≤ x} = 2 √ e−t /2 dt.
0 2π
Hence,
1
f (x) = F 0 (x) = √ e−x/2 .
2πx
If n is a positive integer, then the Gamma distribution with parameters λ = 1/2 and α = n/2 is called
the χ2 -distribution with n degrees of freedom. It is the distribution of
Z12 + · · · + Zn2
where Z1 , . . . , Zn are independent standard normal random variables. This distribution is very important in
statistics.
Nt = k, if τk ≤ t < τk+1 .
Then Nt is a Poisson process with parameter λ. We think of τk as the time until k events have occured
where events happen “randomly” with an average rate of λ events per time unit. The random variable Nt
denotes the number of events that have occurred by time t.
The process Nt has the following properties.
• For each 0 ≤ s < t, Nt − Ns has a Poisson distribution with mean λ(t − s).
38
• If 0 < t1 < t2 < . . . < tn , then the random variables
are independent.
• With probability one t 7→ Nt is nondecreasing, right continuous, and all jumps are of size one. For
every t > 0,
P{Nt+∆t = k + 1 | Nt = k} = λ ∆t + o(∆t), δt → 0 + .
39