Lecture Notes
Lecture Notes
Lecture Notes
Perla Sousi∗
October 10, 2011
Contents
1 Conditional expectation 1
1.1 Discrete case . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.2 Existence and uniqueness . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1 Conditional expectation
Let (Ω, F , P) be a probability space, i.e. Ω is a set, F is a σ-algebra on Ω and P is a
probability measure on (Ω, F ).
1. Ω ∈ F
Let A, B ∈ F be two events with P(B) > 0. Then the conditional probability of A given
the event B is defined by
P(A ∩ B)
P(A|B) = .
P(B)
∗
University of Cambridge, Cambridge, UK; [email protected]
1
Definition 1.3. The Borel σ-algebra, B(R), is the σ-algebra generated by the open sets in
R, i.e., it is the intersection of all σ-algebras containing the open sets of R. More formally,
let O be the open sets of R, then
Informally speaking, consider the open sets of R, do all possible operations, i.e., unions,
intersections, complements. and take the smallest σ-algebra that you get.
{B ⊆ R : X −1 (B) ∈ F }
Recall the definition of expectation. First for positive simple random variables, i.e., linear
combinations of indicator random variables, we define
" n # n
X X
E ci 1(Ai ) := ci P(Ai ),
i=1 i=1
where ci are positive constants and Ai are measurable events. Next, let X be a positive
random variable. Then X is the increasing limit of positive simple variables. For example
So we define
E[X] :=↑ lim E[Xn ].
Finally, for a general random variable X, we can write X = X + − X − , where X + =
max(X, 0) and X − = max(−X, 0) and we define
2
Let X be a random variable with E[|X|] < ∞. Let A be an event in F with P(A) > 0.
Then the conditional expectation of X given A is defined by
E[X 1(A)]
E[X|A] = ,
P(A)
X ′ is G − measurable (1.1)
since X ∈ L1 .
Let G ∈ G. Then it is straightforward to check that
3
Theorem 1.6. [Monotone convergence theorem] Let (Xn )n be random variables such
that Xn ≥ 0 for all n and Xn ↑ X as n → ∞ a.s. Then
E[Xn ] ↑ E[X] as n → ∞.
Let p ∈ [1, ∞) and f a measurable function in (Ω, F , P). We define the norm
kf kp = (E[|f |p ])1/p
and we denote by Lp = Lp (Ω, F , P) the set of measurable functions f with kf kp < ∞. For
p = ∞, we let
kf k∞ = inf{λ : |f | ≤ λ a.e.}
and L∞ the set of measurable functions with kf k∞ < ∞.
Formally, Lp is the collection of equivalence classes, where two functions are equivalent if
they are equal almost everywhere (a.e.). In practice, we will represent an element of Lp by
a function, but remember that equality in Lp means equality a.e..
Theorem 1.7. The space (L2 , k · k2) is a Hilbert space with hf, gi = E[f g]. If H is a closed
subspace, then for all f ∈ L2 , there exists a unique (in the sense of a.e.) g ∈ H such that
kf − gk2 = inf h∈H kf − hk2 and hg, f − gi = 0.
(a) Y is G-measurable;
Remark 1.10. We could replace (b) in the statement of the theorem by requiring that for
all bounded G-measurable random variables Z we have
Remark 1.11. In a later section we will show how to construct explicit versions of the
conditional expectation in certain simple cases. In general, we have to live with the indirect
approach provided by the theorem.
4
Proof of Theorem 1.9. (Uniqueness.) Suppose that both Y and Y ′ satisfy (a) and (b).
Then, clearly the event A = {Y > Y ′ } ∈ G and by (b) we have
Notice that the left hand side is nonnegative, while the right hand side is non-positive,
implying that P(Y < 0) = 0.
2nd step: Suppose that X ≥ 0. For each n we define the random variables Xn = X ∧ n ≤
n, and hence Xn ∈ L2 . Thus from the first part of the existence proof we have that for
each n there exists a G-measurable random variable Yn satisfying for all A ∈ G
Since the sequence (Xn )n is increasing, from (1.3) we get that also (Yn )n is increasing. If we
now set Y =↑ limn→∞ Yn , then clearly Y is G-measurable and by the monotone convergence
theorem in (1.4) we get for all A ∈ G
since Xn ↑ X, as n → ∞.
In particular, if E[X] is finite, then E[Y ] is also finite.
3rd step: Finally, for a general random variable X ∈ L1 (not necessarily positive) we
can apply the above construction to X + = max(X, 0) and X − = max(−X, 0) and then
E[X|G] = E[X + |G] − E[X − |G] satisfies (a) and (b).
5
Remark 1.12. Note that the 2nd step of the above proof gives that if X ≥ 0, then there
exists a G-measurable random variable Y such that
i.e., all the conditions of Theorem 1.9 are satisfied except for the integrability one.
Definition 1.13. Sub-σ-algebras G1 , G2 , . . . of F are called independent, if whenever Gi ∈
Gi (i ∈ N) and i1 , . . . , in are distinct, then
n
Y
P(Gi1 ∩ . . . ∩ Gin ) = P(Gik ).
k=1
E[Xn ] → E[X].
Theorem 1.17. [Jensen’s inequality] Let X be an integrable random variable and let
ϕ : R → R be a convex function. Then
E[ϕ(X)] ≥ ϕ(E[X]).
6
Proposition 1.18. Let G ⊂ F be a σ-algebra and let X ∈ L1 (Ω, F , P).
1. Conditional monotone convergence theorem: If (Xn )n≥0 is an increasing se-
quence of nonnegative random variables with a.s. limit X, then
E[Xn |G] ր E[X|G] as n → ∞, a.s..
2. The sequence inf k≥n Xk is increasing in n and limn→∞ inf k≥n Xk = lim inf n→∞ Xn .
Thus, by the conditional monotone convergence theorem we get
lim E[inf Xk |G] = E[lim inf Xn |G].
n→∞ k≥n n→∞
Clearly, E[inf k≥n Xk |G] ≤ inf k≥n E[Xk |G]. Passing to the limit gives the desired
inequality.
3. Since Xn +Y and Y −Xn are positive random variables for all n, applying conditional
Fatou’s lemma we get
E[X + Y |G] = E[lim inf(Xn + Y )|G] ≤ lim inf E[Xn + Y |G] and
n→∞
E[Y − X|G] = E[lim inf(Y − Xn )|G] ≤ lim inf E[Y − Xn |G].
n→∞
7
4. A convex function is the supremum of countably many affine functions: (see for
instance [1, §6.6])
ϕ(x) = sup(ai x + bi ), x ∈ R.
i
So for all i we have E[ϕ(X)|G] ≥ ai E[X|G] + bi a.s. Now using the fact that the
supremum is over a countable set we get that
References
[1] David Williams. Probability with martingales. Cambridge Mathematical Textbooks.
Cambridge University Press, Cambridge, 1991.