Probabilityspace
Probabilityspace
Probabilityspace
University of Munich
Probability Spaces
Kolmogorov set down a clear mathematical foundation for probability theory
in 1933. The basic ingredient is a triple (Ω, F, P), where
i) Ω ∈ F
ii) if A ∈ F then Ac = Ω \ A ∈ F
S
iii) if A1 , A2 , . . . ∈ F then n An ∈ F
Examples:
1
i) Z = {B1 , B2 , . . .} a countable partition of Ω. Then
( )
[
σ(Z) = Bi |I subset of N .
i∈I
Random Variables
Let (E, E) be a measurable space (the ”state space”, e.g. E = R, Rd , C[0, 1], . . .).
A map X : Ω → E is an (E-valued) random variable (or (F/E)-measurable)
if
X −1 (B) ∈ F for all B ∈ E.
The σ-field generated by X is the smallest σ-field on Ω that makes X mea-
surable:
\
σ(X) := G = X −1 (E).
G σ-field
X G/E-measurable
For (E, E) = (R, B(R)) we have σ(X) = σ({X ≤ x}, x ∈ R); and X : Ω → R
is a random variable if and only if {X ≤ x} ∈ F for all x ∈ R. The
distribution function of X is
F (x) = P [X ≤ x] , x ∈ R.
A map Y : Ω → R is σ(X)-measurable if and only if Y = f (X) for some
measurable function f : R → R (this also hold for more general state spaces).
Independence
A collection G1 , G2 , . . . ⊂ F of σ-fields is independent if for every choice
Ai ∈ Gi , i = 1, 2, . . . , the events A1 , A2 , . . . are independent:
P [A1 ∩ A2 ∩ . . . ∩ An ] = P[A1 ] . . . P[An ] for all n ∈ N.
The random variables X1 , X2 , . . . are independent if σ(X1 ), σ(X2 ), . . . are
independent.
2
Expectation
Let X be a non-negative random variable. Then there exists a sequence
(Xn ) of simple random variables (taken only finitely many values) such that
0 ≤ X1 ≤ X2 ≤ . . . ≤ Xn ↑ X a.s. Then 0 ≤ E[X1 ] ≤ E[X2 ] ≤ . . . (the
definition of E[Xn ] is clear) and the limit
E[X] := lim E[Xn ] (monotone convergence!)
n
Idea: realize Xn on Ωn = ([0, 1], B[0, 1], Leb), take infinite product Ω =
Ω1 × Ω2 × . . . (see Chapter 8 in D. Williams, Probability with Martingales,
Cambridge University Press, 1995).
Conditional Expectation
Let G ⊂ F be a σ-field, and X an F-measurable, integrable random variable.
We consider G as the collection of events that can be ”observed” (the available
information). What is the best prediction, say Y , of X given G?
3
• If G = {Ω, ∅} then Y = E[X].
In both cases we clearly have E[Y 1A ] = E[X1A ] for all A ∈ G. This is in fact
the defining property: Y is called conditional expectation of X with respect
to G if
i) Y is G-measurable
4
vi) (cMCT) If 0 ≤ X1 ≤ X2 ≤ . . . then
E[E[X|G]| H] = E[X|H].
ix) (”Taking out what is known”) If Y is G-measurable and |XY | and |X|
are integrable then
E[X|G] = E[X].
References
[1] W. Feller: An Introduction to Probability Theory and Its Applications.
Vol. I. Wiley, New York (1968).