Draft Notes
Draft Notes
IN
STOCHASTIC DIFFERENTIAL EQUATION
Contents
1 Week1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Tuesday 1
1.2 Thursday 7
2 Week2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.1 Tuesday 13
2.2 Thursday 22
3 Week3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
3.1 Tuesday 27
3.1.1 Reviewing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
v
3.2 Thursday 35
4 Week4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
4.1 Tuesday 41
4.2 Thursday 48
5 Week5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53
5.1 Tuesday 53
5.2 Thursday 60
6 Week6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
6.1 Tuesday 65
6.1.1 Localization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
6.2 Thursday 69
7 Week7 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
7.1 Tuesday 73
vi
7.2 Thursday 79
8 Weak 8 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
8.1 Thursday 83
9 Week9 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
9.1 Tuesday 87
9.2 Thursday 92
10 Week10 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
10.1 Tuesday 99
11 Week11 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
vii
12 Week12 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
13 Week13 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
viii
Chapter 1
Week1
1.1. Tuesday
We first discuss the difference between deterministic differential equations and stochas-
where N (t) denotes the size of the population at time t; a(t) is the given (deterministic)
constant.
with r (t) being a deterministic function of t, and the “noise” term models something
random. The question arises: How to rigorously describe the “noise” term and solve it?
1
Problem 2: Electric Circuit. Let Q(t) denote the charge at time t in an electrical
circuit, which admits the following ODE:
8
> 1
< LQ00 (t) + RQ0 (t) + Q(t) = F (t),
C
>
: Q (0) = Q , Q 0 (0) = Q 0
0 0
where L denotes the inductance, R denotes the resistance, C denotes the capacity, and
Now consider the scenario where F (t) is not completely known, e.g.,
dX (t)
= rX (t) + aX (t) · noise
dt
where r, a are given constants. The goal of this person is to maximize the expected
selling price:
sup E [ X (t )]
t 0
2
Problem 4: Portfolio Selection Problem. Suppose a person is interested in two
types of assets:
dX1 (t)
= rX1 (t),
dt
dX2 (t)
= µX2 (t) + sX2 (t) · noise
dt
The policy of the investment is as follows. The wealth at time t is denoted as v(t).
This person decides to invest the fraction u(t) of his wealth into the risky asset, with
the remaining 1 u(t) part to be invested into the safe asset. Suppose that the utility
function for this person is U (·), and his goal is to maximize the expected total wealth
where the decision variable is the portfolio function u(t) along whole horizon [0, T ].
Problem 5: Option Pricing Problem. The financial derivates are products in the
market whose value depends on the underlying asset. The European call option is
a typical financial derivative. Suppose that the underlying asset is stock A, whose
price at time t is X (t). Then the call option gives the option holder the right (not the
obligation) to buy one unit of stock A at a specified price (strike price) K at maturity
date T. The task is to inference the fair price of the option at the current time. The
c0 = E [( X ( T ) K )+ ]
3
1.1.3. Reviewing for Probability Space
1. W 2 F ;
measure on (W, F ) if
• P (W) = 1;
• P ( A) 0, 8 A 2 F ;
✓ • ◆ •
[
P Ai = Â P ( A i ).
i =1 i =1
follows:
4
Definition 1.4 [Almost Surely True] A statement S is said to be almost surely (a.s.)
• B := {w : S(w) is true} 2 F
• P ( F ) = 1.
1. ∆, X 2 T
T
2. U, V 2 T implies U V2T
S
3. If Ua 2 T for all a 2 A, then a2A Ua 2T.
Definition 1.6 [Borel s-Algebra] Consider a topological space W, with U being the
containing U :
B(W) , s(U ).
1
f ( B ) = {w | f (w) 2 B} 2 F ,
5
Definition 1.8 [Generated s-Algebra] Suppose X is a random variable on (W, F , P ).
W to make X measurable. ⌅
to finish the proof, which is true since B(R n ) = s(U ), with U being the topology of
Rn .
6
1.2. Thursday
Reviewing for Probability Space.
• (W, F , P );
• Random variable;
• Generated s-algebra;
variabe X isdefined as
1
µX ( B ) = P(X ( B )),
Z
E[X] = X (w ) dP (w )
W
Z
E[X] = y dµ X (y)
Rn
Note that the expectation of the random variable X is well-defined when X is integrable:
Z
| X (w)| dP (w) < •.
W
7
R
⌅ Example 1.1 If f : R n ! R is Borel measurable, and W
| f ( X (w ))| dP (w ) < •, then
Z Z
E [ f ( x )] = f ( X (w )) dP (w ) = f (y) dµ X (y).
W Rn
• Define L p -norm of X as
✓Z ◆1/p
p
kXk p = | X (w )| dP
W
If p = •, define
Z
| X (w )| p dP (w ) < •,
W
denoted as X 2 L p (W, F , P ).
Z ✓Z ◆q/p ✓Z ◆ 1p ·q
q q
k X kq = | X |q dP (| X |q ) p/q dP = | X | p dP = kXk p .
W W W
8
Definition 1.13 [Independence]
T
1. Two events A1 , A2 2 F are said to be independent if P ( A1 A2 ) = P ( A1 )P ( A2 ).
for 8 F1 2 F1 , F2 2 F2
R The independence defined above can be generalized from two events into
finite number of events.
Proposition 1.3 If X and Y are two independent random variables, and E [| X |] <
E [ XY ] = E [ X ]E [Y ] < •.
Proof. The first step is to simplify the probability distribution for the product random
Z ZZ
E [ XY ] = xy dµ X,Y ( x, y) = xy dµ X ( x )µY (y)
Z Z Z
= y x dµ X ( x ) µY (y) = E [ X ]yµY (y) = E [ X ]E [Y ].
9
1.2.2. Stochastic Process
Consider a set T of time index, e.g., a non-negative integer set or a time interval [0, •).
cot, left-cot, resp.) a.s., if t ! Xt (w ) is continuous (right-cot, left-cot, resp.) a.s., i.e.,
✓ ◆
P {w : t ! Xt (w ) is continuous (right-cot, left-cot, resp.)} = 1.
variables with Possion distribution with intensity l > 0. Let T0 = 0, and Tn = Ânj=1 x j .
https://github.com/WalterBabyRudin/Courseware/tree/master/MAT4500/week1
10
125
0
0 20 40 60 80 100 120
Figure 1.1: One simulation of { Xt } with intensity l = 1.2 and 500 samples
11
Chapter 2
Week2
2.1. Tuesday
1
{ X 2 F } , {w : X (w ) 2 F } = X ( F ).
is defined as
⌅ Example 2.1 [Brownian Motion] Consider a probability space (W, F , P ). Define the
function
✓ ◆
1 (y x )2
P(t, x, y) = p exp , x, y 2 R, t > 0
2pt 2t
13
t1 , t2 , . . . , tk is given by:
P ( Bt1 2 F1 , . . . , Btk 2 Fk ) =
Z
P(t1 , 0, x1 )P(t2 t1 , x1 , x2 ) · · · P( t l tk 1 , xk 1 , xk ) dx1 dx2 · · · dxk
F1 ⇥···⇥ Fk
a now consider only the Brownian motion with independent, normally distributed increment.
Definition 2.2 [Measurable Set] Let (S, F ) be a pair, with S being a set and F is a
The issue is that this event F is not necessarily B(R [0,•) )-measurable. Some-
times we need some extra conditions on the stochastic process to make F
measurable. The significance of F will also discussed in the future.
{ Xt 2 B} := {w : Xt (w ) 2 B} is measurable.
For the second assertion, it suffices to check that h 1 ([ •, a)) = \t2T { Xt < a} is
measurable. ⌅
14
However, when T is uncountable, it is problematic to show the measurability. It is
even difficult to show that for almost all w, t 7! Xt (w ) is continuous. In order to obtain
Definition 2.3 [Equivalent random variables] Let { Xt }t 0 and {Yt }t 0 be two stochastic
R It is easy to see that when { Xt }t 0 is a version of {Yt }t 0, they have the same
finite-dimensional distributions, but their path properties may be different,
e.g., for almost all w, t 7! Xt (w ) may be continuous while t 7! Yt (w ) may not.
1. E [ X | G] is G -measurable;
In other words,
E [E [ X | G]1 A ] = E [ X1 A ].
15
In other words, suppose that Y is another random variable satisfying the
condition mentioned in Definition 2.4, i.e.,
• Y is G -measurable;
Conditional expectation has many of the same properties that ordinary expectation
does:
1. E [E [ X | G]] = E [ X ]
E [ a1 X1 + a2 X2 | G] = a1 E [ X1 | G] + a2 E [ X2 | G], a.s.
4. (Positivity) If X 0, then E [ X | G] 0.
E E [ X | G] | H = E [ X | H], a.s.
In particular, E [ X | H] = E [ X ] if H is independent of X.
16
Proof. 1. Recall the definition of E [ X | G] and take A = W,
E [E [ X | G]] = E [E [ X | G]1W ] = E [ X ].
2. It suffices to verify that X satisfies 1) and 2) in Definition 2.4, and the result holds
3. Again, verify the RHS satisfies 1) and 2) in Definition 2.4, and the result holds by
4. For fixed w 2 W,
L = { x 7 ! f ( x0 ) + gT ( x x0 ) | x0 2 dom(f), g 2 ∂f( x0 )}
f( x ) = sup L2L L( x ).
6. It suffices to show that E [ X | H] is a version of E E [ X | G] | H . The key is to
E [E [ X | H]1 A ] = E [E [ X | G]1 A ].
17
7. It suffices to show that E [ X | G] is a version of E X s(G , H) , i.e., for any
A 2 s(G , H),
E [ X1 A ] = E E [ X | G]1 A .
A key issue is that two sets may loss orders. For instance, it is possible that neither
[ \
Em = En , Em = En
n m n m
limits:
•
n=1 P ( En ) < •, then ✓ ◆
P lim sup En = 0.
n!•
Proof. Define Em as above, and thus lim supn!• En = \m Em . As a result, for any m,
✓ ◆ •
P lim sup En = P (\m Em ) P ( Em )
n!•
 P(En ).
n=m
18
Because of the condition •
n=1 P ( En ) < •, as m ! •,
• ✓ ◆
 P(En ) ! 0 =) P lim sup En
n!•
= 0.
n=m
Then
Z Z
f dµ = lim f n dµ.
S n!• S
The proof for the monotone convergence theorem (MCT) can be found in the website
Daniel Wong, Jie Wang. (2019) Lecture Notes for MAT3006: Real Analy-
https://walterbabyrudin.github.io/information/Updates/Updates.html
We can apply MCT to show the Fatou’s lemma, in which the required condition is
weaker:
Z Z
lim inf f n dµ lim inf f n dµ.
S n!• n!• S
19
non-negative functions. Then
Z Z Z
lim inf f n dµ = lim gn dµ = lim gn dµ
n!• n!• n!•
Z
= lim inf gn dµ
n!•
Z
lim inf f n dµ
n!•
where the second equality is by MCT, and the last equality is because that gn
f n , 8n. ⌅
⌅ Example 2.2 In general the integral of the limit-inf on a sequence of functions is smaller.
Then
Z
lim inf f n = 1{1/2} =) lim inf f n dm = 0,
n!• n!•
R
while [0,1] f n dm = 1/2 for each n. ⌅
R We also have the reversed fatou’s lemma, saying that in general the integral
of the limit-sup on a sequence of functions is bigger:
Z Z
lim sup f n dµ lim sup f n dµ.
S n!• n!• S
| f n ( x )| g( x )
20
R
for almost all x 2 S, with S
| g| dµ < •.
2. f n converges to f almost everywhere for some measurable function f .
R
Then f is integrable and f n ! f in L1 , i.e., limn!• S
| fn f | dµ = 0, which implies
that
Z
f dµ = lim f n dµ.
S n!•
Z Z
lim sup | fn f | dm lim sup | f n f | dm = 0.
Daniel Wong, Jie Wang. (2019) Lecture Notes for MAT3006: Real
https://walterbabyrudin.github.io/information/Updates/Updates.html
21
2.2. Thursday
L1 if
Z
lim | fn f | dµ = 0.
n!• S
L1 (W, F , P ), then for any # > 0, there exists d > 0 such that for any F 2 F with P ( F ) < d,
we have
Z
E [| X |; F ] , E [| X |1F ] = | X | dP < #
F
Proof. Suppose on the contrary that there exists some # 0 > 0, and a sequence of events
1
P ( Fn ) < , but E [| X |; Fn ] #0.
2n
As a result, •
n=1 P ( Fn ) < •. By applying theorem 2.2,
Z Z
E [| X |; H ] = | X |1 H dP lim sup | X |1Fn dP = lim sup E [| X |; Fn ] #0
n!• n!•
22
Corollary 2.1 Suppose that X 2 L1 (W, F , P ), then for any # > 0, there exists K > 0,
such that
Z
E [| X |; | X | > K ] := | X | dP < #.
{| X |>K }
Proof. The idea is to construct K such that {| X | > K } happens with small probability.
• Firstly we have the Markov inequality P ({| X | > K }) K1 E [| X |], since the follow-
E [| X |] = E [| X |; | X | > K ] + E [| X |; | X | K ]
E|X|
• Applying Proposition (3.1), we choose K large enough such that K < d, which
implies P (| X | > K ) < d. The desired result follows immediately.
and only if for any given # > 0, there exists a K 0 such that
E [| X |; | X | > K ] < #, 8X 2 C .
Proof. Choose # = 1, then there exists K > 0 such that for any X 2 C ,
E [| X |] = E [| X |; | X | > K ] + E [| X |; X K ] # + K = 1 + K,
However, the converse of this statement is not necessarily true. See Example 2.3 for
a counter-example.
23
⌅ Example 2.3 Consider the probability space (W, F , P ) = ([0, 1], B([0, 1]), Leb), and
• However, C is not UI. Take # = 1, and for any K > 0, as long as n > K,
E [| Xn |; | Xn | > K ] = 1
a.s., but
Z
| Xn 0| dP = 1, 8n.
Although L1 -boundedness does not imply UI, the L p -boundness for p > 1 does.
Theorem 2.6 Let p > 1. Suppose that a class C of random variables are uniformly
bounded in L p , i.e.,
Z
E [| X | p ] = | X | p dP < M < •, 8 X 2 C ,
W
where M is some finite constant. Then the class C is uniformly integrable (UI).
Proof. Choose some K > 0, the idea is to bound the term E [| X |; | X | > K ], for any X 2 C :
Z Z
|X| p
| X | dP = p 1
dP
{| X |>K } {| X |>K } | X |
Z Z
|X| p 1
p 1
dP = p 1
| X | p dP
{| X |>K } K K {| X |>K }
Z
1
p 1 | X | p dP
K W
M
p 1.
K
where the last inequality is by the L p -boundedness. Therefore, for any given # > 0, the
24
M
desired result holds by choosing K large enough such that Kp 1
#.
⌅
| X (w )| Y (w ), 8w 2 W, E |Y | < •,
Proof. The idea is to bound the term E [| X |; | X | > K ] to show the UI:
Z Z Z
| X | dP | X | dP |Y | dP
{| X |>K } {|Y |>K } {|Y |>K }
where the first inequality is because that {| X | > K } ✓ {|Y | > K }, and the second is
because that | X | < Y. The desired result holds by applying Corollary 3.1 such that
Z
|Y | dP < #.
{|Y |>K }
25
Chapter 3
Week3
3.1. Tuesday
3.1.1. Reviewing
Definition 3.1 For p 1, we say a random variable X 2 L p if
p
k X k p , E [| X | p ] < •.
Proposition 3.1 Suppose that a random variable X is integrable, then for any # > 0,
there exists d > 0 such that for any F 2 F with P ( F ) < d, we have
Z
E [| X |; F ] , E [| X |1F ] = | X | dP < #
F
Since {| X | > K } happens with small probability, we have the following corol-
lary:
Corollary 3.1 Suppose that X 2 L1 (W, F , P ), then for any # > 0, there exists K > 0,
such that
Z
E [| X |; | X | > K ] := | X | dP < #.
{| X |>K }
27
Definition 3.2 Consider a collcetion of random variables instead, denoted as C :
E [| X | p ] < M, 8X 2 C .
that
Recall the proof stated in Theorem 2.6 and Theorem 2.7 in detail.
following formula:
Z Z
1
E [| X |1{|X |>K} ] = | X | dP 1
| X | p dP.
{| X |>K } Kp {| X |>K }
Z Z
E [| X |1{|X |>K} ] = | X | dP |Y | dP.
{| X |>K } {|Y |>K }
28
3.1.2. Necessary and Sufficient Conditions for UI
Our first result is about sufficient conditions for the UI on a collection of conditional
expectations:
Theorem 3.1 Suppose that X 2 L1 (W, F , P ) and {Ga }a2A is a collection of s-algebras
⇢
C = E [ X | Ga ] : a 2 A
is uniformly integrable.
Proof. • Apply proposition 3.1 on X: For given # > 0, there exists d > 0 such that
Z
E |E [ X | Ga ]|; |E [ X | Ga ]| > K = |Ya | dP
{|Ya |>K }
Z
E [| X | | Ga ] dP
{|Ya |>K }
Z
E [| X | | Ga ] dP
{E [| X ||Ga ]>K }
Z
= | X | dP
{E [| X ||Ga ]>K }
where the last equality is because of the definition for conditional expectation
E [E [| X | | Ga ]] E [| X |]
P {E [| X | | Ga ] > K } = ,
K K
29
the result on the first part, we have
Z
E |E [ X | Ga ]|; |E [ X | Ga ]| > K | X | dP #.
{E [| X ||Ga ]>K }
Z
lim sup | X | dP = 0.
k!• X 2C {| X |>K }
theory.
✓ ◆
lim P | Xn X| > # = 0.
n!•
✓n o◆
P w : lim Xn (w ) = X (w ) = 1.
n!•
lim k Xn X k1 = 0.
n!•
30
R
• Xn ! X in L1 implies Xn ! X in prob.;
9
a.s. >
Xn ! X >
>
>
=
L1
| Xn | < Y > ) Xn ! X
>
>
>
E (Y ) < • ;
gence in L1 :
| Xn (w )| M, 8w 2 W, n 1. Then { Xn } converges to X in L1 :
lim E [| Xn X |] = 0.
n!•
1. Xn ! X in probability, and
2. { Xn } is uniformly integrable.
31
Proof for the Reverse Direction. For K > 0, construct a function fK : R ! [ K, K ]:
8
>
>
>
> K, if x > K
>
<
fK ( x ) = x, if | x | K
>
>
>
>
>
: K, if x < K
| Xn X | | Xn fK ( Xn )| + |fK ( Xn ) fK ( X )| + |fK ( X ) X |.
It suffices to upper bound three terms on the RHS for the following formula:
• For the first term, by choosing sufficiently large K, by Corollary 3.1, it can be
• For the third term, when K is large enough, by the uniform integrability of { Xn },
which means that P ({|fK ( Xn ) fK ( X )| > #}) P ({| Xn X | > #}). As a result,
1 Following the similar method, we can show that as long as f is continuous and Xn ! X in prob., we
have f ( Xn ) ! f ( X ) in prob.
32
Thus for sufficiently large n,
#
E [|fK ( Xn ) fK ( X )|] <
3
Combining these three bounds above, for fixed # > 0, we can pick K > 0 and there
E [| Xn X |] #.
Proof for the Forward Direction. • Firstly we show that { Xn } is L1 -bounded, which
vation:
E [| Xn |] E [| X |] E || Xn | | X || E | Xn X | ! 0.
• Then we show the uniform integrability result. By the L1 -convergence, for fixed
#
E | Xn X| < , 8n > N0 .
2
Similar as the previous proof for the uniform integrability results, we should
apply proposition 3.1 on finitely many random variables: for fixed # > 0, there
#
E [| X |1F ] < (3.2a)
2
#
E [| Xn |1F ] < , 8n N0 (3.2b)
2
E | Xn | supn E | Xn |
P (| Xn | > K )
K K
33
supn E | Xn |
Therefore, we choose K such that K < d, and then P (| Xn | > K ) < d.
• Now we can conclude the uniform integrability result: For n N0 , by the con-
For n > N0 ,
E | Xn |1{|Xn |>K} E | X Xn |1{|Xn |>K} + E | X |1{|Xn |>K}
E |X Xn | + E | X |1{|Xn |>K}
# #
< + = #.
2 2
✓ ◆
E [| Xn X |]
P | Xn X| > # ! 0, as n ! •.
#
34
R A typical example of F is defined by generated s-algebra:
FnX , s( Xt : t n), 8n 0.
predictable process.
3.2. Thursday
{ T n} , {w 2 W : T (w ) n} 2 Fn , 8n.
35
Proof. (a) Suppose that { T n} 2 Fn , 8n, then
{T n 1} 2 F n 1 ✓ Fn =) { T = n} = { T n} \ { T n 1} 2 F n .
[
{ T = k} 2 Fk ✓ Fn , 8k n =) { T = n} = { T = k } 2 Fn .
kn
T (w ) , inf {n 0 : Xn ( w ) 2 B }.
Here T denotes the first time that { Xn }n 0 enters into set B. Define inf(∆) = • by
{ T = n } = { X0 2 B c , X1 2 B c , X2 2 B c , . . . , X n 1 2 B c , Xn 2 B }
= { Xn 2 B} [ ([0kn 1 { Xk 2 Bc })
Definition 3.7 [Stopping Time s-algebra] Define the stopping time s-algebra for a given
F T = { A 2 F : A \ { T n } 2 F n , 8 n }.
36
Here FT represents the information available up to a random time T. ⌅
2. T is FT -measurable;
✓ ◆c
c
A \ { T n} = A \ { T n} \ { T n} 2 Fn .
✓ ◆
[ [
Ak \ { T n} = ( Ak \ { T n}) 2 Fn .
k 1 k 1
2. It suffices to show that { T m} 2 FT for any m. This is true because for any n,
{ T m } \ { T n } = { T m ^ n } 2 Fm^n ✓ Fn .
{ T2 n} ✓ { T1 n} =) A \ { T2 n} ✓ A \ { T1 n} 2 Fn ,
X T ( w ) , X T ( w ) ( w ), 8w 2 W.
Then XT is FT -measurable.
37
s-algebra, it suffices to check
[
{ XT a} \ { T n} 2 Fn , 8n (= { Xk a} \ { T = k} 2 Fn , 8n.
0 k n
k } 2 Fk , 8k. Therefore,
{ Xk a } \ { T = k } 2 F k ✓ F n .
1. Xn 2 L1 , 8n;
E [|Yk |] < • and E [Yk ] = 0, 8k. Define Fn = s(Y1 , Y2 , . . . , Yn ) for n 1 and F0 = {∆, W}.
38
2. Check that
E [ Xn+1 | Fn ] = E [ Xn + Yn+1 | Fn ]
= E [ Xn | Fn ] + E [Yn+1 | Fn ]
= Xn + E [Yn+1 ] = Xn ,
where the third equality is because that Xn is Fn -measurable and Yn+1 is independent
of Fn .
a.s. and E [Yk ] = 1, 8k. Define Fn = s (Y1 , Y2 , . . . , Yn ) for n 1 and F0 = {∆, W}. Define
2. Check that
E [ Xn+1 | Fn ] = E [ Xn · Yn+1 | Fn ]
= Xn · E [Yn+1 | Fn ]
= Xn · E [Yn+1 ] = Xn ,
where the second equality is because that Xn is Fn -measurable; the third equality is
39
Chapter 4
Week4
4.1. Tuesday
E [| Xn |] = E [|E [z | Fn ]|]
E [E [|z | | Fn ]]
= E [|z |] < •
E [ Xn +1 | F n ] = E [ E [ z | F n +1 ] | F n ]
= E [ z | F n ] = Xn .
41
⌅ Example 4.2 [Martingale Transform] Let Cn be the stake to be bet on game n, and
Xn Xn 1 be the net with per stake in game n, with n 1. Suppose that the process
M· Â E [| Xk Xk 1 |]
1 k n
M· Â E [| Xk |] + E [| Xk 1 |] < •,
1 k n
• Moreover,
E [Yn+1 | Fn ] = E Â Ck ( Xk Xk 1) Fn
1 k n +1
= Yn + Cn+1 E [ Xn+1 Xn | Fn ] = Yn ,
where the third equality is by the Fn -measurability of Cn+1 ; the fourth equality is
a Here the boundedness means that |Cn (w )| M for some M > 0 and almost all w
42
Theorem 4.1 Suppose that { Xn } is a martingale with respect to Fn , and f : R ! R
E [ f ( Xn +1 ) | F n ] f(E [ Xn+1 | Fn ]) = f( Xn ).
an increasing convex function such that f( Xn ) is integrable for all n, then {f( Xn )}
Theorem 4.3 Let { Xn } be a martingale and T be a stopping time. Define the stopped
process { Xn^T } as
Xn ^ T ( w ) , Xn ^ T ( w ) ( w ), 8w 2 W, 8n.
E [ X n ^ T ] = E [ X0 ] , 8n.
Proof. We will show this result by applying the Martingale transform technique men-
43
Note that {CnT }n 0 is predictable since {CnT = 0} = { T (w ) n 1} 2 F n 1. Now we
+ · · · + (1{ T n 1} 1{ T n}) Xn 1 + 1{ T n } Xn
n 1
= 1{ T n } Xn 1{ T 1 } X0 + Â 1 { T = i } Xi
i =1
✓ n 1 ◆ ✓ ◆
= 1{ T n } Xn + Â 1 { T = i } Xi 1{ T 1 } X0 1 { T = 0 } X0
i =0
= Xn ^ T X0 .
By the boundedness of {CnT }, and the result in Example 4.2, we can show that { Xn^T
E [ Xn ^ T ] = E [E [ Xn ^ T | F n 1 ]] = E [ X( n 1)^ T ] = · · · = E [ X0 ^ T ] = E [ X0 ] .
Note that E [ XT ] does not equal to limn!• E [ Xn^T ] = E [ X0 ]. The following provides
a counter-example:
T , inf{n 0 Xn = 1}.
Then E [ Xn^T ] = E [ X0 ] = 0, 8n. Since the random walk is recurrent, P ( T < •), and
1 = E [ XT ] 6= lim E [ Xn^T ] = E [ X0 ] = 0.
n!•
44
⌅
E [ X0 ] :
conditions hold:
1. T is bounded a.s.;
Proof. 1. Suppose that T is bounded a.s., which means that there exists K such
2. Suppose that T is finite a.s., then we can show that Xn^T ! XT a.s.: Note that
P { T < •} = 1, and
E [ XT ] = lim E [ Xn^T ] = E [ X0 ].
n!•
T ^n T ^n
| XT ^n X0 | = Â ( Xk Xk 1) Â | Xk Xk 1| M · ( T ^ n) MT,
k =1 k =1
where the second inequality is by the boundedness of {| Xn Xn 1 |}, i.e., for any
45
XT ^n ! XT in L1 , and
E [ XT ] = lim E [ Xn^T ] = E [ X0 ].
n!•
Proof. We only show the result based on the assumption that { Xn } is a sub-martingale,
since the remaining part follows the similar logic. Since S, T are bounded a.s., random
Z N Z
A
( XT XS ) dP = Â ( XT XS ) dP (4.1a)
j=0 A\{S= j}
N Z N Z
=Â ( XT XS ) dP + Â ( XT XS ) dP
j=0 A\{S= j}\{ T > j} j=0 A\{S= j}\{ T = j}
(4.1b)
N Z
=Â ( XT XS ) dP (4.1c)
j=0 A\{S= j}\{ T > j}
where (4.1a) is by the assumption that S is bounded a.s., i.e., |S| N a.s.; (4.1b)
46
is by the assumption that T S; (4.1c) is because 1{S = j, T = j} · ( XT XS ) = 0.
Z Z
(Xj X j+1 ) dP 0 =) ( XS XT ) dP 0, 8 A 2 FS . (4.2)
A\{S= j}\{ T > j} A
R
For two F -measurable random variables, if A
(X Y ) dP 0, 8 A 2 F , then one
· · · R N T a.s., with
0 R1 S 1, 0 Rj Rj 1 1, 8 j 0 T R N 1.
Z
( XS XR1 ) dP 0.
A
A \ {S j} \ { R1 j} = A \ { R1 j} 2 F j =) A 2 F R1 .
Z
( X R1 XR2 ) dP 0.
A
Similarly,
Z Z
( XR j 1
XR j ) dP 0, j = 2, . . . , N, ( XR N XT ) dP 0.
A A
R
Adding those integrals above, A
( XS XT ) dP 0, 8 A 2 FS , i.e.,
E [ X T | F S ] XS , a.s.
47
⌅
4.2. Thursday
R(w ) = inf{k 0 : Xk ( w ) l }, 8w 2 W.
Theorem 4.5,
E [ X0 ] E E [ X T | F0 ] = E [ X T ]
Z ⇢ Z ⇢
= 1 sup Xk l XT dP + 1 sup Xk < l XT dP
k N k N
✓ ◆ Z ⇢
l · P sup Xk l + 1 sup Xk < l X N dP
k N k N
where the first inequality is because that X0 E [ XT | F0 ] a.s.; and the last
⇢
inequality is because that conditioned on the event sup Xk < l , XT ⌘ X N .
k N
Thus the desired result holds.
48
2. Let Yn = Xn , and {Yn } is a sub-martingale. Define the stopping time
R(w ) = inf{k 0 : Yk (w ) l }, 8w 2 W.
Theorem 4.5,
✓ ◆ ✓ ◆
E [YN ] E [YT ] lP sup Yk l + E YN 1 sup Yk < l
k N k N
It follows that
✓ ◆ ✓ ◆ ✓ ◆
l·P inf Xk l = l · P sup Yk l E [YN ] E YN 1 sup Yk < l
k N k N k N
✓ ◆
= E ( X N )1 inf Xk l .
k N
✓ ◆ " ( )# ⇢
l · P sup | Xk | l E [ X0 ] E X N · 1 sup Xk < l E XN · 1 inf Xk l
k N k N k N
E [ X0 ] + 2E [ X N ],
Theorem 4.7 Let { Xn } be a martingale. Choose some N > 0 and let X N 2 L2 , i.e.,
✓ ◆
1
P sup | Xk | > l E [ X 2N ].
k N l2
49
As a result, E [ Xk2 ] E [ X 2N ] < •, 8k N, i.e., { Xk2 }k N is a super-martingale. Apply
✓ ◆ ✓ ◆
2
l ·P inf ( Xk2 ) l 2 2
= l · P sup | Xk | l
k N k N
⇢
E X 2N · 1 inf ( Xk2 ) l2
k N
E [ X 2N ].
!p ✓ ◆p
p
E sup Xk+ E [( Xn+ ) p ].
kn p 1
✓ ◆p
p p
E sup | Xk | E [| Xn | p ].
kn p 1
Z
E [f( Z )] = f( Z (w )) dP (w )
W
Z Z Z (w )
= df(y) dP (w )
W 0
Z Z
= 1{y Z (w )} df(y) dP (w )
W [0,•)
Z Z
= 1{y Z (w )} dP (w ) df(y) (4.3a)
[0,•) W
Z
= P(Z y) df(y)
[0,•)
50
Take f(y) ⌘ y p and define Xn⇤ = supkn Xn 1 for notation simplifcation. As a result,
Z
E [( Xn⇤ ) p ] = P ( Xn⇤ l) dl p
[0,•)
Z
" ( )#
1
E | Xn |1 sup | Xk | l dl p (4.3b)
[0,•) l kn
Z • Z
1
= Xn (w )1 { Xn⇤ (w ) l} dP (w ) dl p (4.3c)
0 l W
Z Z •
1
= Xn ( w ) 1 { Xn⇤ (w ) l} dl p dP (w ) (4.3d)
W 0 l
Z Z X ⇤ (w )
n
2
= Xn ( w ) pl p dl dP (w )
W 0
Z
p 1
= Xn ( w ) [ Xn⇤ (w )] p dP (w )
W p 1
p ⇣ ⌘1/q
(E [( Xn ) p ])1/p E [( Xn⇤ )( p 1) q
] , with 1/q = 1 1/p
p 1
(4.3e)
p
= (E [( Xn ) p ])1/p (E [( Xn⇤ ) p ])( p 1)/p
(4.3f)
p 1
where (4.3b) is by the Doob’s maximal inequality; (4.3c) is by the assumption that
2. The second inequality follows by applying the first and replacing { Xn } with
{| Xn |}.
⌅
✓ ◆
e +
E sup Xk 1 + sup E [ Xk log Xk ] ,
kn e 1 kn
51
Tale f(y) = (y 1)+ , then
Z •
E [( Xn⇤ 1) + ] = P ( Xn⇤ l) df(l)
0
Z •
1
E [ Xn 1{ Xn⇤ l}] df(l) (4.4a)
0 l
Z Z X⇤
n 1
= Xn df(l) dP (4.4b)
W 0 l
Z Z X⇤
n 1
= Xn 1{l 1} dl dP
W 0 l
Z
= Xn 1{ Xn⇤ 1} log Xn⇤ dP
W
= E [ Xn log+ Xn⇤ ].
where (4.4a) is by the Doob’s maximal inequality, and (4.4b) is by Fubini’s theorem. As a
result,
b
a log+ b a log+ a + , (4.4c)
e
52
Chapter 5
Week5
5.1. Tuesday
53
Proof. The increasing property is trivial. To check Tk is a stopping time, observe that
0 1
m\1
{ T2k 1 = m} = @ { Xt < b } A \ { Xm b} 2 Fm ,
t= T2k 2 +1
0 1
m\1
{ T2k = m} = @ { Xt > a } A \ { Xm a } 2 F m .
t= T2k 1 +1
interval [ a, b] by k times.
We can also assert that XT2j a if T2j < • a.s.; and XT2j+1 b if T2j+1 < • a.s.
⇣ ⌘ 1 h i
P Uab [ X; n] k+1 E ( Xn a) 1{Uab [ X; n] = k} .
b a
1
As a result, E [Uab [ X; n]] b a E [( Xn a ) ].
⇣ ⌘ 1 h i
P Uab [ X; n] k E ( Xn a)+ 1{Uab [ X; n] = k } .
b a
1
As a result, E [Uab [ X; n]] b a E [( Xn a ) + ].
54
Proof. 1. Considering that { Xn } is a super-martingale and T(2k+1)^n , T2k^n are two
0 E [ XT(2k+1)^n XT2k^n ]
= E [( XT(2k+1)^n XT2k^n )1{n < T2k }] + E [( XT(2k+1)^n XT2k^n )1{ T2k n < T2k+1 }]
1
P {n T2k+1 } E [( Xn a) 1{ T2k 1 n < T2k+1 }]. (5.1b)
b a
{Uab [ X; n] k + 1} = [ j b
k +1 {Ua [ X; n ] = j} ✓ { T2k+1 n}.
Therefore, applying these two conditions on (5.1b) gives the desired inequality:
1
P {Uab [ X; n] k + 1} P { n T2k+1 } E [( Xn a) 1{Uab [ X; n] = k }].
b a
1
 P{Uab [X; n] k + 1}
b a
E [( Xn a ) ].
k 0
55
The LHS is essentially E [Uab [ X; n]]:
• • j 1
 P{Uab [X; n] k + 1} =   P {Uab [ X; n] = j} =   P{Uab [X; n] = j}
k 0 k 0 j = k +1 j =1 k =0
•
= Â jP {Uab [ X; n] = j} = E [Uab [ X; n]].
j =1
2. We may use the similar technique to finish the proof on the second part. Apply
0 E [ XT(2k 1)^n
XT2k^n ]
= E [( XT(2k 1)^n
XT2k^n )1{n < T2k 1 }] + E [( XT(2k 1)^n
XT2k^n )1{ T2k 1 n < T2k }]
+ E [( XT(2k 1)^n
XT2k^n )1{n T2k }]
= E [( XT2k 1
Xn )1{ T2k 1 n < T2k }] + E [( XT2k 1
XT2k )1{n T2k }]
= E [( a Xn )1{ T2k 1 n < T2k }] + E [(b a)1{ T2k 1 n < T2k }] + E [(b a )1{ n T2k }]
Or equivalently,
1
P {n T2k 1} E [( Xn a)+ 1{ T2k 1 n < T2k }]
b a
T2k 1 }, we imply
1
P {Uab [ X; n] k} E [( Xn a)+ 1{Uab [ X; n] = k }]
b a
56
R The upcrossing of { Xn } on the interval is the same as teh upcrossing of
{ Xn } on [ b, a]. Using this fact, we can assert that
⇣ ⌘ 1 h i
P Uab [ X; n] k E ( Xn b) 1{Uab [ X; n] = k } .
b a
⇣ ⌘ 1 h i
P Uab [ X; n] k+1 E ( Xn b)+ 1{Uab [ X; n] = k } .
b a
From the upcrossing inequality, we can easily get the result for the convergence of
a martingale.
martingale which is L1 -bounded, i.e., supn E [| Xn |] < •. Then there exists a random
Proof. • Firstly, we study the limit of {Uab [ X; n]}n 1, which is guaranteed to exist
where (5.2a) is by the Fatou’s lemma and (5.2b) is by the Doob’s upcrossing
• Note that the result in the first part holds for any rational a, b with a < b, which
57
means that P (Uab [ x ] < •) = 1, 8 a, b 2 Q, a < b. Therefore, we can show that
[
P [W ] = 0, where W = {lim inf Xn < a < b < lim sup Xn }.
a,b2Q,a<b
• Given that { Xn } is lower bounded by zero, the remaining result can be shown by
R
upper bounding the integral A E [ X• | Fn ] for any A 2 Fn :
Z Z
E [ X• | F n ] = X• dP
A A
Z
lim inf Xm dP (5.3a)
m!• A
Z
= lim inf E [ Xm | Fn ] dP (5.3b)
m!• A
Z Z
lim inf Xn dP = Xn dP. (5.3c)
m!• A A
continuous-time.
58
1. Xt 2 L1 , 8t;
t 0.
t > 0.
It is easy to check that a stopping time is always an optional time. Now we discuss an
⌅ Example 5.1 Let T be an optional time. For each n 1, define the step-function
mapping 8
> k
< n , if (k 1)/2n T < k/2n , k = 1, 2, . . .
Tn = 2
>
: •, if T = •
⇢
[ k 1 k
{ Tn t} = { Tn t} \ T<
k 1
2n 2n
b t · 2n c ⇢ b t · 2n c ⇢
[ k 1 k [ k 1 k
= { Tn t} \ T< n = T< .
k =1
2n 2 k =1
2n 2n
⇢
k 1 k
T< 2 Fk/2n ✓ Ft , 8k t · 2n =) { Tn t} 2 Ft , 8t.
2n 2n
59
• The result for Tn # T can be found in MAT3006 knowledge:
Daniel Wong, Jie Wang. (2019) Lecture Notes for MAT3006: Real
https://walterbabyrudin.github.io/information/Updates/Updates.html
T
Definition 5.4 1. Let {Ft }t 0 be a filteration. Define Ft+ , s>t Fs . Then {Ft+ }t 0
Ft+ , 8t.
F.
3. A filteration {Ft }t 0 is called an augmented filteration, or said to satisfy the
5.2. Thursday
F T , { A 2 F : A \ { T t} 2 Ft , 8t 0}.
60
Proposition 5.1 Let {Ft } be a filtration. Define
F T + , { A 2 F : A \ { T t } 2 Ft+ , 8 t 0}
G T , { A 2 F : A \ { T < t } 2 F t , 8 t > 0}
1
P ( Xt⇤ l) E [| Xt |].
l
L1 and Xn ! X• a.s.
and E [ X• | Ft ] = Xt a.s. for any t 0. Let S T be two {Ft }-optional times. Then
particular, E [ XT ] = E [ X0 ].
Let’s first show a necessary and sufficient condition for the existence of X• :
61
Proof. Assume that { Xt }t 0 is uniformly integrable, which implies that { Xt }t 0 is L1 -
bounded. By the martingale convergence theorem 5.3, there exists a random variable
P
X• such that Xt ! X• a.s. and X• 2 L1 . In particular, Xt ! X• . Together with the
Z Z
X• dP = lim Xt dP
A t!• A
Z Z Z
= lim E [ Xt | Fu ] dP = lim Xu dP = Xu dP.
t!• A t!• A A
Now we assume that the last element exists. Then { Xt } is a collection of conditional
Proof. • Firstly, construct the approximation of S, T and argue the similar optional
8
> k
< n, if (k 1)/2n S < k/2n , k = 1, 2, . . .
Sn = 2
>
: •, if S = •
8
> k
< n, if (k 1)/2n T < k/2n , k = 1, 2, . . .
Tn = 2
>
: •, if T = •
theorem,
R R
Therefore, for any A 2 FSn , A
XTn dP = A
XSn dP.
62
• We claim that FS+ = \n 1 F Sn . Therefore, for any A 2 FS+ ,
Z Z
XTn dP = XSn dP. (5.4)
A A
Z Z
XSn dP = XSn+1 dP.
A A
in L1 and XSn ! XS in L1 .
• Then we can show that E [ XT | FS+ ] = XS a.s. as the follwoing. For any A 2 FS+ ,
Z Z
E [ XT | FS+ ] dP = XT dP
A A
Z
= lim XTn dP
n!• A
Z
= lim XSn dP
n!• A
Z
= XS dP
A
where the second and the last equality is because of the L1 convergence, and the
63
Chapter 6
Week6
6.1. Tuesday
At the beginning of this lecture, let’s fill the gap for the Theorem 5.3.
Proposition 6.1 Suppose that Tn is a positive stopping time and T < Tn conditioned
F T + , { A 2 F : A \ { T t } 2 Ft+ , 8 t 0}
, { A 2 F : A \ { T < t } 2 F t , 8 t > 0}
= [n { Tn > t}c = [n { Tn t}
• Suppose that A 2 \•
n=1 F Tn , then A \ { Tn t } 2 Ft , 8 t, 8 n 1. As a result,
[
Ft 3 { A \ { Tn t}} = A \ [[n { Tn t}]
n
= A \ { T < t }, 8t.
In other words, A 2 FT + .
65
Therefore,
In other words, A 2 \•
n=1 FTn .
F ◆ F1 ◆ · · · ◆ F n ◆ · · · .
Suppose that { Xn }• •
n=1 is a backward submartingale w.r.t {Fn }n=1 , i.e., i) Xn is Fn -
measurable, ii) E [| Xn |] < •, iii) E [ Xn | Fn+1 ] Xn+1 , a.s. If limn!• E [ Xn ] > •, then
the sequence { Xn }•
n=1 is UI.
Proof. Note that the limit of E [ Xn ] exists since it is decreasing. Denote c , limn!• E [ Xn ].
E [| Xn |]
8l > 0, P (| Xn | > l)
l
1
= 2E [ Xn+ ] E [ Xn ]
l
1
2E [ X1+ ] c < •.
l
where the first inequality is because that E [ Xn ] # c and { Xn+ } is a backward submartin-
for any # > 0, there exists d such that for any A 2 F1 , P ( A) < d,
Z
| X1 | dP < #. (6.1)
A
66
We can choose l sufficiently large such that P (| Xn | > l) < d, 8n. As a result,
Z Z
| Xn | dP | Xn 1 | dP ···
{| Xn |>l} {| Xn |>l}
Z
| X1 | dP < #
{| Xn |>l}
where the first inequality is because that E [| Xn 1 |Fn ] | Xn |, and the last inequality is
R
because of (6.1). Therefore, {|Xn |>l} | X1 | dP < # for any n. The proof is complete. ⌅
6.1.1. Localization
The concepts of stopping times provide a tool of “localizing” quantities.
(W, F , {Ft }, P ), and T is a stopping time. Define the stopped process { Xt^T }t 0 such
that
Xt ^ T ( w ) = Xt ^ T ( w ) ( w ), 8w 2 W.
Note that a martingale is a local martingale. Now we give a sufficient condition such
Theorem 6.1 Suppose that { Xt }t 0 is an {Ft }-adapted local martingale, and there
67
Proof. Considering that i) Xt^Tn ! Xt a.s. because Tn " •, ii) | Xt^Tn | supn | Xt^Tn |, with
the random variable supn | Xt^Tn | integrable, we can apply the dominated convergence
L1
theorem to show that Xt^Tn ! Xt , 8t, as n ! •.
Z Z
E [ Xt | Fs ] dP = Xt dP
A A
Z
= lim Xt^Tn dP
n!• A
Z
= lim Xs^Tn dP
n!• A
Z
= Xs dP
A
Robert Brown. Now we give a way for constructing the Brownian motion.
1. P ( B0 = 0) = 1;
✓ ◆
P Bt2 Bt1 x1 , . . . , Btk Btk 1
xk 1 = ’ P ( Bt j Bt j 1
xj 1)
2 j k
68
R In some situations, the first condition may not be satisfied. Instead, the process
may start at a non-zero point x. Then we write such a process { x + Bt }.
Definition 6.4 [Canonical Wiener Measure] Let the sample space be W = C [0, •), and
continuos function with support [0, •). Define Bt (w ) = w (t). A probability measure P
on (C [0, •), B) is called a Wiener measure if conditions (1)-(3) in Definition 6.3 are
(C [0, •), B , P ). ⌅
Theorem 6.2 — Existence and Uniqueness of Wiener Measure. For each s > 0, there
6.2. Thursday
1. Joint distribution: Fix 0 t1 < t2 < · · · < tk . Given x1 , x2 , . . . , xk 2 R, the joint den-
sity of ( Bt1 , Bt2 , . . . , Btk ) in ( x1 , x2 , . . . , xk ) is equal to the joint density of ( Bt1 , Bt2
!
k
1 ( x j x j 1 )2
’q exp
2( t j t j 1 )
.
j =2 2p (t j tj 1)
2. Stationary: For any s > 0, define Bts = Bt+s Bs , t 0. Then { Bts }t 0 is a Brownian
motion.
3. Scaling:
69
• For each c > 0, { Bt/c }t 0 is a Brownian motion with variance 1/c;
p
• (Scaling invariance / self-similarity) By previous two properties, { cBt/c }t 0
{ B̂t } with B̂t = tB1/t for t > 0, and B̂0 = 0. Then { B̂t } is a standard Brownian
motion.
Proof on the first four parts. 1) can be shown by the indepedent increments and normal
cov( Bt , Bs ) = E [ Bt Bs ] E [ Bt ]E [ Bs ]
= E [( Bt Bs + Bs ) Bs ]
= E [( Bt Bs ) Bs ] + E [ Bs2 ]
= E [ Bt Bs ]E [ Bs ] + E [ Bs2 ]
= s.
Proof on the time reversal part. We need to check those four conditions in Definition 6.3
Since B1/t B1/s ⇠ N (0, 1/s 1/t), we imply s( B1/t B1/s ) ⇠ N (0, s2 (1/s
s2 (1/s 1/t) = t s.
70
• In order to check condition (2), fix t1 < t2 < t3 . It suffices to check B̂t3 B̂t2 and
B̂t2 B̂t1 are independent. Considering that these two r.v.’s are jointly normal, it
= Var( B̂t3 B̂t2 ) + Var( B̂t2 B̂t1 ) + 2Cov( B̂t3 B̂t2 , B̂t2 B̂t1 )
• Finally we check the condition (4). Since the continuity of { B̂t } holds at any
n ! •.
– Then we show that B̂t ! 0 for other values of t. Fix any s 2 (n, n + 1), note
that
Bs Bn Bs Bn Bn Bn
+
s n s s s n
1 1 1
= | Bs Bn | + | Bn |
s s n
1 1
sup | Bs Bn | + 2 | Bn |
n n s n +1 n
Bn |, then
Bs Bn Zn 1
sup + 2 | Bn |
n < s < n +1 s n n n
71
We claim that for any # > 0,
⇢
Zn (w )
P w2W: > #, infinitely often = 0.
n
Zn (w )
Then n ! 0 for almost all w 2 W. As a result,
Bs Bn
sup ! 0, a.s.,
n < s < n +1 s n
– Then we need to show the correctness of our claim. By the stationary and
Z •
E [ Z0 ] = E [ sup | Bs |] = P ( Z0 > x ) dx
0 s 1 0
• Z ( n +1) #
= Â P ( Z0 > x ) dx
n=0 n#
• Z ( n +1) #
 P ( Z0 > (n + 1)#) dx
n=0 n#
• •
= Â #P(Z0 > (n + 1)#) = # Â P(Z0 > n#)
n =0 n =1
•
= # Â P ( Z0 /n > #)
n =1
•
= # Â P ( Zn /n > #)
n =1
We claim that E [ Z0 ] < • (which will be shown in the next lecture), which
implies that
•
 P(Zn /n > #) < •.
n =1
72
Chapter 7
Week7
7.1. Tuesday
{Ft }t 0 be the natural filteration, i.e., Ft = s( Bu : u t). Suppose that a > 0, define
Ta , inf{t 0 : Bt = a}.
Then Ta is the first time that Brownian motion hits level a. By convention, inf ∆ = +•.
P ( Ta < •) = 1.
✓ ◆
q2 t
Ztq = exp qBt , t 0.
2
As a result,
2 s2 /2
• Since E [eqX ] = eq for X ⇠ N (0, s2 ),
✓ ◆
q2 t
E [| Ztq |] = E exp qBt = 1, 8t
2
73
• For any 0 u < t, we have
✓ ◆
q2
E [ Ztq | Fu ] = E exp qBt Fu
t
✓ ◆ ✓ ◆
q 2 (t u) q2 u
= E exp q ( Bt Bu ) exp qBu Fu
2 2
✓ ◆ ✓ ◆
q2 u q 2 (t u)
= exp qBu · E exp q ( Bt Bu ) Fu
2 2
✓ ◆ ✓ ◆
q2 u q 2 (t u)
= exp qBu · E exp qBt u = Zuq
2 2
Therefore, { Ztq } is a martingale w.r.t. {Ft }t 0. Now we compute limt!• E [ ZTq a 1{ Ta <
q 2 t/2
Ztq 1{ Ta = •} eqa ! 0,
It follows that
q 2 Ta /2
= E [ ZTq a 1{ Ta < •}] = E [eqa 1{ Ta < •}].
74
Therefore, E [e q 2 Ta /2 1{ T < •}] = e qa . Since e q 2 Ta /2 1{ T < •} is increasing, by mono-
a a
qa q 2 Ta /2
1 = lim e = lim E [e 1{ Ta < •}] = E [1{ Ta < •}] = P [1{ Ta < •}].
q !0 q !0
Z •
2 x2
P ( Mt a) = 2P ( Bt a) = p e 2t dx.
2pt a
P ( Bt a) = P ( Bt a, Mt a) + P ( Bt a, Mt < a)
= P ( Bt a, Mt a)
= P ( Bt a | Mt a ) P ( Mt a)
= P ( Bt a | Ta t)P ( Mt a)
= P ( Bt BTa 0 | Ta t)P ( Mt a)
P ( Bt BTa 0 | Ta t) = 12 . Therefore,
Z •
1 1 x2
P ( Bt a ) = P ( Mt a) = p e 2t dx.
2 2pt a
75
7.1.2. Distributions of Brownian Motion
Theorem 7.3 The joint distribution of the Brownian motion and the running maxi-
Z •
1 x2
P ( Mt a, Bt a y) = P ( Bt a + y) = p e 2t dx.
2pt a+y
P ( Bt a + y) = P ( Bt a + y | Mt a ) P ( Mt a)
= P ( Bt a + y | Ta t)P ( Mt a)
= P ( Bt BTa y | Ta t)P ( Mt a)
= P ( Bt BTa y | Ta t)P ( Mt a)
= P ( Bt a y | Mt a ) P ( Mt a ).
p
lTa 2la
E [e ]=e .
✓ ◆
a 1 a2
f Ta (t) = p p exp .
t t 2p 2t
R•
Computing the integral 0
f Ta (t)e lt dt gives the desired result. ⌅
p q 2 Ta /2 1{ T
Another Quick Proof. Since P ( Ta < •) = 1, substituting q with 2l for E [e a <
•}] = e qa gives the desired result. ⌅
76
Theorem 7.5 Consider the Brownian motion with drift:
Xt , µt + sBt ,
where µ 6= 0, s > 0.
✓ ◆
2| µ |
P ( M• > y) = exp y , y 0.
s2
The first two parts are trivial, and we give a proof for the last part:
Proof. Choose some q 6= 0 and define the random process {Vtq }t 0 such that
✓ ◆
s2 q 2
Vtq = exp qXt µqt t .
2
✓ ◆
s2 q 2
E [Vtq | Fu ] = E exp qXt µqt t
Fu
2
✓ ◆
s2 q 2 s2 q 2
= E exp q ( Xt Xu ) (µq + )(t u) e(qXu (µq + 2 )u) Fu
2
✓ ◆
2 2
(qXu (µq + s 2q )u) s2 q 2
=e E exp q ( Xt Xu ) (µq + )(t u)
2
= Vuq .
2µ
For a < 0 < b, define Ta,b = inf{t 0 : Xt = a or Xt = b}. Define q = s2
such that
a.s.
µq + 12 s2 q 2 = 0. Considering that Ta,b Ta with Ta being finite a.s., we imply Vtq^Ta,b !
77
VTqa,b . Moreover, |Vtq^Ta,b | max(e qa , eqb ). By dominated convergence theorem,
Therefore,
eqb 1
P ( Ta,b = a) = .
eqb eqa
a2 < a1 < 0 =) A a2 ✓ A a1 .
\
L , { M• < b} = {w 2 W : Xt (w ) never hits level b} = Aa .
a<0,a2Q
Thus
✓ ◆
\
qb
P Aa = lim P ( A a ) = lim P ( Ta,b = a) = 1 e .
a!• a!•
a<0,a2Q
2|µ|/s2
P ( M• < b ) = P ( L ) = 1 e b.
78
7.2. Thursday
f : [ a, b] 7! R.
where the supremum is taken over all the possible partitions P on the interval [ a, b]. Since
TV ( f )[ a, b] = lim
kPk!0 k
 | f (tk ) f (tk 1 )|
R The real analysis shows that a bounded variation function, i.e., the function
whose total variation is bounded, is differentiable almost everywhere. Then if
the function is nowhere differentiable, it is not of bounded variation.
Theorem 7.6 Brownian motion is not of bounded variation almost surely, i.e.,
79
This result is based on the fact that the Brownian motion is nowhere differentiable
alomst surely.
✓⇢ ◆
B (w ) Bt (w )
P w 2 W : lim sup t+h = •, t 2 [0, •) =1
n!• h
⇢
( M) Bt+h (w ) Bt (w )
A , w 2 W : 9t 2 [0, T ] such that lim sup M .
h !0 h
that when n n0 ,
Bu (w ) Bt (w )
2M, 8u 2 (t 2/n, t + 2/n).
u t
⇢
( M)
An , w 2 W : 9t 2 [0, T ] such that | Bu (w ) Bt (w )| 2M |u t |, 8u 2 (t 2/n, t + 2/n) .
(7.1)
( M) ( M)
Then i) A( M) ✓ [n An , and ii) { An } is monotone: An ✓ An+1 .
Z : j/n t}, which means k is close enough to t. Define Yk as the maximal of three
independent increments:
n o
Yk = max B(k+2)/n B(k+1)/n , B(k+1)/n Bk/n , Bk/n B(k 1)/n .
80
( M)
We can show that Yk (w ) 6M/n, 8w 2 An as follows. Firstly,
where the last inequality is because that k/n t < (k + 1)/n. Following the similar
6M 6M
B(k+1)/n (w ) Bk/n (w ) , Bk/n (w ) B(k 1)/n ( w ) =) Yk (w ) .
n n
( M)
Now define the new set based on the consequence of the claim about An :
⇢
( M) 6M
En , w 2 W : 9 j 2 [1, Tn ] \ Z such that Yj (w ) .
n
with
n o
Yj = max B( j+2)/n B( j+1)/n , B( j+1)/n Bj/n , Bj/n B( j 1)/n .
( M) ( M) ( M)
Directly An ✓ En for each n. Now we begin to upper bound P ( En ):
( M) 6M
P ( En ) Â P (Yj
n
)
1 j Tn
✓ n o ◆
6M
Tn P max B( j+2)/n B( j+1)/n , B( j+1)/n Bj/n , Bj/n B( j 1)/n
n
✓ ◆
6M
= Tn · ’ P B(i+1)/n B(i)/n (7.2a)
i = j 1:j+1
n
✓ ◆ 3
6M
= Tn · P | B1/n | , (7.2b)
n
where (7.2a) is because of the independent increments of Brownian motion, and (7.2b)
81
is because of its stationary increment property. In particular,
✓ ◆ ✓ ◆
6M 6M 6M
P | B1/n | =P B1/n
n n n
✓ ◆
6M 6M
=P p B1 p (7.2c)
n n
Z 6M
1 p
n x2 /2
=p e dx
2p 6M
p
n
2 6M
p p (7.2d)
2p n
( M) ( M)
P ( A( M) ) P ([n An ) = lim P ( An ) = 0.
n!•
82
Chapter 8
Weak 8
8.1. Thursday
h i
lim E ( Q(P(n) ) T )2 = 0.
n!•
83
We claim that q j , qk are uncorrelated for j 6= k:
⇣ ⌘
E [q j qk ] = E ( Bt j Bt j 1 )2 (t j tj 1) ( Btk Btk 1 )2 (tk tk 1)
h i ⇥ ⇤
= E ( Bt j Bt j 1 )2 ( Btk Btk 1 )2 (t j t j 1 )E ( Btk Btk 1 )2
h i
(tk tk 1 )E ( Bt j Bt j 1 )2 + (t j t j 1 )(tk tk 1 )
h i ⇥ ⇤ ⇥ ⇤
= E ( Bt j Bt j 1 ) E ( Btk Btk 1 )2
2
(t j t j 1 )E ( Btk Btk 1 )2
h i
(tk tk 1 )E ( Bt j Bt j 1 )2 + (t j t j 1 )(tk tk 1 )
= (t j tj 1 )( tk tk 1) (t j tj 1 )( tk tk 1)
(tk tk 1 )( t j tj 1) + (t j tj 1 )( tk tk 1) = 0.
" #
h i
E ( Q ( P(n) ) T )2 = E ( Â q k )2
k
⇥ ⇤ ⇥ ⇤
= Â E qk2 + Â E q j qk
k j6=k
h 2
i
= ÂE ( Btk Btk 1 )2 (tk tk 1)
k
h i
= Â E ( Btk Btk 1 )4
k
⇥ ⇤
2 Â(tk tk 1 )E ( Btk Btk 1 )2 + Â(tk tk 1)
2
k k
= 3 Â(tk tk 1)
2
2 Â(tk tk 1)
2
+ Â(tk tk 1)
2
k k k
= 2 Â(tk tk 1)
2
2k P ( n ) k  ( t k tk 1)
k k
= 2T · kP(n) k ! 0.
R The Theorem 8.1 shows that the quadratic variation of Brownian motion on
the interval [0, T ] converges to T in L2 for any T. This implies that Q(P(n) ) !
T in probability as kP(n) k ! 0. Then there exists a subsequence of P(n) , such
84
that Q(P(n) ) ! T almost surely.
lim n2 · kP(n) k = 0,
n!•
h i
⇣ ⌘ E ( Q ( P(n) ) T )2 2T kP(n) k T
P ( Q ( P(n) ) T )2 > 2dn = 2.
2dn 2dn n
⇣ ⌘
ÂP ( Q(P (n)
) 2
T ) > 2dn < •.
n
By Borel-Cantelli Lemma,
⇣ ⌘
P ( Q ( P(n) ) T )2 > 2dn , infinitely often = 0.
p p
| Q ( P(n) , w ) T| > 2dn , for finite n =) | Q(P(n) , w ) T| 2dn , n ! •.
| Q ( P(n) , w ) T | ! 0.
The Brownian motion is nowhere differentiable almost surely and does not have
bounded variation. However, it turns to have finite quadratic variation limit. Using this
quadratic variation property, we can go back to show that the Brownian motion does
85
not have bounded variation.
A Direct Proof for Theorem 7.6. Since the Brownian motion is almost surely continuous,
Define the set L = {w 2 W : TV ( B)[0, t] < •}. Then for any partition P, if w 2 L,
 | Bt (w )
k
Btk 1 (w )| TV ( B· (w ))[0, t] < •
k
Then P (limn!• Q(P(n) ) = 0) P (L) > 0, where kP(n) k ! 0. Choose some e 2 (0, t).
86
Chapter 9
Week9
9.1. Tuesday
nian motion on (W, F , P ) and {Ft }t 0 be the natural filtration, i.e., Ft , s({ Bu : u t}).
Z t
Xs dBs , t 0. (9.1)
0
Z t
0
Xs dBs , lim
kp k!0 k
 Xt k 1
· ( Btk Btk 1 ), (9.2)
where the limit is taken in L2 sense along the partition p defined on the interval [0, t],
It is reasonable to study how the limit for (9.2) looks like, by considering a simple
87
Definition 9.1 [Simple Stochastic Process]
Z T
E Xt2 dt < •, 8 T > 0.
0
Xt ( w ) = Xt j ( w ), t 2 [t j , t j+1 ), j = 0, 1, . . . ,
Yt (w ) = Xt j+1 (w ), t 2 ( t j , t j +1 ] .
process.
Definition 9.2 [Ito Integral for Simple Stochastic Process] Suppose that { Xt }t 0 is a
Z T n 1
0
Xt (w ) dBt (w ) = Â Xt (w ) · [ Bt + (w )
k k 1
Btk (w )] + Xtn (w ) · [ BT (w ) Btn (w )]
k =0
(9.3)
where n , max{ j 2 N : t j T }. ⌅
Proposition 9.1 The Ito Integral for Simple Stochastic Process admits the following
properties:
88
1. Linearity: Suppose that { Xt }t 0 , {Yt }t 0 2 L20 and a, b 2 R, then
Z T Z T Z T
(aXt + bYt ) dBt = a · Xt dBt + b · Yt dBt
0 0 0
Rt
3. Define the new random variable It [ X ] , 0
Xu dBu , then the process { It [ X ]}t 0 is
E [( It [ X ])2 ] < •, 8t 0.
(1) (2)
Proof for Part 1). Denote {tk }k 0 and {tk }k 0 as the paritions corresponding to sim-
these two partitions. With respect to this new sequence {tk }, processes { Xt }t 0 , {Yt }t 0
are still simple. Then {aXt + bYt } is also a simple process corresponding to {tk }. The
Proof for Part 2). Note that the left side in this part can be expanded as the following:
"✓Z 2 !2 3
T
◆2 # n 1
E
0
Xt dBt = E4 Â Xt k
· [ Btk+1 Btk ] + Xtn · [ BT Btn ] 5
k =0
h i
= Â E Xtk1 Xtk2 ( Btk1 +1 Btk1 )( Btk2 +1 Btk2 )
0k1 ,k2 n 1
⇥ ⇤ n 1 ⇥ ⇤
+ E Xt2n · ( BT Btn )2 + 2 Â E Xtk Xtn · ( Btk+1 Btk )( BT Btn )
k =0
⇥ ⇤
where the second term equals E Xt2n · ( T tn ) by the independent incremental prop-
⇥ ⇤ ⇥ ⇤
E Xtk ( Btk+1 Btk ) = E E [ Xtk ( Btk+1 Btk ) | Fk
⇥ ⇤ ⇥ ⇤
= E Xtk E [( Btk+1 Btk ) | Fk ] = E Xtk E [( Btk+1 Btk )] = 0.
89
Following the similar trick, for 0 k1 < k2 n 1, we have
h i h i
E Xtk1 Xtk2 ( Btk1 +1 Btk1 )( Btk2 +1 Btk2 ) = E E [ Xtk1 Xtk2 ( Btk1 +1 Btk1 )( Btk2 +1 Btk2 ) | Ftk2 ]
h i
= E Xtk1 ( Btk1 +1 Btk1 ) Xtk2 E [( Btk2 +1 Btk2 )] = 0,
where the second equality is because that all random variables except Btk2 +1 are Ftk2 -
⇥ ⇤ ⇥ ⇤
E Xt2k ( Btk+1 Btk )2 = E Xt2k E [( Btk+1 Btk )2 | Ftk ]
⇥ ⇤
= E Xt2k (tk+1 tk ) .
Therefore,
"✓Z ◆2 #
T n 1 ⇥ ⇤ ⇥ ⇤
E
0
Xt dBt = ÂE Xt2k (tk+1 tk ) + E Xt2n · ( T tn )
k =0
" # Z
n 1 T
=E Â Xt2k (tk+1 tk ) + Xt2n · (T tn ) = E
0
Xt2 dt .
k =0
✓Z t
◆2 Z t
2
E [( It [ X ]) ] = E [ Xu dBu ]=E Xu2 du .
0 0
Now we begin to show that { It [ X ]}t 0 is a martingale with respect to {Ft }t 0. For
" #
n 1
E [ It [ X ] | Fs ] = E Â Xt ( Bt +
k k 1
Btk ) + Xtn ( Bt Btn ) Fs .
k =0
90
Considering separate the summation from 0 to n 1 into two parts, we further have
" #
n0 1
E [ It [ X ] | Fs ] = E Â Xt ( Bt + k k 1
Btk ) + Xtn0 ( Bs Btn0 ) Fs
k =0
h i
+ E Xtn0 ( Btn0 +1 Bs ) Fs
" 0 #
n 1
+E Â Xtk ( Btk+1 Btk ) + Xtn ( Bt Btn ) Fs
k = n 0 +1
n0 1
 Xt ( Bt +
k k 1
Btk ) + Xtn0 ( Bs Btn0 ),
k =0
n0 1 ⇥ ⇤
Â0 E Xtk E [ Btk+1 Btk | Ftk ] Fs + E [ Xtn E [ Bt Btn | Ftn ]|Fs ] = 0.
k = n +1
As a result,
n0 1
E [ It [ X ] | Fs ] = Â Xt ( Bt + k k 1
Btk ) + Xtn0 ( Bs Btn0 )
k =0
Z s
= Xu dBu .
0
91
9.2. Thursday
⇢ !
P w 2 W : sup | Xt (w )| M and Xt (w ) continuous on t 0 = 1.
t 0
(n)
Then for given T > 0, there exists a sequence of simple processes { Xt }t 0 2 L20
such that
Z T
(n)
lim E ( Xt Xt )2 dt = 0.
n!• 0
Z T
(n)
lim E ( Xt Xt )2 dt = 0.
n!• 0
3. Let { Xt }t 0 2 L2 . Then for given T > 0, there exists a sequence of almost surely
(n)
bounded processes { Xt }t 0 2 L2 , 8 n 1 such that
Z T
(n)
lim E ( Xt Xt )2 dt = 0.
n!• 0
(n)
Proof for Part 1). Construct { Xt }t 0 as the following. Pick a sequence of partitions
{P(n) } on the interval [0, T ] with kP(n) k ! 0. For each n, define the stochastic process
(n)
{ Xt } with
(n) (n) (n) (n)
Xt ( w ) = Xt j ( w ), for t 2 [t j , t j+1 ),
92
(n) (n) (n)
where P(n) , {0 = t0 < t1 < · · · < T }. It is clear that { Xt }t 0 is a simple process.
⇢
L= w 2 W : sup | Xt (w )| M and Xt (w ) continuous on t 0
t 0
r
#
| Xs ( w ) Xt (w )| < , for any |s t| < d.
T
r Z T
(n) # (n)
8t 2 [0, T ], | Xt ( w ) Xt (w )| < =) [ Xt ( w ) Xt (w )]2 dt < #.
T 0
RT (n)
Therefore, the random variable 0
[ Xt Xt ]2 dt ! 0 almost surely. For w 2 L, { Xt (w )}
(n)
is bounded by M, and thus { Xt (w )} is bounded by M as well. Thus the random
R T (n)
variable 0 [ Xt Xt ]2 dt is upper bounded by (2M )2 · T almost surely. By the bounded
convergence theorem,
Z T
(n)
lim E ( Xt Xt )2 dt = 0.
n!• 0
(n)
Proof for Part 2). Construct { Xt }t 0 as the following. For each n, pick a non-negative
1
1. fn ( x ) = 0 for x 2 ( •, n] [ [0, •);
R•
2. • n
f ( x ) dx = 1.
(n)
Then define the process { Xt }t 0 with
Z t
(n)
Xt ( w ) ⌘ fn ⇤ X· (w ) |0t , fn (s t) Xs (w ) ds, 8w 2 W.
0
93
Then for each w 2 L,
Z t Z t
(n)
| Xt (w )| fn (s t)| Xs (w )| ds M · fn (s t) ds M, 8t 0, 8n.
0 0
(n)
Therefore, the process { Xt }t 0 is almost surely bounded. Moreover, by definition,
(n)
{ Xt } t 0 is continuous a.s. and {Ft } adapted.
RT (n)
Now we begin to show that 0
[ Xt Xt ]2 dt ! 0 almost surely. Take w 2 L, then
Z T Z T
(n) 2 (n)
[ Xt ( w ) Xt (w )] dt 2M · | Xt ( w ) Xt (w )| dt,
0 0
where the integral on the RHS can be upper bounded as the following:
Z T Z T Z t
(n)
| Xt ( w ) Xt (w )| dt = fn (s t) Xs (w ) ds Xt (w ) dt (9.4a)
0 0 0
Z T Z •
= fn ( s) Xt s (w ) ds Xt (w ) dt (9.4b)
0 0
Z T Z • Z •
= fn ( s) Xt s (w ) ds fn ( s) Xt (w ) ds dt
0 0 0
(9.4c)
Z TZ •
fn ( s)| Xt (w ) Xt s (w )| ds dt (9.4d)
0 0
Z • Z T
= fn ( s) | Xt ( w ) Xt s (w )| dt ds (9.4e)
0 0
R•
where (9.4b) is by the change of variable s0 = t s; (9.4c) is because that • fn (s) ds =
R0 RT
f (s) ds = 1; (9.4e) is by the Fubini’s theorem. We claim that the term 0 | Xt (w )
• n
Xt s (w )| dt is small when s is small, i.e., for any # > 0, there exists d such that when
s < d,
Z T
| Xt ( w ) Xt s (w )| dt < #. (9.5)
0
94
We can further apply (9.5) to upper bound the term (9.4e):
Z • Z T
fn ( s) | Xt ( w ) Xt s (w )| dt ds
0 0
Z d Z T Z • Z T
= fn ( s) | Xt ( w ) Xt s (w )| dt ds + fn ( s) | Xt ( w ) Xt s (w )| dt ds
0 0 d 0
Z • Z •
# fn ( s) ds + 2MT · fn ( s) ds = #.
d d
1
where the last equality holds when we choose n large enough such that d n. Thus
Z T
(n)
| Xt ( w ) Xt (w )| dt ! 0 for w 2 L.
0
Finally, we show the claim in (9.5) by discussing cases for continuous and discon-
tinuous Xt :
• When Xt is continuous on t 2 [0, T ], we show that for any # > 0, there exists d
such that
Z T
8s < d, | Xt Xt s | dt < #.
0
#
| Xt Xt s | < , 8t 2 [s, T ].
2T
As a result,
Z T Z s Z T
| Xt Xt s | dt = | Xt Xt s | dt + | Xt Xt s | dt
0 0 s
#
Md + T · < #.
2T
• When Xt is not continuous on t 2 [0, T ], we can also show the same result:
Because the continuous functions are dense in L p (1 p < •), for any # > 0, there
95
exists a continuous function X̂t such that
Z T
#
| Xt X̂t | dt < . (9.7)
0 3
As a result,
Z T Z T Z T Z T
| Xt Xt s | dt | Xt X̂t | dt + | Xt s X̂t s | dt + | X̂t s X̂t | dt
0 0 0 0
<#
where the first two terms are all bounded by #/3 because of (9.7), and the last
(n)
Proof of Part 3). For each n, we construct the almost surely bounded process { Xt }t 0
(n)
Therefore, | Xt (w )| | Xt (w )|, 8w 2 W. Together with the inequality ( a + b)2 2a2 +
2b2 , we have
Z T Z T Z T Z T
(n) (n)
( Xt Xt )2 dt 2 ( Xt )2 dt + 2 ( Xt )2 dt 4 ( Xt )2 dt < •, 8t.
0 0 0 0
RT (n)
Therefore, 0
Xt )2 dt is dominated by an integrable random variable. Moreover,
( Xt
(n) RT
substituting the form of Xt and considering 0 ( Xt )2 dt < • gives
Z T Z T Z T
(n) 2 2 2
( Xt Xt ) dt ( Xt ) 1{( Xt ) n} dt + ( Xt )2 1{( Xt )2 n} dt ! 0, as n ! •.
0 0 0
96
Combining Theorem 9.1 part 1) to 3), we conclude that for { Xt }t 0 2 L2 , there
(n)
exists a sequence of simple processes { Xt }t 0 2 L20 such that
Z T
(n)
lim E ( Xt Xt )2 dt = 0.
n!• 0
97
Chapter 10
Week10
10.1. Tuesday
Recall the property 3) in proposition 9.1, in which we show that when { Xu }u is a simple
nR o
t
process, 0
X u dBu is a square integrable, almost surely continuous martingale. In
t
this lecture, we first review some basic ideas about square integrability.
2. Let U 2 be the class of square integrable, right-continuous, with left limit exist
martingales.
3. Let Uc2 be the class of square integrable, almost surely continuous martingales.
1/2
k X k , E [ XT2 ] , { Xt } t 0 2 U 2.
99
Theorem 10.1 — Completeness of Square Integrable Martingales. With respect to the
norm k · k,
k X (n) X k ! 0, as n ! •.
• Since { X (n) }n is a Cauchy sequence, for any # > 0, there exists N such that
as long as m, n > N,
(m) (n)
Since { Xt Xt } t 0 is a martingale, by convexity of quadratic function,
(m) (n)
{( Xt Xt )2 } t 0 is a sub-martingale, which implies
(n)
This means that for fixed t T, { Xt }n is a Cauchy sequence in L2 , with
1/2
the metric defined as k X kL2 , E [ X 2 ] . By the completeness of L2 , there
( n ) L2
exists a random variable Xt 2 L2 such that Xt ! Xt as n ! •.
(a) In order to show { Xt }t 0 is {Ft }-adapted, we need the almost sure conver-
100
that
(nk ) a.s.
Xt ! Xt as k ! •.
⇢
(nk )
L= w 2 W : lim Xt ( w ) = Xt ( w ) .
k!•
satisfies the usual condition (see Definition 5.4). It follows that for a < 0,
⇢ ⇢
\[ [ (nk ) 1
w 2 W : Xt ( w ) a = w 2 W : Xt (w ) < a + 2 Ft .
j m k>m
j
For a 0, since Lc 2 Ft ,
2 3
⇢ ⇢
[ \[ [ (nk ) 1 5
w 2 W : Xt ( w ) a = L c 4 w 2 W : Xt (w ) < a + 2 Ft .
j m k>m
j
Z Z Z Z
(n) (n)
Xt dP Xs dP = ( Xt Xt ) dP ( Xs Xs ) dP, 8n.
A A A A
(n)
By the L2 convergence of Xt ,
Z ⇣ ⌘1/2
(n) (n) (n)
| Xt Xt | dP E [| Xt Xt |] E [( Xt Xt )2 ] ! 0, n ! •.
A
R (n)
Similarly, A
| Xt Xt | dP ! 0. The martingability of { Xt }t 0 follows since
Z Z Z Z
(n) (n)
Xt dP Xs dP = lim ( Xt Xt ) dP ( Xs Xs ) dP
A A n!• A A
Z Z
(n) (n)
lim | Xt Xt | dP + | Xt Xt | dP = 0.
n!• A A
101
(c) To make { Xt }t 0 right-continuous with left limit exists, apply part 1) in
Theorem 5.3.
(n)
2. Consider a sequence { Xt }t 2 Uc2 for n = 1, 2, . . .. By result in part 1), there exists
(n)
• Note that { Xt Xt }t is a martingale for all n, then by Doob’s inequality,
!
(n) 1 h (n) i
P sup | Xt Xt | > # E ( XT XT )2 ! 0, as n ! •.
t T #2
(n) P (nk )
Since suptT | Xt Xt | ! 0, there exists a subsequence {suptT | Xt
⇢
\ (nk )
W1 = w 2 W : Xt (w ) is continuous
k
⇢
(nk )
W2 = w 2 W : lim sup | Xt Xt | = 0
k!• t T
W = W1 \ W2 =) P (W ) = 1.
Take w 2 W, then for any # > 0, there exists N > 0 such that as long as k > N
and t T,
(nk ) (nk ) #
| Xt (w ) Xt (w )| sup | Xt (w ) Xt (w )| < .
t T 3
(nk )
In other words, Xt (w ) uniformly converges to Xt (w ) for w 2 W.
(nk )
The continuity of Xt (w ) (and therefore uniform continuity) implies there exists
d such that as long as h < d,
(n ) (nk ) #
| Xt+kh (w ) Xt (w )| < , 8t T.
3
102
Hence we conclude the continuity of Xt (w ) for w 2 W:
(n ) (n ) (nk )
| Xt + h ( w ) Xt (w )| | Xt+h (w ) Xt+kh (w )| + | Xt+kh (w ) Xt (w )|
(nk )
+ | Xt (w ) Xt (w )| < #.
10.2. Thursday
Z T
(n)
lim E ( Xt Xt )2 dt = 0. (10.1)
n!• 0
Z T n 1
0
X̃t (w ) dBt (w ) = Â X̃t (w ) · [ Bt + (w )
k k 1
Btk (w )] + X̃tn (w ) · [ BT (w ) Btn (w )]
k =0
Rt
Denote It ( X̃ ) , { 0
X̃s (w ) dBs (w )}w 2W .
Theorem 10.2 Given a process { Xt }t 0 2 L2 and fixing T > 0, there exists an unique
process { Zt }t 0 2 Uc2 such that
h i
lim E ( It ( X (n) ) Zt )2 = 0, 8t 2 [0, T ]. (10.2)
n!•
103
exists { Z̃t }t 0 2 Uc2 such that
h i
lim E ( It ( X̃ (n) ) Z̃t )2 = 0, 8t 2 [0, T ],
n!•
then
✓ ◆
P Zt = Z̃t for all t 2 [0, T ] = 1.
"✓Z Z T ◆2 #
h i T
( j) (k)
E ( IT ( X ( j ) ) IT ( X (k) ))2 = E Xt dBt Xt dBt (10.3a)
0 0
Z T
( j) (k)
=E ( Xt Xt )2 dt (10.3b)
0
Z T
( j) (k)
=E ( Xt Xt + Xt Xt )2 dt (10.3c)
0
Z T
Z T
( j) (k)
2E ( Xt Xt )2 dt + 2E ( Xt Xt )2 dt
0 0
(10.3d)
where (10.3b) is by the linearity and isometry property in Proposition 9.1; (10.3d)
Recall (10.1), i.e., for any # > 0, there exists N such that as long as j, k > N,
Z T
Z T
( j) 2 # (k) #
E ( Xt Xt ) dt < , E ( Xt Xt )2 dt < .
0 4 0 4
h i
Thus E ( IT ( X ( j) ) IT ( X (k) ))2 < # for large j, k.
and { It ( X (k) )}t 0 are martingales, together with the convexity of quadratic
104
means for large j, k,
h i h i
E ( It ( X ( j) ) It ( X (k) ))2 E ( IT ( X ( j) ) IT ( X (k) ))2 < #, 8t < T.
By part 3) in Proposition 9.1, the process { It ( X (n) )}t 0 2 Uc2 for each n. By the Cauchy
property for { It ( X (n) )}t 0, n = 1, 2, . . . and the closedness of Uc2 with respect to the
norm k X k = (E [ XT 2 ])1/2 , there exists a limit { Zt }t 0 2 Uc2 such that
h i
lim E ( IT ( X (n) ) ZT )2 = 0.
n!•
equality, !
1
8#, P sup | Zt Z̃t | > # E [( ZT Z̃T )2 ].
t T #2
This suggests that in order to show that { Z̃t }t 0 is a version of { Zt }t 0, we can start
where the inequality is by applying (Âin=1 ai )2 n Âin=1 a2i . The first and last term
h i Z T
(n) (n) (n) (n)
E ( IT ( X t ) IT ( X̃t ))2 = E ( ( Xt X̃t ) dBt )2
0
Z T
(n) (n)
=E ( Xt X̃t )2 dt
0
Z T Z T
(n) (n)
2E ( Xt Xt )2 dt + 2E ( X̃t Xt )2 dt
0 0
105
h i
(n) (n)
and thus E ( IT ( Xt ) IT ( X̃t ))2 ! 0 as n ! •. Put things together, we can assert
that !
2
E [( ZT Z̃T ) ] = 0 =) P sup | Zt Z̃t | > # = 0, 8# > 0.
t T
!
•
[ 1
L✓ sup | Zt Z̃t | > ,
n =1 t T n
✓ ◆
P Zt = Z̃t for all t 2 [0, T ] = P (Lc ) = 1.
Definition 10.3 [Ito Integral for Square Integrable Process] For any process { Xt }t 0 2 L2 ,
define the Ito integral
It ( X ) = Zt , t 2 [0, T ], (10.4)
Z T
1⇥ 2 ⇤
Bt dBt = B T .
0 2 T
(n)
Firstly, we need to figure out the simple process { Bt }t 0 that approximating
(n) (n)
the argument in the integral, say { Bt }t 0. Define P(n) = {0, t1 , t2 , . . . , T } the
106
As a consequence,
Z " Z t(n)
# Z t(n)
T j +1 j +1
h i
(n) (n) (n)
E
0
( Bt 2
Bt ) dt = E Â (n)
tj
( Bt Bt ) dt = Â
2
tj
(n)
E ( Bt Bt )2 dt
j i
Z t(n) Z t(n)
j +1 j +1 (n)
=Â (n)
E ( Bt(n) Bt ) 2
dt = Â (n)
(t t j ) dt
j tj j
j tj
(n)
which indicates that as long as we construct { Bt }t 0 such that kP(n) k ! 0,
hR i
T (n)
E 0 ( Bt Bt )2 dt ! 0 as n ! •.
(n)
2. The Ito integral of the simple process { Bt }t
1. 0 is
" ✓ ◆2 #
1 2
Bt(n) ( Bt(n) Bt(n) ) = B (n) B2(n) Bt(n) Bt(n)
j j +1 j 2 t j +1 tj j +1 j
which implies
" ✓ ◆2 #
1
IT ( B ( n ) ) =
2 Â( Bt2(+) n
j 1
B2(n) )
tj
 Bt(n)
j +1
Bt(n)
j
j j
1h 2 i 1h i
= ( BT B02 ) Q ( P(n) ) = B2 Q ( P(n) )
2 2 T
So we conclude that
Z T
1⇥ 2 ⇤
IT ( B ) , Bt dBt = B T .
0 2 T
107
10.2.2. Properties of Ito Integral
RT
Proposition 10.2 For any { Xt }t 0 2 L2 , the Ito integral 0
Xt dBt has the following
properties:
Z T Z T Z T
(aXt + bYt ) dBt = a Xt dBt + b Yt dBt .
0 0 0
"✓Z ◆2 # Z
t t
E Xu dBu Fs = E Xu2 du Fs .
s s
3. { It ( X )}t 0 2 Uc2 .
Proof. The linearity property is trivial to show, and the third property comes from
(n)
Theorem 10.2. It suffices to show the second propperty. Let { Xt }t 0, n = 1, 2, . . . be
the sequence of simple processes approximating { Xt }t 0, and A 2 FS . It follows that
"✓Z ◆2 #
t h i
2
E Xu dBu 1 A = E ( It ( X ) Is ( X )) 1 A
s
⇣ ⌘2
=E It ( X ) It ( X (n) ) + It ( X (n) ) Is ( X (n) ) + Is ( X (n) ) Is ( X ) 1A
⇣ ⌘2 ⇣ ⌘2
(n)
=E It ( X ) It ( X ) 1A + E It ( X (n) ) Is ( X (n) ) 1A
| {z } | {z }
( a) (b)
⇣ ⌘2
+E Is ( X (n) ) Is ( X ) 1A
| {z }
(c)
h⇣ ⌘⇣ ⌘ i
+ 2E It ( X ) It ( X (n) ) It ( X (n) ) Is ( X (n) ) 1 A
| {z }
(d)
h⇣ ⌘⇣ ⌘ i
+ 2E It ( X ) It ( X (n) ) Is ( X (n) ) Is ( X ) 1 A
| {z }
(e)
h⇣ ⌘⇣ ⌘ i
+ 2E It ( X (n) ) Is ( X (n) ) Is ( X (n) ) Is ( X ) 1 A
| {z }
(f)
108
It is easy to show that ( a), (c) vanishes as n ! •:
⇣ ⌘2 ⇣ ⌘2
(n)
E It ( X ) It ( X ) 1A E It ( X ) It ( X (n) ) ! 0.
h⇣ ⌘⇣ ⌘ i h i
E It ( X ) It ( X (n) ) It ( X (n) ) Is ( X (n) ) 1 A E It ( X ) It ( X (n) ) It ( X (n) ) Is ( X (n) )
⇣ h i⌘1/2 ⇣ h i⌘1/2
E ( It ( X ) It ( X (n) ))2 E ( It ( X (n) ) Is ( X (n) ))2 ! 0,
h i
as E ( It ( X ) It ( X (n) ))2 ! 0 as n ! •.
h⇣R ⌘ i
t 2
It remains to show that the term (b) in fact converges to E X du
s u
1 A as
n ! •:
⇣ "✓Z ◆2 # ✓Z ◆
⌘2 t t
(n) (n)
E It ( X (n)
) Is ( X (n)
) 1A = E Xu dBu 1A = E ( Xu )2 du 1 A ,
s s
and
✓Z t
◆ ✓Z t
◆
(n)
E ( Xu )2 du 1A E ( Xu ) du 1 A 2
s s
✓Z t
◆
(n) (n)
= E ( Xu + Xu )( Xu Xu ) du 1 A
s
✓Z t
◆
(n) (n)
E | Xu + Xu || Xu Xu | du
s
Z t 1/2 Z t 1/2
(n) 2 (n) 2
E ( Xu + Xu ) du E ( Xu Xu ) du
s s
Z t
Z t
(n) (n)
E ( Xu + Xu )2 du = E ( Xu Xu + 2Xu )2 du
s s
Z t
Z t
(n) 2
2E ( Xu Xu ) du + 2E (2Xu )2 du < •.
s s
Together with the fact that the second term vanishes as n ! •, we assert that
✓Z t
◆ ✓Z t
◆
(n)
E ( Xu )2 du 1A E 2
( Xu ) du 1 A ! 0.
s s
109
Or equivalently,
⇣ ⌘2 ✓Z t
◆ ✓Z t
◆
(n)
lim E It ( X (n)
) Is ( X (n)
) 1 A = lim E ( Xu )2 du 1A = E 2
( Xu ) du 1 A .
n!• n!• s s
110
Chapter 11
Week11
11.1. Tuesday
Z t
It ( X ) = Xu dBu .
0
directly from definition is sophisticated. The quadratic variation formula of Ito integral
plays a central role for simplifying the calculation of Ito integral. In this lecture, we
Definition 11.1 [Qudaratic Variation] Suppose that X is a function of time t, define its
qudaratic variation over the interval [0, t] as the limit (when it exists)a
n
Q(P(n) ; X )[0, t] = Â [ Xt(n) Xt (n) ], (11.1)
i i 1
i =1
(n) (n)
with kP(n) k = max j (t j (t j 1) ! 0. ⌅
a To simplify the notation, sometimes we write X [t] for Xt , and X [t](w ) for Xt (w ).
111
We first discuss how to compute the quadratic variation for { Xt }t 0 2 Uc2 , by
applying the Doob-Meyer Decomposition trick:
a non-negative sub-martinalge. The process X·2 admits the unique Doob-Meyer decompo-
sition:
Xt2 = Mt + At , 0 t < •,
h X i , A, h X i0 = 0.
⌅ Example 11.1 Consider the Brownian motion { Bt }t 0 2 Uc2 , then the Doob-Meyer
Decomposition is
Bt2 = ( Bt2 t) + t,
quadratic variation:
with kP(n) k ! 0. Then for any t > 0, then the quadratic variation
process of X.
We first consider the bounded process X• and h X i• , and then extend the results into
112
Proof. 1. We first consider the bounded process X· on [0, t], i.e., | Xs | K almost
surely for any s 2 [0, t], and h X is K almost surely for all s 2 [0, t]. In order
to show the desired result, we will show the stronger result in the sense that
h i2
E Q(P(n) ; X )[0, t] h X it ! 0.
" #2
h i2 ⇣ ⌘2
(n) (n)
E Q(P(n) ; X )[0, t] h X it =E Â X [ t j +1 ^ t ] X [t j ^ t] h X it
j
" #2
⇣ ⌘2 ⇣ ⌘
(n) (n) (n) (n)
=E Â X [ t j +1 ^ t] X [t j ^ t] Â h X i[t j+1 ^ t] h X i[t j ^ t]
j j
" ⇢⇣ #2
⌘2 ⇣ ⌘
(n) (n) (n) (n)
=E Â X [ t j +1 ^ t] X [t j ^ t] h X i[t j+1 ^ t] h X i[t j ^ t]
j
⇢⇣ ⌘2 ⇣ ⌘ 2
(n) (n) (n) (n)
=ÂE X [ t j +1 ^ t] X [t j ^ t] h X i[t j+1 ^ t] h X i[t j ^ t]
j
⇢⇣ ⌘2 ⇣ ⌘
(n) (n) (n) (n)
+2ÂE X [ t j +1 ^ t ] X [t j ^ t] h X i[t j+1 ^ t] h X i[t j ^ t]
j<k
⇢⇣ ⌘2 ⇣ ⌘
(n) (n) (n) (n)
⇥ X [ t k +1 ^ t ] X [tk ^ t] h X i[tk+1 ^ t] h X i[tk ^ t]
(11.2)
Now we show that the second part actually vanishes. For any 0 s < t u < v,
2
E ( Xt Xs ) (h X it h X is ) ( Xv Xu )2 (h X iv h X iu )
⇢ ✓ ◆
2 2
=E E ( Xt Xs ) (h X it h X is ) ( Xv Xu ) (h X iv h X iu ) Ft
⇢ ✓ ◆
=E ( Xt Xs )2 (h X it h X is ) E ( Xv Xu )2 (h X iv h X iu ) Ft
(11.3)
Moreover,
⇥ ⇤ ⇥ ⇤
E ( Xv Xu )2 | Ft = E Xv2 + Xu2 2Xv Xu | Ft
⇥ ⇤
= E Xv2 + Xu2 2E [ Xv Xu | Fu ] | Ft
⇥ ⇤ ⇥ ⇤
= E Xv2 + Xu2 2Xu2 | Ft = E Xv2 Xu2 | Ft
= E [h X iv h X iu | Ft ] ,
113
term in (11.3) vanishes, and therefore the second part in (11.2) vanishes.
By the elementary inequality ( a b)2 2a2 + 2b2 , the first part in (11.2) can be
upper bounded as
⇢⇣ ⌘2 ⇣ ⌘ 2
(n) (n) (n) (n)
ÂE X [ t j +1 ^ t ] X [t j ^ t] h X i[t j+1 ^ t] h X i[t j ^ t]
j
⇣ ⌘4 ⇣ ⌘2
(n) (n) (n) (n)
2 Â E X [ t j +1 ^ t ] X [t j ^ t] + 2 Â E h X i[t j+1 ^ t] h X i[t j ^ t]
j j
(11.4)
We claim that:
⇣ ⌘4
(n) (n)
(a) Â j E X [t j+1 ^ t] X [t j ^ t] ! 0 as n ! •;
⇣ ⌘2
(n) (n)
(b) Â j E h X i[t j+1 ^ t] h X i[t j ^ t] ! 0 as n ! •.
h i2
L2
Then E Q(P(n) ; X )[0, t] h X it ! 0 as n ! •, i.e., Q(P(n) ; X )[0, t] ! h X it . The
desired result holds since L2 convergence implies convergence in probability.
2. Then consider the case where { Xt }t 0, hXi are unbounded. We argue for this
Tk = inf{t 0 : | Xt | K or h X it K }.
n o
(k)
( X ( k ) )2 = ( Xt )2 h X (k) it + h X (k) it .
114
Hence X (k) ⌘ X [t ^ Tk ] admits the Doob-Meyer decomposition
✓ ◆ ✓ ◆
(n) (k) (k) (n) (k) h
P | Q(P ;X )[0, t] hX i[t]| > # = P | Q(P ;X )[0, t] h X i[t ^ Tk ]| > # < .
2
(11.5)
P
Now we begin to show that Q(P(n) ; X )[0, t] ! h X it :
✓ ◆
P | Q(P(n) ; X )[0, t] h X i[t]| > #
✓ ◆ ✓ ◆
=P | Q(P(n) ; X )[0, t] h X i[t]| > #, Tk (n)
t + P | Q(P ; X )[0, t] h X i[t]| > #, Tk < t
✓ ◆
P | Q(P(n) ; X )[0, t] h X i[t]| > #, Tk t + P ( Tk < t) (11.6a)
✓ ⇣ ⌘2 ◆
(n) (n)
=P Â X [ t j +1 ^ t] X [t j ^ t] h X i[t ^ Tk ] > #, Tk t + P ( Tk < t)
j
✓ ⇣ ⌘2 ◆
(n) (n)
=P Â X [ t j +1 ^ Tk ^ t] X [t j ^ Tk ^ t] h X i[t ^ Tk ] > #, Tk t + P ( Tk < t)
j
✓ ⇣ ⌘2 ◆
(n) (n)
P Â X [ t j +1 ^ Tk ^ t] X [t j ^ Tk ^ t] h X i[t ^ Tk ] > # + P ( Tk < t)
j
✓ ◆
=P Q(P(n) ; X (k) )[0, t] h X i[t ^ Tk ] > # + P ( Tk < t)
h h
+ = h. (11.6b)
2 2
where the upper bound presented (11.6a) is because P ( Tk < •) ! 0, and (11.6b)
In order to complete the proof, it remains to show the following two claims are
correct.
115
Proposition 11.1 Let { Xt }t 0 2 Uc2 with | Xs | K almost surely for any s 2 [0, t]. Let
{ P(n) }•
n=1 be a sequence of partitions on [0, t ] with k P
(n) k ! 0. Then
" #
(n) (n)
lim E
n!•
 ( X [ t j +1 ] X [t j ])4 = 0.
j
Proof. 1. We first show that E [ Q(P; X )[0, t]]2 is bounded for any partition P:
2 2
E Q(P; X )[0, t] =E Â ( X [ t j +1 ] X [t j ])2
j
" #
=E Â ( X [ t j +1 ] X [t j ])4 + 2E Â Â ( X [ t j +1 ] X [t j ])2 ( X [tk+1 ] X [tk ])2
j k j>k
(11.7)
In particular,
" # " #
E Â ( X [ t j +1 ] X [t j ])2 Ftk+1 = E Â ( X [ t j +1 ] 2 + X [ t j ] 2 2X [t j+1 ] X [t j ]) Ftk+1
j>k j>k
" #
⇣ ⌘
=E Â X [ t j +1 ] 2
+ X [t j ] 2
2E X [t j+1 ] X [t j ] Ft j F t k +1
j>k
" #
⇥ ⇤
=E Â X [ t j +1 ] 2 X [ t j ] 2 F t k +1 E X [ t k + 2 ] F t k +1 K 2 .
j>k
h i
Taking k = 0, we have E Â j 1 ( X [ t j +1 ] X [t j ])2 K2 . We can apply these in-
" #
E Â Â ( X [ t j +1 ] X [t j ])2 ( X [tk+1 ] X [tk ])2
k j>k
" ( )#
=E Â E Â ( X [ t j +1 ] 2
X [t j ]) ( X [tk+1 ] 2
X [tk ]) Ftk+1
k j>k
" ( )#
=E Â ( X [ t k +1 ] X [tk ])2 E Â ( X [ t j +1 ] X [t j ])2 Ftk+1
k j>k
" #
E Â ( X [ t k +1 ] X [tk ])2 K2 K4 .
k
116
Then we upper bound the first term in (11.7):
E Â ( X [ t j +1 ] X [t j ])4
j
=E Â ( X [ t j +1 ] X [t j ])2 ( X [t j+1 ] X [t j ])2
j
E Â 4K2 (X [t j+1 ] X [t j ])2 4K4 ,
j
h i
where the first inequality is because E Â j 1 ( X [ t j +1 ] X [t j ])2 K2 . Then we
2
can assert that E [ Q(P; X )[0, t]] 6K4 .
2. In order to show the desired result, we start with simplifying  j ( X [t j+1 ] X [t j ])4 :
Cauchy-Schwarz inequality,
E Â ( X [ t j +1 ] X [t j ])4
j
" #2
E A2P Â ( X [ t j +1 ] X [t j ]) 2
j
8 " #2 91/2
n o1/2 < =
: Â
EA4P E ( X [ t j +1 ] X [t j ])2
j
;
n o1/2 n o1/2
2
= EA4P E [ Q(P; X )[0, t]]
n o1/2 n o1/2
EA4P 6K4
117
⌅
Proposition 11.2 Let { Xt }t 0 2 Uc2 and h X is K almost surely for any s 2 [0, t]. Sup-
pose that {P(n) } is a sequence of partitions on [0, t] such that kP(n) k ! 0, then we
have
⇣ ⌘2
(n) (n)
lim E
n!•
 h X i[t j+1 ] h X i[t j ] = 0.
j
⇣ ⌘2
(n) (n)
Proof. The proof is by direct computing and upper bounding the term E Â j h X i[t j+1 ] h X i[t j ] :
⇣ ⌘
(n) (n) 2
E Â h X i[t j+1 ] h X i[t j ]
j
⇣ ⌘
(n) (n) (n) (n)
E sup h X i[t j+1 ] h X i[t j ] · Â h X i[t j+1 ] h X i[t j ]
j j
(n) (n)
=E sup h X i[t j+1 ] h X i[t j ] · h X it E [ ÂP · h X it ],
j
continous, and therefore uniformly continuous on [0, t], when kPk ! 0, ÂP ! 0. It is
⇣ ⌘2
(n) (n)
E Â h X i[t j+1 ] h X i[t j ] E [ ÂP · h X it ] ! 0.
j
X• 2 L2 . However, we don’t prove this result by taking the limit of the formula
presented in Theorem 11.1 (because it is a little bit complicated), but making use of the
Doob-Meyer decomposition.
118
Proof. The Doob-Meyer decomposition for { It ( X )}t 0 2 Uc2 is
= E [ It2 ( X ) | Fs ] Is2 ( X ).
Z t
2
E [( It ( X ) Is ( X )) | Fs ] = E Xu2 du | Fs
s
Z t Z s
=E Xu2 du | Fs Xu2 du.
0 0
Z t Z s
E [ It2 ( X ) Xu2 du Fs = Is2 ( X ) Xu2 du.
0 0
Rt Rt
Hence, { It2 ( X ) 0
Xu2 du}t 0 is a martingale. Moreover, 0
Xu2 du is a predictable
Z t
h I ( X )it = Xu2 du.
0
119
11.2. Thursday
single process. Now we aim to quantify the sum of product of jumps for two processes
{ Xt }t 0 , {Yt }t 0 2 Uc2 :
as
⇣ ⌘⇣ ⌘
(n) (n) (n) (n)
Qc (P; X, Y )[0, t] = Â X [t j+1 ^ t] X [t j ^ t] Y [ t j +1 ^ t ] Y [t j ^ t]
j
1
h X, Y i = (h X + Y i hX Y i),
4
Similar as in Theorem 11.1, we can show that the quadratic covariational process is
1
Qc (P; X, Y )[0, t] = ( Q(P; X + Y )[0, t] + Q(P; X Y )[0, t]) , (11.8)
4
and the remainings can be shown by taking the limit of the quadratic variation
Q(P(n) ; X + Y )[0, t] and Q(P(n) ; X Y )[0, t]. Here we begin to simplify the RHS in
120
(11.8):
2
Q(P; X + Y )[0, t] = Â X [t j+1 ^ t] + Y [t j+1 ^ t] X [t j ^ t] Y [ t j +1 ^ t ]
j
2 2
= Â X [ t j +1 ^ t ] X [t j ^ t] + Â Y [ t j +1 ^ t ] Y [ t j +1 ^ t ]
j j
+ 2 Â X [ t j +1 ^ t ] X [t j ^ t] Y [ t j +1 ^ t ] Y [ t j +1 ^ t ] .
j
Similarly,
2 2
Q(P; X Y )[0, t] = Â X [t j+1 ^ t] X [t j ^ t] + Â Y [ t j +1 ^ t ] Y [ t j +1 ^ t ]
j j
2 Â X [ t j +1 ^ t ] X [t j ^ t] Y [ t j +1 ^ t ] Y [ t j +1 ^ t ] .
j
1. Symmetry: h X, Y i = hY, X i;
3. Cauchy-Schwarz inequality: |h X, Y it |2 h X it hY it .
Proof. Part 1) and part 2) can be shown by definition. For part 3), we first apply
!2
⇣ ⌘⇣ ⌘
(n) (n) (n) (n)
( Qc (P; X, Y )[0, t])2 = Â X [ t j +1 ^ t ] X [t j ^ t] Y [ t j +1 ^ t ] Y [t j ^ t]
j
! !
⇣ ⌘2 ⇣ ⌘2
(n) (n) (n) (n)
 X [ t j +1 ^ t] X [t j ^ t]  Y [ t j +1 ^ t] Y [t j ^ t]
j j
Then taking the limit on both sides gives the desired result. ⌅
R Recall that a monotone function has the bounded variation over any finite
121
intervals. Then h X i has the bounded variation since it is increasing. It follows
that the quadratic covariational process h X, Y i has the bounded variation
since it is a linear combination of two quadratic variational processes.
Next, we show that the product process XY admits the decomposition that is
ing:
1
[( X + Y )2 h X + Y i] [( X Y )2 h X Y i]
4
1
= [ X 2 + 2XY + Y2 X 2 + 2XY Y2 ] [h X + Y i hX Y i]
4
1
= XY [h X + Y i h X Y i].
4
with the corresponding partition P = {0 = t0 < t1 < · · · < tn T }. Then the Ito integral
of Xt w.r.t. Mt is defined as
Z T n 1
0
Xt dMt = Â X [ t j ] · ( M [ t j +1 ] M[t j ]) + X [tn ]( M [ T ] N [tn ]),
j =1
1. { It ( X )}t 0 2 Uc2 ;
122
2. The quadratic variational process of I• ( X ) is
Z t
h I ( X )it = Xu2 dh Miu .
0
n Rt o
The process It2 ( X ) X 2 dh M i u
0 u
is a martingale;
t 0
3. Ito isometry: for any t 0 we have
Z t 2 Z t
E Xu dMu =E Xu2 dh Miu .
0 0
We need some conditions, such as integrability results about X, to define the Ito
RT
integral 0 Xt dMt for general stochastic processes X• .
Definition 11.4 [Ito Integrable Space] The set L2 ( M) denotes the space containing
all adapted stochastic processes { Xt }t 0, such that, there exists a sequence of simple
(n)
processes { Xt }t 0 satisfying
Z T
(n)
lim E ( Xt Xt )2 dh Mit = 0, 8 T > 0.
n!• 0
a sequence of finite stopping times {tn } with tn " • so that X tn ⌘ { X (tn ^ t)}t 0 is a
123
Definition 11.6 [Local Bounded] A process X• is said to be locally bounded if there is
a sequence of finite stopping times {sn } with sn " • so that X sn is bounded for each
n 1. ⌅
The answer to the question above is yes. We take t̃n = tn ^ sn , then t̃n " •, and Mt̃n 2 Uc2 .
(n) (n)
Construct Xt = Xt 1{t t̃n }, then X• 2 L2 ( Mt̃n ). Then we define the Ito’s integral
Z T Z T
(n)
Xt dMt = Xt dMt̃n , T 2 [0, t̃n ].
0 0
Note that this definition is consistent, i.e., does not depend on the particular choice of
decomposition
X t = A t + Mt ,
•
| A|(t) ⌘ sup sup  1(tn t)| A[tn ] A[tn 1 ]| < •, 8t 0.
d>0,t0 =0 tn tn 1 d n =1
Z T Z T Z T
Xt dYt = Xt dMt + Xt dAt .
0 0 0
Note tht a left-continuous adapted process is locally bounded. Hence the first term
RT
0
Xt dMt is the Ito integral w.r.t. a continuous local martingale. The second term is
124
Chapter 12
Week12
12.1. Tuesday
Now we show how to compute the Ito integral for Xt2 based on this identity:
⌅ Example 12.1 Suppose that { Xt }t 0 2 Uc2 , and P = {t0 < t1 < · · · < t} is a partition
on the interval [0, t]. Then take the summation on 12.1 both sides yields
Z t
Xt2 X02 = 2 Xs dXs + h X it .
0
125
Theorem 12.1 Let { Bt }t 0 be a Brownian motion and f 2 C 2 (R ). Then
Z t Z t
1
f ( Bt ) = f (0) + f 0 ( Bu ) dBu + f 00 ( Bu ) du,
0 2 0
Proof. Let P = {t0 < t1 < · · · < tn = t} be a partition on [0, t], then f ( Bt ) admits the
expansion
n 1
f ( Bt ) = f (0) + Â [ f ( B[t j+1 ]) f ( B[t j ])]. (12.2)
j =0
By Taylor expansion on the RHS, there exists qt j (w ) 2 [ B[t j ](w ), B[t j+1 ](w )] such that
1 00
f ( B[t j+1 ]) f ( B[t j ]) = f 0 ( B[t j ])( B[t j+1 ] B[t j ]) + f ( q t j ) · ( B [ t j +1 ] B[t j ])2 . (12.3)
2
1
f (0) = Â f 0 ( B[t j ])( B[t j+1 ]
2Â
f ( Bt ) B[t j ]) + f 00 (qt j ) · ( B[t j+1 ] B[t j ])2 .
j j
Rt
As kPk ! 0, the term  j f 0 ( B[t j ])( B[t j+1 ] B[t j ]) ! 0
f 0 ( Bu ) dBu in probability. Then
" #2
E Â f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 Â f 00 ( B[t j ]) · (t j+1 tj )
j j
" #2
=E Â f 00 ( B[t j ]) · ( B [ t j +1 ] B[t j ])2 ( t j +1 tj )
j
⇥ ⇤2
= Â E f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 ( t j +1 tj )
j
+ 2 Â E f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 ( t j +1 tj )
j<k
⇥ f 00 ( B[tk ]) · ( B[tk+1 ] B[tk ])2 ( t k +1 tk )
126
The second term vanishes because for any j < k,
E f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 ( t j +1 tj )
⇥ f 00 ( B[tk ]) · ( B [ t k +1 ] B[tk ])2 ( t k +1 tk )
=EE f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 ( t j +1 t j ) Ft j
⇥ f 00 ( B[tk ]) · ( B [ t k +1 ] B[tk ])2 ( t k +1 tk )
00
=E f ( B[t j ])E ( B[t j+1 ] B[t j ])2 ( t j +1 t j ) Ft j
⇥ f 00 ( B[tk ]) · ( B [ t k +1 ] B[tk ])2 ( t k +1 tk ) = 0,
where the first equality is by tower property and the last equality is because
⇥ ⇤2
E f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 (t j+1 t j )
h ⇣ ⌘i2
=E f 00 ( B[t j ]) · E ( B[t j+1 ] B[t j ])2 (t j+1 t j ) Ft j
⇥ ⇤2 ⇥ ⇤2
=E f 00 ( B[t j ]) · 2(t j+1 t j )2 = 2 ( t j +1 t j )2 E f 00 ( B[t j ])
It follows that
⇥ ⇤2
ÂE f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 ( t j +1 tj )
j
⇥ ⇤2
= 2 Â ( t j +1 t j )2 E f 00 ( B[t j ])
j
⇥ ⇤2 ⇥ ⇤2
2 k P k  ( t j +1 t j )E f 00 ( B[t j ]) = 2tkPkE f 00 ( B[t j ]) .
j
⇥ ⇤2
ÂE f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 ( t j +1 tj ) 2tkPkK2 ! 0,
j
L2
which means that  j f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 !  j f 00 ( B[t j ]) · (t j+1 tj )
127
and therefore in probability. Also, by the Lebesgue integration knowl-
Z t
2 P
Âf 00
( B[t j ]) · ( B[t j+1 ] B[t j ]) !
0
f 00 ( Bu ) du.
j
(b) Now consider the case where f 00 is unbounded, then for any K, we wish to
show that
⇥ ⇤ P
 f 00 ( B[t j ]) · ( B [ t j +1 ] B[t j ])2 ( t j +1 t j ) ! 0.
j
⇥ ⇤
f 00 ( B[t j ]) · ( B[t j+1 ] B[t j ])2 ( t j +1 tj )
⇥ ⇤
= f 00 ( B[t j ])1{| f 00 ( B[t j ])| K } · ( B[t j+1 ] B[t j ])2 (t j+1 t j )
⇥ ⇤
+ f 00 ( B[t j ])1{| f 00 ( B[t j ])| > K } · ( B[t j+1 ] B[t j ])2 (t j+1 t j )
⇥ ⇤
 f 00 ( B[t j ])1{| f 00 ( B[t j ])| K} · ( B [ t j +1 ] B[t j ])2 ( t j +1 t j ) ! 0.
j
⇥ ⇤
 f 00 ( B[t j ])1{| f 00 ( B[t j ])| > K} · ( B [ t j +1 ] B[t j ])2 ( t j +1 tj )
j
max f 00 ( B[t j ])1{| f 00 ( B[t j ])| > K } · Â ( B[t j+1 ] B[t j ])2 ( t j +1 tj )
j j
" #
0 u t
00 00
max f ( B[t j ])1{| f ( B[t j ])| > K } · Â ( B [ t j +1 ] 2
B[t j ]) + t
j
Applying Theorem 8.1 gives  j ( B[t j+1 ] B[t j ])2 ! t in probability. By the
128
continuity of f 00 and Bt on the interval [0, t],
Hence,
⇥ ⇤ P
 f 00 ( B[t j ])1{| f 00 ( B[t j ])| K} ( B [ t j +1 ] B[t j ])2 ( t j +1 t j ) ! 0, as kPk ! 0.
j
As a result,
Z t
2 P P
Âf 00
( B[t j ]) · ( B[t j+1 ] B[t j ]) ! Â f ( B[t j ]) · (t j+1
00
tj ) !
0
f 00 ( Bu ) du.
j j
• Now consider the case where qt j 6= Bt j . It remains to show that  j f 00 (q [t j ])( B[t j+1 ]
Rt
B[t j ])2 converges to 0 f 00 ( Bu ) du in probability:
 f 00 (q [t j ])( B[t j+1 ] B[t j ])2  f 00 ( B[t j ])( B[t j+1 ] B[t j ])2
j j
P
By Theorem 8.1 we can see that  j ( B[t j+1 ] B[t j ])2 ! t. For almost all w 2 W, by
max f 00 (q [t j ]) f 00 ( B[t j ]) ! 0.
j
Hence,
Z t
P P
 f 00 (q [t j ])( B[t j+1 ] B[t j ])2 !  f 00 ( B[t j ])( B[t j+1 ] B[t j ])2 !
0
f 00 ( Bu ) du.
j j
129
Definition 12.1 [Ito Processes] An Ito process is a stochastic process X• on (W, F , F, P )
of the form
Z t Z t
X t = X0 + µ[u] du + s[u] dB[u], 0 t T, (12.4)
0 0
Z t
P |µu | du < •, for all t 0 = 1;
0
Z t
P su2 du < •, for all t 0 = 1.
0
dXt = µt dt + st dBt .
Theorem 12.2 — Ito Formula for 1-dimension. Let { Xt }t 0 be an Ito process with
stochastic differential
dXt = µt dt + st dBt .
Let f 2 C 2 (R ), then f ( Xt ) is again an Ito process, with the following equality holds
almost surely:
Z t Z t
1
f ( X t ) = f ( X0 ) + f 0 ( Xu ) dXu + f 00 ( Xu )( dXu )2 .
0 2 0
130
So we further have the representation
Z t Z t Z t
1
f ( X t ) = f ( X0 ) + f 0 ( Xu )µu du + f 00 ( Xu )su2 du + f 0 ( Xu )su dBu , 8t 0
0 2 0 0
1 00
d f ( Xt ) = f 0 ( Xt ) µ t + f ( Xt )st2 dt + f 0 ( Xt )st dBt .
2
The proof follows the similar idea as stated in Theorem 12.1. Here we only provide
Outline of Proof. Let P = {t0 < t1 < · · · < tn = t} be a partition on [0, t], then f ( Xt )
f ( Xt ) = f ( X0 ) + Â f ( X [t j+1 ]) f ( X [t j ])
j
(12.5)
1
= f ( X0 ) + Â f ( X [t j ])( X [t j+1 ]
0
X [t j ]) + f 00 (q [t j ])( X [t j+1 ] X [t j ]) 2
j
2
(a) First consider the case where {µt } and {st } are simple processes, then
P
Assume that q [t j ] = X [t j ], then we first show  j f 00 ( X [t j ])( X [t j+1 ] X [t j ])2 !
Z t
1 a.s.
Âf 00 2
( X [t j ])s[t j ] (t j+1 tj ) !
2 0
f 00 ( Xu )su2 du.
j
131
For the case where q [t j ] 6= X [t j ], we will show that
a.s.
 f 00 (q [t j ])(X [t j+1 ] X [t j ])2  f 00 (X [t j ])(X [t j+1 ] X [t j ])2 ! 0.
j j
(b) For general processes {µt }t 0 and {st }t 0, we will use the approximation of
simple processes.
Xt = µt + Bt , µ 2 R.
Then we apply the Ito’s formula to compute the stochastic differential form of X 2 :
d f ( Xt ) = dXt2
(1) (d)
Theorem 12.3 — Ito’s Formula. Let { Xt }t 0 , . . . , { Xt }t 0 be continuous semi-
132
Recall that in discrete case, any real-valued process is a semi-martingale, while it is
not true for continuous case. We define the semi-martingale for continuous case as the
following:
decomposition
X t = A t + Mt ,
•
| A|(t) ⌘ sup sup  1(tn t)| A[tn ] A[tn 1 ]| < •, 8t 0.
d>0,t0 =0 tn tn 1 d n =1
( j) ( j) ( j)
R Suppose that X ( j) admits the decomposition Xt = At + Mt , then
d Z t d Z t
∂f (1) (d) ( j) ∂f (1) (d) ( j)
 ∂X j
( Xu , . . . , Xu ) dXu = Â
∂X j
( Xu , . . . , Xu ) dMu
j =1 0 j =1 0
d Z t
∂f (1) (d) ( j)
+Â ( Xu , . . . , Xu ) dAu
j =1 0
∂X j
(1) (d) f f
f ( X t , . . . , X t ) = A t + Mt ,
where
d Z t
f (1) (d) ∂f (1) (d) ( j)
Mt = f ( X0 , . . . , X0 ) + Â ∂X j
( Xu , . . . , Xu ) dMu ,
j =1 0
d Z t
f ∂f (1) (d) ( j)
At = Â ∂X j
( Xu , . . . , Xu ) dAu
j =1 0
Z t
1 d ∂2 f (1) (d)
+ Â ( Xu , . . . , Xu ) dh M ( j ) , M ( k ) i u
2 j,k=1 0 ∂X j ∂Xk
133
⌅ Example 12.3 We can also apply Theorem 12.3 to obtain an “integration by parts”
formula. Suppose that X• and Y• are semi-martingales, then by direct computation with
f ( Xt , Yt ) = Xt Yt ,
Z t Z t
Xt Yt X0 Y0 = Xs dYs + Ys dXs + h X, Y it
0 0
X t = Mt + A t , Yt = Nt + Wt ,
Z t Z t Z t Z t
Xt Yt = X0 Y0 + Xs dNs + Ys dMs + Xs dWs + Ys dAs + h M, N it
0 0 0 0
⌅ Example 12.4 We aim to sovle the following stochastic differential equation with
This equation is called the stochastic exponential of X, which can be re-written as the
integral form:
Z t
Zt = 1 + Zu dXu , (12.7)
0
where the integration refers to the Ito’s integral. We guess the solution should be Zt =
134
(z = X + V ) gives
Z t Z t
1
Zt = 1 + Zu d( Xu + Vu ) + Zu dh X + V iu
0 2 0
Z t Z t Z t
1
=1+ Zu dXu + Zu dVu + Zu dh M + V iu
0 0 2 0
1 1
In order to satisfy (12.7), we take Vt = 2 h M it . As a result, exp Xt 2 h M it is the
⌅ Example 12.5 [Levy’s characterization of Brownian Motion] Levy states that the
quadratic variational process of a continuous local martingale will characterize the Brownian
motion uniquely:
Theorem 12.4 — Levy’s Theorem. Consider the filtered probability space (W, F , F, P )
⇣ ⌘
(z ) z2
Proof. In order to show the sufficiency, we construct Zt = exp z Mt 2t . By
(z )
Ito’s formula on Zt and f ( x ) = e x , for 0 s < t, we have
(z ) (z )
dZt = zZt dMt .
(z )
It indicates that Zt is a martingale, by the uniqueness of semi-martingale decom-
(z ) (z )
position. Re-arranging the term E [ Zt | Fs ] = Zs yields
✓ ◆
z2
E [exp(z ( Mt Ms )) | Fs ] = exp (t s) ,
2
is a Brownian motion. ⌅
135
⌅
tt = inf{u : h Miu > t}. Then for 8t 0, tt is a stopping time, and Bt ⌘ Mtt is a
Proof Outline. We call the increasing sequence of stopping time {tt }t 0 the time-change.
Since h Mi• = • almost surely, we can assert that each tt is finite almost surely. By the
{ Mu^tt }u 0 gives
which implies that Bt is a Ftt -local martingale. Applying the optional samping theorem
This implies that { Bt2 t} is a Ftt -local martingale. By Levy’s characterization, together
Z Z
X (w ) dQ(w ) = X (w )z (w ) dP (w ).
W W
Or equivalently, E X ⇠Q [ X ] = E X ⇠P [zX ].
136
dQ
to be Q( A) = E [ ZT 1 A ], A 2 F T . Then we have dP F = ZT . Because { Zt }t 0 is a
T
dQ
= Zt .
dP Ft
martingale on (W, F , F, Q ).
Proof Outline. By the technique of localization, we assume that M, Z, Z1 are all bounded.
It suffices to show that X is a martingale w.r.t. the probability measure Q, i.e., for any
0 u < t T, EQ [1 A ( Xt Xu )] = 0, 8 A 2 Fu . By definition,
E Q [1 A ( Xt Xu )] = EP [1 A ( Zt Xt Zu Xu )].
Z t Z t
Zt Xt = Z0 X0 + Zs dXs + Xs dZs + h X, Z it
0 0
Z t ✓ ◆ Z t
1
= Z0 X0 + Zs dMs dh M, Z is + Xs dZs + h X, Z it
0 Zs 0
Z t Z t
= Z0 X0 + Zs dMs + Xs dZs .
0 0
martingale. ⌅
137
Rt
X0 + f dBs
0 s
is always a martingale w.r.t. Ft .1 The martingale representation theorem
states that the converse is also true: any Ft -martingale can be represented as an Ito’s
integral.
process { f t }t 0 2 L2 so that
Z t
Mt = E [ M0 ] + f u dBu .
0
Z T
z = E [z ] + f t dBt .
0
Proof for Theorem 12.7. Assume the claim is true, then for the martingale M• with
Z T Z T
MT = E [ MT ] + f t dBt = E [ M0 ] + f t dBt .
0 0
Z t
Mt = E [ MT | Ft ] = E [ M0 ] + f u dBu .
0
1 This
can be shown either by directly applying definition or by the uniqueness of semi-martingale
decomposition.
138
Proof on the Claim. Let T > 0 and define the stochastic exponential martingale for any
f 2 L2 ([0, T ]):
⇢Z t Z t
1
E [ f ]t = exp f (u) dBu f 2 (u) du , t 2 [0, T ].
0 2 0
Then we can show that ` ⌘ span{E [ f ]t : f 2 L2 ([0, T ])} is dense in the space L2 (W, FT , P ).
It remains to verify this claim when z = E [ f ]t for some f 2 L2 ([0, T ]). Directly applying
Z T
dE [ f ]t = f (t)E [ f ]t dBt =) E [ f ]t = 1 + f (t)E [ f ]t dBt .
0
139
12.2. Thursday
We will talk about how to solve some simple SDEs. We will also study some important
SDEs in applications that cannot find explicit solutions, and establish the theorem of
dSt
= r dt.
dt
dSt
This simple ODE admits a deterministic solution St . The corresponding SDE is dt =
r dt + s dBt , called Black-Sholes equation, in which the increasing rate is a constant
In this lecture, we will conside the SDE of the Markovian type, i.e., the parameter
The differential for SDE has no meaning, while only the integration has the meaning.
Therefore, the above equation refers to the follwoing equation in integral form:
Z t Z t
Xt X0 = µ(u, Xu ) du + s(u, Xu ) dBu . (12.9)
0 0
140
Definition 12.3 [Solution to SDE] Given the functions µ, s defined above, the solution
Definition 12.4 [Pathwise Uniqueness] The solution to (12.8) has pathwise uniqueness
if given the initial value X0 and the Brownian motion { Bt }t 0 on a filtered probability
equation. In other words, if two solutions ( X, B) and ( X 0 , B0 ) have the same initial value,
i.e., X0 = X00 and B0 = B00 , then Xt = Xt0 almost surely for any t 0. ⌅
Definition 12.5 [Uniqueness in Law] The solution to (12.8) has uniqueness in law if two
solutions X, X 0 with the same initial distribution are equivalent, i.e., X and X 0 have the
motion B• with completion. A solution that is not a strong solution is called a weak
solution. ⌅
R If a strong solution exists, then the SDE always has a solution for any given
probability space and Brownian motion. In other words, when a strong
solution exists, any initial value and Brownian motion corresponds to at least one
solution. If one has the pathwise uniqueness together with the existence of a
141
strong solution, then any initial value and Brownian motion will correspond to a
unique solution.
The following example shows an important concept in SDE: sometimes the solution
exists but there is no strong solution; sometimes we have no pathwise uniqueness but
Z t
Xt = sign( Xu ) dBu , 0 t < •, (12.10)
0
SDE
motion. For any other solution X 0 satisfying (12.10), notice that X 0 is a continuouis
Rt Rt
local martingale and the quadratic variation h X 0 it = 0 sign( Xu0 )2 du = 0 du = t.
2. A weak solution exists: choose X be any Brownian motion {Wt }t 0 , and define
Rt
B̃t by B̃t = 0 sign(Ws ) dWs , i.e., d B̃t = sign( Xt ) dXt . Then B̃• is also a Brownian
motion. Moreover,
a solution.
142
⌅ Example 12.8 [Ornstein-Uhlenbeck Process] Consider solving the SDE
• Take Yt = eat Xt , where eat can be viewed as the integrating factor. Applying Ito’s
= seat dBt
It follows that
Z t Z t
Yt = Y0 + s eau dBu =) Xt = e at
Yt = e at
+s ea(u t)
dBu .
0 0
where µ, s are constants. We claim that there is a unique strong solution Xt = e(µ 1/2s2 )t+sBt .
143
e(µ 1/2s2 )t+sB gives
✓ ◆
1 2 (µ 1/2s2 )t+sBt 1/2s2 )t+sBt
dXt = µ s e dt + se(µ dBt
2
1 1/2s2 )t+sBt
+ s2 e(µ dt
2
1/2s2 )t+sBt 1/2s2 )t+sBt
= µe(µ dt + se(µ dBt
• We can apply the change of probability measure trick to solve this SDE. Let {Wt }t 0
⇢Z t Z t
1
Zt = exp b(u, Wu ) dWu b2 (u, Wu ) du .
0 2 0
✓ ◆
1
dZt = d exp Nt h N it
2
1 1 Nt 1 1 1
= e Nt 2 h N it dNt e 2 h N it dh N it + e Nt 2 h N it dh N it = Zt dNt .
2 2
Z t
Zt = 1 + Zu dNu .
0
144
• We recover the original solution by Girsanov theorem. Define B̃t = Wt W0
Rt 1
0 Zu
dhW, Z iu . In particular,
Z t Z t Z t
hW, Z iu = h dWu , Zu dNu i = Zu dhW, N iu ,
0 0 0
Rt 1
which implies that 0 Zu
dhW, Z iu = hW, N it . Therefore, B̃t = Wt W0 hW, N it
is a martingale w.r.t. the probability measure Q. Moreover, h B̃it = hW it = t. By
Z t Z t Z t
hW, N it = h dWu , b(u, Wu ) dWu i = b(u, Wu ) du.
0 0 0
Z t Z t
Wt = W0 + b(u, Wu ) du + B̃t = b(u, Wu ) du + B̃t .
0 0
As a result, (W, B̃) is a solution on the probability space (W, F , Q). This solution is
a weak solution.
145
Chapter 13
Week13
13.1. Thursday
the strong solution; otherwise the weak solution is enough, if we only concern the
construct different solutions w.r.t. the same Brownian motion, the strong solution is
needed. The first theorem indicates that the existence of solution, together with the
Theorem 13.1 — Existence and Uniqueness for SDEs. Suppose that the SDE in (12.8)
2. The existence of solution implies the existence of strong solution, i.e., there
147
constant L so that
{ Bt }t 0 , satisfying E [ Z2 ] < •.
Then the SDE with the initial condition X0 = Z admits a strong solution { Xt }t 0
adapted to the filtration {Ft }t 0 and X 2 L2 . Furthermore, the solution has pathwise
uniqueness.
Proof. The uniqueness is by Ito’s isometry and Lipschitz condition. The existence of
Z t Z t 2
∂f 1 ∂ f
f ( Xt ) f ( X0 ) = ( Xu ) dXu + ( Xu ) dh X i u .
0 ∂X 2 0 ∂X 2
Z t Z t Z t
X t = X0 + µ( Xu ) du + s( Xu ) dBu =) h X it = s2 ( Xu ) du.
0 0 0
148
It follows that
Z t Z t
∂f 1 ∂2 f ∂f
f ( Xt ) f ( X0 ) = ( Xu ) µ ( Xu ) + ( Xu )s2 ( Xu ) du + ( Xu )s( Xu ) dBu .
0 ∂X 2 ∂X 2 0 ∂X
1 d2 d
L = s2 ( x ) 2 + µ ( x ) .
2 dx dx
Z t
f
Wt = f ( Xt ) f ( X0 ) (L f )( Xs ) ds
0
is a continuous local martingale, then X• solves the SDE (13.1) on the space (W, F , P ).
Z t Z t
X t = X0 + s ( Xu ) dBu + µ( Xu ) du.
0 0
Rt 1 f
We set f ( x ) = x and Bt = 0 s ( Xu )
dWu . Then we show it is a Brownian motion by
Levy’s characterization. ⌅
149