Lecture 02
Lecture 02
Table of Contents 1 | 66
for each n ≥ 1
• Fair game
• It will be tacitly assumed that M0 = 0 throughout
• What does this imply for EMn ?
The past 3 | 66
• The “past” can be seen as a wider amount of information as
time progresses
• Often includes some external information (covariates) in
addition to the previous values of the process itself
• In stochastic process theory the past is formulated as a
σ-algebra of events
• Informal definition: family of events that can be decided to
have happened or not by observing the past
• Denoted by Fn , formal way of representing what is known at
time n
• We will refer to Fn as the history at time n
• Family has to increase, since our past knowledge increases as
time passes
Discover the world at Leiden University
Survival Analysis
Stochastic processes in discrete time: Martingales in discrete time
Back to martingales 4 | 66
• Assume that Fn , for each n, is generated by M1 , . . . , Mn plus
possibly external information
• Technical formulation would be that the process
M = {M0 , M1 , M2 , . . .} is adapted to the history {Fn }
• This means that, for each n, the random variables M1 , . . . , Mn
are measurable with respect to the σ-algebra Fn
• Practical implication:
E(Mm | Fn ) = Mm for all m ≤ n.
More martingales 5 | 66
• So we have:
E(Mm | Fn ) = Mm for all m ≤ n.
• The first displayed formula states that we know the past and
present of the process M, the last states that the expected
value of the process in the future equals its present value
Discover the world at Leiden University
Survival Analysis
Stochastic processes in discrete time: Martingales in discrete time
Properties of martingales 6 | 66
• Since the martingale has mean zero for all n, we say that it is
a mean zero martingale
• By similar argument one may show that
Martingale differences 7 | 66
Variation processes 8 | 66
• Two processes describe the variation of a martingale
M = {M0 , M1 , M2 , . . .}
• The predictable variation process, denoted ⟨M⟩, is defined as
the sum of conditional variances of the martingale differences
n
X n o n
X
⟨M⟩n = E (Mi − Mi −1 )2 | Fi −1 = Var(∆Mi | Fi −1 ),
i =1 i =1
with ⟨M⟩0 = 0
• The optional variation process, denoted [M], is defined by
n
X n
X
[M]n = (Mi − Mi −1 )2 = (∆Mi )2 ,
i =1 i =1
with [M]0 = 0
Discover the world at Leiden University
Survival Analysis
Stochastic processes in discrete time: Variation processes
Stopping times 10 | 66
Stopped martingales 11 | 66
MnT = Mn∧T
Transformations 12 | 66
• Let X = {X0 , X1 , X2 , . . .} be some general process with a
history {Fn }, and let H = {H0 , H1 , H2 , . . .} be a
predictable process
• That is a sequence of random variables where each Hn is
measurable with respect to Fn−1 and hence is known one step
ahead of time
• The process Z defined by
• Formulated as sums:
n
X
⟨H • M⟩n = Hs2 ∆⟨M⟩s
s=1
n
X
[H • M]n = Hs2 ∆[M]s
s=1
Martingale increments 20 | 66
Multiple martingales 23 | 66
• Extension of the concept of covariance to martingales
• We may define the predictable covariation process ⟨M1 , M2 ⟩
for a pair of martingales M1 and M2 as the limit (in
probability) of the sum of conditional covariances
Cov(∆M1k , ∆M2k | F(k−1)t/n ) and the optional covariation
process [M1 , M2 ] as the limit of the sum of the products
∆M1k ∆M2k
• Informally, we may write
d⟨M1 , M2 ⟩(t) = Cov(dM1 (t), dM2 (t) | Ft− )
Multiple martingales 24 | 66
• As a consequence of these results
Cov(M1 (t), M2 (t)) = E(M1 (t)M2 (t)) =
= E⟨M1 , M2 ⟩(t) = E[M1 , M2 ](t)
Stochastic integrals 25 | 66
• Stochastic integrals are analogues to the transformations for
discrete-time martingales
• Let H = {H(t); t ∈ [0, τ ]} be a a predictable stochastic
process
• This means that for any time t, the value of H(t) is known
just before t
• (Formal definition of predictability is a bit intricate, we will
not go into this.)
• Sufficient conditions for H to be predictable are
• H is adapted to the history {Ft }
• The sample paths of H are left-continuous
• Predictability may sound uninteresting and technical, but
turns out to be very important!
Discover the world at Leiden University
Survival Analysis
Processes in continuous time: Stochastic integrals
Stochastic integrals 26 | 66
Stochastic integrals 27 | 66
• The stochastic integral can be defined as a limit of such a
transformation in the following sense:
n
X
I(t) = lim Hk ∆Mk
n→∞
k=1
Doob-Meyer decomposition 29 | 66
• In the discrete-time context we saw how a stochastic process
can be decomposed into a predictable process and a sequence
of martingale differences
• A similar result holds in continuous time and is known as the
Doob-Meyer decomposition
• An adapted process X = {X(t); t ∈ [0, τ ]} is called a
submartingale if it satisfies
E(X(t) | Fs ) ≥ X(s), for all t > s
Doob-Meyer decomposition 30 | 66
• The Doob-Meyer decomposition states that any submartingale
X can be decomposed uniquely as
X = X ∗ + M,
where X ∗ is a nondecreasing predictable process, often
denoted as the compensator of X, and M is a means zero
martingale
• Heuristically we have that
dX ∗ (t) = E(dX(t) | Ft− )
and
dM(t) = dX(t) − E(dX(t) | Ft− )
Doob-Meyer decomposition 31 | 66
Poisson process 32 | 66
• A homogeneous Poisson process describes the distribution of
events that occur independently of each other
• Imagine a basic rate of occurrence, denoted λ, such that the
probability of an event occurring in [t, t + dt) is λdt
• Well known properties:
• The time between events is exponentially distributed with
probability density λe −λt
• The expected value and the variance of the number of events in a
time interval of length h are both equal to λh
• The number of events in a time interval of length h is Poisson
distributed; the probability of exactly k events occurring is
(λh)k e −λh /k!
• The process has independent increments, that, is, the number of
events in nonoverlapping intervals are independent
Counting processes 35 | 66
• Counting process N = {N(t); t ∈ [0, τ ]} is a
right-continuous process with jumps of size 1 at event times
and constant in between
• We assume it is adapted to the history {Ft }
• (Technical way of saying that the history is generated by N and
possibly additional external information)
• Intensity process λ(t) of N (wrt Ft ) is heuristically defined by
λ(t)dt = P(dN(t) = 1 | Ft− ) = E(dN(t) | Ft− )
• More precisely, counting process is nondecreasing, hence a
submartingale. So by the Doob-Meyer decomposition there
exists a unique predictable process Λ(t), called the cumulative
intensity process, such that M(t) = N(t) − Λ(t) is a mean
zero martingale
Discover the world at Leiden University
Survival Analysis
Processes in continuous time: Counting processes
with a partition
defining ∆Mk =[0, t] into n−subintervals
M(kt/n) each of
M((k − 1)t/n) as length
the t/n,
increment of the martingale over the kth of these subintervals
• For the latter we obtain [M](t) = N(t)
and
Xk Z k Z t
X
Hj dMj (t) = Hj2 (s)dNj (s)
j =1 j =1 0
Independent censoring 46 | 66
• In previous lecture we gave an informal discussion of the
concept of independent censoring
• The main point was that independent censoring preserves the
form of the intensity processes of the counting processes
• We will now discuss this more formally
• Three different models are of interest
(i) A model for the hypothetical situation without censoring, where
all occurrences of the event of interest are observed
(ii) A joint model for the situation where all occurrences of the event
of interest as well as the censoring processes are observed
(iii) A model for the situation with censoring, i.e., for the occurrences
of the event actually observed
• Parameters of interest defined according to (i), concept of
independent censoring by (ii), actual inference by (iii)
Discover the world at Leiden University
Survival Analysis
Processes in continuous time: Independent censoring
Joint model 50 | 66
Independent censoring 51 | 66
Covariates 56 | 66
• In order to progress further, an assumption on the observation
of covariates is needed
• Typically, {Ftc } will be generated by the fully observed
counting processes as well as by covariate processes running
in parallel, and the αi (t) may depend on these covariates
• For statistical inference on αi (t), we must assume that these
covariates are available at time t, so they should be
Ft -predictable
• The intensity processes of the observed counting processes
take the form
λFi (t) = Yi (t)αi (t)
Modifications
• Possible to add drift
• Possible to use increments of Wiener process as building
blocks (Ornstein-Uhlenbeck)
Use
• Wiener processes can be used as models for underlying
processes
• Disease is often the result of a long deterioration of a biological
process
• Time-transformations of the Wiener process arise as limits in a
number of applications (in particular for stochastic integrals of
counting process martingales, basis for asymptotic theory of
Discover the and
estimators worldtest
at statistics)
Leiden University
Survival Analysis
Processes with continuous sample paths: The Wiener process and Gaussian
martingales
Gaussian martingales 60 | 66
Intuitive discussion 61 | 66
Intuitive discussion 62 | 66
• Two things to be taken care of to ensure that a sequence of
martingales converges to a Gaussian martingale
i The predictable variation processes of the martingales should
converge to a deterministic function
ii The sizes of the jumps of the martingales should go to zero
• The first assumption implies a stabilization of the sample
space of processes, the second one is a requirement on the
sample paths of the processes
Theorem
Let M̃ (n) be a sequence of mean zero martingales defined on
[0, τ ], and let M̃ε(n) be the martingale containing all the jumps of
P
M̃ (n) larger than a given ε > 0. Let → denote convergence in
probability. Under the conditions:
P
(i) ⟨M̃ (n) ⟩(t) → V (t) for all t ∈ [0, τ ] as n → ∞, where V is
a strictly increasing continuous function with V (0) = 0
P
(ii) ⟨M̃ε(n) ⟩(t) → 0 for all t ∈ [0, τ ] and all ε > 0 as n → ∞
Then, as n → ∞, the sequence of martingales M̃ (n) converges in
distribution to the mean zero Gaussian martingale U given by
U(t) = W (V (t))
martingale
• More generally, we will consider sums of stochastic integrals
k Z t
X (n) (n)
Hj (s)dMj (s),
j =1 0
(n)
where for each j , n, Hj (t) is a predictable process and
(n) (n) (n)
Mj t) = Nj (t) − 0t λj (s)ds is a counting process
R
martingale
Discover the world at Leiden University
Survival Analysis
Processes with continuous sample paths: Asymptotic theory for martingales