Stochastic Processes 2
Stochastic Processes 2
Stochastic Processes 2
Random Process
Denition 1. Random process A random (stochastic) process {Xt, t T } is a collection of random variables on the same probability space (, F , P ). The index set T is usually representing time and can be either an interval [t1, t2] or a discrete set. Therefore, the stochastic process X can be written as a function:
X : R R,
(t, ) X (t, )
Filtration
Observations:
The amount of information increases with time. We use the concept of sigma algebras. We assume that information is not lost with increasing time Therefore the corresponding -algebras will increase over time when there is more information available.
Denition 2. Filtration/adapted process A collection {Ft}t0 of sub -algebras is called ltration if, for every s t, we have Fs Ft. The random variables {Xt : 0 t } are called adapted to the ltration Ft if, for every t, Xt is measurable with respect to Ft.
Filtration
Example 1. Suppose we have a sample space of four elements: = {1, 2, 3, 4}. At time zero, we dont have any information about which has been chosen. At time T 2 we know wether we will have {1 , 2 } or {3, 4}.
r
B r A
r r
r r r
T 2
Let x() be a real, continuously dierentiable function dened on the interval [0, T ]. Its continuous dierentiability implies both a bounded total variation and a vanishing sum of squared increments: 1. Total variation:
dx(t) dt < dt
2. Sum of squares:
) N ( ( ( T) T) 2 lim x k x (k 1) =0 N N N k=1
Random processes do not have either of these nice smoothness properties in general. This allows the desired wild and random behavior of the (sample) noise signals.
Markov Process
A Markov process has the property that only the instantaneous value X (t) is relevant for the future evolution of X . Past and future of a Markov process have no direct relation.
Denition 3. Markov process A continuous time stochastic process X (t), t T, is called a Markov process if for any nite parameter set {ti : ti < ti+1} T we have
Transition probability
Denition 1. Let X (t) be a Markov process. The function P (s, X (s), t, B ) is the conditional probability P (X (t) B | X (s)) called transition probability or transition function.
This means it is the probability that the process X (t) will be found inside the area B at time t, if at time s < t it was observed at state X (s).
Transition probability
45
40
35 B
30
x(t)
25
20
15
10
Xs t s s+t 25 t 30 35 40 45 50
10
15
20
Gaussian Process
A stochastic process is called Gaussian if all its joint probability distributions are Gaussian. If X (t) is a Gaussian process, then X (t) N ((t), 2(t)) for all t. A Gaussian process is fully characterized by its mean and covariance function. Performing linear operations on a Gaussian process still results in a Gaussian process. Derivatives and integrals of Gaussian processes are Gaussian processes themselves (note: stochastic integration and dierentiation).
Martingale Process
A stochastic process X (t) is a martingale relative to ({Ft}t0, P ) if the following conditions hold:
X (t) is {Ft}t0-adapted, E [|X (t)|] < for all t 0. E[X (t)|Fs] = X(s) a.s. (0 s t).
The best prediction of a martingale process is its current value. Fair game model
10
Diusions
A diusion is a Markov process with continuous trajectories such that for each time t and state X (t) the following limits exist
:= :=
For these limits, (t, X (t)) is called drift and 2(t, X (t)) is called the diusion coecient.
11