Piecewise Deterministic Processes in Biological Models
Piecewise Deterministic Processes in Biological Models
Ryszard Rudnicki
Marta Tyran-Kamińska
Piecewise
Deterministic
Processes in
Biological Models
123
SpringerBriefs in Applied Sciences
and Technology
Mathematical Methods
Series editor
Anna Marciniak-Czochra, Heidelberg, Germany
More information about this series at http://www.springer.com/series/11219
Ryszard Rudnicki Marta Tyran-Kamińska
•
Piecewise Deterministic
Processes in Biological
Models
123
Ryszard Rudnicki Marta Tyran-Kamińska
Institute of Mathematics Institute of Mathematics
Polish Academy of Sciences University of Silesia
Katowice Katowice
Poland Poland
v
vi Preface
Acknowledgements
This research was partially supported by the National Science Centre (Poland)
Grant No. 2014/13/B/ST1/00224. The authors are grateful to Michael C. Mackey
and Katarzyna Pichór who read the manuscript and made useful suggestions for
improvements.
1 Biological Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Birth-Death Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.3 Grasshopper and Kangaroo Movement . . . . . . . . . . . . . . . . . . . . . 3
1.4 Velocity Jump Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.5 Size of Cells in a Single Line . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.6 Two-Phase Cell Cycle Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.7 Stochastic Billiard as a Cell Cycle Model . . . . . . . . . . . . . . . . . . 9
1.8 Stochastic Gene Expression I . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
1.9 Stochastic Gene Expression II . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.10 Gene Regulatory Models with Bursting . . . . . . . . . . . . . . . . . . . . 16
1.11 Neural Activity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
1.12 Processes with Extra Jumps on a Subspace . . . . . . . . . . . . . . . . . 19
1.13 Size-Structured Population Model . . . . . . . . . . . . . . . . . . . . . . . . . 21
1.14 Age-Structured Population Model . . . . . . . . . . . . . . . . . . . . . . . . . 24
1.15 Asexual Phenotype Population Model . . . . . . . . . . . . . . . . . . . . . 25
1.16 Phenotype Model with a Sexual Reproduction . . . . . . . . . . . . . . . 26
1.17 Coagulation-Fragmentation Process in a Phytoplankton
Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .... 28
1.18 Paralog Families . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .... 29
1.19 Definition of PDMP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .... 30
2 Markov Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
2.1 Transition Probabilities and Kernels . . . . . . . . . . . . . . . . . . . . . . . 33
2.1.1 Basic Concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
2.1.2 Transition Operators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
2.1.3 Substochastic and Stochastic Operators . . . . . . . . . . . . . . . 38
2.1.4 Integral Stochastic Operators . . . . . . . . . . . . . . . . . . . . . . . 39
2.1.5 Frobenius–Perron Operator . . . . . . . . . . . . . . . . . . . . . . . . 41
2.1.6 Iterated Function Systems . . . . . . . . . . . . . . . . . . . . . . . . . 42
vii
viii Contents
1.1 Introduction
p ji (Δt)
q ji = lim , i = j,
Δt→0 Δt
is called the transition rate from i to j. The process ξ(t) has transition rates qi+1,i =
bi, qi−1,i = di, and q ji = 0 otherwise. We also define
∞
qii = − q ji . (1.2)
j=0, j=i
The infinite matrix Q = [q ji ] is called the transition rate matrix or the infinitesi-
mal generator matrix. The matrix Q plays a crucial role in the theory of Markov
processes. In particular, if pi (t) = Prob(ξ(t) = i) for i ∈ N and t ≥ 0, then the col-
umn vector p(t) = [ p0 (t), p1 (t), . . .]T satisfies the following infinite dimensional
system of differential equations
p (t) = Qp(t).
d1 d2 dn
0 b0 1 b1 2 n–1 bn−1 n
Fig. 1.1 The diagram of the transition between states in the birth-death process
population with i individuals. The problem of the existence of such a process is non-
trivial and we shall consider it in detail in the next chapter. It should be noted that for
given rates we have a family of different birth-death processes and their distributions
depend on the initial distributions (i.e. the distribution of the random variables ξ(0)).
Two special types of birth-death processes are pure birth (if di = 0 for all i) and
pure death (if di = 0 for all i) processes. A Poisson process N (t) with intensity λ is
a pure birth process with bi = λ for all i.
A similar construction based on the transition rate matrix can be applied to define
general continuous-time Markov chains. Let Q = [q ji ] be a transition rate matrix,
i.e. a matrix which has non-negative entries outside the main diagonal and satisfies
(1.2). The matrix Q is a transition rate matrix for a homogeneous continuous-time
Markov chain ξ(t) if the following condition is fulfilled
Continuous-time Markov chains belong to a slightly larger class called pure jump-
type Markov processes. A pure jump-type Markov process is a Markov process,
which remains constant between jumps. For example, the process used in a simple
description of the grasshopper and kangaroo movement [81] is an example of a pure
jump-type Markov process, which is not a Markov chain. A grasshopper jumps at
random times (tn ) from a point x to the point x + ϑn . We assume that jump times
are the same as for a Poisson process N (t) with intensity λ > 0, i.e. N (tn ) = n,
and that (ϑn ) is a sequence of independent and identically distributed (i.i.d.) random
vectors, which are also independent of {N (t) : t ≥ 0}. Then the position ξ(t) of the
grasshopper is given by
N (t)
ξ(t) = ξ(0) + ϑn . (1.3)
n=1
and define
ξn−1 for tn−1 ≤ t < tn ,
ξ(t) =
ξn for t = tn ,
where the nth post-jump position ξn is an X -valued random variable such that
for every Borel subset B of Rd , where x(tn− ) and v(tn− ) are the left-hand side limits of
x(t) and v(t), respectively, at the point tn . Then ξ(t) = (x(t), v(t)), t ≥ 0, is a PDMP
corresponding to our movement and between jumps the pair (x(t), v(t)) satisfies the
following system of ordinary differential equations
x (t) = v(t),
(1.4)
v (t) = 0.
symmetric movement has values in the space R × {−1, 1} and P(x, v, {−v}) = 1 for
v = −1, 1. More advanced examples of velocity jump processes and their comparison
with dispersal of cells, insects and mammals are given in [81, 109]. We can consider
velocity jump processes defined in a bounded domain G. Examples of such processes
are stochastic billiards [36]. Stochastic billiards do not change velocity in the interior
of G but when an individual or a point strikes the boundary, a new direction is
chosen randomly from the directions that point back into the interior of G and
the motion continues. Stochastic billiards can describe not only the movement of
individuals in a bounded domain but also various different biological processes. In
Sect. 1.7, we present Lebowitz-Rubinow cell cycle model which can be identified
with a one-dimensional stochastic billiard. Some examples of PDMPs with jumps
on the boundary that appear in the theory of the gene regulatory systems are presented
in Sect. 1.12.
There have been a number of papers devoted to velocity jump processes, their dif-
fusion approximations and applications to aggregation and chemotaxis phenomena.
The interested reader is referred to [50, 75, 82] and the references therein.
Denote by ϕ(x) the division rate of a cell with size x, i.e. a cell with size x replicates
during a small time interval of length Δt with probability ϕ(x)Δt + o(Δt). Finally,
we assume that a daughter cell has a half size of the mother cell. It means that
where ξ(tn− ) is the left-hand side limit of ξ(t) at the point tn . In this way, we define
a homogeneous PDMP which is not a pure jump-type Markov process.
Given growth and division rate functions and the initial size x0 of a cell it is not
difficult to find the distribution of its life-span and the distribution of its size at the
point of division. Let π(t, x0 ) be the size of a cell at age t if its initial size is x0 , i.e.
π(t, x0 ) = x(t), where x is the solution of the equation x = g(x) with the initial
condition x(0) = x0 . The life-span of a cell is a random variable T which depends
on the initial size x0 of a cell. Let Φx0 (t) = Prob(T > t) be the survival function, i.e.
the probability that the life-span of a cell with initial size x0 is greater than t. Then
6 1 Biological Models
and F(t) = 1 − Φx0 (t) is the distribution function of the life-span. For y ≥ x0 , we
define t (x0 , y) to be the time t such that π(t, x0 ) = y. Since
∂t
· g(π(t, x0 )) = 1,
∂y
we see that
∂t 1
=
∂y g(y)
and
t (x0 ,y)
∂ ϕ(y)
ϕ(π(s, x0 )) ds = .
∂y 0 g(y)
x ϕ(r )
where Q(x) = dr. Let ζ be a random variable with exponential distribution
0 g(r )
i.e. Prob(ζ > x) = e−x for x ≥ 0. Then
Prob(ϑ > y) = exp(Q(x0 ) − Q(y)) = Prob ζ > Q(y) − Q(x0 )
= Prob Q −1 Q(x0 ) + ζ > y ,
which means that the random variables ϑ and Q −1 Q(x0 ) + ζ have the same dis-
tribution. From this it follows that
ξ(tn ) = 21 Q −1 Q(ξ(tn−1 )) + ζ ,
d
(1.9)
1.5 Size of Cells in a Single Line 7
(t4 )
(t1 )
(t3 )
(t0 )
ϕ(x) is the division rate of a cell with parameter x, and P(x, B) is a transition
probability function on X which describes the change of parameters from a mother
to a daughter cell, i.e.
The distribution function of tn − tn−1 is given by F(t) = 1 − Φx0 (t), where Φx0 (t)
is as in (1.7) with x0 := ξ(tn−1 ). PDMPs of this type are called flows with jumps (see
Fig. 1.2).
The cell cycle is a series of events that take place in a cell leading to its replication.
Usually the cell cycle is divided into four phases. The first one is the growth phase G 1
with synthesis of various enzymes. The duration of the phase G 1 is highly variable
even for cells from one species. The DNA synthesis takes place in the second phase
S. In the third phase G 2 significant protein synthesis occurs, which is required during
8 1 Biological Models
the process of mitosis. The last phase M consists of nuclear division and cytoplasmic
division. From a mathematical point of view, we can simplify the model by consid-
ering only two phases: A = G 1 with a random duration t A and B which consists of
the phases S, G 2 and M. The duration t B of the phase B is almost constant. There are
several models of the cell cycle. Let us mention the models by Lasota and Mackey
[61], Tyson and Hannsgen [112], and the generalization of these model given by
Tyrcha [115]. Here, we describe a continuous-time version of the Tyrcha model and
we show that it can be treated as a PDMP.
The crucial role in the model is played by a parameter x which describes the
state of a cell in the cell cycle. It is not clear what x exactly should be. We simply
assume that x is the cell size. We start with a model similar to that from Sect. 1.5
with only one difference. Here ϕ(x) is the rate of entering the phase B, i.e. a cell with
size x enters the phase B during a small time interval of length Δt with probability
ϕ(x)Δt + o(Δt). It is clear that the process ξ(t) is piecewise deterministic but it
is non-Markovian, because its future ξ(t), t ≥ t0 , depends not only on the random
variable ξ(t0 ) but also depends in which phase it is at the time t0 .
Now we extend the process ξ(t), t ≥ 0, to obtain a homogeneous PDMP. A new
process ξ (t), t ≥ 0, is defined on the phase space [0, ∞) × [0, t B ] × {1, 2} in the
following way. First, we add additional jump points sn = tn − t B , n ≥ 1. At time sn
a cell from the n-generation enters the phase B (see Fig. 1.3). Let
ξ (t) = (
ξ1 (t),
ξ2 (t),
ξ3 (t)) = (ξ(t), y(t), i),
sn tB
tn+1
sn+2
tB
tA
sn+1
tA
tA
tn−1 tn tn+1
1.6 Two-Phase Cell Cycle Model 9
⎧
⎪
ξ1 (t) = g(ξ1 (t)),
⎪
⎪
⎨
0, if ξ3 (t) = 1,
ξ2 (t) = (1.10)
⎪
⎪ 1, if ξ 3 (t) = 2,
⎪
⎩
ξ3 (t) = 0.
ξ1 (sn ) =
ξ1 (sn− ),
ξ2 (sn ) =
ξ2 (sn− ) = 0,
ξ3 (sn ) = 2,
and
ξ1 (tn ) = 21
ξ1 (tn− ),
ξ2 (tn ) = 0,
ξ3 (tn ) = 1.
where x0 = ξ1 (tn−1 ) and π(t, x0 ) = x(t) is the solution of equation x = g(x) with
initial condition x(0) = x0 , while that of tn − sn by F(t) = 0 for t ≤ t B and F(t) = 1
for t > t B .
The life-span tn − tn−1 of a cell with initial size x0 has the distribution
0, if t < t B ,
F(t) = t−t B (1.11)
1 − exp − 0 ϕ(π(s, x0 )) ds , if t ≥ t B ,
and the relation between the distributions of the random variables ξ(tn ) and ξ(tn−1 ):
ξ(tn ) = 21 π t B , Q −1 Q(ξ(tn−1 )) + ζ ,
d
(1.12)
One of the oldest models of cell cycle introduced by Rubinow [94] is based on the
concept of maturity and maturation velocity. Maturity, also called physiological age,
is a real variable x from the interval [0, 1] which describes the position of a cell in the
cell cycle. A new born cell has maturity 0 and a cell splits at maturity 1. In Rubinow’s
model x grows according to the equation x = v, where the maturation velocity v
can depend on x and also on other factors such as time, the size of the population,
temperature, light, environmental nutrients, pH, etc. If we neglect environmental
factors, resource limitations, and stochastic variation, and we assume that v depends
10 1 Biological Models
only on x, then all cells will have identical cell cycle in this model, in particular they
have the same cell cycle length l.
However, experimental observations concerning cell populations, cultured under
identical conditions for each member, revealed high variability of l in the popula-
tion. It means that the population is heterogeneous with respect to cell maturation
velocities, and therefore, mathematical models of the cell cycle should take into
account maturation velocities. A model of this type was proposed by Lebowitz and
Rubinow [64]. In their model the cell cycle is determined by its length l, which is
fixed at the birth of the cell. The relation between cycle lengths of mother’s and
daughter’s cells is given by a transition probability. If the maturation velocity of each
individual cell is constant, we can use v instead of l in the description of the model
because lv = 1. Such a model is a special case of the Rotenberg model [93] and we
briefly present it here using PDMP formalism. Each cell matures with individual
velocity v, which is a positive constant. We assume that the relation between the
maturation velocities of mother’s and daughter’s cells is given by a transition proba-
bility P(v, dv ). As in Sect. 1.5, we consider a sequence of consecutive descendants
of a single cell. Let tn be a time when a cell from the n-generation splits and let
ξ(t) = (x(t), v(t)), t ∈ [tn−1 , tn ), be the maturity and the maturation velocity of the
cell from the n-generation at time t. It is clear that ξ(t), t ≥ 0, is a PDMP and between
jumps the pair (x(t), v(t)) satisfies system (1.4). We assume that the process ξ(t),
t ≥ 0, has càdlàg sample paths, i.e. they are right-continuous with left limits. We
have x(tn ) = 0 and Prob(v(tn ) ∈ B |v(tn− ) = v) = P(v, B) for each n ∈ N and each
Borel subset B of (0, ∞). Since the v(t) is constant in the interval (tn−1 , tn ), we have
tn − tn−1 = 1/v(tn−1 ).
Properties of the density of the population with respect to (x, v) in the above model
were studied in [19]. But this model can be also identified with a one-dimensional
stochastic billiard. Namely, consider a particle moving in the interval [0, 1] with a
constant velocity. We assume that when the particle hits the boundary points 0 and 1,
it changes its velocity according to the probability measures P0 (−v, B) = P(v, B)
and P1 (v, −B) = P(v, B), respectively, where v > 0 and B is a Borel subset of
(0, ∞). Observe that the PDMP defined in the Lebowitz–Rubinow model is given
by ξ(t) = (x(t), |v(t)|), where x(t) and v(t) represent position and velocity of the
moving particle at time t. Asymptotic properties of the general one-dimensional
stochastic billiard were studied in [78].
It is not difficult to build a model which connects the classical Rubinow model
with Lebowitz–Rubinow’s one. We still assume that the cell maturity changes from
0 to 1 with velocity v and v depends on x and some specific genetic factor y (say
phenotypic trait) inherited from the mother cell which does not change during the
life of the cell. The trait y of the daughter cell is drawn from a distribution P(y, dy ),
where y is the mother cell trait. Similarly to the previous model, let tn be a time when
a cell from the n-generation splits and let ξ(t) = (x(t), y(t)), t ∈ [tn−1 , tn ), be the
maturity and the trait of the cell from the n-generation at time t. Then between jumps
the pair (x(t), y(t)) satisfies the following system:
1.7 Stochastic Billiard as a Cell Cycle Model 11
x (t) = v(x(t), y(t)),
(1.13)
y (t) = 0.
We have also x(tn ) = 0 and Prob(y(tn ) ∈ B |y(tn− ) = y) = P(y, B). Then the length
of the interval T = tn − tn−1 depends on the trait y of the cell and T is a positive
number such that x(T ) = 1, where x(t) is the solution of the initial problem x (t) =
v(x(t), y), x(0) = 0.
As we mentioned above, the Lebowitz–Rubinow model is a special case of the
Rotenberg model [93]. Now we briefly present this model using PDMP formalism.
In the Rotenberg model, the maturation velocity is random and can change also
during the cell cycle. A new born cell inherits the initial maturation velocity from
its mother as in the Lebowitz–Rubinow model according to a transition probability
P(v, dv ). During the cell cycle it can change its maturation velocity with intensity
q(x, v), i.e. a cell with parameters (x, v) can change the maturation velocity in a
small time interval of length Δt with probability q(x, v)Δt + o(Δt). Let x0 be the
maturity of a cell when it receives the maturation velocity v. From (1.7) it follows
that the probability distribution function of the length of the time interval when the
maturation velocity remains v is given by
t
1 − exp − q(x0 + vs, v) ds .
0
We suppose that if (x, v) is the state of the cell at the moment of division, then
a new maturation velocity is drawn from a distribution P(x, v, dv ). The process
ξ(t) = (x(t), v(t)) describing consecutive descendants of a single cell is a PDMP
which has jumps when cells split and random jumps during their cell cycles. Between
jumps the pair (x(t), v(t)) satisfies system (1.4). If a jump is at the moment of division
then it is given by the same formula as in the Lebowitz–Rubinow model. If a jump
is during the cell cycle, then x(tn ) = x(tn− ) and
Prob(v(tn ) ∈ B |x(tn− ) = x, v(tn− ) = v) = P(x, v, B) for each Borel subset B of (0, ∞).
Gene expression is a complex process which involves three processes: gene activa-
tion/inactivation, mRNA transcription/decay and protein translation/decay. Now we
consider a very simple model when proteins production is regulated by a single gene
and we omit the intermediate process of mRNA transcription (see Fig. 1.4). A gene
can be in an active or an inactive state and it can be transformed into an active state
or into an inactive state, with intensities q0 and q1 , respectively. The rates q0 and q1
depend on the number of protein molecules ξ(t). If the gene is active then proteins
12 1 Biological Models
Auto-regulation
q0
Inactive Active P
gene gene Protein
q1
degradation
are produced with a constant speed P. In both states of the gene, protein molecules
undergo the process of degradation with rate μ. It means that the process ξ(t), t ≥ 0,
satisfies the equation
ξ (t) = P A(t) − μξ(t), (1.14)
where A(t) = 1 if the gene is active and A(t) = 0 in the opposite case. Then the
process ξ (t) = (ξ(t), A(t)), t ≥ 0, is a PDMP. Since the right-hand side
of Eq.
(1.14)
is negative for ξ(t) > μP we can restrict values of ξ(t) to the interval 0, μP and the
process ξ (t) is defined on the phase space 0, μP × {0, 1}.
The process ξ (t) has jump points when the gene changes its activity. Formula
(1.7) allow us to find the distribution of time between consecutive jumps. Observe
that if x0 is the number of protein molecules at a jump point, then after time t we
have
P P −μt
πt0 (x0 ) = x0 e−μt , πt1 (x0 ) = + x0 − e
μ μ
A more advanced (and more realistic) model of gene expression was introduced by
Lipniacki et al. [65] and studied in [17]. In this model, mRNA transcription and
decay processes are also taken into consideration. The rates q0 and q1 depend on the
number of mRNA molecules ξ1 (t) and on the number of protein molecules ξ2 (t). If
the gene is active then mRNA transcript molecules are synthesized with a constant
speed R. The protein translation proceeds with the rate Pξ1 (t), where P is a constant
(see Fig. 1.5). The mRNA and protein degradation rates are μ R and μ P , respectively.
Now, instead of Eq. (1.14) we have the following system of differential equations
ξ1 (t) = R A(t) − μ R ξ1 (t),
(1.15)
ξ2 (t) = Pξ1 (t) − μ P ξ2 (t),
where A(t) = 1 if the gene is active and A(t) = 0 in the opposite case. Then the
process ξ (t) = (ξ1 (t), ξ2 (t), A(t)), t ≥ 0, is a homogeneous PDMP.
As in the previous model the process ξ (t) has jump points when the gene changes
its activity and formula (1.7) allow us to find the distribution of the time between
consecutive jumps. Let x0 and y0 be the initial number of mRNA and protein mole-
cules. Let us denote by πti (x0 , y0 ) = x(t) the solution of the system (1.15) at time
t, where i = 0 and i = 1 correspond to an inactive and an active state, respectively
(see Fig. 1.6). Then the probability distribution function of the length of an inactive
(or active) state is given by
t
1 − exp − qi (πsi (x0 , y0 )) ds .
0
Auto-regulation
q0
Inactive Active R P
mRNA Protein
gene gene
q1
R P
degradation degradation
2 2
1 1
1 1
i=0 i=1
1 1
Fig. 1.6 The flows π i for P = R = μ P = μ R = 1: left i = 0, right i = 1. The set between the
curves χ0 and χ1 is a stochastic attractor of the related PDMP, i.e. almost all trajectories of the
PDMP enter this set and do not leave it
In some cases the process of protein production has an additional primary tran-
script step [100]. After the activation of a gene, the DNA code is transformed into
pre-mRNA form of transcript. Then, pre-mRNA molecules are converted to func-
tional forms of mRNA, which is transferred into the cytoplasm, where during the
translation phase mRNA is decoded into a protein. We consider three-dimensional
model of gene expression with variables ξ1 , ξ2 , ξ3 describing evolution of pre-mRNA,
mRNA and protein levels. This model is described by the following system of dif-
ferential equations ⎧
⎨ ξ1 (t) = R A(t) − (C + μ p R )ξ1 (t),
⎪
ξ2 (t) = Cξ1 (t) − μ R ξ2 (t), (1.16)
⎪
⎩
ξ3 (t) = Pξ2 (t) − μ P ξ3 (t).
Remark 1.1 Examples from Sects. 1.8 and 1.9 show that using PDMPs to model
autoregulated genetic networks is very natural. New results concerning this subject
can be also found in [71, 77, 122, 123].
Models described in the last two sections lead to the following general scheme
called dynamical systems with random switching (see Fig. 1.7). We consider a family
of deterministic processes indexed by a set I ⊂ R. Each process is a solution of a
system of differential equations in some subset X of the space Rd :
1.9 Stochastic Gene Expression II 15
x4
x1 2
x2
1
1 2
x0
where πti (x0 ) is the solution of Eq. (1.17) with the initial condition x(0) = x0 . At
time Ti we leave the state i and using transition probability Px1 (i, B), x1 = πTi i (x0 ),
we choose the next state, say j. We repeat the procedure starting now from the point
x1 and the state j and continue it in subsequent steps. If i(t) is the state of the system
at time t then the process ξ (t) = (x(t), i(t)), t ≥ 0, is a homogeneous PDMP. We
can also treat the process ξ (t) as a solution of a single system of equations:
16 1 Biological Models
x (t) = bi(t) (x(t)),
(1.19)
i (t) = 0,
The central dogma of molecular biology, suggested 60 years ago, proposes that the
information flows from DNA to mRNA and then to proteins. A gene is expressed
through the processes of transcription and translation. First, in the process of tran-
scription, the enzyme RNA polymerase uses DNA as a template to produce an mRNA
molecule which can be translated to make the protein encoded by the original gene.
Recent advances to monitor the behaviour of single molecules allow experimental-
ists to quantify the transcription of mRNA from the gene as well as the translation
of the mRNA into protein. The visualization revealed that in many cases mRNA
and protein synthesis occur in quantal bursts in which the production is active for a
relatively short and random period of time and during that time a random number of
molecules is generated. There is the well-documented production of mRNA and/or
protein in stochastic bursts in both prokaryotes and eukaryotes. For example, in [25]
the authors observe bursts of expression of beta-galactosidase in E. coli, yeast and
mouse cells.
An analysis of the data obtained from such experiments was used in [40] to predict
the size and frequency of the bursts of proteins and to derive some of the first analytical
expressions for protein distributions in steady states across a population of cells.
Including mRNA dynamics in gene regulatory models introduces more complexity.
We assume that the rates of degradation of mRNA and of protein are γ1 and γ2 ,
respectively, and that k1 is the rate of transcription, while k2 is the rate of translation
of mRNA. Let us first consider translational bursting. It was observed in [25, 121]
that the amplitude of protein production through bursting translation of mRNA is
geometric/exponentially distributed with mean number of protein molecules per burst
equal to b = k2 /γ1 and the frequency of bursts being equal to a = k1 /γ2 . If ξ(t)
denotes the number of protein molecules then ξ(t) can be described in terms of
a continuous-time Markov chain with state space equal to N (see [72]). Here, we
follow the approach of [40] (see also [71, 72]) and we assume that ξ(t) denotes
the concentration of protein in a cell, which is the number of molecules divided
by the volume of the cell, being a continuous variable. We consider the general
situation when the degradation of proteins is interrupted at random times tn when
new molecules are being produced with frequency ϕ depending on the current level x
of the concentration of proteins and according to a distribution with density h(x, y).
Thus between the jump times ξ(t) is the solution of the equation
1.10 Gene Regulatory Models with Bursting 17
and at jump time tn we go from the point ξ(tn −) to the point ξ(tn −) + ϑn where ϑn
has a distribution with density h, possibly depending on x, so that
between jump times, and that mRNA are produced in bursts with jump rate ϕ depend-
ing on the current level ξ2 of proteins, see [15]. Regulatory mechanisms and genetic
switches can be modelled using this type of PDMPs [18, 69, 73, 80]. These are
particular examples of the so-called flows with jumps as described at the end of
Sect. 1.5.
V = Vm − Vm,R .
A cell is said to be excited (or depolarized) if V > 0 and inhibited (or hyperpolarized)
if V < 0. The Stein model [107, 108] describes how the depolarization V (t) is
changing in time. The cell is initially at rest so that V (0) = 0. Nerve cells may be
excited or inhibited through neuron’s synapses—junctions between nerve cells (or
18 1 Biological Models
subthreshold phase
λI λE λI λE
0
V − xI V V + xE V − xI V V + xE θ
tR
refractory phase
between muscle and nerve cell) such that electrical activity in one cell may influence
the electrical potential in the other. Synapses may be excitatory or inhibitory. We
assume that there are two non-negative constants x E and x I such that if at time
t an excitation occurs then V (t + ) = V (t − ) + x E and if an inhibition occurs then
V (t + ) = V (t − ) − x I . The jumps (excitations and inhibitions) may occur at random
times according to two independent Poisson processes N E (t), N I (t), t ≥ 0, with
positive intensities λ E and λ I , respectively. Between jumps the depolarization V (t)
decays according to the equation V (t) = −αV (t). When a sufficient (threshold)
level θ > 0 of excitation is reached, the neuron emits an action potential (fires). This
will be followed by an absolute refractory period of duration t R , during which V ≡ 0
and then the process starts again (see Fig. 1.8).
Now, we describe the neural activity as a PDMP. Since the refractory period has
a constant duration we can use a model similar to that of Sect. 1.6 with two phases A
and B, where A is the subthreshold phase and B is the refractory phase of duration
t R . We consider two types of jump points: when the neuron is excited or inhibited
and the ends of refractory periods. Thus, we can have one or more jumps inside
the phase A. Let t0 , t1 , t2 , . . . be the subsequent jump times. We denote by F the
subset of jump points consisting of firing points. Let ξ (t) = (V (t), 1) if the neuron
is in the phase A, and ξ (t) = (y, 2) if the neuron is in the phase B, where y is the
time since the moment of firing. The process ξ (t) = (ξ1 (t),
ξ2 (t)), t ≥ 0, is defined
on the phase space (−∞, θ ) × {1} ∪ [0, t R ] × {2} and between jumps it satisfies the
following system of equations:
−αξ1 (t), if
ξ2 (t) = 1,
ξ1 (t) =
1, if
ξ2 (t) = 2.
If the neuron is in the phase A, i.e. ξ2 (t) = 1, the depolarization can jump with
intensity λ = λ E + λ I . It means that F(t) = 1 − e−λt is the distribution function of
tn − tn−1 if tn−1 ∈
/ F , while F(t) = 0 for t < t R and F(t) = 1 for t ≥ t R if tn−1 ∈ F .
The transition at a jump point depends on the state of the neuron (its phase and the
ξ (tn− ) = (t R , 2) then
value of its depolarization). If ξ (tn ) = (0, 1) with probability
one; if ξ (tn ) = (V, 1) and V < θ − x E then
−
ξ (tn ) = (V + x E , 1) with probability
1.11 Neural Activity 19
λ E /λ and
ξ (tn ) = (V − x I , 1) with probability λ I /λ; while if ξ (tn− ) = (V, 1) and
V ≥ θ − x E then ξ (tn ) = (0, 2) with probability λ E /λ and
ξ (tn ) = (V − x I , 1) with
probability λ I /λ.
Remark 1.2 In [26] a simpler version of Stein’s model is considered without the
refractory period. A PDMP V (t), t ≥ 0, corresponding to this model is defined on
the phase space (−∞, θ ), between jumps it satisfies the equation V (t) = −αV (t),
and F(t) = 1 − e−λt is the distribution function of tn − tn−1 . The transition at a jump
point are given by two formulae:
if V (tn− ) < θ − x E then V (tn ) = V (tn− ) + x E with probability λ E /λ and V (tn ) =
V (tn− ) − x I with probability λ I /λ;
and if V (tn− ) ≥ θ − x E then V (tn ) = 0 with probability λ E /λ and V (tn ) = V (tn− ) −
x I with probability λ I /λ.
Remark 1.3 There are several different models of the neural activity. The interested
reader is referred to the monographs [7, 110] and papers [23, 24, 39, 42, 111].
A similar modification can be made in the model from Sect. 1.9, i.e. we assume
that the gene is activated if the number of protein molecules ξ2 (t) reaches the level θ ,
where 0 < θ ≤ P R/μ R μ P . It can happen that ξ1 (t¯) < μ p θ at the moment of gene
activation t¯, and then ξ2 (t¯) < 0. It implies that the number of protein molecules ξ2 (t)
decreases in some interval (t¯, t¯ + ε), thus it goes below the level θ . It is clear that
the gene should remain active until the number of protein molecules ξ2 (t) crosses
at least the level θ , therefore, we assume that the intensity of inactivation q1 (x1 , x2 )
equals zero if x2 ≤ θ . The corresponding ξ (t) = (ξ1 (t), ξ2 (t), A(t)), t ≥ 0
PDMP
is defined on the phase space X = 0, μRR × 0, μPP μR R × {0, 1} and has extra jumps
at the line segment S = 0, μRR × {θ } × {0, 1}. Here, an extra jump can be inside
the phase space. The probability distribution function of the length of an inactive
state is given by Eq. (1.20) and the intensity of activation is a measure such that
μ(X \ S) = 0 and μ{(x, θ, 0)} = ∞ for x ∈ 0, μRR .
An interesting PDMP with jumps on some subset is used in [20, 21] to describe
the motor-driven transport of vesicular cargo to synaptic targets located on the axon
or dendrites of a neuron. In this model, a particle is transported along the interval
[0, L] to some target x0 . The particle can be in one of three states i ∈ I = {−1, 0, 1}
1.12 Processes with Extra Jumps on a Subspace 21
and at a given state i it moves with speed i. Transitions between the three states
are governed by a time homogeneous Markov chain ϑ(t) defined on the set I with
the transition rate from i to j given by q ji . The particle starts from the point x = 0
and moves to the right. We assume that 0 is a reflecting point and L is an absorbing
point. Finally, we assume that if the particle is in some neighbourhood U of the
target x0 and is in the stationary state i = 0, then the particle can be absorbed by
the target at a rate κ. If the particle is absorbed by the target or by the point L then
the process repeats. The movement of the particle between jumps is described by a
velocity jump process ξ(t) = (x(t), i(t)), t ≥ 0, from Sect. 1.4 defined on the phase
space [0, L] × I , where the pair (x(t), i(t)) satisfies the following system of ordinary
differential equations
x (t) = i(t),
(1.21)
i (t) = 0.
If x ∈ (0, L) then the process ξ(t) can jump from (x, i) to (x, j) with intensity q ji
for i, j ∈ I , and if x ∈ U the process ξ(t) can also jump from (x, 0) to (0, 1) with
intensity κ. The process ξ(t) has two extra jumps from (0, −1) and from (L , 1) to
(0, 1) with probability 1.
There are a lot of different stochastic models of intracellular transport. The inter-
ested reader is referred to the review paper [22].
Remark 1.4 We can consider PDMPs with jump intensities given by measures
defined on the trajectories of the processes. Assume that a process ξ(t) is defined
on the space X ⊂ Rd and between jumps satisfies the equation ξ (t) = g(ξ(t)).
Let π(t, x0 ) = x(t) be the solution of this equation with the initial condition
x(0) = x0 defined on the maximal time interval (a, b), −∞ ≤ a < b ≤ ∞. Let
L = {π(t, x0 ) : t ∈ (a, b)} be the trajectory of x0 . We consider only the case when
the trajectory is not periodic. For x = π(t1 , x0 ) and y = π(t2 , x0 ) with t1 < t2 , we
denote by L(x, y) the arc {π(t, x0 ) : t ∈ (t1 , t2 ]}. The jump intensity on the trajec-
tory L can be a non-negative Borel measure μ on L. If ξ(0) = x, x ∈ L, then the
probability distribution function of the time till the next jump is given by
In particular, if μ(x) = c > 0 and ξ(t − ) = x then the process has a jump at time t
with probability 1 − e−c . In this case we should assume that the probability of jump
at the point y = ξ(t) is zero.
We return back to the model of size distribution from Sect. 1.5 but instead of a single
cell we consider the size distribution of all cells in the population. As in Sect. 1.5 the
size x(t) of a cell grows according to the equation
22 1 Biological Models
x (t) = g(x(t)).
A single cell with size x replicates with rate b(x) and dies with the rate d(x). A
daughter cell has a half size of the mother cell. Let assume that at time t we have k
cells and denote their sizes by x1 (t), x1 (t), . . . , xk (t). We can assume that a state of
the population at time t is the set
and the evolution of the population is a stochastic process ξ(t) = {x1 (t), . . . , xk (t)}.
Since the values of this process are sets of points, the process ξ(t) is called a point
process. Although such approach is a natural one, it has one important disadvantage.
We are not able to describe properly the situation when two cells have the same size.
One of the solution of this problem is to consider ξ(t) as a process whose values are
multisets. We recall that a multiset (or a bag) is a generalization of the notion of set
in which members are allowed to appear more than once. Another solution of this
problem is to consider ξ(t) as a process with values in the space of measures given
by
ξ(t) = δx1 (t) + · · · + δxk (t) ,
where δa denotes the Dirac measure at the point a, i.e. δa is the probability measure
concentrated at the point a. Also this approach has some disadvantages, for example
it is rather difficult to consider differential equations on measures. One solution of this
problem is to consider a state of the system as a k-tuple (x1 (t), . . . , xk (t)). But some
cells can die or split into two cells which changes the length of the tuple. To omit this
problem we introduce an extra “dead state” ∗ and describe the state of the population
at time t as an infinite sequence of elements from the space R+ ∗ = [0, ∞) ∪ {∗}
which has on some positions the numbers x1 (t), . . . , xk (t) and on other positions
has ∗. In order to have uniqueness of states, we introduce an equivalence relation
∼ in the space X of all R+ ∗ -valued sequences x such that x i = ∗ for all but finitely
many i. Two sequences x ∈ X and y ∈ X are equivalent with respect to ∼ if y can be
obtained as a permutation of x, i.e. x ∼ y if and only if there is a bijective function
σ : N → N such that y = (xσ (1) , xσ (2) , . . .). The phase space X in our model is the
space of all equivalence classes with respect to ∼, i.e. X = X/ ∼.
We now can describe the evolution of the population as a stochastic process
ξ(t) = [(x1 (t), x2 (t), . . .)] with values in the space X . The process ξ(t) has jump
points when one of the cells dies or replicates. We define g(∗) = b(∗) = b(∗) = 0
and assume that x(t) = ∗ is the solution of the equation x (t) = 0 with the initial
condition x(0) = ∗. Between jumps the process ξ(t) satisfies the equation
[(x1 (t) − g(x1 (t)), x2 (t) − g(x2 (t)), . . .)] = [(0, 0, . . .)]. (1.22)
For t ≥ 0 and x 0 ∈ R+
∗ we denote by π(t, x ) the solution x(t) of the equation
0
x (t) = g(x(t)) with the initial condition x(0) = x 0 . Let x0 = [(x10 , x20 , . . .)] ∈
X
and define
1.13 Size-Structured Population Model 23
The jump rate function ϕ(x) at the state x = [(x1 , x2 , . . .)] is the sum of rates of
death and division of all cells:
∞
ϕ(x) = (b(xi ) + d(xi )). (1.23)
i=1
If x0 ∈
X is the initial state of the population at a jump time tn , then the probability
distribution function of tn+1 − tn is given by
t
1 − exp − ϕ(π̃ (s, x0 ) ds . (1.24)
0
At time tn one of the cells dies or replicates. If a cell dies we change the sequence
by removing its size from the sequence and we have
di (xi (tn− ))
Prob ξ(tn ) = [(x1 (tn− ), . . . , xi−1 (tn− ), xi+1 (tn− ), xi+2 (tn− ), . . .)] =
ϕ(ξ(tn− ))
for i ∈ N. If a cell replicates, we remove its size from the sequence and add two new
elements in the sequence with the sizes of the daughter cells and we have
Prob ξ(tn ) = [(x1 (tn− ), . . . , xi−1 (tn− ), 21 xi (tn− ), 21 xi (tn− ), xi+1 (tn− ),xi+2 (tn− ), . . .)]
bi (xi (tn− ))
=
ϕ(ξ(tn− ))
for i ∈ N. In this way, we have checked that the point process ξ(t), t ≥ 0, is a
homogeneous PDMP with values in X.
We can identify the space
X with the space N of finite counting measures on R+
by a map η :
X → N given by
η(x) = δxi (1.25)
{i: xi =∗}
of points in a state space which is a metric space (S, ρ). First, we extend the state
space S adding “the dead element” ∗. We need to define a metric on S ∪ {∗}. The best
situation is if S is a proper subset of a larger metric space S . Then we simply choose
∗ as an element from S which does not belong to S and we keep the same metric.
In the opposite case, first we choose x0 ∈ S and define ρ(∗, x) = 1 + ρ(x0 , x) for
x ∈ S. Next, we define a metric d on the space X by
d([x], [y]) = min{d(a, b) : a ∈ [x], b ∈ [y]}.
We consider a population of one sex (usually females) in which all individuals are
characterized by their age x ≥ 0. We assume that an individual with age x dies with
rate d(x) and gives birth to one child with rate b(x) p1 or twins with rate b(x) p2 , where
p1 + p2 = 1. The model is very similar to that of the size-structured population. We
assume that g(∗) = b(∗) = d(∗) = 0 and g(a) = 1 for a ∈ R+ . Then the evolution
of the population is described by a stochastic process ξ(t) = [(x1 (t), x2 (t), . . .)] with
values in the space X with jumps when an individual dies or gives birth. Between
jumps the process ξ(t) satisfies Eq. (1.22) and the jump rate function ϕ(x) is given by
(1.23). Let us assume that at the jump time tn−1 the population consists of k individuals
with ages x1 , x2 , . . . , xk . It is easy to check that the probability distribution function
of tn+1 − tn is given by
t
k
1 − exp − (b(xi + s) + d(xi + s)) ds .
0 i=1
If ξ(tn− ) = x = [(x1 , x2 , . . .)], then ξ(tn ) has one of the following values:
xi . Even in this case the stochastic process ξ(t) is a homogeneous PDMP. Between
jumps, the process ξ(t) satisfies Eq. (1.22). The jump rate function is given by
∞
ϕ(x) = (b(xi , x) + d(xi , x)). (1.27)
i=1
We also need to replace b(xi ) and d(xi ) by b(xi , x) and d(xi , x) in formulas (1.26)
for the transition at a jump time.
As in Sects. 1.13 and 1.14 we consider a population of one sex in which all individuals
are characterized by their phenotype which is described by a vector x from a set
F ⊂ Rd . The phenotype remains constant during the whole life of an individual. As in
the previous model the state space of the population is X . Let x = [(x1 , x2 , . . .)] be the
state of the population. We assume that an individual with phenotype xi dies with rate
d(xi , x) and gives birth to one child with rate b(xi , x) p1 or twins with rate b(xi , x) p2 ,
where p1 + p2 = 1. Phenotypes of progeny can differ from the parent’s phenotype.
We assume that if x is the phenotype of the parent then the progeny phenotype is from
a set B ⊂ F with probability P(x, B), where P : F × B(F) → [0, 1] is a transition
probability on F and B(F) is a σ -algebra of Borel subsets of F. We assume that the
phenotype of twins is the same (but it is not difficult to consider a model which is
based on a different assumption). The point process ξ(t) = [(x1 (t), x2 (t), . . .)] which
describes the phenotype structure of the population is a pure jump-type homogeneous
Markov process, i.e. ξ (t) = 0. The jump rate function ϕ(x) is given by (1.27). Since
ξ(t) is a pure jump-type homogeneous Markov process the probability distribution
function of tn+1 − tn is given by 1 − e−ϕ(x)t . If ξ(tn− ) = x = [(x1 , x2 , . . .)], then
d(xi , x)
Prob(ξ(tn ) = [(x1 , . . . , xi−1 , xi+1 , . . .)]) = for i ∈ N,
ϕ(x)
∞
p1 b(xi , x)P(xi , B)
Prob(ξ(tn ) ∈ [(x, x1 , x2 , . . .)] : x ∈ B) = ,
i=1
ϕ(x)
∞
p2 b(xi , x)P(xi , B)
Prob(ξ(tn ) ∈ [(x, x, x1 , x2 , . . .)] : x ∈ B) = .
i=1
ϕ(x)
It should be noted that in cellular populations we need to change the last two
formulas because progeny consists of two cells ( p1 = 0), and we lose a cell which
replicates. In this case, we have
b(xi , x)P(xi , B)
Prob(ξ(tn ) ∈ [(x, x, x1 , . . . , xi−1 , xi+1 , . . .)] : x ∈ B) = .
ϕ(x)
26 1 Biological Models
Also the assumption that both daughter cells have the same phenotype is generally not
correct. If we consider hematopoietic stem cells—precursors of blood cells living
in the bone marrow—then such cells can be at different levels of morphological
development. After division a daughter cell can remain on the same level as the
mother cell or go to the next level. We can introduce a model with a point process to
describe the evolution of hematopoietic stem cells but a model which is based on a
continuous-time Markov chain seems to be more appropriate.
p(xi ) p(x j )
p(xi , x j ) = k . (1.28)
r =1 p(xr )
d(xi , x)
Prob(ξ(tn ) = [(x1 , . . . , xi−1 , xi+1 , . . .)]) = for i ∈ N,
ϕ(x)
∞
∞
p(xi , x j )P(xi , x j , B)
Prob(ξ(tn ) ∈ [(x, x1 , x2 , . . .)] : x ∈ B) = .
i=1 j=1
ϕ(x)
We now consider assortative mating. Then individuals with similar traits mate
more often than they would choose a partner randomly. Then we can use matching
theory, according to which each participant ranks all potential partners according
to its preferences and attempts to pair with the one with highest-ranking [2], or we
can adapt a model based on a preference function [41] used in two-sex populations
models. In assortative mating, a preference function a(x, y) is usually of the form
a(x, y) = ψ(x − y), where ψ : [0, ∞) → [0, ∞) is a continuous and decreasing
function. We assume that two individuals with phenotypes xi and x j form a pair with
rate
a(xi , x j ) ψ(xi − x j )
p(xi , x j , x) = ∞ = ∞ (1.31)
l=1 a(x i , xl ) l=1 ψ(x i − xl )
and the rest of the model is similar to that with semi-random mating.
If we consider a two-sex population (i.e. each individual is exclusively male or
female) then the mating process is more complex and there are only a few mathe-
matical models of it (see e.g. [60] for insect populations). Since the role of males
and females is different, it is clear that in such models we should characterize an
individual by its sex and phenotype. An important factor is a sexual selection [57]
which often depends on the social behaviour of individuals. Males usually produce
enough sperm to inseminate many females and in the case when they do not take part
in parental care their reproductive success depends on their phenotype. We give some
hints how to build a model in this case. First each individual is described by a pair
(x, s), where x is its phenotype, s = 0 if it is a male and s = 1 if it is a female. Let
x = [((x1 , s1 ), (x2 , s2 ), . . .)] be the state of the population. We assume that a female
with phenotype xi gives birth with rate b(xi , x) and a male with phenotype x j has
the competition rate p(x j , x). To simplify notation we set b(xi , x) = 0 if si = 0 and
p(x j , x) = 0 if si = 1. A female with phenotype xi and a male with phenotype x j
form a pair with rate
b(xi , x) p(x j , x)
p(xi , x j , x) = ∞ . (1.32)
r =1 p(xr , x)
The rest of the model is similar to the model of hermaphrodite population and we
omit it here.
28 1 Biological Models
d(xi )
Prob(ξ(tn ) = [(x1 , . . . , xi−1 , xi+1 , . . .)]) = ,
ϕ(x)
b(xi ) k(y, xi ) dy
Prob(ξ(tn ) ∈ [(x1 , . . . , xi−1 , x, xi − x, xi+1 , . . .)] : x ∈ B) = B
,
ϕ(x)
p(xi , x j )
Prob(ξ(tn ) = [(xi + x j , x1 , . . . , xi−1 , xi+1 , . . . , x j−1 , x j+1 , . . .)]) =
ϕ(x)
for i, j ∈ N.
0 1 2 3 n–1 n
Fig. 1.9 The diagram of the transition between states in the paralog families model
d xi
Prob(ξ(tn ) = [(x1 , . . . , xi−1 , xi + 1, xi+1 , . . .)]) = ,
ϕ(x)
r xi
Prob(ξ(tn ) = [(x1 , . . . , xi−1 , xi − 1, xi+1 , . . .)]) = ,
ϕ(x)
mxi
Prob(ξ(tn ) = [(1, x1 , . . . , xi−1 , xi − 1, xi+1 , . . .)]) = ,
ϕ(x)
for i ∈ N.
Remark 1.6 The main subject of the paper [98] is the size distribution of paralogous
gene families in a genome. Let sk (t) be the number of k-element families in our model
at time t. It means that sk (t) = #{i : xi (t) = k}. It follows from the description of
our model that
∞
s1 (t) = −(d + r )s1 (t) + 2(2m + r )s2 (t) + m js j (t), (1.35)
j=3
sk (t) = d(k − 1)sk−1 (t) − (d + r + m)ksk (t) + (r + m)(k + 1)sk+1 (t) (1.36)
for k ≥ 2.
We call Γ the active boundary of the set X . If the active boundary Γ of a set X is
non-empty then we can also consider the following survival function
32 1 Biological Models
t
Φx (t) = 1[0, t∗ (x)) (t) exp − ϕ(π(r, x)) dr , t ≥ 0, (1.41)
0
and set
π(t − τn−1 , ξn−1 ) for τn−1 ≤ t < τn ,
ξ(t) =
ξn for t = τn ,
where the nth post-jump position ξn is an X Δ -valued random variable such that
and ξ(τn− ) = limt↑τn ξ(t). Thus, the trajectory of the process is defined for all t <
τ∞ := limn→∞ τn and τ∞ is called the explosion time. To define the process for all
times, we set ξ(t) = Δ for t ≥ τ∞ . The process {ξ(t)}t≥0 is called the minimal PDP
corresponding to (π, Φ, P). It has right-continuous sample paths, by construction. In
the next chapter, we will show that the minimal PDP exists and is a Markov process.
Chapter 2
Markov Processes
In this chapter, we provide a background material that is needed to define and study
Markov processes in discrete and continuous time. We start by giving basic examples
of transition probabilities and the corresponding operators on the spaces of functions
and measures. An emphasis is put on stochastic operators on the spaces of integrable
functions. The importance of transition probabilities is that the distribution of a
stochastic process with Markov property is completely determined by transition
probabilities and initial distributions. The Markov property simply states that the past
and the future are independent given the present. We refer the reader to Appendix
A for the required theory on measure, integration, and basic concepts of probability
theory.
In this section, we introduce transition kernels, stressing the interplay between ana-
lytic and stochastic interpretations. Let X be a set and let Σ be a σ -algebra of subsets
of X . The pair (X, Σ) is called a measurable space. A function P : X × Σ → [0, 1]
is said to be a transition kernel if
(1) for each set B ∈ Σ, P(·, B) : X → [0, 1] is a measurable function;
(2) for each x ∈ X the set function P(x, ·) : Σ → [0, 1] is a measure.
If P(x, X ) = 1 for all x ∈ X then P is said to be a transition probability on (X, Σ).
Let (Ω, F , P) be a probability space. A mapping ξ : Ω → X is said to be an
X -valued random variable if it is measurable, i.e. for any B ∈ Σ
{ξ ∈ B} = ξ −1 (B) = {ω ∈ Ω : ξ(ω) ∈ B} ∈ F .
Given an X -valued random variable ξ on the probability space (Ω, F , P) the family
σ (ξ ) = {ξ −1 (B) : B ∈ Σ}
E(g(ξ, ϑ)|G ) = E(g(ξ, ϑ)|ξ ) and E(g(ξ, ϑ)|ξ = x) = E(g(x, ϑ)) (2.1)
for all x ∈ X and for any measurable non-negative function g; see Lemma A.3.
Let Y be a metric space. Consider a probability measure ν on the Borel σ -algebra
B(Y ). Given a measurable mapping κ : X × Y → X , we define
P(x, B) = 1 B (κ(x, y))ν(dy), x ∈ X, B ∈ Σ, (2.2)
where 1 B is the indicator function, equal to one on the set B and zero otherwise.
Clearly, P is a transition probability. For a Y -valued random variable ϑ with distrib-
ution ν and for every x ∈ X , the X -valued random variable κ(x, ϑ) has distribution
P(x, ·), since
P(κ(x, ϑ) ∈ B) = E(1 B (κ(x, ϑ))) = 1 B (κ(x, y))ν(dy) = P(x, B), B ∈ Σ.
As a particular example, we can take Y to be the unit interval [0, 1] and ν to be the
Lebesgue measure on [0, 1]; in that case ϑ is said to be uniformly distributed on the
unit interval [0, 1]. Now let ξ0 be an X -valued random variable independent of ϑ and
let ξ1 = κ(ξ0 , ϑ). We can write
and if we take g(x, y) = 1 B (κ(x, y)) in (2.1) then E(g(x, ϑ)) = P(x, B). Thus
which gives the conditional distribution of ξ1 given ξ0 . Moreover, if μξ0 is the dis-
tribution of ξ0 then the joint distribution μ(ξ0 ,ξ1 ) of ξ0 and ξ1 satisfies
μ(ξ0 ,ξ1 ) (B0 × B1 ) = P(ξ0 ∈ B0 , ξ1 ∈ B1 ) = P(x, B1 )μξ0 (d x)
B0
for all B0 , B1 ∈ Σ.
We need an extension of the concept of transition probabilities. Let (X 0 , Σ0 ) and
(X 1 , Σ1 ) be two measurable spaces. A function P : X 0 × Σ1 → [0, 1] is said to be
a transition probability from (X 0 , Σ0 ) to (X 1 , Σ1 ) if P(·, B) is measurable for every
B ∈ Σ1 and P(x, ·) is a probability measure on (X 1 , Σ1 ) for every x ∈ X 0 . The next
result ensures existence of probability measures on product spaces.
which is clearly a measure on the product space. Since the product σ -algebra is gen-
erated by rectangles, i.e. Σ0 ⊗ Σ1 = σ (C ), where C = {B0 × B1 : B0 ∈ Σ0 , B1 ∈
Σ1 }, and C is a π -system, the result follows from the π − λ lemma.
We now show that, for a large class of measurable spaces, any transition probability
is as in (2.2). We call a measurable space (X, Σ) a Borel space if X is isomorphic to
a Borel subset of [0, 1], i.e. there exists a Borel subset Y of [0, 1] and a measurable
bijection ψ : X → Y such that its inverse ψ −1 is also measurable. Any Borel subset
of a Polish space, i.e. a complete separable metric space, with Σ = B(X ) is a Borel
space (see e.g. [55, Theorem A1.2]). Recall that a metric space (X, ρ) is complete
if every sequence (xn ) satisfying the Cauchy condition limm,n→∞ ρ(xn , xm ) = 0 is
convergent to some x ∈ X : limn→∞ ρ(xn , x) = 0, and it is separable if there is a
countable set X 0 ⊂ X which is dense in X , so that for any x ∈ X there exists a
sequence from X 0 converging to x.
36 2 Markov Processes
For each x the function t → Fx (t) is right-continuous with Fx (1) = 1. Consider its
generalized inverse
with the convention that the infimum of an empty set is equal to +∞, and define
κ(x, q) = ψ −1 (Fx← (q)) if Fx← (q) ∈ ψ(X 1 ) and κ(x, q) = x1 if Fx← (q) ∈
/ ψ(X 1 ),
where x1 is an arbitrary point from X 1 . We have
It is convenient to associate with each transition kernel two linear mappings acting in
two Banach spaces, one which is a space of bounded functions and the other which
is a space of measures.
Let P be a transition kernel on a measurable space (X, Σ). Consider first the
Banach space B(X ) of bounded, measurable, real-valued functions on X with the
supremum norm
g u = sup |g(x)|.
x∈X
Tg u ≤ g u , g ∈ B(X ).
Then Pμ is a finite measure and Pμ(X ) ≤ μ(X ). The space M + (X ) is a cone, i.e.
c1 μ1 + c2 μ2 ∈ M + (X ), if c1 , c2 are non-negative constants and μ1 , μ2 ∈ M + (X ),
and we have
P(c1 μ1 + c2 μ2 ) = c1 Pμ1 + c2 Pμ2 .
Therefore we can extend P so that it is a linear operator on the vector space of all
finite signed measures
M (X ) = {μ1 − μ2 : μ1 , μ2 ∈ M + (X )},
Pμ TV ≤ μ TV, μ ∈ M (X ).
In this section, we suppose that a measurable space (X, Σ) carries a σ -finite measure
m. The set of all measurable and m-integrable functions is a linear space and it
becomes a Banach space, written L 1 (X, Σ, m), by defining the norm
f = | f (x)| m(d x), f ∈ L 1 (X, Σ, m),
X
and by identifying functions that are equal almost everywhere. Given a measurable
function f : X → R the essential supremum of f is defined by
The set of all functions with a finite essential supremum is denoted by L ∞ (X, Σ, m)
and it becomes a Banach space when we identify functions that are equal almost
everywhere and take · ∞ as the norm.
A linear operator P : L 1 (X, Σ, m) → L 1 (X, Σ, m) is called substochastic if it
is positive, i.e. if f ≥ 0 then P f ≥ 0 for f ∈ L 1 (X, Σ, m), and if P f ≤ f for
f ∈ L 1 (X, Σ, m). A substochastic operator is called stochastic if P f = f for
f ∈ L 1 (X, Σ, m) and f ≥ 0. We denote by P ∗ : L ∞ (X, Σ, m) → L ∞ (X, Σ, m)
the adjoint operator of P, i.e. for every f ∈ L 1 (X, Σ, m) and g ∈ L ∞ (X, Σ, m)
P f (x)g(x) m(d x) = f (x)P ∗ g(x) m(d x).
X X
dμ
P f = g, if f = ,
dm
which is the density of the measure ν = Pμ
Pμ(B) = P f (x) m(d x).
B
D = { f ∈ L 1 : f ≥ 0, f = 1}.
Remark 2.1 There exists a stochastic operator which does not have a transition ker-
nel [32]. But if X is a Polish space (a complete separable metric space) and Σ is a
σ -algebra of Borel subsets of X , then each substochastic operator on L 1 (X, Σ, m)
has a transition kernel [48]. The same result holds if X is a Borel subset of a Polish
space and Σ is the σ -algebra of Borel subsets of X .
If the state space is discrete and the measure m is the counting measure then (X, Σ, m)
is a σ -finite measure space and every kernel is an integral kernel.
We now consider the case of X = N = {0, 1, . . .}, where we use the notation l 1 =
L (X, Σ, m) with m being the counting measure on X . We represent
1
∞ any function
f as a sequence u = (u i )i∈N . We have (u i )i∈N ∈ l 1 if and only if i=0 |u i | < ∞. If
∞
ki j = 1 for all j,
i=0
is a stochastic operator. It can be identified with a matrix P = [ki j ]i, j∈N , called a
stochastic matrix. It has non-negative elements and the elements in each column
sum up to one. The transposed matrix
is called a transition matrix and it is such that the elements in each row sum up to
one. Note that k ji is the probability of going from state i to state j.
and the transformation S which satisfies (2.6) is called non-singular. This operator
can be extended to a bounded linear operator PS : L 1 → L 1 , and PS is a stochastic
operator. The operator PS is called the Frobenius–Perron operator for the transfor-
mation S or the transfer operator or the Ruelle operator.
We now give a formal definition of the Frobenius–Perron operator. Let (X, Σ, m)
be a σ -finite measure space and let S be a measurable non-singular transformation
of X . An operator PS : L 1 → L 1 which satisfies the following condition
PS f (x) m(d x) = f (x) m(d x) for B ∈ Σ and f ∈ L 1 (2.7)
B S −1 (B)
42 2 Markov Processes
d(m ◦ S −1 )
PS f (x) = 1 S(X ) (x) f (S −1 (x)) (x) for m-a.e. x ∈ X,
dm
n
= f (ψi (x))| det ψi (x)| d x
i=1 B∩ S(Ui )
= f (ψi (x))| det ψi (x)| d x = P f (x) d x.
B i∈I B
x
the action of the system is given by S j (x). Thus we consider the following transition
probability
k
P(x, B) = p j (x)δ S j (x) (B)
j=1
k
k
Pμ(B) = p j (x)δ S j (x) (B)μ(d x) = p j (x)μ(d x).
j=1 X j=1 S −1
j (B)
k
k
Pμ(B) = p j (x) f (x) m(d x) = PS j ( p j f )(x) m(d x),
j=1 S −1
j (B) j=1 B
k
Pf = PS j ( p j f ), f ∈ L 1.
j=1
We can extend the iterated function system in the following way. Consider a family
of measurable transformations S y : X → X , y ∈ Y , where Y is a metric space which
carries a Borel measure ν, and a family of measurable functions p y : X → [0, ∞),
y ∈ Y , satisfying
p y (x)ν(dy) = 1, x ∈ X,
Y
is given by
P f (x, i) = pi (x, j) f (x, j).
j=i
ξ
The sequence Fn = σ (ξ0 , . . . , ξn ), n ≥ 0, is called the natural filtration or the his-
tory of the process (ξn ). We have
The monotone class theorem and properties of the conditional expectation imply that
if (ξn ) has the weak Markov property then for each k ≥ 0 and all g0 , . . . , gk ∈ B(X )
the following holds
We can extend it further as follows. A sequence (ξn ) has the weak Markov property
if and only if the future σ (ξm : m ≥ n) and the past σ (ξm : m ≤ n) are conditionally
independent given the present σ (ξn ) (see Lemma A.4), i.e. for each n and all A ∈
σ (ξm : m ≥ n) and F ∈ σ (ξm : m ≤ n) we have
for all B ∈ Σ and n ≥ 0. Taking the conditional expectation of (2.11) with respect
to σ (ξn ) we obtain
P(ξn+1 ∈ B|ξn ) = P(ξn , B)
which implies that ξ = (ξn )n≥0 has the weak Markov property. Moreover, we have
X N = {x = (x0 , x1 , . . .) : xn ∈ X, n ≥ 0}
To see this let for each ω the mapping n → ξn (ω) denotes the sequence ξ(ω) =
(ξn (ω))n≥0 ∈ X N . Thus we obtain the mapping ξ : Ω → X N and ξ is measurable
if and only if ξn : Ω → X is measurable for each n ∈ N. The distribution of the
process ξ is a probability measure μξ on (X N , Σ N ) and it is uniquely defined through
its values on finite-dimensional rectangles. In other words, the finite-dimensional
distributions μ0,1,...,n = μ(ξ0 ,...,ξn )
μξ (B0 × · · · × Bn × X N ) = P(ξ ∈ B0 × · · · × Bn × X N )
= μ(ξ0 ,...,ξn ) (B0 × · · · × Bn )
P(ξ0 ∈ B0 , ξ1 ∈ B1 , . . . , ξn ∈ Bn )
= ... P(xn−1 , Bn )P(xn−2 , d xn−1 ) . . . P(x0 , d x1 ) μ(d x0 )
B0 B1 Bn−1
(2.14)
Suppose that (ϑn )n≥1 is a sequence of i.i.d. random variables with values in a metric
space Y . Let ξ0 be an X -valued random variable independent of (ϑn )n≥1 and with dis-
tribution μ. Consider a measurable function κ : X × Y → X and define a sequence
2.2 Discrete-Time Markov Processes 47
ξn = κ(ξn−1 , ϑn ), n ≥ 1. (2.15)
ξn = ξn−1 + ϑn , n ≥ 1.
From Proposition 2.2 it follows that a typical discrete-time Markov process can
be defined by a recursive formula as in (2.15).
Theorem 2.2 Let P be a transition probability on a Borel space (X, Σ). Then there
exists a measurable function κ : X × [0, 1] → X such that for any sequence (ϑn ) of
independent random variables with uniform distribution on [0, 1] and for any x ∈ X
the sequence defined by (2.15) is a discrete-time Markov process with transition
probability P and initial distribution δx .
To define a discrete-time Markov process with transition probability P on a Borel
space (X, Σ), we can take Ω, by Theorem 2.2, to be the countable product [0, 1]N of
the unit interval [0, 1] with the product σ -algebra F = B([0, 1])N and the measure P
as a countable product of the Lebesgue measure on [0, 1]. Then we define a sequence
of i.i.d. random variables ϑn : Ω → [0, 1], n ≥ 0, by ϑn (ω) = ωn for ω = (ωn )n≥0 ∈
Ω, and ξn by (2.15).
Px (B0 × · · · × Bn × X N )
= 1 B0 (x) ... P(xn−1 , Bn )P(xn−2 , d xn−1 ) . . . P(x, d x1 )
B1 Bn−1
Ω = X N = {ω = (x0 , x1 , . . .) : xn ∈ X, n ≥ 0}
A = {ω : ξ0 (ω) ∈ B0 , . . . , ξn (ω) ∈ Bn }
then we have
Pμ (A) = Pμ (ξ0 ∈ B0 , ξ1 ∈ B1 , . . . , ξn ∈ Bn )
= ... P(xn−1 , Bn )P(xn−2 , d xn−1 ) . . . P(x0 , d x1 )μ(d x0 ).
B0 B1 Bn−1
Consequently, for each probability measure Pμ on (Ω, F ) the process ξ = (ξn )n≥0
is homogeneous Markov with transition probability P and initial distribution μ. We
write Ex for the integration with respect to Px , x ∈ X , and we have
Eμ (η) = Ex (η) μ(d x)
X
By the monotone class theorem the Markov property (2.16) holds if and only if
for all g ∈ B(X ), where B(X ) is the space of bounded and measurable functions
g : X → R. Note that a process is Markov if and only if for any 0 ≤ s1 < · · · < sn
and g ∈ B(X )
for all g ∈ B(X ), s, t ≥ 0. We interpret P(t, x, B) as the probability that the sto-
chastic process ξ moves from state x at time 0 to a state in B at time t. The transition
50 2 Markov Processes
on the Banach space B(X ) with supremum norm · u . It follows from the Chapman–
Kolmogorov equation (2.17) that T (t), t ≥ 0, is a semigroup, i.e.
If τ1 and τ2 are two stopping times and τ1 ≤ τ2 then Fτ1 ⊆ Fτ2 . A Markov process
ξ is said to be strong Markov if for any stopping time τ , the strong Markov property
at τ holds
for all B ∈ Σ and all times t. If X is a Borel space and the process has right-continuous
sample paths then ξ(τ ) is Fτ -measurable on {τ < ∞} for any stopping time τ for
the filtration (Ft ).
As in the discrete-time case we can consider a canonical space Ω = X R+ , which
now is the space of all functions ω : R+ → X , with product σ -algebra Σ R+ = σ (C )
which is the smallest σ -algebra of subsets of X R+ containing the family C of all
cylinder sets:
and
2.3 Continuous-Time Markov Processes 51
Px (ξ(t1 ) ∈ B1 , . . . , ξ(tn ) ∈ Bn )
= ... P(tn − tn−1 , xn−1 , Bn )P(tn−1 − tn−2 , xn−2 , d xn−1 ) . . . P(t1 , x, d x1 )
B1 Bn−1
for B ∈ B(Rd ), since ξ(s + t) = ξ(s + t) − ξ(s) + ξ(s) and ξ(s + t) − ξ(s) is
independent of Fs , which shows that ξ has the Markov property.
The process ξ has stationary increments, if the distribution of ξ(s + t) − ξ(s) is
the same as the distribution of ξ(t) − ξ(0) for all s, t ≥ 0. In that case
for all x and s, t. In particular, we have ξ(s + t) − ξ(0) = ξ(s + t) − ξ(t) + ξ(t) −
ξ(0) and the random variables ξ(s + t) − ξ(t) and ξ(t) − ξ(0) are independent.
Hence, if μt is the distribution of ξ(t) − ξ(0), then μs+t is the convolution of μs and
μt , i.e.
μs+t (B) = (μs ∗ μt )(B) = μs (B − x)μt (d x),
and
P(t, x, B) = 1 B (x + y)μt (dy).
Rd
ε > 0 we have
lim P(|ξ(t)| > ε) = 0.
t→0
Let (ξn )n≥0 be a discrete-time Markov process with transition probability P̄ and
(σn )n≥1 be a sequence of independent random variables, exponentially distributed
with mean λ−1 , and independent of the sequence (ξn )n≥0 . We set τ0 = 0 and we
define
n
ξ(t) = ξn−1 if τn−1 ≤ t < τn , where τn = σk , n ≥ 1.
k=1
It follows from the strong law of large numbers that τn → ∞, as n → ∞, a.s. The
sample paths of the process ξ are constant between consecutive τn and the random
variables σn are called holding times.
Let N (t) be the number of jump times τn in the time interval (0, t], i.e. we have
∞
N (t) = max{n ≥ 0 : τn ≤ t} = 1(0,t] (τn ).
n=0
(λt)n
P(N (t) = n) = e−λt , n ≥ 0.
n!
P(N (t) = n) = P(τn ≤ t < τn+1 ) = P(τn+1 > t) − P(τn > t),
2.3 Continuous-Time Markov Processes 53
since {τn > t} ⊆ {τn+1 > t}. The random variable τn , being the sum of n indepen-
dent exponentially distributed random variables with parameter λ, has a gamma
distribution with density
λn x n−1 −λx
f τn (x) = e for x ≥ 0.
(n − 1)!
Hence,
∞
λn x n−1 −λx (λt)n
P(τn > t) = e d x = −e−λt + P(τn+1 > t).
t (n − 1)! n!
We can write ξ(t) = ξ N (t) for t ≥ 0. In particular, if we take X = N and the trivial
Markov chain ξn = n for all n, then we have ξ(t) = N (t) for all t ≥ 0 and N (t),
t ≥ 0, is a Poisson process with intensity λ > 0. If (ξn )n≥0 is a random walk then ξ(t)
is the compound Poisson process as in (1.3). The Poisson process N has stationary
independent increments, equivalently, for any n, t, s and A ∈ σ (N (r ) : r ≤ s)
Consider again the general process ξ(t) = ξ N (t) for t ≥ 0. We denote by ξx (t)
the particular process ξ(t) starting at ξ(0) = ξ0 = x. We can easily calculate the
distribution of ξx (t). By independence of N (t) and (ξn ), we obtain
∞
P(ξx (t) ∈ B) = P(ξ(t) ∈ B|ξ0 = x) = P(N (t) = n)P(ξn ∈ B|ξ0 = x).
n=0
Since the random variable N (t) is Poisson distributed with parameter λt, we have
∞
(λt)n n
P(t, x, B) = e−λt P̄ (x, B), (2.21)
n=0
n!
E(1 B (ξ(t + s))|Fs ) = E(1 B (ξ N (t+s) )|Fs ) = E(1 B (ξ N (t+s)−N (s)+N (s) )|Fs ),
which gives
∞
E(1 B (ξ(t + s))|Fs ) = E(1{N (t+s)−N (s)=n} 1 B (ξn+N (s) )|Fs ).
n=0
54 2 Markov Processes
The family
Assume that ξ(t) is a Markov process with state space X defined on (Ω, F , Px ). Let
T (t)g(x) = Ex g(ξ(t)) = g(y)P(t, x, dy), g ∈ B(X ), x ∈ X, t ≥ 0,
X
where Ex is the expectation with respect to Px . Consider the class D(L) of all bounded
and measurable functions g : X → R such that the limit
Ex (g(ξ(t))) − g(x)
lim
t↓0 t
2.3 Continuous-Time Markov Processes 55
exists uniformly for all x ∈ X . We denote the limit by Lg(x) and we call L the
infinitesimal generator of the Markov process ξ . In particular,
1
Lg(x) = lim (g(y) − g(x))P(t, x, dy)
t↓0 t X
which shows that D(L) consists of all bounded measurable functions and that
Lg(x) = λ (g(y) − g(x)) P̄(x, dy), x ∈ X, g ∈ B(X ).
X
Using (2.20) we conclude that the infinitesimal generator of this process is of the
form
Lg(x) = ϕ(x) (g(y) − g(x))P(x, dy), x ∈ X, g ∈ B(X ).
X
is a martingale, i.e. the random variable η(t) is integrable, Ft -measurable for each
t ≥ 0, and
E(η(t + s)|Fs ) = η(s)
for all t, s ≥ 0. In particular, the Dynkin formula holds if ξ(t) is a Markov process
with a metric state space and right-continuous paths. To see this take g and Lg
bounded and observe that
t+s
Ex (η(t + s)|Fs ) = Ex (g(ξ(t + s))|Fs ) − g(x) − Ex (Lg(ξ(r ))|Fs ) dr.
0
which gives the claim, by using the identity (see (3.4) in Sect. 3.1.2)
t
T (t)g = g + T (r )(Lg) dr, t ≥ 0.
0
There are several different versions of generators for Markov processes. For exam-
ple, one can consider instead of the uniform convergence in B(X ), the pointwise
convergence or the so-called bounded pointwise convergence (see [35]). Another
approach, given by [28], introduces the extended generator using the concept of
local martingales and allowing unbounded functions in the domain of the generator;
here we adopt this definition. Let M(X ) be the space of all measurable functions
g : X → R. An operator L is called the extended generator of the Markov process
ξ , if its domain D( L) consists of those g ∈ M(X ) for which there exists f ∈ M(X )
such that for each x ∈ X , t > 0,
t
Ex (g(ξ(t))) = g(x) + Ex f (ξ(r )) dr
0
and t
Ex (| f (ξ(r ))|) dr < ∞,
0
In this section, we consider the general setting from Sect. 1.19. We assume that (X, Σ)
is a Borel space and that (π, Φ, P) are three characteristics representing, respectively,
a semiflow, a survival function, and a jump distribution, being a transition probability
from X ∪ Γ to X where Γ is the active boundary. We assume that P(x, X \ {x}) = 1
for all x ∈ X ∪ Γ . Since (X, Σ) is a Borel space, we can find a measurable mapping
κ : (X ∪ Γ ) × [0, 1] → X such that
Let (Ω, F , P) be a probability space and let (ϑn )n≥1 be a sequence of independent
random variables with uniform distribution on [0, 1]. We define τ0 = σ0 = 0, ξ0 = x
and
σ1 = Φξ← 0
(ϑ1 ), τ1 = σ1 + τ0 ,
and we set ξ(t) = π(t − τ0 , ξ0 ) for t < τ1 . Since the function t → π(t, x) has a
left-hand limit, which belongs to the set X Δ ∪ Γ , we can define
On the set {τ1 = ∞} the process is defined for all times t. On {τ1 < ∞} we continue
the construction of the process inductively. We define σ2 = Φξ← 1
(ϑ3 ), τ2 = σ2 + τ1 ,
and we set
and so on. Consequently, we define the minimal process {ξ(t)}t≥0 starting at ξ(0) = x
by
π(t − τn , ξn ), if τn ≤ t < τn+1 , n ≥ 0,
ξ(t) = (2.24)
Δ, if t ≥ τ∞ ,
where
τn = σn + τn−1 , σn = Φξ←
n−1
(ϑ2n−1 ), ξn = κ(ξ(τn− ), ϑ2n ), n ≥ 1. (2.25)
Let N (t) be the number of jump times τn in the time interval [0, t]
Then N (t) = 0 if t < τ1 , N (t) = n if and only if τn ≤ t < τn+1 , and N (t) = ∞ for
t ≥ τ∞ . If we set ξ∞ = Δ and τ∞+1 = ∞ then we have
We outline the main steps of the proof. It is similar to the proof given in [28, 29]
for processes with Euclidean state space. Let Ft be the filtration generated by ξ(t).
Then each τn is a stopping time. Observe that for any t ≥ 0
Thus we obtain
and leads to the weak Markov property. To show the strong Markov property, we
take a stopping time τ . We have
and for each n there exists Fτn -measurable random variable ζn such that
Note that if
Ex (N (t)) = Ex 1(0,t] (τn ) < ∞ for all t > 0, x ∈ X, (2.29)
n
then the process ξ is non-explosive, i.e. τ∞ = ∞ a.s. In that case we have P(t, x, X ) =
1 for all t > 0 and x ∈ X .
Remark 2.2 If P(x, {x}) = 0 for some x ∈ X then we can extend the state space X
to by
X = X × {0, 1} and define a transition probability P
P((x, i), B × {1 − i}) = P(x, B),
P((x, i), B × {i}) = 0, (x, i) ∈ (X ∪ Γ ) × {0, 1}.
κ:
It follows from (2.22) that X × [0, 1] →
X given by
κ ((x, i), r ) = (κ(x, r ), 1 −
i) satisfies
P((x, i), B × { j}) = Leb{r ∈ [0, 1] :
κ ((x, i), r ) ∈ B × { j}}).
We consider the minimal PDMP ξ with characteristics (π, Φ, P) and jump times
(τn ) as given in Sect. 2.3.5. Let Px be the distribution of the process ξ(t) starting at
x. For any non-negative measurable functions h defined on X Δ × [0, ∞], we have
Ex [h(ξ(τ1 ), τ1 )] = h(y, s)P(π(s − , x), dy)Φx (ds).
X Δ ×[0,∞]
for B ∈ ΣΔ , J ∈ B([0, ∞]). The strong Markov property of the process ξ(t) at τn
implies that the sequence (ξ(τn ), τn ), n ≥ 0, is a Markov chain on X Δ × [0, ∞] such
that for B ∈ ΣΔ and J ∈ B([0, ∞])
Since ξ(t) = π(t − τn , ξ(τn )) on {τn ≤ t < τn+1 }, it follows form (2.28) that
∞
P(t, x, B) = 1 B (π(t − s, y))Φ y (t − s)K n (x, dy, ds) (2.30)
n=0 X ×[0,t]
We now suppose that the semiflow is continuous in X , i.e. π(s − , x) = π(s, x) for
all s < t∗ (x), x ∈ X . We consider Φ as in (1.38) defined with the help of a jump rate
function ϕ. Then
t t
T (Tn−1 (t − s)g)(π(s, x))Φx (ds) = T (Tn−1 (t − s)g)(π(s, x))ϕ(π(s, x))Φx (s) ds
0 0
and we obtain
t
Tn (t)1 B (x) = T0 (t)1 B (x) + T0 (s)(ϕT (Tn−1 (t − s)1 B ))(x) ds (2.34)
0
then g ∈ D( L) and
Lg(x) = L 0 g(x) + ϕ(x) (g(y) − g(x))P(x, dy), x ∈ X, (2.35)
X
with L 0 g defined by
t
g(π(t, x)) − g(x) = L 0 g(π(s, x)) ds, t < t∗ (x), x ∈ X.
0
A more general condition instead of (3) characterizes all elements of the domain
of the extended generator as defined and showed in [28, 29]. Note that if g is bounded
and condition (2.29) holds, then g satisfies condition (3).
Chapter 3
Operator Semigroups
Semigroups of linear operators provide the primary tools in the study of continuous-
time Markov processes. They arise as the solutions of the initial value problem
for the differential equation u (t) = Au(t), where A is a linear operator acting on a
Banach space. We describe what is generally regarded as the basic theory. We provide
basic definitions, examples and theorems characterizing the operators as being the
generators of semigroups. The aim here is to provide necessary foundations for
studying semigroups on L 1 spaces in the next chapter.
f = 0 if and only if f = 0,
c f = |c| f , whenever c ∈ R and f ∈ X ,
f + g ≤ f + g, whenever f, g ∈ X ,
linear operator A with D(A) = X is bounded if and only if it is continuous, i.e. the
mapping f → A f is continuous for all f ∈ X . The operator A is a contraction if
A ≤ 1. The operator A is said to be densely defined if its domain D(A) is dense
in X , so that every f ∈ X is a limit of a sequence of elements from D(A). The
operator (A, D(A)) is closed if its graph {( f, A f ) : f ∈ D(A)} is a closed set in the
product space X × X or equivalently if f n ∈ D(A), n ≥ 1,
α, A f = β, f , f ∈ D(A), (3.1)
n
S(Δ, u) = u(sk )|tk − tk−1 |
k=1
where sk ∈ (tk−1 , tk ]. Let f ∈ X . If for each ε > 0 there exists δ > 0 such that for
every partition Δ with supk |tk − tk−1 | < δ we have f − S(P, u) < ε, then the
3.1 Generators and Semigroups 65
Most properties of the Riemann integral for real-valued functions can be proved
for normed space-valued functions. It is easy to see that we have the following: if
u : [a, b] → X is continuous then u is uniformly continuous on [a, b] and Riemann
integrable on every interval [a, t] for t ∈ [a, b]; in particular
t
w(t) = u(s) ds
a
1
D(A) = { f ∈ X : lim (S(t) f − f ) exists in X },
t↓0 t
1
A f = lim (S(t) f − f ), f ∈ D(A).
t↓0 t
Proposition 3.1 Let {S1 (t)}t≥0 and {S2 (t)}t≥0 be strongly continuous semigroups
on a Banach space X with generators (A1 , D(A1 )) and (A2 , D(A2 )), respectively.
If A1 ⊆ A2 then S1 (t) f = S2 (t) f for all f and t ≥ 0.
Proof Let f ∈ D(A1 ) and t > 0. Define u(s) = S2 (s)S1 (t − s) f for s ∈ [0, t]. The
function s → u(s) is differentiable at every s ∈ (0, t) with the derivative
d
u (s) = (S2 (s)S1 (t − s) f ) = S2 (s)A2 S1 (t − s) f + S2 (s)(−A1 S1 (t − s) f ).
ds
Remark 3.1 Suppose that (A, D(A)) is the generator of a strongly continuous semi-
group. If we let u(t) = S(t) f , t ≥ 0, for a fixed f ∈ D(A) then it follows from
(3.4) that u(t) ∈ D(A), the function t → u(t) is differentiable with a continuous
derivative, and u(t) is the solution of the initial value problem
Similar arguments as in the proof of Proposition 3.1 show that u is the only solution
of (3.5) with values in D(A). Consequently, for each f ∈ D(A) equation (3.5) has
one and only one classical solution and it is given by u(t) = S(t) f for t ≥ 0. For
f ∈ X the function t → S(t) f is only a mild solution of (3.5), i.e. it is continuous
and satisfies (3.3).
3.1 Generators and Semigroups 67
Remark 3.2 Observe that if a strongly continuous semigroup {S(t)}t≥0 satisfies (3.2)
then passing to the rescaled semigroup T (t) = e−λt S(t), t ≥ 0, where λ ≥ ω,
yields a bounded semigroup on (X , · ), i.e. T (t) ≤ M for all t ≥ 0. Moreover,
(A, D(A)) is the generator of {S(t)}t≥0 if and only if (A −λI, D(A)) is the generator
of {T (t)}t≥0 , since
1 1 1
lim (T (t) f − f ) − (S(t) f − f ) = lim (e−λt − 1)S(t) f = −λ f.
t↓0 t t t↓0 t
Let (A, D(A)) be a linear operator. We say that λ ∈ R belongs to the resolvent
set ρ(A) of A, if the operator λI − A : D(A) → X is invertible. The operator
R(λ, A) := (λI − A)−1 for λ ∈ ρ(A) is called the resolvent operator of A at λ.
Proposition 3.2 Suppose that μ ∈ ρ(A) and that |λ − μ| < R(μ, A)−1 . Then
λ ∈ ρ(A) and
∞
R(λ, A) = (λ − μ)n (R(μ, A))n+1 ,
n=0
which implies that R(λ, A), λ ∈ ρ(A), satisfies the resolvent identity
∞
(I − (λ − μ)R(μ, A))−1 = [(λ − μ)R(μ, A)]n ,
n=0
We now provide the integral representation of the resolvent operator of the gen-
erator of a strongly continuous semigroup. Let {S(t)}t≥0 be a strongly continuous
semigroup with generator (A, D(A)) and let ω ∈ R and M ≥ 1 be constants such
that (3.2) holds. For each λ > ω we consider
∞ r
R(λ) f = e−λt S(t) f dt := lim e−λt S(t) f dt, f ∈X. (3.6)
0 r →∞ 0
We will show that R(λ) is the resolvent operator of A at λ. Since the mapping
t → e−λt S(t) f is continuous and e−λt S(t) f ≤ Me−(λ−ω)t f for f ∈ X , we
see that R(λ) is well defined and that
∞ ∞
−λt M
R(λ) f ≤ e S(t) f dt ≤ Me−(λ−ω)t dt f = f .
0 0 λ−ω
Thus R(λ) is a bounded linear operator and it is the Laplace transform of the semi-
group. Observe that for h > 0 and f ∈ X we have
1 1 ∞ −λt
(S(h) − I )R(λ) f = e (S(h + t) f − S(t) f ) dt
h h 0
∞ h
1 1
= (eλh − 1) e−λt S(t) f dt − eλh e−λt S(t) f dt,
h 0 h 0
since (3.4) holds and the operator A is closed. Consequently, (ω, ∞) ⊂ ρ(A) and
R(λ) = R(λ, A) for λ > ω.
We close this section with a simple result which will be useful to identify gener-
ators through their resolvents.
3.1 Generators and Semigroups 69
Lemma 3.1 Let (A2 , D(A2 )) be an extension of the operator (A1 , D(A1 )). Assume
that 1 ∈ ρ(A1 ) and that the operator I − A2 is one to one. Then D(A2 ) = D(A1 )
and A2 = A1 .
Proof Let f ∈ D(A2 ) and consider g = f − A2 f . Since 1 ∈ ρ(A1 ), the operator
(I − A1 , D(A1 )) is invertible. Thus we can find h ∈ D(A1 ) such that g = h − A1 h.
We have A1 ⊆ A2 which implies that f − A2 f = h − A1 h = h − A2 h. Since the
operator I − A2 is one to one, we conclude that f = h, showing that D(A2 ) ⊆ D(A1 ).
is Cauchy, thus it converges and its limit is denoted by et A f . The family of operators
∞ n
t
S(t) = et A = An
n=0
n!
Moreover, we have
S(t) − I ≤ etA − 1,
lim S(t) − I = 0.
t↓0
Observe that
∞ n
t
S(t) f − f − t A f ≤ An f ≤ (etA − 1 − tA) f ,
n=2
n!
which implies that there is δ > 0 such that T (t) − I < 1/2 for 0 ≤ t < δ. Hence,
the operator T (t) is invertible for a sufficiently small t > 0. Since for any f ∈ X
we have
1
(S(t) f − f ) = AT (t) f,
t
Remark 3.3 Note that the use of L ∞ spaces is quite limited in the theory of operator
semigroups. It follows from the result of [66] that if the Banach space X is L ∞ =
L ∞ (X, Σ, m) and {S(t)}t≥0 is a strongly continuous semigroup on L ∞ , then the
generator of {S(t)}t≥0 is a bounded operator. Thus strongly continuous semigroups
on L ∞ are necessarily uniformly continuous.
We have
−tϕ(x)
S(t) f = |e f (x)| m(d x) ≤ | f (x)| m(d x) = f
X X
A f = −ϕ f, f ∈ L 1ϕ = { f ∈ L 1 : ϕ f ∈ L 1 }.
To this end denote by (A1 , D(A1 )) the generator of {S(t)}t≥0 . If f ∈ D(A1 ) then
1
lim (S(t) f − f ) = A1 f ∈ L 1 ,
t↓0 t
1
lim (e−tn ϕ(x) − 1) f (x) = A1 f (x)
n→∞ tn
Here we give an example showing that not every semigroup is strongly continuous.
Let X be either L 1 (R), the space of Lebesgue integrable functions on R, or B(R),
the space of bounded functions on R. Define the operator S(t) on X by
lim f (x − t) = f (x)
t↓0
for every f ∈ Cc (R). Since the set Cc (R) is a dense subset of L 1 (R), the semigroup
is strongly continuous on L 1 (R).
The generator of the translation semigroup on L 1 (R) is given by A f = − f with
domain
To see this let us denote by (A1 , D(A1 )) the generator of the translation semigroup
{S(t)}t≥0 . Recall that we have S(t) f (x) = f (x − t), x ∈ R, t ≥ 0, f ∈ L 1 (R). First
take f ∈ D(A1 ) so that
1
lim (S(t) f − f ) = g ∈ L 1 (R).
t↓0 t
Since
1
lim f (x) d x = f (b)
t↓0 t [b−t,b]
for a.e. a, b ∈ R. Thus f is absolutely continuous and its derivative being equal to
−g in the L 1 (R) space is integrable, which shows that the operator (A, D(A)) is an
extension of the generator (A1 , D(A1 )). We have 1 ∈ ρ(A1 ). Note that the general
solution of the differential equation f (x) + f (x) = 0 is of the form f (x) = ce−x
for x ∈ R, where c is a constant. Thus the operator I − A is one to one which implies
that A = A1 by Lemma 3.1.
Consider now the semigroup {S(t)}t≥0 on the space X = B(R). Then {S(t)}t≥0
is a semigroup of contractions, since
However, it is no longer strongly continuous on the space B(R). To see this take
f (x) = 1[0,1) (x), x ∈ R. We have
|S(t) f (x) − f (x)| = |1[t,1+t) (x) − 1[0,1) (x)| = |1[1,1+t) (x) − 1[0,t) (x)|
X0 = { f ∈ X : lim S(t) f − f = 0}
t↓0
resolvent set ρ(A) of A contains (0, ∞), and for every λ > 0
λR(λ, A) ≤ 1.
In that case, the semigroup {S(t)}t≥0 with generator (A, D(A)) is given by
∞
λn t n
S(t) f = lim e−λt (λR(λ, A))n f, f ∈ X , t ≥ 0. (3.7)
λ→∞ n!
n=0
For a proof see [34] or [35]. To check that an operator A with a dense domain is the
generator by using the Hille–Yosida theorem, we need to show that for each λ > 0
and g ∈ X there exists a unique solution f ∈ D(A) of
λ f − A f = g and λ f ≤ g.
Lemma 3.2 Suppose that (A, D(A)) is dissipative. Then for each λ > 0 the opera-
tor I − λ−1 A is one to one and its inverse (I − λ−1 A)−1 : R(λI − A) → D(A) is a
contraction, where R(λI − A) is the range of λI − A. Moreover, if R(λI − A) = X
for some λ > 0 then R(λI − A) = X for all λ > 0 and (0, ∞) ⊆ ρ(A).
Proof The linear operator λI − A is one to one for each λ > 0, since from 0 =
λ f − A f ≥ λ f , it follows that f = 0. For g ∈ R(I − λ−1 A) we let
f = (I − λ−1 A)−1 g which gives
g = (I − λ−1 A)(I − λ−1 A)−1 g ≥ (I − λ−1 A)−1 g.
|λ − λ0 | < λ0
3.3 Generators of Contraction Semigroups 75
belongs to ρ(A), which implies that (λI − A)(D(A)) = X for λ ∈ (0, 2λ0 ).
Repeating the argument we conclude that (λI − A)(D(A)) = X for all λ > 0.
We come to the key result, the Lumer–Phillips reformulation [67] of the Hille–
Yosida theorem.
Theorem 3.2 A linear operator (A, D(A)) on a Banach space X is the generator of
a contraction semigroup if and only if D(A) is dense in X , (A, D(A)) is dissipative,
and the range of the operator λI − A is X for some λ > 0.
To check that a densely defined operator A is the generator by using this refor-
mulation of the Hille–Yosida theorem, we need to show that A is dissipative and that
there is λ > 0 such that for each g ∈ X there exists a solution f ∈ D(A) of the
equation
λ f − A f = g.
α, f = f and α, A f ≤ 0. (3.9)
It is easy to prove the “if” part, since for each λ > 0 and f ∈ D(A) condition (3.9)
implies
λ f = α, λ f ≤ α, λ f − α, A f = α, λ f − A f ≤ αλ f − A f .
For the proof of the converse see, e.g. [34, Proposition II.3.23].
It follows from the Hahn–Banach theorem that the normalized duality set
J ( f ) = {α ∈ X ∗ : α ≤ 1 and α, f = f }
α, A f = α, P f − f = α, P f − α, f ≤ α f − f ≤ 0.
Theorem 3.3 A linear operator (A, D(A)) generates a strongly continuous semi-
group on X satisfying (3.2) with constants M ≥ 1 and ω ∈ R if and only if D(A)
is dense in X , the resolvent set ρ(A) of A contains (ω, ∞), and
An operator A with resolvent operator R(λ, A) satisfying (3.10) for some constants
M, ω is called a Hille–Yosida operator.
Remark 3.5 If an operator A is not densely defined, then we can still generate a
semigroup on a closed subspace of the Banach space X . In particular we have the
following; see [34, Chapter III]. If an operator (A, D(A)) is dissipative and the range
of the operator λI − A is X for some λ > 0, then the part A| of the operator A in
the subspace X0 = D(A), i.e.
G (A) = {( f, g) ∈ X × X : f ∈ D(A), g = A f }
and A f = g. It is easily seen that if (A, D(A)) is dissipative and with dense domain,
then it is closable and its closure (A, D(A)) is dissipative and the range of λI − A
is the closure of the range of λI − A. This gives an extension of the Hille–Yosida
theorem.
We now show how to check the range condition from the Lumer–Phillips theorem
using the notion of the adjoint operator. It is easy to see that if A is a densely defined
linear operator and λ ∈ ρ(A) then λ ∈ ρ(A∗ ) and R(λ, A)∗ = R(λ, A∗ ), where A∗
is the adjoint of A. Using the Lumer–Phillips theorem we now prove the following
3.3 Generators of Contraction Semigroups 77
Corollary 3.1 Let (A, D) be a dissipative operator with dense domain D. Then
the closure of (A, D) is the generator of a contraction semigroup if and only if the
operator λI ∗ − A∗ is one to one for some λ > 0, where I ∗ is the identity in X ∗ .
Proof Suppose that the closure of the set (λI − A)D is not equal to X . From the
Hahn–Banach theorem it follows that there must exist α ∈ X ∗ , α = 0, such that
α, λ f − A f = 0
for all f ∈ D. This implies that α ∈ D(A∗ ) and A∗ α = λα. Thus λI ∗ − A∗ is not
one to one.
In general, it is difficult to identify the whole domain of the generator. The follow-
ing concept is useful. If (A, D(A)) is a closed linear operator then a linear subspace
D of D(A) is called a core for A if the closure of the restriction of A to D, denoted
by A|D , is equal to A; in symbols, A|D = A. We give a sufficient condition for D
to be a core.
Theorem 3.5 Let {S(t)}t≥0 be a contraction semigroup with generator (A, D(A)).
Suppose that D0 ⊆ D ⊆ D(A) are subspaces of X such that S(t)(D0 ) ⊆ D for all
t > 0. Then D0 ⊆ R(λI − A|D ) for all λ > 0. In particular, if D0 is dense in X
then D is a core for (A, D(A)).
Proof Let g ∈ D0 and μ > 0. The operator (μA, D(A)) is the generator of the
rescaled semigroup {S(μt)}t≥0 . From (3.6) it follows that we have
r
g = (I − μA)−1 (I − μA)g = lim e−t S(μt)(I − μA)g dt
r →∞ 0
r − jr/n
n−1
= lim lim e S(μjr/n)(I − μA)g,
r →∞ n→∞ n
j=0
which implies the first claim, since S(μt)(I − μA)g = (I − μA)S(μt)g ∈ R(I −
μA|D ) for any t ≥ 0. The second claim follows from the Lumer–Phillips theorem
and Lemma 3.1.
Finally, the following result gives the uniqueness of the semigroup generated by
an extension of a given operator.
Proof Let A1 be a closed extension of (A, D). Since the closure A of (A, D) is the
smallest closed extension of (A, D), we have A ⊆ A1 . If A and A1 are generators
of strongly continuous semigroups, then the semigroups are equal. If D is not a core
for the generator of {S(t)}t≥0 , then one can find an infinite number of extensions of
A which are generators (see [5, A-II, Theorem 1.33]).
An important tool for the construction of a semigroup is perturbation theory; see [34,
Chapter III]. Here we only give the Phillips perturbation theorem which concerns
bounded perturbations and the variation of parameters formula. First, we need the
following result.
Lemma 3.3 Assume that (A, D(A)) and (B, D(B)) are linear operators such that
D(A) ⊆ D(B). Let λ ∈ ρ(A) and let B R(λ, A) be a bounded operator on X . Then
the operator (A + B, D(A)) satisfies
Proof Since λ ∈ ρ(A), we have R(R(λ, A)) = D(A) and (λI − A)R(λ, A) f = f
for f ∈ X , which gives
Theorem 3.7 (Phillips Perturbation) Let (A, D(A)) be the generator of a strongly
continuous semigroup and let B be a bounded operator. Then (A + B, D(A)) is the
generator of a strongly continuous semigroup.
Proof We can assume that (A, D(A)) is the generator of a contraction semigroup.
Then for each λ > 0 the operator B R(λ, A) is bounded and B R(λ, A) ≤ B/λ.
If we take λ > B, so that 1 ∈ ρ(B R(λ, A)), then it follows from Lemma 3.3 that
λ ∈ ρ(A + B) and
3.3 Generators of Contraction Semigroups 79
1 1 1
R(λ, A + B) ≤ = .
λ 1 − B/λ λ − B
for any f ∈ D(A) and t ≥ 0. To see this observe that for f ∈ D(A) the function
u(s) = P(t − s)S(s) f is continuously differentiable with derivative
since S(s) f ∈ D(A) and C S(s) f = AS(s) f + B S(s) f . This implies that
t t
P(t) f − S(t) f = u(0) − u(t) = − u (s)ds = P(t − s)B S(s) f ds.
0 0
If the operator B is bounded then the semigroup {P(t)}t≥0 is also given by the
Dyson–Phillips expansion
∞
P(t) f = Sn (t) f, (3.13)
n=0
We close this section by giving the Phillips perturbation theorem for perturbations
of contraction semigroups.
80 3 Operator Semigroups
Corollary 3.2 Let P̄ be a contraction on X and let (A0 , D(A0 )) be the generator
of a contraction semigroup {P0 (t)}t≥0 . If λ > 0 is a constant then the operator
(A0 − λI + λ P̄, D(A0 )) is the generator of a contraction semigroup {P(t)}t≥0 and
∞
P(t) = e−λt λn Sn (t) f, f ∈ L 1 , t ≥ 0, (3.15)
n=0
We denote by (A0 , D(A0 )) the restriction of the operator (A, D) to the zero boundary
condition
A0 f = A f, f ∈ D(A0 ) = { f ∈ D : Ψ0 ( f ) = 0}. (3.17)
We now describe how we can rewrite the operator (A, D(A)) using perturbations.
We define two operators A , B : D(A ) → X × ∂X with D(A ) = D × {0} by
We have
(A + B)( f, 0) = (A f, Ψ f − Ψ0 f ), f ∈ D.
Using ideas of Greiner [44] we are able to compute the resolvent operator of the
operator (A, D(A)) as stated next.
Lemma 3.4 Let (A0 , D(A0 )) be as in (3.17) and let λ ∈ ρ(A0 ). Assume that the
operator Ψ0 : D → ∂X restricted to the nullspace N (λI − A) = { f ∈ D : λ f −
A f = 0} is invertible with bounded inverse Ψ (λ) : ∂X → N (λI − A). For the
operator (A, D(A)) with D(A) as in (3.16) we have λ ∈ ρ(A) if and only if I∂X −
Ψ Ψ (λ) is invertible, where I∂X is the identity operator on ∂X . In that case, if
Ψ R(λ, A0 ) and Ψ Ψ (λ) are bounded then the resolvent operator of A at λ is given
by
Thus,
Remark 3.6 It is assumed in [44] that the operators (A, D) and (Ψ0 , D) are closed,
the operator Ψ is bounded, the range of the operator Ψ0 is equal to ∂X , and that
there are constants γ > 0 and λ0 ∈ R such that
These conditions imply that λΨ (λ) ≤ γ −1 for all λ ∈ ρ(A0 ) and that one can
apply the Hille–Yosida theorem to show that the operator A generates a strongly
continuous semigroup. In fact, operators I − Ψ (λ)Ψ and I − Ψ Ψ (λ) are invertible
and (I − Ψ (λ)Ψ )−1 = I + Ψ (λ)(I − Ψ Ψ (λ))Ψ . In Sect. 4.1.5 we shall provide an
extension of Grainer’s result to unbounded perturbations in L 1 space.
Chapter 4
Stochastic Semigroups
In this chapter, we assume that the Banach space X is L 1 = L 1 (X, Σ, m), where
(X, Σ, m) is a σ -finite measure space, with the norm
f = | f (x)| m(d x), f ∈ L 1.
X
We define the positive cone L 1+ to be the set of f ∈ L 1 which are positive f ≥ 0. A lin-
ear operator A : D(A) → L 1 is said to be positive if A f ≥ 0 for f ∈ D(A)+ , where
D(A)+ = D(A) ∩ L 1+ , and we write A ≥ 0. A positive and everywhere defined oper-
ator is a bounded operator and its norm is determined through values on the positive
cone, as we show next.
(A f )+ ≤ f + , f ∈ L 1,
A = sup A f .
f ≥0, f =1
∞
fn
f =
n=1
n2
n 3 ≤ A f n ≤ n 2 A f ,
Let us outline the proof. Note that if A is dispersive and f ∈ D(A) then − f ∈ D(A)
and, for any λ > 0,
The operator A is dispersive if and only if for each f ∈ D(A) there exists g ∈ L ∞ ,
g ≥ 0, with g||∞ ≤ 1 such that
g, f = f + and g, A f ≤ 0.
Remark 4.1 The Phillips theorem is valid in Banach lattices; see [5]. It can be used
to generate a positive contraction semigroup on a closed subspace X of the space
B(X ) of bounded measurable functions with the supremum norm · u . In particular,
suppose that the operator (A, D(A)) satisfies the positive maximum principle of
the form: for each f ∈ D(A) such that f + u > 0 there exists x0 ∈ X satisfying
f (x0 ) = f + u and A f (x0 ) ≤ 0. Then the operator A is dispersive, since
Thus the resolvent operator of A at each λ > 0 is positive. Following Arendt [4], we
call a linear operator A resolvent positive if there exists ω ∈ R such that (ω, ∞) ⊆
ρ(A) and R(λ, A) ≥ 0 for all λ > ω. The spectral bound of A is defined as
s(A) = inf{ω ∈ R : (ω, ∞) ⊆ ρ(A) and R(λ, A) ≥ 0 for all λ > ω}.
R(λ, A) ≥ R(μ, A) ≥ 0.
implies that R(λ, A) ≥ R(μ, A) if μ > λ > s(A). Suppose now that s(A) > −∞
and that μ = s(A) ∈ ρ(A). Then R(μ, A) ≥ 0 and for λ < μ such that (μ −
λ)R(μ, A) < 1 we see, by Proposition 3.2, that λ ∈ ρ(A) and
4.1 Aspects of Positivity 87
∞
R(λ, A) = (μ − λ)n R(μ, A)n+1 ≥ 0,
n=0
The following unpublished perturbation result [30] for generators of positive semi-
groups is only valid on an L 1 space. For its proof we refer the reader to the framework
of Miyadera–Voigt perturbation theory (see [10] and [117]).
Theorem 4.3 (Desch) Suppose that (A, D(A)) is the generator of a positive semi-
group on L 1 and that B with D(B) ⊇ D(A) is a positive operator. If (A + B, D(A))
is resolvent positive, then (A + B, D(A)) is the generator of a positive semigroup.
Theorem 4.4 A linear operator (A, D(A)) is the generator of a substochastic semi-
group on L 1 if and only if D(A) is dense in L 1 , the operator A is resolvent positive,
and
A f (x) m(d x) ≤ 0 for all f ∈ D(A)+ . (4.2)
X
Proof Suppose first that (A, D(A)) is the generator of a substochastic semigroup
{S(t)}t≥0 . We know that (A, D(A)) is resolvent positive. We also have for any f ∈
D(A)+
1
A f (x) m(d x) = lim (S(t) f (x) − f (x)) m(d x)
X t→∞ t X
1
= lim (S(t) f − f ) ≤ 0.
t→∞ t
To show that (A, D(A)) is the generator we make use of the Hille–Yosida theorem.
Since A is resolvent positive, A is closed and the resolvent R(λ, A) is a positive
88 4 Stochastic Semigroups
operator for all λ > s(A), by Lemma 4.1. Thus, it remains to check that s(A) ≤ 0
and that λR(λ, A) is a contraction for all λ > 0. Take f ∈ D(A)+ . Then, for any
λ > 0, we have
(λ − A) f ≥ (λ − A) f (x) m(d x) = λ f − A f (x) m(d x) ≥ λ f .
X X
and λR(λ, A) ≤ 1. Suppose now that s(A) > 0. Let λn > s(A) and λn ↓ s(A).
We have
1 1
R(λn , A) ≤ ≤ , n ≥ 0.
λn s(A)
Since s(A) > −∞, we have s(A) ∈ / ρ(A). Thus R(λn , A) → ∞, which gives a
contradiction and completes the proof.
Corollary 4.1 A linear operator (A, D(A)) is the generator of a stochastic semi-
group on L 1 if and only if D(A) is dense in L 1 , the operator A is resolvent positive,
and
A f (x) m(d x) = 0 for all f ∈ D(A)+ .
X
In this section, we provide sufficient conditions for the sum of two operators to be the
generator of a substochastic semigroup. We assume that (A, D(A)) is the generator
of a substochastic semigroup and that (B, D(B)) is an operator with D(B) ⊇ D(A).
4.1 Aspects of Positivity 89
We now suppose that (A, D(A)) is resolvent positive and (B, D(A)) is a positive
operator. Then R(λ, A) f ≥ 0 for every λ > s(A) and B R(λ, A) f ≥ 0 for f ∈ L 1+
which implies that the positive operator B R(λ, A) being defined everywhere is
bounded. If, additionally, (4.3) holds then the operator B R(λ, A) is substochastic
for every λ > 0. To see this take λ > 0 and observe that
for f ∈ L 1 . Thus
showing that B R(λ, A) ≤ 1. If there exists λ > 0 such that B R(λ, A) < 1 then
(A + B, D(A)) generates a substochastic semigroup, as we show next. The case
when B R(λ, A) = 1 for all λ > 0 is discussed in the next section.
Theorem 4.5 Let (A, D(A)) be resolvent positive and (B, D(A)) be a positive
operator such that (4.3) holds. If B R(λ, A) < 1 for some λ > 0 then
∞
R(λ, A + B) = R(λ, A) (B R(λ, A))n
n=0
and is a positive operator. The formula for R(λ, A + B) follows from Lemma 3.3.
Thus the operator (A + B, D(A)) is resolvent positive, which together with Theo-
rem 4.4 completes the proof.
We show that positive bounded perturbations of generators of substochastic semi-
groups lead to positive semigroups.
90 4 Stochastic Semigroups
Corollary 4.2 Suppose that (A, D(A)) is the generator of a substochastic semi-
group and that B with D(B) = L 1 is a positive operator. Then (A + B, D(A)) is
the generator of a positive semigroup.
Proof Condition (2) implies that the operator I − Ψ Ψ (λ) is invertible for λ > c.
Since (A0 , D(A0 )) is resolvent positive, the operators Ψ and Ψ (λ) are positive and
the operator I − Ψ Ψ (λ) is invertible for all sufficiently large λ > 0, Lemma 3.4
implies that the operator (A, D(A)) is resolvent positive. Consequently, we have
B R(λ, A) < 1 for some λ > 0 and the result follows from Theorem 4.5.
Theorem 4.7 Suppose that (A0 , D(A0 )) is the generator of a stochastic semigroup
on L 1 = L 1 (X, Σ, m) and that P is a stochastic operator on L 1 . Then for any
bounded measurable function ϕ : X → [0, ∞) the operator
A1 f = A0 f − ϕ f, f ∈ D(A0 ),
A f = A1 f + P(ϕ f ), f ∈ D(A0 ),
A1 f = A0 f − ϕ f = A0 f − λ f + (λ − ϕ) f for f ∈ D(A0 ).
for any f ∈ D(A0 )+ . Since the operator f → (λ − ϕ) f is positive and bounded, the
operator (A1 , D(A0 )) generates a substochastic semigroup, by Theorem 4.5. Finally,
the operator B f = P(ϕ f ) for f ∈ L 1 is bounded and positive. Since P preserves
the integral on the cone L 1+ , we have
B f (x) m(d x) = ϕ(x) f (x) m(d x), f ∈ L 1+ .
X X
Consequently, the operator (A1 + B, D(A0 )) is resolvent positive and is the generator
of a stochastic semigroup, by Corollary 4.1.
Remark 4.4 Note that if A0 generates a substochastic semigroup and P is a substo-
chastic operator then A as defined in Theorem 4.7 generates a substochastic semi-
group.
In this section we describe a generation result of Kato [56] (see also [9, 10, 116])
ensuring that some extension of (A + B, D(A)) is the generator of a substochastic
semigroup when the operator B is positive and unbounded.
Theorem 4.8 Assume that (A, D(A)) is the generator of a substochastic semigroup
on L 1 and that (B, D(A)) is a positive operator such that (4.3) holds. There exists
an extension (C, D(C)) of the operator (A + B, D(A)) which is the generator of a
substochastic semigroup {P(t)}t≥0 . The resolvent operator of C is
N
R(λ, C) f = lim R(λ, A) (B R(λ, A))n f, f ∈ L 1 , λ > 0. (4.6)
N →∞
n=0
92 4 Stochastic Semigroups
The semigroup {P(t)}t≥0 from Theorem 4.8 is called the minimal semigroup
related to A + B. To justify the name we show that the semigroup {P(t)}t≥0 is the
smallest substochastic semigroup whose generator is an extension of (A + B, D(A)),
t≥0 with generator (C,
i.e. for any positive semigroup { P(t)} D(C))
being an exten-
f ≥ P(t) f for all f ∈ L 1+ , t > 0.
sion of the operator (A + B, D(A)) we have P(t)
To this end for each r ∈ (0, 1) consider the operator Ar = A + r B with domain
D(A). We have r B R(λ, A) ≤ r < 1 and the operator (Ar , D(A)) is resolvent
positive with
∞
R(λ, A + r B) f = R(λ, A) (r B R(λ, A))n f, f ∈ L 1 , λ > 0.
n=0
f = (1 − r )B f
(λI − A − r B) f − (λI − C)
which leads to
N
f ≥ R(λ, A + r B) f ≥ R(λ, A)
R(λ, C) (r B R(λ, A))n f
n=0
f ≥ P(t) f , by (3.7).
which implies that P(t)
The minimal semigroup {P(t)}t≥0 can be defined by
where, for each r ∈ (0, 1), {Pr (t)}t≥0 is a substochastic semigroup generated by
(A + r B, D(A)). The semigroup {P(t)}t≥0 satisfies the integral Eq. (3.12) and it is
also given by the Dyson–Phillips expansion (3.13), where {S(t)}t≥0 is the semigroup
generated by (A, D(A)). Theorem 4.8 and the Lumer–Phillips theorem imply that the
closure of the operator (A + B, D(A)) is the generator of a substochastic semigroup
if and only if the range of λI − (A + B) is dense in L 1 for some λ > 0. Equivalently,
by Corollary 3.1, the operator λI ∗ − (A + B)∗ is one to one. By Theorem 3.6, what
we are missing is the characterisation of a core of the generator of the minimal
semigroup. Necessary and sufficient conditions for the generator of the minimal
4.1 Aspects of Positivity 93
B f = P(ϕ f ) for f ∈ L 1ϕ = { f ∈ L 1 : ϕ f ∈ L 1 },
where Px is the distribution of ξ(t) starting at x, then P(t) f is the density of ξ(t) if
the distribution of ξ(0) has a density f .
In this section, we relate a stochastic semigroup {P(t)}t≥0 to the transition
semigroup {T (t)}t≥0 on B(X ) associated to a homogeneous Markov process ξ =
{ξ(t) : t ≥ 0} with transition function P = {P(t, ·) : t ≥ 0} as described in Sect. 2.3.4.
The transition semigroup {T (t)}t≥0 on the Banach space B(X ) with supremum norm
· u associated to the process ξ is given by
T (t)g(x) = Ex (g(ξ(t)) = g(y)P(t, x, dy), g ∈ B(X ).
X
In particular, if
T (t)g(x) f (x) m(d x) = g(x)P(t) f (x) m(d x) (4.10)
X X
holds for all f ∈ L 1+ and all g ∈ B(X ), then (4.9) holds and {P(t)}t≥0 corresponds
to the transition function P.
The transition semigroup {T (t)}t≥0 on B(X ) is strongly continuous on the closed
subspace of B(X )
where · ∞ denotes the essential supremum norm in L ∞ . It is called the sun dual
(or the semigroup dual) space. The semigroup given by the part of P ∗ (t) in L , i.e.
is called the sun dual semigroup. It is strongly continuous, thus a contraction semi-
group on L . Let (A, D(A)) be the generator of the semigroup {P(t)}t≥0 and let
(A∗ , D(A∗ )) be its adjoint. It is easy to see that D(A∗ ) ⊂ L and that A∗ is an exten-
sion of the generator A of the sun dual semigroup {P (t)}t≥0 . In fact, (A , D(A ))
is the part of (A∗ , D(A∗ )) in L , i.e.
Consequently, if Eq. (4.10) holds for all g ∈ B0 (X ) and f ∈ L 1 then {T (t)}t≥0 is the
part of the sun dual semigroup {P (t)}t≥0 in B0 (X ) and we have L ⊆ A ⊆ A∗ .
We would like to obtain (4.10) using generators. We now assume that there exists
a closed subspace X of B0 (X ) such that T (t)g ∈ X for g ∈ X and that D L is a
core for the part of the generator (L , D(L)) of {T (t)}t≥0 in X . We suppose that the
following condition holds
Lg(x) f (x) m(d x) = g(x)A f (x) m(d x) (4.11)
X X
for g ∈ D L and f ∈ D A , where D A is a core for the generator (A, D(A)) of the
stochastic semigroup {P(t)}t≥0 on L 1 . By approximations, condition (4.11) extends
to all functions from the domains of the generators. It is easy to see that the function
v(s) = T (s)g(x)P(t − s) f (x) m(d x), s ∈ [0, t], g ∈ D(L), f ∈ D(A),
X
is continuous on [0, t] and differentiable with v (s) = 0 for s ∈ (0, t). This implies
that (4.10) holds for g ∈ D(L) and f ∈ D(A). Since domains of the generators are
dense in the corresponding spaces, we conclude that P ∗ (t)g = T (t)g for g ∈ X .
If X is large enough so that we can extend equality (4.10) to all g being indicator
functions, then {P(t)}t≥0 corresponds to the transition function P. In particular, if X
is a metric space and Σ = B(X ), then, by the uniqueness theorem A.2, if condition
96 4 Stochastic Semigroups
(4.9) holds for all closed subsets B, then it holds for all Borel sets. Since each indicator
function of a closed set can be approximated by a sequence of (globally) Lipschitz
continuous functions, it is enough to check (4.10) for uniformly continuous functions
g : X → R, or only for Lipschitz continuous functions.
To check condition (4.11) we can use the extended generator ( L, D(L)), which
for a class of PDMPs is characterized at the end of Sect. 2.3.6. The next result may
be useful in showing that L is a generator. The notion of resolvent positive operators
as introduced in Sect. 4.1.3 can also be used for operators defined on X ⊂ B(X )
with the supremum norm.
Proof The assumption 1 X ∈ D(L) implies that the operator L is cofinal, i.e. for each
non-negative g ∈ X there exists a non-negative h ∈ D(L) such that g(x) ≤ h(x) for
all x ∈ X . As in the proof of [4, Theorem 2.1] we conclude that L is a Hille–Yosida
operator. Consequently, L generates a positive semigroup on the closure of D(L)
(see Remark 3.5).
In Sect. 1.3 we considered a pure jump-type Markov process. This is a Markov process
on the space (X, Σ, m) which remains constant between jumps; see Sect. 2.3.3. In
the case when the state space X is at most a countable set, we obtain continuous-time
Markov chains. We assume that a measurable function ϕ : X → [0, ∞) is a jump
rate function. Let a transition probability function P(x, B) describe the new position
of the point after a jump from x. We assume that P(x, B) satisfies (2.4) and that P
is the stochastic operator on L 1 (X, Σ, m) corresponding to P(x, B).
Suppose first that ϕ is bounded. We now show that if the random variable ξ(t) is
the position of the moving point at time t and ξ(0) has density u(0), then ξ(t) has
density u(t) which satisfies the evolution equation
The stochastic semigroup {P(t)}t≥0 defined by P(t) f = u(t) with f = u(0) corre-
sponds to the transition function P(t, x, B), induced by the process ξ = {ξ(t) : t ≥ 0}
(see Sect. 4.1.7) and given by (2.21), since
P̄ f (x) m(d x) =
k
P̄ k (x, B) f (x) m(d x)
B X
A f = −ϕ f + P(ϕ f ) for f ∈ L 1ϕ = { f ∈ L 1 : ϕ f ∈ L 1 }.
ϕ(P ∗ g − g) = λg
(i) i j ≥ 0 for i = j,
q
(ii) i∈I qi j = 0 for j ∈ I ,
and it is called a sub-Kolmogorov matrix if it satisfies condition (i) and the condition
(ii ) i∈I qi j ≤ 0 for j ∈ I .
qi j
pi j = if ϕ j > 0 and pi j = 1 if ϕ j = 0, i = j.
ϕj
We have
(Qu)i = qii u i + qi j u j = −ϕi u i + P(ϕu)i .
j=i
If Q is a sub-Kolmogorov matrix then condition (4.3) holds and the minimal substo-
chastic semigroup {P(t)}t≥0 related to Q is generated by an extension of the operator
Q. Note that the matrix Q is a Kolmogorov matrix if and only if
ϕj = qi j , j ∈ I,
i= j
In this case the solution of (4.12) is given by (4.13). If ϕ is unbounded then we have
the following result.
Theorem 4.11 (Kato) Let Q be a Kolmogorov matrix and let λ > 0 be a posi-
tive constant. We denote Q ∗ = (qi∗j )i, j∈I , where qi∗j = q ji for i, j ∈ I . The minimal
semigroup related to Q is a stochastic semigroup on l 1 if and only if the equation
Q ∗ w = λw has no nonzero solution w ∈ l ∞ and w ≥ 0.
We assume that for each t > 0 the transformation πt is non-singular with respect
to m. Let P0 (t) be the Frobenius–Perron operator on L 1 corresponding to πt (see
Sect. 2.1.5). It satisfies
1 B (πt (x)) f (x) m(d x) = P0 (t) f (x) m(d x), B ∈ Σ, f ∈ L 1 ,
X B
for all measurable and non-negative g, as well as for all g ∈ L ∞ and f ∈ L 1 . Since
{πt }t≥0 is a semigroup on X , it follows from (4.14) that P0 (0) = I and P0 (t + s) =
P0 (t)P0 (s) for all s, t ≥ 0. Thus the family {P0 (t)}t≥0 is a semigroup on L 1 .
We give sufficient conditions for the semigroup of Frobenius–Perron operators to
be a stochastic semigroup. The proof of the following is based on [45].
Theorem 4.12 Assume that L 1 is separable. If the semigroup {P0 (t)}t≥0 of Frobe-
nius–Perron operators corresponding to {πt }t≥0 satisfies P0 (t)| f | = |P0 (t) f | for all
f ∈ L 1 and t > 0, then {P0 (t)}t≥0 is a stochastic semigroup.
Proof We first prove that for any f ∈ L 1 the function t → P0 (t) f is continuous at
each s > 0. If g ∈ L ∞
+ and f ∈ L + then the function (t, x) → g(π(t, x)) f (x) is
1
is Borel measurable, by Fubini’s theorem. This together with (4.14) implies that for
any f ∈ L 1 the function
t → g, P0 (t) f = g(x)P0 (t) f (x) m(d x)
X
for any t, s > 0 and f ∈ L 1 . Since π(s, x) ∈ X for all x ∈ X and s > 0, we have
P0 (t) f − f = 1 X (π(s, x))|P0 (t) f (x) − f (x)| m(d x),
X
We now assume that X is a separable metric space and that Σ = B(X ) is the
σ -algebra of Borel subsets of X . In that case the space L 1 = L 1 (X, B(X ), m) is
separable for any σ -finite measure m on (X, B(X )). If π : [0, ∞) × X → X is a
semiflow as in Sect. 1.19, then the mapping (t, x) → π(t, x) is measurable. More-
over, if πt is non-singular and one to one then the Frobenius–Perron operator P0 (t)
for the transformation πt is of the form given in Sect. 2.1.5 and satisfies the condition
|P0 (t) f | = P0 (t)| f | for f ∈ L 1 .
Recall that the survival function t → Φx (t) is right-continuous with Φx (0) = 1
and that for each t ≥ 0 the mapping x → Φx (t) is measurable. Define the operators
S(t), t ≥ 0, on L 1 by
S(t) f = P0 (t)(Φ(t) f ), (4.15)
for t ≥ 0, f ∈ L 1 , where we write Φ(t) for the function x → Φx (t). From (4.14) it
follows that
Φx (t)1 B (πt (x)) f (x) m(d x) = S(t) f (x) m(d x) (4.16)
X B
for all t ≥ 0, f ∈ L 1+ , B ∈ Σ.
Theorem 4.13 Suppose that {S(t)}t≥0 is defined as in (4.15) and satisfies (4.16). If
{P0 (t)}t≥0 is a substochastic semigroup then {S(t)}t≥0 is a substochastic semigroup.
0 ≤ S(t) f ≤ P0 (t) f
for f ∈ L 1+ . Thus, each operator S(t) is substochastic. To show the semigroup prop-
erty, we use the equation
4.2 Stochastic Semigroups for PDMPs 101
This and (4.14) imply that for any measurable set B and any f ∈ L 1+ we have
P0 (s)(Φ(t + s) f )(x) m(d x) = 1 B (πs (x))Φπs (x) (t)Φx (s) f (x) m(d x)
B
X
= 1 B (x)Φx (t)P0 (s)(Φ(s) f )(x) m(d x).
X
for f ∈ L 1 ( X̃ , Σ̃, m̃) and g ∈ L ∞ ( X̃ , Σ̃, m̃), then the restriction of { P̃0 (t)}t≥0 to
X is a substochastic semigroup {P0 (t)}t≥0 on L 1 = L 1 (X, Σ, m). In particular, we
have
1[0,t∗ (x)) (t)g(πt (x)) f (x) m(d x) = g(x)P0 (t) f (x) m(d x), f ∈ L 1, g ∈ L ∞,
X X
where t∗ (x) is the exit time from the set X as defined in (1.39). If the survival function
Φ is given by (1.41) then the semigroup {S(t)}t≥0 satisfies (4.16).
In this section we show how we can define stochastic semigroups for a large class
of PDMPs. Let (X, Σ, m) be a σ -finite measure space, where X is a separable
metric space and Σ = B(X ) is the σ -algebra of Borel subsets of X . The character-
istics (π, Φ, P) define a PDMP process ξ = {ξ(t) : t ≥ 0} with state space X , where
102 4 Stochastic Semigroups
where ϕ is a bounded measurable function such that for each x ∈ X , ϕ(πt (x)) →
ϕ(x) as t → 0.
Theorem 4.14 Let {P0 (t)}t≥0 be a stochastic semigroup of Frobenius–Perron oper-
ators corresponding to {πt }t≥0 . Suppose that {S(t)}t≥0 is a substochastic semigroup
as defined in (4.15) with Φx (t) as in (4.17). Then the generator of {S(t)}t≥0 is given
by
A1 f = A0 f − ϕ f for f ∈ D(A0 ), (4.18)
1
lim (Φ(t) − 1) f + ϕ f = 0.
t↓0 t
Since {P0 (t)}t≥0 is a stochastic semigroup with generator A0 , we conclude that the
generator of the semigroup {S(t)}t≥0 is an extension of the operator (A1 , D(A0 )).
Let the transition probability P correspond to a stochastic operator P(x, B) on
L 1 = L 1 (X, Σ, m). It follows from Theorem 4.7 that the operator
A f = A0 f − ϕ f + P(ϕ f ), f ∈ D(A0 ),
∞
t
P(t) f = Sn (t) f, where Sn (t) f = Sn−1 (t − s)P(ϕ S(s) f ) ds (4.19)
n=0 0
for every n ≥ 1, and S0 (t) = S(t) with {S(t)}t≥0 as in (4.15). We now show that the
semigroup {P(t)}t≥0 corresponds to the transition function induced by the PDMP
ξ as constructed in Sect. 2.3.5. To this end it is enough to show (4.9). Recall from
Sect. 2.3.6 that
where Tn (t) is defined in (2.33). Since T0 (t)1 B (x) = 1 B (π(t, x))Φx (t), x ∈ X ,
Eq. (4.16) implies that
T0 (t)1 B (x) f (x) m(d x) = 1 B (x)S(t) f (x) m(d x)
X X
on some Borel subset X ⊆ R d with non-empty interior and with boundary of the
Lebesgue measure zero. We assume that b is a locally Lipschitz function on X and that
for each x0 ∈ X Eq. (4.20) with initial condition x(0) = x0 has a solution x(t) ∈ X
for all t > 0 and denote it by πt (x0 ). We assume that a point moves according to this
flow, i.e. if x0 ∈ X is the initial position of a point, then x = πt (x0 ) is its location at
time t. The point can jump with intensity which is a bounded continuous function
ϕ : X → [0, ∞). Let a transition probability function P(x, B) describe the new
position of the point after a jump from x and let L 1 = L 1 (X, B(X ), m) where m is
the Lebesgue measure on X . We assume that P(x, B) satisfies (2.4) and defines a
stochastic operator P on L 1 , which corresponds to the transition probability P(x, B).
If the random variable ξ(t) is the position of the moving point at time t and ξ(0) has
density u(0), then ξ(t) has density u(t) which satisfies the evolution equation
d
∂
u (t) = Au(t) = A0 u(t) − ϕu(t) + P(ϕu(t)), A0 u(t) = − (bi u).
i=1
∂ xi
(4.21)
As in Sect. 4.2.1 we define
λ = sup{ϕ(x) : x ∈ X }, P̄ f (x) = λ−1 P(ϕ f )(x) + (λ − ϕ(x)) f (x) .
where the mapping x → π−t (x) is the inverse of the one to one mapping πt : X →
πt (X ). The semigroup {P(t)}t≥0 with the generator A can be given by the Dyson–
Phillips expansion (3.15) in Corollary 3.2.
We consider here the simplest case when the number of different dynamical systems is
finite. Precisely, we consider a family of systems of differential equations x = bi (x),
i ∈ I = {1, . . . , k}, defined on some open set G ⊆ Rd . We assume that all functions
bi : G → Rd are locally Lipschitz functions on G. Let X ⊆ G be a Borel set with
non-empty interior and with boundary of Lebesgue measure zero. Let x0 ∈ X and
t ≥ 0. We denote by πti (x0 ) the solution x(t) of the system x = bi (x) with the initial
condition x(0) = x0 provided that such a solution exists in the time interval [0, t]
and x(s) ∈ X for s ∈ [0, t]. The state of the system is a pair (x, i) ∈ X = X × I or
the dead state ∗ which is an extra point outside the space X.
4.2 Stochastic Semigroups for PDMPs 105
If the system is at state (x, i) then we can jump to the state (x, j) with a bounded
and continuous intensity q ji (x). Between two consecutive jumps the trajectory start-
ing from the state (x, i) is deterministic and at time t it reaches the state (πti (x), i)
or the dead state ∗ if πti (x) is not defined and then the system spends the rest of the
time in the dead state.
Consider the differential equation
d d
x(t) = bi(t) (x(t)), i(t) = 0. (4.22)
dt dt
Its solutions are solutions of
d
(x(t), i(t)) = b(x(t), i(t)), (4.23)
dt
where the mapping b defined by b(x, i) = (bi (x), 0) for (x, i) with x ∈ G is locally
Lipschitz continuous. We take the semiflow π : [0, ∞) × X → X ∪ {∗} to be
as long as πti (x) exists and belongs to X ; otherwise we set π(t, x, i) = ∗. The jump
rate function ϕ : X → [0, ∞) is given by
ϕ(x, i) = qi (x) = q ji (x),
j=i
and Φ(x,i) (t) = 0, otherwise. The jump distribution P is described at the end of
Sect. 2.1.6 with the probability of jump from a state (x, i) to the state (x, j) with
j = i defined by
1, qi (x) = 0, j = i,
p ji (x) = q ji (x)
qi (x)
, qi (x) > 0, j = i,
One can also give an analytic definition of the semigroup {P(t)}t≥0 by using a per-
turbation theorem and an analytic formula for a substochastic semigroup generated
by a single semiflow (see e.g. [96]).
In this section we show how the perturbation result from Sect. 3.3.4 can be used in
an example with only forced jumps. Consider the Lebowitz–Rubinow model from
Sect. 1.7. We take X = [0, 1) × (0, 1] and X̃ = R2 . The solution of (1.4) with initial
condition (x(0), v(0)) = (x, v) is of the form
We have
1−x
Γ = {1} × (0, 1], t∗ (x, v) = , x, v ∈ (0, 1].
v
The survival function is given by Φ(x,v) (t) = 1[0,t∗ (x,v)) (t) for (x, v) ∈ X and the
transition probability satisfies
P((1, v), {0} × B) = P(v, dv ) = αδv (B) + β k(v , v) dv , B ∈ B((0, 1]),
B B
∂g(x, v)
Lg(x, v) = v , (x, v) ∈ [0, 1] × (0, 1],
∂x
for all bounded functions which are continuously differentiable in the first variable,
supplemented with the boundary condition
1
g(1, v) = g(0, v )P(v, dv ). (4.24)
0
We show that the process ξ(t) = (x(t), v(t)) induces a stochastic semigroup
{P(t)}t≥0 on the space L 1 = L 1 (X, Σ, m), where Σ = B(X ), and m(d x × dv) =
d x × dv. The infinitesimal generator of this semigroup is of the form
∂f
A f (x, v) = −v (x, v) (4.25)
∂x
and the functions f from the domain of A satisfy the boundary condition
4.2 Stochastic Semigroups for PDMPs 107
1
v h(0, v ) = αv h(1, v ) + β vk(v , v)h(1, v) dv. (4.26)
0
∂f
A f (x, v) = −v (x, v), f ∈ D = { f ∈ L 1 : A f ∈ L 1 },
∂x
f A = f + A f , f ∈ D.
Observe that P preserves the integral on the positive cone so that we have
Ph(x, v) m 0 (d x, dv) = h(x, v) m 1 (d x, dv), h ∈ L 11 , h ≥ 0.
Γ0 Γ1
The operator
∂f
A0 f (x, v) = −v (x, v)
∂x
108 4 Stochastic Semigroups
with domain
D(A0 ) = { f ∈ D : γ0 f = 0}
For each t > 0 the operator S(t) is the restriction to X of the Frobenius–Perron
operator P0 (t) for the transformation πt : R2 → R2 given by πt (x, v) = (x + tv, v)
and defined on L 1 (R2 , B(R2 ), m).
For any λ > 0 the resolvent of (A, D(A)) at λ with
D(A) = { f ∈ D : γ0 f = Pγ1 f }
where C[0, 1] is the space of continuous functions on [0, 1]. Consider the operator
L on the maximal domain
0 , Ψ
We define the boundary operators Ψ : B(X ) → B(Γ ), where Γ = Γ1 , by
4.2 Stochastic Semigroups for PDMPs 109
1
0 g(1, v) = g(1, v), Ψ
Ψ g(1, v) = g(0, v )P(v, dv ), v ∈ (0, 1], g ∈ B(X ).
0
It is a positive operator and the operator I − ΨΨ(λ) is invertible for each λ > 0.
The operator L 0 being the restriction of 0 ) is resolvent positive, since
L to N (Ψ
1
λx λr
R(λ, L 0 )g(x, v) = e v e− v f (r, v) dr, g ∈ X .
x
is resolvent positive. Since the boundary condition (4.24) holds for g = 1 X and
L1 X = 0, we conclude from Theorem 4.10 that L generates a positive contraction
semigroup.
In this section we show how we can generate a stochastic semigroup for a particular
example of a PDMP described in Sect. 1.11 where we considered the Stein model. As
in [91] we introduce an additional 0-phase, which starts at the end of the refractory
period and lasts until depolarization takes place. Thus we consider three phases
0, 1, 2. Recall that in phase 1 the depolarization x(t) decays according to the equation
x (t) = −αx(t) with α > 0 and phase 2 denotes the refractory phase of duration t R
(see Fig. 4.1). We take X̃ = R × {0, 1, 2}. The flow π : R × X̃ → X̃ is given by
The process is defined on the space X = {(0, 0)} ∪ (−∞, θ ] × {1} ∪ [0, t R ) × {2}
and the active boundary is Γ = {(t R , 2)} with t∗ (x, 2) = t R − x for x ∈ [0, t R ). We
take
Φ(x,i) (t) = e−(λ E +λ I )t , i = 0, 1,
and Φ(x,2) (t) = 1[0,t∗ (x,2)) (t), so that the jump rate function ϕ is defined by ϕ(x, i) =
λ E + λ I for i = 0, 1 and ϕ(x, 2) = 0. The jump distribution P is given by
1-phase
pI pE pI pE
−xI 0 xE
0-phase (0, 0)
(tR , 2)
2-phase tR
(0, 2)
where
λI λE
pI = , pE = , S I (x, i) = (x − x I , 1),
λE + λI λE + λI
and
Lg(0, 0) = −(λ E + λ I )g(0, 0) + λ I g(−x I , 1) + λ E g(x E , 0),
∂g(x, 1)
Lg(x, 1) = −αx − (λ E + λ I )g(x, 1) + λ I g(x − x I , 1)
∂x
+ λ E g(x + x E , 1)1(−∞,θ] (x + x E ) + λ E g(0, 2)1(θ,∞) (x + x E ),
∂g(x, 2)
Lg(x, 2) = ,
∂x
for all bounded functions which are differentiable in the first variable, supplemented
with the boundary condition
[0, t R ] × {2} is the product of the Lebesgue measure on the interval [0, t R ] and the
Dirac measure at {2}. We define the operator P on L 1 = L 1 (X, B(X ), m) by
P f (x, 1) = p E f (x − x E , 1) + p I f (x + x I , 1)1(−∞,θ) (x + x I ),
and tR
A1 f (x, 2) d x = f (0+ , 2) − f (t R− , 2).
0
We take
Γ− = {(0, 2), (x E , 1), (−x I , 1)}, m − = δ(0,2) + αx E δ(x E ,1) + αx I δ(−x I ,1) ,
and
Γ+ = {(t R , 2), (x E , 1), (−x I , 1)}, m + = δ(t R ,2) + αx E δ(x E ,1) + αx I δ(−x I ,1) .
Note that Γ+ contains all points through which the flow π can exit the segments
(−∞, −x I ) × {1}, (−x I , x E ) × {1}, (x E , θ ) × {1}, (0, t R ) × {2}, in a finite positive
time, while the set Γ− contains points through which the flow can exit these segments
in a finite negative time. We define the traces γ− : D → L 1 (Γ− ) by
and γ+ : D → L 1 (Γ+ ) by
112 4 Stochastic Semigroups
The operator
A0 f = A1 f, f ∈ D(A0 ) = { f ∈ D : γ− f = 0},
and
⎧ r −1
⎪
⎨ f (x E , 1)
⎪ , x ∈ (0, x E )
x
xE
r −1
Ψ (λ) f (x, 1) = f (−x I , 1) x
, x ∈ (−x I , 0)
⎪
⎪ −x I
⎩
0, x ∈ (−∞, −x I ) ∪ (x E , θ ),
where
λI λE
a(−x I , 1) = , a(x E , 1) = .
αx I αx E
λE λE
0 ≤ Ψ Ψ (λ) f (0, 2) ≤ αx E f (x E , 1) ≤ f L 1 (Γ− ) ,
λ + λE + λI λ + λE + λI
which implies that λΨ Ψ (λ) ≤ λ E . From Theorem 4.6 it follows that the operator
(A1 , N (Ψ − Ψ0 )) is the generator of a substochastic semigroup.
The operator
with domain
4.2 Stochastic Semigroups for PDMPs 113
D(A) = { f ∈ D : Ψ0 ( f ) = Ψ ( f )}
since
θ θ
A1 f (0, 0) + A1 f (x, 1) d x = −(λ E + λ I ) f (x, 1) d x,
−∞ −∞
and tR
f (t R− , 2) + A1 f (x, 2) d x = λ E f (x, 1) d x.
0 θ−x E
It follows from Theorem 4.9 that (A, D(A)) is the generator of a stochastic semi-
group.
It remains to show that a restriction of the extended generator
L of the process ξ
generates a positive semigroup on a sufficiently large subspace of B(X ). We consider
the subspace X ⊂ B(X ) of all functions g such that g restricted to (−∞, θ − x E ) ×
{1} is uniformly continuous function on (−∞, θ − x E ], g restricted to (θ − x E , θ ] ×
{1} belongs to C[θ − x E , θ ] and g(·, 2) belongs to C[0, t R ]. We define
L 1 on the
domain
by
L 1 g(0, 0) = −(λ E + λ I )g(0, 0),
∂g(x, 1)
L 1 g(x, 1) = −αx − (λ E + λ I )g(x, 1),
∂x
∂g(x, 2)
L 1 g(x, 2) = .
∂x
and
x
1 −r x 1 ry
R(λ,
L 1 )g(x, 1) = er (θ−x E −x) g((θ − x E )− , 1) + e e g(y, 1) dy
α θ−x E y
Bg(x, i) = λ I g(x − x I , 1) + λ E g(x + x E , 1)1(−∞,θ−x E ] (x) + g(0, 2)1(θ−x E ,θ] (x)
Let the triple (X, Σ, m) be a σ-finite measure space. In the whole chapter, we will
consider substochastic operators and semigroups on L 1 = L 1 (X, Σ, m) correspond-
ing to transition kernels. The support of any measurable function f is defined up to
a set of measure zero by the formula
but since P ∗ (t)1 X ≤ 1 X and f ∗ > 0 a.e., we conclude that P ∗ (t)1 X = 1 X for all
t ≥ 0.
A substochastic semigroup {P(t)}t≥0 is called quasi-asymptotically stable if
there exist an invariant density f ∗ and a positive linear functional α defined on
L 1 (X, Σ, m) such that
lim P(t) f − α( f ) f ∗ = 0 for f ∈ L 1 . (5.1)
t→∞
We usually consider sweeping from a family of sets, e.g. from sets of finite measure
m or from all compact subsets of X if X is a topological space.
Sweeping is also called zero type or null property [79]. Asymptotic stability is also
called strong mixing or exactness, especially in the context of ergodic theory [62].
We can also introduce quasi-asymptotic stability, asymptotic stability and sweeping
of a substochastic operator P if in (5.1) and (5.2) we replace P(t) with the iterates
P t of the operator P.
There are several results concerning asymptotic stability [62]. Some of them are based
on assumptions concerning boundedness from below of a stochastic semigroup. One
of them is the lower function theorem of Lasota and Yorke [63]. A function h ∈ L 1 ,
h ≥ 0 and h = 0, is called a lower function for a stochastic semigroup {P(t)}t≥0 if
5.1 Asymptotic Stability and Sweeping 117
Here we use the notation f − (x) = 0 if f (x) ≥ 0 and f − (x) = − f (x) if f (x) < 0.
Condition (5.3) can be equivalently written as: there are functions ε(t) ∈ L 1 such
that limt→∞ ε(t) = 0 and P(t) f ≥ h − ε(t). Observe that if the semigroup is
asymptotically stable then its invariant density f ∗ is a lower function for it. Lasota
and Yorke [63] proved the following converse result.
Theorem 5.1 Let {P(t)}t≥0 be a stochastic semigroup. If there exists a lower func-
tion h for a stochastic semigroup {P(t)}t≥0 then this semigroup is asymptotically
stable.
for some t > 0. If equality holds in (5.4) then {P(t)}t≥0 is said to be an integral
semigroup. The following theorem is only valid for continuous-time semigroups.
Theorem 5.2 Let {P(t)}t≥0 be a partially integral stochastic semigroup. Assume
that the semigroup {P(t)}t≥0 has a unique invariant density f ∗ . If f ∗ > 0 a.e., then
the semigroup {P(t)}t≥0 is asymptotically stable.
The proof of Theorem 5.2 is given in [87] and it is based on the theory of Harris
operators [38, 54].
We now formulate corollaries which are often used in applications. A substochas-
tic semigroup {P(t)}t≥0 is called irreducible if
∞
P(t) f dt > 0 a.e. for every density f.
0
Corollary 5.1 A partially integral and irreducible stochastic semigroup which has
an invariant density is asymptotically stable.
Corollary 5.2 A partially integral stochastic semigroup which overlaps supports
and has an invariant density f ∗ > 0 a.e. is asymptotically stable.
Remark 5.1 The assumption in Theorem 5.2 that the invariant density is unique
can be replaced by an equivalent one: there does not exist a set E ∈ Σ such that
m(E) > 0, m(X \ E) > 0 and P(t)E = E for all t > 0. Here P(t) is the operator
acting on the σ-algebra Σ defined by: if f ≥ 0, supp f = A and supp P(t) f = B
then P(t)A = B. The definition of P(t)A does not depend on the choice of f
because if supp f = supp g a.e. then supp P(t) f = supp P(t)g a.e. Irreducibility
is equivalent to the following: the semigroup has no nontrivial invariant subset. We
recall that a measurable set A is invariant with respect to {P(t)}t≥0 if P(t)(A) ⊆ A
for t ≥ 0 (in terms of the transition kernel it means that P(t, x, X \ A) = 0 for
m-a.e. x ∈ A and t ≥ 0).
We now present results concerning the sweeping property which are based on an
assumption about the existence of a σ-finite positive invariant function. These were
proved in [58] and in [95] for discrete-time stochastic semigroups (iterates of a sto-
chastic operator) but similar results also hold for continuous-time stochastic semi-
groups. Before formulating some results we need to extend a substochastic oper-
ator beyond the space L 1 . If P is a substochastic operator and f is an arbitrary
non-negative measurable function, then we define P f as a pointwise limit of the
sequence P f n , where ( f n ) is any monotonic sequence of non-negative functions
from L 1 pointwise convergent to f almost everywhere (see [38, Chap. I]). Observe
that we can also define an extension of a substochastic operator using a transition
kernel. If P(x, B) is the transition kernel then for any non-negative measurable func-
tion f we have P f = dμ/dm, where μ(B) = X f (x)P(x, B) m(d x) and dμ/dm
is the Radon–Nikodym derivative.
The following condition plays a crucial role in results concerning sweeping:
∗ ∗ ∗
∗ ∗
∗ f such that: 0 < f < ∞ a.e., P(t) f ≤
(KT) There exists a measurable function
f for t ≥ 0, f ∈ / L and Z f (x) m(d x) < ∞ for some set Z ∈ Σ with
1
m(Z ) > 0.
Theorem 5.3 ([58]) Let {P(t)}t≥0 be an integral substochastic semigroup which
has no invariant density. Assume that the semigroup {P(t)}t≥0 and a set Z ∈ Σ
satisfy condition (KT). Then the semigroup {P(t)}t≥0 is sweeping with respect to Z .
Unfortunately, substochastic semigroups induced by PDMPs are usually not inte-
gral and we need versions of Theorem 5.3 without assumption (KT). The problem
5.1 Asymptotic Stability and Sweeping 119
is discussed in details in [95] and we present here only some conclusions. In order
to formulate the next theorem we need some auxiliary definitions.
We observe that if a substochastic semigroup {P(t)}t≥0 has a subinvariant function
f ∗ > 0 then we can define a new substochastic semigroup { P̄(t)}t≥0 on the space
L 1 (X, Σ, μ) with dμ = f ∗ dm given by
1
P̄(t) f = P(t)( f f ∗ ) (5.5)
f∗
and P̄(t)1 X ≤ 1 X . Thus, all operators P̄(t) and P̄ ∗ (t) are contractions on
L 1 (X, Σ, μ) and L ∞ (X, Σ, μ), respectively, and, consequently, they are positive
contractions on L 2 (X, Σ, μ). Let us define
and
Σ1 = {B ∈ Σ : 1 B ∈ K }.
The definition of C does not depend on the choice of f . The substochastic semigroup
{P(t)}t≥0 is called conservative if C = X a.e. and dissipative if C = ∅ a.e. From this
120 5 Asymptotic Properties of Stochastic Semigroups—General Results
definition it follows immediately that if {P(t)}t≥0 has an invariant density f ∗ > 0 a.e.,
then {P(t)}t≥0 is conservative. We also have P ∗ (t)1C ≥ 1C for t > 0 or, equivalently,
P(t, x, C) = 1 for x ∈ C. In particular, a conservative substochastic semigroup is
a stochastic semigroup. A substochastic semigroup can be neither conservative nor
dissipative, but if it is an irreducible semigroup then it has one of these properties.
If a stochastic
∞ semigroup is dissipative then for any positive density f the function
f ∗ (x) = 0 P(t) f (x) dt is a positive and subinvariant function. The function f ∗ is
also not integrable because in the opposite case the semigroup will be conservative.
If {P(t)}t≥0 is a conservative pre-Harris semigroup, then {P(t)}t≥0 has a strictly
positive invariant function. Even if we know that a semigroup has an subinvariant
function f ∗ , a problem still remains because we need to identify sets Z such that
∗
Z f (x) m(d x) < ∞. In order to do this we need to have more information about
f ∗ —not only its existence. We cannot apply directly the formula A f ∗ ≤ 0 to find f ∗
as in the case of a density because f ∗ ∈ / D(A), where A is the infinitesimal generator
of the semigroup and D(A) is its domain.
The main problem with applying results of Sects. 5.1.3 and 5.1.4 to study asymptotic
stability is that we need to have an invariant density and a positive subinvariant
function to check asymptotic stability and sweeping, respectively. Now we present
a general result concerning asymptotic decomposition of substochastic semigroups
and some corollaries which do not have any assumption on invariant functions. We
hope that these results will prove useful to study asymptotic stability and sweeping
of semigroups induced by PDMPs.
From now on, we assume that X is a separable metric space and Σ = B(X ) is the
σ-algebra of Borel subsets of X . We will consider partially integral substochastic
semigroups {P(t)}t≥0 which satisfy the following condition:
(K) For every y0 ∈ X there exist an ε > 0, a t > 0 and a measurable function
η ≥ 0 such that η(x) m(d x) > 0 and
from compact sets. Condition (K) is satisfied if, for example, for every point y ∈ X
there exist a t > 0 and an x ∈ X such that the kernel k(t, ·, ·) is continuous in a
neighbourhood of (x, y) and k(t, x, y) > 0.
The following theorem states that a substochastic semigroup can be decomposed
into asymptotically stable and sweeping components.
Theorem 5.5 Let {P(t)}t≥0 be a substochastic semigroup on L 1 (X, Σ, m), where
X is a separable metric space, Σ = B(X ), and m is a σ-finite measure. Assume
that {P(t)}t≥0 satisfies (K). Then there exist an at most countable set J , a family of
invariant densities { f j∗ } j∈J with disjoint supports {A j } j∈J , and a family {α j } j∈J of
positive linear functionals defined on L 1 (X, Σ, m) such that
(i) for every j ∈ J and for every f ∈ L 1 (X, Σ, m) we have
(ii) if Y = X \ A j , then for every f ∈ L 1 (X, Σ, m) and for every compact set
j∈J
F we have
lim P(t) f (x) m(d x) = 0. (5.8)
t→∞
F∩Y
Theorem 5.5 was proved in [89] for stochastic semigroups and generalized in [90]
for substochastic semigroups. Theorem 5.5 unifies a part of the theory of substochas-
tic semigroups related to asymptotic stability and sweeping and generalizes some
earlier results [85, 95].
Since f j∗ is an invariant density with supp f j∗ = A j , the restricted semigroup
{PA j (t)}t≥0 is a stochastic semigroup. Hence, condition (i) implies that {PA j (t)}t≥0 is
asymptotically stable and limt→∞ P(t) f − f j∗ = 0 for f ∈ D such that supp f ⊆
Aj.
Remark 5.2 The sets A j , j ∈ J , which occur in the formulation of Theorem 5.5,
are not only disjoint but also their closures are disjoint. Indeed, let m j be the Borel
measures on X given by
Theorem 5.5 has a number of corollaries concerning asymptotic stability and sweep-
ing. We start with a corollary concerning sweeping:
Corollary 5.3 Let {P(t)}t≥0 be a substochastic semigroup on L 1 (X, Σ, m), where
X is a separable metric space, Σ = B(X ), and m is a σ-finite measure. Assume
that {P(t)}t≥0 satisfies (K) and has no invariant density. Then {P(t)}t≥0 is sweeping
from compact sets.
Condition (K) and irreducibility allow us to express asymptotic properties in the
form of the Foguel alternative, i.e. the stochastic semigroup is asymptotically stable
or sweeping [17]. We use the notion of the Foguel alternative in a narrow sense, when
sweeping is from all compact sets.
Corollary 5.4 Let X be a separable metric space, Σ = B(X ) and m be a σ-finite
measure. If {P(t)}t≥0 is a substochastic semigroup on the space L 1 (X, Σ, m), it
satisfies (K) and it is irreducible, then {P(t)}t≥0 is asymptotically stable or sweeping
from compact sets.
If a substochastic semigroup {P(t)}t≥0 satisfies conditions (WI) and (K) then the set
J defined as in Theorem 5.5 is an empty set or a singleton.
(ii) if Y = X \ A, then for every f ∈ L 1 (X, Σ, m) and for every compact set F
we have
lim P(t) f (x) m(d x) = 0. (5.11)
t→∞
F∩Y
It is enough to check (5.14) for densities. Then the function β(t) = P(t) f (x) m(d x)
Y
is decreasing. Suppose that, contrary to our claim,
lim P(t) f 1 (x) m(d x) = c > 0 (5.15)
t→∞
Y
have
P(t0 ) f 1 (x) m(d x) < c + ε.
Y
for t ≥ 0. Condition (ii) of Corollary 5.5 implies that for each compact set F we
have
P(t) f 3 (x) m(d x) < κ/4 (5.17)
F∩Y
If we know that a stochastic semigroup satisfies the Foguel alternative then we can
eliminate sweeping by constructing a Hasminskiı̆ function [86]. Now we show how
a slight modification of this method can be applied to study weak tightness. It is
clear that a weakly tight stochastic semigroup is non-sweeping from compact sets.
We precede the definition of a Hasminskiı̆ function by some observations.
Let A be a generator of a substochastic semigroup {P(t)}t≥0 and let
∞
R := R(1, A) = (I − A)−1 = e−t P(t) dt (5.19)
0
t n e−t
where rn (t) = . Suppose that, on the contrary, there exist t0 > 0 and ε < ε
n!
such that
P(t) f (x) m(d x) ≤ ε for t ≥ t0 .
Z
∞ t0
Since 0 rn (t) dt = 1, 0 rn (t) dt → 0 as n → ∞, and
t0 ∞
R f (x) m(d x) ≤
n
rn (t) dt + rn (t) P(t) f (x) m(d x) dt,
Z 0 t0 Z
we obtain lim supn→∞ Z R n f (x) m(d x) ≤ ε , a contradiction.
Since m is a σ-finite measure, the set DV is dense in D. Assume that there exist
M > 0 and ε > 0 such that
V (x)P f (x) m(d x) ≤ −ε + V (x) f (x) m(d x) + M P f (x) m(d x) (5.20)
X X Z
for f ∈ DV . Then the function V will be called a Hasminskiı̆ function for the operator
P and the set Z .
Lemma 5.2 Let P be a stochastic operator and Z ∈ Σ. Assume that there exists a
Hasminskiı̆ function for the operator P and the set Z . Then for f ∈ D we have
lim sup P n f (x) m(d x) ≥ ε/2M, (5.21)
n→∞ Z
Proof Let f ∈ DV . Then from (5.20) it follows that P n f ∈ DV for every integer
n ≥ 1. Thus
5.2 Asymptotic Decomposition of Stochastic Semigroups 127
V (x)P n+1 f (x) m(d x) ≤ V (x)P n f (x) m(d x) − ε + M P n+1 f (x) m(d x)
X X Z
(5.22)
for f ∈ DV and n ≥ 1. Suppose, contrary to our claim, that (5.21) does not hold. Then
there exists a positive integer n 0 = n 0 ( f ) such that Z P n+1 f (x) m(d x) < ε/(2M)
for n ≥ n 0 . From (5.22) it follows that
ε
V (x)P n+1 f (x) m(d x) ≤ V (x)P n f (x) m(d x) − (5.23)
X X 2
which is impossible. Thus (5.21) holds for f ∈ DV . Since the set DV is dense in D,
(5.21) is satisfied for all f ∈ D.
Consider a stochastic semigroup {P(t)}t≥0 with the generator A and let R =
(I − A)−1 . A measurable function V : X → [0, ∞) is called a Hasminskiı̆ function
for the semigroup {P(t)}t≥0 and a set Z ∈ Σ if (5.20) holds for P = R. Lemmas 5.1
and 5.2 imply the following
Corollary 5.8 Let {P(t)} be a stochastic semigroup generated by the equation
u (t) = Au(t).
Assume that there exists a Hasminskiı̆ function for this semigroup and a set Z . Then
lim sup P(t) f (x) m(d x) ≥ ε/2M for f ∈ D. (5.24)
t→∞ Z
Then condition (5.24) holds and the stochastic semigroup {P(t)}t≥0 is non-sweeping
with respect to the set Z . If Z is a compact set then {P(t)}t≥0 is weakly tight.
The function V is called the Hasminskiı̆ function because he has showed [47]
that the semigroup generated by a non-degenerate Fokker–Planck equation has an
invariant density if and only if there exists a positive function V such that A∗ V (x) ≤
−c < 0 for x ≥ r . The main difficulty in using this method is to define in a
proper way A∗ V because usually V does not belong to the domain of the operator
A∗ . The method of Hasminskiı̆ function has been applied to continuous-time Markov
chains [92], multi-state diffusion processes [86] and diffusion with jumps [88], where
inequality (5.20) was proved by using some generalization of the maximum principle.
This method was also applied in [84] to flows with jumps (see the end of Sect. 1.5)
but the proof of inequality (5.20) is based on an approximation of V by a sequence
of elements from the domain of the operator A∗ .
Chapter 6
Asymptotic Properties of Stochastic
Semigroups—Applications
Condition (I) implies that the semigroup {P(t)}t≥0 is irreducible and Corollary 5.4
gives the following theorem.
Then the stochastic semigroup {P(t)}t≥0 related to Q is non-sweeping from the set
{0, 1, . . . , k}. In particular, if the matrix Q satisfies conditions (I) and (6.4), then the
semigroup {P(t)}t≥0 is asymptotically stable.
The proof of Theorem 6.2 based on the idea of Hasminskiı̆ function is given in [92,
Theorem 10].
As an example of application we consider a birth–death process (see Sect. 1.2). The
evolution of densities of this process is given by the following system of equations:
u i (t) = −ai u i (t) + bi−1 u i−1 (t) + di+1 u i+1 (t) (6.5)
for all i ∈ N. If w is a non-zero sequence then from the last inequality it follows that
limi→∞ wi = ∞. It means that w ∈ / l ∞ and according to Theorem 4.11 the matrix Q
is non-explosive and, consequently, it generates a stochastic semigroup.
We now consider again a birth–death process with bi > 0 and di+1 > 0 for all
i ≥ 0. Let us assume that there exists ε > 0 such that bi ≤ di − ε for i ≥ k. Then
the system (6.5) generates a stochastic semigroup and condition (I) holds. Let vi = i
for i ≥ 0, then
∞
vi qi j = ( j − 1)d j − j (b j + d j ) + ( j + 1)b j = b j − d j ≤ −ε
i=0
for j ≥ k, which implies condition (6.4). Hence, the stochastic semigroup generated
by the system (6.5) is asymptotically stable.
We now consider a non-explosive Markov chain such that condition (I) holds for all
i, j except of j = 0. It means that states i, j ≥ 1 communicate, qi0 = 0 for all i ≥ 0,
and for some j ≥ 1 we have q0 j > 0. Then 0 is an absorbing state, u 0 = (1, 0, 0, . . . )
is an invariant density, and condition (WI) holds for u 0 . According to Corollary 5.5
there exists w ∈ l ∞ with wi ≥ 0 for all i such that
∞ ∞
i=k ρi
wk = ∞ if ρi < ∞,
1 + i=1 ρi i=1
132 6 Asymptotic Properties of Stochastic Semigroups—Applications
In Sect. 1.18 we considered a PDMP which describes the evolution of paralog families
in a genome. We recall that if sk (t) is the number of k-element families at time t then
∞
s1 (t) = −(d + r )s1 (t) + 2(2m + r )s2 (t) + m js j (t), (6.7)
j=3
sk (t) = d(k − 1)sk−1 (t) − (d + r + m)ksk (t) + (r + m)(k + 1)sk+1 (t) (6.8)
for k ≥ 2. The following result [98] characterizes the long-time behaviour of the
sequences (sk (t)).
Theorem
∞ 6.3 Let X be the space of sequences (sk ) which satisfy the condition
k=1 k|s k | < ∞. There exists a sequence (sk∗ ) ∈ X such that for every solution
(sk (t)) of (1.35) and (1.36) with (sk (0)) ∈ X we have
for every k = 1, 2, . . . and C dependent only on the sequence (sk (0)). Moreover if
d = r then
βi
lim sk (t) = C , (6.10)
t→∞ k
r
where β = .
r +m
The idea of the proof of Theorem 6.3 is to replace system (6.7)–(6.8) by a new
one which generates a stochastic semigroup. Let
Then
∞
u 1 = −(2d + m)u 1 + (m + r )u 2 + mu k , (6.11)
k=1
u k = −(d + r + m + d−r
k
)ku k + dku k−1 + (r + m)ku k+1 for i ≥ 2. (6.12)
This system is of the form u (t) = Qu(t). It is easy to check that Q is a Kolmogorov
matrix. We apply Theorem 4.11 to prove that Q is non-explosive. The case d = 0 is
trivial, so we can assume that d > 0. The sequence w = (wk )k≥1 satisfies equation
6.1 Continuous-Time Markov Chains 133
Q ∗ w = λw if and only if
λ (n − 1)(r + m) m
wn+1 = 1 + wn + (wn − wn−1 ) + (wn − w1 )
(n + 1)d (n + 1)d (n + 1)d
and consequently
n λ
wn ≥ w1 1+ for n ≥ 2.
i=2
di
∞
Since the product (1 + λd −1 k −1 ) diverges, we have w ∈
/ l ∞ and according to
k=1
Theorem 4.11 the matrix Q generates a stochastic semigroup.
We next apply Theorem 5.1 to prove that the semigroup {P(t)}t≥0 generated by the
system (6.11)–(6.12) is asymptotically stable. From Eq. (6.11) applied to densities
we obtain
u 1 (t) ≥ −(2d + m)u 1 (t) + m.
rable metric space and Σ = B(X ). Let us assume that inf{ϕ(x) : x ∈ X } > 0. Let
Q(x, B) = ∞ n=0 P (x, B), where the sequence of transition probablities P (x, B)
n n
We assume that there exists a point x0 ∈ X such that for each neighbourhood U
of x0 we have Q(x, U ) > 0 for x-a.e. Then from formula (4.13) it follows that the
semigroup {P(t)}t≥0 is weakly irreducible, i.e. it satisfies condition (WI). Moreover,
we suppose that for every y0 ∈ X there exist an ε > 0, an n ∈ N, and a measurable
function η ≥ 0 such that η(x) m(d x) > 0 and
Again from (4.13) the semigroup {P(t)}t≥0 satisfies condition (K). In order to give a
condition which guarantees asymptotic stability, we can use a Hasminskiı̆ function.
Let V : X → [0, ∞) be a measurable function. We define
This definition is formally correct (though it can happen that A∗ V (x) = ∞ on a set
of positive measure m), and formula (5.26) holds. If for some compact set F we
have A∗ V (x) ≤ M for x ∈ F and A∗ V (x) ≤ −ε for x ∈ X \ F, then there exists
a Hasminskiı̆ function for the semigroup {P(t)}t≥0 and the set F, and consequently
{P(t)}t≥0 is weakly tight. Since the semigroup {P(t)}t≥0 satisfies conditions (WI),
(K) and (WT) of Theorem 5.6, it is asymptotically stable.
Example 6.1 Consider a pure jump-type Markov process on X = [0, ∞) such that
we jump from a point x to the point ϑx , where the random variable ϑx has values in
X . We assume that there exist r > 0 and δ > 0 such that
We also assume that for each x ≥ 0 the random variable ϑx has a continuous density
f x and that there exists a point x0 ∈ X such that f x (x0 ) > 0 for x ≤ r . Since the
densities f x are continuous, condition (K) is fulfilled. From (6.15) we deduce that
for each x ≥ r there is a point y < x − ε such that y ∈ supp f x . Using induction
argument we check that P n (x, [0, r ]) > 0 for sufficiently large n, and since f x (x0 ) >
0 for x ≤ r we finally have P n+1 (x, U ) > 0 for every neighbourhood of x0 , i.e. (WI)
holds. We check that V (x) = x is a Hasminskiı̆ function. We have
From (6.15) we see that A∗ V is bounded from above and A∗ V (x) ≤ −ε for x ≥ r ,
where ε = δ inf ϕ(x). Since the semigroup {P(t)}t≥0 satisfies conditions (WI), (K)
x∈X
and (WT), it is asymptotically stable.
Based on the paper [90] we present here a general method of studying asymptotic
properties of semigroups induced by dynamical systems with random switching
introduced in Sect. 1.9 and described in Sect. 4.2.5. The family of processes ξ(t)
induces a substochastic semigroup {P(t)}t≥0 on the space L 1 (X, Σ, m), where Σ is
the σ -algebra of the Borel subsets of X = X × I and m is the product measure of the
Lebesgue measure on X and the counting measure on I . The semigroup {P(t)}t≥0
corresponds to the transition function P given by
where ξx,i (t) is a Markov process from the family ξ(t) starting from the state (x, i).
We now recall how to check condition (K) for the semigroup {P(t)}t≥0 . Let n ∈ N,
i = (i 1 , . . . , i n ) ∈ I n , i p = i p−1 for p = 2, . . . , n, and t = (t1 , . . . , tn ) be such that
t p > 0 for p = 1, . . . , n. Take y ∈ X and assume that the sequence y0 , . . . , yn given
i
by the recurrent formula y0 = y, y p = πt pp (y p−1 ) for p = 1, . . . , n, is well defined
and qi p+1 i p (y p ) > 0 for p = 1, . . . , n − 1. The function
is called a cumulative flow along the trajectories of the flows π i1 , . . . , π in which joins
the state (y, i 1 ) with the state (yn , i n ). For i ∈ I d+1 , y ∈ X , and t > 0 we define the
function ψ y,t on the set Δt = {τ = (τ1 , . . . , τd ) : τi > 0, τ1 + · · · + τd ≤ t} by
k(t, ·, ·) is the kernel of P(t) and x0 = ψ y,t (τ10 , . . . , τd0 ). It means that condition (K)
holds for the state (y0 , i 1 ).
Condition (6.16) can be checked by using Lie brackets. Let f (x) and g(x) be two
vector fields on Rd . The Lie bracket [ f, g] is a vector field given by
d
∂g j ∂fj
[ f, g] j (x) = f k (x) (x) − gk (x) (x) .
k=1
∂ xk ∂ xk
b2 (x) − b1 (x), . . . , bk (x) − b1 (x), [bi , b j ](x)1≤i, j≤k , [bi , [b j , bl ]](x)1≤i, j,l≤k , . . .
span the space Rd . Let (x0 , i), (y0 , j) ∈ X and assume that the cumulative flow joins
(y0 , j) with (x0 , i). If Hörmander’s condition holds at x0 and ql l (x0 ) > 0 for l = l ,
then condition (K) is satisfied at the state (y0 , j) (see [8, Theorem 4]).
We now show how to check condition (WI). Let (x0 , i) be given. We assume that
for each ε > 0 and each state (y0 , j) we can find x ∈ B(x0 , ε/2) such that (y0 , j)
can be joined with (x, i) by a cumulative flow π it0 (y0 ), where i 1 = j, i n = i and
t0 = (t10 , . . . , tn0 ). From continuous dependence of solutions on initial conditions
and continuity of intensity functions qi p+1 ,i p it follows that there exists δ > 0 such
that each state (y, j) with y ∈ B(y0 , δ) can be joined with a state (x , i) with x ∈
B(x0 , ε/2), i.e. πti0 (y) ∈ B(x0 , ε/2) for y ∈ B(y0 , δ). We now can find γ > 0 such
that πti (y) ∈ B(x0 , ε) if t − t0 < γ . It means that
Prob πti (y) ∈ B(x0 , ε) and t1 + · · · + tn = t > 0
P(t, (y, j), B(x0 , ε) × {i}) > 0 for y ∈ B(y0 , δ). (6.17)
Then inequality (6.18) follows from (6.17) and (6.19). Thus, condition (WI) holds if
there exists a state (x0 , i) such that starting from any state we are able to go arbitrarily
close to (x0 , i) by a cumulative flow. For example, if x0 is an asymptotically stable
stationary point of a semiflow πti with the basin of attraction U ⊆ X and each state
6.2 Dynamical Systems with Random Switching 137
(y, j) can be joined by a cumulative flow with some state (x, i), where x ∈ U , then
condition (WI) holds.
We now explain how to apply Corollary 5.7 to dynamical systems with ran-
dom switching. We assume that we have an increasing sequence
of bounded open
sets G n ⊆ G with smooth boundaries ∂G n , such that X ⊆ ∞ n=1 c1G n , where c1G n
denotes the closure of G n . Let n(x) be an outward pointing normal field of the
boundary ∂G n . We assume that
Then the solutions of each system x = bi (x) starting from points in c1G n remain in
the set c1G n . In particular the sets X n = X ∩ c1G n are invariant with respect to the
semigroup {P(t)}t≥0 .
We summarize the above discussion. Consider a family of systems of differential
equations x = bi (x), i ∈ I , such that for each point x̄ ∈ X the initial problem x =
bi (x), x(0) = x̄, has a unique solution x(t) ∈ X for all t ≥ 0. Then a dynamical
system with random switching created by this family of systems and some continuous
and bounded intensity functions q ji (x) generates a stochastic semigroup {P(t)}t≥0
on L 1 (X, B(X), m). Assume that there exist x0 ∈ X and i 0 ∈ I such that starting
from any state (x, i) ∈ X we are able to go arbitrarily close to (x0 , i 0 ) by a cumulative
flow and that Hörmander’s condition holds at x0 and ql l (x0 ) > 0 for all l, l , l = l .
We also assume ∞that there is an increasing family of bounded open sets G n ⊆ Rd
such that X ⊆ n=1 c1G n and condition (6.20) holds. Then the semigroup {P(t)}t≥0
is asymptotically stable.
From now on we consider only models given by (6.21), (6.22) and (6.23). In all
these models we have two dynamical systems, which correspond, respectively, to
i = 0 and i = 1. Each system has a unique stationary point, either 0 = (0, 0, 0), or
1 = (1, 1, 1), which is asymptotically stable. Assume that intensities q0 and q1 of
transformations of a gene into an active state and into inactive state, respectively, are
non-negative continuous functions of x. We assume that q0 (0) > 0 and q1 (1) > 0.
This assumption implies two properties of the relative PDMP. The first one is that
if the process starts from any point x ∈ Rd then it enters the invariant set X =
[0, 1]d × {0, 1} and the process visits any neighbourhood of the point (0, 0) and the
point (1, 1) for infinitely many times. It means that condition (WI) holds for x 0 = 0
and for x 0 = 1. Moreover, if we assume additionally that q0 (1) > 0 or q1 (0) > 0
then we can check that condition (K) holds by applying Hörmander’s condition. We
consider here only the most advanced three-dimensional model. Let v = (1, 0, 0).
Then b2 (x) = v + b1 (x). Hence
b2 − b1 = (1, 0, 0),
[b1 , b2 ] = [b1 , v] = [(−x1 , α(x2 − x1 ), β(x3 − x2 )), (1, 0, 0)] = (1, −α, 0),
[b1 , [b1 , b2 ]] = (1, −(α 2 + α), αβ).
The vectors b2 − b1 , [b1 , b2 ], [b1 , [b1 , b2 ]] span the space R3 , and consequently
Hörmander’s condition holds. Since the set X is compact, the semigroup {P(t)}t≥0
is asymptotically stable according to Corollary 5.6.
The one-dimensional model from Sect. 1.8 is simple enough to find the invariant
density, which allows us to give more information on asymptotic behaviour of distri-
butions of this process. In this case the functions u i (t, x) = P(t) f (x, i), i ∈ {0, 1},
satisfy the following Fokker–Planck system:
∂u 0 ∂
+ (−xu 0 ) = q1 u 1 − q0 u 0 ,
∂t ∂x
∂u 1 ∂
+ ((1 − x)u 1 ) = q0 u 0 − q1 u 1 .
∂t ∂x
If u i , i ∈ {0, 1}, do not depend on t, then
(−xu 0 (x)) = q1 u 1 − q0 u 0 ,
(6.24)
((1 − x)u 1 (x)) = q0 u 0 − q1 u 1 .
Hence
−xu 0 (x) + (1 − x)u 1 (x) = c for some c ∈ R.
6.2 Dynamical Systems with Random Switching 139
c x
From the last equation we find u 1 (x) = + u 0 (x). If we suppose that
1−x 1−x
u(x, i) is a density then both functions u 0 and u 1 are non-negative and integrable,
but this case is possible only if c = 0. Indeed, if c < 0 then
x
su 0 (s) x
u 0 (s) |c|x
ds ≤ x ds <
0 1−s 0 1−s 2
Thus
q0 (s) − 1 q1 (s)
x
u 0 (x) = C exp − ds,
0 s 1−s
x (6.25)
Cx q0 (s) − 1 q1 (s)
u 1 (x) = exp − ds,
1−x 0 s 1−s
1
where we choose C such that 0 (u 0 (x) + u 1 (x)) d x = 1.
Example 6.2 Consider the case when q0 and q1 are constants. Then
Remark 6.1 Observe that if q0 (0) > 0 and q1 (1) > 0 then the functions u 0 and u 1
given by (6.25) are integrable. That means that there exists a unique density for this
semigroup and since the semigroup is partially integral, it is asymptotically stable
according to Theorem 5.2. Also in two- and three-dimensional cases the semigroup
remains asymptotically stable when q0 (1) = 0 and q1 (0) = 0, but the proof is more
advanced because we need to check directly condition (6.16).
Remark 6.2 It is rather difficult to find a general formula for an invariant density
in two- and three-dimensional cases. But we can find a formula for the support of
invariant density if q0 and q1 are strictly positive functions (see [100]). Also in this
paper we can find a discussion concerning adiabatic limit of the system, i.e. the
behaviour of our model, when both rates q0 and q1 tend to infinity, unlike their ratio.
In this case, our process “tends” to a deterministic model, which can exhibit two
specific types of behaviour: bistability and the existence of the limit cycle.
140 6 Asymptotic Properties of Stochastic Semigroups—Applications
Remark 6.3 If q0 (0) = 0 or q1 (1) = 0 then the system can loose asymptotic stability
even in the case when qi (x) are positive for all other x. Consider one-dimensional
example with q0 (x) = x r , r > 0 and q1 (x) = 1. −1Then from (6.25) it follows that
−1 r
u 0 (x) = C 1−x
x
exp r x and u 1 (x) = C exp r x r . In this case u 0 is not inte-
grable in 0, and therefore an invariant density does not exist. In this example the
distributions of the process converge weakly to δ(0,0) . It means that the average time
when the gene is inactive tends to 1. If q0 (0) = 0 and q1 (1) = 0, then the distribution
of the process can converge to c1 δ(0,0) + c2 δ(1,1) , where the constants c1 and c2 can
depend on the initial density. Such properties can be checked by using Theorem 5.3.
We now generalize the method considered for pure jump-type Markov processes to
flows with jumps described in Sect. 4.2.4. We show how we can apply Theorem 5.6 to
prove asymptotic stability of the stochastic semigroup {P(t)}t≥0 . Define R(t, x, B) =
ϕ(πt x)P(πt x, B) for t ≥ 0, x ∈ X , B ∈ Σ and by induction
d ∂V
where A∗0 V (x) = bi (x) (x). If for some compact set F we have A∗ V (x) ≤ M
i=1 ∂ xi
for x ∈ F and A∗ V (x) ≤ −ε for x ∈ X \ F, then V is a Hasminskiı̆ function for
the semigroup {P(t)}t≥0 and the set F. The proof of this fact can be found in [84]
and it is based on an approximation of V by a sequence of elements from the domain
of the operator A∗ . Thus {P(t)}t≥0 is weakly tight. Since the semigroup {P(t)}t≥0
satisfies conditions (WI), (K) and (WT) of Theorem 5.6, it is asymptotically stable.
We now investigate a size-structured model from Sect. 1.5. We keep the notation from
Sect. 1.5, i.e. x is the size of a cell, it changes according to the equation x = g(x);
ϕ(x) is the intensity of division; if the mother cell has size x at the moment of division,
then the new born daughter cell has size x/2. Here the jump is deterministic and given
by S(x) = x/2. Let xmin be the minimum cell size. We assume that ϕ is a continuous
function and that ϕ(x) = 0 for x ≤ 2xmin and ϕ(x) > 0 for x > 2xmin , because 2xmin
is the minimum cell size when it can divide. Thus we can consider our process in the
phase space X = [xmin , ∞). We also assume that the function g : X → (0, ∞) has
a bounded derivative. Moreover, we assume that
We can join any point y0 ∈ X with x0 by the action of the dynamical system and
transformation S. Indeed, if y0 < x0 then there exists t > 0 such that x0 = πt (y0 ).
If y0 > x0 then S n (y0 ) < x0 for some positive integer n and, therefore, it can be
joined with x0 . Hence condition (K) holds. The semigroup {P(t)}t≥0 induced by this
process is irreducible because we can join any two points from the set (xmin , ∞).
From Corollary 5.4 it follows that the semigroup {P(t)}t≥0 is asymptotically stable
or sweeping from compact sets. Therefore, if {P(t)}t≥0 has an invariant density f ∗ ,
then {P(t)}t≥0 is asymptotically stable and f ∗ > 0 a.e.; if {P(t)}t≥0 has no invariant
density, then for all c > xmin and f ∈ L 1 we have
c
lim P(t) f (x) d x = 0.
t→∞ xmin
∂u ∂
= − (g(x)u(t, x)) − ϕ(x)u(t, x) + 2ϕ(2x)u(t, 2x). (6.28)
∂t ∂x
6.3 Cell Maturation Models 143
Thus, if there exists an invariant density f ∗ , then it satisfies the following equation:
g(x)
lim sup ϕ(x) > 0, and κ = lim sup < log 2. (6.30)
x→∞ x→∞ xϕ(x)
Let h(r ) = (1 − 2−r )/r . Since limr →0+ h(r ) = log 2, we find r > 0 such that h(r ) >
κ. From (6.30) and (6.31) it follows that there exists ε > 0 such that A∗ V (x) ≤ −ε
for sufficiently large x. Thus if (6.27) and (6.30) hold, then the semigroup {P(t)}t≥0
is asymptotically stable.
Remark 6.4 Age- and size-structured models play an important role in population
dynamics. The first one was introduced in [74, 105] and the second in [12]. Both
models were intensively studied and generalized [31, 46, 76, 97, 118]. In [11] it
is considered an evolution equation u = Au which can be used to study both types
of models. This equation generates a semigroup of positive operators {T (t)}t≥0 on
some space L 1 (X, Σ, μ). The main result of this paper is asynchronous exponential
growth of the population. Precisely, under suitable assumptions there exist λ ∈ R,
an integrable function f ∗ and a positive linear functional α such that
We can use the technique developed for stochastic semigroup in Chap. 5 to prove
(6.32). In order to do it we first prove that there exist λ ∈ R and function w
such that A∗ w = λw and 0 < c1 ≤ w(x) ≤ c2 for x ∈ X . Then we introduce a sto-
chastic semigroup {P(t)}t≥0 on the space L 1 (X, Σ, m) with m(d x) = w(x) μ(d x)
given by P(t) = e−λt T (t). We check that {P(t)}t≥0 is asymptotically stable, which
gives (6.32).
144 6 Asymptotic Properties of Stochastic Semigroups—Applications
We now consider the two-phase cell cycle model presented in Sect. 1.6. This is a
simplified version of the model introduced in [68]. We assume that the size x of
a cell changes according to the equation x = g(x), where the function g has a
bounded derivative; the intensity of division ϕ(x) is a continuous function, ϕ(x) = 0
for x ≤ x B and ϕ(x) > 0 for x > x B , where x B is the minimum cell size when
it can enter the phase B; the constant t B > 0 is the duration of the phase B. The
minimum cell size is xmin = 21 πt B x B and it is clear that we need to assume that
πt B x B < 2x B to have xmin < x B . Thus we can consider our process on the space
X = [xmin , ∞) × {1} ∪ [x B , ∞) × [0, t B ] × {2}. As in Sect. 6.3.3 we assume that
Since this model is not exactly a special case of the general model considered in
Sect. 6.3.2, we explain how to check condition (K). Let z̄ = π−t B (2 x̄). Then z̄ > x B
and ϕ(z̄) > 0. First, we check that (K) holds for y0 = (z̄, 1). Precisely, we show that
there exist t > t B , ε > 0, x0 ∈ (xmin , ∞) × {1} and neighbourhoods U of y0 and V
of x0 , such that the operator P(t) is partially integral with the kernel
Let z > x B and t > t B . We introduce a function ψz,t defined on the interval [t B , t]
by
ψz,t (τ ) = πt−τ 21 πτ z .
We have
dψz,t ∂πt−τ 1
(τ ) = −g πt−τ 21 πτ z + πτ z · 21 g(πτ z),
dτ ∂x 2
dψz,t
lim+ lim (t B ) = −g 21 πt B z̄ + 21 g πt B z̄ = −g x̄ + 21 g 2 x̄ = 0.
t→t B z→z̄ dτ
for z from some neighbourhood U1 of z̄, for some t > t B and τ ∈ [t B , t]. Now we
consider our stochastic process starting from the point (z, 1). Let the random variable
T be the first time when the process jumps to the phase B. Since ϕ(z̄) > 0, T has a
density distribution h such that h(τ )≥ β for τ ∈[0,t − t B ], where β is a positive
constant. Let τ0 = 21 (t B + t), x0 = πt−τ0 21 πτ0 z̄ , 1 and V = (x0 − r, x0 + r ) ×
{1} with r = 21 δ(t − t B ). Then (6.34) holds with ε = β/M. Since every point x ∈ X
can be joined with (z̄, 1), condition (K) is fulfilled. The semigroup {P(t)}t≥0 induced
6.3 Cell Maturation Models 145
by this process also satisfies condition (WI) because we can join any point from X
with y0 . The infinitesimal generator of {P(t)}t≥0 is of the following form:
∂
A f (x, 1) = − (g(x) f (x, 1)) − ϕ(x) f (x, 1) + 2 f (2x, t B , 2),
∂x (6.36)
∂ ∂
A f (x, τ, 2) = − (g(x) f (x, τ, 2)) − ( f (x, τ, 2)),
∂x ∂τ
and the functions from the domain of A satisfy the condition
We find that
∂f
A∗ f (x, 1) = g(x) (x, 1) − ϕ(x) f (x, 1) + ϕ(x) f (x, 0, 2),
∂x (6.37)
∂f ∂f
A∗ f (x, τ, 2) = g(x) (x, τ, 2) + (x, τ, 2),
∂x ∂τ
We take V (x, 1) = x r and V (x, τ, 2) = Ce−aτ x r , where r , a and C are some positive
constants. Then
Then taking a = κr , we check that there exist ε > 0 and a sufficiently small r > 0
such that A∗ V (x, 1) ≤ −ε and A∗ V (x, τ, 1) ≤ −ε for τ ∈ [0, t B ] and sufficiently
large x, which means that the semigroup {P(t)}t≥0 is weakly tight. Summarizing, if
g(2 x̄) = 2g(x̄) for some x̄ > xmin and
g(x) g(x)
lim sup < log 2 − t B lim sup ,
x→∞ xϕ(x) x→∞ x
We recall that in this model a maturity of a cell is a real variable x ∈ [0, 1]. A new born
cell has maturity 0 and a cell splits at maturity 1. The maturity x grows according to the
equation x = v, where the maturation velocity v of each individual cell is constant.
The relation between the maturation velocities of mother’s v and daughter’s cells v
is given by a transition probability P(v, dv ). As we have mentioned in Sect. 1.7, this
model is a special case of one-dimensional stochastic billiards investigated in [78] and
based on that paper we briefly present results concerning its asymptotic properties.
We assume that v ∈ (0, 1] and P(v, dv ) = αδv (v ) + βk(v , v) dv , where β ∈ (0, 1],
α + β = 1, and 0 k(v , v) dv = 1 for v ∈ (0, 1]. We showed in Sect. 4.2.6 that the
1
process ξ(t) = (x(t), v(t)) induces a stochastic semigroup {P(t)}t≥0 on the space
L 1 (X, Σ, m), where X = (0, 1]2 , Σ = B(X ) and m(d x × dv) = d x × dv. Since
the boundary condition (4.26) contains a kernel operator, one can check that the semi-
group {P(t)}t≥0 is partially integral. Observe that if this semigroup has an invariant
density f ∗ , then A f ∗ = 0 and from (4.25) it follows that f ∗ does not depend on
x. From (4.26) we deduce that v f ∗ = K (v f ∗ ), where K is a stochastic operator on
L 1 [0, 1] given by
1
K h(v ) = k(v , v)h(v) dv. (6.39)
0
We assume that K is irreducible which is equivalent to say that there does not exist a
set B ⊆ (0, 1] of the Lebesgue measure 0 < |B| < 1 such that k(v , v) = 0 for v ∈ B
and v ∈/ B. From irreducibility it follows that if an invariant density f ∗ exists then
it is unique and f ∗ (x) > 0 for a.e. x. The question of the existence of an invariant
density is nontrivial. If for example we assume that there exist C > 0 and γ > 0
such that
k(v, v ) ≤ C|v|γ for v, v ∈ (0, 1], (6.40)
follows that
ε 1
lim P(t) f (x, v) d x dv = 1 (6.41)
t→∞ 0 0
for every density f and every ε > 0. For example if k ≡ 1, then h ∗ ≡ 1 and f ∗ (v) =
v−1 is not integrable. Thus, the semigroup has no invariant density and consequently
(6.41) holds. It is interesting that in this example we have
c
P(t) f (x, v) ∼ (log t)−1 as t → ∞
|v|
6.3 Cell Maturation Models 147
In Sect. 1.11 we considered Stein’s model which describes how the depolarization
V (t) changes in time. Now, based on the paper [91], we present its asymptotic
properties. First, we introduce an extra 0-phase, which begins at the end of the
refractory period and ends when depolarization jumps from 0 for the first time. Thus
the process is defined on the space X = {(0, 0)} ∪ (−∞, θ ] × {1} ∪ [0, t R ] × {2}.
Define a measure m on the σ -algebra B(X ) of the Borel subsets of X by m =
δ(0,0) + m 1 + m 2 , where δ(0,0) is the Dirac measure at (0, 0), m 1 is the Lebesgue
measure on the segment (−∞, θ ] × {1} and m 2 is the Lebesgue measure on the
segment [0, t R ] × {2}. The process induces a stochastic semigroup {P(t)}t≥0 on the
space L 1 (X, B(X ), m).
The stochastic semigroup introduced by Stein’s model is asymptotically stable.
The proof of this result is given in [91] and it is based on Theorem 5.6. Since the
semigroup is asymptotically stable it would be useful to find an invariant density.
Unfortunately, this is rather a difficult task and we only present the fact that the
invariant density f ∗ of the process with the refractory period can be expressed by
the invariant density f¯∗ of the process without the refractory period (i.e. t R = 0) and
the same other parameters. Let
−1
c = 1 + t R (λ E + λ I ) f¯∗ (0, 0) .
Then
f ∗ (0, 0) = c f¯∗ (0, 0),
f ∗ (x, 1) = c f¯∗ (x, 1) for x ∈ (−∞, θ ],
f ∗ (x, 2) = c(λ E + λ I ) f¯∗ (0, 0) for x ∈ [0, t R ].
Appendix A
Measure and Probability Essentials
Our objective in this appendix is to introduce basic definitions and results from
measure and probability theory required in the text. For the general background and
proofs we refer the reader to [55].
Note that the intersection of two measurable rectangles is a measurable rectangle. The
σ-algebra Σ1 ⊗ Σ2 is called the product σ-algebra on X 1 × X 2 , and the measurable
space (X 1 × X 2 , Σ1 ⊗ Σ2 ) is called the product of (X 1 , Σ1 ) and (X 2 , Σ2 ).
f −1 (B) = {x ∈ X 1 : f (x) ∈ B} ∈ Σ1 .
where
(k − 1)2−n , if (k − 1)2−n ≤ f (x) < k2−n , 1 ≤ k ≤ n2n ,
f n (x) =
n, if n ≤ f (x).
Theorem A.2 (Uniqueness) Let μ and ν be two finite measures on (X, Σ). Suppose
that C is a π-system such that X ∈ C and σ(C ) = Σ. If μ(B) = ν(B) for all B ∈ C
then μ = ν.
Lemma A.2 (Fatou) Let (X, Σ, μ) be a measure space and let ( f n ) be a sequence
of non-negative measurable functions. Then
lim inf f n (x) μ(d x) ≤ lim inf f n (x) μ(d x).
X n→∞ n→∞ X
(2) Dominated convergence: If g is integrable, | f n (x)| ≤ g(x) for all n and x, and
the sequence ( f n ) is convergent pointwise, then (A.1) holds.
The next theorem, referred to usually as the Fubini theorem, is about the existence
of a product measure and the change of the order of integration in double integrals.
The result remains true for any μ1 × μ2 -integrable f . The measure μ1 × μ2 is called
the product measure of μ1 and μ2 .
For a given measure space (X, Σ, μ) a finite signed measure ν on (X, Σ) is called
absolutely continuous with respect to μ if ν(B) = 0 whenever μ(B) = 0. In this
case ν is also said to be dominated by μ, written ν μ.
Theorem A.5 (Radon–Nikodym) Let (X, Σ, μ) be a σ-finite measure space and let
ν be a finite signed measure on (X, Σ) absolutely continuous with respect to μ. Then
there exists f ∈ L 1 (X, Σ, μ) such that
ν(B) = f (x) μ(d x) for all B ∈ Σ.
B
Appendix A: Measure and Probability Essentials 153
A. 3 Random Variables
k
k
|F(b j ) − F(a j )| < ε if (b j − a j ) < δ.
j=1 j=1
If ζ = h(ξ) with measurable h and if the X -valued random variable ξ has distribution
μξ then
154 Appendix A: Measure and Probability Essentials
E(h(ξ)) = h(x) μξ (d x).
X
Theorem A.7 (Strong law of large numbers) Let ξ1 , ξ2 , . . . be i.i.d. random vari-
ables such that E(|ξ1 |) < ∞. Then
1
n
P lim ξk = E(ξ1 ) = 1.
n→∞ n
k=1
P(A ∩ B)
P(B|A) = , B ∈ F.
P(A)
It is denoted by E(ζ|G ).
P(A|G ) = E(1 A |G ).
In particular, the random variable P(ζ ∈ B|ξ), being σ(ξ) measurable, can be repre-
sented as h B (ξ) for some measurable h B , since we have the following.
Theorem A.10 (Representation theorem) Let ξ be an X -valued random variable.
Then a real-valued random variable η is measurable with respect to σ(ξ) if and only
if there exists a measurable h : X → R such that η = h(ξ).
The next result is useful for the computation of conditional expectations.
for every A ∈ G , or, equivalently, E(ηg(ξ, ζ)) = E(ηh(ξ)) for every positive G -
measurable random variable η. Let μζ and μ(ξ,η) denote the distributions of ζ and
(ξ, η). We have
h(x) = E(g(x, ζ)) = g(x, y) μζ (dy).
Proof We show that (1) implies (2). Since for any A the random variable P(A|H )
is K -measurable, it is enough to show that for every B ∈ K we have
P(B ∩ A|H ) = E(1 B∩A |H ) = E(E(1 B∩A |K )|H ) = E(1 B E(1 A |K )|H )
= E(1 B E(1 A |H )|H ) = E(1 A |H )E(1 B |H ).
References
17. Bobrowski, A., Lipniacki, T., Pichór, K., Rudnicki, R.: Asymptotic behavior of distributions
of mRNA and protein levels in a model of stochastic gene expression. J. Math. Anal. Appl.
333, 753–769 (2007)
18. Bokes, P., King, J.R., Wood, A.T.A., Loose, M.: Transcriptional bursting diversifies the behav-
iour of a toggle switch: Hybrid simulation of stochastic gene expression. Bull. Math. Biol.
75, 351–371 (2013)
19. Boulanouar, M.: A mathematical study for a Rotenberg model. J. Math. Anal. Appl. 265,
371–394 (2002)
20. Bressloff, P., Newby, J.: Directed intermittent search for hidden targets. New J. Phys. 11,
023033 (2009)
21. Bressloff, P., Newby, J.: Quasi-steady state reduction of molecular motor-based models of
directed intermittent search. Bull. Math. Biol. 72, 1840–1866 (2010) Rev. Mod. Phys. 85,
135–196 (2013)
22. Bressloff, P., Newby, J.: Stochastic models of intracellular transport. Rev. Mod. Phys. 85,
135–196 (2013)
23. Buckwar, E., Riedler, M.G.: An exact stochastic hybrid model of excitable membranes includ-
ing spatio-temporal evolution. J. Math. Biol. 63, 1051–1093 (2011)
24. Burkitt, A.N.: A review of the integrate-and-fire neuron model: I Homogeneous synaptic
input. Biol. Cybern. 95, 1–19 (2006)
25. Cai, L., Friedman, N., Xie, X.S.: Stochastic protein expression in individual cells at the single
molecule level. Nature 440, 358–362 (2006)
26. Capasso, V., Bakstein, D.: An Introduction to Continuous-Time Stochastic Processes. Theory,
Models and Applications to Finance, Biology and Medicine. Birkhäuser, Boston (2005)
27. Cassandras, C.G., Lygeros, J. (eds.): Stochastic Hybrid Systems. Control Engineering Series,
vol. 24, CRC Press, Boca Raton, FL (2007)
28. Davis, M.H.A.: Piecewise-deterministic Markov processes: a general class of nondiffusion
stochastic models. J. R. Stat. Soc. Ser. B 46, 353–388 (1984)
29. Davis, M.H.A.: Markov Models and Optimization. Monographs on Statistics and Applied
Probability, vol. 49, Chapman & Hall, London (1993)
30. Desch, W.: Perturbations of positive semigroups in AL-spaces. Unpublished (1988)
31. Diekmann, O., Heijmans, H.J.A.M., Thieme, H.R.: On the stability of the cell size distribution.
J. Math. Biol. 19, 227–248 (1984)
32. Dieudonne, J.: Sur le théorème de Radon-Nikodym. Ann. Univ. Grenoble 23, 25–53 (1948)
33. Dunford, N.: On one parameter groups of linear transformations. Ann. Math. 39, 569–573
(1938)
34. Engel, K.J., Nagel, R.: One-parameter semigroups for linear evolution equations. Graduate
Texts in Mathematics, vol. 194. Springer, New York (2000)
35. Ethier, S.N., Kurtz, T.G.: Markov Processes. Characterization and Convergence. Wiley, New
York (1986)
36. Evans, S.N.: Stochastic billiards on general tables. Ann. Appl. Probab. 11, 419–437 (2001)
37. Feller, W.: An introduction to Probability Theory and Applications. Wiley, New York (1966)
38. Foguel, S.R.: The Ergodic Theory of Markov Processes. Van Nostrand Reinhold Comp, New
York (1969)
39. Fox, R.F.: Stochastic versions of the Hodgkin-Huxley equations. Biophys. J. 72, 2068–2074
(1997)
40. Friedman, N., Cai, L., Xie, X.S.: Linking stochastic dynamics to population distribution: an
analytical framework of gene expression. Phys. Rev. Lett. 97, 168302 (2006)
41. Gavrilets, S., Boake, C.R.B.: On the evolution of premating isolation after a founder event.
Am. Nat. 152, 706–716 (1998)
42. Genadot, A., Thieullen, M.: Averaging for a fully coupled piecewise-deterministic Markov
process in infinite dimensions. Adv. Appl. Probab. 44, 749–773 (2012)
43. Golding, I., Paulsson, J., Zawilski, S., Cox, E.: Real-time kinetics of gene activity in individual
bacteria. Cell 123, 1025–1036 (2005)
References 161
44. Greiner, G.: Perturbing the boundary conditions of a generator. Houston J. Math. 13, 213–229
(1987)
45. Gwiżdż, P., Tyran-Kamińska, M.: Densities for piecewise deterministic Markov processes
with boundary. Preprint
46. Gyllenberg, M., Heijmans, H.J.A.M.: An abstract delay-differential equation modelling size
dependent cell growth and division. SIAM J. Math. Anal. 18, 74–88 (1987)
47. Hasminskiı̆, R.Z.: Ergodic properties of recurrent diffusion processes and stabilization of
the solutions of the Cauchy problem for parabolic equations (in Russian). Teor. Verojatn.
Primenen. 5, 196–214 (1960)
48. Hennequin, P., Tortrat, A.: Théorie des probabilités et quelques applications. Masson et Cie,
Paris (1965)
49. Hille, E.: Functional Analysis and Semi-Groups. American Mathematical Society Colloquium
Publications, vol. 31. American Mathematical Society, New York (1948)
50. Hillen, T., Hadeler, K.P.: Hyperbolic systems and transport equations in mathematical biology.
In: Warnecke, G. (ed.) Analysis and Numerics for Conservation Laws, pp. 257–279. Springer,
Berlin, Heidelberg (2005)
51. Hu, J., Lygeros, J. Sastry, S.: Towards a theory of stochastic hybrid systems. In: Lynch, N.,
Krogh, B,H. (eds.) Hybrid Systems: Computation and Control. LNCS vol. 1790, pp. 160–173.
Springer, Berlin (2000)
52. Hu, J., Wu, W.C., Sastry, S.S.: Modeling subtilin production in bacillus subtilis using stochastic
hybrid systems. In: Alur, R., Pappas, G.J. (eds.) Hybrid Systems: Computation and Control.
LNCS, vol. 2993, pp. 417–431. Springer, Berlin (2004)
53. Jacod, J., Skorokhod, A.V.: Jumping Markov processes. Ann. Inst. H. Poincaré Probab. Statist.
32, 11–67 (1996)
54. Jamison, B., Orey, S.: Markov chains recurrent in the sense of Harris Z. Wahrsch. Verw.
Gebiete 8, 41–48 (1967)
55. Kallenberg, O.: Foundations of Modern Probability, 2nd edn. Springer, New York (2002)
56. Kato, T.: On the semi-groups generated by Kolmogoroff’s differential equations. J. Math.
Soc. Jpn 6, 1–15 (1954)
57. Kokko, H., Jennions, M.D., Brooks, R.: Unifying and testing models of sexual selection.
Annu. Rev. Ecol. Evol. Syst. 37, 43–66 (2006)
58. Komorowski T., Tyrcha, J.: Asymptotic properties of some Markov operators. Bull. Pol. Ac.:
Math. 37, 221–228 (1989)
59. Kouretas, P., Koutroumpas, K., Lygeros, J., Lygerou, Z.: Stochastic hybrid modeling of bio-
chemical processes. In: Cassandras, C.G., Lygeros, J. (eds.) Stochastic Hybrid Systems. Con-
trol Engineering Series, vol. 24, pp. 221–248. CRC Press, Boca Raton, FL (2007)
60. Kuno, E.: Simple mathematical models to describe the rate of mating in insect populations.
Res. Popul. Ecol. 20, 50–60 (1978)
61. Lasota, A., Mackey, M.C.: Globally asymptotic properties of proliferating cell populations.
J. Math. Biol. 19, 43–62 (1984)
62. Lasota, A., Mackey, M.C.: Chaos, Fractals, and Noise, Applied Mathematical Sciences, vol.
97. Springer, New York (1994)
63. Lasota, A., Yorke, J.A.: Exact dynamical systems and the Frobenius-Perron operator. Trans.
AMS 273, 375–384 (1982)
64. Lebowitz, J.L., Rubinow, S.L.: A theory for the age and generation time distribution of micro-
bial population. J. Math. Biol. 1, 17–36 (1974)
65. Lipniacki, T., Paszek, P., Marciniak-Czochra, A., Brasier, A.R., Kimmel, M.: Transcriptional
stochasticity in gene expression. J. Theoret. Biol. 238, 348–367 (2006)
66. Lotz, H.P.: Uniform convergence of operators on L ∞ and similar spaces. Math Z 190, 207–220
(1985)
67. Lumer, G., Phillips, R.S.: Dissipative operators in a Banach space. Pacific J. Math. 11, 679–698
(1961)
68. Mackey, M.C., Rudnicki, R.: Global stability in a delayed partial differential equation describ-
ing cellular replication. J. Math. Biol. 33, 89–109 (1994)
162 References
69. Mackey, M.C., Santillán, M., Tyran-Kamińska, M., Zeron, E.S.: Simple Mathematical Models
of Gene Regulatory Dynamics. Lecture Notes on Mathematical Modelling in the Life Sciences.
Springer, Cham (2016)
70. Mackey, M.C., Tyran-Kamińska, M.: Dynamics and density evolution in piecewise determin-
istic growth processes. Ann. Polon. Math. 94, 111–129 (2008)
71. Mackey, M.C., Tyran-Kamińska, M., Yvinec, R.: Molecular distributions in gene regulatory
dynamics. J. Theoret. Biol. 274, 84–96 (2011)
72. Mackey, M.C., Tyran-Kamińska, M., Yvinec, R.: Dynamic behavior of stochastic gene expres-
sion models in the presence of bursting. SIAM J. Appl. Math. 73, 1830–1852 (2013)
73. Mackey, M.C., Tyran-Kamińska, M.: The limiting dynamics of a bistable molecular switch
with and without noise. J. Math. Biol. 73, 367–395 (2016)
74. McKendrick, A.G.: Application of mathematics to medical problems. Proc. Edinb. Math. Soc.
14, 98–130 (1926)
75. Méndez, V., Campos, D., Pagonabarraga, I., Fedotov, S.: Density-dependent dispersal and
population aggregation patterns. J. Theoret. Biol. 309, 113–120 (2012)
76. Metz, J.A.J., Diekmann, O. (eds.): The Dynamics of Physiologically Structured Populations.
Springer Lecture Notes in Biomathematics, vol. 68. Springer, New York (1986)
77. Miȩkisz, J., Szymańska, P.: Gene expression in self-repressing system with multiple gene
copies. Bull. Math. Biol. 75, 317–330 (2013)
78. Mokhtar-Kharroubi, M., Rudnicki, R.: On asymptotic stability and sweeping of collisionless
kinetic equations. Acta Appl. Math. 147, 19–38 (2017)
79. Nummelin, E.: General Irreducible Markov Chains and Non-negative Operators, Cambridge
Tracts in Mathematics, vol. 83. Cambridge University Press, Cambridge (1984)
80. Ochab-Marcinek, A., Tabaka, M.: Transcriptional leakage versus noise: a simple mechanism
of conversion between binary and graded response in autoregulated genes. Phys. Rev. E 91,
012704 (2015)
81. Othmer, H.G., Dunbar, S.R., Alt, W.: Models of dispersal in biological systems. J. Math. Biol.
26, 263–298 (1988)
82. Othmer, H.G., Hillen, T.: The diffusion limit of transport equations II: Chemotaxis equations.
SIAM J. Appl. Math. 62, 1222–1250 (2002)
83. Phillips, R.S.: Semi-groups of positive contraction operators. Czechoslovak Math. J. 12, 294–
313 (1962)
84. Pichór, K.: Asymptotic stability of a partial differential equation with an integral perturbation.
Ann. Polon. Math. 68, 83–96 (1998)
85. Pichór, K.: Asymptotic stability and sweeping of substochastic semigroups. Ann. Polon. Math.
103, 123–134 (2012)
86. Pichór, K., Rudnicki, R.: Stability of Markov semigroups and applications to parabolic sys-
tems. J. Math. Anal. Appl. 215, 56–74 (1997)
87. Pichór, K., Rudnicki, R.: Continuous Markov semigroups and stability of transport equations.
J. Math. Anal. Appl. 249, 668–685 (2000)
88. Pichór, K., Rudnicki, R.: Asymptotic behaviour of Markov semigroups and applications to
transport equations. Bull. Pol. Ac.: Math. 45, 379–397 (1997)
89. Pichór, K., Rudnicki, R.: Asymptotic decomposition of substochastic operators and semi-
groups. J. Math. Anal. Appl. 436, 305–321 (2016)
90. Pichór, K., Rudnicki, R.: Asymptotic decomposition of substochastic semigroups and appli-
cations. Stochastics and Dynamics, (accepted)
91. Pichór, K., Rudnicki, R.: Stability of stochastic semigroups and applications to Stein’s neu-
ronal model. Discrete Contin. Dyn. Syst. B, (accepted)
92. Pichór, K., Rudnicki, R., Tyran-Kamińska, M.: Stochastic semigroups and their applications
to biological models. Demonstratio Mathematica 45, 463–495 (2012)
93. Rotenberg, M.: Transport theory for growing cell populations. J. Theoret. Biol. 103, 181–199
(1983)
94. Rubinow, S.I.: A maturity time representation for cell populations. Biophys. J. 8, 1055–1073
(1968)
References 163
95. Rudnicki, R.: On asymptotic stability and sweeping for Markov operators. Bull. Pol. Ac.:
Math. 43, 245–262 (1995)
96. Rudnicki, R.: Stochastic operators and semigroups and their applications in physics and biol-
ogy. In: Banasiak, J., Mokhtar-Kharroubi, M. (eds.) Evolutionary Equations with Applications
in Natural Sciences. Lecture Notes in Mathematics, vol. 2126, pp. 255–318. Springer, Hei-
delberg (2015)
97. Rudnicki, R., Pichór, K.: Markov semigroups and stability of the cell maturation distribution.
J. Biol. Syst. 8, 69–94 (2000)
98. Rudnicki, R., Tiuryn, J., Wójtowicz, D.: A model for the evolution of paralog families in
genomes. J. Math. Biol. 53, 759–770 (2006)
99. Rudnicki, R., Tiuryn, J.: Size distribution of gene families in a genome. Math. Models Methods
Appl. Sci. 24, 697–717 (2014)
100. Rudnicki, R., Tomski, A.: On a stochastic gene expression with pre-mRNA, mRNA and
protein contribution. J. Theoret. Biol. 387, 54–67 (2015)
101. Rudnicki, R., Tyran-Kamińska, M.: Piecewise deterministic Markov processes in biological
models. Semigroups of Operators Theory And Applications. Springer Proceedings in Math-
ematics & Statistics 113, pp. 235–255. Springer, Cham (2015)
102. Rudnicki, R., Wieczorek, R.: Mathematical models of phytoplankton dynamics. In: Russo R.
(ed.) Aquaculture I. Dynamic Biochemistry, Process Biotechnology and Molecular Biology,
vol. 2 (Special Issue 1), pp. 55–63 (2008)
103. Rudnicki, R., Wieczorek, R.: Phytoplankton dynamics: from the behaviour of cells to a trans-
port equation. Math. Model. Nat. Phenom. 1, 83–100 (2006)
104. Rudnicki, R., Zwoleński, P.: Model of phenotypic evolution in hermaphroditic populations.
J. Math. Biol. 70, 1295–1321 (2015)
105. Sharpe, F.R., Lotka, A.J.: A problem in age-distributions. Philos. Mag. 21, 435–438 (1911)
106. Singh, A., Hespanha, J.P.: Stochastic hybrid systems for studying biochemical processes.
Philos. Trans. R. Soc. A 368, 4995–5011 (2010)
107. Stein, R.B.: A theoretical analysis of neuronal variability. Biophys. J. 5, 173–194 (1965)
108. Stein, R.B.: Some models of neuronal variability. Biophys. J. 7, 37–68 (1967)
109. Stroock, D.W.: Some stochastic processes which arise from a model of the motion of a
bacterium. Z. Wahrscheinlichkeitstheorie verw. Gebiete 28, 305–315 (1974)
110. Tuckwell, H.C.: Stochastic Processes in the Neurosciences. SIAM, Philadelphia (1989)
111. Tuckwell, H.C.: Synaptic transmission in a model for stochastic neural activity. J. Theoret.
Biol. 77, 65–81 (1979)
112. Tyson, J.J., Hannsgen, K.B.: Cell growth and division: a deterministic/probabilistic model of
the cell cycle. J. Math. Biol. 23, 231–246 (1986)
113. Tyran-Kamińska, M.: Substochastic semigroups and densities of piecewise deterministic
Markov processes. J. Math. Anal. Appl. 357, 385–402 (2009)
114. Tyran-Kamińska, M.: Ergodic theorems and perturbations of contraction semigroups. Studia
Math. 195, 147–155 (2009)
115. Tyrcha, J.: Asymptotic stability in a generalized probabilistic/deterministic model of the cell
cycle. J. Math. Biol. 26, 465–475 (1988)
116. Voigt, J.: On substochastic C0 -semigroups and their generators. Transp. Theory Statist. Phys.
16, 453–466 (1987)
117. Voigt, J.: On resolvent positive operators and positive C0 -semigroups on AL-spaces. Semi-
group Forum 38, 263–266 (1989)
118. Webb, G.W.: Theory of Nonlinear Age-Dependent Population Dynamics. Marcel Dekker,
New York (1985)
119. Yin, G.G., Zhu, C.: Hybrid Switching Diffusions: Properties and Applications. Stochastic
Modelling and Applied Probability, vol. 63. Springer, New York (2010)
120. Yosida, K.: On the differentiability and the representation of one-parameter semi-group of
linear operators. J. Math. Soc. Jpn 1, 15–21 (1948)
121. Yu, J., Xiao, J., Ren, X., Lao, K., Xie, X.: Probing gene expression in live cells, one protein
molecule at a time. Science 311, 1600–1603 (2006)
164 References
122. Zeiser, S., Franz, U., Müller, J., Liebscher, V.: Hybrid modeling of noise reduction by a
negatively autoregulated system. Bull. Math. Biol. 71, 1006–1024 (2009)
123. Zeiser, S., Franz, U., Liebscher, V.: Autocatalytic genetic networks modeled by piecewise-
deterministic Markov processes. J. Math. Biol. 60, 207–246 (2010)
Index
A Characteristics of PDMP, 32
Absolutely continuous Closable operator, 76
function, 153 Closed operator, 64
measure, 154 Closure of a linear operator, 76
Active boundary, 31 Coagulation-fragmentation process, 28
Adiabatic limit, 139 Compact operator, 93
Adjoint Complete metric space, 35
operator, 38, 39, 42, 64, 76, 93, 94 Composition operator, 42
semigroup, 95 Compound Poisson process, 3, 53
Age- structured population, 24 Condition
Asymptotic stability, 116 (I), 130
Asynchronous exponential growth, 143 (K), 120, 129, 134, 135, 138, 141, 144
(WI), 123, 131, 134, 136, 137, 141, 145
(WT), 124
B Conditional
Banach space, 63 distribution, 35
Birth expectation, 34, 154, 155
pure process, 3 independence, 156
rate, 2 probability, 155
Birth–death process, 2, 130, 131 Cone, 37
Borel Conservative semigroup, 119
σ-algebra, 149 Continuous operator, 64
set, 149 Continuous-time
space, 35 Markov chain, 2, 96, 129
Bounded operator, 63 stochastic process, 49
Contraction
operator, 64
C semigroup, 65
Càdlàg, 10, 30, 49 Core, 77
Canonical space, 48 C0 -semigroup, 65
Cell cycle model, 7, 144
Chapman–Kolmogorov equation, 45, 49
D
Death
pure process, 3
rate, 2
© The Author(s) 2017 165
R. Rudnicki and M. Tyran-Kamińska, Piecewise Deterministic Processes
in Biological Models, SpringerBriefs in Mathematical Methods,
DOI 10.1007/978-3-319-61295-9
166 Index
L
H Law of large numbers, 154
Hahn–Banach theorem, 64 Lebesgue
Hasminskiı̆ function, 125–127 integral, 151
Index 167