0% found this document useful (0 votes)
209 views41 pages

Probability and Statistics With Reliability, Queuing and Computer Science Applications: Chapter 7 On

The document summarizes key concepts about discrete time Markov chains (DTMCs). It defines DTMCs and their properties, including: (1) the Markov property that the future state depends only on the present state; (2) transition probabilities between states; (3) homogeneous and non-homogeneous DTMCs; (4) state classification into transient, recurrent, absorbing, and periodic states; and (5) properties of irreducible DTMCs such as having a unique stationary distribution. Performance measures like expected cycle time and throughput are also discussed.

Uploaded by

Akshta Kaushal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
209 views41 pages

Probability and Statistics With Reliability, Queuing and Computer Science Applications: Chapter 7 On

The document summarizes key concepts about discrete time Markov chains (DTMCs). It defines DTMCs and their properties, including: (1) the Markov property that the future state depends only on the present state; (2) transition probabilities between states; (3) homogeneous and non-homogeneous DTMCs; (4) state classification into transient, recurrent, absorbing, and periodic states; and (5) properties of irreducible DTMCs such as having a unique stationary distribution. Performance measures like expected cycle time and throughput are also discussed.

Uploaded by

Akshta Kaushal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
You are on page 1/ 41

Probability and Statistics with

Reliability, Queuing and Computer


Science Applications: Chapter 7 on
Discrete Time Markov Chains

Kishor S. Trivedi
Visiting Professor

Dept. of Computer Science And Engineering.


Indian Institute of Technology, Kanpur
Discrete Time Markov Chain
 Dynamic evolution is such that future depends only on the present
(past is irrelevant); can depend on time step.
 Markov Chain  Discrete state space.
 DTMC : time (index) is also discrete, i.e., system is observed only
at discrete epochs of time.
 X0, X1, .., Xn, .. :observed state at discrete times, t0, t1,..,tn, ..
 Xn = j  system state at time step n is j.
 P(Xn = in| X0 = i0, X1 = i1, …, Xn-1 = in-1)
= P(Xn = in| Xn-1 = in-1) (Markov Property)

 pjk(m,n)  P(Xn = k | Xm = j) , (conditional pmf)

 pj(n)  P(Xn = j) (unconditional pmf) (first order)


Transition Probabilities
 pjk(m,n): transition probability function of a DTMC.
 Homogeneous DTMC: pjk(m,n) = pjk(n-m)
 1-step transition prob, p = p (1) = P(X = k| X
jk jk n n-1 = j) ,
 Assuming 0-step transition prob as:

 Joint pmf (nth order) : P(X0 = i0, X1 = i1, …, Xn = in)


= P(X0 = i0, X1 = i1, …, Xn-1 = in-1). P(Xn = in| X0 = i0, X1 = i1, …, Xn-1 =
in-1)
= P(X0 = i0, X1 = i1, …, Xn-1 = in-1). P(Xn = in| Xn-1 = in-1) (due to
Markov property)
= P(X0 = i0, X1 = i1, …, Xn-1 = in-1).pin-1, in
: = pi0(0)pi0, i1 (1) …pin-1, in (1) = pi0(0)pi0, i1 …pin-1, in = pi0(0)pi0, in(n)
The Beauty of Markov Chains
 Given the initial probabilities
 And
 the repeated use of one-step transition
probabilities
 Or n-step transition probability
 We can determine the nth order pmf for
all n
Transition Probability Matrix
 The initial prob. pi0(0) = P(X0 = i0 ). In general,
 p0(0) = P(X0 = 0 ), …, pk(0) = P(X0 = k ) etc, or,
 p(0) = [p0(0), p1(0), … ,pk(0), ….] (initial prob. row
vector)
 Let the transition probability Matrix (TPM):

 Sum of ith row elements pi,0(0)+ pi,1(0)+ … ?

 Such a square matrix with probabilities as entries and with


row sums =1 is called a stochastic matrix (prop)
State Transition Diagram
 Node with labels i, j etc. and arcs labeled
pij i
p
j
ij
 Example: 2-state DTMC for a cascade of
binary comm. channels. Signal values: ‘0’
or ‘1’ form the state values.
xn-1 = 0 1-a xn = 0 1-a 1-b
a
a

b
0 1

xn-1 = 1 1-b xn = 1
b
(n-1)th stage nth stage
Unconditional Probability
 Finding unconditional pmf:
n-Step Transition Probability
 For a DTMC, find
 Events: state reaches k (from i) & reaches j (from k)
are independent due to the Markov property (i.e. no
history)
pik(m)
pkj(n) j
k

0 m m+n

 Invoking the theorem of total probability :

 Let P(n) : n-step prob. transition matrix (i,j) entry is


pij(n). Making m=1, n=n-1 in the above equation,
Marginal (unconditional) pmf
 Quite often we are not interested in the
joint pmf
 But only the marginal pmf at step n
 Given the initial pmf
 And either the 1-step TPM or the n-step
TPM
 Find the marginal pmf at step n
Marginal (unconditional) pmf

 j, in general can assume countable values, 0,1,2, …. Defining,

 pj(n) for j=0,1,2,..,k,… can be written in the vector form as,

 Or,

 P n can be easily computed if I is finite. However, if I is countably


infinite, it may be difficult to compute P n (and p(n) ).
Marginal pmf Example
 For a 2-state DTMC described by its 1-step transition prob.
matrix,

the n-step transition prob. Matrix is given by,

 Proof follows easily by using induction, that is, assuming that the
above is true for Pn-1. Then,
Pn = P. Pn-1
Computing Marginal pmf
 Example of a cascade of digital comm. channels: each stage described
by a 2-state DTMC, We want to find p(n) (a=0.25 & b=0.5),

 The ’11’ element for n=2 and n=3 are,

 Assuming initial pmf as, p(0) = [p0(0) p1(0)] = [1/3 2/3] gives,

 What happens to Pn as n becomes very large ( infinity)?


DTMC State Classification
 From the previous example, as n approaches infinity, pij(n) becomes
independent of n and i ! Specifically,

 Not all Markov chains exhibit such a behavior.


 State classification may be based on the distinction that:
 Average number of visits to some states may be infinite while other
states may be visited only a finite number of times (on average)
 Transient state: if there is non-zero probability that the system will
NOT return to this state (or the average number of visits is finite).
 Define Xji to be the # of visits to state i, starting from state j, then,

 For a transient state (i), visit count needs to finite, which requires pji(n)
 0 as n  infinity
DTMC State Classification
(contd.)
 State i is a said to be recurrent if, starting from state i, the process
eventually returns to the state i with probability 1.

 For a recurrent state, time-to-return is a relevant measure. Define fij(n) as


the cond. prob. that the first visit to j from i occurs in exactly n steps.
 If j = i, then fii(n) denotes the prob. of returning to i in exactly n steps.
 Known result:

 Let,

 Mean recurrence time for state i is


Recurrent state

 Let i be recurrent and pii(n) > 0, for some n > 0.
 For state i, define period di as GCD of all such +ve n’s that
result in pii(n) > 0
 If di=1,  aperiodic and if di>1, then periodic.
 Absorbing state: state i absorbing if pii=1.
 Communicating states: i and j are said to be communicating if
there exist directed paths from i and j and from j and i.
 Closed set of states: A commutating set of states C forms a
closed set, if no state outside of C can be reached from any
Closed set c1

state in C.

Closed set c2
Closed set ck
Irreducible Markov Chains
 Markov chain states can be partitioned into k distinct subsets:
c1, c2, .., ck-1, ck , such that
 ci, i=1,2,..k-1 are closed set of recurrent nun-null states.
 ck is the set of all transient states.
 If ci contains only one state, then ci is an absorbing state
 If k=2 & ck empty, then c1 forms an irreducible Markov chain
 Irreducible Markov chain: is one in which every state can be
reached from every other state in a finite no. of steps, i.e., for
all i,j ε I, for some integer n > 0, pij(n) > 0. Examples:
 Cascade of digital comm. channels DTMC is irreducible
 0 1
Irreducible DTMC (contd.)
 If one state of an irreducible DTMC is recurrent aperiodic, then
so are all the other states. Same result if periodic or transient.
 For a finite aperiodic irreducible Markov chain, pij(n) becomes
independent of i and n as n goes to infinity.

• All rows of Pn become identical


Irreducible DTMC (contd.)
 Law of total probability gives,

 Substitute in the 1st equation to get,

 Or in the vector-matrix form,

 Since v is a probability vector, we impose

 Self reading exercise (theorems on pp. 351)


 For an aperiodic, irreducible, finite state DTMC,
Eigenvalue & Eigenvector
 λ is an eigenvalue of P iff
det(P- λI) = 0
 λ =1 is an eigenvalue of a stochastic
matrix P
 x is an eigenvector of P corresponding
to eigenvalue λ iff
x P=x λ
Measures of Interest
 Attach reward ri (cost or penalty) to state
i enabling computation of various
interesting measures
 The steady-state expected reward is the
weighted average of state probabilities:
Irreducible DTMC Example
 Typical computer program: continuous cycle of compute & I/O
q0 1
q1
1
q2 1
0 2
1
q3

 The resulting DTMC is irreducible with period =1. Then from,


Performance Measures
 Let tj be the time to execute node j in
the previous DTMC
 Expected cycle time is obtained as the
expected steady state reward by
assigning rj = tj
 Expected thruput is the reciprocal of the
expected cycle time
Sojourn Time; HDTMC
 If Xn = i, then Xn+1 = j should depend only on the current
state i, and not on the time spent in state i.
 Let Ti be the time spent in state i, before moving to state j
 DTMC will remain in state i at the next step with prob. pii
and,

 Next step (n+1), BT, ‘0’ Xn+1 = i, ‘1’Xn+1 # i


 Then Ti is the number of trials up to and including the first
success :
Bernoulli Arrival Process
 Many systems can be considered as discrete-
time queues
 Instead of a Poisson arrival process, we can
use a Bernoulli arrival process
 At every time step we have an arrival with
probability c and no arrival with prob. 1-c
 Generalize to MMBP, non-homogeneous BP,
generalized BP
Markov Modulated Bernoulli
Process (MMBP)
 Generalization of a Bernoulli process: the Bernoulli process
parameter is controlled by a DTMC.
 Simplest case is Binary state (on-off) modulation
 ‘On’ Bernoulli parameter = c0; ‘Off’  c1’ (or =0)
1-a 1-b
a

0 1

 Modulating process is an irreducible DTMC, and,

 Reward assignment, r0 = c0 and r1=c1. Then cell arrival prob. is


Slotted ALOHA DTMC
Backlogged Requests
New Requests
 New and backlogged requests +

 Successful channel access if:


1. Exactly one new req. and no backlogged req.
2. Exactly one backlogged req. and no new req.

DTMC state: # of backlogged requests .
backlogged
new n
++
m-n + x x
x
+

Σ Channel
Slotted Aloha contd.

• In a particular state n, successful contention occurs with prob. rn

•rn may be assigned as a reward for state n.


Software Performance Analysis
 Control structure point of view
 Chapter 5, also later in chapter 7
 Data structure point of view
 Stacks, queues, trees etc.
 Probability of insertion b, probability of
deletion d (generalized BP)
 Keep track of the number of items in the
data structure (can be a vector)
Discrete-time Birth-Death Process
 Special type of DTMC in which the TPM P is tri-
diagonal
DTMC solution steps

Solving for v = vP, gives the steady state probabilities .
Data Structure Oriented Analysis
 Can consider finite storage space and
thence compute the probability of an
overflow
 Can be generalized two stacks sharing a
common storage space or not sharing
 More general data structures
 Can consider the elapsed time between
two requests to the data structure
Software Performance Analysis
 Back to the control structure
 But now allow arbitrary branching
 Consider the control flow graph as a
DTMC
 Consider a terminating application
DTMC with Absorbing States
 Example: Program having a set of interacting modules.
Absorbing state: completion

s1
0.6 0.4
0.2
s2 s3
0.6 0.4
s4

0.4 0.6 0.4


0.4

s5

1
DTMC with Absorbing States
 M contains useful information.
 Xij : rv denoting the number to visits to j starting from i
 E [Xij] = mij (for i, j = 1,2,…, n-1) . Need to prove this
statement.
 There are three distinct situations that can be enumerated

{
δij , occurs with prob. pij
si sj
Xij = Xkj + δij , occurs with prob. Pik
i

sk k=1,2,..n
(δij : term accounts for i=j case)
sn

 Let rv Y denote the state at step #2 (initial state: i)


 E[Xij| Y = n] = δij
 E[Xij| Y = k] = E[Xkj + δij]= E[Xkj]+ δij
DTMC with Absorbing States
 Since, P(Y=k) = pik , k=1,2,..n, total expectation rule gives,

 Over all (i,j) values, we need to work with the matrix,

 Therefore, fundamental matrix M elements give the expected #


of visits to state j (from i) before absorption.
 If the process starts in state “1”, then m1j gives the average #
of visits to state j (from the start state) before absorption.
Software Performance/Reliability
Analysis
 By assigning rewards to different states, a variety of
measures may be computed.
 Average time to execute a program
 s is the start state; r : execution time/visit for s
1 j j
 V = m
j 1j is the average # times statement block sj
is executed
 We need to calculate total expected execution time,

I.e. until the process gets absorbed into stop state


(s5 )

 Software reliability: Rj: Reliability of sj .Then,


Terminating Applications

Architecture: DTMC
pij  i j
Pr{transfer of control from module to module }
Ri
Failure behavior: component reliability
Solution method: Hierarchical
Compute the expected number
of times Vi each component is executed using

n
Vi   i   V j p ji
j 1
Equation (7.76)
Terminating Applications
Cont’d

V
 Ri i can be considered as the equivalent reliability
of the component that takes into account the
component utilization n
V
R   Ri i
i 1
 System reliability becomes
Architecture-Based
Analysis
Example
Terminating application
1  architecture described by
1 DTMC with transition
probability matrix P=[pij]
p23 2 p24
 component reliabilities are:

R1 R2 R3 R4 R5
3 1 4
0.999 0.980 0.990 0.995 0.999
1

5
Architecture-Based Analysis
Example (contd.)
Solution method - Hierarchical
Vi is a clear indication of
p24 0.8 0.2 component usage
V1 1 1  when p24=0.8 components 2
and 4 are invoked within a loop
V2 5 1.25 many times which results in
V3 1 1 a significantly higher expected
number of executions compared
V4 4 0.25 to the case when p24=0.2

V5 1 1 Application reliability is highly


R 0.87536 0.97218 dependent on the components
usage
Architecture-Based Analysis
Example (contd.)
Solution method - Composite
1 1-R1
R1 P24 0.8 0.2
1-R2
R 0.88056 0.96227
P23 R2 2
P24 R2

R4 1-R4
3 4 F
1-R3
R3
1-R5
5
R5 S

You might also like