Prob. Distri.
Prob. Distri.
The collection of pairs [xi, p(xi)], i = 1,2,…, is called the probability distribution
of X, and p(xi) is called the probability mass function (pmf) of X.
4
Continuous Random Variables [Probability Review]
3. f ( x) 0, if x is not in RX
• Properties
x0
1. P( X x0 ) 0, because x0
f ( x)dx 0
2. P(a X b) P (a X b) P (a X b) P (a X b)
5
Continuous Random Variables [Probability Review]
1 x/2
e , x 0
f ( x) 2
0, otherwise
• All probability questions about X can be answered in terms of the cdf, e.g.:
7
Cumulative Distribution Function [Probability Review]
– The probability that the device lasts for less than 2 years:
P (0 X 2) F (2) F (0) F (2) 1 e 1 0.632
8
Expectation [Probability Review]
– If X is continuous
E ( x) xf ( x)dx
9
Expectations [Probability Review]
1 2 x/2 x/2
E ( X ) x e dx x e
2 2 2 xe x / 2 dx 8
2 0 0
• Hence, the variance and standard deviation0 of the device’s life are:
V ( X ) 8 2 2 4
V ( X ) 2
10
Useful Statistical Models
• In this section, statistical models appropriate
to some application areas are presented. The
areas include:
– Queueing systems
– Inventory and supply-chain systems
– Reliability and maintainability
– Limited data
11
Queueing Systems [Useful
Models]
• In a queueing system, interarrival and service-time patterns
can be probablistic
– For more queueing examples, see Chapter 2; Chapter 6 is all queueing systems
12
Inventory and supply chain [Useful
Models]
• In realistic inventory and supply-chain systems, there are at
least three random variables:
– The number of units demanded per order or per time period
– The time between demands
– The lead time
13
Reliability and maintainability
[Useful Models]
14
Other areas [Useful Models]
15
Discrete Distributions
• Discrete random variables are used to
describe random phenomena in which only
integer values can occur.
• In this section, we will learn about:
– Bernoulli trials and Bernoulli distribution
– Binomial distribution
– Geometric and negative binomial distribution
– Poisson distribution
16
Bernoulli Trials
and Bernoulli Distribution [Discrete
• Bernoulli Trials: Dist’n]
– Consider an experiment consisting of n trials, each can be a success or
a failure.
• Let Xj = 1 if the jth experiment is a success
• and Xj = 0 if the jth experiment is a failure
– The Bernoulli distribution (one trial):
p, x j 1, j 1,2,..., n
p j ( x j ) p ( x j ) 1 p q, x j 0 ,j 1,2 ,...,n
0, otherwise
– where E(Xj) = p and V(Xj) = p (1-p) = p q
• Bernoulli process:
– The n Bernoulli trials where trails are independent:
p(x1,x2,…, xn) = p1(x1) p2(x2) … pn(xn)
17
Binomial Distribution [Discrete Dist’n]
18
Geometric & Negative
Binomial Distribution [Discrete
Dist’n]
• Geometric distribution
– The number of Bernoulli trials, X, to achieve the 1st success:
q x 1 p, x 0,1,2,..., n
p ( x)
0, otherwise
– E(x) = 1/p, and V(X) = q/p2
19
Poisson Distribution [Discrete
Dist’n]
• Poisson distribution describes many random processes quite
well and is mathematically quite simple.
– where a > 0, pdf and cdf are:
e x x
e i
p ( x) x! , x 0,1,... F ( x)
i 0 i!
0, otherwise
– E(X) = a = V(X)
20
Poisson Distribution [Discrete
Dist’n]
21
Continuous Distributions
• Continuous random variables can be used to
describe random phenomena in which the
variable can take on any value in some
interval.
• In this section, the distributions studied are:
– Uniform
– Exponential
– Normal
– Weibull
– Lognormal 22
Uniform Distribution
[Continuous Dist’n]
• A random variable X is uniformly distributed on the interval
(a,b), U(a,b), if its pdf and cdf are:
0, x a
1 x a
, a x b
f ( x) b a F ( x) , a x b
0, otherwise b a
1, x b
• Properties
– P(x1 < X < x2) is proportional to the length of the interval [F(x2) – F(x1) =
(x2-x1)/(b-a)]
– E(X) = (a+b)/2 V(X) = (b-a)2/12
• U(0,1) provides the means to generate random numbers,
from which random variates can be generated.
23
Exponential Distribution [Continuous
Dist’n]
• A random variable X is exponentially distributed with
parameter l > 0 if its pdf and cdf are:
e x , x 0 0, x 0
f ( x) F ( x) x t
0, elsewhere
0
e dt 1 e x , x 0
• Memoryless property
– For all s and t greater or equal to 0:
P(X > s+t | X > s) = P(X > t)
– Mean:
– Variance: 2 0
– Denoted as X ~ N(m,s2)
• Special properties:
– lim x f ( x) 0, and lim x f (. x) 0
– f(m-x)=f(m+x); the pdf is symmetric about m.
– The maximum value of the pdf occurs at x = m; the mean and mode are
equal.
26
Normal Distribution [Continuous
Dist’n]
• Evaluating the distribution:
– Use numerical methods (no closed form)
– Independent of m and s, using the standard normal distribution:
Z ~ N(0,1)
– Transformation of variables: let Z = (X - m) / s,
x
F ( x) PX x P Z
( x ) / 1 z2 / 2
e dz
2
( x ) /
( z )dz ( x ) z
, where ( z )
1 t2 / 2
e dt
2
27
Normal Distribution [Continuous
Dist’n]
• Example: The time required to load an oceangoing vessel, X, is
distributed as N(12,4)
– The probability that the vessel is loaded in less than 10 hours:
10 12
F (10) ( 1) 0.1587
2
• Using the symmetry property, F(1) is the complement of F (-1)
28
Weibull Distribution [Continuous
Dist’n]
• A random variable X has a Weibull distribution if its pdf has the form:
x 1
x
f ( x)
exp , x
0, otherwise
• 3 parameters:
– Location parameter: u, ( )
– Scale parameter: b , (b > 0)
– Shape parameter. a, (> 0)
• Example: u = 0 and a = 1:
When b = 1,
X ~ exp(l = 1/a)
29
Lognormal Distribution [Continuous
Dist’n]
• A random variable X has a lognormal distribution if its pdf has
the form:
1 ln x μ 2 m=1, s2=0.5,1,2.
exp , x 0
f ( x) 2π σx 2σ 2
0, otherwise
30
Poisson Process
• Definition: N(t) is a counting function that represents the
number of events occurred in [0,t].
• A counting process {N(t), t>=0} is a Poisson process with
mean rate l if:
– Arrivals occur one at a time
– {N(t), t>=0} has stationary increments
– {N(t), t>=0} has independent increments
• Properties
e t ( t ) n
P[ N (t ) n] , for t 0 and n 0,1,2,...
n!
– Equal mean and variance: E[N(t)] = V[N(t)] = lt
– Stationary increment: The number of arrivals in time s to t is also
Poisson-distributed with mean l(t-s)
31
Interarrival Times [Poisson
Process]
• Consider the interarrival times of a Possion process (A1, A2, …), where Ai is
the elapsed time between arrival i and arrival i+1
– The 1st arrival occurs after time t iff there are no arrivals in the interval [0,t],
hence:
P{A1 > t} = P{N(t) = 0} = e-lt
P{A1 <= t} = 1 – e-lt [cdf of exp(l)]
– Interarrival times, A1, A2, …, are exponentially distributed and independent with
mean 1/l
32
Splitting and Pooling [Poisson
Dist’n]
• Splitting:
– Suppose each event of a Poisson process can be classified as Type I,
with probability p and Type II, with probability 1-p.
– N(t) = N1(t) + N2(t), where N1(t) and N2(t) are both Poisson processes
with rates l p and l (1-p)
lp N1(t) ~ Poi[lp]
l
N(t) ~ Poi(l)
N2(t) ~ Poi[l(1-p)]
l(1-p)
• Pooling:
– Suppose two Poisson processes are pooled together
– N1(t) + N2(t) = N(t), where N(t) is a Poisson processes with rates l1 +
l
l2 N1(t) ~ Poi[l ]
1 1 l + l
1 2
N(t) ~ Poi(l + l )
1 2
N2(t) ~ Poi[l ] l
2 2
33
Nonstationary Poisson
Process (NSPP) [Poisson Dist’n]
• Poisson Process without the stationary increments, characterized by l(t), the
arrival rate at time t.
• The expected number of arrivals by time t, L(t):
t
Λ(t) λ(s)ds
0
• Relating stationary Poisson process n(t) with rate l=1 and NSPP N(t) with
rate l(t):
– Let arrival times of a stationary process with rate l = 1 be t1, t2, …, and
arrival times of a NSPP with rate l(t) be T1, T2, …, we know:
ti = L(Ti)
Ti = L-1(ti)
34
Nonstationary Poisson
Process (NSPP) [Poisson Dist’n]
• Example: Suppose arrivals to a Post Office have rates 2 per minute from 8
am until 12 pm, and then 0.5 per minute until 4 pm.
• Let t = 0 correspond to 8 am, NSPP N(t) has rate function:
2, 0 t 4
(t )
0.5, 4 t 8
Expected number of arrivals by time t:
2t , 0 t 4
(t ) 4 t t
0
2 ds 4 0 .5 ds
2
6, 4 t 8
35
Empirical Distributions [Poisson
Dist’n]
• A distribution whose parameters are the observed values in a
sample of data.
– May be used when it is impossible or unnecessary to establish that a
random variable has any particular parametric distribution.
– Advantage: no assumption beyond the observed values in the sample.
– Disadvantage: sample might not cover the entire range of possible
values.
36