Unit3-Probability and Stochastic Processes (18MAB203T)
Unit3-Probability and Stochastic Processes (18MAB203T)
Unit3-Probability and Stochastic Processes (18MAB203T)
Department of Mathematics ,
College of Engineering and Technology
SRM Institute of Science and Technology
Markov Inequality
Theorem
If X is any nonnegative random variable, then
E (X )
P (X ≥ c) ≤ , for any c > 0. (1)
c
Proof ( X is discrete):
Σ
E (X ) = x P (X = x )
Σ Σ
= x P (X = x ) + x P (X = x )
x≥ c x<c
Σ
≥ x P (X = x ) (for any c > 0)
x≥ c
Σ
≥ c P (X = x ) (since x ≥ c)
x≥ c
Σ
= c P (X = x )
x≥c
= c P (X ≥ c).
Hence,
E (X )
P (X ≥ c) ≤ .
c
Proof ( X is continuous):
∫ ∞
E (X ) = x f (x )dx
−∞
∫ ∞
= x f (x )dx (since X is positive-valued)
∫ 0∞
≥ x f (x )dx (for any c > 0)
c
∫ ∞
≥ c f (x )dx (since x > c in the integrated region)
c
∫ ∞
= c f (x )dx
c
= c P (X ≥ c).
Hence,
E (X )
P (X ≥ c) ≤ .
c
Other form of Markov’s inequality
When X is at least k times of mean E ( X ) ,
1
P (X ≥ k E(X)) ≤ , for any k > 0. (2)
k
Example 1
Suppose that the average grade on the upcoming math exam is 70%.
Give an upper bound of the proportion of the students who score at
least 90%.
E (X ) = 7 .
P (X ≥ 90) ≤
90 9
Example 2
Suppose X ∼ Binomial (100, 12 ). Find an upper bound of P (X ≥ 75).
E (X ) = 50 = 2 .
P (X ≥ 75) ≤
75 75 3
General form
If U ( X ) is a function of the random variable X , then
E (U (X ))
P (U (X ) ≥ c) ≤ , for any c > 0. (3)
c
Chebyshev’s Inequality
Theorem
If X is a random variable with E ( X ) = µ and Var (X) = σ 2 , then
σ2
P {|X −µ|≥ c} ≤ , for any c > 0. (4)
c2
or,
σ2
P {|X −µ|< c} ≥ 1 − 2 , for any c > 0. (5)
c
Alternative forms
If we put c = kσ, where k > 0, then If X is a random variable with
E ( X ) = µ and Var (X) = σ 2 , then
1
P {|X −µ|≥ kσ} ≤ , (6)
k2
or,
1
P {|X −µ|< kσ} ≥ 1 − . (7)
k2
Example 3
If X is a RV with E ( X ) = 3 and E ( X 2 ) = 13. Find the lower bound
for P (−2 < X < 8).
σ 2 = E (X 2 ) −{ E (X )} 2 = 4
σ2 21
P (−2 < X < 8) = P (|X −3|< 5) ≥ 1 − =
52 25
Example 4
Two dice are thrown once, if X is the sum of the number showing up,
prove that P ( | X −7|≥ 3) ≤ 35
54
and compare this value with exact
probability.
X 2 3 4 5 6 7 8 9 10 11 12
1 2 3 4 5 6 5 4 3 2 1
P (X) 36 36 36 36 36 36 36 36 36 36 36
Σ
E (X ) = x P (X = x ) = 7
x
Σ 1974
E (X 2 ) = x 2 P (X = x ) =
x
36
35
Var(X) = σ 2 =
6
By Chebyshev’s inequality
35/6
P (|X −7|≥ 3) ≤ = 0.6481
9
Exact Probability:
Given that,
X
P 0.4 < n < 0.6 ≥ 0.9. (8)
σ2 σ2
P . X —µ . < c ≥ 1− = ⇒ P (µ −c <
X
< µ + c) ≥ 1 − 2
.n . c2 n c
σ2 1
P (|X −µ|> c) < = ⇒ P (|X −1|> 2) < .
c2 4
Actual probability:
= e−3.
Weak law of large number
Convergence in Probability
If P { | X n −X | > ϵ} → 0 as n → ∞, then the sequence of RVs { X n } is
said to convergence to X in probability.
Proof:
X1 X2 Xn
E ( X n) = E + E + ··· + E
n n n
= 1 { E (X 1 ) + E (X 2 ) + · · · + E (X n )} = µ.
n
X1 X2 Xn
Var(X n ) = Var + Var + · · · + Var
n n n
1 σ2
= {Var(X 1 ) + Var(X 2 ) + · · · + Var(X n )} = .
n 2 n
By Chebyshev’s inequality we have,
σ2
P (|X n −µ|> ϵ) <
nϵ2
= ⇒ P (|X n −µ|> ϵ) → 0 as n → ∞
Central Limit Theorem
Central Limit Theorem
Let X 1 , X 2 , X 3 , . . ., X n , . . . be the sequence of independent and
identically distributed random variables with E ( X i ) = µ < ∞ and
0 < Var(X i ) = σ 2 < ∞.
Then S n = X 1 + X 2 + · · · + X n follows a normal distribution with
mean nµ and variance nσ 2 as n tends to infinity, i.e., S n ∼ N (nµ, nσ 2 )
Example 7
Let X 1 , X 2 , X 3 , . . ., X 10 be independent Poisson random variables
with mean 1. Using CLT find P { ( X 1 + X 2 + · · · + X 1 0 ) ≥ 15}
E ( X i ) = var(X i ) = 1, S 1 0 = X 1 + X 2 + · · · + X 1 0 ∼ N (10, 10)
S 10 −10 15 −10
P ( S 10 ≥ 15) = P √ ≥ √
10 10
√
= P (z ≥ 1.58), where z = (S 10 −10)/ 10 ∼ N (0, 1)
= 0 . 5 - P ( 0 < Z < 1 . 5 8 ) = 0.0571
Example 8
A bank teller serves customers standing in the queue one by one.
Suppose that the service time X i for customer i has mean E ( X i ) = 2
and Var(X i ) = 1. Assume that service times for different bank
customers are independent. Let S be the time the bank teller spends
serving 50 customers. Find P (90 < S < 110).
Example 9
The life time of a certain brand of an electric bulb may be considered
as a random variable with mean 1200 hours and standard deviation
250 hours. Find the probability using central limit theorem, that the
average life time of 60 bulbs exceeds 1250 hours.
X 1 + X2 + · · · + n
→ µ as n → ∞,
X n
X 1 + X 2 + ···+ X n
i.e., P lim = µ = 1.
n→ ∞ n
One Sided Chebyshev’s Inequality
σ2
P (X ≥ µ + c) ≤ ,
σ2 + c2
σ2
P (X ≤ µ −c) ≤ ,
σ2 + c2
Example 11 (Markov inequality weaker than one-sided Chebyshev’s
inequality)
If the number of items produced in a factory during a week is a
random variable with mean 100 and variance 400, compute an upper
bound on the probability that this week’s production will be at least
120.
E (X ) = 5 ,
P (X ≥ 120) ≤
120 6
which is far weaker than the Chebyshev’s.
Cauchy Schwartz inequality
For any two random variables X and Y , we have
√
| E ( X Y )|≤ E ( X 2 ) E ( Y 2 ),
Example 12
Using the Cauchy-Schwartz inequality, show that for any two random
variables X and Y
|ρ(X , Y )|≤ 1.
X − E (X ) Y − E (Y )
Let U = σX
and V = σY
.
Then E ( U ) = E ( V ) = 0, and Var(U ) = Var(V ) = 1. Now applying
Cauchy Schwartz inequality for U and V , we get
√
|E (U V )|≤ E (U 2 )E (V 2 ) = ⇒ |ρ(X , Y )|≤ 1.
E { (X −E (X ))(Y −E (Y ))} )
as ρ(X , Y ) = = E(UV )
σXσY
Chernoff Bounds
Chernoff Bounds
If M (t) = E ( e t X ) be the moment generating function of the random
variable X , then
Jensen’s Inequality
If f (x ) is convex function, then
E (f (X )) ≥ f (E (X )),
Example 14
Show that the variance of a random variable X is non-negative.