9 CLT
9 CLT
Faculty of Engineering
Electrical and Electronics Engineering Department
Salma Elkawafi
[email protected]
Goals
Understand the following:
Probability Bounds
• The Union Bound and Extension
• Markov and Chebyshev Inequalities
Limit Theorems
• Low of large number
• Central Limit Theorem
Probability Bounds
Probability bounds: These inequalities usually apply to a
general scenario. There are several scenarios in which we
resort to inequalities. Sometimes we do not have enough
information to calculate a desired quantity (such as the
probability of an event or the expected value of a random
variable). In other situations, the problem might be
complicated and exact calculation might be very difficult. In
other scenarios, we might want to provide a result that is
general and applicable to wide range of problems.
Probability Bounds
The Union Bound and Extension:
The union bound or Boole's inequality is applicable when you
need to show that the probability of union of some events is
less than some value. Remember that for any two events A and
B we have
𝑛 𝑛
𝑃 ራ 𝐴𝑖 ≤ 𝑃(𝐴𝑖 )
𝑖=1 𝑖=1
Probability Bounds
The Union Bound and Extension:
Example:
Your friend tells you that he had four job interviews last week. He says that based on
how the interviews went, he thinks he has a 20% chance of receiving an offer from
each of the companies he interviewed with. Nevertheless, since he interviewed with
four companies, he is 90% sure that he will receive at least one offer. Is he right?
Solution:
Let 𝐴𝑖 be the event that your friend receives an offer from the 𝑖th company, 𝑖
= 1,2,3,4. Then, by the union bound:
4 4
Thus, we conclude
𝐸𝑋
𝑃 𝑋≥𝑎 ≤ , for any 𝑎 > 0.
𝑎
We can prove the above inequality for discrete similarly (using the generalized PDF).
Probability Bounds
Chebyshev Inequality
Let 𝑋 be any random variable. If you define 𝑌 = 𝑋 − 𝐸 𝑋 2 , then 𝑌 is a nonnegative
random variable, so we can apply Markov's inequality to 𝑌. In particular, for any positive
real number 𝑏, we have
𝐸𝑌
𝑃(𝑌 ≥ 𝑏2 ) ≤ 2
𝑏
But note that
𝐸 𝑌 = 𝐸[ 𝑋 − 𝐸𝑋 2 ] = 𝑉𝑎𝑟 𝑋 ,
𝑃(𝑌 ≥ 𝑏2 ) = 𝑃( 𝑋 − 𝐸 𝑋 2 ≥ 𝑏2 ) = 𝑃(|𝑋 − 𝐸𝑋| ≥ 𝑏).
Thus, we conclude that
𝑉𝑎𝑟 𝑋
𝑃(|𝑋 − 𝐸[𝑋]| ≥ 𝑏) ≤ .
𝑏2
Chebyshev's inequality states that the difference between 𝑋 and 𝐸𝑋 is somehow limited
by 𝑉𝑎𝑟(𝑋). This is intuitively expected as variance shows on average how far we are
from the mean
Probability Bounds
Markov and Chebyshev Inequalities
Markov's Inequality
If 𝑋 is any nonnegative random variable, then,
𝐸𝑋
𝑃 𝑋≥𝑎 ≤ , for any 𝑎 > 0.
𝑎
Chebyshev's Inequality
If X is any random variable, then for any b>0 we have
𝑉𝑎𝑟 𝑋
𝑃 𝑋−𝐸 𝑋 ≥𝑏 ≤ , for any 𝑏 > 0
𝑏2
Probability Bounds
Markov and Chebyshev Inequalities
Example:
Let 𝑋 ∼ 𝐵𝑖𝑛𝑜𝑚𝑖𝑎𝑙(𝑛, 𝑝). Using Markov's inequality, find an upper bound on 𝑃(𝑋 ≥ 𝛼𝑛),
1 3
where 𝑝 < 𝛼 < 1. Evaluate the bound for 𝑝 = 2 and 𝛼 = 4 .
Solution:
𝐸𝑋
𝑃 𝑋≥𝑎 ≤ ,
𝑎
𝐸 𝑋 = 𝑛𝑝
𝑎 = 𝛼𝑛
𝑛𝑝 𝑝
𝑃 𝑋 ≥ 𝛼𝑛 ≤ = ,
𝛼𝑛 𝛼
Definition: For i.i.d. random variables 𝑋1 , 𝑋2 , … , 𝑋𝑛 , the sample mean, denoted by 𝑋ത , is defined as
𝑋 +𝑋 +⋯.+𝑋𝑛
𝑋ത = 1 2 𝑛
𝑋ത is a random variable
Another common notation for the sample mean is 𝑀𝑛 (𝑋).
𝐸 𝑋 +𝐸 𝑋2 +⋯.+𝐸 𝑋𝑛 𝑛𝐸[𝑋]
𝐸 𝑋ത = 1 𝑛
= 𝑛
= 𝐸[𝑋]
Using Chebyshev's Inequality: (If 𝑋 is any random variable, then for any 𝑏 > 0 we have)
𝑉𝑎𝑟 𝑋
𝑃(|𝑋 − 𝐸[𝑋]| ≥ 𝑏) ≤
𝑏2
𝑉𝑎𝑟 𝑋ത 𝑉𝑎𝑟 𝑋
𝑃 𝑋ത − 𝜇 ≥ 𝜖 ≤ =
𝜖2 𝑛𝜖 2
Central Limit Theorem
The central limit theorem (CLT) is one of the most important results in probability theory. It
states that, under certain conditions, the sum of a large number of random variables is
approximately normal.
Suppose that 𝑋1 , 𝑋2 , … , 𝑋𝑛 are i.i.d. random variables with expected values 𝐸[𝑋𝑖 ] = 𝜇 < ∞
𝑋 +𝑋 +⋯.+𝑋𝑛
and variance 𝑉𝑎𝑟(𝑋𝑖 ) = 𝜎 2 < ∞, and the sample mean 𝑋ത = 1 2 𝑛
𝜎2
has mean 𝐸[𝑋]
ത = 𝜇 and variance 𝑉𝑎𝑟(𝑋)
ത = . Then, the random variable
𝑛
𝑋ത − 𝜇 𝑋1 + 𝑋2 + ⋯ . +𝑋𝑛 − 𝑛𝜇
𝑍𝑛 = =
𝜎/ 𝑛 𝑛𝜎
converges in distribution to the standard normal random variable as 𝑛 goes to infinity, that is
lim 𝑃(𝑍𝑛 ≤ 𝑥) = Φ(𝑥), 𝑓𝑜𝑟 𝑎𝑙𝑙 𝑥 ∈ 𝑅,
𝑛→∞
where Φ(𝑥) is the standard normal CDF.
Central Limit Theorem𝑍 = 𝑋ത − 𝜇 = 𝑋 + 𝑋 + ⋯ . +𝑋
𝑛
1 2 𝑛 − 𝑛𝜇
𝜎/ 𝑛 𝑛𝜎
• Central Limit Theorem
Let's assume that 𝑋𝑖 ’s are 𝐵𝑒𝑟𝑛𝑜𝑢𝑙𝑙𝑖 𝑝 .
Then
𝐸 𝑋𝑖 = 𝑝,
𝑉𝑎𝑟(𝑋𝑖) = 𝑝(1 − 𝑝).
Also, 𝑌𝑛 = 𝑋1 + 𝑋2 + ⋯ + 𝑋𝑛 has
𝐵𝑖𝑛𝑜𝑚𝑖𝑎𝑙(𝑛, 𝑝) distribution.
Thus,
𝑌𝑛 − 𝑛𝑝
𝑍𝑛 =
𝑛𝑝(1 − 𝑝)
where 𝑌𝑛 ∼ 𝐵𝑖𝑛𝑜𝑚𝑖𝑎𝑙(𝑛, 𝑝)
Central Limit Theorem
• Central Limit Theorem
How to Apply The Central Limit Theorem (CLT)
❑ Write the random variable of interest, 𝑌, as the sum of n random variable 𝑋𝑖 ′s:
𝑌 = 𝑋1 + 𝑋2 +. . . +𝑋𝑛
𝑌−𝐸[𝑌] 𝑌−𝑛𝜇
❑ According to the CLT, conclude that = 𝑛𝜎
is approximately standard normal; thus, to
𝑉𝑎𝑟 𝑌
find 𝑃(𝑦1 ≤ 𝑌 ≤ 𝑦2 ), we can write
𝑦1 − 𝑛𝜇 𝑌 − 𝑛𝜇 𝑦2 − 𝑛𝜇 𝑦2 − 𝑛𝜇 𝑦1 − 𝑛𝜇
𝑃(𝑦1 ≤ 𝑌 ≤ 𝑦2 ) = 𝑃( ≤ ≤ ) ≈ Φ( ) − Φ( ).
𝑛𝜎 𝑛𝜎 𝑛𝜎 𝑛𝜎 𝑛𝜎
Central Limit Theorem
Example
A bank teller serves customers standing in the queue one by one. Suppose that the service time 𝑋𝑖 for
customer 𝑖 has mean 𝐸[𝑋𝑖 ] = 2 (minutes) and 𝑉𝑎𝑟(𝑋𝑖 ) = 1. We assume that service times for different
bank customers are independent. Let 𝑌 be the total time the bank teller spends serving 50 customers.
Find 𝑃(90 < 𝑌 < 110)
Solution
𝑦1 − 𝑛𝜇 𝑌 − 𝑛𝜇 𝑦2 − 𝑛𝜇 𝑦2 − 𝑛𝜇 𝑦1 − 𝑛𝜇
𝑃(𝑦1 ≤ 𝑌 ≤ 𝑦2 ) = 𝑃( ≤ ≤ ) ≈ Φ( ) − Φ( ).
𝑛𝜎 𝑛𝜎 𝑛𝜎 𝑛𝜎 𝑛𝜎
𝑌 = 𝑋1 + 𝑋2 +. . . +𝑋𝑛 ,
Where,
𝑛 = 50,
𝐸 𝑋𝑖 = 𝜇 = 2,
𝑉𝑎𝑟 𝑋𝑖 = 𝜎 2 = 1.
90 − 100 𝑌 − 𝑛𝜇 110 − 100
𝑃 90 < 𝑌 < 110 = 𝑃 < <
50 𝑛𝜎 50
𝑌−𝑛𝜇
By the CLT, is approximately standard normal, so we can write
𝑛𝜎
𝑃 90 < 𝑌 < 110 ≈ Φ − 2 − Φ 2 = 0.8427
Central Limit Theorem
Example
In a communication system each data packet consists of 1000 bits. Due to the noise, each bit may be
received in error with probability 0.1. It is assumed bit errors occur independently. Find the probability
that there are more than 120 errors in a certain data packet.
Solution
Let us define 𝑋𝑖 as the indicator random variable for the 𝑖th bit in the packet. That is, 𝑋𝑖 = 1 if the ith bit
is received in error, and 𝑋𝑖 = 0 otherwise. Then the 𝑋𝑖 ′s are i.i.d. and 𝑋𝑖 ∼ 𝐵𝑒𝑟𝑛𝑜𝑢𝑙𝑙𝑖(𝑝 = 0.1). If 𝑌 is
the total number of bit errors in the packet, we have
𝑌 = 𝑋1 + 𝑋2 +. . . +𝑋𝑛 ,