Module 5
Module 5
MB+SCP
Mani Bhushan, Sachin Patwardhan
Department of Chemical Engineering,
Indian Institute of Technology Bombay
Mumbai, India- 400076
mbhushan,[email protected]
Spring 2021
X1 + · · · + Xn E [X1 ] + · · · + E [Xn ]
E [X̄ ] = E =
n n
nµ
= =µ
n
MB+SCP (IIT Bombay) CL202 Spring 2021 10 / 42
Why do we infer µ using the sample mean?
E [X̄ ] = µ
X̄ is an unbiased estimator of µ.
For a statistic θ̂ to be an unbiased estimator of θ,
E [θ̂] = θ
σ2
E [X̄ ] = µ var (X̄ ) =
n
σ2
P (|X − µ| ≥ k) ≤
k2
1
P (|X − µ| ≥ kσ) ≤ 2
k
Allows derivation of probability bounds when only mean and variance
are known.
Theorem
Let X1 , X2 , ..., Xn be a sequence of independent and identically
distributed random variables each having mean µ and variance σ 2 .
Then for large n, the distribution of
X1 + X2 + .... + Xn
is approximately normal with mean nµ and variance nσ 2 .
Thus,
!
X − (450)(0.3) 150.5 − (450)(0.3
P(X > 150.5) = P p ≥ p
(450)(0.3)(0.7) (450)(0.3)(0.7)
CLT does not tell us how large sample size n needs to be for the
normal approximation of X̄ to be valid.
n depends on population distribution of the sample data.
For binomial np(1 − p) ≥ 10, for normal any n ≥ 1 is ok.
Rule of thumb: sample size n ≥ 30 works for almost all
distributions, i.e. no matter how nonnormal the underlying
population is, the sample mean of a sample of size atleast 30
will be approximately normal.
In most cases, normal approximation will be valid for much
smaller sample sizes.
(Xi − X̄ )2 = (Xi − µ + µ − X̄ )2
= (Xi − µ)2 + (µ − X̄ )2 + 2(Xi − µ)(µ − X̄ )
n
X Xn X n
2 2
(Xi − X̄ ) = (Xi − µ) + (µ − X̄ )2
i=1 i=1 i=1
n
X
+2 (Xi − µ)(µ − X̄ )
i=1
Taking expectations
Pn 2
Pn 2
2 i=1 (Xi − X̄ ) i=1 (Xi − µ)
E [S ] =E =E −
n−1 n−1
n 2
E (X̄ − µ)
n−1
But,
Pn 2
σ2
(Xi − µ) n
E i=1
= σ 2 , E [(X̄ − µ)2 ] = var (X̄ ) =
n−1 n−1 n
MB+SCP (IIT Bombay) CL202 Spring 2021 33 / 42
E [S 2] (continued)
This implies
Pn
− X̄ )2 σ2
i=1 (Xi n 2
E = σ −
n−1 n−1 n−1
2
= σ
or
E [S 2 ] = σ 2