Handout 14: Unbiasedness and MSE
Handout 14: Unbiasedness and MSE
We begin making the transition from the ideas and tools of probability
theory to the theory and methods of statistical inference. For simplicity,
suppose that we have the data X1 , X2 , . . . , Xn in hand, where the X’s
are assumed to be independent random variables with a common dis-
tribution Fθ indexed by the parameter θ whose value is unspecified. In
practice, we would often be willing to assume that the distribution Fθ has
a density or probability mass function of some known form. For exam-
ple, physical measurements might be assumed to be random draws from
an N (µ, σ 2 ) distribution, counts might be assumed to be Binomial(n, p)
or P oisson(λ) variables, and the failure times of engineered systems in a
life-testing experiment might be assumed to be Γ(α, β) variables. Note
that such assumptions do not specify the exact distribution that applies
to the experiment of interest, but rather, specifies only its “type.”
θb = θ(X
b 1 , X 2 , . . . , Xn )
The MSE has a useful decomposition that we will often use in the
analyses of various estimators.
b = V ar(θ)
M SE(θ) b + Bias(θ)
b2
Where:
b = E(θb − θ)
Bias(θ)
θb − θ = (θb − Eθ)
b + (Eθb − θ)
Taking expectations on both sides, notice that the cross-term drops out:
iid
Theorem 2 If X1 , . . . Xn ∼ F for some distribution F with finite mean
µ and variance σ 2 , then
∑
Xi
X̄ = i
n
1 ∑
s2 = (Xi − X̄)2
(n − 1) i
i=1 i=1
[ ]
∑
n ∑
n
=E (Xi − µ)2 + 2(µ − X̄) (Xi − µ) + n(X̄ − µ)2
i=1 i=1
[ ]
∑
n
=E (Xi − µ) − 2n(X̄ − µ) + n(X̄ − µ)
2 2 2
i=1
[ n ]
∑
=E (Xi − µ) − n(X̄ − µ)
2 2
i=1
∑
n
= E(Xi − µ)2 − nE(X̄ − µ)2
i=1
∑
n ( )
σ2
= σ2 − n
i=1
n
= (n − 1)σ 2