Statistical Inference: Dr. Venkataramana B
Statistical Inference: Dr. Venkataramana B
Dr. Venkataramana B
Population
Sample
Parameter
Statistic
Definition 1.1
Any function of the random sample x1 , x2 , . . . ,xn that are being observed, say
Tn (x1 , x2 , . . . ,xn ) is called a statistic. Clearly, a statistic is a random variable.
If it is used used to estimate an unknown parameter θ of the distribution, it is
called an estimator.
Definition 1.2
A particular value of the estimator, say Tn (x1 , x2 , . . . ,xn ) is called an estimate
of θ.
How to assess the performance of estimators and to choose the best one?
How to assess the performance of estimators and to choose the best one?
Are there methods for obtaining estimators other than "ad-hoc" methods.
Unbiasedness
Consistency
Efficiency
Sufficiency
E(Statistic) = P arameter
n
"P # " #
n
− X̄)2
i=1 (Xi 1 X
E(S 2 ) = E = E (Xi − X̄)2
n−1 n−1 i=1
n
" #
1 X
= E (Xi2 + X̄ 2 − 2X̄Xi )
n−1 i=1
n n n
" #
1 X X X
= E Xi2 + X̄ 2 − 2X̄ Xi
n−1 i=1 i=1 i=1
n
" #
1 X
= E Xi2 + nX̄ 2 − 2nX̄ 2
n−1 i=1
n n
" # " #
1 X 1 X
= E Xi2 − nX̄ 2 = E(Xi2 ) − nE(X̄ 2 )
n−1 i=1
n − 1 i=1
Solution:
Given, E(Xi ) = µ, V (Xi ) = 1, ∀ i = 1, 2, . . . , n
=⇒ E(Xi2 ) = V (Xi ) + {E(Xi )}2 = 1 + µ2 {∵ V (x) = E(X 2 ) − (E(X))2 }
n n n
" #
1X 1X 1X
E(t) = E Xi2 = E(Xi2 ) = (1 + µ2 )
n i=1 n i=1 n i=1
n(1 + µ2 )
=
n
E(t) = 1 + µ2
Solution: Since x1 , x2 , . . . ,xn is a random sample from Bernoulli population with parameter θ,
n
X
Let T = xi ∼ B(n,θ)
i=1
=⇒ E(T ) = nθ and V ar(T ) = nθ(1 − θ)
n o
X ∼ B(1, p) =⇒ E(X) = p & V ar(X) = pq
h P i
xi − 1)
P
xi (
T (T − 1)
E =E
n(n − 1) n(n − 1)
1 h i
= E(T 2 ) − E(T )
n(n − 1)
1 h i
= V ar(T ) + {E(T )}2 ) − E(T )
n(n − 1)
1 h i
= nθ(1 − θ) + n2 θ2 − nθ
n(n − 1)
nθ2 (n − 1)
= = θ2
n(n − 1)
hP P i
xi ( xi −1)
=⇒ n(n−1) is an unbiased estimator of θ2 .
Solution:
e−θ θx
Since X ∼ P oisson(θ) =⇒ P (X = x) = x! ; X = 0, 1, . . . , ∞
x=0
x!
−θ −kθ
=e ·e
−(k+1)θ
E{T (X)} =e
Definition 2.2
An estimator Tn = T (x1 , x2 , . . . ,xn ) based on a random sample of size n, is
said to be consistent estimator of γ(θ), θ ∈ Θ, the parameter space,
p
if Tn converges to γ(θ) in probability, i.e., if Tn −→ γ(θ) as n −→ ∞.
where ε, η are arbitrarily small positive numbers and m is some large value of
n.
Theorem 2.3
If Tn is a consistent estimator of γ(θ) and ψ{γ(θ)} is a continuous function
of γ(θ), then ψ(Tn ) is a consistent estimator of ψ{γ(θ)}.
Theorem 2.4
Let {Tn } be a sequence of estimators such that for all θ ∈ Θ,
Solution:
Ṡince X1 , . . . , Xn are i.i.d Bernoulli random variables with the parameter ‘p0 ,
n
X
T = Xi ∼ B(n, p)
i=1
E(T ) = np and V ar(T ) = npq
E(X̄) =
V ar(X̄) =
Definition 2.5
If Tn is the most efficient estimator with variance V1 and T2 is any other
estimator with variance V2 , then the efficiency E of T2 is defined as :
V1
E= V2
Theorem 2.7
An M.V.U.E is unique in the sense that if T1 and T2 are M.V.U. estimators for
γ(θ), then T1 = T2 almost surely.
Theorem 2.8
Let T1 and T2 be unbiased estimators of γ(θ) with efficiencies e1 and e2
√ p and ρ = ρθ be the correlation
respectively
√
coefficient
p between them. Then
e1 e2 − (1 − e1 )(1 − e2 ) ≤ ρ ≤ e1 e2 + (1 − e1 )(1 − e2 )
Corollary 2.9
√ √ √
If we take e1 = 1 and e2 = e, we get e≤ρ≤ e =⇒ ρ = e
Solution:
Let x1 , x2 , . . . , xn be random sample from a Bernoulli
population with the parameter
‘p‘, 0 < p < 1, i.e.,
1, with probability p
xi =
0, with probability q=(1-p)
i=1
L = gθ [t(x)].h(x)
where gθ [t(x)] is depends on θ and x only through the value of t(x) and
h(x) is independent of θ.
Solution:
Let us write θ = (µ, σ 2 ); −∞ < µ < ∞, 0 < σ 2 < ∞
Then
n 1
Pn
Y 1 n
− (x −µ)2
i=1 i
2σ 2
L= fθ (xi ) = √ ·e
i=1 σ 2π
h P i
n
− 12
P
1 n i=1
xi 2 −2µ xi +nµ2
= √ · e 2σ
σ 2π
=gθ [t(x)] · h(x)
h i
1
1 n − {t2 (x)−2µt1 (x)+nµ2 }
2σ 2
gθ [t(x)] = √ e
σ 2π
Where, X X
t(x) =[t1 (x), t2 (x)] = xi , xi 2 and h(x) = 1
X
Thus t(x) = xi is sufficient for µ and
Solution:
n n
θ xiθ−1
Y Y
L(x, θ) = f (xi , θ) =
i=1 i=1
n
Y
n θ−1
=θ xi
i=1
Yn θ 1
=θn xi · Q
n
i=1 i=1 xi
h i2 h i2
d
dθ · γ(θ) γ 0 (θ)
V ar(t) ≥ i2 =
I(θ)
h
∂
E ∂θ logL
Corollary
If t is an unbiased estimator for θ, i.e.
h 2 i
∂
Where I(θ) = E ∂θ logL is called by R. A. Fisher as the amount of information on θ supplied by the
sample (x1 , x2 , . . . , xn ) and its reciprocal 1/I(θ), as the information limit to the variance of estimator
t = t(x1 , x2 , . . . , xn ).