Psp-Unit-6 Estimation Theory PDF
Psp-Unit-6 Estimation Theory PDF
• Here, the estimator is a point estimator and it is the formula for the
mean.
• Notice the use of different symbols to distinguish estimators and
parameters.
• Point estimates are calculated from the data; parameters are not.
ˆ X 1 X 2 ...... X n
X
n
INTERVAL ESTIMATION
The sample mean is called a point estimate because it assigns a single value
to the estimate. Thus, a point estimate is a single value that best estimates
the population parameter
Another type of estimate is the interval estimate.
• In interval estimation, the parameter being estimated lies between a
certain interval, called the confidence interval, with a certain probability.
• Thus, an interval estimate consists of two numbers between which the
population parameter can be expected to lie with a certain level of
confidence.
• The endpoints of a confidence interval are called the confidence limits
• If a sample of a normally distributed random variable with known standard
deviation is used to estimate the true mean E[X] Then the 95% confidence
interval for the sample mean are given by
X −1.96σX ≤ E[X] ≤ X + 1.96σX.
• There are infinitely many possible estimators for θ, so how can we
make sure that we have chosen a good estimator?
X
X i E( X i )
n
ˆ 2
(X i ) 2
E( X i )
2 2
n
How to find estimators?
Method of Estimation
• This is called least squares estimation because it gives the least value
for the sum of squared errors. Finding the best estimates of the
coefficients is often called “fitting” the model to the data, or sometimes
“learning” or “training” the model.
• Suppose X1,X2,X3…..Xn is a random sample of the population whose
parameter θ is the mean of the population which is unknown.
• To minimize S consider
θ 0 1 2 3 4 5 6 7 ……
S 89 62 41 26 17 14 17 26 ……
d i X 2 X i n 0
i 1 i 1
n
0 2 X i 2 n 0
i 1
n
X i
i 1
X
n
Count 0 1 2 3 4 5 >5
Frequen 11 37 64 55 37 24 12
cy
Here precise record about the sample is not kept. Therefore Method of least squares
is not applicable as mean can not be calculated.
• Furthermore, if the sample is large, the method will yield an excellent estimator of
θ. For these reasons, the method of maximum likelihood is probably the most
widely used method of estimation in statistics.
Method of Maximum Likelihood Estimation
• Suppose we have a random sample X1, X2,..., Xn whose assumed probability
distribution depends on some unknown parameter θ
• If Xi's are discrete, then the likelihood function is defined as the product
• In some problems, it is easier to work with the log likelihood function given by
e . x
P( X x)
x!
The likelihood function is
e . xi
L ( )
xi !
The Log likelihood function is
e . xi
L( ) Log xi Log Logxi !
xi !
l ( ) 1 Log xi Logxi !
l ( ) n Log xi Logxi !
To maximize it
xi
x
n
Exponential Distribution
f ( x, ) e x , x 0
The likelihood function is
L(( xi , ) e xi ( e x1 ).( e x2 )..........( e xn )
n ( x1 x2 ..... xn )
L(( xi , ) e e
n xi
Maximize it
d n 1 xi n xi
L(( xi , ) n e e . xi 0
d
n 1 xi 0
1n
xi Mean
Problems
Suppose that we have observed the random sample X1, X2, X3, ..., Xn,
where Xi∼N(θ1,θ2), so
A coin is flipped 100 times. Given that there were 55 heads, find the
maximum likelihood estimate for the probability p of heads on a
single toss.
Problems
A city dispute center that settles disputes per month follows a Poisson
distribution. The following table gives the number of disputes settled
in a month. Determine the maximum likelihood estimator.
Number of Disputes 1 2 3 4 5 6 7 8 9 10
Frequency 2 11 18 6 24 3 10 9 7 10