Lecture5 Module2 Anova 1
Lecture5 Module2 Anova 1
MODULE II
LECTURE - 5
Analysis of variance
Analysis of variance is a body of statistical methods of analyzing the measurements assumed to be structured as
xij
be identically and independently distributed with mean 0 and variance 2 . It may be noted that the i s can be assumed additionally to follow a normal distribution N (0, 2 ). It is needed for the maximum likelihood estimation of parameters from the beginning of analysis but in the least squares estimation, it is needed only when conducting the tests of hypothesis and the confidence interval estimation of parameters. parameters The least squares method does not require any knowledge of distribution like normal upto the stage of estimation of parameters. We need some basic concepts to develop the tools.
S 2 = i2 = ' = ( y X )( y X )
i =1
= yy 2 X ' y + X X
y = ( y1 , y2 ,..., yn ) . Differentiating i minimum is i i where h Diff i i S2 with i h respect to and d substituting b i i i it to b be zero, the h normal l
equations are obtained as
dS 2 = 2 X X 2 X y = 0 d
or X X = X y.
If X has full rank then ( X X ) has a unique inverse and the unique least squares estimate of is
= ( X X ) 1 X y
which is the best linear unbiased estimator of in the sense of having minimum variance in the class of linear and unbiased
) estimator If rank of X is not full, estimator. full then generalized inverse is used for finding the inverse of ( X X ).
If L is a linear parametric function where L = ( 1 , 2 ,..., p ) is a non-null vector, then the least squares estimate of L
. is L
L admits A question ti arises i th t what that h t are the th conditions diti under d which hi h a linear li parametric t i function f ti d it a unique i l least t
squares estimate in the general case.
Estimable functions
A linear function of the parameters with known is said to be an estimable parametric function (or estimable) if there exists a linear function L Y of Y such that
E ( LY ) = for all Rb .
Theorem 1
A linear parametric function L admits a unique least squares estimate if and only if L is estimable.
g minimum variance in the class of all linear and unbiased is the best linear unbiased estimater of L in the sense of having estimators of L .
Theorem 3
If the linear parametric function is also estimable.
1 , 2 ,..., k
Theorem 4
All linear parametric functions in are estimable if and only if X has full rank rank.
If X is not of full rank, then some linear parametric functions do not admit the unbiased linear estimators and nothing can be inferred about them. The linear parametric functions which are not estimable are said to be confounded. A possible solution to this problem is to add linear restrictions on so as to reduce the linear model to a full rank.
Theorem 5
' ' be their least squares estimators. Then and L'2 be two estimable parametric functions and let L1 and L'2 Let L1
X can be used in place of unique inverse. assuming that X is a full rank matrix. If not, the generalized inverse of X
Estimator of
Consider an estimator of 2 as 1 )( y X ) 2 = (y X n p
1 [ y X ( X X ) 1 X ' y ][ y X ( X X ) 1 X y ] n p 1 y [ I X ( X X ) 1 X ][ I X ( X X ) 1 X ] y = n p 1 = y [ I X ( X X ) 1 X ] y n p =
2) = E (
2
n p
tr[ I X ( X X ) 1 X ]
= 2
and so
2 is an unbiased estimator of 2 .
intervals. For maximum likelihood estimation, we need the distributional assumption from the beginning. Suppose y1 , y2 ,..., yn are independently and identically distributed following a normal distribution with mean E ( yi ) = j xij and variance Var ( yi ) = 2 (i = 1, 2,, n). Then the likelihood function of
p
y1 , y2 ,..., yn is
j =1
L( y | , 2 ) =
1 (2 ) ( )
n 2 n 2 2
1 exp 2 ( y X )( y X ) 2
where
L = X y, = 0 X X 1 L 2 )( y X ). = 0 = (y X ) 2 n
8
Assuming the full rank of X, the normal equations are solved and the maximum likelihood estimators are obtained as