Lecture Notes GLS
Lecture Notes GLS
II Effects of Heteroskedasticity:
1. OLS is no longer efficient, since it treats all observations as being of equal importance,
while there is much more information about the line to be had from some observations
than from others, and when estimating the line more attention should be paid to the
observations having small variances than those with larger variances.
2. The standard errors are biased, and thus hypothesis tests and confidence intervals based
on the t-distribution are likely to be invalid.
Let us assume that the variance of (t ) is directly proportional to the variable Z (t ) and the model
for heteroskedasticity is
2 2 Z (t ) ,
(1)
2
where is an unknown parameter and Z (t ) is a variable on which we have observations.
(2)
where
E ( (t )) 0,
E ( 2 (t )) 2( t ) 2 Z (t ),
and E ( (t ), (t )) 0 when t .
Now divide equation (2) by
Z (t ) to obtain
y (t )
1
x(t )
a
b
,
Z (t )
Z (t )
Z (t )
(3)
(t )
~ (t )
.
Z (t )
where
If we introduce the notation
y (t )
~
y (t )
, xo (t )
Z (t )
1
x (t )
,~
x (t )
,
Z (t )
Z (t )
~y (t ) a ~
xo (t ) b ~
x (t ) ~(t ).
The OLS estimates of the unknown parameters a and b in the transformed model are BLUE,
because the disturbance term of the transformed model satisfies the classical disturbance
assumptions:
(4)
(a)
(t )
E (~ (t )) E
Z
(
t
)
1
E ( (t ))
Z (t )
E ( (t )) 0),
(b)
(c)
2 (t )
1
1
E (~ 2 (t )) E
E ( 2 (t ))
2 Z (t ) 2 ,
Z (t )
Z (t )
Z (t )
(t )
(t )
1
E (~ (t ), ~ (t )) E
Z (t ) Z (t ) E ( (t ), (t )) 0.
Z
(
t
)
Z
(
t
Note: dividing by
(t ) may be regarded as summarizing all the influences on Y not accounted for by the
explanatory variables X. As successive values of economic variables are nearly always highly
correlated, it is reasonable to assume that the disturbances are autocorrelated.
optimistic impression of the worth of the model for prediction. Parameter estimates
associated with irrelevant variables may be highly significant.
(5)
where
(t ) (t 1) u (t ).
(6)
The value of u in each observation is assumed to be independent of its value in all other
observations. Assume for the moment that is known, although in practice this is never the case.
Our strategy is to transform (5) so that the disturbance for the transformed model has no
autocorrelation. OLS is then used on the new model to produce parameters estimators that are
BLUE.
We lag (5) by one period and multiply by :
y (t 1) a b x (t 1) (t 1), t 2, T .
(7)
(8)
(9)
~y (t ) y (t ) y (t 1), x (t ) 1 , ~
x (t ) x(t ) x (t 1),
o
then the transformed model is
~y (t ) a ~
xo (t ) b ~
x (t ) u (t ).
(11)
The disturbance term for this model was originally assumed to exhibit no autocorrelation. But
the number of observations to be used for fitting the line is only T-1 instead of T. This might
cause the loss of efficiency that overweighs the gain due to overcoming autocorrelation.
2
Therefore it may be useful to regain the first observation y(1), x(1) with its residual variance .
The problem is that the latter is different from the residual variance in the remaining
2
observations, u . So we have solved the problem of autocorrelation at the expense of
2 in terms of u2 :
(12)
2
since (t 1) and u(t) are independent and since var (t ) var( (t 1)) . Hence
u2
.
1 2
(13)
2
This allows us to eliminate heteroskedasticity by weighting the 1st observation by 1 .
2
The disturbance term in that observation becomes 1 (1). Its variance is
2 (1 2 ) 2 u2 (see (13)).
So after regaining the first observation and eliminating heteroskedasticity we obtain a classical
linear regression model, which means that applying OLS will produce parameters estimates
which are BLUE (Best Linear Unbiased Estimates).