Chapter 4
Chapter 4
1
Violation of the Assumptions of the CLRM
1.
2.
3.
4. The X matrix is non-stochastic or fixed in repeated samples
5. N(0,2)
2
Investigating Violations of the
Assumptions of the CLRM
• We will now study these assumptions further, and in particular look at:
- How we test for violations
- Causes
- Consequences
In general we could encounter any combination of 3 problems:
- the coefficient estimates are wrong
- the associated standard errors are wrong
- the distribution that we assumed for the
test statistics will be inappropriate
- Solutions
- the assumptions are no longer violated
- we work around the problem so that we
use alternative techniques which are still valid
3
Assumption 1: E( ) = 0
• The mean of the residuals will always be zero provided that there is a
constant term in the regression.
4
Assumption 2: Var( ) = 2 <
• We have so far assumed that the variance of the errors is constant, 2 - this
is known as homoscedasticity. ˆ +
t
-
5
Detection of Heteroscedasticity: The GQ Test
• Graphical methods
• Formal tests: There are many of them: we will discuss Goldfeld-Quandt
test and White’s test
6
The GQ Test (Cont’d)
3. The test statistic, denoted GQ, is simply the ratio of the two residual
variances where the larger of the two variances must be placed in
the numerator.
s12
GQ 2
s2
4. The test statistic is distributed as an F(T1-k, T2-k) under the null of
homoscedasticity.
5. A problem with the test is that the choice of where to split the
sample is that usually arbitrary and may crucially affect the
outcome of the test.
7
Detection of Heteroscedasticity using White’s Test
4. Decision rule: If the 2 test statistic from step 3 is greater than the
corresponding value from the statistical table then reject the null
hypothesis that the disturbances are homoscedastic.
9
Consequences of Using OLS in the Presence of
Heteroscedasticity
• OLS estimation still gives unbiased coefficient estimates, but they are
no longer BLUE.
The Ordinary Least Squares (OLS) estimator is the Best Linear Unbiased
Estimator. This means:
• Whether the standard errors calculated using the usual formulae are
too big or too small will depend upon the form of the
heteroscedasticity.
11
How Do we Deal with Heteroscedasticity?
• If the form (i.e. the cause) of the heteroscedasticity is known, then we can
use an estimation method which takes this into account (called generalised
least squares, GLS).
• A simple illustration of GLS is as follows: Suppose that the error variance is
related to another variable zt by
var t 2 zt2
• To remove the heteroscedasticity, divide the regression equation by zt
yt 1 x2 t x3t
1 2 3 vt
zt zt zt zt
t
where vt is an error term.
zt
t var t 2 zt2
• Now varvt var 2
2
2
for known zt.
• So the disturbances zt from
zt zt
the new regression equation will be
homoscedastic. 12
Background –
The Concept of a Lagged Value
t yt yt-1 yt
1989M09 0.8 - -
1989M10 1.3 0.8 1.3-0.8=0.5
1989M11 -0.9 1.3 -0.9-1.3=-2.2
1989M12 0.2 -0.9 0.2--0.9=1.1
1990M01 -1.7 0.2 -1.7-0.2=-1.9
1990M02 2.3 -1.7 2.3--1.7=4.0
1990M03 0.1 2.3 0.1-2.3=-2.2
1990M04 0.0 0.1 0.0-0.1=-0.1
. . . .
. . . .
. . . .
13
Autocorrelation
• If there are patterns in the residuals from a model, we say that they are
autocorrelated.
14
Positive Autocorrelation
+
û t û t
+
- +
uˆ t 1 Time
15
Negative Autocorrelation
+ ût
û t
+
- +
uˆ t 1 Time
- -
- +
uˆ t 1
17
Detecting Autocorrelation:
The Durbin-Watson Test
• The coefficient estimates derived using OLS are still unbiased, but
they are inefficient, i.e. they are not BLUE, even in large sample sizes.
• Thus, if the standard error estimates are inappropriate, there exists the
possibility that we could make the wrong inferences.
21
“Remedies” for Autocorrelation
22
Multicollinearity
• This problem occurs when the explanatory variables are very highly correlated
with each other.
• Perfect multicollinearity
Cannot estimate all the coefficients
- e.g. suppose x3 = 2x2
and the model is yt = 1 + 2x2t + 3x3t + 4x4t +
23
Measuring Multicollinearity
• Note that high correlation between y and one of the x’s is not
muticollinearity.
• Method 2: Variance inflationary factor
24
Measuring Multicollinearity
Solutions to the Problem of Multicollinearity
26
Adopting the Wrong Functional Form
• Essentially the method works by adding higher order terms of the fitted values
(e.g. yt2 , yt3 etc.) into an auxiliary regression:
Regress on powers of the fitted values:
•
Obtain R2 from this regression. The test statistic is given by TR2 and is
distributed as a 2 ( p 1) .
• So if the value of the test statistic is greater than a 2 ( p 1) then reject the null
hypothesis that the functional form was correct.
27
But what do we do if this is the case?
yt Axt e t ln yt ln xt t
28
Testing the Normality Assumption
29
Testing for Normality
• Bera and Jarque formalise this by testing the residuals for normality by
testing whether the coefficient of skewness and the coefficient of excess
kurtosis are jointly zero.
• It can be proved that the coefficients of skewness and kurtosis can be
expressed respectively as:
𝐸[ ] 𝐸[ ]
( ) / ( )
6 24
• We estimate b1 and b2 using the residuals from the OLS regression, .
30
What do we do if we find evidence of Non-Normality?
• Could use a method which does not assume normality, but difficult and
what are its properties?
• Often the case that one or two very extreme residuals causes us to reject
the normality assumption.
ût
+
Oct Time
1987
33
Parameter Stability Tests
• We can test this implicit assumption using parameter stability tests. The
idea is essentially to split the data into sub-periods and then to estimate up
to three models, for each of the sub-parts and for all the data and then to
“compare” the RSS of the models.
34
The Chow Test
35
The Chow Test (cont’d)
where:
RSS = RSS for whole sample
RSS1 = RSS for sub-sample 1
RSS2 = RSS for sub-sample 2
T = number of observations
2k = number of regressors in the “unrestricted” regression (since it comes
in two parts)
k = number of regressors in (each part of the) “unrestricted” regression
3. Perform the test. If the value of the test statistic is greater than the critical
value from the F-distribution, which is an F(k, T-2k), then reject the null
hypothesis that the parameters are stable over time.
36
A Chow Test Example
• Consider the following regression for the CAPM (again) for the
returns on Glaxo.
• Say that we are interested in estimating Beta for monthly data from
1981-1992. The model for each sub-period is
• 1981M1 - 1987M10
0.24 + 1.2RMt T = 82 RSS1 = 0.03555
• 1987M11 - 1992M12
0.68 + 1.53RMt T = 62 RSS2 = 0.00336
• 1981M1 - 1992M12
0.39 + 1.37RMt T = 144 RSS = 0.0434
37
A Chow Test Example - Results
H0 : 1 2 and 1 2
• The unrestricted model is the model where this restriction is not imposed
00434
. 00355
. 000336
. 144 4
Test statistic
00355
. 000336
. 2
= 7.698
• We reject H0 at the 5% level and say that we reject the restriction that the
coefficients are the same in the two periods.
38