Lecture 8
Lecture 8
13
ECONOMETRIC
MODELING: MODEL
SPECIFICATION AND
DIAGNOSTIC TESTING
1
Keith Cuthbertson, Stephen G. Hall, and Mark P. Taylor, Applied Econometrics Techniques,
Michigan University Press, 1992, p. X.
2
David F. Hendry, Dynamic Econometrics, Oxford University Press, U.K., 1995, p. 68.
3
Peter Kennedy, A Guide to Econometrics, 3d ed., The MIT Press, Cambridge, Mass., 1992,
p. 82.
506
Gujarati: Basic II. Relaxing the 13. Econometric Modeling: © The McGraw−Hill
Econometrics, Fourth Assumptions of the Model Specification and Companies, 2004
Edition Classical Model Diagnostic Testing
4
D. F. Hendry and J. F. Richard, “The Econometric Analysis of Economic Time Series,”
International Statistical Review, vol. 51, 1983, pp. 3–33.
5
Milton Friedman, “The Methodology of Positive Economics,” in Essays in Positive Eco-
nomics, University of Chicago Press, Chicago, 1953, p. 7.
Gujarati: Basic II. Relaxing the 13. Econometric Modeling: © The McGraw−Hill
Econometrics, Fourth Assumptions of the Model Specification and Companies, 2004
Edition Classical Model Diagnostic Testing
It is one thing to list criteria of a “good” model and quite another to actu-
ally develop it, for in practice one is likely to commit various model specifi-
cation errors, which we discuss in the next section.
Note that we have changed the notation to distinguish this model from the
true model.
Since (13.2.1) is assumed true, adopting (13.2.2) would constitute a speci-
fication error, the error consisting in omitting a relevant variable (Xi3 ).
Therefore, the error term u2i in (13.2.2) is in fact
Now assume that yet another researcher postulates the following model:
Yi* = β1* + β2* Xi* + β3* Xi*2 + β4* Xi*3 + ui* (13.2.7)
Yi = β Xi ui (13.2.8)
where the stochastic error term enters multiplicatively with the property
that ln ui satisfies the assumptions of the CLRM, against the following
model
Yi = α Xi + ui (13.2.9)
where the error term enters additively. Although the variables are the same
in the two models, we have denoted the slope coefficient in (13.2.8) by β and
the slope coefficient in (13.2.9) by α. Now if (13.2.8) is the “correct” or “true”
model, would the estimated α provide an unbiased estimate of the true β?
That is, will E(α̂) = β? If that is not the case, improper stochastic specifica-
tion of the error term will constitute another source of specification error.
To sum up, in developing an empirical model, one is likely to commit one
or more of the following specification errors:
1. Omission of a relevant variable(s)
2. Inclusion of an unnecessary variable(s)
3. Adopting the wrong functional form
4. Errors of measurement
5. Incorrect specification of the stochastic error term
Before turning to an examination of these specification errors in some
detail, it may be fruitful to distinguish between model specification errors
and model mis-specification errors. The first four types of error discussed
above are essentially in the nature of model specification errors in that we
have in mind a “true” model but somehow we do not estimate the correct
model. In model mis-specification errors, we do not know what the true
model is to begin with. In this context one may recall the controversy
Gujarati: Basic II. Relaxing the 13. Econometric Modeling: © The McGraw−Hill
Econometrics, Fourth Assumptions of the Model Specification and Companies, 2004
Edition Classical Model Diagnostic Testing
between the Keynesians and the monetarists. The monetarists give primacy
to money in explaining changes in GDP, whereas the Keynesians emphasize
the role of government expenditure to explain changes in GDP. So to speak,
there are two competing models.
In what follows, we will first consider model specification errors and then
examine model mis-specification errors.
Yi = α1 + α2 X2i + vi (13.3.2)
6
But see exercise 13.32.
Gujarati: Basic II. Relaxing the 13. Econometric Modeling: © The McGraw−Hill
Econometrics, Fourth Assumptions of the Model Specification and Companies, 2004
Edition Classical Model Diagnostic Testing
E(α̂2 ) = β2 + β3 b3 2 (13.3.3)
ILLUSTRATIVE EXAMPLE: CHILD MORTALITY regress FLR on PGNP (regression of the excluded
REVISITED variable on the included variable), the slope coefficient
in this regression [b3 2 in terms of Eq. (13.3.3)] is
Regressing child mortality (CM) on per capita GNP 0.00256.8 This suggests that as PGNP increases by a
(PGNP) and female literacy rate (FLR), we obtained the unit, on average, FLR goes up by 0.00256 units. But if
regression results shown in Eq. (7.6.2), giving the partial FLR goes up by these units, its effect on CM will be
slope coefficient values of the two variables as −0.0056 (−2.2316) (0.00256) = β̂3 b3 2 = −0.00543.
and −2.2316, respectively. But if we now drop the FLR Therefore, from (13.3.3) we finally have ( β̂2 +
variable, we obtain the results shown in Eq. (7.7.2). If we β̂3 b3 2 ) = [−0.0056 + (−2.2316)(0.00256)] ≈ −0.0111,
regard (7.6.2) as the correct model, then (7.7.2) is a mis- which is about the value of the PGNP coefficient ob-
specified model in that it omits the relevant variable FLR. tained in the incorrect model (7.7.2).9 As this example il-
Now you can see that in the correct model the coefficient lustrates, the true impact of PGNP on CM is much less
of the PGNP variable was −0.0056, whereas in the (−0.0056) than that suggested by the incorrect model
“incorrect” model (7.7.2) it is now −0.0114. (7.7.2), namely, (−0.0114).
In absolute terms, now PGNP has a greater impact
on CM as compared with the true model. But if we
7
For an algebraic treatment, see Jan Kmenta, Elements of Econometrics, Macmillan,
New York, 1971, pp. 391–399. Those with a matrix algebra background may want to consult
J. Johnston, Econometrics Methods, 4th ed., McGraw-Hill, New York, 1997, pp. 119–112.
8
The regression results are:
FLR = 47.5971 + 0.00256PGNP
se = (3.5553) (0.0011) r2 = 0.0721
9
Note that in the true model β̂2 and β̂3 are unbiased estimates of their true values.
Gujarati: Basic II. Relaxing the 13. Econometric Modeling: © The McGraw−Hill
Econometrics, Fourth Assumptions of the Model Specification and Companies, 2004
Edition Classical Model Diagnostic Testing
σ2 σ2
var (β̂2 ) = 2
= VIF (13.3.5)
1 − r223 2
!
x2i x2i
10
To bypass the tradeoff between bias and efficiency, one could choose to minimize the
mean square error (MSE), since it accounts for both bias and efficiency. On MSE, see the sta-
tistical appendix, App. A. See also exercise 13.6.
Gujarati: Basic II. Relaxing the 13. Econometric Modeling: © The McGraw−Hill
Econometrics, Fourth Assumptions of the Model Specification and Companies, 2004
Edition Classical Model Diagnostic Testing
Yi = β1 + β2 X2i + ui (13.3.6)
11
Note, though, α̂1 is still biased, which can be seen intuitively as follows: We know that
β̂1 = Ȳ − β̂2 X̄ 2 − β̂3 X̄3 , whereas α̂1 = Ȳ − α̂2 X̄2 , and even if α̂2 = β̂2 , the two intercept estima-
tors will not be the same.
12
For details, see Adrian C. Darnell, A Dictionary of Econometrics, Edward Elgar Publisher,
1994, pp. 371–372.
Gujarati: Basic II. Relaxing the 13. Econometric Modeling: © The McGraw−Hill
Econometrics, Fourth Assumptions of the Model Specification and Companies, 2004
Edition Classical Model Diagnostic Testing
and
σ2
var (α̂2 ) = 2
1 − r223
!
x2i (13.3.9)
Therefore,
var (α̂2 ) 1
= (13.3.10)
var (β̂2 ) 1 − r223
Since 0 ≤ r223 ≤ 1, it follows that var (α̂2 ) ≥ var (β̂2 ); that is, the variance of α̂2
is generally greater than the variance of β̂2 even though, on average, α̂2 = β2
[i.e., E(α̂2 ) = β2 ].
The implication of this finding is that the inclusion of the unnecessary
variable X3 makes the variance of α̂2 larger than necessary, thereby making
α̂2 less precise. This is also true of α̂1 .
Notice the asymmetry in the two types of specification biases we have
considered. If we exclude a relevant variable, the coefficients of the vari-
ables retained in the model are generally biased as well as inconsistent, the
error variance is incorrectly estimated, and the usual hypothesis-testing
procedures become invalid. On the other hand, including an irrelevant vari-
able in the model still gives us unbiased and consistent estimates of the co-
efficients in the true model, the error variance is correctly estimated, and
the conventional hypothesis-testing methods are still valid; the only penalty
we pay for the inclusion of the superfluous variable is that the estimated
variances of the coefficients are larger, and as a result our probability infer-
ences about the parameters are less precise. An unwanted conclusion here
would be that it is better to include irrelevant variables than to omit the rel-
evant ones. But this philosophy is not to be espoused because addition of
unnecessary variables will lead to loss in efficiency of the estimators and
may also lead to the problem of multicollinearity (why?), not to mention the
loss of degrees of freedom. Therefore,
In general, the best approach is to include only explanatory variables that, on
theoretical grounds, directly influence the dependent variable and that are not
accounted for by other included variables.13
13
Michael D. Intriligator, Econometric Models, Techniques and Applications, Prentice Hall,
Englewood Cliffs, N.J., 1978, p. 189. Recall the Occam’s razor principle.
Gujarati: Basic II. Relaxing the 13. Econometric Modeling: © The McGraw−Hill
Econometrics, Fourth Assumptions of the Model Specification and Companies, 2004
Edition Classical Model Diagnostic Testing
However, we are not totally sure that, say, the variable Xk really belongs in
the model. One simple way to find this out is to test the significance of the
estimated βk with the usual t test: t = β̂k/se (β̂k). But suppose that we are not
sure whether, say, X3 and X4 legitimately belong in the model. This can be
easily ascertained by the F test discussed in Chapter 8. Thus, detecting the
presence of an irrelevant variable (or variables) is not a difficult task.
It is, however, very important to remember that in carrying out these tests
of significance we have a specific model in mind. We accept that model as
the maintained hypothesis or the “truth,” however tentative it may be.
Given that model, then, we can find out whether one or more regressors are
really relevant by the usual t and F tests. But note carefully that we should
not use the t and F tests to build a model iteratively, that is, we should not
say that initially Y is related to X2 only because β̂2 is statistically significant
and then expand the model to include X3 and decide to keep that variable in
the model if β̂3 turns out to be statistically significant, and so on. This strat-
egy of building a model is called the bottom-up approach (starting with
a smaller model and expanding it as one goes along) or by the somewhat
pejorative term, data mining (other names are regression fishing, data
grubbing, data snooping, and number crunching).
14
James Davidson, Econometric Theory, Blackwell Publishers, Oxford, U.K., 2000, p. 153.