EC501 Lecture 04
EC501 Lecture 04
Marcus Chambers
Department of Economics
University of Essex
02 November 2023
1 / 30
Outline
Review
Selecting regressors
Non-nested models
An example
2 / 30
Review
Our focus so far has been estimation and inference in the linear
regression model, y = Xβ + ϵ.
The ordinary least squares (OLS) estimator is:
• BLUE under strong assumptions; and,
• consistent under weaker ones.
One of these assumptions is that the model is correctly
specified.
This raises the question of how to select regressors and
compare models, and maybe how to test for the correct
functional form.
We have also implicitly assumed that the dependent variable is
continuous.
But sometimes this is not the case and new approaches are
needed to deal with limited dependent variables (LDVs).
3 / 30
Comparing models
Last week we considered the two models
y = X1 β1 + ϵ, (1)
y = X1 β1 + X2 β2 + ϵ, (2)
y:N×1 X1 : N × (K − J) β1 : (K − J) × 1
ϵ:N×1 X2 : N × J β2 : J × 1
We noted that (1) is obtained from (2) by setting the J elements
of the vector β2 equal to zero – (1) is nested within (2).
The J linear restrictions β2 = 0 can be tested using an F-test.
But suppose we estimate (1) when (2) is true – what are the
properties of the OLS estimator?
This is a case of omitted variables (those in X2 ).
4 / 30
OLS with omitted variables
We know that, in the regression (1), the OLS estimator is
5 / 30
OLS with irrelevant variables
6 / 30
Selecting regressors
7 / 30
Comparing models
this
PN provides a trade-off between the fit (as measured by
e 2 ) and the parsimony of the model (as measured by K).
i=1 i
We might select the model with the largest R̄2 value.
8 / 30
Information criteria
Other model comparison methods include information criteria.
These, too, provide a trade-off between model fit and
parsimony.
Examples include Akaike’s Information Criterion (AIC):
N
1 X 2 2K
AIC = log ei + ,
N N
i=1
10 / 30
Non-nested F-test
11 / 30
J-test
12 / 30
Logarithms versus levels
The non-nested approach can also be used to distinguish
between logarithmic and levels regressions.
Consider the two models:
13 / 30
Testing the functional form
H0 : α2 = α3 = . . . = αQ = 0.
14 / 30
Non-linear models
If the RESET test rejects E{yi |xi } = xi′ β then what can we do?
One response would be to estimate a non-linear specification
for the conditional mean of the form
15 / 30
Limited dependent variables (LDVs)
16 / 30
Motivating example
yi = β1 + β2 xi2 + ϵi = xi′ β + ϵi , i = 1, . . . , N,
17 / 30
Rogue probabilities!
= P{yi = 1|xi }.
But E{yi |xi } = xi′ β which implies P{yi = 1|xi } = xi′ β in this
model.
We know that a probability must lie between 0 and 1 and xi′ β is
not restricted to have this property in this model.
Hence predicted probabilities from the model could therefore be
negative or greater than one!
18 / 30
Values of ϵi
19 / 30
Binary choice model
A binary choice model is of the form
20 / 30
Common choices of F(·)
Some common choices of F(·) result in the following models:
Probit model (standard normal distribution function):
Z w
1 1
F(w) = Φ(w) = √ exp − t2 dt;
−∞ 2π 2
21 / 30
The distribution functions
Probit
Logit
0.8 Linear probability
0.6
F(w)
0.4
0.2
0
-5 -2.5 0 2.5 5
w
22 / 30
Parameter interpretation
23 / 30
Latent variables
Sometimes, however, the dependent variable of interest is not
actually observed, or latent.
For example, suppose the utility difference (y∗i ) between having
a job and not having one is a function of (observable)
characteristics (xi ).
The individual chooses to work if y∗i > 0.
The model might be of the form
y∗i = xi′ β + ϵi , i = 1, . . . , N.
24 / 30
A binary choice representation
Let F(·) denote the cdf of ϵi . Then
= 1 − F(−xi′ β).
26 / 30
Linear probability model
The results obtained from estimating the linear probability
model by OLS in R are:
> olsfit <- lm(inlf~educ+exper+expersq+age,data=mroz)
> summary(olsfit)
Call:
lm(formula = inlf ~ educ + exper + expersq + age, data = mroz)
Coefficients:
Estimate Std. Error t value Pr(>|t|)
(Intercept) 0.3273246 0.1374985 2.381 0.017535 *
educ 0.0275965 0.0072596 3.801 0.000156 ***
exper 0.0450286 0.0058559 7.689 4.66e-14 ***
expersq -0.0007236 0.0001922 -3.765 0.000180 ***
age -0.0105285 0.0022045 -4.776 2.15e-06 ***
---
27 / 30
Probit model
Call:
glm(formula = inlf ~ educ + exper + expersq + age,
family = binomial(link = "probit"),data = mroz)
Coefficients:
Estimate Std. Error z value Pr(>|z|)
(Intercept) -0.5209982 0.4130301 -1.261 0.207163
educ 0.0855908 0.0225252 3.800 0.000145 ***
exper 0.1287285 0.0180941 7.114 1.12e-12 ***
expersq -0.0020214 0.0005872 -3.443 0.000576 ***
age -0.0316601 0.0067338 -4.702 2.58e-06 ***
---
28 / 30
Logit model
The results from estimating a Logit model are:
> logmod <- glm(inlf~educ+exper+expersq+age,
family = binomial(link = "logit"),data=mroz)
> summary(logmod)
Call:
glm(formula = inlf ~ educ + exper + expersq + age,
family = binomial(link = "logit"),data = mroz)
Coefficients:
Estimate Std. Error z value Pr(>|z|)
(Intercept) -0.8824469 0.6863688 -1.286 0.198557
educ 0.1441256 0.0379671 3.796 0.000147 ***
exper 0.2114275 0.0306252 6.904 5.07e-12 ***
expersq -0.0033137 0.0009879 -3.354 0.000796 ***
age -0.0524305 0.0112967 -4.641 3.46e-06 ***
---
29 / 30
Summary
• Next week:
• heteroskedasticity
30 / 30