(2021) EC6041 Lecture 4 Inference
(2021) EC6041 Lecture 4 Inference
Ratjomose P. Machema
[email protected]
Department of Economics
National University of Lesotho (NUL)
EC6041:
Econometric Theory and Applications
1 Introduction
ε|X ∼ N 0, σ 2 In
(1)
−1 0
and β̂ = X0 X X y is a linear combination of ε , then
−1
β̂|X ∼ N β, σ 2 X0 X (2)
and the standard normal distribution for OLS estimators can be defined as
β̂j − βj0
∼ Normal(0, 1)
sd(β̂j )
The standardized random variable always has zero mean and variance one.
β̂j − βj0
∼ tn−k−1 = tdf (3)
se(β̂j )
The t distrbution also has a bell shape, but is more spread out than the
Normal(0, 1).
As df → ∞,
tdf → Normal(0, 1)
The difference is practically small for df > 120.
The quantity eqt 3 in is a test statistic and is computed from the estimate β̂j ,
its standard error se β̂j , and the hypothesized value βj0
H0 : β1 = 0 H 0 : β2 = β3 = · · · = βk = 0
R= 1 0 ··· 0 0 1 0 ··· 0
1×k
0 0 1 ··· 0
β0 =
β1 β2 ··· βk 1×k R=
.. .. .. ..
..
c=0
. . . . .
0 0 0 ··· 1 J×k
β0 =
H0 : β2 + β3 = 1 β1 β2 ··· βk 1×k
c= 0 0 ··· 0 1×J
R= 0 1 1 0 ··· 0 1×k
0
β = β1 β2 ··· βk 1×k H0 : β2 + β3 + · · · + βk = 1 and β2 = β3
c= 0 0 ··· 0 1×J
0 1 1 ··· ··· 1
R=
0 1 −1 ··· ··· 0 1×k
β0 =
β1 β2 ··· βk 1×k
1
c=
0 J×1
0
Var Rβ̂ = E R β − β̂ β − β̂ R0 = RVar β̂ R0
−1
= Rσ 2 X0 X R0
and
h i h i
−1
Var [m|X] = Var Rβ̂ − c|X = RVar β̂|X R0 = Rσ 2 (X0 X) R0
therefore
−1
Rβ̂ − c ∼ N 0, Rσ 2 (X0 X) R0
the Wald statistic will be distributed
0 h i−1
2 0 −1 0
W = Rβ̂ − c Rσ (X X) R Rβ̂ − c (4)
∼ χ2 [2]
R.P. Machema, (NUL ) Lecture_4 Inference EC6041 13 / 23
F test
The only unknown quantity in eqtn 4 is σ 2 , but we know that
e0 e
σ̂ 2 = n−k so in large samples we could base our Wald statistic on
0 h i−1
2 0 −1 0
Ŵ = Rβ̂ − c Rσ̂ (X X) R Rβ̂ − c
(n−k)σ̂ 2
σ2 /(n−k)
R.P. Machema, (NUL ) Lecture_4 Inference EC6041 14 / 23
F test
1 0 h i−1
0 −1 0
F = 2 Rβ̂ − c R (X X) R Rβ̂ − c (6)
j σ̂
−1
where σ̂ 2 (X0 X) is the estimated variance matrix of β̂. The test
procedure is then to reject the hypothesis Rβ̂ = c if the
computed F value exceeds a preselected critical value.
0 h i−1
−1
Rβ̂ − c R (X0 X) R0 Rβ̂ − c
F = ∼ F (j, n − k)
j σ̂ 2
Rβ̂ − c
t= h i
ˆ Rβ̂
se
the Lagrangian
0
L = (y − Xβ)0 (y − Xβ) + 2 Rβ̂ − c λ (7)
1
where we have multiplied by 2
in order to simplify the algebra
later on.
R.P. Machema, (NUL ) Lecture_4 Inference EC6041 18 / 23
LM type tests
The first order conditions are
−X0 y − Xβ̂ R + R0 λ̂ = 0 (8)
Rβ̂ R − c = 0
It is clear that if the restriction is valid, the term on the right hand
side should asymptotically converge to zero. It is also plausible that
we should be able to apply some central limit theorem to this vector
to show that it is asymptotically normal with convariance matrix
σ 2 (X0 X).
R.P. Machema, (NUL ) Lecture_4 Inference EC6041 19 / 23
LM type tests
We should therefore be able to base a test of the hypothesis that λ is
zero on the statistic
1
LM = λ̂R (X0 X) R0 λ̂
se
ˆ
Equivalently we can use the fact that R0 λ̂ = X0 y − Xβ̂ R to write
the statistic as
1 R
0 0
R
LM = y − Xβ̂ X (X X) X y − Xβ̂
se
ˆ
The residuals from the restricted regression are standardised (by
dividing through by the estimated standard error of the restricted
regression) and then regressed on the full set of explanatory variables.
If the restriction is valid the explained sum of squares should be small.
R.P. Machema, (NUL ) Lecture_4 Inference EC6041 20 / 23
Asymptotic LR test
In order to implement this, we will initially consider the case where we assume
normality of the errors. Assume also that we know σ 2 but need to estimate β.
Under the assumption of normality, the maximum likelihood estimator will
be the least squares estimator β̂.
The LR test compares the log likelihoods of the two models and tests whether
this difference is statistically significant. If the difference is statistically
significant, then the less restrictive model (the one with more variables) is said to
fit the data significantly better than the more restrictive model.
The likelihood ratio statistic will be given 2 (lU − lR ) that is
0 0 0 0
y − Xβ̂R y − Xβ̂R − y − Xβ̂ y − Xβ̂
LR =
σ2
e0 eR − e0 e0
= R
σ2
Under the assumption of normality (as well as asymptotically), it is
distributed as χ2 (j).
R.P. Machema, (NUL ) Lecture_4 Inference EC6041 21 / 23
The delta method
Rβ̂ = c