Exercise 5.3 Solution Guide
Exercise 5.3 Solution Guide
Solution Guide
(1) (OLS) State the minimal conditions for consistency of the OLS estimator
in the regression model (5.3). How is this related to the interpretation of
(5.3) as a conditional expectation: E[yt | xt ] = x0t β ?
Discuss how this assumption can be used to construct moment conditions,
Now assume that the K2 variables in x2t are endogenous, in the sense
E[x2t t ] 6= 0.
1
Solution: The miniman condition for consistency of the OLS estimator in
the linear regression model (5.3) is that the moment condition E(xt t ) = 0 holds.
The assumption of predeterminedness, E(t | xt ) = 0, implies that (5.3) can
be interpreted as a conditional expectation, E(yt | xt ) = x0t β. The assumption of
predeterminedness also implies the moment condition,
which corresponds to the simple linear expression for the function f (wt , zt , β) =
xt t , where wt are the model variables, wt = (yt , x0t )0 , and zt are the instruments
(here we have zt = 0 as there are no instruments). As there are K moment
conditions and K parameters in β we have exact identification and we can find
a unique set of parameters, β, which satisfies the population moment condtions
in (5.4). By pluggin in, t = yt − x0t β, we get,
2
βM M . Again, we plug in, t = yt − x0t β, to get,
T
b = 1
X
gT (β) xt t
T
t=1
T
1 X
= xt (yt − x0t β)
b
T
t=1
T T
1X 1X
= x t yt − xt x0t βb
T T
t=1 t=1
= 0, (5.7)
which corresponds to the OLS estimator, βbOLS . To compute the estimator the
matrix ( Tt=1 xt x0t , must be non-singular, so that it can be inverted, so we assume
P
E(x2t t ) 6= 0.
That implies that the moment condition, E(xt t ) 6= 0, where xt = (x01t , x02t )0 , so
the MM/OLS estimator is inconsistent.
(2) (IV) Now we assume the existence of K2 new instrumental variables, z2t ,
with the property
E[z2t t ] = 0. (5.9)
Should the new instruments, z2t , fulfill other requirements besides (5.9) for
being valid and relevant instruments?
Define the K × 1 vector of instruments, zt = (x01t , z2t
0 )0 . State the pop-
3
Solution: The K2 instruments in z2t are valid instruments if they satisfy the
moment condition, E(z2t t ) = 0, and they are relevant instruments if they are
correlated with the endogenous variables, so that T1 Tt=1 z2t x02t is non-singular.
P
the K1 model variables x1t (we say x1t are instruments for themselves) and the
K2 new instruments z2t . We have the two sets of valid moment conditions,
E(x1t t ) = 0 (5.10)
E(z2t t ) = 0, (5.11)
g(β) = E(zt t )
= E(zt (yt − x0t β))
= E(zt yt ) − E(zt x0t β)
= E(zt yt ) − E(zt x0t )β
= 0, (5.13)
= 0, (5.16)
4
which gives the estimator,
T −1 XT
1X 0 1
βbM M = zt x t zt yt . (5.17)
T T
t=1 t=1
What is the role of the weight matrix WT , and how should it be optimally
chosen?
State the sample moments, gT (β), for the case R > K. Insert the moment
conditions in (5.19) and derive the GMM estimator for a given weight
matrix, βbGM M (WT ), as the solution to
∂QT (β)
= 0.
∂β
and K parameters in β.
5
The corresponding sample moment conditions are,
T T
1X 1X
gT (β) = f (wt , zt , β) = zt t , (5.21)
T T
t=1 t=1
6
Defining the matrices,
y1
y2
Y = ..
(5.25)
(T ×1) .
yT
0
x1 x11 x12 ... x1K
0
x2 x21 x22 ... x2K
X = .. = ..
.. .. .. (5.26)
(T ×K) . . . . .
x0T xT 1 xT 2 ... xT K
0
z1 z11 z12 ... z1R
0
z2 z21 z22 ... z2R
Z = .. = ..
.. .. .. (5.27)
(T ×R) . . . . .
zT 0 z T 1 zT 2 ... zT R
1
2
= .. ,
(5.28)
(T ×1) .
T
Y = βX + . (5.29)
(T ×1)
= T −2 (Y 0 Z − β 0 X 0 Z)WT (Z 0 Y − Z 0 Xβ)
= T −2 (Y 0 ZWT Z 0 Y − Y 0 ZWT Z 0 Xβ − β 0 X 0 ZWT Z 0 Y + β 0 X 0 ZWT Z 0 Xβ)
= T −2 (Y 0 ZWT Z 0 Y − 2β 0 X 0 ZWT Z 0 Y + β 0 X 0 ZWT Z 0 Xβ), (5.32)
7
where the last step follows as (Y 0 ZWT Z 0 Xβ) and (β 0 X 0 ZWT Z 0 Y ) are identical
scalar terms.
We differentiate the quadratic form, QT (β) with respect to K parameters in
the vector β,
∂QT (β)
= 0 − 2T −2 X 0 ZWT Z 0 Y + T −2 (X 0 ZWT Z 0 X + (X 0 ZWT Z 0 X)0 )β
∂β
= −2T −2 X 0 ZWT Z 0 Y + T −2 (X 0 ZWT Z 0 X + X 0 ZWT Z 0 X)β
= −2T −2 X 0 ZWT Z 0 Y + T −2 (2X 0 ZWT Z 0 X)β
= −2T −2 X 0 ZWT Z 0 Y + 2T −2 X 0 ZWT Z 0 Xβ (5.33)
where we have used (7∗) and (8∗) in Lecture Note 3, Introduction to Vector and
Matrix Differentiation in the first line.
We find the GMM estimator as the solution to the first-order conditions,
(5.34)
The efficient weight matrix (or the optimal weight matrix ) minimizes the
asymptotic variance of the GMM estimator by setting,
(4) (GIV) Discuss how the optimal weight matrix, WTopt , can be estimated if
t is identically and independently distributed, IID.
Insert the optimal weight matrix in the formula for the estimator to
obtain βbGM M (WTopt ) and show that it simplifies to the generalized IV esti-
mator. Show that the GIV estimator, βbGIV , can be derived as a two-stage
least squares estimator.
8
Solution: If we assume that t is identically and independently distributed
(IID) over time with variance σ 2 , then the asymptotic variance of the moments,
S, is given by,
S = T · V (gT (β))
X T
1
=T ·V f (wt , zt , β)
T
t=1
X T
1
= ·V f (wt , zt , β)
T
t=1
T
1 X
= · V f (wt , zt , β)
T
t=1
T
1 X 0
= · E f (wt , zt , β)f (wt , zt , β)
T
t=1
T
1 X 0
= · E zt t (zt t )
T
t=1
T
1 X 0 0
= · E zt t t zt
T
t=1
T
1 X 2 0
= · E t zt zt
T
t=1
T
σ2 X
0
= · E zt zt
T
t=1
σ2 0
= ZZ (5.36)
T
Given that t is assumed IID with variance, σ 2 , a natural consistent estimator of
b2 = T −1 Tt=1 2t , and a consistent estimator of S is,
σ 2 is given by, σ
P
b2 0
σ
ST = Z Z. (5.37)
T
The optimal weight matrix becomes,
WTopt = ST−1 = (T −1 σ
b2 Z 0 Z)−1 , (5.38)
9
and we find the GMM estimator as,
−1
0 0
βbGM M = X ZWT Z X X 0 ZWT Z 0 Y
−1
0 −1 2 0
−1 0 −1 0
XZ T σb ZZ ZX X 0 Z T −1 σ
b2 Z 0 Z ZY
−1
0 2 0
−1 0 −1 0
XZ σ
b ZZ ZX X 0Z σ
b2 Z 0 Z ZY
−1
0 0
−1 0 −1 0
XZ ZZ ZX X 0Z Z 0Z Z Y, (5.39)
Finally, we want to show that the GIV estimator can be derived as a two-
stage least squares (2SLS) estimator. The 2SLS estimator is intuitive easy to
understand. In the first step, we regress X on the instruments Z and save the
predicted values X.b In the second step, we regress Y on the predicted values
from the first step, X,
b to get an estimate of the model parameters β.
We consider again the model in matrix notation,
Y = Xβ + , (5.40)
X = Zγ + U, (5.41)
b = (Z 0 Z)−1 Z 0 X,
γ (5.42)
X γ = Z(Z 0 Z)−1 Z 0 X.
b = Zb (5.43)
Y = XB
b + E, (5.44)
B b 0 X)
b = (X b −1 X
b 0 Y. (5.45)
10
Pluggin in the expression for X,
b we find the 2SLS estimator,
B
b2SLS = (X b 0Y
b −1 X
b 0 X)
−1
0 −1 0 0 0 −1 0
= (Z(Z Z) Z X) Z(Z Z) Z X (Z(Z 0 Z)−1 Z 0 X)0 Y
−1
0 0 −1 0 0 −1 0
= (X Z(Z Z) Z )Z(Z Z) ZX (X 0 Z(Z 0 Z)−1 Z 0 )Y
−1
0 0 −1 0
= X Z(Z Z) Z X X 0 Z(Z 0 Z)−1 Z 0 Y, (5.46)
which is identical to the GIV estimator and the GMM estimator based on an
optimal weight matrix derived under the assumption of IID residuals.
11