Chapter 2. Dynamic Panel Data Models
Chapter 2. Dynamic Panel Data Models
University of Orléans
April 2018
Remark
In a dynamic panel model, the choice between a …xed-e¤ects formulation
and a random-e¤ects formulation has implications for estimation that are
of a di¤erent nature than those associated with the static model.
Objectives
1 The LSDV estimator is consistent for the static model whether the
e¤ects are …xed or random.
2 On the contrary, the LSDV is inconsistent for a dynamic panel data
model with individual e¤ects, whether the e¤ects are …xed or random.
αi = α + αi
Assumptions
jγj < 1
2 The initial condition yi 0 is observable.
3 The error term satis…es with E (εit ) = 0, and E (εit εjs ) = σ2ε if j = i
and t = s, and E (εit εjs ) = 0 otherwise.
b LSDV 6= γ
plim γ dynamic panel bias
n !∞
b LSDV = γ
plim γ
n,T !∞
b
αi = y i b LSDV y i ,
γ 1
! 1
n T
b LSDV
γ = ∑ ∑ (yi ,t 1 y i, 1)
2
i =1 t =1
!
n T
∑ ∑ (yi ,t 1 y i, 1 ) (yit yi )
i =1 t =1
T T T
1 1 1
xi =
T ∑ xit yi =
T ∑ yit y i, 1 =
T ∑ yi ,t 1
t =1 t =1 t =1
De…nition (bias)
The bias of the LSDV estimator is de…ned by:
! 1
n T
b LSDV
γ γ = ∑ ∑ (yi ,t 1 y i, 1)
2
i =1 t =1
!
n T
∑ ∑ (yi ,t 1 y i, 1 ) ( εit εi )
i =1 t =1
Let us consider the numerator. Because εit are (1) uncorrelated with αi
and (2) are independently and identically distributed, we have
n T
1
plim
n !∞ nT ∑ ∑ (yi ,t 1 y i, 1 ) ( εit εi )
i =1 t =1
T n T n
1 1
= plim
n !∞ nT ∑ ∑ yi ,t 1 εit plim
n !∞ nT ∑ ∑ yi ,t 1 εi
t =1 i =1 t =1 i =1
| {z } | {z }
N1 N2
T n
1 1 T n
plim
n !∞ nT ∑ ∑ y i, 1 εit + plim ∑ ∑ y i,
n !∞ nT t =1 i =1
1 εi
t =1 i =1
| {z } | {z }
N3 N4
1 m p
m i∑
Xi ! E (Xi ) = µ
=1
or
1 m
m i∑
plim Xi = E (Xi ) = µ
m !∞ =1
E (yi ,t 1 εit ) =0
and …nally
n T
1
N1 = plim
n !∞ nT ∑ ∑ yi ,t 1 εit =0
i =1 t =1
n T n
1 1 1 n
N4 = plim
n !∞ nT ∑ ∑ y i, 1 εi = plim
n !∞ nT
T ∑ y i, 1 εi = plim
n !∞ n i∑
y i, 1 εi
i =1 t =1 i =1 =1
Solution
The numerator of the expression of the LSDV bias satis…es:
n T
1 1 n
plim
n !∞ nT ∑ ∑ (yi ,t 1 y i, 1 ) ( εit εi ) = plim
n !∞ n i∑
y i, 1 εi
i =1 t =1 =1
Remark
n T
∑ ∑ (yi ,t 1 y i, 1 ) ( εit εi ) / (nT )
i =1 t =1
b LSDV
γ γ= n T
2
∑ ∑ (yi ,t 1 y i, 1) / (nT )
i =1 t =1
n T
1 1 n
plim
n !∞ nT ∑ ∑ (yi ,t 1 y i, 1 ) ( εit εi ) =
n !∞
plim
n i∑
y i, 1 εi
i =1 t =1 =1
b LSDV is biased when n
If this plim is not null, then the LSDV estimator γ
tends to in…nity and T is …xed.
1 n
n i∑
plim y i, 1 εi
n !∞ =1
We know that
For yi ,t 1, we have:
yi ,t 1 = εi ,t 1 + γεi ,t 2 + γ2 εi ,t 3 + ... + γt 2
εi 1
1 γt 1
+ α + γt 1 yi 0
1 γ i
2 t 2 1 γt 1
yi ,t 1 = εi ,t 1 + γεi ,t 2 + γ εi ,t 3 + ... + γ εi 1 + α + γt 1
yi 0
1 γ i
2 t 2 1 γt 1
yi ,t 1 = εi ,t 1 + γεi ,t 2 + γ εi ,t 3 + ... + γ εi 1 + α + γt 1
yi 0
1 γ i
Proof: We have (each lign corresponds to a date)
T
∑ yi ,t 1 = yi ,T 1 + yi ,T 2 + .. + yi ,1 + yi ,0
t =1
T 2 1 γT 1
= εi ,T 1 + γεi ,T 2 + .. + γ εi 1 + αi + γT 1 yi 0
1 γ
1 γT 2
+εi ,T 2 + γεi ,T 3 + ... + γT 3
εi 1 + αi + γT 2 yi 0
1 γ
+..
1 γ1
+εi ,1 + α + γyi 0
1 γ i
+yi 0
C. Hurlin (University of Orléans) Advanced Econometrics II April 2018 27 / 209
2. The dynamic panel bias
αi
1 γ+1 γ2 + ... + 1 γT 1
1 γ
αi
= T 1 γ γ2 .. γT 1
1 γ
αi 1 γT
= T
1 γ 1 γ
αi T T γ 1 + γT
=
(1 γ )2
So, we have
T
1
y i, 1 =
T ∑ yi ,t 1
t =1
1 1 γ2 1 γT 1
= εi ,T 1 + εi ,T 2 + ... + εi 1
T 1 γ 1 γ
!
T Tγ 1 + γT 1 γT
+ αi + yi 0
(1 γ )2 1 γ
1 n
n i∑
plim y i, 1 εi
n !∞ =1
1 n 1 1 γ2 1 γT 1
n i∑
= plim εi ,t 1 + + ... +
εi ,t 2 εi 1
n !∞ =1 T 1 γ 1 γ
!
T Tγ 1 + γT 1 γT 1
+ αi + yi 0 (εi 1 + ... + εiT )
(1 γ )2 1 γ T
1 n
n i∑
plim y i, 1 εi
n !∞ =1
1 n 1 1 γ2 1 γT 1
n i∑
= plim εi ,T 1 + + ... +
εi ,T 2 εi 1
n !∞ =1 T 1 γ 1 γ
!
T Tγ 1 + γT 1 γT 1
+ αi + yi 0 (εi 1 + ... + εiT )
(1 γ )2 1 γ T
σ2ε 1 γ 1 γ2 1 γT 1
= + + ... +
T2 1 γ 1 γ 1 γ
T
σ2ε T Tγ 1 + γ
=
T2 (1 γ )2
Theorem
If the errors terms εit are i.i.d. 0, σ2ε , we have:
n T
1
plim
n !∞ nT ∑ ∑ (yi ,t 1 y i, 1 ) ( εit εi )
i =1 t =1
n
1
n i∑
= plim y i, 1 εi
n !∞ =1
σ2ε T Tγ 1 + γT
=
T2 (1 γ )2
b LSDV
By similar manipulations, we can show that the denominator of γ
converges to:
n T
1
plim
n !∞ nT
∑ ∑ (yi ,t 1 y i, 1)
2
i =1 t =1
!
σ2ε 1 2γ T T γ 1 + γT
= 1
1 γ2 T (1 γ )2 T2
So, we have :
b LSDV
plim (γ γ)
n !∞
n T
1
nT ∑ ∑ (yi ,t 1 y i , 1 ) (εit εi )
i =1 t =1
= plim n T
n !∞ 1
nT ∑ ∑ (yi ,t 1 y i , 1 )2
i =1 t =1
σ2ε (T T γ 1 + γT )
T2 (1 γ )2
=
σ2ε 1 2γ (T T γ 1 + γT )
1 γ2
1 T (1 γ )2 T2
b LSDV
plim (γ γ)
n !∞
T Tγ 1 + γT
=
1 γ 2γ
T2 T (T Tγ 1 + γT )
1 +γ (1 γ )2
(1 + γ ) T Tγ 1 + γT
=
2γ
(1 γ) T 2 T (T Tγ 1 + γT )
(1 γ )2
Fact
If T also tends to in…nity, then the numerator converges to zero, and
denominator converges to a nonzero constant σ2ε / 1 γ2 , hence the
LSDV estimator of γ and αi are consistent.
Fact
b LSDV and
If T is …xed, then the denominator is a nonzero constant, and γ
b
αi are inconsistent estimators when n is large.
(1 + γ ) T Tγ 1 + γT
b LSDV
plim (γ γ) =
n !∞ 2γ
(1 γ) T 2 T (T Tγ 1 + γT )
(1 γ )2
1+γ 1 1 γT
b LSDV
plim (γ γ) = 1
n !∞ T 1 T 1 γ
1
2γ 1 γT
1 1
(1 γ ) (T 1) T (1 γ )
Fact
b LSDV is caused by having to eliminate the individual
The dynamic bias of γ
e¤ects αi from each observation, which creates a correlation of order
(1/T ) between the explanatory variables and the residuals in the
transformed model
0 1
B C
(yit y i ) = γ @yi ,t 1 y i, 1 A
| {z }
depends on past value of εit
0 1
+ @εit εi A
|{z}
depends on past value of εit
If we approximate yit by εit (in fact yit also depend on εit 1, εt 2 , ...) then
we have
1
cov (y i , 1 , εi ) = cov ((yi 1 + ... + yiT 1 ) , (εi 1 + ... + εiT ))
T2
1
' (cov (εi ,1 , εi ,1 ) + ... + (cov (εi ,T 1 , εi ,T 1 )))
T2
(T 1) σ2ε
' 6= 0
T2
(T 1) σ2ε
cov (y i , 1 , εi ) =
T2
By taking into account all the interaction terms, we have shown that
1 n σ2ε (T 1) γ 1 + γT
n i∑ 1 εi = cov (y i , 1 , εi ) =
plim y i,
n !∞ =1 T2 (1 γ )2
Remarks
1+γ 1 1 γT
b LSDV
plim (γ γ) = 1
n !∞ T 1 T 1 γ
1
2γ 1 γT
1 1
(1 γ ) (T 1) T (1 γ )
-0.05
Semi-asymptotic bias
-0.1
-0.15
-0.2
T=10
T=30
-0.25
T=50
T=100
-0.3
0 0.2 0.4 0.6 0.8 1
semi-asymptotic bias
semi-asymptotic bias
0.6 0.6
0.4 0.4
0.2 0.2
0 0
-0.2 -0.2
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
T=50 T=100
1 1
True value of True value of
plim of the LSDV estimator 0.9 plim of the LSDV estimator
0.8
0.8
0.7
semi-asymptotic bias
semi-asymptotic bias
0.6
0.6
0.4 0.5
0.4
0.2
0.3
0.2
0
0.1
-0.2 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
-20
-40
relative bias (in %)
-60
-80
T=10
-100 T=30
T=50
T=100
-120
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Step 1: parameters
i .i .d .
Let assume that γ = 0.5, σ2ε = 1 and εit N (0, 1) .
Simulate n individual e¤ects αi once at all. For instance, we can use a
uniform distribution
αi U[ 1,1 ]
b sLSDV .
and compute the LSDV estimates γ
b LSDV based on
Compute the average bias of the LSDV estimator γ
the S Monte Carlo simulations
S
1
av .bias =
S ∑ γbsLSDV γ
s =1
1+γ 1 1 γT
b LSDV
plim (γ γ) = 1
n !∞ T 1 T 1 γ
1
2γ 1 γT
1 1
(1 γ ) (T 1) T (1 γ )
2 Repeat this this experiment for various time dimensions T : when T
increases,the average bias should converge to 0.
300
250
Number of simulations
200
150
100
50
0
0.3 0.31 0.32 0.33 0.34 0.35 0.36 0.37 0.38
at
h
Click me!
-0.15
Theoretical semi-asymptotic bias
MC average bias
-0.155
-0.16
-0.165
-0.17
-0.175
-0.18
0 200 400 600 800 1000
Sample size n
Judson R.A. and Owen A. (1999), Estimating dynamic panel data models: a
guide for macroeconomists. Economics Letters, 1999, vol. 65, issue 1, 9-15.
n T γ b LSDV
Avg. γ Avg. bias
10 10 0.5 0.3282 0.1718
50 10 0.5 0.3317 0.1683
100 10 0.5 0.3338 0.1662
10 50 0.5 0.4671 0.0329
50 50 0.5 0.4688 0.0321
100 50 0.5 0.4694 0.0306
n T γ b LSDV
Avg. γ Avg. bias
10 10 0.3 0.3686 0.0686
50 10 0.3 0.3743 0.0743
100 10 0.3 0.3753 0.0753
10 50 0.3 0.3134 0.0134
50 50 0.3 0.3133 0.0133
100 50 0.5 0.3142 0.0142
b LSDV and b
In this case, both estimators γ βLSDV are biased.
Objectives
y = Xβ + ε
V ( ε j X ) = σ 2 IN
E ( εj X) 6= 0N 1
with
1 0
plim X ε = E (xj εj ) = γ 6= 0K 1
N
where β denotes the true value of the parameters. This bias is called the
endogeneity bias.
b 1
plim β OLS = β + Q γ
where Q = plim N 1 X0 X.
We have:
1
b 1 0 1 0
plim β OLS = β + plim XX plim Xε
N N
1
= β+Q γ 6= β
Remarks
b 1
plim β OLS = β + Q γ
wi N (03 1 , ∆)
with 0 1
1 0.3 0
∆ = @ 0.3 1 0.5 A
0 0.5 1
It means that Cov (εi , xi 1 ) = 0 (x1 is exogenous) but Cov (εi , xi 2 ) = 0.5
(x2 is endogenous) and Cov (xi 1, xi 2 ) = 0.3 (x1 is correlated to x2 ).
yi = β1 + β2 xi 1 + β3 xi 2 + εi
0
b = b
Denote β β1s b
β2s b
β3s the OLS estimates obtained from the
s
simulation s 2 f1, ..S g . (3) compare the true value of the parameters in
the population (DGP) to the average OLS estimates obtained for the S
simulations
De…nition (Instruments)
Consider a set of H variables zh 2 RN for h = 1, ..N. Denote Z the N H
matrix (z1 : .. : zH ) . These variables are called instruments or
instrumental variables if they satisfy two properties:
(1) Exogeneity: They are uncorrelated with the disturbance.
E ( εj Z) = 0N 1
E (xjk zjh ) 6= 0
E Z0 Z is non singular
or equivalently
rank E Z0 Z =H
E ( εj j zj ) = 0 =) E (εj zj ) = 0H
E @ zj yj xj0 β A = 0H
(H ,1 ) (1,1 ) (H ,1 )
De…nition (Identi…cation)
The system is identi…ed if there exists a unique vector β such that:
E zj yj xj0 β =0
Remark
H=K
De…nition (Consistency)
b is
Under the assumption that plim N 1 Z0 ε = 0, the IV estimator β IV
consistent:
p
b !
β β
IV
So, we have:
1
b = β + plim 1 0 1 0
plim β IV ZX plim Zε
N N
where
1 0 1 0
QZZ = plim ZZ QZX = plim ZX
K K N K K N
b σ2
Vasy β IV = Q 1 QZZ QZX1
N ZX
A consistent estimator is given by
b
b asy β 1 1
V IV b 2 Z0 X
=σ Z0 Z X0 Z
QZX = QXZ
the estimator can also written as
b
b asy β 1 1
V IV b 2 Z0 X
=σ Z0 Z Z0 X
ε0b
b ε 1 N 2
b2 =
σ
N K
=
N K ∑ yi b
xi0 β IV
i =1
Relevant instruments
1 0
plim Z X = QZX a …nite H K positive de…nite matrix
N
1 While strictly speaking, this condition is su¢ cient to determine the
asymptotic properties of the IV estimator
2 However, the common case of “weak instruments,” is only barely
true has attracted considerable scrutiny.
b σ2
Vasy β IV = Q 1 QZZ QZX1
N ZX
As soon as N 1 Z0 X ' 0H K , the estimated asymptotic variance
covariance is also very large since
b
b asy β 1 1
V IV b 2 Z0 X
=σ Z0 Z X0 Z
H>K
Introduction
If Z contains more variables than X, then much of the preceding derivation
is unusable, because Z0 X will be H K with
rank Z0 X = K < H
Introduction (cont’d)
The crucial assumption in the previous section was the exogeneity
assumption
1
plim Z0 ε = 0K 1
N
1 That is, every column of Z is asymptotically uncorrelated with ε.
2 That also means that every linear combination of the columns of Z
is also uncorrelated with ε, which suggests that one approach would
be to choose K linear combinations of the columns of Z.
Introduction (cont’d)
Which linear combination to choose?
A choice consists in using is the projection of the columns of X in the
column space of Z:
Xb = Z Z0 Z 1 Z0 X
b for Z, we have
With this choice of instrumental variables, X
1
b
β = b 0X
X b 0y
X
2SLS
1 1 1
= X0 Z Z0 Z Z0 X X0 Z Z0 Z Z0 y
b = Z Z0 Z 1
where X Z0 X corresponds to the projection of the columns of
X in the column space of Z, or equivalently by
1 1 1
b 0 0
Z0 X X0 Z Z0 Z Z0 y
β 2SLS = X Z Z Z
Since
b = Z Z0 Z 1
X Z0 X = PZ X
where PZ denotes the projection matrix on the columns of Z. Reminder:
PZ is symmetric and PZ PZ0 = PZ . So, we have
1
b
β = X0 PZ X
0 b 0y
X
2SLS
1
=
0
X0 PZ PZ X b 0y
X
1
= b 0X
X b b 0y
X
xkj = b
b α1 z1j + b
α2 z2j + .. + b
αH zHj
yj = β1 b
x1j + β2 b
x2j + .. + βK b
xKj + εj
Theorem
If any column of X also appears in Z, i.e. if one or more explanatory
(exogenous) variable is used as an instrument, then that column of X is
b
reproduced exactly in X.
Z = (z1 : .. : zJ : x1 : .. : xK 1)
Then
b = (x1 : .. : xK
X 1 :b
xK )
where b
xK denotes the projection of xK on the columns of Z.
Objectives
Remark
If the vector ω i includes a constant term, the associated parameter can be
interpreted as the mean of the (random) individual e¤ects
0 0
yit = γyi ,t 1 + β xit + ρ ω i + αi + εit
αi = µ + αi E ( αi ) = 0
0 1 0 1
1 µ
B zi 2 C B ρ C
ωi = B @ ... A
C ρ =B 2 C
@ ... A
(K 2 ,1 ) (K 2 ,1 )
ziK 2 ρK 2
Vectorial form:
0
yi = yi , 1γ + Xi β + ω i ρe + αi e + εi
Remarks
De…nition (Instruments)
Anderson and Hsiao (1982) consider two sets of K1 + 1 instruments, in
both cases the system is just identi…ed (IV estimator):
!0
0
zi = yi ,t 2 : (xit xi ,t 1)
(K 1 +1,1 ) (1,1 ) (1,K 1 )
!0
0
zi = (yi ,t 2 yi ,t 3 ) : (xit xi ,t 1 )
(K 1 +1,1 ) (1,1 ) (1,K 1 )
Remarks
1 The …rst estimator (with zit = yi ,t 2 ) has an advantage over the
second one (with zit = yi ,t 2 yi ,t 3 ), in that the minimum number
of time periods required is two, whereas the …rst one requires T 3.
2 In practice, if T 3, the choice between both depends on the
correlations between (yi ,t 1 yi ,t 2 ) and yi ,t 2 or (yi ,t 2 yi ,t 3)
=> relevance assumption.
Third step
0 0
yit = γyi ,t 1 + β xit + ρi ω i + αi + εit
b IV and b
Given the estimates γ βIV , we can deduce an estimate of ρ,
the vector of parameters for the time-invariant variables ω i .
with vi = αi + εi .
0
with hi = y i b IV y i ,
γ 1
b
βIV x i .
De…nition
b IV , b
Given γ βIV , and b
ρ, we can estimate the variances as follows:
T n
1
n (T 1) t∑ ∑ bε2it
b2ε =
σ
=2 i =1
2
1 n 0 1 2
n i∑
b2α =
σ yi b IV y i ,
γ 1
b
βIV x i ρ0 zi
b b
σ
=1 T ε
with
0
bεit = (yi ,t yi ,t 1) b IV (yi ,t
γ 1 yi ,t 2)
b
βIV (xi ,t xi ,t 1)
Theorem
The instrumental-variable estimators of γ, β, and σ2ε are consistent when
n (correction of the Nickell bias), or T , or both tend to in…nity.
b IV = γ
plim γ plim b
βIV = β b2ε = σ2ε
plim σ
n,T !∞ n,T !∞ n,T !∞
The estimators of ρ and σ2α are consistent only when n goes to in…nity.
plim b
ρ=ρ b2α = σ2α
plim σ
n !∞ n !∞
for t = 2, .., T .
θ = β, γ, ρ, σ2α , σ2ε
Arellano, M., and S. Bond (1991). “Some Tests of Speci…cation for Panel
Data: Monte Carlo Evidence and an Application to Employment Equations,”
Review of Economic Studies, 58, 277–297.
such that
0
E yi ,t 2 j ∆yit γ0 ∆yi ,t 1 β0 ∆xit =0
Assumption: exogeneity
We assume that the time varying explanatory variables xit are strictly
exogeneous in the sense that:
0
E xit εis = 0 8 (t, s )
0 0 0
where xi = xi 1 , .., xiT . At time t = 3, we have
! 00 1 1
yi 0
E qi 3 ∆εi 3 = E @ @ yi 1 A ( ε i 3 εi 2 ) A = 0
(2 +TK 1 ,1 )
(2 +TK 1 ,1 ) (1,1 ) xi0
where :
0 1 0 1
yi 2 yi 1 yi 1 yi 0
B yi 3 yi 2 C B yi 2 yi 1 C
∆yi =B
@
C ∆yi ,
A 1 =B
@
C
A
(T 1,1 ) .. (T 1,1 ) ..
yiT yi ,T 1 yiT 1 yi ,T 2
where :
0 1 0 1
xi 2 xi 1 εi 2 εi 1
B xi 3 xi 2 C B εi 3 εi 2 C
∆Xi =B
@
C
A ∆εi =B
@
C
A
(T 1,K 1 ) .. (T 1,1 ) ..
xiT xi ,T 1 εiT εi ,T 1
0 1
qi 2 0 ... 0
B (1 +TK 1 ,1 ) C
B 0 qi 3 C
B C
Wi = B
B (2 +TK 1 ,1 ) C
C
B .. C
@ A
0 .. qiT
(T 1 +TK 1 ,1 )
yi 0
E (qi 2 ∆εi 2 ) = E ( εi 2 εi 1 ) =0
xi0
Example
For T = 10 et K1 = 0 (without explicative variable), we have
T (T 1)
r= = 45 orthogonal conditions
2
Example
For T = 50 et K1 = 0 (without explicative variable), we have
T (T 1)
r= = 1225 orthogonal conditions !!
2
4500
4000
3500
3000
2500
2000
1500
1000
500
0
0 10 20 30 40 50 60 70 80 90 100
T
Assumption: pre-determination
We assume that the time varying explanatory variables xit are
pre-determined in the sense that:
E xit0 εis = 0 if t s
De…nition
The conditions E (qit ∆εit ) = 0 for t = 2, .., T , can be written as
!
E Wi ∆εi = 0
(r ,T 1 )(T 1,1 ) (m,1 )
0 1
qi 2 0 ... 0
B (1 +K 1 ,1 ) C
B 0 qi 3 C
B C
Wi = B
B (2 +2K 1 ,1 ) C
C
B .. C
@ A
0 .. qiT
(T 1 +(T 1 )K 1 ,1 )
r = 1 + K1 + 2 + K1 .. + (T 1) K1 + (T 1)
= (1 + K 1) (1 + 2 + ... + (T 1))
T (T 1)
= ( 1 + K1 )
2
10000
X exogeneous
X pre-determined
5000
0
0 10 20 30 40 50 60 70 80 90 100
T
Fact
Whatever the assumption made on the explanatory variable, the number of
othogonal conditions (moments) r is much larger than the number of
parameters, e.g. K1 + 1. Thus, Arellano and Bond (1991) suggest a
generalized method of moments (GMM) estimator.
Arellano, M., and S. Bond (1991). “Some Tests of Speci…cation for Panel
Data: Monte Carlo Evidence and an Application to Employment Equations,”
Review of Economic Studies, 58, 277–297.
E (Wi ∆εi ) = 0
0
to estimate by GMM the parameters θ = γ, β0 in
De…nition
The standard method of moments estimator consists of solving the
unknown parameter vector θ by equating the theoretical moments with
their empirical counterparts or estimates.
E (m (yt ; θ 0 )) = 0
1 n
n t∑
b (y , θ ) =
m m (yt ; θ )
=1
b y,b
m θ =0
E (m (yt ; θ 0 )) = 0
1 n p
b (y , θ ) =
m ∑
n t =1
m (yt ; θ ) ! E (m (yt ; θ ))
In particular
p
b (y , θ 0 ) ! E (m (yt ; θ 0 )) = 0
m
So, the GMM consists in …nding b
θ such that
p
b y,b
m θ = 0 =) b
θ ! θ0
2E yt2
v=
E (yt2 ) 1
1 n 2 p
n t∑
b2 =
µ yt ! µ2
=1
2b
µ2
vb =
b2
µ 1
v
E (m (yt ; v )) = E yt2 =0
v 2
v
E (m (yt ; v )) = E yt2 =0
v 2
1 n 1 n v
b (y ; v ) =
m ∑
n t =1
m (yt ; v ) = ∑ yt2
n i =1 v 2
2b
µ2 p 2E yt2
b (y ; vb) = 0 ,
m vb = !v =
b2
µ 1 E (yt2 ) 1
b (y ; θ ) 0 S
q (y , θ ) = m 1
b (y ; θ )
m
3v 2
µ4 = E yt4 =
(v 2) (v 4)
The two moment conditions (r = 2) can be written as
!
yt2 v v 2 0
E (m (yt ; v )) = E 2 =
yt4 (v 23v)(v 4 ) 0
b (y ; v ) 0 S
m 1
b (y ; v )
m
then we have
!2
1 n 2 v
n t∑
0 1
b (y ; v ) S
m b (y ; v ) =
m yt
=1 v 2
!2
1 n 2 3v 2
n t∑
+2 yt
=1 (v 2) (v 4)
b (y ; v ) 0 S
It is now possible to …nd vb such that m 1 b (y ; v ) = 0
m
b b (y ; θ ) 0 S
θ = arg min q (y , θ ) = arg min m 1
b (y ; θ )
m
θ 2 Ra (1,1 ) θ 2 Ra (1,r ) (r ,r ) (r ,1 )
b b (y ; θ ) 0 S
θ = arg min q (y , θ ) = arg min m 1
b (y ; θ )
m
θ 2 Ra (1,1 ) θ 2 Ra (1,r ) (r ,r ) (r ,1 )
Remark
The optimal weighting matrix is
∞
S= ∑ E m (yt ; θ 0 ) m (yt j ; θ 0 )0
j= ∞
We can replace the unknow value θ 0 by the GMM estimator θ̂ and the
optimal weighting matrix becomes
∞ 0
S= ∑ E m yt ; b
θ m yt j ; b
θ
j= ∞
n 0
bj = 1 ∑ m yt ; b
Γ θ m yt j ; b
θ
n t =j +2
But, this estimator may be no positive de…nite...
n 0
bj = 1 ∑ m yt ; b
Γ θ m yt j ; b
θ
n t =j +2
If q = 2 then we have
S b0 + 2 Γ
b=Γ b10 + 1 Γ
b1 + Γ b2 + Γ
b20
3 3
3 sin (6πu/5)
K (u ) = 2
cos (6πu/5)
(6πu/5) (6πu/5)
Problem 2
b b (y ; θ ) 0 S
θ = arg min m 1
b (y ; θ )
m
θ 2 Ra
∞ 0
S= ∑ E m yt ; b
θ m yt j ; b
θ
j= ∞
1 In order to compute b
θ, we have to know S 1.
Solutions
Solutions
Two-step GMM
Step 1: put the same weight to the r moment conditions by using an
identity weighting matrix
S0 = Ir
Compute a …rst consistent (but not e¢ cient) estimator b
θ0
b
θ0 = b ( y ; θ ) 0 S0 1 m
arg min m b (y ; θ )
θ 2 Ra
= b (y ; θ ) 0 m
arg min m b (y ; θ )
θ 2 Ra
Two-step GMM
b1
Step 2: Compute the estimator for the optimal weighting matrix S
q
j
b1 = Γ
S b0 + ∑K q+1
bj + Γ
Γ bj0
j =1
n 0
bj = 1 ∑ m yt ; b
Γ θ 0 m yt j ; b
θ0
n t =j +2
b b (y ; θ ) 0 S
θ 1 = arg min m b 1m
1 b (y ; θ )
θ 2Ra
Various GMM estimators (i.e. moment conditions) have been proposed for
dynamic panel data models
Various GMM estimators (i.e. moment conditions) have been proposed for
dynamic panel data models
Problem
Let us consider the dynamic panel data model in …rst di¤erences
Let us denote
with
E (m (yi , xi ; θ )) = 0r
or equivalently
! !
1 n 1 n
n i∑ n i∑
b
θ = arg min ∆εi0 Wi0 S 1
Wi ∆εi
θ 2 RK 1 + 1 =1 =1
with S = E (m (y ; θ 0 ) m (y ; θ 0 ))0 .
Various GMM estimators (i.e. moment conditions) have been proposed for
dynamic panel data models
vi = yi γy i , 1 β0 x i ρ0 ω i
De…nition
Arellano and Bond (1991) consider the following moment conditions
De…nition
Arellano and Bover (1995) consider additional moment conditions
E (m (yi , xi ; θ )) = E q
ei y i γy i , 1 β0 x i ρ0 ω i =0
Various GMM estimators (i.e. moment conditions) have been proposed for
dynamic panel data models
Apart from the previous linear moment conditions, Ahn and Schmidt
(1995) note that the homoscedasticity condition on E ε2it implies the
following T 2 linear conditions
Various GMM estimators (i.e. moment conditions) have been proposed for
dynamic panel data models
Blundell and Bond, S. (2000): GMM Estimation with persistent panel data:
an application to production functions. Econometric Reviews,19(3), 321-340.
Remarks
Remarks