Vb V ε X = σ Vb = σ Vb = X'X Σx X'X: I X'X X'
Vb V ε X = σ Vb = σ Vb = X'X Σx X'X: I X'X X'
Review
Testing IV assumptions
1 of 9
Important because one is using IV in the first place because there is doubt about
endogeneity, and never obvious that instruments are both exogenous and highly
correlated with X.
H 0 : β̂ 2SLS − b = 0
H A : β̂ 2SLS − b ≠ 0
( ){
′
} (β̂ )
−1
H = β̂ 2SLS − b Est.Asy.Var ⎡⎣ β̂ 2SLS − b ⎤⎦ 2SLS
−b
−1 −1
σ2 ⎛ X̂ ' X̂ ⎞ σ2 ⎛ X'X⎞
Asy.V ⎡⎣ β̂ 2SLS ⎤⎦ − Asy.V ⎡⎣b ⎤⎦ = plim ⎜ − plim ⎜
n ⎝ n ⎠ ⎟ n ⎝ n ⎟⎠
since X̂ is a estimate of X
2 of 9
- it is less correlated with X than X is with itself
- (unless columns of Z perfectly predict columns of X )
−1 −1
⎛ X̂ ' X̂ ⎞ ⎛ X 'X ⎞
> plim ⎜ so Asy.Var ⎡⎣ β̂ 2SLS ⎤⎦ > Asy.Var [ b ]
⎝ n ⎟⎠
plim ⎜
⎝ n ⎟⎠
Second, he proved that
the Cov between an efficient estimator (b)
and the difference with an inefficient estimator ( β̂ 2SLS )
for the same parameter is zero.
( ) − s2 ( X ' X )
−1 −1
Est.Asy.V ⎡⎣ β̂ 2SLS − b ⎤⎦ = s 2 X̂ ' X̂
(
H = β̂ 2SLS − b )′ (V ⎡⎣β̂ 2SLS
⎤ − V ⎡b ⎤ β̂
⎦ )(
⎣ ⎦ 2SLS − b )
2. Overidentification test
- only possible if L > K
E ( z i ε i ) = 0 othogonality condition
⎛1 n ⎞
E(m) = E ⎜ ∑ z i ε i ⎟ = 0 , even though not exactly true in sample
⎝ n i=1 ⎠
1 n
So test whether ∑ z iε i = 0 when L>K
n i=1
i.e. test m = 0
n
1 1 n
- use m = ∑ i IV ,i n ∑ z i (yi − xi ' β̂ IV )
n i=1
z e =
i=1
3 of 9
β̂ IV already forces first K moment conditions to be exactly equal to
zero
1 n
2 ∑ ( i IV ,i ) ( i IV ,i )
1
- Est.Var ( m ) = ze z e ' = 2 Z'e IV e IV 'Z
n i=1 n
1 n 1
- m = ∑ z i eIV ,i = Z'e IV
n i=1 n
- so Wald stat is χ 2 = e IV 'Z [ Z'e IV e IV 'Z ] Z'e IV
−1
- can view this as a test of whether the instruments give the same answer
as each other.
( X 'X )kk
Rk2 = ,
( X̂ ' X̂ ) kk
R (n − L )
2
then k
∼F
1− Rk2 ( L − 1)
- more complicated with multiple endogenous xk
4 of 9
Measurement Error
yi* = β xi* + ε i
yi = yi* + υi
xi = xi* + ui
- if only error in yi , no problem
yi = β xi* + ε i + υi = β xi* + ε i′
- if error in xi , big problem
yi = β xi + ε i − β ui = β xi + wi
Cov[ xi ,wi ] = Cov ⎡⎣ xi* + u i , ε i − β ui ⎤⎦ = − βσ u2
- violates exogeneity of x
β
plim b = attenuation bias - b too small
1+ σ u / plim(x'x)
2
Panel Data
Have cross section data on units (the “panel”) repeatedly measured over time
- AKA cross-section time-series data (“xt” in Stata)
Nothing inherently problematic, just allows you to correct for more issues
- an opportunity to make more precise estimates
- in particular, to control for all unchanging individual characteristics
- with an individual-specific constant term
yit = x it′ β + ε it
5 of 9
if Ti is different for each individual, “unbalanced panel”
- in general, just complicates the notation a bit
- rarely a substantive issue, unless you are programming estimators
6 of 9
with A1-A4, we can estimate this with OLS
consistent and efficient
known as “fixed effects”, but doesn’t mean that α i are not random variables
- misnomer
Issues:
1) α i not consistently estimated
- each α i just estimated from T observations
- imagine we just had data on 1 individual
- could still estimate that α i
- since T is typically small, too few obs for consistent estimate
- typically less than 25, almost certainly less than 100
- often said that “T is assumed fixed”
- not a good way to say it
- T just too small for accurate estimates
- and asymptotic approximations
- therefore can’t trust value of α i
but we have controlled for all unchanging individual characteristics
7 of 9
instead subtract off individual means:
yit = x it′ β + α i + ε it
1 T
yi ≡ ∑ yit
T t=1
yi = xi′β + α i + ε i n.b. α i = α i
( ) ( ) ( )
V ε it − ε i X = V ε it X + V ε i X − 2Cov(ε it , ε i X)
⎛ ε +!+ ε ⎞ 1 σ2
V (ε X ) = V ⎜ X⎟ =
i1
Tσ iT 2
=
i
⎝ T ⎠ T 2
T
ε i1 +!+ ε iT ε
Cov(ε it , ε i X) = Cov(ε it , X) = Cov(ε it , it X)
T T
because Cov(ε it , ε is X) = 0 ∀ t ≠ s
σ2
Cov(ε it , ε i X) =
T
σ2 σ2 ⎛ 1⎞ 2
so, ( )
V ε it − ε i X = σ +
T
−2 2
= 1−
T ⎜⎝ T ⎟⎠
σ
8 of 9
σ2
(
Cov(ε i , ε i X) = V ε i X = ) T
, so
σ 2 σ 2 −σ 2
Cov ( ε it − ε i , ε is − ε i X ) = −2 + =
T T T
⎡ ⎛ 1 ⎞ 2 −σ 2 −σ 2 ⎤
⎢ ⎜ 1− ⎟ σ ! ⎥ ⎡ ε1 ⎤
⎢ ⎝ T⎠ T T ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ! ⎥
−σ 2 ⎢ ε1 ⎥
⎢ ! ! " ⎥
Let Σ i = ⎢ T ⎥ , and [ ε ] = ⎢ ⎥
! ⎥ , then
⎢ −σ 2 ⎥ i ⎢
⎢ " ! ! ⎥ ⎢ εn ⎥
⎢ T ⎥ ⎢ ⎥
⎢ ! ⎥
⎢ −σ 2 −σ 2 ⎛ 1⎞ 2 ⎥
⎢ ! ⎜⎝ 1− ⎟⎠ σ ⎥ ⎢⎣ εn ⎥
T T T ⎦
⎢⎣ ⎥⎦
⎡ Σ 0 ! 0 ⎤
⎢ 1 ⎥
⎢ 0 Σ2 ! 0 ⎥
V ⎡⎣ε − ⎡⎣ε i ⎤⎦ X ⎤⎦ = ⎢ ⎥
⎢ ! ! " ! ⎥
⎢ 0 0 ! Σn ⎥
⎣ ⎦
How big is this matrix?
nT x nT
Are our OLS assumptions met?
No - Autocorrelation within individual time series
Use GLS - easy to form P matrices
s2 σ2
because just need estimate of
T T
yit = x it′ β + α i + δ t + ε it
1 n
if yt = ∑ yit , then regress yit − yi − yt on xit′ − xi′ − xt′
n i=1
9 of 9