cn2 Multi
cn2 Multi
[1] MULTICOLLINEARITY
(1) Perfect Collinearity:
• When regressors are perfectly linearly related: The X matrix is less than
full rank (Rank(X) < k).
Example 1:
GDPt = β1 + β2Gt + β3Tt + β4DEFt + εt.
→ DEFt = Gt - Tt for any t.
→ Rank(X) = 3.
MULTI-1
(2) Near Multicollinearity
• When regressors are highly (not perfectly) correlated.
MULTI-2
Theorem:
xj = j'th column of X;
Xj* = X with j'th column deleted.
SSEj = SSE from a regression of xj on Xj*.
SSTj = Σt(xtj- x j )2.
1 1
Then, the j'th diagonal of (X′X)-1 = = .
SSE j SST j (1 − R j 2 )
Implication:
σ2
• var( βˆ j ) = → As Rj2↑, var( βˆ j )↑.
SST j (1 − R j )
2
s2
• se( βˆ j ) = → As Rj
2
↑, se( βˆ j )↑.
SST j (1 − R j )
2
βˆ j
• (t statistic for Ho: βj = 0) = → As Rj2↑, |t|↓.
se( βˆ j )
3) Remedies
1. Drop some regressors highly correlated with others (?)
2. Collect a richer data set.
3. Use alternative estimators.
MULTI-3
(3) Alternative Estimators
1) Ridge regression estimator:
βˆr = (X′X + rIk)-1X′y, r > 0.
• Biased but smaller MSE than OLS variances.
Cov( βˆr ) = σ2[X′X + rIk]-1X′X[X′X + rIk]-1.
• This estimator solves multicollinearity problem?
• No clear meaning to statistical inferences.
• What is optimal choice of r?
MULTI-4
• Facts:
3) Suggestion:
Ridge regression or PC estimators could be useful for prediction,
but may not be much helpful for statistical inferences.
MULTI-5
<Example>
Dependent Variable: LWAGE
Sample: 1 935
2) Alternative regression:
Step 1: Regress xt3 on one and xt4, and residuals ut.
Step 2: Regress yt on one, xt2, ut and xt4.
<Example>
Dependent Variable: LWAGE
Included observations: 935
MULTI-6
3) Logic of this treatment:
• Let xt 3 = δ 1 + δ 2 xt 4 + ut . Substitute this into the original regression
model:
yt = β1 + β 2 xt 2 + β 3 (δ 1 + δ 2 xt 4 + ut ) + β 4 xt 4 + ε t
= ( β1 + β 3δ 1 ) + β 2 xt 2 + β 3ut + ( β 4 + δ 2 β 3 ) xt 4 + ε t .
Since ut and xt4 are uncorrelated, this alternative model does not
suffer from M.
• Is it really?
MULTI-7