Econometrics Lecture4 MultipleRegression
Econometrics Lecture4 MultipleRegression
Yaein Baek
Sogang University
Spring 2024
Y = β0 + β1 X1 + β2 X2 + · · · + βk Xk + u
▶ β0 : intercept
▶ βj : slope parameter associated with Xj , j = 1, . . . , k
▶ u: error term (disturbance) that contains factors other than
X1 , X2 , . . . , Xk that affect Y
∆Y
βj =
∆Xj
E (u|X ) = 0
E (u|X1 , X2 , . . . , Xk ) = 0
E (Y |X1 , X2 , . . . , Xk ) = β0 + β1 X1 + · · · + βk Xk
∂E (Y |X1 = x1 , X2 = x2 , . . . , Xk = xk )
βj =
∂xj
where b ≡ (b0 , b1 , . . . , bk )
The OLS estimators (β̂0 , β̂1 , . . . , β̂k ) are obtained from (k + 1) FOCs
X
(Yi − β̂0 − β̂1 Xi1 − · · · − β̂k Xik ) = 0
i
X
(Yi − β̂0 − β̂1 Xi1 − · · · − β̂k Xik )Xi1 = 0
i
..
.
X
(Yi − β̂0 − β̂1 Xi1 − · · · − β̂k Xik )Xik = 0
i
ûi = yi − ŷi
Yi = β0 + β1 Xi1 + β2 Xi2 + ui ,
Frisch-Waugh theorem
It can be shown that β̂j can be obtained from the following procedure
1 Regress Y on X1 , . . . , Xj−1 , Xj+1 , . . . , Xk and obtain the residual û Y
2 Regress Xj on X1 , . . . , Xj−1 , Xj+1 , . . . , Xk and obtain the residual û Xj
3 Regress û Y on û Xj
R-squared
SSE SSR
R2 ≡ =1−
SST SST
R 2 is non-decreasing in the number of independent variables
R-squared never decreases, and usually increases when other independent variable is
included in the regression.
Even if we add an independent variable that is irrelevant to Y, R-squared increases.
MLR Assumptions
MLR.1 (Linear in Parameters) The population model is
Y = β0 + β1 X1 + β2 X2 + · · · + βk Xk + u
MLR Assumptions
MLR.3 (No Perfect Collinearity) In the sample (and therefore in
the population), none of the independent variables is constant, and
there are no exact linear relationships among the independent
variables.
MLR Assumptions
MLR.4 (Zero Conditional Mean) The error u has an expected
value of zero given any values of the independent variables.
E (u|X1 , X2 , . . . , Xk ) = 0.
Notations:
▶ Vector of n observations of the jth explanatory variable:
Xj = (X1j , . . . , Xij , . . . , Xnj )′ , j = 1, . . . , k
▶ X ≡ [1 X1 X2 · · · Xk ], where 1 is a (n × 1) vector of ones
Assumption MLR.4 with MLR.2 implies E (ui |X) = 0 for i = 1, . . . , n
E (β̂j ) = βj , j = 0, 1, . . . , k
for any values of the population parameters βj . In other words, the OLS
estimators are unbiased estimators of the population parameters.
The true population model has two explanatory variables and an error
term
Y = β0 + β1 X1 + β2 X2 + u
and assume this model satisfies Assumptions MLR.1-MLR.4
Suppose we specify the model by excluding X2 :
Y = β̃0 + β̃1 X1 + ũ
X2 = γ0 + γ1 X1 + v ,
Therefore,
E β̃ˆ1 |X1 = β1 + β2 · γ1
h i
E β̃ˆ1 |X1 − β1 = β2 · γ1
h i
MLR Assumptions
MLR.5 (Homoskedasticity): The error u has the same variance
given any value of the explanatory variables.
Var(u|X1 , . . . , Xk ) = σ 2
σ2
Var(β̂j |X) = , j = 1, . . . , k
SSTj (1 − Rj2 )
where SSTj = ni=1 (Xij − X̄j )2 and Rj2 is the R-squared from regressing
P
Xj on all other independent variables (including an intercept).
Yi = β0 + β1 X1 + β2 X2 + U
where SSTj = ni=1 (Xij − X̄j )2 and Rj2 is the R-squared from
P
regressing Xj on all other independent variables (including an
intercept)
df = n − (k + 1)
= (number of observations) − (number of estimated parameters)
σ
q
sd(β̂j ) = Var(β̂j ) = q
SSTj (1 − Rj2 )
σ̂
q
se(β̂j ) = Var(
d β̂j ) = q
SSTj (1 − Rj2 )
Var(β̂j ) ≤ Var(β̃j ), j = 0, 1, . . . , k
Pn
for all β̃j = i=1 Wij Yi for which E (β̃j ) = βj , j = 0, 1, . . . , k.