Lecture - Multiple - Regression - Analysis
Lecture - Multiple - Regression - Analysis
UPB
Semestre II - 2022
Multiple regression
Model with two independent variables
yi
xi1
xi2
β0
β1
β2
ui
Interpretation
ŷ = βˆ0 + βˆk xk
Variation
Partial effect, ceteris paribus
Partial-out
1 Regress x , over intercept and all other variables and obtein the
j
residual
ˆ
xˆj = γ0 + γˆ−j x−j
rˆj = xj − xˆj
2 Regress y P
on the residual with no intercept
n
rˆij yi
βˆ1 = Pi=1
n 2
i=1 rˆij
Partialling out
Goodness-of-Fit
(yi − ȳ )2
P
Total sum of squares (SST),
Explained sum of squares (SSE), (ŷi − ȳ )2
P
P 2
Residual sum of squares (SSR), ûi
SST = SSE + SSR
Unbiasedness assumptions
Required assumptions
Linear parameters
Random sampling
No perfect collinearity
Selection of variables
Too many variables
Include not related variables
No effect on our parameters
OLS remains unbiased
y = β0 + β1 x1 + β2 x2 + u y = β0 + β1 x1 + v
Bias ̸= 0 :
Homoskedasticity
The error u has the same variance given any value of the explanatory
variables
Var (u|x1 , ..., xk ) = σ 2
Var (y |x1 , ..., xk ) = σ 2
Gauss-Markov assumptions (5)
Assumptions for unbiasedness
Linear parameters
Random sampling
No perfect collinearity
Zero conditional mean
Homoskedasticity assumption
σ2
Var (β̂j ) =
SSTj (1 − Rj2 )
Components
σ2
(xij − x̄j )2
P
SSTj =
Rj2 , regressing xj on all other x’s
Standard deviation q
ˆ (β̂j |x
se(β̂j ) = Var
Under Gauss-Markov the OLS estimator β̂j for βj is the best linear
unbiased estimator
BLUE, Best Linear Unbiased Estimator
Best, having the smallest variance
Linear, can be expressed as a linear function of the data on the
dependent variable
n
X
β̃j = wij yi
i=1
Unbiased, assumptions