AE 2023 Lecture3 PDF
AE 2023 Lecture3 PDF
Model
Applied Econometrics
Dr. Le Anh Tuan
1
Introducing Multiple Regression
►“Explains variable ! in terms of variables "1, "2, … , "'
Dependent variable, Independent variables,
Explained variable, explanatory variables,
Response variable,… regressors,…
2
Why Do We Need Multiple Regression?
►Control for many other factors which simultaneously
affect the dependent variable
►Reduce omitted variable bias
3
Why Do We Need Multiple Regression?
► Useful for creating functional relationships between
variables
► Quadratic function
4
Simple Regression vs. Multiple Regression
► Most of the properties of the simple regression model
directly extend to the multiple regression case
→ same principles
► Population regression model
!" = $% + $' ("' + $) (") + ⋯ + $+ ("+ + ,i
► Fitted values of y:
-% + $
!/" = $ -' ("' + $
-) (") + ⋯ + $
-+ ("+
► Residuals
/ " = !" − !
, -% + $
/" = !" − $ -' ("' + $
-) (") + ⋯ + $
-+ ("+
5
OLS model
6
Properties of OLS
7
Interpreting the coefficients
►Interpretation of the multiple linear regression model
By how much does the dependent variable change
if the j-th independent variable is increased by one
unit, holding all other independent variables and the
error term constant
8
Example
►Determinants of college GPA
Grade point
High school grade Achievement test
average at college
point average score
10
Goodness of Fit
►Decomposition of total variation: SST = SSE + SSR
11
Goodness of Fit
12
Goodness of Fit
►Econometricians came up with adjusted R-squared,
which enables a comparison of models:
()*
"! # = 1 − 1 − " # .
()+)*
14
Assumptions of MLR
► Assumption MLR.1 (Linear population model). In the
population, the relationships between dependent and
independent variables are linear.
! = #$ + #& '1 + #) '2 + ⋯ + #, '- + .
where #$ , #& , #) , … , #, are the population intercept and
slope parameters, respectively.
15
Assumptions of MLR
► Assumption MLR.3 (no perfect collinearity). In the
sample (and therefore in the population), none of the
independent variables is constant, and there are no exact
linear relationships among the independent variables.
16
MLR3. no perfect collinearity
► Example for perfect collinearity: relationships between
regressors
17
Assumptions of MLR
► Assumption MLR.4 (zero conditional mean of u).
The error ! has an expected value of zero given any
value of the explanatory variables. In other words,
"[!|%] = 0.
18
Endogenous vs Exogenous
19
Unbiasedness of OLS
► Theorem 3.1 (Unbiasedness of OLS)
20
Including Irrelevant Variables in a
Regression Model
► Estimated model: ! +$ + #
*=# + & '& + , (omitted -2 )
22
Variances of the OLS estimators
23
Variances of the OLS estimators
► Theorem 3.2 (Variances of the OLS estimators)
► Under assumptions MLR.1 – MLR.5: the variance of !
24
Variances of the OLS estimators
The more the total variation in !" Remember that σ2 is the variance of
, the more accurate the OLS the error term u. If the variance of u
estimates of #" will be. reduces, the accuracy of OLS
What can help: adding more estimators increases.
observations increases $$%" . What can help: adding explanatory
variables (taking some factors out of u)
26
Example : Multicollinearity
Average Other ex-
Expenditures Expenditures for in-
standardized penditures
for teachers structional materials
test score of school
27
Variances in misspecified models
► The choice of whether to include a particular variable in a
regression can be made by analyzing the tradeoff between
bias and variance
► True population model:
► Estimated model 1
► Estimated model 2
28
Variances in misspecified models
► Case 1:
► Case 2:
29
Variances of the OLS estimators
► Estimating the error variance:
30
Theorem 2.3
► Calculation of standard errors for regression coefficients
The estimated
sampling variation of
the estimated
Plug in for the unknown
31
Gauss-Markov Theorem
► The Gauss-Markov theorem shows that, within a certain
class of unbiased estimators, OLS is the one that exhibits
the smallest variance among all competing estimators
32
Gauss-Markov Theorem
On the meaning of BLUE:
►Best:
►“best” actually means “the one with the lowest
variance
►Linear:
►An estimator of multiple regression coefficients is
linear
►OLS can be shown to be a linear estimator
►Unbiased
►the expected value of the estimate of "!# is the real
value "#
►Estimator
33
Exercise
34