Lectures - Multiple - Regression - Inference
Lectures - Multiple - Regression - Inference
UPB
Semestre II - 2022
OLS, BLUE
Add another assumption, Normality
u ∼ Normal(0, σ 2 )
The population assumption
y |x ∼ Normal(β0 + β1 x1 + ... + βk xk , σ 2 )
Homoskedastic
Sometimes normality is not the case, but large samples will let us drop
Normality:
βj ∼ Normal[βj , Var (β̂j )]
β̂j − βj
sd(β̂j )
β̂j − βj
= tdf
se(β̂j )n−k−1
Assume t distribution
σ → σ̂
Testing the null hypothesis
Ho : β j = 0
β̂j
tβ̂j ≡
se(β̂j )
Rejection rule
It has the same sign as the estimator
The larger the estimator, the larger the t
Measures how many estimated standard deviations β̂j is away from
zero.
t Test: Alternatives
What is the smallest significance level at which the null hypothesis would
be rejected?
P − value = (|T | > |t|)
Confidence Intervals
Under CLM assumptions, the 95% confidence interval for the population
parameter βj is given by:
H0 : βk−q+1 = 0, ...., βk = 0
H1 : Not true
Can’t just check each t statistic separately
We want to know if the q parameters are jointly significant
Estimate
1 Restricted model
2 Unrestricted model
(SSRr − SSRu )/q
F ≡
SSRu /(n − k − 1)
F statistic
Overall significance
R 2 /k
F ≡
1 − R 2 /(n − k − 1)
H0 : β1 = ... = βk = 0
Close relation to the goodness of fit of the original model
R 2 large, our independent variables do a good job of explaining the
variation in the dependent variable and hence we reject the null
Single linear restriction
F statistic equals the square of the t
p-values will be the same