Lecture 4
Lecture 4
Multiple
Regression Analysis:
Inference
Prepared by Quanquan Liu
Fall 2024
Sampling Distributions of the OLS
Estimators
Assumption MLR.6. Normality
The population error u is independent of the explanatory variables and is normally
distributed with zero mean and variance .
u is independent of the ;
;
.
For cross-sectional regression applications, Assumptions MLR.1-6 are called the classical
linear model (CLM) assumptions.
Under the CLM assumptions, the OLS estimators are the minimum variance unbiased
estimators.
Sampling Distributions of the OLS
Estimators
Under the CLM assumptions
.
Theorem 1. Normal Sampling Distributions
Under the CLM Assumptions MLR.1 through MLR.6, conditional on the sample values of
the independent variables,
,
where . Therefore,
.
Going further: any linear combination of the is also normally distributed, and any subset
of the has a joint normal distribution.
Testing Single Population Parameter:
The t Test
Theorem 2. t Distribution for the Standardized Estimators
Under the CLM Assumptions MLR.1 through MLR.6,
,
where is the number of unknown parameters in the population model ( slope
parameters and the intercept ) and is the degrees of freedom (df).
Assume our primary interest lies in testing the null hypothesis
,
where j corresponds to any of the k independent variables.
Null hypothesis: In classical hypothesis testing, we take this hypothesis as true and require the
data to provide substantial evidence against it.
The t Test
The t-statistic will be used to test the above null hypothesis. The farther the estimated coefficient
is away from zero, the less likely it is that the null hypothesis holds true. But what does “far”
away from zero mean?
Distribution of the t-statistic if the null hypothesis is true
Goal: Define a rejection rule so that, if it is true, H0 is rejected only with a small probability
(= significance level, e.g. 5%)
The t Test
Alternative hypothesis: The hypothesis against which the null hypothesis is tested.
Consider a one-sided alternative of the form
.
We are testing against .
Significance level: The probability of a Type I error (rejecting the null hypothesis when it's actually
true) in hypothesis testing.
A 5% significance level means that we are willing to mistakenly reject H 0 when it is true 5% of the time.
We are looking for a “sufficiently large” positive value of in order to reject in favor of .
The definition of “sufficiently large,” with a 5% significance level, is the 95 th percentile in a t
distribution with n − k − 1 degrees of freedom; denote this by c.
The rejection rule is that: is rejected in favor of at the 5% significance level if .
The t Test
Test against .
t statistic
Conclusion: The effect of experience on hourly wage is statistically greater than zero at the
5% (and even at the 1%) significance level.
The t Test
Test against .
t statistic
Conclusion: One cannot reject the hypothesis that there is no effect of school size on
student performance (not even for a lax significance level of 10%).
The t Test
Suppose that we wish to test the null hypothesis that a parameter is zero against a two-
sided alternative, and with 40 df we get a t statistic equal to 1.85. While and :
.
Question: Given the observed value of the t statistic, what is the smallest significance level
at which the null hypothesis would be rejected?
p-value: The smallest significance level at which the null hypothesis can be rejected.
Equivalently, the largest significance level at which the null hypothesis cannot be rejected.
The t Test
A small p-value is evidence against the null hypothesis because one would reject the null
hypothesis even at small significance levels.
A large p-value is evidence in favor of the null hypothesis.
Rejection rule: If denotes the significance level of the test, then is rejected if p-value .
It is simple to obtain the one-sided p-value: just divide the two-sided p-value by 2.
The t Test
The statistical significance of a variable is determined entirely by the size of , whereas the
economic significance or practical significance of a variable is related to the size (and sign)
of .
Check for statistical significance. If the variable is statistically significant, discuss the
magnitude of the coefficient to get an idea of its practical or economic importance.
If a variable is statistically and economically important but has the “wrong” sign, the
regression model might be misspecified.
If a variable is not statistically significant at the usual levels (10%, 5%, or 1%), you might
still ask if the variable has the expected effect on y and whether that effect is practically
large. If it is large, you should compute a p-value for the t statistic. For small sample sizes,
you can sometimes make a case for p-values a little bit larger than usual levels. With large
p-values, the practically large estimates may be due to sampling error.
Confidence Intervals
Confidence interval (CI): A rule used to construct a random interval so that a certain
percentage of all data sets, determined by the confidence level, yields an interval that
contains the population value.
Using the fact that , simple manipulation leads to a CI for the unknown : e.g., a 95%
confidence interval is given by
Consider a simple model comparing the returns to education at junior colleges and four-
year colleges (“universities”):
Hypothesis of interest: whether one year at a junior college is worth one year at a
university:
against
Rewrite the null and alternative as and
Testing Single Linear Restriction
: The standard error of the difference in parameters is difficult to obtain with standard
regression output
Test: once years in the league and games per year have been controlled for, the statistics
measuring performance - bavg, hrunsyr, and rbisyr - have no effect on salary.
against is not true.
Exclusion restrictions: Restrictions which state that certain variables are excluded from the
model (or have zero population coefficients).
The F Test
: the null hypothesis is overwhelmingly rejected (even at very small significance levels).
A more convenient way to conduct F test in STATA is using the command test:
test bavg hrunsyr rbisyr
Discussion:
The three variables are “jointly significant”.
They were not significant when tested individually.
The likely reason is multicollinearity between them.
The F Test
Because has an distribution, the two approaches lead to exactly the same outcome,
provided that the alternative is two-sided.
The F Test
As a general rule, the SSR form of the F statistic should be used if a different dependent
variable is needed in running the restricted regression.
The F Test