Lecture 3 Classical Linear Regression Model
Lecture 3 Classical Linear Regression Model
2
Regression
• Note that there can be many x variables but we will limit ourselves to the case
where there is only one x variable to start with. In our set-up, there is only one
y variable.
• For simplicity, say k=1. This is the situation where y depends on only one x
variable.
• Suppose that we have the following data on the excess returns on a fund
manager’s portfolio (“fund XXX”) together with the excess returns on a market
index:
45
40
Excess return on fund XXX
35
30
25
20
15
10
5
0
0 5 10 15 20 25
Excess return on market portfolio
‘Introductory Econometrics for
8
Finance’ © Chris Brooks 2013
Finding a Line of Best Fit
x
‘Introductory Econometrics for
11
Finance’ © Chris Brooks 2013
Ordinary Least Squares
• The most common method used to fit a line to the data is known as OLS
(ordinary least squares).
• What we actually do is take each distance and square it (i.e. take the
area of each of the squares in the diagram) and minimise the total sum
of the squares (hence least squares).
yi
û i
ŷ i
xi x
ût
ŷt
t t
y ˆ
y 2
t
ˆ
u 2
0 x
‘Introductory Econometrics for
16
Finance’ © Chris Brooks 2013
The Population and the Sample
20
‘Introductory Econometrics for Finance’ © Chris Brooks 2013
Properties of the OLS Estimator
yˆ t 20.3 0.5091xt
(14.38) (0.2561)
• We can use the information in the sample to make inferences about the
population.
• We will always have two hypotheses that go together, the null hypothesis
(denoted H0) and the alternative hypothesis (denoted H1).
• The null hypothesis is the statement or the statistical hypothesis that is actually
being tested. The alternative hypothesis represents the remaining outcomes of
interest.
• For example, suppose given the regression results above, we are interested in the
hypothesis that the true value of is in fact 0.5. We would use the notation
H0 : = 0.5
H1 : 0.5
This would be known as a two sided test.
• Sometimes we may have some prior information that, for example, we would
expect > 0.5 rather than < 0.5. In this case, we would do a one-sided test:
H0 : = 0.5
H1 : > 0.5
or we could have had
H0 : = 0.5
H1 : < 0.5
• There are two ways to conduct a hypothesis test: via the test of significance
approach or via the confidence interval approach.
• What if the errors are not normally distributed? Will the parameter estimates still
be normally distributed?
• Yes, if the other assumptions of the CLRM hold, and the sample size is
sufficiently large.
yt xt ut
*
26
‘Introductory Econometrics for Finance’ © Chris Brooks 2013
The Test of Significance Approach (cont’d)
3. We need some tabulated distribution with which to compare the estimated test statistics. Test
statistics derived in this way can be shown to follow a t-distribution with T-2 degrees of freedom.
More degree of freedom means more observations available to estimate the unknown population parameters and
calculate the variable of these estimates. In estimations, n-1 means you have n-1 observations available to estimate
the parameter for example sample standard deviation. In simple Linear Regression T-2, means you have T-2
observations available to estimate two unknown population parameters.
As the number of degrees of freedom increases, we need to be less cautious in our approach since we
can be more sure that our results are robust.
4. We need to choose a “significance level”, often denoted . This is also sometimes called the size of
the test and it determines the region where we will reject or not reject the null hypothesis that we are
testing. It is conventional to use a significance level of 5%.
Intuitive explanation is that we would only expect a result as extreme as this or more extreme 5% of the
time as a consequence of chance alone.
Conventional to use a 5% size of test, but 10% and 1% are also commonly used.
27
‘Introductory Econometrics for Finance’ © Chris Brooks 2013
Determining the Rejection Region for a Test of
Significance
5. Given a significance level, we can determine a rejection region and non
rejection region. For a 2-sided test:
f(x)
28
‘Introductory Econometrics for Finance’ © Chris Brooks 2013
The Rejection Region for a 1-Sided Test
(Upper Tail) B>B*
f(x)
95% non-rejection
region 5% rejection region
f(x)
6. Use the t-tables to obtain a critical value or values with which to compare
the test statistic.
7. Finally perform the test. If the test statistic lies in the rejection region then
reject the null hypothesis (H0), else do not reject H0.
( ˆ t crit SE ( ˆ ), ˆ t crit SE ( ˆ ))
• Under the test of significance approach, we would not reject H 0 that = * if the test statistic lies
within the non-rejection region, i.e. if
*
t crit tcrit
SE ( )
• Rearranging, we would not reject if
t crit SE ( ˆ ) ˆ * t crit SE ( ˆ )
• But this is just the rule under the confidence interval approach.
ˆ t crit SE ( ˆ ) * ˆ t crit SE ( ˆ )
yˆ t 20.3 0.5091xt
, T = 22
(14.38) (0.2561)
• Using both the test of significance and confidence interval approaches, test
the hypothesis that =1 against a two-sided alternative.
• The first step is to obtain the critical value. We want tcrit = t20;5%
f(x)
H0 : = 1
H1 : 1
* ˆ t crit SE ( ˆ )
test stat
SE ( )
05091
. 1 0.5091 2.086 0.2561
1917
.
0.2561 (0.0251,1.0433)
Do not reject H0 since
Since 1 lies within the
test stat lies within confidence interval,
non-rejection region do not reject H0
• Note that we can test these with the confidence interval approach.
For interest (!), test
H0 : = 0
vs. H1 : 0
H0 : = 2
vs. H1 : 2
f(x)
-1.725 +1.725
• t20;10% = 1.725. So now, as the test statistic lies in the rejection region, we
would reject H0.
• If we reject the null hypothesis at the 5% level, we say that the result of the
test is statistically significant.
• The probability of a type I error is just , the significance level or size of test we chose.
To see this, recall what we said significance at the 5% level meant: it is only 5% likely
that a result as or more extreme as this could have occurred purely by chance.
• Note that there is no chance for a free lunch here! What happens if we reduce the size
of the test (e.g. from a 5% test to a 1% test)? We reduce the chances of making a type
I error ... but we also reduce the probability that we will reject the null hypothesis at
all, so we increase the probability of a type II error:
less likely
to falsely reject
Reduce size more strict reject null
of test criterion for hypothesis more likely to
rejection less often incorrectly not
reject
• So there is always a trade-off between type I and type II errors when choosing a
significance level. The only way we can reduce the chances of both is to increase the
sample size.
• Suppose that we have the following parameter estimates, standard errors and
t-ratios for an intercept and slope respectively.
Coefficient 1.10 -4.40
SE 1.35 0.96
t-ratio 0.81 -4.63
• If we reject H0, we say that the result is significant. If the coefficient is not
“significant” (e.g. the intercept coefficient in the last regression above), then it
means that the variable is not helping to explain variations in y. Variables
that are not significant are usually removed from the regression model.
• In practice there are good statistical reasons for always having a constant
yt
even if it is not significant. Look at what happens if no intercept is included:
‘Introductory x
t
Econometrics
for Finance’ 47
© Chris
An Example of the Use of a Simple t-test to Test a
Theory in Finance
• Testing for the presence and significance of abnormal returns (“Jensen’s alpha” -
Jensen, 1968).
• The Data: Annual Returns on the portfolios of 115 mutual funds from 1945-1964.