Chapter 3
Chapter 3
Yi X i U i
The above identity represents population regression
function (to be estimated from total enumeration of data
from the entire population).
Ordinary Least Square Method (OLS)
But, most of the time it is difficult to generate population
data owing to several reasons; and most of the time we
use sample data to estimate sample regression function.
Thus, we use the following sample regression function
for the derivation of the parameters and related analysis.
Yˆi ˆ ˆX i
Before discussing the details of the OLS estimation techniques,
let‟s see the classical assumptions that are necessary for the
validity of the analysis, interpretations and conclusions of the
regression function.
Classical Assumptions(CA)
For the validity of a regression function our regression
function should fulfill the following conditions known as
classical assumptions.
1. The error terms ‘Ui’ are the disturbance terms are not
correlated.
relation among the value of the error terms (Ui and Uj);
Where i = 1, 2, 3, …….., j = 1, 2, 3, ……. and .
Cov (Ui , Uj) = 0 for population data
Cov (ei, ej) = 0 for sample data
Otherwise, the error terms do not serve an adjustment
purpose rather it causes an autocorrelation problem.
CA---Cont’d
2. The disturbance terms „Ui‟ have zero mean.
This implies the sum of the individual disturbance terms
is zero.
The disturbance terms are negative; some are zero and
some are positive and the sum or the average is zero.
This is given by the following identities.
E(Ui) = U i 0 . for population
The same argument is true for sample regression
function and so for residual terms given as follows:
for sample
Otherwise the estimated models will be biased and
cause the regression function to shift.
For instance: if (or positive) it is going to shift the
estimation upward from the true representative model.
This is demonstrated by the following fig. Error not zero
CA---Cont’d
3. The disturbance terms have constant variance in each
period.
This is given as follows:
Var U i E ((U i E (U i ))
2
2
= u .
2
Cov U i , X i 0 .
Ui ˜
N 0, u
2
.
This assumption or combination of assumptions is used in testing
hypotheses about significance of parameters.
It is also useful in both estimating parameters and testing their
significance in maximum likelihood method.
CA---Cont’d
8. Explanatory variables should not be perfectly, linearly and/or
highly correlated.
Using explanatory variables which are highly or perfectly
correlated in a regression function causes a biased function or
model. It also results in multicollinearity problem.
9. The relationship between variables (or the model) is correctly
specified.
All the necessary variables are included in the model.
For instance, “Y = f (X 2)” may better reflect the relationship
between Y and X than “Y = f (X )”.
10.The explanatory variables do not have identical value.
This assumption is very important for improving the precision of
estimators.
OLS method of Estimation
• From this identity, we solve for the residual term , ' e ' ,
i
square both sides and then take sum of both sides. These
three steps are given (respectively as follows.
ei Yi Yˆi Yi ˆ0 ˆ1 X i
ei Yi ˆ0 ˆ1 X i
2
2
ei
2
2 Yi ˆ0 ˆX i (1) 0 2.9
ˆ 0
Y X
i i ˆ
X
0 i ˆ
1 i 0
X 2
2
X i Yi ˆ0 X i ˆ1 X i
2
2
^ ^
Note that the equation (Yi 0 1 X i ) 2 is a composite function and we should apply a
chain rule in finding the partial derivatives with respect to the parameter estimates.
Equations from above are together called the system of normal equations. Solving the system
of normal equations simultaneously we obtain:
n XY ( X )( Y )
ˆ1
Or
n X 2 ( X ) 2
_ _
XY n Y X and we have ˆ 0 Y ˆ1 X
1 _
from above
X i2 n X 2
Example 2.4: Given the following sample data of three pairs of „Y‟ (dependent variable) and
„X‟ (independent variable), find a simple linear regression function; Y = f(X).
Yi Xi
10 30
20 50
30 60
n XY ( X )( Y )
3(3100) (140)(60)
ˆ1 0.64
3(7000) (140) 2
n X 2 ( X ) 2
n2
Hypothesis Testing of OLS Estimates
(Yi Yˆi ) 2 ei
2
Or R2 1
Unexplained Variation in Y
1 1
Total Variation inY
i
(Y Y ) 2
yi 2
2
Since y i 1 xi y i
1 xi y i
R
2
y i2
The higher R2 is the better the fit(good model)
1. H 0 : 0 vs H1 : 0
2. H 0 : 0 vs H1 : 0
3. H 0 : 0 vs H1 : 0
Calculate a statistic. ( Zcal or tcal)
Case 1: When sampling is from a normal distribution
with known standard deviation
The relevant test statistic is:
X 0
Z cal
n
Decision
H0 Reject H0 if Accept H0 if Inconclusive if
0 Z cal Z 2 Z cal Z 2 Z cal Z 2 or Z cal Z 2
0 Z cal Z Z cal Z Z cal Z
0 Z cal Z Z cal Z Z cal Z
Calculate a statistic. ( Zcal or tcal)
Case 2: When sampling is from a normal distribution with
unknown standard deviation and small sample size
The relevant test statistic is
X 0
t cal ~ t with n 1 deg rees of freedom.
S n
Decision
H0 Reject H0 if Accept H0 if Inconclusive if
0 tcal t 2 tcal t 2 tcal t 2 or tcal t 2
0 tcal t tcal t tcal t
0 tcal t tcal t tcal t
Calculate a statistic. ( Zcal or tcal)
Case 3: When a population whose functional form is
unknown.
The relevant test statistic is
X 0
Z cal , if 2 is known.
n
X 0
, if 2 is unknown.
S n