ECO - Chapter 2 SLRM
ECO - Chapter 2 SLRM
Econ-2061
Sied Hassen (PhD)
Chapter Two
Economic theories are mainly concerned with the relationships among various
economic variables. These relationships, when phrased in mathematical terms, can
predict the effect of one variable on another. The functional relationships of these
variables define the dependence of one variable upon the other variable (s) in the
specific form. The specific functional forms may be linear, quadratic, logarithmic,
exponential, hyperbolic, or any other form.
13
specific value only with some probability. Let’s illustrate the distinction between
stochastic and non stochastic relationships with the help of a supply function.
Assuming that the supply for a certain commodity depends on its price (other
determinants taken to be constant) and the function being linear, the relationship
can be put as:
Q f ( P) P (2.1)
The above relationship between P and Q is such that for a particular value of P,
there is only one corresponding value of Q. This is, therefore, a deterministic
(non-stochastic) relationship since for each price there is always only one
corresponding quantity supplied. This implies that all the variation in Y is due
solely to changes in X, and that there are no other factors affecting the dependent
variable.
If this were true all the points of price-quantity pairs, if plotted on a two-
dimensional plane, would fall on a straight line. However, if we gather
observations on the quantity actually supplied in the market at various prices and
we plot them on a diagram we see that they do not fall on a straight line.
The deviation of the observation from the line may be attributed to several factors.
a. Omission of variables from the function
b. Random behavior of human beings
c. Imperfect specification of the mathematical form of the model
14
d. Error of aggregation
e. Error of measurement
In order to take into account the above sources of errors we introduce in
econometric functions a random variable which is usually denoted by the letter ‘u’
or ‘ ’ and is called error term or random disturbance or stochastic term of the
function, so called because u is supposed to ‘disturb’ the exact linear relationship
which is assumed to exist between X and Y. By introducing this random variable
in the function the model is rendered stochastic of the form:
Yi X u i ……………………………………………………….(2.2)
Thus a stochastic model is a model in which the dependent variable is not only
determined by the explanatory variable(s) included in the model but also by others
which are not included in the model.
2.2. Simple Linear Regression model.
The above stochastic relationship (2.2) with one explanatory variable is called
simple linear regression model.
The true relationship which connects the variables involved is split into two parts:
a part represented by a line and a part represented by the random term ‘u’.
15
The scatter of observations represents the true relationship between Y and X. The
line represents the exact part of the relationship and the deviation of the
observation from the line represents the random component of the relationship.
- Were it not for the errors in the model, we would observe all the points on the
line Y1' , Y2' ,......, Yn' corresponding to X 1 , X 2 ,...., X n . However because of the random
- The first component in the bracket is the part of Y explained by the changes
in X and the second is the part of Y not explained by X, that is to say the
change in Y is due to the random influence of u i .
The classicals made important assumption in their analysis of regression .The most
importat of these assumptions are discussed below.
b. Yi X i U i
16
2. U i is a random real variable
This means that the value which u may assume in any one period depends on
chance; it may be positive, negative or zero. Every value has a certain probability
of being assumed by u in any particular instance.
Mathematically, E (U i ) 0 ………………………………..….(2.3)
For all values of X, the u’s will show the same dispersion around their mean.
In Fig.2.c this assumption is denoted by the fact that the values that u can
assume lie with in the same limits, irrespective of the value of X. For X 1 , u
can assume any value with in the range AB; for X 2 , u can assume any value
with in the range CD which is equal to AB and so on.
Graphically;
17
Mathematically;
18
Algebraically,
Cov(u i u j ) [(u i (u i )][u j (u j )]
E (u i u j ) 0 …………………………..….(2.5)
( X iU i ) ( X i )(U i )
( X iU i )
0
8. The explanatory variables are measured without error
- U absorbs the influence of omitted variables and possibly errors of
measurement in the y’s. i.e., we will assume that the regressors are error
free, while y values may or may not include errors of measurement.
19
We can now use the above assumptions to derive the following basic concepts.
Proof:
Mean: (Y ) xi u i
X i Since (u i ) 0
X i u i ( X i )
2
(u i ) 2
2 (since (u i ) 2 2 )
var(Yi ) 2 ……………………………………….(2.8)
the distribution of y i .
Yi ~ N( x i , 2 )
Proof:
Cov(Yi , Y j ) E{[Yi E (Yi )][Y j E (Y j )]}
E{[ X i U i E ( X i U i )][ X j U j E ( X j U j )}
(Since Yi X i U i and Y j X j U j )
20
= E[( X i Ui X i )( X j U j X j )] ,Since (u i ) 0
Therefore, Cov(Yi ,Y j ) 0 .
ˆ and ˆ are estimated from the sample of Y and X and ei represents the sample
21
Estimation of and by least square method (OLS) or classical least square
(CLS) involves finding values for the estimates ˆ and ˆ which will minimize the
sum of square of the squared residuals ( ei2 ).
e 2
i (Yi ˆ ˆX i ) 2 ……………………….(2.7)
To find the values of ˆ and ˆ that minimize this sum, we have to partially
differentiate e 2
i with respect to ˆ and ˆ and set the partial derivatives equal to
zero.
ei2
1. 2 (Yi ˆ ˆX i ) 0.......................................................(2.8)
ˆ
ei2
2. 2 X i (Yi ˆ ˆX ) 0..................................................(2.11)
ˆ
Note: at this point that the term in the parenthesis in equation 2.8and 2.11 is the
residual, e Yi ˆ ˆX i . Hence it is possible to rewrite (2.8) and (2.11) as
e i 0 and X e i i 0............................................(2.12)
Equation (2.9) and (2.13) are called the Normal Equations. Substituting the
values of ̂ from (2.10) to (2.13), we get:
22
Y Xi i X i (Y ˆX ) ˆX i2
Y X i ˆXX i ˆX i2
Y Xi i Y X i ˆ (X i2 XX i )
XY nXY = ̂ ( X i2 nX 2)
XY n X Y
ˆ ………………….(2.14)
X i2 n X 2
( X X ) 2 X 2 nX 2 (2.16)
xi yi
ˆ ……………………………………… (2.17)
x i2
23
n
We minimize: ei2 (Yi ˆ ˆX i ) 2
i 1
Subject to: ˆ 0
The composite function then becomes
Z (Yi ˆ ˆX i ) 2 ˆ , where is a Lagrange multiplier.
z
2 0 (iii )
Substituting (iii) in (ii) and rearranging we obtain:
X i (Yi ˆX i ) 0
Yi X i ˆX i 0
2
X i Yi
ˆ ……………………………………..(2.18)
X i2
This formula involves the actual values (observations) of the variables and not
their deviation forms, as in the case of unrestricted value of ̂ .
24
‘Closeness’ of the estimate to the population parameter is measured by the mean
and variance or standard deviation of the sampling distribution of the estimates of
the different econometric methods. We assume the usual process of repeated
sampling i.e. we assume that we get a very large number of samples each of size
‘n’; we compute the estimates ̂ ’s from each sample, and for each econometric
method and we form their distribution. We next compare the mean (expected
value) and the variances of these distributions and we choose among the
alternative estimates the one whose distribution is concentrated as close as
possible around the population parameter.
According to the this theorem, under the basic assumptions of the classical linear
regression model, the least squares estimators are linear, unbiased and have
minimum variance (i.e. are best of all linear unbiased estimators). Some times the
theorem referred as the BLUE theorem i.e. Best, Linear, Unbiased Estimator. An
estimator is called BLUE if:
a. Linear: a linear function of the a random variable, such as, the
dependent variable Y.
b. Unbiased: its average or expected value is equal to the true population
parameter.
c. Minimum variance: It has a minimum variance in the class of linear
and unbiased estimators. An unbiased estimator with the least variance
is known as an efficient estimator.
25
According to the Gauss-Markov theorem, the OLS estimators possess all the
BLUE properties. The detailed proof of these properties are presented below
a. Linearity: (for ̂ )
(but xi ( X X ) X nX nX nX 0 )
x Y xi
ˆ i 2 ; Now, let Ki (i 1,2,.....n)
x i xi2
ˆ K i Y (2.19)
̂ is linear in Y
b. Unbiasedness:
Proposition: ˆ & ˆ are the unbiased estimators of the true parameters &
From your statistics course, you may recall that if ˆ is an estimator of then
E (ˆ) the amount of bias and if ˆ is the unbiased estimator of then bias =0
In our case, ˆ & ˆ are estimators of the true parameters & .To show that they
are the unbiased estimators of their respective parameters means to prove that:
( ˆ ) and (ˆ )
26
Proof (1): Prove that ̂ is unbiased i.e. ( ˆ ) .
k i k i X i k i u i ,
but k i 0 and k i X i 1
xi (X X ) X nX nX nX
ki 0
xi2
xi2
x i2 xi2
k i 0 …………………………………………………………………(2.20)
xi X i ( X X ) Xi
k i X i
xi2 xi2
X 2 XX X 2 nX 2
1
X 2 nX 2 X 2 nX 2
k i X i 1............................. ……………………………………………(2.21)
ˆ k i u i ˆ k i u i (2.22)
( ˆ ) , since (u i ) 0
1
n X i 1 n u i Xk i Xk i X i Xk i u i
1 n u i X k i u i , ˆ 1
n u i X k i u i
1 n Xk i )u i ……………………(2.23)
(ˆ ) (2.24)
27
̂ is an unbiased estimator of .
first obtain variance of ˆ and ˆ and then establish that each has the minimum
variance in comparison of the variances of other linear and unbiased estimators
obtained by any other econometric methods than OLS.
a. Variance of ̂
var( ) ( ˆ ( ˆ )) 2 ( ˆ ) 2 ……………………………………(2.25)
( k i2 u i2 ) (k i k j u i u j ) i j
x i xi2 1
k i , and therefore, k 2
2
x i ( x i ) x i
2 i 2 2
2
var(ˆ ) 2 k i2 ……………………………………………..(2.26)
xi2
b. Variance of ̂
ˆ (2.27)
2
28
1 n Xk i (u i ) 2
2
2 ( 1 n Xk i ) 2
2 ( 1 n 2 2 n Xk i X 2 k i2 )
2 ( 1 n 2 X n k i X 2 k i2 ) , Since k i 0
2 ( 1 n X 2 k i2 )
1 X2 xi2 1
2( ) , Since k 2
2
n xi ( x i ) x i
2 i 2 2
Again:
1 X 2 xi2 nX 2 X 2
n xi2 nxi2 nxi
2
X2 X i2
var(ˆ ) 2 1 n 2 2
n x 2
…………………………………………(2.28)
x i i
We have computed the variances OLS estimators. Now, it is time to check
whether these variances of OLS estimators do possess minimum variance property
compared to the variances other estimators of the true and , other than
ˆ and ˆ .
1. Minimum variance of ̂
Suppose: * an alternative linear and unbiased estimator of and;
Let * w i Y i ......................................... ………………………………(2.29)
where , wi k i ; but: wi k i ci
29
* wi ( X i u i ) Since Yi X i U i
wi var(Yi )
2
30
var( *) var( ˆ ) 2 c i2
Given that ci is an arbitrary constant, 2 ci2 is a positive i.e it is greater than zero.
Thus var( *) var(ˆ ) . This proves that ̂ possesses minimum variance property.
In the similar way we can prove that the least square estimate of the constant
intercept ( ̂ ) possesses minimum variance.
2. Minimum Variance of ̂
We take a new estimator * , which we assume to be a linear and unbiased
estimator of function of . The least square estimator ̂ is given by:
ˆ ( 1 n Xk i )Yi
By analogy with that the proof of the minimum variance property of ̂ , let’s use
the weights wi = ci + ki Consequently;
* ( 1 n Xwi )Yi
* ( 1 n Xwi )( X i u i )
X ui
( Xwi XX i wi Xwi u i )
n n n
* X ui / n Xwi Xwi X i Xwi u i
( wi ) 0, ( wi X i ) 1 and ( wi u i ) 0
i.e., if wi 0, and wi X i 1 . These conditions imply that ci 0 and ci X i 0 .
( 1 n Xwi ) 2 var(Yi )
2 ( 1 n Xwi ) 2
31
2 ( 1 n 2 X 2 wi 2 1 n Xwi )
2
2 ( n n 2 X 2 wi 2 X wi )
2 1
n
var( *) 2
1
n X 2 wi
2
,Since wi 0
1 X2
var( *) 2 2 2 X 2 ci2
n x i
X i2
2 2 X 2 ci2
nxi
2
The first term in the bracket it var(ˆ ) , hence
Therefore, we have proved that the least square estimators of linear regression
model are best, linear and unbiased (BLU) estimators.
32
To use ˆ 2 in the expressions for the variances of ˆ and ˆ , we have to prove
e
2
n2
e
2
To prove this we have to compute i from the expressions of Y,
Yˆ , y, yˆ and ei .
Proof:
Yi ˆ ˆX i ei
Yˆ ˆ ˆx
Y Yˆ ei ……………………………………………………………(2.31)
ei Yi Yˆ ……………………………………………………………(2.32)
Y Yˆ
(Y Y ) (Yˆ Yˆ ) e
y i yˆ i e ………………………………………………(2.34)
From (2.34):
ei y i yˆ i ………………………………………………..(2.35)
33
Y X U
We get, by subtraction
y i (Yi Y ) i ( X i X ) (U i U ) xi (U U )
y i x (U U ) …………………………………………………….(2.36)
Note that we assumed earlier that , (u ) 0 , i.e in taking a very large number
samples we expect U to have a mean value of zero, but in any particular single
sample U is not necessarily zero.
Similarly: From;
Yˆ ˆ ˆx
Y ˆ ˆx
We get, by subtraction
Yˆ Yˆ ˆ ( X X )
yˆ ̂x …………………………………………………………….(2.37)
(u i u ) ( ˆi ) xi
The summation over the n sample values of the squares of the residuals over the
‘n’ samples yields:
ei2 [(u i u ) ( ˆ ) xi ] 2
[(u i u ) 2 ( ˆ ) 2 xi 2(u i u )( ˆ ) xi ]
2
34
( u i ) 2
u i2
n
1
(u i2 ) (u ) 2
n
n 2 1n (u1 u 2 ....... u i ) 2 since (u i2 ) u2
n 2 1n (u i2 2u i u j )
n 2 1n ((u i2 ) 2u i u j ) i j
n 2 1n n u2 n2 (u i u j )
n u2 u2 ( given (u i u j ) 0)
u2 (n 1) ……………………………………………..(2.39)
x i2 . ( ˆ ) 2 u2 ……………………………………………(2.40)
x i u i
= -2 ( x i u i ) xi
2 ,since k i
x i x
2
i
(x u ) 2
2 i 2i
xi
35
xi 2 u i 2 2xi x j u i u j
2
xi
2
x 2 (u i 2 ) 2( xi x j )(u i u j )
2 i j
xi x i
2 2
x 2 (u i )
2
2 ( given (u i u j ) 0)
x i
2
Consequently, Equation (2.38) can be written interms of (2.39), (2.40) and (2.41)
as follows: ei2 n 1 u2 2 2 u2 (n 2) u2 ………………………….(2.42)
From which we get
e 2
i E (ˆ u2 ) u2 ………………………………………………..(2.43)
n2
ei2
Since ˆ u2
n2
ei2
Thus, ˆ 2 is unbiased estimate of the true variance of the error term( 2 ).
n2
The conclusion that we can drive from the above proof is that we can substitute
ei2
ˆ 2 for ( 2 ) in the variance expression of ˆ and ˆ , since E (ˆ 2 ) 2 .
n2
X i2 ei X i
2 2
Var (ˆ ) ˆ
2
……………………………(2.45)
nxi n(n 2) xi
2 2
e can be computed as ei 2 y i ̂ xi y i .
2 2
Note: i
36
After the estimation of the parameters and the determination of the least square
regression line, we need to know how ‘good’ is the fit of this line to the sample
observation of Y and X, that is to say we need to measure the dispersion of
observations around the regression line. This knowledge is essential because the
closer the observation to the line, the better the goodness of fit, i.e. the better is the
explanation of the variations of Y by the changes in the explanatory variables.
We divide the available criteria into three groups: the theoretical a priori criteria,
the statistical criteria, and the econometric criteria. Under this section, our focus
is on statistical criteria (first order tests). The two most commonly used first order
tests in econometric analysis are:
i. The coefficient of determination (the square of the
correlation coefficient i.e. R2). This test is used for judging
the explanatory power of the independent variable(s).
ii. The standard error tests of the estimators. This test is used
for judging the statistical reliability of the estimates of the
regression coefficients.
37
.Y
Y = e Y Yˆ
Y Y = Ŷ Yˆ ˆ 0 ˆ1 X
= Yˆ Y
Y.
X
Figure ‘d’. Actual and estimated values of the dependent variable Y.
As can be seen from fig.(d) above, Y Y represents measures the variation of the
sample observation value of the dependent variable around the mean. However
the variation in Y that can be attributed the influence of X, (i.e. the regression line)
is given by the vertical distance Yˆ Y . The part of the total variation in Y about
Y that can’t be attributed to X is equal to e Y Yˆ which is referred to as the
residual variation.
In summary:
ei Yi Yˆ = deviation of the observation Yi from the regression line.
Now, we may write the observed Y as the sum of the predicted value ( Yˆ ) and the
residual term (ei.).
Yi Yˆ ei
predicted Yi
Observed Yi Re sidual
From equation (2.34) we can have the above equation but in deviation form
y yˆ e . By squaring and summing both sides, we obtain the following
expression:
y 2 ( yˆ 2 e) 2
38
y 2 ( yˆ 2 ei2 2 yei)
y i ei2 2yˆ ei
2
(but e i 0 , ex i 0 )
yˆ e 0 ………………………………………………(2.46)
Therefore;
y i2
yˆ 2 ei2 ………………………………...(2.47)
Total Explained Un exp lained
var iation var iation var ation
OR,
Total sum of Explained sum Re sidual sum
square of square of square
TSS ESS RSS
i.e
TSS ESS RSS ……………………………………….(2.48)
Mathematically; the explained variation as a percentage of the total variation is
explained as:
ESS yˆ 2
……………………………………….(2.49)
TSS y 2
From equation (2.37) we have yˆ ̂x . Squaring and summing both sides give us
yˆ 2 ˆ 2 x 2 (2.50)
xy xi x y
2 2
2 , Since ˆ i 2 i
x y x i
2
39
xy xy
………………………………………(2.52)
x 2 y 2
The limit of R2: The value of R2 falls between zero and one. i.e. 0 R 2 1 .
Interpretation of R2
Suppose R 2 0.9 , this means that the regression line gives a good fit to the
observed data since this line explains 90% of the total variation of the Y value
around their mean. The remaining 10% of the total variation in Y is unaccounted
for by the regression line and is attributed to the factors included in the disturbance
variable u i .
40
Check yourself question:
a. Show that 0 R 2 1 .
b. Show that the square of the coefficient of correlation is equal to ESS/TSS.
Exercise:
Suppose rxy is the correlation coefficient between Y and X and is give by:
xi yi
x i2 y i2
And let ry2 yˆ the square of the correlation coefficient between Y and Ŷ , and is
(yyˆ ) 2
given by: ry2 yˆ
y 2 yˆ 2
41
estimators; because the testing methods or procedures are based on the assumption
of the normality assumption of the disturbance term. Hence before we discuss on
the various testing methods it is important to see whether the parameters are
normally distributed or not.
We have already assumed that the error term is normally distributed with mean
zero and variance 2 , i.e. U i ~ N(0, 2 ) . Similarly, we also proved
2
1. ˆ ~ N , 2
x
2 X 2
2. ˆ ~ N ,
nx 2
To show whether ˆ and ˆ are normally distributed or not, we need to make use of
one property of normal distribution. “........ any linear function of a normally
distributed variable is itself normally distributed.”
ˆ k i Yi k1Y1 k 2 Y2i .... k n Yn
2 2 X 2
~ N , 2
ˆ ; ~ N ,
ˆ
x nx 2
The OLS estimates ˆ and ˆ are obtained from a sample of observations on Y and
X. Since sampling errors are inevitable in all estimates, it is necessary to apply
test of significance in order to measure the size of the error and determine the
degree of confidence in order to measure the validity of these estimates. This can
be done by using various tests. The most common ones are:
i) Standard error test ii) Student’s t-test iii) Confidence interval
42
All of these testing procedures reach on the same conclusion. Let us now see these
testing methods one by one.
i) Standard error test
This test helps us decide whether the estimates ˆ and ˆ are significantly different
from zero, i.e. whether the sample from which they have been estimated might
have come from a population whose true parameters are zero.
0 and / or 0 .
Formally we test the null hypothesis
H 0 : i 0 against the alternative hypothesis H 1 : i 0
SE ( ˆ ) var(ˆ )
SE (ˆ ) var(ˆ )
Second: compare the standard errors with the numerical values of ˆ and ˆ .
Decision rule:
If SE ( ˆi ) 1
2 ˆi , accept the null hypothesis and reject the alternative
If SE ( ˆi ) 1
2 ˆi , reject the null hypothesis and accept the alternative
43
Numerical example: Suppose that from a sample of size n=30, we estimate the
following supply function.
Q 120 0.6 p ei
SE : (1.7) (0.025)
Test the significance of the slope parameter at 5% level of significance using the
standard error test.
SE ( ˆ ) 0.025
( ˆ ) 0.6
1
2 ˆ 0.3
5% level of significance.
Note: The standard error test is an approximated test (which is approximated from
the z-test and t-test) and implies a two tail test conducted at 5% level of
significance.
ii) Student’s t-test
Like the standard error test, this test is also important to test the significance of the
parameters. From your statistics, any variable X can be transformed into t using
the general formula:
X
t , with n-1 degree of freedom.
sx
( X X ) 2
sx
n 1
n sample size
We can derive the t-value of the OLS estimates
44
ˆi
t ˆ
SE ( ˆ )
with n-k degree of freedom.
ˆ
tˆ
SE (ˆ )
Where:
SE = is standard error
k = number of parameters in the model.
Since we have two parameters in simple linear regression with intercept different
from zero, our degree of freedom is n-2. Like the standard error test we formally
test the hypothesis: H 0 : i 0 against the alternative H 1 : i 0 for the slope
parameter; and H0 : 0 against the alternative H 1 : 0 for the intercept.
ˆ 0 ˆ
t*
SE ( ˆ ) SE ( ˆ )
45
there is no need to divide the chosen level of significance by two to obtain the
critical value of to from the t-table.
Example:
If we have H 0 : i 0
against: H1 : i 0
Then this is a two tail test. If the level of significance is 5%, divide it by two to
obtain critical value of t from the t-table.
Step 4: Obtain critical value of t, called tc at and n-2 degree of freedom for two
2
tail test.
Step 5: Compare t* (the computed value of t) and tc (critical value of t)
If t*> tc , reject H0 and accept H1. The conclusion is ̂ is statistically
significant.
If t*< tc , accept H0 and reject H1. The conclusion is ̂ is statistically
insignificant.
Numerical Example:
Suppose that from a sample size n=20 we estimate the following consumption
function:
C 100 0.70 e
(75.5) (0.21)
The values in the brackets are standard errors. We want to test the null hypothesis:
H 0 : i 0 against the alternative H 1 : i 0 using the t-test at 5% level of
significance.
a. the t-value for the test statistic is:
ˆ 0 ˆ 0.70
t* = 3.3
SE ( ˆ ) SE ( ˆ ) 0.21
46
b. Since the alternative hypothesis (H1) is stated by inequality sign ( ) ,it is a
two tail test, hence we divide
2 0.05 2 0.025 to obtain the critical value of
‘t’ at =0.025 and 18 degree of freedom (df) i.e. (n-2=20-2). From the
2
In order to define how close the estimate to the true parameter, we must construct
confidence interval for the true parameter, in other words we must establish
limiting values around the estimate with in which the true parameter is expected to
lie within a certain “degree of confidence”. In this respect we say that with a
given probability the population parameter will be with in the defined confidence
interval (confidence limits).
47
ˆ
any value of t which is equal to at n-2 degree of freedom is
SE ( ˆ )
1 2 2 i.e. 1 .
ˆ
but t* …………………………………………………….(2.58)
SE ( ˆ )
Pr SE ( ˆ )t c ˆ SE ( ˆ )t c 1 by multiplying SE ( ˆ )
Pr ˆ SE ( ˆ )t ˆ SE ( ˆ )t 1 by subtracting ˆ
c c
Pr ˆ SE ( ˆ ) ˆ SE ( ˆ )t 1 by multiplyin g by 1
c
The limit within which the true lies at (1 )% degree of confidence is:
H1 : 0
Decision rule: If the hypothesized value of in the null hypothesis is within the
hypothesis is outside the limit, reject H0 and accept H1. This indicates ̂ is
statistically significant.
Numerical Example:
48
Suppose we have estimated the following regression line from a sample of 20
observations.
Y 128.5 2.88 X e
(38.2) (0.85)
ˆ SE ( ˆ )t c
ˆ 2.88
SE ( ˆ ) 0.85
The results of the regression analysis derived are reported in conventional formats.
It is not sufficient merely to report the estimates of ’s. In practice we report
regression coefficients together with their standard errors and the value of R2. It
has become customary to present the estimated equations with standard errors
placed in parenthesis below the estimated parameter values. Sometimes, the
49
estimated coefficients, the corresponding standard errors, the p-values, and some
other indicators are presented in tabular form.
These results are supplemented by R2 on ( to the right side of the regression
equation).
Y 128 . 5 2 . 88 X
Example: , R2 = 0.93. The numbers in the
( 38 . 2 ) ( 0 . 85 )
parenthesis below the parameter estimates are the standard errors. Some
econometricians report the t-values of the estimated coefficients in place of the
standard errors.
Review Questions
Review Questions
1. Econometrics deals with the measurement of economic relationships which are stochastic
or random. The simplest form of economic relationships between two variables X and Y
can be represented by:
Yi 0 1 X i U i ; where 0 and 1 are regression parameters and
50
a. Estimate the linear regression line ( S ) P
X Y 1,296,836
i i
Y 539,512
i
2
i) Estimate the regression line of sale on price and interpret the results
ii) What is the part of the variation in sales which is not explained by the
regression line?
iii) Estimate the price elasticity of sales.
5. The following table includes the GNP(X) and the demand for food (Y) for a country over
ten years period.
year 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989
Y 6 7 8 10 8 9 10 9 11 10
X 50 52 55 59 57 58 62 65 68 70
a. Estimate the food function
b. Compute the coefficient of determination and find the explained and unexplained
variation in the food expenditure.
c. Compute the standard error of the regression coefficients and conduct test of
significance at the 5% level of significance.
6. A sample of 20 observation corresponding to the regression model Yi X i U i
Y 21.9 Y Y 86.9
2
i i
X 186.2 X X 215.4
2
i i
X X Y i i Y 106.4
a. Estimate and
51
b. Calculate the variance of the estimates
c. estimate the conditional mean of Y corresponding to a value of X fixed at X=10.
7. Suppose that a researcher estimates a consumptions function and obtains the following
results:
C 15 0.81Yd n 19
(3.1) (18.7) R 2 0.99
where C=Consumption, Yd=disposable income, and numbers in the parenthesis are the ‘t-ratios’
a. Test the significant of Yd statistically using t-ratios
b. calculate the standard errors of the parameter estimates
8. State and prove Guass-Markov theorem
9. Given the model:
Yi 0 1 X i U i with usual OLS assumptions. Derive the expression for the error
variance.
52