0% found this document useful (0 votes)
9 views6 pages

Restricted Regression Edit - Removed

Uploaded by

ankushstatsdu
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
9 views6 pages

Restricted Regression Edit - Removed

Uploaded by

ankushstatsdu
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

Chapter 6

Regression Analysis Under Linear Restrictions and Preliminary Test Estimation

One of the basic objectives in any statistical modeling is to find good estimators of the parameters. In the
context of multiple linear y X β + ε , the ordinary least squares estimator
regression model =

b = ( X ' X ) X ' y is the best linear unbiased estimator of β . Several approaches have been attempted in the
−1

literature to improve further the OLSE. One approach to improve the estimators is the use of extraneous
information or prior information. In applied work, such prior information may be available about the
regression coefficients. For example, in economics, the constant returns to scale imply that the exponents
in a Cobb-Douglas production function should sum to unity. In another example, absence of money illusion
on the part of consumers implies that the sum of money income and price elasticities in a demand function
should be zero. These types of constraints or the prior information may be available from
(i) some theoretical considerations.
(ii) past experience of the experimenter.
(iii) empirical investigations.
(iv) some extraneous sources etc.

To utilize such information in improving the estimation of regression coefficients, it can be expressed in the
form of
(i) exact linear restrictions
(ii) stochastic linear restrictions
(iii) inequality restrictions.

We consider the use of prior information in the form of exact and stochastic linear restrictions in the model
y X β + ε where y is a (n × 1) vector of observations on study variable, X is a (n × k ) matrix of
=
observations on explanatory variables X 1 , X 2 ,..., X k , β is a (k ×1) vector of regression coefficients and ε is

a (n ×1) vector of disturbance terms.

Econometrics | Chapter 6 | Linear Restrictions and Preliminary Test Estimation | Shalabh, IIT Kanpur
1
Exact linear restrictions:
Suppose the prior information binding the regression coefficients is available from some extraneous sources
which can be expressed in the form of exact linear restrictions as
r = Rβ
where r is a (q ×1) vector and R is a (q × k ) matrix with rank=
( R) q (q < k ). The elements in
r and R are known.

Some examples of exact linear restriction r = Rβ are as follows:


(i) If there are two restrictions with k = 6 like
β2 = β4
β3 + 2β 4 + β5 =
1
then

0  0 1 0 − 1 0 0 0 
=r =  , R  .
1  0 0 1 2 1 0 0 

(ii) If k = 3 and suppose β 2 = 3, then

=r [3=
] , R [ 0 1 0]
(iii) If k = 3 and suppose β1 : β 2 : β3 :: ab : b :1

0  1 − a 0 
then r =
= 0  , R 0 1
 −b  .
0  1 0 −ab 

The ordinary least squares estimator b = ( X ' X ) −1 X ' y does not uses the prior information. It does not obey
the restrictions in the sense that r ≠ Rb. So the issue is how to use the sample information and prior
information together in finding an improved estimator of β .

Econometrics | Chapter 6 | Linear Restrictions and Preliminary Test Estimation | Shalabh, IIT Kanpur
2
Restricted least squares estimation
The restricted least squares estimation method enables the use of sample information and prior information
simultaneously. In this method, choose β such that the error sum of squares is minimized subject to linear
restrictions r = Rβ . This can be achieved using the Lagrangian multiplier technique. Define the Lagrangian
function
S (β , λ ) =( y − X β ) '( y − X β ) − 2λ '( R β − r )
where λ is a (k ×1) vector of Lagrangian multiplier.

Using the result that if a and b are vectors and A is a suitably defined matrix, then

= ( A + A ')a
a ' Aa
∂a

a ' b = b,
∂a
we have
∂S ( β , λ )
= 2 X ' X β − 2 X ' y − 2 R ' λ =' 0 (*)
∂β
∂S ( β , λ )
= R β − r= 0.
∂λ
Pre-multiplying equation (*) by R( X ' X ) −1 , we have

2 Rβ − 2 R( X ' X ) −1 X ' y − 2 R( X ' X ) −1 R ' λ ' =


0

or Rβ − Rb − R( X ' X ) −1 R ' λ ' =


0
−1
⇒λ'=
−  R( X ' X ) −1 R ' ( Rb − r )

using R ( X ' X ) −1 R ' > 0.


Substituting λ in equation (*), we get
−1
2 X ' X β − 2 X ' y + 2 R '  R( X ' X ) −1 R ' ( Rb − r ) =
0

X ' y − R ' ( R( X ' X ) −1 R ') ( Rb − r ) .


−1
or X 'Xβ =

Pre-multiplying by ( X ' X ) yields


−1

−1
( X ' X ) X ' y + ( X ' X ) R '  R( X ' X )−1 R ' ( r − Rb )
−1 −1
βˆR =
−1
b − ( X ' X ) R '  R ( X ' X ) R ' ( Rb − r ) .
−1 −1
=
 
This estimation is termed as restricted regression estimator of β .
Econometrics | Chapter 6 | Linear Restrictions and Preliminary Test Estimation | Shalabh, IIT Kanpur
3
Properties of restricted regression estimator
1. The restricted regression estimator βˆR obeys the exact restrictions, i.e., r = RβˆR . To verify this,
consider

R b + ( X ' X ) R '{ R( X ' X ) −1} ( r − Rb ) 


−1 −1
RβˆR =
 
= Rb + r − Rb
= r.

2. Unbiasedness
The estimation error of βˆR is

βˆR − β = ( b − β ) + ( X ' X ) −1 R '  R ( X ' X ) R ' ( Rβ − Rb )


−1
 
= I − ( X ' X ) R '{ R( X ' X ) −1 R '} R  ( b − β )
−1 −1

 
= D (b − β )

where
−1
D= I − ( X ' X ) −1 R  R ( X ' X ) R ' R.
−1
 
Thus

( )
E βˆR − β= DE ( b − β )
=0

implying that βˆR is an unbiased estimator of β .

3. Covariance matrix
The covariance matrix of βˆR is

( )
V βˆR =E βˆR − β( )( βˆ R )
−β '
= DE ( b − β )( b − β ) ' D '
= DV (b) D '
= σ 2D ( X ' X ) D '
−1

−1
= σ 2 ( X ' X ) − σ 2 ( X ' X ) R '  R ( X ' X ) R ' R ' ( X ' X )
−1 −1 −1 −1
 
which can be obtained as follows:

Econometrics | Chapter 6 | Linear Restrictions and Preliminary Test Estimation | Shalabh, IIT Kanpur
4
Consider
−1
D (=
X 'X) (X 'X ) − ( X ' X ) R '  R ( X ' X ) R ' R ( X ' X )
−1 −1 −1 −1 −1
 

{ } { }
'
 X ' X −1 − X ' X −1 R ' R X ' X −1 R '
D( X ' X ) D' =
−1
R ( X ' X )   I − ( X ' X ) R ' R ( X ' X ) R '
−1
R 
( ) ( ) ( )
−1 −1 −1 −1

 
−1
( X ' X ) − ( X ' X ) R '  R ( X ' X ) R ' ' R ( X ' X ) − ( X ' X ) R '  R ( X ' X ) R ' R ( X ' X )
−1 −1 −1 −1 −1 −1 −1
=
−1 −1
+ ( X ' X ) R '  R ( X ' X ) R ' R ( X ' X ) R '  R ( X ' X ) R ' R ( X ' X )
−1 −1 −1 −1 −1
   
−1
(X 'X ) − ( X ' X ) R '  R ( X ' X ) R ' R ( X ' X ) .
−1 −1 −1 −1
=
 

Maximum likelihood estimation under exact restrictions:


Assuming ε ~ N (0, σ 2 I ) , the maximum likelihood estimator of β and σ 2 can also be derived such that it
follows r = Rβ . The Lagrangian function as per the maximum likelihood procedure can be written as
n
 1 2  1  ( y − X β ) '( y − X β ) 
L ( β , σ=
2
,λ)  2 
exp  −  − λ ' ( Rβ − r ) 
 2πσ   2 σ 
2

where λ is a ( q ×1) vector of Lagrangian multipliers. The normal equations are obtained by partially

differentiating the log – likelihood function with respect to β , σ 2 and λ and equated to zero as

∂ ln L ( β , σ 2 , λ ) 1
− 2 ( X ' X β − X ' y ) + 2R ' λ =
= 0 (1)
∂β σ
∂ ln L ( β , σ , λ )
2

= 2 ( Rβ − r=
) 0 (2)
∂λ
∂ ln L ( β , σ 2 , λ ) 2n 2 ( y − X β ) ' ( y − X β )
=
− 2+ =
0. (3)
∂σ 2
σ σ4
Let βR , σ R2 and λ denote the maximum likelihood estimators of β , σ 2 and λ respectively which are
obtained by solving equations (1), (2) and (3) as follows:

From equation (1), we get optimal λ as

( r − Rβ )
−1
 R ( X ' X )−1 R '
λ =   .
σ 2

Substituting λ in equation (1) gives

( )
−1
β + ( X ' X ) R '  R ( X ' X ) R '
βR = r − Rβ
−1 −1
 
Econometrics | Chapter 6 | Linear Restrictions and Preliminary Test Estimation | Shalabh, IIT Kanpur
5
where β = ( X ' X ) X ' y is the maximum likelihood estimator of β without restrictions. From equation (3),
−1

we get

σ 2
=
( y − X β ) ' ( y − X β )
.
R
n
The Hessian matrix of second order partial derivatives of β and σ 2 is positive definite at

=β β=
R and σ
2
σ R2 .
The restricted least squares and restricted maximum likelihood estimators of β are same whereas they are

different for σ 2 .

Test of hypothesis
It is important to test the hypothesis
H 0 : r = Rβ
H1 : r ≠ R β
before using it in the estimation procedure.

The construction of the test statistic for this hypothesis is detailed in the module on multiple linear regression
model. The resulting test statistic is
 (r − Rb) '  R ( X ' X ) −1 R ' −1 (r − Rb) 
   
 q 
F=  
 ( y − Xb) '( y − Xb 
 
 n−k 
which follows a F -distribution with q and (n − k ) degrees of freedom under H 0 . The decision rule is to

reject H 0 at α level of significance whenever

F ≥ F1−α (q, n − k ).

Econometrics | Chapter 6 | Linear Restrictions and Preliminary Test Estimation | Shalabh, IIT Kanpur
6

You might also like