Ch4 2945310
Ch4 2945310
The coefficients we calculated from the OLS are estimators and used sampling data not the population
data. It is not the true value of the coefficient according to the PRF. If we change the sample, we will
get different estimates.
Thus, the estimator, β̂ , is a random variable. As a random variable, it must have a specific
distribution which can be explain by its moments. We will focus on the first and second moments i.e.
Expected Value and its Variance.
Expected Value of β̂
E [ β̂ ∣ X ]=E [( X ' X ) X ' Y ]
−1
E [ β̂ ∣ X ]= β
Variance of β̂
Var [β̂ ∣ X ]=E [( β−β)(
̂ ̂
β−β)' ∣ X ] since E [ β̂ ∣ X ]=β
Var [β̂ ∣ X ]=E [( X ' X ) X ' ϵϵ' X ( X ' X )−1∣ X ] since β=β+(
̂
−1
X ' X )−1 X ' ϵ
Var [β̂ ∣ X ]=( X ' X )−1 X ' E [ϵϵ' ∣ X ] X ( X ' X )−1
Var [β̂ ∣ X ]=( X ' X )−1 X ' (σ2 I ) X ( X ' X )−1 ; E [ϵϵ'∣ X ]=(σ 2 I ) by assumptions
Var [β̂ ∣ X ]=σ 2 ( X ' X )−1 X ' X ( X ' X )−1
Var [β̂ ∣ X ]=σ 2 ( X ' X )−1
“Given the assumptions of the classical linear regression model, the least squares estimators, in the
class of unbiased linear estimators, have minimum variance, that is, they are the Best Linear Unbiased
Estimators, BLUE.”, Basic Econometrics, 5th Edition, Gujarati and Porter
Proof:
̃
Let β=CY
−1
, any estimator linear in Y and C =( X ' X ) X ' + D then
̃
E [β]=E [( X ' X )−1 X ' + D][ X β+ϵ]
̃
E [β]=β+ DX β
̃
E [β]=β or unbiased if DX =0
Covariance Matrix of β̃ is
̃
E ( β−β)( ̃
β−β)'=
−1 −1
E [( X ' X ) X ' + D]ϵϵ' [( X ' X ) X ' + D]'
E ( β̃ −β)( β̃ −β)'=σ2 [( X ' X )−1 X ' IX ( X ' X )−1 + DIX ( X ' X )−1 +( X ' X )−1 X ' ID ' + DID ' ]
E ( β̃ −β)( β̃ −β)'=σ2 ( X ' X )−1+ σ2 DD ' since DX = 0
But DD' is a positive semidefinite matrix i.e. at least = 0. Thus Covariance of β̃ cannot be less than
the Covariance of β̂