Classical Normal Linear Regression Model
Classical Normal Linear Regression Model
• Because ki, the betas, and Xi are all fixed, βˆ2 is ultimately a linear function of
ui, which is random by assumption. Therefore, the probability distribution of
βˆ2 (and also of βˆ1) will depend on the assumption made about the
probability distribution of ui .
• OLS does not make any assumption about the probabilistic nature of ui.
This void can be filled if we are willing to assume that the u’s follow some
probability distribution.
THE NORMALITY ASSUMPTION FOR ui
• With ui follow the normal distribution, OLS estimators have the following
properties;.
• 1. They are unbiased.
• 2. They have minimum variance. Combined with 1., this means that they are
minimum-variance unbiased, or efficient estimators.
• 3. They have consistency; that is, as the sample size increases indefinitely,
the estimators converge to their true population values.
• 4. βˆ1 (being a linear function of ui) is normally distributed with
• Mean: E(βˆ1) = β1 (4.3.1)
• var (βˆ1): σ2βˆ1 = (∑ X2i/n ∑ x2i)σ2 = (3.3.3) (4.3.2)
• Or more compactly,
• βˆ1 ∼ N (β1, σ2β ˆ1ˆ1)
• then by the properties of the normal distribution the variable Z, which is
• defined as:
• Z = (βˆ1 − β1 )/ σβˆ1 (4.3.3)
• follows the standard normal distribution, that is, a normal distribution with
zero mean and unit ( = 1) variance, or
• Z ∼ N(0, 1)
• 5. βˆ2 (being a linear function of ui) is normally distributed with
• Mean: E(βˆ2) = β2 (4.3.4)
• var (βˆ2): σ2 βˆ2 =σ2 / ∑ x2i = (3.3.1) (4.3.5)
• Or, more compactly,
• βˆ2 ∼ N(β2, σ2βˆ2)
• Then, as in (4.3.3),
• Z = (βˆ2 − β2 )/σβˆ2 (4.3.6)
• also follows the standard normal distribution.
• Geometrically, the probability distributions of βˆ1 and βˆ2 are shown in
Figure 4.1.
• 6. (n− 2)( ˆσ2/σ 2) is distributed as the χ2 (chi-square) distribution with (n −
2)df.
• 7. (βˆ1, βˆ2) are distributed independently of σˆ2.
• 8. βˆ1 and βˆ2 have minimum variance in the entire class of unbiased
estimators, whether linear or not. This result, due to Rao, is very powerful
because, unlike the Gauss–Markov theorem, it is not restricted to the class
of linear estimators only. Therefore, we can say that the least-squares
estimators are best unbiased estimators (BUE); that is, they have minimum
variance in the entire class of unbiased estimators.
• To sum up: The important point to note is that the normality assumption
enables us to derive the probability, or sampling, distributions of βˆ1 and βˆ2
(both normal) and ˆσ2 (related to the chi square). This simplifies the task of
establishing confidence intervals and testing (statistical) hypotheses.
• In passing, note that, with the assumption that ui ∼ N(0, σ2), Yi , being a
linear function of ui, is itself normally distributed with the mean and variance
given by
• E(Yi) = β1 + β2Xi (4.3.7)
• var (Yi) = σ2 (4.3.8)
• More neatly, we can write
• Yi ∼ N(β1 + β2Xi , σ2) (4.3.9)
END
NEXT