Topic 1 Wble
Topic 1 Wble
Topic 1 Wble
A. Multicollinearity
Existence of linear relationship among some or
all explanatory variable:
(1) Perfect Multicolinearity
(2) Imperfect Multicolinearity
Multicollinearity: consequences
1. Estimates will be still unbiased.
2. Standard errors of the estimates increases
3. t-statistics are small
4. Goodness of fit is not diminished.
5. Estimates vary widely if the specification model is
changed.
6. Slope estimates for independent variables without
multicollinerity are not seriously affected.
Detection of Multicollinearity
(1) High R2 but few significant t-ratio
(2) High pair-wide correlation among xs
(3) Variance inflation Factor, VIF = (1/1-R2aux)
(4) Tolerance (TOL) factors = (1/VIF)
Treating multicollinerity
1.
2.
3.
4.
Example: consumption-income
Assuming consumption expenditure is linearly related to income
and wealth, we obtain the following regression:
1.
Regression estimates
2.
Correlation matrix
10
B. Heteroscedasticity
Violation of the assumption of classical linear
regression model
2
E (ei ) = i
11
Heteroscedasticity: nature
Causes:
1. Error learning model: Learning behaviors across
time
2. Model misspecification - omitted variable or
improper functional form.
3. Outlier.
4. Skewness in the regressors.
5. Changes in data collection or definitions.
12
Heteroscedasticity: consequences
's
unbiased and consistent
Detecting heteroscedasticity
Nature of problems
informal
Graphical
Detection
of
Heteroscedasticity
Park test
formal
Glejser test
Breusch-Pagan test
White test
14
Heteroskedasticity: remedies
15
17
Regression model
18
Graphical analysis
10,000
100
8,000
R
E
S
ID
_S
Q
75
50
25
6,000
4,000
2,000
4,000
8,000
12,000
16,000
20,000
PGNP
-25
10,000
-50
R
E
S
ID
_S
Q
8,000
-75
-100
5
10
15
20
25
30
35
40
CM Residuals
45
50
55
60
6,000
4,000
2,000
0
0
20
40
60
80
100
FLR
19
Park Test
Glejser Test
Breush Pagan
White Heteroscedaticity
test
20
21
22
23
24
C. Autocorrelation
The Classical Linear Regression Model (CLRM)
assumes that the disturbance term is not
autocorrelated, i.e. E (ut us) = 0 for t s.
The term autocorrelation can be defined as
correlation between member of observations
order in time (as in time series data) or space (as
in cross sectional data).
25
Pure autocorrlation
Pure autocorrelation occurs when classical assumption,
which assumes uncorrelated observations of the error
term, is violated in a correctly specified equation.
Exist in the form of first order autocorrelation and
higher order of autocorrelation
26
t = 1,,T
where t s
ut = ut-1 + et
where
-1 < < 1
(
: RHO)
and
(white noise)
27
Positive Autocorrelation
28
Negative Autocorrelation
ut
+
u t
u t 1
Time
No pattern in residuals
No autocorrelation
ut
+
u t
u t 1
30
High order
autocorrelation
Impure autocorrelation
By impure autocorrelation we mean serial correlation
that is caused by a specification error such as omitted
variable or an incorrect functional form.
Pure autocorrelation is caused by underlying
distribution of the error term of the true specifications
of an equation (which cannot be changed by the
researcher), impure serial correlation is cased by a
specification error that often can be corrected.
32
34
Durbin-Watson d Test
Steps involved in the Durbin-Watson Test
1. Run the OLS and obtain the residuals t.
2. Compute the d statistics.
n
(
d =
t 1 ) 2
t=2
n
t2
or
n
t t 1
t=2
and d 2(1 - )
t2
1
35
Durbin-Watson d Test
3.
36
Durbin-Watson d Test
d lies between 0 and 4
d = 2 implies residuals uncorrelated
37
Durbin-Watson d Test
Assumption underlying the d statistics:
1.
2.
3.
4.
5.
38
2.
3.
4.
5.
The Remedies
Only after the specification of the question has
been reviewed carefully should the possibility of
an adjustment for pure serial correlation be
considered.
If you concluded that you have pure serial
correlation, the appropriate response is to
consider the application of Generalized Least
Squares or Newey-West standard errors.
41
42
43
44
47
48
o
o
In practice
(a) For the presence of unnecessary variable
o The t test , the F test
(b) For the omission of variable (s)
o Look for model adequacy (R bar Square., t-ratios,
signs of estimated coefficient, DW)
o If found not adequate, this may be attributed to
one of the following: omission of a relevant
variable, the use of a wrong functional form,
presence of serial correlation, etc .
50
51
52
Omitted variables
Incorrect functional form
Correlation between X and , which may be caused by
measurement error in X, simultaneous equation
considerations, combination of lagged y values and serially
correlated disturbances.
53
54
RESET Test:
55
4.Erros of measurement
If there are errors of measurement
in the regressand only, OLS
estimators are unbiased as well as
consistent but the are less efficient
(estimated variances are now larger
than in the case where there are no
such errors of measurement).
If there are errors of measurement
in the regressors, OLS estimators
are biased as well as inconsistent.
56
Estimated model:
57
6.Non-Normal Errors
CNLRM assumed that the error
term follows the normal
distribution.
Invoked the central limit theorem
(CLT) to justify the normality
assumption.
Because of CLT: we able to
established that OLS estimators are
also normally distributed.
As a results, we able to perform
hypothesis testing using t/F tests
regardless of the sample size.
Jarque-bera test can be used to
detect whether the residuals are
normally distributed.