Regression and Assumptions
Regression and Assumptions
EVIEWS
1
Seven assumptions about a good regression model
2
(Assumption no. 1)
3
Goodness of Data Fit
4
(Assumption no. 2)
**
t- test
5
Individual significance of the variable
6
For Example>>
Variables:
We have four variables, Y, X1, X2 X3
Here Y is dependent and X1, X2 X3 are independent
Here, u is the residual for population regression line while e is the residual for
sample regression line. e is the estimator of u. We want to know the nature of u
from e.
7
Tips
If the sample collection is done as per the
statistical guideline (several random procedures)
then sample regression line can be a
representative of population regression line.
9
Hypothesis Setting
Null hypothesis : B1=0
Alternative hypothesis: B1≠0
Joint Significace
Independent variables should be jointly significant
to explain dependent variable
**
F- test
ANOVA
(Analysis of Variance)
11
Joint significance
12
Joint hypothesis setting
13
Few things
14
Residual Analysis
(Assumption no. 4)
**
Breusch-Godfrey serial correlation LM test : BG test
16
Serial correlation
17
How serial correlation can be formed in the
model?
• Incorrect model specification,
• omitted variables,
• incorrect functional form,
• incorrectly transformed data.
18
Detection of serial correlation
19
Hypothesis setting
20
(Assumption no. 5)
21
• Heteroscedasticity is a term used to the
describe the situation when the variance of the
residuals from a model is not constant.
22
How the heteroscedasticity may form?
– Incorrect model specification,
– Incorrectly transformed data,
23
Hypothesis setting for heteroscedasticity
24
Detection of heteroscedasticity
• Bruesch-Pegan-Godfrey test
• White heteroscedasticity test
25
(Assumption no. 6)
26
Setting the hypothesis:
28
An Emperical
Model Development
29
Our hypothetical model
Variables:
30
DATA
35
Few things
2.0
1.6
1.2
1.2
0.8
0.8
0.4
0.4
0.0
-0.4
-0.8
5 10 15 20 25 30 35
0.8
0.6
0.4
0.2
0.0
-0.2
-0.4
-0.6
5 10 15 20 25 30 35
Y Residuals
(Assumption no. 1)
Goodness of Fit Data
R-square: 0.1684
Since the p-value is more than 5 percent (0.05) we can not reject null and meaning
we accept null meaning B1=0. In other words, X1 can not influence Y in the
population.
Since p-value (0.03035) is less than 5 percent meaning that we can reject null and
accept alternative hypothesis. It means that variable X2 can influence variable Y in
the population but what direction we can not say as alternative hypothesis is ≠.
43
Assumption no 4
SERIAL OR AUTOCORRELATION
Breusch-Godfrey Serial Correlation LM Test:
Since the p-value ( 0.3185) of Obs*R-squared is more than 5 percent (p>0.05), we can not reject null hypothesis meaning
that residuals (u) are not serially correlated which is desirable.
44
Assumption no. 5
Heteroscedasticy Test
The p-value of Obs*R-squared shows that we can not reject null. So residuals do
have constant variance which is desirable meaning that residuals are homoscedastic.
45
Assumption no. 6
Residual (u) Normality Test
6
Series: Residuals
Sample 1 35
5
Observations 35
4 Mean 1.15e-16
Median -0.029701
Maximum 0.972181
3 Minimum -0.505102
Std. Dev. 0.356788
2 Skewness 0.880996
Kurtosis 3.508042
1 Jarque-Bera 4.903965
Probability 0.086123
0
-0.6 -0.4 -0.2 -0.0 0.2 0.4 0.6 0.8 1.0
Jarque Berra statistics is 4.903 and the corresponding p value is 0.08612. Since p
vaue is more than 5 percent we accept null meaning that population residual (u) is
normally distrbuted which fulfills the assumption of a good regression line. 46
Assumption No. 7
Sign of the coefficients
48
References