Lecture 3. Part 1 - Regression Analysis
Lecture 3. Part 1 - Regression Analysis
• Backward stepwise
Starting with all possible predictors (full model), we delete
“insignificant” predictors
Forward stepwise
• Data File. This example is based on the examples data file
Job_prof.sta (from Neter, Wasserman, and Kutner, 1989, page
473). Open this data file by selecting Open Examples from the
File menu (classic menus) or by selecting Open Examples from
the Open menu on the Home tab (ribbon bar); it is in the
Datasets folder. The first four variables (Test1-Test4) represent
four different aptitude tests that were administered to each of
the 25 applicants for entry-level clerical positions in a company.
Regardless of their test scores, all 25 applicants were hired. Once
their probationary period had expired, each of these employees
was evaluated and given a job proficiency rating (variable
Job_prof).
When Test2 was evaluated, the F value was less than the F to enter
value of 1.0, therefore, it was not entered into the model.
Forward stepwise
Now, according to the Forward stepwise regression procedure, the
subset of aptitude tests (independent variables) that best predicts the
job proficiency score (dependent variable) contains Test3, Test1, and
Test4. Therefore, the regression equation appears as follows:
𝑦 = 𝑏0 + 𝑏1 𝑥3 + b2 𝑥1 + 𝑏3 𝑥4
R-squared is 0.9615
Interpretation: The Durbin-Watson is 1.148347 which is below the rule of thumb (1.5-2/5) for job
proficiency rating. The respondents in the data are different individuals with the assumption that
they do not affect the answer of one another. The value 1.14 can be tolerated.
Note: That the assumptions for independence of observation should be satisfied upon data
collection such as the respondents should only answer the questionnaire once, otherwise a
respondent answering more than once can lead to dependence of observation.
Normality of residuals by Histogram
Statistics->Multiple Regression
-> Variables->Dependent variable
(job_prof) –independent variable- (1 2 3 4)-
>Ok
->Advance tab, click Advance Options box-
>Ok-> Ok
Click Residual/assumption/prediction tab->
click performance residual analysis
Click Normal Plot of residuals
Normality of residual by scatter plot of
residuals Normal Probability Plot of Residuals
2.5
H0: the residuals follow a
normal distribution 2.0
1.5
1.0
-1.0
-1.5
-2.0
-2.5
https://online.stat.psu.ed -8 -6 -4 -2 0 2 4 6 8
u/stat501/lesson/4/4.6 Residuals
Normality of residual by histogram
Distribution of Raw residuals
Expected Normal
8
No of obs
4
0
-8 -6 -4 -2 0 2 4 6 8
https://online.stat.psu.edu/stat501/lesson/4/4.6/4.6.1
Normality using
chi-square
Statistics ->Basic Statistics-
>Tables and banners
Specify tables(select
variables)-> Job_prof ->
Test1 Test 2 Test3 Test 4 Ok-
>ok
Normality using chi-square
Interpretation: None of the p-value is less than 0.05 hence not statistically
significant, therefore we fail to reject the null hypothesis. The residuals
follow a normal distribution.
https://www.youtube.com/watch?v=vn5a5lAL54I
Multicollinearity
This is when predictors (independent variables) are correlated with
each other.
Redundance.