Lecture 25 - Multiple Regression
Lecture 25 - Multiple Regression
Error Analysis
Y = a + bX This equation gives the conditional mean of Y at
any given value of X.
X
So… In reality, our line gives us the expected mean of Y given
each value of X
The line’s equation tells you how the mean on your dependent
variable changes as your independent variable goes up.
Error Analysis
Y
Error Analysis
Can tell the improvement of predicting Y when taking X into
account
Conditional standard deviations will be smaller than Y’s original SD
Y
Y Original Conditional
standard standard
deviation deviation
X
Error 2
Error 1 Y
Y
X
R-squared
AKA “coefficient of determination”…
(Square of the Pearson correlation)
Error 2
Error 1 Y
Y
X
R-Squared
Is the improvement obtained by using X in getting as near as possible to
everybody’s value for Y over just using the mean for Y alone.
Falls between 0 and 1
Of 1 means an exact fit
(there is no variation of scores around the regression line)
Of 0 means no relationship
(as much scatter as in the original Y variable and a flat regression line
through the mean of Y)
Would be the same for X regressed on Y as for Y regressed on X
Can be interpreted as the percentage of variability in Y that is explained by
X.
Error Analysis, SPSS
r2
Line to the Mean 215.464 ÷ 3019.978 = .071
Data
points to
the line
Data points
to the mean
Regression, Inferential Statistics
Inferential:
Descriptive:
What can we say about the
The equation for your line is
relationship between your
a descriptive statistic. It
variables in the population???
tells you the real, best-fitted
The inferential statistics are
line that minimizes squared
errors FOR YOUR DATA estimates based on the best-
POINTS. fitted line.
F-test
Regression, Inferential Statistics
The slope
and y-
intercept
can be
found in
the SPSS
output.
The Regression is significant
The Slope is significant.
X2 X1
For
our problem, SPSS says the equation is:
Y = 1.6 + .21X1 - .003X2
Expected Happiness = 1.6 + .21*Job Satisfaction - .003*Family Income
Multiple Regression
Y = 1.6 + .21X1 - .003X2
Multiple Regression
For our problem, SPSS says the equation is:
Y = 1.6 + .21X1 - .003X2
Expected Happiness = 1.6 + .21*Job Satisfaction - .003*Family Income
We can input values and see what the Y would be for those values.
Job Satisfaction = 3
Family Income = 20
For each 1 SD
increase in Family
Income, there is
a .19 SD decrease
in happiness
Put in single DV
Put in 2+ IVs
Click OK
(That’s it!)
Multiple Regresson - SPSS
Let’s take a look at an example in SPSS…
Writing Statistics – Regression
For Predictors, Degrees of
freedom for the t-test is the DF
Residual.
• Inferential Statistics – Regression:
• F(5, 125)=7.11, p=0.015 – Now we have DF Regression and DF Residual reported
here (just like ANOVA).
• Where does this go in a sentence? We have two independent variables
now and a different conclusion. Let’s start with our overall findings:
• A simple linear regression was conducted, which showed that age and income
together significantly predicted happiness, F(5, 125)=7.11, p=0.02.
• Next, add individual variables’ unique contributions:
• On their own, age uniquely predicted happiness [β = -.34, t(125) = 7.12, p = .013]
and income uniquely predicted happiness [β = .52, t(125) = 6.53, p = .001].
Writing Statistics – Regression