Four Assumptions of Multiple Regression That Researchers Should A
Four Assumptions of Multiple Regression That Researchers Should A
2002
Elaine Waters
Recommended Citation
Osborne, Jason W. and Waters, Elaine (2002) "Four assumptions of multiple regression that researchers
should always test," Practical Assessment, Research, and Evaluation: Vol. 8 , Article 2.
DOI: https://doi.org/10.7275/r222-hv23
Available at: https://scholarworks.umass.edu/pare/vol8/iss1/2
This Article is brought to you for free and open access by ScholarWorks@UMass Amherst. It has been accepted for
inclusion in Practical Assessment, Research, and Evaluation by an authorized editor of ScholarWorks@UMass
Amherst. For more information, please contact [email protected].
Osborne and Waters: Four assumptions of multiple regression that researchers should a
The third method of detecting curvilinearity is to routinely run regression analyses that incorporate curvilinear
components (squared and cubic terms; see Goldfeld and Quandt, 1976 or most regression texts for details on how to do
this) or utilizing the nonlinear regression option available in many statistical packages. It is important that the
nonlinear aspects of the relationship be accounted for in order to best assess the relationship between variables.
VARIABLES ARE MEASURED WITHOUT ERROR (RELIABLY)
The nature of our educational and social science research means that many variables we are interested in are also
difficult to measure, making measurement error a particular concern. In simple correlation and regression, unreliable
measurement causes relationships to be under-estimated increasing the risk of Type II errors. In the case of multiple
regression or partial correlation, effect sizes of other variables can be over-estimated if the covariate is not reliably
measured, as the full effect of the covariate(s) would not be removed. This is a significant concern if the goal of research
is to accurately model the “real” relationships evident in the population. Although most authors assume that reliability
estimates (Cronbach alphas) of .7-.8 are acceptable (e.g., Nunnally, 1978) and Osborne, Christensen, and Gunter (2001)
reported that the average alpha reported in top Educational Psychology journals was .83, measurement of this quality
still contains enough measurement error to make correction worthwhile, as illustrated below.
Correction for low reliability is simple, and widely disseminated in most texts on regression, but rarely seen in the
literature. We argue that authors should correct for low reliability to obtain a more accurate picture of the “true”
relationship in the population, and, in the case of multiple regression or partial correlation, to avoid over-estimating the
effect of another variable.
Reliability and simple regression
Since “the presence of measurement errors in behavioral research is the rule rather than the exception” and
“reliabilities of many measures used in the behavioral sciences are, at best, moderate” (Pedhazur, 1997, p. 172); it is
important that researchers be aware of accepted methods of dealing with this issue. For simple regression, Equation #1
provides an estimate of the “true” relationship between the IV and DV in the population:
(1)
In this equation, r12 is the observed correlation, and r11 and r22 are the reliability estimates of the variables. Table 1
and Figure 2 presents examples of the results of such a correction.
Table 1: Values of r and r2 after correction for attenuation
Reliability of DV and IV
Perfect
.80 .70 .60 .50
measurement
Observed
r
https://scholarworks.umass.edu/pare/vol8/iss1/2
r r2 r r2 r r2 r r2 r r2
DOI: https://doi.org/10.7275/r222-hv23 Page 2 of 5 2
Osborne and Waters: Four assumptions of multiple regression that researchers should a
.10 .10 .01 .13 .02 .14 .02 .17 .03 .20 .04
.20 .20 .04 .25 .06 .29 .08 .33 .11 .40 .16
.40 .40 .16 .50 .25 .57 .33 .67 .45 .80 .64
.60 .36 .75 .57 .86 .74 -- -- -- --
.60
Note: for simplicity we show an example where both IV and DV have identical reliability
estimates. In some of these hypothetical examples we would produce impossible values, and so do
not report these.
Figure 2: Change in variance accounted for as correlations are corrected for low reliability
As Table 1 illustrates, even in cases where reliability is .80, correction for attenuation substantially changes the effect
size (increasing variance accounted for by about 50%). When reliability drops to .70 or below this correction yields a
substantially different picture of the “true” nature of the relationship, and potentially avoids a Type II error.
Reliability and Multiple Regression
With each independent variable added to the regression equation, the effects of less than perfect reliability on the
strength of the relationship becomes more complex and the results of the analysis more questionable. With the addition
of one independent variable with less than perfect reliability each succeeding variable entered has the opportunity to
claim part of the error variance left over by the unreliable variable(s). The apportionment of the explained variance
among the independent variables will thus be incorrect. The more independent variables added to the equation with low
levels of reliability the greater the likelihood that the variance accounted for is not apportioned correctly. This can lead
to erroneous findings and increased potential for Type II errors for the variables with poor reliability, and Type I errors
for the other variables in the equation. Obviously, this gets increasingly complex as the number of variables in the
equation grows.
A simple example, drawing heavily from Pedhazur (1997), is a case where one is attempting to assess the relationship
between two variables controlling for a third variable (r12.3). When one is correcting for low reliability in all three
variables Equation #2 is used:
(2)
Where r11, r22, and r33 are reliabilities, and r12, r23, and r13 are relationships between variables. If one is only correcting
for low reliability in the covariate one could use Equation #3:
Published by ScholarWorks@UMass Amherst, 2002 Page 3 of 5 3
Practical Assessment, Research, and Evaluation, Vol. 8 [2002], Art. 2
(3)
Table 2 presents some examples of corrections for low reliability in the covariate (only) and in all three variables.
Table 2: Values of r12.3 and r212.3 after correction low reliability
Reliability of Reliability of All
Examples:
Covariate Variables
.80 .70 .60 .80 .70 .60
Observed
r12 r13 r23 r12.3 r12.3 r12.3 r12.3 r12.3 r12.3
r12.3
Table 2 shows some of the many possible combinations of reliabilities, correlations, and the effects of correcting for only
the covariate or all variables. Some points of interest: (a) as in Table 1, even small correlations see substantial effect
size (r2) changes when corrected for low reliability, in this case often toward reduced effect sizes (b) in some cases the
corrected correlation is not only substantially different in magnitude, but also in direction of the relationship, and (c) as
expected, the most dramatic changes occur when the covariate has a substantial relationship with the other variables.
ASSUMPTION OF HOMOSCEDASTICITY
Homoscedasticity means that the variance of errors is the same across all levels of the IV. When the variance of errors
differs at different values of the IV, heteroscedasticity is indicated. According to Berry and Feldman (1985) and
Tabachnick and Fidell (1996) slight heteroscedasticity has little effect on significance tests; however, when
heteroscedasticity is marked it can lead to serious distortion of findings and seriously weaken the analysis thus
increasing the possibility of a Type I error.
This assumption can be checked by visual examination of a plot of the standardized residuals (the errors) by the
regression standardized predicted value. Most modern statistical packages include this as an option. Figure 3 show
examples of plots that might result from homoscedastic and heteroscedastic data.
Figure 3. Examples of homoscedasticity and heteroscedasticity
Ideally, residuals are randomly scattered around 0 (the horizontal line) providing a relatively even distribution.
Heteroscedasticity is indicated when the residuals are not evenly scattered around the line. There are many forms
https://scholarworks.umass.edu/pare/vol8/iss1/2
heteroscedasticity can take, such as a bow-tie or fan shape. When the plot of residuals appears to deviate substantially
DOI: https://doi.org/10.7275/r222-hv23 Page 4 of 5 4
Osborne and Waters: Four assumptions of multiple regression that researchers should a
from normal, more formal tests for heteroscedasticity should be performed. Possible tests for this are the Goldfeld-
Quandt test when the error term either decreases or increases consistently as the value of the DV increases as shown in
the fan shaped plot or the Glejser tests for heteroscedasticity when the error term has small variances at central
observations and larger variance at the extremes of the observations as in the bowtie shaped plot (Berry & Feldman,
1985). In cases where skew is present in the IVs, transformation of variables can reduce the heteroscedasticity.
CONCLUSION
The goal of this article was to raise awareness of the importance of checking assumptions in simple and multiple
regression. We focused on four assumptions that were not highly robust to violations, or easily dealt with through
design of the study, that researchers could easily check and deal with, and that, in our opinion, appear to carry
substantial benefits.
We believe that checking these assumptions carry significant benefits for the researcher. Making sure an analysis meets
the associated assumptions helps avoid Type I and II errors. Attending to issues such as attenuation due to low
reliability, curvilinearity, and non-normality often boosts effect sizes, usually a desirable outcome.
Finally, there are many non-parametric statistical techniques available to researchers when the assumptions of a
parametric statistical technique is not met. Although these often are somewhat lower in power than parametric
techniques, they provide valuable alternatives, and researchers should be familiar with them.
References
Berry, W. D., & Feldman, S. (1985). Multiple Regression in Practice. Sage University Paper Series on Quantitative
Applications in the Social Sciences, series no. 07-050). Newbury Park, CA: Sage.
Cohen, J., & Cohen, P. (1983). Applied multiple regression/correlation analysis for the behavioral sciences. Hillsdale,
NJ: Lawrence Erlbaum Associates, Inc.
Nunnally, J. C. (1978). Psychometric Theory (2nd ed.). New York: McGraw Hill.
Osborne, J. W., Christensen, W. R., & Gunter, J. (April, 2001). Educational Psychology from a Statistician’s
Perspective: A Review of the Power and Goodness of Educational Psychology Research. Paper presented at the national
meeting of the American Education Research Association (AERA), Seattle, WA.
Osborne, J. W. (2001). A new look at outliers and fringeliers: Their effects on statistic accuracy and Type I and Type II
error rates. Unpublished manuscript, Department of Educational Research and Leadership and Counselor Education,
North Carolina State University.
Pedhazur, E. J., (1997). Multiple Regression in Behavioral Research (3rd ed.). Orlando, FL:Harcourt Brace.
Tabachnick, B. G., Fidell, L. S. (1996). Using Multivariate Statistics (3rd ed.). New York: Harper Collins College
Publishers
Tabachnick, B. G., Fidell, L. S. (2001). Using Multivariate Statistics (4th ed.). Needham Heights, MA: Allyn and Bacon
Contact Information:
Jason W. Osborne, Ph.D
ERLCE, Campus Box 7801
Poe Hall 608,
North Carolina State University
Raleigh NC 27695-7801
(919) 515-1714
[email protected]