Spurious Relationship
Spurious Relationship
Another example of a spurious relationship can be seen by examining a city's ice cream sales. The sales
might be highest when the rate of drownings in city swimming pools is highest. To allege that ice cream
sales cause drowning, or vice versa, would be to imply a spurious relationship between the two. In reality, a
heat wave may have caused both. The heat wave is an example of a hidden or unseen variable, also known
as a confounding variable.
Another commonly noted example is a series of Dutch statistics showing a positive correlation between the
number of storks nesting in a series of springs and the number of human babies born at that time. Of course
there was no causal connection; they were correlated with each other only because they were correlated
with the weather nine months before the observations.[5]
In rare cases, a spurious relationship can occur between two completely unrelated variables without any
confounding variable, as was the case between the success of the Washington Commanders professional
football team in a specific game before each presidential election and the success of the incumbent
President's political party in said election. For 16 consecutive elections between 1940 and 2000, the
Redskins Rule correctly matched whether the incumbent President's political party would retain or lose the
Presidency. The rule eventually failed shortly after Elias Sports Bureau discovered the correlation in 2000;
in 2004, 2012 and 2016, the results of the Commanders' game and the election did not match.[6][7][8] In a
similar spurious relationship involving the National Football League, in the 1970s, Leonard Koppett noted
a correlation between the direction of the stock market and the winning conference of that year's Super
Bowl, the Super Bowl indicator; the relationship maintained itself for most of the 20th century before
reverting to more random behavior in the 21st.[9]
Hypothesis testing
Often one tests a null hypothesis of no correlation between two variables, and chooses in advance to reject
the hypothesis if the correlation computed from a data sample would have occurred in less than (say) 5% of
data samples if the null hypothesis were true. While a true null hypothesis will be accepted 95% of the time,
the other 5% of the times having a true null of no correlation a zero correlation will be wrongly rejected,
causing acceptance of a correlation which is spurious (an event known as Type I error). Here the spurious
correlation in the sample resulted from random selection of a sample that did not reflect the true properties
of the underlying population.
Experiments
In experiments, spurious relationships can often be identified by controlling for other factors, including
those that have been theoretically identified as possible confounding factors. For example, consider a
researcher trying to determine whether a new drug kills bacteria; when the researcher applies the drug to a
bacterial culture, the bacteria die. But to help in ruling out the presence of a confounding variable, another
culture is subjected to conditions that are as nearly identical as possible to those facing the first-mentioned
culture, but the second culture is not subjected to the drug. If there is an unseen confounding factor in those
conditions, this control culture will die as well, so that no conclusion of efficacy of the drug can be drawn
from the results of the first culture. On the other hand, if the control culture does not die, then the researcher
cannot reject the hypothesis that the drug is efficacious.
Disciplines whose data are mostly non-experimental, such as economics, usually employ observational data
to establish causal relationships. The body of statistical techniques used in economics is called
econometrics. The main statistical method in econometrics is multivariable regression analysis. Typically a
linear relationship such as
is hypothesized, in which is the dependent variable (hypothesized to be the caused variable), for
th
j = 1, ..., k is the j independent variable (hypothesized to be a causative variable), and is the error term
(containing the combined effects of all other causative variables, which must be uncorrelated with the
included independent variables). If there is reason to believe that none of the s is caused by y, then
estimates of the coefficients are obtained. If the null hypothesis that is rejected, then the
alternative hypothesis that and equivalently that causes y cannot be rejected. On the other hand,
if the null hypothesis that cannot be rejected, then equivalently the hypothesis of no causal effect of
on y cannot be rejected. Here the notion of causality is one of contributory causality: If the true value
, then a change in will result in a change in y unless some other causative variable(s), either
included in the regression or implicit in the error term, change in such a way as to exactly offset its effect;
thus a change in is not sufficient to change y. Likewise, a change in is not necessary to change y,
because a change in y could be caused by something implicit in the error term (or by some other causative
explanatory variable included in the model).
Regression analysis controls for other relevant variables by including them as regressors (explanatory
variables). This helps to avoid mistaken inference of causality due to the presence of a third, underlying,
variable that influences both the potentially causative variable and the potentially caused variable: its effect
on the potentially caused variable is captured by directly including it in the regression, so that effect will not
be picked up as a spurious effect of the potentially causative variable of interest. In addition, the use of
multivariate regression helps to avoid wrongly inferring that an indirect effect of, say x1 (e.g., x1 → x2 → y)
is a direct effect (x1 → y).
Just as an experimenter must be careful to employ an experimental design that controls for every
confounding factor, so also must the user of multiple regression be careful to control for all confounding
factors by including them among the regressors. If a confounding factor is omitted from the regression, its
effect is captured in the error term by default, and if the resulting error term is correlated with one (or more)
of the included regressors, then the estimated regression may be biased or inconsistent (see omitted variable
bias).
In addition to regression analysis, the data can be examined to determine if Granger causality exists. The
presence of Granger causality indicates both that x precedes y, and that x contains unique information
about y.
Other relationships
There are several other relationships defined in statistical analysis as follows.
Direct relationship
Mediating relationship
Moderating relationship
See also
Causality
Correlation does not imply causation
Illusory correlation
Model specification
Omitted-variable bias
Post hoc fallacy
Statistical model validation
One in ten rule
Literature
David A. Freedman (1983) A Note on Screening Regression Equations, The American
Statistician, 37:2, 152-155, DOI: 10.1080/00031305.1983.10482729
Footnotes
1. Burns, William C., "Spurious Correlations (https://web.archive.org/web/20190925212058/htt
p://www.burns.com/wcbspurcorl.htm)", 1997.
2. Pearl, Judea. "UCLA 81st Faculty Research Lecture Series" (http://singapore.cs.ucla.edu/LE
CTURE/lecture_sec1b.htm). singapore.cs.ucla.edu. Retrieved 2019-11-10.
3. Yule, G. Udny (1926-01-01). "Why do we Sometimes get Nonsense-Correlations between
Time-Series? A Study in Sampling and the Nature of Time-Series" (https://semanticscholar.o
rg/paper/bcaa3dd240555b9e93197f49f34531abecf439e1). Journal of the Royal Statistical
Society. 89 (1): 1–63. doi:10.2307/2341482 (https://doi.org/10.2307%2F2341482).
JSTOR 2341482 (https://www.jstor.org/stable/2341482). S2CID 126346450 (https://api.sema
nticscholar.org/CorpusID:126346450).
4. Granger, Clive W. J.; Ghysels, Eric; Swanson, Norman R.; Watson, Mark W. (2001). Essays
in Econometrics: Collected Papers of Clive W. J. Granger (https://archive.org/details/essaysi
neconomet0001gran). Cambridge University Press. ISBN 978-0521796491.
5. Sapsford, Roger; Jupp, Victor, eds. (2006). Data Collection and Analysis. Sage. ISBN 0-
7619-4362-5.
6. Hofheimer, Bill (October 30, 2012). " 'Redskins Rule': MNF's Hirdt on intersection of football
& politics" (http://www.espnfrontrow.com/2012/10/redskins-rule-mnfs-hirdt-on-intersection-of-
football-politics/). ESPN. Retrieved October 16, 2016.
7. Manker, Rob (November 7, 2012). "Redskins Rule: Barack Obama's victory over Mitt
Romney tackles presidential predictor for its first loss" (http://articles.chicagotribune.com/201
2-11-07/business/ct-talk-redskins-rule-1108-20121107_1_popular-vote-home-game-redskin
s-victory). Chicago Tribune. Retrieved November 8, 2012.
8. Pohl, Robert S. (2013). Urban Legends & Historic Lore of Washington (https://books.google.
com/books?id=rZIVBAAAQBAJ). The History Press. pp. 78–80. ISBN 978-1625846648.
9. Don Peppers. "Big Data. Super Bowl. Small Minds" (http://www.linkedin.com/today/post/artic
le/20130204035821-17102372-big-data-super-bowl-small-minds). Retrieved December 31,
2015.
References
Gumbel, E.J. (1926), "Spurious correlation and its significance to physiology", Journal of the
American Statistical Association, 21 (154): 179–194, doi:10.1080/01621459.1926.10502169
(https://doi.org/10.1080%2F01621459.1926.10502169)
Banerjee, A.; Dolado, J.; Galbraith, J. W.; Hendry, D. F. (1993). Co-Integration, Error-
Correction, and the Econometric Analysis of Non-Stationary Data. Oxford University Press.
pp. 70–81. ISBN 0-19-828810-7.
Pearl, Judea (2000). Causality: Models, Reasoning and Inference (https://archive.org/details/
causalitymodelsr0000pear). Cambridge University Press. ISBN 0521773628.
External links
Spurious correlations (http://www.tylervigen.com/spurious-correlations) – a collection of
examples