Data Analysis With Small Samples and Non-Normal Data - Nonparametrics and Other Strategies
Data Analysis With Small Samples and Non-Normal Data - Nonparametrics and Other Strategies
Data Analysis With Small Samples and Non-Normal Data - Nonparametrics and Other Strategies
Small Samples
and Non-Normal Data
POCKET GUIDES TO
SOCIAL WORK RESEARCH METHODS
Series Editor
Tony Tripodi, DSW
Professor Emeritus, Ohio State University
Data Analysis
with Small Samples and
Non-Normal Data
Nonparametrics and
Other Strategies
1
3
Oxford University Press is a department of the University of Oxford. It furthers
the University’s objective of excellence in research, scholarship, and education
by publishing worldwide. Oxford is a registered trade mark of Oxford University
Press in the UK and certain other countries.
1 3 5 7 9 8 6 4 2
Printed by WebCom, Inc., Canada
Acknowledgments ix
1 Introduction to Nonparametrics 1
2 Analyzing Single Variables and Single Groups 27
3 Comparing Two or More Independent Groups 69
4 Comparing Two or More Related Groups 107
5 Predicting with Multiple Independent Variables 149
vii
Acknowledgments
ix
Data Analysis with Small Samples
and Non-Normal Data
1
Introduction
to Nonparametrics
OVERVIEW
Research training in the social sciences must cover a wide range of
methodological and statistical coursework in a few years of doctoral
study. Frequently, doctoral students take two or three statistics courses
to learn basic analytic techniques, and students who take additional
courses often do so to learn advanced inferential techniques, such
as Hierarchical Linear Modeling (HLM) and Structural Equation
Modeling (SEM). These techniques can be very useful when one is work-
ing with large data sets and normally distributed data.
However, researchers regularly find themselves working with small
data sets, and this is especially commonplace for intervention research
studies, which are the hallmark of social science research. Similarly,
researchers may find themselves with data sets that simply do not meet
the assumptions required for the parametric techniques they were
trained to use. Yet very little, if any, time in their research training is
devoted to the analytic techniques that would allow them to properly
analyze these kinds of data—nonparametric statistics. This gap in edu-
cation leaves individuals without the necessary tools to conduct suitable
1
2 Data Analysis with Small Samples and Non-Normal Data
additional resources that will meet the needs of those interested in the
mathematical foundations of nonparametrics.
Another important topic that is not integrated into the chapters is
the difference between statistical significance (i.e., results from a statis-
tical test) and practical importance. Statistical tests are mathematical
procedures that provide a mathematical test of a hypothesis. Finding
significance from a statistical test is not the same as determining how
important the statistical finding might be for a particular research ques-
tion. Your data analyses may indicate statistical significance, when in
the real world, the difference is practically inconsequential. We strongly
urge anyone conducting statistical analyses, especially those making
decisions from statistical findings, to review fully the topic of practi-
cal importance related to statistics. (For more information on statistical
significance versus practical importance, see Hojat & Xu, 2004.)
Finally, we do not introduce bootstrapping as an option in our chapters.
Bootstrapping is another option for researchers dealing with small data
sets, and in simple terms, it draws inferences about the population under
investigation by resampling the data set from which a sample of the pop-
ulation came. Since the introduction of the bootstrap (Efron, 1979), proce-
dures have been improved and its use has been demonstrated to be a viable
alternative when dealing with challenging data (Davidson & Hinkley, 1998;
Janssen & Pauls, 2003; Wu, 1986). Because bootstrapping is an advanced
technique, we leave the discussion of it and how to use it to others.
ORGANIZATION
This book is organized so you can quickly locate the nonparametric
technique that is optimal for your data, and then follow the clear, step-
by-step instructions for using SPSS to conduct your analysis. Each
chapter includes a scenario typically found in social science research
to contextualize the analytic techniques, including details about the
research design, population, sample, groups within the sample, num-
ber of cases/observations/data points, time (if relevant), variables, level
of measurement, description of the variables’ distribution, and a series
of research questions and hypotheses. Next, using the scenario as the
example, relevant research questions and hypotheses guide the discus-
sion of the nonparametric method(s) that can be used to answer them.
4 Data Analysis with Small Samples and Non-Normal Data
In this way, you can easily match your particular questions to the
appropriate technique, and then follow the example to its conclusion
and interpretation.
Beginning with Chapter 2, each chapter starts with a list of chap-
ter objectives, a scenario that serves as the backdrop for the chapter’s
research questions, and a table of the nonparametric methods covered
in the chapter. Chapter 2 covers nonparametric methods using single
variables or groups. Chapter 3 presents procedures for comparing two or
more independent groups on a particular characteristic, and Chapter 4
describes techniques for comparing two or more related groups. The
fifth chapter describes methods for predicting a particular value or level
based on a set of independent variables, when the data do not meet the
assumptions for least-squares regressions.
The primary software used throughout the book is SPSS, but because
SPSS currently does not include nonparametric regression techniques
in its software, we utilize an add-in to Excel—X LSTAT—for the steps to
conduct the predictive analyses.
Additional useful information is supplied in the Appendices (e.g.,
SPSS syntax, missingness) and from SPSS Help (e.g., how to prepare
your data for analysis). We also suggest additional resources (i.e., web
links, articles, books) that will provide detailed foundational informa-
tion and formulas. On this book’s companion website you will find all
the data sets used in the book’s analyses. After downloading the data
sets, you can follow the steps for each nonparametric method using
the example data to reinforce your learning and understanding of the
process.
You may be tempted to go straight to the data analysis chapters at
this point, but if you do, you will not necessarily understand the foun-
dation for what you are doing. Consequently, you may err in describ-
ing your procedures to others, or in presenting the conclusions and
interpretations of your analyses. We urge you to review carefully all the
information in this chapter before proceeding.
KEY TERMS
Several key terms used are used throughout the book. These terms are
as follows:
PARAMETRIC ASSUMPTIONS
Assumption 1. The Population Data Are Approximately Normally Distributed
That the data are approximately normally distributed may be the most
common parametric assumption (Hoskins, n.d., p. 4). The normal-
ity assumption is important whenever a scale variable (i.e., continu-
ous variable) is used in a parametric statistics test. Most parametric
tests involve a scale variable that is assumed to represent data from
a symmetric bell-shaped curve distribution. Small departures from
the symmetrical bell curve often have limited negative consequences
for an analysis, but larger departures from normality can make
the results wholly incorrect and/or uninterpretable. Furthermore,
although some parametric tests like ANOVA purport to be robust
8 Data Analysis with Small Samples and Non-Normal Data
against mild violations of normality, this does not mean that you can
ignore the normality assumption, particularly with small samples. In
addition, you are left to answer the question as to what, exactly, con-
stitutes a mild departure from normality.
All scale variables must be tested for normality prior to includ-
ing them in a parametric test. This is especially true when conducting
multivariable analyses in which multiple scale variables are included
in a single statistical model (e.g., multivariate regression). Each varia-
ble must demonstrate its association with a smooth, continuous, and
bell-shaped distribution. However, when the sample size is small, the
gaps between values (i.e., the data points on the curve) make the test
of normality more likely to fail. This is particularly true when testing a
Likert-t ype response variable that is being used as a scale variable rather
than an ordinal variable, and even more so when the response options
number five or fewer. Thus, although many social scientists tend to use
Likert-scaled response options as if they represent a continuous varia-
ble, one must be sure to take the size of the sample into consideration.
For example, it is easier to justify the use of a Likert-t ype variable if the
sample size is 1,000 and the response options range from 1 to 7 or 1 to 9
than if the sample size is 50 and the Likert response options are 1 to 4.
Testing to determine if a variable can be treated as a scale variable is
included in upcoming chapters.
You have two primary options when your variable fails the normal-
ity test. Your first option is to use a transformation strategy to normal-
ize your variable so that it will pass a normality test (e.g., log or square
transformation). Not all variables can be transformed into a normal
distribution (e.g., data with zeroes and negative values cannot be log
transformed), and when your transformation attempts fail, you cannot
simply use them anyway, hoping that your test will be robust against the
normality violation. When a non-normal variable does respond well to
transformation, you may use it in a parametric procedure, but then you
cannot interpret your findings as if the variable were not transformed.
You are faced with making sense of the transformed variable in the
analysis and in your findings, and this can be complicated, if not impos-
sible. (For more information on variable transformation, see McDonald,
2014, and Stat Trek, n.d.) Your second choice, when faced with a non-
normal variable, is simply to use a nonparametric test or the aforemen-
tioned bootstrapping.
Introduction to Nonparametrics 9
each data set is unique, assuming your data set is one that conforms
to these earlier simulation studies may place your findings at risk if
you choose to ignore parametric assumptions.
In simple terms, you are assuming that the data points for the relation-
ship between the dependent and independent variable have similar var-
iability along the entire range of the prediction line. For example, if you
were to examine a scatter plot of the prediction line and data points,
homoscedasticity is present if the points along the prediction line show
a similar pattern along the entire range of the prediction line. A viola-
tion of this assumption (i.e., heteroscedasticity) results in some cases
having more pull than others, as demonstrated by scatter plots that
can sometimes look like a cone or a fan. Heteroscedasticity can result
in invalid prediction values or inaccurate estimates of variable rela-
tionships. When faced with heteroscedasticity, you have two options.
First, you can look for variables that are causing the heteroscedastic-
ity and examine the cases that appear to be problematic to see if you
can legitimately replace them. However, it is not appropriate to go “case
shopping,” because any case that represents a real respondent cannot
be excluded without cause. The other option is to use a nonparametric
approach, such as nonparametric or nonlinear regression, as discussed
in Chapter 5. (For more information on testing the homoscedasticity
assumption, see Gaskin, 2015.)
OTHER ASSUMPTIONS
The assumptions considered next apply to both parametric and nonpa-
rametric statistical techniques, and are the product of a well-designed
study. Violating any of the assumptions below weakens or possibly
invalidates any conclusions you draw from your statistical analyses.
that set them apart from parametric procedures. Over the years, many
people have disseminated the misconceptions, while others have refuted
them (e.g., Hunter, 1993; Sawilowsky, 2005). However, the repudiations
do not seem to have made much of an impact in correcting people’s
beliefs. We attribute this, in large part, to the lack of inclusion of nonpa-
rametric training in statistical courses. The confusion about these mis-
conceptions makes it difficult to avoid making errors when choosing
your statistical strategies. The list of misconceptions is quite long, but
a few that seem to come up more frequently than others are as follows:
1. The p-value can indicate how incompatible the data are with a
specified statistical model.
2. The p-values do not measure the probability that the studied
hypothesis is true, or the probability that the data were produced
by random chance alone.
3. Scientific conclusions and business or policy decisions should
not be based only on whether a p-value passes a specific
threshold.
4. Proper inference requires full reporting and transparency.
5. A p-value, or statistical significance, does not measure the size of
an effect or the importance of an effect.
6. By itself, a p-value does not provide a good measure of evidence
regarding a model or hypothesis. (Wasserman & Lazar, 2016)
FIRST STEPS
If you plan to follow the steps in the chapters to analyze your own data,
you will need to get your data into SPSS. Procedures for getting your
data into SPSS are found in SPSS’s help feature (click on the Help menu
option). Once your data are in SPSS, the next step is to verify that all
your variables are assigned to the appropriate level of measurement.
Figure 1.1 displays the data view for the example data used in Chapter 2.
Note that each variable is assigned one of the three levels of measure-
ment values (e.g., nominal, ordinal, or scale).
Once your data are in SPSS, you recognize SPSS notation, and you
are familiar with the key terms, you are ready to move on to Chapter 2.
Analyzing Single
Variables and
Single Groups
CHAPTER OBJECTIVES
This chapter focuses on techniques for assessing single variables and
single groups, organized by the variable’s level of measurement. The fol-
lowing topics are covered:
27
28 Data Analysis with Small Samples and Non-Normal Data
INTRODUCTION
Chapter 1 describes the importance of understanding the data and vari-
ables prior to and while using them in statistical tests. This begins by first
identifying the measurement level of the variables in the data set. The
measurement level dictates which nonparametric procedures can be used
to answer research questions. The second step is to review the quality of
the information provided by each variable. For scale-type variables, the
quality of information is related to capturing the true variability or distri-
bution of the data. In addition, variables that represent concepts or ideas
that have very little variability are not as informative and sometimes con-
tribute less to a statistical model. For example, a variable that has very
little variability (i.e., small variance) does not offer as much information
about the characteristic the variable is attempting to capture as a vari-
able that has more variability. In other words, in most cases, the more
recorded variability, the more information available to answer research
questions. This is not to say that variables with large variances are always
preferred over variables with smaller variances, just that scale (i.e., con-
tinuous) variables with minimal variances do not contribute as much to
the analysis. Therefore, before using any scale variable, you should review
the underlying concept you are measuring to investigate if the observed
variance has theoretical support, and if there is enough variance to war-
rant its inclusion in an analysis.
For categorical and dichotomous variables, the quality of informa-
tion is related to coverage in the variable categories. The more cover-
age, the more information you have for the analyses. For example, if
using a dichotomous variable for ethnicity (i.e., a nominal categori-
cal variable) and the data set contains 30 cases (i.e., a small data set)
with 27 Hispanic/L atino(a) people and three non-Hispanic/L atino(a)
people, you have much more information for Hispanic people than
for non-Hispanic people. In this situation, you must decide if you
have enough information to consider differences between the two
Analyzing Single Variables and Single Groups 29
Dr. Chonody has just completed a study investigating the coming out process
for people who are gender and sexually diverse. She conducted extensive,
semi-structured interviews with 29 LGBTQ youth who were recruited through
websites and newspaper advertisements. Dr. Chonody used grounded theory
and narrative approaches, audio taping the interviews. The tapes were
transcribed and then analyzed using a combination of grounded theory
and cross-case procedures. She ensured the quality of the data through peer
debriefing, independent coding into key categories, and member checking. After
analyzing the qualitative data, Dr. Chonody asked the respondents to fill out
a paper-and-pencil questionnaire that contained several validated measures
that could confirm or extend her findings. In combination, the qualitative and
quantitative data provide improved validity and a more complete picture of
the respondents. Dr. Chonody has hired you as her statistical consultant. Her
research questions follow.
BINOMIAL TEST
The binomial test is used for testing the statistical significance of possible
differences between an expected proportion and an observed proportion.
The binomial test is called an exact test because it provides the exact prob-
ability that the data represent the stated expected proportion. A compa-
rable parametric test is a t-test that uses z scores to test a hypothesis of an
expected proportion, but it is used rarely for proportion testing outside of
introductory statistics courses. The assumptions associated with the bino-
mial test are that the variable used in the test is dichotomous —two cat-
egories representing success and failure —or can represent a dichotomous
32 Data Analysis with Small Samples and Non-Normal Data
1. You recently opened a new health clinic, and you know the
percentage of people with STDs in your city, so you want to test
if the proportion of people coming to your new clinic is similar
to the city percentage or significantly different.
2. A treatment center’s administrators believe that 90%
of their patients are satisfied with their care, and you
suspect that the administrators are overestimating the
satisfaction level.
Prior to conducting any hypothesis test, you must examine the informa-
tion provided by the variables. For Research Question #1 (see Box 2.2),
the variable is dichotomous, representing two nominal categories —
having a confidante or not having a confidante. For dichotomous vari-
ables, the easiest way to examine the variable information is to run a
frequency of the responses.
SPSS Process
To begin your analysis
Because the interest is only in frequencies, leave all the options for the
buttons on the right in their default position.
Analyzing Single Variables and Single Groups 33
SPSS Output
Two tables will appear in the IBM SPSS Statistical Viewer window
(i.e., analysis output window).
• The first table shows the number of valid and missing cases
found in the data set. Use this table to verify that the table
matches the number of cases intended to be part of the analysis.
• The second table shows the frequency at which the 28 youths
report having a confidante (see Table 2.2 —Frequency Table for
Research Question #1).
Table 2.2 shows that 21 out of the 29, or 72.4% of the youths, have
a confidante. Therefore, do the data collected in the follow-up support
Dr. Chonody’s inference that 60% of participants had a confidante? At
first glance, it appears that they do not (60% is too low), but a rigorous
researcher will test further to be sure.
You might be tempted to utilize a parametric test like a t-test using
z scores, despite a sample size of fewer than 30. If you run a one-sample
test in SPSS (i.e., Analyze => Compare Means => One Sample
t-
T-test…), the 95% confidence interval for the percentage is .55 to .90
(55% to 90%). This would support Dr. Chonody’s inference that 60% of
the youths have confidantes. However, because the sample has fewer
than the sample-size guideline, you should use a nonparametric proce-
dure called a binomial test, which is well suited for this analysis because
of the ranking approach used in testing hypotheses.
34 Data Analysis with Small Samples and Non-Normal Data
SPSS Process
To begin this analysis,
have a confidante?
□ Click the move arrow to move the variable over to the Test
Fields: area.
□ Select => Test Options in the Select an item: area if you wish
to change the significance level from the default (0.05) for a test,
or if you wish to change the confidence interval range from
the default (95.0%). Here you can also change the missing data
default (exclude cases test-by-test).
□ Select => Choose Tests from the Select an item: area to see two
The first option (Automatically choose the test based on the data)
allows SPSS to decide automatically which tests are appropriate based
on the recorded SPSS variable definition (i.e., defined level of meas-
urement for each variable). Because it is important to follow the anal-
ysis process closely to understand fully how the analysis is performed,
click to select the Customize tests option to gain access to the test
selection boxes.
□ Click the first option in the list (i.e., binomial test). Figure 2.1
shows the One-Sample Nonparametric Tests window with the
test selected that matches Research Question #1.
Shown in Figure 2.1 is the list of different tests available for analyz-
ing a single variable. The icons shown on the right side of the screen
represent the tests that are appropriate for each measurement level. For
Research Question #1, the binomial test is appropriate because it com-
pares a hypothesized proportion of responses to the observed propor-
tion of responses in the variable categories. For a dichotomous variable,
the default is that half of the respondents selected one category while
the other half selected the other category (i.e., successes are as equally
36 Data Analysis with Small Samples and Non-Normal Data
likely to occur as fails, or 50% success and 50% fails). However, Research
Question #1 proposes a 60% rate for yes responses, so
Categorical Fields.
□ Click in the Success Values: white area and enter a 1 (to indicate
window.
□ Select => Run to conduct the analysis.
SPSS Output
A variety of information shows up on the output window, including:
Findings
When you set the View in the Model Viewer window to Hypothesis
Summary View, it displays the results of the hypothesized 60%. The Sig.
value is .120, which fails to reject the hypothesis that the proportions of
with and without confidantes are .60 and .40, respectively. Whenever
you run a statistical test, you want to verify that your settings in SPSS
were as you intended. You can verify your settings by closely reviewing
the wording in the tables, values of the confidence intervals, the look of
the histograms, and the information below the histograms, etc. You can
find all this information by selecting the different dropdown options in
the Model Viewer.
Next, select the dropdown for View: on the left side of the window
and select Confidence Interval Summary View. The values under the
Lower and Upper columns show the boundary values for the 95% con-
fidence interval for the percent success identified in this data set. The
Analyzing Single Variables and Single Groups 39
calculation methods are slightly different among the three, which is why
the values are slightly different.
In statistical terms, you can say a binomial test of the null hypoth-
esis that the percent of youths with confidantes is equal to 60% fails to
be rejected at an alpha level of .05 with a p-value of .120. Another expla-
nation from the results is that if the youths responding to the follow-up
survey are typical for gay youths, then there is a 12% chance that the pro-
portion of youths who have confidantes is 60%. A Clopper-Pearson cal-
culation provides a 95% confidence interval for proportion rate of 52.8%
to 87.3%. In this analysis, selecting 95% for the confidence interval pro-
vides a range of possible percentages of youths with a confidante in which
we are 95% confident that the true percentage falls. If 90% was selected,
the range would be a bit wider, but the confidence that the true percent-
age falls within the range would be smaller. Note that this finding differs
from that of the (inappropriate) parametric t-test mentioned earlier that
found lower and upper bounds of 55% to 90%.
Practically, the findings suggest that, for this study, the number of
gay youths with a confidante is a bit higher than the number without a
confidante. For example, if you had a group of 100 gay youths who were
similar to the youths in Dr. Chonody’s study, the number of youths with
a confidante would be somewhere between 53 and 87.
CHI-SQUARE TEST
Different kinds of chi-square tests address different data analysis sit-
uations (e.g., 2 × 2 contingency table analysis, Pearson chi-square,
likelihood-ratio chi-square, etc.). Basically, the chi-square test is suited
for testing possible proportion differences between an expected propor-
tion and an observed proportion when more than two combinations of
categories are possible. When proportions among categories are exam-
ined, this is similar to investigating whether the numbers within each
category are similar. The kind of variable used in a chi-square test in
this situation is either a nominal or an ordinal variable (i.e., categorical
variable representing two or more responses —e.g., none, sometimes,
or often for each respondent). The assumptions associated with the chi-
square test are that the variables used in the test are categorical and that
responses are independent. In addition, each category should be repre-
sented with at least five responses to allow for appropriate information
coverage for the respondent categories. The strengths of the chi-square
test are that the results are easy to interpret and you can apply the test to
all kinds of categorical variables. One limitation is that the chi-square
test is much more sensitive to sample sizes and often requires a larger
sample to find significant proportional differences; however, if the sam-
ple is too large, the test finds significance too easily. Lastly, no alter-
native parametric procedure is possible without manipulating the data
into something testable with a t-test.
Other uses for the chi-square test are as follows:
1. You are examining the data from a pilot study, and you want
to investigate the proportions of people subscribing to different
religions (i.e., the expected vs. observed).
2. In your small sample of gay youths, you wish to explore your
suspicion that only 10% have come out to no one and the other
90% are evenly distributed between coming out to one or a few
and coming out to everyone.
SPSS Process
To begin your analysis,
Fields: area.
□ Click the move arrow to move Coming_Out over to Test
Fields: area.
Note: If other variables are listed in the Test Fields: area, select them and
move them back to the Fields: area on the left side of the window.
Now that SPSS knows what variables to use,
In Figure 2.1, the variable measure icons indicate the variables with
the appropriate level of measurement. For the chi-cquare test, the levels
of measurement are nominal and ordinal, and the icons change to color
when you check the Chi-Square Test checkbox.
The options for the chi-square test are different from the binomial
test. The only time a change to the default options is necessary is when
you are testing the category proportions for something other than being
42 Data Analysis with Small Samples and Non-Normal Data
SPSS Output
To see the details of the test, double-click the Hypothesis Test Summary
information (i.e., point the mouse within the Hypothesis Test Summary
table and double-click). The Model Viewer window will open to display
the detailed information.
Findings
The null hypothesis for the chi-square test for Coming_Out is that equal
proportions of responses exist for the three categories. When the Model
Viewer window first opens, it presents a histogram on the right and the test
results on the left. The Sig. value for Coming_Out indicates that the test fails
to reject the null hypothesis (i.e., Sig. value = .185 is larger than the default
alpha level of .05), thereby suggesting that there is not enough information
to identify a proportional difference among youths’ Coming_Out catego-
ries. Another explanation is that the proportions are similar enough to not
be identified statistically as significantly different. In statistical terms, you
may say that the chi-square test for a null hypothesis of equal proportion of
youth Coming_Out categories fails to be rejected at α = .05, with p = .185.
However, when you examine the histograms for Coming_ Out
among gay youth categories (i.e., no one, one or a few, and everyone),
they seem to show possible differences despite not being significantly
different, statistically. More than likely, if the sample size were larger
and followed similar proportions, there would be enough power to
identify differences. This is one important reason always to include as
many respondents in your study as possible, to increase the likelihood of
finding statistical differences. In this situation, you do not have enough
information from the variables to identify statistical significance; that is,
Analyzing Single Variables and Single Groups 43
the statistics are keeping you from drawing a potentially wrong conclu-
sion based on viewing the histograms only.
KOLMOGOROV-SMIRNOV TEST
The Kolmogorov-Smirnov test examines how well a distribution from a
scale variable matches a probability distribution. In SPSS’s menu struc-
ture for this nonparametric procedure, you can test the variable against
four different types of distribution —normal, uniform, exponential,
and Poisson. The normal distribution is most often selected because
of the applications of the normal distribution to parametric assump-
tions. On the parametric side of SPSS’s menu structure, the option to
use the Kolmogorov-Smirnov test exists only for the normal distribution
(Normality plots with tests option within the Explore: Plots window).
You can consider the Kolmogorov-Smirnov test a type of goodness-of-fit
test, because it examines how well the data conform to the selected dis-
tribution. For information on distributions, see http://pages.stern.nyu.
edu/~adamodar/New_Home_Page/StatFile/statdistns.htm. (Be sure not
to include the concluding period in the web link.) The kind of variable
used in a Kolmogorov-Smirnov test is a scale variable, a continuous vari-
able representing responses along a range of potential values —e.g., Age,
Burnout level, etc. If you want to examine a Likert-like variable to see if it
follows a normal distribution, you will need to redefine its level of meas-
urement to scale before attempting to perform the test.
44 Data Analysis with Small Samples and Non-Normal Data
SPSS Process
To change the level of measurement for a variable
□ Select => Variable View tab at the bottom of the IBM SPSS
Statistics Data Editor window.
□ Find the Measure column and the row for Loneliness, and then
Note: Another method for moving variables from one side of the win-
dow to the other side is to point your mouse at the variable name and
double-click. SPSS interprets the double-click as a message to move the
variable to opposite area.
Now that SPSS knows what variable to use,
Note: The decision to use the default is based on the details in Research
Question #3, but in other situations the default may not be appropriate. For
example, if the desire was to investigate if a Likert-like variable can be from
a normal distribution with a mean of 5 and a standard deviation of 1, then
changing the Distribution Parameters for the Normal option is accom-
plished by clicking on Custom and entering the Mean and Std Dev. values.
SPSS Output
Shortly after selecting Run, SPSS’s output window will display and
show the Hypothesis Test Summary results for the analysis. Because all
four distributions were selected to test against the distribution for the
Loneliness variable, four results are presented in the summary.
Findings
According to the Kolmogorov-Smirnov test, there is sufficient informa-
tion to conclude that the data for the Loneliness variable do not come from
a normal distribution. A review of the graph clearly shows a highly right-
skewed distribution for Loneliness. Recall, one of the challenges of using
nonparametric procedures is lower power in rejecting the null hypothesis,
but in this case, it does not seem to be even close with a Sig. value less than
.001. Even so, the decision to use an ordinal variable as continuous (i.e.,
scale) should not be based on a single statistical test alone.
Clicking on the other test summaries on the left side of the window (i.e.,
clicking within the Hypothesis Test Summary table) shows the correspond-
ing information for each test. A review of the histograms and the hypoth-
esized distribution shows the level of conformity of the Loneliness data
Analyzing Single Variables and Single Groups 47
Research Question #4
Is the median score 15 for these respondents’ ability to cope with stressful events?
When examining small samples, the use of a median value is often pre-
ferred over the use of a mean value because a single response can change
the mean score dramatically in a small sample (i.e., an extreme score can
overly influence a mean calculation). When outliers are present, using
a median value is preferred because this removes the outliers’ excessive
influence on the analysis. However, if no outliers are present, the mean is
the preferred choice.
The parametric t-test is similar to the Wilcoxon signed-rank test,
except that the t-test uses a specified mean value rather than a median
value. The kind of variable used in a Wilcoxon signed-rank test is a scale/
continuous variable representing responses along a range of potential val-
ues —e.g., Age, Burnout level —based on a summed scale score from
multiple items in the measure, etc. The Wilcoxon signed-rank test is
the test to replace the parametric t-test when you cannot verify the nor-
mality assumption. However, like many nonparametric procedures, the
Wilcoxon signed-rank test has less power to find statistical significance
than the t-test when the data meet the normality assumption for the t-test.
Other uses for the Wilcoxon signed-rank test are as follows:
For Research Question #4, the continuous variable from the data set is
Coping_Level. The Coping_Level variable in the data set represents a
scale score from a validated measure that examines coping levels with
stressful events. Twenty-nine youths completed the validated measure,
and their responses were used to set the values for Coping_Level.
SPSS Process
To begin your analysis
list of tests.
□ Click the checkbox for the Wilcoxon signed-rank test.
□ Click in the Hypothesized median: area to place focus on the
Unlike the other test examples, the Wilcoxon signed-rank test does not
have any options —it requires only the identification of the hypothesized
median value. The level of measurement icon to the right indicates that
this test is only for interval or scale variables. Therefore, if the variable
is nominal, testing for a hypothesized median value is not appropriate.
SPSS Output
To see the details of the test, double-click the Hypothesis Test Summary
information (i.e., point the mouse within the Hypothesis Test Summary
table and double-click). The Model Viewer window will open to display
the detailed information.
Findings
The histogram displayed on the right side of the Model Viewer
window shows the relative distribution of the Coping_L evel variable,
50 Data Analysis with Small Samples and Non-Normal Data
RUNS TEST
The runs test, also called the Wald-Wolfowitz test, provides an opportu-
nity to see if the values in a variable are random. The runs test in SPSS
is constructed around a dichotomous variable, but SPSS can examine
other levels of measurement variables for randomness by identifying two
groups within the data. Randomness associated with testing for runs is
the hypothesis that the order in which the two values in a dichotomous
variable appear does not show any kind of pattern (i.e., the two values in
the dichotomous variable appear at random). In statistics, many different
Analyzing Single Variables and Single Groups 51
For Research Question #5, the variable from the dataset is Rater_
score.
Part 1 of 2
Because an investigation for randomness may be necessary with any
level of variable (i.e., nominal, ordinal, or scale), a few additional steps
are sometimes required before actually running the tests. For exam-
ple, because SPSS defines Rater_score as a scale variable, you must
first identify a cut point —the value used to demarcate the two sides
on which the variable values fall. The cut point, sometimes called a
threshold, will be used to identify how many times the variable values
switch from above to below (or vice versa) the cut point value, which
is the method used by the Runs test to determine if the values are not
necessarily random.
One way to identify the cut point is to run a descriptive analysis on
Rater_score.
The analysis reveals that the mean = 16.45 and the standard devia-
tion = 9.482, so you must next decide if 16.45 is a practically reason-
able threshold for this measure before using it for your analysis. Using
the mean value, as you did in this procedure, is a statistical option for
determining a cut point value. However, other options are available to
you. For example, you can utilize a cut point commonly found in the
literature (e.g., a cut point of 16 for the CESD depression scale). Another
option is to review the variable’s distribution to see if you can identify
visually a cut point value. When you have identified your cut point/
threshold, you may now conduct the Runs test.
SPSS Process
After you identify your cut point, continue your analysis:
When the Runs Test Options window opens, you will notice that the
default for scale variables is to use the Sample median option. Because
the descriptive information was obtained for Rater_score and the mean
value was identified so it could be used in the test, you will need to make
a change.
Analyzing Single Variables and Single Groups 53
□ Click to select Sample mean option in the Define Cut Pont for
Continuous Fields area.
Note: If a cut point other than the mean value is more appropriate for
the situation being tested, selecting the Custom option provides access
to the Cut point area, where a specific value can be entered.
the analysis.
SPSS Output
Shortly after you select => Run, SPSS’s output window will display and
show the Hypothesis Test Summary results for the analysis. To see the
details of the test, double-click the Hypothesis Test Summary informa-
tion (i.e., point the mouse within the Hypothesis Test Summary table and
double-click). The Model Viewer window will open to display the detailed
information.
Findings for Part 1 of 2
The graph displayed on the right side of the Model Viewer window shows
a normal distribution. The distribution represents the probability of
observing the calculated number of runs based on the expected number
of runs given the selected cut point. The red line on the graph shows the
calculated number of runs and the vertical dashed line shows the expected
number of runs. Therefore, the farther away the red line is from the dashed
line, the less likely the observations for Rater_score are random.
The Runs test for Rater_score, using the sample mean as the cut point,
suggests that you do not have enough information to reject the possibility
that Rater_score values are random. In statistical terms, the Runs test of the
null hypothesis that the Rater_score values are random, based on the sample
mean of 16.448, fails to reject the null hypothesis at α = .05 (p = .577 in this
example).
Practically, patterns in values are sometimes difficult to recognize,
especially when the sample size is relatively small. In this situation, the
Runs test did not have enough evidence to suggest a pattern. If a pattern
does not exist, the natural conclusion is that the values are possibly ran-
dom, which is the case for Part 1. However, do keep in mind that another
explanation is that the small sample size simply did not have enough
information to recognize a pattern. Remember, use all the information
available (e.g., data distribution, graphs, categorical coverage, etc.) to
draw a conclusion for your analysis.
youths reporting to have a confidante is not random — for example, the first 15
report “no” and the next 15 report “yes” — then she will need to review other
aspects of her study (e.g., how she selected respondents, when the respondents
filled out their questionnaires) to learn the potential influences causing the
nonrandom respondent order.
Part 2 of 2
Unlike investigating randomness with a scale variable, the runs test on a
dichotomous variable does not require the selection of a cut point. With
the dichotomous variables’ two response options (e.g., yes or no, positive or
negative), the runs test will examine the number of times the values switch
from one to another. For Part 2 of Research Question #5, the dichotomous
variable to investigate is Do you have a confidante? (1-yes, 0-no).
To begin this part of your analysis,
□ When you have carefully reviewed your choices, select => Run
to begin the analysis
SPSS Output
Shortly after you select => Run, SPSS’s output window will display and
show the Hypothesis Test Summary results for the analysis. To see the
details of the test, double-click the Hypothesis Test Summary informa-
tion (i.e., point the mouse within the Hypothesis Test Summary table
and double-click). The Model Viewer window will open to display the
detailed information.
Note: If you left the Rater_score variable from Part 1 in the Fields:
area when Do you have a confidante? was added, the Hypothesis Test
Summary will present information on two analyses, one for Rater_score
and one for Do you have a confidante? Simply click on the second test
(i.e., the second line in the Hypothesis Test Summary area) to display
the results of the runs test for the dichotomous variable.
Findings for Part 2 of 2
For Confidante, the red line in the graph on the right is farther away
from the dashed line than it was for Rater_score, but the distance
is still not far enough to suggest that Do you have a confidante? is
not random. Clicking between the two tests in the Hypothesis Test
Summary area, if you tested both Rater_score and Do you have a
confidante? at the same time, will show the differences in distance
between the two lines, which is reflected in the Sig. values. For Rater_
score, Sig. = .577, and for Do you have a confidante? Sig. = .164.
Therefore, once again, you do not have enough information to reject
the possibility that youths who have a confidante is random (i.e., you
have no information to question Dr. Chonody’s assumption of youth
randomness).
Practically, it is not a good idea to assume something like ran-
domness and then not check to see if it is true. Part 1 was clear in that
there was not enough information to suggest a pattern in rater scores.
In Part 2, a variable somewhat related to the concern of patterns in
rater scores was tested for randomness. Neither variable indicated a
Analyzing Single Variables and Single Groups 57
two variables are, using two almost identical variables will bias the findings
and invalidate analysis results. Generally, examining bivariate and univari-
ate scatter plots for outliers and linearity is always a good practice when first
examining the data. Another reason to run correlations on all the variables
is to help validate the data, which provides another opportunity to find data-
entry mistakes or to catch an error in how a variable was operationalized.
For Research Question #6, the variables from the data set are Age
and Self_esteem. Because Dr. Chonody is interested in learning more
about the relationship between Age and Self_esteem, and whether this
relationship differs for youth who have come out, a good way to begin is
to examine the information contained in the two variables. You can do
this by having SPSS create histograms of their distributions.
SPSS Process
To request histograms,
side of the window, then click the move arrow to move the
variable to the Variable(s): area.
□ Click to select Self_esteem in the list of available variables on
the left side of the window, then click the move arrow to
move the variable to the Variable(s): area.
□ Select => Charts… to open the Frequencies: Charts window.
□ Click to select the Histograms: option and then click to select
window.
□ Select => OK button at the bottom to close the Frequencies
Shortly after you select OK, SPSS’s output window will display a few
tables, and if you scroll down the window you will see the two histo-
grams for Age and Self_esteem. These two histograms show the variable
distributions for all the youths who completed the follow-up question-
naire. However, along with these two graphs, you want histograms for
60 Data Analysis with Small Samples and Non-Normal Data
the two groups of youths. To do this, you need to set a data filter in SPSS.
First filter to get only youths who have not come out to anyone, request
histograms, and then filter to get youths who have come out to one or
more people.
To filter respondents to select only the youths who have not come out
to anyone, and ignoring youths who have come out to one or more people,
□ Select => Data => Select cases… to open the Select Cases
window.
□ Click to select the If condition is satisfied option in the
Select area.
□ Select => If… button to open the Select Cases: If window.
□ Click to select the Coming_Out variable on the left side of the
You can verify that the filter is working by looking at SPSS’s cases view
datasheet. You can identify the cases that have been filtered by the
diagonal line that appears across some of the line numbers on the left
side of the datasheet. Another way to verify that the filter is working as
expected is to review the value for N (i.e., number of cases used in the
analysis) on the notes in the histogram.
Now that the youths who have come out to one or more people have
not been included, you can rerun the steps for requesting histograms
(see “To request histograms”). After you have obtained the histograms
for your first group of youths, you need to change the SPSS data filter to
select the other group of youths.
To filter respondents to select youths who have come out to one or
more people, ignoring youths who have not come out to anyone,
Analyzing Single Variables and Single Groups 61
□ Select => Data => Select cases… to open the Select Cases
window.
□ Click to select the If condition is satisfied option in the
Select area.
□ Select => If… button to open the Select Cases: If window.
The white area should still have the previous filter command (i.e.,
Coming_Out = 1), but, if not,
Next, to filter youths who have come out to one or a few people (Coming_
Out values of 2) and youths who have come out to everyone (Coming_
Out values of 3), you need to use a “not equal to” symbol instead of the
equal sign. Therefore, you want the filter command in the white area
that will tell SPSS to use all the cases that have something other than a
1 for Coming_Out (i.e., Coming_Out ~=1, where ~= is the same as ≠).
□ Once you set the filter command to use all the cases that do not
have a 1 for Coming_Out, select => Continue button to close
the Select Cases: If window.
□ Select => OK button to close the Select Cases window and filter
Now that your filter is set, you can again rerun the request to get his-
tograms for the second group of youths by following the same steps as
before (i.e., steps under “To request histograms”).
SPSS Output
Figures 2.4, 2.5, and 2.6, show the histograms produced by completing
the steps above (i.e., Age and Self_esteem histograms for all the youths,
youths who have not come out to anyone, and youths who have come
out to one or more people). A review of the values for N in each of the
62 Data Analysis with Small Samples and Non-Normal Data
Age Self_esteem
10 Mean = 13.96 6 Mean = 14.25
Std. Dev. = 2.487 Std. Dev. = 3.855
N = 28 5 N = 28
8
4
Frequency
Frequency
6
3
4
2
2
1
0 0
10 12 14 16 18 20 5 10 15 20 25 30
Age Self_esteem
Age Self_esteem
4 Mean = 13.67 2.0 Mean = 15.67
Std. Dev. = 2.598 Std. Dev. = 3
N=9 N=9
3 1.5
Frequency
Frequency
2 1.0
1 0.5
0 0.0
10 12 14 16 18 20 10 12 14 16 18 20 22
Age Self_esteem
Figure 2.5. Age and Self_esteem histograms for youths who have not come out
to anyone.
Analyzing Single Variables and Single Groups 63
Age Self_esteem
5 Mean = 14.11 4 Mean = 13.58
Std. Dev. = 2.492 Std. Dev. = 4.1
N = 19 N = 19
4
3
3
Frequency
Frequency
2
2
1
1
0 0
10 12 14 16 18 20 5 10 15 20 25 30
Age Self_esteem
Figure 2.6. Age and Self_esteem histograms for youths who have come out to
one or more people.
anyone are reviewed (i.e., Figure 2.5), the distribution appears to follow
a uniform distribution.
Another reason a close review of the data using histograms is
important is that a single statistical test alone is not enough to sup-
port a strong conclusion. For example, if the above distributions for
all respondents were tested for normality (don’t forget to remove the
filter) using the Kolmogorov-Smirnov test (see Research Question #3),
both normal and Poisson distributions are possible for Self_esteem.
The reality is that, when dealing with small samples, the amount of
available information to reject a hypothesis is small, making it more
of a challenge to find significance (i.e., more difficult to reject a null
hypothesis). Therefore, a researcher must use different approaches
to draw conclusions about analysis findings as well as information
to validate assumptions. Because the histograms for Age are non-
normal, the procedures to answer Research Question #6 must be
nonparametric.
Note: When using filters, you must pay extra attention to make
sure your filters are working properly before conducting any analysis
or requesting graphs. SPSS has no way to warn you if your filter is not
working as you intended, so you must look at what cases are and are not
filtered out. In addition, make sure you remove any filters when you are
finished using subsets of your dataset. Forgetting to remove a filter can
64 Data Analysis with Small Samples and Non-Normal Data
be very confusing if you move forward with other analyses and cannot
understand the unexpected analysis results!
SPSS Process
If you still have an active filter from requesting the histograms above,
□ Select => Data => Select cases… to open the Select Cases
window.
□ Click to select All cases option in the Select area.
□ Select => OK button to close the Select Cases window and
Note that Spearman’s rho and Kendall’s tau-b are not found in the
Nonparametric Tests menu structure.
After the Bivariate Correlations window opens,
□ Click to select Age in the list of variables on the left side of the
window, then click the move arrow to move the variable to
the Variables: area.
□ Click to select Self_esteem in the list of variables on the left
SPSS Output
Shortly after you click OK, SPSS’s output window will display and show
the Correlations results for the analyses (see Table 2.3).
Findings
Table 2.3 shows that both Spearman’s rho and Kendall’s tau-b meas-
ure a negative relationship between Age and Self_esteem. Note that the
Kendall’s tau-b finds a significant correlation (p = .042) between the two
variables, while the Spearman’s rho fails to find significance (p = .061)
at α = .05. You can attribute the difference between the two tests to their
differences in dealing with ties within the ranking process. To verify
that you have ties in the data, you can sort the data by Age and look for
multiple occurrences of Age and Self_esteem combinations.
□ Select => Data => Sort cases… to open the Sort Cases window.
□ Click to select Age in the list of variables on the left side of the
window, then click the move arrow to move the variable to
the Sort by: area.
□ Select => OK button to close the Sort Cases window.
Note: When planning to sort cases, it is a good practice to have one var-
iable that will allow you to return the cases to their original order. You
can accomplish this by adding a variable to the dataset that assigns an
ID to the cases, counting from 1 through k. The variable in the data set
for Chapter 2 that identifies the original order is ID.
After you sort the data by Age, a review of the values for Age and
Self_esteem shows multiple occasions in which cases have the same
Age and Self_esteem values —suggesting that the tests managed mul-
tiple ties. Because Kendall’s tau-b handles ties more efficiently than
Spearman’s rho, it makes logical sense that Kendall’s tau-b calculated a
smaller p-value than Spearman’s rho. Again, you should always pursue
multiple ways to verify your findings and the performance of your test
so that you can have confidence in your conclusions.
However, in Research Question #6, Dr. Chonody is interested in
more than just the relationship between Age and Self_esteem for all the
youths. She is interested in the possible differences in the relationships
between youths who have not come out and youths who have. Therefore,
you must run the correlation analysis two more times, once after filter-
ing the cases so that youths who have come out are excluded, and then
again after the youths who have not come out are excluded (i.e., filter on
the variable Coming_Out).
Using the steps discussed earlier to filter out certain cases and the
steps to conduct the Spearman’s rho and Kendall’s tau-b tests, you
Analyzing Single Variables and Single Groups 67
collect the Sig. values for the multiple tests into a single table for review
(see Table 2.4).
Although the negative correlations still exist when you divide the
youths into two groups, youths who have come out and youths who
have not, the Sig. values in Table 2.4 show that the negative relation-
ship between Age and Self_esteem is being heavily influenced by the
youths who have come out (i.e., Sig. values are significant at .032 and
.043). In addition, Table 2.4 shows that the relationship between Age
and Self_esteem is not significant for youths who have not come out,
and the negative relationship between Age and Self_esteem is moderate
(i.e., correlation value is between -.3 and -.6) for youths who have come
out. Therefore, the Sig. values for youths who have come out to anyone
suggest that there is only a 3% to 4% chance that there is no relationship
between Age and Self_esteem.
A more statistical explanation of the results of the correlation anal-
ysis is that a test of the relationship between Age and Self_esteem using
Kendall’s tau-b is -.384 and found to be significant (p = .032) for youths
who have come out to one or more people (n = 19).
Practically, if the gay youths in Dr. Chonody’s study are typical of
all gay youths, the findings suggest that self-esteem is not necessarily
linked to age for youths who have not come out. However, for youth
who have come out to anyone, the findings suggest that their self-
esteem will decrease as they get older. This finding should concern you
and push you to conduct a closer examination of, among other things,
your data, your sample, how the data were collected, and the wording
of the questions provided to the respondents. The reason for the closer
examination is that these findings seem to be in conflict of what you
CHAPTER OBJECTIVES
This chapter covers analyses that compare two or more independent
groups, and the topics presented are as follows:
69
70 Data Analysis with Small Samples and Non-Normal Data
INTRODUCTION
Recall that Chapter 1 describes the importance of understanding the
data and individual variables used in statistical tests, and Chapter 2
explains how to examine closely the individual variables and how to
conduct nonparametric tests that involve variables for a single group.
This includes discussion of how to inspect individual variables to gain
a deeper understanding of exactly how much information the variable
contributes to its conceptual construct and characteristics. Most impor-
tantly, a deeper understanding of the individual variable will help you
select variables for other purposes, such as analyses that involve two or
more groups of respondents.
Chapter 3 explores independent groups and how to use nonpara-
metric statistics that are specific to group comparisons. The term inde-
pendent groups refers to the situations in which a response from an
individual in one group is not influenced by a response from an indi-
vidual in another group. In comparison, related groups, discussed in
Chapter 4, involve responses that have a connection (e.g., examining the
difference between two responses from the same individual, so that the
individual is the connection).
Many of the data concerns from previous chapters remain when
moving from single to multiple groups, but the concerns are more pro-
nounced in this chapter because the information available from the
data is now being shared among a number of groups. This becomes
more of a challenge when dealing with the limited information avail-
able in a small sample. The limited information shared among the
groups makes meeting parametric analysis assumptions much more
difficult, if not impossible. Thus, knowing how to use nonparamet-
ric procedures becomes critical to good research. Becoming familiar
with nonparametric procedures for groups will help avoid the poten-
tial mistake of reporting invalid results due to running a paramet-
ric procedure on data that do not meet the necessary assumptions. In
addition, verifying that the assumptions are met is more challenging
when the data are broken up into even more pieces when involving
even more groups.
Research Questions 7 to 11 illustrate situations that involve com-
paring independent groups and the manner in which data are analyzed.
Comparing Two or More Independent Groups 71
After each research question, detailed steps are provided for how to
perform an analysis, issues to be aware of when conducting the anal-
ysis, and how to interpret test results. Simply stated, Chapter 3 covers
nonparametric statistics for independent group comparison, includ-
ing the researchers’ responsibility to understand the variables individ-
ually as well as how variables are used to investigate potential group
differences.
SPSS Process
To begin your analysis,
□ Select => Fields tab to see the list of available variables for the
analyses.
Comparing Two or More Independent Groups 75
specific tests.
□ Click the check boxes next to Moses extreme reaction.
□ Click the check box next to Hodges-L ehmann estimate.
For Research Question #7, the only test option to consider is the Moses
extreme reaction for identifying outliers. Use the default option (i.e.,
Compute outliers from sample) to use the observed data.
□ When you have reviewed your choices carefully, select => Run
to conduct the analysis.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
Findings
The information for the Moses extreme reaction test shows that the
two groups have similar values for percentage of time using a con-
dom (i.e., Sig. = .665 is larger than the α = .05 significance level). The
Moses extreme reaction test provides an exact significance (Sig. is the
notation SPSS uses to represent statistical significance), indicating
that there is a 66.5% chance that the ranges are the same. By tradi-
tion, statistical significance must be small, with only a 5% or smaller
chance that the medians are similar, before concluding that the groups
are different.
Note: Not all statistical tests provide exact significance values.
Therefore, be sure to translate a Sig. value to “percent chance” interpre-
tation only when the test allows this.
To see the Hodges-Lehmann results, you must change the view. At
the bottom left of the Model Viewer window, one of the dropdown lists
is named View:.
After you select the Confidence Interval Summary View, the informa-
tion for the Confidence Interval Summary is displayed. It reveals a
95% confidence interval for the difference between median values range
from -.20 to .10. This range includes a plausible difference of zero and,
therefore, is in line with the conclusion drawn by the Moses extreme
reaction test.
In addition, a box-a nd-whiskers plot associated with the test is
included with the other analysis results. Box-a nd-whiskers plots are
a graphical representation of a distribution of values. The box part
Comparing Two or More Independent Groups 77
of the graph represents the first and third quartiles, called the inter-
quartile range or IQR, and the thicker dark line represents the
median value. The lines extending from the box are called whiskers
and are calculated differently by different applications. SPSS draws
the end of each whisker as the minimum and maximum values in
the data that are not identified as statistical outliers. SPSS identifies
outliers as values that fall farther away from the first and third quar-
tiles than 1.5 times the size of the IQR. SPSS identifies outliers with
small circles and with asterisks when outliers are three times beyond
the IQR. Figure 3.1 does not identify the presence of any outliers, but
Figure 3.5 does. The similarity of the vertical positions of these areas
for the two groups illustrates and supports the values for the confi-
dence interval and the hypothesis test finding of similar range values
(see Figure 3.1).
In summary, the Moses extreme reaction test did not find a signif-
icant difference in the range of percentage of condom use between the
case and control groups. In addition, the Hodges-Lehmann confidence
interval for plausible median difference identified a lower and upper
bound of −.20 and .10, respectively. Therefore, the two analyses support
the conclusion that there is not enough information to suggest that the
median value or range of values for condom use is different.
1.00
0.80
Percent_Pre
0.60
0.40
0.20
0.00
Received MI Did not Receive MI
Received_MI
KOLMOGOROV-SMIRNOV TEST
The Kolmogorov-Smirnov test calculates the amount of separation of
the empirical distributions for the two groups, and it uses the amount
of separation to decide if the two distributions are significantly dif-
ferent. Said differently, it tests to find out if it is possible that the val-
ues for the two groups come from the same distribution. Because the
focus is on the separation of the two distributions, no prior assump-
tion about the distribution exists, unlike the t-test, in which the data
are assumed to come from a normal distribution. Recall that when the
data clearly meet all parametric assumptions, the parametric proce-
dure does have more power to find significance than its nonparamet-
ric counterpart (emphasis on the word clearly). Importantly for the
Kolmogorov-Smirnov test, not having a normality assumption does
make the test less able to find significant differences between distribu-
tions than the parametric t-test when the data clearly meet the t-test’s
normality assumption. In addition, SPSS restricts the Kolmogorov-
Smirnov test to two groups.
Other examples for uses of the Kolmogorov-Smirnov test are as
follows:
SPSS Process
To begin your analysis,
□ Select => Fields tab to see the list of available variables for the
analyses.
□ Click to select the Use custom field assignments option.
□ Click to select the grouping variable Received_MI.
□ Click the move arrow closest to the Groups: area to move the
variable to that area.
□ Click to select the scale variable (Percent_DIF) that contains
specific tests.
□ Click the check box to the left side of Kolmogorov-Smirnov
You may run into situations in which you want to change the signifi-
cance level for your test. In addition, some of the tests within this list of
nonparametric procedures can estimate confidence intervals. The steps
below show how to change the options associated with significance level
and confidence intervals.
Comparing Two or More Independent Groups 81
test-by-test).
□ After reviewing your choices carefully, select => Run to conduct
the analysis.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
Findings
The Kolmogorov-Smirnov test shows Sig. = .056, which is very close to
the Significance level: set in the Test Options from the Select an item:
area. In the Hypothesis Test Summary information, SPSS is not indi-
cating that this test rejects the null hypothesis (i.e., the null hypothesis is
that distributions between the two groups are the same). We discuss in
Chapter 2 how nonparametric procedures are more conservative when
finding test significance. Therefore, a case can be made that this test
is close enough to rejecting the null hypothesis to report a difference
82 Data Analysis with Small Samples and Non-Normal Data
MANN-WHITNEY U TEST
The Mann-Whitney U test, also called the Wilcoxon rank sum test, uses
ranking of the observed values to determine if the two groups come from
the same distribution. Unlike the parametric equivalent independent t-test
that assumes a normal distribution, the Mann-Whitney has no distribu-
tion assumption. The only restrictions in SPSS are that the test is for only
two groups and the comparison variable is on a scale level of measurement.
The lack of any distribution assumption does make the Mann-Whitney
U appropriate for any type of distribution, but, as with the Kolmogorov-
Smirnov and Wald-Wolfowitz tests, the flexibility does make finding sig-
nificant differences between the distributions a bit more difficult.
WALD-WOLFOWITZ TEST
The Wald-Wolfowitz test analyzes runs for the values while focusing
on group membership. The test’s structure provides an opportunity to
test the hypothesis that values for a scale variable from two groups have
similar value distributions. The test is restricted to two groups for com-
paring the distribution of a scale variable. Because the Wald-Wolfowitz
uses runs, no assumption about the distribution is required. However,
when a sample size is greater than 30, SPSS uses a normal approxima-
tion for a test statistic because as the number of possible runs reaches
30+, the distribution for the number of possible runs is similar to a
Comparing Two or More Independent Groups 83
SPSS Process
To begin your analysis using all three tests,
□ Select => Fields tab to see the list of available variables for the
analyses.
□ Click to select the Use custom fields assignments option.
□ Click the move arrow closest to the Test Fields: area to move
the variable to that area.
specific tests.
Figure 3.2 shows many options for nonparametric tests, some for two
groups and some for more than two groups. Looking closely at Figure
3.2 will show five tests specific to having only two groups.
□ Click to select the three check boxes on the left side of the
Compare Distribution across Groups list of tests —Mann-
Whitney U, Kolmogorov-Smirnov, and Wald-Wolfowitz.
□ Select => Test Options in the Select an item: area to see the
place where to change default settings.
□ Decide on the significance level (default is 0.05) for the test.
□ Decide on the confidence interval range (default is 95.0%).
□ Decide how to handle missing data (default is Exclude cases
test-by-test).
□ After reviewing your choices carefully, select => Run to conduct
the analysis.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
window, you change the information on the right side to that particular
test information.
Another way to switch between test information is to change the
selection from the dropdown lists on the left for View and Field Filter,
and on the right for Test, Field(s), and View. Figure 3.2 shows the
information for the Kolmogorov-Smirnov test because that test is the
one highlighted in the list of tests on the right.
Findings
A review of the Sig. values finds that the Wald-Wolfowitz runs test is
clearly not rejecting the null hypothesis (p = .994). The Mann-W hitley
U test (p = .059) and the Kolmogorov-Smirnov test (p = .056) are similar
in their findings. In other words, the two are close to rejecting the null
hypothesis at α = .05 that the groups are similar in their Percent_DIF
values. Differing values for these three tests is common, and this situa-
tion demonstrates how important it is to review all three. To complicate
matters even further, if you conduct an independent-samples t-test on
the data, the parametric test will return a Sig. value of .012 —leading
you to conclude that the two groups are significantly different for values
of Percent_DIF.
Comparing Two or More Independent Groups 87
When faced with conflicting information, begin with what you know
about the tests. You know that the nonparametric tests are less likely
to find significance than parametric tests when parametric assump-
tions are met. You also know that one guideline for deciding if you
have enough information about a group is 30 per group, although small
group sizes can typically detect only large effect sizes. An a priori power
analysis is another method for understanding the strengths of your
data, but the discussion on how to conduct a power analysis is left for
other resources. For Research Question #8, you again find yourself with
a marginal group size (i.e., two groups of 30). You know also that when
the data for each group resembles normality, the independent-sample
t-test may be appropriate (i.e., after looking for outliers, missingness,
data value gaps, etc.). Figure 3.4 shows histograms of the distributions
for the two groups by filtering on the variable Received_MI. Recall that
the process on how to filter data is discussed in Chapter 2, and it is dis-
cussed later in this chapter.
Neither of the two histograms shows a large departure from normal-
ity, providing additional support for the findings of the independent-
samples t-test. In addition, the two nonparametric tests related to
distribution comparisons were close to being significant, with Sig.
values of .059 and .056. Thus, Dr. Rutledge has enough information to
cautiously reject the null hypothesis and conclude that the MI inter-
vention significantly changed the percentage of unsafe sexual behav-
ior over a two-month period. However, when cautiously rejecting a null
Frequency
4
2
2 1
0 0
–0.40 –0.20 0.00 0.20 0.40 0.60 –.50 –.25 .00 .25 .50
Percent_DIF Percent_DIF
when sample sizes are small. The Kruskal-Wallis one-way ANOVA test
does assume independent groups but has no normality assumption.
This distribution-free characteristic makes the test applicable to many
data analyses and it is best suited for testing data that represent groups
that are nominal, conceptually (i.e., the compared groups do not have a
natural ordering). As in other nonparametric tests, the distribution-free
characteristic comes with the cost of being more conservative, or less
likely to find significant group differences. The Kruskal-Wallis one-way
ANOVA test uses runs of values from two, three, or more independ-
ent groups to determine if the groups come from the same distribu-
tion. Of course, as the number of groups increases, so does the required
size of the sample. The test pools the values from the groups to see how
often the values switch from one group to the other to determine group
differences.
Other examples for use of the Kruskal-Wallis one-way ANOVA test
are as follows:
SPSS Process
To filter the data so that only respondents who received the MI interven-
tion are analyzed,
□ Select => Data => Select cases… to open the Select Cases
window.
□ Click to select the If condition is satisfied option in the
Select area.
□ Select => If… button to open the Select Cases: If window.
□ Click to select the Received_MI variable on the left side of the
□ Select => Fields tab to see the list of available variables for the
analyses.
□ Click to select the Use custom field assignments option.
is in the Test Fields: area, click to select the variable and then click the
move arrow to move it back to the Fields: area.
specific tests.
Figure 3.2 shows many options of nonparametric tests, some for two
groups, and some for three or more groups, frequently referred to as k
samples in published literature. A close review of Figure 3.2 shows three
tests specific to having three or more groups.
Before running the test, you should consider the test option. For the
Kruskal-Wallis test, the dropdown list for Multiple comparisons: should
be set to All pairwise to compare the distributions of all the possible
pairing of groups in the data. When the number of groups is too large,
the comparison of all possible pairwise combinations becomes cumber-
some, and the Stepwise step-down option is helpful. This option sorts
the groups by their calculated group median value and compares only
adjacent groups, making the comparison of groups possible. However,
92 Data Analysis with Small Samples and Non-Normal Data
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
Findings
Figure 3.5 shows the results of the analyses for Research Question #9. The
box-and-whiskers plot in Figure 3.5 shows that the median values among
the three groups are very similar, but the group variability appears differ-
ent. Specifically, the respondents who are exclusively with opposite-sex
partners fall much closely around the median value, while the other two
groups spread out more (i.e., whiskers indicate a wider range of values for
exclusively same-sex and both same-sex and opposite-sex partner groups
than for the exclusively opposite-sex partner group).
Comparing Two or More Independent Groups 93
Even though the box-a nd-whiskers plot in Figure 3.5 shows only
a slight difference in Percent_Post median values along the horizon-
tal axis for the three Partner groups, you still need to verify observed
difference. Another way to examine the difference in median values is
to review the calculated mean ranks for the three groups. To view the
mean rank values, hover the mouse pointer over the black line in the
box-a nd-whiskers plots. Hovering will cause a Mean Rank value to
appear (14.25 for exclusively same-sex, 15.27 for exclusively opposite
sex, and 17.43 for both).
The Kruskal-Wallis test did not find significant distribution differ-
ences among the three groups. The Kruskal-Wallis Sig. value = .775,
larger than the traditionally used threshold value of α = .05. This non-
significant finding is graphically supported by the amount of similar
vertical placement in the box-and-whiskers plots.
In statistical terms, the Kruskal-Wallis test for distribution compar-
ison (p = .775) failed to reject the null hypothesis of equal distributions
at α = .05.
Practically, this means that the evidence suggests that the influence
of an MI intervention is similar among groups of people who are exclu-
sively with same-sex partners, exclusively with opposite-sex partners,
and those who have both same and opposite-sex partners. Of course,
94 Data Analysis with Small Samples and Non-Normal Data
another explanation is that the small sample simply does not pro-
vide enough information to identify differences. Therefore, adding
the information from Research Question #8, we can infer that there is
not enough information to conclude that the intervention benefits one
group over the other.
JONCKHEERE-TERPSTRA TEST
Like the Kruskal-Wallis test, the Jonckheere-Terpstra test uses runs to
determine if the values from the groups come from the same distribution.
However, the alternative hypothesis is that the distributions are somehow
Comparing Two or More Independent Groups 95
SPSS Process
Similar to the steps for Research Question #9, Research Question #10
involves analyzing only the data that represent those who were part
96 Data Analysis with Small Samples and Non-Normal Data
of the treatment group. Therefore, the data filter used earlier on the
Received_MI variable needs to be active before beginning the analysis.
If you do not recall how to set the filter, see the first few steps under the
SPSS Process for Research Question #9.
To begin your analysis,
□ Select => Fields tab to see the list of available variables for the
analyses.
□ Click to select the Use custom field assignments option.
□ Click to select Stage_of_Change variable in the list of variables
in the Fields: area.
□ Click the move arrow closest to the Groups: area to move the
variable to that area.
□ Click to select Percent_DIF scale variable that contains percent
specific tests.
□ Click the check box for Jonckheere-Terpstra for k samples
Before running the test, you should consider a few options. First, set the
dropdown list for Multiple comparisons: to All pairwise to compare
Comparing Two or More Independent Groups 97
the analysis.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
the dropdown lists to find detailed information about the data and the
analysis.
Findings
The results displayed in the Model Viewer window for the Jonckheere-
Terpstra test did not indicate that the distributions were significantly
different among the three groups (Sig. = .055). Because SPSS did not
find significant differences, the multiple comparisons All pairwise
option was ignored. However, a Sig. value of .055 is very close to the
α = .05 threshold. Considering the closeness of the Sig. value to .05
and your understanding that nonparametric tests tend to be more
conservative in finding significance, a .055 may be enough to proceed
further. But how?
One option is to adjust the significance level slightly so that SPSS
will provide the pairwise comparison output, which in turn will give
you more information for you to make a final determination about the
test (despite inflating the experiment-w ise Type I error rate). To adjust
the significance level value,
item: area.
□ Select => Test Options from the Select an item: area.
□ Change the value displayed for the Significance level: area from
0.05 to 0.10.
□ Select => Run button to rerun the analysis.
Now that SPSS recognizes that the Sig. value is smaller than the
Significance level option value, the pairwise comparisons are con-
ducted. Notice that although the significance level value has changed, the
Sig. value remains the same. The reason is that the decision to set the sig-
nificance level at a specific value is theoretical, and that decision affects
only the choice of rejecting or failing to reject a test’s null hypothesis. The
chance that the three groups have similar means is still the same.
Comparing Two or More Independent Groups 99
Precontemplation
11.14
Contemplation
17.60
Preparation
18.50
Each row tests the null hypothesis that the Sample 1 and Sample 2 distributions are the same.
Asymptotic significances (1-sided tests) are displayed. The significance level is .10.
reject the null hypothesis based on the established significance level. The
difference is that there is an adjustment to the value due to the increased
likelihood of finding significance just because of conducting multiple
tests. Recall, in this figure the reason that the pairwise is showing signif-
icance is that earlier you changed the significance level to .10, allowing
SPSS to conduct the pairwise comparisons.
Because the initial test was very close to rejecting the null hypothe-
sis and the review of the pairwise comparisons shows a clear difference
between Precontemplation and Preparation, there is enough evidence
to claim that respondents within this sample that are in different stages
of change do, indeed, respond differently to the MI. The stage difference
is seen in the percent change in condom use prior to and then after the
intervention, with the largest difference being between those in precon-
templation and preparation.
In statistical terms, the Jonckheere-Terpstra test (p = .055) rejects
the null hypothesis of equal distributions at α = .10. A pairwise compari-
son found adjusted p = .082 between Precontemplation and Preparation,
adjusted p = .154 between Precontemplation and Contemplation, and
adjusted p = 1.0 between Contemplation and Preparation.
Practically, this means that the evidence suggests that individuals
in the precontemplation stage of change decrease their level of unsafe
sexual behavior after experiencing an MI intervention, but less of a
decrease than that of individuals in the preparation stage. This informa-
tion makes theoretical sense, and it will greatly benefit the effort to
improve the effectiveness of an MI intervention that targets decreasing
the amount of unsafe sexual activity.
the post-intervention treatment group, and it will support the use of these
measures in the clinical trial.
Research Question #11
Do the median values for self-efficacy differ among respondents who are
unlikely to ask their partner to use condoms, those somewhat likely, and those
very likely to request condom use?
MEDIAN TEST
The median test uses the frequency in which the values for each group
are above and below a median value and then employs these counts
to decide if the group medians are significantly different. In SPSS, the
median value can be either a pooled value, in which SPSS calculates a
median value using all the values from the groups, or a specified value
identified by the person conducting the test. Unlike the parametric
t-test that assumes a normal distribution for the groups, the median test
has no assumption about the group’s distribution. The median test does
calculate a chi-square statistic based on expected vs. observed frequen-
cies, so the test will identify significant median differences when the
sample size is large. Note that “large” for three groups is a sample size
of 300+ per group. The other median test assumption is that the groups
are independent, so the values in one group do not have influence on the
values in another group.
Other examples of uses for the median test are as follows:
The variables needed for the median test are a categorical variable (i.e.,
nominal or ordinal) representing three or more groups of an inde-
pendent variable, and a scale or continuous variable representing the
dependent variable. For Research Question #11, the categorical variable
is Condom_Request and the scale variable is Self_efficacy.
102 Data Analysis with Small Samples and Non-Normal Data
SPSS Process
Similar to Research Questions #9 and #10, Research Question #11 is ana-
lyzing only the data that represent those who were part of the treatment
group. Therefore, the data filter for the Received_MI variable needs to be
active before beginning the analysis. If you do not recall how to set the fil-
ter, see the first few steps under the SPSS Process for Research Question #9.
To begin your analysis,
□ Select => Fields tab to see the list of available variables for the
analyses.
□ Click to select the Use custom field assignments option.
□ Click to select the grouping variable Request_Condom.
□ Click the move arrow closest to the Groups: area to move the
variable to that area.
□ Click to select the scale variable (Self_efficacy) that contains
values representing levels of self-efficacy for the respondents.
□ Click the move arrow closest to the Test Fields: area to move
the variable to that area.
to select specific tests.
□ Click to select the check box next to the Median test on the
Before running the test, you must consider a few options. You must
decide what median value to use when the median test is counting the
larger and smaller values. If you do not have a specific value you iden-
tified from the literature or another source to use as a cut point, then
it is best to allow SPSS to use a pooled sample median, which is the
median value calculated from all the values in the groups. Therefore,
for Research Question #11, you click to select Pooled sample median.
To test a specific median value, you would select the Custom
option that enables you to enter a specific value in the Median area.
To compare the medians of all the possible pairings of groups in the
data, set the dropdown list for Multiple comparisons: to All pairwise.
When the number of groups is too large, making the comparison of all
possible pairwise combinations cumbersome, the Stepwise step-down
option is helpful. This option sorts the groups by their calculated
group median value and compares only adjacent groups, making the
comparison of groups possible. However, with only three groups, the
comparison of all pairwise combinations is completely manageable.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
Findings
The overall median test indicates sufficient evidence to conclude that
there is a difference in median self-efficacy across likelihood of request-
ing condom use categories (Sig. value = .035). The results of the pair-
wise tests show that median self-efficacy between respondents reporting
they are unlikely to request a partner to use a condom and very likely
to request a partner to use a condom is significantly different, with
median values of 14.50 for unlikely and 18.50 for very likely categories.
somewhat likely
16.00
unlikely
14.50
very likely
18.50
Each row tests the null hypothesis that the Sample 1 and
Sample 2 distributions are the same.
Asymptotic significances (2-sided tests) are displayed. The
significance level is .05.
CHAPTER OBJECTIVES
Chapter 4 presents information about comparing two or more groups
that relate to one another. The chapter covers the following topics:
107
108 Data Analysis with Small Samples and Non-Normal Data
INTRODUCTION
Chapter 4 covers information about comparing related samples and the
way you analyze these data. Before beginning, if you are new to research,
it might be helpful to clarify the use of the word sample. In this chap-
ter, when comparing samples that are related, the term samples refers to
collecting data multiple times from the same group of people, not to be
confused with the use of the word sample to mean a subset of a target
population (i.e., a sample of the population). It is very important to rec-
ognize when the term sample relates to a single data collection effort from
one group of respondents, as described in Chapter 3, a sample from a
population, and when the term samples relates to multiple data-collection
activities from one group of respondents.
Prior to analyzing data for related samples, it is very important to
make sure your data are identified with the appropriate level of measure-
ment in SPSS and are in the appropriate format for related-samples kinds
of tests. For example, in Research Question #12, the data format requires
that responses for the two nominal variables for each respondent fall on
the same row on the data spreadsheet, as they might for pre and post
scores for the same respondent. This lets SPSS know that the pre and post
scores are related to a single person or a single respondent.
As a reminder, to avoid any unexpected surprises or worse, report a
finding not supported by the data, be thorough in your approach, and
examine the information provided by each variable individually before
moving forward with an analysis that involves multiple variables. This
includes examining categorical variables to verify that the appropriate
response coverage is present (i.e., a minimum of 5 responses for each
category), and reviewing a graphical depiction of the distribution of
scale variables to examine the theoretical/conceptual validity of the data
in the variable. Chapter 1 describes a few things to consider when exam-
ining variables individually, and Chapter 2 provides multiple analytical
procedures to assist the examination.
Comparing Two or More Related Groups 109
MCNEMAR’S TEST
McNemar’s test uses combination frequencies to determine if yes/
no kinds of responses at two different times are different. The com-
binations are constructed from responses captured by two dichot-
omous variables (i.e., yes-yes, yes-no, no-yes, and no-no). The 2 × 2
contingency table structure of McNemar’s test is similar to Pearson’s
chi-square test; however, the chi-square test is for independent sam-
ples (i.e., unpaired data) and not very effective when dealing with
Table 4.1. Summary of Analyses and Variables in Chapter 4
Research Nonparametric Variable Measure Variable Name
Question Test in DB
small sample sizes (i.e., less than 100 records). Another related test
is the paired-sample t-test, but the paired-sample t-test is only for
continuous variables that meet the normality assumption. Because
McNemar’s test uses two dichotomous variables, the null hypothesis
is that the marginal probabilities for each of the two data point out-
comes are equal. This means, for example, that the probability that a
Hispanic/L atino person will answer yes to a specific question is equal
to the probability that a non-Hispanic/L atino person will answer yes
to the same question.
Other examples of uses for McNemar’s test are as follows:
The variables needed for the McNemar’s tests for Question #12 are two
dichotomous variables representing two related responses —e.g., pre
and post values for respondents. For this question, the dichotomous
variables are victim of a violent crime and joined a gang in wave 2 (i.e.,
Violent_crime_w2 and Joined_gang_w2).
112 Data Analysis with Small Samples and Non-Normal Data
SPSS Process
To begin the analysis,
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the second variable, Joined_gang_w2, found in
the Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
Now that SPSS knows what variables to use, you need to choose the
kind of nonparametric test to conduct.
Figure 4.1 shows many different tests, some for two related observa-
tions, some for three or more related observations, some for categorical
variables, and some for scale variables. A close examination of Figure
4.1 reveals four tests that are specific to having only two related obser-
vations. Because Research Question #12 involves binary data (i.e., two
dichotomous variables), you should select McNemar’s test.
Note: When you select Choose Tests in the Select an item: area, you will
see one option that allows SPSS to decide automatically which tests are
appropriate based on the variable definitions (i.e., identified by the level
of measurement). It is prudent to monitor the analysis process closely
by always choosing Customize tests and then selecting the specific test
you want to conduct.
Before running McNemar’s test, SPSS needs to know which value in
the dichotomous variables is associated with success. SPSS uses the ter-
minology of success and failure to identify the two options in dichoto-
mous variables. You should keep track of the values identified as success
to verify the use of your data and to understand how best to interpret
the test results.
Once the window opens, two options are displayed. The default option
is to let SPSS use the First value found in data. To be fully aware of how
the data are defined and used,
□ Click on the white space below the word Value and enter a 1.
Entering a 1 tells SPSS to associate the number 1 with a success.
□ Select => OK to close the window.
Note: If the two variables had more than two outcomes that represent
success, multiple lines in the white space identify the success outcomes.
Recall, McNemar’s test is based on dichotomous events. Having the
capability to combine values expands the usability of McNemar’s test to
categorical variables with more than two values.
To continue your analysis,
the analysis.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
Model Viewer window will open to display the detailed information for
the analysis (see Figure 4.2).
The Model Viewer window has access to more than the default view
shown in Figure 4.2. At the bottom of the window are dropdown lists on
the left for View: and Field Filter: and dropdown lists on the right for
Test:, Field(s):, and View:.
Findings
The Hypothesis Test Summary information on the left side of the
Model Viewer window reveals the exact significance of the test,
p = .031. The exact significance values suggest only a 3.1% chance that
being a victim of a violent crime is unrelated to the probability of join-
ing a gang. The values in the bar charts in Figure 4.2 reveal that every
respondent who joined a gang in wave 2 was a victim of a violent crime.
Therefore, the data suggest that joining a gang is related to being a vic-
tim of a violent crime.
In statistical terms, McNemar’s test rejects the null hypothesis of
equal marginal probabilities at α = .05 (p = .031). The small sample
and use of the Binomial distribution in the statistical analysis provide
an exact significance value (p = .031). The p-value indicates that if the
null hypothesis is true, there is only a 3.1% chance of equal success
probabilities.
Practically, this means that you found, in Dr. Fleming’s sample of
adolescents, a significant relationship between adolescents’ joining a
gang and having been a victim of a violent crime. Not explained in this
analysis is if their reason for joining the gang is for protection.
Comparing Two or More Related Groups 117
MARGINAL HOMOGENEITY TEST
The marginal homogeneity test is an extension of McNemar’s test in that it
addresses two ordinal categorical variables rather than two dichotomous
variables. The number of categories can be three or more, but, as described
in Chapter 2, the more categories you have, the more responders you need
to be sure each category has enough coverage to be used in an analysis.
The marginal homogeneity test uses response combination frequencies to
determine if the number of positive outcomes has changed, and because
the test uses two categorical variables, the null hypothesis is that the mar-
ginal probabilities for each of the two data point outcomes are equal.
Other examples of uses for the marginal homogeneity test are as
follows:
For Research Question #13 (see Box 4.3), the categorical variables are
Friends in wave 2 and wave 4 (Friends_w2 and Friends_w4). These vari-
ables identify the number of the respondent’s friends who are in a gang,
with categorical response options of few, many, and most. A review of the
two categorical variables shows that coverage is just above the five responses
per category guideline. However, although the data meet the coverage
guideline, the frequency for each category is small, which does not provide
much information for identifying differences. Nevertheless, when analyz-
ing small datasets, your only option is to use the best test that matches your
situation while making sure the data meet all the related assumptions.
SPSS Process
To begin the analysis,
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the second variable, Friends_w4, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
Now that SPSS knows what variables to use, the type of nonparametric
test to conduct needs to be chosen.
Comparing Two or More Related Groups 119
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
Findings
The Hypothesis Test Summary information on the left side of the Model
Viewer window shows a Sig. value of .414. This Sig. value is larger than the
α = .05 threshold for significance. Therefore, the data suggest that there is not
enough information to determine that the number of friends in a gang has
changed from wave 2 to wave 4. However, it is important to note that, with so
little information for the combinations of categories, you would most likely
need a sizable difference (effect size) for any statistical test to find significant
differences. With so few respondents, you simply do not have enough power.
In statistical terms, the marginal homogeneity test fails to reject the
null hypothesis of equal marginal probabilities at α = .05 (p = .414).
Practically, this means that there is not enough information in the
responses from the few adolescents to conclude that there is an increase
in the number of reported friends in gangs over time.
SIGN TEST
The sign test uses the positive and negative differences for the matched
pairs of values from two continuous variables to determine if the two
sample medians have equal value. The null hypothesis is that the median
of the calculated differences for the two samples is zero (i.e., the initial
premise is that the median of one group is equal to the median of the sec-
ond group). The equivalent parametric test is the related-sample t-test,
but this test assumes that the difference scores are normally distributed in
the population, whereas the sign test has no distribution assumption. The
lack of a distribution assumption does make the sign test more applicable
Comparing Two or More Related Groups 121
The variables needed for the sign test are two paired scale or continuous
variables representing two related responses, such as pre and post values
for respondents. For Research Question #14 (see Box 4.4), the scale vari-
ables are Anxiety in waves 2 and 4 (i.e., Anxiety_w2 and Anxiety_w4).
SPSS Process
To begin the analysis,
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the second variable, Anxiety_w4, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
122 Data Analysis with Small Samples and Non-Normal Data
Now that SPSS knows what variables to use, the type of nonparametric
test to conduct needs to be chosen.
Unlike the step in Research Question #12 in which you identify val-
ues related to success, the sign test does not use a success distinction.
Therefore, no options button is available other than the Test Options
found in the Select an item: area.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
Findings
The histogram shows that nine of the paired values had a positive dif-
ference and 10 had a negative difference. In the total of 25 records, one
pair had missing data and five pairs were equal. Note that the statistical
calculations do not include pairs with zero differences. The number of
positive-difference pairs to negative-difference pairs suggests that there
is no information to indicate an increase in anxiety levels. In addition,
the sign test shows a Sig. value (1.000) that is the largest significance
value a test can return. Therefore, the sample of adolescents does not
indicate any increase in anxiety from wave 2 to wave 4.
However, if you look at the histogram on the right side of the Model
Viewer window (see Figure 4.3), the bars for the negative differences are
in a very different pattern from the bars for the positive differences. This
does suggest that something is happening related to anxiety levels.
Figure 4.3 reveals that the distribution of negative differences is
much wider than the distribution of positive differences. Knowing that
the sign test uses rankings and not difference values to test for equal
medians, you might consider that another test may offer more insight
into anxiety values over time (see Research Question #15).
10.0
8.0
Frequency
6.0
4.0
2.0
0.0
–6.00 –4.00 –2.00 0.00 2.00
Anxiety_w4 - Anxiety_w2
For Research Question #14, using statistical terms, the sign test fails
to reject the hypothesis of equal medians at α = .05 (p = 1.00), indicat-
ing that there is not enough information to indicate that the median
anxiety levels for wave 2 are different from the median anxiety levels
for wave 4.
Practically, this means that no information suggests that adoles-
cent anxiety levels are changing. However, a review of the histogram
in Figure 4.3 reveals that adolescents who report drops in their anxiety
level experience a larger drop, on average, than the increase for adoles-
cents reporting an increase in anxiety. Further analysis of these data is
warranted to fully understand what the data suggest.
SPSS Process
To begin the analysis,
□ Select => Analyze => Nonparametric Tests => Related Samples.
126 Data Analysis with Small Samples and Non-Normal Data
0.05 to 0.10.
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
Now that SPSS knows what variables to use, the type of nonparametric
test to conduct needs to be chosen.
Unlike the step in Research Question #12 in which you identify values
related to success, the Wilcoxon matched-pair signed-rank test does not
use a success distinction.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
Findings
The histogram in Figure 4.4 is the same as the histogram in Figure 4.3.
As before, the number of positive difference pairs to negative differ-
ence pairs suggests that there is no increase in median anxiety levels.
However, instead of looking at the pairs, a review of the value differ-
ences suggests a different finding. The Wilcoxon signed-rank test finds
128 Data Analysis with Small Samples and Non-Normal Data
that median of the differences does not equal zero, indicating that there
is a difference between anxiety levels for wave 2 and wave 4 (Sig. value
of .062, which is smaller than α = .10).
To view the upper and lower bounds of a 95% confidence interval for
median difference values,
□ Click on the dropdown list next to the View: area on the bottom
left of the Model Viewer window.
□ Click to change Hypothesis Summary View to the Confidence
Interval Summary View.
signed-rank test. Recall, this assumption is that the distribution for the
median difference values is symmetrical about the median.
To verify that the data meet the assumption, you need to generate a
new variable from the two samples and review its distribution.
SPSS Process
□ Select => Transform from the main SPSS menu.
□ Select Compute Variable… to open the Compute Variable
window.
□ Enter Median_DIFF in the white space below the Target
Variable: area.
□ Click to select the variable Anxiety_w2 from the list of variables
The above steps tell SPSS to create a new variable called Median_DIFF
by calculating the anxiety difference value for each respondent (i.e.,
Anxiety_w2 – Anxiety_w4)
After the new variable is computed, you need to ask SPSS to graph the
new computed values.
Note: The steps above include a request to draw a normal curve on the
graph. The assumption you are attempting to verify refers to the sym-
metry about the median and NOT an assumption of normality. The
reason for requesting SPSS to draw a normal curve is only to assist in
visually interpreting the symmetry and NOT to compare the distribu-
tion against a normal curve.
SPSS Output
Figure 4.5 shows the graph of the computed variable Median_DIFF.
A visual examination of the graph reveals a departure from symmetry
Histogram
10 Mean = 1.13
Std. Dev. = 2.383
N = 24
8
6
Frequency
0
–2.00 0.00 2.00 4.00 6.00 8.00
Median_DIFF
Figure 4.5. Distribution of value differences for two related samples.
Comparing Two or More Related Groups 131
COCHRAN’S Q TEST
Cochran’s Q test uses frequency of success and failure for k-related sam-
ples/observations to determine if the number of successes/failures is dif-
ferent among the samples. For Conchran’s Q, success could represent
testing positive on a medical test, answering yes to a yes/no question, or
endorsing, rather than not endorsing, a candidate. Comparing samples
to find potential differences is similar to repeated measures ANOVA, but
with dichotomous variables. Remember that repeated measures ANOVA
compares mean values and assumes normality for the sample values
within groups. Before conducting Cochran’s Q test, you must verify that
you have enough data for all the responses to have enough information
for a statistical test, and that the values are independent within groups
(i.e., the relationships among the values exist only across groups/waves of
data). The variables needed for Cochran’s Q test are three or more dichot-
omous variables representing three or more related response values for
respondents (e.g., observation 1, observation 2, observation 3). Because
Cochran’s Q test uses dichotomous variables, the null hypothesis is that
the distribution of frequencies for each of the k data point outcomes is
similar.
Other examples of uses for Cochran’s Q test are as follows:
For Research Question #16 (see Box 4.6), the dichotomous variables are
Committed an Assault for waves 2, 3, and 4 (i.e., Assault_w2, Assault_w3,
and Assault_w4). These three variables represent three time points in which
data were collected from adolescents, asking them if they had participated
in an assault on another individual since the last time interviewed.
SPSS Process
To begin your analysis,
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the second variable, Assault_w3, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the third variable, Assault_w4, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
Note: Another method to move variables from one side of the window to
the other side is to point your mouse at the variable name and double-
click. SPSS interprets the double-click as a message to move the variable
to the opposite area.
Now that SPSS knows what variables to use, you must choose the
kind of nonparametric test to conduct.
134 Data Analysis with Small Samples and Non-Normal Data
Before running the Cochran’s Q test, SPSS needs to know which value in
the dichotomous variables are associated with a success.
After the window opens, two options are displayed. The default option
is to let SPSS use the First value found in data.
□ Click on the white space below the word Value and enter a
1. Entering a 1 tells SPSS to associate the number 1 with a
success.
□ Select => OK to close the window.
Note: If the variables have more than two outcomes that represent
success, then the success outcomes are identified on multiple lines
in the white space. Recall, Cochran’s Q test is based on dichotomous
events. Having the capability to combine values expands the usabil-
ity of Cochran’s Q test to categorical variables with more than two
values.
You must set one more option for the Cochran’s Q test, the Multiple
comparisons: option. The default is for All pairwise and is shown in
the white area for the dropdown list under Multiple comparisons:. The
Comparing Two or More Related Groups 135
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
□ Using the View: dropdown list at the bottom right of the Model
Viewer window, select Pairwise Comparisons to display the
information shown in Figure 4.6.
136 Data Analysis with Small Samples and Non-Normal Data
Pairwise Comparisons
Assault_w2
7.00
Assault_w3
12.00
Assault_w4
15.00
Each row tests the null hypothesis that the Sample 1 and Sample 2 distributions
are the same.
Asymptotic significances (2-sided tests) are displayed. The significance level is .05.
Findings
The Adj. Sig. values shown in Figure 4.6 indicate that the only signif-
icant pairwise difference for assault frequencies is from wave 2 (i.e.,
Assault_w2) to wave 4 (i.e., Assault_w4). The other pairwise differences
are not significant (Adj. Sig. values of .377 and 1.00, both greater than
the α = .05 threshold).
Comparing Two or More Related Groups 137
Hovering the mouse over the lines in the Model Viewer’s Pairwise
Comparisons graph will display the related Adj. Sig. values. Therefore,
what you can draw from the Adj. Sig. values is that the number of respon-
dents committing assaults increased at each wave and by wave 4 adoles-
cents committing assaults had significantly increased from wave 2.
In statistical terms, Cochran’s Q test (p = .047) rejects the null hypoth-
esis that the number of respondents committing assaults has not changed
at α = .05 for the three waves of collecting data from the sample of gang
members. In addition, a test of equal probabilities was conducted for all
pairwise comparisons. The adjusted significance for multiple tests was
calculated, and it revealed that a single pair —wave 2 and wave 4 —was
significantly different (p = .043). The remaining pairs, wave 2 and wave 3
(p = .377), and wave 3 and wave 4 (p = 1.00) were not significant at α = .05.
Practically, this means that as a group in the sample, the number par-
ticipating in assaults significantly increased from wave 2 to wave 4. The
frequency changes from wave 2 to 3 and wave 3 to 4 are not large enough
to claim a significant difference. In addition, viewing the frequencies at
each wave indicates the change is from fewer assaults in wave 2 to signif-
icantly more in wave 4. Another interesting relationship to investigate
would be to look at joining a gang and committing assaults.
The value range for the coefficient is from 0 for no similarity to 1 for com-
plete similarity. Kendall’s coefficient requires three or more continuous/
scale variables that represent three or more samples, in which the value (i.e.,
row) in the sample is somehow related to the same row in the other sam-
ples. The commonly used parametric procedures associated with Kendall’s
coefficient is the Pearson correlation coefficient, but the Pearson correlation
coefficient assumes that the variables come from a normally distributed
population of values, and it can compare only two variables in a test. As
mentioned, Kendall’s coefficient handles three or more variables and its only
assumption is that the samples are dependent. Kendall’s coefficient uses
rankings and the number of ties to look for trends among the observations
to determine if the trends in the ranks are different among the samples. The
null hypothesis is that the distribution of the samples is the same.
Other examples of uses for Kendall’s coefficient of concordance are
as follows:
For Research Question #17 (see Box 4.7), the scale variables are Levels of
Support in waves 1, 2, 3, and 4 (i.e., Support_w1, Support_w2, Support_
w3, and Support_w4), which represent the values recorded for social
support for the four waves of data collections.
SPSS Process
To begin your analysis,
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the second variable, Support_w2, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the third variable, Support_w3, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the fourth variable, Support_w4, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
Note: Another method to move variables from one side of the window to
the other side is to point your mouse at the variable name and double-
click. SPSS interprets the double-click as a message to move the variable
to the opposite area.
Now that SPSS knows what variables to use, you need to choose the
type of nonparametric test to conduct.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
□ Click the dropdown list for the View: dropdown list at the
bottom right of the Model Viewer window and select Pairwise
Comparisons to display the information shown in Figure 4.7.
Findings
The Adj. Sig. values shown in Figure 4.7 indicate that one of the six pair-
wise comparisons is significant —Support_w1 vs. Support_w4 (p =.031).
Pairwise Comparisons
Support_w2 Support_w1
2.58 1.86
Support_w3
Support_w4
2.68
2.88
Each row tests the null hypothesis that the Sample 1 and Sample 2 distributions are the same.
Asymptotic significances (2-sided tests) are displayed. The significance level is .05.
The other five pairwise comparisons have Adj. Sig. values larger than the
Significance level: set in Test Options. Hovering the mouse over the lines
in the Pairwise Comparisons graph shown in the Model Viewer window
will cause the related Adj. Sig. values to display. Therefore, from the Adj.
Sig. values, you can conclude that the level of support reported by the ado-
lescents is significantly different from wave 1 to wave 4. If you review the
descriptive information for the four variables, you will see that reported
support levels are larger in wave 4 than in wave 1, indicating that the dif-
ference is an increase in reported support. In addition, you can see this
increase in the Pairwise Comparisons graph under each variable name.
In statistical terms, Kendall’s coefficient of concordance (p < .016)
at α = .05 rejected the null hypothesis of similar observed trends for the
four samples. This analysis says that if the null hypothesis is true, there
is less than a 1.6% chance that all four trends are similar. The test of
equal responses was conducted for all pairwise comparisons. Results of
the statistical analysis identify only one pair that is significantly differ-
ent (Support_w1 and Support_w4), with adjusted p = .031.
Practically, this means that reported support levels are changing
over time for the adolescents, and that a significant increase in the sup-
port levels exists for the adolescents in the sample between wave 1 and
wave 4. Understanding why the support levels increase over time will
require additional data collection and analyses.
along with a calculated mean rank for each sample, is used to draw a
conclusion about the null hypothesis that the sample distributions are
similar. Because the comparable related-samples ANOVA has a nor-
mality assumption that is difficult to meet and sometimes to verify
when samples are small, you can use the Friedman’s test in a wider
range of data situations. Both Kendall’s coefficient of concordance
and Friedman’s two-way ANOVA test examine sample similarities.
However, Friedman’s test is calculating the possibility that the samples
all come from one population, and Kendall’s coefficient is looking only
for similarities among the samples, a slight but recognizable difference
when selecting which procedures to use when analyzing your data.
The variables needed for Friedman’s two-way ANOVA test are three
or more continuous/scale variables representing three or more related
responses (e.g., pre-observation, second observation, third observa-
tion, fourth observation for respondents).
Other examples of uses for the Friedman’s two-way ANOVA test are
as follows:
For Research Question #18 (see Box 4.8), the scale variables are reported
Support Levels at waves 1, 2, 3, and 4 (i.e., Support_w1, Support_w2,
Support_w3, and Support_w4).
SPSS Process
To begin your analysis,
If you just completed the analysis for Research Question #17, the vari-
ables may still be in the Test Fields: area. If they are, skip down to the
Note: below. If they are not, SPSS needs to know which variables to use
in the analysis.
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the second variable, Support_w2, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the third variable, Support_w3, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
□ Click to select the fourth variable, Support_w4, found in the
Fields: area.
□ Click the move arrow to move the variable to the Test
Fields: area.
Note: Another method to move variables from one side of the window to
the other side is to point your mouse at the variable name and double-
click. SPSS interprets the double-click as a message to move the variable
to opposite area.
Now that SPSS knows what variables to use, you need to choose the
type of nonparametric test to conduct.
Unlike the step in Research Question #12 in which you identified val-
ues related to success, Friedman’s test does not use a success distinction.
Therefore, no buttons to select options are available. However, you can
change the Multiple comparisons option. The default is All pair-
wise, shown in the white area for the dropdown list under Multiple
comparisons: for the test. The other options for Multiple compari-
sons: are Stepwise step-down or None. When the number of groups
is too large, requesting the comparison of all possible pairwise com-
binations can be cumbersome. The Stepwise step-down option is
helpful in these situations. This option sorts the groups by their
calculated group median value and compares only adjacent groups,
making the comparison of groups possible. However, with only four
groups, the comparison of all pairwise combinations is completely
manageable.
SPSS Output
After you click Run, a variety of information shows up on SPSS’s output
window, including:
Within the Model Viewer window, on the right side, you will see the
distributions for each sample and their individual Mean Rank score.
Also on the right side, you will see the Total N count for the number
of responders, the Test Statistic, the Degrees of Freedom, and the
Asymptotic Sig. (2-sided test) value. You can use this information in
the test to decide if the distributions are dissimilar enough to reject the
possibility that they come from the same population.
Findings
The Asymptotic Sig. value (.016) tells us to reject the possibility that they
come from the same population. A review of the horizontal histograms
in the Model Viewer window clearly reveals different distributions for
the four variables because the shapes of the histograms are different.
However, there is more analysis information to review than just the sig-
nificance test. To see the descriptive information and the distribution
for the individual waves, change the selected option in the dropdown
list for View: at the bottom of the window.
□ Click the View: dropdown list at the bottom right of the Model
Viewer window.
□ Select Pairwise Comparisons to display the pairwise
information.
the statistical analysis identify only one pair that is significantly differ-
ent (Support_w1 and Support_w4), with adjusted p = .031.
Practically, this means not only that the support levels from one
wave to another increased over time, but also the reported levels clus-
tered differently. More adolescents reported lower levels of support in
wave 1 and more reported higher levels in wave 4.
5
Predicting with Multiple
Independent Variables
CHAPTER OBJECTIVES
This chapter presents nonparametric models that allow for prediction of
a dependent variable based on values for a set of independent variables.
The chapter covers the following:
INTRODUCTION
Chapter 5 covers prediction using both dependent and independent
variables in a nonparametric model. Similar to parametric models like
linear regression, nonparametric models are available that allow for pre-
diction of a dependent variable based on values for a set of independent
variables. As in the earlier chapters, the reason for selecting a nonpa-
rametric procedure is because the data do not meet the assumptions
149
150 Data Analysis with Small Samples and Non-Normal Data
a survey about the personal and occupational stressors that may influence
the quality of these social workers’ practice. After assembling her questions,
receiving comments from expert content and measurement reviewers, and
revising her items, she was ready to pilot test her questionnaire. She sent a
questionnaire to every kth person on the membership list of a professional
organization for adoption and foster care workers, following up with reminder
postcards and duplicate questionnaires at regular intervals, as recommended in
the survey research literature (e.g., Dillman, 2014). She mailed questionnaires
to 50 professionals and achieved a 70% response rate (N = 35). However, eight
professionals did not answer all the questions, reducing her usable sample to 27.
Dr. Northcott included many variables for the study, knowing from the literature
that both personal and occupational issues might be relevant. A sample size of
27 does not provide sufficient power to conduct the planned regression analyses,
nor does it meet the guidelines that states 10 to 15 cases are needed for each
independent variable in the regression model. She hires you to assist her in
answering her questions.
For Research Question #19 (see Box 5.2), three variables are used in
the analysis —two continuous independent variables, Burnout and
Stressors, and one continuous dependent variable, Impairment.
Before beginning a nonparametric procedure, you want to check
if the data do, indeed, violate parametric assumptions, which would
prevent you from using the customary linear regression. This is espe-
cially true when choosing between linear regression and nonparametric
regression. The major objective when building a nonparametric regres-
sion statistical model is to find the model that best fits your data. Thus,
unlike linear regression, nonparametric regression does not provide
beta values that you can use to interpret variables’ contribution to the
model. The reason lies in the technique used to estimate the requested
model for prediction. Therefore, viewing the results from a nonpara-
metric regression analysis is very different, and the process of identify-
ing the best model with the available variables is a bit more repetitive,
as it requires manually moving variables in and out of the model while
monitoring R 2 and prediction charts.
The first step is to examine the variables individually to fully under-
stand the quality and type of information they contain. Begin by run-
ning frequency distributions on each so you can review, among other
things, levels of skewness and kurtosis. For Research Question #19, the
distributions of both Burnout and Stressors are left-skewed and the dis-
tribution of Impairment is somewhat bimodal (see Figure 5.1).
SPSS Process
To obtain frequency distributions,
the window.
□ Click the move arrow to move the variable to the
Variable(s): area.
Note: Another method to move variables from one side of the window to
the other side is to point your mouse at the variable name and double-
click. SPSS interprets the double-click as a message to move the variable
to opposite area.
Now that SPSS knows what variables to use in the descriptive
analysis,
histogram.
□ Select => Continue to close the window.
□ Select => OK to run the analysis, which will open the Model
Viewer window.
SPSS Output
The histograms in Figure 5.1 are displayed in the Model Viewer
window.
If these variables were normally distributed, you would continue
assessing the assumptions. However, given the variables’ departure
from normality and the small data set (N = 27), using parametric
regression is not an option. In addition, one of the cases has a miss-
ing value (ID 11), and XLSTAT requires you either to drop the case
from the analysis or to select a missing value replacement option
(e.g., mean, nearest neighbor replacement). Because of the repetitive
nature of conducting nonparametric regression, it is probably best to
Impairment
10 Mean = 2.93
Std. Dev. = 1.269
N = 27
8
6
Frequency
0
0 1 2 3 4 5 6
Impairment
Burnout
6 Mean = 16.15
Std. Dev. = 9.134
5 N = 27
4
Frequency
0
0 10 20 30 40
Burnout
Stressors
6 Mean = 7.04
Std. Dev. = 3.117
5 N = 26
4
Frequency
0
2 4 6 8 10 12 14
Stressors
After you click on the button, Excel will take a few minutes to bring
up the toolbar and add a new tab named XLSTAT to the list of tabs at
the top of the Excel window.
XLSTAT Process
The nonparametric regression analysis to run is found under the
Modeling Data button .
□ Select => XLSTAT tab in the Excel menu options at the top of
the Excel window.
158 Data Analysis with Small Samples and Non-Normal Data
At the top of the Nonparametric regression window are seven tabs (General,
Options, Validation, Prediction, Missing data, Outputs, and Charts).
The information and selections you provide within each of these tabs tell
XLSTAT how to conduct the nonparametric regression (see Figure 5.2).
For robust LOWESS, XLSTAT dims some of the options to indicate that
this specific nonparametric regression model does not use them.
To begin the analysis,
By clicking in the white space, you draw focus to that area, which allows
for the identification of the dependent variables. The dependent variable
in the spreadsheet for Impairment is in column B, so find that column
on the spreadsheet and
When you click on the B, you tell XLSTAT where to find the values
for the dependent variable, Impairment. Your click will both high-
light the column in the traditional way Excel selects a column and add
text to the Y/Dependent variables: field (i.e., Data!$B:$B). Data!$B:$B
tells XLSTAT to use all the data in that column for your dependent
variable.
□ While holding down the “Ctrl” key on the keyboard, click and
drag the mouse from C to D at the top of the two columns.
Doing this tells XLSTAT which data to use for the independent vari-
ables and will place text in the field to identify the two columns (i.e.,
Data!$C:$D).
Note: Using the mouse to select the data for the analysis is only one
method of telling XLSTAT what to use. If you want to avoid using the
mouse, you can type the information directly into the fields that repre-
sent the dependent and independent data.
Because this research question does not include categorical data, the
Qualitative checkbox should be unchecked. Next,
For Research Question #19, and for many data situations, the robust
LOWESS will meet the prediction requirements.
Note: For more information on the above options available, including
more information on model procedures, select => Help button at the
lower right side of the Nonparametric regression window. Selecting
Help will open a separate window named HTML Help that provides
access to help information for XLSTAT and available procedures.
To proceed with the analysis,
The reason for this is that the first time you run your model, you start with
a linear polynomial (i.e., value of 1). Starting with 1 in this example tells
XLSTAT not to investigate the squared values of Burnout or Stressors for
predicting Impairment. Once you review the fit indices, you can return to
this option and try higher levels of polynomials to see if model fit can be
improved, as long as increasing the polynomial makes theoretical sense.
For example, does adding the squared values of Stressor (polynomial = 2)
improve model fit? However, keep in mind that variables with influen-
tial outliers can suggest a curvilinear relationship, especially with small
samples. This is another example of how important it is to understand the
data thoroughly before building regression models or testing hypotheses.
As you explore all the options provided by XLSTAT for nonpara-
metric regression, you will recognize many different analytic possi-
bilities. However, for this analysis most of the default options fit with
Research Question #19.
With larger data sets, validation is an option to further validate the fit of
a model. This is like splitting a data set so you can conduct an explora-
tory factor analysis (EFA) on one half and then validate the model with a
confirmatory factor analysis (CFA) using the other half. Unfortunately,
with smaller data sets, taking some of the records to validate the esti-
mated model is impractical. This is the case for Research Question #19;
therefore, due to the limited number of records, you should leave the
Validation option unchecked.
□ Select => Missing data tab to verify that cases with missing data
will be dropped.
Dropping the cases with missing data is the safest way to avoid adding
unrecognized bias. If you replace the missing values, you introduce the
possibility of unnaturally increasing the power of the model to predict
the independent variable, and it may not be a true reflection of the con-
cepts you are analyzing.
□ Select => Outputs tab to verify that all four options are
selected.
162 Data Analysis with Small Samples and Non-Normal Data
For Research Question #19, we want to request all the information avail-
able to help understand the data and the model.
□ Select => Charts tab to request that the analysis output includes
charts.
Be sure that the Data and predictions option is checked and that
the As a function of X1 option is selected. In addition, be sure to
select the Residuals option to add residual information to the anal-
ysis output.
Note: It is important that you set these options as described above for
both the Outputs and Charts tabs, because these options provide a clear
graphical view into the quality of the data in the model as well as model
fit (e.g., predicted vs. observed).
Now that the options are set,
□ Select => Continue button and a new Excel sheet will appear
with all the requested analysis output.
The search for the best-fitting model involves trying different model set-
tings (i.e., options) while monitoring the model fit information. Before
using any results from a statistical model, you must assess model fit to
make sure your model represents your data. If you are familiar with
factor analysis or latent class analysis (LCA), this exploratory approach
will sound familiar.
The first step in an examination of nonparametric regression model
fit in XLSTAT is to consider the amount of variance for the depend-
ent variables explained by the independent variables, identified as the
determination coefficient R 2. R 2 is part of the Goodness of fit statistics:
information found on the new Excel tab at the bottom of the Excel win-
dow (see Figure 5.3). R 2 is a calculated value that falls between 0 and
1. Zero represents the worst possible fit, in which the data do not follow
the model at all, and 1 represents a model that supports every data point
exactly.
4
Impairment
0
0 5 10 15 20 25 30 35 40
Burnout
Impairment Pred(Impairment)
Pred(Impairment)/Impairment
6
4
Impairment
0
0 1 2 3 4 5 6
Pred(Impairment)
and model prediction values. Like the scatter plot, histograms that show
a pattern as you review the bars from left to right suggest the possibility
of a poor-fitting model.
In addition, the bar chart provides the opportunity to identify cases
that have unusually large error (i.e., a bar on the chart that is much
larger than other bars). In these cases, you need to verify that you do not
a have a data-entry problem and that the case does follow all the original
guidelines for inclusion in the target population. In extreme situations,
you may remove the case if the case clearly does not belong theoretically
or conceptually, but we do not recommend this practice unless you have
irrefutable information to support the fact that the respondent is outside
of the parameters of the study.
Fortunately, for Research Question #19, the scatter plot in Figure 5.4
does not show a pattern along an envisioned diagonal line. In addition,
the bars in Figure 5.6 appear relatively random and do not indicate a
pattern or problem with cases’ being dependent. Therefore, you can be
Predicting with Multiple Independent Variables 167
Residuals
1.5
0.5
0
Residual
Obs1
Obs2
Obs3
Obs8
Obs9
Obs13
Obs14
Obs19
Obs20
Obs21
Obs22
Obs23
Obs26
Obs27
Obs10
Obs7
–0.5
Obs5
Obs17
Obs4
Obs15
Obs16
Obs18
–1
Obs6
Obs24
Obs12
–1.5
Obs25
–2
Observations
Verify that the information for each tab is correct in that it matches
what you intended for the final model, and that options are set the same
as when you found the best-fitting model.
□ Select => Prediction tab and click to select the checkbox next to
the Prediction option.
168 Data Analysis with Small Samples and Non-Normal Data
By clicking in the white space, you draw focus to the area that allows
the identification of the predictor values. The predictor values represent
values for Burnout and Stressors variables to predict impairment. In the
data set, the values are stored in two columns named Pred_Burnout and
Pred_ Stressors.
Doing this will place text in the field to identify the data to use for the
prediction (i.e., Data!$J$2:$K$3). Unlike the process used to select a col-
umn for the independent and dependent variables, the selection of the
predictor values cannot include the column headings.
The use of the mouse to select the data for the predictors is only one
method of telling XLSTAT what to use. If you want to avoid using the
mouse, you can type the information directly into the field.
Important: The order from left to right of the independent variables
in the spreadsheet must match the order of the predictor values. For
example, the Burnout variable column in the spreadsheet is to the left
of the Stressors column, which means that the values for Burnout and
Stressors in a prediction model must be Burnout values on the left and
Stressors values on the right.
Now that XLSTAT knows what information to use for the prediction,
The XLSTAT —Selections window appears, giving you one last chance
to see what data you have selected for the prediction.
XLSTAT Output
A new Excel sheet appears that contains all the LOWESS results
requested and a scatter plot graph that shows the two prediction val-
ues along with the other data points. Figure 5.7 shows the two pre-
dicted Impairment values (5.038 and 4.466) for the two sets of values
for Burnout (27.000 and 15.000) and Stressors (5.000 and 15.000).
A review of the predicted values shown in the scatter plot reveals that
two are on the fringe of other predicted values, but not so far out that
they appear to be outliers. When predicted values are not within the
proximity of other data values, this suggests the predictor values may
be outside of the model’s prediction capability. When this happens,
you need to use caution in using these predictions for decision-making
purposes.
4
Impairment
0
0 5 10 15 20 25 30 35 40
Burnout
Findings
In statistical terms, a nonparametric regression model using the robust
LOWESS estimation method achieved model fit with R 2 = .618 explain-
ing 62% of the variance in Impairment using two independent variables,
Burnout and Stressors. Two predictions for Impairment with values for
Burnout (27 and 15) and Stressors (5 and 15) were 5.038 and 4.466 and
were found to be within the model’s range for effective prediction.
Practically, this means that the new Impairment scale is showing
signs of reliability for use with two previously validated scales of Burnout
and Stressors. This means that you can estimate level of Impairment
with values for Burnout and Stressors.
Note: For comparison, we ignored the issue around meeting assump-
tions and conducted a parametric linear regression on the Research
Question #19 data. The results established an R 2 of 0.604, indicating that
the parametric model explained 60% of the variance in Impairment, in
comparison to the 62% we found for the nonparametric model. In addi-
tion, the RMSE for the linear regression was 0.809, while the RMSE for
the nonparametric regression as 0.747, meaning that there is a bit more
variation in error values for the linear regression model than the error
in the LOWESS model. The difference may be due to a variety of things,
including how close the data are to violating assumptions. However,
this finding does not mean that nonparametric models will always show
improved model fit when the data violate the assumptions. Rather, it just
means that you must consider many important factors when selecting
a statistical approach so that you are not presenting results that are not
supported statistically.
her independent variable. Dr. Northcutt asks if you can find a regression model
that could verify this relationship in her pilot sample. If she finds a relationship,
she will use this information in her upcoming class.
Research Question #20
Are depressive symptoms significantly related to an ordinal measure of
education and training?
NONLINEAR REGRESSION
Nonlinear regression is a method that models a nonlinear relationship
between a dependent variable and a single variable or multiple inde-
pendent variables. Unlike the frequently used multiple regression, the
relationship is not assumed to be linear —the relationship between the
dependent and the independent variables is not assumed to represent
a straight line. The dependent and independent variables can be either
categorical/ordinal or continuous, and it is assumed that the responses
captured by the dependent variable are independent and that the inde-
pendent variables are not overly correlated (i.e., multicollinearity). As in
nonparametric regression, the goal in conducting a nonlinear regres-
sion is to find a well-fitting model.
Other examples of uses for nonlinear regression are as follows:
After you click on the button, Excel will take a few minutes to bring
up the toolbar and add a new tab named XLSTAT to the list of tabs at
the top of the Excel window.
To request the plots,
The information and selections provided within each of these tabs tells
XLSTAT how to produce the two plots.
Now that the options are set,
A new Excel sheet will appear with all the requested descriptive output.
EXCEL Output
The Q-Q plots help identify when distributions among the groups are
too dissimilar to use in a nonlinear regression. Fortunately, the Q-Q
Predicting with Multiple Independent Variables 175
plots in Figure 5.8 show only minor differences among the three groups,
which is important because the distribution must be similar for the
three groups if you are going to use nonlinear regression with a cate-
gorical dependent variable. In addition, the Q-Q plots show a specific
0
0 1 2 3 4 5
Depression | Edu_Level-1
14
12
10
2
2 4 6 8 10 12 14 16
Depression | Edu_Level-2
20
15
10
0
0 5 10 15 20 25
Depression | Edu_Level-3
Figure 5.8. Q-Q plots.
176 Data Analysis with Small Samples and Non-Normal Data
pattern along the diagonal line (i.e., first above and then below), which
does support the possibility of a nonlinear model.
The requested box-and-whiskers plots show the data distributions
for each group, which are the categories for the dependent variable. Like
the Q-Q plots, the differences among the box-and-whiskers plots are not
enough to indicate that we cannot continue.
As stated for other statistical procedures, another important step
before moving on is to make sure you have enough information for
each group, which in this situation translates into enough represen-
tation in all three groups (i.e., n = 8, 12, and 7 for bachelor’s degree,
master’s degree with no license, and master’s degree with clinical license,
respectively). The minimum guideline we recommend is to have at least
5 responses for each category in a statistical procedure, but remember
that you should use any parameter estimates with care due to the lack of
precision from using a very small sample.
Note: The Q-Q plots presented in Figure 5.8 are those that should be
produced prior to conducting a linear regression to verify that the nor-
mality assumption is not violated.
Now that you have examined the data closely and have veri-
fied similarity among the categories, you can move on to finding a
model that fits your data. So how many different models are possi-
ble? Unfortunately, the list is extremely long. One of the strengths of
nonlinear regression is that a researcher can explore many different
models with many different levels of complexity. Up to now, we have
avoided discussing formulas, and we will continue to leave formula
development to other discussions outside of this book. However, to
accomplish the nonlinear regression, you will need to use three dif-
ferent formulas we provide for this procedure. The three formulas
represent three possible probability function models that fit nonlin-
ear relationships. Many different probability functions are built into
XLSTAT for conducting nonlinear regression. In addition to the built-
in functions, XLSTAT offers a process that allows for the development
of other function models.
While answering Research Question #20, both the built-in and the
user-defined methods are described. You may ask, “With all this flexi-
bility, how do I know where to start?” Fortunately, the three function
options cover the use of two-, three-, and four-parameter models (i.e.,
probability models predicting a value for the dependent variable) that
Predicting with Multiple Independent Variables 177
will meet the needs of most people who are not statisticians but have
the occasion to use nonlinear regression. These three examples will get
you well on your way to conducting a nonlinear regression analysis. We
present the functions related to the examples later when discussing the
steps for the XLSTAT procedure.
Exploring different models is the main process for identifying an
effective nonlinear relationship (unlike nonparametric regression, in
which the data identify the relationship). The first model relates to one
that uses only two parameters to identify the relationship between a
dependent and an independent variable. If you took a basic statistics
class, a simple linear regression formula included an intercept and
a slope (i.e., β0 and β1, respectively). The intercept and slope are two
parameters that help explain the relationship between a dependent and
independent variable. The two-parameter nonlinear model attempts to
do the same thing, which is to find estimates for the two parameters
that appropriately explain the relationship. You can add parameters
when the two parameters are not enough to explain a more complex
situation. Each model has a unique way of using the parameters in a
formula and explaining the relationship. As you would expect, using
more parameters than are necessary adds complexity and possibly
unwanted estimate error to your prediction. Therefore, the goal is to
find an effective model by finding the formula that has the fewest num-
ber of parameters while still effectively representing the relationship
between the variables.
A nice feature of XLSTAT is that two of the three functions that
include two and three parameters are already in a list of available func-
tions; however, SPSS offers no list of functions to get you started. In
addition, XLSTAT provides a very friendly way to include user-defined
functions —which is demonstrated as part of answering Research
Question #20.
After you click on the button, Excel will take a few minutes to bring
up the toolbar and add a new tab named XLSTAT to the list of tabs at the
top of the Excel window. You will find the nonlinear regression analysis
to run under the Modeling Data button.
At the top of the Nonlinear regression window are seven tabs. The
information and selections provided within each of these tabs tell
XLSTAT how to conduct the nonlinear regression. For Research
Question #20,
By clicking in the white space, you shift your focus to the area that
allows for the identification of the dependent variable. The dependent
variable in the spreadsheet for Edu_Level is in column F, so find that
column on the spreadsheet and
Doing this will both highlight the column in the traditional way Excel
selects a column and add text to the Y/Dependent variables: area (i.e.,
Data!$F:$F). Data!$F:$F indicates that you want to use all the data in
that column for the dependent variable.
□ Make sure you check the Variable labels option on the right side
of the window.
Predicting with Multiple Independent Variables 179
□ Find the E column on the spreadsheet and click on the “E” at the
top of the column.
Doing this will place text in the field to identify the column (i.e.,
Data!$E:$E).
Note: The use of the mouse to select the data for the analysis is only one
method for telling XLSTAT what to use. If you want to avoid using the
mouse, you can type the information directly into the fields.
The next step is to select the probability model to use.
□ Select => Options tab and verify that none of options on the left
side is selected.
□ Select => Validation and Prediction tabs and verify that none
and a new Excel sheet will appear with all the requested analysis
output.
Predicting with Multiple Independent Variables 181
analysis.
Note: Excel will not overwrite the previous output appearing in sepa-
rate Excel sheets. Each run of an analysis will create a new Excel sheet.
Keeping track of the sheets by renaming them will save you time later
when you are writing up your findings.
It is fair to say that the results from the nonlinear regression model
containing three parameters show that you have sufficient evidence to
conclude that the levels of depressive symptoms are different for people
with different education and training levels. You can draw this conclu-
sion primarily because you could find a model that showed appropriate
fit and prediction capabilities. If you were not able to find an effective
model, you could not draw this conclusion, and you would need to test
the probability model with four parameters. To be thorough in present-
ing the process and to complete the introduction to nonlinear regres-
sion, we next describe the four-parameter model option.
Predicting with Multiple Independent Variables 183
Edu_Level
2.5
1.5
1
0 5 10 15 20 25
Depression
Active Model
3.5 Pred(Edu_Level)/Edu_Level
2.5
Edu_Level
1.5
0.5
0.5 1 1.5 2 2.5 3 3.5
Pred(Edu_Level)
Residuals
2
1.5
1
Residual
0.5
0
Obs1
Obs2
Obs5
Obs7
Obs8
Obs10
Obs13
Obs16
Obs17
Obs18
Obs22
Obs23
Obs24
Obs27
Obs6
Obs11
Obs25
–0.5
Obs21
Obs4
Obs26
Obs19
Obs3
Obs9
Obs15
Obs12
Obs14
Obs20
–1
Observations
Said differently, if you have not yet downloaded the Chapter 5 data
set, open an Excel sheet that contains no information, type the four
formulas exactly as written above into the first four lines in column
A in the empty sheet and label it derivatives in the label space at the
bottom of the sheet. These four formulas are the same formulas you
can use with other data when examining a model with four param-
eters. Again, you do not need to understand the workings of these
Predicting with Multiple Independent Variables 185
formulas for now, just know that XLSTAT needs these four to sup-
port your effort to examine a four-parameter model that XLSTAT
did not predefine. Once you enter the four, and after verifying that
you typed them correctly, return to XLSTAT and set up the new
probability model.
To continue your analysis in XLSTAT, you’ll need to reopen the
Nonlinear regression window.
□ Click on the Add button toward the bottom right of the window.
When you click the Add button, the two white areas for Function: Y =
and Derivatives: become active.
Next, we want to let XLSTAT know where to find the four formu-
las entered earlier into that blank Excel sheet that you have named
derivatives.
□ Navigate to where the four formulas are, and click and drag to
highlight all four of them (i.e., click and drag adds the location
information into the area —derivatives!$A$1:$A$4).
□ Click to move the focus to the white space for the Function:
Y = area.
□ Type the formula pr3/(1 + exp(-pr1-pr2*X1))^(1/pr4)
Note: The formula starts with “p” and ends with “)” —a closed paren-
thesis symbol. Also, once you add this as a user-defined function,
186 Data Analysis with Small Samples and Non-Normal Data
you will not have to add it again for future analyses unless you update
XLSTAT with a new version.
□ After you verify the formula is correct, click the Save button to
add the formula to the User defined functions list.
□ Select => Continue button and a new Excel sheet will appear
with all the requested analysis output.
(
Edu_Level = 2.80 / 1+ Exp ( +0.61 - 0.24* Depression ) )
Predicting with Multiple Independent Variables 187
The above formula can be found in the Excel sheet that was generated
from the analysis. The formula is located on, or close to, line 41 on the
Excel sheet. Because the model fit is adequate, you can enter a value
for Depression into the above equation to predict what category is most
likely —the education level category in which the respondent will most
likely be. One way to accomplish this is to use Excel.
(
= 2.80 / 1 + exp (.61 − .24 * A1) )
□ After typing in the formula above, press the Enter key on your
keyboard to calculate a value for Edu_Level that will appear in
cell B1. The value that appears in cell B1 is most likely a number
that you can round to the nearest integer. The integer number
is the predicted Edu_Level category for the value of Depression
you entered in A1.
Findings
A three-parameter nonlinear model with an ordinal dependent vari-
able Edu_Level and a continuous independent variable Depression
was identified for estimating education and training level (Edu_Level)
using scores on a depressive symptoms measure (R 2 = .571). The three-
parameter fit indices for RMSE and R 2 were much improved over a
two-parameter model. In addition, R 2 for a four-parameter model was
the same as the three-parameter model, and the RMSE showed only a
slight increase, results that support the three-parameter model as most
efficient.
Practically, this means that the nonlinear model has identified a
relationship between Depression and Edu_Level (level of education
and training). These findings from this sample suggest that levels of
depressive symptoms do predict variations in education and training,
but more data collection and analyses are necessary to conclude a causal
relationship. A different and perhaps more intuitive regression model
188 Data Analysis with Small Samples and Non-Normal Data
could reverse the roles of the variables (i.e., education and training as
a predictor of depressive symptoms level). Knowing which to choose as
the dependent variable centers on the research question you are attempt-
ing to answer and the type of data available. This alternative regression
model would explore the possibility that education and training lower
depressive symptoms levels.
Note: Another SPSS analysis method for models with ordered catego-
rical dependent variables is PLUM (Polytomous Universal Model).
For more information on this nonparametric strategy, see the website
https://s tatistics.laerd.com/spss-t utorials/ordinal-regression-u sing-
spss-statistics-2.php
FACTORIAL ANALYSIS
Factorial analysis, also called general linear model (GLM) analysis, is
a method for conducting a regression analysis in which the dependent
variable is continuous (in this example) and the independent variables
are categorical (i.e., the factors in factorial analysis). This analysis exam-
ines the effects factors have on the dependent variable. In other words,
Predicting with Multiple Independent Variables 189
accepted guideline of five per category, you do not have enough males
representing the male category —the variable does not have enough
information to compare males and females.
The five-count guideline becomes more difficult to meet as you
add more categorical variables into a factorial model. For example, if
Research Question #21 aimed at knowing if gender and education level
were predictors of depressive symptoms, the data do not have enough
coverage to pursue this analysis (see Table 5.2). Remember that Edu_
Level is an ordinal measure of education and training (1 = bachelor’s
degree only, 2 = master’s degree with no license, and 3 = master’s degree
with clinical license).
Review of Table 5.2 reveals that cross-tabulation of males with edu-
cation level does not meet the five-count guideline. In other words,
you do not have enough males in Education Levels 1 and 2 to represent
properly all the potential males who are at those education levels, and
the same is true for both males and females in the Education Level 3
category.
Sometimes the coverage for the various categories is such that col-
lapsing a category is necessary. For example, one option for Research
Question #21 would be to merge education categories 1 and 2 to com-
pensate for the low number of men, but the few numbers and therefore
the lack of information for category 3 would still be unresolved, and
just managed for analysis purposes. It is important to note that SPSS
will perform a factorial analysis on a model with Gender and Edu_Level
unchanged, but the potential of coming to a flawed conclusion based on
these variables is high. SPSS does not always warn you when cell cov-
erage is too low. Therefore, with small data sets, at times you may not
Edu_Level 1 6 2 8
2 6 4 10
3 3 4 7
Total 15 10 25
Predicting with Multiple Independent Variables 191
You select the Univariate option because you have only one dependent
variable, named LifeSatisfaction. Once the Univariate window opens,
you need to tell SPSS which variable to analyze. The list on the left
includes the variables in the data set.
Verify that the Full factorial option is selected because this will test
potential differences for Gender, Employment, and their interaction
(Gender * Employment) for LifeSatisfaction.
After the window opens, you will see the two factors Gender and
Employment in the Factors: area. With only two factor variables, we
suggest that you request two plots —one with Gender on separate lines
and one with Employment on separate lines.
interactions: area.
□ Click the move arrow to copy (OVERALL) to the Display
Means for: area.
□ Select four options in the Display area (Descriptive
Note: You can click the Help button on this window to get additional
information on all the available options.
SPSS Output
After you select OK, a variety of information shows up on the SPSS out-
put window. On the left side of the output window, you will see an index
showing the names of the information sections created from running the
analysis. The names are arranged in an index, and clicking on the name
displays the details of that section on the right. On the right side of the out-
put window, you will see all the detailed information within each section.
If you have been running other analyses before your factorial model,
you may have other information in your output window. If this is the
case, simply click on Univariate Analysis of Variance in the index on
the left side of the output window. This will cause the left side of the win-
dow to scroll to the detailed information for the analysis output.
Findings
In the Descriptive Statistics table in the IBM SPSS Statistics Viewer win-
dow, you can verify that you selected the correct variables for the anal-
ysis and that the data provide sufficient coverage for the combination of
categories (i.e., values for n meet the guideline of five or more per cell).
The next table, Levene’s Test of Equality of Error Variance, sug-
gests (p = .213) that the statistical assumption about homogeneity of
variances among the groups has not been violated. A Sig. value smaller
than α =.05 would indicate the variances are potentially different among
the factors and therefore may require a different analysis approach,
because an assumption has been violated.
When you scroll down to the next section of output information,
the Test of Between-Subjects Effects table shows the results of your
194 Data Analysis with Small Samples and Non-Normal Data
Plots button prior to running the analysis. The differences in how the
lines are drawn using the Estimated Marginal Means values is larger for
Gender than it is for the Employment Settings. Using the vertical case for
Estimated Marginal Means will help in comparing the two graphs. The
observed differences in the two graphs support the p-values for Gender
(p = .043) and Employment Setting (p = .951).
Another important observation is the intersection of the two lines
for Gender (see Figure 5.11). Intersecting lines can indicate a potential
significant factor interaction. However, in our analysis, the p-value for
Gender*Employment was not significant (p = .167). Because the sam-
ple is very small, there may not be enough information to identify a
significant interaction of Gender and Employment, especially with
Employment categories involved (p = .951).
A good way to test the effects of the nonsignificant interaction on the
model is to remove the interaction term altogether and rerun the anal-
ysis. To accomplish this,
16 16
15 15
14 14
13 13
12 12
f m Private Public
Gender Employment
Employment Gender
Private f
Public m
Model: area.
□ Click to select Gender in the Factors & Covariates: area.
□ Click the move arrow to move Horizontal Axis: to the
Model: area.
With the updated model, repeat your review of the Levene’s test and
check the assumptions as previously described. For this analysis, after
repeating the process, the Tests of Between-Subjects Effects table shows
that Gender is significant (p = .031) and that the Adjusted R Squared of
.152 is just .003 smaller than the previous model. Therefore, this simpler
model, with just a trivial drop in Adjusted R Squared explaining 15% of
LifeSatisfaction variance, suggests a more efficient model for analyzing
Research Question #21. The Parameter Estimates table shows a signif-
icant parameter estimate for Gender (β = -3.067, p = .031), providing
evidence that mean differences for LifeSatisfaction exist between males
and females (see Table 5.3).
Findings
In statistical terms, a GLM model examining LifeSatisfaction for factors
Gender and Employment found that Employment and the interaction
of Employment and Gender were not significant. A model with only
Gender was significant (p = .031) at α = .05 with an adjusted R2 = .152.
Predicting with Multiple Independent Variables 197
Dependent variable: LifeSatisfaction.
b
Practically, this means that the GLM analysis shows that you do not
have enough evidence to suggest a difference in life satisfaction between
individuals who have public or private industry employment. However,
there is sufficient evidence to suggest that, in general, males and females
have different life satisfaction levels, and that females, on average, score
three points lower than males when measured by a validated profes-
sional life satisfaction scale. In addition, gender accounts for 15% of the
variance in life satisfaction in the way life satisfaction was measured for
this study.
Appendix A
SPSS Syntax
Appendix A presents the SPSS syntax for each of the procedures pre-
sented in this book, identified by chapter and by research question.
Working with the syntax is more comfortable for some researchers,
as they find it easier to follow the specific characteristics of each ana-
lytic procedure. In addition, the capability and usefulness of SPSS can
greatly increase when a researcher moves beyond simply using menus
and point-and-click procedures.
CHAPTER 2
Research Question #1
*Frequency Analysis:
FREQUENCIES VARIABLES=Confidante
/ORDER=ANALYSIS.
*Parametric T-Test
T-TEST
/TESTVAL=0
/MISSING=ANALYSIS
199
200 Appendix A
/VARIABLES=Confidante
/CRITERIA=CI(.95).
*Binomial Analysis:
*Nonparametric Tests: One Sample.
NPTESTS
/
ONESAMPLE TEST (Confidante) BINOMIAL(TESTVALUE=0.6
CLOPPERPEARSON JEFFREYS LIKELIHOOD
SUCCESSCATEGORICAL=LIST(1) SUCCESSCONTINUOUS=CUTPOINT(MIDP
OINT))
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #2
*Chi-square test:
*Nonparametric Tests: One Sample.
NPTESTS
/
ONESAMPLE TEST (Coming _
Out) CHISQUARE(EXPECTED=EQUAL)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #3
*Kolmogorov-Smirnov test:
*Nonparametric Tests: One Sample.
NPTESTS
/
ONESAMPLE TEST (Loneliness)
KOLMOGOROV _
SMIRNOV(NORMAL=SAMPLE UNIFORM=SAMPLE
EXPONENTIAL=SAMPLE POISSON=SAMPLE)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #4
*Wilcoxon signed-
rank test:
*Nonparametric Tests: One Sample.
NPTESTS
/
ONESAMPLE TEST (Coping _
Level) WILCOXON(TESTVALUE=15)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Appendix A 201
Research Question #5
*Descriptive Analysis:
DESCRIPTIVES VARIABLES=Rater _
score
/
STATISTICS=MEAN STDDEV MIN MAX.
*Runs test with mean cutpoint:
*Nonparametric Tests: One Sample.
NPTESTS
/
ONESAMPLE TEST (Rater _
score) RUNS(GROUPCATEGORICAL=SAMPLE
GROUPCONTINUOUS=CUTPOINT(SAMPLEMEAN))
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
*Runs test with dichotomous variable:
*Nonparametric Tests: One Sample.
NPTESTS
/
ONESAMPLE TEST (Rater _
score Confidante)
RUNS(GROUPCATEGORICAL=SAMPLE
GROUPCONTINUOUS=CUTPOINT(SAMPLEMEAN))
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #6
*Histograms for Age and Self _ esteem
FREQUENCIES VARIABLES=Age Self _ esteem
/HISTOGRAM NORMAL
/ORDER=ANALYSIS.
*Filter cases (use only youths who have not come out to
anyone)
USE ALL.
COMPUTE filter _
$=(Coming _
Out=1).
VARIABLE LABELS filter _ $ ‘Coming _
Out=1 (FILTER)’.
VALUE LABELS filter _
$ 0 ‘Not Selected’ 1 ‘Selected’.
FORMATS filter _
$ (f1.0).
FILTER BY filter _
$.
EXECUTE.
*Filter cases (use only youths who have come out to one or
more people)
USE ALL.
COMPUTE filter _
$=(Coming _
Out<>1).
VARIABLE LABELS filter _ $ ‘Coming _
Out<>1 (FILTER)’.
VALUE LABELS filter _
$ 0 ‘Not Selected’ 1 ‘Selected’.
FORMATS filter _
$ (f1.0).
202 Appendix A
FILTER BY filter _
$.
EXECUTE.
*Remove any active filter
FILTER OFF.
USE ALL.
EXECUTE.
*Correlation with all youths
NONPAR CORR
/
VARIABLES=Age Self _ esteem
/
PRINT=BOTH TWOTAIL NOSIG
/
MISSING=PAIRWISE.
*Sort cases
SORT CASES BY Age(A).
*filter and correlation with only youths who have not come
out to anyone
USE ALL.
COMPUTE filter _
$=(Coming _
Out=1).
VARIABLE LABELS filter _ $ ‘Coming _
Out=1 (FILTER)’.
VALUE LABELS filter _
$ 0 ‘Not Selected’ 1 ‘Selected’.
FORMATS filter _
$ (f1.0).
FILTER BY filter _
$.
EXECUTE.
NONPAR CORR
/
VARIABLES=Age Self _ esteem
/
PRINT=BOTH TWOTAIL NOSIG
/
MISSING=PAIRWISE.
*filter and correlation with only youths who have come out
to one or more people
USE ALL.
COMPUTE filter _
$=(Coming _
Out<>1).
VARIABLE LABELS filter _ $ ‘Coming _
Out<>1 (FILTER)’.
VALUE LABELS filter _
$ 0 ‘Not Selected’ 1 ‘Selected’.
FORMATS filter _
$ (f1.0).
FILTER BY filter _
$.
EXECUTE.
NONPAR CORR
/
VARIABLES=Age Self _ esteem
/
PRINT=BOTH TWOTAIL NOSIG
/
MISSING=PAIRWISE.
*Remove any active filter
FILTER OFF.
USE ALL.
EXECUTE.
Appendix A 203
CHAPTER 3
Research Question #7
*Independent Group Comparison and CI
*Nonparametric Tests: Independent Samples.
NPTESTS
/
INDEPENDENT TEST (Percent _
Pre) GROUP (Received _
MI)
MOSES(TRIMOUTLIERS=SAMPLE) HODGES _
LEHMANN
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #8
*Independent Group Analysis:
*Nonparametric Tests: Independent Samples.
NPTESTS
/INDEPENDENT TEST (Percent _ DIF) GROUP (Received _ MI)
KOLMOGOROV _ SMIRNOV
/MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
*Independent Group Analysis using three tests:
*Nonparametric Tests: Independent Samples.
NPTESTS
/INDEPENDENT TEST (Percent _ DIF) GROUP (Received _ MI)
MANN _
WHITNEY KOLMOGOROV _ SMIRNOV WALD _ WOLFOWITZ
/MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
*Frequency for MI=1 respondents
USE ALL.
COMPUTE filter _
$=(Received _
MI=1).
VARIABLE LABELS filter _ $ ‘Received _
MI=1 (FILTER)’.
VALUE LABELS filter _
$ 0 ‘Not Selected’ 1 ‘Selected’.
FORMATS filter _
$ (f1.0).
FILTER BY filter _
$.
EXECUTE.
FREQUENCIES VARIABLES=Percent _ DIF
/HISTOGRAM NORMAL
/ORDER=ANALYSIS.
*Frequency for MI=2 respondents
USE ALL.
COMPUTE filter _
$=(Received _
MI=2).
VARIABLE LABELS filter _ $ ‘Received _
MI=2 (FILTER)’.
204 Appendix A
Research Question #9
*Filter to get only respondents in the treatment group:
USE ALL.
COMPUTE filter _
$=(Received _
MI=1).
VARIABLE LABELS filter _ $ ‘Received _
MI=1 (FILTER)’.
VALUE LABELS filter _
$ 0 ‘Not Selected’ 1 ‘Selected’.
FORMATS filter _
$ (f1.0).
FILTER BY filter _
$.
EXECUTE.
*Nonparametric Tests: Independent Samples.
NPTESTS
/
INDEPENDENT TEST (Percent _ Post) GROUP (Partner)
KRUSKAL _
WALLIS(COMPARE=PAIRWISE)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #10
*Filter to get only respondents in the treatment group:
USE ALL.
COMPUTE filter _
$=(Received _
MI=1).
VARIABLE LABELS filter _ $ ‘Received _
MI=1 (FILTER)’.
VALUE LABELS filter _
$ 0 ‘Not Selected’ 1 ‘Selected’.
FORMATS filter _
$ (f1.0).
FILTER BY filter _
$.
EXECUTE.
*Jonckheere _ Terpstra test:
*Nonparametric Tests: Independent Samples.
NPTESTS
/
INDEPENDENT TEST (Percent _ DIF) GROUP (Stage _
of _
Change)
JONCKHEERE _
TERPSTRA(ORDER=ASCENDING COMPARE=PAIRWISE)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
*Nonparametric Tests: Independent Samples.
Appendix A 205
NPTESTS
/
INDEPENDENT TEST (Percent _
DIF) GROUP (Stage _
of _
Change)
JONCKHEERE _
TERPSTRA(ORDER=ASCENDING COMPARE=PAIRWISE)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.1 CILEVEL=95.
Research Question #11
* set filter to only use respondents in treatment group
USE ALL.
COMPUTE filter _
$=(Received _ MI=1).
VARIABLE LABELS filter _ $ ‘Received _ MI=1 (FILTER)’.
VALUE LABELS filter _
$ 0 ‘Not Selected’ 1 ‘Selected’.
FORMATS filter _
$ (f1.0).
FILTER BY filter _
$.
EXECUTE.
*Nonparametric Tests: Independent Samples.
NPTESTS
/
INDEPENDENT TEST (Self _ efficacy) GROUP (Condom _ Request)
MEDIAN(TESTVALUE=SAMPLE COMPARE=PAIRWISE)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
CHAPTER 4
Research Question #12
* Analysis using McNemar test:
*Nonparametric Tests: Related Samples.
NPTESTS
/
R ELATED TEST(Violent _
crime _
w1 Violent _
crime _
w2)
MCNEMAR(SUCCESS=LIST(1))
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #13
*Descriptive Analyses:
FREQUENCIES VARIABLES=Friends _
w2 Friends _
w4
/ORDER=ANALYSIS.
*Nonparametric Tests: Related Samples.
NPTESTS
206 Appendix A
/
R ELATED TEST(Friends _
w2 Friends _
w4)
MARGINAL _
HOMOGENEITY
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #14
* Analysis using Sign test:
*Nonparametric Tests: Related Samples.
NPTESTS
/
R ELATED TEST(Anxiety _
w2 Anxiety _
w4) SIGN
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #15
*Analysis using Wilcoxon Signed- Rank test and CI:
*Nonparametric Tests: Related Samples.
NPTESTS
/
R ELATED TEST(Anxiety _
w2 Anxiety _
w4)
WILCOXON HODGES _
LEHMAN
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.1 CILEVEL=95.
COMPUTE Median _
DIFF=Anxiety _
w2 -Anxiety _
w4.
EXECUTE.
Research Question #16
*Analysis using Cochran’s Q test:
*Nonparametric Tests: Related Samples.
NPTESTS
/
R ELATED TEST(Assault _
w2 Assault _
w3 Assault _
w4)
COCHRAN(SUCCESS=LIST(1) COMPARE=PAIRWISE)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #17
*Analysis using Kendall’s Coefficient of Concordance:
*Nonparametric Tests: Related Samples.
NPTESTS
/
RELATED TEST(Support _
w1 Support _
w2 Support _
w3 Support _
w4)
Appendix A 207
KENDALL(COMPARE=PAIRWISE)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
Research Question #18
*Analysis using Friedman’s Two- Way ANOVA test:
*Nonparametric Tests: Related Samples.
NPTESTS
/
RELATED TEST(Support _
w1 Support _
w2 Support _
w3 Support _
w4)
FRIEDMAN(COMPARE=PAIRWISE)
/
MISSING SCOPE=ANALYSIS USERMISSING=EXCLUDE
/
CRITERIA ALPHA=0.05 CILEVEL=95.
CHAPTER 5
Research Question #19
n/a
Research Question #20
n/a
Research Question #21
*Factor Analysis:
CROSSTABS
/TABLES=Edu _
Level BY Gender
/FORMAT=AVALUE TABLES
/CELLS=COUNT
/COUNT ROUND CELL.
UNIANOVA LifeSatisfaction BY Gender Employment
/M ETHOD=SSTYPE(3)
/INTERCEPT=INCLUDE
/POSTHOC=Gender(LSD BONFERRONI)
/PLOT=PROFILE(Gender*Employment Employment*Gender)
/EMMEANS=TABLES(OVERALL)
/PRINT=LOF HOMOGENEITY DESCRIPTIVE PARAMETER
/PLOT=RESIDUALS
/CRITERIA=ALPHA(.05)
208 Appendix A
Most researchers will have to deal with missing data at some point in
their careers. The challenge of missing data is that there are many reasons
for the missingness, and probably even more ways in which to address
missingness in a data set. As you may remember from your introduc-
tion to statistics course, missing data fall into three categories—missing
completely at random (MCAR), missing at random (MAR), and missing
not at random (MNAR). MCAR and MAR are, for the most part, man-
ageable when it comes to dealing with the gaps in the information (i.e.,
the missing data).
However, MNAR can be problematic. For example, the gaps in
information might be due to how a few respondents reacted to the data-
collection procedures—say they interpreted a survey question differ-
ently from the majority of other respondents, or perhaps they elected
to skip a response for personal reasons outside of the study issues. This
systematic missingness undermines the accuracy of your findings, and
in some cases can invalidate your study altogether, without your even
recognizing it. For this reason, whenever missing data are present, you
must conduct an extensive review of the reliability and validity of the
study methods. MNAR can be caused by issues in the data collection
methods, such as bias in sampling methods, uneven interpretation of
209
210 Appendix B
211
References
CHAPTER 1
Aguinis, H., Gottfredson, R. K., & Joo, H. (2013). Best-practice recommen-
dations for defining, identifying, and handling outliers. Organizational
Research Methods, 16(2), 270–301. doi:10.1177/1094428112470848
Barnett, V., & Lewis, T. (1978). Outliers in statistical data. Chichester: Wiley.
Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2d ed.).
Hillsdale, NJ: Lawrence Earlbaum Associates.
Cohen, J. (1992). A power primer. Psychological Bulletin, 112, 155–159.
doi:10.1037/0 033-2909.112.1.155
Corder, G. W., & Foreman, D. I. (2009). Nonparametric statistics for non-
statisticians. Hoboken, NJ: Wiley.
Davidson, A. C., & Hinkley, D. V. (1998). Bootstrap methods and their applica-
tion. Cambridge, U.K.: Cambridge University Press.
Efron, B. (1979). Bootstrap methods: Another look at the jackknife. The Annals
of Statistics, 7, 1–26. https://doi.org/10.1214/aos/1176344552
Elliott, A. C. (2015). Determine equality of variance for two groups. http://stat-
tutorials.com/EXCEL/EXCEL-EQUAL-VARIANCE.html
Gaskin, J. (2015). Plotting homoscedasticity in SPSS [Video file]. Available from
https://w ww.youtube.com/watch?v=WrQ1O_He63Q
Hojat, M., & Xu, G. (2004). A visitor’s guide to effect sizes: Statistical sig-
nificance versus practical (clinical) importance of research findings.
Advances in Health Sciences Education, 9, 241–249. https://doi.org/10.1023/
b:ahse.0000038173.00909.f6
213
214 References
CHAPTER 2
Cochran, W. G. (1952). The χ2 test of goodness of fit. Annals of Mathematical
Statistics, 25, 315–345. https://doi.org/10.1214/aoms/1177729380
References 215
CHAPTER 3
Tsangari, H., & Akritas, M. G. (2004). Nonparametric ANCOVA with two and
three covariates. Journal of Multivariate Analysis, 88, 298–319. https://doi.
org/10.1016/s0047-259x(03)00098-8
CHAPTER 5
Chatterjee, S., & Hadi, A. S. (2012). Regression analysis by example. Hoboken,
NJ: Wiley.
Cleveland, W. S. (1979). Robust locally weighted regression and smoothing
scatterplots. Journal of the American Statistical Association 74, 368, 829–836.
doi:10.2307/2286407
Cleveland, W. S. (1981). LOWESS: A program for smoothing scatterplots by
robust locally weighted regression. The American Statistician 35, 54. https://
doi.org/10.2307/2683591
Cleveland, W. S., & Devlin, S. J. (1988). Locally- weighted regression: An
approach to regression analysis by local fitting. Journal of the American
Statistical Association, 83, 596–610. doi:10.2307/2289282
Dillman, D. A., Smyth, J. D., & Christian, L. M. (2014). Internet, phone, mail,
and mixed-mode surveys: The Tailored Design Method. Hoboken, NJ: Wiley.
Grande, T. L. (2015). Testing the assumption of independent errors with
ZERESID, ZPRED, and Durbin-Watson using SPSS. Retrieved from https://
www.youtube.com/watch?v=Lyk_S9T-oCw
Raudenbush, S.W. (1993). Hierarchical linear models and experimental design.
In L. K. Edwards (Ed.), Applied analysis of variance in behavioral science.
New York: Marcel Dekker.
APPENDIX B
Allison, P. D. (2001). Missing data. Thousand Oaks, CA: Sage Publications, Inc.
Enders, C. K. (2010). Applied missing data analysis. New York: Guilford Press.
Rubin, D. B., & Roderick J. A. L. (2002). Statistical analysis with missing data.
New York: Wiley.
APPENDIX C
Books and Book Chapters
Corder, G. W., & Foreman, D. I. (2009). Nonparametric statistics for non-
statisticians: A step-by-step approach. Hoboken, N.J: Wiley.
216 References
Articles
Carroll, R. J., Ruppert, D., Crainiceanu, C. M., Tosteson, T. D., & Karagas, M. R.
(2004). Nonlinear and nonparametric regression and instrumental variables.
Journal of the American Statistical Association, 99, 736–750. http://d x.doi.
org/10.1198/016214504000001088
References 217
Friedman, M. (1937). The use of ranks to avoid the assumption of normal-
ity implicit in the analysis of variance. Journal of the American Statistical
Association, 32, 675–701. http://d x.doi.org/10.1080/01621459.1937.10503522
Friedman, M. (1939). A correction: The use of ranks to avoid the assumption
of normality implicit in the analysis of variance. Journal of the American
Statistical Association, 34, 109. http://d x.doi.org/10.2307/2279169
Geyer, C. J. (2005). Fuzzy p-values and ties in nonparametric tests. Retrieved
from http://w ww.stat.umn.edu/geyer/f uzz/ties.pdf
Han, L. (2008). Calculating the point estimate and confidence interval of Hodges-
Lehmann’s median using SAS software. Proceedings from the Southeastern
SAS Users’ Group (SESUG). Retrieved November 20, 2016 from http://ana-
lytics.ncsu.edu/sesug/2008/ST-154.pdf
Kenny, D. (November 24, 2015). Measuring model fit. Retrieved from http://
davidakenny.net/cm/fit.htm
Kraemer, H. C. (1976). The small sample nonnull properties of Kendall’s coef-
ficient of concordance for normal populations. Journal of the American
Statistical Association, 71(355), 608–613. http://d x.doi.org/10.1080/
01621459.1976.10481536
Park, H. M. (2003). Estimating regression models for categorical dependent vari-
ables using SAS, Stata, LIMDEP, and SPSS. Retrieved from https://w ww.ikhe-
beenvraag.be/mediastorage/FSDocument/101/LogitProbit.pdf
Sawilowsky, S. S. (1990). Nonparametric tests of interaction in experimental
design. Review of Educational Research, 60(1), 91–126. http://d x.doi.org/
10.3102/0 0346543060001091
Steele, J. M. (2005). Darrell Huff and fifty years of how to lie with sta-
tistics. Statistical Science, 20, 205–209. http://d x.doi.org/10.1214/
088342305000000205
Warner, C. B., & Meehan, A. M. (2001). Microsoft Excel as a tool for teach-
ing basic statistics. Teaching of Psychology, 28, 295–298. http://d x.doi.org/
10.1207/S15328023TOP2804_11
Wilcox, R. (2009). Comparing Pearson correlations: Dealing with heterosce-
dasticity and non-normality. Communications in Statistics—Simulation and
Computation, 38, 2220–2234. http://d x.doi.org/10.1080/03610910903289151
Wilcox, R. R. (2005). Depth and a multivariate generalization of the Wilcoxon-
Mann-W hitney test. American Journal of Mathematical and Management
Sciences, 25, 343–361. http://d x.doi.org/10.1080/01966324.2005.10737655
Websites
Data analysis examples by Institute for Digital Research and Education at
UCLA http://w ww.ats.ucla.edu/stat/dae/
218 References
A C
assumptions, 2, 4, 5, 7, 10–15, 40, 43, 58, CFA, 161
73, 82, 88, 150, 172, 189, 193 coverage, 28, 40, 55, 117–8, 189–90, 193
equal interval, 17 response, 6, 17, 108
equal variances, 10 cut point, 51–3, 55, 103
heteroscedasticity, 13
homogeneity of variance, 7, 10, D
88, 193 dependent and independent variables
homoscedasticity, 7, 11, 150, dependent variable, 11–3, 101, 149, 153,
153, 189 171, 184, 188
independence, 6, 150, 172, 189 independent variable, 5, 11–3, 101, 149,
independent groups, 69–70, 88–9, 95 153–4, 171
independent observations, 7, 15 distributions, 73, 79, 83, 93–5, 108,
linear relationship, 7, 11–12, 116, 131
57–8, 150 bell-shaped curve, 7–8
multicollinearity, 171 binomial, 116
nonlinearity, 57 curvilinear, 12, 160
normality, 7–9, 47–8, 57, 88 distribution-f ree, 5, 89
parametric assumptions, 5, 7, exponential, 43, 45, 47, 184
10, 151 frequency, 154
unequal variances, 10, 57 kurtosis, 73, 154
non-normal, 13, 20, 57
B normal, 7–8, 12–3, 43, 54, 95, 101, 130,
bias, 59, 68, 157, 161, 189, 210 132, 153, 155, 176, 189
book conventions, 22 Poisson, 43, 45, 47, 63
bootstrapping, 3 probability distribution, 43
219
220 Index
distributions (cont.) M
skewness, 73, 131, 154 marginal probabilities, 111, 117
uniform, 43, 45, 47 measures of central tendency
mean, 5–6, 9, 16, 48, 52
E median, 16, 47–50, 73–4, 77, 91–3, 97, 101,
EFA, 161 103–4, 120, 123–5, 128, 131
effect size, 6, 19, 21, 87, 120 mode, 16
expected vs. observed, 40, 101 missing cases/missingness
missing, 4, 33–4, 87, 115, 123, 157, 161
F Monte Carlo simulation, 9, 19
filter, 60–1, 63–4, 87, 90, 96, 102 MSE, 164
frequency, 32–3, 88, 101, 118, 124, 132
N
G Nonlinear relationship, 150, 171
GLM, 188, 194, 196–7 nonparametric statistical procedures
group comparison, 29, 70–1 binomial test, 27, 31–3, 35–6, 39–41
guideline, 5, 18, 29, 33, 87, 118, 176, chi-square test, 27, 31, 40–2 , 109
190, 193 Clopper-Pearson, 36, 39
Cochran’s Q test, 108, 110, 132,
H 134, 137
HLM, 1 Durbin-Watson test, 150
hypothesis Friedman’s 2-way ANOVA, 108, 110,
alternative, 5, 94 142–3, 145–6
null, 5–6, 18, 21 Hodges-Lehmann confidence interval,
test, 2–3, 32, 37 73, 77, 110, 125, 128, 131
Hodges-Lehmann estimate, 69, 73
I Jonckheere-Terpstra for k samples, 69,
independent groups, 4, 68–9 72, 94–8, 100
independent samples, 69, 86–8, 109 Kendall rank correlation coefficient, 57
interventions, 10, 21, 78, 88, 93–4, 100 Kendall’s coefficient of concordance,
research, 1 110, 137–9, 142–3
IQR, 77 Kendall’s tau-b, 28, 31, 57–8, 64–7
Kendall’s W, 137
L Kolmogorov-Smirnov test, 27, 31, 43–7,
LCA, 163 63, 69, 72, 79–84, 86, 88
levels of measurement, 5, 7, 15, 41, 50, Kruskal-Wallis one-way ANOVA, 69,
108, 113, 189 72, 88–9, 91, 93–5, 97
categorical, 16, 22–4, 28–9, 36, 40, 44, Levene’s test, 193–4, 196
101, 117–8, 134, 171, 175, 188–90 Mann-W hitney U test, 69, 82–4, 86, 88
continuous, 7–8, 16–7, 23, 27, 36, Marginal homogeneity test, 107, 109,
43–4, 171 117, 119–20
dichotomous, 16, 28–9, 31–2 , 35, 40, 50, McNemar’s test, 107, 109–17
55, 111, 132 median test, 69, 101–4
interval, 5, 16–17 moses test of extreme reactions,
nominal, 5, 16, 22, 49, 58, 89 69, 73, 77
ordinal, 5, 16–19, 22, 40, 46–7, 57–8, 97 Nonparametric regression (LOWESS),
ratio, 5, 16, 23 149, 153, 165
scale, 7–9, 16, 23, 28–9, 43, 46, 48, Pearson’s chi-square test, 109
52, 107–8 rank order, 16, 58
Index 221