Simplifying Statistics For Graduate Students.
Simplifying Statistics For Graduate Students.
All rights reserved. No part of this book may be reproduced in any form or by
any electronic or mechanical means, including information storage and retrieval
systems, without written permission from the publisher, except by a reviewer who
may quote passages in a review.
Introduction ix
Chapter 1 Variables and How to Measure Them 1
Chapter 2 Intelligently Managing Your Data 11
Chapter 3 Graphing Techniques to Support Your Data 17
Chapter 4 Means, Medians, and Modes and When to Use Each 27
Chapter 5 Measuring Variability: An Important Role in Data 35
Chapter 6 Random Sampling and Other Useful Sampling Strategies 47
Chapter 7 Stating Hypotheses and Hypothesis Testing 55
Chapter 8 t-Test Procedures 63
Chapter 9 ANOVA Procedures 75
Chapter 10 Chi-Square Procedures 85
Chapter 11 Correlation Procedures 93
Chapter 12 Regression Procedures 103
Chapter 13 Practical Tips for Graduate Students 113
About the Authors 121
vii
Introduction
ix
x • Introduction
and more importantly, examining them in detail, are some of the most valu-
able steps in data management.
Chapter 3 presents graphing techniques that are most beneficial to gradu-
ate students’ research. Correctly choosing and then setting up graphs are
discussed in this chapter along with some mistakes to avoid. Graphing can
be an asset in the ability for graduate students to communicate their results
in a simple, informative manner.
Chapter 4 begins the statistics presentation with descriptive statistics. The
mean, median, and mode are used often every day. This chapter explains the
three measures and when to use each one. Chapter 5 is a complement as it
takes chapter 4 and explains why variability is critical as a partner to de-
scriptive statistics. The benefit of understanding, and then using, the normal
curve and percentiles is discussed in this chapter.
Chapter 6 shows graduate students the different sampling strategies. It
explains the process of random sampling and also includes several other
sampling strategies that might be of use in different situations. Especially
important is the presentation of nonrandom sampling options, which may
be more practical for graduate student research.
Chapter 7 reviews hypothesis testing as a prerequisite to running statisti-
cal analyses. It is imperative to state the research question and corresponding
null hypothesis that statistics will be used to test. Type I and Type II errors,
along with the Bonferroni correction, are examined.
Chapters 8 and 9 begin the introduction of inferential statistics. T-tests
and ANOVA (analysis of variance) procedures are examined in a straightfor-
ward manner so that graduate students can confidently execute them. These
parametric statistics have application in many graduate research designs
where two or more groups are compared on some measure.
In chapter 10 the nonparametric statistic chi-square analysis is presented.
Because this procedure examines categorical variables such as demographics,
its usage to graduate students is immense.
Correlation and multiple regression are presented in chapters 11 and 12.
Correlation is a descriptive statistic that tests relationships among variables.
This allows graduate students to avoid speculating about what is and is not
related. The correlation coefficients do the work. Multiple regression is a
complex, multivariate procedure presented in a basic and uncomplicated
manner. This statistic has many applications but may require external exper-
tise to help execute it.
Chapter 13 gives graduate students advice on determining their research
topic, using library references, selecting a research design, and writing up re-
Introduction • xi
sults from their data. Strategies to do this with finesse, accuracy, and honesty
are offered to help graduate students make the most of their data.
Expertise with data is expected of graduate students. It is part of the rea-
son you are earning an advanced degree. Simplifying Statistics for Graduate
Students: Making the Use of Data Simple and User-Friendly is an antidote for
the research and statistics blues. It can provide graduate students with the
confidence they need when signing on to the research methods and statistics
courses required in their programs. Graduate students should enjoy the jour-
ney and this book can help.
CHAPTER ONE
1
2 • Chapter One
deaths from COVID-19. These are represented by whole numbers. If they are
continuous variables, they can take on fractions or decimals. Body weight,
credit card balance, body temperature, height, or race time are examples of
continuous variables, where race time in the 100-meter dash might be 13.2
seconds. See table 1.1 for a classification of variables.
Table 1.2. Three Nominal Scale Options for the Variable Residential Location
Option One Option Two Option Three
Numbers Numbers Numbers
Variable Assigned Variable Assigned Variable Assigned
Rural 1 Rural 2 Rural 3
Urban 2 Urban 3 Urban 2
Suburban 3 Suburban 1 Suburban 1
Some nominally scaled variables have only two categories. These are
called dichotomous variables. This means that only two numbers can be as-
signed to each of the categories, respectively. Some examples of dichotomous
variables may include primary language spoken at home (English/Spanish),
parent (father/mother), staff (supervisor/trainee), student (special education/
general education), loan applicant (accepted/rejected), patient (inpatient/
outpatient), and many others. You may use two numbers (such as 1 and 2 or
0 and 100 or 200 and 300) as the nominal scaling applied to each category.
Nominal scaling is ideal to categorize schools. The type of school the child
attends might be categorized as Elementary (1), Middle (2), and Secondary
(3). Or, if we have five elementary schools, we could nominally scale all
seven of the schools that the children in a town attend as in table 1.3.
Being able to nominally scale our variables this way allows us to drill down
our data or inspect our data much more thoroughly.
One pitfall threatens. Many individuals get confused about what the ac-
tual variable is and what are the respective categories. Sometimes, you might
think the variable is female instead of gender with categories. This differen-
tiation is very important, especially when you are investigating statistical dif-
ferences with independent and dependent variables, which will be discussed
later in the book. Ask yourself: What variable are the individual categories
representing? The answer should be the variable itself and not its categories.
in the ordinal scale. Unlike nominal scaling, the numbers in ordinal scales
have meaning. Ordinal scales give more information and more precise data
than nominal scales do. Here are three examples of ordinal-scaled variables:
Variable: Frequency
Never (1)
Rarely (2)
Sometimes (3)
Frequently (4)
Always (5)
Variable: Satisfaction
Dissatisfied (1)
Satisfied (2)
Very Satisfied (3)
Variable: Attainment
Below Expectations (1)
Meets Expectations (2)
Exceeds Expectations (3)
One of the most common uses of ordinal scaling is with ratings, preferences,
rankings, goal attainment, satisfaction, degrees of quality, and agreement
levels—typical of the Likert scale found on questionnaires. Here is an ex-
ample of ordinal scaling using a rubric to measure student performance:
Beginning (1)
Developing (2)
Proficient (3)
Advanced (4)
Chapter Summary
As a summary of the four measurement scales, they can be remembered by
the following characteristics:
Frequency Distributions
One of the most useful expenditures of time is to produce a frequency distribu-
tion for each of the variables in your data set. This is a systematic arrange-
ment of numeric values from the highest to the lowest—with a count of the
number of times each value was obtained. It is a procedure for organizing and
summarizing data into a meaningful representation. It does not tell you ev-
erything about your data, but it provides a beginning and convenient way of
grouping data so that meaningful patterns can be found. Statistical packages
like SPSS can do this for you, but here is how you do it if you did it yourself.
11
12 • Chapter Two
Table 2.1. Data Set for Daily Website Clicks for 50 Days
Week Monday Tuesday Wednesday Thursday Friday
1 699 459 450 450 445
2 420 420 420 420 420
3 430 420 445 435 467
4 420 320 420 320 420
5 689 445 479 450 467
6 435 430 430 430 420
7 420 467 467 450 435
8 320 320 320 320 420
9 430 445 430 435 467
10 420 420 320 420 467
The lowest number of clicks is 320 and the highest is 699. The frequency
distribution orders the 50-day click rates from lowest to highest and then
tally marks are placed to show the frequency each was obtained. Percentages
(the relative frequencies) are calculated to provide even more information to
you. Finally, cumulative frequencies and cumulative percentages are added to
complete the picture (review table 2.2).
From this frequency distribution, you can draw some important conclusions:
• The most frequently occurring click rate was 420—the most typical for
30% of the days.
• More than half (56%) of the click rates were 430 or lower.
Intelligently Managing Your Data • 13
• There is a broad range of click rates from highest (699) to lowest (320).
• Two click rates (699 and 689) were very different from the other
rates—outliers.
Class Intervals
Let’s say that you have twice as many days in your sample: 100 instead of 50.
You might use a shorthand called class intervals. The number of classifications
of data (classes) can be reduced by combining several of the actual clicks
into an interval or band of values. You are essentially consolidating the data
points into bundles to make them more manageable and comprehensible.
A good rule of thumb is to have between 10 to 20 class intervals, alto-
gether. This is particularly true if you intend to graph your frequency distri-
bution data into a frequency polygon or histogram, which will be discussed in
the next chapter. Ten to 20 class intervals will summarize your data without
distorting the shape of your graph. Too few intervals compress the data and
thus conceal meaningful changes in the shape of your graph. Too many in-
tervals stretch out the data, so they are not summarized enough for a clear
visualization.
There are two general rules for using class intervals:
• The class interval should be of such size that between 10 and 20 inter-
vals will cover the total range of values. This provides for a manageable
number of intervals without losing the general shape of the distribution.
• Whenever possible, the width of the class interval should be an odd
rather than an even number. Under this rule, the midpoint of the
14 • Chapter Two
interval will be a whole number rather than a fraction. This will be-
come important when you graph your data.
Table 2.3 displays what our frequency distribution with class intervals would
look like.
Table 2.3. Frequency Distribution of Daily Website Clicks with Class Intervals
(N = 100 days)
Class Intervals Frequency Relative Cumulative Cumulative
of 45 Tallies (f) Frequency Frequency Frequency
300–344 / 1 1% 1 1%
345–390 ////// 6 6% 7 7%
391–436 ///////////////////////// 25 25% 32 32%
437–482 ///////////////// 18 18% 50 50%
483–528 /////////// 11 11% 61 61%
529–574 //////////// 12 12% 73 73%
575–620 ///////// 9 9% 82 82%
621–666 ////// 6 6% 88 88%
667–712 ////// 6 6% 94 94%
713–758 ///// 5 5% 99 99%
759–804 / 1 1% 100 100%
Intelligently Managing Your Data • 15
As a note, using class intervals is less cumbersome when your data are con-
tinuous and have a wide range of values. Think about the variable of age
and the adults in your community if you were conducting a survey of city
residents. Listing all possible ages might begin at 18 and go beyond 100 or
more years! Doing a tally of each and every age would require a lot of paper,
never mind time. This is a case when categorizing your variable into class
intervals is a good idea. You might use intervals of five years in this case so
those class intervals would look like this. (There would be less than 20 in-
tervals for adult ages.)
18–22
23–27
28–32
33–37
38–42
43–47
48–52
53–57
58–62
63–67
68–72 and so forth.
On the other hand, there are times when using class intervals shortchanges
the picture; you lose valuable data. Think about the variable “years that
nurses have been employed by your hospital.” If you used class intervals, you
would lose key information. A nurse who is brand-new, one who has been
with you for only one year, and one who has been on staff for three years may
be very different from each other. If you collapsed the “years employed” into
intervals of five years as shown with the age data, you would lose important
insights. Much of this is common sense. What is the information you want
to uncover with your data?
16 • Chapter Two
• You can identify atypical or odd values like that in table 2.2 (689 and
699). You have to be able to identify outliers before you can determine
what you should do with them. And it is important to decide whether
to keep or discard outliers. They can skew your results!
• You can compare the spread the values, visually. If you set up two fre-
quency distributions representing two segments of your data (such as
Weeks 1–5 and Weeks 6–10), you might be able to infer why the click
rates varied.
• By indicating the frequency that each value occurred in the distribu-
tion, this information could help you to make decisions about statistical
analyses you will execute later. This is particularly true with categorical
variables. For example, if you wanted to categorize click rates into high,
moderate, and low, you might look at the percentages to classify them.
• Finally, the frequency distribution may reveal data entry errors that you
typed in accidentally. If you had a click rate of 9 or 9999 that showed
up in your frequency distribution, you would know that that was an
error. You would be able to make your correction and not run the data
blindly, compromising the purity of your data set. This happens often
when frequency distributions are not run on all the data before the sta-
tistical analyses are executed. Take the time to do the basic work before
you plunge into the analytics.
Chapter Summary
The process of becoming an expert with your data takes work. Setting up the
frequency distribution for your data set, graphing it, and then interpreting
what it conveys, are often-skipped steps in data management. The time it
takes to do the “grunt” work as a preliminary step is a meaningful invest-
ment of time. There are advantages to be gained in many areas. You will be
able to reduce errors, make wise decisions about statistical techniques, and
ultimately provide a better foundation for interpreting your results. Take
the time to know your data set. It will be well worth the hours spent. Don’t
shortchange yourself and compromise your research quality.
CHAPTER THREE
Graphing Techniques
to Support Your Data
• If the curve has a hump in the middle, its symmetry means that you
have most values in the middle of the range—few high and few low.
(Normal curve, as in figure 3.1.)
17
18 • Chapter Three
• If the hump of the curve is inclined toward the left, you have more low
values in your data set. The tail or end of this type of curve indicates
atypical or odd values in your data. In this case, there was a minority of
high values. (Positively skewed, as in figure 3.2.)
• If the hump of the curve is to the right, you have more high values in
your data set. The tail or end of this type of curve indicates atypical or
odd values in your data. In this case, there was a minority of low values.
(Negatively skewed, as in figure 3.3.)
Graphing Techniques to Support Your Data • 19
• If the hump of the curve is flat, you have values that are spread out a
great deal. Values are all over the map. (Platykurtic as in figure 3.4.)
• If the hump of the curve is peaked, your values are very, very similar.
There is not much difference among values. (Leptokurtic as in figure
3.5.)
20 • Chapter Three
The discussion about the shapes of graphs will have more meaning in subse-
quent chapters.
Types of Graphs
Frequency Polygons
To construct a graph from a set of data, we start with the frequency distribu-
tion. A frequency distribution can be graphed into a frequency polygon. This
is how you convert a frequency distribution into a frequency polygon.
Draw a vertical side or ordinate axis of your graph. This is called the Y axis.
On the Y axis it is an accepted practice for the frequencies to be plotted. The
horizontal axis, called the abscissa, is used to plot the variable that you are
displaying data for. The horizontal axis is the X axis. As a rule of thumb for
good visual presentation, the vertical axis should be roughly two-thirds the
length of the horizontal axis.
Connect the points with a dot that intersects the values on the abscissa
axis to the frequencies on the ordinate axis. If class intervals are used, con-
nect the midpoint of the interval on the horizontal axis with the frequency
on the vertical axis. (This is why it is helpful to use odd numbers for the
span of your class interval. The midpoint is easy to identify.) As a note, if
the range in frequencies is large, you may want to start the vertical axis with
a value that is not zero (0).
Plotting the data this way allows the graph to take on a shape. When
you look at that shape, you can make a conclusion about the data from the
Graphing Techniques to Support Your Data • 21
frequency distribution. When there are a lot of data from many values, the
.
frequency polygon looks like a smooth curve. When there are fewer values,
it is jagged. Either way, the picture tells a story. The graph is a picture that
gives us an immediate message about our data. We look at it and we can infer
something about our data.
For our daily website click graph with class intervals, the simple message
is that there are many low values in this distribution, and it is a positively
skewed curve (see figure 3.6).
Bar Graph
A bar graph is a very convenient graphing device that is particularly useful
for displaying nominal data such as gender, ethnicity, and other categorical
variables. The various categories are located along the horizontal axis. The
frequency or count, as always, is on the vertical axis. The height of each bar
is equal to the frequency for that category. This is very helpful for viewing
the differences in the data for individual groups on some variable.
If your dissertation or thesis were presenting demographics on households
with and without children, and there were 200 households in your sample,
the results in the bar graph would be obvious. Even without the actual num-
bers inside the bars, the two columns of the bar graph show that twice as
many respondents did not have children as compared with those who did.
Please refer to figure 3.7.
22 • Chapter Three
Histograms
A histogram is another pictorial representation of a frequency distribution.
While bar graphs are useful for nominal-scaled, categorical variables, the
histogram is preferred if there is a quantitative variable where categories have
gradations such as age or income. A general rule in laying out the histogram
is to make the height of the vertical axis equal to approximately two-thirds
the length of the horizontal axis. Otherwise, the histogram may appear to be
out of proportion.
A vertical bar is constructed above each respective category equal in
height to its frequency count. All the rectangles have equal width. This is
a very common way to display data. Figure 3.8 displays what a histogram
would look like if you were reporting age segments. The age differences in
the respondents can immediately be seen. There are many respondents who
were older.
Pictographs
A pictograph uses pictures or symbols in place of numbers. Figure 3.9 portrays
a pictograph of the number of books read by a fourth-grade class during the
holiday vacation week. It uses a book icon, representing five books read for
each day Monday through Friday.
Pie Charts
Another method to display data is by using a pie chart. This graph is very easy
to generate and is user-friendly. It shows the categories of a variable by divid-
ing up “the pie.” Use 100% of the area in the circle and divide it up propor-
tionally to the categories of interest. Figures 3.10 and 3.11 display pie charts
that reflect the same data on Household Composition and Age Segments for
respondents, as depicted on the histogram and bar graph, previously shown.
The graph in figure 3.12 reports a “steep decline in SAT values.” But look
how the vertical scale has been adjusted to begin at 492, distorting the
picture. The drop over a three-year period is negligible—from 502 to 496!
An initial glance at the graph would cause board of education members to
faint—before they could find out the real story in the actual data.
Figure 3.12. “Steep” Decline in SAT Scores over Three Years. Source: Author developed
26 • Chapter Three
Chapter Summary
Graphs are most beneficial to graduate students because they can provide a
summary data sheet when presenting data. However, it is important to keep
in mind that a graph is a picture that is helping to tell a story about the data.
Those who will be viewing the graph are not as involved in your data and
analysis as you are and have other thoughts competing for their attention.
Don’t waste their time with graphs that are either too simplistic or overly
complex. Use graphs judiciously and with deliberation.
CHAPTER FOUR
The importance of getting to know your data has been discussed. The
frequency distributions and the graphing techniques both produce a basic
profile of your data set. Taking the time to construct them helps to indicate
the characteristics of your data. There are also certain statistics that are
generated for the purpose of describing the distributions of your data or the
relationships between your variables. They are called descriptive statistics.
These very useful statistics bring together large amounts of data so they can
be presented and comprehended with minimal effort.
Descriptive statistics are widely applied. A good example of real-life ap-
plications is the US Census. By using some of the popular descriptive statis-
tics, we get a sense of important characteristics of households in the United
States. For example, descriptive statistics that are available in census data
may indicate:
27
28 • Chapter Four
The Mean
The mean represents a whole data set with one single number! The mean is
the arithmetic average in your distribution of values. To obtain the mean you
add up all of the values (Xs) and divide by the total number of values (n) in
your distribution. As a rule, your data set should tend to cluster together and
not be spread all over. The beauty of the mean is that you utilize each and
every value in your data set to calculate it. As a result, it is the most stable
measure of central tendency and the one used most often. Since it is an aver-
age, you probably have been calculating a mean often in your everyday life.
In professional journals, the mean is often noted as M.
Here is a simple example to illustrate the use of the mean. During this
past school year, the board of education was contemplating the creation
of additional elementary school classroom space. The three neighborhood
elementary schools were brimming with children. This was due to recent
out-migrations, where families were moving from cities to a more rural lo-
cation. The board members decided to find out what the average class size
was in order to arm themselves with supporting data for overcrowding and
ultimately for additional classroom space.
Although class sizes were compiled for all grade levels, the data set for ten
third-grade classrooms was reported as follows in table 4.1.
The mean was calculated at 27 students per classroom—larger than what
was desirable. Here was the evidence the board of education needed. With
just one number representing all 10 classrooms, the board had a solid piece
of information.
The mean is a very good measure of central tendency.
Means, Medians, and Modes and When to Use Each • 29
Outliers
An outlier is an extremely high or low value in your distribution. It is atypi-
cal and does not resemble most of the other numbers in your distribution. It
affects the calculation of the mean and makes the mean less representative.
While the mean is looking to create a typical value in your data set, the
outlier is atypical.
What happens is that outliers cause the mean to shift in the direction of
the outlier. If the outlier is a high value, the mean is calculated higher than
it should be. If it is a low value, the mean is lower than it should be. Outliers
30 • Chapter Four
Here, the outliers caused the mean to shift down to 23 children per class-
room. The presence of the outliers decreased the case for additional class-
room space, but it is not a good representation of average class size in your
school system. Most class sizes in your data set distribution are around 28 to
31. If you used the mean as the measure of central tendency in this case, you
would be presenting an underestimated picture of need.
What should you do? Examine your data through the basic work of devel-
oping a frequency distribution, and then use more than one measure of central
tendency. One of the best, when your data has outliers, is the median.
Means, Medians, and Modes and When to Use Each • 31
The Median
When there are extreme values or outliers in your distribution, the median
is the preferred measure of central tendency. The median is the counting
average. It is simply the midpoint in your distribution of ranked, ordered
values. The median can be used with ordinal, interval, and ratio-scaled data.
Nominal-scaled data are inappropriate.
To calculate the median, list all of the values in your distribution from the
lowest to the highest and then find the midpoint—the place where it divides
your distribution into equal halves. That is, 50% of all values are above and
50% are below.
If you had 11 values (an odd number), the formula would be (11 + 1) ÷ by 2
= 6. The median would be the sixth value in the set of data where the values
are listed in order from lowest to highest.
Here is an example where we have an odd number (11) of values. The
midpoint is the sixth value or the exact middle of the distribution. In this
distribution, the median is 29, and 29 is an actual value in the data set. There
are five values above and below it in table 4.3.
The median and midpoint of the distribution is 30. However, note that
the median is not an actual value in the set of data. As you can see, the
median is not affected one bit by the outliers. While the mean was 23, the
median of 30 is much more representative of the data set. This is the greatest
benefit of using the median. Oddball values in the data set do not affect it.
Although the median is not calculated by using all the values in the distribu-
Means, Medians, and Modes and When to Use Each • 33
tion, as the mean is, the median plays an important role in stating the central
tendency of your data set—especially if there are outlier values.
The Mode
The mode is an unsophisticated measure of central tendency. It is the most
frequently occurring value in your distribution. It is a simple but rough statistic
to calculate the central tendency of your data. The mode does not need to
be calculated. If you look at your frequency distribution, a simple eyeball
inspection of your data can tell you which value occurred most often. Look
at the tallies or frequencies. It is quick and can be obtained with a glance.
We did that in chapter 2 with our first data set of values in table 2.2. We
looked at the frequency distribution and could see that the mode was 420
for click rates.
The mode provides little information beyond simply identifying the value
in your data set that appears with the greatest frequency. Therefore, it should
only be used when you have a large data set of values, and not just a few. A
small set of values would not have enough frequency of occurrence to de-
velop a mode. You need many values so that you can be certain which value
turned up most often.
When there is one value that occurs with the most frequency in a distri-
bution, we say the distribution is unimodal or when graphed it has one hump
due to the fact that there is one mode. Many frequency distributions have
more than one mode. That is, more than one value turns up at the same high
level of frequency. Two modes in a frequency distribution create a bimodal
frequency distribution. This might occur in a set of data where two groups
are very different on the variable measured. If graphed, there would be two
humps in the curve or shape. If there are more than two modes, the distribu-
tion is called multimodal and the graphing shows several humps or curves.
Chapter Summary
Each measure of central tendency has a role to play in descriptive statistics.
The mean is the preferred method of calculating the center of your data set.
But when outliers emerge, the median is an alternative that is not affected
by these unrepresentative values. If you want to get a ballpark feeling for the
average, the mode will work with a quick glance at the frequency distribution
table. Here is a summary of the three measures and when to use each:
34 • Chapter Four
The Mean
• If you want the greatest reliability
• If you will be calculating variability and other statistical computations
• If your distribution has no outliers
• If your data are interval, ratio, or ordinal scaled
The Median
• If your distribution is skewed by outliers
• If your data are interval, ratio, or ordinal scaled
The Mode
• If you need a quick estimate
• If you have interval, ratio, or ordinal scaled data
• If you can eyeball the data from a frequency distribution
CHAPTER FIVE
Measuring Variability
An Important Role in Data
So far, a case has been made about the importance of constructing a fre-
quency distribution and plotting the data into a graph for an immediate
snapshot of findings. Similarly, the value of the three measures of central
tendency was established. They can indicate to us with a single number the
general characteristics of our set of data. There is another essential step:
measuring the variability of data sets. Many times statistics are reported with-
out mention of the variance or spread of values. The media is guilty of this.
Doing so can misrepresent data. This is why.
The variance is the manner in which your data are spread in either direc-
tion from the center of your distribution. It is important to know whether the
values tend to be homogeneous or whether they vary from each other. The
measures of variability tell us how representative our mean (M) is. Are the
values similar or are they spread out?
For example, look at the glucose levels for patients exposed to prediabetes
interventions, either Treatment A or Treatment B, at a hospital. The mean
blood sugar level in both cases is 100, so both treatments seem to be equally
effective. Closer inspection of the data shows that, while those in Treatment
A were very much the same at the end of the treatment, Treatment B had
some very high and very low results. For Treatment A, the lowest result was
98 and the highest was 102, not too far apart. For Treatment B, the lowest
result was 79 while the highest was 140. The spread or dispersion was great in
Treatment B, where there was high variability. Conversely, the low variabil-
ity in Treatment A was evident from the similar values. There are practical
35
36 • Chapter Five
There are several descriptive statistics that measure variability: the range,
the variance, and the standard deviation.
The Range
The range is a very simple statistic and the most unsophisticated measure of
variability. In this regard, it possesses characteristics analogous to the mode in
central tendency. It is a rough estimate, quickly computed, but not tremen-
dously stable. This is because it is computed with only two values: the highest
value and the lowest value in your distribution of values. You subtract one
from the other. The range for Treatment A would be 4 (102 – 98) while the
range for Treatment B would be 61 (140 – 79). Even this crude measure of
variability is very helpful in providing us with information that the mean
value for Treatment B is not stable.
The range is influenced by the size of your data set. The larger the data set,
the greater the likelihood of extreme values because you have more potential
for outliers. This is a limitation of the range; outliers affect it. Since you are
using only two values in your data set to calculate the range, if there is an
outlier at either end, it will influence the calculation.
histograms that are constructed from the frequency distribution can tell us
the story about the spread of values at a glance. This is another asset of the
frequency distribution (and their respective graphs), and a further reason to
use them in getting to know your data set.
If the hump or curve in your graph is markedly flat, the values are spread
out; there is a great deal of variability. Subjects are not alike. This type of
curve, called platykurtic, means that your values are spread out around the
mean value of the distribution. On the other hand, if the hump or curve in
your graph is peaked and tight, there is little spread in values; the variability
is low. Subjects are alike. This type of curve, called leptokurtic, means that
your values are very close to the mean value of the distribution. (Please refer
to the corresponding figures in chapter 3.)
Another example is illustrated in two graphs that display the number of
pounds that bank employees lost during a Get Healthy Competition. The
intervention group participated in a special exercise and nutrition program
and the comparison group did not.
For both groups the mean number of pounds lost was 6 pounds. As you can
see, employees in the intervention group performed similarly. The members
lost between 4.5 and 8.5 pounds each. On the other hand, the comparison
group lost between 1 pound and 11 pounds. In fact, the range was 10 (11 – 1)
for the comparison group and only 4 (8 – 4) for the intervention group! This
difference in spread of pounds lost is different between the two groups. But
the mean is 6 pounds for each group. Please refer to figures 5.1 and 5.2, which
present visuals that indicate the variability or spread of values.
Conceptually, each of these statistics is based upon how much each value in
your data set deviates from the mean, and then putting the deviation val-
ues into a formula for computation. The variance is the standard deviation
squared, and the standard deviation is the square root of the variance. They
are more or less “siblings” in the realm of variance. The symbol for reporting
the variance is “sigma squared” (ơ²). The symbol for standard deviation is SD
or sigma (ơ). These are what you might see in professional journals.
Because both the variance and standard deviation are calculated by de-
viations from the mean in your data set, their value is tremendous. With
one single number, you can tell whether most of the values in your data set
cluster closely around the mean or are spread out. The larger the standard
deviation, the more spread out are your values in your data set.
For example, let’s consider customer satisfaction scores over a two-year
period for a restaurant chain where 900 was the highest possible score. Last
year they had a mean of 600 and a SD of 50. This year they had a mean of
610 and SD of 100. Our mean (M) customer satisfaction scores rose from
600 to 610. What appears to be good news for our restaurant chain is in
fact a hidden problem. The standard deviation of 100 shows outlier values
have shifted the mean upward and artificially created the image that all of
our values were rising. While we can celebrate the seemingly good news this
year, it puts a strain on next year when we have to explain to the CEO why
the values are back down to (or below) 600. If the standard deviation were
calculated in addition to the mean, there would be a more accurate presenta-
tion of our restaurant’s performance.
Here is sound advice: Whenever you report the mean, you should report the
standard deviation. They are the statistical couple. The standard deviation
tells you a great deal: how representative your mean is as a measure of central
tendency for your data set.
Look at the data for the number of motor vehicle accidents occurring in three
towns of about the same size and demographic profiles over a five-year period
(see table 5.2):
At face value three towns have the exact same number of accidents
among residents: 60. So, you assume that this is typical or average. However,
the standard deviations tell a very different story. For Torrington and Litch-
field, the means are very representative of the average. The SD is small for
both. If you actually drew a frequency polygon from the frequency distribu-
tion data per month, these two towns would have leptokurtic curves. The
values would be close to the mean.
But this is not true for Winsted. This town has a greater spread of values
around its mean of 60, as signified by the standard deviation of 15. The fre-
quency polygon for Winsted data would be flat, not peaked. And actually,
the median might be the better measure of central tendency in this case.
When you draw a normal curve, you observe these characteristics. The
normal curve, when drawn in most statistics books, has ordinates or vertical
lines from the top of the curve to the baseline. When you look at the picture
of a normal curve, visualize thousands of people standing under the curve;
imagine faces looking at you. This is what the normal curve is representing—
data of individuals on whatever variables you are examining.
In the normal curve, the vertical lines are drawn from the top of the curve
to the baseline at 0 and at numbers 1, 2, and 3 to the left and to the right of 0.
Zero (0) represents the mean value or dead center. These vertical lines mark
Measuring Variability • 41
off areas under the curve and represent standard deviation units (1, 2, and 3)
or distance from the mean. The standard deviation (SD) units act like a ruler
and divide up the area under the normal curve as illustrated in figure 5.3.
Figure 5.3. Normal Curve, Standard Deviations, and Areas Under the
Curve. Source: Author developed
The base of the normal curve is divided into six units with three SD
above and three SD below the mean. Entries within the graph indicate the
proportion of the total area (or number of values) that falls into each of the
six segments or demarcations. These are the proportions of area under the
SD cutoffs:
• between the mean (0) and +1 SD units are 34.1% of all values
• between the mean (0) and –1 SD units are 34.1% of all values
• between ±1 SD are 68% of all values
Since we know that the mean is dead center, it is at the 50th percentile
mark. The next bit of information is important in interpreting values:
42 • Chapter Five
• One standard deviation above the mean is an area of 34%. This indicates
that if your value is 1 SD above the mean, you are higher than 84% of
all other values in the distribution (50% plus 34%).
• One standard deviation below the mean is an area of 34%. This indicates
that if your value is 1 SD below the mean, you are higher than only
16% (50% – 34%).
• One standard deviation above and below the mean is an area of 68%. Over
two-thirds are within one standard deviation of the mean (±1 #SD). If
you look at the normal curve and pretend again that people are stand-
ing underneath it, then this is very logical. It is where most values fall,
creating the large hump.
• Two standard deviations above the mean cover an area of 48%. Between
the center (0) and +1 SD there is 34% of the area under the curve.
Between +1 SD and +2 SD there is 14% of the area under the curve.
As you can see, the further you get from the center, the less area there
is under the curve. This is because most values are in the center where
the hump is. Again, think of people standing underneath the curve. If
you obtain a value that is two standard deviations above the mean, that
is a select few. This value is better than 98% of all others (50% + 48%).
• Two standard deviations below the mean cover an area of 48%. Between
the center and –1 SD there is 34% of the area under the curve. Between
–1 SD and –2 SD there is 14% of the area under the curve. If you get
a value that is two standard deviations below the mean, this is in the
minority. The value is better than only 2% of all others (50% – 48%).
• Two standard deviations above and below the mean is an area of 95%. If
you look at the normal curve and pretend again that people are stand-
ing underneath it to create the hump in it, then the percentage of 95%
is very logical. Almost all people perform within two standard deviation
units of the mean (±2 SD).
The standard deviation units, together with the percentage of area under the
normal curve, can be used to determine the relative position of values above
and below the mean. The information is defined as the “distance from the
mean.” Extremely useful, this is why.
All standardized measurement tools have means and standard deviations
reported. You can compare your data to the national norms. The normal
curve gives you the ability to make a call on where your sample of subjects
stands in relation to all others. This is because the percentiles under the
normal curve indicate the relative position of those falling above and below
a certain designation.
Measuring Variability • 43
Standard or Z Scores
A standard or z score is another useful tool. Its symbol is lowercase z. It tells
you how many standard deviations a raw score is from the mean. You cal-
culate a z score by taking a subject’s raw score (X) minus the mean (M) and
then dividing it by the standard deviation (SD) of your data. This is your z
score. If the z value is negative, then your z score is below the mean. If it is
positive, it is above the mean. If it is zero, it is the mean.
For example, the IQ test has a mean of 100 and a standard deviation of
15. Your IQ was 130. If you apply the formula, your standard score would be
+2.00 or two standard deviations above the mean. Referring back to area
under the normal curve, your score would be higher than 98% of those who
took the IQ test.
Most statistics books have a table that helps you convert your raw data to
z values, and some statistical programs will even do the conversion for you.
Skewed Distributions
Before closing this chapter, we need to address the fact that often our distri-
butions or data sets are not normally distributed. Don’t panic. This is the case
more often than not with thesis or dissertation data. Just as a review, normal
distribution curves have the following characteristics:
• Our frequency distributions will confirm this because they will have
outliers in them.
• The mean, median, and mode will not be equal.
• The frequency polygon will extend outward either left of right. This
extension is called a tail.
Chapter Summary
Reporting only part of the story about your data set occurs when measures
for variance are missing. You are looking at part of the information puzzle
but leaving off an important dimension. If your thesis or dissertation research
hinges on descriptive statistics, then using both central tendency measures
and measures of variation will be considered vital.
CHAPTER SIX
47
48 • Chapter Six
We use samples all the time to represent the whole. We take a sample
of exotic food, maybe a bite or two to see if it pleases our palette. We use a
sample of new software that comes in the mail to see if it really works as well
as the manufacturer purports it does. We take a sample of client references
to see if the consultant is as good as the proposal suggests. We visit a sample
of classrooms to determine if behavioral issues are increasing. We sample
some of the households in our community to see what languages are spoken
at home.
Why not use the entire population to draw our conclusions? The short
answer is it is a waste of time. For most purposes we can obtain suitable ac-
curacy, quickly and inexpensively, from a sample. Assessing all individuals
may be impossible, impractical, expensive, or even inaccurate. It is usually
not feasible to include an entire population in a study except if the number
of those in our population is small and manageable. Furthermore, statistics
make it unnecessary. A sign of ineptitude in data management is using an
entire population when a sample will provide the same results. You save
money, time, make fewer mistakes, and achieve the same end.
What is great about random samples is that you can generalize to the
population that you are interested in. If you randomly sample 300 households
in your community, you can generalize to the 50,000 households that live
there. If you match some of the demographic characteristics of the 300 with
the 50,000, you will see that they are surprisingly similar. Technically speak-
ing, if you calculated a mean for these 300 randomly selected households
and then drew another sample of 300 different households from the same
population as the first, your mean values (M) would be the same. This is the
beauty of random samples. If drawn scientifically, they represent the entire
population in which you are interested.
Many a doubting Thomas will still be skeptical about whether a sample
can truly represent the population. This is understandable. Yet if properly
conducted, sampling does work. Think about all the polls reported in the
Random Sampling and Other Useful Sampling Strategies • 49
media. They do not survey all of the nation’s households. Instead, a random
sample is drawn and surveyed. The results are pretty accurate. If you would
like to get a feeling for the size of samples that represent entire populations,
please refer to Krejcie and Morgan (1970; see source for Table 6.1). Table 6.1
suggests the great benefit that randomly selected samples afford. The larger
the population size, the smaller the sample. However, for smaller samples,
you must use almost the entire population. (Population size is noted by up-
percase “N” and sample size by lower case “n.”)
1. The first step in drawing a random sample is to identify all the mem-
bers in your population. You must be able to list them in what is called
a sampling frame. The frame should have the names without order to
them and be nonoverlapping (no duplicates). Alphabetizing the list by
surname is a way to ensure a random order in the sampling frame. Your
computer can sort alphabetically based on the last name, if surname is
entered into a database as a separate field. One reason that you need
to ensure random order of names is that some lists cluster them by
neighborhood location, housing type, income, or some other grouping.
This precludes their “equal chance” of being selected.
2. Give each name an identification number. Start with “1” and
continue.
3. Decide the size for your sample. You can use the table (see table 6.1),
or whatever you believe is right, in order to believe in the results you
obtain. As a rule of thumb, use as large a sample as possible. Whenever
you are calculating means, percentages, or other statistics, the popula-
tion is being estimated. Statistics calculated from large samples are
more accurate than those from small samples. Large samples give the
principle of randomness a chance to work.
4. Get a Table of Random Numbers (table 6.3 displays a partial one). Many
are located at the end of statistical or mathematical textbooks. The
Table of Random Numbers contains numbers generated mechanically
so that there is no discernable order or system to them. Each digit gets
an equal representation. The Table of Random Numbers consists of
rows and columns of numbers arranged at random so that they can be
used at any point by reading in any direction left or right, up or down.
We are now ready to draw our random sample. Here is our sampling frame
of 20 members of our population who are behavioral health counselors. The
director of the behavioral health clinic wants to select five counselors for an
overnight stay in the clinic with teens who are inpatients. In order to be fair
in the selection process, the director uses random sampling and lists the 20
counselors’ names in alphabetical order and assigns an ID number to each
from 1 through 20. Table 6.2 has the sampling frame.
The director knows that the largest ID number has two digits (20). So,
they are going to need a two-digit column in the Table of Random Numbers.
They put their finger down at any place in the Table of Random numbers.
This is their starting point. They have to decide in advance whether they
would use two digits going up, down, left, or right.
Random Sampling and Other Useful Sampling Strategies • 51
This works the exact same way as using the Table of Random Numbers and
is useful if your sample is small in size.
There are also free random number generators on the internet that you
can use: two are www.random.org and www.randomizer.org.
• For example, if you had a population of 1,500 (N) and needed to have
a sample of 306 (n), the key number (k) would be 5.
• Then, you would randomly pick any number between 1 and 5. Let’s say
you picked “4.” That is your first ID number. ID #4 is the first member
of your sample from the list of 1,500.
• Then, systematically add 5 (your key number) to the first ID number of
4 and you get #9. ID #9 is the second member of your sample. Continue
adding your key number.
• The next ID number is ID #14, then ID #19, ID #24, ID #29 until you
get your sample size up to the 306 you intended.
Chapter Summary
Except when the population of interest is small, it is imprudent to survey
an entire population. Sampling strategies are available for your use. By us-
ing simple random, systematic, cluster, stratified, convenience, purposive or
snowball sampling procedures we can get a handle on what we need to know,
save time, and get the thesis or dissertation completed expeditiously. One
of the most important uses of samples is that they afford us the opportunity
to make inferences without spending a fortune in time and money to do so.
CHAPTER SEVEN
Stating Hypotheses
and Hypothesis Testing
Descriptive statistics, which describe our data, are very useful. Other statistics
are also very useful but take a little more comprehension in implementation.
They are called inferential statistics. This is because we can draw conclusions
or inferences from them.
In the course of your graduate school research, you pose guesses or
hunches that you believe to be true. In order to tell whether these hypotheses
are true or false, you have to subject them to a test. You may draw a randomly
selected sample from the population of interest, compute statistics on the
sample, and test whether the hunch is true or false. Central to the discussion
of inferential statistics is the concept of probability. When your statistical
analysis reveals that the probability is rare that your statistical result is due
to chance alone, we call this a statistically significant result. It means that our
observed outcome is so unique that it is unlikely to have occurred by luck.
The role of the sample, discussed in the previous chapter, is essential
to the process and makes the task of inferring a possibility. The inferential
statistics most commonly used by graduate students to draw these conclu-
sions include t-test, one-way analysis of variance, chi-square analysis, and
correlational techniques including regression, among many others. These
will be discussed in this book since they are considered to be user-friendly,
pragmatic, and applicable in research conducted by most graduate students.
55
56 • Chapter Seven
Hypothesis Testing
Hypotheses are suppositions presumed to be true. Because of the importance of
posing these hypotheses, it is crucial to accept those that are true, and reject
those that are false. But how do we do this?
All research, whether it is in the graduate program or in the laboratory,
begins with the research question or hypothesis. It is always stated in the null
(negative) and is called the null hypothesis. This is because it is incumbent
on you, the researcher, to prove something is indeed true. It is similar to
the philosophy in the courtroom: innocent until proven guilty. The onus of
proof is on you to prove that the null is false. Technically speaking, the word
hypothesis is derived from a Greek word that means “an assumption subject
to verification.” The symbol for the null hypothesis is H0. Basically, there are
two types of null hypotheses: one that tests for differences and one that tests
for relationships between or among variables.
be a superficial guess. You are actually stating which mean will be greater
when you calculate your statistics.
In this case, you really believe that newly hired employees have higher
levels of job satisfaction than long-term employees do. Remember that the
hypothesis must still be stated in the opposite of what you think is true. It
has to be stated in the null (which you really think is false). Then you have
to prove it to be false. When we accept the null (or fail to reject the null),
we are saying that our results are not statistically significant. When we reject
the null (or the null is false) we are saying that our results are statistically
significant and likely due to factors or conditions other than chance.
For practical utility this book will focus on these two types of null hypoth-
eses: those that test for differences and those that test for relationships.
Alternative Hypotheses
If hypothesis testing shows that the null hypothesis is false, and is rejected,
the alternative hypothesis is stated. The symbol for the alternative hypothesis
is Ha. It asserts that the opposite of what the null hypothesis stated is true.
Almost always, graduate students believe that the alternative hypothesis is
true and that the null is false. But statistics are used to prove it. For the pre-
vious null hypotheses, these are the corresponding alternative hypotheses:
Probability Levels
How do graduate students determine whether a null hypothesis should be
accepted or rejected? Data are collected and statistics are executed. But the
key is the probability value associated with the statistical procedure that allows
the decision to be made. The probability value is also called the p value. It is
obtained once you execute your statistical analysis. You compare the signifi-
cance level you chose with the p value you obtained to determine whether
there are statistically significant results.
Probability applies exclusively to the likely occurrence of a future event.
In statistics, probability provides a quantitative measurement of the likeli-
hood that event will occur. In other words, probability is used to measure
the certainty or uncertainty of the outcome of an event. If it is sure to occur,
the probability is 100%. If it will never occur, the probability is 0%. A prob-
ability of .50 means the event should occur 50% of the time. Any event that
may or may not occur has a probability of between 0 and 1.00.
Professional journals often state with the results from a statistical pro-
cedure that “the probability level is less than .05” (p < .05). This means
that observed difference is likely to be real rather than having occurred by
chance. Or the statistical results may have occurred by chance only five out
of 100 times. Although somewhat arbitrary, the probability level is the mag-
nitude of error that one is willing to accept in making the decision to reject
the null hypothesis.
The conventional levels for rejecting the null hypothesis are either .05 or
.01. One (.01) is more conservative than the other (.05) because with .01
there is less willingness to have results due to chance alone; only one time in
100 will results be due to chance. A .05 significance level is more generous in
accepting a statement as true. A very conservative significance level is .001
or one time in 1,000. This may be used in scientific studies such as drug ef-
ficacy. For graduate students, the commonly accepted significance levels for
rejecting the null hypothesis are .05 and .01.
As a note, the significance levels must be chosen a priori. This is a Latin
term meaning “in advance.” Decide the significance level early on, before
collecting data, and accept it when the statistical testing is completed. It is
unethical to choose the conservative .01 significance level and then find out
that statistical significance would have occurred if the more liberal choice
of .05 was made.
Type I Errors
A Type I error is when the null is rejected when it was actually true, con-
cluding falsely that there were differences or relationships when there were
none. This can be problematic when someone else replicates the study and
the results do not hold up.
Making a Type I error is based on the level of statistical significance. If a
.05 significance level was selected, then five times out of 100 the results will
be due to chance. Unfortunately, one of the five times occurred. A more con-
servative significance level is .01 where only one time out of 100 the results
will be due to chance alone. A liberal significance level is .10 where 10 times
out of 100 chance results are OK. This level is not practiced in most graduate
student research and can largely contribute to Type I errors.
Type II Errors
A Type II error occurs when accepting the null when it was in fact false,
concluding that there were no differences or relationships when in fact there
were. If the null should have been rejected, but it was not, the maintenance
of the status quo may be a mistake. For example, you use conservative sig-
nificance level such as .01 to determine if a pre-K program attendance had
an effect on academic performance for first graders. The statistical result sug-
gests no impact when in fact there was one. Not rejecting the null (accepting
the status quo) might limit funding for a program with important benefits
because of the choice of a conservative significance level.
60 • Chapter Seven
do not always rely on means and standard deviations, they lack the
precision that the parametric statistics possess.
Generally speaking, if the data come from a population that is nor-
mally distributed, use the parametric test. If not, use the nonparametric
test. In a later chapter, chi-square analysis, an essential nonparametric
statistic for use by graduate students, will be discussed.
• If using a directional hypothesis: One-tailed tests (used with directional
null hypotheses) are more powerful than two-tailed tests (used with
nondirectional hypotheses).
• If using large samples: These are more powerful than small sample sizes.
• If using a more liberal significance level: If choosing .05 instead of .01, the
chances to reject the null are greater.
1. State the hypothesis in the null form. The null can be stated for either
differences or relationships. If for differences, the null is either non-
directional or directional, but awareness of the type is a prerequisite.
2. Select your significance level, either .05 (p ≤ .05) or .01 (p ≤ .01). A sig-
nificance level of .05 establishes a 95% confidence level and is more
liberal but may result in a Type 1 error. A significance level of .01
establishes a 99% confidence level and is more conservative but may
result in a Type II error.
3. Conduct the statistical analysis and obtain the p value from the statistical
procedure. Compare the p value to the significance level you chose to
determine whether there is a statistically significant result. The p value
obtained must be less than or equal to the significance level you set in
Step 2. If you chose p ≤ .05, the p value obtained must be .05 or less.
4. Accept or reject the null.
No statistically significant result: accept the null as true.
Yes, a statistically significant result: reject the null as false.
62 • Chapter Seven
Chapter Summary
The use of inferential statistics allows graduate students to make sound con-
clusions based on data. There is no guesswork. However, a method to the
madness must be observed. There is a scientific nature to hypothesis testing:
its principles must be respected. But it is a small price to pay for the increase
in knowledge that we get in exchange.
CHAPTER EIGHT
t-Test Procedures
For graduate students, many decisions are focused on comparisons such as the
following examples:
When two groups are compared, the statistic that is most useful is the t-test.
It is both an inferential and a parametric statistic.
The purpose of a t-test is to determine if there is a statistically significant
difference between the mean scores of two groups. This is where the mean,
the queen of central tendency, is vital. The mean scores of two groups are
compared via the formula for a t-test. Because the t-test is a parametric sta-
tistic, it is powerful. If there are differences, even slight ones, the t-test will
uncover them. There are a few basic facts about t-tests:
1. A t-test is used if there are only two groups or two testing times (pre/
post or before/after) to compare.
63
64 • Chapter Eight
Preliminary Assumptions
In the last chapter parametric statistics were described. This group of statistics
is more powerful than nonparametric statistics in rejecting the null hypoth-
esis. But they depend on your meeting a few assumptions before you use them.
Since the t-test is a parametric statistic, there are a few preliminary steps that
you must try to abide by before you implement this statistical procedure:
• The two groups should have equal variances on the dependent variable. The
variability of the individual groups’ mean scores should be equivalent.
They should have the same degree of variability. This is called a test of
homogeneity of variance and done as a preliminary step, a kind of insur-
ance program for your data. This is particularly true when your groups
are of different sizes (n)—common in real-world settings versus labora-
tory settings. The test of homogeneity of variances will be done for you
with statistical program packages like SPSS, as a preliminary step for
conducting the t-test.
66 • Chapter Eight
• The two groups should have an equal number of subjects. If the two groups
are unequal—like 20% more subjects in one group than in the other—
look at the standard deviation. If they are similar, go ahead and use
the t-test. If it is not, use the Mann-Whitney U test, a nonparametric
counterpart, discussed at the end of this chapter.
• Groups should be equivalent on all other variables except the dependent vari-
ables. If you are examining training approaches in two leadership devel-
opment programs by comparing scores on a leadership test, make sure
that the two groups are similar on other characteristics. These might
include gender breakdown, years working at the company, position in
the company, educational levels, age, and other variables. These factors
could account for the differences and not the leadership development
programs.
In your effort to meet the assumptions of the t-test, keep in mind that t-tests
are robust. They can hold up even when the assumptions are violated. Al-
though this robustness lets us off the hook to some extent, it is a good idea
to approximate to the greatest extent possible the rules of the game, the as-
sumptions. This will ensure excellent data management, avoid errors, and ulti-
mately produce sound data. As a note, many graduate research data sets do not
meet these assumptions. The operative word is to try to meet them. Your thesis
or dissertation is a learning opportunity as a researcher. If your data set violates
these assumptions, do not quit. Acknowledge the shortcomings and proceed.
Table 8.3. Data Set for Two Groups of Teachers and Their
Mathematics Scores
ID Teachers Category Group Math
1 Mr. Ramirez 9–12 1 88
2 Ms. DeJohn 9–12 1 89
3 Ms. Santos 9–12 1 88
4 Ms. Considine 9–12 1 89
5 Ms. Ponte 9–12 1 76
6 Ms. Marchand 9–12 1 89
7 Mr. Ross 9–12 1 88
8 Ms. Beyer 9–12 1 89
9 Ms. Gallagher 9–12 1 88
10 Ms. Magistrali 9–12 1 89
11 Ms. Strand 9–12 1 89
12 Mr. Fischer 9–12 1 88
13 Ms. Bowen 9–12 1 89
14 Ms. Peak 9–12 1 88
15 Mrs. Pacheco 9–12 1 89
Mean for Group 1 87.73
16 Ms. French K–8 2 89
17 Mr. Dickinson K–8 2 78
18 Ms. Addazio K–8 2 75
19 Ms. Saidel K–8 2 78
20 Mr. Dwyer K–8 2 79
21 Ms. Latavia K–8 2 89
22 Ms. Matarese K–8 2 78
23 Ms. Kozlak K–8 2 75
24 Mr. Fitz K–8 2 78
25 Ms. Hernandez K–8 2 79
26 Mr. Guzman K–8 2 78
27 Ms. Santo K–8 2 75
28 Ms. Strawson K–8 2 78
29 Ms. Vega K–8 2 79
30 Ms. Hebert K–8 2 70
Mean for Group 2 78.53
Correlated t-tests are used most often for this purpose—if you want
to determine impact, efficacy, and persistence over time. Some ex-
amples of when a correlated t-test might be used include the following:
• When assessing the efficacy of a health care intervention, treat-
ment, or drug over two time periods
• When tracking community involvement from last year to this
year
• When monitoring the effect of professional development training
on knowledge acquisition, pre and post
2. If you have matched your sample on some other variable so that the two
groups are alike.
One of the primary tenets of the t-test is that the two groups you are
studying are alike except for your dependent variable. You want to
ensure that something else called extraneous variance or error does not
70 • Chapter Eight
account for differences between the two groups instead of your depen-
dent variable.
For example, look at the data set of teachers, discussed previously.
What might account for the difference in mathematics knowledge
besides whether a teacher is K–8 or 9–12? The answers might be the
years in teaching, the amount of professional development, confidence
with the subject area, attitude toward the profession, or other factors.
If there was one particular factor that you felt might compromise your
data, you might try to match the groups on that variable so that they
are equivalent except for the dependent variable you were exploring.
Matched Samples
Here is an example of a matched sample. Using the same set of data that we
used in the t-test, the two teacher groups were matched on the number of
professional development courses they had taken. As you can see in table
8.5, each teacher in Group One is matched with another teacher in Group
Two on the number of professional development courses that they took. For
each ID number in Group One there is a member of Group Two that has
the exact same number of professional development courses. Ramirez with
French, DeJohn with Dickinson, and all the way through, there are pairs of
teachers with the same number of professional development courses in each
of the two groups.
Table 8.5. Teachers Matched on Professional Development (PD) for the Correlated
t-Test
Pair 9–12 Teachers PD Group Math K–8 Teachers PD Group Math
1 Mr. Ramirez 1 1 88 Ms. French 1 2 89
2 Ms. DeJohn 1 1 89 Mr. Dickinson 1 2 78
3 Ms. Santos 2 1 88 Ms. Addazio 2 2 75
4 Ms. Considine 3 1 89 Ms. Saidel 3 2 78
5 Mr. Ponte 4 1 70 Mr. Dwyer 4 2 79
6 Ms. Marchand 1 1 89 Ms. Latavia 1 2 89
7 Mr. Ross 4 1 88 Ms. Matarese 4 2 78
8 Ms. Beyer 5 1 89 Ms. Kozlak 5 2 75
9 Ms. Gallagher 3 1 88 Mr. Fitz 3 2 78
10 Ms. Magistrali 4 1 89 Ms. Hernandez 4 2 79
11 Ms. Strand 5 1 89 Ms. Guzman 5 2 78
12 Mr. Fischer 1 1 88 Ms. Santo 1 2 75
13 Ms. Bowen 2 1 89 Ms. Strawson 2 2 78
14 Ms. Peak 2 1 88 Ms. Vega 2 2 79
15 Ms. Pacheco 3 1 89 Ms. Hebert 3 2 70
t-Test Procedures • 71
Degrees of Freedom
Independent sample and Correlated sample t-tests have different degrees
of freedom associated with each t-test form. Degrees of freedom are noted in
professional journals as df. You need to know how the degrees of freedom are
attained so that you can report your statistical results.
• If you use an Independent samples t-test, the degrees of freedom (df) are
equal to the number in your total sample minus 2 or [N – 2]. In the first
example of K–8 and 9–12 teachers, the degrees of freedom would be 28
[or 30 teachers – 2 = 28].
• If you are using a Correlated samples t-test, the degrees of freedom (df)
are equal to the number of pairs minus 1 [N of pairs – 1]. In the second
example of teachers, you used matched pairs on professional develop-
ment courses; the degrees of freedom would be 14 [or 15 pairs – 1].
Correlated samples t-tests make it more difficult for you to reject the null.
Since the data for the groups are related, you have to ensure that the differ-
ence you find is real and not due to the relationship between the sets of data.
Therefore, the bar is raised with Correlated t-tests in the computation for
statistical significance with respective degrees of freedom and critical cutoff
points for rejecting the null. (Statistical textbooks will have a deep discus-
sion of degrees of freedom along with statistical computations of t-tests for
graduate students who are interested.)
This is how you might report your t-test findings in a table format. For
a Correlated samples t-test, there are 25 workshop participants tested twice,
pre and post. The two mean scores on a leadership test and the respective
standard deviations are reported. The t-value that is calculated with the
SPSS statistical package is 4.69, there are 24 degrees of freedom, and the
calculated p-value is .00. This is below the cut off level of significance you
chose, which was .05. You have statistically significant results; the null is
rejected (see table 8.7).
Nonparametric Counterparts
There are two nonparametric statistical procedures that answer the same
question that the t-test does. If the assumptions mentioned in the begin-
ning of this chapter are violated, then you may want to consider using a
nonparametric alternative. The Mann-Whitney U test and the Wilcoxon
matched-pairs signed ranks test are the counterparts to the Independent and
Correlated samples t-tests, respectively. They are not bound by normal dis-
tributions or equal variances. They can be used with ordinal data or ranked
data instead of interval/ratio scaling for the dependent measure. These non-
parametric counterparts can be executed in a statistical program like SPSS.
The downside is that they are not as powerful as the parametric t-tests in
rejecting the null hypothesis.
t-Test Procedures • 73
Chapter Summary
T-tests are very useful tools. Often, you are comparing two groups of students,
managers, patients, employees, customers and so forth. The t-test is a simple
and straightforward statistic that allows you to get beyond a frequency count.
One group’s mean score may look different from another’s mean score, or it
may look the same. The litmus test for statistically significant two group com-
parisons is called the t-test.
CHAPTER NINE
ANOVA Procedures
The t-test is very useful when you have two groups or testing times (pre/post
or before/after) to compare. But what happens if you have more than two?
The statistical technique that is analogous to the t-test is called Analysis of
Variance or ANOVA. The purpose of ANOVA is understood quickly if it is
thought of as an extension of the t-test.
ANOVA is an inferential statistic. It is also a parametric statistic and as
such it is very powerful. It can reject the null hypothesis or find differences if
they exist among groups. The assumptions of homogeneity of variance, equal
group sizes, and normal distribution of scores should be adhered to, just as the
t-test should meet these assumptions. Yet as a robust statistic, ANOVA can
sustain having the assumptions violated and still perform its function. There
is a nonparametric version of ANOVA called the Kruskal-Wallis H test. If
you have serious violations of the assumptions, use it.
ANOVA is a little more complicated than a t-test even when using a
statistical software package. Although there are many versions of ANOVA
(which will be described briefly), this book will focus on a simple version,
One-Way ANOVA. For the purpose of discussion, the main point is that you
use ANOVA as you would a t-test but when there are more than two groups.
There is one dependent and one independent variable but with more than
two levels or categories.
Here are some facts about One-Way ANOVAs:
75
76 • Chapter Nine
With ANOVA there are more than two mean scores. You cannot eyeball
the means to see which is higher and then make deductions. You have to
conduct an additional statistical procedure. A follow-up test, called a post
hoc procedure or a multiple comparison test, is designed just for this purpose.
It shows exactly where the significant differences lay after (post) a significant
ANOVA Procedures • 79
• Fisher’s LSD
• Duncan’s new multiple range test
• Newman-Keuls
• Tukey’s HSD
• Scheffé test
Some are more conservative than others are, making it harder to reject the
null. A liberal procedure will find a significant difference between two mean
scores that are relatively close together. A conservative procedure will indi-
cate that two mean scores are significantly different only when the means are
far apart. The Fisher LSD is the most liberal and the Scheffé test, the most
conservative of the post hoc tests.
You may wonder why t-tests are not used to compare each of the com-
binations above. That would be logical except for the fact that conducting
multiple t-tests increases the chances of getting spurious results with error.
As the number of t-tests increase, the probability of getting a statistically
significant difference by chance alone also increases. This means you are ap-
proaching Type I error territory. Post hoc or multiple comparison tests adjust
the level of significance to reduce the influence of error.
For the Town example, you conducted a Scheffé multiple comparison test.
It showed that the level of cigarette-smoking in Town B was significantly
different from that in both Town A and Town C on number of cigarettes
smoked per month. However, the levels for Town A and Town C were not
different from each other. Textbox 9.1 shows how you might report your
findings in table format. There are other ways to report ANOVA findings
that can be found in professional journals in different formats.
For the ANOVA example, the degrees of freedom are these two numbers:
2 and 36. You will see it often reported like this (df = 2, 36) with a comma
separating the between from the within.
Factorial ANOVAs
Two-Way ANOVA
ANOVA designs, called factorial ANOVAs, compare more than one in-
dependent variable. Let’s use the same example that we had in One-Way
ANOVA, but add another independent variable, also referred to as a “fac-
tor.” Besides town of residency, we will add school (middle or secondary).
This design is called a Two-Way ANOVA.
Now we have Town of residency as one independent variable, and School
as a second independent variable. Instead of three groups that we had in
One-Way ANOVA, we have six groups because of the two independent
variables. Sometimes, a design such as this would be listed as a 3 × 2 design.
This means there are two independent variables: one with three categories
and one with two categories. Each of the six groups’ mean scores (M) on your
one dependent variable of “Number of Cigarettes Smoked” will be compared
by the Two-Way ANOVA (see table 9.2).
Three-Way ANOVA
This can further become a Three-Way ANOVA by adding yet another in-
dependent variable or third factor. Let’s add “coffee usage.” There are now
three independent variables with twelve separate groups on which the de-
pendent variable (M) will be measured. The three independent variables will
have three categories (Town), two categories (School), and two categories
(coffee usage). This may be called a 3 × 2 × 2 ANOVA design (see table 9.3).
82 • Chapter Nine
Chapter Summary
One-Way ANOVA is a very useful statistic and a great partner to the t-test.
It functions quite the same but affords us the opportunity to expand the
number of groups beyond two. This is often called for in many settings where
more than two categories exist for graduate student research. The conceptual
steps to One-Way ANOVA follow the t-test with the exception of the need
to conduct post hoc tests after significant results are obtained. Many statisti-
cal packages like SPSS offer this option as part of One-Way ANOVA.
For statistical questions that are more complex (in that they add more
variables to the research design), there are statistical models to accommodate
them. Factorial ANOVA, ANCOVA, and Repeated Measures ANOVA
are sophisticated statistical procedures that have been described in the most
elementary fashion here. Reference to statistical or mathematical textbooks
will present a more detailed discussion for interested graduate students.
CHAPTER TEN
Chi-Square Procedures
85
86 • Chapter Ten
(40.00), and the significance level (.00) are reported. The results can be as
simple in a narrative like this. X² (3, N = 80) = 40.00, p ≤ .05. Or you can
develop a table of results that looks like table 10.4 or table 10.5.
Table 10.4. Reporting the Chi-Square (X2) Results with Expected and Actual
Frequencies
New Hires X2
N = 80 Frequency Novice Apprentice Practitioner Expert df = 3
Business Expected 10 (25%) 10 (25%) 10 (25%) 10 (25%) 40.00*
40 = n Actual 10 (25%) 10 (25%) 20 (50%) 0 (00%)
Engineering Expected 10 (25%) 10 (25%) 10 (25%) 10 (25%)
40 = n Actual 10 (25%) 10 (25%) 0 (00%) 20 (50%)
* p ≤ .05
Table 10.5. Reporting the Chi-Square (X2) Results with Actual Frequencies Only
New Hires
N = 80 Novice Apprentice Practitioner Expert X2 df p
Business 10 (25%) 10 (25%) 20 (50%) 0 (00%) 40.00 3 .00*
(n = 40)
Engineering 10 (25%) 10 (25%) 0 (00%) 20 (50%)
(n = 40)
* p ≤ .05
• The Fisher’s exact test: This test is used to compare two independent
samples with respect to a dichotomous variable. This is similar to the
chi-square test just described, but the Fisher’s exact test works better
when there are few subjects in the data set. If that is a limitation, this
is the statistic that helps with a solution.
• The Mann Whitney U test: This statistic is used to compare groups by
using rankings on variables of interest. Both groups are independent
just as with the Independent samples t-tests. This statistic does not use
mean scores but rather ranking of data to determine differences.
• The Kruskal-Wallis H test: This test is an extension of the Mann Whit-
ney U test but used with three or more groups, parallel to the one-way
ANOVA procedure. It uses a similar methodology to the Mann Whit-
ney U Test with rankings.
• The Wilcoxon-matched pairs signed-ranks test: Similar to the correlated
t-test, this statistic uses data where there is repeated measurement. If
the data are ordinal or the assumptions of the t-test cannot be met, this
statistic is useful.
The sign test, the binomial test, and the median test are among other nonpara-
metric procedures that are useful to consider. Again, they are less powerful
than their parametric counterparts to reject the null hypothesis. This means
a Type 2 error has more likelihood of occurring. Certainly, they should be
considered as part of the array of strategies that graduate students can avail
themselves of, if the situation calls for alternatives.
Chapter Summary
The chi-square statistic, which relies on nominal and ordinal scaling de-
scribed in chapter 1, is used very often because graduate research has categori-
cal variables. The chi-square statistical procedure allows us to compare groups
against each other on important variables. Although it may not have the
sophistication of its parametric counterparts, its utility may exceed theirs. It
is a statistical technique that provides us with insights that would be impos-
sible to diagnose if you had to depend on statistics that required interval or
ratio scaling.
CHAPTER ELEVEN
Correlation Procedures
There are statistics that help to determine if relationships do exist, and if so,
what are the characteristics of those relationships? This is where correlations
are useful statistical techniques. They test the extent to which two variables
occur together, and how related they are. As statistics, correlations can be
descriptive and inferential at the same time. They can describe your data and
can also infer relationships from samples to populations.
93
94 • Chapter Eleven
1. If one variable [x] increases does the other variable [y] increase?
Example: Greater parent-to-child reading time increases school
readiness.
2. If one variable [x] decreases does the other variable [y] decrease?
Example: The less time patients spend out of bed, the less their
mobility.
3. If one variable [x] increases does the other variable [y] decrease?
Example: Increased sports participation by teens decreases cell
phone usage.
4. If one variable [x] decreases does the other variable [y] increase?
Example: Employee satisfaction declines when overtime require-
ments increase.
Strength
The actual numeric value of the correlation coefficient tells us the strength
of the relationship. The nearer the number is to either +1.00 or –1.00, the
stronger the relationship is between the two variables. Correlations of –.88
and +.88 have the same strength. A correlation of –.88 is stronger than a
correlation of +.87 by only an infinitesimal amount.
A zero correlation (r = 0) indicates absolutely no correlation whatsoever.
The relationship between the amount of rainfall and an employee’s salary
would have no correlation (r = 0). Correlations of –.07 or +.02 are negli-
gible. Although there is no hard and firm interpretation of what constitutes
strength of all correlation coefficients, here are some suggestions of strength.
Direction
Relationships between two variables can either be positive or negative.
That is what is meant by direction. Therefore, correlation coefficients can be
either positive or negative. A plus or a minus sign before the numeric value
indicates direction.
If a correlation is positive, it means that
Positive correlations sometimes have a plus sign before the decimal, but many
times it is implied. A positive correlation of +.67 might also look like .67.
If a correlation is negative, it means that
correlation coefficient (.92), you can see that there is a strong, positive cor-
relation between attendance and extracurricular activities for preteens.
1. Examine the computer output to compare the p value that is associated with
the correlation coefficient.
The degrees of freedom for a Pearson product moment correlation
coefficient is (the number of pairs – 2). In this case there were twenty
pairs of data with the two variables so the degrees of freedom was 20
minus 2, or 18. As calculated, your actual p value for the data was .00
and your level of significance was p ≤. 05.
2. Accept or reject the null hypothesis.
You reject the null hypothesis. There is a significant relationship
between after-school/extracurricular activities and school attendance.
The direction and strength tell you that as one variable increases, so
does the other. Specifically, participation in extracurricular activities
boosts school attendance for preteens at risk for dropping out. Fur-
thermore, the shared variance (r²) is 84%—the correlation coefficient
squared. This means that 84% of school attendance can be explained
by the number of extracurricular activities a preteen in this community
is involved in.
Scattergrams are very useful because they tell us a lot about our two
variables and their relationship. Here are the pieces of information that scat-
tergrams supply.
1. Scattergrams can tell the strength of the relationship between the two variables.
• If the dots cluster close to an imaginary line, there is a very strong
relationship or correlation between the two variables.
• When the dots are scattered in an ellipse or cigar-shape, there is a
moderate correlation.
• When the dots are scattered randomly around the graph, there is a
low or negligible correlation.
The scattergram in figure 11.1 shows the dots clustering close to an
imaginary line or the line penciled in on the scatter diagram. This
illustrates a very strong relationship between the two variables. The
tighter the clustering of dots, the stronger the relationship. In the case
of our data set, the number of extracurricular activities might be con-
sidered a predictor of school attendance. The dots are clustered into a
fairly tight line.
2. Scattergrams can tell the direction of the relationship between the two variables.
• If the slope of the line falls from left to right, there is a negative
correlation.
• If the slope of the line rises from left to right, there is a positive
correlation.
The latter is the case with the data in our scattergram. The slope of
the line rises from left to right signifying a positive correlation. The
slope of the line indicates that as extracurricular activities increase, so
does school attendance. Conversely, a positive relationship also means
that as after-school and extracurricular activities decrease, so does
attendance. Just looking at the slope tells us a great deal about the
relationship between these two variables.
3. Scattergrams can also show outlier scores.
Values completely out of range with all the others should be consid-
ered for elimination in the calculation of a correlation coefficient.
They might contaminate the data set and give misleading results if a
correlation coefficient is calculated. In our example, there are no out-
lier scores. All of our x/y intersections fall close to the line. But look at
figure 11.2, which has some outliers in it.
Correlation Procedures • 101
Figure 11.2. Scattergram for Attendance and Extracurriculars with Outliers. Source:
Author developed
1. The most useful purpose is to see if two variables are correlated. This
is what chapter 11 has focused on.
2. Another important function of the correlation procedure is with re-
spect to tests or instruments we measure with. The reliability of a test is
reported as a correlation coefficient. It measures whether the test con-
sistently measures what it purports to measure. You can be confident of
that when you examine the correlation coefficient, reported in the test
manual as the “reliability coefficient.” If it is r = .80 or r = .75, we can
be certain that it is a pretty reliable measurement tool. Medical instru-
ments have high reliability coefficients; personality tests or vocational
interest assessments would have lower reliability coefficients.
3. A third function is performed by correlational statistics. If you are in-
terested in determining whether two raters or two observers are seeing
the same thing when they collect information, you are looking for inter-
rater reliability. For example, if two of your managers were evaluating
the customer service for the same group of bank tellers, you would want
their observations (recorded by a checklist or rubric) to be very similar.
By correlating their data, you would create an inter-rater reliability
coefficient. Hopefully, the correlational numeric value would be strong.
102 • Chapter Eleven
Chapter Summary
There is one last and very important point to make when correlations are
discussed. No attempt should be made to say that one variable (x) causes the
other variable (y). This is untrue. A correlation only suggests that a relation-
ship exists between the variables. It does not mean that one variable causes
the other. The use of correlational statistics is a great asset. It suggests that
relationships exist. To use them to pinpoint causality is a misuse of a highly
valuable statistical technique.
CHAPTER TWELVE
Regression Procedures
Another difference is the labeling of the variables. The variables that ex-
plain or predict are the independent variables. The variable that is explained
or predicted is the dependent variable. In the previous explanation example, the
independent variables are salary, hours, hospital resources, patient mix, sup-
plies, and building maintenance. The dependent variable is nurse satisfaction.
103
104 • Chapter Twelve
Ŷ = a + b times X
The variable Ŷ is the unknown dependent variable being predicted: the at-
titude toward math. The variable X is the known independent variable, the
actual midterm grade for the 2023 students. The bivariate regression statisti-
cal output that the teacher used with the 2022 class data provides the num-
ber for the a (the constant) and b (the regression coefficient or slope). (As
106 • Chapter Twelve
graduate students you need to know where to find a and b on the computer
output like SPSS.)
For each student in her 2023 class, the math teacher uses the equation by
putting in the value for a (the constant), the value for b, and the freshman’s
actual 2023 midterm grade (X) to predict their attitude toward math (Ŷ).
The equation is calculated like this: you add the value for a with b times the
midterm grade (X) and you get a predicted score on attitude toward math
(Ŷ). The equation is below.
The teacher can now identify those freshmen whose attitudes are predicted
to be negative, and then provide support as they begin the second half of
their freshmen year of high school. The bivariate regression equation devel-
oped from the 2022 class of freshmen will help to identify those 2023 fresh-
men whose attitude toward math may be low.
Ŷ = a + b1 × (X1) + b2 × (X2)
(X1) and (X2) are the two independent variables. There is always a constant
(a). There is a regression coefficient or slope (b) for each independent vari-
able (X).
Regression Procedures • 107
There could be a third independent variable and the equation would look
similar. There is always one constant (a). There is a regression coefficient or
slope (b) for each independent variable (X).
These statistics are provided in statistical packages like SPSS as part of the
regression procedure. Table 12.1 displays the descriptive statistics on the
three variables under study. The means (M) and standard deviation (SD) are
produced for each variable X1, X2, and Ŷ.
Table 12.2 produces the correlations among the three variables. This is an
important piece of information. It indicates that the independent variables
Regression Procedures • 109
are correlated with the dependent variable to differing degrees, and the in-
dependent variables have a low correlation with each other. That is what is
desired when a multiple regression procedure is used.
*p ≤ .05
Ŷ = a + b1 × (X1) + b2 × (X2)
(Ŷ) = 6.16 (a) + .91 (b1) × (job satisfaction X1) + .14 (b2) × (years empl. X2)
Chapter Summary
The occasion to undertake a regression procedure may be commonplace
due to the nature of correlated variables in research settings. Given that
fact, graduate students should consider using this statistic if the situation
warrants. Some expert assistance may be worth the investment, but the in-
sights produced are invaluable. The concepts behind this statistic are fairly
straightforward. However, some understanding of this statistic is necessary
because a click of the mouse will produce output, but it may be misinforma-
tion if done incorrectly.
CHAPTER THIRTEEN
• Consider your own interest in the topic. Are you genuinely interested
in it and willing to spend days, weeks, months, and maybe even years
focused on it? Are you willing to do the work to become an expert on
this topic? Is it one you would really enjoy exploring?
• Evaluate the importance of the topic. Is it something that will be of
benefit to the field you are part of? Will your work on this topic add to
the body of knowledge? Is there some practical value that other profes-
sionals in your field can use or apply that makes your topic of actionable
value?
• Is your topic relevant and timely? Is it something that others in your
field have written about in the recent past? Is it an area of discussion
right now, or is it passé? Is there a new treatment that you would like to
compare to another one or one that is older? Is there a gap in the body
of knowledge that you would like to fill? Are there contradictions in
113
114 • Chapter Thirteen
results published in the literature that you would like to address? Does it
build upon someone else’s work but has a different twist? Is it an original
idea or has it already been done with conclusive results?
• Is the time required to complete the research on this topic realistic? Is
the topic something you can research within a timeframe or is the com-
mitment of time too great for you to undertake as a graduate student?
• What is the level of difficulty? Is your topic one that has sensitivity,
geography, or complexity that will make the process of investigating it
filled with extra stress? Does controversy surround it? Do you have to
travel far and wide to access the data sources? Is it so complex that you
will drown in the process of doing the research?
• What are the out-of-pocket expenses associated with your topic? Ana-
lyze the costs of pursuing it. Is the topic one that fits within a student’s
limited budget or will the cost exceed the pocketbook?
Use multiple library and online resources to establish keywords to get the ball
rolling; add keywords as you move along. “Research the research” in the litera-
ture review process. Here are a range of resources that can help you search
the literature, thoroughly and efficiently. Remember to add keywords as you
go along with the purpose of “researching the research.”
Take your time in the library and online when reviewing the literature.
Avoid costly and embarrassing mistakes because you were rushing. It is in-
credibly easy to make literature review errors when you are first starting out
on this task. Listed below are some of the most common ones that graduate
Practical Tips for Graduate Students • 115
students make during their review of the literature. You will see that avoiding
them usually requires just a little slowing down, and thoughtful planning on
your part. Literature review is oftentimes done in a hurry instead of system-
atically. Sources are overlooked or skipped because
Use index cards to take notes and to guide your actual writing. It seems
antiquated but you can shuffle the cards. They are physical and can be moved
in a different order for easy writing of narratives.
Chapter Summary
Research is hard work, and it requires a commitment over time. One of the
greatest barriers to completing the thesis or doctorate is implementing the
research methods, including statistics. This book is intended to help you
move through the barriers that seem formidable but are not. With a little
persistence and time, you can master the research challenge and become the
success that you set out to be when you signed on to your academic journey.
In the process, you will find that conducting research can be gratifying in so
many ways. The content of this book is to help you find joy in that process,
and maybe even to continue to do research after your degree is obtained.
Good luck.
About the Authors
Since 1984, Words & Numbers Research, Inc. has conducted numerous evalu-
ation and research studies in the fields of education, healthcare, manufactur-
ing, banking, human services, and others. External evaluation services have
been provided to the National Science Foundation and the US Department of
Education. The firm conducts surveys, focus groups, and in-depth interviews.
The office can be reached by phone at 860-489-5639, by email at wordsnum
@optonline.net or on website at www.wordsandnumbersresearch.com.
121