Research Methods Notes
Research Methods Notes
RESEARCH METHODS
Research: Is a process of enquiry and investigation that is systematic, methodical and ethical
(www.brad.ac.uk/intrduction-to-research)
According to Orodho and Kombo (2002) research is a process of arriving at dependable solutions
to problems through planned systematic collection, analysis and interpretation of data.
Research project or thesis: Is a written scientific report that deals with the concerns related to
the problem or series of problems in one’s area of research (Kombo and Tromp 2006). It
comprises of five chapters:-
Chapter one: Introduction
Chapter two: Literature review
Chapter three: Research methodology
Chapter four: Research findings
Chapter five: Summary, Conclusions, Recommendations
VARIABLES
You won't be able to do very much in research unless you know what variables are. A variable is
any entity that can take on different values. The term variable is derived from variations. It
refers to differences, attributes, qualities of the cases that we measure or record (Kombo and
Tromp 2006).
A variable is anything that takes differing or varying values e.g. Age, height, price, income, level
of education, motivation, performance, etc.
Anything that can vary can be considered a variable. For instance, age can be considered a
variable because age can take different values for different people or for the same person at
different times. Similarly, country can be considered a variable because a person's country can
be assigned a value.
TYPES OF VARIABLES
1. Independent variables (also known as predictor variable)
An independent variable is what the researcher (or nature) manipulates, a treatment or
program or cause.
It is the intervention or a condition implemented by the researcher to see if it will create
change on the dependent variable. This could be a program, method, system, or other action.
Independent variables are predictors or explanatory variables, those that cause changes to
the dependent variables. They influence the dependent variables positively or negatively.
1
Jane Muchembi
e.g. The relationship of availability Computer Aided Design (CAD) software (independent
variable) on the number of defective product i.e. rejects (dependent variable) in fashion
manufacturing firm may be affected by the designers ability to use the CAD technology
(moderating variable).
Whenever there is independent variable there is also a dependent variable.
An increase or decrease in units of independent variable will lead to increase or decrease in
dependent variable i.e. variance in dependent variable is accounted for by variance in
independent variables.
3. MODERATING VARIABLE
Moderating variable is one that has a strong contingent effect on the independent variable-
dependent variable relationship.
2
Jane Muchembi
Its presences modifies the relationship between the independent and dependent variables e.g. The
designers ability or skills in using CAD technology will interfere with the relationship between
availability of CAD software and number of defective products (rejects).
Moderating variable
4. INTERVENING VARIABLE
Intervening variable is one that surfaces between the time independent variable starts operating
to the time their impact is felt on it. There is therefore a temporal quality or time dimension to
the intervening variable.
It surfaces as a function of the independent variable(s) operating in any situation and helps to
conceptualize and explain the influence of independent variable(s) on the dependent variable.
e.g. in a study where the independent variable is workforce diversity which influences the
dependent variable is organizational effectiveness, creative synergy (collaboration, cooperation,
interaction) may surface as the intervening variable. Creative synergy may surface due to
multiracial, multiethnic and multifaceted expertise in problem solving.
NB: creative synergy in the above example surfaces at time t2 as a function of workforce
diversity which was in place at time t1 to bring about organizational effectiveness at time t3.
The intervening variable of creative synergy helps us to understand how workforce diversity
brings about organizational effectiveness.
Time t2 t3
t1
:me
Time t1 t2 t3
:me
Independent variable
Workforce diversity Creative synergy Organizational effectiveness
Managerial expertise
Moderating variable
NB: Inclusion of a moderating variable will affect the relationship between the other variables
e.g. creative synergy may not occur if the management is not able to harness the different
problem solving skills in the diverse workforce.
Abstract: is a summary of the whole research project including the objectives, methodology
used, findings and recommendations. It should not exceed 500 words.
RESEARCH TOPIC: Refers to the subject, issue or area under discussion or investigation e.g.
fashion marketing, African dress, interior design.
4
Jane Muchembi
A researcher should choose a topic which interests him/her as this will help sustain the
motivation to complete the research. This is because one will enjoy reading materials in the area,
collecting and analyzing data.
RESEARCH TITLE: is the heading, label or tag. It describes what the research is about. It
should be formulated after selecting the topic.
QUALITIES OF A GOOD TITLE
i. It should be brief i.e. not more than 15 words
ii. It should be specific i.e. it should be easy to identify independent and dependent variable from
the title.
iii. It should be in line with the objectives i.e. it should portray the aims and objectives of the
research.
iv. It should be clear and not ambiguous i.e. it should not be subject to different interpretations. It
should reflect a relationship between independent and dependent variables.
v. It should portray an issue that is researchable i.e. the aspects in question should be
measurable.
REFERENCING STYLES/SYSTEMS
There are several referencing styles to choose from e.g.
i. Vancouver which uses numbers in the text and footnotes. Each number in the text represents a
reference or publication whose details are given in the footnotes.
5
Jane Muchembi
6
Jane Muchembi
Available from <URL: http://www. remainder of full internet address>e.g. Illingworth, N. (2001)
‘The internet matters’ Sociological Research Online 6:2, (online) (cited 20 March 2002)
Available from < URL: http://www.scoresonline.org.uk/6/2/illing th.html>
NB:
i. All the authors cited in the text should be included in the reference list.
ii. Authors not cited in the text should not be in the reference list.
iii. Single space within a reference and double space between references.
iv. Use hanging indents in references
v. The publication in the reference list should be arranged in alphabetical order by the
authors’ names i.e. authors whose surnames start with letter “A” should be listed before
those whose surnames start with “B” and so on.
Babbie, E. (2001). The practice of social Research: 9th Ed. CA: Wadworth Thomson.
Saunders, M, Lewis, P. & Thornhill, A. (2007) Research Methods for Business Students (4th
Edition.) Harlow: Prentice Hall.
7
Jane Muchembi
8
Jane Muchembi
9
Jane Muchembi
who have gone through a diploma course in design and those who have gone through a degree
program.
2. The alternative hypothesis: is the opposite of the null hypothesis and states that a value or
relationship in the null is not true. It symbolized as H1 e.g. H1 e.g. H1: there is a significant
difference in performance of designers who have gone through diploma and those who have
gone through a degree program.
DEFINITION OF TERMS
Significant terms which are used whose meanings are not obvious are defined. Start this section
in a new page.
10
Jane Muchembi
iv. Helps the researcher to develop a significant problem of the study which will provide further
knowledge in the field of study.
v. It exposes the researcher to a variety of approaches or methods of conducting research. This
contributes to a well-designed methodology because the researcher can avoid methods that have
been reported in the literate to be weak or have failed.
vi. It helps in developing analytical framework or basis for analyzing and interpreting data. The
researcher will be able to compare and contrast their findings with the findings of the other
researchers.
vii. Helps to convince the reader that:-
a) The researcher is knowledgeable about the problem area and has done the preliminary home
work that is necessary to conduct research
b) Theoretical framework will be structured on work already done and will add to the solid
foundation of existing knowledge.
(Sekaran 2003)
11
Jane Muchembi
2.2 The body: this may be structured into subtopics based on either the:- a) objectives or
b) Variables
2.3 Summary and identified gaps: summarizes the key and relevant findings from the reviewed
literature and brings out what the previous studies have not covered on the topic and needs to be
researched.
NB:
When writing the literature review, read the publication critically and analytically, understand
and then summarize the idea you find relevant in your own words. Then cite the author.
If there is need to use direct quotation put them into quotes and cite the authors name, date and
page number. If the quotation is from 40 words and above indent and single space the whole block
and do not put them in quotes, cite the author, date and page number. Desist from using quotes after
quotes and copy pasting.
Bring out some discussions on every idea you write about based on what different authors have
written i.e. explain where different authors agree or disagree, contradictions and gaps of what they
have not explained about the idea (missing gaps)
Only include relevant literature. Not every material you will read is relevant to your study.
Reasons for citing other peoples work when conducting research in clothing, textile and
interior design industry.
i. The body of knowledge is not built on vacuum but on accumulated knowledge over generations
and so the researcher should try to fit his/her research within already existing knowledge.
ii. It is courteous to acknowledge use of other peoples’ work.
iii. It provides evidence of thorough examination of literature and shows that the researcher has
read.
iv. It gives support to the points that the researcher has stated in the research.
v. Helps to identify gaps that should be met by the researcher.
13
Jane Muchembi
data.
Basic research Applied research
- Aims at improving or adding knowledge without - Is designed right from the start to apply its
any particular applied purpose in mind at the findings to a particular situation or to solve a
onset. specific problem.
Deductive research: testing theory Inductive research: building theory
- Moves from general ideas to specific, particular - Moves from particular situations to make or infer
situations. broad general ideas or theories.
- The particular is deduced from the general e.g. - E.g. if you wanted to know the meaning of the
broad theories word fashion to a range of people may be of
- E.g. if you wanted to know the meaning of the word different ages or gender, you could start by
fashion to a range of people may be of different checking the definitions with the range of
ages or gender, you could start by checking the people and then compare with definition given
definitions given by professional associations or by professional associations or sources and
sources and then test the definition with the range then test the definition
of people. - Then collate the data collected and the results
- Then collate the data collected and the results analyzed and presented.
analyzed and presented. - This may lead to new definition or not
14
Jane Muchembi
V. Case study
i. Seeks to study a unit in detail, in context and holistically.
ii. Enables the researcher to deep information and understanding on the unit being studied
iii. The findings may not be easily generalized.
iv. The researcher should give clear justification for using this design.
15
Jane Muchembi
Let's imagine that you wish to generalize to urban homeless males between the ages of 30 and 50
in the Kenya. If that is the population of interest, you are likely to have a very hard time
developing a reasonable sampling plan. You are probably not going to find an accurate listing of
this population, and even if you did, you would almost certainly not be able to mount a national
sample across hundreds of urban areas. So we probably should make a distinction between the
population you would like to generalize to (theoretical population) and the population that will
be accessible to you (the accessible population). In this example, the accessible population
might be homeless males between the ages of 30 and 50 in six selected urban areas across the
Kenya
SAMPLING FRAME: is the list of all the members of the population under investigation. The
researcher can obtain one from the relevant bodies or develop one in case where such a list does
not exist.
Once you've identified the theoretical and accessible populations, you have to do one more thing
before you can actually draw a sample -- you have to get a list of the members of the accessible
population. (Or, you have to spell out in detail how you will contact them to assure
representativeness). The listing of the accessible population from which you'll draw your sample
is called the sampling frame.
16
Jane Muchembi
You actually draw your sample (using one of the many sampling procedures). The sample is the
group of people who you select to be in your study. Notice that I didn't say that the sample was
the group of people who are actually in your study. You may not be able to contact or recruit all
of the people you actually sample, or some could drop out over the course of the study. The group
that actually completes your study is a subsample of the sample -- it doesn't include non-
respondents or dropouts. The problem of nonresponse and its effects on a study will be addressed
when discussing "mortality" threats to internal validity.
People often confuse what is meant by random selection with the idea of random assignment.
You should make sure that you understand the distinction between random selection and random
assignment.
At this point, you should appreciate that sampling is a difficult multi-step process and that there
are lots of places you can go wrong. In fact, as we move from each step to the next in identifying
a sample, there is the possibility of introducing systematic error or bias. For instance, even if you
are able to identify perfectly the population of interest, you may not have access to all of them.
And even if you do, you may not have a complete and accurate enumeration or sampling frame
from which to select. And, even if you do, you may not draw the sample correctly or accurately.
And, even if you do, they may not all come and they may not all stay.
SAMPLING FORMULAR
N
n =______________________
1 + Ne2
Where;
N- the population/Number of cases in the sampling frame
e- the margin of error
n- the sample size/number of cases in the sample
17
Jane Muchembi
SAMPLING DESIGNS
Are categorized as:-
1. PROBABILITY SAMPLING
Probability sampling techniques: apply random selection where each member of population is
given equal chances of being in the sample. They don’t give all the members of the population
equal chances of being in the sample.
A probability sampling method is any method of sampling that utilizes some form of random
selection. In order to have a random selection method, you must set up some process or procedure
that assures that the different units in your population have equal probabilities of being chosen.
Humans have long practiced various forms of random selection, such as picking a name out of a
hat, or choosing the short straw. These days, we tend to use computers as the mechanism for
generating random numbers as the basis for random selection.
2. NON-PROBABILITY SAMPLING
Non-probability sampling techniques: The researcher is interested in the representativeness of
the concepts in their varying forms maximizing the scope or range of variations of respondents in
the sample.
1. First, we have to get the sampling frame organized. The researcher obtains or makes a list of all
the members of the population and assigns to each of them a number or letter i.e. if the whole
population comprises of 1,000 people make a list of 1,000.
2. Decide on the number of people you would like to have in the final sample depending on the size
of the population and the research design. For example, if you want to select 100 clients to survey
and that there are 1000 members of the population. Then, the sampling fraction is f = n/N =
100/1000 =10 or 10%.
3. Write those numbers in small papers (tickets) similar to those used in a raffle i.e. 1-1,000.
4. Fold each of them and mix them nicely in a container.
18
Jane Muchembi
5. Pick the number of tickets equal to the sample size chosen e.g. if a sample of 10% is required
pick 100 tickets and individuals represented by the numbers in the 100 tickets picked will be the
ones to make up the sample.
But this mechanical procedure would be tedious and the quality of the sample would depend on
how thoroughly you mixed them up and how randomly you reached in.
With the development of inexpensive computers there is a much easier way. Here's a simple
procedure that's especially useful if you have the names of the clients already on the computer.
Many computer programs can generate a series of random numbers. Let's assume you can copy and
paste the list of client names into a column in an EXCEL spreadsheet. Then, in the column right
next to it paste the function =RAND() which is EXCEL's way of putting a random number between
0 and 1 in the cells. Then, sort both columns -- the list of names and the random number -- by the
random numbers. This rearranges the list in random order from the lowest to the highest random
number. Then, all you have to do is take the first hundred names in this sorted list. Pretty simple.
You could probably accomplish the whole thing in under a minute.
Simple random sampling is simple to accomplish and is easy to explain to others. Because simple
random sampling is a fair way to select a sample, it is reasonable to generalize the results from the
sample back to the population. Simple random sampling is not the most statistically efficient
method of sampling and you may, just because of the luck of the draw, not get good representation
of subgroups in a population. To deal with these issues, we have to turn to other sampling methods.
ii. Systematic random sampling
It involves drawing every kth elements in the population starting with a randomly selected element
between 1 and N.
Here are the steps you need to follow in order to achieve a systematic random sample:
number the units in the population from 1 to N (N= number the units in the population)
decide on the sample size that you want or need (n=sample size)
determine the interval size (k ). The interval size =N/n
randomly select an integer between 1 to N
then take every kth unit
e.g. if 10% sample is required from a population of 1,000 people (i.e. 100 people)
1. The researcher obtains or makes a list of all the members of the population and assigns to each
of them a number or letter i.e. if the whole population comprises of 1,000 people make a list of
1,000.
Pick the first element (number) at random
2. Pick every 10th (1000/100) number from the randomly selected number until you get the 100
required e.g. if 1 (one)is the randomly selected number then the next number picked is, 10,
then 20, then 30, the 40 etc. until a sample of 100 is picked.
1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22,23, 24, 25, 26,27,27,
29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53,
54, 55-----------1,000
19
Jane Muchembi
Objective: Divide the population into non-overlapping groups (i.e., strata) N1, N2, N3, ... Ni, such
that N1 + N2 + N3 + ... + Ni = N. Then do a simple random sample of f = n/N in each strata.
There are several major reasons why you might prefer stratified sampling over simple random
sampling. First, it assures that you will be able to represent not only the overall population, but
also key subgroups of the population, especially small minority groups. If you want to be able to
talk about subgroups, this may be the only way to effectively assure you'll be able to. If the
subgroup is extremely small, you can use different sampling fractions (f) within the different
strata to randomly over-sample the small group (although you'll then have to weight the within-
group estimates using the sampling fraction whenever you want overall population estimates).
When we use the same sampling fraction within strata we are conducting proportionate stratified
random sampling. When we use different sampling fractions in the strata, we call this
disproportionate stratified random sampling. Second, stratified random sampling will generally
have more statistical precision than simple random sampling. This will only be true if the strata
or groups are homogeneous. If they are, we expect that the variability within-groups is lower
than the variability for the population as a whole. Stratified sampling capitalizes on that fact. For
example, let's say that the population of clients for our agency can be divided into three groups:
Caucasian, African-American and Hispanic-American. Furthermore, let's assume that both the
African-Americans and Hispanic-Americans are relatively small minorities of the clientele (10%
and 5% respectively).
If we just did a simple random sample of n=100 with a sampling fraction of 10%, we would
expect by chance alone that we would only get 10 and 5 persons from each of our two smaller
groups. And, by chance, we could get fewer than that! If we stratify, we can do better.
20
Jane Muchembi
First, let's determine how many people we want to have in each group. Let's say we still want to
take a sample of 100 from the population of 1000 clients over the past year. But we think that in
order to say anything about subgroups we will need at least 25 cases in each group.
Our within-stratum sampling fraction for them is 50/850 = about 5.88%. Because the groups are
more homogeneous within-group than across the population as a whole, we can expect greater
statistical precision (less variance). And, because we stratified, we know we will have enough
cases from each group to make meaningful subgroup inferences.
21
Jane Muchembi
sampling methods that are considerably more complex than these simple variations. The most
important principle here is that we can combine the simple methods described earlier in a variety
of useful ways that help us address our sampling needs in the most efficient and effective manner
possible. When we combine sampling methods, we call this multi-stage sampling. The sample
is selected in multiple steps e.g. in a study on the factors that influence the choice of shoes by the
university students in Kenya, The researcher can follow the following steps:-
1. Cluster sampling of counties in which the universities are located as the first stage of the
process.
2. We could then have a two-stage sampling process with stratified samples within cluster
samples to select the counties.
3. Randomly select the universities in each sub-county in the sample
4. Randomly select the schools in each university sampled
5. Randomly select the faculties
6. Randomly select the departments
7. Randomly select the classes
8. Randomly select the individuals to be n the sample in each class
v. Snowballing
Used when the population is difficult to locate or identify e.g. when studying a sensitive
population like prostitutes, gays or lesbians.
22
Jane Muchembi
i. The researcher identifies a few individuals, gathers information from them and requests
them to help identify others in the same category or group.
ii. The researcher then collects information from the ones referred by the previous group
and then asks them to help identify others and the process continues until the sample is
enough.
vi. Volunteer/ self-selection: The research collect information from anybody willing to participate.
This may introduce bias since volunteers are usually different from non-volunteers.
VALIDITY: The validity is the extent to which the data accurately measures what they were
intended to measure.
ii. Construct Validity: Construct validity refers to the degree to which inferences can legitimately
be made from the operationalization in your study to the theoretical constructs on which those
operationalization were based.
Like external validity, construct validity is related to generalizing. But, where external validity
involves generalizing from your study context to other people, places or times, construct validity
involves generalizing from your program or measures to the concept of your program or measures.
You might think of construct validity as a "labeling" issue. When you implement a program that
you call a "Head Start" program, is your label an accurate one? When you measure what you term
"self esteem" is that what you were really measuring?
iii. Conclusion Validity: Conclusion validity is the degree to which conclusions we reach about
relationships in our data are reasonable. Conclusion validity is only concerned with whether there
is a relationship. For instance, in a program evaluation, we might conclude that there is a positive
relationship between our educational program and achievement test scores -- students in the
program get higher scores and students not in the program get lower ones. Conclusion validity is
essentially whether that relationship is a reasonable one or not, given the data. But it is possible
that we will conclude that, while there is a relationship between the program and outcome, the
program didn't cause the outcome. Perhaps some other factor, and not our program, was
23
Jane Muchembi
responsible for the outcome in this study. For instance, the observed differences in the outcome
could be due to the fact that the program group was smarter than the comparison group to begin
with. Our observed posttest differences between these groups could be due to this initial difference
and not be the result of our program. This issue -- the possibility that some other factor than our
program caused the outcome -- is what internal validity is all about. So, it is possible that in a
study we can conclude that our program and outcome are related (conclusion validity) and also
conclude that the outcome was caused by some factor other than the program (i.e., we don't have
internal validity).
iv. External Validity: External validity is related to generalizing. That's the major thing you need
to keep in mind. Recall that validity refers to the approximate truth of propositions, inferences, or
conclusions. So, external validity refers to the approximate truth of conclusions the involve
generalizations. Put in more pedestrian terms, external validity is the degree to which the
conclusions in your study would hold for other persons in other places and at other times.
24
Jane Muchembi
RELIABILITY: Reliability is the extent to which the data collection method will yield consistent
findings if replicated by others.
NB: Validity and reliability of the data you collect depend on the design of the questionnaire and
the words that that are used by the researcher.
2. SECONDARY DATA: Is gathered from secondary sources such as books, journals, magazines
etc. it is often referred to as desk research.
1. QUESTIONNAIRE
A questionnaire is a data collection method that uses an instrument or tool called a
questionnaire.
This instrument comprises of a list of written down questions based on the research problem,
objectives and questions which if answered will help the researcher to answer the research
questions and meet the research objectives.
The researcher gives the questionnaire to the respondents to fill on their own either in print form
or online and then analysis the data to make conclusions.
25
Jane Muchembi
To ensure effectiveness of the instrument a pre-test or pilot test should be carried out on a small
representative sample to enable the researcher find out if the questions are measuring what they
are supposed to measure and if the wording is clear i.e. All questionnaires should be piloted, if
possible, with a small group before the main research to assess their value, validity and
reliability.
The questions used may be:-
a) Closed questions: where a limited number of alternative responses to the set question are
provided. These can be in list, category, ranking, scale/rating, grid or other quantitative form.
They can be pre- coded on a questionnaire to facilitate analysis. e.g. Please tick the right answer
from the boxes shown below i.e. (specific choices of answers to choose from are provided) or
b) Open ended questions i.e. a question is posed, but space is left for the respondent’s own answer
i.e. no suggested answers.
Advantages of open ended questions Disadvantages of open ended questions
I. Enable you to get below the surface, I. The responses can be hard to collate
explore and probe ii. The research may be difficult for others to
ii. Encourages respondents to think and offer reproduce, so your findings may be open to
considered answers doubt or
iii. Encourages respondents to give honest
opinions
26
Jane Muchembi
27
Jane Muchembi
iv. Each question included has a specific purpose, and contributes to the study.
v. There are no leading questions.
vi. There is a balance of questions per topic.
BASIC PRINCIPLES TO BE OBSERVED WHEN PREPARING A QUESTIONNAIRE
i.Begin by giving a short rationale explaining the objectives of the study.
ii.Explain the potential benefits to the target population in terms they can easily understand.
iii.Make it as brief and simple to fill as possible and avoid irrelevant questions.
iv. Begin with straightforward non-threatening questions which the respondents will find easy to
answer. Personal questions should be left until the end.
v. Make the questionnaire attractive and spacious as a crowded questionnaire will discourage
responses.
2. INTERVIEW
An interview is a method of collecting data in which questions are asked orally and answers
recorded by the researcher or research assistant. This gives the interviewer an opportunity to
probe or ask for further clarification to get deeper understanding of the topic. The interviewer
can also clarify questions that may be misunderstood or misinterpreted.
FORMS OF INTERVIEWS
a) UNSTRUCTURED INTERVIEW
It is a method of collecting data in which:-
The researcher has some idea and may be a list of the topics to be covered but no specific
questions to be asked to the respondents.
They are informal and conversational with aim of encouraging the respondents to feel free to
open up and give information freely.
It is suitable when studying sensitive topics such as sexuality, some cultural issues etc.
Since no possible answers are suggested respondents give a wide variety of answers and
therefore data analysis can be difficult.
28
Jane Muchembi
c) STRUCTURED INTERVIEW
It is a method of collecting data in which:-
The interviewer is guided by an instrument called interview guide or schedule.
All the respondents are subjected to the same questions.
29
Jane Muchembi
30
Jane Muchembi
DEMEANOUR OF INTERVIEWER
Generally speaking, we often like people who appear to like us!
The interviewer should then, appear to be interested in the interviewee, but in a neutral and
detached way: ‘neutrally interested’ is perhaps the best way of describing the best interviewer-
interviewee relationship.
The tone of voice of the interviewer is important, as it should project an impression of quiet
confidence and quiet enthusiasm in the topic under discussion.
Torrington (1991) suggests that an open posture is best, where the interviewer sits slightly
forward toward the interviewee, keeps regular eye contact, and avoids folded arms.
The interviewer must avoid appearing shocked, disbelieving or astonished by comments made by
interviewees.
The interviewer wear clothes similar to those of the interviewees: too scruffy, or too over-
dressed can affect the credibility of the interviewer.
SUSPICION OF THE INTERVIEWER
People are increasingly suspicious of interviewers and their motives.
Most people have experience of being stopped in the street by an interviewer who appears to be
asking questions in a neutral way but is really seeking to make a marketing contact for a
commercial organization.
The true purpose of the interview should be carefully explained to the interviewee and how the
data collected will be used.
Wherever possible, the student researcher should have a letter from a University tutor explaining
the research initiative, e.g. it is part of a legitimate first or post-graduate course.
Wherever possible, the interviewer should send details of the interview process and agenda to
interviewees in advance, for example explaining the estimated length of time it will take, the
aim and purpose of the questions to be asked and the range of questions likely to be asked.
CONFIDENTIALITY
Part of the suspicion shown to ward interviewers concerns the issue of confidentiality, and
interviewees may be worried about disclosing sensitive information in case it has negative
repercussions in some way against them.
Interviewees may also be concerned that their personal details would be passed on to commercial
organisations and that they would be subsequently pestered to buy things.
Interviewees must be given a complete reassurance about confidentiality and told who will see
the data obtained – and don’t forget this is likely to include at least two university tutors and
possibly one external examiner
The interviewee should be assured that his or her views will be generalized in the final report or
that any direct quotes used would remain anonymous unless the interviewee wishes otherwise.
The interviewer needs subsequently to take pains to avoid revealing the identity of respondents
by using false names and not giving any clues to the identity of any interviewee.
3. OBSERVATION
31
Jane Muchembi
It a method or technique of collecting data that uses tool or instrument called observation schedule
or guide. It provides information about the actual behavior of the respondents.
FORMS OF OBSERVATION
a) Participant observation: is a qualitative data collection method in which the researcher
becomes an active functioning member of the population being investigated e.g. a designer
wishing to study university students dressing behavior enrolls as a student. Its emphasis is on
discovering meanings that people attach to their actions (Saunders, Lewis and Adrian 2004 and
Kombo and Tromp 2006).
b) Structured observation: I quantitative and more concerned with frequency of those actions.
The observer/researcher is an onlooker (Saunders, Lewis and Adrian 2004).
32
Jane Muchembi
Saunders, Lewis and Thornhill (2003, p. 131) summarise the main issues
33
Jane Muchembi
to consider, although the ethical issues surrounding these items are not
always clear-cut:
i. The rights of privacy of individuals
ii. Voluntary nature of participation – and the rights of individuals to withdraw partially or
completely from the process
iii. Consent and possible deception of participants
iv. Maintenance of the confidentiality of data provided by individuals or identifiable
participants and their anonymity
v. Reactions of participants to the ways in which researchers seek to collect data
vi. Effects on participants of the way in which data is analyzed and reported
vii. Behaviour and objectivity of the researcher
Data analysis involves examining data in ways that reveal the relationships, patterns, trends, etc. that
can be found within it. That may mean subjecting it to statistical operations that can tell you not only
what kinds of relationships seem to exist among variables, but also to what level you can trust the
answers you’re getting. It may mean comparing your information to that from other groups (a
control or comparison group, statewide figures, etc.), to help draw some conclusions from the
data. The point, in terms of your evaluation, is to get an accurate assessment in order to better
understand your work and its effects on those you’re concerned with, or in order to better understand
the overall situation.
There are two kinds of data the researcher may be working with, although not all evaluations will
necessarily include both:-
i. Quantitative data
This refers to the information that is collected as, or can be translated into, numbers, which can then
be displayed and analyzed mathematically. Some examples include:
The frequency (rate, duration) of specific behaviors or conditions
Test scores (e.g., scores/levels of knowledge, skill, etc.)
Survey results (e.g., reported behavior, or outcomes to environmental conditions; ratings
of satisfaction, stress, etc.)
Numbers or percentages of people with certain characteristics in a population (e.g. have
design skills, have carpeted offices, wear high heeled shoes, buy second hand clothes etc.)
Data can also be collected in forms other than numbers, and turned into quantitative data for
analysis. Researchers can count the number of times an event is documented in interviews or
records, for instance, or assign numbers to the levels of intensity of an observed event or
34
Jane Muchembi
behavior. For instance, community initiatives often want to document the amount and intensity of
environmental changes they bring about – the new programs and policies that result from their
efforts. Whether or not this kind of translation is necessary or useful depends on the nature of what
you’re observing and on the kinds of questions your evaluation is meant to answer.
Quantitative data is usually subjected to statistical procedures such as calculating the mean or
average number of times an event or behavior occurs (per day, month, year). These operations,
because numbers are “hard” data and not interpretation, can give definitive, or nearly definitive,
answers to different questions. Various kinds of quantitative analysis can indicate changes in a
dependent variable related to – frequency, duration, timing (when particular things happen),
intensity, level, etc. They can allow you to compare those changes to one another, to changes in
another variable, or to changes in another population. They might be able to tell you, at a particular
degree of reliability, whether those changes are likely to have been caused by your intervention or
program, or by another factor, known or unknown. And they can identify relationships among
different variables, which may or may not mean that one causes another.
Qualitative data can sometimes be changed into numbers, usually by counting the number of times
specific things occur in the course of observations or interviews, or by assigning numbers or ratings
to dimensions (e.g., importance, satisfaction, ease of use). The challenges of translating qualitative
into quantitative data have to do with the human factor. Even if most people agree on what 1
(lowest) or 5 (highest) means in regard to rating “satisfaction” with a program, ratings of 2, 3, and 4
may be very different for different people. Furthermore, the numbers say nothing about why people
reported the way they did. One may dislike the program because of the content, the facilitator, the
time of day, etc. The same may be true when you’re counting instances of the mention of an event,
such as the onset of a new policy or program in a community based on interviews or archival
records. Where one person might see a change in program he considers important another may
omit it due to perceived unimportance.
Qualitative data can sometimes tell you things that quantitative data can’t. It may reveal why
certain methods are working or not working, whether part of what you’re doing conflicts with
participants’ culture, what participants see as important, etc. It may also show you patterns – in
behavior, physical or social environment, or other factors – that the numbers in your quantitative
data don’t, and occasionally even identify variables that researchers weren’t aware of.
35
Jane Muchembi
In most social research the data analysis involves three major steps, done in roughly this order:
36
Jane Muchembi
In most social research, quality of measurement is a major issue. Assuring that the data
collection process does not contribute inaccuracies will help assure the overall quality of
subsequent analyses.
ii. Handling blank responses: If many questions e.g. 25% of the questions are not answered you
may decide to discard the questionnaire and not include it in data set for analysis and report or
account for them in your report. If there are only a few blank responses then decide how to
handle them.
iii. Coding data: involves allocating codes to possible answers e.g. in a question where possible
answers are yes or no, you may assign number one (1) to “yes” and (2) to “No”. Where there are
five possible answers you can have codes 1-5.
iv. Categorizing data: This involve grouping questions addressing the same variable together.
vi. Programming :
NB:
Organizing the data correctly can save a lot of time and prevent mistakes.
Most researchers choose to use a database or statistical analysis program (Microsoft Excel,
SPSS) that they can format to fit their needs in order to organize their data effectively.
A good researcher enters all of the data in the same format and in the same database, as doing
otherwise might lead to confusion and difficulty with the statistical analysis later on.
Once the data has been entered, it is crucial that the researcher check the data for accuracy.
This can be accomplished by spot-checking a random assortment of participant data groups, but
this method is not as effective as re-entering the data a second time and searching for
discrepancies.
37
Jane Muchembi
This method is particularly easy to do when using numerical data because the researcher can
simply use the database program to sum the columns of the spreadsheet and then look for
differences in the totals.
Perhaps the best method of accuracy checking is to use a specialized computer program that
cross-checks double-entered data for discrepancies (as this method is free from error), though these
programs can be hard to come by and may require extra training to use correctly.
In every research project, you should generate a printed codebook that describes the data and
indicates where and how it can be accessed. Minimally the codebook should include the
following items for each variable:
variable name
variable description
variable format (number, data, text)
instrument/method of collection
date collected
respondent or group
variable location (in database)
notes
The codebook is an indispensable tool for the analysis team. Together with the database, it
should provide comprehensive documentation that enables other researchers who might
subsequently want to analyze the data to do so without any additional information.
Data Transformations
Once the data have been entered it is almost always necessary to transform the raw data into
variables that are usable in the analyses. There are a wide variety of transformations that you
might perform. Some of the more common are:
missing values
Many analysis programs automatically treat blank values as missing. In others, you need to
designate specific values to represent missing values. For instance, you might use a value of -99
to indicate that the item is missing. You need to check the specific program you are using to
determine how to handle missing values.
item reversals
38
Jane Muchembi
On scales and surveys, we sometimes use reversal items to help reduce the possibility of a
response set. When you analyze the data, you want all scores for scale items to be in the same
direction where high scores mean the same thing and low scores mean the same thing. In these
cases, you have to reverse the ratings for some of the scale items. For instance, let's say you had
a five point response scale for a self-esteem measure where 1 meant strongly disagree and 5
meant strongly agree. One item is "I generally feel good about myself." If the respondent
strongly agrees with this item they will put a 5 and this value would be indicative of higher self
esteem. Alternatively, consider an item like "Sometimes I feel like I'm not worth much as a
person." Here, if a respondent strongly agrees by rating this a 5 it would indicate low self esteem.
To compare these two items, we would reverse the scores of one of them (probably we'd reverse
the latter item so that high values will always indicate higher self esteem). We want a
transformation where if the original value was 1 it's changed to 5, 2 is changed to 4, 3 remains
the same, 4 is changed to 2 and 5 is changed to 1. While you could program these changes as
separate statements in most program, it's easier to do this with a simple formula like:
In our example, the High Value for the scale is 5, so to get the new (transformed) scale value, we
simply subtract each Original Value from 6 (i.e., 5 + 1).
scale totals
Once you've transformed any individual scale items you will often want to add or average across
individual items to get a total score for the scale.
categories
For many variables you will want to collapse them into categories. For instance, you may want to
collapse income estimates (in dollar amounts) into income ranges.
ii. Describing the data (Descriptive Statistics) : Descriptive Statistics are used to describe the
basic features of the data in a study. They provide simple summaries about the sample and the
measures. Together with simple graphics analysis, they form the basis of virtually every quantitative
analysis of data. With descriptive statistics you are simply describing what is, what the data shows.
iii. Testing Hypotheses and Models (Inferential Statistics): Inferential Statistics investigate
questions, models and hypotheses. In many cases, the conclusions from inferential statistics extend
beyond the immediate data alone. For instance, we use inferential statistics to try to infer from the
sample data what the population thinks. Or, we use inferential statistics to make judgments of the
probability that an observed difference between groups is a dependable one or one that might have
happened by chance in this study. Thus, we use inferential statistics to make inferences from our
data to more general conditions; we use descriptive statistics simply to describe what's going on in
our data.
In most research studies, the analysis section follows these three phases of analysis. Descriptions of
how the data were prepared tend to be brief and to focus on only the more unique aspects to your
study, such as specific data transformations that are performed. The descriptive statistics that you
actually look at can be voluminous. In most write-ups, these are carefully selected and organized
into summary tables and graphs that only show the most relevant or important information. Usually,
the researcher links each of the inferential analyses to specific research questions or hypotheses that
39
Jane Muchembi
were raised in the introduction, or notes any models that were tested that emerged as part of the
analysis. In most analysis write-ups it's especially critical to not "miss the forest for the trees." If
you present too much detail, the reader may not be able to follow the central line of the results.
Often extensive analysis details are appropriately relegated to appendices, reserving only the most
critical analysis summaries for the body of the report itself.
40
Jane Muchembi
two numbers in the middle. It is useful when the data set has an outlier and values distribute very
unevenly.
c) “Mode” is the value which is observed most often. It is useful when the data is non-numeric
or when asked to find the most popular item.
v. “Range” is the difference of the maximum value and the minimum value for the variable. For
the “age example”, the maximum value is 54 and the minimum value is 19. So the range is 35.
vi. “Standard deviation” shows how much variation the value exits from the mean. Variance is
the average of squared difference from the mean. Standard deviation is the square root of
variance.
2. Inferential Statistics
Investigate questions, models and hypotheses.
In many cases, the conclusions from inferential statistics extend beyond the immediate data
alone. For instance, we use inferential statistics to try to infer from the sample data what the
population thinks.
Or, we use inferential statistics to make judgments of the probability that an observed difference
between groups is a dependable one or one that might have happened by chance in this study.
Thus, we use inferential statistics to make inferences from our data to more general conditions;
we use descriptive statistics simply to describe what's going on in our data.
Inferential statistics allow the researcher to begin making inferences about the hypothesis on the
basis of the data collected. This means that, while applying inferential statistics to data, the
researcher is coming to conclusions about the population at large. Inferential statistics seek to
generalize beyond the data in the study to find patterns that ostensibly exist in the target
population. This course will not address the specific types of inferential statistics available to the
researcher, but a succinct and very useful summary of them, complete with step-by-step examples
and helpful descriptions, is available
i. T-Tests
T-Tests are used to test if the difference of means is statistically significant. It tests if the sample
is representative of the populations. For example, if the mean for variable 1 is 20 and the mean
for variable 2 is 28, you may say the means are different. T-Tests may show you that they are not
significantly different, however, and you can’t base your conclusion on the means’ difference
since the difference in the sample is not representative for the population.
41
Jane Muchembi
This analysis is appropriate whenever you want to compare the means of two groups, and
especially appropriate as the analysis for the posttest-only two-group randomized experimental
design.
ii. Correlation
The correlation is one of the most common and most useful statistics. A correlation is a single
number that describes the degree of relationship between two variables. Let's work through an
example to show you how this statistic is computed.
Correlations are used when you want to know about the relationship between two variables. For
example, you want to know consumers’ willingness to pay and their ratings for the product quality.
If the correlation is 1, meaning the willingness to pay and the ratings for the product quality are
completely positively correlated and if the correlation is 0, meaning there is no correlation between
these two variables. If the correlation is -1, it shows they are completely negatively correlated,
meaning the higher one variable, the lower the other variable. If the absolute value of the variables
is bigger than 0.5, they are usually significant.
Correlation is one of the most often used (and most often misused) kinds of descriptive statistics.
It is perhaps best described as a single number that describes the degree of relationship between
two variables.
If two variables tend to be “correlated,” that means that a participant’s score on one tends to
vary with a score on the other. For example, people’s height and shoe size tend to be positively
correlated. This means that for the most part, if a given man is tall, he is likely to have a large
shoe size. If short, he is likely to have a smaller shoe size.
Correlation can also be negative. For example, the temperature outside in Fahrenheit may be
negatively correlated with the number of hot chocolates sold at a local coffee shop. This is to
say that as the temperature goes down, hot chocolate sales tend to go up.
Although causality may seem to be implied in this situation, it is important to note that on a
statistical level, correlation does not imply causation.
A good researcher knows that there is no way to assess from correlation alone that a causal
relationship exists between two variables.
In order to assert that “X caused Y”, a study should be experimental, with control groups and
random sampling procedures.
Determining causation is a difficult thing to do, and it is a common mistake to assert a cause-
and-effect relationship when the study methodology does not support this assertion.
For y = ax + b, y is the dependent variable, x is the causal variable and the intercept is a,
indicating the correlation between x and y. If “a” is 0.2 for example, it means when x variable
increases 1 unit, y increases 0.2 units. If “a” is negative, meaning y decreases as x increases.
For y = ax12 + bx2 + c, y is the dependent variable, x1 is causal variable 1 and x2 is causal variable
2. “a” is the intercept for variable 1 and “b” for variable 2. For example, if y = 0.6 x12 – 0.4 x2 +
42
Jane Muchembi
0.23, it means when x1 increases 1 unit, y increases 0.6 units and when x2 increases 1 unit, y
decreases 0.4 units. (Given the variables are statistically significant.)
Regression is a more accurate way to test the relationship between the variables compared with
correlations since it shows the goodness of fit (Adjusted R Square) and the statistical testing for
the variables. The formulas for one-variable regressions is y = ax + b and for multiple
regressions is y = ax12 + bx2 + c.
For y = ax + b, y is the dependent variable, x is the causal variable and the intercept is a,
indicating the correlation between x and y. If “a” is 0.2 for example, it means when x variable
increases 1 unit, y increases 0.2 units. If “a” is negative, meaning y decreases as x increases.
For y = ax12 + bx2 + c, y is the dependent variable, x1 is causal variable 1 and x2 is causal variable
2. “a” is the intercept for variable 1 and “b” for variable 2. For example, if y = 0.6 x12 – 0.4 x2 +
0.23, it means when x1 increases 1 unit, y increases 0.6 units and when x2 increases 1 unit, y
decreases 0.4 units. (Given the variables are statistically significant.)
Statistical Significance
Researchers cannot simply conclude that there is a difference between two groups in a well-
constructed study. This difference must be due to the manipulation of the independent variable.
No matter how well a researcher designs the study, there always exists a degree of error in the
results.
This error can be due to individual differences both within and between experimental groups, or
the error can be due to systematic differences within the researcher’s sample.
Irrespective of its source, this error acts as a kind of “noise” in the data. It affects participants’
scores on study measures even though it is not the variable of interest.
Statistical significance is aimed at determining the probability that the observed result of a study
was due to the influence of the independent variable rather than by chance.
A result is “statistically significant” at a certain level. For example, a result might be significant
at p<.05. “P” represents the probability that the result was due to chance, and .05 represents a 5%
probability that the result was due to chance.
Therefore, p<.05 means that inferential statistical analysis has indicated that the observed results
have over a 95% probability of being due to the influence of the independent variable.
The 5% cutoff is generally thought of as the standard for most scientific research. Note that it is
theoretically impossible to ever be entirely certain that one’s results are not due to chance, as the
nature of science is one of falsification, not immutable proof.
iv. Analysis of Variance (ANOVA): It helps to examine the significant mean differences among
more than two groups on an interval or ratio scaled dependent variable.
43
Jane Muchembi
Data analysis procedures (DAP) help the researcher to arrive at the data analysis. The uses of such
procedures put the research project in perspective and assist the researcher in testing the
hypotheses with which he/she has started the research. Hence with the use of DAP, one can:-
NB: Understanding of the data analysis procedures/techniques helps the researcher to:-
appreciate the meaning of the scientific method, hypotheses testing and statistical significance in
relation to research questions
realize the importance of good research design when investigating research questions
have knowledge of a range of inferential statistics and their applicability and limitations in the
context of your research
be able to devise, implement and report accurately a small quantitative research project
be capable of identifying the data analysis procedures relevant to your research project
show an understanding of the strengths and limitations of the selected quantitative and/or
qualitative research project
demonstrate the ability to use word processing, project planning and statistical computer packages
in the context of a quantitative research project and report
be adept of working effectively alone or with others to solve a research question/ problem
quantitatively.
44
Jane Muchembi
The analytical techniques will determine the recording style that will be used during the data
collection.
2. Thematic Analysis: Analyzing data using themes. Themes refer to topics or major subjects that
come up in discussions. In this method the researcher categorizes related topics and identifies
major concepts or themes and then does the following:-
i. Peruses the data collected and identifies the information that is relevant to the research questions
and objectives.
ii. Develops a coding system based on the samples of the collected data
iii. Classifies majors issues or topics covered
iv. Rereads the text and highlights key quotations/insights and interpretations.
v. Indicates major themes in the margins
vi. Places coded materials under major themes or topics identified
vii. Develops a summary report identifying major themes and associations between them
viii. Uses graphics and direct quotations to present the findings.
ix. Reports the intensity, which refers to the number of times certain words or phrases or
descriptions are used in the discussion. NB: the frequency with which an idea, or a word, or
description appears is used to interpret its importance, attention or emphasis.
NB: The analysis of quantitative data varies from simple to more elaborate analysis techniques. The
analysis varies with the objectives of the research, its complexity and extent to which conclusions
can be easily reached. Data analysis in quantitative research depends on the type of study for
example in:-
i. Correlation studies
45
Jane Muchembi
46
Jane Muchembi
In the case of causal-comparative research the independent variable is established by the identity
of the groups chosen and not under experimental control.
In experimental designs the observer should decide before carrying out the experiment the
analytical process.
The analytical process in experimental studies mainly involves the calculation of effect size.
Effect size is the mean of experimental group minus the control group.
The idea is to calculate the effect size across a number of studies to determine the relevance of
the test, treatment, or method.
DATA PRESENTATION
Data may be presented using frequency tables, graphs or charts to make it easier for the readers
to understand what has been found out by the research.
All the tables, graphs should be clearly labeled to make the readers understand them at a glance.
Data can be presented under in line with the objectives or research questions.
Frequency Distribution
Frequency distribution is a simple data analysis technique which allows you to get a big picture
of the data. From frequency distribution, you can see how frequently the specific values are
observed and what their percentages are for the same variable. For example: for variable of
“age,” you can use frequency distribution to figure out how many people in the survey are aged
18 to 25, and how many are aged 26 to 33, etc.
47
Jane Muchembi
5.2 Discussions
Explain how the results of your research relate with the results of other studies in the literature.
Explain whether they support or refute or contradict the previous theories.
Identify any weaknesses or limitations in your study and indicate how they could be eliminated
in future studies e.g. an ethical issue that became apparent after you had already stated the
research, or a weakness in the research design etc.
5.3 Conclusions
Make inferences on based on the findings.
5.4 Recommendations
Make recommendations or suggestions on what can be done to make the situation you
investigated better. Also make recommendations or suggestions on areas or topics that need
further research.
Ethical Issues
There are a number of key phrases that describe the system of ethical protections that the
contemporary social and medical research establishment have created to try to protect better the
rights of their research participants.
The principle of voluntary participation requires that people not be coerced into participating
in research. This is especially relevant where researchers had previously relied on 'captive
audiences' for their subjects -- prisons, universities, and places like that.
Ethical standards also require that researchers not put participants in a situation where they
might be at risk of harm as a result of their participation. Harm can be defined as both
physical and psychological.
There are two standards that are applied in order to help protect the privacy of research
participants. Almost all research guarantees the participants confidentiality -- they are assured
that identifying information will not be made available to anyone who is not directly involved
in the study.
The stricter standard is the principle of anonymity which essentially means that the participant
will remain anonymous throughout the study -- even to the researchers themselves. Clearly,
the anonymity standard is a stronger guarantee of privacy, but it is sometimes difficult to
accomplish, especially in situations where participants have to be measured at multiple time
points (e.g., a pre-post study).
48
Jane Muchembi
Increasingly, researchers have had to deal with the ethical issue of a person's right to service.
Good research practice often requires the use of a no-treatment control group -- a group of
participants who do not get the treatment or program that is being studied. But when that
treatment or program may have beneficial effects, persons assigned to the no-treatment
control may feel their rights to equal access to services are being curtailed.
TIME/WORK PLAN: outlines the main activities of the research and when each will be
carried out. It should be on its own page.
EXAMPLE OF TIME PLAN
Activity YEAR 2021 YEAR 2022
March April may June July Aug Sep Oct Nov Dec Jan Feb March
Identify topic and write
title
Background to the study
Problem statement
Research objectives and
questions
Complete chapter one
Literature review
Research methodology
Compiling and handing
in research proposal
Corrections on the
proposal
Handing in corrected
proposal
Data collection and
analysis
Research findings,
conclusions and
recommendations
49
Jane Muchembi
BUDGET
Shows the approximated cost of the research project from the time one starts writing the proposal
to the time the project is completed. It should be as detailed and realistic as possible i.e. Show all
the items you will require to complete the project, quantities, cost per item and total cost. You
can also include 10% to cater for inflation. It should be on its own page.
Example of a budget
Item No. of units price@ item Kshs) Total cost (Kshs)
Printing papers 2 realms 400 800
Printing the proposal 30 pages 10 300
Photocopying proposals 30 pages 2 60
Internet services 1000
Transport 2 trips 500 1000
Total 3160
APPENDICES
Includes all the materials that do not fit well within the project or proposal e.g. sample
Instruments.
REFERENCES
It refers to materials that have been cited in the proposal or project. Lists all the Publications
cited in the research proposal or project. It should be arranged in alphabetical order. There are
several referencing styles but for social sciences APA style is recommended (Refer to the most
current edition of APA Manual). It should start on a fresh page.
REFERENCES
Babbie, E. (2001). The practice of social Research: 9th Ed. CA: Wadworth Thomson.
50
Jane Muchembi
Saunders, M., Lewis, P. & Thornhill, A. (2007). Research Methods for Business Students (4th
edition) Harlow: Prentice Hall.
Trochim, W. M. K. “Data Preparation” Research Methods Knowledge Base 2nd Edition.
Accessed 2/24/09.
Trochim, W. M. K. “Descriptive Statistics” Research Methods Knowledge Base 2nd Edition.
Accessed 2/24/09.
Trochim, W. M. K. “Descriptive Statistics” Research Methods Knowledge Base 2nd Edition.
Accessed 2/24/09.
Trochim, W. M. K. “Inferential Statistics” Research Methods Knowledge Base 2nd Edition.
Accessed 2/24/09.
Pitfalls of Data Analysis (or How to Avoid Lies and Damned Lies) by Clay Helberg, M.S.,
Research Design and Statistics Unit
51