Note-For-Exam - Market Research UHasselt
Note-For-Exam - Market Research UHasselt
Note-For-Exam - Market Research UHasselt
1. Introduction
2. Qualitative research techniques:
+ focus group: how to organize, conducting the focus group , participant have interest in research topic
+ interview: the sample size, need to design your sample (who should be included, how many people -
to realize the situation), guide - how do you do structure of interview (need some ideas, need to write
down the topic, dont try to formulate the questions because want people to talk about it, how feel
question asked, get people's time; guide is the simply list of topics in interest)
Brand personification is a Projective Technique that asks people to think about brands as if they were
people and to describe how the brands would think and feel.
- Advantages:
o May elicit responses that participants would be unwilling or unable to give if they knew the
purpose of study.
o Frames the actions as those of “someone else”.
o Increasing the validity by disguising the purpose of study.
o Helpful when underlying motivations, beliefs, and attitudes that operate at a subconscious
level (not aware of them).
- Disadvantage:
o Unstructured direct technique at a greater level require experienced interviewers who are
expensive.
o Risk of interpretation bias.
o Some participants may lack self-confidence or the ability to express themselves fully with
some techniques that show unusual behavior (e.g., role playing)
- Projective techniques are used less frequently than other unstructured direct techniques, except
word association should be used when required info can’t be accurately obtained by direct
questioning, participants find difficult to conceive and express issues, engaging participants in the
subject with an interesting and novel way.
2.4.Ethnography
- Ethnography is a research approach based upon the observation of the customs, habits, and
differences between people in everyday situations. Observation may be direct or indirect (e.g. via
written material)
- Ethnography is about entering repondents’ natural life worlds (home, shopping etc.). This provides a
holistic, natural, and nuanced view.
- Ethnography is a mix of observation and interviewing, but always takes place in the object’s natural
environment (no laboratory settings)
- People are natural habitat
ADVANTAGES
o Many objects simultaneously efficiency
o Versatile / Many behaviors can be observed
o Usually fast (can ask direct questions)
o Relatively inexpensive
DISADVANTAGES
o Limited generalizability
o Usually limited to overt behavior (need to learn techniques, done in an open way)
o Usually limited insight in the object’s attitudes, motivations etc.
o Ethical concerns
3. Online qualitative research
3.1.ONLINE FOCUS GROUPS
- Participants use the technology of the Internet to approximate the interaction of a face-to-face
focus group. Typically respondents are at different locations as well as the moderator.
- RELATIVE DISADVANTAGES ONLINE FGs
o Lack of group dynamics
o Conservation less fluent and slower (less probing, fewer insights)
o Non-verbal communication cannot be observed
o It is difficult to ascertain who actually participates
o Respondents may be engaged in other activities
- RELATIVE ADVANTAGES ONLINE FGs
o Cheaper
o No geographical boundaries in terms of participants
o No traveling; busy respondents more likely to join
o Instant data recording and data analysis
o Openness due to anonymity (unknown to other people)
3.2.NETHNOGRAPHY
- Netnography, or ethnography on the Internet, is a new qualitative research methodology that
adapts ethnographic research techniques to study the cultures and communities that are emerging
through computer-mediated communications.
- Consumers increasingly turn to computer-mediated communication (Internet) for information to
help them in their purchase decisions.
- Usefulness depends on setting
Suitable online communities
4. Qualitative research: what to ask?
-
- Fit
- Keep eye on ball
- Iterative
5. Wrap up
- Depending on the marketing research problem qualitative research may be preferred over
quantative research (and vice versa, of course)
- A wide variety of tools is available. Again, there should be a match between research problem/topic
and qualitative research technique employed
- Qualitative research is based on at least 2 intellectual traditions: a set of ideas and associated
methods from the in-depth psychology and motivational research, and from sociology.
- “exploratory vs conclusive” and “qualitative vs. quantitative” are parallel but not identical.
- Not possible to say that one technique is better or worse than other to choose which, based on
the confidence of marketing decision makers in using technique.
MEASUREMENTS
1. Theory
- A theory is a proposed description, explanation, or model of the manner of interaction of a set of ..
phenomena, capable of predicting future occurrences or observations of the same kind, and capable
of being tested through experiment or otherwise falsified through empirical observation.
- Theories:
o ...in Marketing are (usually) not that theoretical (as opposed to practical)!
o …are better, the more they prohibit (following Popper).
o …are not tautological (saying the same thing twice in different words) they explain
something.
o …ought to be empirically testable or falsifiable.
o …consist of constructs (concepts, phenomena, variables) and hypotheses (on their interactions
or relationships).
o Popper: the quality of a theory is its falsifiability, or refutability, or testability
2. Hypothesis:
- Usually consists of two parts:
o a condition
o a consequence
- Each of the parts contains a construct
o the independent variable (condition)
o the dependent variable (consequence)
- Examples:
o The higher A, the higher B.
o A leads to B.
3. Construct
- A construct:
o … is ‘‘a conceptual term used to describe a phenomenon of theoretical interest’’
(Edwards/Bagozzi, 2000, pp. 156–157).
o … is quantifiable and directly, or indirectly observable
o An indirectly observable construct is called ‘latent’
o Example: IQ
- Constructs MUST be defined in terms of:
o Object (e.g., somebody is loyal)
o Attribute (loyalty)
o Rater entity (extreme high or low value of loyalty)
- Researchers in marketing usually want to investigate relationships between constructs
o direct causal relationships
Usually, but not necessarily, a linear effect is meant.
A is called an exogenous variable; B is called an endogenous variable.
AB
o (fully or partially) mediated (indirect) causal relationships
A appears statistically to have a direct effect on B. Logically, however, A influences Z, and Z
influences B.
Z is called a mediator (variable).
Mediation is called partial if the effect between A and B remains significant after inclusion of
the mediator.
o spurious relationships
A third variable (here: Z) influences A as well as B.
Example:
There is a relationship between the temperature of the pavement in Chicago and the
fertility of birds in Norway.
Via statistics, may see relationship between A and B, but not, it’s because Z influence both A
and B.
- Indicators never measure exactly the constructs, good research try to minimize errors
- Measurement model: Try to measure different constructs
- Structural model: Show relations between constructs
- Nomenclature:
o Indicators are normally represented as squares. For questionnaire based research, each x
indicator would represent a particular question. 11
o Latent variables are normally drawn as circles or ovals. Latent variables are used to represent
phenomena that cannot be measured directly. Examples would be beliefs, intention, motivation. η1
o Few researchers expect their models to perfectly explain reality. They therefore explicitly model
structural error.
o In the case of error terms, for simplicity, the circle is often left off. ε21
- Multi-item measurement:
o increases reliability and validity of measures
o allows measurement assessment
measurement error
reliability
validity
o two forms of measurement models:
formative (emerging)
reflective (latent)
- reflective measurement model:
o direction of causality is from construct to measure
o indicators expected to be correlated
o dropping an indicator from the measurement model does not alter
the meaning of the construct
o takes measurement error into account at the item level
o similar to factor analysis
o typical for consumer research constructs (e.g. attitudes)
- formative measurement model:
o direction of causality is from measure to construct
o no reason to expect indicators to be correlated
o dropping an indicator from the measurement model may alter the
meaning of the construct
statistical tests of reliability and validity do not make any sense
o based on multiple regression
beware of multicollinearity!
o typical for success factor research
MEASUREMENT
1. Introduction
Measurement
- In data collection stage
- Measurement is crucial for all research designs!
- In research, measurement consists of assigning numbers
to empirical events (that is, properties, attributes or
characteristics of an object) in compliance with a set of
rules.
- Basically, measurement is a 3-part process consisting of:
o 1) Selecting empirical events: that is decide on what
property/characteristic/attribute you want to measure
(gender, preference, perceptions, height, income)
o 2) Developing a set of mapping rules or a scheme for
assigning numbers or symbols to represent aspects of
the event being measured.
o 3) Apply the mapping rule(s) to each observation of the
empirical event
- Illustrate with example of last meeting’s in-class assignments. To make the distinction between
research questions and the measurement questions. Also link to “detailed” figure of problem
definition.
Scaling
- Is an extension of measurement
- Creating a continuum upon which the measured object is located
- The terms measurement and scaling are often used interchangeably
2. Scaling
2.1. Primary levels of measurement scales
- Respondents’ answers can be recorded in many different ways, but one of the
four primary scaling levels always applies.
- Nominal scale (e.g., gender, nationality):
o Nominal scales do not reflect the amount of an attribute
o Few analytical possibilities: frequencies and percentages
o Classification:
Labels for classes or categories
Mutually exclusive and collectively exhaustive (not overlaps, but when
integrating, it shows the whole population)
Same class is same number
No two classes with the same number
o Identification
Each number is uniquely assigned to an object
Each object has only one number assigned to it
- Ordinal scale (e.g., ranking, preference)
o Basic property is ranking
o Numbers should be interpreted relative (more than/less than)
o In practice used for assessing preferences
o More analytical possibilities than the nominal scale
- Ratio scale (e.g., money, time, weight)
o Quantity of an attribute, differences in quantities, ratios
o Fixed zero point
o Typical marketing examples: income, revenues, market share
o All statistics are possible
- Interval scale
o Degree of attribute, difference in degree
o Arbitrary zero point (as a result, no ratio) (not based on a reason)
o Typically used in marketing research to measure attitudes, feelings, intentions,…
o Nearly all statistical tests are possible
Thus, if you are asked to determine which level of measurement you are dealing with, start with
nominal and see whether the property of the next level is present or not!
- The measurement level/primary scale level is determined by the answering format. Not the
question itself.
2.2.Different scale types
Constant sum
- Respondents are asked to allocate a constant sum of points (usually 100) among a set of objects or
attributes according to some criterion.
o Unimportant objects/attributes receive value 0
o Equally important objects/attributes receive equal value
o Twice as important = double amount of points
o Restricted to number of objects/attributes
o Use with restricted number of objects/attributes
- Although constant sum scaling is comparative in nature, it has some characteristics of metric scaling.
o Frequencies, percentages, and averages
o Rank order
o Distances
- Constant sum should be considered an ordinal scale because of its comparative nature and lack of
generalizability.
Itemized rating scale: Scale consists of various categories marked with a number and/or a brief
description
- Likert scale
o Respondent expresses (dis)agreement with a series of statements
o Originally 5 points (7,9,11 points also possible)
o Ease of use is an advantage
o Effort asked from respondent is a relative disadvantage
High points (9,10): to have more precise, more details; disadvantage: time more time to think
and give answers (take more energy, effort to answer)
o Analysis per statement
o Summed scores
o If 5 or more scale points are used, the data is considered interval (metric)
o This opens up many possibilities for statistical analysis
- Semantic differential: Itemized rating scale in which the anchors (extreme values) are associated
with bipolar labels that have semantic meaning.
o Characteristics rather similar to Likert scale
o Originally 7 points (5,9, 11 points also possible)
o Difficult when the respondent cannot see the scale (e.g., phone)
o Not always possible to come up with adjectives
o Mostly used for promotional studies and NPD perceptions
o Analysis per statement
o Summed scores
o If 5 or more scale points are used, the data is considered interval (metric)
o This opens up many possibilities for statistical analysis
- NUMBER OF CATEGORIES
o Trade-off between discrimination and required effort
o General guideline: between 5 and 9 categories
o Factors that influence the decision on number of categories:
Involvement and knowledge about object
Nature of object (need fine discrimination or not)
Mode of data collection (e.g., telephone interviews restrict number of categories)
Type of statistical analysis
o Although it is possible to use an even number of categories this in typically not
recommended
- (UN)BALANCED:
o (Un)balanced scale: (un)equal number of favourable and unfavorable scale categories
o In general, balanced scales are used
- (NON)FORCED
o Forced scale: respondent is forced to express his opinion; there is no “no opinion”/”no
knowledge” alternative provided
o Nonforced scale does include such a “no opinion” alternative
o If it is to be expected that a significant proportion of the sample will have no opinion or is
unwilling to give their opinion, a non-forced scale will yield better data
o When using a nonforced scale be careful when analyzing the data
- DESCRIPTIONS
o Many possibilities exist for describing the scale points
o Typically we use numbers and a description for the anchors (and midpoint)
o Relative to the all-verbal descriptions this does not influence data accuracy
o Description anchors does influence data distribution
o Use strong anchors like “Strongly (dis)agree” or “Very (un)likely”
3. Measurement
- Measurement is about asking the right questions
Reflective Formative
High correlations among indicators No requirements regarding the correlations among indicators
- Reliability and validity are important for all measurements (both objective and subjective properties)
- The specific types of reliability and validity differ per type of measurement instrument
Reliability
- Extent to which a scale produces consistent results if repeated measurements are made (you
compare multiple measurement on a single object)
- It is the degree to which the measurement results are free of random error (XR)
- A measure is perfectly reliable when XR=0
- Systematic error (XS) does not have an impact on the reliability
- XO = Observed score; XT = True score; XS = Systematic error; XR = Random error
- Reliability is assessed by determining the (co)variation present in the different measurements
occasions
- Thus, data needs to be collected and analyzed to draw conclusions about a measurement
instrument’s reliability
- Different types of reliability
o Test-retest reliability
Each respondent is measured at two
different times under as equivalent
conditions possible
the degree of correlation between the two measurements reflects the reliability of the
measurement instrument: Higher correlation = higher reliability
DRAWBACKS:
Sensitive to time interval
Not always possible
Carry-over effects (where the evaluation of a particular scaled item significantly
affects the participant’s judgement of subsequent scaled items)
Influence of external sources
o Alternative forms reliability
Two equivalent measurement instruments (content wise) of a certain characteristic are
designed
Every respondent is measured by the two instruments
The correlation between the respondent scores obtained by the two instruments is
indicative of the degree of reliability
DRAWBACKS:
Difficult and/or time consuming to construct two equivalent instruments
Low correlation: low reliability or unequivalent instruments?
o Internal consistency / reliability
Only applicable for reflective multi-item measurement instruments
Analysis per construct
As the indicators all reflect the same underlying construct, they should produce
consistent results
IC reliability assesses the extent to which the various indicators lead to consistent results
Using the collected data, two IC reliability measures can be computed (split-half and
coefficient alpha)
split-half reliability
item scores are split into two groups (e.g., 1 construct is measured by 4 items
split into 2 groups, 2 items/ group)
For each group of items the summed score is calculated
The correlation between the two summed scores is the split half reliability
coefficient for internal consistency
DRAWBACKS: Correlation coefficient is influenced by the way the items are split into
groups
Even numbered items vs odd numbered items
First x items vs last x items
Much more possibilities (to split into groups)
Coefficient alpha
Average of all possible split-half correlation coefficients
Thus, only for multiple item measurements
Cut-off value 0.60 (possible range 0-1)
Done with statistical software
Also often referred to as Cronbach’s alpha
Wrap up:
oAll sampling starts with a definition of the target population: I am going to develop insight in
the attitude, behavior of... (Belgian citizens; girls between the ages of 10-15; male adults; Ford
drivers; etc.)
o Then you need to either describe the population, or list all of its members, which helps you to
determine the ‘sampling frame’.
o Then you choose how you will select your sample from the population.
o You need to determine how many elements your sample needs to contain (sample size).
o Then you collect your data from the sample.
2.1.Defining target population
- A good definition of the target population should contain information about
o The sampling elements (person)
o The sampling units (individual, couple, group)
o The area of coverage (extent, geographical, period, etc.)
o Time (when)
2.2.Sampling frame
- Representation of the elements of the target population
o List
o Set of directions (description)
- Sampling frame error: when the sampling frame contains more or less of a specific type of
individual than its proportion in the target population (more males, more females, more elderly
people, more Ford drivers etc.)
o Ignore We can choose to ignore the problem (if we think it is not affecting the results)
o Screen We can also screen the sample (post hoc determination of the sample composition:
report frequencies)
o Correct Sometimes we can also correct it (delete the disproportionate parts – too large or too
small - so that we end up with a perfectly representative sample)
The sampling frame allows you to select your sample.
This can be a listing of individual elements representative of the target population, or a description.
2.3.Sampling techniques
- Sampling strategies can be divided into two major techniques: probability and non-probability
sampling.
o Non-probability sampling means that the probability of an element to be included in the sample
is not the same for each element (e.g., I only use a list of students from Hasselt to sample
university students).
o Probability sampling means that the probability of an element to be included in the sample is
the same for each element (I use a list of all students to sample from).
- Probability sampling is generally considered to lead to superior results, but non-probability
sampling is more common (for reasons of efficiency):
o Convenience sampling: I select from elements that are in my neighborhood, that I can reach
easily.
o Judgmental sampling: I pick the elements that I judge most appropriate.
o Quota: I only use a limited number (or percentage), and stop when I have reached that
number (2 stages: the first stage consists of developing control categories or quotas of
population elements, the second stage selects sample elements by convenience or
judgement).
o Snowball sampling: I ask respondents to invite their friends to respond to my survey
- PROBABILITY SAMPLING
o Objective is to select a representative, unbiased sample
o Sampling procedure wherein each element has a known, fixed, (but not necessarily equal)
probabilistic, chance of being selected for the sample.
o Researcher specifies an objective procedure; selection of elements is independent of
researcher
Precise definition of population
Precise definition of sampling frame
o Simple random sampling: I ‘pick without watching’. As you can see, this may lead to a
misrepresentation of one element
o COMPARING PROBABILITY SAMPLING TECHNIQUES
Probability sampling techniques vary in terms of sampling efficiency (tradeoff between
precision and costs).
Optimize sampling efficiency; maximize level of precision subject to the budget
constraints
Efficiency of probability sampling techniques is assessed by comparing it with that of
Simple Random Sampling (SRS)
o Systematic Sampling is not random sampling, but sampling according to a specific rule (I select two
columns, and all members of a column).
o Stratified sampling is an attempt to make the sample random and representative. Make sure that
all categories are included in the ’right proportion’ (but members from each category have the
same chance to be included).
This happens when we really want to make reliable predictions about a complex
population.
Increasing efficiency by increasing precision
Two stages
Mutually exclusive and collectively exhaustive strata
From each stratum, respondents by SRS
The number of stratum should not larger than 6 (increasing precision with increased cost)
Ensuring homogeneity between elements in one strata, heterogeneity between elements
of different stratum, stratification variables should be close to characteristics of interest,
decreasing cost with easiness in measure
o Cluster sampling is probabilistic, but more efficient than random sampling.
Increasing efficiency by decreasing cost
Stages:
Mutually exclusive and collectively exhaustive clusters (each cluster contains diversity
of participants in the target population)
Randomly select clusters
From selected clusters, sample all elements (one-stage cluster sampling) or randomly
sample subset of elements (two-stage cluster)
2.4.Sampling size
- SAMPLING SIZE DETERMINATION-THEORETICALLY
We can theoretically determine the necessary size of a sample by the following formula.
However, in practice this doesn’t work, because we generally do not know the variance in the
population (we first need to establish it in a random sample) and it works only in the case of
Simple Random Sampling.
- Qualitative factors should be taken into account when deciding sample size:
o The importance of decision
o The nature of research
o The number of variables
o The nature of analysis
o Sample sizes of similar researches
o Incidence rate (occurrences of behavior or characteristics in population)
o Completion rate
o Resource constraints.
- SAMPLING SIZE DETERMINATION-PRACTICALLY
o The formula looks very handy, but
Only applies to SRS
Population parameter (σ) not readily available.
Conduct a pilot study
Use secondary data (e.g., previous studies on the topic)
Judgment of an expert
Are based on normal distributions
o Often rules of thumb or experience is used to determine sample size.
for regression: 20 observations per variable
for an ANOVA or t-test: 20 observations per group
o for the same reliability as the simple random sampling, sample size are the same for
systematic sampling, smaller for the stratified sampling, and larger for cluster sampling.
2.5.handling non-response bias
- The more the people-who-do-not-respond differ from the people-who-respond the bigger the
problem
o Non response bias (the fact that not every invited respondent fills out your questionnaire) can
be a problem. Especially when specific groups of respondents refuse to answer. For example
‘highly frustrated customers’ refuse to fill in a satisfaction survey: you get a very positive bias.
o Or when lazy respondents do not respond... etc. So, in general we do our best to make sure
that as many respondents from as many as possible categories reply.
- What you can do to increase response rates:
o Incentives
o Reminder
o Designing a motivating research instrument
- Assessing the extent of the non-response problem
o Compare profile of respondents and non-respondents
o Contact the non-respondents in a different way
o Use the people who responded late as a proxy for the non-respondents
Often we try to compare the respondents to non-respondents. How: well, we know that late
respondents resemble non-respondents, so we compare the last 10% of respondents with the
first 10% (t-test, compare the means: if you find significant difference, there is a problem...).
3. Hypothesis testing
- The idea behind hypothesis testing is simple. You have an idea about a population and you test
whether the data you collected from a sample of that population statistically supports that idea.
3.1.Formulate hypothesis
- RESEARCH QUESTION -> Content-based hypotheses -> Statistical hypotheses
hypothesis is the answer to research question look back at your research questions, and
provide the ‘answer’ that is most logical, and for which you have good arguments.
For example: RQ: What is the effect of quality on satisfaction. Hypothesis: There is a positive
relationship between quality and satisfaction Then translate it into a statistically testable
formulation: there is a positive correlation between quality and satisfaction.
WHAT IS YOUR HYPOTHESES?
These are the distributions of test statistics we use a lot: T-test (normally distributed), Chi-
square test, F-test.
3.2.Select appropriate statistical test
Choose the right STATISTISTICAL TEST
- The test you choose depends on : the purpose you use it for:
o differences in value (t-test, ANOVA, etc.)
o differences in frequency, fit (Chi-square)
o differences in effect size (regression), etc.)
Keep in mind that your data need to fit the test-statistic (or the other way round, of course)
- Regardless of the statistical test used, the remainder of the hypothesis testing
procedure stays the same.
- The value for the test statistic is based on your sample data
- This value will be compared to some theoretical standard (distribution)
3.3.Choose level of significance
- Usually, we use p = 0.05
- It represents the probability of committing a Type I error
- Type I error: reject the Ho when it is actually true
- Confidence level is 1 - p ; the probability that you fail to reject the H0 when you indeed should not
reject it (i.e. H0 is true)
- Thus, the significance level represents the degree of risk that you are willing to take in rejecting H0
when it is actually true
Choose a level of significance (before you do any test!!!). This represents the probability of
making an error that you will accept. It depends on how certain you need to be about the
acceptance or rejection of your hypothesis.
- BECAUSE THERE IS ALSO THE PROBABILITY OF MAKING A TYPE II ERROR
Sometimes it is worse to accept an untrue hypothesis than to reject a true one. In that case we
need to decide conscientiously about the acceptable error levels
We do not want to send an innocent person into jail!
We do not want to set a guilty person free.
Although beta is unknown, it is related to alpha. An extremely low value of alpha (e.g. 0.001)
will result in intolerably high beta errors So it is necessary to balance the two types of
errors. As a compromise, a is often set at 0.05; sometimes it is 0.01
3.4.Calculate statistic and its p-value
- Hypothesis testing involves comparing our data to some known or given standard.
- This standard is typically a so-called distribution (t-distribution, F-distribution etc.)
- To achieve this, the data needs to be transformed into a value that is related to this distribution
(data analysis)
o Conducting a t-test -> gives you a sample t-value
o Conducting an F-test -> gives you a sample F-value
o And many more possibilities
- SPSS does all the calculations, but it is not ‘idiotproof’!
o So, based on the appropriate test I need to compare my sample value of the test statistic to
the known distribution
And it will tell how ‘much’ your sample value of the test statistic differs from the test statistic.
Wrap up
- Sampling
o From whom do you get the needed information
o Sampling is important, as a bad sample yields bad population estimates (especially in the case of
quantitative research)
- Hypothesis testing
o Putting your thoughts to the test – what do the data say about them?
o Hypothesis testing is done using statistical tests
o Regardless of the statistical test used, the procedure stays the same: comparing the p-value of
the sample statistic to the alpha level
- Nominal: mode
- Ordinal: mode, median
- Interval, ratio: mode, median, mean.
Which tests we can conduct depends on the type of data, and the samples (groups) we want to
compare.
If the data are non-metric, the number of tests is very limited.
When they are metric, it depends on the number of groups we want to compare
- NATURE OF THE TEST VARIABLE: The test variable is the variable we investigate
o Variable on which you want to draw conclusions
o Variable on which you compare (the different) sample(s)
o When your test variable is:
Metric (interval and ratio) -> parametric tests
Nonmetric (nominal and ordinal) -> nonparametric tests
- ASSUMPTIONS PARAMETRIC TESTS:
Before we can use the parametric tests we need to check a few things:
5.1.One-sample t test:
- Compare whether the sample mean is different from some test value.
-
- μ is a predetermined number , it is the value to which you want compare your sample
mean μ0.
- ANALYZE > COMPARE MEANS > ONE SAMPLE T-TEST
o TEST VARIABLE: the variable you wish to investigate
o TEST Value: the value you wish to compare with
5.2.Independent samples t-test
- Compare whether two independent groups have a different mean score on a test
variable
- Samples are independent when the scores on the test variable in one sample are not
influenced by (“independent of”) the scores on the test variable in another sample
-
- μ1 is the mean of group 1 and μ2 is the mean of group 2
- ANALYZE > COMPARE MEANS > INDEPENDENT SAMPLES T-TEST
o Enter the test variable (the variable of which you want to compare the means
between the groups (exam scores, IQ, etc.).
o Define the groups with a new variable: E.g., GENDER (0,1) group variables
- SELECTING THE APPROPRIATE IND. SAMPLES T-TEST
o How the appropriate independent samples t-test is calculated depends on
whether or not the variance in the two samples is equal.
o Different formulas apply
o Again, SPSS computes both and leaves the decision which t-value is the
appropriate one to you
o To choose the appropriate t-value use Levene’s test
o Hypotheses Levene’s test (F-test)
-
o Indicates the differences between two related sample means. Other forms are
possible (e.g., μ0 – μ1 = 0).
- ANALYZE > COMPARE MEANS > PAIRED SAMPLES T-TEST
o Enter the two test variables (there are always two!) that you wish to compare
(e.g., wine A score, wine B score)
5.4.One-way ANOVA
- Compare whether the sample means differ between two or more independent samples
o Dependent variable: metric variable
o Independent variable: must be categorial variable (non-metric), if including both
categorial and metric variables ANCOVA (analysis of covariance)
o If all independent variables are metric regression.
- In case of:
o two unrelated samples it does not matter whether you use an independent
samples t-test or a one-way ANOVA
o More than two unrelated samples you must use a one-way ANOVA (a series of
independent t-tests may yield erroneous results) can compare many
different samples
-
o This is the only appropriate form of hypotheses for One-way ANOVA
- ANALYZE > COMPARE MEANS > ONE-WAY ANOVA
o Enter the test variable (s) (you can do ANOVA between the same groups on
several variables: preference, attitude, satisfaction, etc.)
o Define and Enter the grouping variable (e.g., Nationality (0,1,2,3)
o Tick the post-hoc tests in the options box (always tick all of them!)
- POST HOC COMPARISON TESTS
o If the null hypothesis is rejected, we conclude that at least one group is different.
o But which group? Post hoc comparisons are used to determine which groups differ
from each other.
o They have the following format:
o If you open the “post hoc” dialogue box in the ANOVA menu you see that there
are a lot of different post hoc tests
Which one to pick?
Homogeneity of variance test
H0: all groups have equal variances
H1: groups do not have equal variances (at least one differs)
Equal variances: Tukey
Unequal variances: Dunnett’s C
To avoid double work always click the homogeneity of variance test (“options”) and both post
hoc tests
6. Translation into statistical hypotheses
-
- Always trans late your hypothesis (the answer to your research question) into a
statistical hypothesis (one that can be tested!)
7. Wrap up
- Data analysis
o helps you to make sense out of your data
o formulates answers to your research questions
- Data analysis techniques
o Can be viewed as necessary evil, but are crucial
o And will continue to gain importance
o Are numerous and the trick is to know
What techniques are available
When to use what technique
-
- in the domain of ‘dependence’ and ‘interdependence’ between variables, there are two
techniques:
o Correlation is bi-directional, and therefore called an ‘interdependence’
technique (Variable A ‘varies along’ with variable B and variable B ‘varies along’
with variable A)
o Regression is directional (For example Variable A (independent variable, e.g.
Satisfaction) has a positive effect on Variable B (dependent variable, e.g.,
Loyalty), but not the other way around)
2. Pearson product moment correlation
- ASSUMPTIONS / WHEN TO USE
o Two metric variables If you do not have metric variables, you may need
different techniques (Chi-square, etc.)
o Normally distributed variables
SPSS offers options for non-normally distributed variables, but Pearson
Correlation is pretty ‘robust’ (even if your variables are not ‘perfectly
normally distributed’ you will still get a reasonable estimate)
Per definition a Pearson product moment correlation takes into account the relation between
only two variables (i.e., ignores the impact of other possible factors)
- EXAMPLES OF HYPOTHESES
o How strongly are sales related to advertising expenditures?
o Is customer satisfaction positively related to customer loyalty?
o Is attitude towards a technology positively related to usage intentions of that
technology?
- STATISTICAL HYPOTHESES
o H0: Variables X and Y are not related (r=0)
o H1: Variables X and Y are related (r≠ 0)
Hypotheses that can be tested are the hypotheses that do not specify a direction in the effects.
- CHARACTERISTICS
o The Pearson product moment correlation (r) summarizes the strength of the
linear relationship between two variables.
The correlation coefficient is ‘r’.
o r ranges from –1 to +1
It varies between 1 (full positive linear relationship) and -1 (full,
negative linear relationship).
o r is a symmetric measure of association. This means that rxy = ryx
o r = 0 means that is no linear relationship. It is does not mean that X and Y are
not related. There might well be a nonlinear relation, which is not captured by
r. To examine this use scatterplots.
Scatterplots (one variable on X-axis and the other on the Y-axis: plotting
the relationships (pairs, (x,y)) are very useful to explore the nature of
the relationship between variables
o A PEARSON PRODUCT MOMENT CORRELATION OF 0? (r=0)
- Analyze>Correlate>Bivariate
o Pearson is selected by default
o Select the relevant variables
- SPSS OUTPUT CORRELATION ANALYSIS
oProvided are the correlation coefficient, the accompanying p-value, and the
sample size
3. REGRESSION ANALYSIS
3.1.Bivariate regression
- GENERAL BIVARIATE REGRESSION MODEL
- Regression models have the same formula as a straight line. In this case the formula expresses a
linear relationship between two variables: the dependent (explained) variable and the explaining
variable (independent variable).
- The software will try to find a straight line through your observations.
- The straight line may not go through the origin (0,0), but it may hit the Y-axis at a positive or
negative value (this is called the Intercept).
- The regression coefficient says something about the strength of the effect of the independent
variable on the dependent value. It is a value between 0-1. It is visible in the slope of the line: Beta =
0 means (flat, horizontal line), Beta = 1 means 45° line.
Same for a multidimensional (multiple)regression. Only thing is, that this represents a line in a
multidimensional space; As many dimensions as variables. Coordinates look like vectors!
Calculations in SPSS take the form of matrix calculations (matrix algebra).
- SOME MAIN ASSUMPTIONS
o Dependent variable is metric
o Independent variable(s) are metric or dummy-coded
o Model is correctly specified: form and variable(s)
o No exact linear relationship between Xi’s (only for multiple regression)
- The discussion on least squares estimation and model evaluation is the same for bivariate and
multivariate regression analysis! For the sake of clarity the least squares procedure is explained in a
bivariate context.
3.3.Regression analysis
- MY DATA POINTS (COME DIRECTLY FROM QUESTIONNAIRE)
Look at the dots: every dot represents a data pair (x, y), where y is the value of the dependent
variable for a specific x value of the independent variable: every dot is one observation. For
example ‘my dot’ is: (7,6), if I have answered 7 on the satisfaction measurement and 6 on the
loyalty measurement.
- GRAPHICAL PRESENTATION OF REGRESSION PARAMETERS
o Error = distance between the observation (dot) and the optimal line.
o You can see that the algorithm tries to globally minimize the error (difference
between the line and the dots)
- THREE SOURCES OF VARIATION:
This shows in the ANOVA output of your regression in SPSS. (error = residual).
The Anova says something about the ‘quality of your model’: is the straight line really a good
representation of the underlying pattern in the data.
3.4.Interpretation (2 stages)
-
Only if the overall model performance (stage 1) is good (i.e., significant) we proceed with stage 2.
The ANOVA represents the quality of your model (does your model ‘fit’ the data structure?) NB the
fit does need to be good (your Anova needs to be significant!) to be able to say something
meaningful about the relationships between the variables!
The regression coefficients say something about the effects of each variable on the dependent
variable.
3.4.1. STAGE 1: OVERALL MODEL PERFORMANCE
- COEFFICIENT OF DETERMINATION (R2)
o Expresses the relative amount of variance (RSS) of the dependent variable that
is explained by the independent variable(s) (how much of the variance in your
dependent variable is explained by the variance in the independent variables)
Or equivalently
The overall model fit hypothesis (stating that r-square adjusted is bigger than 0) should be
significant. Once we have established this (in the ANOVA reported in the regression output: the
F-value should be significant), we can starting to test the model hypotheses.
In some cases we would like to see if a categorical variable has an effect on a dependent
variable. To investigate this, we need to code dummy variables for the categories.
- In SPSS: TRANSFORM > RECODE INTO DIFFERENT VARIABLES
o Select the variable you want to recode into a dummy (or dummies)
o Specify the new variable (for example dummy01)
o Specify which values of the original variable (old values) will correspond with
the dummy (new values)
R-square does not equal zero and F is significant So the regression is valid.
3.7.Multigroup analysis
- We assumed that the relationship between the dependent and independent variables
was equal for the entire sample
- BUT WHAT IF……..
We are interested in comparing this relationship across several groups?
To address this question we need to do a group comparison (or a multigroup analysis)
compare Betas between groups
- Comparing regression equation: 4 possibilities
o Option 1: same Beta, but different constant
o Option 2: same constant, different Betas
o Option 3: Betas are different, and Constants are different
o Option 4: there is no difference between the groups.
- TEST SEQUENCE
o
o ESS = Error sum of squares of the different models (pooled and per group)
o K = number of parameters in underlying model
o N = sample sizes of the different groups
Can easily be extended to accommodate more than two groups
You need to do the regressions for each group and the ‘pooled’ groups (full sample)
3.7.2. Equality of intercepts
- In a regression context this is a fairly complicated test to perform
- As a proxy perform an independent samples mean-differences test (independent
sample t-test) on the outcome variable.
- Results independent samples t-test (groupvar = gender; testvariable = irritation ad)
t(67) = 10,385 (p < 0,001)
- We conclude that the intercepts are different
3.7.3. Equality of slopes (“Paralellism”)
- To be sure that Betas are the same (between groups), we need to examine this by
means of a t-test.
- MULTIGROUP HYPOTHESES-EQUALITY OF SLOPES
o Estimate the regression equation separately for each group (you did this already
for the Chow test)
o Subsequently, apply the test below to assess whether the coefficients are
significantly different
- If you need to do analysis for a part of your sample, you can use this command in SPSS to ‘split’ the
file Data > Split File
o Do not forget to turn the split file procedure off when you are finished with your analysis
WRAP UP:
- Suppose we have following model:
-
- must do the multivariate regression, and NOT two correlations
- WHY? ONLY MULTIPLE REGRESSION IS VALID!
o Takes into account relationships among the different IV’s
o As such each coefficient reflect the unique contribution of each independent
variable on Y.
THUS: If your model tells you that there are several variables “causing” an outcome (i.e.,
dependent variable); estimate a model that analyzes these multiple variables simultaneously.
- BIVARIATE REGRESSION AND CORRELATION COEFFICIENTS
o In case of a single dependent and independent variable, the Pearson product moment
correlation coefficient (r) equals the standardized regression coefficient (Beta).
o Even though correlation coefficients and standardized bivariate regression analysis provide
you with the same information, bivariate regression allows you to make predictions.
OBSERVATION
1. Introduction
- We are going to discuss the stages of research design (i.e., the answer to the question:
How should we set up the experiment?) and the collection of the data.
- Data analysis techniques that are used in most experiments have been discussed before:
for most experiments, we use t-tests, ANOVA, or MANOVA.
- In many so-called field experiments, we simply observe. The result of the analysis of the
data is a description of behavior (observation descriptive design).
2. Observation
2.1.What is observation?
o Recording and quantification of behavioral patterns
o In a systematic way
o To obtain information about a phenomenon
o Without direct communication with subjects
All experiments rely on observation. In so-called field experiments, researchers try to observe
the phenomena of interest as ‘flies on the wall’: The observed subjects do not notice that they
are being observed, so as to be able to observe their ‘natural behavior’.
2.2.Observation methods
There are several ways to collect and analyze observational data:
- Personal observation: the researcher him or herself observes the object of interest. This can be
taken an active form (mystery shopper), or a more passive form.
o Mystery shopping
Monitoring the competition
Assessing employees` performance
o Shopping pattern studies
Way through aisles
Time spent at places
o Shopper behavior studies
Effect of shelf height, positioning of products
Reading of labels
- MECHANICAL (DIGITAL) OBSERVATION: high tech devices (Smart watches, Mobile phones, smart
cameras, microphones, etc.) are used, instead of personal observation, to ‘mechanically’ observe
the phenomena of interest
o Eye-tracking monitors
Reading of ads
Viewing of commercials
o Skin response
Evaluation of commercials
Package design
o Voice pitch analysis
Analysis of emotional responses
Product preferences
- Audit: use the digital traces that customers leave behind as data
o Pantry audit: Pantry audit is possible, because many new fridges are equipped
with detectors and the products with RFID tags
o Loyalty cards
o Commercial scanning data
- Content analysis: is a way to try and understand the ‘meaning’ of the observed phenomena
o Skin color of actors in food commercials
o Use of words (and frequency) in ads
o Role of women in commercials
o Trending topics on twitter
- Trace analysis: is the explicit use of traces left by the consumer during their customer journey.
Digital and visual data can be combined to get a more complete picture.
o Finger prints
o Parking lots
o Internet cookies
o Location-based Services
2.3.Necessary conditions
If we simply wish to ‘describe’ or explore behavior we can use either ‘observation’ or surveying
(using a questionnaire)
Surveying is used when it is difficult to get access to the ‘situation’ (place-wise, time-wise,
situation-wise) we want to observe.
Surveying is cheaper…
We have several ways to design experiments in such a way that the effect of extraneous variables is
neutralized, or controlled for
o Randomization
o Matching
o Statistical control
o Design control
4. EXPERIMENTAL DESIGNS
Important warnings:
Here the 2*2 design is clearly visible. It leads to a matrix with four combinations of levels: LL,
LH, HL, and HH (H = high, L=low).
- By also measuring the extraneous (confounding) variables that could also have
influenced the dependent variable, we can ‘control’ for them
- FACTORIAL DESIGNS: MAIN EFFECTS
o Main effect (per factor!) is the effect of the factor on the outcome while
ignoring the effect of the other experimental factors
o For every experimental factor there is a main effect
- FACTORIAL DESIGNS: INTERACTION EFFECTS
o An interaction effect is present when the effect of a factor is not constant over
all levels of the other factor
o An interaction effect means that the main effect of a factor varies with the
different levels of the other factor (can be attenuated or reinforced).
o In contrast, if there is no interaction between the factors that means that the
difference in cell means caused by a factor is constant over all levels of the other
factor
The presence of an interaction effects means that you have to be careful when interpreting the
main effects (“depends on”)
No interaction effects:
Interaction effects:
WRAP UP
- Causal research is about establishing cause-effect relationships
- Descriptive studies offer too little control, experiments are needed
- In order to have valid results, one needs to rule out the effects of so-called extraneous
variables
- In practice, factorial designs are most used
- In general, experiments are considered a double-edged sword
- Expensive
- But it’s the only game in town available