ARM Lecture 3 - Moderation vs. Mediation
ARM Lecture 3 - Moderation vs. Mediation
Moderators are typically assessed with an interaction term; figuring out whether or not the relation
between the predictor and the outcome differ for different levels of the moderator.
For mediators, they act as a second type of third variable - the predictor predicts the mediator which
predicts the outcome.
Note that the moderator/mediator do nothing other than predicting the outcome.
This is different from a confounder; these affect both the predictor and the outcome - which is why
they're a problem. The predictor is no longer the start-point, but an outcome.
With moderators/mediators, the predictor stays as the start-point.
Moderation
A moderator (Mod) is a variable that alters the strength of the linear relationship between a
predictor (X) and an outcome (Y).
E.g. psychotherapy might reduce depression more for men than for women, so gender (Mod)
moderates the effect of psychotherapy (X) on depression (Y).
A moderator analysis is a test of external validity, in that, the question is how universal is the effect
(i.e. does the effect of X on Y differ as a function of Mod).
Moderation tests if effects are conditional.
Moderator is a test for universality; is the relationship between X and Y the same for everyone. If the
relationship is the same (i.e. if this is a universal association applying to all individuals), the
interaction effect of the moderator is non-significant. If the association does differ for different
variables of the Mod variable, it's called a conditional effect - the association between X and Y is
conditional on different levels of Mod.
Moderators can be categorical or continuous :)
Testing moderation
The primary way of testing moderation is with interactions, which are typically calculated by
multiplying the focal predictor (X) by Mod.
The exact procedure depends on how the predictor, moderator and outcome are measured.
The type of statistical analysis you can perform depends on the variables :)
For ANOVA:
If X and Mod (Z) are categorical and Y is continuous.
For MLR:
If X, Z and Y are continuous
If X/Z are categorical and Y is continuous
For logistic regression:
If X and Z are continuous and Y is categorical
If X is continuous and Z and Y are categorical
Moderation in ANOVA
When X and Z are categorical.
This is generally the easiest technique; it examines group differences.
Linear regression can also be used to address similar questions :)
SPSS automatically puts the interaction term in if you specify two IVs in your ANOVA :DD
If the interaction is significant (as indicated by p) use the (a priori) contrasts or pairwise comparisons
and the plot function to interpret the interaction via the means of the groups.
Example:
Gender as Z - is romantic involvement related to depression in the same way for male/female
adolescents
o RI and gender as IVs, depression as DV
o Interaction was significant :)
o
You can see that the female line is very diagonal (which means that the
interaction effect occurs there; that's where the difference is).
I.e. gender moderates the effect of RI on depression via females
having higher depression scores if they are romantically involved.
o You can then test group differences with simple effects (pairwise comparisons)
Which does require changing the syntax!
Linear regression
When X, Z and Y are all continuous.
Requires more work :) SPSS doesn't automatically put the interaction in :(
Follow a few steps:
1. Centre X and Z (i.e. take individual score minus the mean/average score)
a. Standardising is not the same as centring!
2. Calculate the interaction using the centred scores (i.e. multiply cX * cZ)
3. Perform analysis with the centred scores of X and Z as well as the newly created interaction
effect as predictors.
4. If the interaction is significant, plot the simple slopes via Aiken & West (1991) (what)
a. It's more complicated due to the continuous nature of the X/Z
Example:
Age as Z - Is the romantic partner relationship quality (RPqual) associated with depression for
older and younger adolescents?
X = Rpqual / Y = depression / Z = age
o Step 1:
Identify the means of both predictors (for centring)
Subtract the means from each individual score
Centred scores for RPqual and age acquired :)
o Step 2:
Multiply the two centred scores to create the interaction
o Step 3: Perform the MLR
Some notes:
Adjusted R2 = adjusting for sample size/# of predictors. Always
smaller than R2 because it's compensating for stuff
ANOVA table tells you whether your basic R2 is significant or not -
that's what the p-value there is telling you. Do the predictors
collectively explain a significant amount of difference on the
outcome variable?
Regression weights :)
RP*age = significant - so follow-up analysis time!
o Step 4a: Calculate simple slopes
So you get one formula for younger age and one for older age.
do math
End up with a simple slope and simple intercept for both younger and older
children.
Older:
Simple intercept = 3.657
Simple slope = -.025
Younger:
Simple intercept = 3.411
Simple slope = -.173
Now we have a better understanding of what's going on :) This tells you
what's happening already (sort of)
The slope for older children is pretty flat; a slope of -.025 is pretty flat
(and thus not that interesting)
The slope for younger children is steeper; so you can already
estimate that the action takes place in the younger children.
o Step 4d: Calculate points to plot by solving for predictor values
We want to know what are the values that we should put into the figure. To
do so, you need high and low levels of X (i.e. RPqual). Since this is for plotting
purposes, we want a wide range of values on the x-axis. So instead of using
+/- 1 SD, we use +/- 2 SD for the x-axis of X!
Note that there are some automatic features in SPSS using 1 SD for
creating these points to plot (so that we can see how "Process" will
give us the answer as well).
To get the SD, go back to the descriptive stats table; see calculations below.
Logistic regression
When Y is categorical
The same idea, but the interpretation differs slightly.
These models instead predict the likelihood of group membership
Realise that all of this generalises and scales up to 3- and 4+-way interactions!
In summary
Moderation describes conditional effects
o We're testing whether the association is universal (no moderation) or conditional
(yes moderation).
Moderation is typically tested with an interaction between the predictor and the moderator
o For plotting and testing you can just use modules :)
Moderators are typically measured at the same time as the predictors
o Even in longitudinal data; we're interested in the association between predictor and
outcome. If the predictor is measured before the outcome, you also want the
association between moderator/outcome to be assessed at the same interval!
o This is an important distinction from mediators!
Mediation
A mediation model seeks to identify and explain the mechanism or process that underlies an
observed relationship between an IV and a DV via the inclusion of a third variable - the mediator.
So, a mediation model proposes that the IV influences the mediator variable, which in turn influences
the DV.
The mediator serves to clarify the nature of the relationship between the IV and the DV by
examining causal mechanisms (processes) that explain the IV-DV relationship.
You start with X, predicting the mediator, which predicts the outcome.
If you include the mediator, the association between the X and Y becomes 0. What a mediator is
trying to do is identifying an explanatory mechanism as to why the predictor is related to the
outcome.
You start with your total effect and break it apart into a direct effect (c') and an indirect effect (a-b).
The sum of the direct and indirect effect leads to the total effect.
One of the issues here is that we have to take coefficients from different models. There's a solution
for this! Structural equation models :)
we don't do that here tho
Importantly - we're trying to come up with a more robust way of testing these mediation effects
instead of performing different regression models, doing manual calculations, etc.
PROCESS example
Testing mediation can also be easily accomplished using PROCESS.
Example:
Does social anxiety mediate the link between self-esteem and loneliness in a sample of
adolescents?
Note that self-esteem was assessed at T1, social anxiety at T2 and loneliness at T3 (all 1 year
apart).
o boom process time
Lots of output. Look at the outcome variables to see which path you're investigating!
Outcome social = a path (esteem)
Outcome loneline = c' (esteem) and b path (social)
All paths are significant here yibbee
There's also the c path (last image; doesn't include mediator) which needs to be (and
is) significant as well.
In summary
Mediation describes indirect effects and is used to examine "causal"/explanatory mechanisms.
Mediation is tested with either a Sobel test or preferably with bootstrapping methods
bootstrapping better lole
Mediators should be measured after the predictor and before the outcome.
This is an important distinction with moderators! You need 3 time points to be able to assure
the reader you have the directional associations between the mediator and the outcome.