Reviewer For Psych Stats
Reviewer For Psych Stats
Use parametric if your data follows a normal distribution and meets the assumptions.
Use non-parametric if your data is skewed, small, or doesn't meet assumptions.
BROAD CATEGORIES OF DATA
NOMINAL (Non-parametric) - Data consists of categories or labels without any meaningful
order. No numerical value or ranking. Used for classification.
Examples:
Gender (Male, Female, Non-binary)
Blood Type (A, B, AB, O)
ORDINAL (non-parametric) - Data is categorized with a meaningful order, but the
differences between categories are not uniform or measurable. Ranking exists, but the
gaps between ranks are not equal.
Examples:
Survey Responses (Strongly Disagree → Disagree → Neutral → Agree → Strongly
Agree)
Socioeconomic Status (Low, Middle, High)
INTERVAL (Parametric) - (Numerical - No True Zero) Data has a meaningful order with
equal intervals between values. No true zero (zero does not mean "none" or "absence").
Addition & subtraction make sense, but multiplication & division do not.
Examples:
Temperature in Celsius or Fahrenheit (0°C does not mean "no temperature")
IQ Scores (An IQ of 140 is higher than 120, but 140 is not "twice as intelligent" as 70)
RATIO (Parametric) - Data has a meaningful order with equal intervals and has a true zero
(zero means "none" or "absence"). Supports all arithmetic operations (addition, subtraction,
multiplication, division). Most precise level of measurement.
Examples:
Age (A 40-year-old is twice as old as a 20-year-old)
Height & Weight (A person weighing 80kg is twice as heavy as a person weighing 40kg)
In summary: The mean is the most commonly used measure, but the median is better when
dealing with skewed data or outliers, and the mode is useful for categorical data.
Summary:
Range = Simple but affected by outliers.
IQR = Best for skewed data because it ignores outliers.
Variance = Good for precise calculations but harder to interpret.
Standard Deviation = Best for normal distributions, easy to interpret.
MEASURES OF SYMMETRY
MEASURES OF SYMMETRY – Measures of symmetry describe whether a dataset is
symmetrically distributed or skewed. Symmetry means that the left and right sides of a
distribution (around the center) look the same. If the data is not symmetric, it is called
skewed.
SKEWNESS - Measures the asymmetry of data distribution. A skewness of 0 means the data
is perfectly symmetrical (normal distribution). Positive or negative values indicate
skewness.
Interpretation:
Sk = 0 → Symmetric Distribution
Sk > 0 → Positively Skewed (Right Skewed)
o Tail is longer on the right
KURTOSIS - Measures the "tailedness" of the distribution (how extreme the outliers are).
Interpretation:
K = 3 → Mesokurtic (Normal Distribution)
o Moderate tails, same as a normal distribution
SKEWED DISTRIBUTION - In skewed distributions, the mean, median, and mode are
usually in different positions, and their order can be predicted based on the direction of the
skew.
Why is Symmetry Important?
Helps choose the right measure of central tendency:
o If data is symmetric, use mean.
Affects statistical tests: Many tests (e.g., t-tests, ANOVA) assume normality, so
skewness and kurtosis help determine if normality assumptions are met.
Summary:
Skewness measures how much a distribution is shifted left or right.
Kurtosis measures how peaked or flat the distribution is.
Normal distributions have Sk = 0 and K = 3.
Skewed distributions impact data interpretation and statistical analysis.
Null Hypothesis (H₀): There is no significant difference between the means of the two
groups.
Alternative Hypothesis (H₁): There is a significant difference between the means of the
two groups.
CHECK ASSUMPTIONS
The data should be independent (from different participants).
The data should be normally distributed in each group.
The variances of both groups should be equal (use Levene’s Test to check).
CALCULATE THE T-STATISTIC
Use the formula to compute t.
Compare the computed t-value with the critical t-value from the t-table.
MAKE A DECISION
Repeated measures design (e.g., testing the same group under two
conditions).
OTHER ASSUMPTIONS IN DEPENDENT SAMPLES:
NORMALITY: The differences between paired observations should be approximately
normally distributed.
INDEPENDENCE: The pairs of observations should be independent of each other. This
means that the data points within each pair are related, but the pairs themselves are
independent of each other.
SCALE OF MEASUREMENT: The dependent variable should be measured on a
continuous scale (interval or ratio).
RANDOM SAMPLING: The sample should be randomly selected from the population to
generalize the results properly.
EXAMPLES:
Effect of Exercise: Testing participants' weight before and after a 3-month
fitness program.
🏫 Teaching Methods: Comparing students' test scores before and after
implementing a new teaching strategy.
💊 Medical Study: Measuring blood pressure before and after taking a new
drug.
STEPS TO CONDUCT AN DEPENDENT T-TEST
STATE THE HYPOTHESES
Null Hypothesis (H₀): There is no significant difference between the means of the two
groups.
Alternative Hypothesis (H₁): There is a significant difference between the means of the
two groups.
CHECK ASSUMPTIONS
The data should be independent (from different participants).
The data should be normally distributed in each group.
The variances of both groups should be equal (use Levene’s Test to check).
CALCULATE THE T-STATISTIC
Use the formula to compute t.
Compare the computed t-value with the critical t-value from the t-table.
MAKE A DECISION
If the data in a dependent (paired) samples t-test are not normally distributed, the
best alternative is the Wilcoxon Signed-Rank Test.
Why?
The paired t-test assumes normality of the differences between paired
observations.
The Wilcoxon Signed-Rank Test is a nonparametric test that does not
require normality and instead ranks the differences between paired values.
When to use the Wilcoxon Signed-Rank Test?
✔ When your paired differences are not normally distributed.
✔ When dealing with ordinal data or small sample sizes.
✔ When data contain outliers that could skew parametric results.
This test compares the median difference instead of the mean, making it more robust
against non-normality and outliers.
If an independent samples t-test violates the assumption of homogeneity of
variance (meaning the two groups have unequal variances), the best alternative is
Welch’s t-test.
Why?
The standard independent t-test assumes equal variances (homogeneity of
variance).
Welch’s t-test is a modified version of the independent t-test that does not
assume equal variances, making it more reliable when variances are unequal.
When to use Welch’s t-test?
✔ When Levene’s test (or another variance test) shows significant inequality of
variances.
✔ When sample sizes are unequal, making variance differences more impactful.
✔ When you still meet normality assumptions but violate homogeneity of variance.
It’s basically the best fix when variance inequality is an issue while keeping the power
of the t-test. If normality is also violated, you might need a nonparametric alternative
like the Mann-Whitney U test.
IN CONCLUSION
For Independent T test, if assumptions are violated
If the data in an independent t-test are NOT normally distributed, I will use the
Mann-Whitney U Test (Wilcoxon Rank-Sum Test).
If the data in an independent t-test FAILS the homogeneity of variance
assumption (based on Levene’s test), I will use Welch’s t-test.
If the data is BOTH not normally distributed AND fails homogeneity of variance
(based on Levene’s test), I will use the Mann-Whitney U Test (Wilcoxon Rank-
Sum Test).
For Dependent T test, if assumptions are violated
If the dependent t-test fails normality → Use Wilcoxon Signed-Rank Test.
If the dependent t-test fails homogeneity of variance (even though it’s rarely an
issue) Still use Wilcoxon Signed-Rank Test.
When your dependent variable is continuous (numerical) (e.g., test scores, height,
reaction time).
NULL HYPOTHESIS (H₀): All group means are equal. (No significant difference)
If p-value < 0.05 → Reject H₀ (At least one group mean is significantly different).
The groups should have the same shape distribution (though variances can be different).
What If Kruskal-Wallis is Significant? 🤔
If KRUSKAL-WALLIS finds a difference, you still don’t know which specific groups are different.
You need post hoc tests like:
✔ Dunn’s Test – Adjusts for multiple comparisons.
✔ Mann-Whitney U Test – For pairwise group comparisons.
If the two factors interact and influence the dependent variable together (Interaction
Effect).
WHEN TO USE TWO-WAY ANOVA?
You have two independent categorical variables (factors).
You have one continuous dependent variable (e.g., test scores, height, sales).
You want to analyze both individual and combined effects of the factors.
Imagine you’re studying how teaching method and gender affect students' exam scores.
Transforms data into ranks, then runs a standard ANOVA on the transformed data.
Tests for Group Differences More Accurately – Adjusts group means by accounting for a
covariate's influence.
EXAMPLE SCENARIO:
A researcher wants to study the effect of teaching method (Lecture, Group Work, Online
Learning) on final exam scores, but some students started with different IQ levels.
INDEPENDENT VARIABLE (IV): Teaching Method (Categorical: 3 groups).
DEPENDENT VARIABLE (DV): Final Exam Scores (Continuous).
COVARIATE: IQ Score (Continuous).
ANCOVA helps remove the influence of IQ so that the true effect of teaching method is
clearer.
ASSUMPTIONS OF ANCOVA:
LINEARITY – The relationship between the covariate and DV must be linear. (Use Scatter
Plot)
HOMOGENEITY OF REGRESSION SLOPES – The effect of the covariate should be similar
across all groups. (Check the regression coefficient)
INDEPENDENCE & NORMALITY – Residuals should be normally distributed. (Shapiro-Wilk
Test, Q-Q Plot for better representation)
HOMOSCEDASTICITY – Equal variance across groups. (Levene’s Test)
When a covariate might affect the dependent variable and needs to be adjusted for.
When pre-test scores are used to control for baseline differences (e.g., in educational
research).
Interpretation of Pearson’s r:
+0.70 to +1.00: Strong positive correlation.
+0.30 to +0.69: Moderate positive correlation.
+0.01 to +0.29: Weak positive correlation.
0.00: No linear correlation.
-0.01 to -0.29: Weak negative correlation.
-0.30 to -0.69: Moderate negative correlation.
-0.70 to -1.00: Strong negative correlation.
EXAMPLE:
If you measure the height and weight of a group of people and compute Pearson’s r, a result of
+0.85 suggests that there’s a strong positive correlation—as height increases, weight tends to
increase too.
Measures the strength and direction of the monotonic relationship between two
variables.
Works with ordinal data or continuous data that isn’t normally distributed.
It’s more robust than Spearman’s rho for smaller datasets or tied ranks.
When to Use:
Similar to Spearman’s rho, but especially when you have small sample sizes or many
tied ranks.
Measures the strength and direction of the monotonic relationship between two
variables.
Works with ordinal data or continuous data that isn’t normally distributed.
It’s more robust than Spearman’s rho for smaller datasets or tied ranks.
When to Use:
Similar to Spearman’s rho, but especially when you have small sample sizes or many
tied ranks.
rXY⋅Z>0r_{XY \cdot Z} > 0rXY⋅Z>0 → Positive relationship between X and Y even after
controlling for Z.
o Then, compute the partial correlation formula but using Spearman’s rank
correlation (ρ) instead of Pearson’s correlation (r).
Interpretation: Similar to Pearson’s partial correlation, but based on ranked data.
KENDALL’S TAU PARTIAL CORRELATION (FOR SMALL SAMPLE SIZES)
Use when:
o Non-normality + small sample size (n < 20).
Why use it? More robust than Spearman when the data contains ties or outliers.
Example: Predicting house price (Y) based on square footage (X₁), number of rooms
(X₂), and location (X₃).
ASSUMPTIONS OF REGRESSION ANALYSIS
LINEARITY – The relationship between X and Y is straight. (Scatter plot, Residual Plot)
Independence of observations
IN CONCLUSION: You reject the null hypothesis in a Chi-Square test when the p-value is
smaller than the chosen significance level (usually 0.05). This indicates that the observed
frequencies are significantly different from what you would expect under the null hypothesis,
implying a relationship or association between the variables.
MCNEMAR TEST - A statistical test used to compare paired nominal data to assess if there
is a significant difference between two related groups. Evaluates changes or effects in
dichotomous (binary) outcomes, like Yes/No or Success/Failure. Utilized in fields like
psychology, medicine, and education to determine the impact of interventions, therapies, or
treatments.
OTHER DEFINITION: The McNemar Test is a non-parametric test used to analyze paired
nominal data, specifically when you have two related dichotomous variables. It tests
whether there is a significant difference in the proportion of cases between two related groups
(before and after treatment, or two different conditions).
NULL HYPOTHESIS:
There is no significant difference between the two related groups.
The changes in responses are symmetrical.
Example: The number of participants who improve is equal to those who
worsen.
ALTERNATIVE HYPOTHESIS:
If p < 0.05, there is a significant difference between the two related
groups.
The changes in responses are asymmetric.
Example: More participants improve than worsen after treatment.
ASSUMPTIONS OF MCNEMAR:
Paired data (before/after, treatment/control, etc.).
Binary data (two categories).
Symmetry of categories.
Sufficient sample size for each outcome category (at least 5).
Independence of pairs.
The test detects binary changes between two conditions.
When to Use McNemar's Test?
1. Paired Data: You have paired observations (same subjects tested twice or matched
groups).
o Example: A pre-test/post-test design (e.g., testing opinions on a product before
and after using it).
2. Two Dichotomous Variables: You are working with two categorical variables that have
two categories each (Yes/No, Success/Failure, etc.).
SUMMARY:
McNemar's Test is used to analyze paired nominal data.
It compares the proportion of changes between two conditions (before and after).
It is commonly used in pre-test/post-test designs.
The test statistic is based on the difference between the "failure to success" and
"success to failure" pairs.
If p-value < 0.05, you reject the null hypothesis, indicating a significant difference
between the two conditions.
RISK RATIO (RR), also called RELATIVE RISK (RR), it is non-parametric where it measure
used in epidemiology and medical research to compare the probability (risk) of an event
occurring in two different groups.
It is commonly used in cohort studies to determine if exposure to a certain factor (e.g.,
smoking, medication, lifestyle) increases or decreases the risk of an outcome (e.g., disease,
recovery, death).
ASSUMPTIONS:
Proper Study Design (Cohort or Prospective Study)
Dichotomous (Binary) Outcome
Independence of Observations
Adequate Sample Size
No Confounding Variables (or Controlled for Confounders)
INTERPRETING RISK RATIO (RR)
RR = 1 → No difference in risk (exposure has no effect).
RR > 1 → Exposure increases risk (e.g., smoking increases the risk of lung cancer).
RR < 1 → Exposure decreases risk (e.g., a vaccine reduces the risk of infection).
SUMMARY:
Risk Ratio (RR) compares the risk of an event occurring in an exposed group vs. an
unexposed group.
Used in cohort studies to measure relative risk.
RR = 1 → No effect, RR > 1 → Increased risk, RR < 1 → Reduced risk.
RR is different from Odds Ratio (OR) and is preferred for prospective studies.
Use McDonald’s Omega (ω) instead, which is a better alternative when tau-
equivalence is violated.
SUMMARY
Cronbach’s Alpha measures internal consistency (reliability).
Higher α = More reliable scale, but extremely high values (≥ 0.95) may indicate
redundancy.
TRANSFORMING SCORES - Score transformations are used to convert raw scores into
transformed scores that can be compared and interpreted. The purpose of transforming scores
is to give meaning to the scores and to allow for direct comparison between scores.
When you want to reduce the number of variables while keeping meaningful
information.
When you need to check if survey items measure the same construct (e.g., personality
traits, psychological scales).
SUMMARY
Partial Mediation – When X still affects Y, but M significantly reduces this effect.
No Mediation – If M does not significantly influence Y or if indirect effect (a × b) is not
significant.
SUMMARY
Mediation analysis checks if an independent variable (X) affects a dependent
variable (Y) through a mediator (M).
Direct effect (c') and indirect effect (a × b) are tested to determine mediation strength.