ANOVA (Analysis of Variance)
ANOVA (Analysis of Variance)
ANOVA (Analysis of Variance)
ANOVA is a statistical technique that assesses potential differences in a scale-level dependent variable by a nominal-
level variable having 2 or more categories. For example, an ANOVA can examine potential differences in IQ scores
by Country (US vs. Canada vs. Italy vs. Spain). The ANOVA, developed by Ronald Fisher in 1918, extends the t and
the z test which have the problem of only allowing the nominal level variable to have two categories. This test is
also called the Fisher analysis of variance.
General Purpose of ANOVA
Researchers and students use ANOVA in many ways. The use of ANOVA depends on the research
design. Commonly, ANOVAs are used in three ways: one-way ANOVA, two-way ANOVA, and N-way ANOVA.
One-Way ANOVA
A one-way ANOVA has just one independent variable. For example, difference in IQ can be assessed by
Country, and County can have 2, 20, or more different categories to compare.
Two-Way ANOVA
A two-way ANOVA refers to an ANOVA using two independent variables. Expanding the example above, a 2-
way ANOVA can examine differences in IQ scores (the dependent variable) by Country (independent variable
1) and Gender (independent variable 2). Two-way ANOVA can be used to examine the interaction between the
two independent variables. Interactions indicate that differences are not uniform across all categories of the
independent variables. For example, females may have higher IQ scores overall compared to males, but this
difference could be greater (or less) in European countries compared to North American countries. Two-way
ANOVAs are also called factorial ANOVAs.
N-Way ANOVA
A researcher can also use more than two independent variables, and this is an n-way ANOVA (with n being the
number of independent variables you have). For example, potential differences in IQ scores can be examined
by Country, Gender, Age group, Ethnicity, etc, simultaneously.
General Purpose – Procedure
Omnibus ANOVA test:
The null hypothesis for an ANOVA is that there is no significant difference among the groups. The alternative
hypothesis assumes that there is at least one significant difference among the groups. After cleaning the data,
the researcher must test the assumptions of ANOVA. They must then calculate the F-ratio and the associated
probability value (p-value). In general, if the p-value associated with the F is smaller than .05, then the null
hypothesis is rejected and the alternative hypothesis is supported. If the null hypothesis is rejected, one
concludes that the means of all the groups are not equal. Post-hoc tests tell the researcher which groups are
different from each other.
So what if you find statistical significance? Multiple comparison tests
When you conduct an ANOVA, you are attempting to determine if there is a statistically significant difference
among the groups. If you find that there is a difference, you will then need to examine where the group
differences lay.
At this point you could run post-hoc tests which are t tests examining mean differences between the
groups. There are several multiple comparison tests that can be conducted that will control for Type I error
rate, including the Bonferroni, Scheffe, Dunnet, and Tukey tests.
Types of Research Questions the ANOVA Examines
One-way ANOVA: Are there differences in GPA by grade level (freshmen vs. sophomores vs. juniors)?
Two-way ANOVA: Are there differences in GPA by grade level (freshmen vs. sophomores vs. juniors) and
gender (male vs. female)?
Data Level and Assumptions
The level of measurement of the variables and assumptions of the test play an important role in ANOVA. In
ANOVA, the dependent variable must be a continuous (interval or ratio) level of measurement. The
independent variables in ANOVA must be categorical (nominal or ordinal) variables. Like the t-test, ANOVA is
also a parametric test and has some assumptions. ANOVA assumes that the data is normally distributed. The
ANOVA also assumes homogeneity of variance, which means that the variance among the groups should be
approximately equal. ANOVA also assumes that the observations are independent of each other. Researchers
should keep in mind when planning any study to look out for extraneous or confounding variables. ANOVA has
methods (i.e., ANCOVA) to control for confounding variables.
Testing of the Assumptions
1. The population from which samples are drawn should be normally distributed.
2. Independence of cases: the sample cases should be independent of each other.
3. Homogeneity of variance: Homogeneity means that the variance among the groups should be approximately
equal.
These assumptions can be tested using statistical software (like Intellectus Statistics!). The assumption of
homogeneity of variance can be tested using tests such as Levene’s test or the Brown-Forsythe
Test. Normality of the distribution of the scores can be tested using histograms, the values of skewness and
kurtosis, or using tests such as Shapiro-Wilk or Kolmogorov-Smirnov. The assumption of independence can be
determined from the design of the study.
It is important to note that ANOVA is not robust to violations to the assumption of independence. This is to say,
that even if you violate the assumptions of homogeneity or normality, you can conduct the test and basically
trust the findings. However, the results of the ANOVA are invalid if the independence assumption is
violated. In general, with violations of homogeneity the analysis is considered robust if you have equal sized
groups. With violations of normality, continuing with the ANOVA is generally ok if you have a large sample size.
Related Statistical Tests: MANOVA and ANCOVA
Researchers have extended ANOVA in MANOVA and ANCOVA. MANOVA stands for the multivariate analysis
of variance. MANOVA is used when there are two or more dependent variables. ANCOVA is the term for
analysis of covariance. The ANCOVA is used when the researcher includes one or more covariate variables in
the analysis.
he ANOVA Test
An ANOVA test is a way to find out if survey or experiment results are significant. In other words, they help you to figure out if
you need to reject the null hypothesis or accept the alternate hypothesis. Basically, you’re testing groups to see if there’s a
difference between them. Examples of when you might want to test different groups:
A group of psychiatric patients are trying three different therapies: counseling, medication and biofeedback. You want to see if
one therapy is better than the others.
A manufacturer has two different processes to make light bulbs. They want to know if one process is better than the other.
Students from different colleges take the same exam. You want to see if one college outperforms the other.
Types of Tests.
There are two main types: one-way and two-way. Two-way tests can be with or without replication.
One-way ANOVA between groups: used when you want to test two groups to see if there’s a difference between them.
Two way ANOVA without replication: used when you have one group and you’re double-testing that same group. For example,
you’re testing one set of individuals before and after they take a medication to see if it works or not.
Two way ANOVA with replication: Two groups, and the members of those groups are doing more than one thing. For
example, two groups of patients from different hospitals trying two different therapies.
Back to Top
What is MANOVA?
Analysis of variance (ANOVA) tests for differences between means. MANOVA is just an ANOVA with several dependent
variables. It’s similar to many other tests and experiments in that it’s purpose is to find out if the response variable (i.e. your
dependent variable) is changed by manipulating the independent variable. The test helps to answer many research questions,
including:
Do changes to the independent variables have statistically significant effects on dependent variables?
What are the interactions among dependent variables?
What are the interactions among independent variables?
MANOVA Example
Suppose you wanted to find out if a difference in textbooks affected students’ scores in math and science. Improvements in
math and science means that there are two dependent variables, so a MANOVA is appropriate.
An ANOVA will give you a single (“univariate”) f-value while a MANOVA will give you a multivariate F value. MANOVA tests the
multiple dependent variables by creating new, artificial, dependent variables that maximize group differences. These new
dependent variables are linear combinations of the measured dependent variables.
Disadvantages
1. MANOVA is many times more complicated than ANOVA, making it a challenge to see which independent variables are affecting
dependent variables.
2. One degree of freedom is lost with the addition of each new variable.
3. The dependent variables should be uncorrelated as much as possible. If they are correlated, the loss in degrees of freedom
means that there isn’t much advantages in including more than one dependent variable on the test.
Reference:
(SFSU)
Back to Top
Variability
In a one-way ANOVA, variability is due to the differences between groups and the differences within groups. In factorial ANOVA,
each level and factor are paired up with each other (“crossed”). This helps you to see what interactions are going on between
the levels and factors. If there is an interaction then the differences in one factor depend on the differences in another.
Let’s say you were running a two-way ANOVA to test male/female performance on a final exam. The subjects had either had 4,
6, or 8 hours of sleep.
1. Is sex a main effect? In other words, do men and women differ significantly on their exam performance?
2. Is sleep a main effect? In other words, do people who have had 4,6, or 8 hours of sleep differ significantly in their performance?
3. Is there a significant interaction between factors? In other words, how do hours of sleep and sex interact with regards to exam
performance?
4. Can any differences in sex and exam performance be found in the different levels of sleep?