0% found this document useful (0 votes)
19 views4 pages

Analysis of Variance

The document provides an overview of variance and One-Way ANOVA, explaining how variance measures the spread of data points in a dataset and the significance of high vs. low variance. It details the steps to calculate variance, the role of independent and dependent variables, and the process of conducting One-Way ANOVA to compare means across groups. Additionally, it discusses the null hypothesis, p-values, and methods for mean multiple comparisons following ANOVA results.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
19 views4 pages

Analysis of Variance

The document provides an overview of variance and One-Way ANOVA, explaining how variance measures the spread of data points in a dataset and the significance of high vs. low variance. It details the steps to calculate variance, the role of independent and dependent variables, and the process of conducting One-Way ANOVA to compare means across groups. Additionally, it discusses the null hypothesis, p-values, and methods for mean multiple comparisons following ANOVA results.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 4

ANALYSIS OF VARIANCE Step 2: Deviation from the

ONE-WAY ANOVA Mean:


Calculate the difference
VARIANCE between each data point and
 A statistical measurement of the spread the mean. These are called
between numbers in a data set. deviations.
 It is a measure that tells us how spread out or
scattered the values in a dataset are. Step 3: Square the
 It quantifies the average squared difference Deviations:
between each data point and the mean of the Square each deviation to
dataset. eliminate negative
values. This is because
HIGH VARIANCE VS LOW VARIANCE we're interested in the
If data is widely spread out, it means the variance is magnitude of the
high. This indicates that the data points are far apart spread, not the direction.
from each other and from the mean (average) of the
dataset. Step 4: Average the Squared Deviations:
Here’s what a high variance tells us: Sum up all the squared deviations and divide by the
number of data points minus one (n−1). This gives you
the variance.

VARIANCE plays a crucial role in various statistical


concepts like hypothesis testing, confidence
intervals, and regression analysis.
It helps us understand the variability within a dataset
and make informed decisions based on that
Greater Variability: The data points are less information.
consistent and show a larger range of values.
Less Predictability: It becomes harder to predict the
value of a data point based on the overall trend.
Potential Outliers: A high variance might suggest the
presence of outliers, which are extreme values
significantly different from the rest of the data.

Examples:
Student Test Scores: If test scores in a class have a
high variance, it means some students scored very
high, while others scored very low.
Stock Prices: If stock prices have a high variance, it
ONE-WAY ANOVA
means they fluctuate a lot, making it risky to invest.
 One-Way ANOVA ("analysis of variance")
compares the means of two or more
HIGH VARIANCE VS. LOW VARIANCE
independent groups to determine whether
In contrast, low variance means data points are
there is statistical evidence that the associated
clustered closely together. The data is more consistent
population means are significantly different.
and predictable, with less variability.

Example: In a class, if the variance in test scores is


low, it means students are performing similarly well.
This could indicate effective teaching, a well-aligned
curriculum, or a group of students with similar learning
abilities.

WHAT IS VARIANCE?
 ANOVA is used to compare the means of two
or more independent variables to test
whether the difference between the means is
INDEPENDENT VARIABLE VS. DEPENDENT
statistically significant.
VARIABLE
 A high variance indicates that the data points
INDEPENDENT VARIABLE (THE CAUSE):
are widely scattered, while a low variance
 This variable is the one that is manipulated
suggests that the data points are clustered
or changed by the researcher.
closely around the mean.
 It’s the "input" or the factor believed to
influence the outcome.
UNDERSTANDING VARIANCE.
Example: Heights of Students
Example: In a study examining the effect of fertilizer
Let's say we have the heights of five students: 160
on plant growth, the amount of fertilizer is the
cm, 170 cm, 165 cm, 175 cm, and 180 cm.
independent variable. The researcher controls how
much fertilizer is applied to different groups of plants.
Step 1: Calculate the Mean: Find the average
(mean) of your data set: DEPENDENT VARIABLE (THE EFFECT):
 This variable is measured or observed in media use to find out if there is a difference in hours of
response to changes in the independent sleep per night.
variable.
 It’s the "output" or the outcome of interest. Null Hypothesis
There is no significant difference in the hours of sleep
Example: In the plant growth study, the height of the per night among individuals with low, medium, and
plants is the dependent variable. high levels of social media use.

P-VALUE Step 1. Gather data on hours of sleep per night for


 Used in hypothesis testing to help decide each group (low, medium, and high social media use)
whether to reject the null hypothesis.
 The smaller the p-value, the more likely you
are to reject the null hypothesis.

 A p-value less than 0.05 is typically considered


statistically significant, in which case the null
hypothesis should be rejected. Step 2. Encode our data in Microsoft excel
 A p-value greater than 0.05 means that deviation
from the null hypothesis is not statistically
significant, and the null hypothesis is not
rejected.

The commonly used threshold for statistical


significance is 0.05. If the p-value is less than 0.05,
we reject the null hypothesis (there’s a significant
effect). If it’s greater than 0.05, we fail to reject the
null hypothesis (no significant effect).

P value = less than 0.05 (null hypothesis is rejected) Step 3. Select Data Analysis Tool
P value = more than 0.05 (null hypothesis is Go to the "Data" tab in Excel.
accepted) Look for the "Data Analysis" tool.

NULL HYPOTHESIS
 All statistical tests have a null hypothesis.
 For most tests, the null hypothesis is that there
is no relationship between your variables of
interest or no difference among groups.

Example: If you don't see it, you may need to enable the
Effect of Foliar Fertilizer on the Growth and Yield of Analysis ToolPak:
Purple Corn Go to File > Options > Add-Ins.
In the Manage box, select "Excel Add-ins" and click
Null hypothesis: Foliar fertilizer has no significant "Go..."
effect on the growth and yield of purple corn. Check the "Analysis ToolPak" box and click "OK".

When to Use One-Way ANOVA Step 4. Run ANOVA Analysis


The One-Way ANOVA is often used to analyze data Click on "Data Analysis" in the "Data" tab.
from the following types of studies: Select "ANOVA: Single Factor" from the list of analysis
 Field studies tools and click "OK".
 Experiments
 Quasi-experiments

Use a one-way ANOVA when you have collected data


about one categorical independent variable and one
quantitative dependent variable.

The One-Way ANOVA is commonly used to test


the following:
 Statistical differences among the means of two
or more groups
 Statistical differences among the means of two
or more interventions
 Statistical differences among the means of two
or more change scores

Both the One-Way ANOVA and the Independent


Samples t-Test can compare the means for two
groups. However, only the One-Way ANOVA can
compare the means across three or more groups.

Example:
Your independent variable is social media use, and you
assign groups to low, medium, and high levels of social
Step 5. Input Range field
Select the range of your data
For "Alpha", enter the significance level (commonly
0.05).

Range Field

MEAN MULTIPLE COMPARISON


In statistics, mean multiple comparison refers to a
set of techniques used to identify which specific group
means are significantly different from each other after
finding a significant overall effect (like in ANOVA).

Since ANOVA tells you that at least one group differs,


multiple comparison tests help pinpoint the exact
differences.

When to Use Multiple Comparison Tests:


 After finding a significant result from ANOVA.
 When you have three or more groups and want
to understand the specific differences.

Common Methods for Mean Multiple


Comparisons:
 Tukey’s Honestly Significant Difference (HSD)
 Bonferroni Correction
 Scheffé’s Test
 Dunnett’s Test
 Newman-Keuls Test

TUKEY’S HONESTLY SIGNIFICANT DIFFERENCE


(HSD):
Step 1: Calculate the Means

Step 2: Calculate the Overall Mean


Step 3: Calculate the ANOVA
 Calculate the sum of squares total (SST)
 The sum of squares between groups (SSB)
 And the sum of squares within groups (SSW)

Step 4: Conduct the ANOVA


 dfbetween= 3
 dfwithin= 8
 Sum of squares between groups (SSB) =
28.1667
 Sum of squares within groups (SSW) = 28.75
 Total sum of squares = 56.9167
 Mean Square Calculations (MSB) = 9.39
 Mean Square Within (MSW) = 3.59

You might also like