Chapter Three Mean Difference Analysis (T-Test, Analysis of Variance)
Chapter Three Mean Difference Analysis (T-Test, Analysis of Variance)
Chapter Three Mean Difference Analysis (T-Test, Analysis of Variance)
Then sX 5 5
est. M .35
N 200 14.14
If (11 10)
X 11 z 2.83; 2.83 1.96 p .05
.35
The t Distribution
We use t when the population variance is unknown (the
usual case) and sample size is small (N<100, the usual
case).
The t distribution is a short, fat relative of the normal.
The shape of t depends on its df. As N becomes infinitely
large, t becomes normal.
Single-sample t-test
X tˆ M
Interval =
11 2.064(1) [8.936, 13.064]
Interval is about 9 to 13 and contains 10, so n.s.
Independent Samples t (1)
( x1 x2 ) ( 1 2 )
Looks just like z: t diff est diff
df=n1-1+n2-1=n1+n2-2
If SDs are equal, estimate is:
2 2 2 1 1
diff
n1 n2 n1 n2
Pooled variance estimate is weighted average:
2 [(n1 1) s12 (n2 1) s22 ] /[1 /(n1 n2 2)]
Pooled Standard Error of the Difference (computed):
H 0 : 1 2 0; H1 : 1 2 0
( x1 x2 ) ( 1 2 )
t diff est diff
x1 18; s12 7; n1 5
x2 20; s22 5.83; n2 7
tcrit = t(.05,10)=2.23
Dependent t
d
(d ) i
i
( d d ) 2
est. MD
sd
n s
2
d
n 1 n
d D0
t D f =n(pairs)-1
sd / n
Analysis of Variance
ANOVA (Analysis Of Variance) is a statistical method
for determining the existence of differences among
several population means.
ANOVA is designed to detect differences among means
from populations subject to different treatments
ANOVA is a joint test
The equality of several population means is tested
simultaneously or jointly.
ANOVA tests for the equality of several population means
by looking at two estimators of the population variance
(hence, analysis of variance).
One-Way Analysis of Variance
10-13
One-Way ANOVA Assumptions
Continued
Normality
The p populations of values of the response
variable all have normal distributions
Independence
The samples of experimental units are randomly
selected, independent samples
10-14
Testing for Significant Differences
Between Treatment Means
10-15
Testing for Significant Differences
Between Treatment Means Continued
10-16
Partitioning the Total Variability in the
Response
p ni p p ni
xij x
2
ni xi x 2
xij xi 2
i 1 j 1 i 1 i 1 j 1
10-17
Mean Squares
10-18
F Test for Difference Between
Treatment Means # 1
10-19
F Test for Difference Between
Treatment Means #2
10-20
F Test for Difference Between
Treatment Means #3
10-21
Example
22
Cont’d
SST = 5(56.8-58)2 +5(58-58)2+ 5(60.2-58) = 31.4
SSE = (40-56.8)2 + ….+(67-56.8)2+(67-58)2+……+
(56-58)2+(66-60.2)2+…..+(56-60.2)2 =3441.6
TSS = SST +SSE = 3473
MST=SST/p-1 = 31.4/3-1=15.7
MSE=SSE/n-p = 3441.6/15-3= 286.8
F= MST/MSE=15.7/286.8=0.055
Fα (2,12)= F.05 (2,12) =3.89. Since F =.055 < Fα = 3.89,
do not reject H0.(μA= μB= μC)
23
F Test for Block Effects
H0: No difference between block effects
Ha: At least two block effects differ
MSB SSB/ p 1
Test Statistic : F=
MSE SSE/ p-1 b-1
Reject H0 if
F > For
p-value <
Fis based on b-1 numerator and (p-1)
(b-1) denominator degrees of freedom
10-24
Two-Way Analysis of
Variance
A two factor factorial design compares the mean response for a levels
of factor 1 (for example, display height) and each of b levels of factor 2
( for example, display width.) A treatment is a combination of a level of
factor 1 and a level of factor 2
Factor 2
1 2 3… b
1
2 xijk = response for the
Factor 1
10-27
Estimation of Treatment Differences
Under Two-Way ANOVA, Factor 2
10-28
MANOVA
MANOVA can be thought of as ANOVA for
situations in which there are several dependent
variables.
The principles of ANOVA extend to
MANOVA in that we can use MANOVA when
there is only one independent variable or when
there are several, we can look at interactions
between independent variables, and we can
even do contrasts to see which groups differ
from each other.
10-29
Cont’d
10-30
10-31
Chi-Square: χ 2
32
Different Scales, Different Measures of
Association
34
Cont’d
35
Steps in Test of Hypothesis
36
1. Determine Appropriate Test
Chi Square is used when both variables are measured
on a nominal scale.
It can be applied to interval or ratio data that have
been categorized into a small number of groups.
It assumes that the observations are randomly
sampled from the population.
All observations are independent (an individual can
appear only once in a table and there are no
overlapping categories).
It does not make any assumptions about the shape of
the distribution nor about the homogeneity of
variances.
37
2. Establish Level of Significance
α is a predetermined value
The convention
α = .05
α = .01
α = .001
38
3. Determine The Hypothesis:
Whether There is an Association or Not
39
4. Calculating Test Statistics
( Fo Fe ) 2
2
Fe
41
4. Calculating Test Statistics
Ob
fre se
qu rve
en d
cie
s
( Fo Fe ) 2
2
Fe
Ex que
fre
pe ncy
ct e
d
qu e d
cy
fre pect
en
Ex
42
Example
43
Bivariate Frequency Table or
Contingency Table
Favor Neutral Oppose f row
Democrat 10 10 30 50
Republican 15 15 10 40
f column 25 25 40 n = 90
44
Bivariate Frequency Table or
Contingency Table
Favor Neutral Oppose f row
Democrat 10 10 30 50
Republican 15 15 10 40
f column ed 25 25 40 n = 90
r v ies
b se n c
O que
fre 45
Row frequency
Bivariate Frequency Table or
Contingency Table
Favor Neutral Oppose f row
Democrat 10 10 30 50
Republican 15 15 10 40
f column 25 25 40 n = 90
46
Bivariate Frequency Table or
Contingency Table
Favor Neutral Oppose f row
Democrat 10 10 30 50
Republican 15 15 10 40
f column 25 25 40 n = 90
Column frequency
47
1. Determine Appropriate Test
Degree of freedom
(row-1) (column-1)
48
3. Determine The Hypothesis
49
4. Calculating Test Statistics
50
4. Calculating Test Statistics
51
4. Calculating Test Statistics
52
4. Calculating Test Statistics
= 11.03
53
6. Compare computed test
statistic against a tabled/critical
value
α = 0.05
df = 2
Critical tabled value = 5.991
Test statistic, 11.03, exceeds critical value
Null hypothesis is rejected
Democrats & Republicans differ significantly in
their opinions on gun control issues
54
The end
Thank you!
55