STA4C04 - Statistical Inference and Quality Control
STA4C04 - Statistical Inference and Quality Control
QUALITY CONTROL
(STA4C04)
STUDY MATERIAL
IV SEMESTER
B.Sc. MATHEMATICS
(2019 Admission)
UNIVERSITY OF CALICUT
SCHOOL OF DISTANCE EDUCATION
CALICUT UNIVERSITY P.O.
MALAPPURAM - 673 635, KERALA
19562
School of Distance Education
University of Calicut
Study Material
IV Semester
B.Sc. MATHEMATICS
(2019 Admission)
Prepared by:
Dr. APARNA ARAVINDAKSHAN M.,
Assistant Professor,
Department of Statistics,
St. Joseph’s College, Devagiri.
Scrutinized by:
Ms. JEENA M.P.,
Assistant Professor,
Department of Statistics,
Thunjan Memorial Government College, Tirur.
DISCLAIMER
"The author(s) shall be solely responsible
for the content and views
expressed in this book".
STA4C04 : Statistical Inference and Quality Control
CONTENTS
1 Estimation Theory 1
2 Testing of Hypothesis 49
ii
REFERENCES 164
MODULE
ONE
Estimation Theory
1. Data is quantitative.
1. Estimation of a parameter
2. Testing of hypothesis
1. Point estimation
2. Interval estimation
Parametric Space
T1 = X̄
T2 = X1
X1 + X2
T3 =
2
Xn n
X
T4 = aj Xj , 0 ≤ aj ≤ 1 aj = 1
j=1 j=1
will give us more information at the lowest cost and so on. There
are four desirable characteristics of a good estimator. They are,
i) Unbiasedness
ii) Consistency
iii) Efficiency
iv) Sufficiency
Unbiasedness
10
estimator if
E(T ) = θ ∀ n.
Problems
11
Consistency
12
lim P (|Tn − θ| ≤ ) = 1
n→∞
or
lim P (|Tn − θ| ≥ ) = 0.
n→∞
P
And we write Tn → θ.
V (Tn )
P [|Tn − E(Tn )| > ] ≤ .
2
Therefore,
P [|Tn − θ| > ] → 0 as n → ∞.
13
Problems
14
P
n → ∞. That is, γ(Tn ) → γ(θ). Hence, γ(Tn ) is a consistent
estimator of γ(θ).
Efficiency
We know that
M.S.E.θ (T ) = V (T ) + (Bias)2 .
15
Vθ (T2 )
R.E.θ (T1 |T2 ) = .
Vθ (T1 )
Note:
Problems
1. P.T. in estimating the mean of a normal population, sam-
ple mean is more efficient than the sample median.
16
Sufficiency
17
Note: The sample mean x̄ satisfies all the above four prop-
erties when it is being used as an estimator for estimating the
population mean θ.
Problems
1. A r.s. of size 2 is taken from B(1, θ). S.T. X1 + X2 is
sufficient for θ.
18
19
Problems
1. A r.s. of size 2 is taken from B(1, θ). S.T. X1 + X2 is
sufficient for θ using factorisation theorem.
Minimal Sufficiency
20
Completeness
E[g(T )] = 0 ⇒ g(T ) = 0.
Likelihood Function
21
L(θ; x1 , x2 , . . . , xn ) = f (x1 , x2 , . . . , xn ; θ)
= f (x1 ; θ) f (x2 ; θ) . . . f (xn ; θ)
Yn
= f (xi ; θ)
i=1
22
23
Cramer-Rao Inequality
Then,
[ψ 0 (θ)]2
V (T ) ≥ ,
I(θ)
where I(θ) is called the Fisher information about θ contained in
the sample given by
2
∂ ln L
I(θ) = E
∂θ
2
∂ ln L
= −E
∂θ2
2
∂ ln f
= nE
∂θ
2
∂ ln f
= −n E
∂θ2
24
[ψ 0 (θ)]2
V (T ) = = C.R.L.B.,
I(θ)
∂ ln L
= A(θ)[T − ψ(θ)])
∂θ
ψ 0 (θ)
and V (T ) = A(θ) .
Problems
25
where σ is known.
Methods of Estimation
Method of Moments
26
Properties
Problems
27
∂L ∂2L
= 0 and if for that value of θ, < 0.
∂θ ∂θ2
The value of θ that maximises L(θ) will also maximise ln L(θ).
So M.L.E. can also be obtained by solving
∂ ln L ∂ 2 ln L
= 0 and if for that value of θ, < 0.
∂θ ∂θ2
Properties
28
statistic exists.
Problems
29
30
31
Case I : σ is known
X̄−µ
When σ is known Z = √σ ∼ N (0, 1). Then, from the area
n
property of standard normal distribution, the 100(1-α)% C.I.
32
for µ is given by
h i
P |Z| ≤ z α = 1 − α
2
h i
i.e., P −z α ≤ Z ≤ z α = 1 − α
2 2
" #
X̄ − µ
i.e., P −z α ≤ ≤ zα =1−α
2 √σ 2
n
σ σ
i.e., P −z α √ ≤ X̄ − µ ≤ z α √ =1−α
2 n 2 n
σ σ
i.e., P −X̄ − z α √ ≤ −µ ≤ −X̄ + z α √ =1−α
2 n 2 n
σ σ
i.e., P X̄ − z α √ ≤ µ ≤ X̄ + z α √ =1−α
2 n 2 n
33
34
X̄−µ
When σ is unknown and n is small S
√
∼ tn−1 . Now, by the
n
symmetry of t - distribution
h i
P |t| ≤ tn−1, α = 1 − α
2
h i
i.e., P −tn−1, α ≤ t ≤ tn−1, α = 1 − α
2 2
" #
X̄ − µ
i.e., P −tn−1, α ≤ ≤ tn−1, α =1−α
2 √S 2
n
S S
i.e., P −tn−1, α √ ≤ X̄ − µ ≤ tn−1, α √ =1−α
2 n 2 n
S S
i.e., P −X̄ − tn−1, α √ ≤ −µ ≤ −X̄ + tn−1, α √ =1−α
2 n 2 n
S S
i.e., P X̄ − tn−1, α √ ≤ µ ≤ X̄ + tn−1, α √ =1−α
2 n 2 n
35
Note:
σ2
X̄ → N µ, as n → ∞.
n
h i
Then the 100(1-α)% C.I. for µ is X̄ − z α √σ , X̄ + zα √σ .
2 n 2 n
36
Problems
Let X̄1 be the sample mean of a r.s. of size n1 from N (µ1 , σ12 )
and X̄2 be the sample mean of a r.s. of size n2 from N (µ2 , σ22 ).
37
σ12 σ2
X̄1 − X̄2 ∼ N µ1 − µ2 , + 2
n1 n2
and
(X̄1 − X̄2 ) − (µ1 − µ2 )
Z= r ∼ N (0, 1).
σ12 σ22
n + n
1 2
h i
P |Z| ≤ z α = 1 − α
2
h i
i.e., P −z α ≤ Z ≤ z α = 1 − α
2 2
−z α ≤ (X̄1 −r
X̄2 ) − (µ1 − µ2 )
i.e., P 2 ≤ zα
=1−α
σ 2 σ 2 2
n
1
+ n
2
1 2
38
" s
σ12 σ2
i.e., P −z α + 2 ≤ (X̄1 − X̄2 ) − (µ1 − µ2 )
2 n1 n2
s #
σ12 σ2
≤ zα + 2 =1−α
2 n1 n2
" s
σ12 σ2
i.e., P −(X̄1 − X̄2 ) − z α + 2 ≤ −(µ1 − µ2 )
2 n1 n2
s #
σ12 σ2
≤ −(X̄1 − X̄2 ) + z α + 2 =1−α
2 n1 n2
" s
σ12 σ2
i.e., P (X̄1 − X̄2 ) − z α + 2 ≤ (µ1 − µ2 )
2 n1 n2
s #
σ12 σ2
≤ (X̄1 − X̄2 ) + z α + 2 =1−α
2 n1 n2
39
Let
(n1 − 1)S12 + (n2 − 1)S22 ∗2
=σ .
n1 + n2 − 2
40
Then
(X̄1 − X̄2 ) − (µ1 − µ2 )
r ∼ t(n +n −2) .
∗2 1 2
σ n1 + n1
1 2
Case I : µ is unknown
When µ is unknown,
(n − 1)S 2
∼ χ2(n−1) .
σ2
41
(n − 1)S 2
2 2
i.e., P χ(n−1),1− α ≤ ≤ χ(n−1), α = 1 − α
2 σ2 2
χ2(n−1),1− α χ2(n−1), α
" #
1
i.e., P ≤ 2 ≤
2 2
=1−α
(n − 1)S 2 σ (n − 1)S 2
" #
(n − 1)S 2 2 (n − 1)S 2
i.e., P ≥ σ ≥ =1−α
χ2(n−1),1− α χ2(n−1), α
2 2
" #
(n − 1)S 2 2 (n − 1)S 2
i.e., P ≤ σ ≤ =1−α
χ2(n−1), α χ2(n−1),1− α
2 2
Case II : µ is known
When µ is known,
Pn
(Xi − µ)2
i=1
∼ χ2(n) .
σ2
42
p0 − p
Z=q → N (0, 1)
p0 q 0
n
43
h i
P |Z| ≤ z α = 1 − α
2
h i
i.e., P −z α ≤ Z ≤ z α = 1 − α
2 2
−z α p0 − p
i.e., P ≤q ≤ zα = 1 − α
2 p0 q 0 2
n
" r r #
p0 q 0 p0 q 0
i.e., P −z α ≤ p0 − p ≤ z α =1−α
2 n 2 n
" r r #
0 p0 q 0 pq
i.e., P −p − z α ≤ −p ≤ p0 + z α =1−α
2 n 2 n
" r r #
0 p0 q 0 p0 q 0
i.e., P p − zα ≤ p ≤ p0 + z α =1−α
2 n 2 n
44
Note:
45
where p01 and p02 are the point estimate of p1 and p1 respectively.
By refereing to normal table we can find z α such that
2
h i
P |Z| ≤ z α = 1 − α
2
h i
i.e., P −z α ≤ Z ≤ z α = 1 − α
2 2
0 0
− p2 ) − (p1 − p2 )
−z α ≤ (p1 r
i.e., P ≤ zα
=1−α
2 0 0
p1 q1 0 0
p2 q2 2
n + n
1 2
" s
p01 q10 p0 q 0
i.e., P −z α + 2 2 ≤ (p01 − p02 ) − (p1 − p2 )
2 n1 n2
s #
p01 q10 p0 q 0
≤ zα + 2 2 =1−α
2 n1 n2
" s
p01 q10 p0 q 0
i.e., P −(p01 − p02 ) − z α + 2 2 ≤ −(p1 − p2 )
2 n1 n2
s #
0 0 p01 q10 p0 q 0
≤ −(p1 − p2 ) + z α + 2 2
2 n1 n2
=1−α
46
" s
p01 q10 p0 q 0
i.e., P (p01 − p02 ) + z α + 2 2 ≥ (p1 − p2 )
2 n1 n2
s #
0 0 p01 q10 p0 q 0
≥ (p1 − p2 ) − z α + 2 2
2 n1 n2
=1−α
" s
p01 q10 p0 q 0
i.e., P (p01 − p02 ) − z α + 2 2 ≤ (p1 − p2 )
2 n1 n2
s #
0 0 p01 q10 p0 q 0
≤ (p1 − p2 ) + z α + 2 2
2 n1 n2
=1−α
47
1+r
Zr = 0.5 ln
1−r
Zr − µr
Z= ∼ N (0, 1)
σr
h i
P |Z| ≤ z α = 1 − α
2
h i
i.e., P −z α ≤ Z ≤ z α = 1 − α
2 2
48
Zr − µr
i.e., P −z α ≤ ≤ zα = 1 − α
2 σr 2
h i
i.e., P −z α σr ≤ Zr − µr ≤ z α σr = 1 − α
2 2
h i
i.e., P −Zr − z α σr ≤ −µr ≤ −Zr + z α σr = 1 − α
2 2
h i
i.e., P Zr − z α σr ≤ µr ≤ Zr + z α σr = 1 − α
2 2
1+ρ
i.e., P Zr − z α σr ≤ 0.5 ln ≤ Zr + z α σr = 1 − α
2 1−ρ 2
1+ρ
i.e., P a ≤ 0.5 ln ≤b =1−α
1−ρ
e2a − 1 e2b − 1
P 2a ≤ ρ ≤ 2b =1−α
e −1 e −1
MODULE
TWO
Testing of Hypothesis
49
50
Hypothesis
51
Statistical Hypothesis
Parametric Hypothesis
Non-parametric Hypothesis
52
population is 10.
53
Null Hypothesis
Alternative Hypothesis
Test of Hypothesis
54
Critical Region
Acceptance Region
Test Statistic
55
True State
Decision
H0 H1
Reject H0 Type I Error Correct Decision
Accept H0 Correct Decision Type II Error
56
57
Level of Significance
The fixed level of the probability of type I error is called the level
of significance. The significance level, denoted as α, is a measure
of the strength of the evidence that must be present in your
sample before you will reject the null hypothesis and conclude
that the effect is statistically significant. The significance level
or level of significance is the probability of rejecting the null
hypothesis when it is true. For example, a significance level of
0.05 indicates a 5% risk of concluding that a difference exists
when there is no actual difference. Lower significance levels
indicate that you require stronger evidence before you will reject
the null hypothesis. It is also called level of the test or size of
the test or size of the critical region or producer’s risk. The
researcher determines the significance level before conducting
the experiment.
58
Best test
Among all the tests whose size is less than or equal to α, the
one for which P(Type II error) is minimum is called the best
test.
Power of a Test
= P(rejecting H0 |H1 )
= 1 − P(accepting H0 |H1 )
= 1 − P(type II error)
= 1−β
59
The larger the value of 1 − β for fixed α, the better is the test
in general. Sensitiveness of a test is its ability to ascertain the
correctness of the alternative hypothesis when it is true for fixed
α. Thus, power is a measure of sensitiveness of the test.
Among all the tests whose size is less than or equal to α, the
one for which β is maximum is called the most powerful test.
Critical Value
The value of the test statistic which separates the critical re-
gion and acceptance region is called critical value. It is usually
referred as ∆α (Zα , tα etc.) depending on the sampling distri-
bution of the test statistic and the level of significance used. It
is also called the tolerance limit.
p-value
60
61
Neymann-Pearson Lemma
The best critical region or the most powerful critical region for
testing a simple H0 : θ = θ0 against the simple H1 : θ = θ1 ,
is given by the well known Neymann Pearson Lemma. That is,
the theorem specifies the critical region which has pre assigned
probability of type I error and a minimal probability of type II
error. This is the same as maximizing the power subject to the
condition that probability of type I error is a pre determined
constant (α).
62
Parametric Tests
Y − E(Y )
Z= p .
V (Y )
63
64
statistic.
Any test where the critical region consist only one tail of
the sampling distribution of the test statistic is called a ‘one
tailed test’, particularly they are called ‘left tailed test’ and
‘right tailed test’ respectively.
65
66
H0 : µ = µ0
1. H1 : µ < µ0 or
2. H1 : µ > µ0 or
3. H1 : µ 6= µ0
x̄ − µ0
Z= √
σ/ n
For the significance level α, the best critical regions are respec-
tively,
67
1. Z < −Zα
2. Z > Zα
3. |Z| ≥ Zα/2
Using the sample data, calculate the value of the test statistic
Z . If it lies in the critical region, reject H0 , otherwise accept
H0 .
x̄ − µ0
Z= √ .
S/ n
68
x̄ − µ0
Z = √
σ/ n
65 − 60
= √ = 2.5
10/ 25
69
and found that they have an average life of 191 hours. Is the
claim of the manufacturer justified?
70
H0 : µ1 = µ2
1. H1 : µ1 < µ2
2. H1 : µ1 > µ2
3. H1 : µ1 6= µ2 .
x¯1 − x¯2
Z=q .
σ1 2 σ2 2
n1 + n2
1. Z < −Zα
2. Z > Zα
3. |Z| ≥ Zα/2
71
x¯1 − x¯2
Z=q .
S1 2 S2 2
n1 + n2
72
H 0 : p = p0
1. H1 : p < p0
2. H1 : p > p0
3. H1 : p 6= p0 .
73
p0 − p
Z= q 0
p0 q0
n
1. Z < −Zα
2. Z > Zα
3. |Z| ≥ Zα/2
74
H 0 : p1 = p2
1. H1 : p1 < p2
75
2. H1 : p1 > p2
3. H1 : p1 6= p2
n1 p1 0 +n2 p2 0
where p∗ = n1 +n2 and q ∗ = 1 − p∗ .
1. Z < −Zα
2. Z > Zα
3. |Z| ≥ Zα/2
76
77
t, F and χ2 .
1X 2
s2 = (xi − x̄)
n
t - tests
78
79
H0 : µ = µ0
1. H1 : µ < µ0
2. H1 : µ > µ0
3. H1 : µ 6= µ0
Or
x̄ − µ0
t= ∼ tn−1 .
√s
n−1
1. t < −tn−1,α
80
2. t > tn−1,α
3. |t| ≥ tn−1,α/2
81
82
3. The sample sizes are small (n1 < 30, n2 < 30)
1. H1 : µ1 < µ2
2. H1 : µ1 > µ2
3. H1 : µ1 6= µ2 .
x¯1 − x¯2
t= r ∼ tn1 +n2 −2 .
(n1 −1)S12 +(n2 −1)S22
1 1
n1 +n2 −2 n1 + n2
Or
x¯1 − x¯2
t= r ∼ tn1 +n2 −2
n1 s21 +n2 s22
1 1
n1 +n2 −2 n1 + n2
83
3. |t| ≥ tn +n −2,α/2
1 2
84
Sample 1 : 14,18,12,9,16,24,20,21,19,17
Sample 2: 20,24,18,16,26,25,18.
85
H0 : µ = 0
1. H1 : µ < 0
2. H1 : µ > 0
3. H1 : µ 6= 0
Or
d¯
t= s ,
√ d
n−1
P q P
d
where d¯ = n i and sd = n1 ¯ 2 . For a given signifi-
(di − d)
cance level α, the best critical regions are respectively,
1. t < −tn−1,α
2. t > tn−1,α
86
3. |t| ≥ tn−1,α/2
87
H0 : ρ = 0
1. H1 : ρ < 0
2. H1 : ρ > 0
3. H1 : ρ 6= 0
88
1. t < −tn−2,α
2. t > tn−2,α
3. |t| ≥ tα/2,n−2
89
90
H0 : σ 2 = σ 0 2
against
H1 : σ 2 > σ0 2
ω : χ2 > χα 2 ,
91
Problems
This test is used for testing the equality of variances (or stan-
dard deviations) of two normal populations. Draw a random
sample of size n1 from N µ1 , σ1 2 and a sample of size n2 from
S1 2 and S2 2 be their sample variances. Here
N µ2 , σ2 2 . Let
we want to test
H0 : σ 1 2 = σ 2 2
92
against
H1 : σ1 2 > σ2 2 .
S12
F = ∼ F(n −1)(n −1) .
S22 1 2
Notes :
1. Before starting the F test, compute S12 and S22 . Take the
larger value as the numerator in the F ratio and n1 corre-
sponds to the value in the numerator.
93
Problems
1. Sample of sizes 10 and 18 taken from two normal popula-
tions gave s1 = 14 and s2 = 20. Test the hypothesis that
the samples have come from populations with the same
standard deviation.
94
95
from the data exceeds limit, we reject H0 . Thus the chi square
test for goodness of fit is basically a right tailed test. The best
critical region is χ2 > χα 2 , where χα 2 is obtained by referring
the chi square table for (n−1) degrees of freedom. Now calculate
the value of the test statistic and if lies in the critical region,
reject H0 or otherwise.
Problems
96
X : 0 1 2 3 4 5 6
F : 275 72 30 7 5 2 1
Let there be two attributes A and B, ‘A’ divided into ‘m’ classes
A1 , A2 , . . . Am and ‘B’ divided into ‘n’ classes B1 , B2 , . . . Bn .
The various cell frequencies can be expressed as in the following
table, having ‘m’ rows and ‘n’ columns.
97
98
m X
n 2
2
X Oij − Eij
χ =
Eij
i=1 j=1
Problems
(Hint : χ2 = 0.249)
99
is given by,
2
N (ad − bc)
χ2 = ,
(a + b) (c + d) (a + c) (b + d)
where N = a + b + c + d.
Yates’ Correction
100
101
102
Assumptions:
One-Way Classification
103
Pk
into k classes of sizes n1 , n2 , . . . , nk , N = i=1
ni as exhibited
in the table below.
Class/ Sample Observations Total
Treatment
1 y11 y12 ... y1n1 T1
2 y21 y22 ... y2n T2
2
.. .. .. .. ..
. . . . .
i yi1 yi2 ... yin Ti
j
.. .. .. .. ..
. . . . .
k yk1 yk2 ... ykn Tk
k
Grand Total G
This scheme of classification according to 1 factor or criterion
is called one-way classification and its analysis is called one-way
ANOVA.
104
yij = µ + αi + ij
That is,
H0 : µ1 = µ2 = . . . = µk
105
component parts.
Note 2.4.1. N − 1 = (k − 1) + (N − k)
106
T.S.S.
M.S.T. =
N −1
S.S.Tr.
M.S.Tr. =
k−1
S.S.E.
M.S.E. =
N −k
E(M.S.E.) = σ 2
Otherwise,
E(M.S.Tr.) > E(M.S.E.)
107
M.S.Tr.
F = ∼ Fk−1,N −k .
M.S.E.
Under H0 , F = 1, otherwise F > 1. So, smaller values of
F support H0 and larger values of F support H1 . Hence, the
critical region is
F > Fα .
ANOVA Table
Total T.S.S. N −1
G2
Correction Factor, C.F. =
N
108
(T1 )2 (T )2 (T )2
S.S.Tr. = + 2 + . . . + k − C.F.
n1 n2 nk
Two-Way Classification
109
110
That is,
H0A : µA1 = µA2 = . . . = µAk
That is,
H0B : µB1 = µB2 = . . . = µBh
111
where, S.S.A. is the S.S. due to criterion A (i.e., the row S.S.),
S.S.B. is the S.S. due to criterion B (i.e., the column S.S.), and
S.S.E. is the S.S. due to error. The degrees of freedom (d.f.)
due to various sum of squares are as follows:
S.S. d.f.
T.S.S. kh − 1
S.S.A. k−1
S.S.B. h−1
S.S.E. (k − 1)(h − 1)
T.S.S.
M.S.T. =
kh − 1
S.S.A.
M.S.A. =
k−1
S.S.B.
M.S.B. =
h−1
112
S.S.E.
M.S.E. =
(k − 1)(h − 1)
E(M.S.E.) = σ 2
Otherwise,
E(M.S.A.) > E(M.S.E.)
Otherwise,
E(M.S.B.) > E(M.S.E.)
M.S.A.
FA = ∼ Fk−1,(k−1)(h−1) .
M.S.E.
113
M.S.B.
FB = ∼ Fh−1,(k−1)(h−1) .
M.S.E.
ANOVA Table
Total T.S.S. N −1
114
G2
Correction Factor, C.F. =
N
(R1 )2 (R )2 (R )2
S.S.A. = + 2 + . . . + k − C.F.
h h h
(C1 )2 (C )2 (C )2
S.S.B. = + 2 + . . . + h − C.F.
k k k
S.S.E. = T.S.S. − S.S.A. − S.S.B.
MODULE
THREE
Non-parametric Tests
115
116
Advantages
1. Easily understandable.
2. Short calculations.
Disadvantages
117
118
Test Procedure
119
Normal Approximation
120
121
1. H1 : M > M0
2. H1 : M < M0
3. H1 : M 6= M0
Test Procedure
122
Since the median divides the data in such a way that the
number of observations above it is equal to the number of ob-
servations below it, the probability of a positive sign is same as
the probability of the negative sign.
1
H0 : p =
2
against one of the possible alternatives
1
1. H1 : p > 2
1
2. H1 : p < 2
1
3. H1 : p 6= 2
We reject H0 , in favor of
123
1
:: either N − is too small or alterna-
H1 : M > M 0 p > 2
tively N + is too big
1
H1 : M < M0 p < 2 :: if either N + is too small or alter-
natively N − is too big
1
:: if any one of N + or N − is too big
H1 : M 6= M0 p 6= 2
or too small.
124
H 0 : M1 = M2
1. H1 : M1 > M2
2. H1 : M1 < M2
3. H1 : M1 6= M2
125
Test Procedure
1
H0 : p =
2
against one of the possible alternatives
1
1. H1 : p > 2
1
2. H1 : p < 2
1
3. H1 : p 6= 2
We reject H0 , in favor of
126
1
:: if any one of N + or N − is too
H1 : M1 6= M2 p 6= 2
big or too small.
127
Median Test
H0 : M1 = M2
H1 : M1 6= M2 .
128
Test Procedure
and
V ∼ B(n2 , p2 ),
H 0 : p1 = p2
H1 : p1 6= p2 or p1 − p2 6= 0.
U V
−
n1 n2
129
U V
and the null is rejected if n1 − n2 > Cα .
where N = a + b + c + d.
Now, calculate
n1
n2
a
p= n
b .
a+b
130
N (|ad − bc| − N2 )2
χ2 = ∼ χ2 (1).
(a + b)(c + d)(a + c)(b + d)
Wilcoxon rank sum test is used to test whether the two popu-
lation have the same distribution. That is,
H0 : FX (x) = FY (x).
H 0 : M1 = M2
1. H1 : M1 > M2
2. H1 : M1 < M2
3. H1 : M1 6= M2
131
Test Procedure
Normal Approximation
n1 (N +1)
W− 2
Z= n1 n2 (N +1)
∼ N (0, 1),
12
132
1. H1 : M1 > M2 if Z > zα
3. H1 : M1 6= M2 if |Z| ≥ zα/2
Mann-Whitnney U Test
H0 : FX (x) = FY (x) ∀x
against
H1 : FX (x) = FY (x − θ) ∀ x and θ 6= 0.
133
H0 : M1 = M2
1. H1 : M1 > M2
2. H1 : M1 < M2
3. H1 : M1 6= M2
Test Procedure
134
and
n 1 n2
X X
U2 = (1 − T (xi , yj )).
i=1 j=1
135
Normal Approximation
2. H1 : M1 < M2 if Z > zα
3. H1 : M1 6= M2 if |Z| ≥ zα/2
136
Kruskal-Wallis Test
in other words
H 0 : M1 = M2 = . . . = Mk
against
137
Test Procedure
Combine all the samples and rank them. Let Ri denote the sum
of the ranks in ith group. The test statistic is
k
X R2
12
H= i
− 3(n + 1),
n(n + 1) ni
i=1
138
MODULE
FOUR
Quality Control
139
140
Here, quality does not mean the highest standard, but the
quality which conforms to the standards specified for measure-
ment. In many cases, it will be lower than the highest possible
standard. In fact, consistency in quality standard is expected
rather than the absolute standard. Quality control, therefore,
covers all the factors and processes of production which may be
broadly classified as follows:
141
142
143
144
145
146
ance.
Acceptance Sampling
147
Control Charts
148
149
150
151
3σ Control limits
or
P [|T − µ| > 3σ] = 0.0027.
152
X̄ - Chart
153
U.C.L.x̄ = µ0 + Aσ 0
L.C.L.x̄ = µ0 − Aσ 0
¯ + A2 R̄
U.C.L.x̄ = x̄
154
¯ − A2 R̄
L.C.L.x̄ = x̄
¯ − A1 s̄.
L.C.L.x̄ = x̄
R - Chart
155
U.C.L.R = D2 σ
L.C.L.R = D1 σ
U.C.L.R = D4 R̄
L.C.L.R = D3 R̄.
s or σ - Chart
156
s1 +s2 +...+sk
2. Compute the C.L., C.L.s = s̄ = k
U.C.L.R = B2 σ
L.C.L.R = B1 σ
U.C.L.R = B4 s̄
L.C.L.R = B3 s̄.
157
158
1. They are charts for variables only, i.e., for quality charac-
teristics which can be measured and expressed in numbers.
159
p - chart
160
d
Let the sample proportion of defective is p = n. Then,
1 1
E(p) = E(d) = np = p
n n
and
1 1 pq
V (p) = V (d) = 2 npq = , q = 1 − p.
n2 n n
p1 +p2 +...+pk
2. Compute the C.L., C.L.p = p̄ = k
161
np - chart or d - chart
162
c - Chart
163
164
BIBLIOGRAPHY
[4] Mood, A.M. Graybill, F.A. and Boes, D.C. (2007): Intro-
duction to the Theory of Statistics, 3rd Edn., (Reprint),
Tata McGraw-Hill Pub. Co. Ltd.
165
166