Statical Chapman
Statical Chapman
for technology
A COURSE IN APPLIED STATISTICS
Third ed itio n (R evised)
Statistics
for technology
A COURSE IN APPLIED STATISTICS
Third edition (Revised)
Christopher Chatfield
Reader in Statistics
Bath University, UK
This book contains information obtained from authentic and highly regarded sources. Reprinted material
is quoted with permission, and sources are indicated. A wide variety o f references are listed. Reasonable
efforts have been made to publish reliable data and information, but the author and the publisher cannot
assume responsibility for the validity o f all materials or for the consequences o f their use.
Apart from any fair dealing for the purposes o f research or private study, or criticism or review, as
permitted under the UK Copyright Designs and Patents Act, 1988, this publication may not be reproduced,
stored or transmitted, in any form or by any means, electronic or mechanical, including photocopying,
microfilming, and recording, or by any information storage or retrieval system, without the prior per
m ission in writing o f the publishers, or in the case o f reprographic reproduction only in accordance with
the terms o f the licenses issued by the Copyright Licensing Agency in the UK, or in accordance with
the terms o f the license issued by the appropriate Reproduction Rights Organization outside the UK.
The consent o f CRC Press LLC does not extend to copying for general distribution, for promotion,
for creating new works, or for resale. Specific permission must be obtained in writing from CRC Press
LLC for such copying.
Direct all inquiries to CRC Press LLC, 2000 N.W. Corporate Blvd., Boca Raton, Florida 33431.
Tradem ark Notice: Product or corporate names may be trademarks or registered trademarks, and
are used only for identification and explanation, without intent to infringe.
Preface 11
1 Outline of statistics 15
4 Discrete distributions 56
4.1 The discrete probability distribution 56
4.2 The binomial distribution 57
4.3 The binomial model 59
4.4 Types of distribution 64
4.5 The mean and variance of the binomial distribution 64
4.6 The Poisson distribution 69
4.7 Bivariate discrete distributions 75
5 Continuous distributions 81
5.1 Definitions 81
5.2 The mean and variance of continuous distributions 86
5.3 The normal distribution 87
5.4 Uses of the normal distribution 92
5.5 Normal probability paper 95
5.6 The exponential distribution 98
5.7 Bivariate continuous distributions 103
5 Contents
6 Estimation 106
6.1 Point and interval estimates 106
6.2 Properties of the expected value 107
6.3 The sampling distribution of x 111
6.4 The sampling distribution of 116
6.5 Some properties of estimators 118
6.6 General methods of point estimation 121
6.7 Interval estimation 126
6 Contents
10.3 The number of replications 227
10.4 Randomization 230
10.5 The analysis of a randomized comparative experiment 231
10.6 The range test 235
10.7 One-way analysis of variance 237
10.8 Follow-up study of the treatment means 241
10.9 Verifying the model 243
10.10 The randomized block experiment 244
10.11 Two-way analysis of variance 248
10.12 Latin squares 252
10.13 Balanced incomplete block designs 253
7 Contents
Appendix A The relationships between the normal, t- and
F-distributions 332
Index 377
8 Contents
‘It’s p erfectly in te llig ib le ,’ the ca p ta in said , in an o ffen d ed
to n e, ‘to a n y o n e th at u n d erstan d s su ch th in g s.’
Preface
10 Preface
getting too involved in mathematical details. 4 few results are stated
without proof where this is unlikely to affect the student’s comprehen
sion. However, I have tried to explain carefully the basic concepts of
the subject, such as probability and sampling distributions; these the
student must understand. The worst abuses of statistics arise in the
‘cook-book’ approach to the subject where scientists try to analyse
their data by substituting measurements into statistical formulae which
they do not understand.
Many readers will have access to a computer, or at least to a
microcomputer or sophisticated pocket calculator. The statistical
software which is readily available has taken much of the drudgery
out of statistics. There are some remarks on using a computer in
Appendix E.5. I have amended the text in a number of places to make
it clear that the reader need no longer worry too much about some
computational details, although there is much to be said for working
through some of the methods in detail at least once in order to
understand fully what is going on.
I am grateful to many people for constructive comments on earlier
editions and I am always pleased to hear from any reader with new
ideas.
I am indebted to Biometrika trustees for permission to publish
extracts from Biometrika Tables for Statisticians (Appendix B, Table 6)
and to Messrs Oliver and Boyd Ltd, Edinburgh for permission to
publish extracts from Statistical Tables for Biological, Agricultural and
Medical Research (Appendix B, Table 7).
The quotations are, of course, from the works of Charles Lutwidge
Dodgson, better known as Lewis Carroll.
C h r is t o p h e r C h a t f ie l d
Bath University
May, 1989
11 Preface
Part one
Introduction
Example 1
The thrust of a rocket engine was measured at ten-minute intervals
while being run at the same operating conditions. The following thirty
observations were recorded (in newtons x 10 ^):
15 Outline of statistics
Example 2
The numbers of cosmic particles striking an apparatus in forty con
secutive periods of one minute were recorded as follows.
3 0 01 0 2 1 0 1 1
0 3 4 1 2 0 2 0 3 1
10 12 0 2 1 0 1 2
3 1 00 2 1 0 3 1 2
The observations vary between zero and four, with zero and one
observed more frequently than two, three and four. Again there is
experimental uncertainty since we cannot exactly predict what the
next observation would be. However, we expect that it will also be
between zero and four and that it is more likely to be a zero or a one
than anything else. In Chapter 4 we will see that there is indeed a
pattern in this data even though individual observations cannot be
predicted.
Example 3
Twenty refrigerator motors were run to destruction under advanced
stress conditions and the times to failure (in hours) were recorded as
follows.
16 Outline of statistics
A scientific experiment has some or all of the following character
istics.
(1) The physical laws governing the experiment are not entirely under
stood.
(2) The experiment may not have been done before, at least success
fully, in which case the instrumentation and technique are not fully
developed.
(3) There are strong incentives to run the smallest number of the
cheapest tests as quickly as possible.
(4) The experimenter may not be objective, as for example when an
inventor tests his own invention or when a company tests competitive
products.
(5) Experimental results are unexpected or disappointing. (Engineers
explain disappointing results as illustrating Murphy’s law. The
corresponding law for statistics might be phrased ‘if two events are
equally likely to occur, the worse will happen’.)
(6) Although experimental uncertainty may be present, many industrial
situations require decisions to be made without additional testing or
theoretical study.
To illustrate how statistics can help at different stages of an experi
ment, let us assume that we have been given the task of improving the
performance of a space pump. The basic design of this machine is
illustrated in Figure 1: gas is passing through the pump which is
driven by an electric motor.
The first step is to study the current technology. Statistical methods
are never a substitute for understanding the physical laws governing
inlet
temperature..^
pressure
outlet
temperature
pressure
Figure 1 Diagram of a space pump
17 Outline of statistics
the problem under consideration; rather statistics is a tool for the
scientist in the same sense as differential equations and digital
computers.
The second step is to define the objective of the test program as
precisely as possible. For example we may be interested in improving
flow, pressure rise or efficiency, or reducing weight or noise. Alterna
tively we may be interested in improving the reliability of the machine
by increasing the time between breakdowns.
The objective being defined, a list of variables or factors can be
made which will vary or be varied during the test program. In the space-
pump experiment, the variables include power, inlet temperature, inlet
pressure and speed. The test program must be designed to find the best
way of choosing successive values for the different factors. The problems
involved in designing an experiment are discussed in Chapters 10 and
11.
In order to see if the objective is being achieved, we look at the outputs
or responses. In the space-pump example, these include pressure rise,
flow, efficiency and reliability. Note that these responses are probably
interconnected. For example it may be pointless to increase the flow if,
as a result, the efficiency drops or the reliability falls off (that is, the
pump breaks down more frequently).
During the experiment, measurements of the factors and responses
will be made. The article being tested should be instrumented to obtain
the most precise and accurate data. The problems of instrumentation
are discussed in Chapter 9. The analysis of the resulting data should
attempt to determine not only the individual effect of each variable on
the responses, but also the joint effects of several variables, the inter
actions. The process of estimating these effects is called inference. These
estimates should have known statistical properties such as a lack of
systematic error (called unbiasedness). Finally any conclusions or
recommendations should be given together with a measure of the risk
involved and/or the possible error in the estimates.
In order to carry out the above procedure successfully it is usually
helpful to set up a mathematical model to describe the physical situation.
This model is an attempt to formulate the system in mathematical
terms and may, for example, consist of a series of formulae connecting
the variables of interest. This model may involve unknown coefficients
or parameters which have to be estimated from the data. After this
has been done, the next step is to test the validity of the model. Finally,
if the model is sound, the solution to the problem can be obtained from
it.
18 Outline of statistics
The key point to note in this Chapter is that experimental
uncertainty is present in most practical situations and that there is
nothing abnormal about it. Despite its presence, the study of statistics
enables us to collect, analyse and interpret data in a sensible way.
Additional reading
Many other books provide an introduction to statistical techniques.
The references in the list given below have been selected to provide a
non-technical easy-to-read approach. They motivate the subject,
provide examples where statistics has been used successfully, and also
give examples where statistics has been used in a misleading way. The
book by Darrell Huff is particularly amusing and could profitably be
read by everyone. The book by Peter Sprent explains the key prin
ciples of Applied Statistics, and would serve as a useful complement
to a more formal text book such as this one. The text book by Andrew
Ehrenberg adopts a refreshing non-technical approach with the
emphasis on improving numeracy.
19 Outline of statistics
Chapter 2
Simple ways of summarizing
data
2.1 Introduction
The subject ‘Statistics’ is concerned with collecting reliable data and then
analysing and interpreting them. We begin by introducing some simple
methods for summarizing data which are often given the collective title
‘Descriptive Statistics’. They include plotting appropriate graphs and
calculating various summary statistics, possibly using a computer package
such as MINITAB. Note that some supplementary material has been
added in Appendices D .l to D.4 and E.3. The idea is to get a ‘feel’ for
the data and pick out their most important features. The quality of the
data should also be assessed by looking for errors, outliers (see section
11.9), missing observations and other peculiarities. This preliminary
analysis is always advisable before choosing a more advanced statistical
technique, and may prove sufficient anyway.
It may seem strange to consider data analysis before data collection but
it turns out that the topics are easier to learn in this order. Nevertheless,
it is clear that the selected sample must be chosen to be representative of
the population from which it is drawn. Statistics will then enable us to
make deductions about the whole population from the information in the
sample.
The data usually consist of a series of measurements on one or more
quantities of interest, called variates. A discrete variate can only take a
sequence of distinct values which are usually integers. For example, the
number of cosmic particles striking a recorder in one minute may be 0, 1 ,
2, . . . . The number U cannot be observed. On the other hand, a
continuous variate can take any value in a specified interval on a
continuous scale. For example, the thrust of a rocket engine can be any
non-negative number. In practice the recorded values of a continuous
variate will be rounded to a specified number of significant figures and
Number of
cosmic particles Frequency
0 13
I 13
2 8
3 5
4 I
Total 40
21 Graphical methods
> i5r
^0h
....... 4 ..............
number of cosmic particles
Example 1
The heights of 100 students were measured to the nearest inch and
tabulated as follows.
Table 2
Height Number of
(inches) students
60-62 6
63-65 15
66-68 40
69-71 30
72-74 9
Total 100
The data were divided into five groups as shown and the frequencies
with which the different groups occur form a frequency distribution.
The data can be plotted as a histogram in which rectangles are con
structed with areas proportional to the frequency of each group.
Example 2
Plot a histogram of the data given in Example 1, Chapter 1. The
smallest thrust observed is 989*4 and the largest 1014*5. The difference
between them is about twenty-five units so that three units is a reason
able class interval. Then we will have about ten class intervals. Group
the observations into the ten intervals as in Table 3.
23 Graphical methods
Table 3
Frequency distribution of thrust
Number of
Class interval observations
987-990 2
990-993 1
993-996 3
996-999 5
999-1002 7
1002-1005 6
1005-1008 3
1008-1011 1
1011-1014 1
1014-1017 1
(a)
(b)
'liilL v
25 Graphical methods
Frequency curve. Where there are a large number of observations
the histogram may be replaced with a smooth curve drawn through
the midpoints of the tops of each box. Such a curve is called a frequency
curve.
Example 3
Plot the cumulative frequency diagram of the data given in Example 1,
Chapter 1.
Using Table 3 it is easy to calculate the cumulative frequencies at
the divisions between the class intervals.
Cumulative Relative
Thrust frequency frequency
987 0 0000
990 2 0-067
993 3 0-100
996 6 0-200
999 11 0-367
1002 18 0-600
1005 24 0-800
1008 27 0-900
1011 28 0-933
1014 29 0-967
1017 30 1-000
27 Graphical methods
2.3 Summary statistics
In addition to the graphical techniques, it is useful to calculate
some figures to summarize the data. Any quantity which is calculated
from the data is called a statistic (to be distinguished from the subject
statistics). Thus a statistic is a function of the measurements or observa
tions.
Most simple statistics can be divided into two types; firstly quantities
which are ‘typical’ of the data and secondly quantities which measure
the variability of the data. The former are usually called measures of
location and the latter are usually called measures of spread.
The mean. Suppose that n measurements have been taken on the variate
under investigation, and these are denoted by Xj, X2, . . . , The
(arithmetic) mean of the observations is given by
Xi + X 2 + . . .-hX,
X = ”= y ^ 2.1
n’
Example 4
Find the average thrust of the rocket engine from the data in Example 1,
Chapter 1.
999* 1 +1003-2+ ... +1000*9
We find X = _
- 1000- 6.
Z
i= 1
\ * (see also Appendix D.l)
Z /
i= 1
Example 5
In Table 1 we have Xi = 0 occurs with frequency fi = 13. Similarly
Xi = 1 occurs with frequency /2 = 13 and so on. Then we find
(Ox 13) + (1 X 13) + (2 x 8) + (3 x5) + (4x 1)
13+13 + 8 + 5 + 1
_ 48
“ 40
= 1- 2.
Thus an average of 1*2 cosmic particles are emitted every minute.
Example 6
When continuous data are tabulated as in Table 2, the sample mean
can be estimated by assuming that all the observations in a class interval
fall at the class mark.
In this case we find
(61 x 6) + (64x 15)-h(67x40) + (70x30) + (73 x9)
100
67-6
= mean height of the students.
At this point it is worth repeating the fact that a set of data is a sample
drawn from the population of all possible measurements. Thus it is
important to realize that the sample mean, x, may not be equal to the
true population mean, which is usually denoted by p. In Chapter 6
we will see that x is an estimate of p. Similarly the other sample statistics
described here will also be estimates of the corresponding population
characteristics.
29 Summary statistics
The median. This is occasionally used instead of the mean, particularly
when the histogram of the observations is skewed (see Figure 5). It is
obtained by placing the observations in ascending order of magnitude
and then picking out the middle observation. Thus half the observations
are numerically greater than the median and half are smaller.
Example 7
The weekly wages of twelve workers, selected at random from the
payroll of a factory, are (to the nearest pound)
83, 111, 127, 96, 124, 103, 82, 99, 173, 137, 102, 106.
The data give ic = £lll*9.
Rewriting the observations in ascending order of magnitude we
have
82, 83, 96, 99, 102, 103, 106, 111, 124, 127, 137, 173.
As there are an even number of observations, the median is the average
of the sixth and seventh values, namely £104-5.
As eight of the observations are less than the sample mean, it could
be argued that the median is ‘more typical’ of the data. The outlying
observation, 173, has a considerable effect on the sample mean, but
not much effect on the median.
The mode. This is the value of the variate which occurs with the
greatest frequency. For discrete data the mode can easily be found by
inspection. For continuous data the mode can be estimated by plotting
the results in a histogram and finding the midpoint of the tallest box.
Thus in Example 1 the mode is 67 inches.
Range. This is the difference between the largest and smallest observa
tions. It is useful for comparing the variability in samples of roughly
equal size, but unfortunately it depends on the sample size and tends to
get larger as the sample size increases.
Interquartile range. This is the difference between the upper and lower
quartiles (see page 361) and thus includes the middle half of the
observations when put into ascending order of magnitude. Unlike the
range, it does not depend on the sample size and is not affected by a few
outlying values. Thus it is a useful ‘robust’ statistic which is particularly
useful for describing skew distributions, perhaps supplemented by the
values of the median and the highest and lowest observations.
However for distributions which are approximately symmetric, we will
see that there are strong theoretical and practical considerations for
preferring a statistic called the standard deviation, which is essentially the
root-mean-square deviation about the mean. The square of the standard
deviation is called the variance and has many uses, though not as a
descriptive statistic.
31 Sunnmary statistics
Variance and standard deviation. The sample variance 5^ of n observa
tions, Xi, X2, . . . , is given by
(x, -x )^ + (X 2 -x )^ + . ..+ (x „ -x )^
=
(n-I)
(x .-x )
2.2
n 1
The standard deviation s of the sample is obtained by taking the square
root of the variance.
I= 1 n *
However, since is not generally known, the sample mean x is used
instead. Then it can be shown theoretically that it is better to change
the denominator from n to {n—\). (This point is discussed in Section
6.5, where is found to be an unbiased estimate of the true population
variance, a^.) It is worth pointing out that the difference between n
and (n —1) is only important for small samples.
The standard deviation is in the same units as the original measure
ments and for this reason it is preferred to the variance as a descriptive
measure. However it is often easier from a theoretical and computa
tional point of view to work with variances. Thus the two measures are
complementary (see Appendix D.2 for further remarks on the
interpretation of a standard deviation).
It is sometimes more convenient to rearrange 2.2 in the form
-2
2 1(1/ 0- nx
s = 2.3
(n-l)
and this formula is used in many pocket calculators. But if x is much
bigger than s, then 2.3 may give rounding errors unless the data is
coded (see Exercise 3 below). Thus 2.2 is preferred for use on
computers. The formula for calculating the standard deviation of a
frequency distribution is given in Appendix D.l.
X x" = 7-54.
, (7.54-7*26)
s = -------^------ = 0056,
5 = 0-237.
(coefficient of variation) = - .
X
33 Summary statistics
Example 9
Using 2.2 (or 2.3 with coding - see Exercise 3 below), the standard
deviation of the data from Example 1, Chapter 1, was found to be
s = 60.
It is not immediately clear if this is to be judged ‘large’ or ‘small’.
However the coefficient of variation is given by
60
= 0006
X 1000-6
and from this we Judge that the variation in the data is relatively small.
Exercises
1. The weight of an object was measured thirty times and the following
observations were obtained.
(a) 997 995 998 992 995 (here 990 is a suitable constant)
(b) 131 1 137-2 136-4 133-2 139 1 140-0
34 Summary statistics
Part two
Theory
Example 1
An air-to-air missile has a ‘kill’ ratio of 2 in 10. If 4 missiles are launched,
what is the probability that the target will not be destroyed? Is this a
statistics or probability problem?
The physical situation in this problem can be described by one
quantity: the ‘kill’ ratio. As this is known from previous tests we can
calculate the probability that a given number of missiles will hit the
target. Thus we have a probability problem. The solution is obtained
by considering the ‘miss’ ratio which must be 8 in 10. The probability
that all 4 missiles will miss the target, assuming that they are indepen
dently launched, is obviously less than the probability that just one
missile will miss the target. Later in the chapter we will see that the
probability is given by
(0-80)^ = 0-4096.
Example 2
A hundred missiles are launched and eleven ‘kills’ are observed. What
is the best estimate of the ‘kill’ ratio?
As the ‘kill’ ratio was unknown before the test was performed, we
have to estimate this quantity from the observations. The ratio of kills
to launches is the intuitive estimate, namely 0-11. This is a statistical
problem.
39 Some definitions
Example 4
Throw a fair die.
Sample space 1 2 3 4 5
Probability
The probability of an event is the sum of the probabilities of the sample
points which constitute the event.
Example 5
If two fair coins are tossed, obtaining at least one head is an event,
denoted by E, which consists of the three sample points (HH), (HT),
and (TH).
Probability of this event = P(E)
= P(HH) + P(HT) + P(TH)
4 + 4 + ^ ^ 4.
Example 6
If two fair dice are tossed, the sample space consists of the thirty-six
combinations shown below.
The notation (Ei + E 2) means that at least one of the events occurs;
as applied to mutually exclusive events it means that one event or the
other occurs.
Example 7
From a fair pack of well-shuffled cards one card is dealt. Define the
event E^ as ‘the card is a spade’ and event E 2 as ‘the card is a heart’.
These two events are mutually exclusive as one card cannot be both a
heart and a spade.
Then P { E ,^ E 2 ) = M + M 1
2-
41 Types of events
The general addition law may be demonstrated pictorially. Let the
area of a unit square represent the probability of the entire sample space.
Inside this draw two figures with areas equal to the respective proba
bilities of El and E 2 and such that the common area is equal to the
probability that both the events will occur.
The area of the event (Ei + E 2) is the sum of the areas of Ei and E 2
minus their common area Ei E 2 , which is included in both Ei and E 2 .
Thus we have
P(E i +E2) == P(E i ) + P (E 2)-P (E i E2).
Note that if Ej and E 2 are mutually exclusive, the probability of the
joint event (E 1 E 2) is zero - that is, P(E 1 E 2) = 0 - and then equation 3.2
reduces to equation 3.1. Mutually exclusive events have no common
area and can be depicted as in Figure 10.
Example 8
What is the probability of rolling two dice to obtain the sum seven
and/or the number three on at least one die? Let event Ej be ‘the sum
is T and event E 2 be ‘at least one 3 turns up’. By inspection of the table
in Example 6, we find
P (E i)^ ^ P(E2) = ih
P(E i E2) = ^ .
Thus P(E i + E 2) = P(E i ) + P(E 2) - P ( E i E 2) = A + =
This can be checked by looking at the table.
Events that are not mutually exclusive may be further classified as
dependent or independent events. Dependence between events is treated
by the notion of conditional probability.
3.4
P{p2)
The effect of the conditional information is to restrict the sample space
to the sample points contained in event E 2.
Example 9
Given that a roll of two fair dice has produced at least one three, what
is the probability that the sum is seven? Event Ei is ‘the sum is 7’ and
event E 2 is ‘at least one 3 turns up’.
P(E i E3
f (e JE2
P(^l)
2/36
(see Example 8)
11/36
11 •
43 Types of events
E x a m p l e 10
Given that the last roll of two dice produced the sum seven, what is
the probability that the next roll will also produce the sum seven?
Careful! If the dice are fair dice and the roll is preceded by a good
shake in a dice cup, the play is fair, which implies that dice have no
memory. Thus there is no connection between the results of successive
rolls, and they are independent events rather than dependent events.
In other words, the knowledge that the last roll was a seven contributes
nothing to our ability to predict the next roll.
Example 10 (continued)
As two successive rolls are independent the probability of the sum
seven on the next roll is still Te. The last result is irrelevant. This is a
subtle point.
Some people interpret the ‘law of averages’ as meaning that if the
last roll was a seven the next roll is less likely to be a seven; that the
dice will act to ‘even out’ the results. The implication is that the dice
have some kind of memory.
A diametrically opposed notion is that if sevens have been ‘popular’
they will continue to be. This idea is based on the belief that most real
gambling equipment will be slightly biased and therefore some events
will occur more often than expected. For this reason honest gambling
casinos make extreme efforts to keep their games of chance fair. It is
in their best interest. Most serious studies of casino data discredit both
the above theories. Games of chance do not have memories.
P(E,E,)
P(E,|E2) =
P(E,) ’
Example 11
What is the probability of rolling the sum seven with two fair dice,
one of which will be a three? Event Ej is ‘the sum is 1\ event E 2 is
‘at least one 3 occurs’.
P (E ,E 2) = P(E 2)P(E,|E 2)
= XA (see Example 9)
_ 1
—
Example 12
What is the probability of drawing two aces from a well-shuffled pack?
Event El = ace on the first card
Event E 2 = ace on the second card
P(E 2|Ei) is obtained by noting that if the first card drawn is an ace,
then there will be fifty-one cards left in the pack of which three are
aces.
Thus P(E i E 2) = P(E i )P(E2|E i )
- X 3T
= 00045.
Although two cards are drawn in the above experiment, the student
may think of the situation as a single trial. We can write
El = ace on first card together with any second card,
E 2 = ace on second card together with any first card,
El E 2 = ace on first and second cards.
45 Types of events
Alternatively the student may think of the experiment as two trials.
The result of the second trial (the second card) depends on the result of
the first trial (the first card). The laws of probability give the same
result whichever way the problem is studied.
Example 13
A fair die is thrown. What is the probability that it will show 6 on the
first throw and 2 on the second throw?
El = 6 on first throw,
E 2 == 2 on second throw.
But the occurrence or non-occurrence of event Ei has no bearing on
whether or not E 2 occurs. Thus the two events are independent.
A .
mdependisnt ij depymleni
P{B. EA'^PiPAPiPABA
Example 14
If there are three events then similar reasoning and diagrams will
provide all the required relationships.
By inspection,
F(Ei +E2 + E3) = P(E,) + P(E2) + P(E3)-P(E,E2)-P(E i E3)
-P(E2E3) + P(E,E2E3),
PÍE
P n ,)- ^( e ,E 2 E 3)
Ir ,E
(E ,|E
P{E,E2E,) = P{E,)P{E2)P{E,).
47 Types of events
E x a m p l e 15
P(Ei) = 0-81,
P(Ei) = l- P ( E i)
- 019.
P(Ei E 2) = P(E 2|E i )P(E i ) from equation 3.4. But a catastrophic failure
(E2) can only occur if a mission failure (E J occurs. Thus we have
E 1 E 2 = E 2.
= P(E3|E2)P(E2)
= 0-90 X01805
= 016245.
0 -9 5 005
P(E^) = P(E,) + P(E 3)
= 0-81+016245
= 0-97245
P(E^) = 1 -P(E 4) 0-9 0-1
= 0-02755.
P(E3) = P(E3E2)
El E3 E3 E2
= P(E3 E2)P(E2) -------------- ----- ^-------
E4 E4
= 0-10x0-1805
= 0-01805.
Check: P(E 4> = P(E 2) + P(E 3)
0-02755 = 0-0095 + 0-01805.
3.4.1 Permutations
The number of ways in which r items can be selected from n distinct
items, taking notice of the order of selection, is called the number of
permutations of n items taken r at a time, and is denoted by or
P(n, r).
= n ( n -l)...( n - r- h l)
n\
{n -r)l
For example, the number of permutations of the letters a, b, c taken
two at a time is
= 3 x 2 = 6.
These are ab, ac, be, ba, ca, cb.
3.4.2 Combinations
The number of ways in which r items can be selected from n distinct
items, disregarding the order of the selection, is called the number of
combinations of n items taken r at a time. This is denoted by "C^, or
I n\
This is very small indeed (<10 ^^). Most such events which are
recorded in the newspapers are probably hoaxes.
Example 17
An inspector draws a sample of five items from a batch of a hundred
valves which are numbered from one to a hundred to distinguish them.
How many distinct samples can he choose? If there is one defective
item in the batch, how may distinct samples of size five can be drawn
which contain the defective item? What is the probability of drawing
a sample which contains the defective item?
As we are not concerned with the order in which the valves are
selected, the total number of distinct samples is the number of combina
tions of five from a hundred, that is, In the second case one of
the items is fixed and we need to find the total number of ways of
selecting four more items from the remaining ninety-nine valves. This
is ^^€ 4.
Thus the probability of drawing a sample which contains the defective
item is
^^C4 J_
100^ m '
This probability can of course be written down straight away by noting
that any valve has a probability of of being in any particular sample.
51 Random variables
For example, if a penny is tossed twice, the number of heads which turn
up can be either 0, 1 or 2 according to the outcome of the experiment.
We say that the number of heads is a random variable, since it expresses
the result of the experiment as a number. Other simple random variables
are the temperature of a chemical process and the resistance of a
thermocouple.
More generally a random variable is a rule for assigning a particular
number to a particular experimental outcome. For example, if the
result of an experiment is recorded as say ‘the experiment was success
ful’ or ‘the experiment was not successful’, then to obtain a random
variable we must code the results so that, for example, a success corres
ponds to 1 and a failure to 0. We can also consider more complicated
random variables. For example, if a penny is tossed twice, the random
variable X = (number of heads)^ can take the values 0,1 or 4 according
to the outcome of the experiment. The sample variance is another
example of a random variable, since for any given series of observations,
Xi, X2 , . . . , there is a corresponding number
Example 18
The items in a sample of ten electrical units are classified as defective
or non-defective. Is the number of defectives in the sample, X, a discrete
or continuous random variable?
The number of defectives in the sample can be any integer between
0 and 10. Thus X is a discrete random variable (it cannot take the value
1^ for example).
Exercises
1. Two parts of a machine are manufactured independently. These two
parts have probabilities pi and p 2 respectively of failing. Find the
probability that
(a) neither part fails
(b) at least one part fails
(c) exactly one part fails.
2. A component A which has 90 per cent reliability (it will fail one in
ten times) is vital to the running of a machine. An identical component
is fitted in parallel to form system Si, and the machine will work
provided that one of these components functions correctly.
53 Exercises
3. In two rolls of a fair die what is the probability that
(a) the two outcomes are the same?
(b) the second outcome is larger than the first?
(c) the second outcome is a five?
= Pi^n)
= p {e ,) p (e ,\E2)p {e 4 e ,E2 ,.P (E „ |E „ _ ,...E 2).
9. If a batch contains ninety good and ten defective valves, what is the
probability that a sample size five drawn from the batch will not
contain any defective items?
11. You are playing Russian roulette with a gun which has six chambers.
What is the probability of someone being killed during the first six
pulls of the trigger? (The chamber is spun after each firing and you
may assume that the probability of hitting a loaded chamber = ^ This
‘game’ is not recommended!)
55 Exercises
Chapter 4
Discrete distributions
Example 1
Toss two fair coins. Consider the random variable X = number of
heads which turn up. Then X can take the values 0, 1 or 2 according
to the outcome of the experiment. From Example 3, Chapter 3 we
know P{X = 0) = i , P{X = 1) = P{X = 2) = The sum of these
probabilities is one. Thus this random variable follows a discrete
probability distribution.
Example 2
Throw two fair dice. Consider the random variable X = (sum of the
pips on the two dice). Then X can take the values 2, 3 ,4 ,..., 11,12
according to the outcome of the experiment. By inspection of the
table in Example 6, Chapter 3, we find
P{2) = ie P(3) = ^ P(4) = ^ P{5) = A i ’(6) = TO
PC) = ^ P(8) = ^ P(9) = ^ i’(lO) = ^ P{n) = ^
P {n) =
56 Discrete distributions
These probabilities sum to unity and we have another discrete distribu
tion.
The discrete probability distribution is a very important tool in
probability and statistics and in the remainder of the chapter we will
consider two valuable distributions which frequently arise in practice.
= ”C , / ( l - p r - ^ (r = 0 , l , . . . , 4
r=0
0-2
0^0^.......
:0-4r (c)
0-2
00
9 10 11 12 13 14 15
F ig u ré is Binomial distributions
(a) reverse J-shaped, a? = 3, p = ¿ (b) positively skewed or skewed to the right,
/? = 15, p = 0-1 (c) negatively skewed or skewed to the left, a? = 15, p = 0-9
(d) symmetric, a? = 20, p = 0-5
58 Discrete distributions
4.3 The binomial model
The binomial distribution is applicable whenever a series of trials is
made which satisfies the following conditions.
(1) Each trial has only two possible outcomes, which are mutually
exclusive. We will usually call the two outcomes success (denoted by S)
and failure (denoted by F). Other possible pairs of outcomes are
defective or non-defective, go or no-go, heads or tails.
(2) The probability of a ‘success’ in each trial is a constant, usually
denoted by p. Thus the probability of a ‘failure’ is (1 —p).
(3) The outcomes of successive trials are mutually independent.
A typical situation in which these conditions will apply (at least
approximately) occurs when several items are selected at random
from a very large batch and examined to see if there are any defective
items (that is, failures). The number of defectives in a sample size n
is a random variable, denoted by X, which can take any integer value
between 0 and n. In order to find the probability distribution of X,
we first consider the simple case when n = 2.
4.3.1 2 trials
If two trials are made, the sample space consists of the four points
SS, SF, FS and FF.
Probability (2 successes) = P(SS)
= P(S)P(S) by the assumption of indepen
dence
Example 3
Toss two fair pennies. Let X = number of heads showing. If we think
of a head as a success and a tail as a failure we have
p ~ (probability of a head in one throw)
_ 1
— 2-
Substituting this value of p into the above formulae we have
P(0) = i P{1) = i P(2) = I
as already obtained in Example 1.
4.3.2 n trials
More generally we will consider the case where n trials are carried
out. The probability of getting r successes in n trials can be found as
follows;
F(exactly r successes) = P{r successes and (n —r) failures)
There are many possible sequences which can give exactly r successes.
For example, we can start with r successes and finish with {n —r) failures.
Since successive trials are independent we have
P(SS... S F F ... F) = / ( I - pT ~ \
Every possible ordering of r successes and {n —r) failures will have
the same probability, p''(l —p)"“ '*. The number of possible orderings is
the number of combinations of r out of n. Thus we have
P(exactly r successes) — —p)”“'' (r = 0 ,1 ,2 ,..., n)
and so the number of successes follows the binomial distribution.
In order to derive the binomial distribution in the above way, we
have carefully stated certain assumptions about the behaviour of
successive trials. These assumptions constitute what is called the
binomial model. This model is a typical example of a mathematical
model in that it attempts to describe a particular physical situation in
mathematical terms. Such models may depend on one or more para
meters which specify how the system will behave. In the binomial model
60 Discrete distributions
the quantities n and p are parameters. If the values of the model para
meters are known, it is a simple matter to calculate the probability of
a particular outcome. We will now illustrate the use of the binomial
distribution with some examples.
Example 4
Roll three dice. Find the distribution of the random variable
X = (number of twos which turn up).
X can take the values 0, 1, 2 or 3. Now the probability of getting
a two for one die is An incorrect but popular deduction is that the
probability of getting at least one two from three dice is Let us call
the event ‘getting a two’ a success so that the probability of a success is
As the results from the three dice are independent, the binomial model
is applicable and the distribution of the number of twos is a binomial
distribution with parameters n = 3, p = ^.
P(0) = = 0-58,
i’(l) = = 0-35,
P(2) = ^C2(i)^(|) = 0-07,
P(3) = = 0-005.
Note that the probability of getting at least one 2 is 0-42 which is
considerably less than | .
In effect this means that if we roll 3 dice 100 times we would expect
to get the following results.
Number of occurrences
58
35
7
0
Example 5
On past evidence, an electrical component has a probability of 0*98 of
being satisfactory. What is the probability of getting two or more
defectives in a sample size five?
Example 6
The latest recoilless rocket launcher has a kill probability of 0-60. How
many rockets would you have to fire to have at least a 95 per cent
probability of destroying the target? What is the probability distribu
tion of salvos of this size?
Let the required salvo contain n rockets. The probability of success
of each rocket is given by p = 0-60. Thus the number of rockets which
actually hit the target has a binomial distribution with parameters n,
and p = 0-60. Thus the probability that no rockets hit the target is
given by (1 —p)” = 0-40".
Hence the probability that the target is destroyed (that is, at least
one rocket hits the target) is given by
1-0-40".
We are told that this must be greater than 0-95. Therefore,
1-0-40" > 0-95
0-05 > 0-40"
giving n = 4 since n must be an integer.
Then the probability distribution for salvos of this size is given by
62 Discrete distributions
P{r hits) = ^C(0'60)'‘(0-40)'^"'' (r = 0, 1,2, 3,4)
P(0 hits) = 0-026
P(1 hit) = 0-154
P{2 hits) = 0-346
P(3 hits) = 0-346
P(4 hits) = 0-130.
Example 7
An inspector takes a random sample of ten items from a very large
batch. If none of the items is defective he accepts the batch, otherwise he
rejects the batch. What is the probability that a batch is accepted if
the fraction defective is 0, 0 01, 0-02, 0 05, 01, 0-2, 0-5, 1? Plot this
probability against the fraction defective.
Figure 14
Table 5
Discrete Continuous
Theoretical
(parameters known) type I type III
Empirical
(parameters unknown) type II type IV
64 Discrete distributions
N
Z fi^i
X = ~N •
Expected Number of
frequency Observed observed
= P, X30 frequency even numbers
0 0-001 0-0 0 0
1 0-010 0- 3 0 0
2 0-044 1-3 2 4
3 0-117 3-5 3 9
4 0-205 6-2 5 20
5 0-246 7-4 9 45
6 0-205 6-2 3 18
7 0-117 3-5 6 42
8 0-044 1-3 1 8
9 0-010 0-3 1 9
10 0-001 0-0 _0 0
30-0 30 155
66 Discrete distributions
If X is a discrete random variable which can take values , X2, . . . ,
with probabilities Pi,P2,- - ^Pn then the average or expected value of
X is given by
It may not be immediately obvious that this formula will give a quantity
which is the average value of X in a long series of trials. However there
is a clear analogy with the formula for the sample mean if we use the
N
E XiPi
i= 1
E(X)
E p .-
E (X )= E " C ^ p V - p f- 'r
r =0
n- 1
= np E
r =0
= np[p + ( l- p ) ] " “ ‘
= np.
Example 8
Toss two coins and let X = number of heads. It is clear that the expected
number of heads will be one. This is confirmed by applying the above
formula which gives
E(X) = 0 x P { X = 0) + l xP(X - l) + 2 x P (X = 2)
= 1 x ^ + 2 xj:
= 1.
4.5.2 Variance
The idea of an expected value is also extremely useful when calculating
the theoretical variance, and hence the theoretical standard deviation,
of a probability distribution.
We have already seen in Chapter 2 how to calculate the sample
variance of a set of data. If n observations , . . . , are taken from a
population which has mean /i, the sample variance is given by
= E 4.2
68 Discrete distributions
variance (X) = —
In other words the variance of X is the average or expected value of
(X — in a long series of trials.
From equation 4.2 the theoretical variance of the binomial distribu
tion is given by
which after a lot of algebra gives the simple formula np{l —p). From this
result it follows that the standard deviation of the binomial distribution
is given by J[np{\-p)].
r=0 r=0 ^•
00 r
r =0 r\
—e ^ e
fiPjr)
P(r-hl) =
(r+l)‘
X denotes an event
Figure 15
70 Discrete distributions
Rearranging this equation we obtain
P(0,í + A í)-P(0, t)
At
The left hand side of this equation can be recognized as the definition of
dP{0, t)
as Ai 0.
dt
This simple first order diiferential equation for P(0, t) can be solved to
give
m t) =
using the initial condition P(0,0) = 1.
A similar differential equation for P(l, i) can be obtained by noting
that
P(l, i-f Ai) = P(l, t)P(0, Ai) + P(0, t)P{l At)
= P ( l,í ) ( l - M í ) + ^"^'/lAí.
This gives
P (l,í + A í)-P (l,í) _ d P (lt)
At dt
= - X P { l t ) + Á e-^\
P{r, t) = (r = 0, 1 , 2,...).
If we put fi = h
= (average number of accidents in time i),
then we have a Poisson distribution as previously defined.
This model for the Poisson distribution depends on a random
physical mechanism called a Poisson process, further details of which
can be obtained from any book on stochastic (that is, random) processes
(see the Further Reading list).
= '■ 2 rl
r — yj
= p.
= Y, f i r — l)-\-r —2pr-j-p^}P{r)
r=0
variance ^ np = mean,
as for the Poisson distribution. Then it can be shown (see for example,
Hoel, 1962) that if n ^ oo and p ^ 0 in such a way that the mean,
p = np, stays fixed, then the binomial probabilities tend to the Poisson
probabilities. This is of practical importance because it is much easier
to calculate Poisson probabilities than the corresponding binomial
probabilities. Generally speaking if n is large (>20) and the mean,
p = np, is small (<5) then the approximation will be sufficiently
accurate.
72 Discrete distributions
06
0-4
(a) (b)
0-2
11 =
^ «
» . •« :* ....
04\ 00 "" 1
0 1 2 0 1 2 3 4 5
: 0-3^
0-2[' (c)
0-1 ['
ool
0
■III
1 2 3 4 5 6 7 8
Example 10
Suppose that the number of dust particles per unit volume in a certain
mine is randomly distributed with a Poisson distribution and that the
average density is p particles per litre.
A sampling apparatus collects a one-litre sample and counts the
number of particles in it. If the true value of p is six, what is the proba
bility of getting a reading less than two?
Let X = (number of particles in a one-litre sample).
= 0-0174.
= 0-366 + 0-370
= 0*736.
As n is large and p is small it is possible to approximate this binomial
distribution with a Poisson distribution which has the same mean
p = np = \.
The probability that the cable has r defective strands is given approx
imately by e~ ^/r\. Thus the probability that the cable is not defective is
given approximately by
- 0*368-hO-368
= 0*736.
The answers obtained from the binomial and Poisson distributions are
virtually identical, and since it is much easier to calculate e~^ than
0-99100^it is preferable to use the Poisson distribution approximation.
Example 12
The number of cosmic particles striking an apparatus in forty con
secutive periods of one minute were given in Example 2, Chapter 1.
On the assumption that cosmic particles arrive randomly at a constant
74 Discrete distributions
over-all rate, one expects the Poisson distribution to describe the
frequency distribution of the number of particles arriving in a one-
minute period.
The total number of particles observed is
13 + (2x8) + (3x5) + ( 4xl ) - 48.
(The average number of particles observed per minute) = ^ = 1-2.
The Poisson distribution with this mean is given by
P{r) = e- (r = 0, l , 2,...)
rl
= (probability of observing r particles in a one-minute period if
^ = 1 -2).
These probabilities are tabulated in column 3 of Table 7.
Table 7
0 13 0-301 12-0
1 13 0-361 14-4
2 8 0-216 8-7
3 5 0-087 3-5
4 1 0-026 1-0
5-h 0 0-011 0-4
P{x,y) = P{X = x , Y = y ) .
Y.P(x,y) = L
x,y
Px(x) = X >')’
Py(y) = Y.P(x,y)
X
P{x\y)=Px{x)
and then the knowledge that a particular value of 7 has been observed
does not affect the probability of observing a particular value of X.
The conditional distribution of X is then the same for all values of 7
(and vice versa).
76 Discrete distributions
E x a m p l e 13
Marginal
distribution
0 1 2 of Y
1 3
0 i F1 F
1
y 1 1 0 ïï3
2 1 1 0 1
F
Marginal
1
distribution of X 2
1
4 41
The grand total of the joint probabilities is one. The row totals form
the marginal distribution of Y ; the column totals form the marginal
distribution of X.
If y = 0 the conditional distribution of X is given by
P(0\0) = ^ = 1/3,
n m = ^ = 1/3,
P(2|0) = I ? = 1/3.
77 Exercises
For one particular process it has been found in the past that 2 per cent
of the packets are underweight. An inspector takes a random sample of
ten packets. Calculate
(a) the expected number of packets in the sample which are under
weight,
(b) the probability that none of the packets in the sample are under
weight.
(c) the probability that more than one of the packets in the sample is
underweight.
3. It has been found in the past that 4 per cent of the screws produced
in a certain factory are defective. A sample of ten is drawn randomly
from each hour’s production and the number of defectives is noted.
In what fraction of these hourly samples would there be at least two
defectives? What doubts would you have if a particular sample con
tained six defectives?
4. One per cent of a certain type of car has a defective tail light. How
many cars must be inspected in order to have a better than even chance
of finding a defective tail light?
78 Discrete distributions
decision? If the aircraft was a three-engine jet that will fly on one engine
would you agree? Comment on the comparative reliability of the three-
and four-engine Jets in this situation.
7. The new ‘Tigercat’ sports car has an idle loping problem as about
10 per cent of the Tigercats have an unstable fluctuating engine speed
when they are idling. An engineering ‘fix’ is put in a production pilot
lot of a hundred cars.
(a) If the fix has no eflfect on the problem, how many cars would you
expect to have the fault?
(b) If only two of the pilot lot have the idling fault, and the other
ninety-eight cars are not defective, would you conclude that the fix has
a significant effect? (H int: Show that the probability of getting two or
less defectives in a sample size of a hundred, given that the fix has no
effect, is very small indeed.)
10. It has been found in the past that one per cent of electronic com
ponents produced in a certain factory are defective. A sample of size one
hundred is drawn from each day’s production. What is the probability
of getting no defective components by using
(a) the binomial distribution,
(b) the Poisson distribution approximation.
79 Exercises
11. A construction company has a large fleet of bulldozers. The average
number inoperative at a morning inspection due to breakdowns is two.
Two standby bulldozers are available. If a bulldozer can always be
mended within twenty-four hours of the morning inspection find the
probability that at any one inspection
(a) no standby bulldozers will be required,
(b) the number of standby bulldozers will be insufficient.
12. There are many other discrete distributions apart from the binomial
and Poisson distributions. For example the hypergeometric distribution
arises as follows. A sample size n is drawn without replacement from a
finite population size N which contains Np successes and N ( \ —p)
failures. Show that the probability of getting exactly x successes in the
sample is given by
N(1 -
P(x) = (x = 0, 1 , 2,..., minimum of n and Np).
Reference
H o e l , P. G. (1962), Introduction to Mathematical Statistics, W iley, 3rd edn, (4th edn
published in 1971).
80 Discrete distributions
Chapter 5
Continuous distributions
5.1 Definitions
In Chapter 4 we considered discrete distributions where the random
variable can only take a discrete set of values. In this chapter we con
sider continuous distributions where the random variable can take
any value in some specified interval.
In Section 2.2 we described how observations on a continuous
variate can be plotted as a histogj:am. As more and more observations
are taken, and the class interval is made smaller, the histogram tends
to a smooth curve called a frequency curve.
J
00
f(x ) d x = 1 .
81 Definitions
It is important to realize that f(x) is not the probability of observing
X. When the variate is continuous we can only find the probability of
observing a value in a certain range. For example, / ( xq)Ax is the
probability of observing a value between Xq and X q + Ax. In other
words it is the area of the shaded strip in Figure 18. Also note that
the probability of observing a single value is zero.
Figure 18
Figure 19
82 Continuous distributions
Example 1
A random variable is said to be uniformly distributed between a and b
if it is equally likely to occur anywhere between a and b. Thus the height
of the probability curve of the uniform distribution is a constant
between a and b. Elsewhere the probability density function is zero.
f{x}-
The total area under the curve is equal to C{b —a). But this is equal to
unity by definition. Thus we have
C =
ib-aY
a < X < b,
elsewhere.
83 Definitions
For discrete distributions it is a step function which increases from
zero to one. However, from a practical point of view, the function is
more useful for problems involving continuous variables, which are
the concern of this chapter. Since, for continuous distributions, the
probability of observing a single value is zero, it is immaterial whether
we write P ( X ^ x) or P {X < x) in the continuous case.
Figure 21
F{xo) = J/(.v)i/x
—X
84 Continuous distributions
The function must increase from zero to one since we have
F (-o o ) - 0
and
F(+ oo) = 1.
It is often S-shaped as in Figure 22.
Example 2
Find the cumulative distribution function of the uniform distribution
(see Example 1). The random variable cannot take a value less than a.
Figure 23
Thus we have
F{x) = 0 X < a.
F{x) = 1 X > h.
x —a
F(x) =
b —a
85 Definitions
Thus we have
0 X < a,
x —a
F{x) = a^ X h,
b -a
1 X > b.
This function is illustrated below.
=J
XO
F{xo) f(x ) dx
and
dF{x)
f{y^)
dx
(Mathematical note: F{x) may not be differentiable at isolated points
and the p.d.f. is not defined at these points.)
86 Continuous distributions
00
{x -n ff(x )d x . 5.2
- i
Example 3
Find the mean and variance of the uniform distribution which has
p.d.f.
—1 < X < 1,
f{x) =
elsewhere.
By inspection the distribution is symmetric about the point x = 0,
which must therefore be the mean of the distribution. This can be
confirmed by equation 5.1 which gives
1
E(X) = J x j d x = 0.
J x ^ jd x = y.
where /i, a are parameters such that —oo < < oo and (T > 0.
J dz — yj2n
88 Continuous distributions
f{>()
The shape of the normal curve depends on the standard deviation, rx,
since the larger this is, the more spread out the distribution will be.
Figure 26 shows two normal distributions, both of which have mean
zero. However one has a = 1 and the other has cr = 3.
Whatever the values of and rr, the normal distribution is such that
about one observation in three will lie more than one standard deviation
from the mean, and about one observation in twenty will lie more than
two standard deviations from the mean. Less than one observation in
300 will lie more than three standard deviations from the mean.
In practical problems, the height of the normal curve is of little
direct interest, and instead interest centres on the cumulative distribu
tion function. For example, if the weights of a batch of screws are
known to be approximately normally distributed with mean 2T0 grams
and standard deviation 0T5 grams, we might want to find the propor
tion of screws which weigh less than 2*55 grams. This can be found from
the cumulative distribution function.
F(x) = probability (X ^ x)
1
e-*“-") du
J{2n)(x
= probability
= probability (X ^ ju + za).
The full table of F{z) is given in Table 1, Appendix B. Some particularly
useful values are also given in Table 8.
Table 8
z 0 1 2 2-5 3-0
90 Continuous distributions
5.3.1 Notation
If a random variable, X, is normally distributed with mean and
variance cr^, we write
X is N{fi,
The cumulative distribution function of the standard normal distribu
tion, N{0,1), is often denoted by the special notation 0(z) to distinguish
it from the c.d.f. of any other normal distribution. However in this
book we will use the notation F(z) on the understanding that the c.d.f.
of a non-standard normal distribution will be described by
probability(A" ^ x) and not F(x).
Example 4
Find the probability that an observation from a normal distribution
will be more than two standard deviations from the mean.
If X is N{p, then we want to find
probability [X > + 2cr) +probability (X < p —2a\
By making the transformation Z = ( X - /i)/(7 , it can be seen that this
probability is the same as
probability (Z > 2) + probability (Z < —2).
But from Table 8 we have
F(2) = probability (Z 2)
= 0-977
therefore probability (Z > 2) = 0 023.
By symmetry (see Figure 27), this is equal to probability (Z < —2).
Example 5
The individual weights of a batch of screws are normally distributed
with mean p = 2-10 grams and standard deviation g = 0-15 grams.
What proportion of the screws weigh more than 2-55 grams?
In order to evaluate probability(X > 2-55) we make the transforma
tion Z = (X —p)la. This gives
/Z -2 -1 0 2-55-2-10
probability (X > 2-55) = probability I— —
= probability (Z > 3)
= 1—probability (Z ^ 3)
= 1-F (3)
= 1-0-9987 (from Table 8)
= 0-0013.
Thus 0-13 per cent of the screws weigh more than 2-55 grams.
92 Continuous distributions
Indeed non-normality is so rare that it is a useful clue when it does
occur. Particular care is required when describing non-normal distribu
tions particularly if the distribution is skewed.
(2) Some physical phenomena are not normally distributed but can
be transformed to normality. For example the fatigue lives of a batch
of electrical motors give a distribution which is skewed to the right.
However if log(fatigue life) is plotted, then the distribution will be
closer to a normal distribution.
(3) It can be shown that the normal distribution is a good approxima
tion to the binomial distribution for large n, provided that p is not close
to 0 or 1. (For example if n > 20 the approximation is valid for
0*3 < p < 0-7. For larger values of n a wider range for p is permissible).
For a given binomial distribution, the corresponding normal distribu
tion is found by putting
p = np and <j^ = np{ \ —p).
X -24
=P > -1*33
Example 7
A die is tossed 120 times. Find the probability that a Tour’ will turn
up less than fifteen times.
Let X = number of Tours’ which turn up. This random variable will
follow a binomial distribution with parameters n = 120 and p = ¿.
l)r/5\ 120-r
Thus P{X < 15) = 2
6 6
which is very tedious to sum.
However, as n is large and p = ^ we can use the normal approxima
tion with
p = np = 20
= np(l —p) = 16-67
(7 = 4-08.
94 Continuous distributions
Now for the discrete binomial distribution we have
P(X ^ 14) = P{X < 15),
which is not of course true for the continuous normal approximation.
We compromise by finding P{X < 14|) with the normal approxima
tion. (This is often called the continuity correction.)
1 4 Í-2 0
P{X < 14i) = P p . <
408 ' 4-08
= P(Z < -1-35)
= F (-l-3 5 )
= l - F ( + l-35)
= 0*0885 from Table 1, Appendix B.
Xi ^ X2 ^ ^
At each point calculate the observed cumulative frequency
number of observations ^ x,
P M = ------------------
n +n \----------------
n -\-1
96 Continuous distributions
Note that there is a mathematical reason for putting (n+ 1) rather than
n in the denominator. This also makes it possible to plot the point
P(x„) = n/(n+ 1) on the graph.
After choosing a suitable scale for the variate, x, the values of
P(x,), (/ = 1 to n), can be plotted on the graph paper. If the data really
is normal, the points will lie approximately on a straight line. Con
versely if the data is not normal, as for example if the distribution is
skewed, then the points will not lie on a straight line.
If the data does appear to be normal, the mean of the distribution
can be estimated by fitting a straight line to the data and finding the
value of X whose estimated cumulative frequency is 50 per cent. The
standard deviation of the distribution can be estimated by finding the
difference between the two values of x whose estimated cumulative
frequencies are 50 per cent and 84 per cent respectively.
Example 8
The cumulative frequencies of the data from Example 1, Chapter 1,
are given in Table 9.
Table 9
i p(^i) i P {X i) i P {X i)
thrust
Figure 30
frequencies are 50 per cent and 84 per cent is 6*2. These values, which
are estimates of the mean and standard deviation of the distribution,
are very close to the sample mean and sample standard deviation as
calculated in Chapter 2.
98 Continuous distributions
x >0 where the parameter A > 0
fix ) =
10 X < 0.
A typical example of an exponential distribution is given iji Figure 31.
It is easy to show that the total area under the exponential curve is
equal to unity. The probability of observing a value between a and b
is given by
D
)ie~ dx = e —e- X b
i
The cumulative distribution function of the exponential distribution
is given by
X ^ 0,
F{x) =
10 X < 0.
E{X)
=i xX e dx =
variance(X) = e Ix —^
= J dx
0
Example 9
The lifetime of a certain electronic component is known to be exponen>
tially distributed with a mean life of 100 hours. What proportion of
such components will fail before 50 hours?
As the mean life is 100 hours, we have X = 1/100.
Thus the c.d.f. of the distribution of lifetimes is given by
F(t) =
The probability that a component will fail before 50 hours is given by
F(50) = 1-¿>-50/100
= 0*393.
5.6.3 Applications
The exponential distribution is sometimes used to describe the distribu
tion of failure times when studying the reliability of a product. However
the reader is warned that many products are such that the longer they
survive the more likely they are to fail, in which case the exponential
distribution will not apply. An alternative continuous distribution, called
the Weibull distribution, will be introduced in Chapter 13.
The exponential distribution is also useful in the study of queueing
theory, and a brief introduction to this subject is given here. Many
situations can be described as a queueing process; this makes it
possible to tackle problems in areas such as production engineering
and road traffic flow. Two familiar examples of queues are customers
making purchases at a shop, and calls arriving at a telephone exchange
which only has a limited number of lines. A typical industrial situation
occurs when one or more mechanics are available to repair machine
breakdowns. The mechanic can be thought of as the server and a dis
abled machine as a customer. At a given time there may be several
machines being repaired or waiting to be repaired, and these machines
form the queue.
The simplest type of queue is a single-server queue, in which
'customers’ arrive randomly at a single service station and receive
attention as soon as all previous arrivals have been serviced. As
customers arrive randomly, the number of arrivals in a given time-
interval is a Poisson random variable. If the average arrival rate is
denoted by and Ai is a very small time interval, we have
probability(new arrival between t and t + At) = M i
and this is independent of previous arrivals.
The service time is occasionally constant, but more commonly it is
a random variable and several distributions have been proposed to
describe the distribution of service times. It is often found that the
service times are approximately exponentially distributed. Denoting
the exponential parameter by p, the average service time is equal to
l/p. The number of service completions in a given time interval is then
a Poisson random variable, and so the probability that a service
terminates between t and i -h Ai is given by pAt which is independent of
previous service completions.
= Po(i)(l-AAi) + Pi(i)M i.
This equation can be rearranged to give
P o(i + A i ) - P o ( i )
Ai
-XPo(t)+fiPM
P^ = limit Pr(t)
t -*oo
(r = 0, l , 2,...).
.=0
and the probability that the server is busy is given by
J
00 00
j f i x ,y ) d x d y = \.
ou
fxM = J
-X
f(x, y) dy,
X
M y) = J f i x , y) dx
Exercises
1. Find the area under the standard normal curve (/â = 0, a = 1 )
(a) outside the interval ( —1, + 1 ),
(b) between —0-5 and -h0*5,
(c) to the right of 1 *8.
3. The mean weight of 500 students at a certain college is 150 lb and the
standard deviation 15 lb. Assuming that the weights are approximately
normally distributed, estimate the proportion of students who weigh
(a) between 120 and 1801b,
(b) more than 180 lb.
105 Exercises
Chapter 6
Estimation
106 Estimation
a suitable mathematical model is that each observation is randomly
selected from a normal distribution of mean /i and standard deviation
o. Then the sample mean x is an intuitive estimate of the population
mean and the sample standard deviation,
s =>
108 Estimation
probability density function. The summation (or integral) is taken over
all possible combinations of x and y.
In proving the following results we only consider the discrete case.
The proof in the continuous case follows immediately. The quantities
b and c are constants.
6.2.1 If X is any random variable then
E{bX + c) = bE(X) + c. 6.1
Proof
= ft X ^ iP i + C S Pi
i= 1 i= 1
= bE{X)-^c.
= X xP(x, y)+ X y)
i:P ix ,y) + l l y l P i x , y )
= J]xPxix)+ Y^yPriy),
= \^Y.^Pxix)j\YyPy{y)j
= EiX)E{Y).
This result is one of the many reasons why statisticians prefer the
variance (or its square root the standard deviation) as a measure of the
spread of a distribution in most situations.
Variance (X + 7) = E[{X + Y - p ^ - p 2 f ]
= E[(X - p ^ f ] + E [ ( Y -p 2 f] + 2E[(X - p ,){Y-P2)1
110 Estimation
But the third term in this expression is given by
= 0.
Thus variance (X + 7) = variance (A^)-h variance (7).
This result can be extended to any number of independent random
variables. We find
Proof
(i7) The mean. We want to find the average value of the sample mean, x,
in a long series of trials. It is convenient to use the concept of the
expected value. By definition we know that the expected value of any
observation is equal to the population mean jx.
1 +X2+.
Thus E(x) = E
1 n
=E p +£ — using 6.2
1 \ «;
, ,
=- +- using 6.1
n n n
= P-
Thus the mean value of the sampling distribution of x is equal to //.
Note that when we write E{x) or E(xi\ the quantity in the brackets is
a random variable and not a particular sample value.
(b) The standard error.
112 Estimation
Thus variance (x) = \ . -j-a
a
n
Example 1
The percentage of copper in a certain chemical is to be estimated by
taking a series of measurements on small random quantities of the
chemical and using the sample mean percentage to estimate the true
percentage. From previous experience individual measurements of this
type are known to have no systematic error and to have a standard
deviation of 2 per cent. How many measurements must be made so that
the standard error of the estimated percentage is less than 0-6 per cent?
Assume that n measurements are made. The standard error of the
sample mean will then be l/^Jn per cent. If the required precision is
achieved we must have
Theorem 2
If random samples, size n, are taken from a normal distribution, the
sampling distribution of x will also be normal, with the same mean and
a standard error given by a/^n.
What happens if the parent population is not normal? The rather
surprising result is that provided reasonably large samples are taken
Example 2
The diameters of shafts made by a certain manufacturing process are
known to be normally distributed with mean 2-500 cm and standard
deviation 0 009 cm. What is the distribution of the sample mean of
nine such diameters selected at random? Calculate the proportion of
such sample means which can be expected to exceed 2-505 cm.
From Theorem 2 the sampling distribution of x will also be normal
with the same mean 2-500 cm but with a standard deviation (or standard
error) equal to 0-009/^9 = 0-003 cm.
In order to calculate probability (x > 2-505) we standardize in the
usual way to obtain
f x - 2-500
2-505-2-500’
probability
L 0-003 0-003
X - 2-500
= probability > 1-66
0-003
Example 3
The sampling distribution of x is illustrated by constructing a distribu
tion from successive samples of ten random numbers. The mean of
114 Estimation
each sample is calculated and the frequency distribution of the sample
means is obtained.
In this example, a random number is such that it is equally likely
to be any integer between zero and nine. Thus the population proba
bility distribution is given by
Pr = ^ ( r - 0, 1 , . . . , 8,9).
This distribution is sometimes called the discrete uniform distribution.
Random numbers of this type have been tabulated in a number of
books, or can easily be generated on a computer. Alternatively they
can be generated with the simple device shown in Figure 32.
116 Estimation
2 ^
n- 1 ■
and we say that the estimate of has (n— l)d.f. We will say more
about this in section 6.5.
The p.d.f. of the distribution is too complicated to give here, but
the percentage points of the distribution are tabulated in Appendix B,
Table 3, for dififerent values of v. The distribution is always skewed to
the right and has a mean value equal to the number of degrees of
freedom. For large values of v, the distribution tends towards the
normal distribution.
An example of a x^ distribution is shown in Figure 33. The per
centage point Xa,v is chosen so that the proportion of the distribution,
with Vd.f., which lies above it, is equal to a.
As =
(n - l)s^
we have E = n -l
E(fi) = fi.
E(x) = n.
var(/i2
var(/ii
118 Estimation
But it is important to realize that there are many good estimators
which are not unbiased. An estimator with a small bias and a small
standard error may be better than an unbiased estimator with a large
standard error. The idea of efficiency can be extended to consider biased
estimators by calculating the expected mean square error, £(/i —
This is equal to the variance of fi only when fl is unbiased. The relative
efficiency of with respect to /¿2 is given quite generally by
2*
unbiased biased
inefficient
Example 4
If a random sample of n observations is taken from N{p, a^), two
possible estimates of p are the sample mean and the sample median.
We know
var (x) = — .
n
71(7
var(median) =
2n
Thus the efficiency of the median is 2/ tz or about 64 per cent. Thus it
is better to use the sample mean.
Note that for skewed distributions the median will give a biased
estimate of the population mean. Moreover this bias does not tend
to zero as the sample size increases and so the median is not a consistent
estimate of the mean for skewed distributions.
We now state the following important results. If a random sample,
size n, is taken from a normal distribution with mean p and variance
(7^ then
E{s^) =
n —l
120 Estimation
in the denominator of the formula for 5^, then we will get a biased
estimate of This fact, which puzzles many students, arises because
the sample values are ‘closer’ on average to x than they are to jn. How
ever, in the rare case where the population mean jj. is known, but the
variance is not, the unbiased estimate of would indeed be
—jiif/n, with n in the denominator. The denominator in the
formula for is called the number of degrees of freedom of the estimate.
When fi is unknown and the observations are compared with x, there
is one linear constraint on the values of (xj —x), since ^(x^ —x) = 0, and
so one degree of freedom is ‘lost’.
More generally in Chapter 10 we will see that the number of degrees
of freedom can be thought of as the number of independent comparisons
available. With a random sample size n, each observation can be
independently compared with the (n—1) other observations and so
there are {n — 1)d.f. This is the same as the number of degrees of freedom
of the estimate ^(x, —x)^/(n—1). If in addition we know the true mean
fi, then each of the n observations can be independently compared with
jLi and so there are n d.f. This is the same as the number of degrees of
freedom of the estimate ^(x, —
The fact that x and are minimum variance estimates of ¡li and
requires a mathematical proof which is beyond the scope of this book.
^ik = E(X^)
Example 5
Suppose we suspect that a variable has a distribution with p.d.f.
i(fc+l)x‘ (0 < x < l , k > -l),
f(x ) =
(otherwise).
and that the following values are observed.
0-2 0-4 0-5 0-7 0-8 a -8 0-9 0-9.
Estimate the unknown parameter k.
As there is only one unknown parameter, the method of moments
consists simply of equating the sample mean with the population mean.
1
The population mean = j (k-\-l)x^x dx
0
_ (fc+l)
(k + 2)'
Thus the estimate of k is given by
( ^ + 1)
= 0-65
(k + 2)
giving
ic = 0-86.
122 Estimation
6.6.2 The method of maximum likelihood
This second method of estimation gives estimates which, besides being
consistent have the valuable property th at for large m they are the
most efficient. However, the estimates may be biased as may those
obtained by the method of moments.
We introduce the idea of the likelihood function by means of an
example. Suppose that we want to estimate the binomial parameter p,
the probability of a success in a single trial, and that ten trials are
performed resulting in seven successes. We know that the probability
of observing x successes in ten trials is This is a
probability function in which p is fixed and x varies between 0 and 10.
However, in this particular experiment we observe x = 7 and would
like to estimate p. If the value x = 7 is inserted in the probability
function, it can now be thought of as a function of p in which case it
is called a likelihood function. We write
L {p )= ^ ^ C ,p f\-p )\
The value of p, say p, which maximizes this function, is called the
maximum likelihood estimate. Thus the method of maximum likeli
hood selects the value of the unknown parameter for which the proba
bility of obtaining the observed data is a maximum. In other words
the method selects the ‘most likely’ value for p. The true value of p
is of course fixed and so it may be a trifle misleading to think of p as a
variable in the likelihood function. This difficulty can be avoided by
calling the likelihood a function of the unknown parameter.
In order to maximize the likelihood function it is convenient to take
logs, thus giving
log^L(p) = lo g /° C 7+ 71ogeP+31og^(l-p)
^ log L(p) _ 7 ___ ^
dp p l-p'
The likelihood function is maximized when [d log L{p)]/dp is zero. Thus
we have
= 0
P (1 -p )
giving
fi = X
.2 _
52
n -l
is preferred for small samples. (Actually is the marginal likelihood
estimate of cr^, but this method of estimation will not be discussed in
this book.)
Example 6
The Poisson distribution is given by
P.= (r = 0, 1 , 2 ,...).
rl
n‘
1= 1 X.!
(where FI denotes a product).
124 Estimation
Thus the likelihood function is given by
i-w = f ] -
X , !I ’
+ ... + ( - A+ x„ log^A-log<,x„!)
Example 7
Find the maximum likelihood estimate for the problem in Example 5.
For a particular value of /c, the joint probability density function
of observing a series of values Xj, X2, . . . , Xg is given by
(/c -h 1)x^i X (/c -f 1)X2 X ... X (/c + 1)Xg .
L (/ c) = (/ c + 1 ) « ( x , ) W - - ( ^ 8 ) ' ‘
¿ lo g ,L W 8 , ,
126 Estimation
and
Example 8
If the sample mean, w, of the twelve measurements in Example 1 were
found to be 12*91 per cent, give a 95 per cent confidence interval for the
true percentage w.
The 95 per cent confidence interval for w is given by
= 12*91 + 1*14.
By a similar argument we can obtain the confidence intervals
corresponding to stronger or weaker degrees of confidence. For
3C+ 2-58
Figure 35
128 Estimation
If we denote the 95 per cent confidence interval for //, for a sample
size n, by
x + t. 6.8
Table 10
Values of for a
95 per cent Confidence Interval
4 3-18
8 2-36
12 2-20
20 209
00 1*96
Example 9
A sample of eight observations is taken from a distribution with
unknown mean p. The sample mean x is 7*91 and the sample standard
deviation s is 0-67. Thus an estimate of the standard error of the sample
mean is 0 67/^8 = 0*237. From Table 10 we find that for a sample
size 8 the 95 per cent confidence interval for the true mean p is given
by x + 2-36 x s/^n.
Thus the 95 per cent confidence interval is given by 7*91 ±0*56.
In equation 6.8 is actually the percentage point of a distribution
called the i-distribution. We have already seen that if random samples,
size n, are drawn from a normal distribution, mean p and variance
then the sampling distribution of the random variable
X — p .
z = — ^ IS A(0, 1 ).
(T/^n
t =
Figure 36
130 Estimation
We are now in a position to find a confidence interval for fx for any
degree of confidence we wish. If x and are calculated from a sample
size n, the 100(1 —a) per cent confidence interval for ¡j. is given by
Example 10
Use the data given in Example 9 to establish a 99 per cent confidence
interval for the true mean p.
There are eight observations, so that the sample estimate s is based
on (8-1) d.f. From Table 2, Appendix B, we find
^0 005,7 “ 3‘50.
From Example 9 we have s /^ n = 0-237. Thus the 99 per cent
confidence interval for p is given by
7-91 ±3-50x0-237 = 7-91 ±0-83.
This is about half as wide again as the 95 per cent confidence interval
for p.
Exercises
1. A random sample is drawn from a population with a known standard
deviation of 2 0. Find the standard error of the sample mean if the
sample is of size (a) 9; (b) 25 and (c) 100. What sample size would give
a standard error equal to 0*5?
2. If X, Y are independent random variables show that
variance (X —Y) = variance (X) +variance (7).
Two random samples of size and «2 are taken from a population
with variance The two sample means are and X2 - Show that
the standard error of Xj —X2 is given by
1 1
Ml tl2
3. A random sample size 10 is taken from a standard normal distribu
tion. Find the values a and b such that
P(a < x) = 0*95
P(x < b) = 0-95
i.e.
P(a < x < b) = 0*90.
Also use the distribution to find the values c and d such that
P(c < s^) = 0-95
P(s^ < d) = 0-95
i.e.
P{c < s^ < d) = 0-90.
Also find the values of a, b, c and d if the sample was taken from a
normal distribution with a mean of six and a variance of four.
132 Estimation
4. The probability density function of the exponential distribution is
given by
f(x) = Àe~ X > 0.
A random sample of n observations is taken from this distribution.
Show that the method of moments estimate of A is given by
1
X = - where x is the sample mean,
v
Also show that the maximum likelihood estimate of a is the same as
the method of moments estimate.
5. The percentage of copper in a certain chemical is measured six times.
The standard deviation of repeated measurements is known to be 2*5:
the sample mean is 141. Give a 95 per cent confidence interval for the
true percentage of copper, assuming that the observations are approxi
mately normally distributed.
6. The percentage of copper in a certain chemical is measured six times.
The sample mean and sample standard deviation are found to be 14T
and 2*1 respectively. Give a 95 per cent confidence interval for the true
percentage of copper, assuming that the observations are approxi
mately normally distributed.
7. A random sample, size «i, is taken from and a second
random sample, size « 2, is taken from If the sample means
are denoted by x^, x 2 it can be shown that
- w o ,I ),
(X i-X :
where 5 denotes the combined estimate of a from the two samples (see
section 7.3.2).
133 Exercises
Chapter 7
Significance tests
7.1 Introduction
We have seen that statistics is concerned with making deductions from
a sample of data about the population from which the sample is drawn.
In Chapter 6 we considered the problem of estimating the unknown
parameters of the population. In this chapter we shall see how to carry
out a significance test in order to test some theory about the population.
The commonest situation is when the population can be described
by a probability distribution which depends on a single unknown
parameter fi. On the basis of the experimental results we might wish,
for example, to accept or reject the theory that ¡ has a particular value
jl
Example 1
One of the commonest problems facing the engineer is that of trying
to improve an industrial process in some respect. In particular he may
wish to compare a new process with an existing process. However,
because of changeover costs, he will need to be reasonably certain that
the new process really is better before making the necessary changes.
As an example let us suppose that the strength of steel wire made by
an existing process is normally distributed with mean = 1250 and
standard deviation a = 150. A batch of wire is made by a new process,
and a sample of 25 measurements gives an average strength of 3c = 1312.
(We assume that the standard deviation of the measurements does not
change.) Then the engineer must decide if the difference between x and
Pq is strong enough evidence to justify changing to the new process.
The method of testing the data to see if the results are considerably or
significantly better is called a significance test.
Example 2
If a new industrial process is compared with an existing process, the
choice of the null hypothesis depends on a number of considerations and
several possibilities arise.
(a) If the existing process is reliable and changeover costs are high, the
burden of proof is on the new process to show that it really is better.
Then we choose
H q: existing process is as good as or better than the new process,
and H^ : new process is an improvement.
This is the situation discussed in Example 1.
(b) If the existing process is unreliable and changeover costs are low,
the burden of proof is on the existing process. Then we choose
135 Introduction
H q: new process is as good as or better than the existing process.
and : existing process is better than the new process.
Example 3
A new drug is tested to see if it is effective in curing a certain disease.
The Thalidomide tragedy has emphasized that a drug must not be put
on the market until it has been rigorously tested. Thus we must assume
that the drug is not effective, or is actually harmful, until the tests
indicate otherwise. The null hypothesis is that the drug is not effective.
The alternative hypothesis is that the drug is effective.
A second factor to bear in mind when choosing the null hypothesis is
that it should nearly always be precise, or be easily reduced to a precise
hypothesis. For example when testing
H q: Po
against H i'. p > po,
the null hypothesis does not specify the value of p exactly and so is not
precise. But in practice we would proceed as if we were testing
H q\ p = Po
against H^ \ p >Po
z= ■/^O
r/V25-
1312-1250
Zq = . = 2-06.
150/725
137 Introduction
7.1.3 The interpretation of a significant result
Example 4
According to a certain chemical theory, the percentage of iron in a
certain compound should be 12T. In order to test this theory it was
decided to analyse nine different samples of the compound to see if the
measurements would differ significantly from 12T per cent.
Before carrying out the analyses, we must specify the null and alterna
tive hypotheses and also decide if a one-tailed or two-tailed test is
appropriate.
Denote the unknown mean percentage of iron by p. Then the null
hypothesis is given by
12 *1 %.
The alternative hypothesis is given by
12 -1 %.
As we are interested in significantly high or low results, a two-tailed
test is appropriate. The actual measurements and the resulting analysis
will be given later in the chapter.
For tests based on the normal distribution, the level of significance
of a result in a two-tailed test can be obtained by doubling the level of
significance which would be obtained if a one-tailed test was carried
out on the same result (see section 7.2).
139 Introduction
level of significance is very small, then a decision can be made to reject
the null hypothesis. If very large then a decision can be made to accept
the null hypothesis. But if the observed level of significance is around
5 per cent, the results are rather inconclusive and the experimenter
may decide to take some more measurements.
However, if a definite decision has to be taken, the experimenter
should choose a particular level of significance and reject the null
hypothesis if the observed level of significance is less than this. The
5 per cent and the 1 per cent levels are commonly used. This value
should be chosen before the observations are taken. For example, if
the 5 per cent level is chosen and the observed level of significance is
less than this, then we say the result is significant at the 5 per cent level.
A critical value of the test statistic can be calculated which will
correspond to the chosen significance level. Thus in Example 1 we
could choose to reject H q if the observed level of significance is less than
5 per cent. But probability (z > 1*64) = 0 05. Thus the critical value of
the test statistic, z, is 1-64. If the observed value of z is greater than 1-64,
then H q must be rejected.
A variety of significance tests exist each of which is appropriate to a
particular situation. In the remainder of this chapter we will describe
some of the more important tests.
7.2.1 (7 known
This situation has been discussed in detail in Example 1. The test
statistic is given by
z=
<j/yjn'
(b)
zo
/Vi :m <M o
Table 11
T w o-tailed O n e-tailed
-^ 0
t =
s/J n ■
(ii) Small samples - the t-test. When n is less than about twenty-five
the sample standard deviation is not such a good estimate of , so g
Example 4 continued
The analysis of the nine diflferent samples of the compound gave the
following results:
11-7 12*2 10-9 114 11-3 120 IM 10-7 11-6
Is the sample mean of these measurements significantly different from
12T per cent?
0-0 2 5 ,8 = 2-31.
As |iol is larger than io-o25,8? the probability of observing a result
which is as extreme or more extreme than this is certainly less than
2 X0-025 = 0-05. Thus the result is significant at the 5 per cent level.
Moreover we have io oos,8 3-36 so that the result is also significant
at the 1 per cent level. Thus we have fairly conclusive evidence that the
percentage of iron is not 12-1 per cent.
7.3.1 a known
Common sense suggests that the larger the diflference between and
X2, the less likely H q is to be true. A standardized test statistic can
be obtained by dividing (x j—X2) by its standard error, which
can be found as follows. We know variance ( x j = and variance
(X2) = (J^ln2 . In addition we know that the variance of the sum or
difference of two independent random variables is equal to the sum of
the variances of the two variables (see section 6.2). Thus
(T^(l/ni + l/n2
If H q is true it can be shown that the random variable z is a standard
normal variable. The level of significance of a particular result Zq can
be found as in section 7.2.
1.32 G unknown
Following the method adopted in section 7.2 we replace a with the
sample standard deviation, s, to obtain the test statistic
t =
s 7 (v « i+ v « 2 )'
The standard deviation of the first sample is given by
V {X i-X if
1
*^2 = V ^ 1
Then the combined unbiased estimate of is given by
2 = ...................................................
( » l - l )Sl+ («2-l)S2
«1 + « 2 - 2
X(^i
-\-ri2-2
This result follows from the fact that if X, Y are independent random
variables with , V2 degrees of freedom then X + 7 is also a x^ random
variable with (v^ + V2) degrees of freedom. Thus
( n i- l) s f («2 - 1)52
-+- IS X (ni-l) +(n2-l)
If the null hypothesis is true, it can be shown that the test statistic, i,
follows a i-distribution with 4-/12 —2) degrees of freedom. Thus the
level of significance of a particular result, ìq, can be found as before.
Example 5
Two batches of a certain chemical were delivered to a factory. For
each batch ten determinations were made of the percentage of man
ganese in the chemical. The results were as follows :
Batch 1: 3-3 3-7 3-5 44 3-4 3-5 4-0 3*8 3-2 3-7
Batch 2: 3-2 3-6 34 3-4 3-0 3-4 2-8 34 3-3 3-6
Hence
.Vi = 3-62 X, = 3-25
, 0-776
-------- = 0-0862
9
0-605
si = 0-0672.
probability > 2 88
- ) = 0 01
- .
Thus the result is significant at the 1 per cent level and we have strong
evidence that H q is untrue.
Example 6
In order to compare two methods for finding the percentage of iron in
a compound, ten different compounds were analysed by both methods
and the results are given below.
We ask the question ‘is there a significant difference between the two
methods of analysis?’
Note that it would be wrong to calculate the average percentage for
method A and for method B and proceed as in section 7.3 because
the variation between compounds will swamp any difference there
may be between the two methods. Instead we compute the difference
for each compound as below.
Compound Compound
number Difference number Difference
1 01 6 0-3
2 0-3 7 00
3 00 8 01
4 -0-2 9 01
5 0-2 10 0-4
If the two methods give similar results, the above differences should be a
sample of ten observations from a population with mean zero.
Generally we have k pairs of measurements X2 j, (j = 1, 2, . . . , k)
which are independent observations from populations with means
Pij, p 2 j' The null hypothesis is that each pair of means are equal.
Ho’ j = Pij (for all j).
t =
sJs/k'
If H q is true, the distribution of t will be a i-distribution with (k—\)
degrees of freedom, as the estimate is calculated from k differences.
Example 6 continued
d = average difference = 0T3
E
sj = ^---- = 0-031 = 0176
k —l
d
tn — = 2-33
W io
The alternative hypothesis is of the form H^: pij ^ for all j, and
so a two-tailed test is required. From Table 2, Appendix B, we find
^0 025,9 ~ 2*26.
Thus P(\t\ > 2-26) = 0 05.
As to is greater than 2-26, the result is significant at the 5 per cent level,
and so we have reasonable evidence that H q is untrue.
Theorem
If Oj, . . . , Ofc and np^,.. .,npk are the observed and expected frequencies
for the k possible outcomes of an experiment, then, for large n, the
distribution of the quantity
ft
{Oj-nPi)
2 npi
S Of = n = Y.
i= l i= l
Example 7
Assuming that a die is fair we have
Pi = probability that an i turns up
=i {i= 1,2, 3,4, 5 , 6).
02 = 26 = 21 Og = 14.
Observed Poisson
r frequency frequency
0 13 120
1 13 14-4
2 8 8-7
34 6 4-9
Zo05,2 = 5’99.
The observed value of x^ is much smaller than the critical value so we
can accept the null hypothesis that the Poisson distribution gives a
good fit.
No service Service
call required
Factory A 80 32
Factory B 63 33
Table 12
rx c Two-way Table
«1 n.2 . .. n.. n
Let
n,. = Y, ^ij = number of observations in iih row,
j
n.j = Y ^ij = number of observations in jih column.
i
Thus
«=Ii =S
j
"r
Generally speaking we are interested in testing the independence of
the two types of classification. For this reason two-way tables are
often called contingency tables because we may ask if the presence of
ni. n.j
npij = npi.p.j = n------^
n n
ri:.n.
This can be compared with the observed value ttij. The test statistic is
given by
i j
so only (r + c —2) independent estimates have to be made. Thus the
number of degrees of freedom is given by
rc - 1 - ( r - f c - 2) = ( r - l) ( c - 1 ).
We find
i'll = ^12 = 35 -22 = 30.
C2i = 66
(80-77)2 (32-35)2 (63-66)2 (33_3Q)
Thus Xo = -+ +-
77 35 66 30
= 0 -8 2 t
7.6 TheF-test
This significance test is widely used for comparing different estimates
of variance, particularly in the analysis of variance which is described
in Chapter 10.
Suppose that we have a normal distribution with variance Two
random samples, sizes and ^ 2 , are drawn from this population and
the two sample variances s\ and S2 are calculated in the usual way.
As 5? and si are both estimates of the same quantity we expect the
ratio s\ls\ to be ‘close’ to unity, provided that the samples are reasonably
large. If we take repeated pairs of samples, size and ri2 , it can be
shown that the ratio F = s\lsl will follow a distribution called the
F-distribution.
A random variable which follows the F-distribution can be obtained
from two independent x^ random variables (see Appendix A). The
distribution depends on two parameters, and V2, which are the
number of degrees of freedom of si and 52 respectively. Note that
the estimates si and s | must be independent. In the above situation
we have = n^ — l and V2 = «2 —1 .
An F-test is carried out in the following way. Let si and S2 be
independent estimates of al and al respectively, and assume that the
observations in the two samples are normally distributed. Then we are
often interested in testing the hypothesis that and S2 are both
estimates of the same variance In other words we want to test the
null hypothesis H q \ gI = al = For an alternative hypothesis of
the form H^ :al > al, a one-tailed test would be appropriate. If the
ratio sl/sl is much greater than one then we will be inclined to reject
H q. In order to see if the observed ratio sl/sl is significantly large, it
is compared with the upper percentage points of the F-distribution
which are given in Table 4, Appendix B. The point F^^ vi,v2 is the point
on the F-distribution, with Vj and V2 d.f., such that a proportion a of
Figure 39 An F-distribution
Example 10
In Example 5 we assumed that the two sample variances were estimates
of the same population variance. We can now verify this statement with
an F-test.
Let si be an estimate of , and let si be an estimate of . Then the
two hypotheses are given by
). (Jj — (72,
H^ \ o \ ^ g \.
Example 11
The ten differences between the results of method A and method B are
0-1 0-3 0-0 -0-2 0-2 0-3 0-0 0-1 0-1 0-4
Thus seven of the differences have a positive sign, one has a negative
sign and two are zero.
Applying the null hypothesis that the two methods give similar
results, the differences are equally likely to have a positive or negative
sign. If we disregard the two zero differences, the number of positive
Example 12
In Example 1 we showed that in order to test
H o:/i = 1250
against
H ^ \ p > 1250
it was necessary to consider the test statistic
X—1250
150/725 ‘
If H q is true, this is a standard normal variable and
probability (z > 1-64) = 0-05.
Thus H q is rejected at the 5 per cent level if the observed value of z is
greater than 1-64. Therefore 1*64 is the critical value of z. This is
equivalent to rejecting H q if x > 1250 + (l-64x 150)/725 = 1299-2. So
the corresponding critical value of x is 1299*2. Here we have chosen
the critical value in such a way that the probability of an error of type I
is 5 per cent (see Figure 40).
In general the probability of getting a significant result when H q is
true is denoted by a. The above choice of a = 0*05 was quite arbitrary.
If changeover costs are high, so that the experimenter requires strong
evidence of an improvement, then it would be better to choose a = 0-01.
If H q is true we have
x -1 2 5 0
probability < 2-33 = 0-99
(l50/V25
x -1 3 2 0 1319*9-1320
probability (x < 1319*9) = probability < —
30 30
Thus, with the lower level of significance, there is a much higher chance
of making an error of type II.
From Example 12 it can be seen that the two types of errors are
dependent on one another. For example, if the critical value is in
creased in order to reduce the probability of a type I error, then the
probability of a type II error will increase. A ‘good’ test of significance
is one which minimizes the probabilities of these two types of error in
some way. For a given value of a we want to choose the significance
test which minimizes jS.
An alternative way of looking at this problem is to look at the power
of the test. For a specific alternative hypothesis, the power of a
significance test is obtained by calculating the probability that H q is
rejected when it is false (that is, the correct decision is made). Then
H q: p = po
against
H^ \ p > po,
Exercises
1. Test the hypothesis that the random sample
163 Exercises
obtained from this manufacturer and a series of measurements are to be
made on the strength of different pieces of paper from the batch.
Assuming that the standard deviation of these measurements is also
3 p.s.i., how large a sample size should be chosen so as to be 95 per cent
certain of detecting a mean increase in strength of 2 p.s.i. with a one-
tailed test at the 5 per cent level?
5. For a certain chemical product it is thought that the true percentage
of phosphorus is 3 per cent. Ten analyses give x = 3*3 per cent and
s = 0-2 per cent. Is the sample mean significantly different from 3 per
cent? (This question differs from question 3 because we are interested
in departures from 3 per cent in either direction.)
6. One sample of fifteen observations has = 82 and = 5. A
second sample of ten observations taken by a different scientist has
X2 = 88 and S2 = 7. Is there a significant difference between the two
sample means at the (a) 0*05 and (b) 0 01 level of significance? (You may
assume that the two populations have equal variances.)
7. Test the hypothesis that the following set of 200 numbers are
‘random digits’, that is, each number is equally likely to be 0, 1, 2, . . . , 9.
r 0 1 2 3 4 5 6 7 8 9
Frequency 22 16 15 18 16 25 23 17 24 24
8. The following figures show the number of accidents to 647 women
in a period of five weeks while working on the manufacture of shells.
(Source: M. Greenwood and G. U. Yule, Journal of the Royal Statistical
Society, 1920.)
Number of accidents 0 6-f
Find the Poisson distribution with the same mean. Test the hypothesis
that the Poisson distribution gives a good fit to the data.
9. In order to test the effectiveness of a new drug in treating a particular
disease, seventy patients suffering from the disease were randomly
divided into two groups. The first group was treated with the drug and
the second group was treated in the standard way. The results were as
follows.
Recover Die
Drug 20 15
No drug 13 22
Addendum
S ig n ifica n ce te sts can so m e tim e s b e very h elp fu l, but I am d istu rb ed at th e p resen t
te n d e n c y o f n o n -sta tistic ia n s to o v e r d o sig n ifica n ce tests. F or e x a m p le th e re is n o p oin t
in testin g null h y p o th e s e s w h ich are o b v io u sly u n tru e, as w h en the a u th or w as rece n tly
a sk ed to a n a ly se so m e b io lo g ica l d a ta and test th e null h y p o th e sis that w ater has n o
e ffe ct o n plan t gro w th . T h e d a n g er h ere is that th e test m ay b e carried o u t in correctly
or that an in a d eq u a te sa m p le siz e m ay lea d to th e null h y p o th e sis b e in g in correctly
a ccep ted .
T h er e is a lso n o p o in t in testin g a null h y p o th e sis w h ich is o b v io u sly g o in g to b e
a cce p te d , as w h en th e au th o r w as r ece n tly a sk ed to test th e null h y p o th e sis that fx^ / iq ,
w h ere pu w as an u n k n o w n p o p u la tio n m ea n and /jlqa sp ec ific v a lu e , and th e sam p le
m ea n w as less than /jlq.
A n o th e r d a n g er w ith sig n ifica n ce te sts is that th e a ssu m p tio n s o n w h ich th e y are
b a sed m ay n o t b e sa tisfied , if for ex a m p le th e d ata are co rr e la te d , rather than
in d e p e n d e n t.
In so m e jo u rn a ls, particu larly m ed ica l jo u rn a ls, it has b e c o m e virtu ally im p o ssib le to
p resen t any resu lts w ith o u t carrying o u t a test and ‘g ivin g a P -v a lu e ’. T h is is very
d a n g ero u s. It sh o u ld b e rea lised that in g en era l (th o u g h n o t a lw ays), estim a tio n is m ore
im portan t than sig n ifica n ce testin g .
(1) A non -sign ifican t difference is n o t n ecessarily the sam e thin g as n o difference.
(2) A significant difference is not necessarily the sam e thin g as an in terestin g difference.
165 Exercises
Chapter 8
Regression and correlation
Example 1
An experiment was set up to investigate the variation of the specific
heat of a certain chemical with temperature. Two measurements of
the specific heat were taken at each of a series of temperatures. The
following results were obtained.
tS1-75
M -70
1-65
1.60t— 60 70 80 90 100
50
temperature
Our task is to find estimates of Gq and such that the line gives a good
fit to the data. One way of doing this is by the 'method of least squares'.
At any point Xj the corresponding point on the line is given by
a Q - y a ^ X i , so the difference between the observed value of y and the
predicted value is given by
ei = yi-{ao+aiXi).
Figure 42
S = t e f
i= l
i= 1
5ao
Z - «0 - a 1 ^.)( - ^i) = 0.
These can be rearranged to give
Mao + 5, Z ^ i = Z 3'.’
8.1
¿ o Z ^ i + ^ i Z ^ f = Z-x.y.-
These two simultaneous equations in Gq and aj are often called the
normal equations. They can be solved to give
«0 =
2 -^/0 / - y ) S ( x ,- x ) ( y ,- y )
«1 = since x^{yi —y) = 0.
Ex/(jC/ —x)
It is easy to see that the least squares regression line passes through
the centroid of the data, (x, y), with slope . Thus it is often convenient
to write the equation in the equivalent form
y - y = fli(x-x).
Many pocket calculators with a facility for regression use an
algebraically equivalent formula for given by
8.2
Example 2
Fit a straight line to the data of Example 1 and estimate the specific
heat when the temperature is 75°C.
By inspection we have x = 75 = average temperature.
12
By calculation we have ^ y¡ = 20T6 giving y = 1-68.
i= 1
so that if log y is plotted against log x, the points will lie on a straight
line. One advantage of such a transformation is that it is easier to fit
a straight line than any other type of curve. A more detailed account of
such transformations is given by Wetherill (1981, Chapter 8).
In general it will more often be necessary to try and fit a non-linear
curve to the data. Before describing how this is done, we will discuss
the problem of regression in more general terms.
8.3 Regression
We have seen that if several measurements are made on the dependent
variable, y, at the same value of the controlled variable, x, then the
results will form a distribution. The curve which joins the mean values
of these distributions is called the regression curve of y on x, and an
example is given in Figure 43. The problem of finding the most suitable
form of equation to predict one variable from the values of one, or more,
other variables is called the problem of regression.
Figure 43 Regression curve. The locus of the mean values of the /-distributions
171 Regression
The functional form may be selected from theoretical considerations.
For example, the experiment may have been designed to verify a
particular relationship between the variables. Alternatively the
functional form may be selected after inspecting the scatter diagram,
as it would be pointless for example to try and fit a straight line to some
data if the relationship was clearly non-linear. In practice the experi
menter may find it necessary to fit several different types of curve to the
data and to choose the one which gives the best fit (see section 8.7).
After the functional form has been selected, the next problem is to
estimate the unknown parameters of the curve. If, for example, a
quadratic relationship is thought to be appropriate, the regression
curve is given by
y = aQ-\-aiX-\-a2 X^
and then the quantities «o» ^2 n^ust be estimated from the data.
A general method of estimating the parameters of a regression curve
is by the method of least squares. We have already described how this
is done when the regression curve is a straight line. A similar technique
can be adopted with other regression curves. For example, let us
suppose that the regression curve of y on x is given by
y = aQ-\-aiX-\-a2 X^.
At any point x,, the corresponding point on the curve is given by
aQ+ aiXi~\-a2 xf, so the difference between the observed value of y and
the predicted value is
e¡ = yi-(ao + aiX i+ a2 x f).
náo + ái + = Z>'i’
áo X + ái X + ^2 Z = Z
á o Z ^ . - + ^ i Z ^ f + ^ 2 Z ^ ? = 'Ly¡^f-.
These are the normal equations for quadratic regression. They can be
solved to give the least squares estimates áo, áj and ü2 .
173 Regression
problem is considered by Weisberg (1985) Chapter 4. A somewhat
similar problem is also considered in this book in section 9.5.
_2
that variance (â,) = ^y\x
Z (x .-x )^
and that both Gq and are normally distributed, as each is a linear
combination of the observed values of y which are themselves normally
distributed. Thus in order to obtain confidence intervals for Gq, ai and
flo 4-^1 X we must first obtain an estimate of the residual variance,
The sum of squared deviations of the observed points from the
estimated regression line is given by
„2 „ Z iy i-^ O -^ lX ,)^
n -2
The denominator, n —2, shows that two degrees of freedom have been
lost. This is because the two quantities do and di were estimated from
the data, so there are two linear restrictions on the values of
y i~ — A computer will routinely calculate Sy\x, but, for a
«2 - Z y f - ^ o Z y .- ^ i Z^.>'.
n —2
Then it can be shown that the 100(1 —a) per cent confidence interval
for a 1 is given by
for Gq by
^0 ± ha.n - 2 XS,|, J ~+ _ - ^ 2]
^y\x
and follows the i-distribution with n —2 degrees of freedom if H q is true.
Example 3
In Example 2 the estimated slope of the regression line is given by
= 000226.
Is this value significantly different from zero?
33*8894- 1*510833 x 20*16-0*00225556 x 1519*9
We have =
10
= 0*00034,
therefore = 0*018,
Thus the result is significant at the 5 per cent level and so the slope of
the regression line is significantly different from zero, even though it
appears, at first sight, to be very small.
. , / fl , 1 , (^0-^ )^
ao + aiXo±t^,.„-2S,|^ ^ + -+ ~ n --------------
i= 1
This interval is often called a prediction interval.
. .............................................. , . d S d S d S , d S
This quantity can be minimized by setting - —, - —, - — and — equal
VGQ VGj yG2 CG3
to zero, to obtain four simultaneous equations in ^ 0, « 1 , «2 ^ 3*
^0« + « l Z ^ l i + « 2 E ^ 2 i + a 3 E ^ 3 i = E > 'n
+ Z ^ l i + «2 Z ^ 2i^li + «3Z^3i^li =
8.4
^ o Z ^ 2 i + « l Z ^ U ^ 2 i + ^ 2 Z ^ l i + « 3 Z ^ 3 i ^ 2 i = Z yi^2i,
^ o Z ^ 3 , + ai Z ^ l i ^ 3 . + a 2 Z ^ 2 i ^ 3 i + ^ 3 Z ^ 3 i = Z
These four equations, called the normal equations, can be solved to
give the least squares estimates of Gq, g^, G2 and g^. We shall not
discuss numerical methods for solving these equations as multiple
regression programs are now available with the majority of computers
and they have taken most of the hard work out of regression. Some
remarks on the dangers of multiple regression and the selection of
variables are given in the addendum at the end of the chapter.
y =
8.6.2 Mixtures
y =
y = ao-yaiX-\-a 2 Z-ya^xz.
i m = 0.
X
= X/k(2i)y,-
All the other terms are zero because the polynomials are orthogonal;
all summations are for i = 1 to n. The quantities given in
the above tables. Thus the only quantities which have to be calculated
are ^X(Zj)y, ; r = \ to k. This can easily be done as the numerical
values of/^(z,) are also given in Fisher and Yates (1963) and Pearson and
Hartley (1966). The estimates follow immediately.
in general if c observations on y are made for n different values of x
then the normal equations are modified by multiplying the terms on
the left hand side of the equations by c. Finally the regression equation
can be obtained as a polynomial in x by substituting
(x —x)
^ d '
Another advantage of orthogonal polynomials is that the estimates
of the regression parameters are independent. This fact is particularly
useful when the order of the polynomial is not known beforehand.
The problem then is to find the lowest order polynomial which fits the
data adequately. In this case it is best to adopt a sequential approach,
by first fitting a straight line, then fitting a quadratic curve, then a
cubic curve, and so on. At each stage it is only necessary to estimate
one additional parameter as the earlier estimates do not change.
^2 = Z [ y - ^ 0 -^'l/l(z)-^2/2(z)]^-
It can be shown that this reduction is given by
R1- R 2 =
In general, after fitting a polynomial of degree /c, the residual sum of
squares is given by
F =
R J in -k -iy
On the other hand if a polynomial of degree k does fit the data better
than a polynomial of degree (fc —1), then R^ will be substantially less
than R k -i, and the F-ratio may be significantly large.
This sequential process is continued until two non-significant
F-ratios in a row are obtained. This is necessary because even-order
polynomials may give non-significant results even though odd-order
polynomials do give significant results - and vice versa.
The method is illustrated in Example 4.
Then we find
i /i(z,));, = 52-6,
i= 1
= 44-2,
= i'4,
= 5-8,
X y = 66-1 .
Thus we have
flo = y = 9-44,
_ 52-6
= 1-878,
28
44-2
ai = — = 0-526,
^ 84
= ^ = 0037.
Table 13
Neither the cubic nor the quartic terms give a significant F-ratio.
In any case, the residual sum of squares is so small after fitting linear
and quadratic terms that it is really unnecessary to try higher order
terms in this case. Thus a quadratic polynomial describes the data
adequately.
The next stage is to compute the regression equation in terms of the
original controlled variable, x. For this we need to know the orthogonal
polynomials as functions of z. These are also given in Fisher and Yates
(1963) and Pearson and Hartley (1966). We find
/i(z) = with = 1
Example 5
The following pairs of (coded) measurements were taken of the tem
perature and thrust of a rocket engine while it was being run under the
same operating conditions. Plot the results on a scatter diagram.
X y X y X y
19 1-2 33 2-1 45 2-2
15 1-5 30 2-5 39 2-2
35 1-5 57 3-2 25 1-9
52 3-3 49 2*8 40 1-8
35 2*5 26 1-5 40 2-8
Data of this type occurs frequently in biology and the social sciences.
For example measurements are often made on two different character
istics of the same human being, such as his height and weight. Both
these variables are subject to considerable random fluctuation.
Nevertheless we expect to find some relationship between them as a
man who is taller than average is also likely to be heavier than average.
The problem in this sort of situation is to see if the two variables are
inter-related, and, if so, to find a measure of the degree of association or
correlation between them. We will only be concerned with linear
correlation, when the relationship between the variables appears to be
linear.
3
• •
• •
• •
10 20 30 40 50 60
temperature (x)
(a) (b)
(c) (d)
€ 3
• •
2
1 IV
0
10 20 30 40 50 60
temperature (x)
Figure 47
Now any point in quadrant II will be such that both (Xj —x) and
(yj —y) are positive. Thus the product (Xj —x)(yj —y) is positive. In
quadrant III both (x; —x) and (yj —y) are negative so that the product
(Xj —x)(yj —y) is positive. But any point in quadrants I or IV will be
such that the product of (x^—x) and (yj —y) is negative.
The covariance is obtained by summing all the products and dividing
by (n —l). If the variables are positively correlated, as in Figure 47,
then most of the observed points will lie in quadrants II and III, so that
the covariance will be ‘large’ and positive. On the other hand if the
variables are negatively correlated then most of the observed points will
lie in quadrants I and IV, so that the covariance will be ‘large’ and
negative. But if there is no correlation between the variables, then the
number of points in each quadrant will be approximately the same, so
that the covariance will be close to zero. Thus the covariance is a
measure of the association between the two variables.
and Sy =
n -l
Then the required measure of correlation is given by
r=
It can be shown that the value of r must lie between —1 and +1.
For r = + 1, all the observed points lie on a straight line which has a
positive slope; for r = —1, all the observed points lie on a straight line
which has a negative slope.
Example 6
Calculate the correlation coefficient of the data from Example 5.
«=15, 1276-1, ^ x = 540, ^_v = 33,
= 21426, 78-44.
Hence
88-1
r= = 0-82
7(1986x5-84)
Thus the correlation is high and positive; this was to be expected after
inspecting Figure 47.
Example 7
Is the correlation coefficient in Example 6 significantly different from
zero?
. 9-0
V d - - ’)
But ^0 025,13 “ 216.
Thus the correlation between temperature and thrust is significant
at the 5 per cent level.
Table 14 gives the 95 per cent critical points for the absolute value
of the correlation coefficient for different sample sizes. When the
sample size is small a fairly large absolute value of r is required to show
significant correlation.
Table 14
5 0-88 25 0-39
10 0-63 30 0-36
15 0-51 50 0^8
20 0-44 100 0-20
K y i-y ? ■
The two regression lines will only coincide when the observations all
lie on a straight line. Both lines pass through the centroid of the data,
but their slopes are different. If there is no correlation between the
variables then the two regression lines will be at right angles. In this case
the best prediction of y for any value of x is simply y. Conversely the best
prediction of x for any value of y is simply x.
Example 8
Find the regression lines for the data of Example 5.
We have x = 36, y = 2-2.
Z / - ^ = 5-84.
= 0*044(x-36).
88-1
^ - 3 6 = ^ ( . - 2-2)
= 15-l(y-2*2).
= r./.V,
—(x —x) using equation 8.5.
Thus
1 ^ jy-i^yV
X exp <— dx dy,
2(1 -p ^ )
where jHy, (Ty denote the mean and standard deviation of X
and Y respectively. The parameter p is the theoretical correlation
coefficient and is defined by
p = E
and variance
^)\x =
This important result, which will not be proved here, says that the
conditional means of the y-distributions lie on the straight line
<T,
• •
• •
Exercises
1. The following measurements of the specific heat of a certain chemical
were made in order to investigate the variation in specific heat with
temperature.
Temperature °C 0 10 20 30 40
Plot the points on a scatter diagram and verify that the relationship
is approximately linear. Estimate the regression line of specific heat
on temperature, and hence estimate the value of the specific heat when
the temperature is 25°C.
2. When the values of the controlled variable are equally spaced, the
calculation of the regression line can be considerably simplified by
coding the data in integers symmetrically about zero. When the values
of the controlled variable sum to zero, the estimates of the regression
coefficients are given by
^ Z ^iy¡
^0 =
197 Exercises
For example, with five equally spaced values of the controlled variable
the coded values could be - 2, - 1, 0, + 1, + 2 ; with six values the
coded measurements could be —5, —3, —1, + 1, 3, 5.
Use this technique to find the regression line in the following case.
Height (inches) 63 71 72 68 75 66 68 76 71 70
Weight (pounds) 145 158 156 148 163 155 153 158 150 154
Test Xi X2
1 1-6 1 1
2 21 1 2
3 2-4 2 1
4 2-8 2 2
5 3-6 2 3
6 3-8 3 2
7 4-3 2 4
8 4.9 4 2
9 5-7 4 3
10 50 3 4
Addendum
199 Exercises
Part three
A p p lic a tio n s
The next three chapters will be concerned with the design and analysis
of experiments. This chapter, entitled the planning of experiments,
deals with preliminary considerations, particularly that of finding the
precision of the response variable. Chapter 10 covers the design and
analysis of comparative experiments, while Chapter 11 deals with the
design and analysis of factorial experiments.
9.2 Measurements
Once the problem has been well defined the next problem is to determine
the quantities to be measured. The response variables in which we are
really interested cannot always be measured directly but may be a
function of the measurements. These measurements are basic quantities
such as length, weight, temperature or the resistance of a thermo
couple. But the response variables may be more complex quantities
such as efficiency, fuel consumption per unit of output etc. If the
functional relationship between the response variable and the measure
ments is known, the measurements can be transformed by this function
to obtain the required quantities.
Unfortunately the measurements are almost certain to contain an
error component. By now you should be convinced that no two experi
ments are identical except in an approximate sense. There are an
infinite number of reasons why small variations will occur in repeated
tests. The power supply changes, the technician is tired, weather
conditions vary; and so on. If the variation is small the experiment
is said to be precise or repeatable. Precision is expressed quantitatively
as the standard deviation of results from repeated trials under identical
conditions. This precision is often known from past experiments, how
ever if it is not known then one of the experimental test objectives
should be to obtain an estimate of it.
Another useful property of ‘good’ measurements is that they should
be accurate or free from bias. An experiment is said to be accurate or
unbiased if the expected value of the measurement is equal to the true
value. In other words there is no systematic error. Measuring instru
ments should be checked for accuracy by relating them to some defined
standard, which is accepted, for example, by the National Bureau of
Standards in the U.S.A.
If the measured values are plotted against the standard reference
values, a calibration curve results. This curve enables us to estimate the
It is clear from the above that ‘good’ measurements have the same
properties as ‘good’ estimators, namely precision and accuracy (see
section 6.5). Four types of distributions of measurements are illustrated
in Figure 52. Because there is variation in the measurements, there will
also be variation in the corresponding values of the response variable.
The effect of the functional relationship with the measurements on the
error in the response variable is sometimes called the propagation of
error. The size of the error in the response variable will depend on the
precision and accuracy of the measurements. In the following sections
we will usually make the assumption that the measurements are
accurate. This is reasonable provided that all meters are recalibrated at
regular intervals.
205 Measurements
9.3 The propagation of error
We have seen that the response variable is often some function of the
observed measured variables. Our primary concern is to find the
distribution of the response variable so that, for example, we can
judge if the variation is unacceptably large. This section will be con
cerned with methods for finding the distribution of the response
variable (see also Box et ai, 1978, Section 17.2).
Z= +X 2 + . . .+x„.
Denote the true values of these measured variables by jui, . • ■, If
the measurements are accurate and have respective precisions
cTj, (72, then, using the results of section 6.2, we have
+ + - -+ = l^z
and
variance (z) = -b a j = of
For the general case, where the response variable is any linear function
of the measured variables,
Z = UiXi-hU2 X2 +---+<^nX„,
where the u,s are constants. Then we have
E[z) = Ui/li-ha 2 A^2 + ---+«n/^n =
and
variance (z) = a \ o \ - \ - a \ o \ ^ .. .-b a„ o^ = ■
Thus having found the mean and variance of the distribution of z, the
next question that arises is to find the type of distribution. The following
important result is stated without proof. If the response variable is a
linear function of several measured variables, each of which is normally
distributed, then the response variable will also be normally distributed.
So given observations on the measured variables, the corresponding
Example 1
An item is made by adding together three components whose lengths
are Xj, X2 and X3. The over-all length is denoted by z.
z = Xt H-x^ +X 3.
1
z
Figure 53
The tolerance limits (in inches) of the lengths of the three components
are known to be 1-960 ±0*030, 0-860 ±0-030 and 1-865 ±0-015, res
pectively. Thus the respective precisions of Xj, X2, and X3 are 0-010,
0-010, and 0-005 in.
Example 2
Suppose that the tolerance limits for the over-all length, which were
calculated in Example 1, were judged to be too wide. If the tolerance
limits for X2 and X3 cannot be narrowed, find reduced tolerance limits
for X| so that the over-all tolerance limits are 4*685 + 0 036.
Denote the revised precision of x^ by . The revised precision of the
over-all length is given by 0*012. Thus we have
The revised tolerance limits for x^ are 1*960 + 0*013. This, of course,
means that a considerable improvement must be made in the manu
facturing standards of this component in order to achieve the required
precision.
df
f(x) = f(n) + {x-tJ.f + R3
dx
where df/dx and d^f/dx^ are evaluated at x = ju. R 3 is the remainder
term which can be disregarded if higher derivatives are small or zero
and the coefficient of variation of x is reasonably small.
The above equation can be rewritten in the equivalent form
Then E(z) k E
But
E{x -ju) = 0
and
E{z) ^ +
Example 3
The measurement x, which is an unbiased estimate of p, is transformed
by the equation z = ^ x . Find the bias in using z as an estimate of yjp.
dz 1 d^z 1
dx l^ x dx^ 4x^/2-
/^z
Thus the percentage bias in estimating can be calculated for differ
ent values of the coefficient of variation of x, a/p.
Percentage bias
01 0-125%
0-2 0- 5%
0-3 1- 125%
The important point to notice here is that this is now a linear function
of X . In other words we have approximated / (x) by a linear function
in the area of interest. This means that if x is normally distributed.
and z - f i , % 1 ^ 1 (x-ju).
dz
(7^
Example 4
It is instructive to check the above formula when f(x) is a linear
function.
Let z = ax-\-b,
dz
then —- = a.
dx
Therefore variance (z) = ; this was the result obtained earlier.
Example 5
The measurement x, which is an unbiased estimate of p, is transformed
by the equation z == ^ x . Find the coefficient of variation of z in terms
of the coefficient of variation of x.
1
V i“
variance (z) — where a is the precision of x.
4n
V/^
_ a
2n
Example 6
Four measurements, 4 01, 4 08, 4T4 and 4*09 are made on a measured
variable whose unknown true value is denoted by p. These measure
ments are known to be accurate and to have precision a = 0*06. Find
a 95 per cent confidence interval for
The observed average measurement, denoted by Xo, is equal to 4 08.
Thus a point estimate oi ^ pis given by
^Xo = 2-02.
The precision of this estimate depends on {dz/dx)^, where z = ^ x .
Unfortunately this cannot be evaluated exactly as we do not know the
true value of p. However it is sufficiently accurate to evaluate (dz/dx)
at the point Xq. Thus the precision of the estimate of yjp is given by
Zjinxo)
00075.
Example 7
Assume that x and y are independent measured variables with mean
Py, and precision cr^, cr^, respectively. Also assume that (rjpx and
(Ty/py are less than about 0-15. Calculate the expected value and precision
of the response variable, z, in the following cases:
(a) A linear function
z = ax-\-by—where a, b are positive or negative constants.
dz
=a
dx dy
All higher partial derivatives are zero.
(b) Products
z = xy,
E{z) ^ fix
dz dz
= X.
dx dy
(c) Ratios
X
z=
y
E(z) ^ fi. = ^
Py
dz 1 dz X
dx dy 7^2•
.V
(■; ’ 1
^2 _2
2 ^ 2 2
or C i+ C ,^
But for the series-parallel arrangement it is not clear what the best
estimate of the total flow should be. Nor is it clear what the best
estimate of the total flow would be in the series arrangement if the
precisions of the three meters were not the same. Thus in the next
section we will outline a general method for combining different
measurements to obtain the best estimate of the response variable. This
method can be used in the above type of situation or wherever informa
tion is available from several sources.
Xf - + {i = l , . . . , n),
i= 1
n
dS
dfi
=I-=-^ n
Thus the sample mean is also the least squares estimate of jn.
The least squares method can be used whenever we have a set of
measurements, each of which has an expected value which is a function
of the unknown parameters. The sum of squared differences between
the measurements and their expected values is formed, and then
minimized with respect to the unknown parameters. The basic assump
tion necessary to apply the least squares method in this way is that the
precision of the observations should be the same. Later on we shall
see how to modify the method when the precisions are unequal. The
procedure is best illustrated with an example.
Example 8
Suppose that we want to determine the flow in a system with one input
and two output streams, all three being metered. This situation is
similar to the series-parallel arrangements depicted in Figure 56.
input
output 2
Figure 56
The relations between the true values and the measurements are given
by
Ml = 112 + ^2 ,
where 62, are the measurement errors associated with the three
meters. We begin by assuming that the meters have the same precision
so that
The method of least squares can then be used to estimate and ¡^2 -
For particular values of and /¿2, the sum of squared errors is
given by
d S
dni
Equating the two partial derivatives to zero and solving the two simul-
taneous equations, we get
p.1 = i(2m i-m 2+m 3),
fl2 = i(2m 2-m i+m 3),
/¿I + jU2 = (least squares estimate of total flow)
= ^ m , + m 2 + 2m3).
If we assume that the measurement errors are independent, then the
precision of the least squares estimates can be found from
= ^[4(7^ + + iJ^] =
3 ’
2(7^
^ ¡2 ^[4iT^ + i7^ + cr^] =
T ’
21 = 2(7^
{i = 1 , . . . ,n),
Example 9
Two chemists each make three measurements on the percentage of an
important chemical in a certain material. The observations are
Xi2, Xi3 and X21 , X22, ^ 23- Suppose that both chemists make unbiased
measurements but that the variance of measurements made by the
second chemist is four times the variance of measurements made by
the first chemist. Find the least squares estimate of the true percentage
of the chemical.
Denote the true percentage of the chemical by p. Because the
variances are unequal we do not take the simple average of the six
observations. Instead we want to give more weight to the first three
observations.
The constants P^, P2 j are given by
= 1 (i = l,2,3) and P^2 j (7 = 1,2,3).
The adjusted sum of squared deviations is given by
3 ^ 3
^ j=l
dS
dp ¡=1 ^j= l
Putting this derivative equal to zero we obtain the least squares
estimate of p.
Thus four times as much weight is given to the first three observations
as to the other three observations.
Example 10
The system depicted in Figure 56 is similar to the system used to
measure the flow of liquid oxygen through a rocket engine. There are
three meters which measure the total flow and its two components,
pre-burner flow and main burner flow. The precision of these three
meters is known to be the following:
where
p .= !S = i-o
(2-85r
(2-85)^
Pi = (20-37)2 = 0-0196
(2-85)2
F. = 0-0122.
(25-82)2
dS
= -2 P i(m i-/ii)-2 F 3 (m 3 -/ii-/^ 2 )>
dni
dS
= - 2P2Ìni2-H i i - l P ^ i m ^ - f i i - H i ) -
dn2
^2 + ^3) + P i ^ 3 (^ 3 - ^ 2 )
r r iiiP ^? 2 + ^2 ^3) + P j ^ 1)
=
P 1 P2 + P1 P3 + P2 P3
Note that if = P2 = P3 = 1, the precisions are equal and the
equations are the same as those obtained in Example 8.
Substituting the values of P^, P2 and P 3, we find
/¿I = 0*9925mi-h0'0075(m3 —m2),
/¿2 = 0'6193m2 + 0-3808(m3 —mj,
P1 + P 2 0'6118(mi-hm2) + 0*3883m3.
The precision of these three estimates turn out to be
(7^^ = 2-84,
(7^^ = 16*03,
+ = 16-09.
The precision of the least squares estimate of fii is only fractionally
less than the precision of mi, but the precision of the "'"ast squares
estimates of fi2 iPi ’^Pi) substantially less than the precision
of m2 and m3, and this makes the least squares procedure worthwhile.
Exercises
1. The measurement, x, is an unbiased estimate of ^ and has precision
(7. Expand in a Taylor series about the point x = fi. Hence show
that, if the coefficient of variation of x is small, the expected value
of z = x^ is given by -h and that the coefficient of variation of z is
approximately twice as large as that of x.
2. The three measured variables Xi, X2 and X3 have small independent
errors. The precisions are related by
^XI ~ ^X2 ~ ^^X3*
The response variable, z, is given by
Xi -hX2
z=
3x3
Figure 57
3. A system has one input and three output streams. Let mj, m2, m3
and m4 denote the reading of the respective meters. Assuming that
the meters are unbiased and have equal precision, find the least squares
estimate of the total flow.
4. Two flowmeters are connected in series. Let m j, m2, be the readings
of meters A, B respectively. Assuming that both meters are unbiased
and that the precision of meter A is ± 3 Ib/hr and that of meter B
± 5 Ib/hr, find the least squares estimate of the total flow.
5. The tolerance limits for the measured variables x and y are 1 06 ± OT2
and 2*31 ±0T 5 respectively. Find the tolerance limits of the response
variable z = xy.
Figure 58
References
B ox, G. E. P ., H unter, W. G ., and H unter, J. S. (1978), Statistics for
Experimenters, W iley.
M a n d e l , J . (1 9 6 4 ), The Statistical Analysis of Experimental Data, In terscien ce.
T o p p i n g , J , (1 9 6 2 ), Errors of Observation and Their Treatment, 3rd ed n . C h ap m an
and H a ll.
223 Exercises
Chapter 10
The design and analysis of
experiments-
1 Comparative experiments
10.1 Some basic considerations in experimental design
The design and analysis of experiments is an extensive subject to which
numerous books have been entirely devoted (e.g. Box et a/., 1978,
Davies, 1956, Cochran and Cox, 1957 and Cox, 1958). The problems of
design are of course inseparable from those of analysis and it is worth
emphasizing from the outset that, unless a sensible design is employed,
it may be very difficult or even impossible to obtain valid conclusions
from the resulting data. In the next two chapters we will try to illustrate
the basic principles of experimental design and try to provide an intro
duction to the maze of designs which have been proposed. In addition
the main methods of analysis will be discussed, including the analysis
of variance. Nevertheless the scientist should still be prepared to seek
the advice of a statistician if the experiment is at all complicated, not
only regarding the analysis of the results but also to select the appropriate
design.
Much of the early work on the subject was connected with agricul
tural experiments, such as comparing the yield from several varieties
of wheat. However so many other factors are involved, such as weather
conditions, type of soil, position in field, and so on, that unless the
experiment is carefully designed, it will be impossible to separate the
effects of the different varieties of wheat from the effects of the other
variables. The reader is referred to the pioneering work of Sir Ronald
Fisher (1951).
In general the techniques described in the next two chapters are
useful whenever the effects being investigated are liable to be masked by
experimental variation which is outside the control of the scientist.
The physicist or chemist working in a laboratory can often keep
unwanted variation to a minimum by working with pure materials and
by the careful control of variables other than those of direct interest.
In this case the methods described here may be of little value. However
for large scale industrial experiments complete control is impossible.
There is inevitably considerable variation in repeated experiments
_ 4 J ~
The standard error of the estimate of the treatment difference is
given by
/ / I 1
-+
number of observations on A number of observations on B
— ^2-
Thus the value of the i-test statistic is 8|/2^ = 3-4. But the estimate of
s is based on just two degrees of freedom (see below) and io o25,2 = 4*30
so that the result is not significant at the 5 per cent level. Nevertheless we
still suspect that there is a difference between the two methods and so it
would be advisable to make some more observations in order to improve
the power of the test.
It is difficult to give general realistic advice about the number of
replications required in a specific situation. The larger the number of
replications, the smaller will be the standard error of the difference
T a b le 15
T r ea tm en t 1 T, Xi /^i
We have
T>=i X,,.
j= l
,2 ^ Y (X j J - X j f
‘ A,
J= 1
n -l
It is easy to show that the average observation on treatment i is the
best point estimate of
Thus Xi = fii.
Our over-all aim is to compare these observed treatment means.
If they are ‘close together’, the differences may be simply due to the
residual variation. On the other hand if they are widely separated, then
there probably is a significant difference between the treatments. We
will describe two methods of data testing used to decide whether all
the theoretical treatment means, fii, are equal. However it will some
times be ‘obvious’ just by looking at the data that one treatment is
much better than all the other treatments. If we are simply concerned
with finding the ‘best’ treatment, rather than with finding good esti
mates of the treatment differences, then there may be no point in
Example 1
Steel wire was made by four manufacturers A, B, C and D. In order to
compare their products ten samples were randomly drawn from a
batch of wire made by each manufacturer and the strength of each piece
of wire was measured. The (coded) values are given overleaf and are
plotted in Figure D.2 in Appendix D.4 as box plots.
In terms of model lOA we will denote the mean strength of wire made
by A, B, C and D to be /t 2, /ta and respectively. The estimates of
these parameters are given by fl^ = 62-5, /¿2 = 86-0, = 70-5 and
fi^ = 90-5.
In terms of the equivalent model lOB we will denote the over-all
mean strength by and the effects of A, B, C and D by , i 2» ^3 and
The estimates of these parameters are given by
fl = 17A i i = - 1 4 - 9 ?2 = 8-6 i3=-6-9 = 1 3 -1
Note that the sum of these treatment effects is zero - except for a
round-off error of 0*1 .
We have now completed the initial stages of the analysis which
consists of setting up a mathematical model to describe the particular
situation and obtaining point estimates of the unknown parameters.
The next problem is to decide whether or not there is a significant
difference between the observed treatment means. In other words we
want to test the hypothesis
n — 1
j= 1
and this is an estimate of Thus the combined estimate of from
the variation within groups is given by
I S1
i= 1
^1? c (n - \)
Example 1 continued
We have the following information
Sample mean Sample variance
Manufacturer A 6 25 212-5
Manufacturer B 860 204-4
Manufacturer C 70-5 235-8
Manufacturer D 90-5 274-7
Largest treatment mean = 90-5, smallest treatment mean = 62*5,
range = 90*5 —62-5 = 280,
= 231-8.
This is based on thirty-six degrees of freedom.
28-0
(¡0 = 5-82
V(231-8/10)
^0 05(4’ 36) = 3-80.
Thus the result is significant at the 5 per cent level. The data suggests
that there really is a significant difference between the strength of wire
made by the four manufacturers.
i= 1
2
i= 1
1
"2 c —\
Table 16
One-way Anova
The two mean squares, and s^, are obtained by dividing the
appropriate sum of squares by the appropriate number of degrees of
freedom.
An important feature of the A nova table is that the total sum of
squares is equal to the sum of the between-group and within-group sum
of squares. This can be shown as follows. We have
X ..-X = (x ,-.-X f) + ( X f - x ) .
Z - ^)^= Z (^0--
i j ij
+Z - ^)^+ 2z (^0--
ij ij
Z Z (^0—
i= i j = i
= iZ= i L Z
j = i
n
and Y, i^ij ~ ^i) = ^ i.
7=1
Also ZZ
i = l j = l
= niZ= l
and we have the required result.
Another important feature of the A nova table is that the total
number of degrees of freedom is equal to the sum of the between group
and within group degrees of freedom. The details of the computation
are of some interest although most analyses are now carried out using
a computer package. Firstly calculate the group totals 7J(i = 1 to c)
and the over-all total T. Hence find the group means x.(i = 1 to c) and
the over-all mean x. Also calculate Y ^5» which is sometimes called the
iJ
total uncorrected sum of squares. Secondly calculate the total sum of
squares (sometimes called the total corrected sum of squares) which is
given by
Z (^o—
T2
= z^ cn ’
^ Y =^ Y ~
T2
- -n i n -
cn '
Note that the same correction factor is present. Lastly the residual
sum of squares can be obtained by subtraction.
1717
F-ratio =
231-8
= 7-41.
^ 0 0 1 , 3 ,3 6 = 4 *3 9 .
Thus the result is significant at the 1 per cent level and we reject the
hypothesis that there is no difference between wire made by the four
manufacturers. This, of course, is the same conclusion that was ob
tained with a range test. The reader will also notice that the residual
mean square is the same as the estimate of which was obtained in
the range test. However the way it is computed is completely different.
It is useful to know the average or expected values of the treatment
and residual mean squares. Since is an unbiased estimate of we
have
tf
= (7^ + n 2 c —1
Thus E{sl) = only if H q is true. If H q is not true, then the larger the
treatment effects are, the larger we can expect the treatment mean
square and hence the F-ratio to be.
var(x,-
n n
Thus an estimate of the standard error of the difference between the
two treatments is given by
Example 1 continued
The four means are
= 62*5 Xg = 860
Xc = 70-5 Xj) = 90*5.
The estimate 5 of cr is given by
5 = 7231-8 = 15-2
and is based on thirty-six degrees of freedom.
Choose a significance level of 5 per cent. Then the least significant
difference is given by
but (xd —Xg) and (xc —x^) are both less than 13*8, whereas (xg —Xc) is
greater than 13*8. Thus we conclude that manufacturers B and D are
better than A and C but that there is not a significant difference between
B and D. We can express these results as follows.
Xa Xc Xg Xd
The astute reader will notice that the above procedure is nothing
more than a multiple i-test. Earlier in the chapter, we decided that this
could not be used to test the hypothesis that the treatment means are
not significantly different. However it is often used as a follow-up
technique to see how the treatments are grouped, once the null
hypothesis has been rejected by an analysis of variance or range test.
The method will still tend to give too many significant differences,
but it is simple to use and for this reason is often preferred to such
Example 1 continued
^0Q25,36 ~ ^ ~ \5’2.
The 95 per cent confidence intervals for I^c are given by
62-5 ±9-8 86-0 ±9-8
70-5 ±9-8 90-5 ±9-8.
In the above statements the risk that a particular Pi will be outside
its confidence interval is a. But we have formed c such intervals. Thus
the expected number of mistakes will be ecu. In other words the risk
that at least one of the above statements is false will be much higher
than a and is in fact close to 1 —(1 —a)"". In order to form confidence
intervals so that the overall risk of making one or more incorrect
statements is say a', we must calculate the corresponding value of a
from the relation
a = l-(l-a)^
i: = 1 0 = 0-
i=l j=l
Notice that the model assumes that the treatment and block effects
are additive. It is also convenient to assume that the errors are normally
distributed with mean zero and constant variance and that the
errors are independent.
The data can be tabulated as follows.
Table 17
Column
total T, .. Ti.f.
Column
average X.i .. X.,
Column total 74 81 89
Column average 18-50 20-25 22-25
j= 1
Z (^-j
Total rc—i
Source of
variation Sum of squares d.f. E(mean square)
cr
j= 1
Thirdly the block (row) sum of squares is given by
r^ r^ Ti
c2
i= 1
(Xi.-x)^^^2 cr
‘ - i n - cr
Notice that the correction factor appears in all three equations. The
residual sum of squares can now be obtained by subtraction.
Each sum of squares is now divided by the appropriate number of
degrees of freedom to give the required mean squares. Two F-ratios
can now be calculated by dividing first the treatment mean square,
and secondly the block mean square, by the residual mean square. The
observed treatment F-ratio can then be compared with the upper
percentage point of the appropriate F-distribution in order to test the
hypothesis that all the treatment effects are zero. If the block effect is of
interest we can also test the hypothesis that all the block effects are
zero by considering the observed block F-ratio.
Example 2 continued
'^ijk = tj + £ijk
Table 19
Total nrc—1
ij,k
Time of day
Time 1 Time 2 Time 3
Engine 1 B2 Bj
Engine 2 B3 B, B2
Engine 3 B2 B3 Bi
The engines and times of day form the two types of block. The significant
point is that each burner only appears once in each row and in each
column. A design in which each treatment appears exactly once in each
row and each column is called a Latin square design.
It is easy to construct a Latin square. For example, the following
is a 4 X 4 Latin square.
A B C D
D A B C
C D A B
B C D A
The experiment can be randomized by randomly ordering the rows and
columns. The disadvantage of Latin square designs is that the number
of both types of block must equal the number of treatments, a condition
which is rather restrictive.
The analysis of results from a Latin square design is an extension of
the two-way analysis of variance. In addition to calculating the row
The treatment mean square can now be compared with the residual
mean square by means of an F-test.
Exercises
1. In a simple comparative experiment, show that the expected value of
c c
+ « Z i.-A f-l)-
i= 1
2. If the number of observations on each treatment in a one-way
analysis of variance is unequal and there are n, observations on treat
ment /, show that
Z Z = Z Z Z n i(X i-x)\
i= 1 j= 1 1= 1 j= 1 i= 1
Also verify that the number of degrees of freedom corresponding to
the total sum of squares, the treatment sum of squares and the residual
sum of squares are N —1, c —1 and N —c respectively where
N = t
i= 1
3. In a two-way analysis of variance show that
Z Z (Xij-xf
i= 1 j=l
= C Z
i= 1
{X i.-x ^ +r Z
j= 1
( x . j - x f
+Z Zi=lJ=1
{X ij-X i.-X .j + x f .
4. Test the hypothesis that the following four sets of data are homo-
geneous.
Set 1 : 7, 8, 7, 10.
Set 2 : 6, 8, 9, 7.
Set 3: 11 , 10, 12, 9.
Set 4: 12, 10, 9, 10.
5. The following data are the results from a randomized block experi
ment, which was set up to compare four treatments.
T r e a tm e n t
1 2 3 4
1 6 8 7 11
Block 2 5 6 6 8
3 9 9 11 12
255 Exercises
6. The following data resulted from an experiment to compare three
burners, Bi, B2 and B 3. A Latin square design was used as the tests
were made on three engines and were spread over three days.
Engine 1 Engine 2 Engine 3
Day 1 B. 16 B2 17 B3 20
Day 2 B2 16 B3 21 Bi 15
Day 3 Bj 15 B. 12 B2 13
References
Box, G. E. P ., H un t e r , W. G . , and H un t e r , J. S. (1978), Statistics for
Experimenters, Wiley.
C o c h r a n , W . G . , and C o x , G . M . ( 1957), Experimental Designs, W iley, 2nd edn.
C o x , D . R . ( 1 9 5 8 ), Planning of Experiments, Wiley.
D avi es, O . L. (e d.) ( 1 9 5 6 ) , Design and Analysis of Industrial Experiments, O liv e r &
B o y d , 2 n d edn.
F i s h e r , R . A . (1951), Design of Experiments, Oliver & B oyd , 6th edn.
J o h n , J. A . , and Q ue noui lle , M . H. (1 977), Experiments: Design and Analysis,
Griffin.
W e theri ll , G . B. (1981), Intermediate Statistical Methods, C h a p m a n and Hall.
11.1 Introduction
We now turn our attention to experiments where the response variable
depends on several different factors. For example, the yield of a chemical
reaction may be affected by changes in the temperature, the pressure
and the concentration of one of the chemicals. It may also be affected
by whether or not the mixture is agitated during the reaction and by the
type of catalyst employed.
A factor is any feature of the experimental conditions which is of
interest to the experimenter. Factors are of two types. Firstly a quantita
tive factor is one where possible values can be arranged in order of
magnitude. Any continuous variable, such as temperature or pressure,
is a quantitative factor. Conversely a qualitative factor is one whose
possible values cannot be arranged in order of magnitude. In the above
example the type of catalyst employed is a qualitative factor.
The value that a factor takes in a particular test is called the level
of the factor. Except for nuisance factors, we will assume that the levels
of the factors of interest can be determined by the experimenter. By
analogy with comparative experiments, a specific combination of factor
levels is called a treatment or treatment combination.
It is clear that comparative experiments can be thought of as a
special case of factorial experiments. In a simple comparative experi
ment the individual treatments are the different levels of the one and
only factor. A randomized block comparative experiment can also be
thought of as a factorial experiment in which two factors are involved.
The individual treatments form one factor and the blocks form a
second factor.
The over-all objective of a factorial experiment may be to get a
general picture of how the response variable is affected by changes in
the different factors or to find the over-all combination of factor levels
which gives a maximum (or minimum) value of the response variable.
The latter problem is discussed in section 11.10.
257 Introduction
A common type of test programme is the classical one-at-a-time experi
ment in which the value of the response variable is found for a particular
treatment combination, after which the factors are altered one-at-a-time
while keeping the other factors at their initial values. Such experiments
suffer from several serious defects and we shall see that it is often
preferable to run what is called a complete factorial experiment, in which
one or more trials is made at every possible combination of factor levels.
Such an experiment is often simply called a factorial experiment.
E T 105
To 100 108
pressure
(a) (b)
^0 T,
temperature temperature
Design one
One-at-a-time test
T e st n u m ber B u rn er F uel
1,2 Standard ( B 3 ) Standard (Fj)
3,4 Burner 1 (BJ Standard (Fj)
5,6 Burner 2 (B2) Standard (Fj)
7,8 Best burner Low additive (F2)
9,10 Best burner High additive ( F 3 )
Each pair of results is averaged and the best of the three burners with
no fuel additive is selected for the last four tests in order to try out the
fuel additive.
Two important criticisms of this design can be made. Firstly,
because of the sequential nature of the experiment, it is not possible
to randomize the order of the tests and so a nuisance factor may
systematically affect the results. Secondly this design cannot detect
interactions between the burners and the fuel, so there is a distinct
possibility that the experiment will not find the best burner-fuel
combination.
Two possible relationships between the burner efficiency, the burners
and the fuel additive are shown in Figure 61. In the first graph the
response curves are parallel so that there is no burner-fuel interac
tion. In this case the one-at-a-time test would find the best burner-fuel
combination. But in the second graph it is clear that the effect of the
fuel additive depends on the type of burner employed. In this case the
one-at-a-time test would select as the best combination while
in fact F 2B 2 is better. Since it is well known that there is usually an
optimum burner for a particular type of fuel (in other words the design
of the burner interacts with the type of fuel), it would be quite wrong to
employ a one-at-a-time test in this case.
The second design we will consider is a randomized complete fac
torial experiment. With two factors each at three levels there are nine
treatment combinations. Two observations are made at each treatment
combination so that eighteen tests are required. The order of these tests
is completely randomized and a typical test sequence is shown below.
Design two
Fully randomized factorial experiment
B u rn er
B2 B3
Fuel Fi 5,2 18,3 11,14
Fa 15,13 1,10 6,16
F3 4,7 12,9 17,8
(Numbers indicate the test sequence.)
Design three
Randomized block factorial experiment
B u rn er
Bi B2 B3
F uel F, 5,13 3,18 7,14
F2 2, 15 1,10 6,16
F3 4,11 9,12 8,17
Tests 1 to 9 are in the first block and tests 10 to 18 in the second block.
A similar design may also be used if the test programme is carried out on
two engines or with two batches of fuel. Within each block the order of
the tests is randomized and we have a randomized block factorial experi
ment. This is the best design for the particular experiment we have
described, though a simple randomized factorial experiment will
sometimes be adequate in other situations.
The data from the above experiment is given later in Example 1, and
will be investigated after we have discussed the analysis of a factorial
experiment in general terms.
Row Row
Level 1 Level 2 • Level c
total average
Column
T.i T.2 T
1 .f.
total
Column
average
x.i X.2 x.^
Bj = x .j - x .
The next step is to test the main effects of the two factors and the
interaction effect to see if any of them are significantly large. The
best way of doing this is by an analysis of variance. The total sum of
squares, — is partitioned into four components. Two of these
components, the row sum of squares and the column sum of squares,
have the same value as the two-way A nova described in section 10.11.
In addition the interaction sum of squares is calculated. The required
formulae are given below.
T a b le 21
T w o -fa c to r Anova w ith In tera ctio n
R esid u a l rc{n—l)
ij.k
T o ta l Z (Xijk-x)^ rcn —1
i.J.k
■ 1 nrc
z n T2
i= 1
nc me
z n T2
i=l
nr nrc
■^ij?
zn-
ij
n
The interaction sum of squares can then be obtained by subtraction.
Example 1
Two replications were made in the experiment to compare three
burners and to investigate a new fuel additive. The data is given below.
Burners
Cell Cell Cell Row
B. total B2 total B3 total total
F, 16 34 19 36 23 43 113
18 17 20
Fj 19 39 25 48 19 37 124
20 23 18
F3 22 43 21 45 19 36 124
21 24 17
T o ta l 111 17
_
rc ^ ( x . . ^ - x f =
kt
'
rc nrc
Then the true residual sum of squares can be calculated by subtracting
this quantity from ^ —^ 17)^. Note that once we have separated
ij,k
Table 22
Effect
A verage A B AB c AC BC ABC
(1 ) — — + — + + —
a + - - - + +
b + - + - - + +
ab + + + - - - -
c + - - + + - - +
ac + - - + + - -
be + - + - + - + -
abc + + + + + + + -h
1 1 1 1 1 1 1 1
F a c to r 8 4 4 4 4 4 4 4
d .f .
M a in effects 4
T w o -fa c to r in te ra ctio n s 6
T h re e-fa cto r in te ra ctio n s 4
F o u r -fa cto r in te ra ctio n s J
T o ta l l5
One way round this difficulty is to use the fact that interactions involv
ing more than two factors are rarely of practical significance. Thus the
residual mean square can be estimated by combining the sums of
squares of interactions involving three or more factors and dividing
by the corresponding number of degrees of freedom. The main effects
and two-factor interactions each have one d.f., and so the mean square of
each of these effects is the same as the corresponding sum of squares.
A series of F-ratios can now be obtained by dividing the main effect
and two-factor interaction mean squares by the residual mean square.
This enables us to determine which effects are significantly large.
Example 2
The following data are the results from a 2"^ factorial experiment.
Bi B2 B2
0, 18 16 12 10
^1
D2 14 15 8 9
16 19 12 13
^2 D, 14 11 7 8
The three and four factor interaction sums of squares are combined
to give
(residual sum of squares) = 0-25+ 2-25+ 0*0+ 9*0+ 2-25
- 13-75.
This quantity is based on five degrees of freedom so that
13-75
(residual mean square) = = 2-75.
The F-ratios are as follows:
E ffe c t F -r a tio
A 44-0
B 0-0
C 0-1
D 20-5
AB 0-1
AC 0-4
AD 0-0
BC 0-4
BD 0-0
CD 2-3
^ 0-05,1 - 6-61 ^001,1,5 — 16-26
fixed effects.
E (treatment mean square) =
^(j^ + naf random effects.
In the fixed-effects model the quantity ^ tf/{c— 1) can be thought of as
measuring the spread of the i,s, and is therefore analogous to af.
We will consider the case where all the variables are continuous.
In the case of one or two controlled variables, it is convenient to think
of this function in geometrical terms. With just one variable, x^, the
relation between y and can be represented by a curve whereas with
two variables, Xj and X2, the relation between y and Xj and X2 can be
represented by a surface whose height, at particular values of the
controlled variables, is equal to the corresponding value of the response
variable. These values of the response variable generate what is called
a response surface. It is often convenient to represent this response
surface by a contour diagram, similar to a geographical map. On each
contour the value of the response variable is a constant.
The problem of maximizing y is complicated by several factors.
Firstly there may be restrictions on the values which the controlled
variables can take; secondly we may not know the type of functional
relationship between the response variable and the controlled variables,
and thirdly there will be measurement errors in the observed values of
the response variable. However in chemical processes the coefficient
of variation of observations made under identical conditions is usually
reasonably small.
Methods of maximizing y are discussed in Davies (1956). A straight
forward though laborious procedure is as follows. With a single
controlled variable, x, measure the response variable over a wide range
X = —
2d2‘
If there are k controlled variables, the response variable can be mea
sured at a grid of points throughout the region of interest. In other
words a complete factorial experiment is performed. A response surface
can be fitted to the data and this enables us to estimate the optimum
conditions. Unfortunately this procedure may require a large number
of tests.
An alternative, and perhaps more efficient, approach is to proceed
iteratively (e.g. Cochran and Cox, 1957; Box and Draper, 1987).
With one controlled variable, the response variable is measured at
what is considered to be a reasonable staring point, say Xq, and also at
a somewhat higher value Xq 4- a. The position of the third measure
ment depends on the first two results. It is made above Xq + a, if the
second observation is the larger, and below Xq, if the first observation
is the larger. However if the first two observations are about the
same, the third measurement can be made between Xq and Xo + fl.
This step procedure continues until a maximum is found.
In the case of two controlled variables a procedure called the
method of steepest ascent can be employed. If the initial conditions
at the point P are not too close to the maximum, the response surface
can be approximated locally by a plane.
y = Uo-f-UiXi -fU2^ 2 -
If the experimental error is small, these parameters can be estimated
with a 2^ experiment around P. Then an estimate of the path of steepest
ascent can be obtained by making changes in the {xj which are
proportional to the {<5J. The reader is warned that this path will
depend on the scales in which the controlled variables are measured.
Figure 65 (a) Path of steepest ascent goes close to Q (b) Path of steepest ascent
does not go close to Q
Best results are obtained if the contour lines are approximately circular;
if the contour lines are elongated ellipses then the results will be rather
poor. The reason for this is that the path of steepest ascent is not
preserved when a linear transformation is made on one of the controlled
variables. The best procedure is to standardize the controlled variable
in such a way that unit changes in each variable can be expected to have
about the same effect. This is best illustrated by an example.
Example 3
The percentage of the theoretical yield of a chemical process, y, depends
on the temperature Xj and the percentage of one of the chemicals, X2 .
The objective of the experimental programme is to maximize y. A reason
able starting point is known to be x^ = 425°C and X2 = 11 per cent.
A change of 25°C in Xi and a change of 1 per cent in X2 are thought
to be roughly equivalent. A 2^ experiment was carried out and the
results are given below.
400 450
10 50 52
^2 12 54 56
-1 50 52
+1 54 56
Assume that the response surface can be represented locally by the
plane
y = +
Then estimates of «o, and ^2 can be obtained by the method of
least squares, as shown in section 8.6. Since the experiment is orthogonal
and the controlled variables have been standardized we have
Y j ^'u ~ Z ^ 2i = Z = 0* [Note: / = 1 to 4 in all summations].
This considerably simplifies the least squares normal equations to give
do = y = 53,
Y yi i 52 + 56 —54 —50
a, = = 1,
4
54 + 5 6 - 5 2 - 5 0
ai = 2.
Z 4
Thus the estimated plane is given by
y = 53 + x'i + 2x 2 .
The path of steepest ascent can be obtained by changing x\ and x '2 in
the ratio 1:2. But a change of one unit in x\ is equivalent to a 25°C
change in Xj, and a change of two units in x '2 is equivalent to a change
of 2 per cent in X2 . Thus the path of steepest ascent can be obtained by
starting at x^ = 425°C and X2 = 11 per cent, and changing Xj and X2
in the ratio 25°C :2 per cent. Some possible points for further analysis
are given below.
^2
Starting point 425 11
437^ 12
Possible points 450 13
462i 14
475 15
285 Summary
a considerable amount of information available. In other words the
physical laws governing the behaviour of the system are well under
stood. In this sort of situation the statistician is usually only required
to improve the precision and accuracy of the data. Thus when the
state of knowledge is high, the experimental uncertainty is minimized
and the need for statistics is not so acute.
Experiments which do not fall in this category are fewer in number
but much more important. The true research experiment is by defini
tion not founded on well established knowledge. For example, there
are many instances in the history of science in which an experiment has
given a surprising result or has led to an accidental discovery. This
type of experiment, which has a high degree of experimental uncertainty,
will require all the skill of both the engineer and the statistician. The
experimental designs, which have been described in the last two
chapters, will then be of tremendous value and considerable care should
be taken to choose the most suitable design.
Exercises
Z Z E = nc j] (Xi.-x)^ + nr f (x.j-x)^
i= I j = 1 k= I i =1 j= l
+nE Z (^¡r
/= 1 1
-X.J + X Ÿ + ZZZ
i= 1 j = 1 k = 1
i^ ijk -X ij)
El E2 E3
A 44 49 50
46 48 50
B 46 51 51
45 47 52
A, A,
17 18 16 14
10 13 6
Calculate the sum of squares for each effect. Test the main effects by
comparing them with the mean square obtained by combining the two-
and three-factor interactions.
References
B arn et t , V . and L e w i s , T . (1 9 8 4 ), Outliers in Statistical Data, 2n d e d n , W iley .
B o x , G . E . P. and D raper, N . R . (19 6 9 ), Evolutionary Operations, W iley.
B o x , G . E . P. and D r a p e r , N . R . (1 9 8 7 ), Empirical Model-Building and Response
Surfaces, W iley .
B o x , G . E. P., H unter, W. G ., and Hunter, J. S. (1978), Statistics for Experimenters.
W iley.
C h a t f i e l d , C. (1 9 8 8 ), Problem-Solving: A Statistician s Guide, C h ap m an and H a ll.
Cochran, W. G . and C o x , G . M . (1957), Experimental Designs, 2nd ed n , W iley.
D aniel, C . (1976), Applications of Statistics to Industrial Experimentation, W iley.
D avies, O . L. (e d .) (1 9 5 6 ), The Design and Analysis of Industrial Experiments, 2n d ed n ,
O liv e r and B o y d .
Low e, C . W . (1 9 7 4 ), ‘E v o lu tio n a r y o p e r a tio n in a c tio n ’. Applied Statistics, vol. 2 3 , pp.
2 1 8 -2 6 .
Wetherill, G . B. (1981), Intermediate Statistical Methods, C h ap m an and Hall.
287 Exercises
Chapter 12
Quality control
Example 1
What is the O-C curve for a sampling scheme such as that quoted in
section 12.1 in which a single sample is taken and the batch accepted if
there is not more than one defective item?
Let p denote the proportion of defective items in a batch. Let us
also assume that the sample size is small compared with the batch size.
Then the number of defective items in a sample will be a random
variable which follows the binomial distribution with parameters
n = 100 and p. (Strictly speaking it will follow a hypergeometric
distribution, see exercise 12, Chapter 4.)
Thus P(0 defectives) = (1 —p)^^^,
P(1 defective) = 100p(l —p)^^.
Thus the probability of accepting a batch is given by
(l_p)ioo_^100p(i_p)99
, reject
1 -M
log
P 2 (l-P l)
log
Pl(l-P2)_
1-Pl
log
I-P 2
P 2 (l-P l)
log
P l ( l - P 2)
After each item is inspected, the cumulative sample size n and the
cumulative number of defective items d are known. If d > sn-\-h2 , then
the batch is rejected and if d < sn —/zj, the batch is accepted. Otherwise
another item is inspected. Continue sampling until the batch is accepted
or rejected.
Sequential sampling was developed in the 1940’s by A. Wald and G.
A. Barnard; for further information see for example Duncan (1974,
Chapter 8). Note that a sequential sampling scheme is more efficient
than the equivalent double sampling scheme which has the same
specifications.
The consumer will receive two types of batches. The first type will
contain some defective items, but have been accepted by the sampling
inspection. The second type will contain no defective items as they
have been subjected to 100 per cent inspection and rectification. From
these two types of batches we can calculate the average outgoing
quality (abbreviated A.O.Q.), which is the average proportion of
defectives in batches received by the consumer.
As before, let p denote the proportion of defective items in a batch
before the sampling inspection. When p = 0 it is clear that A.O.Q. = 0.
When p = 1, all the batches will be subjected to 100 per cent inspection
and rectification so that again we have A.O.Q. = 0. In between these
two values, the A.O.Q. will have a maximum which is called the average
outgoing quality limit (abbreviated A.O.Q.L.).
Figure 70
• •
Ji___ t- target value
• • •• •
The most commonly used tool for detecting changes is the control
chart or Shewhart chart. Choose a variable which is characteristic of
the quality of the process and plot this against time. For example, a
chemical process may have a certain optimum temperature. In this
case the actual temperature can be measured at regular intervals and
compared with the target value by plotting it on a control chart. Of
course successive measurements will vary to a greater or lesser degree
even when the process is on target. Some of the causes of variation are
outside the control of the manufacturer and these combine to form
what is called the residual variation. However some causes of variation
can be identified and removed and these are called assignable causes.
A process is said to be under control if all the assignable causes of
variation have been removed.
Table 24
n 2 3 4 5 6 7 8 9 10 11 12
ki M3 1-69 2-06 2*33 2-53 2-70 2-85 2-97 3-08 3-17 3-26
ki 3-17 3-68 3-98 4-20 4-36 4.49 4-61 4-70 4-79 4-86 4-92
kj, 4-65 506 5-31 5-48 5-62 5-73 5*82 5-90 5-97 604 609
r^W
The values of A and W can easily be found by evaluating the binomial
probabilities, or, if the Poisson approximation is appropriate, they can
be found by evaluating the Poisson probabilities for which tables are
readily available. So long as the number of defectives in a sample is less
than W, it can be assumed that the process is in control. However one
observation greater than or equal to A, or two consecutive results
greater than or equal to W would mean that some action must be
taken.
Control charts based on the number of defective items in a sample
are sometimes called p-charts. However, with a complex product,
it is possible that several defects may occur in the same item. Then the
total number of defects in a sample can be calculated and plotted on a
control chart which is sometimes called a c-chart (see, for example,
Duncan, 1974, Chapter 20).
Example 2
The upper and lower specification limits for the width of a bolt are
IT 20 inches and MOO inches. Fifteen samples of five measurements
give the following readings for x and R. (Range values are multiplied
by 10^.) Set up control charts and see if the process is under control.
(Note that in practice it is advisable to have at least twenty-five samples
in order to set up control charts.)
This gives X = M l 30 and R = 70. Action lines for the x chart are
placed at
3R
x-h - M 130 + 0-0040.
k j5
All the observed values of x are inside these control limits. The upper
action line for the i?-chart is placed at
M = 16-4.
The values of R from samples 1 and 2 are greater than this and so the
process was then out of control.
The values of x and R were recomputed from samples 3-15. These
were x = 1•1126 and R = 5-4. The control limits for the x and K-charts
Q. 10
E • •
•
5 • •
0
3 4 5 6 7 8 9 10 11 12 13 14 15
sample number
3 (a) x-chart (b) /?-chart
* * ^ ^ w ^
^target
# ^ #
lower
^action
line
5,
• • •
•• • • •
---- -------------------
= S i+ (x 2 -n
In general
5, = 5 ,_ i+ (x ,-T ).
The next problem is to show that the local process mean depends
upon the slope of the cusum graph. We will show that the slope of
the line joining to measures the average difference from T of
to x„.
The local process mean, Xl , is given by
Xl = y X.-
i=m+in-m
1 ix-T)
= T + ‘-
n —m
12.10.2 Post-mortems
Cusum charts can also be used to analyse data retrospectively. Instead
of deducting some target value, we deduct the mean of the observations
over the period in question. This means that the cusum starts and
finishes at zero. The main point of interest is to determine if and when
changes in the local mean occurred. There are several ways of doing this.
The method we describe is suitable for computers.
First any freak values are removed from the data and replaced by
the average of the two adjacent values. The residual variance may be
estimated by
(Xj+i-Xj)
=2 2 (n-l)
with {n— 1) d.f.
Then Student’s i-test is used to locate the turning points; that is the
points at which the local mean changes.
This is done in the following way. Move along the series by connect
ing each observation with the last turning point (or initially with the
first observation), and find the maximum absolute distance between
this chord and the intervening cusum readings. The position of this
largest difference is used to divide the intervening readings into two
groups size n^ and ^2 respectively. The means 3ci and X2 of the two
time
Figure 76 A Manhattan diagram
i:
process
^2
process
controller
where a is a constant such that 0 < a < 1. The weights are intuitively
appealing as they sum to one and give progressively less weight to
observations further in the past. Furthermore the formula can be
rewritten in the simple recursive form:
x(N, l) = a x s - ^ { l - a ) x ( N - l , 1).
Box and Jenkins (1970) propose a general class of linear models
called autoregressive integrated moving average models (abbreviated
ARIMA models). We do not have space to introduce these here but a
description is given elsewhere by the author (Chatfield, 1989). The
Box-Jenkins approach includes exponential smoothing as a special
case. We also note that the approach can cope with non-stationary
series by differencing the data until any trend has been removed. Box
Exercises
1. The target value for a particular process is given by T = 50. Samples
of size five are taken at regular intervals and the average sample range
when the process is under control is given by 5 0. Set up control charts
for the process.
2. When a new process is put into operation the true target value
may be unknown. If the process is under control, the grand over-all
mean denoted by x, can be used as a target value. Set up control charts
for samples of size four when x = 10*80 and R = 0*46.
3. In the past the average proportion of defective items produced
by a certain process has been 3 per cent and this is a satisfactory quality
level. If samples of size thirty are taken at regular intervals, set up an
attributes control scheme.
4. The temperature of a chemical process is read every ten minutes.
The target value is T and the residual standard deviation is a. The
observations are plotted on a control chart which has action lines at
T ± 3(7. If the process mean shifts to T-h a, find the probability that an
observation lies outside the action lines. Hence find the average number
of observations which will be taken before an observation lies outside
the action lines. [H int: use the geometric distribution which is given by
= 6{\—9 y ~ ' for r = 1, 2, . . . and which has mean 1/0.]
5. Warning lines are placed at T ± 2 a on the control chart discussed
in question 4. Find the probability that two points in succession lie
outside the warning lines when the process is on target. Also calculate
this probability if the process mean shifts to T + (x.
6. The specifications for a certain quality characteristic are 15 0 ± 6 0
(in coded values). Fifteen samples of four readings gave the following
values for x and R.
Use the results from all fifteen samples to set up x and R control charts.
Hence decide if the process is in control. If not, remove the doubtful
samples, and recompute values for x and R, for use in succeeding
samples. Estimate the tolerance limits and see if the process will meet
the required specifications.
References
A n in tr o d u ctio n to tra d itio n a l q u a lity co n tro l is g iv en by W eth erill (1 9 7 7 ) and m ore
d e ta ile d g en e r a l a c co u n ts by D u n c a n (1 9 7 4 ) and G rant and L e a v e n w o r th (1 9 8 0 ). A n
in tr o d u ctio n to sa m p lin g in sp e c tio n is g iv en by G u e n th e r (1 9 7 7 ) and a m ore a d v a n ced ,
c o m p r e h e n siv e a cco u n t by H a ld (1 9 8 1 ).
T he paper by Bissell and P ridm ore (1981) is o f particular interest to British readers as
it review s the effect o f the 1979 Weights-and-Measures Act w hich p rovid es for the
ch an g e-o v er in the U n ited K in gd om from a p red om inan tly ‘m in im u m ’ system to an
‘avera g e’ system as practised elsew here in the C o m m o n M arket. F or exam p le, a b ox o f
m atches will now state the average num ber o f m atches included rather than the
m inim um num ber and this has m any im p lica tio n s for packers, trad ing standards
in sp ectors and consum ers.
T h ro u g h o u t the literature on quality co n tro l, it is usually assu m ed that quality
in sp ectors are ‘perfect’. H ow ever, a co m p a riso n o f in sp ectors (B issell, 1977) revealed an
u n a ccep tably large variation in in sp ectio n standards. S om e in sp ectors assessed m uch
low er defect rates than others on the sam e batch o f m aterial. It is also m y experience that
in sp ection standards vary con sid erably, either because o f n egligen ce or because the
in sp ectors gen u in ely apply different standards as to w hat con stitu tes a defective. T his
sh o u ld be borne in m ind in settin g up a quality co n tro l schem e. It appears that qu ality
co n tro l o f in sp ectors is required as well as qu ality co n trol o f the process!
T w o r e fe r e n c e s o n ‘m o d e r n ’ q u a lity c o n tro l p r o ced u res are Ish ik aw a (1 9 8 5 ) and B o x
(1 9 8 9 ). T h e la tter e m p h a siz e s th e r e v o lu tio n a r y sh ift from q u ality c o n tro l to q u ality
improvement, th e u se o f v ery sim p le to o ls , such as c h e c k -sh e e ts and grap h s, to
co n tin u a lly im p r o v e a sy ste m , and the u se o f ex p e r im e n ta l d esig n to im p r o v e p rod u ct
d e sig n . In p articu lar. B o x critically r ev iew s th e u se o f Taguchi methods w h ich p ro v id e a
sy ste m a tic a p p ro a ch , u sin g d e sig n e d e x p e r im e n ts, to se t a p ro d u ct’s ch aracteristics at
their target v a lu e s (o r o p tim iz e so m e d esir a b le q u a lity ch aracteristic) at th e sa m e tim e
as m in im izin g va ria b ility a ro u n d th e se targets and m akin g th e sy stem in se n sitiv e to
e n v ir o n m e n ta l ch a n g e s. A n o th e r k ey figure in m o d e rn q u ality co n tro l is W . E . D e m in g
w h o se ‘14 p o in ts for su c c e ss’ are r e v ie w e d by N e a v e (1 9 8 7 ). T h e y in clu d e such m axim s
as ‘D r iv e o u t fear, so that e v e r y o n e m ay w ork e ffe c tiv e ly for th e c o m p a n y ’, and
317 Exercises
e m p h a siz e the im p o r ta n ce o f w o rk in g as a team and a v o id in g the trad ition al
au th o rita ria n and h ierarch ical sty le o f m a n a g e m en t stru ctu re. O th e r fea tu re s o f
m o d e rn q u a lity co n tro l are that c o n sta n t im p r o v e m e n t sh o u ld b e a w ay o f life and that
th e sta tisticia n sh o u ld act as a q u a lity d e te c tiv e rather than as a p o lic e m a n . W h ile th ere
m ay be little that is sta tistica lly n ew in all th is, th e re is a fu n d a m en ta l d iffe r e n c e in
p h ilo so p h y and in a ttitu d e , particu larly by m a n a g e m e n t, and this n e e d s to b e clearly
r e c o g n iz e d .
Bissell, a . F . (1 9 7 7 ), ‘In co n sisten cy o f in sp ectio n stan d ard s - a case stu d y'. Bulletin in
Applied Statistics, vol. 4, pp. 1 6 -2 7 .
Bissell, A . F . and P ridmore, W. A . (1 9 8 1 ), ‘T he U K average q u an tity system and its
statistical im p lica tio n s’ (w ith d iscu ssio n ). Journal of the Royal Statistical Society. Series
A , vol. 144, pp. 3 8 9 -4 1 8 .
B ox, G. E. P. (1 9 8 9 ), ‘Q u a lity im p r o v em en t: An e x p a n d in g d o m a in for th e
a p p lica tio n o f scien tific m e th o d ’, Phil. Trans, of the Roy. Soc., A , 3 2 7 , 6 1 7 -6 3 0 .
B o x , G . E . P. and Jenkins , G . M . (1 9 7 0 ), Time Series Analysis, Forecasting and
Control, H o ld e n -D a y (revised edn pu b lish ed in 1976).
B o x , G . E . P. and G . M . (1 9 7 9 ), Practical Experiences with Modelling and
Jenkins ,
Forecasting Time Series, G w ily m Jen k in s and Partners (O verseas) Ltd.
C hatfi eld , C. (1 9 8 9 ), The Analysis of Time Series: An Introduction, 4th e d n .
C h a p m a n and H a ll.
Dodge, H . F . and Romig, H . G. (1 9 5 9 ), Sampling Inspection Tables, 2nd ed n , W iley.
D uncan , A . J. (19 7 4 ), Quality Control and Industrial Statistics, 4th ed n , Irwin.
G ran t , E . L. and L eavenworth , R . S. (1 9 8 0 ), Statistical Quality Control, 5th
e d n , M cG ra w -H ill.
G uenther, W. C . (1977), Sampling Inspection in Statistical Quality Control, G riffin.
Hald, a . (1981), Statistical Theory of Sampling Inspection by Attributes, A cad em ic
Press.
Harris, C. J. (1976), ‘P roblem s in system id en tification and c o n tr o l’, Bulletin of the
Institute of Mathematics, vol. 12, pp. 1 3 9 -1 5 0 .
I.C .I. (1964), Cumulative Sum Techniques, I.C .I. M o n o g ra p h no. 3, O liver and B oyd.
IsHiKAWA, K. (1 9 8 5 ), What is Total Quality Control? The Japanese way, tran slated
by D . J. L u , P ren tic e -H a ll.
Jacobs, O . L. R . (1 9 7 4 ), Introduction to Control Theory, C laren d on Press.
Lj u n g , L. (1 9 8 7 ), System Identification: Theory for the User, P r e n tic e -H a ll.
N e a v e , H . R . (1 9 8 7 ), ‘D e m in g ’s 14 p o in ts for s u c c e ss’. The Statistician, 3 6 , 5 6 1 -5 7 0 .
Pearson, E . S . (1 9 6 0 ), Application oj Statistical Methods to Industrial Standardization
and Quality Control, B .S. 600, British S tan d ard s In stitu tion (reprinted).
U .S . D epartment of D efense (1 9 6 3 ), Military Standard I05D, U .S . G overn m en t
Printing Office.
Wetherill, G . B. (1977), Sampling Inspection and Quality Control, 2nd ed n . C h ap m an
and H all.
F(t) = jfiu)du.
0^ ^ -
Thus it appears that our problem is to estimate the related functions
f{t), F{t) and R{t) from the observed failure times. However if the
type of mathematical distribution is unknown it is often expedient to
estimate another function called the conditional failure rate function
(or hazard function or instantaneous failure rate function) which is
given by
m
Z{t) =
R{ty
Then the probability that an item will fail in the interval from t to
i-l-Ai, given that it has survived until time i, is given by Z{t)At. It is
most important to realize that this is a conditional probability and
that Z{t) is always larger than f{t), because R(t) is always less than 1.
For example, the proportion of cars which fail between fifteen and
sixteen years is very small because few cars last as long as fifteen years.
But of the cars which do last fifteen years a substantial proportion will
fail in the following year. In fact the function Z(t) describes the distribu
tion of failure times just as completely as f{t) or F(t\ and is perhaps
the most basic measure of reliability.
It is useful to establish a direct relation between Z{t) and R(t). This
is done in the following way. The functions f{t) and F{t) are related
by the equation
dF{t)
m = (see section 5.1)
~dT
and R{t) = l- F ( i) ,
dR{t)
fit)= -
dt
Thus Z(t) = m
R(t)
= e x p [-/li],
F{t) = 1—exp[ —/li],
dF{t)
m =
dt
= X Qxp[-Xt].
We notice that / (t) is the probability density function of the exponential
distribution.
The exponential model states that chance alone dictates when a
failure occurs and that the product’s age does not affect the issue.
There are a few products of this type such as electron tubes and electric
fuses. In addition the exponential distribution describes the times
between failures of a complex piece of equipment (see Barlow and
Proschan, 1965, Chapter 2). However the reader is warned that many
products deteriorate with age and then the exponential model will not
apply.
tX e dt = T-
9=
Nl
9=
Example 1
Estimates of Z (i) were obtained for the data of Example 3, Chapter 1,
and the results are given below. Since N is 20 and is 691 •5 hours, the
‘obvious’ choice for T is 100 hours, giving 7 intervals. Then we
estimate Z (t) at i = 0, 100, 2 0 0 ,..., 600 hours, as shown below.
No. of items
c Interval No. of failures surviving until ¿(lOOc)xlOO
cT
0 (0, 100) 0 20 0/20 = 0-0
1 ( 100, 200) 3 20 3/20 = 0-15
2 (200, 300) 4 17 4/17 = 0-24
3 (300, 400) 8 13 8/13 = 0-61
4 (400, 500) 2 5 2/5 = 0-40
5 (500, 600) 2 3 2/3 =0-67
6 (600, 700) 1 1 1 /1 = 1-0
The estimates of (Z (t) x 100) range from zero to one and are plotted
in Figure 78. These estimates are fairly crude, particularly when the
remaining sample size sinks to less than 5. Nevertheless it seems clear
from the graph that Z (i) is not constant but rather increases with t. This
means that the exponential model will not apply. We will describe a
more suitable model in the next section.
An increase in Z{t) is a common phenomenon and means that the
product is more likely to fail the older it gets. Another common type
7c5
.9
c
o
o 1-0"
•O
B
to
E
time (hours)
Figure 78 Estimated conditional failure rate
Using formula 13.1 we can find the reliability function of the Weibull
model,
R(t) = e x p [-A r].
Hence F(r) = 1 —exp[ —
and f(t) = ^ exp[ —
R{t) = e x p [ - 2 r ] ,
giving In R(t) = —At"’
N items are tested without replacement. Let t, denote the iih failure
time. Then it can be shown that an unbiased estimate of F(tj) is given by
i/{N-\-l). Thus an unbiased estimate of R{ti) is given by (N + 1—i)/(N + 1).
But
N + 1- N+1
-In = In
N+\ N + l-i
Example 2
The smallest failure time is 104*3 hours, so L = 104*3. This is subtracted
from all the other observations and the analysis is carried out on the
remaining nineteen observations. Thus we take N = 19 and then
will be the (i —l)th failure after L. The values of ln[20/(20 —(i —1))] are
given below and the resulting values are plotted on log-log paper in
Figure 81.
20
i r , - 104-3
2 1 -/ \2 l -
1 0
2 54-4 1-05 0-049
3 89-4 M l 0-104
4 97-0 1-17 0-156
5 101-9 1-25 0-222
6 123-5 1-33 0-285
7 144-8 1-43 0-358
8 203-5 1-54 0-432
9 207-2 1-67 0-513
10 225-3 1-82 0-600
11 254-2 2-00 0-691
12 260-0 2-22 0-798
13 266-1 2-50 0-912
14 276-2 2-86 1-05
15 290-3 3-33 1-20
16 321-9 4.00 1.38
17 329-8 5-00 1-61
18 448-3 6-67 1-90
19 489-7 10-00 2-30
20 587-2 20-00 3-00
20
and when In = 1,
21-i
the corresponding value of t enables us to estimate A by
th ln(t —L) = —\nX.
From the graph we see t —L = 290 hours when ln[20/(21 —i)] = 1.
Hence X = 290~^^^
= 0 000059.
The exponential and Weibull distributions are only two of the many
possible types of model for describing the failure times of manufactured
products. Other distributions which have been suggested include the
gamma and log-normal distributions. The choice of a suitable distribu
tion is by no means automatic, and the reader is referred to Gross and
Clark (1975, especially Chapter 1), Buckland (1964) and Mann et al
(1974). The book by Gross and Clark is concerned mainly with
problems in the biological sciences, where one is concerned with
survival times for animals and humans, but the problems which arise
there are in many ways similar to the problems arising in life-testing.
A d iscu ssio n o f p ro b a b ility m o d els in reliability theory is given by B arlow and P rosch an
(1 9 6 5 , 1975). M ann et al. (1974) discu ss the fitting o f statistical failure m o d e ls,
accelerated life-testin g and system reliability. G ro ss and C lark (1975) are co n cern ed w ith
fitting survival d istrib u tio n s to b io m ed ica l life-tim e d ata. K alb fleisch and P rentice (1980)
are m ainly co n cern ed w ith fitting m o d e ls o f survival tim es in the p resence o f exp lan atory
variables, w hen the regression m eth o d s discu ssed in C h ap ter 8 o f this b o o k are n ot
a p p licab le beca u se survival tim es are generally n o t n orm ally d istributed.
The distribution
If X is N (0,1) then the random variable Y = is said to be a
random variable with one degree of freedom. In the general case,
n
if X i , ^ 2, . . . , are independent N (0,1), then Y = Y,
i= 1
be a random variable with n degrees of freedom. Thus, if
X i , X 2, . . . , X„ are independent N(/z, then
n
X t-A
=1Z
=1
is also a x^ random variable with n degrees of freedom.
This distribution is useful in many situations including the follow
ing; if Xi , . . . , x„ is a random sample size n from N (p i,a \ where
is unknown, then the sampling distribution of
X, —x\^ («-ir
2
332 Appendix A
The p.d.f. of a distribution with v d.f. is given by
( v - 2 ) / 2 ^ - ( > ’/2)
i
0 otherwise.
This distribution, in which v must be a positive integer, is a special
case of the gamma distribution. Its mean and variance are equal to
Vand 2v respectively. It is always skewed to the right, but tends to the
normal distribution as v ^ oo. Percentage points are given in Table 3,
Appendix B.
The t-distribution
If X is N{0,1), y is with Vd.f., and X and Y are independent, then
the random variable
A
JiY /v)
is said to have a t-distribution with v degrees of freedom.
This distribution is useful in a number of situations, including the
following: a random sample, size n, is taken from a normal distribution,
mean jj, and variance cr^. The sample mean, x, and sample variance, s^,
are calculated in the usual way. We know that the sampling distribu
tion of
X — ¡1
a /^ n
is N (0,1), and that the sampling distribution of (n—l)s^/a^ is with
(n—1) d.f. Moreover it can be shown that the random variables x and 5^
are independent, even though the values are calculated from the
same sample. Thus the statistic
a /J n
t=
{n —l)s^/(T^
n —1
sl-Jn
has a t-distribution with (n—1) d.f.
333 Appendix A
The p.d.f. of a i-distribution with v d.f. is given by
The F-distribution
If , ^2 are independent random variables with , V2 d.f. respec
tively, then the random variable
F=
X 2 /V2
is said to have an F-distribution with and V2 d.f. respectively.
This distribution is useful in a number of situations including the
following: two random samples, size and «2 respectively, are taken
from a normal distribution^ mean jj, and variance cr^. The two sample
variances, s j , and S2 , are calculated in the usual way. Then the sampling
distribution of {ni — l)sl/(T^ is x^ with {n^ (n^-— l) d.f., and the sampling
distribution of (/I2 - \)s\I g^ is x^ with («2 “ 1) Thus the statistic
« 1 -1
F=
{n2-\)sllG^
«2- 1
F
■* 1 - a , v i , V 2 = -17 L _
334 Appendix A
Appendix B Statistical tables
Table 1 Areas under the normal curve
0-0 -5000 -5040 -5080 5120 5160 5199 5239 5279 5319 -5359
•5398 -5438 •5478 5517 •5557 •5596 5636 5675 5714 5753
•5793 -5832 •5871 5910 •5948 •5987 6026 6064 6103 6141
•6179 -6217 •6255 6293 •6331 •6368 6406 6443 6480 6517
•6554 6591 •6628 6664 •6700 •6736 6772 6808 6844 6879
•6915 -6950 •6985 7019 •7054 •7088 7123 7157 7190 7224
■7257 -7291 •7324 7357 •7389 •7422 7454 7486 7517 7549
•7580 -7611 •7642 7673 •7703 •7734 7764 7793 7823 7852
•7881 -7910 •7939 7967 •7995 •8023 8051 8078 8106 8133
•9 •8159 -8186 •8212 8238 •8264 •8289 8315 8340 8365 8389
10 •8413 8438 •8461 8485 •8508 •8531 8554 8577 8599 8621
1-6 •9452 -9463 •9474 9484 •9495 •9505 9515 9525 9535 9545
1-7 •9554 *9564 •9573 9582 •9591 •9599 9608 9616 9625 9633
1-8 •9641 -9648 •9656 9664 •9671 ■9678 9686 9693 9699 9706
1-9 •9713 -9719 •9726 9732 •9738 •9744 9750 9756 9761 9767
20 •9772 9778 •9783 9788 •9793 •9798 9803 9812 9817
21 •9821 -9826 •9830 9834 •9838 •9842 9846 9850 9854 9857
2- •9861 2-9864 9871 •9875 •9878 9884 9887 9890
2-3 •9893 9896 9901 •9904 •9906 9911 9913 9916
2-4 •9918 -9920 •9922 9924 •9927 •9929 9930 9932 9934 9936
2-5 •9938 9940 •9941 9943 •9945 •9946 9948 9949 9951 9952
2-6 •9953 -9955 •9956 9957 •9959 •9960 9961 9962 9963 9964
2-7 •9965 -9966 •9967 9968 •9969 •9970 9971 9972 9973 9974
2*8 •9974 9975 •9976 9977 •9977 •9978 9979 9979 9980 9981
2- •9981 9-9982 •9982 9983 •9984 9985 9985 9986 9986
3- •9986 09987 •9987 9988 •9988 9989 9989 9990 9990
3-1 •9990 9991 •9991 9991 •9992 •9992 9992 9992 9993 9993
3-2 •9993 -9993 •9994 9994 •9994 •9994 9994 9995 9995 9995
3-3 •9995 -9995 •9995 9996 •9996 ■9996 9996 9996 9996 9996
3-4 •9997 -9997 •9997 9997 •9997 •9997 9997 9997 9997 9998
3-5 •9998 -9998 9998 •9998 •9998 9998 9998 9998 9998
3-6 -9998 -9998 -9998 9999 9999 -9999 9999 9999 9999 9999
335 Appendix B
Table 2 Percentage points of Student's f-distribution
336 Appendix B
Table 3 Percentage points of the distribution
V
1 0000 00002 0001 00039 0-45 1 64 271 384 502 663 788
2 0010 0020 0051 0 103 1 39 322 4 61 599 738 921 10-60
3 0-072 0 115 0216 0352 237 464 625 781 935 11 34 12 84
4 0207 030 0484 071 336 5-99 778 949 11 14 13 28 14 86
5 0 412 055 0 831 1 15 435 7 29 924 11 07 12 83 15 09 1675
40 20 71 22 16 244 3 26 51 39 34 47 27 51 81 55 76 59 34 63 69 66 77
50 27 99 29 71 3236 3476 4933 58 16 63 17 67 50 71 41 76-15 7949
60 35 53 37 48 40 48 43 19 5933 6897 74 40 79 08 83 30 8838 91 95
70 43 28 4544 48 76 51 74 6933 79 71 85 53 90 53 95 02 100 43 1042
80 51 17 53 54 57 15 6039 79 33 9041 9658 101 88 106 63 112 33 1163
337 Appendix B
Table 4 Upper percentage points of the f-distribution
a
( ) « = 001
1 2 3 4 5 6 7 8 9 10 12 15 20 24 30
V2
1 4052 2 4999 5 5403 4 5624 6 5763 6 5859 0 5928 4 5981 1 6022 5 6055 8 6106 3 6157 3 6208 7 62346 62606
2 9850 99-00 9917 9925 99 30 9933 9936 9937 9939 9940 99-42 99-43 99-45 9946 9947
3 3412 3082 2946 2871 2824 2791 2767 2749 2735 2723 2705 26 87 2669 2660 2650
4 2120 1800 1669 15 98 15 52 1521 1498 1480 1466 1455 1437 14-20 1402 13-93 1384
5 16 26 13-27 12 06 1139 1097 10-67 10-46 10-29 1016 10-05 9-89 972 9-55 9-47 9 38
6 13 75 10-92 978 9-15 875 8-47 8-26 810 798 787 772 756 740 731 723
7 12 25 9-55 8-45 785 746 719 699 684 672 662 647 6-31 616 607 599
8 1126 8-65 759 701 663 637 618 6-03 591 581 5-67 552 536 528 520
9 10 56 802 6-99 642 606 580 561 547 535 526 5-11 4-96 4-81 473 465
10 1004 7-56 6-55 599 5-64 539 520 506 494 4-85 471 456 4-41 433 425
11 9-65 721 622 5-67 532 5-07 489 474 463 4-54 440 4 25 4-10 402 394
12 933 6-93 595 541 506 4-82 4-64 450 439 430 416 401 386 378 3 70
13 907 6-70 574 5-21 486 4-62 44 4 430 419 4 10 3-96 382 3-66 359 3-51
14 8-86 6-51 5-56 504 4-69 446 428 414 403 3 94 3-80 3-66 351 343 335
15 868 6-36 542 4-89 4-56 432 4-14 400 389 3-80 3-67 352 337 329 321
16 853 623 529 477 4-44 420 403 389 378 3-69 355 3-41 3 26 318 310
17 8-40 6-11 518 467 4-34 410 3-93 379 368 359 346 3-31 316 3-08 300
18 829 6-01 5-09 458 425 4-01 3-84 371 360 3-51 337 323 308 300 292
19 818 593 5-01 45 0 417 3-94 377 363 352 343 330 3-15 300 292 284
20 8-10 585 4-94 443 41 0 3-87 370 356 346 337 323 309 2 94 286 278
21 802 578 487 437 404 3-81 3-64 351 340 331 317 303 288 280 272
22 795 572 482 4-31 399 376 3-59 345 335 326 312 298 283 275 2-67
23 788 566 476 426 3-94 371 354 341 330 321 307 293 278 270 262
24 782 561 472 422 390 3-67 350 336 326 317 303 289 274 266 258
25 777 5-57 468 418 385 363 346 332 322 313 299 285 270 262 254
26 772 553 464 4-14 382 359 3-42 329 318 309 296 281 266 258 250
27 768 549 460 4-11 378 3-56 339 326 315 306 293 278 263 255 247
28 764 5-45 457 407 375 353 336 323 312 303 290 275 260 252 244
29 760 5-42 4-54 4-04 373 350 333 320 309 300 287 273 257 249 241
30 756 539 451 402 370 3-47 330 317 307 298 2 84 270 255 247 239
40 731 518 431 383 3-51 329 312 299 289 280 266 252 237 229 220
60 708 498 4-13 365 3-34 3-12 295 282 272 263 250 235 220 212 203
120 685 479 3-95 3-48 317 296 279 266 256 2-47 2 34 219 203 195 186
00 663 4-61 378 332 302 280 264 251 241 232 2-18 204 188 179 170
338 Appendix B
Table 4 (continued)
i b )n = 0025
1 2 3 4 5 6 7 8 9 10 12 15 20 24 30
V2
1 647 79 799-50 86416 899 58 921 85 937 11 948 22 956 66 963 28 968 63 976 71 984 87 993 10 997 25 1001 4
2 3851 39-00 3917 39-25 3930 3933 3936 3937 3939 3940 3941 3943 3945 3946 3946
3 17 44 16-04 1544 K 10 1488 1473 1462 14-54 1447 14-42 1434 1425 1417 1412 1408
4 12 22 10-65 9-98 960 936 920 9-07 898 890 884 875 866 856 8-51 846
5 10-01 8-43 776 739 715 698 685 676 668 662 652 643 633 628 623
6 881 726 6-60 623 599 582 570 560 552 546 537 527 517 512 507
7 8-07 654 589 552 529 512 499 490 482 476 4-67 4-57 4-47 4-41 436
8 757 606 542 505 482 465 453 4-43 436 4-30 420 410 400 395 389
9 721 5-71 508 472 448 432 420 410 403 396 387 377 367 361 356
10 6-94 5-46 483 447 4-24 407 395 385 378 372 362 3-52 342 337 331
11 672 5-26 4-63 428 4-04 388 3-76 3-66 359 353 343 333 323 317 312
12 655 5-10 4-47 412 389 373 3-61 351 344 337 3 28 318 3-07 302 296
13 641 497 435 400 377 360 348 339 331 325 315 3-05 295 289 284
14 630 486 424 389 366 350 338 329 321 3-15 305 295 284 279 273
15 620 477 415 3-80 358 341 329 320 312 306 296 286 276 270 264
16 6-12 4-69 408 373 3-50 3-34 322 3-12 305 299 289 279 268 263 257
17 6-04 462 401 366 344 328 3-16 306 298 292 282 272 262 256 250
18 598 4-56 395 3-61 338 322 3-10 3-01 293 287 277 267 256 250 244
19 592 4-51 390 356 333 317 305 296 288 282 272 262 2-51 245 239
20 587 446 3-86 351 329 313 3-01 291 284 277 268 257 246 241 235
21 5-83 4-42 382 348 325 309 297 287 280 273 2 64 253 242 237 231
22 579 438 378 3-44 322 305 293 284 276 270 2-60 250 239 233 227
23 575 4-35 375 3-41 318 302 290 281 273 2-67 257 247 236 230 224
24 572 432 372 338 315 299 287 278 270 2-64 2 54 244 233 227 221
25 5-69 429 3-69 3-35 313 297 285 275 268 2-61 2-51 2-41 230 224 218
26 5-66 427 367 333 310 294 282 273 265 259 249 239 228 222 216
27 5-63 424 3-65 331 308 292 280 271 263 2-57 2-47 236 225 219 213
28 5-61 422 363 329 306 290 278 269 261 255 245 2 34 223 217 211
29 5-59 4-20 361 327 304 288 276 267 259 253 243 232 221 215 209
30 5-57 4-18 3-59 325 3-03 287 275 265 257 2-51 2-41 231 220 214 207
40 542 405 346 313 290 274 262 253 245 239 229 218 207 2-01 1 94
60 529 393 334 301 279 263 251 2-41 233 227 2-17 206 194 188 182
120 515 380 323 289 267 252 239 230 222 216 205 194 1 82 1 76 169
00 502 3-69 312 279 257 2-41 229 219 211 205 1-94 183 1-71 1-64 1 57
339 Appendix B
Table 4 (continued)
ca
( ) = 005
1 2 3 4 5 6 7 8 9 10 12 15 20 24 30
^2
1 161 45 199 50 215 71 22458 23016 23399 236 77 23888 240 54 241 88 24391 245 95 248 01 24905 25010
2 1851 19-00 1916 19 25 1930 19 33 19 35 1937 1938 19-40 1941 1943 1945 1945 19-46
3 1013 9-55 928 912 901 894 889 885 881 879 874 8 70 8 66 864 862
4 771 694 659 639 626 616 609 604 600 596 591 586 580 577 575
5 661 579 5-41 5-19 505 495 488 482 477 474 468 4 62 456 4 53 450
6 599 5-14 476 453 439 428 421 415 410 4-06 400 394 387 384 381
7 559 47 4 435 412 3-97 387 379 373 368 364 357 351 3-44 3-41 338
8 532 446 407 38 4 369 3-58 350 344 339 335 328 3 22 3-15 312 3 08
9 512 426 386 363 348 337 329 323 3-18 314 307 301 294 290 286
10 496 410 371 348 333 322 3-14 307 302 298 291 284 277 274 270
11 484 398 359 336 320 309 301 295 290 285 279 272 265 261 257
12 4-75 389 349 326 311 300 2-91 285 280 275 269 262 254 251 247
13 4-67 3-81 341 318 303 292 283 277 271 2-67 260 253 246 242 238
14 4-60 3-74 33 4 311 296 285 276 270 265 260 253 246 239 2 35 231
15 454 368 329 306 290 279 271 264 259 2 54 248 240 233 229 2 25
18 449 363 324 301 285 2-74 266 259 25 4 249 242 235 228 224 219
17 4-45 3-59 320 296 281 270 261 255 249 245 238 231 223 219 215
18 4-41 355 316 293 277 2-66 258 251 246 241 234 227 219 215 211
19 43 8 352 313 290 274 263 254 248 242 238 231 223 216 211 207
20 435 349 310 287 271 260 2-51 245 239 235 228 220 212 208 204
21 432 3-47 307 28 4 268 2-57 249 242 237 232 225 218 210 205 201
22 4-30 3-44 3-05 282 266 255 246 2-40 234 230 223 215 207 203 198
23 42 8 342 303 280 264 253 244 237 232 227 220 213 205 201 196
24 426 340 3-01 278 262 2-51 242 236 230 225 218 211 203 198 194
25 42 4 339 299 276 260 249 240 234 228 224 216 209 201 196 192
26 423 337 298 274 259 247 239 232 227 222 215 207 199 195 190
27 421 335 296 273 257 246 237 231 225 2 20 213 2 06 197 193 188
28 42 0 334 2-95 271 256 245 236 229 224 2 19 212 204 196 191 187
29 418 333 293 270 255 243 235 228 222 218 210 2 03 194 190 185
30 417 332 292 269 253 242 233 227 221 216 209 201 193 189 184
40 408 323 284 261 245 234 225 218 21 2 208 200 192 184 179 174
60 40 0 315 276 253 237 Z 25 2-17 2-10 2-04 199 192 184 175 170 1-65
120 392 3-07 268 245 229 218 209 2<)2 196 1-91 183 175 166 1-61 155
00 3*84 300 260 237 221 2-10 m 194 188 183 175 167 157 152 146
340 Appendix B
Ta b les Values of e
-O OU -0 u -OOb• -Ob —X
e e e X e e e
341 Appendix B
Table 6 Percentage points of the distribution of the Studentized range
(a) a =0 01
10 11 12 13 14 15 16 17 18 19 20
1 90-0 135 164 166 202 216 227 237 246 253 260 266 272 277 282 286 290 294 298
14-0 19-0 223 247 20-6 282 29 5 307 31-7 326 33-4 341 348 35-4 360 36-5 37-0 375 379
2
3 826 108 122 13-3 14-2 15-0 15-6 16-2 16-7 17-1 17-5 17-9 18-2 18-5 18-8 19-1 19 3 19-5 19-8
4 6-51 1812 9-47 9-96 10-0 11-1 .115 11-9 12-3 12-6 12 8 13-1 13 3 135 13-7 13-9 14 1 142 14-4
5 5-70 6 97 7 80 8 4 2 89 1 932 967 997 10 24 10 48 10 70 10-80 11 08 11-24 11 40 11 55 11 68 11 81 11 93
6 5-24 633 703 7-56 797 832 8-61 8-87 9-10 930 9-49 9-65 9-81 9-95 10-08 1021 10 32 10 43 10-54
7 4-95 592 6-54 r-01 7-37 768 7-94 8 17 8-37 8-55 871 8-86 900 9-12 9-24 9-35 946 9-55 965
8 4-74 563 620 663 696 724 7-47 768 7-87 803 8 18 8-31 8-44 8-55 8-66 8-76 8-85 894 9-03
9 4-60 5-43 5-96 635 6-66 6-91 7-13 7 32 7-49 7 65 7 78 7-91 803 8-13 823 832 841 849 8-57
10 4-48 5 27 577 6 14 6-43 6-67 687 705 7-21 736 748 7 60 7 71 7-81 7 91 7 99 8-07 8 15 822
11 439 5-14 562 5-97 625 6-48 667 6-84 6-99 713 725 7-36 7-46 7-56 7-65 7 73 781 788 7-95
12 432 5-04 5-50 584 6-10 632 651 6-67 6-81 6-94 706 7 17 726 736 7-44 752 7 59 766 7-73
13 426 496 5-40 5-73 598 6-19 637 6-53 6-67 679 6-90 7-01 7-10 7-19 727 734 7-42 748 7-55
14 421 4-89 532 5-63 588 608 626 6-41 6 54 666 6-77 6-87 696 7-05 7 12 7 20 7 27 733 739
15 4 17 483 525 5-56 580 599 6 16 631 6-44 6-55 666 6 76 684 693 7 00 707 7 14 720 7 26
16 4-13 4-78 5-19 5-49 5-72 5-92 608 622 6-35 6-46 6-56 666 674 682 690 697 703 7-09 7-15
17 4-10 4-74 5-14 5-43 5-66 585 6-01 6-15 6-27 6-38 648 6-57 6-66 673 680 687 694 700 705
18 4-07 470 500 5-38 5-60 579 5-94 6-08 6-20 631 641 6-50 6-58 6-65 6 72 6 79 685 691 696
19 405 467 505 5-33 5-55 5-73 5-89 6-02 6-14 6-25 6-34 6-43 651 658 665 672 678 684 689
20 402 4-64 502 529 5-51 569 584 5-97 6-09 6 19 629 637 645 652 659 665 6 71 6 76 682
24 396 454 4-91 5-17 5-37 5-54 5-69 5-81 5-92 602 6-11 6-19 626 6-33 639 6-45 651 656 6-61
30 3-89 445 480 5-05 524 5-40 5-54 5-65 5-76 585 5-93 6 01 608 6-14 6-20 626 631 636 6-41
40 382 437 470 493 5-11 5-27 530 5-50 5-60 569 5-77 5-84 590 5-96 602 607 6-12 6 17 621
60 3 76 428 4-60 482 499 5 13 525 536 5-45 5 53 560 567 5 73 5 79 584 589 593 5 98 602
120 370 420 4-50 4-71 4-87 5-01 5 12 521 5-30 538 5-44 5-51 556 5-61 566 571 5 75 579 583
364 4-12 440 460 476 488 499 5-08 5-16 523 529 5-35 5-40 545 549 5-54 557 5 61 565
00
c is the size of the sample from which the range is obtained and v is the number of degrees of freedom of s.
( )
b a =005
10 11 12 13 14 15 16 17 18 19 20
1 18 0 270 328 37 1 404 43-1 45-4 474 491 506 52-0 532 54-3 55-4 56-3 572 58-0 588 596
2 609 83 98 10-9 1 17 12 4 13-0 13-5 14-0 14-4 14 7 15-1 15-4 15 7 15 9 16 1 16 4 16 6 168
3 4-50 5-91 682 7-50 8-04 8-48 8-85 9-18 9-46 972 9-95 1015 10-35 10-52 10 69 10 84 10 98 11-11 11 24
4 393 504 576 629 6-71 705 7-35 760 783 803 821 837 852 866 879 891 9-03 9 13 923
5 364 460 522 5-67 603 633 658 680 690 7 17 7 32 747 760 7 72 7 83 7 93 803 8 12 821
6 3-46 4-34 4-90 5-31 563 589 6-12 6-32 649 665 679 6-92 703 7 14 724 734 743 7-51 7 59
7 334 4 16 468 5-06 5-36 5 61 5-82 600 6 16 630 643 655 666 676 685 6-94 702 709 7 17
8 326 404 453 4-89 5 17 540 5-60 577 592 605 6 18 629 639 648 657 665 673 680 687
9 320 395 442 4-76 502 524 5-43 560 5 74 587 598 609 6 19 628 636 644 651 658 6-64
10 3 15 388 433 4-65 491 5 12 530 546 560 5 72 583 593 603 6-11 620 627 634 640 647
11 3 11 382 4-26 457 482 503 5-20 5-35 5-49 5-61 5-71 581 590 599 606 6 14 620 626 633
12 3-08 377 420 4-51 4’75 495 5-12 527 540 5-51 5-62 5-71 5-80 588 5-95 603 609 6 15 621
13 306 3-73 415 4-45 469 488 5-05 5 19 532 543 553 563 5-71 579 586 593 600 605 6-11
14 3-03 3 70 4-11 4-41 4-64 4-83 4-99 5-13 525 536 5-46 5-55 5-64 572 579 585 592 597 603
15 301 367 408 437 460 478 4-94 5-08 520 531 540 549 558 565 572 579 585 590 596
16 3-00 365 4-05 433 456 4-74 4-90 503 5-15 526 535 5-44 552 5-59 566 572 579 584 590
17 2-98 363 4-02 430 4-52 4-71 4-86 499 5-11 5 21 5-31 539 5-47 5-55 5 61 568 574 5 79 584
18 297 3-61 4-00 428 4-49 4-67 4-82 4-96 5-07 5-17 527 535 543 550 557 5-63 569 574 5-79
19 296 359 3-98 4-25 4-47 4-65 4-79 492 5-04 5-14 523 532 539 5-46 553 559 565 5 70 575
20 295 3 58 396 423 4-45 462 4-77 490 5 01 5-11 520 528 536 5-43 549 555 561 566 5 71
24 292 353 3-90 4-17 4-37 454 4-68 4 81 4-92 5-01 5 10 5-18 525 532 538 5-44 550 554 559
30 289 349 3-84 4-10 430 4-46 4-60 472 4-83 492 500 508 5 15 5-21 527 533 538 543 5-48
40 286 3-44 379 4-04 423 439 4-52 463 4-74 4-82 4-91 498 5-05 5-11 5-16 522 5-27 531 536
60 283 340 374 398 4 16 431 444 455 4-65 473 481 488 494 500 506 5 11 5 16 520 524
120 280 336 369 3-92 4-10 4-24 4-36 448 456 4-64 4-72 478 484 4-90 495 500 5-05 509 5-13
00 2-77 3-31 3-63 386 4-03 417 4-29 4-39 4-47 455 4-62 4-68 4-74 480 485 489 493 497 5-01
c is the size of the sample from which the range is obtained and v is the number of degrees of freedom of s.
Table 7 Random numbers
19211 73336 80586 08681 28012 48881 34321 40156 03776 45150
94520 44451 07032 36561 41311 28421 95908 91280 74627 86359
70986 03817 40251 61310 25940 92411 34796 85416 00993 99487
65249 79677 03155 09232 96784 17126 50350 86469 41300 62715
82102 03098 01785 00653 39438 43660 02406 08404 24540 80000
91600 94635 35392 81737 01505 04967 91097 02011 26642 38540
20559 85361 20093 46000 83304 96624 62541 41722 79676 98970
53305 79544 99937 87727 32210 19438 58250 77265 02998 02973
57108 86498 14158 60697 41673 18087 46088 11238 82135 79035
08270 11929 92040 37390 71190 58952 98702 41638 95725 22798
90119 23206 75634 60053 90724 29080 69423 66815 11896 18607
45124 69607 17078 61747 15891 69904 79589 68137 19006 19045
83084 02589 37660 63882 99025 34831 92048 23671 68895 73795
04685 31035 93828 16159 05015 54800 76534 22974 13589 01801
61349 04538 89318 27693 02674 34368 24720 40682 20940 37392
14082 65020 49956 01336 41685 01758 49242 52122 01030 60378
82615 53477 58014 62229 72640 32042 73521 14166 45850 02372
50942 78633 16588 19275 62258 20773 67601 93065 69002 03985
76381 77455 81218 02520 22900 80130 61554 98901 26939 78732
05645 35063 85932 22410 31357 54790 39707 94348 11969 89755
76591 83750 46137 74989 39931 33068 35155 49486 28156 04556
31945 87960 04852 41411 63105 44116 95250 04046 59211 67270
08648 89822 04170 38365 23842 61917 57453 03495 61430 20154
32511 07999 18920 77045 44299 85057 51395 17457 24207 02730
79348 56194 58145 88645 84867 41594 28148 84985 89949 26689
61973 03660 32988 70689 17794 61340 58311 32569 23949 85626
92032 60127 34066 28149 22352 12907 53788 86648 57649 07887
74609 71072 63958 58336 67814 40598 12626 30754 75895 42194
98668 76074 25634 56913 88254 41647 05398 69463 49778 31382
65248 72078 58634 88678 21764 67940 45666 84664 35714 43081
82002 96916 94138 74739 99122 03904 46052 97277 60243 37424
79100 55938 23211 10111 17115 90577 94202 01063 85522 64378
30923 71710 70257 05596 42310 02449 31211 50025 99744 78084
90513 50966 78981 70391 45932 13535 21681 66589 94915 08855
94474 79356 16098 95806 79252 14190 88722 39887 15553 58386
65236 62948 19968 22071 49898 96140 80264 57580 56775 63138
80502 04192 84287 32589 50664 63846 71590 67220 71503 27942
01315 04632 50202 89148 41556 11584 35916 13979 25016 32511
81525 76670 88714 28681 56540 84963 85543 69715 86192 79373
19500 41720 79214 20079 42053 29844 02294 11306 78537 65098
25812 77090 45198 98162 13782 60596 99092 50188 65405 63227
80859 94220 92309 01998 45090 24815 13415 01677 39092
41107 33561 04376 40072 78909 61042 04098 73304 21892 63112
00465 00858 22774 80730 07098 80515 09970 40476 10314 24792
58137 02454 15657 24957 48401 02940 92828 26372 31071 58192
32013 97147 69725 78867 73329 74935 69276 46001 04181 38838
17048 84788 12531 01773 43551 34586 61239 87927 03232 31312
33935 07944 98456 11922 96174 24100 00307 85697 06527 34381
47633 49394 38673 22281 68096 76599 38462 16662 819,59 03358
82161 92521 10712 58839 18546 32920 89220 90493 737^5 22327
344 Appendix B
Table 7 (continued)
99 P50 30876 80821 14955 11495 25666 37656 91874 93051 64664
08090 84688 36332 86858 73763 62534 93378 54809 97076 09077
67619 00352 32735 56954 97851 57350 33068 35393 75938 86086
63779 66008 02516 93878 67930 38445 44166 20168 55128 65337
03259 72119 04797 95593 02754 87120 68167 04455 75318 93127
92914 02066 97320 00328 51685 89729 27446 32599 82486 01718
80001 70542 01530 63033 64348 01306 75419 90348 34717 05147
387.15 09824 86504 14817 74434 80450 95086 73824 40550 14266
15987 74578 12779 69608 76893 94840 36853 00568 35697 00783
06193 94893 24598 02714 69670 06153 97835 71087 58193 97912
40134 12803 33942 46660 05681 35209 65980 77899 38988 75580
88480 27598 48458 65369 81066 02000 68719 90488 50062 10428
49989 94369 80429 97152 67032 62342 96496 91274 71264 45271
62089 52111 92190 85413 95362 33400 03488 84666 99974 01459
01675 12741 94334 86069 71353 85566 16632 97577 18708 99550
04529 19798 47711 63262 06316 00287 86718 33705 31645 70615
63895 63087 91886 43467 55559 35912 39429 18933 75931 18924
17709 21642 56384 85699 24310 85043 00405 59820 54228 58645
11727 83872 22553 17012 02849 39794 50662 32647 67676 95488
02838 03160 92864 29385 63585 46055 41356 96398 70904 87103
62210 02385 73776 03171 83842 94602 31540 96071 55024 87629
16825 05535 99451 81864 99410 81211 62781 55121 62268 48522
05985 62766 58215 61900 53065 85082 88200 74393 24100 88379
14184 86400 41788 82932 27183 44744 14964 71718 76499 37364
95315 04537 85490 90542 42519 35659 87983 51941 20420 56828
65578 64820 95644 98074 72032 53443 92722 96373 36030 78053
18444 28477 01846 95805 91166 74383 55926 92971 99743 04905
03577 99361 21047 21971 71191 70493 70210 87051 94715 88924
49752 47015 09472 20089 90924 03674 73181 81104 95411 00656
32489 04936 30628 99512 40891 39832 25101 71757 77503 82112
76548 92824 53738 65890 78297 50705 96792 56841 41063 92875
26545 68726 06476 57444 35455 46706 40388 79728 99747 75076
67651 97346 75509 50270 27943 71144 15397 04565 95265 52236
67879 04880 01478 97239 32611 85024 37275 46399 59303 28341
96329 85824 79954 96263 91873 37394 45728 32769 72930 82361
87421 32587 32890 79171 54734 60628 53702 06741 98558 19167
22447 888^3 21866 39773 26018 28765 01876 03776 51523 89095
79589 92914 06964 43330 01726 30504 24797 52657 44098 22006
92123 79976 31751 68549 06147 38138 58792 80966 59767 24564
85909 35590 89231 75271 34409 48770 08980 54457 26022 29742
43162 44793 39006 76661 02000 14571 73986 96351 02276 47746
47549 41709 52412 40595 40397 38883 20843 90121 74897 96286
71711 75690 50441 41322 16497 36962 88880 45374 29836 82096
51091 24078 13706 27315 69918 06628 99964 09477 59496 90825
94981 73799 35590 58944 36581 94509 17508 31203 97030 28541
23778 02351 44843 28005 63835 69611 91360 20756 70188 02554
36324 01285 47959 40386 10284 03089 95441 77955 70381 04689
31710 55804 18079 15172 27321 93535 81303 97488 94531 61924
84106 55010 57902 09150 59719 52718 96632 22555 72411 85957
27527 60618 02688 95261 20022 88691 20488 93189 33658 49237
345 Appendix B
Appendix C
Further reading
Probability
Chapter 3 of this book provides an introduction to probability theory.
Those readers, who wish to pursue this topic, will find a more complete
treatment of it in many books including the following:
C hung, K. L. (1979), Elementary Probability Theory with Stochastic
Processes, 3rd edn, Springer-Verlag.
F e l l e r , W. (1968), A n Introduction to Probability Theory and its Applica
tions, vol. 1, 3rd edn, Wiley.
H o e l , P. G ., P o r t , S. C. and S t o n e , C. J. (1971), Introduction to
Probability Theory, Houghton Mifflin.
M e y e r , P. L. (1970), Introductory Probability and Statistical Applications,
2nd edn, Addison-Wesley.
Statistical methods
Statistical methods are discussed in many books, ranging in quality
from the very good to the not-so-good. Some of the older texts give
more detail regarding computation than is necessary today. The
following list gives some of the better applied texts. The author has
found the comprehensive book by Snedecor and Cochran particularly
useful. The books edited by Davies, and by Davies and Goldsmith, are
especially relevant to problems in the chemical industry.
B o w k e r , A . H . and L i e b e r m a n , G . J. (1972), Engineering Statistics, 2nd
edn, Prentice-Hall.
346 Appendix C
Box, G. E. P., H u n t e r , W . G. and H u n t e r , J. S. (1978), Statistics for
Experimenters, Wiley.
D a v i e s , O . L. (ed.) (1956), Design and Analysis of Industrial Experiments,
2nd edn, Oliver and Boyd.
D a v i e s , O . L. and G ol ds mi th , P. L. (eds) (1972), Statistical Methods in
Research and Production, 4th edn, Oliver and Boyd.
S n e d e c o r , G. W. and C o c h r a n , W. G. (1980), Statistical Methods, 7th
edn, Iowa Stale University Press.
S t o o d l e y , K. D. C., L e w i s , T. and S t a i n t o n , C. L. S . (1980), Applied
Statistical Techniques, Ellis Horwood.
W e t h e r i l l , G. B., (1981), Intermediate Statistical Methods, Chapman and
Hall.
W i n e , R. L. (1964), Statistics for Scientists and Engineers, Prentice-Hall.
Statistical theory
There are many books on statistical theory which provide a more
rigorous treatment of statistics than that attempted in this book. These
books include the following:
H oel , P. G . (1984), Introduction to Mathematical Statistics, 5th edn,
Wiley.
L i n d g r e n , B. W. (1976), Statistical Theory, 3rd edn, Macmillan.
Statistical tables
Most text books provide the more commonly used statistical tables
(see Appendix B), but there are many other sets of tables which are
readily available in libraries. Some of these books of tables are for one
specific purpose (e.g. Tables of the Binomial probability distribution),
while others are collections of different tables, which include the
following:
F i s h e r , R. A. and Y a t e s , F. (1963), Statistical Tables for Biological,
Agricultural and Medical Research, 6th edn, Oliver and Boyd.
N eave, H. R. (1978), Statistics Tables, George Allen and Unwin.
P e a r s o n , E. S. and H a r t l e y , H. O. (1966), Biometrika Tables for
Statisticians, vol. 1, 3rd edn, Cambridge University Press.
Operations research
In industrial and commercial organizations, many decisions have to be
made in the presence of competing demands and of uncertainty.
Operations research is an approach to problem-solving which helps in
choosing an effective course of action. The general approach is as
347 Appendix C
follows. First, an approximate mathematical model of the system
under study is developed. Then a measure of the effectiveness of the
system has to be chosen. Finally, the system variables are varied in the
model until an ‘optimum’ solution to the problem is found.
Both statisticians and technologists will find it useful to know
something about operations research. In addition to a knowledge of
probability and statistics, the study of operations research involves a
variety of other topics, particularly optimization techniques. Linear
programming is a mathematical technique for maximizing (or mini
mizing) a linear function of several variables, when these vari
ables are subject to linear constraints. This technique is useful, for
example, in allocating limited resources to competing demands in the
most effective way. Non-linear programming and dynamic pro
gramming are two other optimization methods.
Simulation is another important technique used in operations
research. This involves the construction of a mathematical model of
the relevant portion of the system under study. Then the model is used
to examine the effect of altering the variables which describe the
system without having to make costly changes to the real system. These
experiments are usually carried out on a computer.
Other topics of interest to the operations researcher include stock
control (or inventory control), queueing theory (see Section 5.6.3),
critical path scheduling, and game theory.
The following books are general texts on operations research:
H i l l i e r , F. S. and L i e b e r m a n , G. J. (1980), Introduction to Operations
Research, 3rd edn, Holden-Day.
T a h a , H. a . (1979), Operations Research - An Introduction, 2nd edn,
Prentice-Hall.
W a g n e r , H. M. (1975), Principles o f Operational Research, 2nd edn,
Prentice-Hall.
The following books deal with specialized aspects of operations
research:
B azaraa, M. S . and S h e t t y , C . M. (1979), Non-linear Programming -
Theory and Algorithms, Wiley.
C o o p e r , L. and C o o p e r , M. W. (1981), Introduction to Dynamic
Programming, Pergamon.
C ox, D. R. and S m i t h , W. L. (1961), Queues, Chapman and Hall.
H a s t i n g s , N. A. J. (1973), Dynamic Programming with Management
Applications, Butterworth.
K o l m a n , B. and B e c k , R. E. (1980), Elementary Linear Programming with
Applications, Academic Press.
348 Appendix C
L ew is, C. D . (1970), Scientific Inventory Control, Butterworth.
M organ, B. J. T. (1983), Elements o f Simulation, Chapman and Hall (to be
published).
P r i c e , W. L .(1971), Graphs and Networks: An Introduction, Butterworth.
W illiam s, H. P. (1978), Model-Building in Mathematical Programming,
Wiley.
The study of time series, and also of control theory (see Section
12.11), requires a knowledge of stochastic processes, which are physical
processes whose structure involves a random mechanism.
349 Appendix C
function of the data. The linear function is derived from data for a set
of individuals where it is known which group each individual belongs
to. Principal component analysis makes a linear transformation of the
variables to a new uncorrelated set of variables which are obtained in
order of decreasing importance from first to last, the aim being to
reduce the dimensionality of the situation. Factor analysis has a similar
aim but assumes an underlying model with a specified number of
factors which are linear combinations of the original variables. Multi
dimensional scaling is also concerned with reducing the number of
dimensions but is appropriate when the data arise directly or indirectly
in the form of ‘distances’ between the individuals. This technique can
cope with situations where the ‘distances’ are calculated from ordinal
data. Multivariate methods can be helpful but need to be used with
caution. For example factor analysis, in the author’s experience, rarely
gives results which can be interpreted in a useful way and is widely
misused in the social sciences.
An introduction to multivariate analysis is given in the following
books:
350 Appendix C
Appendix D
Some additional topics
This appendix considers a few topics which were omitted from the first
edition of this book. Most of them come broadly under the heading of
‘Descriptive Statistics’ and extend the work of Chapter 2. First, we
show how to calculate the mean and standard deviation of a frequency
distribution. Second, we discuss the relationship between sample
range and sample standard deviation, and its help in interpreting the
sample standard deviation as a measure of variability. Third, we
discuss how to round off numbers. Fourth, we introduce stem-and-leaf
and box plots.
The final topic discussed in this appendix is that of estimating and
testing a proportion, which extends the work of Chapters 6 and 7.
jc = ^ Xiln (D.l)
i=i
5 = X /(«-!) (D.2)
1= 1
X =
i H i f " (D 3)
351 Appendix D
Note that the summations in (D.3) are from 1 to capital N, and not
lower case n as in (D .l). The denominator in (D.3) is of course the
sample size, n.
It is easy to show that the sample variance of a frequency distribution
is given by
(D.4)
'■=1 i= i
=[ (n - 1) (D.5)
where
N
n = Y . f ‘-
1= 1
Table D .l.
Number o f
cosmic particles Frequency
Xi fi Xifi
0 13 0 0
1 13 13 13
2 8 16 32
3 5 15 45
4 1 4 16
Totals 40 48 106
Then
jc = 4 8 / 4 0 = 1* 2
5^ = ( 1 0 6 - 4 0 x 1*2^39
= 1-24
5 = 1-1.
352 Appendix D
In the continuous case, x and are calculated by assuming that all
the observations in a class interval fall at the mid-point of the interval.
As in the discrete case, two extra columns are added to the frequency
table to calculate the sums X fiXi and X /¿jc? where {x,} are now the class
marks. However this is usually achieved by employing a technique
called ‘coding’, which is an extension of the technique described in
Exercise 3 of Chapter 2.
First the mean of the given frequency distribution is guessed by eye.
The midpoint of the selected class interval is coded zero. If the class
intervals are all of equal width, then the mid-points of the two
neighbouring class intervals are coded +1 and —1 respectively.
Subsequent class intervals are coded ±2 and so on. Then the mean and
standard deviation of the coded values is calculated. From these it is
easy to see that the mean and standard deviation of the original data
can be calculated by the following formulae:
mean = (mid-point of the class interval which is coded zero)
-h (coded mean x class interval width),
standard deviation = coded standard deviation x class interval width.
Note that variance = coded variance x (class interval width)^. Also
note that the above formulae assume the class intervals are of equal
width. If they are not, then the coding technique needs to be modified
in an ‘obvious’ way by letting the most common class width correspond
to a code change of unity.
As an example let us consider the height data of Example 1 Chapter
2 (see p. 22). The distribution is roughly symmetric about the centre
class interval of (66-68) in. We therefore guess 67 in to be the trial
mean and code it zero. The mid-points of the other class intervals are
coded as shown in Table D.2. as the class intervals are of equal width.
Then two columns are needed to calculate X rfr and X r^fr where fr is
the frequency in the class interval whose mid-point is coded with the
value r. The coded mean is given by
Z/-/r/Z/. = 21/100
= 0-21
Now the class interval is the distance between successive class marks,
which is three inches in this case. Thus
353 Appendix D
The coded variance, on rewriting (D.5), is given by
= [1 0 5 -(2 1 )V l0 0 ]/9 9
= 1- 02 .
Thus the coded standard deviation is given by TOl. Then the standard
deviation of the original data is given by (TOl x 3) = 3-0 in.
Note that, although we have carried two decimal places in the
calculations, there is no point in presenting the final values for the
mean and standard deviation to more than one decimal place, as the
original data were measured to the nearest inch. (See the remarks in
Section D.3 on rounding.)
Table D.2.
60-62 -2 6 -12 24
63-65 -1 15 -1 5 15
66-68 0 40 0 0
69-71 +1 30 30 30
72-74 +2 9 18 36
354 Appendix D
look reasonable. This gives a check against arithmetical slips. Most
people can guess the mean of a set of data fairly accurately, but may
have trouble assessing the standard deviation. One way to do this is to
look at the range.
In Chapter 2 we pointed out that the range tends to get larger as the
sample size increases so that it does not provide a reliable measure of
spread by itself, particularly as it depends on just two outlying values.
However, it does provide a quick guide to the likely value of the
standard deviation. In particular, s must always be less than the range.
For a distribution which is roughly symmetric, the following guide
can be used:
where the symbol ’ means ‘is approximately equal to’. If, for
example, the sample size is 15, then interpolating in the above guide,
we have
5 —range/32
Problem 1
The Annual Report of a certain Building Society shows that their
assets rose from £697 432 057 in 1975 to £723 486 872 in 1976. Is it
really possible that a Building Society can specify its assets correctly to
the nearest pound, and would it not, in any case, be clearer to round the
numbers off in some way?
355 Appendix D
Problem 2
Seven observations are made on the time to failure, in hours, of a
certain type of machine. The results are as follows:
104-6, 152-3, 146-5, 130-7, 163-9, 115-6, 121-8.
The mean was calculated on a pocket calculator and was found to be
jc = 133-62857
as shown on the display. How many of the decimal places are worth
recording? These 2 examples highlight a topic which gives trouble to
many students and also to more advanced workers. That is the
question of deciding how many significant digits are needed for any
given number. This problem has grown worse in recent years with the
advent of pocket calculators, as these usually display far more digits
than are actually required. Yet surprisingly few statistics text-books
give any guidance on how to round off numbers. One exception is the
book by A. S. C. Ehrenberg {Data Reduction, 1975, Wiley), whose
work we draw on here.
The rounding problem may arise at three separate stages of a
statistical investigation, namely:
(a) Collecting the data. How many digits should be recorded?
(b) Carrying out the calculations. How many digits need to be ‘carried’
in order to get the results to a specific accuracy?
(c) Presenting the results. How many digits need to be given when
summarizing the analysis of the results?
When presenting data. Ehrenberg {Journal of the Royal Statistical
Society, Series A, 1977, p. 277) has stressed the need to give clear
summary tables and in particular has suggested the two-variable-digits
rule for rounding numbers. This says that data in general and summary
statistics in particular should be rounded to two variable digits, where a
variable digit is defined as one which varies in the kind of data under
consideration.
Let us see how this rule works in practice. Consider the number
181-633. This has six significant figures and three decimal places. But
how many variable digits does it contain? In fact it is impossible to
answer this question in isolation. Rather we must look at the other
observations in the sample. Suppose we have a sample consisting of the
following five observations:
181-633, 182-796, 189-132, 186-239, 191-151.
356 Appendix D
Here the first digit is always ‘1’, and so is not a variable digit. The
second digit is ‘8’ in four cases and ‘9 ’ in the fifth observation.
Although it varies a little, it does not vary over the whole range from ‘0 ’
to ‘9’ and so will not be counted as a variable digit. But the remaining 4
digits in each number could be anything from ‘0’ to ‘9’ and so are
variable digits. Using the two-variable-digit rule, this suggests that the
mean value should be given with just one decimal place, even though
the original data are recorded to three decimal places. Given the
variability in the data (a range of about 10), the two-variable-digit rule
is effectively saying that the digits in the second and third decimal
places give us very little information as we are unlikely to be interested
in variation of smaller magnitude than the range divided by 100.
Applying the rule to Problem 1, the first digit of the Building
Society’s assets would probably not be regarded as a variable digit as it
will change slowly over the years, but all subsequent digits are variable.
This suggests that the numbers should be rounded to the nearest
million pounds for presentation in the annual report. This is probably
what the reader did in the first place when he looked at the figures.
Admittedly some people may find the ‘exact’ figures more comforting,
but one should realise that it is highly unlikely that they are exactly
correct anyway. In addition, since nearly everyone reading the report
will mentally round the figures, it seems sensible to present the data in
rounded format to start with as this makes it easier for the reader. The
original ‘exact’ figures can be stored just in case they are needed.
Applying the two-variable-digit rule to Problem 2, we see that the
first digit is fixed, while the second varies between ‘0’ and ‘6’. The third
and fourth digits are certainly variable, but it is a matter of subjective
judgement as to whether the second digit is regarded as variable or not.
The mean should therefore be given with either zero or one decimal
place depending on one’s judgement. In the matter of rounding, a
degree of subjective judgement is inevitable and all guidelines will
have exceptions. Indeed some people may argue that it is impossible to
give general rules for rounding as there will always be occasional
exceptions. But surely it is better to have general guidelines, which one
can apply with judgement, than to have no guidelines at all. In problem
2 many people will feel ‘safer’ if they give the mean to one decimal
place, but, as we see below, a strong case can be made for rounding the
mean to the nearest integer.
The two-variable digit rule can also be applied to probabilities, as
these should normally be quoted to two decimal places. Occasionally,
additional digits may be desirable, as for example for probabilities
357 Appendix D
close to 0 or 1 when one may wish to see more accurately how close the
values are. Correlation coefficients should also usually be quoted to
two decimal places.
One situation where the two-variable digit rule does not apply,
concerns the presentation of test statistics such as i-values, F-ratios
and ;t^-values. There is no point in giving more than one or two
decimal places as this is the accuracy of the published tables of
percentage points. The observed test statistic is usually compared with
the tabulated percentage points to get an approximate level of
significance.
Having dealt with the presentation of data, let us now consider the
collection of data. I suggest that the two-variable-digit rule can also be
generally (though not always) used. Consider again the sample:
We have already suggested that the mean of the data should be given to
one decimal place. If the numbers were rounded to one decimal place
before carrying out any calculations, one would lose virtually no
information. In addition the data would be somewhat easier to handle
and much easier to understand if presented in the form:
358 Appendix D
calculate the standard deviation of a sample correct to one decimal
place, then one needs to calculate the sample variance correct to two
decimal places. In other words, we will generally need extra working
digits, so that we can round off accurately at the end of the calculations.
This does not mean that one will arbitrarily add half-a-dozen digits ‘to
be on the safe side’, but rather consider the computational aspects of
each problem as in Example 2 below. Usually one or two extra
significant digits will be adequate.
Let us ‘round off’ this section with two more examples.
Example 1
The cosmic particle data of Example 2, Chapter 1, consists of integers
between 0 and 4. Here there is only one recorded variable digit. The
mean and standard deviation need to be calculated to one decimal
place in order to give two-variable-digit accuracy.
Example 2
Consider a more sophisticated type of analysis, such as the regression
example on page 170. Here we want to estimate the two coefficients of
the regression line with appropriate accuracy. Although we want to see
how many significant digits are required, this is not really a rounding
problem but rather a numerical analysis problem. Nevertheless, it is
convenient to consider it here. The approximate argument presented
below is intended to clarify the general approach rather than be precise
numerical analysis.
The response variable, y, is recorded to two decimal places in the
data and contains less than two variable digits, so we want the fitted
regression line to give an estimate of the mean value of y which is
correct to at least two decimal places. The regression line has two
components; the intercept, d o, and ‘slope (a i)x controlled variable
(jc)’. We require each component to be accurate to three decimal places
in order that the sum is accurate to at least two decimal places. The
intercept d o , should therefore be calculated to three decimal places.
Now values of the controlled variable, x, range as high as 100. So we
need to calculate the slope correct to five decimal places in order that
its product with x is accurate to three decimal places. The slope itself is
a ratio of a sum of products and a sum of squares (see Equation 8.2). As
the denominator is exactly 3500, the numerator need only be cal
culated correct to two decimal places. In fact the numerator is exactly
7.9, so there is only one decimal place.
359 Appendix D
D .4 Stem -and-leaf plots and box plots
360 Appendix D
Table D.3
A stem-and-leaf plot of the thrust data of Example 1, Chapter 1
98 9 9
99 3 2 4
99 9 9 6 5 8 '7 6 8 9
100 3 2 0 3 2 1 4 0 2 0
100 6 8 6 7
101 2 4
361 Appendix D
40 50 60 70 80 unihs
Figure D,1 A box plot
observations, namely 47*5. The upper quartile is 77*5. The box plot is
then as shown in Figure D.l.
This plot demonstrates that the distribution may have rather short
‘tails’ in that the inter-quartile range is over half the range. The real
value of the plot is in comparing several groups. The sample given
above is the first of four groups given in Example 1 of Chapter 10
where we compared steel wire made by four manufacturers A, B, C and
D. The four box plots are shown in Figure D.2.
40 50 60 70 80 90 100 110
strength of wire
Figure D.2 Four box plots to compare the four manufacturers of steel wire
362 Appendix D
outlined in Appendix E.3. However, Tukey’s book contains much
material which is hard to follow or which seems too complicated for a
preliminary analysis. The reader is therefore advised to consult
Erickson and Nosanchuck (1977), McNeil (1977) or Tukey and
Mosteller (1977) for further details of these methods.
References
E rickso n , B. H. and N o s a n c h u c k , T. A. (1977), Understanding Data,
McGraw-Hill Ryerson.
M c N e i l , D. R. (1977), Interactive Data Analysis, Wiley.
T u k e y , J. W. (1977), Exploratory Data Analysis, Addison-Wesley.
T u k e y , J. W. and M o s t e l l e r , F. (1977), Data Analysis and Regression,
Addison-Wesley.
(A " -n p )M n p (l - p) ] = ( “ p) / ^ [P (l - p ) / n ] .
Thus
X
---- P
363 Appendix D
By a similar argument to that used on p. 126 to derive a confidence
interval for a mean, we find that a 100(1 - a )% confidence interval for
p is given approximately by
x/n±Za/2 ^[p{l-p)/n].
A difficulty with this formula is that the standard error of x/n, namely
V[p(l —p)/n], involves the quantity p which we are trying to estimate.
For p not close to 0 or 1, we usually replace p by its point estimate, x/n,
to give the confidence interval
x / n ± Z a / 2 ^[x{n - x) / n^] .
Example
In a random sample of 400 British voters, 160 people say that they
intend to vote Conservative at the next election. Find a 95%
confidence interval for the proportion in the whole population.
The sample proportion is 160/400 = 0*4.
Its estimated standard error is 7(160x240/400^) —0-025.
As n is large and p is not close to 0 or 1, we can use the normal
approximation. Thus an approximate 95% confidence interval for p is
given by
0-4±l-96x0-025.
The interval is approximately 0*35 to 0-45.
In order to test the hypothesis that p takes a particular value, say po,
we again use the normal approximation if n is large. The test statistic is
given by
z = ( x - npo)/^/[«Po(l -Po)]-
It can easily be shown that, if the null hypothesis is true, then z will
have a standard normal distribution. It is now straightforward to carry
out an appropriate test, depending on whether a one- or two-tailed
alternative is specified (see Chapter 7).
364 Appendix D
Appendix E
Some general comments on
tackling statistical problems
The reader who has studied the theory and methods described in this
book will now hope to be ready to tackle real statistical problems.
However, the reader will soon discover that life is not as simple as it
may appear in an introductory statistics course, where students are
usually shown only how to apply standard techniques to prepared
data. While such exercises are an essential step in learning statistics,
they do not prepare the student for the possibility that data may not
arrive in a neatly-arranged form with precise instructions as to the
appropriate form of analysis. In practice the data are often ‘messier’
than one would like, and the appropriate method of analysis may not
be obvious. Now that the computer has taken away much of the
drudgery in applying statistical techniques, the analyst can afford to
concentrate more on general strategy, such as selecting the most
appropriate method of analysis.
In this final Appendix, I will try to make some general comments on
how to tackle statistical problems. Of course experience is the best
teacher, but I hope that these remarks will be helpful. I also mention
topics such as ‘Presenting the Results’, which, while indispensable, are
often not covered in conventional statistics courses - see Chatfield (1988)
for more detailed coverage.
365 Appendix E
be carried out in the library. When similar studies have been done
before, a new study may be unnecessary or alternatively may be useful
so that established results can be generalized or updated.
The question of costs must be borne in mind at all stages.
366 Appendix E
Then, because the order of the experiments has not been randomized,
the effect of the explanatory variable on the response cannot be
separated from any time effect (the two effects are said to be
confounded). If time has no effect, then this does not matter, but if later
observations are systematically higher or lower, the estimates of the
regression equation will be biased. Unfortunately it is very easy for bias
effects to creep in unnoticed.
367 Appendix E
best om itted or treated separately because they are out of line with the
rest of the data. It is also worth getting a complete printout of the data as
the hum an eye is very efficient at spotting suspect values in a data array.
Sum m ary statistics may then be calculated for the d ata as a whole
and for im p o rtan t subgroups of the data. The statistics include the
m ean and stan d ard deviation of each m easured variable and the
correlation between each pair of variables. The distribution of each
variable may be exam ined by m eans of histogram s or stem -and-leaf
plots to see w hat distributional assum ptions are reasonable. For
example, an observed distribution which is approxim ately sym m etric
and bell-shaped may be approxim ated by a norm al distribution. Box
plots are useful for com paring the variability in different groups. The
scatter plots for pairs of variables should be plotted, and these may
help the analyst spot obvious relationships between variables, detect
outliers and also detect any clusters of observations. A variety of m ore
com plicated graphical procedures are available for m ultivariate data
(see Chatfield and Collins, 1980, Section 3.3). The possibility of
transform ing one or m ore of the variables should also be considered.
U nfortunately the various m otives for m aking a transform ation, such
as stabilizing variance, m aking effects additive and achieving norm ality
may be conflicting.
Sum m ary statistics and graphs are im portant not only in a
prelim inary analysis but also in the presentation of the conclusions.
Some obvious, but much ignored, rules are th at sum m ary statistics
should not contain too m any significant figures (see Appendix D.3),
th at graphs and tables should have a clear title and not be too spread
out, th at all axes in graphs should be labelled and th at when plotting
one variable against another, the scales should be chosen so th at the
variability in one direction is com parable to the variability in the other
direction.
A fter the prelim inary d ata analysis has been carried out, the analyst
should have a good ‘feel’ for the d ata and perhaps a better idea as to
w hat statistical techniques are appropriate. Indeed, in the a u th o r’s
experience, the prelim inary d ata analysis is often all th at is required.
To take just one example, the au th o r was once asked to carry out
regression and analysis of variance on the d ata shown in Figure E.l.
Two treatm ents were being com pared at different values of a regressor
variable, x, and the response variable, y, was m easured in each case.
The scientist who had collected the d ata wanted to dem onstrate that
treatm ent A was significantly different from treatm ent B. I pointed out
th at this was obvious in Figure E.l as there was no overlap at any
368 Appendix E
value of X. This m ade a significance test unnecessary (see the
addendum to C hapter 7). Of course, not all situations are as clear-cut
as this, but it is surprising how m any are.
JQ
to
fo
>
(U
U)
c
o
Q.
U
Q))
predictor variable (X ) ,
369 Appendix E
may need to be made to standard procedures in order to cope with
particular situations.
Of course, in theory it is desirable to have some idea of a suitable
method of analysis before collecting the data, but this is not always
possible before seeing the data. Sometimes the data will show
unexpected features of obvious importance. Alternatively, having
analysed the data, some discrepancies may result which question the
original choice of model or analysis procedure.
The main objective in many statistical analyses is the construction of
a probability model. The main motives for building probability models
are that they enable the analyst to make inferences from a sample to a
larger population and assess the uncertainty in any conclusions, and
also that they may provide a general parsimonious description of
different sets of data. When setting up a first attempt at a model, prior
knowledge and preliminary data analysis are both very useful. It
should be emphasized that all models are tentative and approximate:
tentative in the sense that they may be modified in the light of
subsequent information, and approximate in that no model is exactly
true. Having fitted a model and looked at the residuals, the model may
well have to be modified, sometimes by complicating it, sometimes by
simplifying it, and sometimes by making major changes in structure.
Indeed there are often a number of cycles in model-fitting as defects in
the original model assumptions are recognized, further data are
acquired, and the model is gradually improved.
One approach to statistical inference which has not been covered at
all in this book is Bayesian inference. This approach requires the
analyst to put his prior information about the problem into a
mathematical form and then use a probability theorem, called Bayes’
theorem, to update this prior information in the light of the data.
While some statisticians are strong advocates of this approach, this
author has rarely found the approach to be practicable. It is often
difficult or impossible to put one’s prior information into the
formalized mathematical form required for this approach. The
approach has therefore been deliberately omitted, although I recognize
that some specialists will not agree with this.
370 Appendix E
a set of data, he or she must decide whether to use a computer package,
to augment a published algorithm, or to write a special program. In the
latter case a choice must be made between a standard programming
language, such as FORTRAN, and a specialized language such as
GENSTAT which is particularly appropriate for complex statistical
problems. A choice must also be made between making the program
suitable for the given set of data only, or making it more general so
that it will cope with similar subsequent sets of data.
As regards computer packages, there are many available including
the SPSS and BMDP packages. If you have interactive facilities, then I
particularly recommend the MINITAB package which is easy to use and
covers many routine statistical techniques. It is worth stressing that most
packages are not intelligent enough to tell the user when the data are
unsuitable for the technique being tried. It is also worth pointing out that
even well-established packages may still have ‘bugs’ in them or do things
which are controversial to say the least.
371 Appendix E
1. Begin by planning the structure of the report by dividing the work
into sections arranged in a suitable order. Each section should be
given an appropriate title.
2. Include a clear summary.
3. Use simple English with appropriate punctuation.
4. Devote extra care to the clear presentation of graphs, tables and
computer output. It is amazing, for example, how often the scales of
graphs are not clearly labelled. Masses of computer output should
generally not be presented. Rather it is the writer’s job to extract the
important features of the output and present them in a clear way.
5. Be prepared to revise the report several times.
6. Leave the completed report for at least twenty-four hours and then
read it through again trying to put yourself in the place of a reader
who has not seen it before; if possible, get someone else to read the
report as well.
7. If the report is typed, check it carefully for errors as it is futile to try
to blame the secretary for any misprints.
References
C h a t f i e l d , C. (1988), Problem-Solving: A Statistician s Guide, Chapman
and Hall.
and C o l l i n s , A. J. (1980), Introduction to Multivariate
C h a t f ie l d , C.
Analysis, Chapman and Hall.
M o s e r , C. A . and FCa l t o n , G. (1971), Survey Methods in Social
Investigation, Heinemann Educational Books.
372 Appendix E
Answers to exercises
Chapter 2
2. (a) 4-6, 5-3, 2-3. (b) 104-6, 5-3, 2-3. (c) 0-46, 0.053, 0.23.
3. (a) 995-4, 5-3 (b) 130 is a suitable constant, ,x = 136-2, .s^ =11-8
Chapter 3
1 . (a) ( 1 - / 7,)( 1 - / 72) (b) l - ( l - p , ) ( l - / ^ 2)
(C) 1 - d - P l) ( l - P 2) - P lP 2
2. 0-99
3. (a) 5 (b) ^ (c) 5 (look at sample space)
4. (a) -j^ (b) (write out all possible outcomes)
5. (a) f (b) i (c )
6. 60 X 59
7. 26-’
8. 23
9. 0-583
10. (a) 0-72 (b) 0-921 (c) 0-950
11. l-dl^" = 0-67
Chapter 4
1. Binomial distribution; (a) 0*2 (b) 0-817 (c) 0*016
2. (a) P(0) = ( |) \ F (l) = 4 (|f(i). etc.
3. Number of defectives follows a binomial distribution with n = \0
and p = 0*04; 5*8%; very grave doubts — take another sample
immediately.
4. n ^ 69
5. 0-0085
6. Yes, P{2 additional failures) ^ 0 00000003
Yes, P(2 additional failures) ^ 0 00000001
So three-engine jet is three times as reliable in this situation.
Chapter 5
1. (a) 0-317 (b) 0-383 (c) 0-036
2. (a) 0-5 (b) 0-841 (c) 0159
3. (a) 0-954 (b) 0-023
4. 0-017
5. (a) 13-3 (b) 14-7;/c = 3-92
6. 97-7% will have a life-time >400 hours. Therefore purchaser will
be satisfied.
7. 13 questions (11-58 + 0-5, rounded up) using normal approxim
ation. The pass mark is also 13 using exact binomial values.
8. Expected proportion ofbolts less than 2-983 in. = 0 015:£(ni) = 4-5,
£(« 2) ^ 30-6. In second case solve /i-l-8 4 (j = 2-983 and
//+ 1-96(7 = 3 021, giving /i = 3-001, o = 0-010.
Chapter 6
1. (a) 0-66 (b) 0-4 (c) 0-2; sample size = 16
3. a = -0-52, h = 0-52, c - 0-37, d = 1-88
a = 4-95, h = 7-05, c = 1-48, d = 7-52
5. 14-1 ±2-0
6. 14-1 ±2-2
Chapter 7
1. Zq = 1-33; two-tailed level of significance > 5 % ; therefore accept
H q. 95% confidence interval for f.i is 12-22± 1-96 x 0-4/^6; note that
this interval contains 12-0.
2. s = 0-37, to = 1-45; two-tailed level of significance > 5 % ; accept
f/o- 93% confidence interval for ¡a is 12-22 ±2-57 xO-37/^6.
3. H o’.fA = 3 / / i : / i < 3 % ; one-tailed test; io = 3-16. But
'0-01 9 = 2-82; therefore reject H q at 1 % level.
4. n is smallest integer above 3^(1-64+ l -64)^/(32 —30)^; n = 25
Chapter 8
\. p = 0-514+ 00028x; y - 0-584
2. y = 13-46+1-13 Xcoded year ; y = 16-85
3. (a) 0*780 (b) height = -19*2 + 0.579 x weight
(c) weight = 81 *0 + 1*04 X height
4. y = -0*074 + 0*941x, + 0*598^2
Chapter 9
3^2-1 + = 0-381<
^3/
(mj + ^ 2 + ^3 + 3 ^ 4)
3.
^ 25/ti, 9^2
34 34
5. E{z) ^ 1-06 X 2-31; a,' ^ (0-04)^ x (2-31)' + (1 06)' x (0-05)';
tolerance limits are 2-45 + 0-32
Chapter 10
Total 233-6 23
Catalyst and agitation effects are highly significant; interaction not
significant. By inspection CE 3 is best combination, but this is only
marginally better than CE 2 which would probably be preferred on
economic grounds.
3. A : - 3 i B:0, C : - 7 , AB: - 2 , /1 C ;-1 , ABC:-\; A
effect significant at 5 % level, B effect not significant, C effect significant
at 1 % level
Chapter 12
1. x-chart has action lines at 50±15/(2-33 x ^ 5 ); K-chart has action
limit at 5 X 5-48/2-32.
2. x-chart has action lines at 10-80+3x0-46/(2-06x2); R-charX has
action limit at 0-46 x 5-31/2-06.
3. P(r defectives in sample size 30) ^ e~^^{0 9Y/r\; warning limit = 4
defectives. Action limit = 6 defectives.
4. 0-023;
P{r observations taken before one lies outside action lines) = (0-977)''" ^
X0-023;
mean = 1/0-023 = 43-5
5. (0-046)^ - 0-0021; (0160)^ = 0-026
6. Remove samples 11 and 12; revised values: x-chart: 15-2 + 2-5,
K-chart: 3-5. Roooi ^ 90; tolerance limits: 15-2 +51.
377 Index
C o n fid en ce interval (co n t.) D istrib u tio n (c o n t.)
for slo p e o f line 175 co n tin u o u s 64, 81
C o n fid en ce lim its 127 d iscrete uniform 115
C o n fo u n d in g 278 e x p o n en tia l 9 8 -1 0 3 , 323
C o n siste n t estim a to r 119 F - 155-7
C o n su m e r ’s risk 291 g a m m a 333
C o n tin g e n c y ta ble 152 g eo m etric 80, 102, 316
C o n tin u ity co rrectio n 95 h y p ergeom etric 80, 291
C o n tin u o u s m arginal 76
d a ta 20 n orm al, see N o r m a l d istrib u tio n
d istrib u tio n 64, 81 P o isso n , see P o is so n d istrib u tion
C o n to u r diagram 281 t- 1 2 9 -3 1 ,3 3 3
C o n tra st 272 uniform 83, 87
C o n tro l chart 299 W eib u ll 327
C o n tro l lim its 300 D istrib u tion -free tests 157
C o n tro lled variable 167 D o u b le sam p lin g 293
C o rrela tio n coefficient 1 8 5 -9 0 D y n a m ic p rogram m in g 3 4 8
in terp retation o f 196
theoretical 195 E fficiency 118
C o v a ria n ce 188 Error
C ritical p ath sc h e d u lin g 3 4 8 o f type 1 159, 162
C ritical v alu e 140, 158 o f type II 159, 162
C ro ssed d esig n 277 E stim ate or e stim a to r 29, 118
C u m u la tiv e d istrib u tio n fun ctio n 83 co n siste n t 119
C u m u la tiv e frequency diagram 2 6 -7 interval 107, 126
C u su m chart 3 0 6 -1 2 p oin t 107, 121
C u rve-fitting 167-71 u n b iased 32, 1 1 8 -1 9
E vents 39
D a ta d ep en d en t 4 2 - 6
c o n tin u o u s 20 in d ep en d en t 4 2 - 6
d iscrete 20 jo in t 44
p resen ta tio n o f 2 0 -3 4 m utu ally exclu siv e 4 0
D e c isio n interval 310 E v o lu tio n a ry o p era tio n (E vop ) 285
D eg re es o f freedom 116, 121, 229 E x p ec ta tio n or ex p ected valu e 6 6 -8 , 86,
for d istrib u tio n 116, 145, 149, 155 107-11
E xp erim en tal unit 225
for F -d istrib u tio n 155
E xp loratory d ata a n alysis 362
for i-d istr ib u tio n 130
E x p o n en tia l d istrib u tion 9 8 -1 0 3 , 323
in a n a ly sis o f variance 238, 2 4 8 ,2 6 6 ,2 7 4
E x p o n en tia l sm o o th in g 3 1 4
o fs^ 117, 121, 145
D ep en d en t variable 167 F -d istrib u tio n 1 5 5 -7
D esig n o f ex p erim en ts 185, 224, 260 rela tio n sh ip w ith x^ d istrib u tio n 334
D escrip tive sta tistics 20 F -test 1 5 5 -7
D iscrete F a cto r 225, 257
d a ta 20 F a cto rial exp erim en t, see C o m p le te
d istrib u tio n 56, 64 factorial exp erim en t
D istrib u tio n fun ction 83 F a cto rial n 50
D istrib u tio n F eed b ack c o n tro l system 312
b in o m ia l 5 7 -6 8 F ixed effects m odel 276
ch i-sq u are, see C h i-sq u a re d istrib u tio n F orw ard selection 199
378 index
F raction al factorial design 278 M arginal p rob ab ility d en sity fun ction 103
F requency curve 26, 30, 81 M a th em atical m odel 18, 37, 60, 106, 226
F requency d istrib u tion 21, 66, 75, 3 51 M axim u m lik elih o o d , m eth od o f 123
M ean
G a in 315
a rith m etic 28
G a m e theory 3 4 8 p o p u la tio n 65, 86
G a m m a d istrib u tio n 333 sam ple, see S am p le m ean
G a u ssia n d istrib u tio n , see N o rm a l M edian 30
d istrib u tio n
M ilitary standard plan 297
G e o m etr ic d istrib u tio n 80, 102, 316
M in im u m varian ce u nbiased estim ate
G o o d n ess-o f-fit test 14 8 -5 5 , 162
118, 120
G raphical m eth o d s 20
M I N IT A B 371
M ixed effects m odel 277
H azard fun ction 322
M o d e 30
H isto g ra m 2 2 -5
H o m o g en eity o f variances 244
M o d el, see M ath em atical m odel
M o m en ts, m eth od o f 121
H yp erg eo m etric d istrib u tio n 80, 291
M u ltip le regression 1 7 8 -8 0
H y p o th esis testing, see S ign ifican ce tests
M u ltiv a riate a n alysis 3 4 9
In co m p lete block design 253 M u tua lly exclu siv e even ts 40
In d ep en d en t even ts 4 2 - 6
In d e p e n d e n t v aria b le 167, 199 N ested d esign 277
Inference 18, 106 N o ise 312
Interaction 18, 179, 258, 262, 272 N o n -p a r a m etric tests 157
In terq u a rtile ran ge 3 1 , 361 N o rm a l d istrib u tion 8 7 -9 8 , 3 3 2 -4
Interval estim a te 107, 126 bivariate 187, 194 -6
m ean o f 88
Joint p ro b a b ility 76 standard 90, 130
uses o f 9 2 -5
Latin square 252 variance o f 88
L east significant difference m eth od 241 N o rm a l eq u a tio n s 169, 172, 178
L east squares, m eth od o f 168, 172, 2 1 6 -2 2 N o rm a l p rob ab ility paper 9 5 -8 , 280
L evel 257, 271 N u isa n ce factor 230
Level o f sign ifican ce 137 N u ll h yp o th esis 135
Life testin g 319
L ik elih o o d fun ction 123 O n e-a t-a -tim e exp erim en t 258
Linear p ro gram m in g 3 4 8 O n e-ta iled (o n e-sid ed ) test 139
L inear regression 177 O n e-w a y a n alysis o f variance 237-41
co n fid e n c e intervals and sign ifican ce O p era tin g ch aracteristic cu rve 2 8 9 -9 3
tests 1 7 4-7 O p erations research 347
fitting a straight line 1 6 8 -7 0 O p tim iz a tio n 2 8 0 -8 5
m odel 171, 173 O r th o g o n a l design 273
w ith tw o random variables 1 9 1 -4 O r th o g o n a l p o ly n o m ia ls 176, 180-85
Lot 37, 288 O u tlier 280
to lera n ce percen tage d efective 290
379 Index
P o isso n d istrib u tio n 6 9 -7 5 , 124, 151 R esid u als 243, 279
m ean o f 72 R esp o n se surface 281
variance o f 72 R esp o n se variab le 167
P o is so n p ro cess 71, 99 R o b u stn ess 158, 244
P o ly n o m ia l regression 1 7 9 -8 5 R o u n d in g 21, 3 5 5 - 9
P o p u la tio n 20
P o w er 1 5 8 -6 3 S a m p le 20
cu rve 161, 291 m ean 2 8 -9 , 6 4 -5 , 3 51
P recise m ea su rem en ts 204 d istrib u tio n o f 1 1 1 -1 6
P red ictio n interval 177 standard error o f 112
P r o b a b ility 39 p o in t 39
cu rv e 81 size, c h o ic e o f 162, 2 2 7 -3 0
d istrib u tio n 56 sp a ce 38
P ro b a b ility d en sity fu n ctio n 81 S a m p lin g d istrib u tio n 107
P ro b a b ility p aper (norm al) 9 5 -8 , 280 o f the sam p le m ean 1 1 1 -1 6
P ro cess avera g e p ercen ta g e d efectiv e 297 o f the sa m p le varian ce 1 1 6 -1 7
P ro cess co n tr o l 288, 299 S catter d iagram 166
P r o d u c e r ’s risk 290 S creen in g exp erim en t 271
P ro d u c t law 45 S eq u en tial sa m p lin g 294
P r o p a g a tio n o f error 2 0 5 -1 5 Service tim e 101
P ro p o rtio n 289, 363 S h ew h art chart 299
Sign test 157
Q u a lita tiv e factor 257 S ig n ifican ce tests 134
Q u a lity circles 288 co m p a rin g tw o sa m p le m ean s 1 4 3 -6
Q u a lity co n tr o l 2 8 8 -3 1 8 d istrib u tion -free 1 5 7 -8
Q u a n tita tiv e factor 257 g o o d n e ss-o f-fit 1 4 8 -5 5
Q u e u e in g th eo ry 101, 3 4 8 o n a sam p le m ean 14 0 -4 3
paired co m p a r iso n s 147 -8
S ign ifican t resu lt 1 3 7 - 8
R a n d o m effects m o d e l 276
S im u la tio n 3 4 8
R a n d o m n u m b er 115
S in g le sa m p lin g 293
g en era to r 115
S k ew d istrib u tio n s 25, 30
R a n d o m v a ria b le 5 1 -3
S p ecific a tio n lim its 304
R a n d o m iz a tio n 185, 230, 26 0
S tan d ard d e v ia tio n 3 1 -3 , 3 5 1 - 5
R a n d o m ize d b lo ck exp erim en t 244 p o p u la tio n 68
R a n g e 31, 235, 3 0 1 ,3 5 5
S tan d ard error 112
R ecta n g u la r d istrib u tio n , see U n iform S tan d ard n orm al d istrib u tio n 90, 130
d istrib u tio n
S ta tistic 28
R ectify in g sc h e m e 289, 296
S teep est ascen t, m eth o d o f 282
R ed u n d a n t m eter 2 1 5 -1 6 S te m -a n d -lea f p lots 21, 360
R eg ressio n 171 S to c h a stic p rocess 71, 3 4 9
cu rv e 167, 171 S tu d en tized range test 235
linear, see L inear regression S tu d e n t’s i-d istr ib u tio n 1 2 9 -3 1 , 333
m u ltip le 1 7 8 -8 0 Sum m ary sta tistics 28
p o ly n o m ia l 1 7 9 -8 5 Survival tim es 330
R egressor v ariab le 167 S y stem id e n tifica tio n 3 1 2 - 6
R elia b ility 18, 319
fu n ctio n 322 r-D istrib u tio n 129 -3 1 , 333
R ep lica tio n s, n u m b er o f 227 r-Test 142, 1 4 7 -8 , 162
380 Index
Taguchi m ethods 278, 317 V-mask 308
Taylor series 209 Variables, sam pling by 289, 298, 301
Test of significance, see Significance tests Variance 31-3
Test statistic 136 of a sum of independent random
Time-series analysis 349 variables 110
Tolerance limits 207, 304 of the sample mean 112
Total quality control 288 population 68, 87
Traffic intensity 103 Variate 20
Transformations 93, 95, 170, 180, 244
Treatment 225
Treatment com bination 257 Warning lines 3(X)
Two-tailed (two-sided) test 139 Weibull distribution 327
Two-way analysis of variance 248-52 Weights-and-Measures Act 317
Tukey’s T-method 242 W ilcoxon signed rank test 158
Tw o-sam ple tests 143-6 W insorization 280
381 Index
9780412253409
780412 253409