Robust Estimation of The Process Standard Deviation For Control Charts (Tatum 1997)
Robust Estimation of The Process Standard Deviation For Control Charts (Tatum 1997)
REFERENCES
Linked references are available on JSTOR for this article:
https://www.jstor.org/stable/1270901?seq=1&cid=pdf-
reference#references_tab_contents
You may need to log in to JSTOR to access the linked references.
JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide
range of content in a trusted digital archive. We use information technology and tools to increase productivity and
facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected].
Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at
https://about.jstor.org/terms
Taylor & Francis, Ltd. , American Statistical Association and are collaborating with JSTOR to
digitize, preserve and extend access to Technometrics
Control charts are tools used to detect aberrant behavior in manufacturing processes. The X chart
plots subsample averages as a function of time, and the R chart plots subsample ranges. Both
of these charts rely on an estimate of the standard deviation of the process when it is operating
correctly. The estimate has traditionally been based on the average range of 20-40 subgroups, but
this will produce an estimate that is biased high when outliers are present. One standard solution is
to construct a range chart for the original subgroups and estimate the standard deviation only from
those subgroups within the control limits, repeating the procedure as necessary. Proposals have
also recently been made to use a trimmed mean of the subsample ranges with a fixed percentage
of trimming, as well as the trimmed mean of the subsample interquartile ranges. This article
presents a new approach to robust estimation of the process standard deviation. The procedure first
centers each subsample on its own median and then applies a modified biweight A estimator to
the pooled residuals. This method combines the strengths of the previous methods-the relatively
high efficiency of the range-based methods when no disturbance is present, together with the strong
resistance to disturbances of the trimmed interquartile range method.
127
This article is organized as follows: The new estimatorofisthe subsample means for n = 5 and g = 30 for the iid
introduced in Section 2. A comparison of the robustnessNormal
of case was roughly a 5% increase in the estimation
the new method to that of existing methods is performed variance.
in
Section 3, based on the concept of breakdown bound. The This is not a trivial loss in efficiency, but it is small com
estimation efficiency of the methods under a wide varietypared to the cost of centering the subsamples, which sa
of situations is studied in Section 4. In Section 5, therifices
per- g - 1 df to protect against the possibility that t
subsamples have differing population means. For examp
formance of range charts based on the estimators is studied.
for n = 5 and g = 30, the standard deviation of the me
Three real datasets are analyzed in Section 6, and Section
7 gives recommendations and conclusions. subsample variance is more than 20% higher than that
the ordinary sample variance for the iid Normal case. Thus
2. THE PROPOSED METHOD my version of the biweight A estimator sacrifices efficien
The new method is constructed around a variant of the under the iid Normal case both by individually centeri
biweight A estimator. As given by Lax (1985), the biweight each subsample and by centering with the subsample m
A estimator of the standard deviation for a sample of size dian rather than the subsample mean.
n is Applying the biweight A to the pooled residuals gives
Sc r ( <n (xi- <l T)2(1 - U2)4) 2 m' (Efk EIuk,il<l Yki( - k,i)4) (5)
(1) c-(m -1)I/2 S i 2
w (n-1)1/2 a Eli< (1 -i)(1 -52) ' ' Z^k \ w here)(I k, )
where
where c is a tuning constant, T is the sample median,
Uk,j = Yk,j/(cM) (6)
ui = (x - T)/(c MAD), (2)
and
and MAD is the median absolute deviation from the med lian.
A = mediank,i |lk,i (7)
The biweight A estimator was well presented by Iglev vicz
(1983) and was shown in a major study by Lax (1985 ) to The process standard deviation, o, is then estimated as
perform well compared to other robust univariate scale es-
(T= S /kn.g-c
c n _q,c (8)
timators. Due to the (x- T)2 term in the numerator, the
biweight A could be termed a squared deviation-based Ies- where kn,g,c is a factor chosen that E(u-) = a. Let B7 denote
timator, as distinguished from the range- and IQR-bc ised S7, the biweight A estimator with tuning constant c = 7,
estimators described in Section 1. One advantage of w( ork- applied to the pooled residuals.
ing with a robust squared deviation-based estimator ra ther It will be shown in simulation studies that B7 performs
than a robust range-based estimator is that the former st arts very well compared to the earlier methods when dealing
with higher efficiency for the iid Normal case. with diffuse disturbances. When a localized disturbance is
The new method begins by removing the median vz alue present, however, the relative performance of B7 suffers be-
of each subsample from that subsample. When n is an e yven cause it does not make use of information on the location
number, this will yield a median-centered kth subsamp] le of the disturbed subsample(s). To find these subsamples,
I could use an estimator that examines and compares the
Yk, -- Xk,z -Xk, i= 1,...,n, (3) subsample ranges or variances, but this would confuse the
handling of localized disturbances with that of diffuse dis-
where xk is the subsample median.
turbances because the sample range and sample variance are
When n is an odd number, subtracting outsensitive
lian highly the to mecoutliers. Instead, I shall make use of the
will produce one zero value, which is dropped, givin
g a subsample IQR's, thus employing a measure of variability
centered subsample of size - 1, that is insensitive to occasional outliers. I shall use Rocke's
definition of the IQR, given in Equation (4).
Yk. X, -Xk, i< (n + 1)/2, ) Information concerning the location and severity of local-
x l Xk,+l-xk, (n + 1)/2 < i < n.
ized disturbances is communicated to the new estimator by
Will including avalues
The total number of median-centered subsample term hk in^the definition of Uk,j that will down-
be rn' = ng when n is even and m' n =
, weight
(n -subsamples
)g whei for which the IQR is unusually large:
is odd.
Uk,j = hkyk,3/(cAI), (9)
Removing the subsample medians as ea-a prelude to m
red where
suring variability ensures that the variability is measu
By subsamples.
within the subsamples and not between the 1, Ek < 4.5,
removing the median rather than the mean,
ing the center
hk= - (Ek 4.5), 4.5 < Ek < 7.5,
step is somewhat protected against outliers.
this Of course, t c, Ek > 7.5, (10)
protection must be paid for by a reduction
ion in the estimati
and,
efficiency for the iid Normal case. In ay, simulation
I stud'
found that the cost of using the subsample
ace medians in pl
Ek = IQRk/M. (11)
The second
Table 1. Normalizing Constants for the Modified Biweight A Estimator value, c = 7, gives an estimator that sacri-
fices some efficiency when no disturbances are present to
Tuning constants
gain efficiency when disturbances are present. In my Monte
7 10
Carlo studies, this choice of c gave the best estimator when
judged on overall performance.
Number of sub
Size 20 30 40 20 30 40
Values of dn,g,c are given ingreat care because of its obvious sensitivity to disturbances.
Table 1, for n =
The adaptive trimmed range, RATM, is often normalized
4,5,...,15,g = 20,30,40, and c = 7 and 10. The table
with D4, the appropriate factor only for R. This will re-
values are estimated from 100,000 simulations for each en-
sult in a small downward bias in the iid Normal case. For
try, running programs compiled in Turbo C? + + on an
my studies I shall employ a factor that produces an un-
NEC` 100-MHz Pentium4, using the random-number gen-
biased estimator of ca in the iid Normal case. Letting ATR
erator ran2 from Press, Teukolsky, Vetterling, and Flannery
denote the resulting estimator, then ATR = RATR/dATR.
(1992). The standard deviation of the estimates is less than
.0004. Because dATR is somewhat smaller than D4, ATR will
Each value of the tuning constant, c, yields a differentbe somewhat larger than R when no ranges have been
deleted.
estimator. I have studied the behavior of the estimator for
c= 7 and c = 10, denoted as D7 and D10. The motivation
for selecting c = 10 is that it leads to an estimator of a Table 2. Breakdown Bound of Scale Estimators (/%), g = 28
that slightly improves on the efficiency of the traditional -~Subsample ,Scale estimator
estimator, R, under the best of conditions for R-that is,Subsample
size (n) R ATR R25 RM IR IR25 M
when there are no disturbances in the data. The fact that
I can replace a traditional, nonrobust estimator with a ro- 4 .0 11.6 6.3 11.6 .9 13.4 24.1
5 .0 9.3 5.0 9.3 .7 10.7 29.3
bust estimator while improving efficiency in the iid Normal
6 .0 7.7 4.2 7.7 .6 8.9 24.4
case is rather surprising. The explanation is simply that our 7 .0 6.6 3.6 6.6 .5 7.7 28.1
method exploits the inefficiency of range-based estimators 8 .0 5.8 3.1 5.8 .9 10.3 24.6
for the iid Normal case as compared to squared deviation- 9 .0 5.2 2.8 5.2 .8 9.1 27.4
based estimators. 10 .0 4.6 2.5 4.6 .7 8.2 24.6
IR25-the
fall below (1/4) - 1/m, which is slightly less than 25% of 23.3% trimmed mean of the subsample IQR's
the observations. A proof is given in the Appendix.
Table 2 compares the breakdown bound for g = 28
Note that a = .233 for R25 and IR25 due to the integer
for R, ATR, R25, MR, IR, IR25, and M. The best break-
trimming method used.
down bound is achieved by M, with IR25 a distant second.
To provide an unbiased estimate of a for the iid Normal
Among the range methods, the highest breakdown bound
case, the values provided by each of the nine methods must
was shared (for this value of g) by RM and ATR, followed
be divided by a normalizing constant. The constants for D 10
by R25, R10, and R. It is somewhat surprising that R25
does not achieve the breakdown bound of ATR. Note that and D7 were provided by Table 1 and for R by standard ta-
bles, and the remainder were estimated from 100,000 ran-
the breakdown bounds for the range methods decrease as
dom simulations. The following values were used for the
n increases. The bounds on the IQR methods suffer a sim-
n = 5 case: B7, 1.069; D10, 1.054; D7, 1.069; R, 2.326;
ilar decrease, although there is a small improvement when
ATR, 2.314; R10, 2.297; R25, 2.272; RM, 2.263; IR, .9907;
moving from n = 7 to n = 8.
IR25, .9239. For n = 10, the constants were B7, .963; D10,
The drawback to working with IR25 is that it is a rel-
.956; D7, .963; R, 3.077; ATR, 3.035; R10, 3.057; R25,
atively inefficient estimator of the standard deviation of a
3.037; RM, 3.026; IR, 1.312; IR25, 1.280.
Normal distribution. A simulation study in Section 4 finds
Comparisons were made under four types of distur-
that the mean squared estimation error for IR25 is sev-
bances-
eral times greater than that of the trimmed range methods.
Hence, when no disturbance is present and the core pro- 1. a 95-5 mixture model of symmetric variance dis
cess can be modeled as Normally distributed, the use of bances in which each observation Xk,i has a 95% prob
IR25 will be costly in terms of estimation error compared ity of being drawn from the N(0, 1) distribution and
to other methods. Used alone, M would suffer from theprobability of being drawn from the N(0, a) distribu
same problem, and hence I employ it only as a preliminarywhere a = 1.0,1.5,...,5.5,6.0;
estimator. RM and ATR will be found to highly efficient2. a model for asymmetric variance disturbances in
compared to IR25 for the iid Normal case and quite com- which each observation is drawn from the N(0, 1) distri-
petitive with IR25 in many of the situations studied. Indeed,
bution and then has a 5% probability of having a multiple
in my studies ATR gave one of the strongest performances of a X2(1) random variable added to it (the multiplier takes
of the existing methods. values of 0, .5, 1.0, 1.5, ..., 5.0);
3. a model for localized, symmetric variance distur-
bances in which the observations in one subsample are
4. SIMULATION STUDIES OF
drawn from the N(0, a) distribution, a = 1.0,..., 6.0, and
ESTIMATION EFFICIENCY
the remaining subsamples contain observations drawn from
In the following simulation studies, I shall compare the
the N(0, 1) distribution;
mean squared error (MSE) of the estimators for g4. =a 95-530 mixture model of diffuse mean disturbances
subsamples and for n = 5 and n = 10. The MSE in will
whichbe each observation has a 95% probability of be-
estimated as ing drawn from the N(O, 1) distribution and a 5% prob-
ability of being drawn from an N(b, 1) distribution, with
b = 0,.5,...,9.0.9.5.
MSE1 o)2 The following results are based on 10,000 simulations for
MSE - Z- (i=1i=1
each level of disturbance for each type of disturbance.
.020
.024 -
.020
.016
.016
2 .012
.012
.008
.008
.004 .004
2 3 4 5 6 1 2 3 4
R10 R25
.012 - i RM
.010 -
.008 -
w
vt
.006 -
.004 -
.002
1 2 3 4 5 6 0 1 2 3 4 5
Figure 1. Mean Squared Error of Estimators (MSE) Versus the Stan- Figure 2. Mean
dard Deviation of a Diffuse, Symmetric Variance Disturbance, Modeledple of a Diffus
Mixture
by 95-5 Mixture of Normal(0, 1) and Normal(0, a) Distributions. Values of the
of a are plotted on the horizontal axis. The estimators are D10, a mod-
X2 (1) Distribut
The
ified biweight A with tuning constant c = 10; D7, c = 7; B7, a biweight estimators
A estimator with c = 7; R, the average subsample range; ATR, adaptive
trimmed average of subsample ranges; R10, 10% trimmed average Type of II: Asy
subsample ranges; R25, 25% trimmed ranges; RM, median subsample
range; IR, average subsample IQR; IR25, 25% trimmed average sub- Figure 2, (a
sample IQR. n = 10 for a
quite similar
the strong p
and IR25. The MSE of the range-based methods increases peated. Wit
much more quickly than for n = 5, which appears to reflect
inates all pr
the fact that the breakdown bound of these methods actu- although D7
ally decreases as the number of observations per subsample This catego
goes up. For n = 10, the 5% of observations subject to dis-
monoff (1987
turbance is now equal to the breakdown bound of RM, the well for asy
most robust of the range-based methods. Note also that B7 tion studies.
and D7 gave nearly identical performances. the competin
.007
signed to handle. Although B7 and D7 were indistinguish-
able in previous types of disturbances, for localized distur-
bances D7 has a decided advantage over B7.
In Figure 3(b), with n = 10, D10 is able to outperform
R10 and all other methods for almost the entire range of the
standard deviation of the disturbance. Again, B7 performs
.006
relatively poorly compared to D7 for n = 10, and ATR does
not perform as well as with n = 5.
.080
.070
1 2 3 4 5 6
.050
.004
.040
.030
B7
.020
_.- _ -
_ - _- ", _
v: .003 .010
.000
0 1 2 3 4 5 6 7 8 9
.022
.002
1 2 3 4 5 6
.018
(b) Standard Deviation of Disturbance
Scale estimator
performances werestantial
given by
improvements D7 methods
over existing andin B7. F
terms of th
n, IR25 outperformed D10
estimation by a
of the underlying scalelarge mar
parameter, but it remai
ate shift in mean, but D10
to be shown performed
that this translates into improved much
control-cha
and large shifts. IR25 also
performance. outperforms
To do so, I examined the performance o
range charts4(a),
brief stretch in Figure based on the various robustwas
which estimatorsthof
happened in my studies.
I have limited the analysis to a subsample size of n =
The only change in the definition of the disturbances
Relative that the Type
Squared Mean III now increases the variance of 3 out of
Error
subsamples rather
To compare the overall performance than just 1 out of 30. This wasof
done
an effort to put the
tors, I found the relative mean squared etrimmed range methods on a stronge
footing.
each level of disturbance for each type of d
RMSE of
estimator The two most
an was common methods
found of comparing the
by per- divid
formance of control
that estimator into the MSE of the methodcharts is to compare the average rate
at which positive signals areThe
best under those conditions. encountered results
and the average are
Table 3 by reportingrun length
the (ARL) until
lowesta positive signal is RMSE
given. A positive achie
signal is deemed to
timator across the levels of disturbance,have occurred if the range of a subsam- by
of n. For example, ple exceeded theentry
the UCL. The weakness ofin
these comparisons
the first
is that they attempt to distinguish the behavior of random
column indicates that for the Type I distur
the RMSE quantities by
did of nota comparisonfall
D10 of the average outcome, thus
below 64%
failing to take into account the
deviation of the disturbance variability of the random from
increased
The best performancequantities involved.
in In terms
eachof the analysis
row done in Sec-of Ta
in boldface and wastion 4, supplied
this would be equivalent to limiting the comparison
either by D
to only the mean error (or bias) of the estimators.
As an overall measure of performance for a
last row 3 My of basis of comparison willthe
reports Table be the meanminimum
squared de-
viation (MSD)
down the columns of the four typesbetween the rate of positive signals for an
of dist
estimator
two subsample sizes. The and the rate
D7 of positive signals that would have
estimator had
occurred if ao was known. That is, I treat the latter rate as
performance withan 86% minimum RMSE,
with a 75% a target rate and measure
minimum RMSE the quality of theand
estimators by D10 a
method had the bestthe performance
squared deviation from that target. Unlike of measuringthe pr
the squared deviation from a fixed target, such as was done
achieving a minimum RMSE of 26%. ATR
performance except using
for r in the previous
the study,Type
here I shall be using
IV a mov-disturb
ing target because
The overall efficiency of the rate
86%of positive signals
for will depend
D7 is r
on the sample realization even when a is known.
sistent with the results of comparative s
weight A estimator The first
as step given
in comparing the estimators
both was to adjust by La
their control limits
Iglewicz (1983). In both ofso that they had approximately
these equal av-
studies,
weight A estimator erage false positive signal rates under the
achieved a iidminimum
Normal case.
across three When a is known, setting the UCL at UCL(a) = 4.918o
distributions.
yields a false positive rate of .602% (see Duncan 1974, p.
436). Consequently, I modified the estimators so that each
5. COMPARISONS OF CONTROL-CHART
PERFORMANCE
had a false positive rate of approximately .602%. This was
accomplished by setting the UCL equal to the following
Section 4 demonstrates that the new methods offer sub-
multiples of the estimators: D10, 4.925; D7, 4.940; B7,
TECHNOMETRICS, MAY 1997, VOL. 39, NO. 2
1000 -
800 -
600
400
200
1 2 3 4 5 6 0 1 2 3 4 5
280
230
180
130
80
30
1 2 3 4 5 6 0 1 2 3 4 5 6 7 8 9
4.940; R, 4.92; ATR, 4.94; R10, 4.98; R25, 5.037; RM, uing the ith trial, 1,000 subsamples of size 5 were generated
5.057; IR, 5.114; IR25, 5.215. These values were arrived from the same type of disturbance and the same level of dis-
at by numerous simulation studies. turbance. Let S(D7, i) denote the number of times that the
The MSD of each of the four types of disturbances subsample range exceeded UCL(D7, i) during the ith trial.
and each level of disturbance was found by the followingIn addition, let S(a, i) denote the number of times that the
method: For each of 1,000 simulation trials, 30 subsamplessubsample range exceeded the "correct" UCL of UCL(a)
of size 5 were generated from a particular type of distur-during the ith trial. The MSD for D7 is then given by
bance and level of disturbance. Estimates of a were then 1,000
formed by each of the 10 estimators, and these estimators MSD(D7) = (1/1, 000) E [S(D7, i)-S(n, i)]2. (21)
i=l
were used to form upper control-chart limits. For example,
let D7i denote the value of D7 on the ith simulation trial, Overall, because 1,000 simulations of 1,00
with corresponding UCL of UCL(D7, i) = 4.94D7i. Contin-were involved, the value of each MSD was based on re-
TECHNOMETRICS, MAY 1997, VOL. 39, NO. 2
Table 4. Minimum Relative Mean Squared Deviation (RMSD) showing the residuals from removing the subsample medi-
ans. Singular residual observations are denoted with a cir
Estimator
cle, and multiple residuals are denoted with a square. Two
Type D10 D7 B7 R ATR R10 R25 RM IR IR25 kinds of graphical diagnostics have been added to the resid-
I 71 92 76 13 76 31 47 41 45 27 ual plots: Individual residuals that are more than three stan-
II 87 87 88 12 72 45 61 41 40 27 dard deviations from 0 as estimated by D7 are marked by
III 69 82 66 14 60 82 79 48 10 23
filled circles, and a double-pointed arrow indicates the pres-
IV 26 84 73 0 2 0 1 7 2 14
ence of an excessive IQR within a subsample. The latter are
Min 26 82 66 0 2 0 1 7 2 14
instances in which Ek > 4.5, where Ek is the standardized
NOTE: Each entry represents the IQRlowest
for the kth subsample,mean
relative as given squared
in Equation (16).
devia
the range chart based on an estimator, for a type of disturbance an
One of the difficulties with these datasets is the high level
measures the variation between the rate of positive signals for a UCL b
for a UCL based on knowledge of of discretizing
a. The best of theperformance
data, which I presume inreflects
each mea
ro
an overall measure of performance, the last row gives the minimum o
surement problems rather
of disturbance and the estimators are as given in Table 3.
than data simplification. The dis-
cretizing is of a large magnitude relative to the variability
of the observations, leading to numerous duplications. In
alizations of one million subsamples. In th
each subsample was addition,
given for one dataset the discretizing was heavilyof
a probability bi-
being disturbed. ased toward trailing digits of 0 or 5. Thus, there are sev-
eral problems present in the real data that did not exist
The results for the Type I, II, III, and IV dist
in
shown in Figure 5, (a), my simulation
(b), studies-measurement
(c), and error, (d), rounding
res or
discretizing error, frequent duplication of observations, and
parisons were limited to the case of n = 5
biased rounding.
figures are not identical to those of the M
I first consider a set of measurements of the melt in-
of Section 4, the same general results are fou
dex of a polyethylene compound analyzed by Wadsworth,
D7 and B7 give the best overall performances
Stephens, and Godfrey (1986, pp. 207-209). The authors
I, Type II, and Type IV disturbances, as see
will hereafter be referred to as WSG. The data consist of
(a), (b), and (d). Again, the trimmed range me
well for the 20 subsamples
localized of size 4. The residuals (from
disturbance case removing
in the
subsample medians) are shown in Figure 6(a), and the sub-
though it is difficult to distinguish the meth
5(c), D7 again sample rangeswell
performs are plotted in Figure 6(b) with four
against B7 candidate
fo
UCL's from estimators D7, R, ATR, and IR25.
disturbance. Finally, ATR performs remarkab
The range of the third subsample is an outlier even when
for the Type IV case. In Figure 5(d), the ve
the the control
log of the MSD to limit is given by R.the
improve Using auxiliary
presentrules for de-
tecting out-of-control behavior, WSG identified the fourth
60 - pie are equal in value (as shown by the use of the square
symbol) and the last value seems to be an average distance
4
away. Subsample 4 also seems to have an inflated range
40 -
caused by an outlier. I would also point out that D7 did not
downweight any subsample as having an excessive IQR-a
further suggestion that changes in process variance are not
the problem in this dataset.
20- 0o
o Hence, I have, in this real dataset, evidence that diffuse
0 0 0 0 disturbances exist in actual industrial settings. Given the
o0O0
oo o o
o
0 o o o oo presence of diffuse disturbances, eliminating all observ
0
8 'C 0 0 008T } o ?B tions in a subsample simply because it contains a sing
0 <u 0 outlier will reduce the efficiency of estimators. In addition
0
0 in terms of operating a control chart it would be mislea
-20- ing to treat all inflated subsample ranges as indicating an
increased process variance if in fact some of these reflected
I
the existence of isolated outliers. The sources of the distur-
-40
bances will be different and so will be the remedies.
3 -
-15 I 4 ?
1 3 5 7 9 11 13 15 17 19 C
0
2-
(a) Sample 0
0
0
0
o
1 - 0 o o0
20
o 0 ?
>o
0 c zoO ?
'0 0 0 < 0a
cn o0 a? 0 >0
c/ 0 o o
15 0 0
? 0 C ) 0 o
R o o
-1- 0 0 0 0
A
0 0 0
D7 c
0 -2
10- -IR25 -- -- - -.. . .. .-- =
ATR 0
0
0 -3
0 O 1 3 5 7 9 11 13 15 17 19 21 23
5 0 0
o o 00 0 (a) Sample
0 0
0 0
0
5- ATR
1 3 5 7 9 11 13 15 17 19
o R
(b) Sample
4- ........... . ....................... .... .....
D7
Figure 7 (a) Pitch Diameter of Threads; Residuals From Removing
the Subsample Medians. Squares denote multiple values, solid circles 0 00
2.16; and IR25, 2.02. The UCL values were set as 4.918&
0 0
using the factor D4 of table H of WSG.
0
My third dataset is a set of measurements of cotton yarn 0 0
count from pages 229-230 of WSG. This dataset was also 0 0 0
It can be shown that the median is much more sensi- control events for which an assignable cause can be found.
tive to rounding errors than are M estimators of location A corollary would seem to be that no subsample range
(see chap. 11D of Hoaglin et al. 1983). Although I shall not
should be deleted if no assignable cause can be found. Al-
offer a formal analysis, I suggest that a parallel situation ex-
though the human-directed investigation of out-of-control
ists in this context: Like the median, estimators such as IR,points is essential to improving a process, it is important to
IR25, and M will be more sensitive to rounding errors thanhave a measure of the variability of the core process that is
are estimators based on M estimators or on the asymptotic independent of the skill or reliability with which operators
variance of M estimators.
are able to determine the special cause for each suspected
7. CONCLUSIONS AND RECOMMENDATIONS out-of-control event. There must be limits on the investiga-
tive skill of operators: Will operators ever fail to find a
The modified version of the biweight A estimator pro-
special cause for a false positive signal? And, if investiga-
vides estimators that combine the best features of the pre-
tion of an unusual subsample range fails to find a special
vious methods-the relatively high efficiency of the range-
cause, must that range be included in the estimator?
based methods when no disturbance is present, together
There is no inherent reason why the estimation of ua is
with the strong resistance to disturbances of the trimmed
best done in imitation of the steps taken when running an
IQR method. By focusing attention on the pooled residu-
established control chart. The classification of events as ei-
als, I am able to achieve a much higher breakdown bound
ther in-control or out-of-control encourages the use of es-
than in previous robust methods. I am also able to gather
timators based on all-or-nothing techniques of deletion. It
together sufficient numbers of points to exploit the inherent
is generally true that such abrupt treatment of the data pro-
advantages of a squared deviation-based estimator found
duces estimators that lack the efficiency of methods that
when working in a basically Normally distributed setting.
allow a smooth downweighting of extreme events; my sim-
I believe that the types of disturbances studied were rea-
ulation studies confirm that this holds true in the present
sonable facsimiles of several major types of problems
context.
that
arise when initializing control charts. Given the strong per-
formance of the new methods under these conditions and
the fact that little or no efficiency is lost when there are ACKNOWLEDGMENTS
no disturbances, I would recommend the use of D10 or D7
I thank the referees, the associate editor, and the editor
over previous methods.
In general, I would recommend D7 over D10 because of for their help in improving the original submission.
its superior overall performance. In a situation in which it
is known that the production process is seldom out of con-
APPENDIX: BREAKDOWN BOUNDS
trol, however, D10 would be preferred because it is a better
estimator of u when few or no disturbances are present. The lowest breakdown bound of M is 1/4 - /m, given
One surprise for me was that the adaptive trimmer ATR, n+g > 3.
a variant of a venerable technique, generally outperformed Proof: The breakdown bound of M will be b* = p/m,
the more recently proposed trimmed range methods such where p is the largest number of observations that can
as RIO and R25. This approach deserves greater study, al- be manipulated to take arbitrary values while M1 remains
though it seems highly unlikely that any subsample range- bounded. The MAD's will break down, roughly speaking,
based method could be competitive when faced with awhen half the values are manipulated. The fastest way to
mean-shift disturbance. One promising line of attack would break down M is to swamp half the subsamples by manip-
be a version of the backward-stepping methods of Simonoff ulating the minimum number of points needed to control
(1987), in which one would begin with a presumably reli- the subsample median in each. This, in turn, takes approx-
able central core of subsample ranges and work out fromimately half the points in each such subsample so that the
there. This approach avoids the masking problem commonbreakdown bound will be about 25%. More precisely, the
to ordinary adaptive trimmers. bound depends on whether n and g are even or odd, and
A robust estimator such as D7 has the ability to perform these four cases are analyzed here.
well under adverse circumstances, but it is not intended The median of an even number of values will be defined
to lessen interest in the causes of those adverse circum- as the average of the middle two ordered observations.
stances. Improving a production process requires getting to Case I: n Even, g Even. The total number of centered
the bottom of each problem as efficiently as possible; I have observations, m', will be even, and therefore the median of
demonstrated a robust estimator that gets to the bottom of the absolute values of the centered observations, M, will
the question of the variability of the core process. This will remain bounded as long as fewer than m'/2 centered ob-
not answer the question of what is causing the disturbances servations are manipulated. To control Xk, the median of
nor will it make the disturbances go away; it will help tothe kth subsample, will require the control of n/2 points
locate the problems faster and more reliably and aid in dis- in that subsample. Thus, manipulation of n/2 points in the
tinguishing diffuse from localized problems. kth subsample will suffice to drive the absolute values of all
An underlying theme found in several textbooks is that n values of the kth centered subsample beyond any bound.
the variance of the core process should be estimated fromTherefore, m'/2 centered observations can be controlled
the average of the ranges remaining after deleting all out-of- with (n/2)(g/2) original observations so that the break-
down bound will be as n + g > 3 because this ensures that 1/(4g) + 1/(4n) -
7/(4m) > -l/m. Therefore, the lowest breakdown bound
b*
b -(n/2)(g/2)-1
::=-m- 1 (A.1) (A.1)
1/rn. for M is 1/4 - /m, for n + g > 3.
mrn 4
b*
[(n+ 1)/2](- 1)/2 + (n- 1)/2- Communications in Statistics, Part A-Theory and Methods, 6, 879-894.
m
Shoemaker, L. H., and Hettmansperger, T. P. (1982), "Robust Estimates
and Tests for the One- and Two-Sample Scale Models," Biometrika, 69,
ng- n + g - 1 + 2n- 6 47-53.