Lectura 1
Lectura 1
Author(s): D. R. Cox
Reviewed work(s):
Source: International Statistical Review / Revue Internationale de Statistique, Vol. 52, No. 1
(Apr., 1984), pp. 1-24
Published by: International Statistical Institute (ISI)
Stable URL: http://www.jstor.org/stable/1403235 .
Accessed: 25/07/2012 12:48
Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at .
http://www.jstor.org/page/info/about/policies/terms.jsp
.
JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of
content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms
of scholarship. For more information about JSTOR, please contact [email protected].
International Statistical Institute (ISI) is collaborating with JSTOR to digitize, preserve and extend access to
International Statistical Review / Revue Internationale de Statistique.
http://www.jstor.org
International Statistical Review, (1984), 52, 1, pp. 1-31. Printed in Great Britain
? International Statistical Institute
Interaction
D.R. Cox
Department of Mathematics, Imperial College, London SW7 2BZ, U.K.
Summary
A broad review is given of various aspects of interaction. After a brief discussion of the relation
between absence of interaction and constancy of variance, the main part of the paper considers
interaction in the usual sense, connected with the structure of an expected value in multiply
classified data. Interactions are first classified depending on the character of the factors involved.
Techniques for detecting interactions are reviewed, including aspects of model formulation,
especially for unbalanced data. Special methods for interpreting interactions are described. Finally,
the relation between interactions and error estimation is considered and some short remarks made
on extensions to binary and other essentially nonnormal models.
Key words: Balanced design; Component of variance; Contrast; F test; Factorial experiment;
Interaction;Latin square;Log linear model; Main effect; Transformation;Unbalanceddesign.
1 Introduction
The notion of interaction and indeed the very word itself are widely used in scientific
discussion. This is largely because of the relation between interaction and causal connex-
ion. Interaction in the statistical sense has, however, a more specialized meaning related,
although often in only a rather vague way, to the more general notion.
The object of the present paper is to review the statistical aspects of interaction. The
main discussion falls under three broad headings:
(i) the definition of interaction and the reasons for its importance;
(ii) the detection of interaction;
(iii) the interpretation and application of interaction.
For the most part the discussion is in terms of the linear and quadratic statistics familiar in
the analysis of variance and linear regression analysis, although most of the ideas apply
much more generally, for instance to various analyses connected with the exponential
family, including logistic and other analyses for binary data. First, however, in ? 2 some
brief comments are made on the connexion between constancy of distributional shape and
constancy of variance and the absence of interaction.
The paper presupposes some familiarity with the broad issues involved. Emphasis is
placed on matters, for example of interpretation, that, however widely known, are not
extensively discussed in the literature. References are confined to key ones.
Although some emphasis will be put on data from balanced experimental designs, in
principle the ideas apply equally to observational data and to unbalanced data.
variance may suggest that a more elaborate method of analysis involving weighting is
desirable, but does not by itself change the central objectives of the analysis. In some
contexts, however, the form of the frequency distribution is of some primary interest and
there is then a relation with notions of interaction in the following sense.
Consider first an experiment with just two treatments. Under the simplest assumption
of unit-treatment additivity, the response obtained on a particular experimental unit
under treatment 2 differs by a constant 7, say, from the observation that would have been
obtained on that unit under treatment 1. This implies that if F2(.) and FI(.) are the
corresponding cumulative distribution functions of response then
F2(y) = - 7). (2.1)
FI(y
That is, the two distributions differ only by translation and in particular have the same
variance. Any departure from (2.1) implies that simple unit-treatment additivity cannot
hold; in other words, to match the two distributions in (2.1), 7 must be a function of y,
7(y) say. That is, the treatment effect depends on the initial level of response and this is
one form of interaction.
Consistency with (2.1) can be examined by inspecting histograms or cumulative dis-
tribution functions, but preferably via a o-- plot (Wilk & Gnanadesikan, 1968) in which
corresponding quantiles of the two distributions are plotted against one another to
produce under (2.1) a line of unit slope.
Doksum & Sievers (1976) have given formal procedures based on this and in particular
a confidence band for 7(y) defined by
F2(y) = F1{y- (y)}, r(y)= y-F71F2(y).
If (2.1) does not hold, it is natural to consider whether a monotonic transformation from y
to h(y) is available such that h(Y) satisfies (2.1). G.E.H. Reuter (unpublished) has shown
that such a transformation exists if F, and F2 do not intersect. If there are more than two
'treatments' transformation to a location family representation exists only exceptionally.
A more traditional approach, very reasonable with modest amounts of data insufficient
for detailed examination of distributional shape and with distributions that are not very
long-tailed, is to consider only the variance. That is, we replace (2.1) by the condition of
constant variance. The condition for the existence of an approximate variance stabilizing
transformation is, for any number of treatment groups, that the variance is determined by
the mean. Then if
var (Y)= v(fk), E(Y)= fk
as the appropriate transformation. That is, evidence of the kind of interaction under
discussion here could come only by showing that
(i) two or more distributions with the same mean have different variances; or
(ii) aspects of distributional shape, e.g. third moments, are not adequately stabilized
by (2.2); or
(iii) the linearizing approximations involved in deriving (2.2) are inadequate and that
a more careful examination of variance shows that it cannot be reasonably
stabilized.
Points (ii) and (iii) are probably best studied numerically in a particular application,
Interaction 3
although some results can be given by higher order expansion. Thus, if py3(t) denotes the
standardized third cumulant of Y, a higher-order approximation is
var{h(Y)} = v()I){h'(L)}2py3+P( ){v([)}?h'(It)h"(tk). (2.3)
Thus, on writing g(k)= {h'(tk)}2, we have a first-order linear differential equation for the
variance-stabilizing transformation. If the solution does not lead to a monotonic h,
variance cannot be stabilized. An alternative use of (2.3) is to retain the transformation
(2.2) when we have that
var {h(Y)} =1- py(G/)v(?G )//v(G), (2.4)
and the magnitude of the correction term can be examined.
For the special choice (2.2) we have that
3v'(t)
Ph(Y),3(t) = PY3(AL)- v PY4 3() 2}
to the next order and if this is strongly dependent on fk there is evidence that distribu-
tional shape is not adequately stabilized, point (ii) above.
For multivariate responses the formulation analogous to that leading to (2.2) is that for
each vector fk there is a covariance matrix fl(fk). It is known (Holland, 1973; Song, 1982)
that only exceptionally does there exist a vector transformation from y to h(y) such that
the covariance matrix of h(Y) is, in the linearized theory, a constant matrix, say the
identity matrix; certain compatability conditions have to be satisfied. Practical techniques
for finding suitable approximate transformations do not seem to be available.
In the remainder of the paper we concentrate on the structure of E(Y), making the
simplest assumptions about variances and distributions. The essence of the present section
is that constancy of variance is a special form of absence of interaction. If variance or
distributional shape vary to an extent that is judged scientifically important, the descrip-
tion of the effect of 'treatments' on response should not be confined to the examination of
means.
appears identical to (3.2) but the interpretation is different, essentially because the roles of
x and z are now asymmetrical. Here the concern is with stability of the x-effect as z
varies, such stability both enabling the treatment effect to be simply specified in a single
function and also providing a basis for rational extrapolation of conclusions to new
individuals and environments. The function -r2(z) itself is in the present context of no
direct interest.
Finally, there are treatment x nonspecific interactions. We discuss these in more detail
in ? 7. Briefly, however, such interactions have either to be given a rational explanation or
regarded as extra sources of unexplained variation, i.e. error.
Similar remarks apply to higher order interactions.
model, often but not necessarily that used to present the conclusions, i.e. in effect by some
pooling of sums of squares. Unless the degrees of freedom for error in the 'full' model are
rather small, decisions on pooling error mean squares usually have relatively little effect
on the conclusions, and use of the error estimate from the 'full' model will often be
simplest. Of course, where there are several layers of error, choice of the appropriate one
can be of critical importance.
In unbalanced data, explicit formulation and fitting of a model will normally be required
both for parameter estimation and for estimating error variance. In this it is usually
reasonable to err on the side of overfitting; for instance, treatment main effects should not
be omitted just because they are insignificant.
In formulating a 'full' model, the broad procedure is, roughly speaking, to identify all
the factors specifying the explanatory variables for each observation and then to set out
the 'types' of these factors (treatment, intrinsic and nonspecific) and the relationships
between the factors in the design (completely cross-classified, hierarchical, incomplete): all
interactions among completely cross-classified variables are included in the 'full' model,
with the exception of those with nonspecific factors which are often amalgamated for
error; see, however, ? 7.
Table 1
Outline analysis of variance table
Table 2
Two-periodcross-overdesign
(a) Simple design (b) Extended design
Subject Subject
group Period 1 Period 2 group Period 1 Period 2
I T, T2 I T1 T2
pt-T-7+y + 7T+T- p + T --7rT-T+ +7rT+T-p-S+Y
II T2 T1 II T2
T L+7R--+p-Y
fL-ITr+T-y -+1/TT+r-T+P-S-/
TL++T- A T, T,
-Trr-T+T
/+Trr-T-p+S-
B T2 T2
--7+7--y R+Tr+7+p+8+y
General mean, pt; period effect, ir; direct treatment effect, 7; residual treatment effect, p; direct
treatment x period interaction, y; direct x residual treatment interaction, S.
8 D.R. Cox
It follows immediately that from the within-subject analysis, the only treatment parameter
that can be estimated is 7-?Ip, which is of interest only if p can be assumed negligible. If
information from between-subject comparisons is added, y -p can be estimated, so that
residual effects and direct treatment x period interaction cannot be separated except by a
priori assumption. Thus from the combined analysis only linear combinations of - p and
y -(P can be estimated and the only such combination of general interest is 1 - -, the
treatment effect in period 1. The conclusion that the design is advisable for 'routine'
comparison of treatment only when there is strong a priori reason for thinking p negligible
seems fairly widely accepted (Brown, 1980; Hills & Armitage, 1979; Armitage & Hills,
1982).
If we use the extended design of Table 2b, subjects are randomized between the four
groups, typically with equal numbers in groups I and II and equal numbers in groups A
and B. The new feature is that residual effect is cross-classified with direct effect and
hence we should introduce for the second period a further parameter for direct x residual
treatment interaction, -8 if the two treatments are different, 8 if the two treatments are
the same. Table 2b gives the expected responses. It follows that in the four groups the
expected values are:
differences: 2 + 2 - - , 2rr -27 + - ,
2w-p + -2y, 2?+p?+8+2y;
sums: 2w -p-8 + 2y, 2 + p- -2y,
2gt-p +-2r, 2~ +p+8+27.
Thus if only within-subject comparisons are used, i.e. differences analysed, the parameters
that can be estimated are 7, p, y +-1, w. That is, a direct treatment effect and a residual
treatment effect can be separately estimated but direct treatment x period interaction and
direct x residual treatment interaction cannot be disentangled. If, however, between-
subject comparisons are introduced all contrasts become estimable.
Three and four period designs are dealt with similarly, although of course the details get
more complicated. The precision of between-subject comparisons can often be increased
by an initial period of standardization supplemented by the measurement of a concomit-
ant variable. We return later to the questions of interpretation; here the initial model
formulation is the aspect of concern. In many applications, use of such designs is advisable
only when residual effects and treatment x period interactions are very likely to be
negligible. There are contexts, however, where the occurrence and nature of residual
effects are of direct interest.
corresponding sets of derived values less directly related to the 'raw' data and having a
more complicated error structure.
There is probably no uniquely optimal way of addressing the problem of formulation,
but the following seems a reasonable general procedure.
(i) Set out the skeleton analysis of variance table for the balanced form analogous to
the data under study. Consider whether 'empty cells' will make any of the contrasts
nonestimable.
(ii) Formulate a baseline model containing, usually, all main effects and any other
contrasts likely to be important, or found in preliminary analysis to be important.
(iii) In isolating the sums of squares for each set of contrasts, not in the baseline model,
include the parameters of the baseline model, the parameters under test, and any
parameters 'above' the parameters under test in the natural hierarchy. Thus, in the
notation used above, if we are examining the three-factor interaction, all the associated
two-factor interactions and, of course, main effects should be included, as well as all
parameters in the baseline model.
Because of the possibility mentioned in (ii) that the initial baseline model will need
modification, the procedure is in general iterative. The final baseline model will usually be
the one underpinning the presentation of conclusions. The procedure is close to a severely
constrained form of stepwise regression.
5 Detection of interaction
is a self-explanatory notation.
10 D.R. Cox
and F2 are not too disparate a factor 3 is a rough approximation adequate for most
purposes (Azzalini & Cox, 1984).
It is important that checks of specification should be carried out, not necessarily by
formal calculation of significance levels, although these are needed in the background for
calibration. Analysis of variance is, among other things, a powerful exploratory and
diagnostic technique and total commitment to an analysis in the light of a priori
assumptions about the form of the response can lead to the overlooking of important
unanticipated effects, defects in the data or misconceptions in formulation.
There are many generalizations of the above discussion, for instance to models for
nonlinear treatment effects: here the interaction with intrinsic variables should, in princi-
ple, be decomposed corresponding to the meaningful parameters in the nonlinear model.
5.3 Graphicalmethods
When, as in a highly factorial treatment system, there are a considerable number of
component interactions, graphical representation may be helpful:
(a) in presentation of conclusions,
(b) in isolating large contrasts for detailed examination,
(c) in determining which contrasts can reasonably be used to estimate error,
(d) in detecting initially unsuspected error structure, such as the hierarchical error
involved in a split plot design (Daniel, 1976, Chapter 16), or important correla-
tions between errors.
Graphical methods work most easily when all mean squares under study have the same
number of degrees of freedom. If most of the contrasts have d degrees of freedom, it may
be worth transforming a mean square with d' degrees of freedom, via the probability
integral transformation and an assumed value 62 for variance, into the corresponding
point of the distribution for d degrees of freedom. Pearson & Hartley (1972, Tables 19
and 20) give expected order statistics corresponding to degrees of freedom 1, 2,..., 7 and
this covers most of the commonly occurring cases. As a simple example, suppose that all
but one of the mean squares has one degree of freedom and that a preliminary estimate of
variance is a2. Suppose that there is in addition a mean square of 1-2a2 with 3 degrees of
freedom. The corresponding value of = 3-6 is at the 0-69 quantile and the correspond-
X32
ing point of 1 is 1-05 so that the single mean square is assigned value 1-052.
It is good to show main effects distinctively: if a primary object is (c), the estimation of
error, it is best to omit the main effects, and any other important effects, in the calculation
of effective sample size for finding plotting positions. That is, the expected order statistics
used for plotting refer to a sample size equal to the number of contrasts used for
estimating error, not to the total number of contrasts. When the contrasts have single
degrees of freedom, there is a choice between plotting mean squares, or absolute value of
contrasts, the half normal plot (Daniel, 1959) or signed values of contrasts. For many
purposes the last seems preferable, partly because the signs of important contrasts can be
given a physical interpretation and partly because if levels are defined so that all sample
main effects are positive, interactions have an interpretable sign, a positive interaction
indicating a reinforcement of the separate effects. Examples of the use of these plots are
to be found in the books of Daniel (1976) and Daniel & Wood (1971).
It is well known that in some approximate large sample sense probability plotting is not
systematically affected by dependence between the points plotted, although the precision
of the points will be altered. Because the shape of 'reasonable' distributions is settled by
the first four moments, a condition for the validity of probability plotting methods is that
12 D.R. Cox
the dependence shall be sufficiently weak that the first four sample moments converge in
probability to their population values. Thus, in unbalanced analyses independence of the
various mean squares to be plotted is not necessary.
A rather extreme example of the case of nonindependent contrasts concerns factors
with nominal levels. For simplicity, consider unreplicated observations y~j(i=
1,..., ml;j== 1,..., m2).
Every elementary interaction contrast
-
Yili Y i21+ Yi2 (il i2;jl: 2) (5.3)
has, under the hypothesis of noYili2-
two-factor interaction and homogeneous variance, zero
mean and constant variance. This suggests that all lm1(mI - 1)m2(m2 -1) contrasts (5.3)
should be plotted against normal order statistics, this being appropriate for detection of
interaction confined to a relatively small number of cells. Interaction produced by outlying
observations, or outlying rows or columns may be detected in this way, although the plot
is highly redundant, especially if mi and m2 are large; the redundancy measured by the
ratio of the number of points plotted to the number of independent contrasts is
lm1m2
and so is large if either or both mi and m2 are large. This technique has been studied in
detail by Bradu & Hawkins (1982) and in an as yet unpublished report by R. Denis and E.
Spjotvoll.
Table 3
Data and outline analysis of 210 system (Fedorov et al., 1968)
xl x2 x3 x4 x5 x6 X7 x8 x9 X10 Y
NH4CI KH2PO4 MgC12 NaC1 CaCI2 Na2S Na2S203 NaHCO3 FeCl3 micro-
9H20 elements
Levels + 1500 450 900 1500 350 1500 5000 5000 125 15 Yield
- 500 50 100 500 50 500 1000 1000 25 5
1 - + + + - + - + - + 14-0
2 - - + + - + + - - + 4-0
3 + - - + + + - - - - 7-0
4 - - + - + + - + + + 24-5
5 + - + + + + + - - - 14-5
6 + - + - + + + + + + 71-0
7 - - - - - - - - - - 15-5
8 + + - + + - - + + - 18-0
9 - + - + - - + - - + 17-0
10 + + + + + - - - + - 13-5
11 - + + - + - + + + + 52-0
12 + + + - - - + + - - 48-0
13 + + - - + - + - + - 24-0
14 - + - - - + - - + - 12-0
15 + - - - - - - + + + 13-5
16 - - - + - + + + - + 63-0
All the concentrations are given in mg/1, with the exception of factor 10, whose central level (10 ml of solution
of micro-elements per 11 of medium) corresponds to 10 times the amount of micro-element in Larsen's
medium. The yield has a standard error of 3.8.
Table 4, is not orthogonal to the rows of the square. Because the analysis of variance of
the derived variable has sums of squares:
columns and treatments, 0; rows, 80; residual, 160; total, 240,
the variance of the estimated interaction is 0.2/160, compared with 0.2/240 for a corres-
ponding orthogonal analysis. That is, in the present instance interaction is estimated with 3
Interaction 15
Table 4
Interaction detection in a 4 x 4 Latin square
'Time'
Subjects 1 2 3 4
1 B, 3 C, -3 A, -1 D, -3
2 D, 3 A, 1 B,-1 C,9
3 A, 3 D, 1 C,3 B,-3
4 C, -9 B, 1 D, -1 A, -3
The derived variable is formed from the product of
the treatment effect C versus A, B, D (-1, -1, 3, -1)
and the linear column effect (-3, -1, 1, 3).
efficiency. If the interaction component of interest were known a priori a particular Latin
square for optimal estimation could be used. This form of test is best regarded as one of
specification. If an interaction were detected the whole basis of the analysis would need
reconsideration.
The principle involved here is that of Tukey's (1949) degree of freedom for nonadditiv-
ity, except that here the isolation of defining contrasts is based partly on a priori
considerations rather than being entirely data-dependent.
6 Interpretationof interaction
with small standard errors of order 1 kg. The treatment effect (1 versus 0) is thus 20 kg for
males and -10 kg for females and interaction, in fact qualitative interaction in the sense of
? 5.2, is present. The interpretation in such cases is normally via these separate effects.
The main effect, i.e. the treatment effect averaged over the levels of the intrinsic factor, is
of interest only in the following circumstances.
(a) The treatment effects at the various levels of the intrinsic factor are broadly
similar, so that an average treatment effect gives a convenient overall qualitative
summary. Clearly, this is not the case in the illustrated example, nor in any
application with qualitative interaction.
(b) An application is envisaged in which the same treatment is to be applied to all
experimental units, e.g. to all animals regardless of sex. Then an average
treatment effect is of some interest in which the weights attached to the different
levels of the intrinsic factor are proportions in the target population. In the
particular example these might be (1,?) regardless of the relative frequencies of
males and females in the data. Note also that if the data frequencies are regarded
as estimating unknown population proportions the standard error of the esti-
mated main effect should include a contribution from the sampling error in the
weights.
Interaction 17
Note especially that in case (b) the precise definition of main effect depends in an
essential way on having a physically meaningful system of weights. This formulation is in
any case probably fairly rarely appropriate and there is some danger of calculating
standard errors for estimates of artificial parameters.
can be applied. One way of viewing (6.1) is that, with fitting by least squares, the model
leads to a rank one approximation to the two-way matrix of interaction terms, and so on if
further terms are added; this establishes a connexion with the singular-value decomposi-
tion of a matrix (Rao, 1973, p. 42).
Important restricted cases of this form of model are obtained by replacing (6.1) by one
of
f 4y+
mO?i +OKi,
?ii=Pt+? 4+?0i (6.2)
/- + 0i4+Wio,
where E 6, = Ei = Ki= yi = 0, or by the symmetrical form
expected response at the mean , then 6 -0. Note, however, that it would rarely make
sense to test the null hypothesis6, - 0 when the f/3differ,for the reasonsgiven in ? 4.1.
The above discussionis appropriatewhen the levels of both factorsare nominal.If, for
example, the rows in (6.1) have quantitativelevels and models with a linear effect are
under examination,it will be naturalto take both the O6and the yi to be linear in the
carriervariable, and this leads to the decompositioncorrespondingto (5.1).
8 Generalizations
Most of the above discussion generalizes immediately to any analysis in which a
parameter describing the distribution of a response variable is related to explanatory
variables via a linear expression. The generalized linear models of Nelder & Wedderburn
(1972) embrace many of the commonly occurring cases. The more indirect the relation
between the original data and the linear representation, the less the interpretative value of
the absence of interaction.
For such generalizations, the method of analysis will usually be either maximum
likelihood or a preliminary transformation followed by least squares with empirical
weights. The main technical difficulty concerns systems with several layers of error. For
these the direct use of weighted least squares will often be simpler, although no generally
useful methods are available at the time of writing.
The definition of interaction within exponential family models involves a linear rep-
resentation of some function of the defining parameter. Thus, for binary data the defining
parameter is the probability of 'success', in this case the moment parameter of the
exponential family. The canonical parameter is the logistic transform. In this case the
canonical parameter has unrestricted range. For exponential distributions, the canonical
parameter is the reciprocal of the mean; the logarithm is the simplest function of the
moment parameter (or of the canonical parameter) having unrestricted range. Interaction
can thus be defined in various ways, in particular relative to linear representations for the
moment parameter, for the canonical parameter, or for the simplest unconstrained
parameter. Especially when the interaction between a treatment and an intrinsic factor is
involved, use of the unconstrained parameter will often be the most appropriate, since
such representations have the greatest scope for extrapolation. Representations in terms
of the canonical parameter lead to the simplest statistical inference, and those in terms of
the moment parameter have the most direct physical interpretation. Of course, often there
will not be enough data to show empirically that one version is to be preferred to another;
with extensive data there is the possibility of finding that function, if any, of the defining
parameter leading to absence of interaction but often familiarity and ease of interpreta-
tion will restrict the acceptable representations.
Darroch (1974) has discussed the relative advantages of so-called additive and multi-
plicative definitions of interaction for nominal variables. His account in terms of associa-
tion between multidimensional responses is easily adapted to the present context of a
single response and several explanatory variables.
Throughout the paper only univariate response variables have been considered. The
formal procedures of multivariate analysis of variance are available for examining interac-
tions affecting several response variables simultaneously. While such procedures may be
used explicitly only relatively rarely, any demonstration that a particular type of simple
representation applied to several response variables simultaneously makes such a rep-
resentation more convincing.
22 D.R. Cox
Acknowledgements
This paper contains the union of invited papers given at the European Meeting of Statisticians, Wroclaw, the
York Conference of the Royal Statistical Society, the Gordon Research Conference, and the Highland Group,
Royal Statistical Society. I am very grateful for various interesting points made in discussion and thank also E.J.
Snell and 0. Barndorff-Nielsen for helpful and encouraging comments on a preliminary written version.
Appendix
Some open technicalissues
The discussionin the main part of the paper has concentratedon general issues and
deliberately put rather little emphasis on details of technique. Nevertheless there are
unsolved technicalmattersinvolved and there follows a list of a few problems.They are
open so far as I am aware, althoughany reader thinkingof workingon any of them is
strongly advised to look through the recent (and not so recent) literaturenot too long
after starting.
Problem 1 [? 2]. Given randomsamples from m populations,test the hypothesisthat
there exists a translation-inducingtransformation.Estimate the transformationand as-
sociated differences.
Problem 2 [? 2]. Calculate the efficiencyof your procedure for 1 as comparedwith
some simple fully parametricformulation.
Problem 3 [?? 2, 5]. Extend the discussionof 1 and 2 to the examinationof inter-
actions.
Problem4 [? 3]. Find some interestingspecial cases satisfyingScheff6'scondition for
transformationto additivity.
Problem4 [? 4]. Extend the notion of order-basedinteractionsto higher order inter-
actions.
Problem6 [? 4]. Develop suitabledesigns and methodsof analysis,and make practical
recommendationsconcerningthe use of, p period, t treatmentcrossoverdesigns (i) for
p = t, (ii) for p > t, (iii) for p < t, when the treatmentsare (a) unstructured,(b) factorial
with quantitativelevels, (c) factorial with qualitativelevels and when interest focuses
primarilyon (I) directtreatmenteffectswith some possibilityof examiningresidualeffects,
(II) estimatingand testing residualeffects, (III) estimating 'total' treatmenteffects.
Problem7 [?? 4, 5]. What is the role of AIC,C, and so on in the study of interactions,
especially from unbalanceddata?
Problem 8 [?? 4, 5]. What is the role of rank invariantproceduresin the study of
interactions?
Problem 9 [? 5]. What is the effect of nonnormalityon the plot of ordered mean
squares?
Problem 10 [? 5]. Are there useful significance tests associated with plots of ordered
mean squares?
Problem 11 [? 5]. Examine in some generality the effect of dependence on probability
plotting methods.
Problem 12 [? 5]. In the two-way arrangement with nominal factors, is there a useful
set of primitive contrasts less redundant than the full set of tetrad differences?
Interaction 23
Problem 13 [? 5]. Develop a systematic theory of testing for treatment x column in-
teraction in an m x m Latin square when both, one or none of the defining contrasts are
known a priori. What is the efficiency of a randomized design for detecting interaction:
what is the lowest efficiency that could be encountered? What are the implications, if any,
for design?
Problem 14 [? 5]. Develop and implement an IKBS(intelligent knowledge-based system)
that would deal with the 2"1 example of ? 5 and similar but much larger problems.
Problem 15 [? 6]. Suppose that Yi (j = 1,... , n) are independently normally distributed
with E(Y) = gi and unit variance. Initially interest lies in the null hypothesis x1
=...=are
= 0 but overwhelming evidence against this is found. Two alternative explanations
1n
contemplated:
(a) Li= 0 (j • 7), fp~# 0, 7 being unknown;
(b) g1, , in are independently normally distributed with zero mean and unknown
....
variance 02.
Develop procedures for examining which if either of (a) and (b) is satisfactory and which is
preferable.
Problem 16 [? 6]. Extend the discussion of Problem 15 to the examination of inter-
actions.
Problem 17 [? 7]. Develop a simple and elegant theory for finding confidence limits for
contrasts where error has to be estimated via synthesis of variance.
Problem 18 [? 7]. Develop from first principles confidence limits for an upper variance
component from balanced and from unbalanced data.
Problem 19 [? 7]. Examine critically the role of time series models for experiments in
which time is a nonspecific factor.
Problem 20 [? 8]. Give a general discussion for analysing interactions in log linear
models for Poisson and binomial data in the presence: (a) of overdispersion, (b) of
underdispersion.
References
Armitage, P. & Hills, M. (1982). The two period crossover trial. Statistician 31, 119-131.
Atkinson, A.C. (1982). Regression diagnostics, transformations and constructed variables (with discussion). J. R.
Statist. Soc. B 44, 1-36.
Azzalini A. & Cox, D.R. (1984). Two new tests associated with analysis of variance. J. R. Statist. Soc. B 46. To
appear.
Bloomfield, P. (1974). Linear transformations for multivariate binary data. Biometrics 30, 609-617.
Box, G.E.P. & Cox, D.R. (1964). An analysis of transformations (with discussion). J. R. Statist. Soc. B 26,
211-252.
Bradu, D. & Hawkins, D.M. (1982). Location of multiple outliers in two-way tables, using tetrads. Technometrics
24, 103-108.
Brown, M.B. (1975). Exploring interaction effects in the analysis of variance. Appl. Statist. 24, 288-298.
Brown, B.W. (1980). The crossover experiment for clinical trials. Biometrics 36, 69-79.
Cochran, W.G. & Cox, G.M. (1957). Experimental Designs, 2nd edition. New York: Wiley.
Cook, P.D. & Weisberg, S. (1982). Residuals and Influence in Regression. London: Chapman and Hall.
Cox, D.R. (1972). The analysis of multivariate binary data. Appl. Statist. 21, 113-120.
Cox, D.R. (1974). Discussion of paper by M. Stone. J. R. Statist. Soc. B 36, 140-141.
Cox, D.R. (1977). Discussion of paper by J.A. Nelder. J. R. Statist. Soc. A 140, 71-72.
Cox, D.R. & Snell, E.J. (1981). Applied Statistics. London: Chapman and Hall.
Daniel, C. (1959). Use of half-normal plots in interpreting factorial two-level experiments. Technometrics 1,
311-341.
Daniel, C. (1976). Applications of Statistics to Industrial Experimentation. New York: Wiley.
24 D.R. Cox
Daniel, C. & Wood, F.S. (1971). Fitting Equations to Data, New York: Wiley.
Darroch, J.N. (1974). Multiplicative and additive interaction in contingency tables. Biometrika 61, 207-214.
Draper, N.R. & Hunter, W.G. (1969). Transformations: Some examples revisited. Technometrics 11, 23-40.
Doksum, K.A. & Sievers, G.L. (1976). Plotting with confidence: graphical comparisons of two populations.
Biometrika 63, 421-434.
Fedorov, V.D., Maximov, V.N. & Bogorov, V.G. (1968). Experimental development of nutritive media for
micro-organisms. Biometrika 55, 43-51.
Fisher, R.A. & Mackenzie, W.A. (1923). Studies in crop variation. II. The manurial response of different potato
varieties. J. Agric. Sci. 13, 311-320.
Freeman, G.H. (1973). Statistical methods for the analysis of genotype-environment interactions. Heredity 31,
339-354.
Gentleman, J.F. (1980). Finding the k most likely outliers in two-way tables. Technometrics 22, 591-600.
Gentleman, J.F. & Wilk, M.B. (1975). Detecting outliers in a two-way table. I. Statistical behaviour of residuals.
Technometrics 17, 1-14.
Hills, M. & Armitage, P. (1979). The two-period crossover clinical trial. Brit. J. Clinical Pharmacol. 8, 7-20.
Holland, P.W. (1973). Covariance stabilizing transformation. Ann. Statist. 1, 84-92.
Johnson, N.L. & Leone, F. (1964). Statistics and Experimental Design, 2. New York: Wiley.
Krishnaiah, P.R. & Yochmowitz, M.G. (1980). Inference on the structure of interaction in two-way classification
models. Handbook of Statist. 1, 973-994.
Lindley, D.V. (1972). A Bayesian solution for two-way analysis of variance. Col. Math. Soc. Jdnos Bolyai 9,
475-496.
Mandel, J. (1971). A new analysis of variance for non-additive data. Technometrics 13, 1-18.
Nelder, J.A. (1977). A representation of linear models (with discussion). J. R. Statist. Soc. A 140, 48-76.
Nelder, J.A. & Wedderburn, R.W.M. (1972). Generalized linear models, J. R. Statist. Soc. A 135, 370-384.
Pearson, E.S. & Hartley, H.O. (1972). Biometrika Tables for Statisticians, 2. Cambridge University Press.
Rao, C.R. (1973). Linear Statistical Inference and its Applications, 2nd edition. New York: Wiley.
Scheff6, H. (1959). The Analysis of Variance. New York: Wiley.
Snee, R.D. (1982). Nonadditivity in a two-way classification: is it interaction or nonhomogeneous variance? J.
Am. Statist. Assoc. 77, 515-519.
Song, C.C. (1982). Covariance stabilizing transformations and a conjecture of Holland. Ann. Statist. 10,
313-315.
Tukey, J.W. (1949). One degree of freedom for non-additivity. Biometrics 5, 232-242.
Wilk, M.B. & Gnanadesikan, R. (1968). Probability plotting methods for the analysis of data. Biometrika 55,
1-17.
Wishart, J. (1938). Growth-rate determinations in nutrition studies with the bacon pig, and their analysis.
Biometrika 30, 16-28.
Yates, F. (1972). A Monte-Carlo trial on the behaviour of the nonadditivity test with nonnormal data.
Biometrika 59, 253-261.
Yates, F. & Cochran, W.G. (1938). The analysis of groups of experiments. J. Agric. Sci. 28, 556-580.
Resum6
Le r61leen la statistique de I'action reciproque (interaction) est discut6. On classifie les interactions en fonction
de la nature des facteurs qui sont engages. Les tests pour l'interaction sont passes en revue. Les methodes
d'interpretation d'une interaction sont discut6s et la connexion avec l'estimation d'erreur est decrite. Enfin
problemes non resolus sont catalogues.