1982 When the Data Are Functions
1982 When the Data Are Functions
DECEMBER,1982
J. O . RAMSAY
MCGILL UNIVERSITY
A datum is often a continuous function x(t) of a variable such as time observed over some
interval. One or more such functions are observed for each subject or unit of observation. The
extension of classical data analytic techniques designed for p-variate observations to such data is
discussed. The essential step is the expression of the classical problem in the language of function-
al analysis, after which the extension to functions is a straightforward matter. A schematic device
called the duality diagram is a very useful tool for describing an analysis and for suggesting new
possibilities. Least squares approximation, descriptive statistics, principal components analysis,
and canonical correlation analysis are discussed within this broader framework.
Key words: continuous data, functional analysis, duality diagram.
Introduction
Sophisticated data collection hardware often produce data which are a set of con-
tinuous functions. I am sure that all of us have seen such data: E E G and E M G records,
learning curves, paths in space, subject responses continuous in time, speech production
measurements during vocalization, bioassay data, and so on. Consider as a further exam-
ple the curves displayed in Figure 1. These indicate the height of the tongue d o r s u m
during ten different utterances of the sound " a h - k a h " by a single subject [Keller & Ostry,
N o t e 1]. It is natural to consider each curve as a single observation, to summarize the ten
curves in terms of an average curve, and to measure in some way the variation of the
curves a b o u t this average.
This paper considers the extension of classical statistical techniques to include func-
tional data. It will be an elementary and simplified treatment, which m a y a n n o y those
wanting more subtlety and rigor. I must warn you, however, that a fundamental change of
point of view about what data are will be required, and if y o u leave my address aware
that an altered state of statistical consciousness is possible, I shall be content.
In dealing with functional data I will refer frequently to two lines of development.
The first is the expression of traditional data analytic technology in the language of func-
tional analysis. M u c h o f this w o r k has taken place in France and is not available in
English. I a m particularly indebted to the m o n g r a p h s of Cailliez and Pages 1-1976] and
Dauxois and Pousse [1976]. We are very fortunate to have with us for these meetings a
n u m b e r of those associated with this work, and in part m y talk is only an introduction to
t o m o r r o w ' s symposium.* The second line of development that has fascinated m y col-
league Suzanne Winsberg and I in recent years has been statistical applications of spline
functions. I feel that splines are destined to play a fundamental role in the analysis of
functional data, but I will try to show h o w in only a vague way at this point. Finally, this
* "New glances at principal components and correspondence analysis" was a symposium at the 1982 Joint
Meetings of the Classification Society and Psychometric Society, Montreal, Canada.
Presented as the Presidential Address to the Psychometric Society's Annual Meeting, May, 1982. I wish to
express my gratitude to my colleagues in France, especially at the University of Grenoble, for their warm
hospitality during my sabbatical leave. Preparation of this paper was supported by Grant APA 0320 from the
Natural Sciences and Engineering Research Council of Canada.
Requests for reprints should be sent to: J. O. Ramsay, Dept. of Psychology, 1205 Dr. Penfield Ave., Mon-
treal, Qurbec, Canada H3A 1B1.
0033-3123/82/1200-5004500.75/0 379
© 1982 The Psychometric Society
380 PSYCHOMETRIKA
B.O
5.5
I--
I,iJ
',r-
E
(f) 5.0
0
hJ
Z
ID
I..-
q.5
tl.0
I I ....... ! I
0.0 0.2 O.q 0.6 0.8 1.0
TIME
FIGURE 1
The height of the tongue dorsum over a 400 millisecond interval of time during which the sound "ah-kah" was
uttered. Each curve represents a single utterance. The same subject was involved in all ten replications. The
average curve is represented by a dashed line and was computed by averaging the ten curves at each point
in time. The time units have been arbitrarily scaled to the interval [0, 1].
Domain o f X
Number of Variables
p<a) p=Co
(/)
Q)
o o e 0 O
t/) O O Q O Q
t3 n <(30 O ~ e S Q
¢,3 $ $ 0 D 0
b e e o o
Q $ O $ O
Z
E
::3 n=a) //
Datum Xil , - - - , Xip xi(t), O t=T
FIGURE 2
Possible domains for statistical observations. These domains depend on whether the number of replications or
cases (n) is finite or infinite and whether the n u m b e r of variables or points of observation (p) is finite or
corresponds to the points on a continuum.
situation where the variables define a continuum. As the upper right corner of Figure 2
indicates, the data now offer a number for each point on a continuum for each subject,
and it becomes natural to use function notation x,.(t) to designate the value assigned to
individual i at point t on this continuum. The lower right corner shows that one may even
have an infinity of subjects or cases to consider, although in this talk I will confine my
attention to finite n.
When presented with a continuous function, classical statisticians have tended to
respond in two ways. The first is to sample the continuum at a limited number of points
tj, j = 1. . . . . p. In effect each point is considered as a variable in the classical multivariate
problem. There are a number of disadvantages to this approach, however. Continuity and
higher orders of smoothness, which are usually important aspects of functional data, are
ignored. Information between sampling points is lost. Covariance parameters proliferate
with murderous speed in p so that a reduction of information does not achieve a re-
duction in model complexity.
The second classical response has been to postulate a family of functions which ap-
proximate the data but which depend on a limited number of parameters. The literature
on the learning curve and the item characteristic curve illustrates this process nicely. Until
382 PSYCHOMETRIKA
recently this approach ran into problems with the lack of flexibility of most parametric
curve families, but the advent of spline function technology has been a great break-
through. Nevertheless, summarizing the data in terms of a point in parameter space is not
the same thing as summarizing directly as functions. This problem becomes particularly
obvious when one tries to express variation in the functional data in functional rather
than point terms.
The major step required to accommodate functional data is to express traditional
statistical ideas in functional analytic terminology. This involves thinking of the data as
defining a mapping rather than a set of points. That is, the data must be viewed as an
element in a space of possible functions taking a domain space into a range space. After
describing what this means first for the p-variate ease and then for the functional data
case, I will look at least squares estimation, principal components analysis, and canonical
correlation analysis from this perspective. In each case the transition from classical multi-
variate data to functional data is very simple, involving essentially replacing a summation
by an integral. The essential and perhaps most difficult step is the change from a static to
a dynamic conceptualization of the data.
with the vector space of real-valued functions defined over a closed interval and for which
the integral of their squares is finite.
p-variate Data
Let X represent the set of observations x~j, j = 1. . . . . p, i = 1. . . . . n. X can be viewed
as a funcion or a mapping from one vector space into another. The two vector spaces in
question are:
1. Subject space E of dimension p. In this space any subject or case, observed or
hypothetical, is represented by a point corresponding to its relation to each of the
p variables. In this space two vectors ei and ek will have an inner product b~(e~,
ek), which for purposes of simplicity will be e~-ek, and the norm I1 e It of a vector e
is the square root of the inner product of e with itself. The space can be spanned
by a set of p orthonormal vectors e.
2. Variable space F of dimension n. In this space any variable, observed or hypotheti-
cal, is represented by the values assigned to the n subjects at that position. It also
has an inner product b r ( . , . ) and a set of n orthonormal spanning vectors.
The mappings that X represents are then
1. X: E ---, F. Let e be any element of E. Then with X arranged as an n by p matrix
the matrix p r o d u c t f = Xe is an n-vector and thus a position in variable space.
2. Xt: F ---, E. For any element f ~ F the product X~f is an element of E. The map-
ping X is called the transpose mapping associated with X. This has a broader
meaning in functional analytic terms* for any elements e of E and f of F a map-
ping and its transpose satisfy br(f,, Xe) = be(X'f, e).
It can be useful to think of spaces E and F as themselves spaces of functions. For example,
F can be considered the space of possible mappings of n individuals into the real number
line. One then has a complete expression of the problem in function space terminology.
As unfamiliar as this may be, it can pay handsome dividends when considering some new
problems. However, the main temptation to be resisted at this point is to consider a
column or row of X as an element of F or E, respectively; X must be viewed as a
mapping rather than a set of vectors.
In addition to the two mappings above associated with X there are two other im-
portant mappings determined by X. Both of these are called operators since they are
mappings from a space into itself.
3. V: E --- E. The fact that X m a p s a vector from E to F and X t maps it back again
means that the composite X ' o X maps E into itself. Denoting the matrix X~X by
V, this mapping corresponds to the matrix product Ve. When the matrix X has
zero column means, the matrix n - 1 V is the variance-covariance matrix, and the
corresponding operator is also called the variance-covariance operator. Figure 3
shows an example for n = 3 and p = 2 of how a particular vector is transformed
by V.
4. W: F - - * F. Along with operator V for E goes a corresponding operator for F
which results from applying X t to an element f and then X to the resulting image
of f i n E. In matrix terms W = X X ' .
* When one or both spaces are equipped with something other than the identity metric it is important to
distinguish between the transpose of a mapping and its adjoint. The appendix discusses this.
384 PSYCHOMETRIKA
-X-
X t"
toXe
FIGu~ 3
Two views of the transformation of the vector e = (2, 5)1 by the operator V = X'X, where X is the data matrix
X= 1 .
0 .
J. O. RAMSAY 385
This is a simplified version of what Jean-Pierre Pag6s [Pag6s & Tenenhaus, Note 2] will
describe for us tomorrow* as the duality diagram. It is a very important symbolic tool for
the presentation of data analytic problems in functional analytic terms. In this talk I will
ignore any considerations involving metrics for spaces E and F. Appendix A explains how
the duality diagram is extended to accommodate metrics other than the identity by the
use of the dual spaces for E and F.
We may now say that the data determine a subspace of F in the following way. Any
vector e in E is mapped by X into F. If we choose the p orthonormal vectors ej in E and
map them into F, the result will be p vectors in F; and the image of any e in F can be
represented by a linear combination of these p images. Thus the subspace of F is in effect
the image of E under the mapping X, and will usually be of dimension min {p, n}. If an
arbitrary elementfwithin F does not lie within this subspace, its image under operator W
certainly will since W is a result of first m a p p i n g f i n t o E and then its image back into F.
Functional Data
The next task is to extend the functional analytic discussion of the data given above
to the situation in which the datum for individual or case i is a function x,(t), 0 < t < T.
As Figure 2 suggests, one may imagine this arising from allowing the number of variables
to become so large that the index j, j = 1. . . . . p, can be treated as a continuum and is
renamed t. The continuum will be referred to as time in this paper since this is so often
the case in practice. In general, where summation over j takes place in the classical case,
integration over t will now be required. The data will still be indicated by X but it will no
longer be worthwhile thinking of X as a matrix. Nevertheless, X defines mappings among
the following spaces.
1. Subject space E of infinite dimension. Any subject, observed or hypothetical, can be
associated with a function e(t). Since an infinite variety of functions is possible the
space has infinite dimensionality. An inner product bE(e~, ek) for functions %(0 and
ek(t) is given by
and again the norm of a function is the inner product of the function with itself;
that is, the integral of its square. If attention is confined to functions having finite
norms, then the resulting space is a ttilbert space. It is fundamental to Hilbert
spaces that any element can be represented in them by a weighted sum of a count-
able number of orthonormal functions, so that although the dimensionality is in-
finite, it is at least countable. E in the p-variate case is also a Hilbert space.
2. Time space F of dimensionality n. This space has exactly the same characteristics
as F in the p-variate case. Any point in time, observed or hypothetical, is repre-
sented in this space by the values of the n functions at that time point.
The mappings that X represents are now
1. X: E --* F. Let e(t) be any function in space E. Then the v e c t o r f w h o s e ith element
is given by
f~ = ;/ x,{t)e(t) dt (2)
what one would obtain in the p-variate case if the summation for the ith element
in the product Xe were replaced by an integration. Thus, it may be helpful to
imagine X in the functional data case as represented by a matrix with completely
dense rows.
2. Xt: F --- E. L e t f b e any vector in F. Then the function
These mappings are still described schematically by the duality diagram (1). The image of
E in F determined by the transformation X is now in general of dimension n and thus
coincides with F itself. This does not mean, however, that a W maps a vector f i n t o itself,
and a central problem is now to study the consequences of any or all of these mappings.
E /~. (6)
P
One set of approximating functions which have spectacular properties are piece-wise
polynomials or splines, which are not only very flexible with only a modest number of
parameters but remarkably easy to handle computationally. Winsberg and myself 1-1980;
1981; Winsberg & Ramsay, Note 3] have been working with monotone splines, and she
will report on another application in these meetings [Winsberg & Ramsay, 1982, Note 4].
Monotone spline approximation involves projecting function space onto a cone and has
turned out to be not at all difficult in a wide variety of problems. A comprehensive
treatment of splines and the various spaces associated with them is to be found in Schu-
maker 1"1981].
Fioua.e 4
T w o views of the transformation of the unit circle by the m a p p i n g of X, where X is the data matrix
X= t 0 .
0 .5
Thus, the search for vector e of unit norm yielding maximum II Xe II is equivalent to the
search for the element maximizing bE(e, Ve). Any inner product satisfies the Cauchy-
Schwarz inequality, bF(ej, ek)2 < b~(ej, ej)b~(ek, ek), and it follows from this that b~(e, Ve)
is maximized when
Ve = 2e. (8)
J. O. R A M S A Y 389
This is of course an eigenequation. From (7) we have that IIXe II2 = 2 and thus that 2 is a
measure of variation in F. One may consider Xe as defining the dominant direction of
variation in the image of E in F, or in the image of F under mapping W. Under fairly
general conditions it can be shown that operator V can be expressed as ~ 2~ej(u)ej(t) with
at most a countable number of terms and where (2j, ei) is the jth solution to (8).
The identification of solutions to this equation is called spectral analysis, and is one
of the central topics in functional analysis. In the finite p case e is an eigenvector of
symmetric positive definite matrix V. In the functional data case e(t) is an eigenfunction of
the operator V. In either case 2 is an eigenvalue of V. Under very general conditions
which are reasonable for practical work it can be shown that the number of solutions to
the eigenequation are at most countable, that the eigenvalues are nonnegative and
distinct, and that the largest eigenvalue is finite.
There are various ways of tackling the problem of computing the eigenfunctions ej(t).
When n is not too large one may perform the matrix eigenanalysis of the matrix W
instead. The required eigenfunctions are then simply X ~ , i = 1. . . . . n. Techniques for
dealing directly with the operator V also exist, some of which involve discrete approxi-
mations for which an adequate convergence theory exists. A simple form of discrete ap-
proximation to operator V involves choosing a sufficiently large number of equally
spaced points in [0, T] and approximating the integrals in (4) by the corresponding sums
divided by the number of points. In this form the problem becomes identical with the
classical multivariate procedure for principal components analysis. However, more
sophisticated quadrature procedures will require much fewer points in general.
An important extension of principal components analysis is when E is mapped into
vet another space G according to the following diagram:
U X
~ ~ '- ------~
C0VARIANCE S U R F A C E
7"0
0
FIGURE 5
The kernel of the variance-covariance operator K(t, u) = xl(t)x~u) computed from the functions in Figure 1
plotted as a surface. The peaks correspond to times of about 0.4 and 0.8 when the tongue dorsum was
decelerating. The hollows occur at the points (0.4, 0.8) and (0.8, 0.4) and indicate that covariance between
points where the tongue dorsum is decelerating is low. The height of the surface along the diagonal is the
variance functinn.
correlation surface defined by K(u, t)/[K(u, u)K(t, 0] 1/2 may also be plotted. The eigen-
functions ej(t) corresponding to dominant eigenvalues indicate the dominant types of vari-
ation from the average function, just as the eigenvectors in p-variate analysis indicate the
dominant directions of variation. Thus they should also be displayed.
The family of techniques known as time series analysis can be linked to the analysis
of functional data at this point. These procedures are based on the concept of stationarity
of covariance structure which implies that the correlation surface K(u, t) can be defined as
a function of I u - t l alone [Doob, 1953-1. It can be shown that the eigenfunctions in this
case are periodic and hence representable by at most a countable combination of sines
and cosines.
The descriptive analysis of functional data can be illustrated by the analysis of the
J. O. RAMSAY 391
CORRELATION SURFACE
"t'O
0
FIGURE 6
The correlation surface K(t, u)/(K(t, t)K(u, u)) ~/2 computed from the function in Figure 1. The hollows at (0.4,
0.8) and (0.8, 0.4) indicate that tongue dorsum movement is uncorrelated with itself at points where the
dorsum is decelerating.
0.6
O.q
0.2
-0.0
-0.2
-0.~
-0.6
I I I ' f"
0.0 0.2 O.q, 0.6 0.8 .0
TIME
FIGURE 7
The first two eigenfunctions of the variance-covariance operator K(t, u) computed from the data in Figure 1.
These correspond to eigenvalues of 0.17 and 0.12, respectively and account for 95% of the total of all
eigenvalues. The first eigenfunction indicates that the dominant form of deviation from the mean function is
a vertical shift, while the second eigenfunction indicates that a second component of variation is under- or
over-shooting of the tongue dorsum at points where it is decelerating.
correspond to the near zero correlation between tongue dorsum height at the two points
of deceleration. It appears that the tongue starts up with precision but has some di~culty
slowing down. The first two eigenvalues account for 95% of the total, and Figure 7
displays the first two eigenfunctions. These may be summarized by saying that the two
dominant modes of variation are a general vertical shift of the function, and an initial
overshooting prior to maximum height followed by an undershooting at minimal height.
Conclusion
I have tried to indicate in a very general way how statistical concepts can be ex-
tended to include functional data. This requires the point of view from which all data are
394 PSYCHOMETRIKA
functional, and finally there is no essential distinction between data domains which are
finite sets and those which are continua. Functional analytic terminology also permits a
natural and rigorous treatment of a number of other familiar and not so familiar situ-
ations, including dual scaling or correspondence analysis, vector-valued continuous func-
tions as data, and tables of data with an arbitrary number of modes. At least one branch
of statistics is already expressed in functional terms: Bayesian inference is essentially the
mapping of the space of density functions defined on a parameter space into itself using a
nonlinear operator determined by the data. In this sense operator V is only a special case.
Unfortunately we in North America are handicapped by the fact that a course in
functional analysis is seldom a part of the preparation of an applied statistician. As was
the case for linear algebra, we pragmatic souls demand a demonstration of practical rele-
vance before making such a commitment. There are now a number of elementary intro-
ductions to functional analysis, including Aubin [1979] and Kreyszig [1978]. However
the classic work by Dieudonn6 [1960] still has few rivals. A very elementary treatment of
this material in a statistical context for the finite p case is a wonderful book by Cailliez
and Pag~g [1976] which is, alas, still only in French.
I would like to conclude with a quote from Dieudonn6 [1960].
The student should as soon as possible become familiar with the idea that a function f is a single object,
which m a y itself "vary" and is in general to be thought of as a "point" in a large "functional space"; indeed,
it m a y be said that one of the main differences between the classical a n d the modern concepts of analysis is
that, in classical mathematics, when one writes f(x)f is visualized as "fixed" and x as "variable" whereas
nowadays bothfand x are considered as "variables" ... (p. 1)
Functional analysis already has revolutionized numerical analysis, so that any issue of a
major journal now has a number of papers using this technology. I claim that this is
about to happen in statistics, and I hope that my talk leads you to speculate on this
possibility.
Appendix
The inner product function used in the text is too simple for many applications. For
example, the use of standardized data is equivalent to the use of inner product
~ ernjemk
bE(e~, ek) = m - a----~ "
Alternatively, one can say that br.(ej, ek) = e~ M e k , where M is a diagonal matrix contain-
ing reciprocals of variances in its diagonal. Matrix M is called the metric for E, and a
more general treatment requires the consideration of metrics M and N for spaces E and
F, respectively. The use of metrics requires the concept of a dual space, and in this appen-
dix the duality diagram (1) is extended to include these dual spaces.
For any vector e there are various possible real-valued functions that can be com-
puted. Weighted sums of elements and the maximal element are two examples. The set of
possible continuous real-valued functions of e which are linear in e is itself a vector space
since it is closed under weighted summation and has the other vector space properties.
This space is denoted by E* and is referred to as the dual of E. For example, if E consists
of the space of column vectors with p elements, then the dual of E is the space of row
vectors of size p. It is usual to use the notation (e, e*~ for the real number which results
when a function e* in E* is applied to an element e of E. F o r example, if e* is a row
vector and e a column vector then the matrix product e*e is denoted by (e, e*). The
symmetry of the notation corresponds to the symmetry of the relation between E and E*
since the dual of E* turns out to be E.
J. O. RAMSAY 395
There is a one-to-one correspondence between the two spaces induced by the equa-
tion
be(e, e0) = (e0, e*), (A1)
where eo is any fixed nonzero element of E. This equation defines a mapping M: E ---, E
for which to any e ~ E is associated the element e* = Me in E which satisfies (1). In the
p-dimensional case both E and E* can be represented by p-tuples of numbers, and
e* = Me, where M is the metric matrix. Thus M is used to denote both a metric and the
mapping from E to E*. If E* is equipped with inner product bE.(e*, eg) = ~"*t,,,~'t- l~.~k and
thus with metric M - l, then the two inner products satisfy br.(e~, e~) = be.(Me~, Mek) =
br(ej, ek).
It is usual to postulate two further properties of E. One of these is completeness,
implying that convergent sequences of vectors converge to a vector in the space. Intuit-
ively this implies that the space has no "holes" in it; the real numbers are a complete set
since they have no gaps between them while the rationals are not because there are se-
quences of rationals converging to irrational numbers. A complete vector space with an
inner product is called a Hilbert space. The other property used in practice is separability.
This says that the space contains a dense countable subset: every neighborhood of any
element contains a member of this dense countable subset. F o r example, every neighbor-
hood of a real number contains a rational number. Intuitively, this condition implies that
the space is "evenly and thinly spread."
Consider now a continuous linear mapping T: E --* F from Hilbert space E into
another Hilbert space F. Then there are two other mappings automatically associated
with T:
1. T*: F --* E: the adjoint mapping defined by the following equation
for any e ~ E a n d f e F.
2. Tt: F* ---, E*: the transpose mapping defined by the following equation
where f * = N f a n d e* = Me.
In the special case where both E and F are equipped with identity metrics, the two map-
pings are equivalent. The transpose mapping has the advantage of being invariant with
respect to changes in metric. In the finite dimensional case with E equipped with metric
matrix M, F with metric matrix N, and the mapping T represented by matrix T, the
transpose mapping is represented by T t and the adjoint by M - ~TtN.
Turning now to the data analytic applicaton of this theory, the data matrix X logi-
cally represents a mapping from E* to F. This arises because of the following two con-
siderations:
1. The ith individual is represented by a vector e~ in E specifying his relationship
with the p variables or with each point in time. If the dual space E* is spanned by
a set of orthonormal basis vectors e*, j = 1. . . . , p, then (e~, e*) = xij, the score of
individual i on variable j. Thus associated with each variable j is the orthonormal
basis vector e*.
2. The jth variable is also represented by v e c t o r fi in F specifying the relationship of
this variable with the n individuals. Again, iff~*, i = 1. . . . . n, is the orthonormal
spanning system of vectors for F* then (fj,f~ *) = xi~, and associated with each
individual i is the orthonormal basis vectorf~ *
396 PSYCHOMETRIKA
(A4)
E )F*.
X'
REFERENCE NOTES
1. Keller, E. & Ostry, D. J. Computerized measurement of tongue dorsum movements with pulsed echo ultra-
sound(a). Manuscript submitted for publication to Journal of the Acoustical Society of America, 1982.
2. Pagrs, J. P. & Tenenhaus, M. Geometry and duality diagram. An example of application: The analysis of
qualitative variables. Paper presented at the Psychometric Society Annual Meeting, Montreal, Canada, 1982.
3. Winsberg, S. & Ramsay, J. O. Monotone spline transformations for dimension reduction. Submitted for
publication in Psychometrika.
4. Winsberg, S. & Ramsay, J. O. Monotone spline transformations for ordered categorical data. Paper presen-
ted at the Psychometric Society Annual Meeting, Montreal, Canada, 1982.
REFERENCES
Aubin, J.-P. Applied Functional Analysis. New York: Wiley, 1979.
Cailliez, F. & Pages, J.-P. Introduction d l'Analyse des Donn~es. Paris: Soci6t6 de Math6matiques Appliqu6es et
de Sciences Humaines, 9 rue Duban, 75016 Paris, 1976.
Dauxois, J. & Pousse, A. Les analyses factorieUes en calcul des probabilit6s et en statistique: Essai d'6tude
synth6tique. ThOse d'6tat, l'Universit6 Paul-Sabatier de Toulouse, France, 1976.
Dieudonn6, J. Foundationsof Modern Analysis. New York: Academic Press, 1960.
Doob, J. L. Stochastic Processes. New York: Wiley, 1953.
Kreyszig, E. Introductory Functional Analysis with Applications. New York: Wiley, 1978.
Schumaker, L Spline Functions: Basic Theory. New York: Wiley, 1981.
Tucker, L. R. Determination of parameters of a functional relationship by factor analysis. Psychometrika, 23,
1958, 19-23.
Winsberg, S. & Ramsay, J. O. Monotonic transformations to additivity using splines. Biometrika, 67, 1980,
669-674.
Winsberg, S. & Ramsay, J. O. Analysis of pairwise preference data using integrated B-splines. Psychometrika, 46,
1981, 171-186.
Final version received 7/12/82