Sensitivity Analysis in Multi-Objective Decision Making: David Rios Insua
Sensitivity Analysis in Multi-Objective Decision Making: David Rios Insua
Systems 347
Sensitivity Analysis
in Multi-objective
Decision Making
Lecture Notes
in Economics and
Mathematical Systems
Managing Editors: M. Beckmann and W. Krelle
347
David Rlos
RIOS Insua
Managing Editors
Prof. Dr. M. Beckmann
Brown University
Providence, RI 02912, USA
Author
Dr. David Rios Insua
School of Computer Studies
University of Leeds
Leeds LS2 9JT, UK
Opt. de Inteligencia Artificial, Facultad de Informatica
Universidad Politecnica de Madrid
Boadilla del Monte, 28660-Madrid, Spain
This work is subject 10 copyright. All rights are reserved, whether the whole or part of the materral
is concerned, spedlcally the rights of translation, reprinting, re·use of illustrations, recitation,
broadcasting, reproduction on microfilms or in other ways, and storage in data banks. Duplication
ofthis publicalion or Darts thereof is only permitted under the provisions of the German Copyright
Law of September 9. 1965, in its current version, and a copyright fee must always be paid.
Violations fall unCer Ihe prosecution act of the German Copyright Law.
© Springer-Verlag Berlin Heidelberg 1990
Originally published by Springer-Verlag Berlin Heidelberg New York in 1990
I wish to thank my supervisor, Professor Simon French, for his permanent en-
couragement, discussions and criticisms, his patience with my sometimes imprecise
approaches and for guiding my interest to my research area and my thinking to
the Bayesian path.
My first steps in Decision Theory were given under the guidance of Professor
Sixto Rios, who has provided invaluable criticisms to my manuscript. I have
received many suggestions for my research from Professor Doug White, Dr. Roger
Cooke and Dr. Les Proll.
My gratitude to staff and fellow students in Manchester University, for taking
part in my experiments, and Leeds University, for helping me with the computers
and supplying an enjoyable environment. Mike Wiper belongs to the intersection
of both sets; in addition, he has discussed many concepts and read most of the
manuscript. Valeria Rios Insua helped me with the figures.
Computer facilities were provided by the University of Manchester Regional
Computer Center and the University of Leeds Computing Service. This research
was supported by a grant of the Bank of Spain.
Finally, I wish to thank my family for their support and understanding, and
Cornelia for showing me that even the impossible is beatable.
Contents
5 Conclusions 164
5.1 Summary . 164
5.2 Topics for further research . 165
5.2.1 Decision making under partial information 165
5.2.2 Sensitivity analysis .. . 167
5.2.3 Computation . . . . . . 168
5.2.4 Group decision making . 170
5.2.5 Imprecision in expert systems 170
5.2.6 Restructuring . . . . . . . . . 170
5.2.7 Integration with other methodologies 171
Bibliography 172
Appendix 186
List oe abbreviations
- p.o. Potentiallyoptimal
- a.p.o. Adjacent potentially optimal
- s.a.p.o Straight adjacent potentially optimal
- a.p.o.w Adjacent potentially optimal according to w
a.w.o. additive weak order
One of the defining capacities of humans is that of being able to make decisions.
We shall consider decision making problems in which adecision maker (DM) has to
choose among a finite number of alternatives according to some criteria. Since this
is not always a routine task, some kind of rational help may be of value to DM's.
Scientists have approached this problem developing more or less sound theories,
models and concepts leading to decision aids. Essentially, these are calculation
procedures or rules such that filled in with the proper data they give us a ranking
of the alternatives, or the best ones, or a sub set of good ones.
Thus, the excellence of an application of adecision aid will depend not only
on the goodness of the calculation procedure, its theoretical foundations and its
numerical precision, but also on the quality of the data input.
Simplifying, data input are of two types:
• 'subjective' data: they refer to the DM's judgemental inputs (weights, pair-
wise comparisons, beliefs, strength of preferences, ... ) required to derive the
decision aid output according to her thought.
States
° 1 .·····Ok
Alternatives ...
. ..
am aml .... ·· amk
fora,bEA.
Additive models are popular in the multi-attribute case. Scott (1964) gives con-
ditions allowing this representation.
Definition 1.1 Let A = f1~lAi, I > 1,dj,bj E A,j = 1, ... ,1. (dl, ... ,d1)E1
(b l , .•• , b,) means that for each i (dL ... , di) is a permutation of (bL ... , bi).
For exarnple, in R 3
«1,3,2), (2, 6, 3), (3,2,4), (4,2,5) )E4 «1, 6,4), (2, 2, 5), (4,2,3), (3,3, 2».
Proposition 1.2 (Scott,1964) Let A = f1~1 Ai, with Ai finite. (A,~) satisfies:
• SC1. Fora,bEA,a~borb~a .
• SC2. Let I> 1, dj , bj E A,j = 1, ... ,1. (d l , ... , d,)El(~, ... , b1) and dj~bj,
for j = 1, ... ,1-1 =? bl~dl.
for a, bE A.
Fishburn (1970) gives other conditions. French (1986) provides a number of results
when A is an infinite set.
4
S3 simply requires that any element of G belong to the same number of Ai and Bi,
for i = 1, ... ,1 + 1. P need not be unique. A unique representation is achieved if
we add an axiom which intro duces as a primitive the notion of a uniform random
variable over [0,1] through a reference experiment, see De Groot (1970), French
(1982). Other ways are reviewed in Fishburn (1986), which includes results for
infinite spaces.
Having seen how to model the DM's beliefs, we associate a lottery with each
alternative. 'vVe assume that the DM may establish her preferences among these
lotteries. More generally, let Po(C) be the set of prob ability measures over C and
:::5 a binary relation on Pa(C) interpreted as folIows: let p, q E Po(C), 'P:::5q' means
'p is at most as preferred as q'. We model this relation via expected utility:
iff
5
• U3. p -< q,q -< r=9ap+(l-a)r -< q, q -< ßp+(l-ß)r forsome a,ß E (0,1).
Results for the cases in which Cis not finite and Po(C) is not only the set of simple
probability measures may be found, for example, in Fishburn (1982).
1.1.3 Comment
This has been a brief review of the foundations of Bayesian decision analysis.
These topics, along with practical ways of building the functions involved, are
fully covered in works such as De Groot (1970), Fishburn (1970, 1982), French
(1986), Rios (1977) or Roberts (1979). This review will facilitate a comparison
with the results we shall give in the next chapter.
These results provide a rational framework in which adecision making prob-
lem can be solved. Any alternative approach has to break this framework in some
direction. We recall the solution suggested by it: if the problem is under certainty,
we model the DM's preferences by a value function v to be maximised; if the prob-
lem is under uncertainty, we model the DM's beliefs by a probability distribution p
and the DM's preferences and risk attitudes by a utility function u whose expected
value has to be maximised. The approach requires building v or p and u. This is a
laborious task of judgemental elicitation requiring precise information. However,
very frequently, the DM is not able to provide it (see e.g. Weber (1987)). More-
over, the assessment procedure assesses the numbers within certain limits (French,
1989) and there is some experimental evidence of the presence of biases and er-
rors in human judgements (see e.g. Kahneman et al. (1982) and the discussions
6
in Berkeley and Humphreys (1982), Bouyssou (1984) and Jaffray (1989)). Also,
the elicitation process is one of measurement and, as such, may involve some error
(see e.g. Lindley et al. (1979)). Thus, in practice, we have to face cases in which:
In any case, if, as we shall explain below, we concei ve of the decision making process
as a constructive one, we should have a framework to criticise that information.
This project develops such a framework.
Observe that the first idea we may have is that the Bayesian approach is inad-
equate, considering the cases stated above. We should look at other approaches,
as we do in the following.
Yager's method
Let S be a linearly ordered finite set, used to indicate preference information. For
instance, Yager illustrates his method with
the criteria specified by Ai. G(A;) = bG(i) ES denotes the importance of the i-th
criterion.
For each x, let Ci(X) = max(bp_G(i)+I, Aj(x)). Reorder the Cj(x)'s, calling them
rj(x), so that i ::; j=}rj(x) ::; rj(x). Then, Vager introduces the relation
where ::;L is the lexicographic order and r(.) = (rl(.),"" rn (.)). In other words,
the alternative proposed by Vager as optimal is that maximising
D(x) = m}n(max(bp_G(i)+I,Ai(x)))
Eventual ties are broken with the second least Cj(x), and so on. For a justification
see Vager (1981).
Example 1.1
Let S = {O,1,2,3,4,5,6} .
al (1,2,3,4)
az (2,3,4,5)
a3 (2,2,3,4)
a4 (1,5,5,,5)
Suppose that the DM thinks all the criteria are equally important: bG(j) =
bG(j), V i, j. Depending on the choice of that same weight, we get the following
rankings:
bG(i) ranking
0 (al, a2, a3, a4)
2 a4,a2,(aI, a3)
5 a2, a3, a4, al
Observe that
- the importance of the weights is the same in all cases, but there are
r anking reversals ,
• b. Suppose that the weights are (5,6,6,6). The ranking is a2, a3, a4, al. Even
giving more importance to all the criteria in which a4 is ranked best, this
alternative may not be the best.
These examples point out some flaws of Yager's method that deserve some analysis.
1. The method is a maximin one, thus having their shortcomings (see Milnor,
1954). Particularly, the property in c is not verified by them. Pessimism is
reflected in b.
4. Once the e;(x)'s are defined, we typically have the situation reflected in the
graph, where the number of criteria is two. The Iined part is the set of
alternatives less preferred than (Xl, X2). This is astrange way of imposing a
preference order.
9
v(a) = ~wivi(ai)
i=l
where .\ E {j, 1/ j}j=1, ... ,9' Its principal eigenvalue is 2. Its principal normalised
eigenvector is
so
'\1 E {jf(j + 1), 1/(j + 1)L=I,... ,9'
10
The table Oll the llext page presellts thirty Olle altematives, the (approximate)
lower bOUlld of the ).1'6 that make them optimal aIld the ).1'S gellerated by the
PEM covered by the respective interval.
For example, au will be optimal if ).1 E (.64156, .67742); 2/3 belongs to this
illterval. Observe that all the alternatives are 1l0t cOllvex domillated but many of
the illtervals do 1l0t cover allY Al gellerated by the PEM: many reasollable solutions
Call1lOt be optimal if we adopt this approach. Note that if ). were allowed to vary
freely in (0,00), we might generate aIlY 1l0t convex domillated altemative. "Saaty
illdicates that it is possible to assigll decimal values to respollses ill llumerical
mode, but the philosophy of the scale mitigates agaillst this" , Scholler alld Wedley
(1989).
The moral is: do 1l0t use this approach to generate the weights of aIl additive
value fUllctioll.
11
The AHP
This observation leads us to find out whether a shnilar phenomenon occurs with
the ARP. The method outlined above to generate weights is used in the ARP to
determine also the vi{ai)'s, see Belton (1986).
With A E {j, 1/ j};=I, ...,9 we may generate 17 Al weights. To follow the previous
idea, we would need at least 18 alternatives. The computational problems would
be tremendous in this case. Thus, we shall modify the ARP to the case j E {1,2}
and conjecture the result for the actual case.
Then, it is Al E {1/3, 1/2, 2/3}. Consider the matrix B ofpairwise comparisons
of ten alternatives according to the first criterion:
1 1 2 2 2 2 2 2 2 2
2 1 2 2 2 2 2 2
.5 .5 2 2 2 2 2 2 2
.5 1 .5 1 2 2 2 2 2 2
.5 .5 .5 .5 1 2 2 2 2 2
B=
.5 .5 .5 .5 .5 1 2 2 2 2
.5 .5 .5 .5 .5 .5 1 2 2 2
.5 .5 .5 .5 .5 .5 .5 1 2 2
.5 .5 .5 .5 .5 .5 .5 .5 1 2
.5 .5 .5 .5 .5 .5 .5 .5 .5 1
(.1604, .1487, .1331, .1223, .0999, .0869, .0756, .0658, .0572, .0498)
For Al = 1/2, a2 and a7 are optimal. For Al = 1/3, a7 is optimal. For Al = 2/3, a2
is optimal. In consequence, a1 and a10, which are not convex dominated, cannot
be generated as optimal by this modification of the AHP. Neither as nor as, which
are nondominated, may be generated as optimal by the AHP, though this should
cause no worries, provided that we have tested that v is additive.
Thus, for this simplification of the AHP, we face the same shortcoming. Other
issues are raised by Belton (1986) and French (1986).
• ai does not outrank a; (..,( aiSa;)) if the arguments in favour of the proposition
'ai is at least as good as aj , are considered insufficient.
More details on these relations canbe seen in Roy and Vincke (1984). Outranking
methods proceed modelling the outranking relation and exploiting it. We shall
14
Example 1.2
Objective 1 2 3 4 5 6
Weight 1/8 1/8 1/8 5/24 5/24 5/24
al 1 1/2 0 1 0
a2 0 1 1/2 1 0 1
a3 1/2 0 1 0 1 1
VetoTh. 9/10 9/10 9/10 2 2 2
Suppose that c = 1/2. We get
a l _a2
and the kernel is N = {ad. Consider now the problem with A = {aJ, a2, a3}.
The graph of S is
and the kernel is N = {a3, a2}. Thus, S is not transitive and ELECTRE I
does not satisfy the principle of irrelevant alternatives.
d\ d2 Graph Kernel
<1 > 1 a\ --+ a2 a\
>1 < 1 a\ f-- a2 a2
<1 < 1 a\ a2 {a\,a2}
>1 > 1 a\ +----+ a2 0
Thus, we may think of very unstable situations in which we may pass from an
'all alternatives acceptable' to a 'no response of ELECTRE 1'. This example
may be connected to the following criticism by Arrow and Raynaud (1987)
to ELECTRE: "The consultant tries different choices for the thresholds and
the weights in order to obtain a selected set as wished by the DM".
So, a number of serious flaws can be found for ELECTRE I. Moreover, "if one
possesses the elements necessary to the building of a linear utility function why
should one try a method as sophisticated as ELECTRE" (Arrow and Raynaud,
1987).
16
Distinct aids differ in the ways of choosing solutions and gathering information
and in the information gathered. We consider three particular approaches.
The first one, as presented in French (1986), assurnes an underlying value
function Vj thus, we try to solve max v( x) s. t. x E A. If v were known, we would
have a mathematical pro gram. Typically, we solve it generating a sequence {XI}
such that v(xd > V(XI_l), stopping when some convergence criteria are satisfied.
As we do not know v, we try to emulate the process as folIows:
mind(x,x"),
rEX
where dis some distance (thus, we are using the surrogate value function -d(., x"),
see Rios and Rios Insua, S. (1986)). Typically, the distances used are lpo A number
of interactive methods follow the scheme:
17
2. Compare them and eliminate the alternatives whieh are known to be worse
than the worst of them.
.ai
Supp~se that aj has the best value in the seeond co ordinate and aj -< a •. As
v is monotonie, we ean improve aj by moving in the directions of J{ = R~. Due
18
to the property of aj, we should only consider a == (1,0) among the directions
of Je Thus, we could determine aj == aj + ßa, ß > 0, such that aj '" a. and
eliminate the alternatives dominated by aj, discarding sorne efficient alternatives
in this operation. We shall call it 'elimination by aj'. Similarly, if aj had had the
best value in the first coordinate then a == (0,1).
Now, if Am designates the available set at iteration m, we shall define
Observe that:
• Condition 5 tests whether the best solution at this iteration coincides with
that of the previous one.
to determine a subset of alternatives which include the best ones and between
which the DM finds not much difference, on a first approximation to the problem.
It helped also to let the DM know what Daellenbach and Buchanan call "the
topography of the problem". To sum up, this approach might be of help as a step
prior to attacking the problem according to the lines described later.
From the critical analyses cited above, we can seleet several points that we
should demand from an !DA.
The problem with some of the IDA's is different. Some of them share the 3)(-
• The DM is not able to locate w precisely, but she may give some constraints
on it, which we shall represent by wES ~ W. This we shall call decision
making under partial information .
• The DM gives an estimate W of w, but, feeling doubtful about it, she asks for
some help to identify which of her judgements are the critical ones in order
to think about them more elosely. In this case, sensitivity analysis comes to
the fore.
n=1
Input to the
y
analysis: Sn, W n
More
precision
Output of the Sensitivity or
analysis: a~ analysis different
structure
required?
(e.g. Berger, 1985). Very few authors try to deal with the problem in its fuH and
true generality (e.g. Kadane and Chuang, 1978).
1 AB is generally understood.
25
ARBORIST
ARBORIST, see Samson (1988), is a product of Texas Instruments to construct
and analyse decision trees of up to 1000 nodes. Its rnain strength is its user-
friendliness.
Sensitivity analyses can be performed with respect to one parameter. The user
defines a range of values around the current value of the parameter and trial values
within this range. Expected values are computed for each of these values. Results
are plot ted.
SUPERTREE
SUPERTREE, see McNamee and Celona (1987), is designed by Strategie Decisions
Group to build and analyse decision trees. Some fiexibility in defining the utility
function is available, including a built in exponential utility function.
The Sensitivity option of SUPERTREE al10ws two sorts of one-dimensional
sensitivityanalyses:
• The Probability command plots the expected utility of the alternatives against
changes in the probability at a certain node. If the node has three or more
branches, the middle ones are set to zero probability. If the selected node
has probabilities that depend on another node, this dependence is deleted
and the same procedure is followed .
HIVIEW
HNIEW, see e.g. Barclay (1987), is a product of the LSE Decision Analysis Unit.
It handles, in a very nice way, multi-attribute models under certainty. Alternatives
are ranked according to a weighted additive value function
v(a) = L wivi(ai ).
i
VISA
Belton and Vickers (1989) developed VISA (Visual Interactive Sensitivity Analy-
sis) adecision aid for problems under certainty, based on a simple multi-attribute
value function
v(a) = LWiai.
i
Their system incorporates two modes of sensitivity analysis. The first one is one-
dimensional sensitivity analyses displayed by line graphs as in HIVIEW. The sec-
ond are interactive sensitivity analyses, where the user inputs a new set of weights
and values are recomputed. Bar charts support the presentation.
A general comment
1.5 Comments
The case is made for adopting a Bayesian approach to decision making. However,
new or more general tools are necessary to cope with decision problems under
partial information and sensitivity analysis in decision problems. These tools are
provided in the next chapters.
Chapter 2
The process of assessing v or u and p is essentially the encoding of the DM's judge-
mental inputs in parameters w E W, which inherit the corresponding uniqueness
properties. However, in practice the DM may not be ready or able to provide
the information necessary to locate w precisely. Instead, she may give some con-
straints on her judgements, which we shall represent by wES ~ W. We called
this decision making under partial information. We study it in this chapter.
We start by reviewing some basic results on quasi orders. Then, we give ax-
iomatic foundations for our problem: namely, how should the DM's judgements
be in our case. Briefly, the Bayesian approach to decision making assurnes that
comparative judgements follow a weak order, which leads to encoding judgemental
inputs in parameters w with some uniqueness properties. We shall see that quasi
orders regulating comparative judgements correspond to sets of feasible parame-
ters and, consequently, to constraints of our kind, wES. The problem is studied
for preference and belief modelling. A parametric model is introduced to analyse
problems under certainty and under uncertainty in parallel. This saves some effort
in defining solution concepts, designing algorithms, etc. Algorithms to compute
the solutions are provided for general problems and particularised for the most
common cases in practice: linear and bilinear models.
As the solutions proposed need not be unique, we analyse some ad hoc rules
proposed in the literature to solve the indeterminacy problem. We pay special
attention to hierarchical approaches.
29
Definition 2.1
Then, Eis an equivalence relation (Roberts, 1979). Let A* be the set of equivalence
classes. Define R* on A* as follows: let a,ß be two equivalence classes and a E a,
bE ßj then, aR*ß iff aRb. (A*, R*) is called reduction of (A, R) and is well-defined
(Roberts, 1979). Moreover,
Definition 2.2 Let (A, R), (A,S) be two bina7-Y relations. 5 is a covering of R
if a, bE A, aRb===}aSb. We write it R ~ S.
From now on, our binary relations shall be written (A, ::S). We shall say that
a ~ b if a::sb, b::Sa.
We give results that provide axiomatic foundations for this case. To do this,
we modify slightly the Bayesian principles. Briefly, these require that comparative
judgements follow a weak order (plus some conditions). Our approach requires
that comparative judgements follow a quasi order (plus some similar conditions).
Values, utilities and probabilities are treated in this way. In summary, we suggest
how to model imprecision within a Bayesian context.
We start by modelling general preferences under certainty. Then, we analyse
additive models, which comprise additive preference and probability models. We
conclude by analysing preferences under uncertainty. An introduction to these
results may be seen in Rios !nsua (1989).
Several authors (see Rios (1975) and the references in Gardenfohrs and Sahlin
(1988)) have proposed to work with families of utilities and probabilities to account
for imprecision in judgements. However, there has not been much work from the
axiomatic viewpoint. Some authors, e.g. Fishburn (1970) or Fine (1973), have
provided axioms for related cases. References are given where pertinent.
The same problem with similar motivations has been dealt within the fuzzy set
framework, see e.g. Freeling (1980) or Watson et al. (1979).
Ior a., aj E A.
I.(aj) = {1 ifada~
o otherwlse
Suppose that adaj. If ada. then adaj, thus Ik(a.) = Ik(aj). If "'(ada.),
then A(a.) = 0::::; min(O, 1) ::::; Ik(aj). Thus, Ma.) ::::; Ik(aj), Vk.
Suppose ..,(adaj). Then I.(aj) = 0< 1 = I.(a.). Thus, "'(A(a;) ::::; A(aj), Vk).
o
Example 2.1
a c d e f 9 h
a 1 1 1 1 1 0
b 0 1 0 0 1 0 0 0
c 0 0 1 0 0 0 1 0
d 0 0 0 0 0 1 0
e 0 0 0 0 0 0 1 0
f 0 0 0 0 0 1 0 1 0
9 0 0 0 0 0 0 1 1 0
h 0 0 0 0 0 0 0 0
I 0 0 0 0 0 0 0
32
Proposition 2.2 Let A be a finite set and ::::; a relation on A. (A, ::::;) is a quasi
order iff there are r real Junctions VI, .•. ,Vr such that
o
Observe that given VI, ... ,V r representing ::::; we have:
ai::::; aj
1=1 1=1 1=1
of Seott (1964) and implies his and Tversky's (1967) eondition. We eharacterise
additive quasi orders, unifying one line of work (Tversky, Adams, Krantz et al. ,
... ), whieh studies when a relation can be eovered by another admitting an additive
representation, and the other (Seott, Kraft et al., ... ), whieh eharacterises additive
weak orders. Our approach includes the weighted additive ease.
Let A = IT:'=I A;, where IAil < 00. The relation ~ on A models the DM's prefer-
enees and is interpreted as before.
where x = (Xl, . .. , x"), y = (yl, ... , y") E A, Vi is areal junction on Ai and '\i is
areal number, i=l, ... ,no We shall write v~(x) = L:'\iVi(Xi).
Initially, we look for the representation (2.1). Suppose the DM is able to provide
the component value functions, but she eannot establish the weights A; preciselYj
however, she ean say that ,\ = ('\1"'" ,\") E A. Thus, we are representing ~ on
A as follows:
x~y {=? I:" '\iVi(X i) ::; I:" '\iVi(yi), V'\ E 11. (2.2)
i=l i=l
X
• = (~C;X X i)"i=l' x. = (rr}li.n x i)"i=l"
We assume that x. -< x'. If it were not, then x. ~ x, Vx E A, and the problem
would be trivial.
Reealling definition 1.1 and proposition 1.2, we shall modify the eondition SC2
there included. First, we give an example:
34
Example 2.2
[(2,1), (2,2), (3, 1), (4,1), (1, 0), (1, O)]E4+2[(1, 2), (1, 1), (3, 0), (4,0), (2, 1), (2,1)]
ky will designate the k-uple (y, y, ... , y). The modification ia:
Definition 2.4 (Property P) (A,~) satisfies property [P] if for all integers
I, k ~ 1 and for all XI, ..• ,X/, XI+I, YI, ••. , yl, YI+I E A such that Xi~Yi, i = 1, ... ,I
and
Clearly, (A,~) is reflexive and transitive. Under the hypothesis of property [P],
V)" E A we have
I I
L V~(Xi) + kV~(XI+d = L V~(Yi) + kV~(YI+l)
I I
I I
LV~(Xi)::; LV,),(Y;)
I I
Proposition 2.4 (A,~) is a waqo iff it satisfies property P and xo~xo for some
Xo E A.
=> Obvious.
<== 'Vx E A, it is (xo,x)E2 (xo,x) and xo~xo, thus x~x and (A,::O is reflexive.
"Ix, y, z E A, (x, y, z)E3 (y, z, x); if x~y, y~z then x~z. Thus, (A,~) is transitive.
o
We shall use the following results and definitions from Rockafellar (1970).
Lemma 2.2 Let S be the system {blx > O,bjx ~ O,btx = o,x E Rn,i E I,j E
J, k E K}, where I, J, K are finite sets. S is consistent iff there are not numbers
J.!., J.!j, J.!~ with i E I,j E J, k E K, J.!" J.!j ~ 0 and at least one J.!' > 0 such that
{ {eb
r ~ O}r } S
eb. > 0
where the b's are vectors whose elements are in {-I, 0, I}. The assertion is true iff
S is consistent. Suppose S is inconsistent, then with K = 0 in lemma 2.2, there
are numbers /Ar ~ 0 and /A. > 0 such that
As the b's are vectors of rational components, the /A 's may be taken to be rationals,
and, consequently, integers. Also, at least one of the /Ar 's has to be positive. As
each inequality is related to a comparison xr :5Y" we deduce that
such that :5~ covers :5. Let A = P : :5 ~ :5~}. We know that A ::j:. 0 since
A= (1,1, ... ,1) E A.
37
Let :::;' = n.~EA :::;,1.. :::;' is a waqo such that :::; ~ :::;'. Suppose x, y E A are
such that x #- y, x:::;'y : VA E A, it is x:::;,l.y. Thus, if we write
a.s
I>i f>~yVi(xD ~ 0,
1=1
with a:: E {-1,0, 1}, this inequality is a consequence of the modified system
{ Ü::: Ai Er~l a~'vi( xi) ~ °O}' } S"
E Ai Er~l a~vi(xi) ~
(Observe that S" adds only to S' the equation
"i
~
l..J Ai ~ il i
l..J a, Vi( x,) = O.
1=1
is a consequence of
As brv #- 0, at least one of the J.I'S has to be positive. As the b's are vectors with
rational components, the J.I'S may be chosen rational. Thus, there are non-negative
integers ß., ß" ßry, with ßrv and at least one of the other ß's positive such that
This implies that there are Xi, Yi E A, i = 1, ... ,.9, with .9 > 0, such that xi:::! Yi, Vi
and
(Xl' ... , x., ßXYY )E,+ß•• (Y1, ... , Y., ßXYX)'
Then, by property P, x:::;y. That is, :::;. ~ :::;.
The :::;,I.'S are weak orders in A, which is finite. Thus, we may choose Al, ... , At
such that
t
:::;=n:::;,l.i.
i=l
o
Clearly, the representation of ::S need not be unique, but every convex set of
weights determining a family of value functions representing ::S is contained in that
defined by the system 5'. We have the following result, which restates corollary
2.2 and this comment
Corollary 2.3 Let Ai be the set 01 weights defining the wawo ::SA' = ::Si' Let A be
the set 01 weights defining the waqo ::S. II::s = ni ::Si, A :l [U Ai].
From this discussion, we may raise two issues. The first one is how to generate the
representation. To do this, we have to set up the system of inequalities generating
the waqo. Sometimes, it may be helpful to apply some algorithm to generate the
vertices of a polyhedron. A review of this topic may be seen in Matheiss and Rubin
(1980). The second one is that of dimension theory as it is done for strict partial
orders, see Roberts (1979). The objective is to find the minimum nu mb er of wawo's
generating the waqo. The solution of the problem might save some computational
eifort, both in terms of generation and storage of::S. Assurne that we start with
pI, ... , At} generating, via ::SA" ::S. We can use a branch and bound technique to
find a smaller number of ::S.\'s generating ::S. To do this, if M designates the bound
(to the number of ::SA's generating ::S) found up to the moment, the backtracking
rule would be 'go back if the current number is M-1'. A possible approach is:
1. L<--0.
2. If all the extreme points of A have been found, go to 4. Else, find a new Ai.
L<--LUPT
These questions of dimension and generation of ::S have a practical interest and
deserve further study. The problem will be illustrated in the paragraph dedicated
to belief modelling.
Kote that, due to monotonicity, we shall have among our statements VA(X') ::;
VA(X) ::; VA(X')' "Ix. This enables us to assurne that Vi(X:) = O.
The next two paragraphs analyse cases that can be adapted to this one.
39
where x = (x\ ... , x"), y = (y\ ... , y") E A and the Vi'S are functions on Ai.
Our initial rum is to find a representation like (2.3). However, the DM cannot
establish the functions Vi precisely. This case is similar to the previous one.
Definition 2.8 (A,~) is an additive quasi order (aqo) if it admits the represen-
tation
"
x~y ~ L:Vi(Xi):::; L:Vi(yi),VV E V
" (2.4)
i::::1 i=l
{= As in theorem 2.1.
* As in theorem 2.1, we associate a system
{
{cbr ~ O}r} S
cb. > 0
to j and prove the existence of an awo covering ~. Let V be the set of additive
value functions v whose associated order ~v covers ~. Define ~. = nvEV ~v' ~.
is an aqo covering j. Consider x, y E A such that x~'y, x =I y. We associate
an inequality cbxy ~ 0 with it. As in theorem 2.1, we see that this inequality is a
consequence of
Observe that the only difference between both results is that once we have
proved there is a jv covering j, theorem 2.1 uses a system in the '\'s and theorem
2.2 uses a system in the v's (or c). We have also
40
Corollary 2.4 Let Vi, .•. , VI be additive value functions such that
Then,
I I
x-j,y <=? 'EAiVi(X)-j,'E.\;vi(y),VA: A ~ O,'EAi = 1.
i=l i=l
Scott (1964) gives necessary and suflicient conditions for an awo when n = 2.
Fishburn (1970) gives necessary and sufficient conditions for an awo using -< as
the primitive relation and studies other kind of representations, different to our
vectorial one. Krantz et al. (1971) give sufficient conditions for the existence of
an awo covering an aqo.
Belief modelling
(1969, 1985) uses them to represent probability se mi orders and interval orders. Our
result weakens Seott's axioms, giving a simple and elegant qualitative treatment
of probability quasi orders. Giron and Rios (1980) provide a eharacterisation in a
more general setting.
Let 8 = {1, ... , k} be the set of states (k ;:: 1), A the algebra of subsets
of 8 and P a probability distribution on (8, A). We assoeiate with eaeh subset
B E A its eharacteristie function XB. The k-uple (xB(i))7=1 is written XB. It
is p( B) = LEB Pi =L PixB. We ean apply the previous results to the case in
whieh the DM eannot establish her beliefs preeisely. Let us first reinterpret the
properties.
The relation will be written :::e and interpreted as in section 1.1. The mono-
tonieity property means here B <;;; C=*B:::eC. The eondition x. --< x' means
°--<e 8 . Property P is read:
Definition 2.9 (Property P) Let 8, A,:::e be as above. (A, :::e) satisfies prop-
erty P if for all integers I, r ;:: 1 and fo1' all Al, BI, ... , A1+I,B1+I E A, such that
Ai:::eBi, i = 1, ... ,1 and
1.0--<e8,0:::e B ,VBEA.
2. Prope1'ty P.
Theorem 2.3 Let 0, A, -jt be as above. (A, -jl) is a pqo iJJ it is the intersection
0/ (a finite number 0/) probability weak orders.
The proof is analogous to that of theorem 2.2. Observe that now we assoeiate with
B-jeC an inequality
As we told, we ean make PirOl = 0 and, ealling Pi(l) = pi, and p = (PI,"" Pk),
we rewrite the system as
2. Property P,
Again, we have:
43
Example 2.3
pi p2
1 .1 .1
2 .2 .3
3 .3 .4
4 .4 .2
1,2,3,4 1,2,3,4
2,3,4 2,3,4
1,3,4 1,2,3
1,2,4 3,4 2,3 1,3,4
1,2,3 2,4 3,4 1,2,4
2,3 1,4 1,3 2,4
1,3 4 2,3 3 1,2
1,2 3 2 1,4
2 4 4
1 1
e e
PI + P3 P4 PI + P4 P2
PI + P2 P3 PI + P2 P3
SI PI + P4 P2 + P3 S2 P2 + P4 PI + P3
Pi ~ 0 Pi ~ 0
LPi l(or > 0) LPi l(or > 0)
44
P2 2: PI
P4 2: Pt
Pt +P2 P3
Pt +P4
2: P2
S12 Pt +P3 2: P4
P2 +P3 2: Pt +P4
P2 +P4 2: Pt +P3
Pi 2: 0
EPi l(or> 0)
The extreme points of the polyhedron defined by S12 are
a (0,1/3,1/3,1/3)
b (.1, .3, .4, .2)
e = (.1, .2, .3, .4)
d = (1/6,1/6,2/6,2/6)
b and e are the initial solutions. Observe that in this example [U1'il f; 1', where
1',1'i are the sets of probability distributions defining ~,~i'
We can also see how the algorithm proposed to generate ~t would work in this
example:
Trial M ~t generated?
a - no
a,d - no
a,d,e - no
a,d,e,b 4 yes
a,d,b 4 no
a,e 4 no
a,e,b 4 yes
e,b 3 yes
b 2 no
Thus,
BID~(CID{::=}p(BID) ~ p(ClD), Vp E P : p(D) > o.
o
• AL j . is a quasi order.
p-<.q==} E( u, p) < E( H, q)
for some real function u on C. \Vhite (1972) gives conditions to represent an order
in Rn bya finite family of linear value f\lnctions, which may bc translated to this
context.
The following simple result establishes the necessity of the main conditions we
shall use to get the representaLon (2.5).
Proposition 2.7 Suppose that (Po(e), j.) admits the repTesentation (2.5). Then,
• B 1. j . is a quasi order.
Definition 2.11 Given a set B, we call K(B) the minimal convex cone contain-
ing Band R( B) the set of rays geneTated by B.
Lemma 2.5 Let C be a convex set of Rn other than Rnitselj. Then, there exists
a closed half-space containing C.
47
Theorem 2.4 Let C, 'Po(C),~. be as before. Suppose that ~. satisfies B1, B2.
Then if:
Let us eali
D = {p - q : q-<.p}, E = {p - q : q~.p}.
Similarly,
1'. D is convex.
Extending the above argument by indudion we may prove that
2. aj > O,L:aj = 1,qj::S.pj=*L:ajQr:5.L:ajpj.
2'. aj> 0, L: aj = 1, qj-<.Pj=* L: ajqj-<. L: ajPJ.
Then:
3. a] > 0, L aj == 1, qd.pj, j == 1, ... , n - 1, L ajqj = L ajpj=*Pn~.qn'
Because of 2, we only have to prove the result for n = 2. If a E (0,1),
48
E(u,q):::; E(u,p).
o
Unfortunately, it is not possible to characterise the cones S in terms of the half
spaces containing them, as can be seen with T = {x : X3 2: 0, Xl > 0, X2 > O}
and S = TU {O}. This causes difficulties in getting the representation (2.5). To
overcome these, we shall add axiom B3. Aumman (1964) provides a related result
of more complicated proof and under stronger conditions.
49
Let E be a.s in theorem 2.4. cl(E) is convex (and closed). Let u E cl(E). There is
Uo E E such that
auo+ (1- alu E E,Va E (0, 1J.
therefore
aqo + (1 - a)q'j,apo + (1 - alp', Va E (0, 1J.
By B3, q'j,p' and u E E. Thus, E is closed. Consequently, S is a closed, convex
set (cone) and it can be characterised by means of its support hyperplanes: there
is a family W of vectors such that
Then,
We have
ai ~ aj <==? p(ai) ~. p(aj), Vp E P
<==? E(u,p(ai)) :::; E(u,p(aj)), Vu E U, Vp E P
(2.6)
E(u,p) == E(u,q),Vu E U.
As before, we have
Corollary 2.7 Under the conditions of proposition 2.8, there is a convex family
U of additiveutility funclions giving the representation (2.6).
• For a given v E V, let w) = vi(aj) and W = (w}, ... ,w;n, ... ,w;:'). Let
5 be as before. 5 is now defined by constraints of the kind wj :S w! and
I: w! :S I: w) (one of the inequalities is strict). The evaluation of the j-th
alternative takes the form Wj(w) = c;w, where Cj is a vector of O's and 1's .
The main thrust of the parametric model is to order the alternatives in a Pareto
sense. We recall that
Clearly,
Observe that if the W/s are continuous there is a neighbourhood N( wo) of Wo such
that 'l1k(W) < 'l1 j (w), for every W E N(wo) n S.
It may seem appealing to propose as solutions those aj that maximise Wj( w)
for some w E 5. They have received several names in the literature, e. g. quasi-
Bayesian (Rios, 1975, in a context in which there is uncertainty in the probabilities)
and potentially optimal (Hazen, 1986, in an uncertain value or utility functions
context). We adopt the latter term.
54
Definition 2.14 aj is potentially optimal (p.o.) if 'h(w) ::; 'l!j(w), Vak E A, for
some wES.
and
Spilzrajn's lemma (Fishburn, 1970) provides the basis for the existence of a con-
sistent value function in White's sense. V will designate the set of all of them.
We only demand the second of the above conditions; proposition 2.2 provides the
basis for the existence of a family V of almost agreeing value functions character-
ising ~. Other results in this chapter provide characterisations of ~ under special
conditions.
Differences between both approaches are not dramatic. The second one extends
Pareto order in Rn and allows characterisations in more cases (e.g. belief quasi
orders or preference quasi orders under uncertainty); however it cannot cope in a
simple way with weak Pareto order in Rn (this is more a technical rather than a
practical shortcoming). The first one can deal with Pareto and weak Pareto orders
very easily, via weighted additive sums.
In any case, the primitive solution set is that of nondominated alternatives,
[(X,~) in White's notation. In his framework, White (1980b) proves that if V is
the set of all consistent value functions (in his sense), this set coincides with
and the set of p.O. alternatives. This does not happen if we work with sets V' C V
and motivates his discussion about the relations and use of those sets, which is
relevant to our case.
55
Thus, our analysis will be in terms of p.o. and nondominated solutions. We shall
explore some relations between those eoneepts, taking into aeeount our par ametrie
representation.
Neither a p.O. alternative needs to be nondominated nor a nondominated al-
ternative needs to be p.O.
Example 2.4
In R Z , eonsider the example of figure 2.1. We assume a linear value function with
unknown weights and the constraints on the parameters defined by S = {Wl, W2 ~
The following results will be used in later seetions. We shall assume that the
W/s are continuous.
56
Proposition 2.11 Let p. be a measure on S such that p.(A) > 0 for every set
A = N(w)nS , where w E S, N(w) is a neighbourhood ofw. If gj = J5 'l1 j(w)p.(dw)
exists (and is finite) for i = j, k and gj ~ g" I a" does not dominate aj.
Suppose a" doroinates aj. There is a set B C S such that p.(B) > 0 and 'l1,,(w) >
'l1 j(w), Vw E B. In Be, it is 'l1,,(w) ~ 'l1 j(w). So
a contradiction. o
Proposition 2.12 Suppose that S is convex and bounded and a" does not domi-
nate aj. Then, there is a measure p. as in proposition 2.11 such that gj ~ 9".
Defining
p.(C) = k(m XB + XBc)djJ
for every measurable set CeS, we get the searched measure. o
Example 2.5
Consider a bi-attribute problem under certainty where the Wj'S are linear: Wi( w) =
w(a; - an + a; with w E [0,1]. Suppose there are three alternatives al = (1,0),
az = (0,1), a3 = (1/4,1/4). Let f be the density of}l. We may assume that
f~ f(w)dw = 1. We have:
gl = l wf(w)dw
g2 = l(l- w)f(w)dw = 1- gl
g3 = 1/4
and 0 ::; gl ::; 1. It cannot be
Under the conditions of proposition 2.12, the result is true for alternatives
which are single optimisers.
Proposition 2.13 Suppose that Wj (w) > Wk( w) I Vk f. j, for some wES. Then,
there is a measure }l such that gj 2: gk, Vk.
Corollary 2.8 Let W E 5 and suppose that aj is optimal for that W. Then:
2. If W E int( 5), the \[I's are differentiable and for every k such that \[I j( w) =
\[Ik(W), V\[Ij(w) -# V\[Ik(W),
aj is nondominated.
Part 1 is immediate. To prove part 2, for those ak such that \[Ij(w) = \[Ik(W) apply
proposition 2.14; for thc others, part 1 applies. 0
Proposition 2.16 5uppose that 5 is convex and compact and \[Ij - /L<I>j is upper
semicontinuous and quasi concave for every /L E iVf = {/L : Li;ij /Li = 1, /Li :::: O}.
If aJ is not p.o., 3/L E 1',;[ such that
thus
59
Then, under the hypothesis of the proposition (see Stoer and Witzgall, 1970)
The same proof applies when aj is not a single optimiser for any wES,
substituting ~ for <. Note that J.Lif!j is not always meaningful.
In view of the previous results, the search should be li mi ted to the nondomi-
nated alternatives which are potentially optimal. We adopt this approach in the
rest of the study.
wES
Then, ak ~ aj iff its optimal value is greater than or equal to O. When the
optimal value is greater than 0, we know that ak -< aj. When it is 0, we should
check whether aj ~ ak. In many cases, however, this is not necessary. Several
distinct algorithms may be used to perform the minimisation problem.
60
1. ft-O,Bt-0.
4. If b(f) = 0, go to 3.
5. Bt-BU{at},b(C)t-O,st-l.
6. s+-s + 1. If s > m, go to 3.
7. Ifb(s)=0,got06.
1. B +-0, f+-O.
2. f+-f + 1.
3. If f > m, stop.
5. Go to 2.
[Pj] mlll Zj
s. t.
[Pi] min Zj
W ES,%j
s. t.
Zj ~ 0
Then, aj is p.o. iff the optimal value is O. For this, we are assuming that aj is not
the only p.o. alternative, the w's are continuous and S is connected.
In principle, we should have to study all these problems for each alternative
with each constraint. Some effort may be saved using the following observations:
• If ak is known to be not p.o., the constraint relative to ak may be dropped
from the problems Pj or Sj.
For example, if we are analysing the consistency of systems, if the system
Example 2.6
Consider our initial problem given by adecision table. Suppose that the utilities
u(aii) = Uii (of action ai under state Bi) are known precisely but we only know
that probabilities satisfy some constraints of the kind
We shall write them pEP. We have that IVj(p) = UjlPI + ... + UjkPk. The
problems we have to solve are:
• Is a, j aj?
Consider the LP problem
• 1s aj p.o.?
Consider the LP problem
mm Zj
P EP,z,
s. t.
Essential!y, this is the way to deal with linear problems. We consider now the
problems model!ed in section 2.2.
Additive models
Example 2.7
These examples fit into the model we are considering, which we analyse now.
66
• Is a/ j aj?
We have to find if
B.t.
a/ j aj iff the optimal value is ~ o. The feasible set is conical: the optimal
value can be 0 or -00. Since w = 0 is a feasible solution, a/ j aj iff w =0
is an optimal solution .
• Is aj p.o.?
We have to find if the system
(Cj-C;)t w ~ O,i-j.j
Cw ~ 0
ctw > 0
is consistent. This can be studied solving the LP problem
s.t.
(Cj -Ci)t w ~ O,i-j.j
Cw ~ 0
Again, it is an LP problem with conical feasible region: its optimal value can
be 0 or 00. aj is p.o. iff w =0 is not optimal.
67
The following results relate dominance and potential optimality. They are partic-
ular instances of results from seetion 2.4.
Proposition 2.17 5uppose that w E int(5), Wi(W) = Wj(w) and Wi(W) 1= Wj(w)
for some W E 5. Then, ~(ai~aj), ~(ada;).
Example 2.8
Corollary 2.10 Assume that 5 is defined as in the model and aj is not p.o. Then
there are Ai ~ 0 such that == Ai = 1 and
Example 2.9
C1Wl ?: a
C;Wl > 0
C2W2 ?: ß
c1 w2 > 0
See sections 2.3 and 2.6.1.
CW?: o.
• Is al ::; aj?
S.t.
Cw 2: 0
c; Wl ?: 0
c1w2 ?: 0
• Is aj p.o.?
We have to determine if the system
S.t.
w;(Bk - B j )W2 :S 0, Vk -# j
Cw 2: 0
C;WI 2: 0
C~W2 2: 0
We assume the appropriate conditions ensuring the existence of the ~(aj)'s (e.g.,
continuity of the evaluation functions and compactness and convexity of the pa-
rameter set). Note also that the W;'s have to correspond to an interval scale and
that intervals be interpretable if we want the Laplace and maximin return crite-
ria to be meaningful, whereas for the other criteria we only require that the Wj 's
correspond to an ordinal scale. Schneller and Sphicas (1983) define the criteria
in their original context. Their example shows that they may lead to different
answers.
None of these solutions are suggested by the proposed axioms: these only imply
that we may associate a certain parametric model to the problem, the solutions be-
ing the nondominated ones; potentially optimal alternatives seem also acceptable,
both on intuitive grounds and because of their relations with the nondominated
ones. It is natural, then, to study the compatibility of the above criteria with
this framework. Proofs and counterexamples will not be provided, since they are
simple.
This corresponds to OUf not ion of consistent value function, see the discussion
in section 2.4. Clearly, all the criteria are compatible with ~.
Only the Laplace solution satisfies this criteria, under certain conditions (sec
2.7.2; if this is the case, Laplace solutions are potentially optimal).
According to the information provided, the DM cannot find any of the p.O. non-
dominated alternatives better than the others. If a single alternative has to be
chosen with the available information, any of them would do as weil. None of
those rules are the right one; pretending the contrary implies assuming informa-
tion that might be incompatible with that available. In short, these criteria are
ad hoc.
where -:::.' is a preference order on A, such that -:::. c:; -:::". Obviously, we have to
be careful in the elicitation of U and (, since the choice depends on them. This
points out the main problem. We are assuming that the DM is unable to provide
more information on w in the primary problem. Observe that :
• ( and U are defined over presumably quite abstract objects for the DM,
In summary, this does not seem the best way to help the D?vI think more carefully
about her problem. It is very likely that we end up with imprccision in thc second
72
2.8 Comments
We have studied how to deal with decision making problems under partial infor-
mation. The same principle allows us to model preferences under certainty, beliefs
and preferences under uncertainty. All problems can be treated with the same
parametric model, whieh associates a function Wj( w) with each alternative aj and
includes some constraints wES on the parameters to represent the partial in-
formation. The solutions suggested by the model are the nondominated ones, in
the Pareto sense. Potential optimality is a reasonable concept whieh allows us
to reduce the final pool of solutions. Algorithms to compute those solutions are
provided. In summary, we have a more robust theory of decision making based
on a weaker set ofaxioms, but embodying coherence, since it essentially implies
carrying out a family of coherent decision analyses.
Very frequently, the set of solutions is not a singleton. To achieve further
reduction, the DM has to provide more information. This rules out the automatie
use of a number of ad hoc criteria. A hierarchical approach is a way of eliciting more
information; however, given the circumstances originating that sort of analyses, we
have to view them as ad hoc, as well.
The results of this chapter will be used in the next one to identify the possible
competitors of a current optimal solution, on a first phase of sensitivity analysis.
Illustrations of the previous ideas are given in chapter 4.
Chapter 3
Sensitivity Analysis in
Multi-objective Decision Making
The basic data in chapter 2 were the evaluation functions Wj and the constraints
wES. In addition, we now have an estimate W of w. We rank the alternatives
according to Wj(w). The current best alternative will be called a. and its evaluation
W.(w). However, we must be ready to criticise that ranking: sensitivity analysis
comes to the fore.
The first section introduces technical results about the maximum ranking func-
tion and the subsets of optimality. The need for sensitivity analysis tools is then
justified. An initial approach is to proceed according to a DPI problem and find
the possible competitors of a•. Certainly, sensitivity analyses are necessary if a.
has competitors, but several arguments try to justify their irrelevance, even in this
case. We present the most important of them, the Flat Maxima Principle, and
refute it on technical, practical and conceptual grounds. The message we convey is
that sensitivity analysis is a constructive way to clarify the DM's mind suggesting
to her the judgements she should consider more carefully. As a result, we mention
some properties that a sensitivity analysis tool should have.
Several concepts of adjacent potential optimality allow us to reduce further the
set of competitors of a •. Then, we introduce and analyse so me sensitivity analysis
tools based on gauge functions. Remarks on errors in judgemental measurement
lead to same other tools.
75
'i'(W) = max'i'j(w).
]
Example 3.1
We have
We have
11<
The example refers to abilinear model. We study these using the notation
W = (Wt,W2) and Wj(w) = Wj(Wt,W2) = W:BjW2. Thus, W(w) = W(Wt,W2). The
following result is easy:
Proposition 3.1 W(Wt, W2) is a piecewise linear convex function 0/ Wl (W2), for
a fixed W2 (wt}.
We can give abound, similar to that of convex functions, for this case. Its proof
is simple.
Proposition 3.2 Suppose that Wj(Wt, W2) = W:BjW2, Vj and J.L E [0,1]. Then
so that
and
The result can be repeated for every pair (w, w') E Sand we get the bound
The hypothesis about K may be strong. As 'Ir is subsmooth, see Clarke (1975),
we can use the following result:
So 'Ir is differentiable except at a set of measure zero. For a given w, let J(w) =
{j: 'Irj(w)::: llI(w)}. Then, 111 is differentiable in w iff 'i7'Irj (w) = 'i7'Ir k (w),Vj,k E
J(w).
The theory of generalised subgradients, see Rockafellar (1981) or Clarke (1983),
gives a general concept of differentiability of 'Ir which will be useful. We need the
following results and definitions:
Definition 3.1 Let I be areal function defined on some vector space X. The
directional derivative 01 I at x in the direction y is
Definition 3.3 Let f be a locally Lipschitzian real function defined on some vector
space X. The generalised subgradient of f at x is
and
5f(x) = [{V<P,\(X),A E J(x)}].
When f is eonvex, it is z E 5f(x)~f(x') 2: f(x) + (x'- x)tz,Vx', whieh is the
c1assieal definition of a subgradient. If f is a 'max' function
Let us apply the previous results to our problem. Results 1 ean be used in
loeal directional approximations; results 3 ean be used to give bounds for the loss
of optimality; results 2 are prior to results 3:
• Linear case
3. As II' is eonvex,
• Bilinear case
W(w;,w~) 2: W(Wj,W2)+
((w; - wd, (w~ - W2))'( I: {Lj(BjW2, wiBj)) + o(lw - w'I),
jEJ(w)
with {Lj 2: 0, L {Lj = 1. Observe that, in this case, we can deduce
another kind of inequalities, taking into account the partial convexity
of W:
W(w;, w~) 2: W(w;, W2) + (w~ - W2)'02 2:
W(Wj, W2) + (w~ - W2)'02 + (w; - wdoj,
where 02 is a sllbgradient of the convex fllnction
Ww;(z) = max(w;'BJ)z
J
W(w;, w~) 2:
• General case
We assume that 'VW)(w) is continllous Vj.
Let wES and suppose Wn->W, with {w n } C Sj. Then, Ilij(w n ) 2: Ilik(w n ), Vn, Vk.
As the Ilik's are continuous, Ilij{w) 2: Ilik(w),Vk. So w E Sj. 0
This example shows that the inverse indusion is not true in general.
Example 3.2
Under the conditions of example 3.1, suppose there are two alternatives:
BI B2
a1 (1,1 ) (0,1)
a2 (1,0) (0,2)
The thickest part belongs to SI nS2' In this case, we have int(S2) = {(p,'\ ):
p< 1/2}. So, int(S2) 1. C2.
Example 3.3
Und er the conditions of example 3.1, suppose there are three alternatives
BI B2
al (1,0) (0,1 )
a2 (25/51,25/51) (25/51,25/51)
a3 (2,-98) (-98,-198)
Then
s,
The figure shows the graph of the subsets of optimality. The Sj's need be
neither convex nor connected.
Convexity requires that the segment linking any two points of Sj is in Sj. We
shall be interested in another property: is there a segment in Sj whose extremes
are a certain point of Sj and any point of Sj nSI, for some I? Example 3.3 allows
us to answer it negatively. It is easy to see that [(.5,0), w] ct. SI, Vw E SI nS3'
Under some conditions, there is a stronger result (proposition 3.11). We need
some previous resuIts and definitions.
We would have a contradiction if w' were an isolated point of {w : ll1 i ( w) = Wj( w)}.
If it were not, since in addition
min(W.(w) - maxWj(w)).
wES J
min(
w
c: W - max cjw)
J
S.t.
Cw~ 0
min(wiB.wz - maxw;Bjwz)
W J
S.t.
Cw ~ 0
CiWl ~ 0
c~wz ~ 0
We may be tempted to use this as a sensitivity measure: if -u' is smalI, losses are
not big when a. is not the optimal alternative. The term close call (Bell et al.,
1988) defines that idea: two alternatives are a elose call if their utilities are elose.
This definition lacks rigour. 1s u' meaningful? Since it is a difference of evalu-
ations, it is meaningful if the w;'s correspond to a measurable value function (see
85
French (1986) for this concept and its relation with various concepts of meaning-
fulness). Then, u' gets its meaning by comparison with differences in the same
value function, that is by comparisons of the kind
which would tell us, informally, how a. would perform if wwere the true parame-
ters. That quantity is invariant under affine transformations. We could study the
function
g(w) = (lII(w) -1II.(w))/(III(w) - m.in lII i (w)).
I
However, gis not very manageable. Von Winterfeldt and Edwards (1987) propose
studying a related function, the relative expected loss, but still this one has to be
stupied numerically.
These sorts of functions or measures have some shortcomings: it may not be
possible to relate their values from one phase of the analysis (when the parameter
set is Sn) to the following (when it is Sn+1); moreover, they do not suggest how
further analysis should be carried out.
We can think of two other approaches to try to explain the implications of u'
to the DM:
• We can use relative errors as in physical measurement, see Pugh and Winslow
(1966). Effectively, we are measuring the value/utility of the optimal alter-
native. lu'l represents the worst (absolute) error. We should consider the
relative error
lu·I/III(w).
We shall use an informal version of this approach, as illustrated in chapter 4: we
announce which is the maximum loss and where do the ratings of the incumbent
alternatives lie. This is far from rigorous but used consistently may have a didactic
effect on DM's.
This topic of relating closeness of alternatives to the preference structure is
very important and deserves much study. Bell et al. (1988) and White (1989)
rruse this issue, which is studied to some extent in Rios and Rios Insua, S. (1986).
86
Example 3.5
--
-~------
If w were the value of the parameter, the DM would choose a •. If the actual
value were w, the loss of optimality would be
where ii; = (iiW j iiw;) Iw. The expression on the right is the first order Taylor
approximation to W(w) centered at w. So W.(w) approximates W(w) near w, the
error of the approximation being given by
h(w) 21j2A.llw-wll.
Assurne that the Iw; - wd are smalI, but they are bigger than a certain L Then,
h(w) 2 1j2A.Eyn.
Then, as the number of parameters increases, the loss of optimality may be big
even if w is near wand W is smooth.
Moreover, W need not be smooth. For example, if therc is a finite number of
alternatives and none of them is s.o., W may be nondifferentiable at some set (of
zero measure). In that case, the argument leading to equation (3.1) is generalised
to the corner points, using directional derivatives. However, the loss of optimality
may be big if w is near a corner point.
Also, in our model, the expression 'w near w' is [ar from convincing: S repre-
sents our 'sure knowledge', so the study of h has to be done not only near w but
in all S.
But, perhaps, in our claim for a general approach to sensitivity analysis, the
most important point is that we should consider sensitivity both with respect to
88
utilities and probabilities. Then, 'lI need not be convex, as we have seen in example
3.1. Thus, one of the basic arguments no longer holds. In connection with this,
suppose the 'lI/s are bilinear, say 'lIj(w) = wiBjw2. Let us study the variations of
g(w) = 'lIj(w) - 'lI.(w) = wi(Bj - B.)W2 =wiDjW2, calling Dj = Bj - B•. If we
limit the analysis to one type of parameter, as assumed by the FMP, and we vary
Wl to Wl + tl we have
Results are similar, if we limit the analysis to the second type of parameters.
However, making a general analysis we get
neighbourhoods of w),
According to this, we should agree that the work introduced before fits par-
tially into this framework, but, certainly, the sensitivity measures were not very
constructive. It seems also that they only provide us with an uninformative list of
those alternatives to which the DM should confine her attention.
More constructive tools are explored in the next sections.
Example 3.6
Consider the example of the figure with four bi-attributed alternatives. We suppose
that the value function is linear with unknown weights and the constraints on the
parameters defined by S = {Wj, Wz ~ 0, Wj + Wz = 1}.
90
S. t.
'lI;(W) - 'lI.(w):::; O,Vi
Example 3.7
(Cj - C.)'w 0
(Ci - C.)'W ::; O,i -# j
Cw ~ 0
C'W > 0
maxc'w
S.t.
(Cj - c.)'w = 0
(Ci - c.)'w ::; 0, i -# j
Cw ~ 0
wi(Bj - B.)w2 0
wi(Bi - B.)W2 ::; 0, i -# j
Cw > 0
ciwI > 0
C;W2 > 0
5.1.
wi(Bj - B.)w2 =0
wirB, - B.)w2::; O,i-#j
Cw ~ 0
ciWI ~ °
C~W2 ~ 0
Say ,\j = 00 if the problem is infeasible. Let ,\! = min'\j and suppose ,\! < 00.
The minimum is achieved at some J. Then, if S is convex, aJ is apow to a •.
The proof follows the same pattern of those in section 3.4, to which we refer.
Conversely, if aJ is apow to a., ,\! = ,\J for every w E T. nSJ.
We can use problems [P!j] (see section 3.3.2) to find the sapo alternatives to a •.
More manageable procedures will be introduced in the next section, where this
concept is justified via perturbations of w within convex sets. Hansen et al. (1989)
have used problems similar to [P!j] in sensitivity analysis in multi-objective linear
programming.
93
To prove the first part just apply the definitions. For the second part, J. = J",
comes from the connectedness of S. and J. = J comes from its convexity.
IM 0
In summary, if the problem is linear the concepts are identical. The example
shows that this is not true in general.
Example 3.8
Then, 1I'4(P,'\ ) = -200p + 200'\ - 198. We sketch the sub sets of optimality.
w E 5,0
s. t.
w -w E [(a
This gives us a Wj, the solution of problem [PKj ], for which aj and a. are indifferent
and which minimises er.
[( is associated to a function 9K : [(00---> R, whose minimum is attained at O.
We can solve the equivalent problem to [P Kj ]:
min 9K(W - w)
wES
s. t.
For this method to be of use, we have to restrict the dass [(. We shall consider
sets [(a which, in addition, are star-shaped with 0 a point of common visibility,
that is, if w E [(a, [w,O] C !{a (see Stoer and Witzgall, 1970). Then
Proposition 3.13 Assume that 5 is convex and the 'Ir j 's are continuous. Let erj
be the optimal value oJ[PKJ !f erj > 0, 'Irj(w) < 'Ir.(w), Vw E (w + [(a) n5, for
a < aj.
Suppose that wE (w + [(a)n5, a < a j and 'Ir j (lV) :::: 'Ir.(w). Then, as 5
is convex, [w,w] C 5 and 'Irj(w) < 'Ir.(w), there must be w' E [w,w) such that
'Ir}( w') = 'Ir. (w') and w' E (w + [(a) n5, against the definition of a]' 0
95
Let p = minj aj. The solution of the problem is achieved at a certain J. Then:
Proposition 3.14 Assume that S is convex and the Wj 's are continuous. aJ is
sapo to a •.
We have that a ::; pimplies W.(w) ::::: Wj( w), Vj, Vw E (w+ I<",) nS. Therefore,
p is an absolute measure of the insensitivity of a •. Let us solve the problem
max a
w E S,a
s.t.
and call its optimal value 8 and its optimal solution w". Note that I<p is the
biggest subset in I< such that (w + I<p) nSeS. and I<. is the smallest subset in
I< such that Sc (w + I<.). So
R = vol(I<p)/vol(I<,)
may be seen as a relative measure of the insensitivity of a. to changes in w. It is
a surrogate of the measure vol(S.)/vol(S). As R may be difficult to compute, we
approximate it by
r=p/8.
We have:
• If ,. = 1, a. is completely insensitivc,
• If r = 0, a. is completely sensitive,
r provides some guidance about how furt her analysis could bc carricd out. To
increase it, we may either increase p or decrease 8. To decrease 8, we have to
reduce S; we have a direction in which we might try the reduction: [w",w], which
gives 8. To increase p, we have to compare (holistically) aJ and a.; if the DM finds
a. superior, we may rule out aJ, thus incrcasing S. and, possibly, p. We should
be careful, however, with such holistic comparisons. Note that both actions have
96
If a :S p, w.(w) 2 Wj(w) Vj. Thus, we do not have to worry about the optimality
of a., Vw E (w +K p) nS, but moving outside (w + K nS means that we need
p)
to worry. WJ marks, in asense, the point at which we need to begin to take care
and can be used as apointer to sensitivity issues.
This approach is very flexible: choices for K or gK are multiple. For example,
we can choose the system of balls of a distance or, equivalently, the distance as gK
as in French and Rios Insua (1989); if W is naturally decomposed as (Wj, wz), we
can choose systems K j , K z for the first and second subsets of parameters and use
the system K = K j X K z; if W refers to probabilities we can use relative information
measures, such as entropies.
Definition 3.10 The gauge /unciion 0/ C is the /unciion ,(.IC) : R t ---> R+ de-
fined by ,(xIC) = inf{oo,'\ 2 0: X E '\C}.
We shall use a specific kind of gauge functions. To motivate the choice, remember
that we had to solve the problems [PZj ] to generate the sapo alternatives; the
DM may not find it easy to specify y. Problems [PKj ] provide the possibility
of solving simultaneously a family of those problems, as weil as to investigate
the perturbation of w within K. As we know, they lead to sapo alternatives
(proposition 3.14). We may wonder if the reciprocal result is true. The dass
we shall choose provides a positive answer under certain conditions. ~Ioreover,
97
mm p=gp(x)
s.t.
x = p L \qi
Ai::: O,I:\ = l,p::: 0
mm gp(w-w)
wES
s.t.
We can rewrite it as
w E 5,>.,g)
s.t.
WJ(w) = W.(w)
Ai ::: 0, I: Ai = 1.
max hJ
w E S,>.,h,
S.t.
hj(w - w ) = L )..,qi
Wj(W) = W.(w)
).., 2:: O,LAi = l,h j 2:: 0
which is simpler. Observe that gj is the minimum A such that the image
of the polyhedron w + P by the application z >---> W + A(Z - w) intersects
{w: Wj(w) = W.(w)}.
Similarly, for the maximum problem we solve
max 9
w E S,>.,g
S.t.
W -w =gLA iqi
A;2::0,LA;=I,g2:0
min h
w E S,>',h
S.t.
h( W - w)= L Aiqi
A; 2:: O,LA; = l,h 2: 0
• pP and 8P play the role of [{p and [{., respectively. In this case, R can be
computed explicitly. P can be partitioned into r (t +1)-edra Pi, with a com-
mon ver tex O. Then val(P) = L val(P;), and val(P;) = l/t!ldet (Pt,··· ,pDI
where Pt, ... ,p~ are the nonzero vertices of Pi. We then have:
'Ne can provide two reciprocal results to proposition 3.14 with the aid of this
dass.
Proposition 3.15 Suppose that [w, w') C C., w' E Sj ns., the W's are CI and
w' - w is not a tangent 10 {w : Wk(W) = W.(w)} al w', Vk E J(w'). Then there is
a P such that aJ = aj.
Under the above conditions, there is r > 0 such that [w, w') C C. Vw E B(w, r) nS
(Proposition 3.11). Choose WI,.'" Wt E B(w, r) such that P +w = [WI"'" Wt, w']
is a (dosed, convex, bounded) polyhedron with 0 E int(P) and ((P + w) \
{w'} ) nSeC•. Thus, gj = 1 and for the rest of ak's gk 2: 1. 0
The next result is simple, taking into account the properties of convex sets.
Thus, this dass has the interesting feature of allowing the characterisation of
sapo alternatives, under certain conditions.
It may be argued that the search should be limited to a certain region. As we
said, we should only use the available information, i.e. wES. It is expected that
through sensitivity analyses, the search might be limited to certain regions in the
next stages. Observe, however, that the solution is P-dependent.
Example 3.9
Suppose that S = {w E R3 : Wi ::: 0, LWi = I}, w = (.25, .5, .25) and there are
three p.o. alternatives al, a2 and a3. a. = a3 for w. S is partitioned into S2, S., SI
as explained below and illustrated in the flgure.
The line separating S, and S. is
(0,0,1)
lllO,f211,l310, L:f; = l.
Thus, for some P's, al is the alternative chosen and for so me others, a2 is the
chosen one.
One way of mitigating against this is by using several polyhedra (PI,"" p.).
Much may be learnt in practice using several gauge functions to find the corre-
sponding a.p.o. alternatives, see French and Rios Insua (1989) and the examples
in chapter 4, specially section 4.2. However, it is interesting to carry efficiency
ideas one stage further, as follows: for each aj find (gfl, ... ,gf') = (gJ, ... ,gJ)
and determine the efficient alternatives of the problem
We have:
Proposition 3.17 Assume th,:: 5 i.s convex and the llIj's are continuous. An al-
ternative aj generated by the aJore procedure is either nondominated or dominated
by an alternative ak such thai g; = gi, i = 1, ... , s.
101
Suppose that aj is dominated by ak. Then, if wj design at es the Wj for the i-th
polyhedron we have IJij(wj) = 1Ji.(wj) ::; IJik(wj). Thus, gj 2: gL Vi, but aj is
efficient for the problem, so gj = gL Vi. o
---
102
at w. Or, if we label the j's as jl, ... ,jm-l so that Wj.(w) ~ Wj,(w) if
k S; I, we can use the steepest direction of W- Wj, , ... , W- Wj, to generate a
polyhedron P and proceed as above. If Wj, (w) = ... = Wj'+k (w) and i +k S; I
we do not have to worry about the possible nondifferentiability of W - Wj,'
When i + k > I, i S; 1 we can use the I - i steepest directions with sm aller
norm from the functions {W -Wj,},=i .... ,i+k.
4. The last approach looks for simple formulations. These are, perhaps, achieved
with Tchebycheff distance (whose polyhedron is P = [{(ai)l=I}], where ai E
{±1}) or city block distance (whose polyhedron is P = [{ ei, !i, i = 1, .. . , t}],
where eij = Oij, !ij = -Oi;' with Oij referring to Kronecker's delta).
We study in more detail the last approach, first seeing that, effectively, the prob-
lems have simpler formulation .
mlll U
w E S,U
S.t.
Wi-U/).i S;Wi
Wi + u/,\; ~ Wi
S.t.
Ui,Vi ~ 0
Wj(W) = W.(w)
Example 3.10
w
y
obtaining a current optimal solution a., how to find the possible competitors of a.
and how to identify sensitivity issues using the information conveyed by sensitivity
measures.
We mentioned that one of the objectives of sensitivity analysis is to detect
errors in modelling. We have not dealt with this problem explicitly. However, we
104
have studied it indirectly since the identified critical judgements may be used as
pointers to possible errors. The eventual robustness of the model may increase
our confidence in it. We briefly address the problem now. As a consequence, we
develop additional sensitivity tools. Section 4.6 explains how to use the previous
ideas to validate a specific model.
Three main topics in error analysis can be identified: sources, detection and
correction of errors. They can be referred to the three elements of our problem:
the functions iflj, the set Sand the estimate w. References to sources of errors are
Kahneman et al. (1982) and those quoted in Farquhar (1984).
The detection and correction of errors will depend on the elements under study.
Given the generality of our approach, problems concerning ifl j seem less impor-
tant (in our finite state, finite consequence spaces context) since we can assign a
parameter to each value (weight, utility, probability, value, ... ) appearing in the
analysis 3 j in later stages, we can reduce the number of parameters, when there is
more confidence in the model.
With respect to S, the problem is essentially one of studying the consistency of
a system of equations for which there is some theory (see Rockafellar (1970) and
Kohlberg (1983)) and algorithms (see Schnabel (1982)), though it is addressed
mainly to the convex case. Pratt and Meyer (1968) and Hazen (1986) analyse
the problem of consistency of preferences and Lindley et a1. (1979) the problem of
consistency of beliefs. Let us mention that mathematical progra.mming libraries
warn when a mathematical program is inconsistent and include devices to deter-
mine, so to speak, the 'nastiest' constraints. It may help to check consistency after
each constraint is added. If inconsistencies are detected then, probably, the only
remedy is to revise the constraints. We shall not study, however, these problems.
Useful references are Sage and White (1984) and White et a1. (1984).
The last problem refers to w. We test whether w rt S by substitution. When
this happens, we can appeal to some sort of projection, e.g. via weighted least
squares, or we can follow a Bayesian approach. Even when wES, we should view
it as a tentative value. We shall discuss these points.
Introducing convenient express ions for <pr(w) and P(aj -< a;), he manages to pre-
serve normality approximately at every stage. He ranks alternatives by E~,(wj).
In classical approaches (1984,1986), he develops a formula for the likelihood <I>(w)
of a statement {all -< al" ... ,al, -< al'+l} and estimates wby the maximum like-
lihood method (MLM). He either ranks the alternatives by 'l1 j (w) or defines the
set
<I> = {w: rp(w)f<P(w) 2: Rcr •• }
for some Her •• , finding the potentially optimal alternatives when w E <I>. He
incorporates these ideas into several interactive methods.
Chaterjee (1986) uses these approaches but his expressions for P(a. -< aj) and
<I>r( w) appear somewhat more natural.
2. The paper by Eliashberg and Hauser (1985) may be seen as classical utility
estimation. They assurne a utility function known up to some parameters u(.,w).
Due to errors, w is modelled as a random variable whose density f is known
up to several parameters'\. ~ is estimated by MLM. Once with f(w, ~), they
compute the distribution of u(x, w), of the expected utility and the prob ability of
choice of an alternative. They do this essentially for a uniparametrie univariate
logarithmic or potential utility function, where w is distributed either normally or
exponen tiall y.
3. Lindley (1986) deals with this problem from a Bayesian viewpoint. Suppose
that a DM has to choose between aj,a2' She does not know their utilities but
has beliefs about them measured in p( Uj , U2 ). If she performs adecision analysis
and VI, V2 are the assessed utilities, her beliefs over the U's are updated via Bayes'
106
theorem
where P(Vl, V2!Ul, U2) is the distribution of the result of the analysis, before it is
done, given that the true utilities are U1 , U2 • The proposed action is that max-
imising E(U;lvl' V2). He also uses the variance of the assessed expected utility to
perform some simple sensitivity analyses. His study is limited to the uniattributed
case, and, essentially, to the normal case.
4. Laskey and Fischer (1987) assurne that preference assessments can be rep·
resented as
where rit, Sill eit are, respectively, the response, systematic part and error term at
time t associated to stimulus i. They use a least squares procedure to fit an additive
value function and report some experiments. A similar decomposition is used in
Hershey et al. (1982), to compare utility assessment methods, and Wallsten and
Budescu (1983), to compare prob ability assessment methods. Dyer and Ravinder
(1986), Ravinder et al. (1988) and Yoon (1989) use it to analyse or estimate the
variance of the assessed additive values, probabilities through decomposition and
multiplicative values, respectively. This is the typical error model and can be
seen in the framework of location parameter estimation models (see Diaconis and
Freedman, 1986) and hierarchicallinear models (see French, 1978). We shall be
interested in models where the systematic part is constrained; some of the few
references to this case are Casella and Strawderman (1981) and Berger (1980,
1985).
Given that the estimate W of the parameter w is subject to error, our purpose is to
find a distribution on w, after having measured w. Let us decompose w = w + v,
where w is the parameter and v the measurement error. Prior to the elicitation,
the DM has beliefs ab out wES, measured by a density 7ro(w), and beliefs about
v. Possibly, the distribution of v depends on w; we call its density fl(vlw). Then,
the joint density of v, w is
f(v,w) = 7ro(w)fl(vlw).
Example 3.11
7 -4w
5+2w
2+7w
and w E [1/6,5/6]. All actions are p.o. nondorninated. Suppose that w has a
uniform distribution over S, i.e. w ~ U(S).
1/6 5/6
Wh at do we do now?
Or, equivalently,
• Ho: wES.
• Ho: aJ ~ a.
We should elicit the loss function to perform the test. This would imply that
the DM would need to provide further information. However, we shall use the
test in a purely exploratory way: since the result of the analysis does not depend
109
directly on the test, the choice of the loss function does not seem so important.
The approach we propose is to carry out several tests. Then, we face a problem
similar to that of multiplicity in classical hypothesis testing, see Diaconis (1985):
if enough tests are computed, some of them will lead to accept Ho and others to
accept H j • Diaconis suggests several solutions but only that of replication seems
to apply in our case: rejecting Ho in one of the tests suggests that the analysis
is not deep enough and that the DM must think more profoundly; if there is no
rejection, we can use them as a support to the current status, in the event that
the DM does not want to provide further information.
Let us see the expressions of some tests:
or, equivalently,
'iJJ(W) - 'iJ.(w)
'iJ.(w) - 'iJJ(w).
Then, we reject Ho if
or
0 if 'iJ.(w) 2: wJ(w)
{
= wJ(w) - 'iJ.(w) otherwise
Similarly for L(w,aJ). We reject Ho if
This is the same result as before: the hierarchical criteria with the identity
as the second order utility.
110
Given the above eomments, we shall earry out the tests with 0-1 and regret
losses, due to their easily interpretable expression.
Example 3.12
Thus, with 0-1 loss we reject the current optimality of a3' With regret loss:
E(1lt 2) = 1 1/6
5/6
(5 + 2w)1ro(wlw)dw <X 4.467
Again, we reject the hypothesis that a3 is optimal. The sensitivity index r is 0.071.
All this demands furt her analysis.
As we eonceive of the decision making process cyclically, we may modify the inputs
w n, Sn to the analysis. We need to extend the previous approach to several cycles.
We assume first that S remains unchanged, and then give a solution when S
ehanges.
At the seeond eyde, the prior distribution over w is
We measure w' = w + v', where v' is the error. The distribution over v' depends
possibly on v and wand we represent it with a density h{v'lv,w). Then we find
that
1r1(wlw') <X 1r1{w)h{w' - wlw - w,w).
111
Summing up, if 7I"i_l represents the prior, Si-l the support, Wi the measurement
and fi(.lw,vl'" . ,Vi-tl the error density on the i-th cycle, i = 1, ... ,n + 1, we
have
and
It may be reasonable to assurne a sort of Markovian property for the errors such
as
and the pertinent changes are made in the formulas. Even the vis may be judged
independent given w, so that
We shall make this hypothesis from now on. Sometimes we can assurne that the
Vi 's are independent of w
The classic error model assurnes that the vis are i.i.d. and S R. Then, the
formula is
See Diaconis and Freedman (1986) for more details about it.
112
Example 3.13
We determine now that w E [.5, .75]. Only a2 and a3 remain in the analysis. a3
is the current optimal solution and a2 is a.p.o. to a3. w = 5/8 is measured with
error v ~ N(O, 1/8). We get that:
Then:
We do not reject the current optimality of a3. The sensitivity index is 0.2. This
might all help to reassure the DM that the result of the analysis was not spurious.
The approximate graphs showing the changes in the distribution over ware
The W distribution
We only know that wES. It would be very strong to assurne that w is normally
distributed, as in Lindley (1986).
We justify the adoption of a uniform distribution over S. Apart from in-
variance arguments (see Villegas, 1977), the typical approach is that one chooses
the most intuitively reasonable parameterisation for the problem and a lack of
prior knowledge should correspond to a constant density in this parameterisation
(Berger, 1985). The counterargument (see O'Hagan, 1988) appeals to the use of
transformations invalidating the constancy of the density over S.
In our case, wES and we know nothing else so, assuming vo/(S) > 0, we put
as a density
7ro(w) = l/vol(S)Is(w).
Observe that our parameterisation is not only reasonable but also deduced from the
axioms introduced. Let us rehearse the transformation counterargument. When
the error distribution is independent of w, w is a location parameter of the dis-
tribution of w (see definitions in Berger, 1985). Then, the dass of admissible
tr ansformations is
T = {sc: sc( w) = w + C, Yw}.
This is so far as probability arguments are concerned. Our parameters also have
a meaning from the point of view of evaluation; to be valid, Sc has to verify
Consider a linear problem, with Wi(W) = clw. Suppose there are nonempty sets
E.,E,Ej eS such that
(c. - Cj )t w > 0, w E E.
(c. - c,)t w = 0, w E E
are valid. In this way, we can generate a full rank system {(Cj-c.)t c = OL, possibly
introducing hypothetical alternatives with associated evaluation Wj(w) = c}w.
(This is valid since the DM's judgements apply to hypothetical alternatives). The
only solution of the system is c = 0: the only valid transformation is the idcntity.
114
A similar argument applies when the iVj's are bilinear. In the general ease, we ean
apply it to the first order Taylor approximations.
When the error distribution is not independent of w, or more gene rally when
w is not a loeation parameter of w, we eannot limit the study to T. Then, we
ean appeal to the uniqueness properties of the probabilities, utilities and values.
For example, if w represents probabilities, w is unique so we eannot talk about
transformations of w apart from the identity; if the w's refer to utilities they are
affine unique but then the transformation eounterargument is not valid, as the
density of the new parameterisation is uniform as well. Similarly if the w's refer
to weights or additive values. The only ease whieh would not be eovered is that
of general values, whieh are unique up to a monotonie transformation. Then, we
have to appeal to the initial argument.
Onee we have 'justified' a uniform prior over S = So, the eorresponding formu-
las are
We eonsider now the error distribution. The arguments given here are mueh
weaker. For each model, we eompute the posterior distribution and give the eor-
responding tests.
A normal model Continuing with the physieal tradition, we shall ass urne
normal errors. We might think that due to a big number of sourees of errors (see
e.g. Kahneman et al. (1982)), the error distribution may be assumed normal. This
argument should be taken with eare.
We shall use the following standard results (see De Groot (1970) and Johnson
and Kotz (1972)):
Lemma 3.4
117:1 exp( -.5( w - ßi)2 / Cii) rx exp( -.5p( w - L)ßd Cii)/ p)2)
with p = L l/ Ci i '
Definition 3.11 A random variable X has a normal distribution with parameters
(J1, (7) truncated at [a, b] if its density is
~ 1~O(("
ifx:S: a
F(,) - p)/.) - O(hll/(O(') - .(")) if a:S: x :s: b
ifx> b
Its mean is
Pt = P + d<j>(h) - <j>(k))/(<I>(k) - <I>(h))
where <j> is the density of the standard normal distribution.
Its variance is
More generally,
• When the /1,'S are known precisely, we can make 'P,( w) = /1, and p,->O. Then
with
with
ir (wlw)
n n
I ()II~
oe Sn W .=1 exp
(_ 5(
. w
_ Wj - 'P,(z,) + 'P:(Z,)Zi) 2j
1 + 'P:(Z;) (1
a;+'P:(z,))2
+ p; )
and applying lemma 3.4
with
117
mo=O .
• When the linear approximations are not valid, we can expand the squares
and apply lemma 3.4 to get
We shall study the cases leading to a truncated normal distribution. The rest
have to be tackled with numerical integration. Typically, in the unidimensional
case, Sn = [en,!n] and ~j(w) = djw + Ij. Suppose that ~.(wn) :::: ~J(Wn) and
~.(Jn):::: ~J(Jn)' Let w = (I. -IJ)/(dJ - d.) so that ~.(w) = ~J(w).
We have that
Then:
or
These distributions are unimodal and symmetrie about 00 , the most likely error.
Berger and Berliner (1986) use them as an E-eontamination dass of the prior. They
point out that this is the dass of mixtures of distributions in
with
• eo == max(eo, (w - 00 ) - a)
• fö == min(fo, (w - 00 ) + a)
We assurne that [e o, föl 1= 0. Suppose first that the loss is 0-1. Let wbe as before.
Then:
119
or
We have:
2(w - 80 ) 1------...
~fo+eo
Let us study the test with the regret 10ss. The quantities involved are:
/;
E('l1i) =
J{••o (diw + li)dw = u; - e~)(di(e~ + !;l/2 + li)'
or, operating, if
Then, we could compute the exact tests. Observe that tests computed with uniform
errors are conservative as indicated in the following result, which says that the
acceptance region is wider for the exact test.
fO W W
~ ( ('l1.-'l1J)dlrg~O
}[2w-.ö.J;l
o
For more than one cyde, the approach is similar. If the error on the i-th cyde
follows a distribution U[O; - Ci, 0; + Ci], the posterior distribution is
with
and e:' j = -00, I:j = 00. The tests have similar expressions. The weighting
functions have now the form I;(c;jw).
Let us give first the expressions of the tests when the pertinent random variable
is ~. Let J.L = E(w) and V = var(w).
When we use the regret loss:
• The general case. Try approximate Wj( w) by its second order Taylor ex-
pansion around J.L
Lemma 3.6
with
The results for the unidimensional case are translated to this case with slightly
more complicated formulas. For example, if we assume that Vi ~ N(Pi' Ei), then
with Zn and Tn as in lemma 3.6, with (w n - f.ln) in place of f.ln. Similar results
hold in the rest of the cases.
The main problem here is the integration. We shall study the ease in which
with
h. _ tP((ei - (.l ;)/(7i) - tP((J; - (.l ;)/(7;)
, - <I>((f; - (.l ;)/(7;) - tP((ei - (.l ;)/(7i)
and the matrix V of variances is diagonal with diagonal elements given as in
definition 3.11.
Then, if we use the regret loss:
When the DM expresses his beliefs over the range of the error distribution, say via
a density fra), the error distribution is:
J-t-1r-a,aj(v)f(a)da
1 == 1 ... 1 fra) dal ... dat
00 00
~II
2 IIa, 2 Ivd lVII ai
We can give explicit formulas for the tests in some special cases. Suppose that
S == IIl=l[ei,J;]. Then
with
• f: == min(fi, w + a;)
Thus, if we use the regret [oss:
When we use the 0-1 loss, we can give general formulas only in the linear case.
We know that
P(c'w :::; c'ö) = .5
Therefore, we reject the optimality of a. if
In the other cases, we have to apply numerical integration. These results are easily
extended to more than one stage.
3.6 Comments
We have argued that sensitivity analysis tools for decision aids are essential as
a means of exploring the DM's mind to identify the most critical judgements
in determining choice. We find first whether the current optimal alternative has
competitors. If this is the case, sensitivity analyses are necessary. Several concepts
of adjacent potential optimality help us to reduce the set of competitors to the
current optimal alternative. A dass of easily implementable sensitivity analysis
tools, based on gauge functions, helps us to identify sensitivity issues in a very
informative way. The problem of error modelling in judgemental measurement
leads to so me other sensitivity tools based on Bayesian hypothesis testing. Chapter
4 illustrates these ideas in various examples.
Chapter 4
Computational experience
• SENSLIN, for linear problems, that is 'itj(w) = CjW and 5 is defined by linear
constraints.
• SENSGEN, for general problems, that is the 'itj's and 5 are general.
1. Read data.
7. Find p.
8. Find 8.
9. Present results (distances, solutions, graphs and ways of improving the cur-
rent status).
They are designed to use NAG library as efficiently as possible. Thus, problems
referred to finding nondominated, p.o., and a.p.o. alternatives are solved with the
optimisation option. The only exception is the computation of a.p.o. alternatives
in the linear case, which is done with the consistency of a system of inequalities
option: the system of linear inequalities is set up faster than the linear program.
We use the forward algorithm of section 2.5, with gi = 'iti(w), to find the non-
dominated alternatives. Steps 3, 4 and 5 filter the alternatives of interest to the
DM. Any alternative discarded in any of these steps is eliminated from the rest of
the analysis. For instance, if aj is dominated we do not look for the alternatives
dominated by aj; we do not check whether aj is p.o. or a.p.o.; we drop, in fact, the
constraints relative to aj in a.p.o. and p.o. computations, by putting their bounds
as -00,00; and we do not compute dj .
As optimisation problems are solved with NAG, distance problems have to be
formulated as smooth:
129
wES
s.t.
W.(W) - Wj(w) = O.
[Pd j ] min dj
w E 5,dj
s. t.
W.(W) -Wj(w) = 0
Wj + dj ~ Wj.
maxWj =!l
wES
and
minwi = fl
wES
giving
.
fJ = ~x(fl- Wj,Wj - In .
130
w E S,.,t
s. t.
Si, ti 2: o.
S. t.
Siti = 0
Si,ti 2: o.
The following results are presented on the screen and, with some more detail, are
also written on a file (see the appendix for a sampie):
• list of a.p.o. alternatives and parameters for which they share optimality
with a.,
• for each aj remaining in the analysis, its dj , parameters for which aj outranks
a., and graphs,
• p, 8, parameters giving the maximum distance and their graph, and the
sensitivity index,
- f~
Negative and positive changes are easily appreciated. The use of colours might
enhance this feature. In our opinion, these graphs provide a faster view of the
changes in the parameters. This is due, perhaps, to a distorting effect of the axis
in bar charts, but this has to be evaluated from a psychological point of view. Both
types of graphs are plotted with the aid of GROST-80. SENSATO also provides
tables indicating patterns of the changes in the parameters leading to outranking
a. by aJ and the greatest changes in S. Graphs and tables are illustrated in section
4.2.3.
132
The following table summarises the NAG routines used in the different parts
of SENSATO routines:
The input to the routines is described in sections 4.2, 4.3 and 4.4.
a1 a 2 a3 a4 a5 a6 a7 a8 a9 a 10
al 6 6 8 8 8 4.3 3.3 1 4 6
a2 5 5 7 7 7 5.9 4 1 4 7
a3 8 2 3 2 4 7.4 5.6 5.3 2 4
a4 7 8 6 6 6 2.7 1 1 6 1
a5 4 4 4 5 5 6.6 4.7 1 3 5
as 6 7 5 4 2 8 2.8 8 7 2
a7 4 3 3 3 4 7.7 2.7 8 1 8
as 1 3 1 1 1 1 8 6 8 2
We shall call this matrix C. The function proposed to rank the alternatives is
10
v{a) = Ea'w"
i=l
133
From the information contained in the text, the only constraints are Wi ~ 0 and
E~~l wi=55 (this is the addition of the initial parameters). To add some fiavour to
the problem, we consider Bome constraints on the w's (which are typica.l of what
might arise in practice). These are:
Par LB UB
Wl 5 20
Wl 4 10
W3 0 15
W4 5 10
Ws 0 10
W6 5 10
W1 5 15
Ws 3 5
W9 0 3
W10 0 5
o :::; W6 + W1 + Ws - W9 - W10
10
L;Wi = 55
1;::1
Dw::;d.
W = (10,7,3,5,6,8,9,4,2,1).
• Number of alternatives 8.
• Matrices C, D.
• w.
The rest of the matrices, dimensions, bounds and constraints are defined/created
by the routine.
4.2.3 Results
We comment in some detail on the results. The rating and ranking of the alter-
natives is:
135
a\ 296.1
a6 293.2
a2 285.2
a3 275.8
a4 257.6
as 245.1
a7 236.9
aB 167.0
For example, the optimal value of [P2d is -14.5 so that al does not dominate
a2; the value of [Psd is 3 and al dominates as. A dash means that the problem
need not be solved. As we see, al dominates as and aB. The set of nondominated
alternati ves is
The p.O. nondominated alternatives are found solving problems [Pi]' Their solu-
tions are shown in the table, where the second column presents the optimal value
of the corresponding problem:
a6 -59.07
a2 -12.41
a3 -46.90
a4 -3.16
a7 1.11
For example, a6 is p.o. and for some wES its value is 59.07 higher than the
best of the rest of the alternatives. a7 is not p.O. as the optimal value is greater
than O. Thus, the p.O. nondominated solutions are
136
That should be the solution of the decision making problem under the partial
information wES. More analysis is required as al is not strongly optimal.
A.p.o. alternatives are then computed studying the consistency of [SJ] with the
results
aS Consistent
a2 Consistent
a3 Consistent
a4 Consistent
These alternatives enter into the distance analysis phase. Results are reported
for lp distances, with p=1,2,3,4,00 (rounded to two decimals). The table gives
the d/s, the maximum distance p and the sensitivity index r for each of the lp
distances:
ds d2 d3 d4 p r
1 0.44 9.70 4.71 20.20 32.02 0.01
2 0.25 3.87 2.04 8.35 14.90 0.02
3 0.20 2.83 1.45 6.57 12.70 0.02
4 0.20 2.41 1.23 6.00 12.18 0.02
00 0.10 1.49 0.76 5.17 12.00 0.01
In all cases, as gives the minimum distance. Observe that r is very small; this
suggests that we have to be very careful in the elicitation of weights. Very small
perturbations of the weights lead as to outrank al. For example, for
w = (10.00,7.10,2.89,5.00,5.89,8.10,8.89,4.10,2.10,0.89),
The trends tables give the sort of changes in the parameters producing the
outranking of a. by aJ and giving the maximum distance. A 0 indicates no change,
a -1 a reduction and a 1 an increase in the parameter value (with respect to wo).
p J Wj Wz W3 W4 Ws W6 W7 Ws Wg WjO
1 6 -1 -1 -1 -1 -1 1 -1 1 1 -1
2 6 -1 1 -1 0 -1 1 -1 1 1 -1
3 6 -1 1 -1 0 -1 1 -1 1 1 -1
4 6 -1 1 1 0 -1 1 -1 1 1 -1
00 6 -1 1 -1 0 -1 1 -1 1 1 -1
We see that J = 6 for all the distances used. The pattern is that changes occur
when Wz, W6, Ws, W9 are increased and Wj, W3, Ws, W7, WjQ are decreased. There need
be little or no change in W4' This information would clarify in the DM's mind the
least changes in her judgements that would lead to a change in the suggested
course of action.
p Wj Wz W3 W4 Ws W6 W7 Ws Wg WjQ
1 -1 -1 1 0 1 -1 -1 -1 1 1
2 -1 -1 1 0 1 -1 -1 -1 1 -1
3 -1 -1 1 1 1 -1 -1 -1 -1 -1
4 -1 -1 1 1 1 -1 -1 1 -1 -1
00 -1 1 1 0 -1 -1 -1 1 -1 1
This table gives the patterns leading to the greatest changes possible in S.
There should be a reduction in Wj, Wz, W6, W7 and an increase in W3, W4, Ws. There
is no pattern for Ws, Wg, WjO' This information would clarify in the DM's mind
the greatest possible changes in her judgements, given the current information
expressed in S.
The DM would probably ask for some more quantitative information. Graphs
seem to be the next step. We give below bar charts and star graphs of the results
when the Euclidean distance is used.
138
,1.000
....
-1.490
-1"
1.000 '.000
0.154
1.0OCQ..5f1
0.000
-0.49(1
~ O.~29
-0.289
·0.269 '0
.0. 002
.0.1198
-O,2H-O.2l19
-D.6"..o,'J'02 -0.6"
139
,
1..19 LOOG Lm "14,' LOQ
, . o.l,.,o.07~.900
... ...
0.0lI01 0.919
1.'\0
UD'
.
I
1.107
.1.101
.. I. Jot
_1.101 _I, XII _1.301 .I.JOI
_1.111
.1.,.,
4.000
.1.000 .1,000
·,m
.,,,,,
.1,000
.'5.~'
140
In this ca.se, minute changes lead to as outranking a\. Compare especia.lly the
star graph associated with as with the others. Changes in favour of a4 seem, on
the contrary, very big. Fina.lly, we observe the possibility of very big variations
in Wb W2, W3, Ws, Ws and W7 within 5. Similar thoughts arise when we look at the
graphs relative to other distances.
The DM may ask for more quantitative information. We can use the numerical
output of SENSLIN; see appendix. The two points of concern are how big are the
least changes leading to as and a4 outranking a\.
Looking at the list of critica.l parameters, we see that deviations from the
origina.l ones do not need to be greater than 0.2 in any of the parameters to cause
the outranking of al by as. In fact, except for Tchebycheff distance only changes
in Ws, Ws and WIO have to be greater than 0.1 to cause that outranking. Changes
in Wl and W4 are very sma.ll. By now, the DM should be sufficiently motivated to
think more carefully about the problem.
More sensitivity issues arise. The biggest difference in rating between a4 and
a\ is only of 13.3 which is not too much. Moreover, the biggest difference in rating
between a4 and the rest of the alternatives is 3.16. And the d4 's seem very big.
Other features which appear in the ana.lysis of the data are the very big changes
possible in parameters Wl, W2, W3, Ws, Ws, W7 confirming the impression acquired by
looking at the graphs.
Sometimes it is not possible to identify sensitivity issues so easily: some help
can be found using tools from multivariate exploratory data ana.lysis. We can
proceed as follows 2 •
If t is the number of parameters, the procedures we have introduced identify
points W of R t (parameters representing least changes for the outranking of a. or
greatest changes in 5). We ca.n reduce the dimension using principal components
analysis. The first two principal components may be regarded as the plane of
best fit (in the least squares sense) to the initial observations: we transform, by
projection into aplane, the points Wj - w into a set of points Cj whose relative
positions and Euclidean distances are approximations of those of the Wj - w.
This idea is ca.rried out with the results from lp-distances (p = 1, ... ,9,00).
We represent the scores of the data relative to Tchebycheff distance in the two
first principal components after an orthogona.l rotation.
lThis part is not incorporated in our library. Computations were done with SPSS PC+.
141
I.
-I.
-I I !
Figure 4.2: Scores of parameters leading to outranking of aj by aj, i = 2,3,4,6
and greatest changes in S(l) for Tchebycheff distance, in the first two principal
components.
This represents the elements of the Tchebycheff distance problem. We can also
display the minimisers for all the distances for the different alternatives and the
maximisers. The following graph was built with the aid of a cluster analysis and
represents these data. Note that the scales are quite distorted.
We get a picture of so me of the issues we have identified: as is the most likely
alternative to be the competitor of aj; very small changes in the parameters lead as
to outrank aj; a4 does not seem to be a competitor of a\; and very big reductions
in S seem possible.
142
1.1
3
1.05
.. 6
.11
-.,11
-.n
-.141
-.n
-.17\
-.14
_.,,\
~~
'"
01
1
.11
.UlS
.'"
.,1
..
.,~
."
.11
•• 01
I.H
,
-1.2
,, I 1
".
1.1
I
•I•
-.,~ -.. ~ -.ll -.' 1
-.14 -.n -.' -.n
1
-, 1
143
More work needs to be done to explore further the possibility of using these
tools. Observe also that we have used them in a conn.rmatory sense. They should
be more useful in an exploratory phase.
The approach seems quite powerful as it allows to identify sensitivity issues in
a very constructive way. It should help DM's to think more carefully about their
problems.
BI B2 B3 B4
al 800 900 1000 1100
a2 600 1000 1000 1000
a3 400 800 1000 1200
a4 600 800 800 1200
as 400 600 1100 1200
a6 800 800 1100 1100
Let us call:
Par LB Wj UB
W1 .20 .30 .50
W2 .20 .30 .40
W3 .10 .20 .40
W4 .10 .20 .60
Ws 1.0 1.0 1.0
Ws .85 .90 .95
Wr .80 .85 .90
Ws .75 .80 .85
Wg .60 .70 .75
WJO .30 .50 .60
Wu .00 .00 .00
i=l
(for example, the first one means that the broker believes that a very low to low
market is at least as likely as a medium to high market). To ensure monotonicity
in the utilities, we add the constraints
W10 :5 Wg
Ws :5 W7
wr:5 ws·
with
(:o : : : : : :1
000 0 0
For the rest of the alternatives
with the Bj's defined in a similar fashion. The constraints can be written
Dw::;d.
W = (.3, .3, .2, .2, .1, .9, .85, .8, .7, .5, 0).
• Dimension of w l 4.
• Dimension of w 2 7.
• Number of alternatives 6.
• Matrices D, BJ •
• w.
Thc rest of the matrices, dimensions, bounds and constraints are defined/ created
by the routine.
146
4.3.3 Results
We comment on the results in less detail. The rating and ranking of the alternatives
IS:
al .800
a6 .780
a2 .745
a4 .700
a3 .580
a5 .530
al is the current optimal alternative. There are four non domina ted alternatives
a6 -.060
a2 -.060
a4 -.035
A.p.o. alternatives are then computed solving [PJJ. The a.p.o. alternatives to
These alternatives enter in the distance analysis phase. Results are reported
for lp distances, with p=1,2,3,4,00:
d6 d2 d4 p r
1 0.06 0.15 0.47 0.92 0.07
2 0.04 0.07 0.18 0.35 0.11
3 0.03 0.06 0.14 0.27 0.12
4 0.02 0.05 0.12 0.24 0.11
00 0.01 0.03 0.08 0.20 0.08
147
w = (0.3,0.29,0.2,0.2,1.,0.91,0.83,0.78, 0.72,0.5,0.0),
1l1 1 (w) = 1l1 6 (w). In this case, the maximum loss due to choosing al instead of ae
is 0.06. This is not very large compared with the values of the alternatives, which
are around 0.8.
ottO.(,Mt-'O"'-J
O.:UO.IO"'O-"·"J
._,'0 , 000
00"
,.....-- .. ".
Figure 4.4: Decision tree for the airliner problem (French, 1986).
Alternatives are: al(N, N, N), a2 (N, Y, Y), a3 (Y, Y, Y), a4 (Y, Y, N),
as (Y, N, Y), aa (Y, N, N), where the first Y-N means (asking or not for the
survey), the second means (buying or not, if the report is favourable) and the third
148
means (buying or not, if the report is unfavourable). Some of the alternatives are
obviously bad: SENSATO will discard them. The directors are asked to provide
their probabilities for the result of the airliner and for the result of the consultant 's
report conditional to the result of the airliner. The utility of the airline depends
on the annual operating profit x!, the annual number of passengers carried X2, the
proportion of flights delayed X3 and the cost of the survey X4' It is established
that
U(Xj, X2, X3, X4) = g(Xj, X2, X3) - kX4.
Let us call
Wj p(vr)
W2 p(mr)
W3 p(ur)
W4 pUlvr)
Ws pUlmr)
W6 pUlur)
W7 p(ulvr)
Ws p(ulmr)
W9 p(ulur)
WjO g(vr)
Wu g(mr)
Wj2 g(nb)
Wj3 g(ur)
Wj4 k
WjS X4
where p(E) is the probability measuring the directors' beliefs in event E and g(B)
is the component utility associated to results (Xj, X2, X3) when B is the decision.
The directors provide the following constraints:
Wj + W2 + W3
Ws + Ws
149
They also provide the following assessments and bounds for the parameters:
Par LB Wi UB
Wl .10 .20 .40
W2 .20 .30 .60
W3 .30 .50 .70
W4 .85 .90 .95
W5 .40 .60 .70
Ws .05 .10 .15
W7 .05 .10 .15
WB .35 .40 .45
W9 .85 .90 .95
WIQ .90 1.0 1.1
Wll .25 .34 .50
W,2 .10 .17 .30
Wl3 .00 .01 .05
WH .05 .10 .15
Wl5 .05 .10 .15
them
W1(W) WI2
W2(W)
W3(W)
W4(W) WIW4WIO + W2WSWll + W3 W S W I3 +
(WIW1 + W2WS + W3W9)W12 - W14 W IS
WIW7WIO + W2WSWll + W3 W 9 W I3 +
Dw S. d
W = (.2, .3, .5, .9, .6, .1, .1,.4, .9, 1., .34, .17, .01, .1, .1).
• Number of alternatives 6.
• w.
The rest of the matrices, dimensions, bounds and constraints are defined/created
by the routine.
151
4.4.3 Results
The rating and ranking of the alternatives is:
a4 .332
a2 .307
a3 .292
al .170
as .160
as .125
P d2 P r
W = (.2, .3, .5, .9, .6, .1, .1,.4, .9, 1., .34, .12763, .01, .1, .1)
(that is, only a change in W12 of 0.0424) W2(W) = W4(W). The loss of optimality
due to choosing a4 instead of a2 may be as bad as 0.1192, which is a lot, since
values are around .3.
(Wl, W2, .•• , WIO) will have the same meaning and will satisfy the same con-
straints of section 4.2. We can fix two of the values of the first and fourth scales,
say at 1 and 8. The remaining four values in the first scale are represented as
(Wll, W12, W13, Wl4); the remaining six values in the fourth scale are represented as
(W15, W16, W17, W18, W19, W20). They are assessed and bounded as folIows:
Par LB Wi VB
Wll 1 4 8
W12 1 5 8
W13 1 6 8
W14 1 7 8
W15 1 2 8
W16 1 3 8
W17 1 4 8
W18 1 5 8
W19 1 6 8
W20 1 7 8
Dw:S d.
There are no nonlinear constraints in this case. We have a general model. The
estimates of the parameters are
W = (10,7,3,5,6,8,9,4,2,1,4,5,6,7,2,3,4,5,6,7).
153
4.5.2 Results
The rating and ranking of the alternatives is as in section 4.2.3. There are now
seven nondominated alternatives:
a2 -27.5
a3 -156.9
a4 -96.8
as -13.5
as -88.5
a7 -23.7
Thus, all nondominated solutions are p.o. They are all a.p.o. to al and enter
into the distance analysis. As an illustration, with 14 distance, the results are
ds d2 d3 d4 ds d7 p r
Analysis 4.5 0.15 0.52 0.81 1.47 2.69 2.16 12.77 0.01
Analysis 4.2 0.20 2.41 1.23 6.00 - - 12.18 0.01
Their performance is measured with fourteen attributes. Weights are assessed and
alternatives are ranked by
Their approach can only be called pseudo-Bayesian, due to the way of assessing
the Wj'S and the Vj's. To do this, they use the scale:
Level Impact
9 V.High+
8 High+
7 Medium+
6 Low+
5 Negligible
4 V.Low-
3 Low-
2 Medium-
1 High-
0 V.High-
The w;'s and the Vj'S are assessed independently on each scale without consid-
ering compensation. We think we cannot appeal to uniqueness properties, apart
horn monotonicity.
Component values are assessed as folIows:
al 1 1 2 2 1 9 3 2 3 5 0 0 2 2
a2 2 3 3 3 1 9 3 2 3 5 0 4 3 2
a3 1 2 2 2 1 7 5 2 2 3 2 2 2 1
a4 1 1 2 2 1 8 5 2 2 2 2 2 2 0
a5 2 2 1 1 3 8 5 3 0 1 1 0 1 1
a6 3 2 0 1 3 8 5 3 1 1 1 0 1 2
a7 2 2 3 3 3 8 5 3 1 0 1 0 3 2
(8,8,3,9,9,9,5,2,7,8,7,9,3,8)
4.6.1 Analysis 1
Following our comments, we assign a parameter to each value on the scale and
use only monotonicity properties. We can fix Wo = 0; also, W6 does not appear in
155
the study so we do not use it. Finally, to bound the space of parameters we make
Wg = 9. Recoding W7, WB as Ws, W7, our estimates and bounds are:
Par LB w. UB
Wl 0 1 9
W2 0 2 9
W3 0 3 9
W4 0 4 9
Ws 0 5 9
W6 0 7 9
W7 0 8 9
Given our feeling that the model is inadequate in general, we study whether
it is inadequate in this particular example. The question to answer here is: Are
there scales conveying the same information which result in a different ranking?
WSW7 + 9W4 + 81
Dw::;d.
There are no non linear constraints in this case. We have a general model. The
estimate of the scale values are
w = (1,2,3,4,5,7,8).
156
4.6.3 Results
The rating and ranking of the alternatives is:
a2 307
ar 237
a3 233
al 232
a6 220
a4 218
as 200
a2 is the current optimal alternative. It was also the implemented one 3 . There
are two nondominated alternatives
For some scales under the stated conditions, a3 is better than a2. Strictly speak-
ing, this is enough information to answer the questions posed. However, it is
informative to continue the analysis.
a3 is potentially optimal with advantage 81 and is a.p.o. to a2. The distance
analysis is:
P d3 p r
1 4.00 33.00 0.12
2 2.20 13.96 0.16
3 1.78 10.83 0.16
4 1.59 9.65 0.16
CXJ 1.23 8.00 0.15
Again, the problem seems to be sensitive. This means that there are scales
relatively elose to the original one for which w3 (w) == w2 (w). The loss due to
choosing a2 instead of a3 may be as bad as 81 (for values around 300).
The model may not be adequate in this case.
4.6.4 Analysis 2
Suppose now that the analysis has been conducted properly: the Wj'S and the v;'s
have been assessed using adequate methods.
3Its implementation did not depend on this selection process. The example is retrospective.
157
As in analysis 1, seven parameters are assigned to the scale values: we fix those
corresponding to Wo = 0 and W9 = 9; as Ws does not appear, we recode W7, WB as
Ws, W7. There are also six different weights, but we can fix one of them, say to 9.
So we have five more parameters Ws to W\2' Estimates and bounds are assessed as
follows:
Par LB W· . UB
W\ 0 1 9
W2 0 2 9
W3 0 3 9
W4 0 4 9
Ws 0 5 9
Ws 0 7 9
W7 0 8 9
Ws 1 2 9
W9 1 3 9
W\O 1 5 9
W11 1 7 9
W\2 1 8 9
w = (1,2,3,4,5,7,8,2,3,5,7,8).
4.6.6 Results(Analysis 2)
The rating and ranking are as in analysis 1. Again, there are two nondominated
alternatives {a2,a3}. a3 is p.o. with advantage 81 and a.p.o. to a2.
P d3 p r
The problem is slightly more sensitive with this analysis. Numerieal eomments
are similar to those of analysis 1. However, the eonclusion in this ease would be
that we have to be very careful in the elicitation of the wis and the vis.
Consider adecision making problem under uncertaintyS with three states and
seven alternatives. Let the payoff matrix be
01 O2 03
al 5 5 5
a2 10 -3 6
a3 -2 11 4
a4 1 -1 12
as -6 12 0
a6 3.5 3 6.75
a7 -2 -2 12
We shall call this matrix C and eta) the payoffs associated to a. Alternatives
are ranked by their expected payoff
3
E(e(a),p) = LP(Oi)Ci(a).
i=l
Wi ;::: 0, i = 1,2,3.
4.7.2 Results
Suppose that
w = (0.23,0.24,0.53).
a4 6.3500
a7 5.4200
aB 5.1025
aj 5.0000
a2 4.7600
a3 4.3000
as 1.5000
These alternatives enter into the distance analysis. For 11 ,1 00 , a3 is the alter-
dj d2 d3 P T
4.8 Comments
In view of the previous description, we believe that this approach is computation-
ally feasible. The following table provides the details and CPU times in microsec-
onds of some of the problems solved.
The first block rows describe the problems: type of evaluation functions and
number of alternatives, parameters, linear and nonlinear constraints. The last
block gives the total time spent in the analysis. The intermediate blocks give
times for the different steps of the analysis: rating, ranking of alternatives and
reordering of data; computation of nondominated alternatives; computation of
p.o. alternatives; computation of a.p.o. alternatives; distance analysis with city-
block, Euclidean, 13 , 14 and Tchebycheff distances. The first row indicates the total
time spent in that step; the second, the number of problems solved6 ; the third,
the average time. For example, in Dallas' problem (section 4.2), we spent 73485
J1.sec in computing the p.o. alternatives; we solved 6 (optimisation) problems; on
average, we spent 73485/6=12247 J1.sec on each of them. Response seems to be in
real time. However, remember that our problems are not very big.
Conclusions
This chapter provides a summary of the ideas introduced and suggests some areas
of potential development.
5.1 Summary
The case was made for adopting a Bayesian approach to decision making. How-
ever, problems arising in practice led us to study decision making under partial
information and sensitivity analysis in decision making.
Decision making problems under partial information are characterised by the
data {{Wj(w)}j,w ES}, where the Wj's are evaluation functions and S are con-
straints on the parameters w modelling judgements. The three questions to ans wer
are:
As far as the first quest ion is concerned, if the Bayesian approach requires a weak
order regulating judgements (plus some conditions), the DPI model requires a
quasi order regulating judgements (plus some similar conditions). This applies
to values, utilities and probabilities. Consequently, alternatives are ordered in a
Pareto sense:
Then, the solutions are the nondominated ones. Potentially optimal alternatives
also seem acceptable, both from an intuitive point of view and because of their
relation with the nondominated ones. Algorithms are provided to compute those
165
solutions in general problems and in some particular cases (linear and bilinear).
Other solution concepts are criticised on the grounds of being ad hoc. In summary,
we have a more robust theory of decision making based on a weaker set ofaxioms,
but embodying coherence, since it essentially implies carrying out a family of
coherent decision analyses.
When we have the data { { {'l1 j( w)} j, wES}, w}, we can rank the alternatives
by 'l1 j (w). However, we have to be ready to criticise this ranking. We argue
that sensitivity analysis is the means to help the DM explore her thoughts in
a constructive way to find which of her judgements are the most influential in
determining the decision. This inc1udes the criticism of the Flat Maxima Principle.
Several concepts of adjacent potential optimality help us to reduce the competitors
of the current optimal alternative. A dass of easily implementable sensitivity tools
based on gauge functions is then introduced which allows us to:
Für finite problems, the main point remaining is the possibility of incorpürating
bounds for parameters in additive models. Specifically, there is a fairly extensive
literature on upper and lower probabilities (see Fine, 1988); however, this litera-
ture tends to put the emphasis on quantitative aspects. Suppes (1974) treats the
behavioural side of the problem but he uses semiorders. See also Fishburn (1986).
166
The aim would be to provide a quasi order representation as in 2.2, but including
upper and lower bounds on the probabilities.
Multiplicative and multilinear quasi order representations might also be con-
sidered.
aj is p.o. iff the optimal value is ~ 0. To compute the a.p.o. alternatives we solve
aj is a.p.o. to a. iff its optimal value is 0. Finally, remember that the distance
problems were nonsmooth (except in the Euclidean case).
Some advantages of this approach are the smaller number of problems to solve,
the smaller number of constraints per problem and the same structure in con-
straints and objective functions. However, the theory and practice of nonsmooth
optimisation problems is not yet as developed as that of smooth ones.
We have not studied multi-stage decision problems, though one of them served
for illustration purposes of SENSGEN. As Bayes' theorem is still valid, we can
see that backward dynamic programming extends to the computation of nondom-
inated and p.o. alternatives under the usual conditions. White and White (1989)
and White and EI Deib (1983, 1986) provide relevant references in this area.
167
Our theory is lirnited to the finite case. There is same theory for the infinite
case: for general preference quasi orders we can use Spilzrajn's lemma but, as we
discussed in section 2.4, this would lead to a representation different to ours; the
procedure in section 2.2 extends to the countable case; Giron and Rios (1980) give
a characterisation of general probability quasi orders; Aumann (1962) and White
(1972) give results for linear value functions. The problem with utility functions
seems more difficult (see Kannai (1963)) but perhaps can be treated as Giron and
Rios treat probabilities. Nondominated and potentially optimal alternatives are
still the solution concepts. As algorithms are concerned, we can appeal to vector
optirnisation results (Yu, 1985, The Luc, 1989). We can also discretise the problem
and apply algorithms for finite problems. It seems useful to apply Hazen's (1988)
differential results to special forms of value functions.
Several results study the stability of these systems. For example, when the in-
equalities are strict for w, if we use Tchebycheff's distance in Robinson's results
(1975, 1976) we get bounds of the kind
where Cl is some constant. This can be used to know how small S has to be if the
DM is ready to accept some loss of optimality due to choosing a. wrongly.
Another point is that a.p.o. concepts are no longer useful, in general, in infinite
problems. Alternatively, we can use t-a.p.o. alternatives to a. in a similar vein as
t- Bayes actions are used in statistical decision theory, see Ferguson (1967).
More precise meaning has to be given to the concept of elose call, see Bell et
al. (1988).
Our emphasis has been on the need for more precision. We have dealt briefly
with some topics on errors in modelling, though we have seen how to use SENSATO
to quest ion the validity of a model. Several questions in error modelling were posed
in section 3.5. We insist on the issue of consistency of S; some ideas can be seen
in White et al. (1984). Some points in our error model need further exploration
such as the use of other distributions or improvements in computations in the
multivariate case.
We can also develop estimators for bounded normal means that should be
compared with others in similar settings such as those of Casella and Strawderman
(1981) or Berger (1985).
5.2.3 Computation
Though SENSATO is based on NAG, it is easily modifiable if another optimisation
library or paekage is finally used. Other routines can also be added by simple
modification of the existing ones: for example, it would be immediate to prepare a
routine not requiring the derivatives of the evaluation functions, based on NAG's
E04UCF.
Due to the advantages mentioned above, we should consider the design of a
nonsmooth version. Moreover, the use of a smooth code led to so me ill-conditioncd
problems, see 4.8. The points raised in that section should be considered, espe-
cially those referred to global optimisation. In bilinear models, we use quadratic
programming heavily; recent developments in bilinear programming might prove
useful, see Thieu (1988).
Some work is required in the formulation of problems. Some possibilities are:
built-in facilities to generate some types of constraints (monotonie, eonvex, ... ) on
the parameters; devices to build and derive symbolically the evaluation functions,
The inclusion of helps to choose the appropriate routines should be useful.
Mueh work is needed to find how to present the results. Apart from the graphs
we have used, there are other possibilities in multivariate data analysis graphie
teehniques, see Chambers el. al. (1983). The use of eolours might enhanee the pre-
169
sentation. When comparing aJ and a. holistically, a graph of {(Ilt J( w), Ilt.( w)) }WES
should be of help. More work should be done to determine the usefulness of the
exploratory data analysis techniques used in section 4.2, even building a small
expert system to interpret the results.
Another possibility is displaying the results via semantic representations. We
have mentioned the process of modelling the judgemental inputs of the DM, which
consists of an iterative protocol of questioningJ answeringJ encoding and leads to
w. Through sensitivity analyses, we get parameters w leading to different optimal
alternatives. Thus, we could think of inverting the protocol, finding the questions
and the answers which would lead to these parameters, pose them to the DM and
use her comments to modify, if necessary, the estimate w andJor the set S. For
this, we could draw upon naturallanguage interfaces currently under development,
as in the COMMUNAL project (Fawcett, 1989).
A possible interface is given below (this would link to the last part of SEN-
SATO's output, see the appendix, which announces that the DM can either reduce
S along some direction or compare holistically aJ and a., to improve the current
status):
170
5.2.6 Restructuring
Restructuring concepts, as in Shien ehen and Yu (1989), are of use in our problems.
In our context they mean: find nontrivial transformations of 5 such that a. is
s.o. or such that the 1055 of optimality is limited by some quantity. Then, these
171
[2] Arrow, K. and Raynaud, H. (1987) Soeial Choiee and Multieriteria Deeision
Making, MIT Press.
[3] Aumann, R. (1962) Utility theory without the completeness axiom, Econo-
metriea, 30, 3, 445-462; (1964) 32, 1,210-212.
[6] Barron, H. and Sehmidt, C. P. (1988b) Entropy based selection with multiple
objectives, Naval Research Logistics Quarterly, 35, 643-654.
[8) Bell, D., Raiffa, H. and Tversky, A. (1988) Decision Making, Cambridge U.P.
[U] Belton, V. and Vickers, S. (1989) VISA - VIM for MCDA, in Lockett (ed),
Improving Deeision Making in Organisations, LNEMS, Springer Verlag.
[12J Berger, J. (1980) A robust generalized Bayes estimator and confidence region
for a multivariate normal mean, Annals 0/ Statisties, 8, 716-761.
173
[13] Berger, J. (1984) The robust Bayesian viewpoint, in Kadane (ed), Robustness
of Bayesian Analyses, North Holland.
[15] Berger, J. and Berliner, M. (1986) Robust Bayes and empirical Bayes analysis
with e-contarninated priors, Annals of Stat., 14,461-486.
[17] Bouyssou, D. (1984) Decision aid and expected utility theory: a critical
survey, in Hagen and Wenstop (eds), Progress in Utility and Risk Theory,
Reidel.
[21] Charnetski, J. and Soland, R. (1978) Multiple attribute decision making with
partial information: the comparative hypervolume criterion, Nav. Res. Log.
Quart., 25, 279-288.
[22] Charnetski, J. and Soland, R. (1979) Multiple attribute decision making with
partial information: the expected value criterion, Nav. Res. Log. Quart., 26,
249-256.
[25] Clarke, F. H. (1975) Generalized gradients and its applications, Trans. Amer.
Math. Soc., 205, 247-262.
[27] Cyert, R. M. and De Groot, M. (1975) Adaptive utility, in Day (ed) Adaptive
Economics, Prentice Hall.
[29] Danskin, J. (1967) The Theory 0/ Max·min and its Applications to Weapon
Allocation Problems, Springer Verlag.
[33] Diaconis, P. (1985) Theories of data analysis: from magical thinking through
classical statistics, in Hoaglin, Mosteller , Tukey (eds) Exploring Data Tables,
Trends, and Shapes, Wiley.
[38] Einhorn, H. and Hogarth, R. (1975) Unit weighting schemes for decision
making, Org. Beh. Hum. PerJ., 13, 171·192.
[39] Eliashberg, J. and Hauser, J.(1985) A measurement error approach for mod·
elling consumer risk preference, Mgt. Science, 31, 1-25.
[41J Farquhar, P. H. (1984) Utility assessment methods, Mgt. Sc., 30, 1283-1300.
[44] Fine, T. L. (1988) Lower prob ability models for uncertainty and nondeter-
ministic processes, Jour. Stat. Plan. InJ., 20, 389-411.
[45) Fischhoff, B. (1980) Clinical decision analysis, Opns. Res., 28, 28-43.
[49) Fishburn, P. C. (1971) One way expected utility with finite consequence
space, Annals of Mathematical Statistics, 42, 572-577.
[53] Fishburn, P. C. (1986) The axioms of subjective prob ability, Statistical Sci-
ence, 1, 335-358.
[55] Freeling, A. (1980) Fuzzy sets and decision analysis, IEEE Trans., SMC-10,
341-354.
[60] French, S. (1987) Fuzzy sets: the unanswered questions, Res. Rep., Stat.
Lab., U. Manchester.
(62] French, S. and Rios !nsua, D. (1989) Partial information and sensitivity anal-
ysis in multi-objective decision making, in Lockett (ed), lmproving Decision
Making in Organisations, LNEMS, Springer Verlag.
[63] Gardenfohrs, N. and Sahlin, P. (1988) Decision, Probability and Utility, Cam-
bridge U. P.
(65] Gill, P., Murray, W. and Wright, M. (1981) Practical Optimization, Prentice
Hall.
[69] Gomez Orea, D., Quintana, A., Aguado, M. and Villarino, M. (1989) Modelo
para La Evaluacion del Impacto Ambiental de Carreteras, ETSIA, Madrid.
[73J Hansen, P., Labbe, M. and WendelI, R. (1989) Sensitivity analysis in multiple
objective linear programming: the tolerance approach, EJOR, 38, 63-69.
[77J Humphreys, P. and Wisudha, A. (1987) Methods and tools for structuring
and analysing decision problems: vols. 1, 2, Tech. Rep.) Decision Analysis
Unit, LSE.
[80] Jaffray, J. Y. (1989) Some experimental findings in decision making and their
implication, EJOR, 38, 301-306.
[8.4] Kahneman, D., Slovic, P. and Tversky, A. (1982) Judgement under Uncer-
tainty: Heuristics and Biases, Cambridge U. P.
178
[88J Keeney, R. and Raiffa, H. (1976) Decisions with Multiple Objectives: Pref-
erences and Value Tradeoffs, Wiley.
[89J Kirkwood, C. (1982) A case history of nuclear power plant selection, JORS,
33, 353-363.
[93J Koopman, B. O. (1940) The bases of probability, Bull. Am. Math. Soc., 41,
269-292.
[94] Kraft, C., Pratt, J. and Seidenberg, A. (1959) Intuitive probability on finite
sets, Ann. Math. Stats., 30, 408-419.
[95J Krantz, D., Luce, D., Suppes, P. and Tversky, A. (1971) Foundations of
Measurement, Acadernic Press.
[97J Laskey, K. and Fischer, G. (1987) Estimating utility functions in the presence
of error, Mgt. Science, 33, 8,965-980.
[101] Leung, P. (1978) Sensitivity analysis of the effect of variations in the form
and parameters of a multi-attribute utility model: a survey, Behav. Sc., 23,
478-485.
[108) Marcotte, O. and Soland, R. (1986) An interactive branch and bound algo-
rithm for multicriteria optimization, Mgt. Science, 32, 61-75.
[110] McNamee, P. and Celona, J. (1987) DecÜlion AnalysÜl for the Professional
with Supertree, Scientific Press.
[112] Milnor, J. (1954) Games against nature, in Thrall, Coombs and Davis (eds)
Decision Processes, Wiley.
[114J NAG FORTRAN Library Reference Manual (Mark 13) (1988) Numerical
Algorithms Ltd., Oxford.
[118] Pierce, D. and Folks, J. L.(1969) Sensitivity of Bayes procedures to the prior
distribution, Opns. Res., 17, 344-350.
[119] Polak, E., Mayne, J. Q. and Ward, Y. (1983) On the extension of constrained
optimization algorithms from differentiable to nondifferentiable problems,
SIAM Jour. Gon. Opt., 21, 179-203.
[120] Pratt, J. W. and Meyer, R. F. (1968) The consistent assessment and fairing
of preference functions, IEEE Trans., SSC-4, 270-278.
[121] Prior, W. J. (1970) Ghost-SO User Manual, Release 7, UKAEA Culham Lab.
[124] Ravinder, H. V., Kleinrnuntz, D. and Dyer, J. (1988) The reliability of sub-
jective probabilities obtained through decomposition, Mgt. Sc., 34, 186-199.
[125J Rios, S. (1975) Ordre quasi-Bayesien des regles de decision, Bull. ISI, 46,
694-698.
[127] Rios, S. and Rios lnsua, S. (1986) Vector value function and vector dis-
tance methods in multicriteria optimization, presented at Int. Gon/. Vector
Optimization, Darmstadt.
[128] Rios, S., Rios lnsua, S. and Rios lnsua, M. J. (1988) Procesos de Decision
Multicriterio, EUDEMA.
181
[129] Rios Insua, S. (1980) Decisiones con Multicriterios, Ph.D. Thesis, UCM.
[135J Rios Insua, D. and French, S. (1989) A framework for sensitivity analysis in
discrete multiobjective decision making, to appear in European Journal of
Operational Research.
[138J Robinson, S. M. (1976) Stability theory for systems of inequalities. Part II:
Differentiable nonlinear systems, SIAM Jour. Num. Anal., 13, 4, 497-513.
[141] Roubens, M. and Vincke, Ph. (198.5) Preference i'vfodelling, LNEMS, Springer
Verlag.
[142] Roy, B. (1973) How outranking relation helps multiple criteria decision-
making, in Cochrane, Zeleny (eds) 111ultiple Criteria Decision kJaking, South
Carolina UP.
[143] Roy, B. and Vincke, Ph. (1984) Relational systems of preferences with one
or more pseudo criteria: so me new concepts and results, Mgt. Sc., 30, 1323-
1355.
182
[149] Sarin, R. K. (1977) Interactive evaluation and bound procedure for selecting
multi-attributed alternatives, TIMS Studies Mgt. Sc., 6, 211-214.
[157] Shien Chien, I. and Yu, P. L. (1989) MCDM decision support using restruc-
turing concept, in Lockett (ed) Improving Decision Making in Organisations,
LNEMS, Springer Verlag.
[164] Stoer, J. and Witzgall, C. (1970) Convezity and Optimization in Finite Di-
mension,. I, Springer Verlag.
[166) Sussman, B., Gremy, J. P., Mare, M. and Buffet, P. (1967) Peut on choisir
en tenant compte des criteres multiples? Une methode (ELECTRE) et trois
applications, METRA, 6, 283-316.
[167] The Luc, D. (1989) The Theory 0/ Vector Optimization, LNEMS, Springer
Verlag.
[169) Tversky, A. (1967) Additivity, l1tility and subjective probability, Jour. Math.
Psychol., 4, 175-201.
[171) Von Winterfeldt, D. and Edwards, W. (1982) Costs and payoffs in perceptual
research, Psych. Bull., 19, 609-622.
[172) Von Winterfeldt, D. and Edwards, W. (1987) Decision Analysis and Be-
havioural Research, Cambridge U.P.
[173) Walley, P. and Fine, T. (1979) Varieties of modal and comparative proba-
bility, Synthese, 41, 321-374.
[177] Watson, S. R., Weiss, J. W. and Donnell, M. L. (1979) Fuzzy decision anal-
ysis, IEEE Trans., SMC-9, 1-9.
[178] Weber, M. (1987) Decision making with incomplete information, EJOR, 28,
44-57.
[181] White, C. C. and EI Deib, H. (1983) Multi-stage decision making with im-
precise utilities, in Hansen (ed) Essays and Surveys on Multiple Criteria
Decision Making, LNEMS, Springer Verlag.
[184] White, C. C., Sage, A. and Scherer, W. (1982) Decision support with par-
tially identified parameters, Large Scale Systems, 3, 177-189.
[185] White, C. C. and White, D. J. (1989) Markov decision processes, EJOR, 39,
1-16.
[191] White, D. J. (1989) Vsing weighting vectors to solve multiple attribute prob-
lems: a framework, EJOR, 293-304.
185
[193] Wolfenson, M. and Fine, T. L. (1982) Bayes-like decision making with upper
and lower probabilities, JASA, 77, 80-88.
[194] Yager, R. (1977) Multiple objective decision making using fuzzy sets, Int.
Jour. Man-Machine Studies, 9, 375-382.
[196] Yoon, K. (1989) The propagation of errors in multi at tribute decision analy-
sis, JORS, 40, 681-686.
[198] Zimmermann, H. J. (1987) Fuzzy Sets, Decision Making and Expert Systems,
Kluwer.
Appendix
This appendix provides the output of the analysis of Dallas' problem with SEN-
SATO. See seetions 4.1 and 4.2 for comments and details on the results.
187
ALTERNATIVE 6
ALTERNATIVE 2
ALTERNATIVE 3
ALTERNATIVE 4
PIlS
THE MINIMUM DISTANCE TO PRODUCE THE OUTRANKING
OF THE CURRENT OPTIMUM BY ALTERNATIVE 6
IS 0.4462
10.00000
7.00000
3.00000
5.00000
5.77692
8.00000
9.00000
4.22308
2.00000
1. 00000
10.0000 5.0000
7.0000 4.0000
3.0000 10.1579
5.0000 5.0000
6.0000 9.9953
8.0000 5.0000
9.0000 5.0000
4.0000 3.0000
2.0000 2.8559
1.0000 4.9908
PI IS 2
TRE MINIMUM DISTANCE TO PRODUCE TRE OUTRANKING
OF TRE CURRENT OPTIMUM BY ALTERNATIVE 6
IS 0.2507
10.0000 5.0000
7.0000 4.0000
3.0000 15.0000
5.0000 5.0000
6.0000 10.0000
8.0000 5.0000
9.0000 5.0000
4.0000 3.0000
2.0000 3.0000
1.0000 0.0000
PI IS 3
10.00000
7.01032
2.96458
5.00000
5.85333
8.07688
9.00000
4.13613
2.07158
0.88718
OR REDUCE THE FEASIBLE SET OF PARAMETERS, ALONG THE DIRECTION
10.0000 5.0000
7.0000 4.0000
3.0000 15.0000
5.0000 6.0365
6.0000 10.0000
8.0000 5.0000
9.0000 5.0000
4.0000 4.0000
2.0000 0.0000
1. 0000 0.9635
***** ••• **************.*******
PI IS 4
*****************************.
PlIS 210