Evaluating The Quality of Approximations To The Non-Dominated Set
Evaluating The Quality of Approximations To The Non-Dominated Set
Andrzej Jaszkiewicz
Institute of Computing Science, Poznan University of Technology, ul. Piotrowo 3a, 60-965 Poznan,
Poland, email: [email protected]
Abstract: The growing interest in hard multiple objective combinatorial and non-linear
problems resulted in a significant number of heuristic methods aiming at generating sets of
feasible solutions as approximations to the set of non-dominated solutions. The issue of
evaluating these approximations is addressed. Such evaluations are useful when performing
experimental comparisons of different multiple objective heuristic algorithms, when defining
stopping rules of multiple objective heuristic algorithms, and when adjusting parameters of
heuristic algorithms to a given problem. A family of outperformance relations that can be used
to compare approximations under very weak assumptions about a decision-makers
preferences is introduced. These outperformance relations define incomplete orders in the set
of all approximations. It is shown that in order to compare approximations, which are
incomparable according to the outperformance relations, much stronger assumptions about the
decision-maker's preferences are necessary. A general framework that can be used to compare
and evaluate approximations under the presence of various types of additional information is
proposed. Some particular comparison and evaluation methods based on this framework are
suggested. The proposed framework is also used to characterize some previously proposed
evaluation methods.
Keywords: Multiple objective optimization; Heuristics; Evaluation
Introduction
Research on multiple objective optimization was traditionally concentrated on linear
programming problems (see e.g. Steuer, 1986 and White, 1990). Efficient solutions of such
problems are in most cases generated by solving some single objective optimization problems,
e.g. by optimizing a scalarizing function on the original set of feasible solutions. As very
effective methods and software for single objective linear optimization exist, most multiple
objective linear programming problems can be effectively solved in this way, except of some
very large scale problems.
2
In recent years, however, researchers have become increasingly interested in other classes of
multiple objective problems, for instance multiple objective combinatorial problems (cf. Ulungu
and Teghem, 1994). This interest is raised by practical applications, e.g. in project scheduling
(see e.g. Slowinski, 1989), vehicle routing (see e.g. Assad, 1988) and engineering design
(Dasgupta and Michalewicz 1997). For example, solutions to vehicle routing problems are
usually evaluated by e.g. total cost, distance, travel time and the number of vehicles used. In
practice, it can therefore be difficult to evaluate a solution to such a problem with only a single
objective. The objectives, however, are traditionally used separately or they are combined into
a single objective.
Multiple objective problems are often hard even in the single objective case. For example, most
vehicle routing problems are extensions of travelling salesperson problem (TSP) which is
already NP-hard. It is also worth mentioning that some combinatorial problems, which are easy
in single objective case, turn hard when multiple objectives are considered. For example, the
single objective shortest path problem is one of the simplest combinatorial problems while the
corresponding multiple objective problem is NP-hard (Serafini, 1987).
For hard multiple objective problems it may be necessary to settle for approximations to the
efficient solutions. As single objective metaheuristic procedures, e.g. simulated annealing
(Cerny, 1982, Kirkpatrick et al., 1983 and Laarhoven and Aarts, 1987), tabu search (Glover,
1989) and genetic algorithms (Goldberg, 1988), often are successful in the single objective
optimization problems it seems natural to use them in the case of multiple objective
optimization.
Several authors have proposed multiple objective metaheuristic procedures that aim at the
effective generation of approximations of the non-dominated set. The methods are based on
ideas of genetic algorithms (Schaffer, 1985, Fonseca and Fleming, 1993, Horn, Nafpliotis and
Goldberg, 1994, Srinivas and Deb, 1995; see also Fonseca and Fleming, 1995, for a review),
simulated annealing (Serafini, 1994, Ulungu et al., 1994 and Czyzak and Jaszkiewicz, 1995) or
tabu search (Gandibleux et. al., 1996, and Hansen, 1997). Authors of such methods usually
state that the methods should generate good approximations of the non-dominated set. The
term good approximation, however, is often only defined intuitively, as being close to the
real non-dominated set and well-dispersed over this set.
The main purpose of this paper is a more rigid definition of the goal of multiple objective
heuristic algorithms. More precisely, we propose some tools that may be used to evaluate and
compare approximations and define what is understood by a good approximation to the set
of non-dominated solutions.
Please note that the issue of measurement is simple if a single objective is considered. In that
case, we evaluate and compare approximate solutions using the obvious quality measure,
which is the value of the objective function. No such natural measure exists in multiple
objective case.
Evaluations of approximation of the non-dominated set may be used to:
3
generate better approximations in a shorter time than others. A natural way to evaluate the
methods is to perform comparative experiments. At present, however, different authors use
different approaches to evaluate quality of obtained approximations. This makes it practically
impossible to perform a fair comparison of different multiple objective heuristics.
Both single and multiple objective metaheuristic procedures involve many parameters that have
to be adjusted for a given class of problems. Settings of the parameters may have crucial
influence on the quality of the algorithm applied to a given class of problems. Although some
general guidelines exist, adjustment of the parameters usually requires some experiments with
different settings and evaluations of the results. In the multiple objective case, it requires
evaluating and comparing approximations obtained with different settings of the parameters.
Stopping rules of single objective metaheuristic algorithms are often based on observations of
the objective function value (see e.g. Laarhoven and Aarts, 1987, and Goldberg, 1988). A
procedure may for example be stopped if in a given number of iterations the improvement of
the objective function value is below a given threshold. In the multiple objective case, such
stopping rules require an evaluation of the quality of the current approximation of the
non-dominated set.
We believe also that clear understanding of the goal of multiple objective heuristic algorithms
is a necessary preliminary step towards any kind of theoretical analysis of such algorithms.
The paper is organized in the following way. In the next Section some basic definitions are
given. In the second Section, we introduce outperformance relations for comparing pairs of
approximations under very weak assumptions about the decision-makers (DMs) preferences.
Quantitative comparison methods using probability distribution of the DMs possible
preferences are presented in the third Section. In the fourth Section, we characterize types of
additional information that may be necessary in quantitative comparison methods. Practical
guidelines for constructing and computing quantitative measures are presented in the fifth
Section. In the sixth Section we comment on some previously used approaches for evaluation
and comparison of approximations to the non-dominated set. The issue of evaluating
approximations under presence of more precise preference information is discussed in the
seventh Section. In the eighth Section, we outline the possibilities of using the preference
information within multiple objective heuristics. Conclusions and directions for further research
are summarized in the last Section.
1. Basic definitions
(P1)
where: solution =
is a vector of decision variables and D is the set of feasible
solutions. The type of the variables, constraints and objective functions may then describe
different classes of problems. If the decision variables are continues, one may have a
multiobjective linear program (MOLP) or a multiobjective non-linear program (MONLP)
depending on the linearity of the constrains and the objective functions. If the variables are
integers, (P1) turns into a multiobjective integer program (MOIP), which again may be both
linear and non-linear. Multiobjective combinatorial optimization (MOCO) problems can often
be formulated as linear multiobjective integer programs.
The image of a solution x in the objective space is a point z x = z1x ,..., z Jx , such that z xj = fj(x),
j=1,..,J. The image of the set D in the criterion space is a set Z composed of attainable points,
i.e. points being images of feasible solutions.
Problem (P1) can also be formulated more succinctly as:
maximize {z}
s.t.
(P2)
z Z
j = 1,..., J .
The point z** composed of the worst attainable objective function values is called the anti-ideal
point:
z** j = min {z j | z Z }
j = 1,..., J .
The point z* composed of the worst objective function values in the non-dominated set is
called the nadir point.
Range equalization factors (Steuer, 1986, sec. 8.4.2) are defined in the following way:
j =
1
, j=1, , J
Rj
(1)
where Rj is the range of objective j in the set N or D. Objective function values multiplied by
range equalization factors are called normalized objective function values.
A utility function u:J, is a model of the DMs preferences that maps each point in the
objective space into a value of utility. It is assumed that the goal of the DM is to maximize the
utility.
A utility function u is compatible with the dominance relation if and only if
z 1 , z 2 J z1 z 2 u z1 u z 2 . The set of all utility functions that are compatible with
the dominance relation is denoted by Uc.
( ) ( )
A utility function U is strictly compatible with the dominance relation if and only if
z 1 , z 2 J z1 z 2 u z1 > u z 2 . The set of all utility functions that are strictly
compatible with the dominance relation is denoted by Usc.
( ) ( )
A convenient way to define a set of utility functions is by the use of parametric utility functions
u(z, r), r D(r ) n , where r is a vector of parameters and D(r) is domain of the parameter
vector. A parametric set of utility functions is then defined in the following way:
U (r ) = {u (z, r )| r D (r )}.
Weighted Lp norms are defined as:
J
L p z , z , = i z 1j z 2j
j =1
1/ p
, p {1,2,...}+ {},
J
u p z, z , , p = i z *j z j
j =1
1/ p
, p {1,2,...}+ {}.
{(
)}
u z, z * , = max i z *j z j .
j
(2)
For p=1 one obtains a parametric set U1 of weighted linear utility functions:
(z, z , ) = (z
J
up
j =1
*
j
zj
) = z + z
J
j =1
*
j
j =1
= const + i z j . (3)
j =1
6
2. Outperformance relations
In the case of a MOO problem, the overall goal of the decision-maker (DM) is to select the
single solution, which is most consistent with his or her preferences, the so-called best
compromise. Generating an approximation to the non-dominated set is only a first phase in
solving the problem. In the second phase, the DM selects the best compromise solution from
the approximation, possibly supported by an interactive procedure. Therefore, the DM may
consider approximation A as being better than approximation B if he or she can find a better
compromise solution in A than in B.
We assume, however, that the DMs preferences are not known a priori. In fact, the use of
heuristics generation of approximations to the full non-dominated set is justified only in this
case. Nevertheless, one may be able to make some general assumptions about possible DMs
preferences.
Using assumptions, we can state that an approximation A outperforms (is better than) B if, for
some possible preferences held by the DM, the DM may find a better compromise solution in A
than may be found in B and for other possible preferences, the solution found in A will be not
worse than those found in B. Specifically, we will assume that all possible preferences of the
DM may be modeled by utility functions belonging to a set U.
Let A and B be two approximations. Let U (A > B ) U denote a subset of utility functions for
which approximation A is better that B, i.e. U (A > B ) = u U | u * (A) > u * (B ) . Then, the
following relation may be defined.
10
3. Quantitative comparison methods
In this Section, we will describe some quantitative comparison methods that can be used to
compare approximations that are incomparable according to the outperformance relations.
Quantitative comparison methods will also be based on some assumptions about the DMs
preferences. It is natural to demand evaluations obtained by a quantitative comparison method
to be concordant with outperformance relations based on these assumptions or on more
general assumptions about the possible preferences. This demand is defined formally below.
Definition 5. (Weak compatibility with an outperformance relation)
A comparison method R is weakly compatible with an outperformance relation O, where
O=O/U, OW, OS or OC, if for each pair of approximations A and B, such that A O B, R will
evaluate approximation A as being not worse than B.
Definition 6. (Compatibility with an outperformance relation)
A comparison method R is compatible with an outperformance relation O, where O= O/U, OW,
OS or OC, if for each pair of approximations A and B, such that A O B, R will evaluate
approximation A as being better than B.
3.1 Comparison based on probability
Assume that each of the considered utility functions has a given probability of being the one
held by the DM. When comparing two approximations A and B, it would then make sense to
consider approximation A as being the better one, if there is a high probability for the utility
functions in which approximation A presents a better solution than approximation B.
We express the probabilities of utility functions by an intensity function p(u) and introduce an
outcome function of the comparison between two approximations using a given utility
function:
1 if u * ( A) > u * ( B)
C ( A, B, u ) = if u * ( A) = u * ( B)
0 if u * ( A) < u * ( B)
We then build the measure R1 to reflect the probability that approximation A is better than
approximation B by integrating over all utility functions:
R1( A, B,U , p) =
C ( A, B, u ) p(u ) du
uU
According to this measure approximation A is better than B if R2(A,B,U,p) > and A is not
worse than B if R2(A,B,U,p) . We notice, that R1(A,B,U,p) = 1 R1(B,A,U,p) and is
therefore not possible, that one approximation can be better than another one while the
opposite also is true.
R1 is weakly compatible with the outperformance relation for any set of utility functions
U Uc. R1 is compatible with outperformance relation subject to a U set of utility functions if
the same set U is used in R1 and if the probability of selecting a utility function uU(A > B) is
11
always greater than zero whenever U(A > B) . The latter condition is, for example, assured
if U is a parametric set and the probability intensity is a continuous and positive function of the
parameter vector.
As an example, Figure 7 shows the u* values of two approximations A={[3,10],[5,7],[9,7]}
and B={[2,9],[5,6],[10,6]} using the Chebycheff utility function with z* = [10,10] and
= [t, 1 t], t 0,1:
Only when t > 0.8 will approximation B contain the best alternative. If we presume that the
utility functions are distributed as follows by letting t belong to the continuous uniform
distribution, t U0,1, we get R1(A,B) = 0.8.
0
0.2
0.4
0.6
0.8
0
-0.5
u*
-1
Approximation A
-1.5
Approximation B
-2
-2.5
-3
parameter t in utility function
0.2
0.4
0.6
0.8
0
-1
u*
-2
Approximation A
-3
Approximation B
-4
Approximation C
-5
-6
-7
parameter t in utility function
12
A different approach to ranking a number of approximation is to compare the approximations
individually with a fixed reference set, R:
R1R ( A,U , p ) = R1( A, R,U , p) =
C ( A, R, u) p(u) du
uU
This measure induces a complete ordering of approximations and cycles can therefore not
occur. The measure is weakly compatible with the outperformance relation for any set of utility
functions U Uc.
However, it is not compatible even with the complete outperformance relation, if e.g. A OC B
but R OC A, then R1R(A,U,p) = R1R(B,U,p) = 0. This also shows that for this measure to be
useful, the reference set R should reflect an attainable quality over the utility functions.
It is also worth mentioning that the resulting ranking of approximations will depend on the
reference set used. With two different reference sets, R and P, we can be in a situation where
R1R(A,U,p) > R1R(B,U,p) and R1P(A,U,p) < R1P(B,U,p), even if R OC P. Still, we consider it
beneficial to use more than one reference set, especially if these are on significantly different
quality levels.
Notice, that the C(A,B,u) function merely performs a ranking of the two approximations to see
which contains the best point with respect to the utility function. This approach can easily be
generalized to encompass more than two approximations, A1, A2, , AK by considering the
rank of an approximation in u*(A1), u*(A2), , u*(AK):
C ( Ai ; A1 , A2 ,..., AK , u ) =
where rank(x; X) is the rank of element x in the set X giving the value 1 is x = min{X} and the
value |K| if x = max{X}. The measure from integrating over all u in U then yields the expected
relative rank of an approximation among the K approximations.
A similar approach is to use a group of reference sets, R1, R2, , RK and calculate the
expected relative rank of an approximation with respect to this set. These two generalizations
are inspired by work of Fonseca and Fleming (1996).
3.2 Comparison based on expected values
Consider the following two approximations A={[1,10],[10,1.8]} and B={[2.2,10],[7,-1]}.
Values of u* are presented in Figure 9, where u is a Chebycheff utility function (2) defined in
the same way as in Section 3.1. When t < 0.5125, approximation B contains the best
alternative. So, applying measure R1 we obtain R1(B) = 0.5215 > R1(A) = 0.4875. However,
in the region where approximation B gives better utility approximation A gives only slightly
worse results, while in the other region the utility given by approximation A is significantly
better. So, intuitively we might evaluate A better that B.
13
0.2
0.4
0.6
0.8
0
-1
Approximation A
-3
Approximation B
u*
-2
-4
-5
parameter t in utility function
) (
R 2(A, B, U , p ) = E u * (A) E u * (B ) =
uU
uU
uU
According to this measure approximation A is better than B if R2(A,B,U,p) > 0 and A is not
worse than B if R2(A,B,U,p) 0. Obviously, R2(A,B,U,p) = R2(B,A,U,p).
In the above example we obtain E(u*(A)) = -2.1239, while E(u*(B)) = -2.5924, and therefore
R2(A,B,U,p) = 0.4684 > 0. So, approximation A is evaluated better than B.
R2 is weakly compatible with the outperformance relation for any set of utility functions U
Uc. R2 is compatible with outperformance relation subject to a U set of utility functions under
the same conditions as R1.
R2 induces complete ranking in the set of all approximations.
Measure R2 is based on an assumption that we are allowed to add values of different utility
functions. Therefore, it is also dependent on the scaling of the different utility functions.
Assume for example that in the above example each Chebycheff function is multiplied by
(1 t )2 (see Figure 10). In this case we obtain E(u*(A)) = -0.6361, while E(u*(B)) = -0.59996
and then R3(A,B,U,p) = -0.03614 < 0. So, approximation B is evaluated better than A.
14
0.2
0.4
0.6
0.8
0
-0.2
u*
-0.4
-0.6
Approximation A
-0.8
Approximation B
-1
-1.2
-1.4
parameter t in utility function
The same method can be used to construct a quality measure that evaluates a single
approximation A with respect to a fixed reference set R:
) (
(u (R ) u (A))p(u )du
*
uU
uU
uU
In this case, the lower the value of the measure is, the higher is the evaluation of A.
In some cases, ratios of best utility values may be more meaningful than their differences.
Therefore, we propose also the following ratio measure:
u * (B ) u * (A )
u * (B ) u * (A)
R3(A, B,U , p ) = E
=
p(u )du .
u * (B )
u * (B )
uU
When a reference set is known we may use the idea of R3 to construct the following measure:
u * (R ) u * (A)
u * (R ) u * (A)
=
p(u )du
R3 R (A, U , p ) = R 3(A, R,U , p ) = E
u * (R )
u * (R )
uU
The idea of R3R follows an approach often used in single objective optimization where
approximate solutions obtained by single objective heuristics are evaluated by the ratio of the
deviation from a fixed bound, for instance the optimal value.
In general, also the values of the ratio measures are dependent on the scaling of the objective
functions.
15
4. Reference information
The dominance-based outperformance relations introduced in Section 2 do not require
information except the points of the two approximations being compared. However, in the case
of outperformance relation subject to a set of utility functions and the quantitative comparison
methods proposed in Section 3 there is a need for further information. In this Section, we
characterize some sources of additional information about a given MOO problem (beyond
those characterizing DMs preferences) that could be useful while comparing approximations
of the non-dominated set.
4.1 Reference levels
In Section 3, we have used a reference set, R, with respect to which we can compare an
approximation. Here we will discuss some of the requirements we have to such a set. First
please notice, that we only use the reference set in order to obtain values of u*(R) for each
given utility function. We may extend this by describing a reference level to each utility
function. This is a generalization, since there might not exist a realization of a reference set, R,
which can give the same values of u*(R) for each of the utility functions.
The reference level may then be described by, for instance, a relaxed correspondent to the
original problem, relaxing the integer/binary constraints on decision variables or relaxing other
constraints. This may result in upper bounds with known maximal (or consistent) deviation
from optimum, insuring some degree of homogeneous quality over the non-dominated frontier.
Lower bounds for the exact reference levels can be found from e.g. approximation algorithms
such as heuristics. Using these, one may be in a situation where an approximation performs
better than the reference level for some utility functions. However, this is no different from the
single-objective case. Again, the most important issue is to insure a homogeneous quality level
over the non-dominated frontier.
Publication of such reference levels can be given as the relaxed model used, the approximation
algorithm used, etc., but it can be useful for other researcher if also some of the actual utility
function values are available. For measure R2R one can publish the integral value (perhaps with
different numerical integration procedures and different sampling intervals) of
u * ( R) p(u)du .
uU
It will often be useful to provide more than one reference level for a problem instance. As
stated in Silver et. al. (1980), the purpose of heuristic methods is not only to be close to the
optimal solution, but also to be far away from poor solutions. Reference levels describing poor
solutions should therefore also be provided, for instance as generated by random solutions. In
between, one may provide mediocre reference levels such as those obtained by generating local
optima from e.g. simple steepest ascent heuristics.
16
4.2 Reference sets
Nevertheless, it can be more convenient to use an actual set of points, a reference set, to
describe the reference levels since it may allow for easy interchange among researchers. Also,
using an actual reference set may allow for fast calculation of some of the measures using
efficient data-structures, such as quad-trees (see Habenicht, 1982, and Sun and Steuer, 1996).
The reference set should ideally be the full (optimal) non-dominated set but for computational
reasons, one may have to settle for or a subset of this, for instance a well dispersed subset, all
or some of the supported non-dominated points, or the like.
Still some problems may be so difficult to solve, that one cannot expect to find any
(guaranteed) non-dominated points at all, to use in the reference set. A second approach is
then to use a set of potentially non-dominated points as generated by approximation
algorithms.
If the reference set does not solely consist of non-dominated points, one must pay special
attention to insure a homogeneous quality of the potentially non-dominated points all-over the
frontier. This must be argued for by the constructor of the reference set. Two main arguments
will here be the deviation from optimality/non-dominance (which again can be measured in
different ways) and the computational/algorithmic effort that has been used to obtain the
points.
When the reference set contains non-dominated points, but only a subset of these, it will also
be necessary to argue for homogeneous quality over the non-dominated frontier; here in terms
of how well dispersed the subset is. The issue of well dispersed-ness is often relevant, since
non-dominated sets can be very large indeed. Filtering techniques may here be useful (Steuer,
1986).
4.3 The ideal point
In the utility functions themselves, we will also often use additional information. Consider for
example the utility functions within Up. These use a weight vector and the ideal point, z*, as an
anchoring point. We will first discuss the ideal point and then the weight vector.
The most used anchoring points are the ideal point and the nadir point. When we prefer using
the ideal point, it is due to the fact, that the nadir point is notoriously difficult to determine for
problems with more than two objectives. One risk to estimate the nadir point terribly wrong
even for otherwise simple problems (Weistroffer, 1985). Even in the relatively simple case of
multiple objective linear programming, the estimation is difficult, especially as the problems
grow in size (Korhonen et. al., 1997).
Hence, we prefer to use the ideal point in the scalarizing functions defining the utility functions.
However, determining the exact ideal point may also cause difficulties, since this imply solving
each of the single objective programs to optimality. This may be possible in some cases
whereas for others one has to estimate the ideal point. In the latter case it is important not to
underestimate the ideal point since this implies that approximation points non-dominated with
respect to the estimated ideal point normally will not contribute fully to the measure, and in
some cases, will not contribute at all.
17
If one has to estimate the Ideal point, or some of its components, it is therefore important that
the estimate is an overestimation. This may be found as upper bounds to the problem, perhaps
from a relaxation. If some of the components of the Ideal point could be the optimal values,
one should in order for the measures to function properly for points attaining the optimal level
on one or more objectives add an -value to each component of the Ideal point (Steuer, 1986,
Section 14.1). The -value can be set as a very small positive constant multiplied with the
range equalization factors.
The utility function values will obviously depend on the coordinates of the ideal point and so
will the measures. One must therefore always accompany the measuring results with the
coordinates of the used ideal point including the -values. For the utility functions of family U1
however, the utility function values only differ by a constant.
4.4 Range scaling
The purpose of the weight vector is in fact twofold. One is to prescribe the importance of
each objective, the higher the weight, the higher the importance of that objective. The second
is to scale the ranges of the objectives, so those objectives with large ranges do not dominate
the objectives with smaller ranges. These two are normally separated into two multiplicative
factors, the importance weight and a range equalization factor. In Section 5 we suggest ways
of using different importance weights in the measures and will for now only consider the range
scaling in terms of the range equalization factors, as defined in formula (1).
The range equalization vector holds as the jth element 1 divided by the range on objective j.
Most naturally would be to use the ideal and nadir point for calculating the ranges, and this we
suggest to do for bi-objective problems. With more than two objectives, the nadir point is as
mentioned previously often so difficult to find, that alternative approaches can be desirable.
Some possibilities will be discussed in the remainder of this Section.
The anti-ideal point (or an approximation for this) can be used in replacement for the nadir
point. This makes most sense if the objectives are not correlated and if it is likely that the nadir
point is placed close to a line between the ideal point and the anti-ideal point.
A probabilistic approach can be used if the objectives are independent and thereby not
correlated. In this case, the levels on each objective will be independent of the level on other
objectives and a fixed (low) percentile on each objective from e.g. a sample of random
solutions can be used to estimate the levels of the nadir point. One must be careful when using
this approach on problems where the objectives only are non-correlated but not independent.
One can omit range scaling in cases where it can be assumed, that the ranges are more or less
equal. This may e.g. be the case for problems where all objective functions are of the same type
and the parameters defining the corresponding single objective instances are generated
independently and in the same way.
The pay-off table, i.e. the matrix formed by the points which define the ideal point, can be used
to obtain a setting for the nadir point as the lowest observed level on each objective. In cases
of multiple optima for the single-objective problem, effort can be put into locating all optima.
This approach may be best for problems with strongly correlated objectives and is often seen
used in practice.
An empirical approach is to build an over-all best approximation from all approximations
obtained in the experiments on the problem instance. From this set, an empirical nadir point
18
can be found and used for the final calculations of measures. This may be the best option in
cases with strong non-linearity in objectives, when the solution space is restricted in special
ways or when there is some correlation among objectives.
A common-sense addition to the other approaches can be desirable. For instance, an objective
may obtain extremely poor values, as can be the case in e.g. bi-criteria project scheduling with
minimization of costs and project completion time. Here, the schedules of low cost may be so
lengthy, that they are without practical interest. A lowest-value-of-interest on some objectives
may therefore be relevant in calculating the range equalization factors. This is obviously also
necessary if the problem has unbounded objectives as can be the case in multiobjective
programs other than MOCO problems.
Again, the range equalization factors should always be given with the measuring results. This
allows other researchers to calculate directly comparable measures (using measures R2 and R3)
for their approximations.
Also, if the nadir point is estimated through feasible solutions (as is the case with the pay-off
table and the empirical approach) the points defining the nadir point should be reported so as
to possibly improve the estimation of the nadir point for other researchers.
19
In this Section we give some general guidelines for setting these parameters and suggest best
practice for a number of cases.
5.1 The choice of the set of utility function
If no additional information about the possible preferences of the DM is available, one should
choose a set of utility functions that does not disregard any non-dominated points. Obviously,
this is assured by the set Uc of utility functions compatible with the dominance relation but this
set cannot be used in practice. We propose to use the set U of weighted Tchebycheff utility
functions in this case.
In other cases, additional assumptions about the DMs preferences may be justified. For
example, one may know that the DMs preferences can be expressed by a weighted sum of
objectives, but where the weights are otherwise unknown. In this case, the set U1 of weighted
linear utility functions should be used.
5.2 The choice of probability distribution of the utility functions
Assume that two approximation A, B are incomparable according to outperformance relation
subject to the selected set of utility functions U. It is then always possible to choose a
probability distribution such that one of the approximations will be evaluated higher than the
other one by the quality measures. This can be done by defining intensity function p(u) that
take on high values in regions of set U where one approximation gives higher utility and low
values in the other regions. Of course, some of the distributions will be more artificial that
the others.
An obvious requirement is that the intensity function p(u) should be greater than 0 for all uU.
Otherwise, functions with intensity function value equal to zero can be excluded from U. The
probability distributions also should not favor any subregion of U.
In the case of weighted Tchebycheff or weighted linear utility functions, the utility functions
are defined by the selected weight vector (see formulas (2) and (3)). Therefore, the
distribution of the utility functions can be defined via the distribution of weights. For practical
reasons, we propose to use normalized weight vectors, i.e. vectors belonging to the following
set:
20
J
J
= | j = 1 and j > 0, j = 1,.., J ,
j =1
We then propose to use a uniform distribution of weights, i.e. a distribution for which:
p( )d / p( )d =V ()/ V ( )
where V() and V() are Euclidean hyper-volumes of and , respectively. In other
words, the probability of a weight vector belonging to should be proportional to the
hyper-volume of . The normalized weight vectors should be applied to normalized objective
values, i.e. original objective values multiplied by range equalization factors (1).
Of course, if there are reasons to assume that some utility functions are more probable, one can
modify the probability distribution to reflect this knowledge, as we will discuss in Section 7.
5.3 Utility functions scaling
Utility functions scaling is crucial for measures R2 and R3 based on expected values. The
measures require that similar changes of the best values of different utility functions reflect
similar changes of the quality of the best solutions, i.e. the utility functions should be linear
with respect to the intuitive interpretation of the quality of solutions.
Having a set U of utility functions, we propose to scale them to the range 0-1, where 0 is the
worst and 1 is the best value. The extreme values should be achieved by points of some
intuitively understood quality, e.g. the quality of a reference set.
If a good reference set R is known, the utility functions should be scaled such that their
maximum in this set is equal to 1. Otherwise, one can use points obtained by solving a relaxed
correspondent to the original problem, if a tight relaxation can be found (cf. Section 4.1). If no
such information is known, we propose to use ideal point or its approximation to scale the
utility functions, so that they achieve the value 1 in the (approximation of) ideal point.
The value of 0 should be associated with some poor points, preferably worse than all the
approximations evaluated. One can for instance use (the approximations of) the anti-ideal or
the nadir point. If no approximation of the nadir point is known but range equalization factors
j are defined through other methods, one can associate value of 0 with the following point:
z = z1 ,..., z J ,
z j = z *j 1 , j = 1,..., J ,
j
where >0 and large enough for z to be dominated by all the points in evaluated
approximations.
5.4 Computational issues
Measures defined in Section 3 are generally difficult to calculate analytically since they require
integration of non-smooth functions. Daniels (1992) proposed a method that can be used to
calculate R3 in the case of linear utility functions.
21
Of course, the values of all the measures can be found by numerical integration. This requires
sampling the set U of utility functions according to its distribution. In the case of weighted
Tchebycheff or weighted linear utility functions this turns into sampling weight vectors . In
Section 5.2 we proposed to use set of normalized weight vectors. We propose to sample
this set by generating all weight vectors in which each individual weight takes on one of the
following values: l , l = 0,..., k , where k is a sampling parameter defining the number of
k
k + J 1
weight
weight levels. With a combinatorial argument, we notice that this produces
J 1
vectors.
For example, for k=3 and J=3, we obtain the following set of 10 vectors: {[0,0,1], [0,1/3,2/3],
[0,2/3,1/3], [0,1,0], [1/3,0,2/3], [1/3,1/3,1/3], [1/3,2/3,0], [2/3,0,1/3], [2/3,1/3,0], [1,0,0]}.
Again, the weight vectors should be applied to normalized objective values.
If the results of numerical integration are reported in the description of an experiment the
authors should precise all the necessary details of the algorithm.
22
A R
R
This measure was used by e.g. Ulungu (1993) and by Morita and Katoh (to appear).
If the reference set does not contain all non-dominated points then the points from A, which
are non-dominated by points contained in R, may actually belong to the non-dominated set. In
this case, it may be more reasonable to use the following measure, which is defined as the ratio
of approximation points non-dominated by R:
C 2 R (A) =
{u A | /r R r
A
u}
23
Figure 11. Reference set composed of all supported solutions and two approximations
obtained for a two objective knapsack problem (100 elements).
Another drawback of the cardinal measures is illustrated by the example presented in figure 12.
The two approximations are composed of 5 non-dominated points, so, their cardinal measures
are equally good. All points composing approximation 3 are, however, very close in the
objective space, i.e. they represent the same region of the non-dominated frontier. The points
of approximation 4, on the other hand, are dispersed over whole reference set. They carry
much richer information, e.g. about the possible ranges of objectives. This example shows that
for the cardinal measures, each point in the approximation has the same weight regardless of
their proximity and information concerning the shape of the non-dominated set.
Figure 12. Reference set composed of all supported solutions and two other
approximations obtained for a two objective knapsack problem (100 elements).
6.2 Distance measure
Czyzak and Jaszkiewicz (1998) proposed the following distance measure based on a reference
set R:
24
D1R (A, ) =
{ (
)} and
1
R
zA
= [1 ,...., J ], j = 1
Rj
25
According to our framework, functions d(r, z) may be interpreted as different utility functions
and the value of D1 as the expected value ever these functions. Distribution of the functions
depends on the distribution of points in the reference set that may be distributed very unevenly.
Please notice that this very well may be the case even if the reference set is composed of all the
non-dominated points.
There is no reason to assume that the regions of the non-dominated set with high density of
points are more important for the DM. For example, the fact that in a bi-objective project
scheduling problem most efficient schedules have long completion time does not mean that
they are more important. Before using this measure one should therefore assure uniform
distribution of points in the reference set or, alternatively, weight the points so as to give low
importance (e.g. low weight) to points which are located close to other points in the objective
space.
26
Logically one should use probabilities, which integrate to one. This could also be omitted,
however, since it only serves to scale the measure.
12
10
8
Normal distribution
Uniform distribution
4
2
0
0.00
0.20
0.40
0.60
0.80
1.00
27
8. The use of measures in multiobjective heuristics
In the beginning of this paper we stated, that the measures are useful information during
execution of a heuristic procedure, namely as stopping rules and as a global objective function.
However, we cannot expect to have reference information such as a reference set or the Ideal
point available during the execution of a heuristic. This information is obviously part of the
solution, which is sought and knowing it in advance would correspond to, in the case of single
objective optimization, if we knew the optimal solution (or a good bound for it) before the
heuristic procedure begins. Still, the heuristic may contain components aiming at acquiring
such information. For example, the knowledge of the ranges of the objectives is often essential
to a multi-objective heuristic aiming at generating an approximation which is equally good all
over the non-dominated frontier. If reference information is used in heuristics for benchmark
experiments, this must be explicitly stated along with the publicized results.
So, while a priori known reference information in general should not be used in the heuristic, it
may to some extent be necessary if preference information is to be included. For instance, it
does not make sense to define preference weights (such as a sub-space of the weight-space) for
the Tchebycheff based utility function without prescribing an anchoring point. However, the
anchoring point may be some point located arbitrarily far away from the ideal point on the
diagonal of the contours defined by the scalarizing function used and going through the ideal
point. In this way, the heuristic can not use the anchoring point for reference information.
Finally, the formula (but without a priori calculated parameters) by which the output of the
heuristic is to be measured can be used in the heuristic to guide the search. The heuristic may
then use the measuring formula with reference information as obtained during the calculation,
to guide the search, for stopping criteria, etc. The computational effort of calculating the
measure should always be included in the evaluation of the heuristic.
28
29
Acknowledgements
The work of the second author was supported by State Committee for Scientific Research,
KBN research grant no. 8T11C 013 13 and by CRIT 2 - Esprit Project no. 20288.
References
Andersen, K., A., Jrnsten, K., Lind, M. (1996). On bicriterion minimal spanning trees: An approximation,
Computers and Operations Research, Vol. 23, No. 12, pp. 1171-1182.
Assad, A.A. (1988). Modelling and implementation issues in vehicle routing. In: B.L. Golden, A.A. Assad
(eds.): Vehicle routing: Methods and Studies, North-Holland, Amsterdam, 7-46.
Cerny V. (1982). A thermodynamical approach to the travelling salesman problem: An efficient simulation
algorithm, Report, Comenius University, Bratislava, Czechoslovakia.
Czyzak P., Jaszkiewicz A. (1996). A multiobjective metaheuristic approach to the localization of a chain of
petrol stations by the capital budgeting model. Control and Cybernetics, 25, 1, 177-187.
Czyzak P., Jaszkiewicz A. (1998). Pareto simulated annealing - a metaheuristic technique for multipleobjective combinatorial optimization. Journal of Multi-Criteria Decision Analysis, 7, 34-47.
Daniels R.L. (1992). Analytical evaluation of multi-criteria heuristics. Management Science, 38, 4, 501-513.
Dasgupta D., Michalewicz Z. (1997). Evolutionary Algorithms in Engineering Applications. Springer Verlag.
Fonseca C.M., Fleming P.J. (1993). Genetic algorithms for multi-objective optimization: Formulation,
discussion and generalization. In: S. Forrest (ed.), Genetic algorithms: Proceedings of the Fifth
International Conference, Morgan Kaufmann, San Mateo, CA, 141-153.
Fonseca C.M., Fleming P.J. (1995). An overview of evolutionary algorithms in multiobjective optimization.
Evolutionary Computation, 3, 1, 1-16.
Fonseca C.M., Fleming P.J. (1996). On the performance assessment and comparison of stochastic
multiobjective optimizers. Lecture Notes in Computer Science, 1141, Springer Verlag, 584-593
Gandibleux, X., Mezdaoui N., Frville A. (1996). A tabu search procedure to solve multiobjective
combinatorial optimization problems, to appear in Proceedings volume of MOPGP 96, R. Caballero and R.
Steuer (eds.), Springer-Verlag.
Glover F. (1989). Tabu Search - Part I, ORSA J. Comput. 1, 190-206.
Goldberg D.E. (1988). Genetic Algorithms in Search, Optimization, and Machine Learning, Addison-Wesley,
Reading, Massachusetts.
Habenicht, W. (1982). Quad trees, A data structure for discrete vector optimization problems, Lecture notes in
economics and mathematical systems, 209, 136-145.
Hansen, M (1997). Tabu search for multiobjective optimization: MOTS, presented at the 13th MCDM
conference, Cape Town, South Africa, January 6-10
Horn. J., Nafpliotis N. (1994). A niched Pareto genetic algorithm for multiobjective optimization. Proceedings
of the First IEEE Conference on Evolutionary Computation, IEEE World Congress on Computational
Intelligence, vol. 1, IEEE, New York, 82-87.
Kirkpatrick S., Gelatt C.D., Vecchi M.P. (1983). Optimization by simulated annealing, Science, 220, 671-680.
Korhonen P., Salo S., Steuer R.E. (1997). A heuristic for estimating nadir criterion values in multiple objective
linear programming. Operations Research, 45, 5, 751-757.
Laarhoven P.J.M., Aarts E.H.L. (1987). Simulated Annealing: Theory and Applications, Reidel, Dordrecht.
Landsowne Z. F. (1996). Ordinal ranking methods for multicriterion decision making. Naval Research
Logistics, 43, 613-627.
Morita H., Katoh N. (to appear). A genetic algorithm for bicriteria scheduling problems. Journal of Heuristics.
Rosenthal R.E. (1985), Principles of multiobjective optimization. Decision Sciences, 16, 133-152.
Serafini P. (1987). Some considerations about computational complexity for multiobjective combinatorial
problems, LNEMS, 294, 222-232, Springer Verlag.
30
Serafini P. (1994). Simulated annealing for multiple objective optimization problems. In: Tzeng G.H., Wang
H.F., Wen V.P., Yu P.L. (eds), Multiple Criteria Decision Making. Expand and Enrich the Domains of
Thinking and Application, Springer Verlag, 283-292.
Silver, E.A., Vidal R.V.V., de Werra D. (1980). A tutorial on heuristic methods, European Journal of
Operation Research, 5, 153-162.
Sun, M., Steuer R.E. (1996). InterQuad: An interactive quad tree based procedure for solving the discrete
alternative multiple criteria problem, European Journal of Operational Research, 89, 462-472.
Steuer R.E. (1986). Multiple Criteria Optimization - Theory, Computation and Application, Wiley, New York.
Schaffer J.D. (1985). Multiple objective optimization with vector evaluated genetic algorithms. In: J.J.
Grefenstette (ed.), Genetic Algorithms and Their Applications: Proceedings of the Third International
Conference on Genetic Algorithms, Lawrence Erlbaum, Hillsdale, NJ, 93-100.
Slowinski R. (1989). Multiobjective project scheduling under multiple-category resource constraint. In:
Slowinski R., Weglarz J. (eds.), Advances in Project Scheduling, Elsevier, Amsterdam, 151-167.
Srinivas N., Deb K. (1994). Multiobjective optimization using nondominated sorting in genetic algorithms.
Evolutionary Computation, 2, 2, 221-248.
Tuyttens D., Teghem J., Fortemps Ph, Van Nieuwenhuyse K. (to appear). Performance of the MOSA method
for the bicriteria assignment problem. Journal of Heuristics.
Ulungu E.L. (1993). Optimisation combinatoire multicritre: dtermination de lensemble des solutions
efficaces et mthodes interactives, Thse de doctorat, Universit de Mons-Hainaut (Belgim).
Ulungu E.L., Teghem J. (1994). Multi-ojective Combinatorial Optimization Problems: A Survey. Journal of
Multi-Criteria Decision Analysis 3, 83-101.
Ulungu E.L., Teghem J., Fortemps Ph. (1995). Heuristics for multi-objective combinatorial optimization by
simulated annealing. Multiple Criteria Decision Making: Theory and Applications. Proceedings of the 6th
National Conference on Multiple Criteria Decision Making, Beijing (China), 228-238.
Viana, A. (1997). Applicao de Meta-heursticas Multiobjectivo ao Problema de Sequenciamento de
Actividades com Restries de Recursos, Master thesis at the Faculty of Engineering, University of Porto,
Portugal (In Portuguese).
Weistroffer H.R. (1985). Careful usage of pessimistic values is needed in multiple objectives optimization.
Operations Research Letters, 4, 1, 23-25.
White D.J. (1990). A bibliography on the applications of mathematical programming multiple-objective
methods. J. Opl Res. Soc., 41, 669-691.