0% found this document useful (0 votes)
26 views34 pages

Pesquisa Operacional: A New Method For Decision Making in Multi-Objective Optimization Problems

This document presents a new method for solving multi-objective optimization problems. The method divides the objective functions into two groups: a control function and performance functions. It then optimizes the control function over the Pareto set that would result from optimizing just the performance functions. This yields a single optimal solution without explicitly calculating the entire Pareto front. The method is tested on problems with increasing complexity. It differs from traditional methods that combine objectives into a single function by not requiring knowledge of designer preferences between objectives.

Uploaded by

John
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
26 views34 pages

Pesquisa Operacional: A New Method For Decision Making in Multi-Objective Optimization Problems

This document presents a new method for solving multi-objective optimization problems. The method divides the objective functions into two groups: a control function and performance functions. It then optimizes the control function over the Pareto set that would result from optimizing just the performance functions. This yields a single optimal solution without explicitly calculating the entire Pareto front. The method is tested on problems with increasing complexity. It differs from traditional methods that combine objectives into a single function by not requiring knowledge of designer preferences between objectives.

Uploaded by

John
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 34

Pesquisa Operacional Services on Demand

Print version ISSN 0101-7438


Journal
Pesqui. Oper. vol.32 no.2 Rio de Janeiro May/Aug. 2012 Epub June 21, 2012
SciELO Analytics
https://doi.org/10.1590/S0101-74382012005000014
Google Scholar H5M5 (2019)

A new method for decision making in multi- Article


objective optimization problems English (pdf)

Article in xml format

How to cite this article

SciELO Analytics
Oscar Brito AugustoI, *; Fouad BennisII; Stephane CaroIII
Curriculum ScienTI
IEscola Politécnica da Universidade de São Paulo. E-mail:
Automatic translation
[email protected]
IIÉcole Centrale de Nantes, Institut de Recherche en Communications et
Indicators
Cybernétique de Nantes. E-mail: [email protected]
IIIInstitut de Recherche en Communications et Cybernétique de Nantes. E- Related links

mail: [email protected] Share

More

More

Permalink
ABSTRACT

Many engineering sectors are challenged by multi-objective optimization


problems. Even if the idea behind these problems is simple and well established, the implementation of any
procedure to solve them is not a trivial task. The use of evolutionary algorithms to find candidate solutions is
widespread. Usually they supply a discrete picture of the non-dominated solutions, a Pareto set. Although it is very
interesting to know the non-dominated solutions, an additional criterion is needed to select one solution to be
deployed. To better support the design process, this paper presents a new method of solving non-linear multi-
objective optimization problems by adding a control function that will guide the optimization process over the
Pareto set that does not need to be found explicitly. The proposed methodology differs from the classical methods
that combine the objective functions in a single scale, and is based on a unique run of non-linear single-objective
optimizers.

Keywords: multi-objective optimization, multi-attribute decision making, engineering design optimization.

1 INTRODUCTION
Many engineering design problems are multi-objective in nature as they often involve more than one design goal to
be optimized. These design goals potentially impose conflicting requirements on the technical and economic
performance of a given system design. To study the trade-offs between these conflicting design objectives and to
explore design options, an optimization problem with multiple objectives has to be formulated. To facilitate the
exploration of different multi-objective formulations, we can incorporate Pareto optimality concepts into
optimization algorithms that require the designer's involvement as a decision maker.

Edgeworth (1881) was the first academic to define an optimization problem involving multiple criteria. The
problem, elaborated in the context of two consumers called P and π can be defined as: "it is required to find a
point (x, y) such that, in whatever direction we take an infinitely step, P and π do not increase together, but, while
one increases, the other decreases".

A few years later, in 1896, Pareto (1971) establishes the optimum for n consumers: "we will say that members of a
collectivity enjoy maximum ophelimity in a certain position when it is impossible to find a way of moving from that
position very slightly in such a manner that the ophelimity enjoyed by each of the individuals of that collectivity
increases or decreases. That is to say, any small displacement in departing from that position necessarily has the
effect of increasing the ophelimity which certain individuals enjoy, and decreasing that which other enjoy, being
agreeable to some and disagreeable to others".

Since then, many researchers have been dedicated to developing methods to solve this kind of problem.
Interestingly, solutions for problems with multiple objectives, also called multi-criteria optimization or vector
optimization, are treated as Pareto optimal solutions or Pareto front, although as Stadler observed (1988), they
should be treated as Edgeworth-Pareto solutions.

On the other hand, thanks to the evolution of computers, the optimization of large scale problems has become a
common task in engineering design processes. Computers, ever faster, with higher capacity of data storage and
Internet connection capability, are revolutionizing the traditional ways of developing engineering designs. Currently,
engineers can rely on a wide range of design alternatives and optimization methods that allow systematic choices
between alternatives when they are based on measurable criteria. Used properly, these methods can improve or
even generate the final solution of a design, as shown in Giassi et al. (2004), which held multi-objective
optimization of a vessel hull shape with collaborative design method where the computational tools for the analysis
of alternatives used in the cycles of optimization were distributed in three different sites. With this approach they
demonstrate that if the optimization is conducted in a distributed way it becomes a powerful tool for the concurrent
design.

Rao et al. (2007) designed a disk for high-pressure turbine aircraft engines, treating the problem as multi-objective
by considering the manufacturing cost and fatigue life as functions to be optimized. Bouyer et al. (2007) performed
the optimal design of a mechanical transmission system, Slide-O-Cam, which turns rotary into translation motion
and replaces the conventional rack-pinion systems, a problem with four objective functions. In both instances, the
optimal solutions were presented by their Pareto fronts.

Plenty of methods for solving design problems with more than one objective are found in scientific and technical
literature. However, since the multi-objective problems have no single solution but a set of points called optimal
solutions, the question that arises is how to choose one alternative among those contained in the set of solutions.
The answer is not simple unless another criterion is added to help choose the alternative that should be deployed.

In this paper we propose a different alternative to solve the problem of optimizing a set of objective functions that
will be divided into two groups: one containing only one function that will be called control function and one in
which the p functions will be called the performance functions group. With this strategy, the optimization problem
is formulated as single-objective, optimizing the control function over the Pareto set that would result from the
optimization problem constituted by the group of performance functions if this multi-objective optimization problem
had been solved. However, since this task is not performed, the Pareto front is not an output of the optimization
process and the computational task is significantly smaller than the traditional multi-objective optimization
algorithms.

Applying this strategy to a multi-objective optimization problem requires the use of an algorithm that can solve a
single-objective problem and work with the functions involved. The result will be a single solution. However, the
methodology differs significantly from other classical algorithms as there is no need to know a value function
relating the objectives that articulates dedesigner preferences.

After this brief introduction, the remaining of the paper is organized into six sections. The first formulates a general
multi-objective optimization problem and defines the nature of optimal solutions from the Pareto perspective and
the necessary conditions to be met. The second presents a brief literature review on multi-objective optimization
problem algorithms. In the sequence, it presents the main contribution of this work, a new strategy to solve the
optimization problem with multiple objectives. Next, it goes on to the implementation of the strategy to solve three
problems with increasing levels of complexity, and finally presents the conclusions onthe work done.

Nomenclature
AIS artificial immune systems

AC ant colony optimization algorithm

DM decision maker

fi (X) ith objective function

fc (X) control function

fiº (X) minimum value of the ith objective function

fimax (X) maximum value of the ith objective function

tolf tolerance for equal values of rms in successive generations

f(X) objective functions vector

fp (X) performance functions vector

ideal point, vector with all objective functions' minima, in criterion


fo
space

GA genetic algorithm

gj (X) jth inequality constraint function

hi (X) ith equality constraint function

k number of objective functions

KKT Karush-Khun-Tucker

l number of equality constraint functions

m number of inequality constraint functions

MOOP multi-objective optimization problem

NSGA II non-dominated sorting genetic algorithm, version two

n dimension of the design space

p number of performance functions

PS particle swarm optimization algorithm

rmsn root mean square of all objective functions at the nth generation

k function or criterion space

n decision variables or design space

SA simulating annealing optimization algorithm

feasible region in the design space

VEGA vector evaluated genetic algorithm

xi ith decision variable

average of the ith decision variable included in the optimal set


X decision variable vector

X* non-dominated solution of a multi-objective optimization problem

Xinf , Xsup lower and upper bounds of the design space

Xextended extended vector for unknowns with X, α, λ and µ

αi weighting factor for ith objective function gradient in KKT condition

α vector of αis

ωi weighting factor for the ith objective function

ω i* normalizing weighting factor for the ith objective function

εj upper bound for the jth objective function in the ε-restricted method

λj weighting factor for jth inequality constraint gradient in KKT condition

λ vector of λjs

σi standard deviation for the ith decision variable included in the optimal
set

µi weighting factor for ith equality constraint gradient in KKT condition

µ vector of µis

∇ gradient operator

2 MULTI-OBJECTIVE OPTIMIZATION PROBLEM


Multi-objective optimization problems (MOOP) can be defined by the following equations:

where

f(X) = [f1, f2, f3, ..., fk]T : X → k

is a vector with the values of objective functions to be minimized. X is the vector containing the design variables,
also called decision variables, defined in the design space n. X
inf and Xsup are respectively the lower and upper
bounds of the design variables. gi (X) represents the i inequality constraint function and hj (X) the jth equality
th

constraint function. The three equations (1b)-(1d), define the region of feasible solutions, , in the design space
n. The constraints g (X) are of type "less than or equal" functions in view of the fact that "greater or equal"
i
functions may be converted to the first type if they are multiplied by -1. Similarly, the problem is the "
minimization" of the functions fi (X), given that functions "maximization" can be transformed into the former by
multiplying them by -1.

2.1 Pareto optimal

The notion of "optimum" in solving multi-objective optimization problems is known as"Pareto optimal." A solution is
said to be Pareto optimal if there is no way to improve one objective without worsening at least one other, i.e., the
feasible point X* ∈ is Pareto optimal if there is no other feasible point X ∈ such that ∀i, j with i ≠ j, fi (X) = fi
(X*) with strict inequality in at least one condition, fj (X) < fj(X*).

Due to the conflicting nature of the objective functions, the Pareto optimal solutions are usually scattered in the
region , a consequence of not being able to minimize all the objective functions simultaneously. In solving the
optimization problem we obtain the Pareto set or the Pareto optimal solutions defined in the design space, and the
Pareto front, an image of the objective functions, in the criterion space, calculated over the set of optimal
solutions.

2.2 Necessary condition for Pareto optimality

In fact, optimizing multi-objective problems expressed by equations Eqs. (1a)-(1d) is of general character. The
equations represent the problem of single-objective optimization when k = 1. According to Miettinen (1998), such
as in single-objective optimization problems, the solution X* ∈ for the Pareto optimality must satisfy the Karush-
Kuhn-Tucker (KKT) condition, expressed as:

where αi is the weighting factor for the gradient of the ith objective function, calculated at the point X*, ∇fi (X*). λj
represents the weighting factor for the gradient of the jth inequality constraint function, ∇gj (X*), and is zero when
the constraint function associated is not active, i.e., gj (X*) > 0. µi represents the weighting factor for the gradient
of the ith equality constraint function, ∇hi (X*).

The set of Eqs. (2a) to (2e) form the necessary conditions for X* to be a Pareto optimal. As described by Miettinen
(1998), it is sufficient for the complete mapping of the Pareto front if the problem is convex and the objective
functions are continuously differentiable in the space. Otherwise, the solution will depend on additional
conditions, as shown by Marler and Aurora (2004).

3 LITERATURE REVIEW IN METHODS FOR SOLVING MOOP


Some researchers have attempted to classify methods for solving MOOP according various considerations. Hwang &
Masud (1979) and later Miettinen (1998) suggested the following four classes, depending on how the decision
maker (DM) articulates preferences: no-preference methods, a priori methods, a posteriori methods, and
interactive methods.

In no-preference articulation methods, the preferences of the DM are not taken into consideration. The problem
can be solved by a simple method and the solution obtained is presented to the DM which will accept or reject it.

In a priori preference articulation methods, the hopes and opinions of the DM are taken into consideration before
the solution process. Those methods require that the DM knows beforehand the priority of each objective
transforming the multi-objective problem in a single-objective problem where the function to be optimized is a
combination of objective functions.

In posteriori preference articulation methods no preferences of the DM are considered. After the Pareto set has
been generated, the DM chooses a solution from this set of alternatives.

In interactive preference articulation methods the DM preferences are continuously used during the search process
and are adjusted as the search continues.

Before explore the basics of some MOOP algorithms, it is opportune to define the ideal point
fo = [f1o, f2o, f3o,..., fkº]T ∈ k

being fiº the minimum of fi (X), X ∈ and i = 1, 2, ..., k. In general, fo is unattainable, i.e., such a point in the
criterion space does not map to a point in the design space.

In many cases it is advantageous to transform the original objective functions. This is especially true with
scalarization methods that involve a priori articulation of preferences. As the objective functions usually have
different scales, it is necessary to transform them in a non-dimensional. The relation

can be used provided that fiº ≠ 0. Such transformation results in a non-dimensional objective function with a lower
limit of one and unbounded upper limit.

Alternatively, a most robust approach would be

with fimax, the maximum value of fi (X) for X ∈ . This operation is referred as normalization as the transformed
function is bounded into the zero-one interval.

3.1 The weighted sum method

One of the most intuitive ways used to obtain a single unique solution for multi-objective optimization is the
weighted sum method. In this approach, the MOOP are converted into a scalar preference function using a linear
weighted sum function of the form,

For this method, the weights ωi reflect, a priori, the designer's preferences. It is simple, but the proper selection of
the weights may be a challenge itself.

The method can be used to find the Pareto set in many individual runs if the weights are consistently chosen for
each run. Nevertheless, varying the weights continuously may not necessarily result in an even distribution of
Pareto optimal points and a complete representation of the Pareto optimal set for non-convex problems as shown
in Miettinen (1998).

3.2 The weighted metric methods

Other means of combining multiple objectives into a single-objective are based on the weighted distance metrics.
Some DMs aim to find a feasible design that minimizes its distance from a pre-defined design as the representation
of a designer's overall preferences. Assuming that the pre-defined design is the ideal design, the weighted p-norm
as is expressed by

where,
is the normalizing weight.

The MOOP is written as

The parameter p may be chosen from 1 to infinite. With p = 2, Eq. (6) yields to the Euclidian distance metric.

A compromise solution, an alternative to the idea of Pareto optimality is a single point that minimizes the Euclidian
distance between the potential optimal point and the ideal point.

3.2.1 The min-max method

The min-max solution was initially proposed by Lightner and Director (1981). The method try to find a feasible
design that minimizes its distance from the ideal design as the representation of a designer's overall preferences.
The weighted ∞-norm is used as distance metric.

In Eq. (6), the limit of lp as p approaches to ∞ is

because the largest ωi* (fi - fiº) will dominate all others when taken to the infinite power.

The min-max problem is then defined by

Given different relative weights of objectives, the min-max method is capable of discovering efficient solutions of a
multi-objective problem whether the problem is convex or non-convex.

3.3 The goal programming

Goal programming, originally proposed by Charnes & Cooper (1977), is a technique which requires preference
information before any efficient solution are generated. In fact, the method requires a designer to set goals, bi for
all objectives, fi, that he wishes to achieve. It adopts the decision rule that the best compromise design should be
the one which minimizes the weighted sum of deviations from the set goals, , where di is the deviation
from the goal bi for the ith objective. To model the absolute values, di is split into positive and negative parts such
that di = di+ + di-, with di+ > 0, di- > 0 and di+ di- = 0. Deviations di- and di+ denote underachievement and
overachievement, respectively, where achievement implies that a goal has been reached. Similarly, the jth-
inequality or equality constraint function are treated like the objective functions setting the goals bj = 0. Then, the
MOOP is formulated as:

The method allows the designer to assign preemptive weights to objectives and to define different achievement
levels of the goals.
3.4 The ε-constraint method

Haimes et al. (1971) introduced the ε-constraint strategy that minimizes the single objective function fi (X). All
other objective functions are used to form additional constraints. The MOOP is formulated as:

The definition of the limits εj requires knowing a priori the designer's preference. A set of Pareto optimal solutions
can be obtained with a systematic variation of εj. However, improper selection of εj ∈ can result in a formulation
with no feasible solution.

3.5 Nature inspired metaheuristic algorithms

The methods for multi-objective optimization presented thus far have involved unique formulations that are solved
using standard optimization engines or single-objective optimization methods algorithms. With those methods, only
one Pareto optimal solution can be expected to be found in one simulation run of a classical algorithm and not all
Pareto optimal solution can be found by some algorithms in non-convex MOOP.

However, other approaches such some heuristics inspired in nature process can solve MOOP getting the Pareto set
directly.

3.5.1 Genetic algorithms

Genetic algorithm (GA) is a metaheuristic inspired by the Darwin evolutionist theory explaining the origin of
species. In nature, weak and unfit species within their environment are faced with extinction by natural selection.
The strong ones have greater opportunity to pass their genes to future generations via reproduction. In the long
run, species carrying the correct combination in their genes become dominant in their population. Sometimes,
during the slow process of evolution, random changes may occur in genes. If these changes provide additional
advantages in the challenge for survival, new species evolve from the old ones. Unsuccessful changes are
eliminated by natural selection.

The concept of GA was originally proposed by Holland (1974) for applications into the control theories and it was
quickly generalized to many different areas of engineering and sciences. The specific mechanics of the algorithm
involve the language of microbiology and, in developing new potential solutions, mimic genetic operations. A
population represents a group of potential solution points. A generation represents an algorithmic iteration. A
chromosome is comparable to a design point, and a gene is comparable to a component of the design vector. Given
a population of designs, three basic operations are applied: selection, crossover, and mutation. The selection
operator involves selecting design vectors, called parents, in the current generation, which are combined together,
by crossover, to form new chromosomes, called offspring. By iteratively applying the crossover operator, genes of
good chromosomes are expected to appear more frequently in the population, eventually leading to convergence to
an overall good solution. The mutation operator introduces random changes into characteristics of chromosomes.
Mutation reintroduces genetic diversity back into the population and assists the search escape from local optima.

Being a population-based approach, GA is well suited to solve multi-objective optimization problems finding a set of
multiple non-dominated solutions in a single run. The ability of GA to simultaneously search different regions of a
solution space makes it possible to find a diverse set of solutions for difficult problems with non-convex,
discontinuous, and multi-modal solutions spaces. In addition, most of multi-objective algorithms do not require the
user to prioritize, scale, or weigh objectives. Consequently, GA has been the most popular heuristic approach to
multi-objective design and optimization problems. Coello (2010) maintains an updated list with more than 5000
titles of publications involving different genetic algorithms.

The first multi-objective genetic algorithm, called vector evaluated genetic algorithm, or VEGA, was proposed by
Schaffer (1985). Afterwards, several multi-objective evolutionary algorithms were developed, as shown by Konak
et al. (2006), including the one that is used in this work to compare problems' results, the non-dominated sorted
genetic algorithm, or NSGA II, proposed by Deb et al. (2000).

In the NSGA II, before selection is performed, the population is ranked on the basis of non-domination: all non-
dominated individuals are classified into one category to provide an equal reproductive potential for these
individuals. Since individuals in the first front have the maximum fitness value, they always get more copies than
the rest of the population when the selection operator is applied. Additionally, the NSGA II estimates the density of
solutions surrounding a particular solution in the population by computing the average distance of two points on
either side of this point along each of the objectives of the problem. This value is called crowding distance. During
selection, the NSGA II uses a crowded-comparison operator which takes into consideration both the non-
domination rank of an individual in the population and its crowding distance. The non-dominated solutions are
preferred over dominated solutions, but between two solutions with the same non-domination rank, the one that
resides in the less crowded region is preferred. The NSGA II uses the elitist mechanism that consists of combining
the best parents with the best offspring obtained. Goldberg (1989) and Deb (2001) are excellent guides for GA
implementations.

3.5.2 Particle swarm

Particle swarm optimization algorithm (PS) was originally proposed by Kennedy & Eberhart (1995). It is a
population-based search algorithm based on the simulation of the social behavior of birds within a flock. Although
originally adopted for balancing weights in neural networks, PS soon became a very popular global optimizer.

There are two main distinctions between PS and GA. Genetic algorithms rely on three mechanisms in their
processing. In contrast, PS only relies on two mechanisms, since PS does not adopt an explicit selection operator.
The absence of a selection mechanism in PS is compensated by the use of leaders to guide the search. Such set of
leaders is usually stored in a different place in an external archive, with all non- dominated solutions found so far.
However, there is no notion of offspring generation in PS as with evolutionary algorithms. A second difference has
to do with the way in which the individuals are manipulated. PS uses an operator that sets the velocity of a particle
to a particular direction. This can be seen as a directional mutation operator in which the direction is defined by
both the particle's personal best and the global best of the swarm. If the direction of the personal best is similar to
the direction of the global best, the angle of potential directions will be small, whereas a larger angle will provide a
larger range of exploration. In contrast, evolutionary algorithms use a mutation operator that can set an individual
in any direction, although the relative probabilities for each direction may be different.

The main algorithm of PS is relatively simple, it only adopts one operator for creating new solutions and its
implementation is, therefore, straightforward. Its main drawback is how to control de size of the external archive
that implies in the computational time expended in the optimization process.

3.5.3 Simulated annealing

Simulated annealing optimization algorithm (SA) is based on an analogy of thermodynamics with the way metals
cool and anneal. If a liquid metal is cooled slowly, its atoms form a pure crystal corresponding to the state of
minimum energy for the metal. The metal reaches a state with higher energy if it is cooled quickly. Kirkpatrick et
al. (1983) and Cerny (1985) showed that a model for simulating the annealing of solids, proposed by Metropolis et
al. (1953), could be used for optimization of problems, where the objective function to be minimized corresponds
to the energy of states of the metal. The first multi-objective version of SA has been proposed by Serafini (1985,
1992). The algorithm of the method is almost the same as the algorithm of single objective SA. The slow
convergence rate for some optimization problems is the main drawback of SA, as mentioned by Suman & Kumar
(2006).

3.5.4 Ant Colony optimization algorithm

The ant colony optimization algorithm (AC) is inspired by the behavior of ants and other insects that live in a
colony that in spite of the simplicity of each individual, present a high level of social organization when observed
together. Some examples of ant colony's capabilities found in Dorigo et al. (1999) are: division of labor and task
allocation, cemetery organization and brood sorting, cooperative transport and finding the shortest path between
two or more locations, often between a food source and a nest.

The first AC algorithm developed was initially applied to the traveling salesman problem, Dorigo (1992). The
algorithm was based on the ant colony capability to find the shortest path between a food source and a nest. The
algorithm uses artificial ants that cooperate on finding solutions to the problem through communication mediated
by artificial pheromone trails.

While moving on the graph associated with the problem, artificial ants deposit pheromone on the edges traversed
marking a path that may be followed by other members of the colony, which then reinforce the pheromone on that
path. With this kind of communication, ants have their activities coordinated. This self-organizing behavior results
in a self-reinforcing process that leads to the formation of a path marked by high pheromone concentration, while
paths that are less used tend to have a diminishing pheromone level due to evaporation.

This concept can be applied to any combinatorial optimization problem for which a constructive heuristic can be
defined. The process of constructing solutions can be regarded as a walk on a construction graph where each edge
of the graph represent a possible step the ant can take. AC is essentially constructive, as ants generate solutions
by adding solution components, corresponding to the edges chosen, to an initially empty solution until the solution
is complete.

3.5.5 Artificial immune systems

The human immune system has as its main task the detection of the infectious foreign elements, called pathogens,
that attack us, and defend us from them, i.e., its main task is to keep our organism healthy. Examples of such
pathogens are bacteria and viruses. Any molecule that can be recognized by our immune system is called antigen.
Such antigens provoke a specific response from our immune system. Lymphocytes of types B and T are special
type of cells that play a major role in our immune system.

Upon detection of an antigen, the B cells that best recognize the antigen are cloned. Some of these cloned cells will
be differentiated into plasma cells, which are the most active antibodies secretors, while others will act as memory
cells. These cloned cells are subject to a high somatic mutation rate in order to increase their affinity level, i.e.,
their matching to the antigens. These mutations experienced by the clones are proportional to their affinity to the
antigen. The highest affinity cloned cells experiment the lowest mutation rates, whereas the lowest affinity cloned
cells have high mutation rates. Due to the random nature of this mutation process, some clones could be
dangerous to the body and are, therefore, eliminated by the immune system itself. Plasma cells are capable of
secreting only one type of antibody, which is relatively specific for the antigen. Antibodies play a key role in the
immune response, since they are capable of adhering to the antigens, in order to neutralize and eliminate them.

Once the antigens have been eliminated by the antibodies, the immune system must return to its normal
conditions, eliminating the in-excess cells. However, some cells remain in our blood stream acting as memory cells,
so that our immune system can 'remember' the antigens that have previously attacked it.

When the immune system is exposed again to the same type of antigen or a similar one, these memory cells are
activated, presenting a faster and perhaps improved response, which is called secondary response.

Based on the previous explanation of the way in which human immune system works, it can be say that, from a
computer science perspective, the immune system can be seen as a parallel and distributed adaptive system.
Clearly, the immune system is able to learn, it has memory, and is able of tasks such as associative retrieval of
information. These features make immune systems very robust, fault tolerant, dynamic and adaptive. All of these
properties can be emulated in a computer.

Artificial immune systems (AIS) are composed of the following basic elements: a) a representation for the
components of the system, e.g., binary strings, vectors of real numbers; b) a set of mechanisms to evaluate the
interaction of individuals with their environment and with each other; such an environment is normally simulated
through an affinity function, which is based on the objective functions in the case of optimization problems; c)
procedures of adaptation, that indicate the way in which the behavior of the system changes over time; these
procedure of adaptation consist of, for example, mutation operators.

The first direct use of the AIS for multi-objective optimization goes back to Yoo & Hajela (1999). In their work they
use a standard genetic algorithm where the immune principle of antibody-antigen affinity is employed to modify
the fitness value. In a first time the population is evaluated versus the problem objectives and different scalar
values are obtained by making reference to different weighting combinations. The best individual with respect to
each combination is identified as antigen. The rest of the population is the pool of antibodies. Then antibodies are
matched against antigens by the definition of a matching score. The best matching antibody fitness is added by
this matching score, evolving the population of antibodies to cover antigens.

4 DECISION MAKING IN MULTI-OBJECTIVE OPTIMIZATION PROBLEMS


Usually, when solving a MOOP, the DM expects, in general, to obtain its Pareto front and its Pareto set. Although it
is very interesting to know these non-dominated solutions, additional criteria are necessary to select a single
solution that will be deployed. Even though useful to better understand the inter-relationships between the
objectives, the Pareto front may become an obstacle since just one solution must be implemented. To make a
selection, the DM will have to use an additional criterion, be it subjective or not.

Then, why not formalize and include this additional criterion into the problem and make it a control function to find
the single and final solution to be deployed among those that belong to Pareto front?

Based on this idea, we propose a new methodology to formulate MOOP involving non-linear differentiable functions.
The method transforms the MOOP from multi-objective into single-objective that can be solved with the aid of any
traditional single-objective optimization engine suitable for the problem in focus.
4.1 Proposal of a new methodology for decision making in MOOP

In the proposed method, the objective functions are divided into two groups:

a) the control function group, or simply, control function, which contains only one function.

b) the performance functions group, which is made up of the functions that will provide the Pareto set.

Which function will be part of each group is a designer's choice and depends on his or her experience and
knowledge related to the design problem. The control function can be an additional objective function or it can be
elected among the problem objective functions.

The performance functions will have an important role in the process as they will providethe Pareto set as a
constraint over which the control function will search the solution that optimizes it.

Thus, the multi-objective optimization problem can be written as:

where fc (X) is the control function,

fp(X) = [f1, f2, ..., fp] : X → p

is the vector composed of the p objective functions in the performance functions group.

To apply the proposed methodology, the performance functions, is substituted by the KKT necessary condition, Eqs.
(2a)-(2e), in such way that the problem's final solution belongs to the Pareto front of the MOOP with the
performance functions only.

It should be noted that the weighting factors of the Eq. (2a), i.e.,

α = [α1, α2, ..., αp]T, λ = [λ1, λ2, ..., λm]T and µ = [µ1, µ2, ..., µl]T

are not known. As unknowns in the problem, they will be incorporated into the vector of design variables, defining
the extended vector of unknowns:

Finally, the problem is formulated as a single-objective optimization problem, with the control function, fc (X), to be
minimized and constrained by the conditions for obtaining the Pareto optimal solutions considering only the
performance functions, f1 (X), f2 (X),..., fp (X).

Mathematically, the optimization problem is formulated as:

Pareto set condition for the performance functions:


The proposed methodology differs from those methods exposed in the items 3.1 to 3.4. Those strategies require
that a complete decision-making structure of the problem is decided a priori. Although the proposed method is also
based on a priori decisions, just one function shall be added or isolated as a control function. The remaining
functions are incorporated into the formulation to ensure that the final solution is within the Pareto set obtained
from the minimization problem of these functions, even though the Pareto set and the Pareto front is not an explicit
outcome of the process.

However, it must be considered that, with the use of Eq. (2a), the methodology applies only to problems where the
functions involved are continuously differentiable. Furthermore, the solution obtained is a local optimum and it can
be a global optimum for convex Pareto fronts, since Eq. (2a) is a necessary condition for the existence of an
extreme point and not a sufficient condition.

5 APPLICATIONS
To validate the proposed multi-objective optimization method, it will be used to solve three examples with
increasing levels of complexity, namely the optimization problem of three quadratic functions, the design of a
cantilever beam and the conceptual design of a bulk carrier.

To solve the single-objective optimization problem originated by the proposed methodology, any algorithm that
works with optimization problems involving nonlinear functions and constraints can be used. Due to its accessibility,
the solver fmincon was used. It is a component of the Optimization Toolbox available in the application MatLab,
Matworks, version 7.1.

The NSGA II was used to compare results and to show that the results of the proposed method lay over the Pareto
set from the MOOP of the performance functions group. A minor adaption was done in the NSGA II in order to
interrupt the evolution if there is no significant difference between consecutive generations. Calling fij the value of
the ith objective function for the jth chromosome at nth generation of a population with pop chromosomes, the root
mean square of the objective functions for the population can be defined as

with fimax and fimin being the maximum and the minimum value, respectively, of the ith objective function
evaluated in the population. The alternative stop criterion was defined as

where tolf is a small number. Sometimes, for a defined tolf value, this criterion may cause a premature interruption
of the evolutionary process. To overcome this situation, the algorithm checks if it occurs in subsequent generations
before discontinue the evolution. As a default, three subsequent generations were adopted with the satisfaction of
Eq. (16) as the condition to interrupt the evolutionary process.
5.1 Minimization of three quadratic functions

Consider the minimization problem defined by:

Assume that the functions f1 (x1, x2) and f2(x1, x2) integrate the performance functions group and the function f3
(x1, x2) is the control function. Accordingly, the problem can be formulated as:

find Xextended = [x1, x2, α1, α2]T

that

The solution of this problem, (x1, x2, α1, α2) = (-1.1142, -0.0226, 0.3019, 0.6981), is shown in Table 1 and in the
Figure (1a). The point that minimizes the control function falls into the Pareto set resulting from the MOOP formed
by the performance functions. The Pareto sets shown in Figures (1a)-(1e) are not obtained by the methodology
and they were computed for comparison only by a suitable algorithm.
Figures (1b) and (1c) show the Pareto set and the corresponding Pareto front for the MOOP involving the
performance functions f1 (x1, x2) and f2 (x1, x2) obtained by the NSGA II. Although the points in the Pareto front
seem evenly distributed, the corresponding points in the Pareto set configure a rough approximation of the
solution. To get this set, a 50 chromosomes population was used with the alternative stop of evolution criteria, Eq.
(16), tolf = 0.001. With theseparameters the Pareto front stabilizes in 177 generations. The same parameter was
used in the other NSGA II examples shown in Table 1.

Figure (1d) shows the Pareto set for the same problem obtained by using the weighted sum approach, with 50
weight vectors chosen at random. This approach returns a well-matched Pareto set. Although the NSGA II is a
powerful algorithm when solving MOOPs, it focuses on the criterion space and consequently can generate rough
solutions in the design space as show the results in Figures (1b) and (1d).

Figure (1e) shows the results when another function is chosen to play the role of the control function. The resultant
points are distinct but they are over the corresponding Pareto set resulting from the MOOP of the corresponding
performance function group. Moreover, they belong to the Pareto set that results from the MOOP composed of the
three quadratic functions, shown in Figure (1f). This Pareto set was obtained by the weighted sum approach with
1,200 sets of weights chosen at random. Observing Figure (1f), where the Pareto set is a region with infinite non-
dominated points, the question that normally arises in the designer's mind is how to choose only one as a solution.
In general, as more functions are aggregated to a MOOP, the more extended the Pareto set on the design space. In
the example, with two functions, the Pareto set is a line. With three functions, the Pareto set is a plane surface.

In addition, comparisons for the computer performance between the algorithms are shown in Table 1. As there are
no other methodologies similar to the one proposed in this paper, the results were compared to two other
algorithms: the weighted sum approach and the NSGA II.For both, the process involved two phases: a) the search
of the non-dominated designs for the MOOP involving the performance functions group and b) the manual search in
the resultant set for the alternative that minimizes the control function. The number of function calls resulted from
the first part of the process. The computational performance measure adopted for all algorithms' comparison is the
number of objective functions calls. Although this measure is not perfect since both algorithms - the genetic
algorithm and weight sum approach - depend on the number of points to be used in the MOOP solver, it has at
least a qualitative value. For sake of comparison the computational times expended to achieve the final solution in
a 2 GHz dual processor computer with 3 Gb RAM are also shown.

Two other methods that return a single solution are shown in Table 1 and figure (1f), the compromise and the min-
max solutions. They are very fast, but they may not reflect the best compromise preferred by the DM.

Although the example is very simple, it shows how the proposed methodology can help decision making processes.
The problem solution falls within the performance functions non-dominated solutions and it is the one that
optimizes the control function.

In the quadratic functions example, the control function was chosen among the problem objective functions. In the
next application, the MOOP is defined as minimizing the mass and tip deflection of a cantilever beam, two
antagonist functions. With the proposed methodology these functions will compose the performance group.
Afterward, it will be included a third function, the manufacturing cost, that will play the role of the control function.

5.2 Design of a cantilever beam

As a second illustrative example, consider the design problem of a cantilever beam, adapted from Deb (2001). The
beam, prismatic and with a circular section, must support the weight P at its free end. The design variables are d
and L, the cross section diameter and beam length, respectively. Also consider two attributes to be minimized, the
beam mass and the beam tip deflection when subjected to the weight P. Moreover, the maximum bending stress
must be below of the material yield stress, σy, and the maximum tip displacement limited to δmax. A schematic
representation of the problem is shown in Figure 2.

It can be demonstrated that the two selected objectives conflict since the minimization of the mass will lead to
lower values of the pair (d, L) and minimization of the tip deflection will result in an increase in the cross sectional
diameter with a reduction in the beam length.

The multi-objective optimization problem can be formulated as:


Adopting the following parameters:

P = 1 kN; δmax = 5 mm; σy = 300 M Pa; E = 210 G Pa; ρ = 7850 kg/m3

the shape of mass, tip deflection, and bending stress surfaces are shown in Figures (3a) and (3b).

A clearer view of the antagonism of the objective functions can be obtained using a method to find the Pareto front
of multi-objective optimization problems. Approximations of the Pareto set and the Pareto front and are shown in
Figures (4a) and (4b), respectively. These results were found by generating 10,000 feasible designs at random and
then selecting the non-dominated ones.
Figure (4a) shows that all alternatives for the beam dimensions have length of 200 mm and cross section diameter
ranging from 19 mm to 50 mm, approximately. These results are expected because, for a fixed diameter, to reduce
the mass and the tip deflection the length decreases and the lower bound for this design variable is 200 mm, as
shown in Figure (3a), but an inverse relationship is observed considering the cross section diameter. Fixing the
beam length, the lower the diameter the higher the tip deflection and the lower the beam mass.

Similar results were obtained by the NSGA II algorithm, using a 50 chromosomes population which has evolved 58
generations, conditioned by the stop criteria, Eq. (16), with tolf = 0.001.
For comparison, Figures (4e) and (4f) show the results found by the weighted sum approach. Fifty random sets of
weights were used to reduce the normalized mass and deflection beam functions to a single scale.

Now suppose that the problem must include the beam manufacturing cost and that it should be minimized.
Additionally, suppose the cost function being defined by the function:

The contours of the cost function are shown in Figure (5a).

If the cost function is included as the third function in the MOOP, there is a growth and dispersion of Pareto optimal
solutions in the design space, as shown in Figure (5b). This spread of the non-dominated alternatives in the design
space makes decision making process even more difficult.

The proposed methodology can help the DM elect one alternative among all non-dominated, while preserving the
concept of multi-objective optimization. The MOOP of the design of a cantilever beam can be re-formulated as:

Among all non-dominated design alternatives for the cantilever beam, where it is searched to minimize
the beam tip deflection, δ(d, L), and minimize the beam mass, m(d, L), find the one that has the
minimum manufacturing cost.

Another interpretation for problem formulation would be: among the solutions that satisfy a technical or
engineering criterion, choose the one that has the lowest manufacturing cost.
Applying the proposed methodology, the MOOP can be rewritten as:

find Xextended = [x1, x2, α1, α2, λ1 . . . λ6]T

that

The fmincon was used to solve this single-objective optimization problem, which returns the following values

Note that the design variables (d, L) and the weighting factors (only λ5 is not zero, revealing that only the
constraint equation g5, Eq. (21k), is active) converged to a point on the Pareto optimal set for the MOOP involving
the performance functions only, as shown in Figure (4a). For the sake of illustration, the values of the
manufacturing cost calculated over the Pareto set obtained for the bi-objective optimization problem are shown in
Figure (5d). The minimum cost is located in the diameter axis at the neighborhood of 35 mm.

Table 2 shows a comparison of the methodologies used to find the single solution for the beam design. Although
the proposed methodology does not return the Pareto front it is very efficient to find the final solution with number
of function calls significantly smaller than the other two methods used for the same purpose.
To conclude, if the cost function is treated globally as the third function in MOOP, the set of Pareto optimal solutions
will no longer be a line, as shown in Figure (4a); instead it is spread over a surface, as shown in Figure (5b). In
fact, the greater the number of conflicting functions in a multi-objective optimization problem, the greater the
region in the space of the design variables that defines the Pareto optimal solutions. Consequently, the decision
making process to elect a single solution is more difficult. How do we choose a non-dominated alternative among
those present in Figure (5b)? The answer is impossible unless a fourth criterion is used to drive the choice. Rather,
in the present proposal, this additional criterion is added to the MOOP as a control function to guide a convenient
solution that satisfies all the constraints in a very efficient way, a solution that is non-dominated inside the group of
the performance functions, and is an appropriate optimal solution obtained by optimizing the control function.

5.3 Conceptual design of a bulk carrier

The third application of the developed methodology was in the conceptual design of a bulk carrier. The conceptual
design of a cargo vessel is not a trivial task. For decades, this problem has been handled in two ways, either by the
adapting a known design aiming to the new requirements, or by the aid of simplified mathematical models driven
by an optimization algorithm that enables obtaining the optimal solution based on technical or economic criteria
previously established.

This work explores the second alternative with the help of the mathematical model for conceptual design of bulk
carriers developed by Pratyush & Yang (1998) and presented in Appendix A. The model is made up of functions
that define the vessel attributes from which are drawn those that constrain the design space, those to be optimized
and those that characterize the technical and economic performance of the vessel and allow the evaluation of each
design alternative. Among them are the annual transportation cost, the annual transported cargo, the ship
lightweight, the ship initial cost and other functions of the vessel design variables such as length, width, depth,
draft, block coefficient and speed, respectively (L, B, D, h, CB, Vk). Pratyush & Yang (1998) proposed a general
problem with explicit bounds on the block coefficient that should range between 0.63 and 0.75, and the ship speed
that varies between 14 and 18 knots, and nine constraints in attributes, presented in Table A1 of Appendix A.

Parsons & Scott (2004) used the same model for comparing optimization methods, but they have altered the
explicit bounds on the design variables. They limited the ship length at 274.32 m (900 ft) and the minimum
deadweight at 25,000 ton, as shown in Table A2 of Appendix A. They also evaluated the problem by constraining
the decision space in order to design ships suitable to cross the Panama Canal, with beam and draft limited to
32.31 m and 11.71 m, respectively, as shown in Table A3 of Appendix A.

Using Parsons and Scott's second model, Hart & Vlahopoulou (2009) constrained the depth to the minimum and
the maximum of 13 m and 25 m, respectively, and extended the acceptable range for ship speed to the values of
11 knots and 20 knots, minimum and maximum respectively, as shown in Table A4 of Appendix A.

The bounds for the constraints feasibility lead to different results in optimization and the constraint bounds
imposed in Pratyush & Yang's (1998) original work may be associated with the mathematical model, based on data
reductions from operating ships and whose extrapolation may undermine not only the quality but even the validity
of the results obtained by using the model. In this paper was used the constraint bounds on design variables
proposed in Pratyush & Yang's (1998) original work. For practical reasons, the limits they have not mentioned were
adopted wide enough to not influence the optimization results. Accordingly, the ranges for the design variables are
described in Table A5 of Appendix A.

5.3.1 Conflicting objectives

Pratyush & Yang (1998) chose to minimize the annual transportation cost, maximize the amount of annual
transported cargo and minimize vessel lightweight. In this work, we replaced the vessel lightweight by the ship
initial cost. Although vessel lightweight and ship initial cost functions are related, the latter has a financial appeal
like the former two objective functions. Annual transported cargo is associated with the annual income, the annual
transportation cost with the annual expenses and the ship cost is related to the capital required for the ship
purchase.

The design alternative that optimizes the single-objective problem can be easily obtained by taking each objective
function separately. Any algorithm running with non-linear optimization problems and constraints can be used to
solve this single-objective optimization problem. The fmincon was used to get the results shown in Table 3.
The results indicate that the objective functions are conflicting as to maximize the annual cargo (AC); the optimum
design will be a large ship with the highest speed of 18 knots. The search reaches the upper bound for the total
deadweight, 500,000 t. The opposite occurs when it minimizes the ship cost (CS). In this case, all vessel
dimensions decrease along with the ship speed, which reaches the minimum of 14 knots with a slender shape
vessel with block coefficient of 0.63, assuring a low power main engine and consequently a lower main engine cost.
The search for even smaller vessels is constrained by the lower bound for the total deadweight of 3,000 tons. To
minimize the annual transportation cost (CT), the solution is situated between the previous two.

5.3.2 Solutions for the MOOP

A clearer view of the antagonism of the objective functions in this problem can be obtained using an algorithm to
find the Pareto front of multi-objective optimization problems. Figure 6 shows 10,000 feasible ship designs and
highlights the non-dominated ones. To get these points, more than 12 million random design alternatives were
generated and analyzed, taking 1097 seconds of computational time. Table 4 shows the design variables statistics
for the feasible points. Although the problem shows a narrow surface in the criterion space, the design variables
are well distributed over the design space.
With the aid of Figure 6, is possible to presume how the Pareto front of this problem should be, but, on the other
hand, the random walk is not adequate to drive the search for the Pareto front, while the number of function calls
becomes prohibitive.

In order to compare the results of the proposed methodology with other methods, the weighted sum approach and
the genetic algorithm were used to do the bulk carrier design optimization task.

Figure 7 and Table 5 show the results of the weighted sum approach algorithm with 100 sets of random weights. As
fmincon fails in find the solution for several weights sets, the process were restarted from 5 different starting
points for each set single run, being chosen the best result as the resultant point. Even though points were
discarded, because they have shown to be dominated solutions.
Better results are obtained from the NSGA II algorithm according to the points shown in Figure 8. These results
were achieved for a population of 100 chromosomes, which has evolved 2000 generations. Table 6 shows the
statistics for the non-dominated designs.
The statistics for the design variables indicate a reasonable dispersion of the values. With the set of non-dominated
solutions, shown in Figure 8, the question that often arises for the DM is: which design alternative should be
implemented?

5.3.3 Applying the proposed methodology

Applying the proposed methodology one attribute function can be chosen as the control function and optimized in
the Pareto front generated by the optimization of the performance functions group. Consider the previous
functions, the annual transportation cost, the annual cargo and the ship cost, incorporating the performance group
and the voyage cost as the control function. Accordingly, the multi-objective optimization problem can be defined
as:

among all the non-dominated solutions that would be obtained by resolving the MOOP for the bulk
carrier design, namely, minimizing the annual transportation cost (CT), minimizing the ship cost (CS)
and maximizing the annual cargo (CA), find the solution that minimizes the voyage cost (CV).

The mathematical formulation of this optimization problem is shown in detail in Appendix B. Note that the problem
still is multi-objective, with four objective functions and no value function involved. However, with the use of the
proposed approach, it can be formulated and solved in a single passage of any algorithm dedicated to single-
objective optimization with inequality and equality constraints and able to work with non-linear functions. In this
manner formulated, the problem will have 30 unknowns (six design variables of the original problem, plus three
weighting factors αk related to the gradients of the performance functions and more 21 weighting factors λj related
to the explicit and implicit constraints).

Table 7 shows the design alternative with some ship attributes obtained by using the fmincon algorithm.
Additionally, the solution is highlighted in Figure 9 over the Pareto front of performance functions obtained thanks
to the NSGA II.
Table 8 compares the results obtained by the proposed methodology with those obtained by the weighted sum
approach and the NSGA II. The proposed methodology is very efficient in finding the optimized design as the
number the function calls is significantly smaller.

It can be observed that the use of the proposed methodology leads to a solution that satisfies all the constraints
and provides a good technical performance and its result is optimal according to the control function.
6 CONCLUSION
Most engineering design problems are multi-objective and the cases where the objective functions do not conflict
are rare. To solve these kinds of problems, many researchers developed methods to search for the solution of
multi-objective optimization without simplifying the problem to single-objective and having to decide a priori how
to group the objective functions into a single scale. The evolutionary methods are frequently used to locate the set
of solutions of multi-objective problems. These algorithms provide a discrete picture of the Pareto front in the
criterion space. It was observed that the greater the number of objective functions, the more scattered the set of
non-dominated solutions is in the decision space and the harder it is for the DM to choose an alternative to be
deployed.

This paper proposes a new methodology to solve multi-objective optimization problems in which one objective
function is proposed or isolated and is treated as a control function that will drive the decision-making process. The
other objective functions will form the performance functions group. Through this strategy, a single-objective
optimization problem is formulated, in which the control function will be optimized over the Pareto set that would
result from the optimization problem established by the performance functions, if this problem was previously
solved.

Then the resultant single-objective optimization problem can be solved by any classical standard optimization
engines, the limitations will be those present in the method used, such as the nature of the functions, whether they
are linear or nonlinear, or continuous, regarding the presence or not of inequality and equality constraints.

With the proposed method a minimum knowledge is needed a priori. There is no need to know a value function
relation to the objectives before starting to solve the problem as other a priori articulation of preferences methods.

As drawbacks, the method doesn't provide a set of non-dominated solutions and as in many other single-objective
optimization algorithms, the solution converges to a local optimum since the conditions of Karush-Kuhn-Tucker,
chief pillar of the proposed methodology, are the necessary conditions but not sufficient to ensure that the solution
is in the "global" Pareto front. If the problem is convex and the functions involved are regular, the local solution
also will be a global solution, ensuring a unique result for the problem. Another disadvantage, the functions
involved in the problem shall be continuously differentiable.

Although these conditions are limiting, the proposed methodology is very efficient in solving engineering design
problems, as demonstrated by the examples solved with its use.

REFERENCES
[1] BOUYER E, CARO S, CHABLAT D & ANGELES, J. 2007. The Multiobjective Optimization of a Prismatic Drive, in
Proceedings of ASME Design Engineering Technical Conferences, September 4-7, Las Vegas, Nevada, U.S.
[ Links ]

[2] CHARNES A & COOPER WW. 1977. Goal programming and multiple objective optimisation - Part 1. European
Journal of Operational Research, 1: 39-54. [ Links ]

[3] CHINCHULUUN A & PARDALOS PM. 2007. A survey of recent developments in multiobjective optimization. Ann.
Oper. Res., 154: 29-50. [ Links ]

[4] COELLO CAC. 2000. An Updated Survey of GA-Based Multiobjective Optimization Techniques. ACM Computing
Surveys (CSUR), 32(2). [ Links ]

[5] COELLO CAC. 2010. List of References on Evolutionary Multiobjective Optimization, in URL
<http://www.lania.mx/~ccoello/EMOO/EMOObib.html> , April 24, 2010. [ Links ]

[6] DEB K, AGRAWAL S, PRATAB A & MEYARIVAN T. 2000. A Fast Elitist Non-Dominated Sorting Genetic Algorithm
for Multi-Objective Optimization, NSGA, KanGAL report 200001. [ Links ]

[7] DEB K. 2001. Multi-objective Objective Optimization using Evolutionary Algorithms, Wiley and Sons Ltd.
[ Links ]

[8] EDGEWORTH FY. 1881. Mathematical Psychic: an essay on the application of mathematics to the moral
sciences, Editor P. Kegan, London, England. [ Links ]

[9] GIASSI A, BENNIS F & MAISONNEUVE JJ. 2004. Multi-Objective Optimization Using Asynchronous Distributed
Applications. J. Mech. Des., 26: 767. [ Links ]
[10] GOLDBERG D. 1989. Genetic Algorithms in Search and Machine Learning, Reading, Addison Wesley.
[ Links ]

[11] HAIMES YY, LASDON LS & WISMER DA. 1971. On a bicriterion formulation of the problems of integrated
system identification and system optimization. IEEE Trans. Syst. Man Cybern. SMC-1, 296-297. [ Links ]

[12] HART CG & VLAHOPOULOS N. 2009. An Integrated Multidisciplinary Particle Swarm Optimization Approach to
Conceptual Ship Design, Struct. Multidisc. Optim. DOI 10.1007/s00158-009-0414-0. Springer-Verlag.
[ Links ]

[13] HOLLAND J. 1975. Adaptation in Natural and Artificial Systems. University of Michigan Press. [ Links ]

[14] HWANG CLM, MASUD AS, in collaboration with Paidy SR & Yoon K. 1979. Multiple objective decision making,
methods and applications: a state-of-the-art survey. In: BECKMANN M AND KUNZI HP (eds.). Lecture Notes in
Economics and Mathematical Systems, 164. Berlin: Springer-Verlag. [ Links ]

[15] KENNEDY J & EBERHART RC. 1995. Particle swarm optimization. In: Proceedings of the IEEE International
Conference on Neural Networks, Piscataway, New Jersey, 1942-1948. [ Links ]

[16] KIRKPATRICK S, GELATT JR CD & VECCHI MP. 1983. Optimization by simulated annealing.Science, 220: 671-
680. [ Links ]

[17] KONAK A, COIT DW & SMITH AE. 2006. Multi-objective optimization using genetic algorithms: A tutorial.
Reliability Engineering and System Safety, 91: 992-1007. [ Links ]

[18] LIGHTNER MR & DIRECTOR SW. 1981. Multiple criterion optimisation for the design of electronic circuits. IEEE
Transactions on Circuits and Systems, 28(3): 169-179. [ Links ]

[19] MARLER RT & ARORA JS. 2004. Survey of Multi-objective Optimization Methods for Engineering. Structural
Multidisciplinary Optimization, 26: 369-395. [ Links ]

[20] MIETTINEN KM. 1998. Nonlinear Multiobjective Optimization, Springer-Verlag. [ Links ]

[21] OSYCZKA A. 1984. Multicriterion Optimiration in Engineering with FORTRAN Programs, Ellis Horwood Ltd,
Chichester, UK. [ Links ]

[22] PARETO V. 1971. Manual of Political Economy, Translated by Schiwier, A.S. from French Edition of 1927,
Augustus M. Kelley Publishers, New York. [ Links ]

[23] PARSONS MG & SCOTT RL. 2004. Formulation of Multicriterion Design Optimization Problems for Solution with
Scalar Numerical Optimization Methods. J. Ship Res., 48(1): 61-76. [ Links ]

[24] PARSONS MG. 2004. Parametric Ship Design, in Ship Design and Construction. Thomas Lamb Editor, Vol. 1,
SNAME, 11, 11.1-11.50. [ Links ]

[25] PRATYUSH S & YANG JB. 1998. Multiple Criteria Decision Support in Engineering Design, Springer-Verlag.
[ Links ]

[26] RAO AR, SCANLAN JP & KEANE AJ. 2007. Applying Multiobjective Cost and Weight Optimization to the Initial
Design of Turbine Disks. J. Mech. Des., 129: 1303. [ Links ]

[27] SCHAFFER JD. 1985. Multiple Objective Optimization with Vector Evaluated Genetic Algorithms, in Proceedings
of the first International Conference on Genetic Algorithms and their Applications, 93-100, Lawrence Erlgaunn,
Pittsburgh, PA. [ Links ]

[28] SERAFINI P. 1994. Simulated annealing for multiple objective optimization problems. In: TZENG GH, WANG
HF, WEN VP AND YU PL (eds.). Multiple Criteria Decision Making. Expand and Enrich the Domains of Thinking and
Application. Springer- Verlag, Berlin, 283-292. [ Links ]

[29] SUMAN B & KUMAR P. 2006. A survey of simulated annealing as a tool for singleand multiobjective
optimization. Journal of the Operational Research Society, 57: 1143-1160. [ Links ]

[30] STADLER W. 1988. Applications of Multicriteria Optimization in Engineering and the Sciences, 1: 1-24, Editor
Stadler W., Editor Plenum Press, New York. [ Links ]

[31] XUEBIN L. 2009. Multiobjective Optimization and Multiattribute Decision Making Study of Ship's Principal
Parameters in Conceptual Design. J. Ship Res., 53(2): 83-92. [ Links ]
[32] ZADEH LA. 1963. Optimality and non-scalar-valued performance criteria. IEEE Trans. Autom.Control AC-8, 59-
60. [ Links ]

Received June 21, 2010


Accepted March 12, 2012

* Corresponding author

APPENDIX A: CONCEPTUAL MODEL FOR BULK CARRIERS

Design Variables (6)

Length (L) (m)

Beam (B) (m)

Draft (T) (m)

Depth (D) (m)

Block coefficient (CB) (m)

Velocity (Vk) (knots)

Objectives (3)

minimize Ship costs (CS) (£)

minimize Transportation cost (CT) (£/T)

maximize Annual cargo (AC) (t)

Ship attributes

Steel weight (ws) 0.034L1.7 B0.7 D0.4 CB0.5 (t)

Outfit weight (wo) 1.00.8 B0.6 D0.3 CB0.1 (t)

a 4977.06 CB2- 8105.61 CB + 4456.51

b -10847.2 CB2- 12817 CB - 6960.32

Salt water specific weight (γ) 1.025 (t/m3)

Gravity (g) 9.81 (m/s2)

Displacement (Δ) γLBTCB (t)

Froud number (Fn) V/ , V = 0.5144 Vk

Power (P) /(a + bFn) (kW)


Machinery weight (wm) 0.17 P0.9 (t)

Ship lightweight (wls) w s + wo + wm (t)

Deadweight (d wt) Δ - wls (t)

Fuel rate (fR) 190 (g/(kWh))

Daily fuel consumption (DFC) 24fR 10-6 P + 0.2 (t/day)

Round trip miles (RTM) 5000 (M)

Sea days (DS) RTM/(24Vk)

Fuel price (FP) 100 (£/t)

Fuel cost (CF) 1.05DFC DSFP (£)

Port cost (CP) 6.3 dwt0.8 (£)

Fuel carried (FC) DFC (DS + 5) (t)

Misc. deadweight (d wtm) 2.0 d wt0.5 (t)

Cargo deadweight (d wtc) d wt - FC - d wtm (t)

Handling rate (HR) 8000 (t/day)

Port days (DP) 2(d wtC/HR +0.5)

Round trips per year (RTPY) 350/(DS + DP)

• Voyage costs (CV) (CF + CP) RTPY (£)

• Ship cost (CS) 1.3(2000ws0.85 + 3500wo + 2400P0.8) (£)

Capital cost (CC) 0.2 CS (£)

Running cost (CR) 40000 d wt0.3 (£)

Annual cost (CA) CC + CR + CV (£)

• Annual cargo (AC) d wtcRTPY (t/year)

• Transportation cost (CT) CA/AC (£/t)

Vertical center of buoyance (K


0.53T (m)
B)

Metacentric radius (B MT) (0.085CB - 0.002) B2/(TCB) (m)

Vertical center of gravity (KG) 1.0 + 0.52D (m)

Constraints
APPENDIX B: BULK CARRIERS DESIGN MODEL WITH THE PROPOSED
METHODOLOGY
To apply the proposed methodology is convenient to define the following entities:

a) The control function:

voyage cost: Cv

b) The performance functions: fp = [CS, CT, AC]T

ship cost: CS

transportation cost: CT

annual cargo: AC

c) The implicit constraints: gc = [g1, g2, ..., g9]T

g1 = 6 - L/B g6 = 0.07B - KB - BMT - KG

g2 = L/D - 15 g7 = d wt - 500000

g3 = L/T - 19 g8 = 3000 - d wt
g4 = T - 0.45d wt0.31 g9 = Fn - 0.32

g5 = T - 0.7D + 0.7

d) The explicit lower bound constraints: ginf = Xinf - X = [g10, g11, ..., g15]T

g10 = 60 - L g13 = 3 - T

g11 = 10 - B g14 = 0.63 - CB

g12 = 4 - D g15 = 14 - Vk

e) The explicit upper bound constraints: gsup = X - Xsup = [g16, g17, ..., g21]T

g16 = L - 600 g19 = T - 30

g17 = B - 100 g20 = CB - 0.75

g18 = D - 40 g21 = Vk - 18

f) The (unknown) design variables:

X = [L, B, D, T, CB, Vk]T

g) The (unknown) performance functions gradient multipliers:

α = [α1, α2, α3]T

h) The (unknown) implicit constraints gradient multipliers:

λc = [λ1, λ2, ..., λ9]T

i) The (unknown) explicit lower bound constraints gradient multipliers:

λinf = [λ10, λ11, ..., λ15]T

j) The (unknown) explicit upper bound constraints gradient multipliers:

λsup = [λ16, λ17, ..., λ21]T

k) The gradient vector:

Let the Karush-Kuhn-Tucker condition be defined by:

F(X, α, λc, λinf, λsup) =∇fkTα+ ∇gcTλc + ∇gTinfλinf +∇gTsupλsup

Then, with the proposed methodology, the multi-objective optimization problem involving the bulk carrier design
can be rewritten as the following single-objective optimization problem:
This single-objective optimization problem was solved by fmincon which returns the following values:

X = [98.78, 12.76, 6.59, 5.22, 0.63, 14.00]T

α = [0.00, 0.00, 1.00]T

λc = [0.40, 0.02, 0.03, 0.13, 0.81, 0.01, 0.00, 0.01, 0.97]T

λinf = [0.00, 0.00, 0.00, 0.02, 1.00, 0.46]T

λsup = [0.00, 0.00, 0.00, 0.00, 0.00, 0.01]T

All the contents of this journal, except where otherwise noted, is licensed under a Creative Commons Attribution License

Rua Mayrink Veiga, 32 - sala 601 - Centro


20090-050 Rio de Janeiro RJ - Brasil
Tel.: +55 21 2263-0499
Fax: +55 21 2263-0501

[email protected]

You might also like