0% found this document useful (0 votes)
48 views

ACO Algorithms For The Traveling Salesman Problem: IRIDIA, Universit e Libre de Bruxelles, Belgium

The document discusses ACO (Ant Colony Optimization) algorithms for solving the Traveling Salesman Problem (TSP). It provides an overview of the available ACO algorithms for the TSP, describing how ACO approaches can be applied to the problem. The key aspects covered are: (1) how ants construct tours probabilistically based on pheromone trails and heuristic information; (2) how the pheromone trails are updated after construction; and (3) how the best performing ACO algorithms for the TSP improve tours using local search methods in a hybrid approach.

Uploaded by

Sait Demir
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
48 views

ACO Algorithms For The Traveling Salesman Problem: IRIDIA, Universit e Libre de Bruxelles, Belgium

The document discusses ACO (Ant Colony Optimization) algorithms for solving the Traveling Salesman Problem (TSP). It provides an overview of the available ACO algorithms for the TSP, describing how ACO approaches can be applied to the problem. The key aspects covered are: (1) how ants construct tours probabilistically based on pheromone trails and heuristic information; (2) how the pheromone trails are updated after construction; and (3) how the best performing ACO algorithms for the TSP improve tours using local search methods in a hybrid approach.

Uploaded by

Sait Demir
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 23

1

ACO Algorithms for the Traveling Salesman Problem


Thomas STUTZLE and Marco DORIGO
IRIDIA, Universit Libre de Bruxelles, Belgium e {tstutzle,mdorigo}@ulb.ac.be

1.1

INTRODUCTION

Ant algorithms [18, 14, 19] are a recently developed, population-based approach which has been successfully applied to several N P-hard combinatorial optimization problems [6, 13, 17, 23, 34, 40, 49]. As the name suggests, ant algorithms have been inspired by the behavior of real ant colonies, in particular, by their foraging behavior. One of the main ideas of ant algorithms is the indirect communication of a colony of agents, called (articial) ants, based on pheromone trails (pheromones are also used by real ants for communication). The (articial) pheromone trails are a kind of distributed numeric information which is modied by the ants to reect their experience while solving a particular problem. Recently, the Ant Colony Optimization (ACO) metaheuristic has been proposed which provides a unifying framework for most applications of ant algorithms [15, 16] to combinatorial optimization problems. In particular, all the ant algorithms applied to the TSP t perfectly into the ACO meta-heuristic and, therefore, we will call these algorithms also ACO algorithms. The rst ACO algorithm, called Ant System (AS) [18, 14, 19], has been applied to the Traveling Salesman Problem (TSP). Starting from Ant System, several improvements of the basic algorithm have been proposed [21, 22, 17, 51, 53, 7]. Typically, these improved algorithms have been tested again on the TSP. All these improved versions of AS have in common a stronger exploita-

To appear in K. Miettinen, M. Mkel, P. Neittaanmki and J. Periaux, editors, Evolua a a tionary Algorithms in Engineering and Computer Science: Recent Advances in Genetic Algorithms, Evolution Strategies, Evolutionary Programming, Genetic Programming and Industrial Applications. John Wiley & Sons, 1999. Presently on leave from FG Intellektik, TU Darmstadt, Germany.

ii

ACO ALGORITHMS FOR THE TSP

tion of the best solutions found to direct the ants search process; they mainly dier in some aspects of the search control. Additionally, the best performing ACO algorithms for the TSP [17, 49] improve the solutions generated by the ants using local search algorithms. In this paper we give an overview on the available ACO algorithms for the TSP. We rst introduce, in Section 1.2, the TSP. In Section 1.3 we outline how ACO algorithms can be applied to that problem and present the available ACO algorithms for the TSP. Section 1.4 briey discusses local search for the TSP, while Section 1.5 presents experimental results which have been obtained with MAX MIN Ant System, one of the improved versions of Ant System. Since the rst application of ACO algorithms to the TSP, they have been applied to several other combinatorial optimization problems. On many important problems ACO algorithms have proved to be among the best available algorithms. In Section 1.6 we give a concise overview of these other applications of ACO algorithms. 1.2 THE TRAVELING SALESMAN PROBLEM

The TSP is extensively studied in literature [29, 31, 45] and has attracted since a long time a considerable amount of research eort. The TSP also plays an important role in Ant Colony Optimization since the rst ACO algorithm, called Ant System [18, 14, 19], as well as many of the subsequently proposed ACO algorithms [21, 17, 52, 53, 7] have initially been applied to the TSP. The TSP was chosen for many reasons: (i) it is a problem to which ACO algorithms are easily applied, (ii) it is an N P-hard [26] optimization problem, (iii) it is a standard test-bed for new algorithmic ideas and a good performance on the TSP is often taken as a proof of their usefulness, and (iv) it is easily understandable, so that the algorithm behavior is not obscured by too many technicalities. Intuitively, the TSP is the problem of a salesman who wants to nd, starting from his home town, a shortest possible trip through a given set of customer cities and to return to its home town. More formally, it can be represented by a complete weighted graph G = (N, A) with N being the set of nodes, representing the cities, and A the set of arcs fully connecting the nodes N . Each arc is assigned a value dij , which is the length of arc (i, j) A, that is, the distance between cities i and j, with i, j N . The TSP is the problem of nding a minimal length Hamiltonian circuit of the graph, where an Hamiltonian circuit is a closed tour visiting exactly once each of the n = |N | nodes of G. For symmetric TSPs, the distances between the cities are independent of the direction of traversing the arcs, that is, dij = dji for every pair of nodes. In the more general asymmetric TSP (ATSP) at least for one pair of nodes i, j we have dij = dji . In case of symmetric TSPs, we will use Euclidean TSP instances in which the cities are points in the Euclidean space and the inter-city distances are

AVAILABLE ACO ALGORITHMS FOR THE TSP

iii

calculated using the Euclidean norm. All the TSP instances we use are taken from the TSPLIB Benchmark library [44] which contains a large collection of instances; these have been used in many other studies or stem from practical applications of the TSP. TSPLIB is accessible on the WWW at the address http://www.iwr.uni-heidelberg.de/iwr/comopt/soft/TSPLIB95/TSPLIB.html

TSPLIB instance att532 -- optimal solution 7000 "att532-opt" 6000

5000

4000

3000

2000

1000

0 0 1000 2000 3000 4000 5000 6000 7000 8000 9000

Fig. 1.1 TSP instance att532 from TSPLIB; it comprises 532 cities in the USA. The given tour is of length 27686, which in fact is an optimal tour.

1.3 1.3.1

AVAILABLE ACO ALGORITHMS FOR THE TSP Applying ACO algorithms to the TSP

In ACO algorithms ants are simple agents which, in the TSP case, construct tours by moving from city to city on the problem graph. The ants solution construction is guided by (articial) pheromone trails and an a priori available heuristic information. When applying ACO algorithm to the TSP, a pheromone strength ij (t) is associated to each arc (i, j), where ij (t) is a numerical information which is modied during the run of the algorithm and t is the iteration counter. If an ACO algorithm is applied to symmetric TSP instances, we always have ij (t) = ji (t); in applications to asymmetric TSPs (ATSPs), we will possibly have ij (t) = ji (t).

iv

ACO ALGORITHMS FOR THE TSP

Initially, each of the m ants is placed on a randomly chosen city and then iteratively applies at each city a state transition rule. An ant constructs a tour as follows. At a city i, the ant chooses a still unvisited city j probabilistically, biased by the pheromone trail strength ij (t) on the arc between city i and city j and a locally available heuristic information, which is a function of the arc length. Ants probabilistically prefer cities which are close and are connected by arcs with a high pheromone trail strength. To construct a feasible solution each ant has a limited form of memory, called tabu list, in which the current partial tour is stored. The memory is used to determine at each construction step the set of cities which still has to be visited and to guarantee that a feasible solution is built. Additionally, it allows the ant to retrace its tour, once it is completed. After all ants have constructed a tour, the pheromones are updated. This is typically done by rst lowering the pheromone trail strengths by a constant factor and then the ants are allowed to deposit pheromone on the arcs they have visited. The trail update is done in such a form that arcs contained in shorter tours and/or visited by many ants receive a higher amount of pheromone and are therefore chosen with a higher probability in the following iterations of the algorithm. In this sense the amount of pheromone ij (t) represents the learned desirability of choosing next city j when an ant is at city i. The best performing ACO algorithms for the TSP improve the tours constructed by the ants applying a local search algorithm [17, 49, 53]. Hence, these algorithms are in fact hybrid algorithms combining probabilistic solution construction by a colony of ants with standard local search algorithms. Such a combination may be very useful since constructive algorithms for the TSP often result in a relatively poor solution quality compared to local search algorithms [29]. Yet, it has been noted that repeating local search from randomly generated initial solutions results in a considerable gap to the optimal solution [29]. Consequently, on the one hand local search algorithms may improve the tours constructed by the ants, while on the other hand the ants may guide the local search by constructing promising initial solutions. The initial solutions generated by the ants are promising because the ants use with higher probability those arcs which, according to the search experience, have more often been contained in shorter tours. In general, all the ACO algorithms for the TSP follow a specic algorithmic scheme, which is outlined in Figure 1.2. After the initialization of the pheromone trails and some parameters a main loop is repeated until a termination condition, which may be a certain number of solution constructions or a given CPU-time limit, is met. In the main loop, rst, the ants construct feasible tours, then the generated tours are improved by applying local search, and nally the pheromone trails are updated. In fact, most of the best performing ACO algorithms for N P-hard combinatorial optimization problems follow this algorithmic scheme [17, 23, 34, 49, 53]. It must be noted that the ACO meta-heuristic [15, 16] is more general than the algorithmic scheme

AVAILABLE ACO ALGORITHMS FOR THE TSP

given above. For example, the later does not capture the application of ACO algorithms to network routing problems.
procedure ACO algorithm for TSPs Set parameters, initialize pheromone trails while (termination condition not met) do ConstructSolutions ApplyLocalSearch % optional UpdateTrails end end ACO algorithm for TSPs

Fig. 1.2

Algorithmic skeleton for ACO algorithm applied to the TSP.

1.3.2

Ant System

When Ant System (AS) was rst introduced, it was applied to the TSP. Initially, three dierent versions of AS were proposed [14, 9, 18]; these were called ant-density, ant-quantity, and ant-cycle. While in ant-density and antquantity the ants updated the pheromone directly after a move from a city to an adjacent one, in ant-cycle the pheromone update was only done after all the ants had constructed the tours and the amount of pheromone deposited by each ant was set to be a function of the tour quality. Because ant-cycle performed much better than the other two variants, here we only present the ant-cycle algorithm, referring to it as Ant System in the following. In AS each of m (articial) ants builds a solution (tour) of the TSP, as described before. In AS no local search is applied. (Obviously, it would be straightforward to add local search to AS.) Tour construction. Initially, each ant is put on some randomly chosen city. At each construction step, ant k applies a probabilistic action choice rule. In particular, the probability with which ant k, currently at city i, chooses to go to city j at the tth iteration of the algorithm is: pk (t) = ij [ij (t)] [ij ] lN k [il (t)] [il ]
i

if j Nik

(1.1)

where ij = 1/dij is an a priori available heuristic value, and are two parameters which determine the relative inuence of the pheromone trail and the heuristic information, and Nik is the feasible neighborhood of ant k, that is, the set of cities which ant k has not yet visited. The role of the parameters and is the following. If = 0, the closest cities are more likely to be selected: this corresponds to a classical stochastic greedy algorithm (with

vi

ACO ALGORITHMS FOR THE TSP

multiple starting points since ants are initially randomly distributed on the cities). If = 0, only pheromone amplication is at work: this method will lead to the rapid emergence of a stagnation situation with the corresponding generation of tours which, in general, are strongly suboptimal [14]. Search stagnation is dened in [19] as the situation where all the ants follow the same path and construct the same solution. Hence, a tradeo between the inuence of the heuristic information and the pheromone trails exists. Pheromone update. After all ants have constructed their tours, the pheromone trails are updated. This is done by rst lowering the pheromone strength on all arcs by a constant factor and then allowing each ant to add pheromone on the arcs it has visited:
m

ij (t + 1) = (1 ) ij (t) +
k=1

k ij (t)

(1.2)

where 0 < 1 is the pheromone trail evaporation. The parameter is used to avoid unlimited accumulation of the pheromone trails and it enables the algorithm to forget previously done bad decisions. If an arc is not chosen by k the ants, its associated pheromone strength decreases exponentially. ij (t) is the amount of pheromone ant k puts on the arcs it has visited; it is dened as follows: if arc (i, j) is used by ant k 1/Lk (t) k (1.3) ij (t) = 0 otherwise where Lk (t) is the length of the kth ants tour. By Equation 1.3, the better the ants tour is, the more pheromone is received by arcs belonging to the tour. In general, arcs which are used by many ants and which are contained in shorter tours will receive more pheromone and therefore are also more likely to be chosen in future iterations of the algorithm. AS has been compared with other general purpose heuristics on some relatively small TSP instances with maximally 75 cities. Despite encouraging initial results, for larger instances AS gives a rather poor solution quality. Therefore, a substantial amount of recent research in ACO has focused on improvements over AS. A rst improvement on the initial form of AS, called the elitist strategy for Ant System, has been introduced in [14, 19]. The idea is to give a strong additional reinforcement to the arcs belonging to the best tour found since the start of the algorithm; this tour is denoted as T gb (global-best tour) in the following. This is achieved by adding to the arcs of tour T gb a quantity e 1/Lgb , where e is the number of elitist ants, when the pheromone trails are updated according to Equation 1.2. Some limited results presented in [14, 19] suggest that the use of the elitist strategy with an appropriate number of elitist ants allows AS: (i) to nd better tours, and (ii) to nd them

AVAILABLE ACO ALGORITHMS FOR THE TSP

vii

earlier in the run. Yet, for too many elitist ants the search concentrates early around suboptimal solutions leading to an early stagnation of the search. 1.3.3 Ant Colony System

Ant Colony System (ACS) has been introduced in [17, 22] to improve the performance of AS. ACS is based on an earlier algorithm proposed by the same authors, called Ant-Q, which exploited connections of ACO algorithms with Q-learning [55], a specic type of reinforcement learning algorithm. ACS and Ant-Q dier only in one specic aspect, which is described below. We concentrate on ACS, since it is somewhat simpler and it is preferred by its authors. ACS diers in three main aspects from ant system. First, ACS uses a more aggressive action choice rule than AS. Second, the pheromone is added only to arcs belonging to the global-best solution. Third, each time an ant uses an arc (i, j) to move from city i to city j it removes some pheromone from the arc. In the following we present these modications in more detail. Tour construction. In ACS ants choose the next city using the pseudorandom-proportional action choice rule: When located at city i, ant k moves, with probability q0 , to city l for which il (t) [il ] is maximal, that is, with probability q0 the best possible move as indicated by the learned pheromone trails and the heuristic information is made (exploitation of learned knowledge). With probability (1 q0 ) an ant performs a biased exploration of the arcs according to Equation 1.1. Global pheromone trail update. In ACS only the global best ant is allowed to add pheromone after each iteration. Thus, the update according to Equation 1.2 is modied to
gb ij (t + 1) = (1 ) ij (t) + ij (t),

(1.4)

gb where ij (t) = 1/Lgb . It is important to note that the trail update only applies to the arcs of the global-best tour, not to all the arcs like in AS. The parameter again represents the pheromone evaporation. In ACS only the global best solution receives feedback. Initially, also using the iteration best solution was considered for the pheromone update. Although for smaller TSP instances the dierence in solution quality between using the global-best solution or the iteration-best solution is minimal, for larger instances the use of the global-best tour gives by far better results.

Local pheromone trail update. Additionally to the global updating rule, in ACS the ants use a local update rule that they apply immediately after having crossed an arc during the tour construction:

viii

ACO ALGORITHMS FOR THE TSP

ij = (1 ) ij + 0

(1.5)

where , 0 < < 1, and 0 are two parameters. The eect of the local updating rule is to make an already chosen arc less desirable for a following ant. In this way the exploration of not yet visited arcs is increased. The only dierence between ACS and Ant-Q is the denition of the term 0 . Ant-Q uses a formula for 0 which was inspired by Q-learning [55]. In Ant-Q the term 0 corresponds to the discounted evaluation of the next state and is set to maxlNjk {jl } [21]. It was found that replacing the discounted evaluation of the next state by a small constant resulted in approximately the same performance and therefore, due to its simplicity, ACS is preferably used. 1.3.4 MAX MIN Ant System

MAX MIN Ant System (MMAS) [52, 51] is a direct improvement over AS. The solutions in MMAS are constructed in exactly the same way as in AS, that is, the selection probabilities are calculated as in Equation 1.1. Additionally, in [53] a variant of MMAS is considered which uses the pseudo-randomproportional action choice rule of ACS. Using that action choice rule, very good solutions could be found faster, but the nal solution quality achieved was worse. The main modications introduced by MMAS with respect to AS are the following. First, to exploit the best solutions found, after each iteration only one ant (like in ACS) is allowed to add pheromone. Second, to avoid search stagnation, the allowed range of the pheromone trail strengths is limited to the interval [min , max ], that is, ij min ij max . Last, the pheromone trails are initialized to the upper trail limit, which causes a higher exploration at the start of the algorithm [49]. Update of pheromone trails. After all ants have constructed a solution, the pheromone trails are updated according to
best ij (t + 1) = (1 ) ij (t) + ij

(1.6)

best where ij = 1/Lbest . The ant which is allowed to add pheromone may be the iteration-best solution T ib , or the global-best solution T gb . Hence, if specic arcs are often used in the best solutions, they will receive a larger amount of pheromone. Experimental results have shown that the best performance is obtained by gradually increasing the frequency of choosing T gb for the trail update [49].

Trail limits. In MMAS lower and upper limits on the possible pheromone strengths on any arc are imposed to avoid search stagnation. In particular, the imposed trail limits have the eect of indirectly limiting the probability

AVAILABLE ACO ALGORITHMS FOR THE TSP

ix

pij of selecting a city j when an ant is in city i to an interval [pmin , pmax ], with 0 < pmin pij pmax 1. Only if an ant has one single possible choice for the next city, then pmin = pmax = 1. Experimental results [49] suggest that the lower trail limits used in MMAS are the more important ones, since the maximum possible trail strength on arcs is limited in the long run due to pheromone evaporation. Trail initialization. The pheromone trails in MMAS are initialized to their upper pheromone trail limits. Doing so the exploration of tours at the start of the algorithms is increased, since the relative dierences between the pheromone trail strengths are less pronounced. 1.3.5 Rank-Based Version of Ant System

Another improvement over Ant System is the rank-based version of Ant System (ASrank ) [7]. In ASrank , always the global-best tour is used to update the pheromone trails, similar to the elitist strategy of AS. Additionally, a number of the best ants of the current iteration are allowed to add pheromone. To this aim the ants are sorted by tour length (L1 (t) L2 (t) . . . Lm (t)), and the quantity of pheromone an ant may deposit is weighted according to the rank r of the ant. Only the (w 1) best ants of each iteration are allowed to deposit pheromone. The global best solution, which gives the strongest feedback, is given weight w. The rth best ant of the current iteration contributes to pheromone updating with a weight given by max{0, w r}. Thus the modied update rule is:
w1

ij (t + 1) = (1 ) ij (t) +
r=1

gb r (w r) ij (t) + w ij (t),

(1.7)

gb r where ij (t) = 1/Lr (t) and ij (t) = 1/Lgb . In [7] ASrank has been compared to AS, AS with elitist strategy, to a genetic algorithm, and to a simulated annealing algorithm. For the larger TSP instances (the largest instance considered had 132 cities) the AS-based approaches showed to be superior to the genetic algorithm and the simulated annealing procedure. Among the AS-based algorithms, both, ASrank and elitist AS performed signicantly better than AS, with ASrank giving slightly better results than elitist AS.

1.3.6

Synopsis

Obviously, the proposed ACO algorithms for the TSP share many common features. Ant System can mainly be seen as a rst study to demonstrate the viability of ACO algorithms to attack N P-hard combinatorial optimization

ACO ALGORITHMS FOR THE TSP

problems, but its performance compared to other approaches is rather poor. Therefore, several ACO algorithms have been proposed which strongly increase the performance of Ant System. The main common feature among the proposed improvements is that they exploit more strongly the best solutions found during the ants search. This is typically achieved by giving a higher weight to better solutions for the pheromone update and often allowing to deposit additional pheromone trail on the arcs of the global-best solution. In particular, in elitist AS and in ASrank the global best solution adds a strong additional reinforcement (in ASrank additionally some of the best ants of an iteration add pheromone); in ACS and MMAS only one single ant, either the global-best or the iteration-best ant, deposit pheromone. Obviously, by exploiting the best solutions more strongly, arcs contained in the best tours receive a strong additional feedback and therefore are chosen with higher probability in subsequent algorithm iterations. Yet, a problem associated with the stronger exploitation of search experience may be search stagnation, that is, the situation in which all ants follow the same path. Hence, some of the proposed ACO algorithms, in particular ACS and MMAS, introduce additional features to avoid search stagnation. In ACS stagnation is avoided by the local updating rule which decreases the amount of pheromone on an arc and makes this arc less and less attractive for following ants. In this way the exploration of not yet visited arcs is favored. In MMAS search stagnation is avoided by imposing limits on the allowed pheromone strength associated with an arc. Hence, since the pheromone trail limits inuence directly the selection probabilities given by Equation 1.1, the selection probability for an arc cannot fall below some lower value. By appropriate choices for the trail limits, it is very unlikely that all ants follow the same path. All proposed improvements over AS show signicantly better performance on the TSP [19, 7, 21, 49] and they all make use of a stronger exploitation of the best solutions found by the ants. It is therefore reasonable to think that the concentration of the search around the best solutions found during the search is the key to the improved performance shown by these algorithms. The characteristics of the TSP search space may explain this phenomenon. In particular, recent research has addressed the issue of the correlation between the solution cost and the distance from very good or from the optimal solution [4, 3] (an obvious distance measure for the TSP is given by the number of dierent arcs in two tours). Similarly, the Fitness-Distance Correlation (FDC) has been introduced in the context of genetic algorithm research [30]. The FDC measures the correlation between the distance of a solution from the closest optimal solution and the solution costs. Plots of the solution cost versus the distance from optimal solutions have been particularly useful to visualize the correlation. In Figure 1.3 we present such a plot for a randomly generated Euclidean TSP instance with 500 cities. The plot is based on 5000 local optima which have been obtained by a 3-opt local search algorithm (see next section). The x-axis gives the distance from the optimal solution, while

AVAILABLE ACO ALGORITHMS FOR THE TSP

xi

on the y-axis the tour length is given. Obviously, a signicant correlation between the solution cost and the distance from the global optimum exists (the correlation coecient is 0.52); the closer a local optimum is to the known global optimum, the better, on average, is the solution. The task of algorithms like ACO algorithms is to guide the search towards regions of the search space containing high quality solutions and, possibly, the global optimum. Clearly, the notion of search space region is tightly coupled to the notion of distance, dened by an appropriate distance measure, between solutions. The most important guiding mechanism of ACO algorithms is the objective function value of solutions; the better a solution the more feedback it is allowed to give. This guiding mechanism relies on the general intuition that the better a solution is, the more likely it is to nd even better solutions close to it. The tness-distance correlation describes this intuitive relation between the tness (cost) of solutions and their distance from very good solutions or from the global best solution. Since for the TSP such a correlation obviously exists, it appears to be a good idea to concentrate the search around the best solutions found.

178000 "dist-opt" 177000

176000

175000 Tour length

174000

173000

172000

171000

170000

169000 60 70 80 90 100 110 120 130 Distance from optimal tour 140 150 160

Fig. 1.3 Fitness distance plot for a Euclidean TSP instances with 500 cities distributed randomly according to an uniform distribution on a square. The plot is based on 5000 3-opt tours. The distance from the optimal tour is measured by the number of dierent arcs.

xii

ACO ALGORITHMS FOR THE TSP

1.4

LOCAL SEARCH FOR THE TSP

Local Search starts from some initial assignment and repeatedly tries to improve the current assignment by local changes. If in the neighborhood of the current tour T a better tour T is found, it replaces the current tour and the local search is continued from T . The most widely known iterative improvement algorithms for the TSP are certainly 2-opt [12] and 3-opt [32]. They proceed by systematically testing whether the current tour can be improved by replacing 2 or at most 3 arcs, respectively. Both local search algorithms are widely studied in the literature [45, 29] and have been shown empirically to yield good solution quality. Local search algorithms using k > 3 arcs to be exchanged are not used commonly, due to the high computation times involved and the low additional increase in solution quality. Already for 2-opt and 3-opt a straightforward implementation would require O(n2 ) or O(n3 ) exchanges to be examined. This is clearly infeasible when trying to solve instances with several hundreds of cities in reasonable computation time. Fortunately, there exist quite a few speed-up techniques [1, 45, 29] achieving, in practice, run-times which grow sub-quadratically. This eect is obtained by examining only a small part of the whole neighborhood. We use three techniques to reduce the run-time of 2-opt and 3-opt implementations. One, consists in restricting the set of moves which are examined to those contained in a candidate list of the nearest neighbors ordered according to nondecreasing distances [1, 33, 45]. Using candidate lists, for a given starting node i we only consider moves which add a new arc between i and one of the nodes in its candidate list. Hence, by using a neighborhood list of bounded length, an improving move can be found in constant time. An additional speed-up is achieved by performing a xed radius nearest neighbor search [1]. For 2-opt at least one newly introduced arc has to be shorter than any of the two removed arcs (i, j) and (k, l). Due to symmetry reasons, it suces to check whether dij > dik . A similar argument also holds for 3-opt [1]. To yield a further reduction in run-time we use dont look bits associated with each node. Initially, all dont look bits are turned o (set to 0). If for a node no improving move can be found, the dont look bit is turned on (set to 1) and the node is not considered as a starting node for nding an improving move in the next iteration. In case an arc incident to a node is changed by a move, the nodes dont look bit is turned o again. For asymmetric TSPs, 2-opt is not directly applicable because one part of the tour has to be traversed in the opposite direction. In case a sub-tour is reversed, the length of this sub-tour has to be computed from scratch. Yet, one of the three possible 3-opt moves does not lead to a reversal of a sub-tour. We call reduced 3-opt this special form of 3-opt. The implementation of reduced 3-opt uses the same speed-up techniques as described before. The local search algorithm producing the highest quality solutions for symmetric TSPs is the Lin-Kernighan heuristic (LK) [33]. The Lin-Kernighan

EXPERIMENTAL RESULTS

xiii

heuristic considers in each step a variable number of arcs to be exchanged. Yet, a disadvantage of the LK heuristic is that its implementation is more dicult than that of 2-opt and 3-opt, and careful ne-tuning is necessary to get it to run fast and produce high quality solutions [45, 29]. 1.5 EXPERIMENTAL RESULTS

In this section we present some computational results obtained with MMAS on some symmetric and asymmetric TSP instances from TSPLIB. Since we apply MMAS to larger TSP instances with several hundreds of cities, we rst present two techniques which decrease the complexity of the ACO algorithm iterations. 1.5.1 Additional techniques

Candidate lists. Each tour construction has complexity O(n2 ), which would lead to substantial run-times for larger instances. A standard technique used when solving large TSPs is the use of candidate lists [29, 38, 45]. Typically, a candidate list comprises a xed number of nearest neighbors for each city in order of nondecreasing distances. The use of candidate lists for the tour construction by the ants has rst been proposed for the application of ACS [22] to the TSP. When constructing a tour an ant chooses the next city among those of the candidate list, if possible. Only if all the members of the candidate list of a city have already been visited, one of the remaining cities is chosen. Note that using nearest neighbor lists is a reasonable approach when trying to solve TSPs. For example, in many TSP instances the optimal tour can be found within a surprisingly low number of nearest neighbors. For example, an optimal solution is found for instance pcb442.tsp, which has 442 cities, within a subgraph of the 6 nearest neighbors and for instance pr2392.tsp (2392 cities) within a subgraph of the 8 nearest neighbors [45]. Fastening the trail update. When applying ACO algorithms to the TSP, pheromone trails are stored in a matrix with O(n2 ) entries (one for each arc). All the entries of this matrix should be updated at every iteration because of pheromone trail evaporation implemented by Formula 1.2 (the update of the whole trail matrix does not happen in ACS). Obviously, this is a very expensive operation if large TSP instances should be solved. To avoid this, in MMAS pheromone evaporation is applied only to arcs connecting a city i to cities belonging to is candidate list. Hence, the pheromone trails can be updated in O(n).

xiv

ACO ALGORITHMS FOR THE TSP

1.5.2

Parameter settings

Suitable parameter settings were determined in some preliminary experiments. We use m = 25 ants (all ants apply a local search to their solution), = 0.2, = 1, and = 2. During the tour construction the ants use a candidate list of size 20. The most important aspect concerning the allowed interval of the trail values is that they have to be in some reasonable interval around the expected amount of pheromone deposited by the trail update. Hence, the interval for the allowed values of the pheromone trail strength is determined 1 to reect this intuition. We set max = T1 , where T gb is the tour length of gb the global best tour found (this setting corresponds to the maximum possible trail level for longer runs). The lower trail limit is chosen as min = max /2 n, where n is the number of cities of an instance. As indicated in Section 1.3.4, the best results with MMAS are obtained when the frequency with which the global-best solution is used to update pheromone trails increases at run-time. To do so, we apply a specic schedule to alternate the pheromone trail update between T gb and T ib . Let ugb indicate that every ugb iterations T gb is chosen to deposit pheromone. In the rst 25 iterations only T ib is used to update the trails; we set ugb to 5 if 25 < t < 75 (where t is the iteration counter), to 3 if 75 < t < 125 to 3 (where t is the iteration counter), to 2 if 125 < t < 250 to 2, and from then on to 1. By shifting the emphasis from the iteration-best to the global-best solution for the trail update, we in fact are shifting from a stronger exploration of the search space to an exploitation of the best solution found so far. Additionally, we reinitialize the pheromone trails to max if the pheromone trail strengths on almost all arcs not contained in T gb is very close to min (as indicated by the average branching factor [21]). After a reinitialization of the pheromone trails, the schedule is applied from the start again. For the local search we use 3-opt, enhanced by the speed-up techniques described in Section 1.4. The size of the candidate list was set to 40 for the local search, following recommendations of [29]. Initially, all the dont look bits are turned o. 1.5.3 Experimental results for symmetric TSPs

In this section we report on experimental results obtained with MMAS, on some symmetric TSP instances from TSPLIB. Most of the instances were proposed as benchmark instances in the First International Contest on Evolutionary Computation (1st ICEO) [2]. We compare the computational results obtained with MMAS to a standard iterated local search algorithm [38, 37, 29] for the TSP using the same 3-opt implementation and the same maximal computation time tmax for each trial as MMAS. Iterated local search (ILS) [38, 37, 29] is well known to be among the best algorithms for the TSP. In ILS a local search is applied iteratively to starting solutions which are obtained by mutations of a previously found local

EXPERIMENTAL RESULTS

xv

optimal solution, typically the best solution found so far. In particular, the ILS algorithm we use for comparison rst locally optimizes an initial tour produced by the nearest neighbor heuristic. Then it applies iteratively 3-opt to initial solutions which are obtained by the application of random, sequential 4-opt moves (called double-bridge moves in [38]) to the best solution found so far. The runs are performed on a Sun UltraSparc II Workstation with two UltraSparc I 167MHz processors with 0.5MB external cache. Due to the sequential implementation of the algorithm only one processor is used. The computational results show that MMAS is able to nd very high quality solutions for all instances and on most instances MMAS achieves a better average solution quality than ILS. This result is very encouraging, since ILS is cited in the literature as a very good algorithm for the TSP [37, 29].

Table 1.1 Comparison of MMAS with iterated 3-opt (ILS) applied to some symmetric TSP instances (available in TSPLIB). Given are the instance name (the number in the name gives the problem dimension, that is, the number of cities), the algorithm used, the best solution, the average solution quality (its percentage deviation from the optimum in parentheses), the worst solution generated, the average time tavg to nd the best solution in a run, and the maximally allowed computation time tmax . Averages are taken over 25 trials for n < 1000, over 10 trials on the larger instances. Best average results are in boldface.
Instance d198 lin318 pcb442 att532 rat783 u1060 pcb1173 d1291 1577 Algorithm MMAS ILS MMAS ILS MMAS ILS MMAS ILS MMAS ILS MMAS ILS MMAS ILS MMAS ILS MMAS ILS Best 15780 15780 42029 42029 50778 50778 27686 27686 8806 8806 224455 224152 56896 56897 50801 50825 22286 22311.0 Average 15780.4 (0.0) 15780.2 (0.0) 42029.0 (0.0) 42064.6 (0.09) 50911.2 (0.26) 50917.7 (0.28) 27707.9 (0.08) 27709.7 (0.09) 8814.4 (0.10) 8828.4 (0.34) 224853.5 (0.34) 224408.4 (0.14) 56956.0 (0.11) 57029.5 (0.24) 50821.6 (0.04) 50875.7 (0.15) 22311.0 (0.28) 22394.6 (0.65) Worst 15784 15784 42029 42163 51047 51054 27756 27759 8837 8850 225131 224743 57120 57251 50838 50926 22358 22505 tavg 61.7 18.6 94.2 55.6 308.9 180.7 387.3 436.2 965.2 1395.2 2577.6 1210.4 3219.5 1892.0 1894.4 1102.9 3001.8 3447.6 tmax 170 450 600 1250 2100 3600 5400 5400 7200

xvi

ACO ALGORITHMS FOR THE TSP

Table 1.2 Comparison of MMAS with iterated 3-opt (ILS) applied to some asymmetric TSP instances (available in TSPLIB). Given are the instance name (the number in the name gives the problem dimension, that is, the number of cities; an exception is instance kro124p with 100 cities), the algorithm used, the best solution, the average solution quality (its percentage deviation from the optimum in parentheses), the worst solution generated, the average time tavg to nd the best solution in a run, and the maximally allowed computation time tmax . Averages are taken at least over 25 trials. Best average results are in boldface.
Instance ry48p ft70 kro124p ftv170 Algorithm MMAS ILS MMAS ILS MMAS ILS MMAS ILS Best 14422 14422 38673 38673 36230 36230 2755 2755 Average 14422.0 (0.0) 14425.6 (0.03) 38673.0 (0.0) 38687.9 (0.04) 36230.0 (0.0) 36542.5 (0.94) 2755.0 (0.0) 2756.8 (0.07) Worst 14422 14507 38673 38707 36230 37114 2755 2764 tavg 2.3 24.6 37.2 3.1 7.3 23.4 56.2 21.7 tmax 120 300 300 600

1.5.4

Experimental results for asymmetric TSPs

We applied the same algorithms also to the more general asymmetric TSP; the computational results are given in Table 1.2. On the asymmetric instances the performance dierence between MMAS and ILS becomes more pronounced. While MMAS solves all instances in all runs to optimality, ILS gets stuck at suboptimal solutions in all the four instances tested. Among the instances tested, ry48p and kro124p are easily solved by MMAS, while on these the relatively poor performance of ILS is most striking. Only the two instances ft70 and ftv170 are somewhat harder. Computational results from other researchers suggest that in particular instance ft70 is, considering its small size, relatively hard to solve [39, 54, 17]. 1.5.5 Related work on the TSP

Because the TSP is a standard benchmark problem for meta-heuristic algorithms, it has received considerable attention from the research community. Here, we only mention some of the most recent work, for a discussion of earlier work we refer to the overview article by Johnson and McGeoch [29]. Currently, the iterated LK heuristic (ILK), that is, ILS using the LK heuristic for the local search, is the most ecient approach to symmetric TSPs for short to medium run-times [29]. Recently, several new approaches and improved implementations have been presented which appear to perform as well or better than ILK for larger run-times. Among these algorithms we nd the

OTHER APPLICATIONS OF ACO ALGORITHMS

xvii

genetic local search approach of Merz and Freisleben [20, 39], an improved implementation of ASPARAGOS (one of the rst genetic local search approaches to the TSP) of Gorges-Schleuter [27], a new genetic local search approach using a repair-based crossover operator and brood selection by Walters [54], a genetic algorithm using a specialized crossover operator, called edge assembly crossover, due to Nagata and Kobayashi [42], and nally a specialized local search algorithm for the TSP called Iterative Partial Transcription by Mbius et.al. [41]. Some of these algorithms [39, 41, 42, 54] achieve better o computational results than the ones presented here. For example, the genetic local search approach presented in [39], which uses the LK heuristic for the local search, reaches on average a solution of 8806.2 on instance rat783 in 424 seconds on a DEC Alpha station 255 MHz. Yet, the computational results with MMAS also would benet from the application of a more powerful local search algorithm like the LK heuristic. In fact, some initial experimental results presented in [49] conrm this conjecture, but still the computation times are relatively high compared to, for example, the results given in [39]. Applied to asymmetric TSP instances, our computational results with respect to solution quality compare more favorably to these approaches. For example, the solution quality we obtain with MMAS is better than that of the genetic local search approach of [39] and the same as reported in [27, 54], but at the cost of higher run-times. 1.6 OTHER APPLICATIONS OF ACO ALGORITHMS

Research on ACO algorithms, and on ant algorithms more in general, has led to a number of other successful applications to combinatorial optimization problems. We call ant algorithm those algorithms that take inspiration by some aspects of social insects behavior. In general, ant algorithms are characterized by being multi-agent systems using only local information to make stochastic decisions and interacting with each other via a form of indirect communication called stigmergy [28]. A general introduction to ant algorithms can be found in Bonabeau, Dorigo, and Theraulaz [5]. In this paper we limit our attention to ACO algorithms, a particular class of ant algorithms which follow the meta-heuristic described in Dorigo, Di Caro, and Gambardella [16] and Dorigo and Di Caro [15]. The most widely studied problems using ACO algorithms are the traveling salesman problem and the quadratic assignment problem (QAP). Applications of ACO algorithms to the TSP have been reviewed in this paper. As in the TSP case, the rst application of an ACO algorithm to the QAP has been that of Ant System [36]. In the last two years several ant and ACO algorithms for the QAP have been presented by Maniezzo and Colorni [35], Maniezzo [34], Gambardella, Taillard, and Dorigo [25], and Sttzle [48]. Currently, u ant algorithms are among the best algorithms for attacking real-life QAP instances. We refer to [50] for an overview of these approaches.

xviii

ACO ALGORITHMS FOR THE TSP

The sequential ordering problem is closely related to the asymmetric TSP, but additional precedence constraints between the nodes have to be satised. Gambardella and Dorigo have tackled this problem by an extension of ACS enhanced by a local search algorithm [23]. They obtained excellent results and were able to improve the best known solutions for many benchmark instances. A rst application of AS to the job-shop scheduling problem has been presented in [10]. Despite showing the viability of the approach, the computational results are not competitive with state-of-the-art algorithms. More recently, MMAS has been applied to the ow shop scheduling problem (FSP) in [47]. The computational results have shown that MMAS outperforms earlier proposed Simulated Annealing algorithms and performs comparably to Tabu Search algorithms applied to the FSP. Costa and Herz [11] have proposed an extension of AS to assignment type problems and present an application of their approach to the graph coloring problem obtaining results comparable to those obtained by other heuristic approaches. Applications of ASrank to vehicle routing problems are presented by Bullnheimer, Hartl, and Strauss [8, 6]. They obtained good computational results for standard benchmark instances, slightly worse than the best performing Tabu Search algorithms. A recent application by Gambardella, Taillard, and Agazzi to vehicle routing problems with time windows, improves the best known solutions for some benchmark instances [24]. A further application of AS to the shortest common supersequence problem has been proposed by Michel and Middendorf [40]. They introduce the novel aspect of using a lookahead function during the solution construction by the ants. Additionally, they present a parallel implementation of their algorithm based on an island model often also used in parallel genetic algorithms. MMAS has recently been applied to the generalized assignment problem by Ramalhinho Lorenou and Serra [43], obtaining very promising compuc tational results. In particular, their algorithm is shown to nd optimal and near optimal solutions faster than a GRASP algorithm which was used for comparison. Applications of ACO algorithms to telecommunication networks, in particular to network routing problems, have recently received a strongly increased interest in the ACO community (see, for example, Schoonderwoerd et al. [46] and Di Caro and Dorigo [13]). The application of ACO algorithms to network optimization problems is appealing, since these problems have characteristics like distributed information, non-stationary stochastic dynamics, and asynchronous evolution of the network status which well match some of the properties of ACO algorithms like the use of local information to generate solutions, indirect communication via the pheromone trails and stochastic state transitions. A detailed overview of routing applications of ACO algorithms can be found in Dorigo, Di Caro and Gambardella [16].

OTHER APPLICATIONS OF ACO ALGORITHMS

xix

Acknowledgments
This work was supported by a Madame Curie Fellowship awarded to Thomas St tzle u (CEC-TMR Contract No. ERB4001GT973400). Marco Dorigo acknowledges support from the Belgian FNRS, of which he is a Research Associate.

REFERENCES 1. J.L. Bentley. Fast Algorithms for Geometric Traveling Salesman Problems. ORSA Journal on Computing, 4(4):387411, 1992. 2. H. Bersini, M. Dorigo, S. Langerman, G. Seront, and L. Gambardella. Results of the First International Contest on Evolutionary Optimisation. In Proceedings of the IEEE International Conference on Evolutionary Computation (ICEC96), pages 611615. IEEE Press, 1996. 3. K.D. Boese. Models for Iterative Global Optimization. PhD thesis, University of California, Computer Science Department, Los Angeles, 1996. 4. K.D. Boese, A.B. Kahng, and S. Muddu. A New Adaptive Multi-Start Technique for Combinatorial Global Optimization. Operations Research Letters, 16:101113, 1994. 5. E. Bonabeau, M. Dorigo, and G. Theraulaz. From Natural to Articial Swarm Intelligence. Oxford University Press, 1999. 6. B. Bullnheimer, R.F. Hartl, and C. Strauss. An Improved Ant System Algorithm for the Vehicle Routing Problem. Annals of Operations Research. To appear. 7. B. Bullnheimer, R.F. Hartl, and C. Strauss. A New Rank Based Version of the Ant System A Computational Study. Central European Journal for Operations Research and Economics. To appear. 8. B. Bullnheimer, R.F. Hartl, and C. Strauss. Applying the Ant System to the Vehicle Routing Problem. In S. Voss, S. Martello, I.H. Osman, and C. Roucairol, editors, Meta-Heuristics: Advances and Trends in Local Search Paradigms for Optimization, pages 285296. Kluwer, Boston, 1999. 9. A. Colorni, M. Dorigo, and V. Maniezzo. Distributed Optimization by Ant Colonies. In Proceedings of the First European Conference on Articial Life (ECAL 91), pages 134142. Elsevier, 1991. 10. A. Colorni, M. Dorigo, V. Maniezzo, and M. Trubian. Ant System for Job-Shop Scheduling. Belgian Journal of Operations Research, Statistics and Computer Science, 34(1):3953, 1994.

xx

ACO ALGORITHMS FOR THE TSP

11. D. Costa and A. Hertz. Ants Can Colour Graphs. Journal of the Operational Research Society, 48:295305, 1997. 12. G.A. Croes. A Method for Solving Traveling Salesman Problems. Operations Research, 6:791812, 1958. 13. G. Di Caro and M. Dorigo. AntNet: Distributed Stigmergetic Control for Communications Networks. Journal of Articial Intelligence Research (JAIR), 9:317365, December 1998. Available at http://www.jair.org. 14. M. Dorigo. Optimization, Learning, and Natural Algorithms (in Italian). PhD thesis, Dip. Elettronica, Politecnico di Milano, 1992. 15. M. Dorigo and G. Di Caro. The Ant Colony Optimization Meta-Heuristic. In D. Corne, M. Dorigo, and F. Glover, editors, New Ideas in Optimization. McGraw-Hill, 1999. 16. M. Dorigo, G. Di Caro, and L.M. Gambardella. Ant Algorithms for Distributed Discrete Optimization. Articial Life. To appear. 17. M. Dorigo and L.M. Gambardella. Ant Colony System: A Cooperative Learning Approach to the Traveling Salesman Problem. IEEE Transactions on Evolutionary Computation, 1(1):5366, 1997. 18. M. Dorigo, V. Maniezzo, and A. Colorni. Positive Feedback as a Search Strategy. Technical Report 91-016, Dip. Elettronica, Politecnico di Milano, 1991. 19. M. Dorigo, V. Maniezzo, and A. Colorni. The Ant System: Optimization by a Colony of Cooperating Agents. IEEE Transactions on Systems, Man, and Cybernetics Part B, 26(1):2942, 1996. 20. B. Freisleben and P. Merz. A Genetic Local Search Algorithm for Solving Symmetric and Asymmetric Traveling Salesman Problems. In Proceedings of the IEEE International Conference on Evolutionary Computation (ICEC96), pages 616621. IEEE Press, 1996. 21. L.M. Gambardella and M. Dorigo. Ant-Q: A Reinforcement Learning Approach to the Traveling Salesman Problem. In Proceedings of the Eleventh International Conference on Machine Learning, pages 252260. Morgan Kaufmann, 1995. 22. L.M. Gambardella and M. Dorigo. Solving Symmetric and Asymmetric TSPs by Ant Colonies. In Proceedings of the IEEE International Conference on Evolutionary Computation (ICEC96), pages 622627. IEEE Press, 1996. 23. L.M. Gambardella and M. Dorigo. HAS-SOP: Hybrid Ant System for the Sequential Ordering Problem. Technical Report IDSIA 11-97, IDSIA, Lugano, Switzerland, 1997.

OTHER APPLICATIONS OF ACO ALGORITHMS

xxi

24. L.M. Gambardella, E. Taillard, and G. Agazzi. MACS-VRPTW: A Multiple Ant Colony System for Vehicle Routing Problems with Time Windows. In D. Corne, M. Dorigo, and F. Glover, editors, New Ideas in Optimization. McGraw-Hill, 1999. 25. L.M. Gambardella, E.D. Taillard, and M. Dorigo. Ant Colonies for the QAP. Journal of the Operational Research Society. To appear. 26. M.R. Garey and D.S. Johnson. Computers and Intractability: A Guide to the Theory of N P-Completeness. Freeman, San Francisco, CA, 1979. 27. M. Gorges-Schleuter. Asparagos96 and the Travelling Salesman Problem. In T. Baeck, Z. Michalewicz, and X. Yao, editors, Proceedings of the IEEE International Conference on Evolutionary Computation (ICEC97), pages 171174, 1997. 28. P. P. Grass. La Reconstruction du Nid et les Coordinations Interindie viduelles chez bellicositermes natalensis et cubitermes sp. La Thorie de la e Stigmergie: Essai dinterprtation du Comportement des Termites Cone structeurs. Insectes Sociaux, 6:4181, 1959. 29. D.S. Johnson and L.A. McGeoch. The Travelling Salesman Problem: A Case Study in Local Optimization. In E.H.L. Aarts and J.K. Lenstra, editors, Local Search in Combinatorial Optimization, pages 215310. John Wiley & Sons, 1997. 30. T. Jones and S. Forrest. Fitness Distance Correlation as a Measure of Problem Diculty for Genetic Algorithms. In L.J. Eshelman, editor, Proceedings of the 6th International Conference on Genetic Algorithms, pages 184192. Morgan Kaufman, 1995. 31. E.L. Lawler, J.K. Lenstra, A.H.G. Rinnooy Kan, and D.B. Shmoys. The Travelling Salesman Problem. John Wiley & Sons, 1985. 32. S. Lin. Computer Solutions for the Traveling Salesman Problem. Bell Systems Technology Journal, 44:22452269, 1965. 33. S. Lin and B.W. Kernighan. An Eective Heuristic Algorithm for the Travelling Salesman Problem. Operations Research, 21:498516, 1973. 34. V. Maniezzo. Exact and Approximate Nondeterministic Tree-Search Procedures for the Quadratic Assignment Problem. Technical Report CSR 98-1, Scienze dellInformazione, Universit di Bologna, Sede di Cesena, a 1998. 35. V. Maniezzo and A. Colorni. The Ant System Applied to the Quadratic Assignment Problem. IEEE Transactions on Knowledge and Data Engineering. To appear.

xxii

ACO ALGORITHMS FOR THE TSP

36. V. Maniezzo, M. Dorigo, and A. Colorni. The Ant System Applied to the Quadratic Assignment Problem. Technical Report IRIDIA/94-28, Universit de Bruxelles, Belgium, 1994. e 37. O. Martin and S.W. Otto. Combining Simulated Annealing with Local Search Heuristics. Annals of Operations Research, 63:5775, 1996. 38. O. Martin, S.W. Otto, and E.W. Felten. Large-Step Markov Chains for the Traveling Salesman Problem. Complex Systems, 5(3):299326, 1991. 39. P. Merz and B. Freisleben. Genetic Local Search for the TSP: New Results. In Proceedings of the IEEE International Conference on Evolutionary Computation (ICEC97), pages 159164. IEEE Press, 1997. 40. R. Michel and M. Middendorf. An Island Based Ant System with Lookahead for the Shortest Common Supersequence Problem. In Proceedings of the Fifth International Conference on Parallel Problem Solving from Nature, volume 1498, pages 692708. Springer Verlag, 1998. 41. A. Mbius, B. Freisleben, P. Merz, and M. Schreiber. Combinatorial o Optimization by Iterative Partial Transcription. Submitted to Physical Review E, 1998. 42. Y. Nagata and S. Kobayashi. Edge Assembly Crossover: A High-power Genetic Algorithm for the Traveling Salesman Problem. In Proc. of ICGA97, pages 450457, 1997. 43. H. Ramalhinho Loureno and D. Serra. Adaptive Approach Heuristics for c the Generalized Assignment Problem. Technical Report Economic Working Papers Series No.304, Universitat Pompeu Fabra, Dept. of Economics and Management, Barcelona, May 1998. 44. G. Reinelt. TSPLIB A Traveling Salesman Problem Library. ORSA Journal on Computing, 3:376384, 1991. 45. G. Reinelt. The Traveling Salesman: Computational Solutions for TSP Applications, volume 840 of LNCS. Springer Verlag, 1994. 46. R. Schoonderwoerd, O. Holland, J. Bruten, and L. Rothkrantz. Ant-based Load Balancing in Telecommunications Networks. Adaptive Behavior, 5(2):169207, 1996. 47. T. Sttzle. An Ant Approach to the Flow Shop Problem. In Proceedings u of the 6th European Congress on Intelligent Techniques & Soft Computing (EUFIT98), volume 3, pages 15601564. Verlag Mainz, Aachen, 1997. 48. T. Sttzle. MAX MIN Ant System for the Quadratic Assignment u Problem. Technical Report AIDA974, FG Intellektik, TU Darmstadt, July 1997.

xxiii

49. T. Sttzle. Local Search Algorithms for Combinatorial Problems Analu ysis, Improvements, and New Applications. PhD thesis, Darmstadt University of Technology, Department of Computer Science, 1998. 50. T. Sttzle. ACO Algorithms for the Quadratic Assignment Problem. In u D. Corne, M. Dorigo, and F. Glover, editors, New Ideas in Optimization. McGraw-Hill, 1999. 51. T. Sttzle and H.H. Hoos. The MAX MIN Ant System and Local u Search for the Traveling Salesman Problem. In T. Baeck, Z. Michalewicz, and X. Yao, editors, Proceedings of the IEEE International Conference on Evolutionary Computation (ICEC97), pages 309314, 1997. 52. T. Sttzle and H.H. Hoos. Improvements on the Ant System: Introducing u the MAX MIN Ant System. In R.F. Albrecht G.D. Smith, N.C. Steele, editor, Articial Neural Networks and Genetic Algorithms, pages 245249. Springer Verlag, Wien New York, 1998. 53. T. Sttzle and H.H. Hoos. MAX MIN Ant System and Local Search for u Combinatorial Optimization Problems. In S. Voss, S. Martello, I.H. Osman, and C. Roucairol, editors, Meta-Heuristics: Advances and Trends in Local Search Paradigms for Optimization, pages 313329. Kluwer, Boston, 1999. 54. T. Walters. Repair and Brood Selection in the Traveling Salesman Problem. In A.E. Eiben, T. Bck, M. Schoenauer, and H.-P. Schwefel, editors, a Proc. of Parallel Problem Solving from Nature PPSN V, volume 1498 of LNCS, pages 813822. Springer Verlag, 1998. 55. C.J.C.H. Watkins and P. Dayan. Q-Learning. Machine Learning, 8:279 292, 1992.

You might also like