Next Article in Journal
Calculation Method of Rotational Constraint Stiffness for a New Tower-Pier Connected System
Next Article in Special Issue
A Novel Filter-Level Deep Convolutional Neural Network Pruning Method Based on Deep Reinforcement Learning
Previous Article in Journal
Wave Diffraction by Metamaterial-Coated Wedges: The UAPO Solution for Skew Incidence
Previous Article in Special Issue
Using Live Spam Beater (LiSB) Framework for Spam Filtering during SMTP Transactions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rank-Based Ant System with Originality Reinforcement and Pheromone Smoothing

by
Sara Pérez-Carabaza
1,*,†,
Akemi Gálvez
2,3,*,† and
Andrés Iglesias
2,3,*,†
1
Department of Applied Mathematics and Computational Sciences, E.T.S.I. Industriales y de Telecomunicación, University of Cantabria, Avda. de los Castros, s/n, 39005 Santander, Spain
2
Department of Applied Mathematics and Computational Sciences, E.T.S.I. Caminos, Canales y Puertos, University of Cantabria, Avda. de los Castros, s/n, 39005 Santander, Spain
3
Faculty of Pharmaceutical Sciences, Toho University, 2-2-1 Miyama, Funabashi 274-8510, Japan
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2022, 12(21), 11219; https://doi.org/10.3390/app122111219
Submission received: 30 September 2022 / Revised: 24 October 2022 / Accepted: 31 October 2022 / Published: 5 November 2022
(This article belongs to the Special Issue Applied Artificial Intelligence (AI))

Abstract

:
Ant Colony Optimization (ACO) encompasses a family of metaheuristics inspired by the foraging behaviour of ants. Since the introduction of the first ACO algorithm, called Ant System (AS), several ACO variants have been proposed in the literature. Owing to their superior performance over other alternatives, the most popular ACO algorithms are Rank-based Ant System (AS R a n k ), Max-Min Ant System (MMAS) and Ant Colony System (ACS). While AS R a n k shows a fast convergence to high-quality solutions, its performance is improved by other more widely used ACO variants such as MMAS and ACS, which are currently considered the state-of-the-art ACO algorithms for static combinatorial optimization problems. With the purpose of diversifying the search process and avoiding early convergence to a local optimal, the proposed approach extends AS R a n k with an originality reinforcement strategy of the top-ranked solutions and a pheromone smoothing mechanism that is triggered before the algorithm reaches stagnation. The approach is tested on several symmetric and asymmetric Traveling Salesman Problem and Sequential Ordering Problem instances from TSPLIB benchmark. Our experimental results show that the proposed method achieves fast convergence to high-quality solutions and outperforms the current state-of-the-art ACO algorithms A S R a n k , MMAS and ACS, for most instances of the benchmark.

1. Introduction

1.1. Motivation

During the last few decades, many artificial intelligence-based optimization methods have been introduced, mostly aimed at solving problems that were either intractable or very complex to deal with using traditional mathematical optimization techniques. Most of such methods are typically labelled as metaheuristics, a global concept encompassing many different high-level procedures to assist partial search algorithms to find a good solution to a challenging optimization problem, often in NP-hard problems or scenarios where only incomplete information about the problem is commonly available. Some of the first metaheuristics were proposed to enhance local search algorithms (e.g., tabu search, scatter search, simulated annealing, guided local search, stochastic local search, VNS, GRASP, and many others). Shortly afterwards, sophisticated nature-inspired population-based metaheuristics for global search were also introduced. Popular families of such methods include evolutionary algorithms (e.g., genetic algorithms, genetic programming, evolutionary programming; see [1,2]) and swarm intelligence methods (e.g., particle swarm optimization, differential evolution, ant colony optimization, artificial bee colony, bat algorithm, firefly algorithm, cuckoo search algorithm, and many others; see [3,4]), to mention just a few. The interested reader is referred to the handbooks of metaheuristics in [5,6,7] and references within for a general overview of the field and its applications to hard optimization problems.
One of the most popular metaheuristics is Ant Colony Optimization (ACO). Rather than a single method, ACO consists of a family of metaheuristics inspired by the foraging activity of natural ants, capable of finding the shortest path between a food source and their nest through a pheromone deposit mechanism. ACO metaheuristics guide their search through the solution space of combinatorial optimization problems using a problem-specific heuristic and the information saved in the pheromone table, which is learned by means of the reinforcement of the pheromone values corresponding to the best decisions in previous iterations of the algorithm.
The first ant colony-based algorithm, called Ant System (AS) [8,9], was originally applied to solve the Traveling Salesman Problem (TSP): finding the shortest closed loop that traverses once a group of cities. Since the introduction of AS, several ACO algorithms have been proposed. Among them, three algorithms stand out owing to their superior performance and, as a result, for being the most often used ACO algorithms: the Rank-based Ant System (AS R a n k ) [10], Max-Min Ant System (MMAS) [11] and Ant Colony System (ACS) [12]. All ACO algorithms guide their search through the solution space by means of a probabilistic decision method that combines the information given by the heuristic and the pheromones. However, ACO algorithms may differ in the rules defined to update the pheromone trails or the different transition rules that combine the heuristic and pheromone information. Additionally, more recent metaheuristics combine ACO with other techniques, for instance, Best-Worst Ant System (BWAS) [13], integrates several components from Evolutionary Computation. Nevertheless, the ACO algorithms (MMAS and ACS) are by far more popular than hybrid methods and are used in a variety of combinatorial optimization problems [14,15,16].
The reinforcement of previous best decisions by means of pheromone deposit mechanism enables ACO algorithms to focus the search on promising regions of the search space, making it possible to find high-quality solutions for high-complexity problems in a reasonable computational time. However, the pheromone reinforcement is also responsible for stagnation, an undesirable situation where the algorithm becomes stuck in local minima, preventing the possibility of finding a better solution.
A distinctive feature of AS extensions (such as AS R a n k , MMAS and ACS) is that they direct the ant’s search in a more aggressive way than AS, exploiting more strongly the best solutions found during the ants’ search [14,17]. This stronger exploitation of the search experience may promote early stagnation situations. Thus, some AS extensions, in particular, MMAS and ACS, introduce additional features to avoid search stagnation (e.g., pheromone limits in MMAS) [18]. However, Rank-based AS generally obtains slightly worse performance than ACS and MMAS [14].
Figure 1 shows the fitness evolution of the tour length of the best-ant tour found by ACO algorithms versus the number of ant tours constructed for a particular instance of TSP called gr48. Although the behaviour of ACO algorithms also depends on the parameter settings, comparative studies usually draw similar conclusions as the ones that can be extracted from Figure 1. The ACO extensions (AS R a n k , MMAS and ACS) have the advantage of showing better performance than the original ACO algorithm, Ant System. The best-performing variants usually are MMAS and ACS, closely followed by AS R a n k . Regarding the performance of ACS, its local pheromone update rule (which applies the evaporation mechanism only to the traversed arcs) has the advantage of preventing ACS from reaching a stagnation situation, and the disadvantage that it precludes the parallel implementation of the construction loop of the ant tours. An advantage of ACS is that it returns the best solution quality for very short computation times, as happens in the instance shown in Figure 1. Differently, as it can be observed in Figure 1, MMAS has the disadvantage of initially producing rather poor solutions, but the advantage of producing solutions whose quality is often the best among the ACO extensions [14]. Lastly, AS R a n k has the advantage of showing better results than MMAS in the initial iterations, but it usually ends in stagnation situations, obtaining slightly worse performance than ACS and MMAS. This early stagnation behaviour suggests that AS R a n k could profit from strategies that diversify the search [11,14]. Consequently, the good results of AS R a n k during the first iterations and its trend to quickly fall in stagnation situations have motivated the present work.

1.2. Aims and Structure of This Paper

In this work, we aim at improving the performance of AS R a n k algorithm for combinatorial optimization problems. Based on the observations described in the previous paragraphs, this paper proposes an extended version of AS R a n k , which includes two diversification strategies:
  • an originality reinforcement strategy that rewards the originality (dissimilarity from already searched space) of the solutions with good fitness; and
  • a pheromone smoothing mechanism that is triggered before the algorithm reaches stagnation, increasing exploration and making possible it to find better solutions.
This new ACO algorithm, called AS O R a n k p s , will be compared with the state-of-the-art ACO algorithms using several instances of the popular TSPLIB95 benchmark. Since our main interest is at improving the state-of-the-art ACO algorithms, in this paper we will focus our comparative work on the best ACO algorithms exclusively. Consequently, we will restrict our analysis to ACO algorithms without any addition of external methods, such as local search procedures or hybridization with other metaheuristics. These procedures, although very promising and potentially successful, would obscure the understanding of the behavior of our algorithm with technicalities totally unrelated to the core and principles of the ACO algorithms. The discussion about how to enhance our new ACO algorithm with local search strategies and/or hybridization with other metaheuristics is out of the scope of this paper and will be part of our future work in the field (see our discussion in Section 5 for further details).
The remainder of the paper is organized as follows. Section 2 introduces the ACO metaheuristics, describing the most important ACO algorithms and presenting an overview of the current research lines in ACO state-of-the-art. Section 3 describes the proposed Rank-based Ant System with originality reinforcement and pheromone smoothing. Section 4 analyses the performance of the proposed ACO algorithm over several TSP and Sequential Ordering Problem (SOP) instances, and compares it with the best-performer ACO algorithms in the literature. Finally, Section 5 summarizes the main conclusions of the work and some futures lines of research.

2. Ant Colony Optimization

This section introduces the ant colony optimization metaheuristics. First, it describes the general structure of ACO algorithms, including the first ACO algorithm proposed in 1992 by Dorigo, and its extensions Rank-based Ant System, Max-Min Ant System and Ant Colony System. Later, it gives a brief overview of the more recent research trends in ACO metaheuristics. A more detailed review of ACO state of the art can be found in [16,17].

2.1. Ant Colony Optimization Algorithms

ACO metaheuristics formulate the combinatorial optimization problems as finding the best ant tour (best solution) in a graph G = ( C , L ) defined by a set of nodes C and arcs L. For instance, in the TSP problem, each node represents a city and each edge has assigned the distance between each pair of cities. Hence, the problem is formulated as finding the closed tour of minimal length that visit once all the nodes in the graph.
ACO metaheuristics are population-based algorithms that at each iteration construct m ant tours (solutions) combining the information learned in previous iterations with a problem-specific heuristic. In order to save the information of previous good ant tours, ACO algorithms use a table that assigns a pheromone value τ i , j to each edge in the graph. These pheromone values are typically initialized with a constant value for all edges, and later reinforced as the ants with better fitness traverse them. The ant tours are constructed component by component according to the probabilities determined by the transition rule, given by Equation (1). This transition rule states the probability p i , j k that the k-th ant moves from node i to node j, assigning higher probabilities to the edges with higher pheromone τ i , j and heuristic values η i , j according to:
p i , j k = τ i , j α η i , j β l N i k τ i , j α η i , j β , j N i k
where α and β are the parameters that control the pheromone and heuristic influence, respectively, and N i k is the feasible neighbourhood of ant k. For instance, in the case of TSP where the cities (nodes) can only be visited once, N i k corresponds to the set of unvisited cities.
The basic operation mode of an ACO algorithm, summarized in Algorithm 1, is as follows: at every algorithm iteration, the m artificial ants start their tours from initial nodes (randomly chosen in TSP) and construct their tours by combining the information of the heuristic η and the pheromones τ . In this way, at every iteration, m candidate solutions (ant tours) are constructed, each one having a corresponding fitness value such as the length of the tour in TSP. Optionally, after the tours of the m ants are completed, a local search procedure may be considered (for example k-exchange neighbourhood for TSP). Next, a positive feedback strategy that rewards the best ant tours is implemented by means of the update of the pheromone table. Finally, when the considered stop criterion is met, the algorithm returns the best-found tour.
Once every ant has finished its tour (which corresponds to a candidate solution to the optimization problem), the pheromone update (evaporation and deposit processes) takes place. In AS, the pheromone update process is given by the following rule:
τ i , j = ( 1 ρ ) τ i , j + k = 1 m τ i , j k , a r c ( i , j ) L
where the parameter ρ (with 0 < ρ < 1 ) is the pheromone evaporation parameter, which controls the rate at which the pheromone is evaporated. The evaporation mechanism affects all the edges in the graph and helps to avoid the unlimited accumulation of the pheromone trails. The second term of Equation (2) corresponds to the pheromone deposit of the m artificial ants. In AS, all the ants deposit a pheromone quantity τ i , j k proportional to their quality in all the arcs belonging to their tour:
τ i , j k = 1 / f ( s k ) , if arc ( i , j ) s k 0 otherwise
where f ( s k ) is the fitness of the k-th ant tour s k , which in TSP corresponds to the length of the tour. In general, arcs that are used by many ants and which are contained in shorter tours will receive higher pheromone deposits and therefore, will more likely be chosen in the next iterations of ACO.
Algorithm 1 ACO metaheuristic
Require: ACO parameters
     τ initialise pheromone trails
    while termination condition not met do
        for  k = 1 to m do                              ▹ for every ant
            for every step until the k-th ant has completed the tour do
                   select node j to visit next according to transition rule               ▹ Equation (1)
            end for
        end for
        Apply local search (optional)
        Update pheromone trails τ
end while
Since the introduction of AS in [8], several ACO algorithms have been proposed, with Rank-based Ant System (AS R a n k ) [10], Max-Min Ant System (MMAS) [11] and Ant Colony System (ACS) [12] standing out. These ACO algorithms differ from AS in the pheromone update process in Equation (3), with the exception of ACS, which also considers a modified transition rule given by Equation (6) from the one given by Equation (1).
The pheromone reinforcement enables ACO algorithms to focus the search on promising regions of the search space allowing them to find high-quality solutions in reasonable computational time. However, the pheromone reinforcement is also responsible for stagnation, an undesirable situation where no better tour is likely to be found anymore. The stagnation situation happens when at each choice point, the pheromone trail is significantly higher for one choice than for all the others. In such a situation, the ants construct the same tours over and over again and the exploration of the search space stops, limiting the possibility of finding a tour with better fitness. One particularity of AS extensions is that they direct the ant’s search in a more aggressive way and that their search is focused on a specific region of the search space [14]. This higher exploitation may, however, induce early stagnation situations. Thus, AS extensions need to be endowed with features intended to counteract stagnation (e.g., the pheromone limits in MMAS and local update rule in ACS). The pheromone update process of the AS extensions, Rank-based AS, MMAS and ACS, are detailed below.
Rank-based Ant System (AS Rank ) is an extension of AS proposed by Bullnheimer et al. [10], which incorporates the idea of ranking to the pheromone update process. Rank-based AS follows a similar procedure to Algorithm 1, but its pheromone update incorporates the idea of ranking of solutions and elitist strategy. While in AS, all ants deposit pheromones according to Equation (2), in Rank-based AS, only w ants deposit pheromones: the best tour found so far, s g b , deposits a pheromone quantity proportional to the algorithm parameter w and the ( w 1 ) best ants of the iteration deposit a pheromone quantity proportional to their ranking r, as expressed by Equation (4):
τ i , j = ( 1 ρ ) τ i , j + r = 1 w 1 ( w r ) τ i , j r + w τ i , j g b
where w is a parameter of the algorithm chosen by the user that determines the number of ants that deposit pheromones, and r corresponds to the ranking of the classification of the ants according to their fitness (where better fitness is associated with lower ranking). In Equation (4), ( 1 ρ ) τ i , j accounts for the pheromone evaporation, which is applied to all pheromone trails. Additionally, the pheromone trails corresponding to the tours of the ( w 1 ) best ants of the iteration are reinforced as by a quantity equal to ( w r ) τ i , j r , where τ i , j r = 1 / f ( s r ) . Finally, the pheromone trails corresponding to the global-best found solution s g b are reinforced with a quantity equal to w τ i , j g b , where τ i , j g b = 1 / f ( s g b ) .
Max-Min Ant System (MMAS), proposed by Stüzle and Hoos [11], is one of the best performing variants of AS [14,15,16]. In MMAS only one ant is used to update the pheromone trails at each iteration. Consequently, the modified pheromone update rule is given by
τ i , j = ( 1 ρ ) τ i , j + τ i , j b e s t
where only the arcs traversed by the best ant are reinforced with a pheromone quantity inversely proportional to the fitness of the best ant tour τ i , j b e s t = 1 / f ( s b e s t ) . Depending on the MMAS implementation, it can be considered the iteration-best ( s i b ) ant or the global-best ant ( s g b ).
Furthermore, MMAS imposes lower and upper limits τ m i n and τ m a x on all the pheromone values in order to avoid stagnation. Hence, at each iteration after the pheromone update process given by Equation (5) takes place, it is ensured that the pheromone values respects the limits, i.e., τ m i n < τ i , j < τ m a x , τ i , j . Recommended values for the pheromone trail limits indicated by [11,14] are τ m a x = 1 / ( ρ f ( s g b ) ) and τ m i n = τ m a x ( 1 0.05 n ) / ( 0.5 n 1 ) , where n is the number of cities (nodes) in the graph. These pheromone limits [ τ m i n , τ m a x ] avoid the situation of strict stagnation (all tours of an iteration being the same). Instead, this may lead to a situation (convergence of MMAS) where the pheromone values of the arcs corresponding to the best-found tour have associated a pheromone value equal to τ m a x , and the remaining arcs to τ m i n . In this convergence situation, all the ants do not follow the same tours (stagnation), but still, the exploration of the search space and the possibility of finding better solutions during the following iterations is quite limited. As a further means of increasing the exploration, some MMAS implementations consider occasional reinitializations of the pheromone trails when the algorithm approaches stagnation or when no improved tour has been generated for a large number of iterations [11].
Ant Colony System (ACS), proposed by Dorigo and Gambardella [12], is considered, along with MMAS, the best performing variant of AS [14,15,16]. ACS implements several modifications with respect to AS. First, it exploits the experience accumulated by the ants more strongly than AS does through the use of a more aggressive action choice rule given by Equation (6):
j = a r g m a x l N i k τ i l [ ν i l ] β q q 0 J otherwise
where q is a random variable uniformly distributed in [0, 1], q 0 is an algorithm parameter ( 0 < q 0 < 1 ), and J is a random variable selected according to the probability distribution given by Equation (1) (with α = 1 ). In this way, the ants select with a probability q 0 the next node j based on the best decision according to the pheromone trails and heuristic information (the node j with maximum associated probability p i , j k given by Equation (1)), while with probability ( 1 q 0 ) , the ants choose the next node according to the probability distribution p i , j k given by Equation (1).
Regarding the pheromone update process, at the end of each algorithm iteration, ACS only applies the pheromone evaporation and reinforcement to those arcs belonging to best-so-far ant, according to Equation (7):
τ i , j = ( 1 ρ ) τ i , j + τ i , j g b a r c ( i , j ) s g b
where ρ is the pheromone evaporation parameter and τ i , j g b = 1 / f ( s g b ) .
In addition to the global pheromone trail update rule given by Equation (7), that is applied after the m ant tours are constructed, ACS also considers a local pheromone update rule that is applied immediately after an ant has crossed an a r c ( i , j ) during the tour construction, according to Equation (8):
τ i , j = ( 1 ξ ) τ i , j + ξ τ 0
where ξ ( 0 < ξ < 1 ) and τ 0 are two algorithm parameters. The value τ 0 is set to be the same as the initial value of the pheromone trails.

2.2. Recent Trends in ACO Algorithms

Among ACO scientific literature, three main research lines can be distinguished. On the one hand, the majority of the state-of-the-art works are clearly focused on the application of ACO metaheuristics to different and computationally challenging applications and domains, such as UAV search missions [19], maximizing influence in social networks [20] or in wireless sensor networks [21]. For more ACO applications, the reader is kindly referred to [17].
On the other hand, there is a challenging research line (in which this work is also included) that aims to improve ACO metaheuristics through the inclusion of modifications or mechanisms that aim to diversify the search to promising regions of the search space while keeping a good balance between diversification/intensification. Within this research line, we can find several hybrid algorithm proposals that combine ACO with other techniques. For instance, the Hybrid Ant Colony Optimization Algorithm (HACO, [22]) introduces an operator similar to the mutation operator of the genetic algorithms. Or the Best-Worst Ant System (BWAS) [13], integrates components inspired by evolutionary computation such as the inclusion of mutations of the pheromone table. The papers in [23,24] consider the hybridization of ACO with genetic algorithms to solve the supplier selection and multiple sequence alignment problems, respectively. Additionally, the authors of [25] consider an annealing ACO with a mutation operator to solve the TSP. Furthermore, BWAS is hybridized with Particle Swarm Optimization (PSO) ideas in [26] and tested over for TSP benchmark. Another example of hybridization of ACO and PSO is given in [27], related to PID parameter optimization on autonomous underwater vehicle control systems. Other approaches include the hybridization of ACO algorithms with differential evolution in [28,29] to solve TSP and cancer data classification problems. Additionally, the ACO metaheuristic proposed in [30] presents a different approach to diversify the solution space based on combining pairs of searching ants. Lastly, another interesting research line is the study of parameter adaptation techniques, that is, the variation of ACO parameter settings while solving an instance of an optimization problem [31].
Finally, a third current line of research concerns the application of the ACO algorithms to dynamic discrete problems [32]. Some particular features of the ACO algorithms, such as the pheromone-based memory, have proven to be remarkably successful for static combinatorial optimization problems, but such features do not seem to be so favorable for dynamic problems, where some parts of the problem can change at runtime without previous knowledge [33]. In such cases, if the ACO algorithm has achieved convergence to a confined region of the search space which becomes irrelevant after the change in the problem, the algorithm can find it difficult to efficiently escape from that region and adapt its memory structure to the new problem change. As a way to overcome such a limitation, a method called P-ACO has been proposed to optimize the reuse of pheromone information after the problem changes [34]. Other alternative methods are based on the use of immigrant schemes for dynamic problems [35,36]. A very recent paper considers the role of parameter setting and the use of enhanced hybridization with local search procedures [37]. The results in [37] show that the hybridization of MMAS (one of the best ACO performers for static problems) with local search can consistently outperform P-ACO for dynamic combinatorial optimization problems.

3. Extension of Rank-Based Ant System

This section presents the proposed extension of the Rank-based Ant System. Firstly, the two contributions of the proposed algorithms are presented; (i) an originality utility function and (ii) a pheromone smoothing mechanism that is triggered before the algorithm reaches stagnation. Next, the proposed variant of the Rank-based AS algorithm is detailed, highlighting the modifications with respect to the original Rank-based AS already described in Section 2.

3.1. Originality Utility Function

With the objective of promoting exploration of the search space, this section presents a utility function that measures the originality or dissimilarity of one solution with respect to a group of solutions.
A way of measuring the distance d i s t ( s , s ) between two tours s and s is to count the number of arcs contained in one tour, but not in the other [14]. The average distance between all possible pairs of tours can be used to measure the amount of exploration of the algorithm, as a decrease in the average distance indicates that preferred paths are appearing [14]. Alternatively, this distance definition (i.e., number of non-common arcs) could be used instead for our purpose of measuring the originality of one ant tour with respect to a group of tours, by considering the average distance of all the possible pairs formed by the ant tour s and the rest of the ant tours. However, a disadvantage of this measure is that it is computationally expensive [14].
The proposed originality function, given by Equation (9), is based on the idea that the tours of those ants that traverse arcs that have been visited by a smaller number of ants are more original. The originality function f o of an ant tour s k is given by the inverse of the number of ants that have traversed each arc of s k , that is:
f o ( s k ) = 1 a r c ( i , j ) s k ν i , j
where ν i , j represents the number of ants that have traversed the edge going from node i to node j.
As an illustrative example, below are calculated the originality values given by Equation (9) for each of the following five tours of a TSP problem with n = 10 cities. The five closed tours are:
s 1 = { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 , 1 } s 2 = { 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 , 1 } s 3 = { 6 , 7 , 8 , 9 , 10 , 1 , 2 , 3 , 5 , 4 , 6 } s 4 = { 1 , 8 , 10 , 9 , 2 , 3 , 4 , 5 , 6 , 7 , 1 } s 5 = { 2 , 9 , 3 , 6 , 5 , 8 , 7 , 10 , 4 , 1 , 2 }
To compute the originality values for each tour, we use the matrix ν corresponding to the five tours:
ν = 0 4 0 0 0 0 0 1 0 0 0 0 4 0 0 0 0 0 1 0 0 0 0 3 1 1 0 0 0 0 1 0 0 0 3 1 0 0 0 0 0 0 0 1 0 3 0 1 0 0 0 0 0 0 1 0 4 0 0 0 1 0 0 0 0 0 0 3 0 1 0 0 0 0 0 0 1 0 3 1 0 1 1 0 0 0 0 0 0 3 3 0 0 1 0 0 0 0 1 0
where each element ν i , j indicates the number of ants that have traversed a r c ( i , j ) . For instance, ν 1 , 2 = 4 indicates that four ants (in this example, those associated with the tours s 1 , s 2 , s 3 and s 5 ) have gone from node 1 to node 2.
According to the originality function given by Equation (9), the following values are obtained: f o ( s 5 ) = 0.076 > f o ( s 4 ) = 0.045 > f o ( s 3 ) = 0.037 > f o ( s 1 ) = f o ( s 2 ) = 0.030 . These results are in accordance with what could be expected, since s 1 and s 2 are identical, s 3 and s 4 have several arcs in common with s 1 and s 2 , while tour s 5 is the most different (original) tour from the rest.

3.2. Pheromone Smoothing Mechanism

This section presents a pheromone smoothing strategy to escape from the stagnation situation. For a better understanding of the stagnation situation, visual representations of the pheromone trails obtained with AS R a n k when solving a TSP instance with 14 cities after 0, 5, 10 and 200 iterations are shown in Figure 2. Note that the pheromone matrix is a square matrix of dimension n × n , where n is the number of nodes (cities) and in this TSP example, n = 14 . Each element of the pheromone matrix τ i , j contains the pheromone quantity associated with the edge that goes from node i to node j. For the representations, the pheromone levels are translated into gray scale, where black and white represent the highest and the lowest pheromone trail values, respectively. At the beginning of the algorithm, all the pheromone values are initialized with the same pheromone value τ 0 , with the exception of the values of the diagonal representing no movement (i.e., i = j ), which are not allowed in TSP (and thus are represented in white in Figure 2).
After each AS R a n k iteration, all the pheromone trails are evaporated and some pheromone trails are reinforced (the ones included in the w 1 best tours of the iteration and the best-found tour). The effect of this pheromone update can be observed in the pheromone representations of Figure 2. Additionally, the symmetry along the main diagonal of the pheromone matrix is due to the fact that the example is a symmetric TSP instance. As the difference among the pheromone values increases, ants are more likely to choose those arcs that were chosen by the best ants from previous iterations. This pheromone reinforcement enables AS R a n k to explore promising regions of the search space obtaining high-quality solutions in a reasonable time. Yet, this usually leads AS R a n k into early stagnation situations, in which all ants follow the same path. Stagnation can be clearly observed in the pheromone trails after 100 iterations shown in Figure 2 (right-bottom). In this case, the pheromone encodes a unique path (only fourteen of all the possible elements of half of the pheromone matrix have a non-negligible pheromone quantity). For this simple TSP instance with only fourteen cities, the encoded path in the pheromone table happens to be the optimal one. However, this is not usually the case and often AS R a n k is trapped into a local minimum.
The proposed pheromone smoothing mechanism, given by Equation (10), rescales the pheromone trails to the range [ γ τ 0 , τ 0 ] , 0 < γ < 1 . Note that if γ equals 1, all the pheromones will be reinitialized to the constant value τ 0 , and previously learned information would be lost.
τ = γ τ 0 + τ m i n ( τ ) m a x ( τ ) m i n ( τ ) ( 1 γ ) τ 0
Figure 3 shows an example of the resulting pheromone trails of A S r a n k before and after applying the pheromone smoothing strategy. The image on the left of Figure 3 shows the pheromone trails when the algorithm is near a stagnation situation. The noticeable difference among the pheromone values can be observed, leading to the construction of very similar ant tours and thus a limited exploration of the search space. The image on the right of Figure 3 shows the pheromone trails after applying the pheromone smoothing strategy, where the pheromones are rescaled [ 0.1 τ 0 , τ 0 ] with τ 0 = m / f ( s s b ) [14]. It can be observed that after applying the pheromone smoothing mechanism, the learned information is still preserved, but the differences among the pheromone trails τ i , j have decreased. In this way, the diversity of the ant tours produced by the algorithm after the pheromone smoothing mechanism is increased, allowing the algorithm to escape from stagnation without losing previously learned information.
In order to incorporate the smoothing strategy into an ACO algorithm, it is also necessary to define when the pheromone smoothing is triggered. For this purpose, it is necessary to select a measure of how close the algorithm is to stagnation and based on this measure, define a criterion that triggers the pheromone smoothing strategy. Several measures can be used to describe the amount of exploration an ACO algorithm performs in order to detect stagnation situations, such as the standard deviation of the length of the tours, the average number of arcs that are not common in every pair of ant tours, and the average λ -branching factor [14]. Despite these last two criteria being good indicators of the size of the search space effectively being explored, they require a high computational cost. For instance, the average λ -branching factor is used in [11] for triggering occasional pheromone initialization of MMAS, but due to its high computational cost, it is only calculated every 100 algorithm iterations. On the contrary, the standard deviation between the fitness of ant’s tours is a more efficient way to measure the diversity among the solutions generated at an iteration. The standard deviation, σ L , of the ant tours of an iteration decreases when the search starts exploiting a concrete region of the search space. Additionally, in the case where σ L reaches a null value, we can assume that all ants are taking the same paths (as it is very unlikely that different tours have the same length) and thus, σ L = 0 is an indicator that the algorithm has fallen in a stagnation situation. Due to its lower computational requirement, σ L is the criterion selected for the proposed ACO algorithm.
Lastly, it is worth clarifying the difference between the discussed dissimilarity functions, which measure the dissimilarity of a group of ants tours, with respect to the originality function presented in Section 3.1, which measures the dissimilarity between a unique solution with respect to several of tours. Additionally, while the former are generally used with the intention of detecting the approach to stagnation situation, the latter is proposed with the intention of promoting solutions that diversify the search.

3.3. Algorithm

Once the originality utility function and the pheromone smoothing strategy have been introduced, below is described the proposed extension of Rank-based Ant System, which incorporates both strategies. Algorithm 2 shows the pseudocode of the algorithm and Figure 4 its general overflow. The algorithm requires similar parameters to AS R a n k : pheromone evaporation rate ρ , the number of ants per iteration m, the number of ants that deposit pheromones w, and the pheromone and heuristic influence parameters α and β .
The steps of the algorithm are the following. First, the algorithm starts with the initialization of the pheromone table to a constant value τ 0 , and of the matrix ν to zero (as ν i , j accounts for the number of ants that have traversed each edge a r c ( i , j ) in the graph G). Within the main iteration loop, the tours of m ants are constructed component by component according to the ACO probabilistic transition rule given by Equation (1). Additionally, in line 7, the counter of ants ν i , j that have traversed a r c ( i , j ) is updated. Once the m tours ( s 1 : m ) have been constructed, they are ranked and selected according to their fitness and originality following the steps sketched in Figure 5 and explained as follows. First, all the tours are evaluated and the ( w 1 ) best are selected (line 11). Then, the ( w 1 ) best tours of the iteration are evaluated and ranked according to the originality utility function given by Equation (9), obtaining in this way a rank of the dissimilarity of the ( w 1 ) best tours with respect all the explored tours (line 13). Moreover, in the case that the best solution of the iteration is better than the global best solution, s g b is updated (line 12). In line 14, the pheromone update process takes place according to Equation (11); all pheromone values are evaporated, and the ( w 1 ) best tours of the current iteration and the global best tour are reinforced.
τ i , j = ( 1 ρ ) τ i , j + r o = 1 w 1 ( w r o ) τ i , j r + w τ i , j g b
The best-so-far tour gives the strongest feedback, as its contribution is multiplied by the weight equal to w, the ( w 1 ) best ants of the iteration deposit a pheromone quantity multiplied by a weight ( w r o ), which is higher for the more original solutions (lower r o ). The update of the pheromones is similar to the original AS R a n k , with the exception that the ranking considered for the reinforcement is a ranking of originality, instead of fitness. In this way, the proposed algorithm rewards the originality of the tours with good fitness values.
Algorithm 2 AS R a n k with originality reinforcement and pheromone smoothing
Require: 
ACO parameters
1:
τ initialise pheromone trails
2:
ν initialise matrix ν
3:
while termination condition not met do
4:
    for  k = 1 to m do                                 ▹ for every ant
5:
        for every step until the k-th ant has completed the tour do
6:
               select next node j according to the transition rule                 ▹ Equation (1)
7:
                ν i , j = ν i , j + 1
8:
               (if symmetric) ν j , i = ν i , j
9:
        end for
10:
    end for
11:
     s 1 : ( w 1 ) E v a l u a t e A n d S e l e c t B e s t ( s 1 : m )                ▹ select ( w 1 ) best ant tours
12:
     s g b S e l e c t B e s t ( s 1 , s g b )                          ▹ update global-best solution
13:
     r o E v a l u a t e O r i g i n a l i t y ( s 1 : ( w 1 ) , ν )               ▹ evaluate originality, Equation (9)
14:
     τ i , j = ( 1 ρ ) τ i , j + r = 1 w 1 ( w r o ) τ i , j r + w τ i , j g b                 ▹ Update pheromones
15:
    (if symmetric) τ j , i = τ i , j
16:
    if approach stagnation then
17:
        pheromone smoothing                               ▹ Equation (10)
18:
    end if
19:
end while
Furthermore, at the end of each algorithm iteration, it is checked whether the algorithm is entering a stagnation situation (line 16). The criterion considered for triggering the pheromone smoothing strategy described by Equation (10) is when the standard deviation of the fitness of the 80% of ants tours of the iteration is zero. Hence, the pheromone smoothing is applied before the algorithm enters stagnation. Finally, once the termination condition is met (maximum number of iterations) the algorithm returns the global best tour.
Furthermore, as in symmetric TSP instances, it is assumed that τ i , j = τ j , i ( i , j ) ; ACO algorithms ensure this by updating the symmetric pheromone element (line 15). Similarly, in the case of symmetric TSP, ν i , j counts the number of ants that have traversed either a r c ( i , j ) or a r c ( j , i ) (line 8).

4. Experimental Results

This section analyses the performance of the proposed extension of Rank-based Ant System. First, the evaluation methodology is described, including the utilized benchmarks and the parameter settings. Next, the performance of the proposed ACO algorithm is analysed and compared with AS, the original AS R a n k , and with MMAS and ACS, which are widely considered the best-performing ACO algorithms according to multiple sources (see, for instance, [14,15,16]).

4.1. TSP and SOP Benchmarks

The following combinatorial optimization problems are used during the analysis:
  • Symmetric and Asymmetric Travelling Salesman Problem (TSP and ASTP) aim to find the Hamiltonian cycle of minimum length given a graph with n cities. In case the distances between the cities are independent of the direction of traversing the edges d i , j = d j , i , ( i , j ) , the problem is known as symmetric TSP; otherwise—as asymmetric TSP.
  • Sequential Ordering Problem (SOP) consists of finding a Hamiltonian path of the minimal length from node 1 to node n taking precedence constraints into account. The precedence constraints impose that some nodes have to be visited before some other nodes of the graph G.
The characteristics of TSP and SOP instances can be expressed by a distance matrix, where each element d i , j states the distance from node i to node j, or in SOP instances it may be equal to 1 , indicating that node j must precede node i. For instance, as SOP imposes that all tours start at node 1 and end at node n, d n , j = 1 , j . Additionally, in symmetric TSP instances, the matrix is symmetric as d i , j = d j , i , ( i , j ) .
Table 1 lists the relevant information about the instances used for the analysis: their name, the type of optimization problem, the fitness of the optimal solution and the number of nodes (dimension of the problem instance). These instances belong to the TSPLIB benchmark library [38], which is accessible at http://comopt.ifi.uni-heidelberg.de/software/TSPLIB95/ (accessed on 24 October 2022).

4.2. Experimental Setup

Due to their stochastic nature, all ACO algorithms were run 30 times for each problem instance. During each run, the length of the best-found solution at each iteration is saved. This information is later used for comparing the performance of the algorithms by means of comparative tables and evolutionary curves (see Section 4.3 and Section 4.4 for details).
Regarding the parameter settings, typically recommended values, already used in several previous studies, were selected [11,13,14,18]. The following pheromone and heuristic influence parameters were set for all ACO algorithms; α = 1 , β = 2 . The number of ants m was set equal to the number of cities n for all ACO algorithms with exception of ACS, leading to good algorithm performance (see [14] for details). In the case of ACS, the number of ants m per iteration was set equal to 10, as recommended in [12,14]. Regarding the values of ACS-specific parameters, the following recommended values were considered: q 0 = 0.9 , ξ = 0.1 [12,14]. Regarding Rank-based ACO variants, the following parameters were considered: w = 6 , ρ = 0.1 and τ 0 = m / f ( s g b ) , also selected in previous studies [14]. Furthermore, in the case of MMAS, the evaporation parameter is set to ρ = 0.02 , the iteration-best solution is considered for pheromone reinforcement, and pheromone trails reinitialization to τ m a x is triggered whenever no improved tour has been found for more than 250 iterations [14].
With regard to the parameter settings of the proposed algorithm, the common parameters with A S R a n k are set with the same values for a fair comparison. Additionally, the pheromone smoothing strategy rescales the pheromone table to [ 0.1 τ 0 , τ 0 ] whenever the standard deviation of the lengths of the 80% of the ant tours of the iteration is zero. This criterion was chosen empirically, after observing that in most of the cases, AS R a n k is not able to find a better solution once 80% of the ants have followed the same path (i.e., have the same fitness). The results of the analysis are shown in Table 2, where for each problem instance, the number of simulations (from a total of 30 runs) is shown in which AS R a n k has been able to find a better solution after different percentages d of solutions (from d = 10%, 20%, , to 100%) have reached null standard deviation (i.e., after the d% of the ants have had the same fitness). The objective of this analysis is to find a good balance, avoiding applying the smoothing mechanism too late (wasting a lot of algorithm iterations with little chance of improving the best solution found so far) or, on the contrary, applying the smoothing mechanism too early (where the pheromone information could still be further exploited to find a better solution). The results of Table 2 show that quite often, AS R a n k is able to find a better solution after 10% of the m solutions have the same fitness, e.g., 22 of the 30 runs for brazil58 instance, or 17 of the 30 runs in kroA100 instance. As we would expect, the higher percentage of solutions considered, the closer it gets to a stagnation situation and the harder it is for the algorithm to find a better solution. For instance, in the case of brazil58, the algorithm was able to find a better solution in 22 of 30 runs after 10% of the m solutions were the same, but only in 6 of the 30 runs after 30% of solutions were the same, and after all the solutions were the same at some iteration, the algorithm was no longer able to improve the best-found-solution in any of the 30 runs.
Moreover, for all ACO algorithms, a maximum number of iterations was considered as a stopping criterion; 800 iterations for problem instances with dimensions lower than 100 nodes and 1200 iterations for bigger instances. Additionally, it is worth mentioning that all implemented ACO algorithms do not consider local search, as the goal is to analyse the performance of the proposed approach. Hence, the results obtained could be improved by the consideration of local search strategies typically used for TSP (e.g., k-exchange neighbourhood), see [11]. This will be part of our future work in the field (see Section 5 for further details).

4.3. Analysis of the Proposed ACO Algorithm

With the purpose of analysing the benefits of the extended Rank-based Ant System with originality reinforcement and pheromone smoothing, denoted onwards as AS O R a n k p s , with respect to the original AS R a n k , this section compares their performance over several TSP and SOP instances. Furthermore, in order to analyse how each of the two new proposed strategies contributes to the performance of the algorithm, we also analyse the performance of the algorithm when only one of the two strategies is considered—either with the originality reinforcement (denoted as AS O R a n k ), or with the pheromone smoothing strategy (denoted as AS R a n k p s ).
The comparative results obtained by these four variants of Rank-based Ant System are shown in Table 3 for the symmetric and asymmetric TSP instances, and in Table 4 for the SOP instances. Both tables are organized in a similar way: the different instances are shown in the first column; for each instance, columns 2–8 show: the algorithm used for execution (second column), the mean length of the best-ant tours and its standard deviation over 30 independent runs of the algorithm (third and fourth columns, respectively), the fitness of the best solution found by the 30 runs (fifth column), the average percentage deviation of the solutions from the optimal solution given by PD a v g = ( A v g O p t ) / O p t · 100 (sixth column) and the percentage deviation of the length of the best solution of the 30 runs from the optimal solution, given by PD b e s t = ( B e s t O p t ) / O p t · 100 (seventh column) and computed according to [30], and lastly, the mean computational time over the independent runs of the algorithms in seconds (eighth column). Furthermore, for each problem instance, the best (lowest) average length tour (Avg.), best-found tour (Best), average (PD a v g ) and best (PD b e s t ) percentage deviation are highlighted in bold.
The comparative results show that the proposed Rank-based extension AS O R a n k p s clearly outperforms the original AS R a n k , obtaining lower mean path lengths in all TSP, ATSP and SOP instances. For example, in the case of the symmetric TSP instance ch130, AS R a n k obtained a mean path length of 6235, which corresponds to an average percentage of deviation from the optimal solution of PD a g v = 2.04 % , while AS O R a n k p s obtained a mean path length of 6170 and PD a g v = 0.98 % . Moreover, for the same problem instance, the fitness of the best solution found during the 30 runs by AS R a n k was 6169, which corresponds to a percentage of deviation from the optimal solution of PD b e s t = 0.97 % , while AS O R a n k p s obtained a path length of 6136 and PD b e s t = 0.43 % . Additionally, no significant differences are observed between the computational times.
Additionally, several conclusions can be drawn from the analysis of the AS O R a n k and AS R a n k p s variants. On the one hand, the results from Table 3 and Table 4 show that the pheromone smoothing strategy by itself (AS R a n k p s ) clearly improves the performance of the Rank-based algorithm, showing better results than AS R a n k in all the problem instances. On the other hand, the inclusion of the originality reinforcement strategy in Rank-based Ant System showed an improvement in the results. AS O R a n k variant obtained better results than AS R a n k in 7 of the 12 problem instances, worse in 4 problem instances (brazil58, ch130, gr48 and tsp95) and similar performance in (ftv35). However, the ACO variant that considers both the originality reinforcement and the pheromone smoothing (AS O R a n k p s ) obtained the best results out of all considered algorithms in the majority of the 12 instances, (with the exception of brazil58, kroA100 and p43.4 where AS R a n k p s obtained better results). Therefore, we can conclude that the originality reinforcement strategy is clearly beneficial when combined with the pheromone smoothing strategy in the proposed AS O R a n k p s .
To sum up, we can conclude that AS O R a n k p s outperformed the original AS R a n k in the fourteen problem instances thanks to both the pheromone smoothing and the originality reinforcement strategies.

4.4. Comparison with ACO Algorithms

The comparative Table 5 and Table 6, and the evolutionary curves in Figure 6 and Figure 7 show the performance of several ACO algorithms: Ant System (AS) [9], Rank-based Ant System (AS R a n k ) [10], Max-Min Ant System (MMAS) [11], Ant Colony System (ACS) [12] and the proposed extension of the Rank-based Ant System (AS O R a n k p s ).
On the one hand, Table 5 and Table 6 allow the comparison of the final solutions obtained by the ACO algorithms, showing that AS O R a n k p s outperforms all analysed ACO algorithms, including MMAS and ACS which are the most widely used and often best performing ACO algorithms [14,15,16]. More concretely, AS O R a n k p s obtained better results than AS and AS R a n k in all problem instances, better results than ACS in all instances with the exception of brazil58, and better results than MMAS in all SOP instances, and all TSP instances with the exception of ftv64 and ftv70, where both algorithms obtained similar results. Regarding the computational times, ACS requires slightly longer computational times than the rest of the algorithms.
On the other hand, the evolutionary curves of Figure 6 and Figure 7, which represent the evolution of the average best-found tour lengths along the number of tours constructed, enable the comparison of the performance of the ACO algorithms along the runs of the algorithms, and not just the comparison of the fitness of final solutions. Note that these graphs allow to compare the performance of the algorithms at different iterations, but due to the scale, the comparison of the fitness of the final solutions is better observed in the comparative Table 5 and Table 6.
The evolutionary curves of Figure 6 and Figure 7 show that during the initial iterations, all ACO algorithms outperform MMAS, which presents a slower convergence to high-quality solutions. For example, at iteration 100 of the TSP instance brazil58, the mean percentage deviations from the optimal solution of AS R a n k and AS O R a n k p s are, respectively, 1.26% and 1.6%, while MMAS clearly presents a higher error equal to 26.2 % . The reason for the observed behaviour is that, while MMAS relies on a high initial exploration of the search space in order to finally converge to high-quality solutions, AS O R a n k p s maintains the rapid convergence of AS R a n k and relies on search diversification of the originality reinforcement and pheromone smoothing mechanism to improve the quality of the solutions. The behaviour of AS O R a n k p s can be especially beneficial in those optimization problems where the available computational time is limited or not known beforehand, as its faster convergence would return a significantly better solution if the number of iterations is restricted.
To sum up the conclusions from comparative tables and evolutionary graphs, AS O R a n k p s is able to find higher-quality solutions than AS and AS R a n k in all the problem instances and in all except one instance when compared with ACS. Moreover, AS O R a n k p s showed a much faster convergence than MMAS, while being able to reach better final solutions in most of the problem instances.

5. Conclusions and Future Work

This work proposes an extension of the Rank-based Ant System that incorporates originality reinforcement and pheromone smoothing strategies in order to avoid early stagnation and improve the quality of the solutions. The performance of the approach is analysed and compared with AS, AS R a n k , MMAS, and ACS, over TSP and SOP benchmarks showing very promising results. Contrary to the occasional reinitialization of pheromone trails employed by MMAS [14,17], the proposed algorithm relies strongly on search diversification through frequent smoothing of the pheromone trails. Therefore, the promising results obtained by the proposed approach are not due only to the pheromone smoothing mechanism, but also to the use of the proposed criterion for triggering the pheromone smoothing. Additionally, the comparative results show that the originality reinforcement strategy helps to improve the quality of the Rank-based Ant System when it is combined with the pheromone smoothing strategy. Together, they enable the proposed approach to maintain the rapid convergence to high-quality solutions during the first iterations of AS R a n k , in addition to reaching higher-quality solutions in the long run. These features can be advantageous in those optimization problems where the available time for computation is limited or not known beforehand.
In this regard, we consider as an interesting future research line the adaptation of the proposed approach to dynamic optimization problems (where the search domain changes over time). For example, the rapid convergence of the proposed ACO algorithm could be advantageous in Dynamic TSP [14,39], where new cities can be added or removed during run time and the new shortest tour after each transition should be found as quickly as possible. We also consider the analysis of the proposed method with other combinatorial optimization problems such as the Vehicle Routing Problem and different modifications of the original TSP [14].
Other interesting potential extension of this work is its hybridization with local search strategies and/or other metaheuristics for better performance. As previously indicated in Section 4.2, the present method does not include any local search, since our primary goal is to enhance the pure state-of-the-art ACO algorithms without the addition of other heuristics. However, it has been remarked by several authors that adding sophisticated local search procedures to the ACO algorithms can improve their performance for certain problems significantly [17]. In line with this, we plan to combine our method with different local search strategies and analyse their performance for the TSP, SOP and other combinatorial optimization problems. We also plan to compare the performance of our method with other ACO implementations after they are all paired with those local search procedures, under the premise that the superior behaviour of our method with respect to best ACO performers without any local search should remain when the local search is added to all methods in a similar fashion. Other interesting extensions are the comparison of this new ACO algorithm with other powerful metaheuristics, such as genetic algorithms, particle swarm optimization or differential evolution, as well as analysing whether the performance of our method can be improved by its hybridization with some of those metaheuristics. These ideas are all part of our plans for future work in the field.

Author Contributions

Conceptualization, S.P.-C. and A.G.; methodology, software, and validation, S.P.-C., A.G. and A.I.; formal analysis, S.P.-C. and A.G.; investigation, S.P.-C. and A.I.; resources, S.P.-C.; data curation, S.P.-C. and A.I.; writing—original draft preparation, S.P.-C.; writing—review and editing, S.P.-C., A.G. and A.I.; supervision, project administration and funding acquisition, A.G. and A.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research work was funded by the European project PDE-GIR of the European Union’s Horizon 2020 research & innovation program (Marie Sklodowska-Curie action, grant agreement No 778035), and by the Spanish government project #PID2021-127073OB-I00 of the MCIN/AEI/10.13039/501100011033/FEDER, EU “Una manera de hacer Europa”.

Data Availability Statement

The data used for the analysis belong to the TSPLIB benchmark library [38], which is accessible at http://comopt.ifi.uni-heidelberg.de/software/TSPLIB95/ (accessed on 24 October 2022).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Holland, J.H. Adaptation in Natural and Artificial Systems; The University of Michigan Press: Ann Arbor, MI, USA, 1975. [Google Scholar]
  2. Goldberg, D.E. Genetic Algorithms in Search, Optimization and Machine Learning; Kluwer Academic Publishers: Norwell, MA, USA, 1989. [Google Scholar]
  3. Kennedy, J.; Eberhart, R.C.; Shi, Y. Swarm Intelligence; Morgan Kaufman: San Francisco, CA, USA, 2001. [Google Scholar]
  4. Engelbretch, A.P. Fundamentals of Computational Swarm Intelligence; John Wiley and Sons: Chichester, UK, 2005. [Google Scholar]
  5. Glover, F.; Kochenberger, G.A. Handbook of Metaheuristics; International Series in Operations Research & Management Science; Springer: Cham, Switzerland, 2003; Volume 57. [Google Scholar]
  6. Gendreau, M.; Potvin, J.Y. Handbook of Metaheuristics; International Series in Operations Research & Management Science; Springer: Cham, Switzerland, 2010; Volume 146. [Google Scholar]
  7. Gendreau, M.; Potvin, J.Y. Handbook of Metaheuristics, 3rd ed.; International Series in Operations Research & Management Science; Springer: Cham, Switzerland, 2018; Volume 272. [Google Scholar]
  8. Dorigo, M. Optimization, Learning and Natural Algorithms. Ph.D. Thesis, Politecnico di Milano, Milan, Italy, 1992. [Google Scholar]
  9. Maniezzo, V.; Dorigo, M.; Colorni, A. The ant system: Optimization by a colony of cooperating agents. IEEE Trans. Syst. Man Cybern. Part 1996, 26, 29–41. [Google Scholar]
  10. Bullnheimer, B.; Hartl, R.F.; Strauss, C. A new rank-based version of the Ant System: A computational study. Cent. Eur. J. Oper. Res. Econ. 1999, 7, 25–38. [Google Scholar]
  11. Stützle, T.; Hoos, H.H. MAX–MIN ant system. Future Gener. Comput. Syst. 2000, 16, 889–914. [Google Scholar] [CrossRef]
  12. Dorigo, M.; Gambardella, L.M. Ant colony system: A cooperative learning approach to the traveling salesman problem. IEEE Trans. Evol. Comput. 1997, 1, 53–66. [Google Scholar] [CrossRef] [Green Version]
  13. Cordón, O.; Viana, I.F.D.; Herrera, F. Analysis of the best-worst ant system and its variants on the QAP. In International Workshop on Ant Algorithms; Springer: Berlin/Heidelberg, Germany, 2002; pp. 228–234. [Google Scholar]
  14. Dorigo, M.; Stützle, T. Ant Colony Optimization; MIT Press: Cambridge, MA, USA, 2004. [Google Scholar]
  15. Blum, C. Ant colony optimization: Introduction and recent trends. Phys. Life Rev. 2005, 2, 353–373. [Google Scholar] [CrossRef] [Green Version]
  16. Cordón García, O.; Herrera Triguero, F.; Stützle, T. A review on the ant colony optimization metaheuristic: Basis, models and new trends. Mathw. Soft Comput. 2002, 9, 141–175. [Google Scholar]
  17. Dorigo, M.; Stützle, T. Ant Colony Optimization: Overview and recent advances. In Handbook of Metaheuristics; Springer International Publishing: New York, NY, USA, 2019; pp. 311–351. [Google Scholar]
  18. Stützle, T.; Dorigo, M. ACO algorithms for the traveling salesman problem. In Evolutionary Algorithms in Engineering and Computer Science; John Wiley & Sons: New York, NY, USA, 1999; pp. 163–183. [Google Scholar]
  19. Pérez-Carabaza, S.; Besada-Portas, E.; López-Orozco, J.A.; Jesus, M. Ant colony optimization for multi-UAV minimum time search in uncertain domains. Appl. Soft Comput. 2018, 62, 789–806. [Google Scholar] [CrossRef]
  20. Singh, S.S.; Singh, K.; Kumar, A.; Biswas, B. ACO-IM: Maximizing influence in social networks using ant colony optimization. Soft Comput. 2020, 24, 10181–10203. [Google Scholar] [CrossRef]
  21. Banerjee, A.; Kumar De, S.; Majumder, K.; Das, V.; Giri, D.; Shaw, R.N.; Ghosh, A. Construction of effective wireless sensor network for smart communication using modified ant colony optimization technique. In Advanced Computing and Intelligent Technologies; Lecture Notes in Networks and Systems; Bianchini, M., Piuri, V., Das, S., Shaw, R.N., Eds.; Springer: Singapore, 2022. [Google Scholar]
  22. Ding, Q.; Hu, X.; Sun, L.; Wang, Y. An improved ant colony optimization and its application to vehicle routing problem with time windows. Neurocomputing 2012, 98, 101–107. [Google Scholar] [CrossRef]
  23. Lee, Z.J.; Su, S.F.; Chuang, C.C.; Liu, K.H. Genetic algorithm with ant colony optimization (GA-ACO) for multiple sequence alignment. Appl. Soft Comput. 2008, 8, 55–78. [Google Scholar] [CrossRef]
  24. Zhao, F.T.; Yao, Z.; Luan, J.; Song, X. A novel fused optimization algorithm of genetic algorithm and ant colony optimization. Math. Probl. Eng. 2016, 2016, 2167413. [Google Scholar] [CrossRef] [Green Version]
  25. Mohsen, A.M. Annealing Ant Colony Optimization with mutation operator for solving TSP. Comput. Intell. Neurosci. 2016, 2016, 8932896. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Qamar, M.S.; Tu, S.; Ali, F.; Armghan, A.; Munir, M.F.; Alenezi, F.; Muhammad, F.; Ali, A.; Alnaim, N. Improvement of traveling salesman problem solution using hybrid algorithm based on best-worst ant system and particle swarm optimization. Appl. Sci. 2021, 11, 4780. [Google Scholar] [CrossRef]
  27. Herlambang, T.; Rahmalia, D.; Yulianto, T. Particle Swarm Optimization (PSO) and Ant Colony Optimization (ACO) for optimizing PID parameters on Autonomous Underwater Vehicle (AUV) control system. J. Phy. Conf. Ser. 2018, 1211, 012039. [Google Scholar] [CrossRef]
  28. Meenachi, L.; Ramakrishnan, S. Differential evolution and ACO based global optimal feature selection with fuzzy rough set for cancer data classification. Soft Comput. 2020, 24, 18463–18475. [Google Scholar] [CrossRef]
  29. Zhang, X.; Duan, H.; Jin, J. DEACO: Hybrid Ant Colony Optimization with Differential Evolution. In Proceedings of the 2008 IEEE Congress on Evolutionary Computation, IEEE CEC 2008, Hong Kong, China, 1–6 June 2008; IEEE Computer Society Press: Los Alamitos, CA, USA, 2008; pp. 921–927. [Google Scholar]
  30. Gao, W. New ant colony optimization algorithm for the traveling salesman problem. Int. J. Comput. Intell. Syst. 2020, 13, 44–55. [Google Scholar] [CrossRef] [Green Version]
  31. Stützle, T.; López-Ibánez, M.; Pellegrini, P.; Maur, M.; de Oca Montes, M.; Birattari, M.; Dorigo, M. Parameter adaptation in ant colony optimization. In Autonomous Search; Springer: Berlin/Heidelberg, Germany, 2011; pp. 191–215. [Google Scholar]
  32. Mavrovouniotis, M.; Yang, S.; Van, M.; Li, C.; Polycarpou, M. Ant colony optimization algorithms for dynamic optimization: A case study of the dynamic travelling salesperson problem. IEEE Comput. Intell. Mag. 2020, 15, 52–63. [Google Scholar] [CrossRef] [Green Version]
  33. Cruz, C.; González, J.R.; Pelta, D.A. Optimization in dynamic environments: A survey on problems, methods and measures. Soft Comput. 2011, 15, 1427–1448. [Google Scholar] [CrossRef]
  34. Guntsch, M.; Middendorf, M. Applying population based ACO to dynamic optimization problems. In Proceedings of the Third International Workshop on Ant Algorithms (ANTS 2002), Brussels, Belgium, 12–14 September 2002. [Google Scholar]
  35. Mavrovouniotis, M.; Yang, S. Ant colony optimization with immigrants schemes for the dynamic travelling salesman problem with traffic factors. Appl. Soft Comput. 2013, 13, 4023–4037. [Google Scholar] [CrossRef]
  36. Mavrovouniotis, M.; Yang, S. Ant colony optimization with memory-based immigrants for the dynamic vehicle routing problem. In Proceedings of the IEEE Congress on Evolutionary Computation, IEEE-CEC 2012, Brisbane, Australia, 10–15 June 2012; pp. 1–8. [Google Scholar]
  37. Oliveira, S.M.; Bezerra, L.C.T.; Stützle, T.; Dorigo, M.; Wanner, E.F.; Souza, S.R. A computational study on ant colony optimization for the traveling salesman problem with dynamic demands. Comput. Oper. Res. 2021, 135, 105359. [Google Scholar] [CrossRef]
  38. Reinelt, G. TSPLIB–A traveling salesman problem library. Orsa J. Comput. 1991, 3, 376–384. [Google Scholar] [CrossRef]
  39. Mavrovouniotis, M.; Müller, F.M.; Yang, S. Ant colony optimization with local search for dynamic traveling salesman problems. IEEE Trans. Cybern. 2016, 47, 1743–1756. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Evolutionary curves of ACO algorithms (AS, AS R a n k , MMAS and ACS) for TSP instance gr48.
Figure 1. Evolutionary curves of ACO algorithms (AS, AS R a n k , MMAS and ACS) for TSP instance gr48.
Applsci 12 11219 g001
Figure 2. A visual representation of the pheromone matrix: (left-right, top-bottom) the pheromone values obtained with AS R a n k applied to the TSPLIB instance burma14 with 14 cities, after 0, 5, 10 and 100 iterations, respectively.
Figure 2. A visual representation of the pheromone matrix: (left-right, top-bottom) the pheromone values obtained with AS R a n k applied to the TSPLIB instance burma14 with 14 cities, after 0, 5, 10 and 100 iterations, respectively.
Applsci 12 11219 g002aApplsci 12 11219 g002b
Figure 3. Pheromone trails before (on the left) and after (on the right) of pheromone smoothing.
Figure 3. Pheromone trails before (on the left) and after (on the right) of pheromone smoothing.
Applsci 12 11219 g003
Figure 4. Algorithm flowchart of the proposed AS O R a n k p s method.
Figure 4. Algorithm flowchart of the proposed AS O R a n k p s method.
Applsci 12 11219 g004
Figure 5. Steps of the process of selection and ranking of solutions for the pheromone reinforcement (where s k i represents a solution component of the tour of length l of the k-th ant).
Figure 5. Steps of the process of selection and ranking of solutions for the pheromone reinforcement (where s k i represents a solution component of the tour of length l of the k-th ant).
Applsci 12 11219 g005
Figure 6. Evolutionary curves of ACO algorithms for symmetric and asymmetric TSP instances.
Figure 6. Evolutionary curves of ACO algorithms for symmetric and asymmetric TSP instances.
Applsci 12 11219 g006aApplsci 12 11219 g006b
Figure 7. Evolutionary curves of ACO algorithms for asymmetric TSP instances (ftv64 and ftv70) and four SOP instances.
Figure 7. Evolutionary curves of ACO algorithms for asymmetric TSP instances (ftv64 and ftv70) and four SOP instances.
Applsci 12 11219 g007aApplsci 12 11219 g007b
Table 1. Problem instances used in this paper for the experiments and analysis.
Table 1. Problem instances used in this paper for the experiments and analysis.
InstanceTypeOptimal SolutionDimension (n)
brazil58TSP25,39558
kroA100TSP21,282100
ch130TSP6110130
tsp225TSP3916335
gr48TSP504648
pr76TSP108,15976
gr202ATSP40,160202
ftv35ATSP147336
ftv64ATSP183965
ftv70ATSP195071
ESC78SOP18,23080
ft70.1SOP39,31371
p43.1SOP28,14044
p43.4SOP83,00544
Table 2. Analysis of the number of times from the 30 runs that AS R a n k was able to find a better solution after different percentages of solutions (ants) have reached null standard deviation.
Table 2. Analysis of the number of times from the 30 runs that AS R a n k was able to find a better solution after different percentages of solutions (ants) have reached null standard deviation.
InstancePercentage (%)
102030405060708090100
brazil5822765321110
kroA100171297433221
ch130181186543330
tsp225211814121074100
gr4814742222100
pr7614965544221
gr202161497543210
ftv351814107444333
ftv64171398775310
ftv70211554221100
ESC789655321000
ft70.11813117522100
p43.1151286653210
p43.42520181513106553
Table 3. Results for symmetric and asymmetric TSP instances obtained by Rank-based ACO variants. The best results for each problem instance are highlighted in bold.
Table 3. Results for symmetric and asymmetric TSP instances obtained by Rank-based ACO variants. The best results for each problem instance are highlighted in bold.
InstanceAlgorithmAvg.Std.BestPD avg ( % ) PD best ( % ) Time (s)
brazil58AS R a n k 25,62812625,4000.920.0214
AS R a n k p s 25,48011825,3950.33014
AS O R a n k 25,67712425,4001.110.0213
AS O R a n k p s 25,48712125,3950.36015
kroA100AS R a n k 21,68325221,3061.890.1145
AS R a n k p s 21,35710021,2820.35045
AS O R a n k 21,59118621,3311.450.2343
AS O R a n k p s 21,37812021,2820.45045
ch130AS R a n k 62353961692.040.97120
AS R a n k p s 61934761411.360.51120
AS O R a n k 62414861542.140.72115
AS O R a n k p s 61703361360.980.43120
tsp225AS R a n k 40263139892.801.86422
AS R a n k p s 39492339160.840422
AS O R a n k 40343339783.021.58407
AS O R a n k p s 39421739160.650424
gr48AS R a n k 51174050661.400.409
AS R a n k p s 51043550541.150.169
AS O R a n k 51353950741.760.559
AS O R a n k p s 50912850490.880.0611
pr76AS R a n k 111,6091064109,3923.191.1460
AS R a n k p s 110,3571187108,2382.030.0761
AS O R a n k 111,507810110,1003.101.7960
AS O R a n k p s 109,922900108,1591.63064
gr202AS R a n k 41,80343540,9604.091.99324
AS R a n k p s 41,09525540,6092.331.12325
AS O R a n k 41,60234041,0223.592.15317
AS O R a n k p s 41,05622840,5542.230.98328
ftv35AS R a n k 1497914731.6505
AS R a n k p s 14881014730.9805
AS O R a n k 14971114731.6405
AS O R a n k p s 14831114730.7105
ftv64AS R a n k 18672118481.500.4917
AS R a n k p s 1859818481.070.4917
AS O R a n k 18621418391.23016
AS O R a n k p s 1858918391.05017
ftv70AS R a n k 20104419573.100.3621
AS R a n k p s 19993519572.490.3621
AS O R a n k 19893019501.98020
AS O R a n k p s 19893719891.990.2021
Table 4. Results for SOP instances obtained by Rank-based ACO variants. The best results for each problem instance are highlighted in bold.
Table 4. Results for SOP instances obtained by Rank-based ACO variants. The best results for each problem instance are highlighted in bold.
InstanceAlgorithmAvg.Std.BestPD avg ( % ) PD best ( % ) Time (s)
ESC78AS R a n k 18,60914718,4152.081.01183
AS R a n k p s 18,4704318,4051.320.96185
AS O R a n k 18,58415818,3801.940.82185
AS O R a n k p s 18,4606918,3001.260.38192
ft70.1AS R a n k 41,40341940,6465.323.3981
AS R a n k p s 41,08229740,5054.503.0381
AS O R a n k 41,05333140,5294.433.0980
AS O R a n k p s 40,76834140,0923.701.9881
p43.1AS R a n k 28,33310728,2200.690.2821
AS R a n k p s 28,2556528,2200.410.2821
AS O R a n k 28,33010728,2200.630.2821
AS O R a n k p s 28,2363428,2200.340.2822
p43.4AS R a n k 83,69311983,4150.830.4967
AS R a n k p s 83,4057383,2950.480.3568
AS O R a n k 83,62012083,4050.740.4867
AS O R a n k p s 83,41611383,2650.490.3168
Table 5. Results for symmetric and asymmetric TSP instances obtained by ACO algorithms. The best results for each problem instance are highlighted in bold.
Table 5. Results for symmetric and asymmetric TSP instances obtained by ACO algorithms. The best results for each problem instance are highlighted in bold.
InstanceAlgorithmAvg.Std.BestPD avg ( % ) PD best ( % ) Time (s)
brazil58AS25,9308925,6852.101.1414
AS R a n k 25,62812625,4000.920.0214
MMAS25,6225125,4800.890.3314
ACS25,46411625,3950.27016
AS O R a n k p s 25,48712125,3950.36015
kroA100AS22,71419822,2216.734.4145
AS R a n k 21,68325221,3061.890.1145
MMAS21,46217321,3300.850.2345
ACS21,55927921,2821.30053
AS O R a n k p s 21,37812021,2820.45045
ch130AS66327164828.546.09120
AS R a n k 62353961692.040.97120
MMAS62023461271.480.28121
ACS62344461452.030.57145
AS O R a n k p s 61703361360.980.42120
tsp225AS437457416611.696.38424
AS R a n k 40263139892.801.86420
MMAS39982039622.101.17420
ACS40224239292.720.33504
AS O R a n k p s 39421739160.650424
gr48AS52273651473.582.009
AS R a n k 51174050661.400.409
MMAS51033550631.130.349
ACS50953550460.96012
AS O R a n k p s 50912850490.880.069
pr76AS115,664793113,9116.945.3260
AS R a n k 111,6091064109,3923.191.1460
MMAS110,5211027109,2712.181.0360
ACS110,1571326108,1591.85070
AS O R a n k p s 109,922900108,1591.63064
gr202AS45,74648244,36813.9010.48327
AS R a n k 41,80343540,9604.091.99324
MMAS42,00441341,3314.592.92327
ACS41,64634040,7203.701.39387
AS O R a n k p s 41,05622840,5542.230.98328
ftv35AS15041014872.110.955
AS R a n k 1497914731.6505
MMAS1493814731.3905
ACS14941914731.4506
AS O R a n k p s 14831114730.7105
ftv64AS19181319024.313.4317
AS R a n k 18672118481.500.4917
MMAS1857718541.000.81518
ACS18662118421.850.1621
AS O R a n k p s 1858918391.05017
ftv70AS214923205110.205.1821
AS R a n k 20104419573.100.3621
MMAS19883019501.95021
ACS20444819674.830.8725
AS O R a n k p s 19893719541.990.2021
Table 6. Results for SOP instances obtained by ACO algorithms. The best results for each problem instance are highlighted in bold.
Table 6. Results for SOP instances obtained by ACO algorithms. The best results for each problem instance are highlighted in bold.
InstanceAlgorithmAvg.Std.BestPD avg ( % ) PD best ( % ) Time (s)
ESC78AS20,63122719,95013.179.43182
AS R a n k 18,60914718,4152.081.01184
MMAS18,4642818,4051.290.96187
ACS18,4779718,2901.350.33196
AS O R a n k p s 18,4606918,3001.260.38192
ft70.1AS44,11046943,22112.209.9481
AS R a n k 41,40341940,6465.323.3981
MMAS40,90344140,1924.052.2481
ACS42,56271541,0148.274.3387
AS O R a n k p s 40,76834140,0923.701.9881
p43.1AS28,7766428,6152.261.6921
AS R a n k 28,33310728,2200.690.2821
MMAS28,2585728,2200.420.2821
ACS28,4618828,2451.400.3724
AS O R a n k p s 28,2363428,2200.340.2822
p43.4AS84,21811583,9501.461.1468
AS R a n k 83,69311983,4150.830.4967
MMAS83,51413083,3600.610.4367
ACS83,60114683,2700.720.3269
AS O R a n k p s 83,41611383,2650.490.3168
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Pérez-Carabaza, S.; Gálvez, A.; Iglesias, A. Rank-Based Ant System with Originality Reinforcement and Pheromone Smoothing. Appl. Sci. 2022, 12, 11219. https://doi.org/10.3390/app122111219

AMA Style

Pérez-Carabaza S, Gálvez A, Iglesias A. Rank-Based Ant System with Originality Reinforcement and Pheromone Smoothing. Applied Sciences. 2022; 12(21):11219. https://doi.org/10.3390/app122111219

Chicago/Turabian Style

Pérez-Carabaza, Sara, Akemi Gálvez, and Andrés Iglesias. 2022. "Rank-Based Ant System with Originality Reinforcement and Pheromone Smoothing" Applied Sciences 12, no. 21: 11219. https://doi.org/10.3390/app122111219

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop