Grand Tour Algorithm: Novel Swarm-Based Optimization for High-Dimensional Problems

: Agent-based algorithms, based on the collective behavior of natural social groups, exploit innate swarm intelligence to produce metaheuristic methodologies to explore optimal solutions for diverse processes in systems engineering and other sciences. Especially for complex problems, the processing time, and the chance to achieve a local optimal solution, are drawbacks of these algorithms, and to date, none has proved its superiority. In this paper, an improved swarm optimization technique, named Grand Tour Algorithm (GTA), based on the behavior of a peloton of cyclists, which embodies relevant physical concepts, is introduced and applied to fourteen benchmarking optimization problems to evaluate its performance in comparison to four other popular classical optimization metaheuristic algorithms. These problems are tackled initially, for comparison purposes, with 1000 variables. Then, they are confronted with up to 20,000 variables, a really large number, inspired in the human genome. The obtained results show that GTA clearly outperforms the other algorithms. To strengthen GTA’s value, various sensitivity analyses are performed to verify the minimal inﬂuence of the initial parameters on e ﬃ ciency. It is demonstrated that the GTA fulﬁls the fundamental requirements of an optimization algorithm such as ease of implementation, speed of convergence, and reliability. Since optimization permeates modeling and simulation, we ﬁnally propose that GTA will be appealing for the agent-based community, and of great help for a wide variety of agent-based applications.


Introduction
Many optimization problems in engineering are of a very complex nature and must be solved in accordance with various, sometimes complicated constraints.As a consequence, finding an optimal solution is often hard.In addition, frequently, a large number of mixed variables, and differential and nonlinear equations, are used to describe the problem.As a result, in many cases, classical optimization procedures based on differential methods cannot be used.Metaheuristic techniques arise to bridge this gap, as they can explore the search space for optimal and feasible solutions in a less restrictive (derivative-free) framework.Since in continuous problems, the set of feasible solutions is infinite, metaheuristic algorithms use empirical iterative search methods, based on various heuristics, to guide the search in a way that the solution is expected to always improve.
As described in [1], metaheuristic algorithms are often inspired by nature, and can be classified into the following groups: (i) evolutionary, inspired from biology, where an initial population evolves over generations; (ii) swarm intelligence, inspired from the social behavior of animals; (iii) physics-based, inspired by natural phenomena; and (iv) human related, inspired from human beings and their physical and mental activities.
In swarm intelligence techniques, the social features of a group are used to define the behavior of its individuals, leading to (local) optimal solutions [2].In general, the position of an individual represents a potential solution, and has a defined score based on the objective function of interest.First, initial positions for individuals are defined randomly, and this defines their initial score.Then, an individual evolves according to: (i) its own preferences, based on its past experience, i.e., on its best performance ever achieved, and (ii) according to the group experience, i.e., the best solution ever found by any individual in the group.Each algorithm has its own rules to express this social behavior.Additional random coefficients are used to guarantee better exploration of the search space, especially in the initial period, and those coefficients are also modified across iterations [3] to achieve better exploitation, mainly at the end of the search.In addition, the historical evolutionary information deployed by swarm algorithms can be efficiently used in dynamic optimization to detect the emergent behaviors of any involved agent or element.Examples are algorithms based on ants [4] and bees [5] looking for food, the breeding behavior of cuckoos [6], the behavior of flocking birds [7], among many others.Extensive lists of algorithms belonging to the four abovementioned categories of metaheuristics algorithms can be found in [1].Specifically, Table 2 in [1] compiles 76 swarm intelligence algorithms, together with some specific details.
Physics-based algorithms mimic the behavior of natural phenomena and use the rules governing these phenomena to produce optimal solutions.Perhaps the most popular of these algorithms is simulated annealing [8], based on crystal development in metallurgy.Table 3 in [1] provides a list of 36 physics-based algorithms, together with basic information about them.
One of the major issues of metaheuristic algorithms is the fact that there is no guarantee that the global optimal solution is obtained.Despite the randomness deployed to build the initial solutions and during each iteration, the search can be easily biased, and the final result may be just a point that is close to a local optimum [9].In addition, to account for possible constraints of the problem, in a single-objective setting, penalty functions have to be used, adding a value to the objective function to hinder unfeasible solutions [10].If the added value is too high, optimal solutions on the boundaries can be disregarded because of the steep path this represents for the individuals and the strong deformation of the objective function landscape close to the boundary.On the other hand, if the penalty is too soft, unfeasible solutions can be considered as optimal [11].The randomness of the process and the penalties added can lead to inconsistent results, so that multiple runs of the algorithm are necessary to verify the quality of the solution.
Another crucial issue of these methods is the computational effort required.Multiple solutions (individuals) must be evaluated in each iteration.This number varies according to the number of variables comprising the problem, and can be very large.In general, the number of individuals has to be at least the same as the number of variables [12].Thus, in complex problems, the number of objective function evaluations can be huge.In addition, many engineering problems use external models to calculate the parameters of the objective function and the constraints, as is the case with hydraulic [13], thermal [14], and chemical models [15], thus increasing even more the computational effort required.
Swarm intelligence in general, and the metaheuristic we develop in this paper in particular, may have a number of benefits for many agent-based models (ABM) and applications.Among the benefits derived from applying optimization in ABM, we briefly mention the following: (a) simulation reuse facility, since there is a generalized demand for improved methods for studying many complex environments as integrated wholes [16][17][18][19]; (b) dynamic optimization due to problem uncertainty, especially in cases of slow changes, i.e., those most commonly occurring in real-world applications [20][21][22]; (c) the ability to identify emergent behaviors and conditions for them [23][24][25]; this provides insights into the landscape of system performance to help identify elements that exhibit emergent behavior where the system performance rapidly improves because it moves from disorganized to organized behavior; (d) simulation cloning to analyze alternative scenarios concurrently within a simulation execution session [26,27]; this would help optimize the execution time for multiple evaluations without repeated computation [28].
A final crucial idea is that there is no such a thing as the best metaheuristic algorithm; rather, each has advantages and disadvantages.In fact, [29,30] propose a multi-agent approach to take the best of multiple algorithms.Each of them considers separated agents searching for their own goals, and also competing and cooperating with others to reach a common goal.In general, an optimization algorithm is evaluated regarding three major aspects [31]: (i) ease of implementation due to the number of parameters to be adjusted (sometimes costly fine-tuned); (ii) computational complexity, which reflects in the convergence speed; and (iii) reliability of the results, with consistent optimal values obtained through a series of tests.
As an attempt to help to solve these problems, a novel, hybrid, swarm-physics-based algorithm that uses the metaphor of cyclists' behavior in a peloton, which we have named the Grand Tour Algorithm (GTA), is proposed in this paper.The main features of the algorithm are: (i) the drag, defined according to the distance to the leading cyclist (embodying the best solution), and (ii) the speed, defined using the difference between two consecutive objective function evaluations.These two elements are used to calculate the coefficients that determine the route of each cyclist.As a result, the peloton will try to follow the cyclist who is closest to the finish line, i.e., the optimal solution, and also, the cyclist who is going fastest at a given point.In Section 2, we describe the methodological aspects first.Then, we set the test conditions.These include the fourteen benchmarking functions used, and the complete setting of parameters for the four classical metaheuristic algorithms used for comparison: Particle Swarm Optimization [7], Simulated Annealing [8], Genetic Algorithm [32] and Harmony Search [33].In Section 3, comparisons are performed under a scenario of 1000 variables, which are used to evaluate the performance of the algorithms under a generally accepted metric.In addition, the algorithms are also applied to the same fourteen functions when their dimension is up to 20,000 variables.Finally, various sensitivity analyses verify the relevance the algorithm parameters have in the optimization convergence and stability of GTA.To finish, conclusions are presented that provide evidence that GTA (i) is superior to other classical algorithms in terms of its speed and consistency when functions with a huge (human genome-like) search space are considered, and (ii) exhibits excellent results regarding ease of implementation, speed of convergence and reliability.

Materials and Methods
In this section, we present a few general ideas about iterative methods.Then, we concisely describe the classical PSO, which has the basic foundations of any swarm-based algorithm.Finally, we fully present GTA.

Iterative Optimization Processes: A General View
Typically, an iterative optimization algorithm can be mathematically described, as in Equation (1): where x t+1 is (the position of) a solution at time step t + 1, α is the step size, and d t is the direction of the movement.Since there is no general a priori clue about which direction to take to progress towards the optimal solution, various mathematical properties are used to define the right track and the step size, better leading to optimal points.First-order derivative methods use the opposite direction of the gradient vector with a certain step size to find minimal points.Despite the fact that first-order methods have fast convergence, sometimes, depending of the step size, an oscillatory process can start, and the optimal point is never found.The well-known gradient method, a first-order method, is shown in Equation (2): Here, ∇ f (x t ) is the gradient vector calculated at x t .
To improve the convergence of first-order methods, second-order methods use the opposite direction of the gradient vector, improved by the information of the Hessian matrix, which provides second-order information.The Modified Newton method, for example, adopts a unitary step size; the method is described by Equation (3).
Even though derivative methods are mathematically efficient, the calculations of the Jacobian vector and the Hessian matrix are computationally demanding, or simply not possible in many cases.
To cope with this drawback, metaheuristic algorithms, many of which are based on natural processes, are widely applied in real-world problems.Among them, swarm-based algorithms use the same search process described by Equation ( 1).However, in contrast with derivative-based algorithms, in metaheuristic algorithms, each solution is brought to a new position (solution) based on a set of rules that replaces the derivative information with another kind of information, while providing a conceptually similar approach.

Particle Swarm Optimization (PSO)
The Particle Swarm Optimization algorithm [10] is one of the most popular swarm-optimization techniques.Each particle represents a candidate solution and has its own position X and velocity V. Initially, these values are randomly defined according to the boundaries of the problem.As time (iterations) goes by, the particles move according to their velocity and reach a new position.The velocity of a particle (see Equation ( 4)) is a linear combination of the following: the particle's inertia, through coefficient ω; its memory, through coefficient c 1 ; and some social interaction with the group, through coefficient c 2 .The best position found by the group, G, and the best position ever reached by the particle, B, are used to update its velocity.In addition, random values, rand 1 and rand 2 , uniformly distributed between 0 and 1, are used to boost the exploitation abilities of the search space.Equations ( 4) and ( 5), used to update the particle positions, describe the complete iteration process.

Grand Tour Algorithm (GTA)
The principle of the Grand Tour Algorithm (GTA) is similar to that of most swarm-based algorithms: each cyclist represents a possible solution, with position and velocity being updated with each iteration.

GTA Fundamentals
The difference hinges on how the velocity is updated.Instead of the social and cognitive aspects, the power spent by cyclists is the main feature of this update.Equation ( 6) shows how the power, P, of a cyclist can be calculated according to its speed, S.
Processes 2020, 8, 980 Three main forces are considered for this calculation: gravitational, F g ; friction, F f ; and drag, F d .In GTA, friction is disregarded, as it represents only a small fraction of the power; also, the losses, l, mainly represented by the friction of the chain and other mechanical components of the bike, are negligible.
The speed used in this equation has no relationship with the velocity used to update a cyclist's position.In cycling, two parameters describe how fast a cyclist is.In flat conditions, the usual concept of speed can be used, describing the distance traveled over a period of time.However, in a hilly terrain, the use of VAM (from the Italian "velocità ascensionale media", i.e., average ascent speed) is useful, and can better describe the cyclist's performance.In our metaphor, the VAM represents how fast the cyclist approaches a better result, while the velocity used to update its position indicates in which direction and how far the cyclist goes.Thus, speed S is calculated through Equation ( 7), representing the cyclist's ascending or descending (vertical) speed, using two consecutive values of the cyclist's objective function (OF).
So, if the objective function is reduced, the cyclist is closer to the finish line, i.e., to the optimal solution.A larger difference in the objective function between two successive iterations means that the cyclist is fast approaching the finish line.

Calculation of Drag Force
The drag force can be calculated by Equation (8).
Here, C d is the drag coefficient, A is the frontal area of the cyclist, ρ is the air density, S the cyclist vertical speed, and W the wind speed.Frontal area and air density may be considered the same for every cyclist, both taken herein as 1 m 2 .The wind speed is considered to be null.The drag coefficient, C d , is the major component of this equation.As shown in an aerodynamic model for a cycling peloton using the CFD simulation [34], in a cycling peloton, the front cyclists have to pedal with much more power than those in the back due to drag.Thus, the cyclist who achieves the best solution is the leader and, according to how far the others are behind, they benefit more or less from the leader's drag.In this sense, cyclists far from the leader find it easy (and are bound) to follow him or her because their paths require less power, while the cyclists closest to the leader are freer because the power difference in following the leader or not is less relevant.Using the physical results obtained by [34], the drag coefficient may be estimated by creating a ranking of the cyclists according to their OF values.The leader has no benefit from the peloton, and has a drag coefficient equal to one.The rear cyclist benefits the most and has a drag coefficient of only 5% of that of the leader.The drag coefficient of the remaining cyclists is obtained by means of a linear interpolation between the first and the last cyclists (Equation ( 9)).
In this expression, OF is the cyclist objective function value, OF best is the leader objective function, and OF worst is the value of the objective function of the last cyclist in the peloton.

Calculation of Gravitational Force
Gravitational force has a great impact on the speed of convergence of the algorithm.It is calculated using Equation (10).
Here, g is the acceleration of gravity, I is the slope, defined as the difference between the values for the objective function in two successive iterations, and m is the cyclist's mass.
As previously described, the objective function represents the (vertical) distance to the finish line.And, as the race is downhill, towards the minimum, the objective function is also the elevation of the cyclist's place on the terrain with respect to the minimum.Thus, when cyclists are going downhill (negative slope), they need to apply less power, because gravity is aiding them.In contrast, a cyclist's weight causes an opposite force to movement.This procedure prevents cyclists from going uphill (positive slope), towards the worst solution, as they would expend too much power in this case.
The value of the gravitational acceleration is constant, as the cyclists' masses are randomly defined at the beginning, in a range from 50 to 80 kg.This procedure allows lightweight cyclists, known as climbers, to be least affected by going uphill, i.e., with a direction towards a worse solution.This guarantees better search capabilities.

Velocity and Position Updating
The power component corresponding to each force is calculated separately and will be used to update the cyclist velocity according to Equation (11), which is used, in turn, to define its new position, as in Equation ( 5).
The elements of this formula are calculated as follows.First, the values of the power spent by the cyclists, according to Equation ( 6), are ranked.Then, the two weight coefficients for drag and gravitational power, k d and k g , shown in Equation (11), are calculated by normalizing F d and F g between 0.5 and 1.0, where 1.0 represents the cyclist with the lowest power.
Regarding the drag component, k d , this procedure reduces the step taken by the leading cyclists which are closer to an optimal solution, thus resulting in a more specific local search around the leader's position, X d , while subsequent cyclists are rapidly pushed towards this point to benefit from the leader's drag.
With the gravitational component, k g , the purpose is to encourage the cyclists that are rapidly converging upon a better solution, X g , to keep going in this direction, while cyclists going uphill or slowly improving are kept to explore with a more local search focus around their respective positions.
Finally, instead of an inertia coefficient, as used in PSO, the gravity coefficient k g is also used to give a weight to a cyclist's current direction.Random values are also used to avoid a biased search.
The flowchart in Figure 1 summarizes the GTA optimization process.
Four classical metaheuristic algorithms are used for comparison: (i) Particle Swarm Optimization (PSO) [7]; (ii) Simulated Annealing (SA) [8]; (iii) Genetic Algorithm (GA) [32]; and (iv) Harmony Search (HS) [33].The characteristic parameters of each algorithm, and their general settings used in this study, are presented in Table 1.We chose these algorithms because each belongs to one of the categories in which metaheuristic algorithms are classified [1], because they are, perhaps, the most used algorithms within their categories, and because the authors have used them for several engineering applications in Urban Hydraulics with good results, and have their codes ready.To evaluate the performance of the five algorithms, each was applied 100 times to fourteen well-known benchmark functions from the literature [36,37] and collections of online of test functions, such as the library GAMS World [38], CUTE [39], and GO Test Problems [40].In addition, [41] provides an exhaustive list of up to 175 functions.We chose the 14 benchmark fitness functions as a random sample among those used by the optimization community, while having formulas involving an arbitrary number of variables, since we were targeting very high-dimensional problems.

GTA
Table 2 shows the name and the equation of each function, the domain considered for the variables, and the minimum value they attained.Let us note here that dimension n was firstly set to 1000 variables, and then to 20,000, a value inspired by human DNA, which may be represented by 20,000 genes [42,43].

Performance
Table 3 summarizes the results of 100 runs for all five algorithms, considering a dimension of 1000 variables.The following parameters are presented: (i) best value found; (ii) mean value; (iii) standard deviation; (iv) success rate, defined as an error less than 10 −8 [34]; (v) mean of objective function (OF) evaluations; and (vi) score.Figure 2a,b graphically presents values (ii) and (v), respectively.This score is a parameter considered for the CEC 2017 competition in [44] to compare different optimization algorithms, and is expressed by Equations ( 12) and (13).
Here, OF is the optimized value of the run i, OF best is the global minimum, and SE min is the minimum value among the five algorithms for SE.The GTA was the only algorithm to reach the minimum value in eleven of the fourteen benchmarking functions.In addition, in all of the 100 runs, the optimization was a success, yielding an error lower than 10 −8 .For the Rosenbrock and Dixon Price functions, none of the algorithms achieved the global minimum.However, the GTA was the closest, with a low standard deviation, which indicates the determination of a local minimum.These two functions are well-known examples of very challenging benchmark optimization functions, which have presented many difficulties for all metaheuristics; see, for example [45][46][47], which use different algorithms and report the same difficulty in reaching the global minimum for these functions despite using less than 100 variables.Finally, for the Salomon function, the GTA presents the worst success rate, while GA presents the best one.However, the mean value reached is very close to the global minimum, and the use of more cyclists, or an increase in the maximum number of iterations, could solve this problem.For the other algorithms, only for the function Sum of Powers, this change of settings could lead to a better success rate, since for the other functions, the value found is too far from the global minimum.The reliability of the GTA is also confirmed by its score, which is not the best only for the Salomon function.To evaluate the convergence speed, Figure 3 shows the OF evolution for each of the benchmarking functions using GTA.It can be seen that a minimum is rapidly achieved, with a finetuning in the following iterations.Comparing the values of the number of OF evaluations in Table 3 for the Salomon and Sum of Powers functions, where the results of all five algorithms are close, the GTA was the one with the lowest number of evaluations, thus requiring less computational effort.
The GTA was the only algorithm to reach the minimum value in eleven of the fourteen benchmarking functions.In addition, in all of the 100 runs, the optimization was a success, yielding an error lower than 10 −8 .For the Rosenbrock and Dixon Price functions, none of the algorithms achieved the global minimum.However, the GTA was the closest, with a low standard deviation, which indicates the determination of a local minimum.These two functions are well-known examples of very challenging benchmark optimization functions, which have presented many difficulties for all metaheuristics; see, for example [45][46][47], which use different algorithms and report the same difficulty in reaching the global minimum for these functions despite using less than 100 variables.Finally, for the Salomon function, the GTA presents the worst success rate, while GA presents the best one.However, the mean value reached is very close to the global minimum, and the use of more cyclists, or an increase in the maximum number of iterations, could solve this problem.For the other algorithms, only for the function Sum of Powers, this change of settings could lead to a better success rate, since for the other functions, the value found is too far from the global minimum.The reliability of the GTA is also confirmed by its score, which is not the best only for the Salomon function.To evaluate the convergence speed, Figure 3 shows the  evolution for each of the benchmarking functions using GTA.It can be seen that a minimum is rapidly achieved, with a finetuning in the following iterations.Comparing the values of the number of  evaluations in Table 3 for the Salomon and Sum of Powers functions, where the results of all five algorithms are close, the GTA was the one with the lowest number of evaluations, thus requiring less computational effort.To finalize the comparison with the classical metaheuristic algorithms, a statistical analysis is performed to evaluate the quality of the solutions, using the Wilcoxon signed-rank test [48] with a 0.05 significance value.Table 4 shows the results of the comparison of GTA with the algorithms.In more than 90% of the solutions, GTA outperformed or equaled the algorithms.The  values found for all the comparisons were null.Thus, the performance difference between GTA and those algorithms is significant.Comparison GTA vs PSO GTA vs SA GTA vs AG GTA vs HS To finalize the comparison with the classical metaheuristic algorithms, a statistical analysis is performed to evaluate the quality of the solutions, using the Wilcoxon signed-rank test [48] with a 0.05 significance value.Table 4 shows the results of the comparison of GTA with the algorithms.In more than 90% of the solutions, GTA outperformed or equaled the algorithms.The p values found for all the comparisons were null.Thus, the performance difference between GTA and those algorithms is significant.For a final test, GTA was applied once more to the fourteen benchmarking functions, but this time, the dimension was increased to 20,000 variables.As can be seen in Table 5, the results remain similar to those obtained with 1000 variables.The success rates are lower for the Exponential and Sum of Powers functions, but both are still close to the global minimum.The Rosenbrock and Dixon-Price functions again achieve a local minimum, and the Salomon function again yields a low success rate.An additional test for these five functions using 500 cyclists was made to try to reach the global minimum.This procedure was effective for the Exponential, Sum of Powers, and Salomon functions, reaching a 100% success rate, albeit with a significant increase in the number of OF evaluations (+100-500%).Nevertheless, the GTA was still robust and fast for this highly dimensional analysis.Figure 4 shows the results for GTA and the other metaheuristics for the 20,000-variable problems.It can be observed that SA significantly increased the number of OF evaluations, while all of them reached a higher average OF value in all cases, except for the Salomon function.Let us note that the GA could not be tested in this scenario due to a lack of memory on the computer used.

Sensitivity Analysis
Here, we report three tests, always using 20,000 variables as the functions dimension.The first test is performed to evaluate the relevance of the number of cyclists and the maximum number of iterations in the optimization process.Figure 5a shows that the increase in the numbers of cyclists and maximum number of iterations has no significant influence on the reduction of the objective function.Only a combination of low numbers of cyclists and iterations yielded poor results.As expected, Figure 5b shows an increase of the number of objective function evaluations when both parameters are increased.Therefore, the algorithm yields the best results after extensive searching,

Sensitivity Analysis
Here, we report three tests, always using 20,000 variables as the functions dimension.The first test is performed to evaluate the relevance of the number of cyclists and the maximum number of iterations in the optimization process.Figure 5a shows that the increase in the numbers of cyclists and maximum number of iterations has no significant influence on the reduction of the objective function.Only a combination of low numbers of cyclists and iterations yielded poor results.As expected, Figure 5b shows an increase of the number of objective function evaluations when both parameters are increased.
Therefore, the algorithm yields the best results after extensive searching, but, as observed in Figure 3, with no significant changes.These results agree with those obtained previously, showing that GTA has a fast convergence to a feasible and high-quality solution.
Processes 2020, 8, x FOR PEER REVIEW 15 of 18 but, as observed in Figure 3, with no significant changes.These results agree with those obtained previously, showing that GTA has a fast convergence to a feasible and high-quality solution.The second sensitivity analysis focused on variations of cyclists' masses.For this purpose, the minimum and maximum values considered were changed to a range from 25 to 125 kg.The results showed no significant difference.Only when all the cyclists had the same weight, regardless of the value, the optimization became slower and yielded worse results.This lack of randomness hampers the exploration of points near the boundaries, since there are no lightweight cyclists, who are less affected by the gravitational force, to venture closer to the boundaries because of the penalties.
Finally, the score used to calculate the  and  coefficients of the ranked cyclists according to the power spent in each iteration was evaluated.The best cyclist, or the one who expends the least energy, receives the highest score, while the worst receives the minimum.Figure 6 shows that only The second sensitivity analysis focused on variations of cyclists' masses.For this purpose, the minimum and maximum values considered were changed to a range from 25 to 125 kg.The results showed no significant difference.Only when all the cyclists had the same weight, regardless of the value, the optimization became slower and yielded worse results.This lack of randomness hampers the exploration of points near the boundaries, since there are no lightweight cyclists, who are less affected by the gravitational force, to venture closer to the boundaries because of the penalties.
Finally, the score used to calculate the k g and k d coefficients of the ranked cyclists according to the power spent in each iteration was evaluated.The best cyclist, or the one who expends the least energy, receives the highest score, while the worst receives the minimum.Figure 6 shows that only with a combination of high values, the results of the objective function got worse.In terms of the number of evaluations of the objective function, no significant difference was observed.In all the analyses made, the GTA presented consistent results despite changes in its parameters.This robustness is very interesting for the user, since there is no need for heavy fine-tuning to adjust the default parameters to obtain the best optimization.with a combination of high values, the results of the objective function got worse.In terms of the number of evaluations of the objective function, no significant difference was observed.In all the analyses made, the GTA presented consistent results despite changes in its parameters.This robustness is very interesting for the user, since there is no need for heavy fine-tuning to adjust the default parameters to obtain the best optimization.

Conclusions
As a swarm-based algorithm, GTA exhibits good results in terms of consistency and speed to find the optimal solution for the benchmarking problems considered.GTA performance is clearly superior to that of the classical algorithms used in this paper for comparison, in terms of speed and consistency when functions with a huge (human genome-like) search space (up to 20,000 decision variables) were tested.The great advantage of GTA is its ability to direct the cyclists to search for the fastest path to improve the objective function.At the same time, points with worst objective function values are explored locally, since they may be just a barrier for big improvements.The search for the fastest path has a similar goal as the first derivative, hence its fast convergence, despite the number of cyclists and iterations used.Finally, a sensitivity analysis shows the robustness of GTA, since a change in the main parameters has little impact on the final result.This feature results in a userfriendly algorithm, as there is no need to adjust the default parameters to achieve a good optimization.Therefore, the three most desirable aspects for an optimization algorithm, i.e., ease of implementation, speed of convergence, and reliability, had good results, confirming the expected improvements of GTA.Having presented the GTA, our target is now to develop it further in several directions so that it may also become efficient with real-world problems; this development will include constrained optimization, multi-objective optimization, and the development of mechanisms for GTA to dynamically accommodate to changing fitness functions.

Figure 1 .
Figure 1.Flowchart of the Grand Tour Algorithm.

Figure 1 .
Figure 1.Flowchart of the Grand Tour Algorithm.

Figure 2 .
Figure 2. Results of 100 runs with 1,000 variables: (a) OF value average; (b) OF evaluations average.Figure 2. Results of 100 runs with 1000 variables: (a) OF value average; (b) OF evaluations average.

Figure 2 .
Figure 2. Results of 100 runs with 1,000 variables: (a) OF value average; (b) OF evaluations average.Figure 2. Results of 100 runs with 1000 variables: (a) OF value average; (b) OF evaluations average.

Figure 3 .
Figure 3. GTA Objective function evolution for the fourteen benchmark functions.

Figure 3 .
Figure 3. GTA Objective function evolution for the fourteen benchmark functions.

Figure 4 .
Figure 4. Results of 100 runs with 20,000 variables: (a) OF value average; (b) OF evaluations average.

Figure 4 .
Figure 4. Results of 100 runs with 20,000 variables: (a) OF value average; (b) OF evaluations average.

Figure 5 .
Figure 5. Sensitivity analysis for the number of cyclists and the maximum number of iterations: (a) Objective function variation; (b) Number of objective function evaluations.

Figure 5 .
Figure 5. Sensitivity analysis for the number of cyclists and the maximum number of iterations: (a) Objective function variation; (b) Number of objective function evaluations.

Figure 6 .
Figure 6.Sensitivity analysis for the score of parameters   and

Table 3 .
Results of 100 runs with 1000 variables.

Table 4 .
Comparison of results between GTA and classical algorithms.

Table 4 .
Comparison of results between GTA and classical algorithms.

Table 5 .
Results of 100 runs with 20,000 variables using GTA.