Next Article in Journal
An Approximate Algorithm for Sparse Distributionally Robust Optimization
Previous Article in Journal
Automated Grading Method of Python Code Submissions Using Large Language Models and Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Open Competency Optimization with Combinatorial Operators for the Dynamic Green Traveling Salesman Problem

IABL, FSTT, Abdelmalek Essaadi University, Tetouan 93000, Morocco
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Information 2025, 16(8), 675; https://doi.org/10.3390/info16080675
Submission received: 13 May 2025 / Revised: 1 August 2025 / Accepted: 4 August 2025 / Published: 7 August 2025
(This article belongs to the Section Artificial Intelligence)

Abstract

This paper proposes the Open Competency Optimization (OCO) approach, based on adaptive combinatorial operators, to solve the Dynamic Green Traveling Salesman Problem (DG-TSP), which extends the classical TSP by incorporating dynamic travel conditions, realistic road gradients, and energy consumption considerations. The objective is to minimize fuel consumption and emissions by reducing the total tour length under varying conditions. Unlike conventional metaheuristics based on real-coded representations, our method directly operates on combinatorial structures, ensuring efficient adaptation without costly transformations. Embedded within a dynamic metaheuristic framework, our operators continuously refine the routing decisions in response to environmental and demand changes. Experimental assessments conducted in practical contexts reveal that our algorithm attains a tour length of 21,059, which is indicative of a 36.16% reduction in fuel consumption relative to Ant Colony Optimization (ACO) (32,994), a 4.06% decrease when compared to Grey Wolf Optimizer (GWO) (21,949), a 2.95% reduction in relation to Particle Swarm Optimization (PSO) (21,701), and a 0.90% decline when juxtaposed with Genetic Algorithm (GA) (21,251). In terms of overall offline performance, our approach achieves the best score (21,290.9), significantly outperforming ACO (36,957.6), GWO (122,881.04), GA (59,296.5), and PSO (36,744.29), confirming both solution quality and stability over time. These findings underscore the resilience and scalability of the proposed approach for sustainable logistics, presenting a pragmatic resolution to enhance transportation operations within dynamic and ecologically sensitive environments.

1. Introduction

The TSP is one of the most well-known and widely studied problems in combinatorial optimization. It consists of finding the shortest possible tour that allows a salesman to visit each city in a given set exactly once and return to the starting point [1]. Due to its simplicity in formulation and complexity in solution, the TSP has become a benchmark problem in various domains such as logistics, transportation planning, and manufacturing. However, real-world routing problems are rarely static. In practical applications, parameters such as travel times, and route constraints are often subject to change. This has led to the development of the dynamic TSP (DTSP), which considers real-time variations in the problem data, making the optimization process significantly more challenging. Among these dynamic extensions, particular attention has recently been given to problems that incorporate both dynamic inputs and environmental considerations. Motivated by growing concerns over climate change and sustainability, the research community has introduced the dynamic green TSP (DG-TSP) [2]. This formulation extends the DTSP by incorporating not only time-dependent travel costs influenced by road gradients and energy consumption models, but also by aiming to minimize fuel usage and greenhouse gas emissions. The DG-TSP is of critical importance in the context of sustainable logistics, where businesses must balance operational efficiency with environmental responsibility. Despite its relevance, solving the DG-TSP remains a complex task. Traditional exact methods and basic heuristics are often insufficient to cope with the stochastic nature of travel conditions—such as fluctuating traffic patterns—and the nonlinearities arising from road gradients and energy constraints. While metaheuristics have demonstrated their capacity to address complex problems, their effectiveness in dynamic and green contexts largely depends on the adaptability of their operators and mechanisms. To address these challenges, we propose a novel metaheuristic framework based on dedicated combinatorial operators specifically tailored to the DG-TSP. These operators are designed to dynamically adapt to changing problem conditions, such as fluctuating demands and environmental constraints, allowing for real-time decision making. The approach is embedded within the OCO framework [3], which provides a flexible structure for integrating adaptive strategies and evolutionary search. Extensive simulations and real-world case studies validate the performance of our method. Compared to well-established metaheuristics such as GA, PSO, GWO, and ACO, our approach yields superior results in both tour length (as a proxy for fuel efficiency) and overall offline performance, confirming its robustness in sustainable routing scenarios.
The key contributions of this article are as follows:
  • Development of dedicated combinatorial operators: We design adaptive operators for permutation-based optimization, capable of reacting to stochastic changes in travel conditions and road gradients.
  • Integration into an adaptive metaheuristic framework: The proposed operators are embedded into the OCO metaheuristic, enabling efficient and scalable dynamic optimization.
  • Rigorous empirical validation: Through dynamic benchmarks, real-world case studies, and comprehensive statistical testing, we demonstrate the superiority of our method over GA, PSO, GWO, and ACO.
  • Formulation of the DG-TSP: We formalize a novel problem definition that integrates dynamic service conditions and environmental sustainability into the classical TSP model.
The remainder of this paper is structured as follows: Section 2 presents the formal definition of the DG-TSP. Section 3 explores the dynamic aspects of the problem and reviews the related literature. Section 4 introduces the metaheuristic framework. Section 5 details the design of our combinatorial operators. Section 6 discusses the experimental setup and results. Section 7 concludes with implications, future work, and a summary of key findings.

2. Green Logistics

Green logistics is a wide-reaching field that balances economic, environmental, and social concerns [4]. Its goal is to reduce harm to ecosystems and living beings while supporting healthier communities and sustainable economic growth. It emphasizes initiatives that produce minimal detrimental impacts on the environment and incorporates methodologies and practices that facilitate societal progress and economic enhancement. The integration of environmental considerations into logistical operations is crucial for promoting sustainable development within society. Consequently, the aims extend beyond merely the economic ramifications of logistical operations to include a wider array of environmental and societal consequences, which encompass the evaluation of environmental contamination, the management of waste, and the analysis of consumer satisfaction. Green logistics functions as a strategic paradigm for organizations to attain sustainability while guaranteeing the enduring profitability of logistical endeavors [5]. The execution of green logistics (GL) entails a broad spectrum of activities and processes [6]. Zhang et al. propose a GL-based framework structured into three key categories: the assessment of green performance, execution of green operations, and the formulation of green strategies [7].
Generally speaking, logistics can be defined as the process of delivering a specified quantity of a product in an optimal condition to a designated customer at a precise location within a given timeframe [8]. Globalization and the decentralization of production have significantly contributed to the enhancement and diversification of logistics services. Today, logistics plays a crucial role across various essential sectors of society. Alongside its rapid development, there has been a growing awareness of environmental concerns in relation to logistics. Recent debates on environmental policies, coupled with rising oil prices, have underscored the necessity of minimizing vehicle movements whenever possible. As a result, research on logistics optimization has emerged as a pivotal field. The primary objective of logistics optimization is to reduce costs, minimize the number of vehicles required, and optimize travel routes. This study aims to address the TSP, a fundamental challenge in green logistics and transportation planning. The TSP is a classic NP-Hard problem [9], as its computational complexity grows exponentially with the size of the problem. Given a set of n cities ( V 1 , V 2 , V 3 , , V n ) and a distance matrix ( D = [ d i , j ] ) , where ( d i , j ) represents the distance between cities V i and V j , the objective is to determine a tour that allows the traveler to visit each of the n cities exactly once and return to the starting city. Naturally, the total distance traveled must be minimized. More formally, the problem involves finding a permutation π of the n cities that minimizes the following sum of distances:
i = 1 n 1 d π ( i ) , π ( i + 1 ) + d π ( n ) , π ( 1 )
To remain consistent with the rest of the paper, we define the problem over n cities, indexed from 1 to n, as reflected in the permutation-based cost function (1). This formulation captures the essence of the Green TSP, where the goal is to minimize a function of fuel consumption by identifying the optimal possible tour that satisfies the constraints of visiting each city exactly once and returning to the origin. Therefore, Equation (1) becomes
i = 1 n 1 C o n s u m p t i o n π ( i ) , π ( i + 1 ) + C o n s u m p t i o n π ( n ) , π ( 1 )
with Consumption representing a function that calculates the fuel consumption between two cities. The article by Moghdani et al. [10] provides a systematic and comprehensive literature review on the green vehicle routing problem. The paper categorizes solution methodologies into six main groups: metaheuristics, heuristics, software-based applications, exact methods, commercial exact solvers, and hybrid approaches. In the same context, Simulated Annealing [11] has been employed to minimize total CO2 emissions and the total traveled distance. Ubeda et al. [12] applied Tabu Search to address routing problems and pollution reduction. Variable Neighborhood Search has been adopted within a hyper-heuristic framework to investigate the routing problem of mixed-energy vehicle fleets [13,14]. S. Zhang et al. [15] noted that Ant Colony Optimization is among the most popular metaheuristics for green logistics. Another metaheuristic, Multi-start Local Search, based on a multigraph reformulation, consists of two phases: the first employs fast constructive operators, while the second improves the solutions obtained in the first phase to address the green fleet problem [16]. The Non-Dominated Sorting Genetic Algorithm II (NSGA-II), a multi-objective genetic algorithm, was used for a Green Vehicle Routing Problem (Green-VRP) involving a three-echelon supply chain with uncertain customer demand [17]. The Firefly Algorithm (FA) was utilized to generate a set of feasible solutions for an asymmetric Green VRP with time windows, variable delivery deadlines, and heterogeneous vehicle dimensions [18].
The problem formulation can be stated as follows:
  • Mathematical Formulation of the Green Traveling Salesman Problem (Green TSP):
    For completeness and consistency with the classical mathematical literature on the TSP, we also provide the standard integer programming formulation.
    Let a complete graph G = ( V , E ) , where
    -
    V = { 1 , 2 , , n 1 } denotes the set of cities (with 1 as the start/return city);
    -
    E denotes the set of arcs between the cities.
    The parameters and decision variables are defined as follows:
    -
    c i j : Environmental cost (e.g., CO2 consumption) associated with traveling from city i to city j;
    -
    x i j { 0 , 1 } : Binary variable equal to 1 if the tour travels from i to j, and 0 otherwise;
    -
    u i : Auxiliary variable used to eliminate subtours (Miller–Tucker–Zemlin formulation).
  • Objective function:
    min i = 0 n 1 j = 0 j i n 1 c i j x i j
  • Constraints:
    j = 0 j i n 1 x i j = 1 i = 0 , 1 , , n 1
    i = 0 i j n 1 x i j = 1 j = 0 , 1 , , n 1
    u i u j + n x i j n 2 , i j , i , j { 1 , , n 1 }
    1 u i n 1 , i { 1 , , n 1 }
  • Variable domains:
    x i j { 0 , 1 } i , j = 0 , , n 1 , i j
  • Principle of the MTZ Method:
    In the TSP, the objective is to find a tour that visits each city exactly once and returns to the starting point. However, some solutions that satisfy degree constraints may contain subtours, i.e., smaller cycles not covering all cities. These are invalid for the TSP. The Miller–Tucker–Zemlin (MTZ) method is a classical technique for eliminating such subtours in integer linear programming formulations. An auxiliary continuous variable u i is introduced for each city i { 1 , , n 1 } (excluding city 0, the depot). This variable estimates the visit order of city i in the tour. Linear constraints are then imposed to enforce a logical progression in the tour, thereby preventing subtours. Let
    -
    x i j { 0 , 1 } be a binary variable indicating whether arc ( i , j ) is used;
    -
    u i be the order of visiting city i.
    The constraints (6), (7) are added with the following implications:
    -
    If x i j = 1 , the constraint becomes u i u j < 0 , hence u i < u j : city j is visited after i.
    -
    If x i j = 0 , the constraint is trivially satisfied: u i u j n 1 always holds.
    This ensures a consistent visit order among cities and prevents the formation of closed cycles (subtours) among subsets of cities.
  • Illustration of a forbidden subtour:
    Information 16 00675 i001
    In this subtour, cities 1, 2, and 3 form a closed cycle without visiting all cities (e.g., city 0 is excluded). The MTZ constraints prevent such configurations, as they would imply u 1 < u 2 < u 3 < u 1 , which is contradictory.
In this work, we consider the environmental cost c i j associated with traveling from node i to node j, which is used as the objective to minimize. This cost may represent fuel consumption or CO2 emissions resulting from vehicle movement. To remain consistent with standard benchmark instances—where detailed information about vehicle load, speed, or road slope is not provided—we adopt a simplified, distance-proportional formulation:
c i j = β · d i j
where d i j denotes the Euclidean distance between nodes i and j, and β > 0 is a conversion coefficient reflecting the average environmental impact per unit distance. This linear model is commonly used in the literature and facilitates fair comparison with existing works on dynamic routing problems. However, it is worth noting that more realistic models of environmental cost exist. For instance, Bektaş and Laporte [19] proposed a load-dependent cost function of the form:
c i j = α · d i j + β · d i j · w i j
where α is the base cost per unit distance (for an unloaded vehicle), β is the marginal cost per unit of load and distance, and w i j is the average load carried along arc ( i , j ) . Such models provide higher fidelity to real-world emissions but require additional input data, which are typically unavailable in dynamic TSP benchmarks. For simplicity and reproducibility, we adopt the linear model in this study, while acknowledging that the framework can easily accommodate more detailed cost functions in future work.
However, in the real case, this green voyage is not always static; instead, it is dynamic. The topology between cities can change, cities can be added or eliminated [20,21], or the distance between cities changes (taking a path because of works, for example) [21]. Consequently, the journey must be re-optimized by making the changes that arise and Equation (1) becomes
f ( T o u r ( t ) ) = i = 1 n ( t ) 1 C o n s u m p t i o n T o u r ( i ) , T o u r ( i + 1 ) ( t ) + C o n s u m p t i o n T o u r ( n ) , T o u r ( 1 ) ( t )
It is important to clarify that the function Consumption i , j ( t ) used in Equation (11) corresponds to the environmental cost incurred when traveling from city i to city j at time t. In this work, we adopt a simplified linear model commonly used in the literature, where the environmental cost is directly proportional to the Euclidean distance between cities, such that
Consumption i , j ( t ) = c i j ( t ) = β · d i j ( t )
Here, d i j ( t ) denotes the time-dependent distance between nodes i and j, and β > 0 is a fixed conversion coefficient representing the average environmental impact (e.g., fuel usage or CO2 emissions) per unit of distance. This formulation allows for dynamic changes in the distance matrix D ( t ) and ensures consistency with standard benchmark assumptions.
The concept of DTSP was initially introduced by Psaraftis [22]. Since then, many variants of DTSP have been implemented, where the set of cities [23] and/or the travel costs of the set of cities [21] or the distance matrix D(t) change during the optimization process [24]. Youness and others [25] introduced a benchmark generator for this dynamic traveler problem with different modes:
  • Change in travel topology [26];
  • Change in distance between cities [20];
  • Variation in number of cities [26].
DG-TSP falls under the general category of real-time fleet management. Much of the current literature is characterized by algorithms that react to new demands only once they arise, while neglecting the available stochastic information [20]. Overviews of these problems can be found in Powell. Others, to react to the changes in the problem due to its dynamic nature, have planned a renewal by the introduction of new solutions. This aspect is known in the literature as the introduction of immigrants [20,26,27]. We also find another line of research, known as anticipatory routing, in which probability distributions are used to predict the change in road traffic and react according to these probabilities [28].

3. Dynamic Green Logistics

The TSP becomes more realistic when it is subject to a dynamic and environmentally friendly setting. For instance, our well-known salesman aims to distribute goods to various cities, departing from their home city and returning after visiting all cities while following a less polluting route. The objective is to optimize CO2 emissions and plan the journey as efficiently as possible while preserving the environment and minimizing pollution. By considering the distances between cities, the salesman can generate the optimal itinerary and commence their tour. Navigating traffic delays while trying to minimize CO2 emissions is a tricky situation. When unexpected congestion complicates the plan, the driver must rapidly adapt their decision making to identify an alternative route that avoids traffic while minimizing emissions. This need for a quick re-evaluation adds complexity to the routing process.
  • Dynamic Green TSP (DG-TSP) with Traffic Factors
    In this paper, we generate Dynamic Green TSPs (DG-TSPs) by incorporating the traffic factor, which directly affects CO2 emissions. We assume that the cost of the link between cities i and j is given by d i j × t i j , where d i j represents the normal travel distance and t i j denotes the traffic factor. Every f iteration of running an algorithm, a random number R [ F L , F U ] is generated to simulate potential traffic congestion, where F L and F U are the lower and upper bounds of the traffic factor t i j , respectively, [29]. It is needless to say that traffic congestion significantly exacerbates air pollution.
    Each link between cities i and j has a probability m of being affected by traffic, wherein a distinct R value is generated to represent low, normal, or high levels of traffic congestion on different roads. Meanwhile, the remaining links are assigned t i j = 1 , indicating an absence of traffic.
    For instance, high-traffic roads are generated by assigning a higher probability to R values closer to F U , whereas low-traffic roads are generated with a higher probability of assigning R values closer to F L . This class of DG-TSP is referred to as random DG-TSP in this paper, since previously visited environments are not guaranteed to reappear [30].
  • Cyclic Dynamic Green TSP
    Another variation of the DG-TSP with traffic factors is the Cyclic DG-TSP, in which dynamic changes follow a cyclic pattern. In other words, previous environments are guaranteed to reoccur in the future. Such environments are more realistic than purely random ones, as they can, for instance, model the 24-h traffic dynamics of a typical day [31].
    A cyclic environment can be constructed by generating different dynamic scenarios with traffic factors as the base states, representing DG-TSP environments with low, normal, or high levels of traffic. The environment then transitions cyclically among these base states in a fixed logical sequence. Depending on the time of day, environments with varying traffic conditions can be generated. For example, during rush hours or on roads with steep slopes, there is a higher probability of generating R values closer to F U , whereas during off-peak evening hours, a higher probability is assigned to R values closer to F L .

3.1. Formulation of the DTSP

Formally, a dynamic optimization problem (DOP) can be defined as follows:
Ω = ( X ( t ) , Λ ( t ) , f ( t ) ) t T
where
  • Ω denotes the optimization problem;
  • X ( t ) is the search space at time t;
  • Λ ( t ) is the set of constraints at time t;
  • f ( t ) is the objective function at time t, which assigns an objective value to each solution x X ( t ) , with all components being time-dependent;
  • T is the set of time values.
For the static TSP, the objective function is to minimize the total distance of a tour:
f ( x ) = min i = 0 n 1 j = 0 n 1 d i j x i j
subject to
x i j = 1 , if ( i , j ) is included in the tour 0 , otherwise
where n is the number of cities and d i j is the distance between city i and city j. In the context of the DTSP with traffic factors, the cost of traveling between cities i and j becomes d i j × t i j . Although the explicit mathematical formulation for the DTSP with traffic factors is not rewritten as a new form of f ( x ) , the introduction of t i j implies that the distances d i j (or travel costs) are now time-dependent, i.e., d i j ( t ) , reflecting changes due to traffic. The goal then becomes to track the optimum as these values d i j ( t ) evolve over time. In summary, the DTSP introduces temporal dynamics and uncertainty into routing problems, making them more complex and more relevant to real-world applications such as route planning under changing traffic conditions.
  • Definitions and Notations:
    Let a dynamic graph be defined as G t = ( V t , E t ) for each time step t T = { 0 , 1 , , H } where H is the time horizon.
    -
    V t : Set of available cities at time t;
    -
    E t : Set of valid arcs at time t;
    -
    x i j t { 0 , 1 } : Equal to 1 if the route from city i to j is taken at time t;
    -
    c i j t : Ecological cost (e.g., CO2 emissions, energy) for the route ( i , j ) at time t;
    -
    d i j t : Distance or travel time between i and j at time t;
    -
    a i t : Equal to 1 if city i is available at time t, 0 otherwise;
    -
    u i : Position of city i in the tour (for subtour elimination);
    -
    n = | V | : Total number of cities to be visited (fixed).
  • Objective Function:
    Minimize the cumulative environmental cost (CO2 emissions) of the dynamic tour:
    min t = 0 H i V t j V t j i c i j t · x i j t
  • Constraints:
    • Each city is visited exactly once over the horizon:
      t = 0 H j V t j i x i j t = 1 i V
      t = 0 H j V t j i x j i t = 1 i V
    • Temporal availability of cities and routes:
      u i u j + n · t = 0 H x i j t n 1 i j , i , j V { 0 } 1 u i n i V { 0 }
    • Subtour elimination (MTZ formulation)
    • Departure from and return to the initial city (city 0):
      t = 0 H j 0 x 0 j t = 1 t = 0 H i 0 x i 0 t = 1
    • Domain of the decision variables:
      x i j t { 0 , 1 } i , j V , t T
      u i Z i V
  • Remarks:
    -
    The graph may evolve over time: certain cities may become available or unavailable, and the costs c i j t may vary.
    -
    The cost c i j t incorporates CO2 emissions.
    -
    This model can be used in real-time or predictive environments (smart logistics, electric vehicles, etc.).

3.2. Integrating Real-World Constraints into DG-TSP for Green Logistics

The DG-TSP is an emerging research area aimed at enhancing the sustainability of logistics by accounting for real-world complexities such as dynamic travel conditions, road topography, and fuel consumption optimization. Here, is a breakdown of the key areas within the DG-TSP:
  • Time-dependent service requests: These add a layer of complexity because the problem needs to adapt in real time. Research has explored ways to handle this uncertainty, using methods like scenario-based planning and reinforcement learning [32]. The focus is on creating algorithms that can efficiently manage these changing requests.
  • Slopes with different inclinations: Factoring in actual road conditions is essential for accurately estimating fuel consumption. Studies show that using detailed road gradient data can significantly improve fuel consumption and overall efficiency [33].
  • Fuel consumption optimization: This is a core aspect of making logistics sustainable. Studies demonstrate that incorporating fuel consumption models into routing algorithms can lead to major reductions in greenhouse gas emissions [34]. Smart technologies also play a role in optimizing routes to lower costs and environmental impact [34].
  • Sustainable logistics and smart technologies: Using smart technologies in logistics is becoming increasingly common for sustainability. This includes decision support systems and key performance indicators that rely on data to optimize operations. Digital twin technology can also make public transport systems more efficient through simulations and real-time insights.
In essence, the DG-TSP aims to create more flexible and environmentally friendly logistics by tackling real-world challenges and using advanced technologies.

4. Open Competency Optimization

A robust metaheuristic requires a careful balance between exploration and exploitation. An algorithm’s exploration capability refers to its ability to investigate diverse regions of the feasible solution space. Conversely, exploitation signifies the algorithm’s ability to guide individuals towards optimal solutions with utmost expediency. Excessive focus on exploration results in a purely stochastic search process, while overemphasis on exploitation yields a search approach prone to becoming trapped in a local optimum. The OCO algorithm [3] is meticulously crafted to achieve a harmonious balance between exploration and exploitation, as it is predicated on a contemporary framework of human learning. Competency-based learning empowers learners to formulate their own educational trajectories in accordance with their competencies while deriving inspiration from their peers. The objective is to steer learners towards the global optimum with maximal efficiency. This paradigm is applicable to the resolution of an open-ended problem within a cohort of learners, which may be restricted to a small classroom setting or extended to an online course wherein multiple students engage in synchronous interactive learning. A salient advantage of this inspiration-based methodology is that it cultivates a population of ideas aimed at addressing a specific problem, in contrast to a population of individuals, as observed in other algorithms modeled after insect or animal behaviors. In our algorithm, the population size is not static; rather, it is dynamic, as learners perpetually respond to and engage with the ideas proffered by their peers. Each learner possesses the capacity to generate one or multiple ideas predicated on their cognitive abilities and the feedback received from others regarding their proposals. The principal objectives of this approach are delineated as follows:
  • Each student formulates their unique educational trajectory predicated upon their specific skill sets;
  • Each student engaged in the learning process collaborates with the most immediate cohort, whether this is defined by geographical closeness or intellectual competencies. It is important to highlight that the size of this group is limited to a maximum of five participants;
  • Students have the opportunity to interact with each other via deliberations or by embracing advanced recommendations (i.e., optimal resolutions) from their colleagues.
Nevertheless, an algorithm may become ensnared in the constructs of optimal local spaces when addressing a multifaceted problem that encompasses numerous local optimal solutions. To mitigate premature convergence and to achieve a harmonious balance between exploration and exploitation capabilities, the proposed algorithm incorporates a variety of corrective measures. On one side, the average component or centroid facilitates the expansion of the search domain of the OCO algorithm. Conversely, learners are enabled to alter specific research concepts by modifying the interpretation of these concepts (adjusting one component of a vector while maintaining the others constant during self-learning updates). Algorithm 1 and Figure 1 provide a comprehensive overview of the proposed OCO algorithm.
Figure 1. Flowchart of OCO.
Figure 1. Flowchart of OCO.
Information 16 00675 g001
Algorithm 1 OCO algorithm: principal steps
1:
Initialize N students (populations of solutions); t = 0 ;
2:
while  t < MaxIteration  do
3:
    Assess each student utilizing the fitness function.;
4:
    if capacity > ThresholdCapacity then
5:
        Update students along self-learning;
6:
    end if
7:
    Update students along neighbor learner groups;
8:
    Update students along leadership interactions;
9:
     t = t + 1 ;
10:
end while
11:
Return the best solution(s);
The Algorithm 1 represents the main optimization framework, while Algorithms 2–4 correspond to procedures that are called conditionally during execution, depending on the context.
As detailed in Section 4.1, Algorithm 2 implements a self-learning operator activated only in dynamic optimization scenarios where the environment or demand evolves over time. A control parameter, ThresholdCapacity, governs the decision to inject new solutions into the population, formalized in Equation (23). When this threshold is exceeded, Algorithm 2 handles the insertion of these candidate solutions to maintain population diversity and adaptability.
Following the decision to inject new solutions, a group of g candidate solutions must be generated. This group is constructed either by Algorithm 3 or Algorithm 4, depending on the diversification strategy employed:
  • Algorithm 3 generates solutions through a neighborhood-based approach leveraging local information for guided exploration.
  • Algorithm 4 constructs solutions randomly to enhance exploration capabilities.
Within the main optimization loop (Algorithm 1), the leadership interaction procedure is designed to enhance the quality of solutions through a guided learning strategy based on high-performing individuals. Specifically, each candidate solution is influenced by one or more leaders, selected based on their superior fitness values. The adjustment of each solution under this mechanism follows a structured update formula that is mathematically expressed in Equation (28). Equation (28) defines how the solution vector is modified by incorporating both exploitation (following the leader) and controlled stochastic perturbations to preserve diversity. This leadership-based learning balances convergence and exploration, allowing the algorithm to efficiently refine promising areas of the search space while avoiding premature stagnation. The implementation of this mechanism significantly contributes to the intensification phase of the algorithm and supports global optimization performance.

4.1. Self-Learning

Each individual engaged in the learning process has the capacity to enhance their understanding by integrating foundational concepts and reinterpreting them through the application of inquiry regarding the novel information presented by the specific problem context. The efficacy of the competency-oriented pedagogical framework resides in its potential to stimulate and motivate learners to generate novel knowledge by interrogating pre-existing understandings, modifying and adapting such knowledge to yield fresh perspectives and juxtaposing these with antecedent insights. The articulation of these cognitive engagements can be delineated as follows for a student X i :
X i , r = X i , r + r a n d r · ( x k X i , r )
where r is an index randomly chosen from 1 to D, D is the dimension of the learning vector (proposed problem solution), r a n d r is a random number between 0 and 1, and x k is a value randomly selected from the domain of values for the element at index r. Nevertheless, certain individuals engaged in the learning process exhibit challenges in autonomously strategizing their educational trajectory. This constraint is contingent upon the individual competencies of each student, prompting us to establish a threshold of capacity. In this study, it should be noted that the Self-Learning operation, illustrated in Figure 1 and described in Algorithm 1, is only triggered under the first condition specified in Algorithm 2. This operator is designed to introduce diversity within the population, which is essential for maintaining adaptability in dynamic optimization scenarios. By selectively applying self-learning, the algorithm can explore new regions of the search space and better respond to environmental changes.
For a learner Xi ( x i 1 , x i 2 ,…, x i D ),
x ir new = x ir old + rand · ( x ir old x i k ) , or x ir old + rand · ( x i l + x i k ) , or x ir old + rand · ( B e s t m + x i k ) , or x i k , iff ( X i new ) < f ( X i old ) x ir old , else
where r, k, l, and m represent indices randomly selected from the set of integers ranging from 1 to D, where D denotes the dimensionality of the learning vector (the proposed resolution of the problem), and rand signifies a stochastic variable constrained within the interval [0, 1]. Equation (23) proposes three different update strategies to improve the quality of the new solution X i new . If none of these strategies leads to an improvement, i.e., if f ( X i new ) < f ( X o i old ) , then the component x i r new is assigned a randomly selected value x i k , where k { 1 , 2 , , D } and D is the problem dimension. More precisely, the algorithm evaluates the three proposed update rules of the candidate solution X i n e w . The strategy that yields the lowest objective function value is selected and assigned to x n e w i r . This selective mechanism ensures that each variable update contributes optimally to improving the overall solution quality. This random assignment is conducted without applying any transformation, unlike the three enhancement formulas proposed in Equation (23). If this substitution does not result in a better solution, the algorithm keeps the original value x i r old .
The ThresholdCapacity parameter plays a critical role in determining whether a learner is eligible for the self-learning operation. A higher threshold (e.g., 0.8) restricts self-learning to individuals with strong internal capacities, thus favoring the exploration of promising solutions. In contrast, a lower threshold (e.g., 0.1) allows a wider range of individuals to engage in self-learning, thereby enhancing population diversity and exploration. This dynamic adjustment enables better adaptability to changes in the optimization landscape, particularly in dynamic problem settings. Conversely, the population size is characterized by its dynamism, attributed to the algorithm’s consideration of learners’ ideas rather than their numerical representation. Thus, each learner is permitted to propose one or more ideas, with the stipulation that this quantity shall not surpass five-fourths of the initial population size. Equation (23) delineates four strategies for the adaptation of solutions and the enhanced exploration of the search space; thus, the algorithm is capable of incorporating one or two solutions into the learners’ population, contingent upon their adaptation (Idea_Adaptation). Algorithm 2 encapsulates the requisite conditions.
Algorithm 2 Self-learning conditions
1:
c a p a c i t y = rand ( 0 , 1 )
2:
I d e a A d a p t a t i o n randomly chosen [ 0 , 1 ]
3:
if  c a p a c i t y > ThresholdCapacity  then
4:
    Update along self-learning
5:
    if  NewPopulationSize < 5 4 × PopulationSize  then
6:
        if  I d e a A d a p t a t i o n < 0.05  then
7:
           Insert one solution;
8:
            NewPopulationSize NewPopulationSize + 1
9:
        else
10:
           if  ThresholdCapacity = 0.1  then
11:
               Insert two solutions;
12:
                NewPopulationSize NewPopulationSize + 2
13:
           end if
14:
        end if
15:
    end if
16:
end if

4.2. Neighbor Learner Groups

The student, in their quest for knowledge acquisition, is influenced by the social network that encompasses them. This influence is dependent on their relative standing within the network as well as their analogous competencies. The membership of this network is confined to a maximum of five participants. Moreover, the configuration of the student roundtable within a network is systematically modified following a predetermined number of generations (the attraction period within a network) to ensure that the dissemination of information encompasses all learners, thereby enhancing exploratory capabilities and facilitating the examination of other promising domains within the search space. This learning roundtable adheres to the following formula:
R T = | α · X g X |
X new = X g β · R T
where g denotes the number of students in the group and is randomly selected in the set 2,…, 5, while α and β (Figure 2) are two coefficients of learning attraction computed in the manner delineated below:
β = r 1 × 2.5 exp ( t 2 2 ( LIter / 3 ) 2 ) × 2 π
α = r 2 t LIter
where r1 are r2 are arbitrarily chosen within the interval [0; 1]. The coefficients α and β exhibit a linear decrement across successive generations, with β adhering to a Gaussian distribution while remaining within the confines of the learning attraction group, whereas LIter denotes the upper limit of iterations.
The variables that have previously been identified enable the student to engage effectively within their educational roundtable.
Figure 2. Variations in α and β over generations.
Figure 2. Variations in α and β over generations.
Information 16 00675 g002
Algorithm 3 Strategy for the student group within a limited neighborhood
1:
g = rand ( 2 , 5 ) /* random number { 2 , , 5 }
2:
R 0
3:
for  j 1 to g do
4:
     R T j = | α i · X i + j X i |
5:
     R = R + X i + j β j · R T j
6:
end for
7:
R = R / g
8:
if  f ( R ) < f ( X i )  then
9:
     X i = R
10:
end if
A student can augment their proficiency not solely by engaging with their immediate environment but also by seeking connections with fellow learners who exhibit analogous competencies. As a result, the algorithm optimally leverages the potential of the neighborhood. The pseudocode of Algorithm 3 was refined at the instruction level to promote educational interactions within a nearest-neighbor cohort for X i (comprising X i + 1 , X i + 2 , , X i + g ), wherein the group size is restricted to a maximum of five randomly selected learners with similar skills. These enhancements are encapsulated in the pseudocode presented in Algorithm 4. Furthermore, an additional strategy was employed in this research, as delineated in Algorithm 4.
Algorithm 4 Strategy of the randomly chosen learner group
1:
g = rand ( 1 , 5 ) /* random number { 1 , 2 , , 5 } */
2:
R 0
3:
for  j 1 to g do
4:
     K = rand ( 1 , LN ) /* random number { 1 , 2 , , LN } */    ▹ LN Number of learners
5:
     R T j = | α i · X K X i |
6:
     R = R + X i + j β j · D j
7:
end for
8:
R = R / g
9:
if  f ( R ) < f ( X i )  then
10:
     X i = R
11:
end if
Both algorithms indeed perform the insertion of g solutions into the population; however, their insertion strategies differ significantly. Algorithm 3 implements a neighborhood-based insertion approach, which aims to insert solutions by exploiting the local structure of the solution space. In contrast, Algorithm 4 generates the g solutions randomly and inserts them without considering neighborhood information.

4.3. Leadership Interaction

In the domain of competency-based education, an individual’s learning is shaped by the paradigms established by the highest-achieving peer and exhibits a favorable reaction to authoritative guidance. This authority figure is not exclusively the individual with the highest achievement but rather the learner who is most centrally located within the spectrum of group performance, effectively representing the overall average of the cohort. This reciprocal engagement cultivates the enhancement and evolution of each learner’s skill set. The repercussions of these interactions, both with the leading performer and the median of the group, reside in the examination of potentially fruitful avenues for exploration. A comprehensive elucidation of this collaborative group methodology is articulated as follows:
  • Each student assimilates or responds to the concept of optimality. Either the optimal solution is retained or alternative avenues for the optimal solution are exploited. The competency-based paradigm is inadequately encapsulated by the notion of the middle class. Within conventional educational frameworks, students are predominantly influenced by their peers who perform at an average level. Those who excel or struggle significantly are often excluded from the benefits of such pedagogical methodologies. The entirety of learning within this paradigm tends to prioritize certain means while disregarding alternative approaches. To mitigate this predicament, each student engages with the average in accordance with their individual capabilities, as they cultivate their own competencies. This dynamic enhances the exploratory essence of the algorithm, and consequently, facilitates the enrichment of diverse ideas (solutions to the problem).
The equation representing this interaction within the framework of competency-based learning can be expressed for a student X i utilizing Formula (28).
x i new = C i · x i old + ε i · ( X best λ · X mean )
where λ is a learning coefficient that belongs to the set { 1 , 2 } , X mean is the average value of all learners, X best is the value of the best learner, C i is the learner’s capability for X i in Formula (29), and r1 is a random number (r1∈ [0,1]).
C i = 1 r 1 + exp f i max ( f i ) Iter
where ϵ i is a modulation coefficient belonging to [ 0 , 1 ] , as shown in Formula (30).
ϵ i = 1 r 1 + exp f i max ( f i ) × Iter
where f i is the competency of the learner X i (the fitness function) and max ( f i ) is the best capability (best value of the fitness function) over the learning cycle. The parameters ϵ i and C i make it possible to control the diversity and improve the speed of convergence.

5. Operators for the TSP

This section introduces a novel toolkit of arithmetic operators tailored towards tackling combinatorial optimization problems, especially the TSP. Instead of using regular addition, multiplication, subtraction, and scalar multiplication, these operators are redefined to work with the discrete, permutation-based nature of TSP routes. The following sections will go into the specifics of these operators and how they are implemented.

5.1. Addition

When we “add” two TSP tours, written as S 1 S 2 , we employ the cycle crossover (CX) operator. CX, a well-established genetic algorithm technique, generates offspring by preserving cycles from parent tours while ensuring feasibility. In this context, CX is applied to S 1 and S 2 to produce a child tour. This ensures that we always obtain a high-quality solution, which is often optimal or very close to it (Figure 3).
The following example highlights this addition operator in the case of the TSP:
  • S 1 = [1, 2, 3, 4, 5, 6, 7, 8]
  • S 2 = [4, 3, 2, 1, 6, 7, 8, 5]
  • S 1 S 2 = [ 1 , 3 , 2 , 4 , 5 , 6 , 7 , 8 ]

5.2. Multiplication

When multiplying two routes, represented as S 1 S 2 , the Partially Mapped Crossover operator [35] comes into play. PMX is useful because it keeps the positions of cities consistent between the “parent” routes while creating new, viable routes. The PMX crossover generates two candidate offspring routes. Among these, the route with the lowest cost—measured in terms of CO2 consumption—is selected as the outcome of the multiplication operator.
This approach takes advantage of PMX’s ability to maintain good sub-route structures while also introducing variety into the search for the best overall route (Figure 4).
The following example illustrates this multiplication operator in the context of the TSP; however, between the two resulting solutions, the best tour is selected:
  • S 1 = [1, 2, 3, 4, 5, 6, 7, 8]
  • S 2 = [5, 6, 7, 8, 1, 2, 3, 4]
  • R 1 = S 1 S 2 = [1, 6, 3, 4, 5, 2, 7, 8]
  • R 2 = S 1 S 2 = [5, 2, 7, 8, 1, 6, 3, 4]
  • f ( X ) = i = 1 7 d π ( i ) , π ( i + 1 ) + d π ( 8 ) , π ( 1 )
We compute this sum for both R 1 and R 2 . If f( R 1 ) < f( R 2 ), then R 1 is selected; otherwise, R 2 is chosen.

5.3. Subtraction

The subtraction operation, denoted as S 1 S 2 , is defined as the addition of S 1 and the negation (inverse) of S 2 . Negation involves reversing the sequence of cities in S 2 to create an inverse tour. Once the negated version of S 2 is obtained, the previously defined addition operation ( S 1 negated S 2 ) is applied to compute the result. This definition extends the concept of subtraction to the combinatorial domain of TSP tours while ensuring consistency within the proposed framework.
For the subtraction operator, it is necessary to invert the second operand and then perform the addition.
  • S 1 = [1, 2, 3, 4, 5, 6, 7, 8]
  • S 2 = [5, 6, 7, 8, 1, 2, 3, 4] reverse( S 2 ) = [4, 3, 2, 1, 8, 7, 6, 5]
  • S 1 S 2 = S 1 reverse( S 2 ) = [1, 3, 2, 4, 5, 7, 6, 8]

5.4. Scalar Multiplication

Scalar multiplication, denoted as k S i , where k is a scalar value, is implemented through a two-step process:
  • Initially, the scalar ( K ) undergoes an initial transformation via one of the nine predefined transformation functions proposed in this study, converting it into an integer value within the set { 1 , 2 , , N } , where N denotes the population size. Following this conversion, the population is systematically ranked based on the fitness function. The swap mutation operator selects two random positions within the solution and exchanges their elements. This simple yet effective operation helps to maintain the permutation structure of the solution, which is crucial for routing problems. For example, if the solution is [1, 3, 2, 4], swapping positions 2 and 3 results in [1, 2, 3, 4].
    A swap mutation is then applied to the tour f ( K ) , where f ( K ) denotes the index of the solution within the population. The outcome of this mutation is a tour, which we shall designate as ( S sm ). Subsequently, we apply the multiplication operation, as defined in this study, with S i ( S i × S sm ).
  • Additionally, a heuristic mutation is applied. This mutation employs a problem-specific improvement strategy, such as 2-opt, to refine the tour further and enhance its quality.
This dual-stage approach ensures that scalar multiplication not only introduces variability but also integrates domain-specific heuristics to maintain solution quality. To illustrate the multiplication of a TSP tour (solution) by a scalar, the following example demonstrates this operation.
  • S 1 = [1, 2, 3, 4, 5, 6, 7, 8]
  • k = 0.25
The majority of functions employed in this study yield values approximating 25 (in this example). Consequently, the selected solution corresponds to the index equal to 25, specifically [2, 1, 3, 6, 5, 4, 7, 8].
After using swap mutation for [2, 1, 3, 6, 5, 4, 7, 8], S s m = [2, 5, 3, 6, 1, 4, 7, 8]. S 1 S s m = [5, 2, 3, 6, 1, 4, 7, 8].
After this, we apply heuristic mutation: [2, 8, 7, 4, 1, 6, 3, 5].
Let x [ 0 , 1 ] be a uniform real number. Nine transformation functions are defined as mappings from the continuous interval [0,1] to the discrete ordinal set { 1 , 2 , 3 , , N } , with N representing the population size.
  • Exponential transformation (parametric control):
    T 2 ( x ) = 100 x k + 1 , k > 0
    For k > 1 , smaller integers are favored (with an enhanced threshold effect).
  • Sigmoidal function (central concentration):
    T 3 ( x ) = 100 1 + e a ( x 0.5 ) + 1 , a 1
    Parameter a controls the transition slope (typically a 10 ).
  • Equidistant partition (indicator function):
    T 4 ( x ) = i = 1 100 i · I [ i 1 100 , i 100 ) ( x )
    Exact implementation of a discrete uniform CDF.
  • Power law (controlled bias):
    T 5 ( x ) = 100 x n , n N
    For n = 2 , P ( y 50 ) 75 % .
  • Logarithmic transform (long tail):
    T 6 ( x ) = min ( 100 ln ( 1 x ) + 1 , 100 )
    Requires special handling of x = 1 (to avoid ln ( 0 ) ).
  • Trigonometric modulation (periodicity):
    T 7 ( x ) = 50 ( sin ( 2 π x ) + 1 ) + 1
    Produces a bimodal distribution (favors extreme values).
  • CDF inverse (distributional adaptation):
    T 8 ( x ) = T 1 ( x ) , T CDF Cumulative Distribution Function .
    Allows the imposition of any arbitrary discrete distribution.
  • Stochastic method (secondary randomization):
    T 9 ( x ) U { 1 , , 100 x }
    Introduces additional variance controlled by x.
  • Linear–nonlinear mixture (compromise):
    T 10 ( x ) = 50 ( x + x 2 ) + 1
    Interpolation between linear (x) and quadratic ( x 2 ) behaviors.
  • For all functions T i , we have T i ( 0 ) = 1 and T i ( 1 ) = 100 , except
    -
    T 6 (logarithmic transform) requires x < 1 to avoid the divergence of ln ( 1 x ) at x = 1 ;
    -
    T 7 (trigonometric modulation) reaches its maximum value (101) at x = 0.25 and x = 0.75 , but not at x = 1 ;
    -
    T 9 (stochastic method) requires x > 0 to ensure the upper bound of the uniform distribution is at least 1.
  • All functions are piecewise continuous except for T 4 (equidistant partition), which is purely discrete.
  • Functions T 4 (uniform partition) and T 8 (CDF inverse) preserve important theoretical properties:
    -
    T 4 ensures fairness via uniform discretization;
    -
    T 8 is inherently invertible if the target CDF is strictly increasing.
  • Functions T 2 (exponential), T 5 (power law), and T 10 (linear–quadratic mixture) allow the explicit parametric control or structural modulation of the output distribution.
  • T 9 is the only non-deterministic transformation, introducing randomness via a uniform integer sampling based on x.
  • All functions require explicit boundary treatment to ensure well-defined behavior over the domain [ 0 , 1 ] .
In the experimental evaluation, the nine transformations are selected probabilistically at each iteration, thus promoting diversity across the search space (Figure 5).

5.5. Rationale and Advantages

To illustrate these operators, which enable techniques for transitioning from real encoding to permutation encoding for the TSP, we will consider the example of Equation (28).
x i new = C i x i old ε i X best ( λ X mean )
The proposed arithmetic operators address the unique challenges of combinatorial optimization problems, particularly those encountered in TSP. By redefining traditional arithmetic operations in terms of crossover and mutation techniques, the methodology facilitates the manipulation of tours while respecting problem constraints. The key advantages of this approach include the following:
  • Feasibility preservation: Each operator ensures that the resulting tours remain valid permutations of cities, avoiding infeasible solutions;
  • Diversity promotion: The integration of crossover and mutation mechanisms enhances exploration of the search space;
  • Heuristic integration: The inclusion of heuristic mutation in scalar multiplication balances exploration and exploitation, thereby improving solution quality;
  • Scalability: The modular design of the operators allows for adaptation to other combinatorial problems beyond TSP.
Empirical evaluation of the proposed operators on benchmark TSP instances will validate their performance and demonstrate their potential applicability to related optimization problems.

6. Experimental Studies

Dynamic optimization can be defined as a sequence of multiple instances of a static problem that are interconnected through certain dynamic rules. The primary aspects of “dynamism” are the frequency and magnitude of environmental changes. Frequency refers to the rate at which changes occur, while magnitude corresponds to the degree of these environmental changes. A change may involve factors such as the objective function, input variables, or problem constraints, for example. Environmental changes are classified as either dimensional or non-dimensional. Both types of changes cause the optimum to shift in position, shape, and value. In a dynamic problem, it is not only necessary to detect the optimum but also to track it over time. However, with such problems, the population of solutions tends to lose diversity over generations.
In [3], we examined the diversity mechanism of OCO in comparison with the random-immigrant genetic algorithm (RI-GA) and the hypermutation genetic algorithm (HMGA). In this study, the concept of novel ideas emerging from the learning process (immigrant insertion: implicitly present) was applied to OCO.
To assess the impact of OCO’s remedial strategies on population diversity, we recorded population diversity at each generation for every OCO run on dynamic optimization problems (DOPs). The mean population diversity of OCO across DOP generations over 30 runs was computed using the following formula:
Div ( t ) = 1 30 k = 1 30 1 ln ( n 1 ) i = 1 n j i n d i , j ( k , t )
where D i , j ( k , t ) denotes the Euclidean distance between the ith and jth individuals at generation t of the k t h run. Our findings revealed that OCO maintained the highest level of population diversity, whereas RIGA and HMGA retained the lowest. The results of the aforementioned study demonstrated that OCO, through parameter control, sustained high diversity levels across diverse dynamic environments.

6.1. Comparison with Other Metaheuristics

In the first experiment, we evaluated OCO against PSO, ACO, GA and GWO. However, the results obtained from these four comparative metaheuristics proved to be significantly suboptimal. The primary reason for this is that naturally, metaheuristics are designed to converge towards the optimum. Convergence simply means a set of solutions that repeats within the population of individuals. Consequently, the algorithm can no longer track solutions effectively.
Comparing the four metaheuristics designed exclusively for stationary optimization with our OCO metaheuristic, which is inherently crafted for both static and dynamic environments (incorporating the concept of new ideas emerging during the learning process), constitutes an inherently unfair comparison.
To address this disparity, we adopted a technique widely employed in dynamic optimization to track shifting optima: random immigrant insertion. This approach mitigates the limitations of static metaheuristics when applied to dynamic environments. The performance of our algorithm is tested on the DG-TSP. We generate the Dynamic TSP by introducing a factor t i j :
t i j = 1 + R if R [ F L , F U ) and q m , 1.0 otherwise .
where F L = 1 , F U = 5 , and q is a uniformly distributed random number in the interval  [ 0 , 1 ] .
In another variant of the dynamic TSP, changes occur in a cyclical manner. This cyclical environment can be constructed by simulating light, normal, or heavy traffic, with these changes occurring across different iterations (representing periods of the day). Here, t i j = c y c l i c i j ( s t a t e ) :
cyclic i j ( f r e q ) = 1 + R if R [ F L , F U ) and q m , 1.0 otherwise .
where ( m ) is a uniformly distributed random number in the interval [ 0 , 1 ] ; f r e q { 5 , 100 } represents the state of environmental change, whether rapid or slow.
In this study, ( m ) takes the values 0.1, 0.25, 0.5, and 0.75, indicating the degree of traffic intensity (light, normal, heavy, or very heavy), while ( f r e q { 5 , 10 , 50 , 100 } ) . These changes are applied to routes derived from instances in the TSPLib library [36], specifically kroA100, kroA150, and kroA200. These instances consist of 100, 150, and 200 cities, respectively, using Euclidean distance as the metric.
Our algorithm, incorporating new ideas (immigrant insertion), is compared with several state-of-the-art algorithms for dynamic optimization—namely ACO, PSO, GWO, and GA—all of which are enhanced with immigrant schemes specifically adapted to dynamic environments. Across the various instances, the performance is evaluated using the over offline performance (OOP) metric as defined in Jin and Branke [37]. This metric is defined as the sum of the measurements of the best solution found by the algorithm after detecting a dynamic change in each iteration (Equation (44)).
P o f f = 1 I t e r i = 1 I t e r 1 R u n j = 1 R u n P i j
Equation (44) defines the offline performance metric P o f f , which evaluates the average quality of the best solutions obtained over multiple runs and iterations. Specifically, R u n independent executions are carried out under identical dynamic scenarios, each for a total of I t e r iterations. At each iteration i of run j, the CO2 consumption of the best solution found is denoted by P i j . The inner summation computes the average best performance at iteration i across all runs, while the outer summation aggregates these values over the entire optimization horizon. This metric reflects the algorithm’s ability to consistently identify high-quality solutions over time in a dynamic environment.
For each algorithm-instance configuration, the reported offline performance (OOP) values in Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 are computed as statistical averages over 30 independent runs, each consisting of 500 iterations. This double-averaging process ensures robust estimation of long-term algorithmic behavior and significantly reduces variance. Although explicit confidence intervals are not shown, this approach provides stable and representative performance metrics suitable for comparative analysis.
The comparatively higher OOP values observed for RI-GWO, RI-PSO, RI-ACO, and RI-GA may be attributed to specific algorithmic characteristics, such as limited population diversity, slower adaptation to environmental changes, or the absence of embedded local search mechanisms. These aspects can affect how effectively each algorithm maintains solution quality over time in dynamic environments. By contrast, the proposed method integrates adaptive learning components and strategic diversification mechanisms, which may explain its comparatively better OOP performance across multiple scenarios.
In this first experimental setup, we set the number of independent runs to R u n = 30 , and the number of iterations to I t e r = 500 for each run.
Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 present the results of this comparison.
To better understand how the algorithms react to varying degrees of environmental change, we provide a detailed analysis of three representative instances: KroA100, KroA150, and KroA200. These instances were selected as they are typical of the small, medium, and large problem sizes in our benchmark set. Table 1 presents the results for the dynamic instance KroA100 under random traffic conditions with a change frequency f r e q = 5 , and varying levels of change magnitude { 0.1 , 0.25 , 0.5 , 0.75 } . The analysis focuses primarily on the offline performance (OOP), which serves as a key indicator of the algorithm’s adaptability and robustness in dynamic environments.
As shown in the table, the proposed OCO algorithm consistently outperforms all comparative approaches (RI-GWO, RI-ACO, RI-PSO, RI-GA) in terms of OOP across all magnitudes. At magnitude= 0.1 , OCO achieves an OOP of 21,290.9, which is significantly lower than that of RI-GWO (122,881.04), RI-ACO (36,957.6), RI-PSO (36,744.29), and RI-GA (59,296.5). This performance gap widens further as the change magnitude increases. In particular, when magnitude = 0.75 , OCO maintains an OOP of 20,677, effectively preserving the solution quality regardless of environmental severity. In contrast, all competing methods suffer from severe degradations; RI-GWO, for example, reaches an OOP of 125,660.76, while RI-PSO and RI-ACO exceed 36,000. These results highlight the superior reactivity and stability of OCO in dynamic scenarios, where the cost of poor adaptability is amplified over time. The ability of OCO to maintain low OOP values under different magnitudes suggests that it effectively integrates change detection and memory mechanisms, which allows it to track the moving optimum more efficiently than reactive-initialized metaheuristics (RI-GWO, RI-ACO, etc.). This confirms that OOP is a sensitive and appropriate metric to assess the dynamic behavior of algorithms in the context of green dynamic TSP. Moreover, OCO’s computational time remains within a reasonable range (approximately 22–25 s), offering a favorable trade-off between solution quality and runtime. By contrast, RI-GWO requires over 120 s with substantially poorer OOP results. In summary, the results from Table 1 clearly demonstrate that OCO exhibits a high degree of robustness and adaptability across varying levels of dynamism. Its superior OOP scores across all settings validate its effectiveness as a dynamic optimizer tailored for the Dynamic Green TSP.
To provide a comprehensive evaluation of the proposed OCO algorithm, the analysis was extended across multiple problem scales—namely KroA100, KroA150, and KroA200—under various settings of change frequency and change magnitude, thereby allowing for an in-depth assessment of its dynamic adaptability and performance stability.
The study of different tables (Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12) evaluates the performance across multiple dynamic scenarios with varying change frequencies (5, 10, 50, 100) and magnitudes (0.1 to 0.75).
To provide a deeper insight into the experimental results (Table 1, Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12), we analyze the impacts of two key dynamic parameters—frequency and magnitude of change—on algorithmic performance, as well as the differences observed between RI-GA and our proposed OCO algorithm.
First, as the change frequency f r e q increases, the dynamic environment becomes more volatile, requiring algorithms to adapt more frequently to maintain solution quality. This is reflected by a general decrease in performance across all methods as f grows, especially for algorithms without memory or adaptive components. OCO mitigates this effect through online adjustment and memory-based mechanisms that enable fast adaptation to frequent changes.
Second, higher change magnitudes introduce more disruptive shifts in the problem landscape. As expected, all algorithms experience a drop in quality when the magnitude increases. However, OCO maintains better stability and responsiveness compared to RI-GA, which suffers from delayed convergence due to its population-based structure.
Interestingly, while RI-GA achieves slightly shorter tour lengths in some instances (e.g., Length = 20,802 vs. 21,535 for OCO, as shown in Table 1), this does not translate into a better overall dynamic behavior. The Offline Performance (OOP) with Iter = 500 and Run = 30 , captures the algorithm’s robustness over time. According to this metric, OCO significantly outperforms RI-GA (e.g., OOP = 21,911.1 for OCO vs. 60,145.1 for RI-GA, as shown in Table 1), demonstrating its superior ability to track the shifting optimum in dynamic environments.
These observations confirm that evaluating algorithms solely on static quality measures (e.g., best tour length) may lead to misleading conclusions in dynamic contexts. Instead, metrics like OOP provide a more meaningful and time-sensitive assessment of adaptive behavior, where OCO shows clear advantages.
Key Observations:
  • Tour Length Performance:
    OCO consistently achieves superior performance with lower distances, maintaining stability even under high dynamics (e.g., magnitude = 0.75; frequency = 100; OCO = 26,786 vs. RI-GA = 64,870, as shown in Table 8).
  • Overall offline performance (OOP):
    For instance, at a change magnitude of 0.1, OCO achieves a tour length of 29,347 and an OOP of 29,654.7, significantly better than those of RI-GWO (OOP: 250,479.01), RI-ACO (OOP: 57,187.85), RI-PSO (OOP: 57,009.21), and RI-GA (OOP: 177,192.2) as shown in Table 12. This trend remains stable across increasing dynamic intensities, highlighting OCO’s robustness and adaptability to dynamic environments with both low and high degrees of change.
  • Computational efficiency (expressed in seconds):
    With respect to computational time, OCO incurs a higher computational cost than RI-ACO, RI-PSO, and RI-GA, yet remains substantially more efficient than RI-GWO. For example, at a change magnitude of 0.5, OCO’s execution time is approximately 333 s, whereas RI-GWO requires over 3000 s. While RI-ACO, RI-PSO, and RI-GA achieve faster runtimes (between 9 and 35 s), these are accompanied by significantly poorer solution quality in both tour length and OOP metrics.
    This trade-off suggests that OCO achieves a superior balance between dynamic performance and computational efficiency, making it particularly suitable for scenarios where solution accuracy and adaptability to environmental changes are critical, such as in green and sustainable logistics contexts.
Dynamic Performance Synthesis:
Resilience: The OCO algorithm consistently maintains a high-quality solution across all levels of environmental dynamism. For example, as the change magnitude increases from 0.1 to 0.75, the OOP values for OCO range only slightly from 27,762 to 27,890.9, and the tour length fluctuates marginally between 26,495 and 27,526. In contrast, RI-GWO exhibits severe degradation in OOP—from 190,501.7 to 187,354.89 (as shown in Table 7)—with relatively minor improvement in solution quality (lengths between 28,414 and 27,430), suggesting poor adaptation to increasing change levels. Similarly, RI-GA’s OOP escalates from 122,030.8 at 0.1 to 121,889.3 at 0.75, with tour lengths deteriorating significantly (from 54,251 to 60,577), showing that its solutions become less effective under dynamic conditions.
Quality–computation trade-off: While OCO’s computational time is higher than that of RI-ACO, RI-PSO, and RI-GA (e.g., OCO: 121–160 s vs. RI-ACO: 17–19 s, RI-PSO: 21–25 s, RI-GA: 12–38 s), this overhead is justified by substantially better performance in both tour length and OOP. For instance, at a change magnitude of 0.5, OCO yields a tour length of 26,495 and an OOP of 26,973.5, compared to RI-GA’s length of 51,720 and an OOP of 125,504.5, or RI-ACO’s length of 43,177 and an OOP of 49,190.58 (as shown in Table 7). Importantly, OCO also outperforms RI-GWO while being six to seven times faster (e.g., 113 s vs. 734–872 s) and producing significantly lower OOP values (OCO: 27,000 vs. RI-GWO: 187,000), affirming its computational efficiency relative to high-cost algorithms.
These findings validate that OCO achieves a robust balance between solution quality and computational time (for example, the average computational time of RI-GWO is approximately 6.6 times higher than that of OCO) making it especially well-suited for deployment in dynamic and environmentally constrained routing scenarios. The performance stability across varying change magnitudes strongly suggests that the algorithm maintains population diversity while avoiding premature convergence. This robustness can be attributed to OCO’s three-phase learning architecture, which enhances exploration through diversification and exploitation via localized refinement, thereby effectively adapting to environmental fluctuations in dynamic Green TSP instances.
Figure 6 compares the average Offline Performance (OOP) of five optimization algorithms, with a focus on CO2 emission minimization:
  • OCO achieves superior performance (OOP = 30,152), outperforming all baseline methods;
  • RI-GWO shows the weakest results (OOP = 254,934);
  • RI-ACO and RI-PSO demonstrate intermediate performance (57,000);
  • RI-GA performs better than RI-GWO but worse than OCO (OOP = 190,490).
The comparative analysis reveals substantial performance differentials among the evaluated algorithms:
  • Superiority of OCO: The algorithm demonstrates remarkable efficiency, achieving an 88% reduction in offline performance (OOP) compared to RI-GWO (30,152 vs. 254,934) and an 84% improvement over RI-GA. This significant margin suggests OCO’s enhanced capability to handle dynamic optimization constraints.
  • Intermediate Performers: RI-ACO and RI-PSO show comparable results, both being approximately 47% less efficient than OCO. Their similar performance profiles (57,000 OOP) indicate comparable limitations in adapting to environmental changes.
  • Relative underperformance: The substantial gap between OCO and RI-GWO (254,934 OOP) highlights fundamental limitations in the latter’s exploration–exploitation balance, particularly when addressing CO2 minimization in dynamic scenarios.
These findings underscore OCO’s algorithmic advantages in
  • Dynamic parameter adaptation;
  • Robustness against environmental variability;
  • Sustainable optimization capability.
The performance hierarchy (OCO ≫ RI-ACO ≈ RI-PSO > RI-GA ≫ RI-GWO) remains consistent across all tested change frequencies and magnitudes, suggesting structural rather than contextual superiority.
The significant performance gaps suggest
  • OCO’s adaptive learning mechanism effectively handles dynamic environmental changes;
  • Traditional metaheuristics (RI-GWO, RI-GA) struggle with frequency and magnitude changes;
  • Population-based methods (RI-ACO, RI-PSO) show moderate adaptability.

6.2. Statistical Tests

In the field of metaheuristic algorithm comparison, statistical validation is essential to assess whether observed performance differences are significant or merely due to stochastic variability. Given the non-deterministic nature of metaheuristics and the lack of normality in performance distributions, non-parametric statistical tests are generally preferred [38,39]. In this study, we adopt a rigorous statistical testing procedure including the Wilcoxon signed-rank test, the Friedman test, and the Nemenyi post hoc test, to compare the proposed OCO algorithm against four reference methods (RI-GWO, RI-ACO, RI-PSO, RI-GA) across several performance indicators.

6.2.1. Wilcoxon Signed-Rank Test

The Wilcoxon signed-rank test is a non-parametric alternative to the paired t-test, designed to compare the performance of two algorithms over multiple problem instances. It evaluates whether the median difference between paired values is significantly different from zero, without assuming the normal distribution of the data. The test operates under the following hypotheses:
  • H 0 : The two algorithms have equivalent performance distributions (median of differences is zero).
  • H 1 : The performance distributions differ significantly (median of differences is non-zero).
For each pair of algorithms, the differences in performance are computed, ranked by their absolute values, and summed separately for positive and negative ranks. The test statistic W is the minimum of these two sums:
W = min i : d i > 0 R i , j : d j < 0 R j
where d i is the difference between paired performances and R i is the rank of | d i | . A small p-value ( p < α = 0.05 ) indicates that the performance difference is statistically significant.
Results:
The Wilcoxon test was applied between OCO and each baseline method for three metrics: tour length, offline performance, and computation time. The results in Table 13 show statistically significant improvements of OCO over all baselines ( p < 0.001 in most cases).
These results confirm that OCO significantly outperforms RI-GWO, RI-ACO, RI-PSO, and RI-GA in terms of all considered metrics.

6.2.2. Friedman Test

To simultaneously compare multiple algorithms across several problem instances, we employ the Friedman test, a non-parametric alternative to repeated-measures Analysis of Variance (ANOVA). It tests the null hypothesis that all algorithms perform equally on average:
H 0 : All algorithms are statistically equivalent H 1 : At least one algorithm differs
The Friedman test statistic is calculated as
χ F 2 = 12 N k ( k + 1 ) j = 1 k R j 2 k ( k + 1 ) 2 4
where N is the number of problems, k is the number of algorithms, and R j is the average rank of algorithm j across all instances.
Results:
The Friedman test was applied for tour length, offline performance, and computation time. As reported in Table 14, the test shows highly significant differences among the algorithms ( p 5.3 × 10 12 for all metrics).

6.2.3. Nemenyi Post Hoc Test

Since the Friedman test indicates that not all algorithms are equivalent, we proceed with the Nemenyi test, which performs pairwise comparisons while controlling the family-wise error rate. It identifies which algorithms differ significantly in rank.
Results:
Table 15 displays the p-values of pairwise comparisons between OCO and the other algorithms for each metric.
The results show that OCO significantly outperforms RI-ACO and RI-GA in most cases. Differences with RI-PSO and RI-GWO are less pronounced and not always statistically significant.

7. Conclusions

This research introduces an innovative and effective methodology for addressing the Dynamic Green TSP, which incorporates time-varying travel conditions and realistic roadway features (e.g., inclines), with the objective of reducing energy consumption and greenhouse gas emissions. The proposed OCO algorithm integrates combinatorial operators within an adaptive metaheuristic framework, enabling robust, low-complexity solutions for dynamic environments. Empirical validation through extensive simulations and case studies confirms significant improvements in fuel efficiency, tour duration, and computational performance compared to state-of-the-art methods, particularly in highly dynamic contexts. Our methodological contribution lies in the direct integration of permutation-based combinatorial operators into existing real-valued metaheuristics, eliminating the need for complex encoding transformations and enabling broad applicability across dynamic optimization problems such as DVRP and adaptive scheduling. Anticipating future endeavors in research, there will be an emphasis on augmenting this theoretical framework to encompass multi-agent systems and integrating it within multi-objective planning paradigms to more effectively address the intricate relationship between economic efficiency, environmental sustainability, and service quality.

Author Contributions

Conceptualization, K.J.; Resources, R.B.; Writing – review & editing, R.B. and M.T.; Supervision, K.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available in a publicly accessible repository.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Osaba, E.; Yang, X.S.; Del Ser, J. Traveling salesman problem: A perspective review of recent research and new results with bio-inspired metaheuristics. In Nature-Inspired Computation and Swarm Intelligence; Elsevier: Amsterdam, The Netherlands, 2020; pp. 135–164. [Google Scholar]
  2. Gao, W.; Luo, Z.; Shen, H. A branch-and-price-and-cut algorithm for time-dependent pollution routing problem. Transp. Res. Part C Emerg. Technol. 2023, 156, 104339. [Google Scholar] [CrossRef]
  3. Ben Jelloun, R.; Jebari, K.; El Moujahid, A. Open Competency Optimization: A Human-Inspired Optimizer for the Dynamic Vehicle-Routing Problem. Algorithms 2024, 17, 449. [Google Scholar] [CrossRef]
  4. Nicoletti, B.; Appolloni, A. Green Logistics 5.0: A review of sustainability-oriented innovation with foundation models in logistics. Eur. J. Innov. Manag. 2024, 27, 542–561. [Google Scholar] [CrossRef]
  5. Wu, Y.; Wang, S.; Zhen, L.; Laporte, G. Integrating operations research into green logistics: A review. Front. Eng. Manag. 2023, 10, 517–533. [Google Scholar] [CrossRef]
  6. Blanco, E.E.; Sheffi, Y. Green logistics. In Sustainable Supply Chains: A Research-Based Textbook on Operations and Strategy; Springer: Berlin/Heidelberg, Germany, 2024; pp. 101–141. [Google Scholar]
  7. Zhang, M.; Sun, M.; Bi, D.; Liu, T. Green logistics development decision-making: Factor identification and hierarchical framework construction. IEEE Access 2020, 8, 127897–127912. [Google Scholar] [CrossRef]
  8. Greco, F. Travelling Salesman Problem; In-teh: London, UK, 2002; pp. 75–115. [Google Scholar]
  9. Halim, A.H.; Ismail, I. Combinatorial optimization: Comparison of heuristic algorithms in travelling salesman problem. Arch. Comput. Methods Eng. 2019, 26, 367–380. [Google Scholar] [CrossRef]
  10. Moghdani, R.; Salimifard, K.; Demir, E.; Benyettou, A. The green vehicle routing problem: A systematic literature review. J. Clean. Prod. 2021, 279, 123691. [Google Scholar] [CrossRef]
  11. Küçükoğlu, İ.; Ene, S.; Aksoy, A.; Öztürk, N. A memory structure adapted simulated annealing algorithm for a green vehicle routing problem. Environ. Sci. Pollut. Res. 2015, 22, 3279–3297. [Google Scholar] [CrossRef]
  12. Úbeda, S.; Faulin, J.; Serrano, A.; Arcelus, F.J. Solving the green capacitated vehicle routing problem using a tabu search algorithm. Lect. Notes Manag. Sci. 2014, 6, 141–149. [Google Scholar]
  13. Zhang, C.; Zhao, Y.; Leng, L. A hyper-heuristic algorithm for time-dependent green location routing problem with time windows. IEEE Access 2020, 8, 83092–83104. [Google Scholar] [CrossRef]
  14. Rodríguez-Esparza, E.; Masegosa, A.D.; Oliva, D.; Onieva, E. A new hyper-heuristic based on adaptive simulated annealing and reinforcement learning for the capacitated electric vehicle routing problem. Expert Syst. Appl. 2024, 252, 124197. [Google Scholar] [CrossRef]
  15. Zhang, S.; Gajpal, Y.; Appadoo, S. A meta-heuristic for capacitated green vehicle routing problem. Ann. Oper. Res. 2018, 269, 753–771. [Google Scholar] [CrossRef]
  16. Andelmin, J.; Bartolini, E. A multi-start local search heuristic for the green vehicle routing problem based on a multigraph reformulation. Comput. Oper. Res. 2019, 109, 43–63. [Google Scholar] [CrossRef]
  17. Zhao, W.; Bian, X.; Mei, X. An Adaptive Multi-Objective Genetic Algorithm for Solving Heterogeneous Green City Vehicle Routing Problem. Appl. Sci. 2024, 14, 6594. [Google Scholar] [CrossRef]
  18. Micale, R.; Marannano, G.; Giallanza, A.; Miglietta, P.; Agnusdei, G.; La Scalia, G. Sustainable vehicle routing based on firefly algorithm and TOPSIS methodology. Sustain. Futur. 2019, 1, 100001. [Google Scholar] [CrossRef]
  19. Bektaş, T.; Laporte, G. The pollution-routing problem. Transp. Res. Part B Methodol. 2011, 45, 1232–1250. [Google Scholar] [CrossRef]
  20. Guntsch, M.; Middendorf, M.; Schmeck, H. An ant colony optimization approach to dynamic TSP. In Proceedings of the 3rd Annual Conference on Genetic and Evolutionary Computation, Málaga, Spain, 14–18 July 2025; Morgan Kaufmann Publishers Inc.: Burlington, MA, USA, 2001; pp. 860–867. [Google Scholar]
  21. Guntsch, M.; Middendorf, M. Pheromone modification strategies for ant algorithms applied to dynamic TSP. In Proceedings of the Applications of Evolutionary Computing, Como, Italy, 18–20 April 2001; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
  22. Psaraftis, H.N. Dynamic vehicle routing: Status and prospects. Ann. Oper. Res. 1995, 61, 143–164. [Google Scholar] [CrossRef]
  23. Stodola, P.; Michenka, K.; Nohel, J.; Rybanskỳ, M. Hybrid algorithm based on ant colony optimization and simulated annealing applied to the dynamic traveling salesman problem. Entropy 2020, 22, 884. [Google Scholar] [CrossRef]
  24. Kang, L.; Zhou, A.; McKay, B.; Li, Y.; Kang, Z. Benchmarking algorithms for dynamic travelling salesman problems. In Proceedings of the 2004 Congress on Evolutionary Computation (IEEE Cat. No.04TH8753), Portland, OR, USA, 19–23 June 2004; Volume 2, pp. 1286–1292. [Google Scholar]
  25. Younes, A.; Basir, O.; Calamai, P. Adaptive control of genetic parameters for dynamic combinatorial problems. In Metaheuristics: Progress in Complex Systems Optimization; Springer: Berlin/Heidelberg, Germany, 2007; pp. 205–223. [Google Scholar]
  26. Mavrovouniotis, M.; Yang, S.; Yao, X. A benchmark generator for dynamic permutation-encoded problems. In Proceedings of the International Conference on Parallel Problem Solving from Nature, Taormina, Italy, 1–5 September 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 508–517. [Google Scholar]
  27. Mavrovouniotis, M.; Yang, S. Ant algorithms with immigrants schemes for the dynamic vehicle routing problem. Inf. Sci. 2015, 294, 456–477. [Google Scholar] [CrossRef]
  28. Touzout, F.A.; Ladier, A.L.; Hadj-Hamou, K. An assign-and-route matheuristic for the time-dependent inventory routing problem. Eur. J. Oper. Res. 2022, 300, 1081–1097. [Google Scholar] [CrossRef]
  29. Asghari, M.; Mirzapour Al-e-hashem, S.M.J. Green vehicle routing problem: A state-of-the-art review. Int. J. Prod. Econ. 2021, 231, 107899. [Google Scholar] [CrossRef]
  30. Daşcıoğlu, B.G.; Yazgan, H.R. Dynamic green location and routing problem for service points. Int. J. Procure. Manag. 2020, 13, 112–133. [Google Scholar] [CrossRef]
  31. Gao, Z.; Xu, X.; Hu, Y.; Wang, H.; Zhou, C.; Zhang, H. Based on improved NSGA-II algorithm for solving time-dependent green vehicle routing problem of urban waste removal with the consideration of traffic congestion: A case study in China. Systems 2023, 11, 173. [Google Scholar] [CrossRef]
  32. Çimen, M.; Soysal, M. Time-dependent green vehicle routing problem with stochastic vehicle speeds: An approximate dynamic programming algorithm. Transp. Res. Part Transp. Environ. 2017, 54, 82–98. [Google Scholar] [CrossRef]
  33. Dündar, H.; Soysal, M.; Ömürgönülşen, M.; Kanellopoulos, A. A green dynamic TSP with detailed road gradient dependent fuel consumption estimation. Comput. Ind. Eng. 2022, 168, 108024. [Google Scholar] [CrossRef]
  34. Lu, Y.; Yuan, Y.; Yasenjiang, J.; Sitahong, A.; Chao, Y.; Wang, Y. An Optimized Method for Solving the Green Permutation Flow Shop Scheduling Problem Using a Combination of Deep Reinforcement Learning and Improved Genetic Algorithm. Mathematics 2025, 13, 545. [Google Scholar] [CrossRef]
  35. Zhang, P.; Wang, J.; Tian, Z.; Sun, S.; Li, J.; Yang, J. A genetic algorithm with jumping gene and heuristic operators for traveling salesman problem. Appl. Soft Comput. 2022, 127, 109339. [Google Scholar] [CrossRef]
  36. TSPLIB95. 2025. Available online: http://comopt.ifi.uni-heidelberg.de/software/TSPLIB95/tsp/ (accessed on 10 April 2025).
  37. Jin, Y.; Branke, J. Evolutionary optimization in uncertain environments-a survey. IEEE Trans. Evol. Comput. 2005, 9, 303–317. [Google Scholar] [CrossRef]
  38. García, S.; Fernández, A.; Luengo, J.; Herrera, F. Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power. Inf. Sci. 2010, 180, 2044–2064. [Google Scholar] [CrossRef]
  39. Derrac, J.; García, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput. 2011, 1, 3–18. [Google Scholar] [CrossRef]
Figure 3. Visualization of the TSP tour before and after applying ⊕ combinatorial operators.
Figure 3. Visualization of the TSP tour before and after applying ⊕ combinatorial operators.
Information 16 00675 g003
Figure 4. Visualization of the TSP tour before and after applying ⊗ combinatorial operators.
Figure 4. Visualization of the TSP tour before and after applying ⊗ combinatorial operators.
Information 16 00675 g004
Figure 5. Transformation functions.
Figure 5. Transformation functions.
Information 16 00675 g005
Figure 6. Overall offline performance.
Figure 6. Overall offline performance.
Information 16 00675 g006
Table 1. Instance KroA100 with random traffic and freq = 5.
Table 1. Instance KroA100 with random traffic and freq = 5.
AlgorithmsChange_MagnitudeLengthOOPComputational_Time
OCO0.121,05921,290.922.3921022415161
RI-GWO0.121,949122,881.04127.165235280991
RI-ACO0.132,99436,957.68.11824822425842
RI-PSO0.121,70136,744.299.33303236961365
RI-GA0.121,25159,296.515.7245984077454
OCO0.2521,53521,911.123.079371213913
RI-GWO0.2523,035123,244.58128.834784030914
RI-ACO0.2531,92836,809.048.31470632553101
RI-PSO0.2523,22437,126.019.31120896339417
RI-GA0.2520,80260,145.117.3999660015106
OCO0.520,49220,775.925.1404733657837
RI-GWO0.521,298126,085.52132.616602420807
RI-ACO0.532,95136,863.788.51267838478088
RI-PSO0.523,11636,728.879.72644925117493
RI-GA0.521,79554,423.216.6300027370453
OCO0.7520,67720,67724.9908838272095
RI-GWO0.7521,709125,660.76121.951491594315
RI-ACO0.7532,76736,773.79.96449542045593
RI-PSO0.7521,59537,064.610.0018041133881
RI-GA0.7521,85357,858.212.9152412414551
Table 2. Instance KroA100 with random traffic and freq = 10.
Table 2. Instance KroA100 with random traffic and freq = 10.
AlgorithmsChange_MagnitudeLengthOOPComputational_Time
OCO0.121,21321,755.420.9276142120361
RI-GWO0.122,415123,082141.25444483757
RI-ACO0.132,64236,868.928.4200701713562
RI-PSO0.123,36636,433.439.23282861709595
RI-GA0.122,26154,110.112.2368183135986
OCO0.2520,94121,319.119.2564599514008
RI-GWO0.2522,445121,673.9135.933153390884
RI-ACO0.2532,02537,099.718.23803329467773
RI-PSO0.2521,93836,932.479.66085481643677
RI-GA0.2521,94460,938.417.2285289764404
OCO0.521,11321,733.417.0954973697662
RI-GWO0.522,685124,461.43125.362582683563
RI-ACO0.532,46837,053.778.51956224441528
RI-PSO0.522,37536,750.329.71863651275635
RI-GA0.521,27265,36016.1961874961853
OCO0.7521,60021,698.620.8990180492401
RI-GWO0.7522,810122,989.42128.998858451843
RI-ACO0.7531,58737,074.318.57306122779846
RI-PSO0.7522,78737,073.639.89743041992188
RI-GA0.7521,33255,11011.2393550872803
Table 3. Instance KroA100 with random traffic and freq = 50.
Table 3. Instance KroA100 with random traffic and freq = 50.
AlgorithmsChange_MagnitudeLengthOOPComputational_Time
OCO0.121,38121,719.813.9666087627411
RI-GWO0.122,508125,408.07131.110047101975
RI-ACO0.132,38437,043.38.42682528495789
RI-PSO0.122,70336,992.849.46939182281494
RI-GA0.122,29757,462.113.7826321125031
OCO0.2521,12321,356.221.1810195446014
RI-GWO0.2522,165124,025.48134.562022924423
RI-ACO0.2533,51536,917.278.55448985099793
RI-PSO0.2522,19737,195.9810.2855339050293
RI-GA0.2521,23759,372.911.4311144351959
OCO0.520,99621,833.324.4424633979797
RI-GWO0.521,806123,893.6127.856323480606
RI-ACO0.532,07136,921.348.39604187011719
RI-PSO0.523,58036,845.2110.1386339664459
RI-GA0.520,82660,281.113.3048303127289
OCO0.7521,49821,90025.3512442111969
RI-GWO0.7522,563125,067.34121.043132781982
RI-ACO0.7533,27437,134.158.59075498580933
RI-PSO0.7521,98036,792.849.66982936859131
RI-GA0.7521,07251,068.412.845165014267
Table 4. Instance KroA100 with random traffic and freq = 100.
Table 4. Instance KroA100 with random traffic and freq = 100.
AlgorithmsChange_MagnitudeLengthOOPComputational_Time
OCO0.121,43221,640.424.6619594097138
RI-GWO0.121,847125,376.59122.259998083115
RI-ACO0.132,99937,034.768.71673130989075
RI-PSO0.122,89036,705.159.7340042591095
RI-GA0.121,20458,76315.9396677017212
OCO0.2520,93321,222.718.1225333213806
RI-GWO0.2522,386120,743.46132.594746589661
RI-ACO0.2532,15336,821.598.32460141181946
RI-PSO0.2522,93237,105.089.6085958480835
RI-GA0.2520,69456,726.313.8298809528351
OCO0.520,73821,019.821.4589228630066
RI-GWO0.522,282124,836.81130.818113327026
RI-ACO0.531,30436,751.178.78234338760376
RI-PSO0.521,86336,829.6810.0047512054443
RI-GA0.520,91761,604.214.5410196781158
OCO0.7520,82721,206.823.4733896255493
RI-GWO0.7521,874123,096.34128.855330705643
RI-ACO0.7532,79036,805.249.08334875106812
RI-PSO0.7522,17336,951.2410.3831281661987
RI-GA0.7521,74658,370.424.3120296001434
Table 5. Instance KroA150 with random traffic and freq = 5.
Table 5. Instance KroA150 with random traffic and freq = 5.
AlgorithmsChange_MagnitudeLengthOOPComputation_Time
OCO0.127,51427,537.497.709445476532
RI-GWO0.127,691192,926.74808.249512910843
RI-ACO0.144,47148,827.6214.8391666412354
RI-PSO0.127,88348,902.1418.997878074646
RI-GA0.151,134115,868.4124.0980129241943
OCO0.2527,53928,145.8100.827348232269
RI-GWO0.2528,224198,925.63761.52139878273
RI-ACO0.2544,69149,477.8515.3931782245636
RI-PSO0.2527,48549,08220.4046738147736
RI-GA0.2563,970119,283.818.8867592811585
OCO0.526,38527,051107.441903352737
RI-GWO0.527,061183,743.64685.366256713867
RI-ACO0.543,17448,899.8915.3056991100311
RI-PSO0.529,01949,057.8318.8881287574768
RI-GA0.566,929125,583.821.7121329307556
OCO0.7527,07927,404.5123.916888237
RI-GWO0.7528,196193,147.6763.206788778305
RI-ACO0.7541,11249,126.1515.0603370666504
RI-PSO0.7527,18749,133.5320.5533874034882
RI-GA0.7564,776122,469.648.262743473053
Table 6. Instance KroA150 with random traffic and freq = 10.
Table 6. Instance KroA150 with random traffic and freq = 10.
AlgorithmsChange_MagnitudeLengthOOPComputation_Time
OCO0.126,37926,659.997.8753378391266
RI-GWO0.127,858193,429.5714.04919552803
RI-ACO0.144,11848,824.1614.9826173782349
RI-PSO0.128,38648,637.320.6628074645996
RI-GA0.157,195117,65114.7551081180573
OCO0.2526,84227,165.5126.872204780579
RI-GWO0.2527,508189,454.851025.27630829811
RI-ACO0.2544,30148,997.6717.3445875644684
RI-PSO0.2527,79449,175.6222.2102901935577
RI-GA0.2558,335112,912.134.976592540741
OCO0.526,54226,662.7134.05682182312
RI-GWO0.527,838190,610.731075.06168675423
RI-ACO0.542,87348,877.3618.8859694004059
RI-PSO0.528,64349,179.1922.0664458274841
RI-GA0.579,180124,260.746.1346783638001
OCO0.7527,81827,818152.461178541184
RI-GWO0.7528,778190,314.64884.227702140808
RI-ACO0.7543,79749,108.4518.1741693019867
RI-PSO0.7528,55049,162.8122.0895233154297
RI-GA0.7564,176121,569.820.1959164142609
Table 7. Instance KroA150 with random traffic and freq = 50.
Table 7. Instance KroA150 with random traffic and freq = 50.
AlgorithmsChange_MagnitudeLengthOOPComputation_Time
OCO0.127,52627,762121.465321063995
RI-GWO0.128,414190,501.7830.968395471573
RI-ACO0.144,37149,287.2817.7421343326569
RI-PSO0.128,01148,972.6921.2658190727234
RI-GA0.154,251122,030.838.3366434574127
OCO0.2526,84927,048.1160.889377355576
RI-GWO0.2527,152187,012.87852.514069080353
RI-ACO0.2542,29649,089.1617.8825562000275
RI-PSO0.2528,02849,131.9822.7740960121155
RI-GA0.2557,388120,709.223.9269785881043
OCO0.526,49526,973.5113.92518901825
RI-GWO0.527,340187,346.31734.989396572113
RI-ACO0.543,17749,190.5819.0754761695862
RI-PSO0.529,01749,260.8224.9999492168427
RI-GA0.551,720125,504.512.8652136325836
OCO0.7526,69127,890.9124.87078166008
RI-GWO0.7527,430187,354.89872.875902891159
RI-ACO0.7544,23649,154.9517.7652485370636
RI-PSO0.7528,56949,329.8122.5915353298187
RI-GA0.7560,577121,889.322.4923396110535
Table 8. Instance KroA150 with random traffic and freq=100.
Table 8. Instance KroA150 with random traffic and freq=100.
AlgorithmsChange_MagnitudeLengthOOPComputation_Time
OCO0.126,48326,967.1151.894057273865
RI-GWO0.127,697180,809.23901.101784229279
RI-ACO0.144,99449,155.8717.162840127945
RI-PSO0.128,52749,025.6621.9899196624756
RI-GA0.160,838120,72232.3962223529816
OCO0.2527,61827,912.8158.637745141983
RI-GWO0.2527,085186,486.95821.483963251114
RI-ACO0.2544,50249,334.3417.9162790775299
RI-PSO0.2528,05249,130.4522.5717957019806
RI-GA0.2568,551136,92833.0491642951965
OCO0.526,63426,822.1106.701067209244
RI-GWO0.527,528188,537.37842.777728319168
RI-ACO0.544,13049,125.6717.9808030128479
RI-PSO0.527,30749,025.2821.8809659481049
RI-GA0.558,353117,597.230.5044939517975
OCO0.7526,78626,896.5142.47221159935
RI-GWO0.7527,566189,727.8914.117618083954
RI-ACO0.7542,76849,019.3218.6860890388489
RI-PSO0.7528,92849,119.6521.5470142364502
RI-GA0.7564,870127,44227.0094890594482
Table 9. Instance KroA200 with random traffic and freq = 5.
Table 9. Instance KroA200 with random traffic and freq = 5.
AlgorithmsChange_MagnitudeLengthOOPComputation_Time
OCO0.129,94430,206.1359.626891613007
RI-GWO0.130,594249,415.272679.29984951019
RI-ACO0.152,04057,023.4621.3514449596405
RI-PSO0.131,23357,408.6429.1911859512329
RI-GA0.1127,401188,598.420.3655452728272
OCO0.2530,34830,588297.62440609932
RI-GWO0.2530,245255,258.472495.94751191139
RI-ACO0.2551,16657,067.1721.998370885849
RI-PSO0.2531,53757,454.1631.5070748329163
RI-GA0.25120,371188,276.6613.6076040267944
OCO0.529,50629,820.4353.145201206207
RI-GWO0.530,662248,967.742924.53922724724
RI-ACO0.551,04356,992.9521.9690067768097
RI-PSO0.531,30957,256.1431.3970522880554
RI-GA0.5108,760190,927.58.91393232345581
OCO0.7529,97130,516.5410.367824316025
RI-GWO0.7529,946251,574.782558.51860165596
RI-ACO0.7552,45657,456.9122.2017307281494
RI-PSO0.7531,08857,208.2431.7390558719635
RI-GA0.75104,103179,700.810.56822681427
Table 10. Instance KroA200 with random traffic and freq = 10.
Table 10. Instance KroA200 with random traffic and freq = 10.
AlgorithmsChange_MagnitudeLengthOOPComputation_Time
OCO0.129,71330,080319.550868988037
RI-GWO0.130,343259,596.232727.19788074493
RI-ACO0.152,37157,396.3722.1682703495026
RI-PSO0.131,33657,121.0533.6118106842041
RI-GA0.1147,503206,709.3140.0141541957855
OCO0.2529,32829,523362.919169902802
RI-GWO0.2530,347253,222.112659.1444709301
RI-ACO0.2552,31457,486.2621.7984554767609
RI-PSO0.2532,08257,334.8229.6725142002106
RI-GA0.25120,751189,406.38.541428565979
OCO0.530,68130,876305.512933254242
RI-GWO0.530,888254,072.172496.83245229721
RI-ACO0.553,07057,826.8822.5814046859741
RI-PSO0.531,82557,389.2530.6847245693207
RI-GA0.5115,006188,956.317.9437294006348
OCO0.7529,51629,894.3415.858951568604
RI-GWO0.7530,352255,688.222832.66897535324
RI-ACO0.7550,24357,443.9222.2953035831451
RI-PSO0.7531,06357,491.2130.4808707237244
RI-GA0.75131,803200,674.718.042248249054
Table 11. Instance KroA200 with random traffic and freq = 50.
Table 11. Instance KroA200 with random traffic and freq = 50.
AlgorithmsChange_MagnitudeLengthOOPComputation_Time
OCO0.129,21029,729.5451.447427988052
RI-GWO0.130,854257,405.182592.22149825096
RI-ACO0.150,96457,112.5821.6153562068939
RI-PSO0.130,81157,450.2832.6675596237183
RI-GA0.1104,405184,189.119.2394804954529
OCO0.2529,04229,800.8415.193927288055
RI-GWO0.2531,585257,293.012587.61305117607
RI-ACO0.2552,74457,805.2122.3461935520172
RI-PSO0.2531,21657,549.130.3804461956024
RI-GA0.25118,791191,227.610.8311760425568
OCO0.530,15930,541.2338.287764072418
RI-GWO0.531,024258,235.323005.51641082764
RI-ACO0.551,17457,097.4223.0797967910767
RI-PSO0.530,73357,288.7632.4898002147675
RI-GA0.5139,844196,514.929.3715693950653
OCO0.7530,77331,107.1353.920367956162
RI-GWO0.7530,745256,092.132695.67001247406
RI-ACO0.7549,63857,178.9323.6201846599579
RI-PSO0.7531,85657,250.2131.6116788387299
RI-GA0.75103,571182,581.412.4816913604736
Table 12. Instance KroA200 with random traffic and freq = 100.
Table 12. Instance KroA200 with random traffic and freq = 100.
AlgorithmsChange_MagnitudeLengthOOPComputation_Time
OCO0.129,34729,654.7459.092036247253
RI-GWO0.130,037250,479.013030.58809566498
RI-ACO0.150,68357,187.8523.9921379089356
RI-PSO0.131,67657,009.2135.8889374732971
RI-GA0.1103,736177,192.219.6413018703461
OCO0.2529,21929,821.6499.806445360184
RI-GWO0.2530,100252,264.762929.30351305008
RI-ACO0.2553,36257,415.5523.0371935367584
RI-PSO0.2532,89657,003.630.6185252666473
RI-GA0.25121,289193,924.824.688747882843
OCO0.529,36730,011.7333.434591770172
RI-GWO0.530,207261,022.393064.4821677208
RI-ACO0.551,70557,332.8823.3675134181976
RI-PSO0.531,39657,218.2934.7962794303894
RI-GA0.5110,632195,980.89.07850098609924
OCO0.7529,65830,261.6478.82949590683
RI-GWO0.7531,108258,356.642811.53076457977
RI-ACO0.7552,29657,220.3223.4361596107483
RI-PSO0.7530,64456,953.9633.5193908214569
RI-GA0.75132,413192,981.324.235230922699
Table 13. Wilcoxon signed-rank test ( α = 0.05 ).
Table 13. Wilcoxon signed-rank test ( α = 0.05 ).
MetricComparisonp-ValueSignificant
Tour lengthOCO vs. RI-GWO0.00015Yes
Tour lengthOCO vs. RI-ACO0.00003Yes
Tour lengthOCO vs. RI-PSO0.00006Yes
Tour lengthOCO vs. RI-GA0.00003Yes
Offline performanceOCO vs. RI-GWO0.00003Yes
Offline performanceOCO vs. RI-ACO0.00003Yes
Offline performanceOCO vs. RI-PSO0.00003Yes
Offline performanceOCO vs. RI-GA0.00003Yes
Computation timeOCO vs. RI-GWO0.00003Yes
Computation timeOCO vs. RI-ACO0.00003Yes
Computation timeOCO vs. RI-PSO0.00003Yes
Computation timeOCO vs. RI-GA0.00003Yes
Table 14. Friedman test results.
Table 14. Friedman test results.
Metric χ 2 p-ValueSignificant
Tour length59 4.7 × 10 12 Yes
Offline performance61 1.79 × 10 12 Yes
Computation time58.75 5.3 × 10 12 Yes
Table 15. Nemenyi post hoc test ( α = 0.05 ).
Table 15. Nemenyi post hoc test ( α = 0.05 ).
MetricComparisonp-Value
Tour lengthOCO vs. RI-ACO 2.69 × 10 6
Tour lengthOCO vs. RI-GA 4.15 × 10 11
Tour lengthOCO vs. RI-GWO0.259
Tour lengthOCO vs. RI-PSO0.056
Offline performanceOCO vs. RI-ACO 7.99 × 10 7
Offline performanceOCO vs. RI-GA0.030
Offline performanceOCO vs. RI-GWO0.10
Offline performanceOCO vs. RI-PSO 8.34 × 10 12
Computation timeOCO vs. RI-ACO0.076
Computation timeOCO vs. RI-GA0.021
Computation timeOCO vs. RI-GWO 2.69 × 10 6
Computation timeOCO vs. RI-PSO0.380
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Benjelloun, R.; Tarik, M.; Jebari, K. Open Competency Optimization with Combinatorial Operators for the Dynamic Green Traveling Salesman Problem. Information 2025, 16, 675. https://doi.org/10.3390/info16080675

AMA Style

Benjelloun R, Tarik M, Jebari K. Open Competency Optimization with Combinatorial Operators for the Dynamic Green Traveling Salesman Problem. Information. 2025; 16(8):675. https://doi.org/10.3390/info16080675

Chicago/Turabian Style

Benjelloun, Rim, Mouna Tarik, and Khalid Jebari. 2025. "Open Competency Optimization with Combinatorial Operators for the Dynamic Green Traveling Salesman Problem" Information 16, no. 8: 675. https://doi.org/10.3390/info16080675

APA Style

Benjelloun, R., Tarik, M., & Jebari, K. (2025). Open Competency Optimization with Combinatorial Operators for the Dynamic Green Traveling Salesman Problem. Information, 16(8), 675. https://doi.org/10.3390/info16080675

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop