Next Article in Journal
Non-Parametric Model-Based Estimation of the Effective Reproduction Number for SARS-CoV-2
Next Article in Special Issue
Solving NP-Hard Challenges in Logistics and Transportation under General Uncertainty Scenarios Using Fuzzy Simheuristics
Previous Article in Journal
Fast Local Laplacian Filter Based on Modified Laplacian through Bilateral Filter for Coronary Angiography Medical Imaging Enhancement
Previous Article in Special Issue
A Framework for Determining Collision Likelihood Using Continuous Friction Values in a Connected Vehicle Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Learnheuristic Algorithm for the Capacitated Dispersion Problem under Dynamic Conditions

1
Research Center on Production Management and Engineering, Universitat Politècnica de València, 03801 Alcoy, Spain
2
Department of Computer Architecture & Operating Systems, Universitat Autònoma de Barcelona, 08193 Bellaterra, Spain
*
Author to whom correspondence should be addressed.
Algorithms 2023, 16(12), 532; https://doi.org/10.3390/a16120532
Submission received: 30 October 2023 / Revised: 14 November 2023 / Accepted: 17 November 2023 / Published: 22 November 2023
(This article belongs to the Special Issue Optimization Algorithms in Logistics, Transportation, and SCM)

Abstract

:
The capacitated dispersion problem, which is a variant of the maximum diversity problem, aims to determine a set of elements within a network. These elements could symbolize, for instance, facilities in a supply chain or transmission nodes in a telecommunication network. While each element typically has a bounded service capacity, in this research, we introduce a twist. The capacity of each node might be influenced by a random Bernoulli component, thereby rendering the possibility of a node having zero capacity, which is contingent upon a black box mechanism that accounts for environmental variables. Recognizing the inherent complexity and the NP-hard nature of the capacitated dispersion problem, heuristic algorithms have become indispensable for handling larger instances. In this paper, we introduce a novel approach by hybridizing a heuristic algorithm with reinforcement learning to address this intricate problem variant.

1. Introduction

Within the evolving dynamics of supply networks, facility capacity limitation has emerged as an undeniable constraint. Deciding upon these capacities is spotlighted as a pivotal strategic decision in supply network design [1,2,3]. However, as challenges such as the facility location problem (FLP) are discussed, a gap is observed: the fluctuating nature of capacity in response to external drivers [4,5,6]. Maximum diversity problems (MDPs) offer an additional dimension, thereby focusing on location without factoring in allocation [4]. These challenges, while relevant, often do not encompass strategic capacity considerations, especially when demand outpaces supply. To address this gap, we introduce an innovative cross between greedy algorithms [7] and reinforcement learning [8]. This amalgamation allows for more dynamic, environment-tuned node selection, thus accounting for shifting factors and unpredictable demands. Our approach surpasses the limitations of traditional models, such as Model (1), by adapting and learning from the environment in real time, thereby ensuring optimal demand fulfillment as discussed in Juan et al. [9].
In supply networks, facilities frequently operate under inherent capacity restrictions. Whether it is a production plant or a distribution center, there is a finite amount of product that can be manufactured or managed within a specified timeframe. The first one to define capacity for this problem was Rosenkrantz et al. [10]. Specifically, the objective of the capacitated dispersion problem (CDP) is to enhance the dispersion of the chosen elements (or sites, as in the context of facility location applications) while adhering to capacity restrictions that represent a minimum storage capacity requirement. A predominant theme in the diversity problem literature revolves around predetermined selection criteria. Specifically, it concerns the determination of the optimal number of elements or facilities to be chosen. The overarching goal is to reflect the innate diversity within the original set of options, thereby ensuring that a function reaches either a maximal or minimal value in the context of combinatorial optimization. Model (1) serves as a representation of such diversity problems. It aims to maximize the smallest distance between any two selected elements, which are otherwise termed as active facilities. In this model, we have the following definitions:
  • V represents the element set.
  • d i j defines the distance between elements i and j within V.
  • E signifies the number of elements chosen.
  • x i is a binary variable that assumes a value of one when element i in V is selected, and it selects zero otherwise.
max min i , j V , i < j { d i j : x i · x j = 1 } s . t . i V x i = E x i { 0 , 1 } i V
While the ubiquity of models like Model (1) is recognized in characterizing diversity problems, their applicability falters in specific strategic logistics scenarios. In particular, situations where a baseline servicing capacity is mandatory present challenges that are not addressed by such models. To solve this, the unique restriction of the model must be modified to include the capacity of each node c i 0 and the minimum required capacity B, as depicted in Model (2):
max min i , j V , i < j { d i j : x i · x j = 1 } s . t . i V c i · x i B x i { 0 , 1 } i V
Accordingly, this paper addresses a dynamic version of the CDP. In this version, the objective is to select a set of elements within a network. What sets this variant apart is the introduction of a dynamic element, since the service capacity of each node is subject to influence by a random Bernoulli component. This means that some elements may unexpectedly have zero capacity, which depends on an underlying black box mechanism that takes into account the environmental variables. To solve this dynamic version, a learnheuristic algorithm [11] is proposed. The novelty of the approach presented in this paper lies in the fusion of a heuristic algorithm with reinforcement learning. This hybrid methodology offers a promising solution to address the complexities introduced by the dynamic CDP. As discussed in Mele et al. [12], combining heuristics with machine learning is an emerging research line with a vast potential with regard to optimization.
The rest of this paper is structured as follows. Section 2 provides a review of related work on diversity problems. Section 3 offers a formal description of the problem being discussed. Section 4 introduces the novel learnheuristic approach that will be used to solve the aforementioned problem. Section 5 illustrates the efficiency of the proposed algorithm using a series of computational experiments. Finally, Section 6 highlights the main contributions of this work.

2. Related Work on Diversity Problems

The diversity problem arose with the idea of maximizing the distance between individuals, where the definition of the distance between elements is customized to specific applications. Once the combinatorial optimization problem was defined by Kuo et al. [13], multiple objective functions with different approaches to the concept of diversity were explored by Sandoya et al. [14]. Additionally, the latter work demonstrates that the diversity problem is NP-hard. Therefore, for larger instances, it is advisable to consider heuristic algorithms.
In the beginning of this century, Rosenkrantz et al. [10] introduced the capacitated diversity problem, thus adding capacity to each node and constraints of minimal capacity. In their work, these authors also proposed the first heuristic approach to the problem with a heuristic algorithm called T1. Peiró et al. [15] proposed a new heuristic for the CDP. In their work, these authors proposed a GRASP algorithm [16], thus taking into account the distance between the solution and the capacity of the node to be selected. The algorithm is completed by combining GRASP with a variable neighborhood descent (VND) framework [17] and a strategic oscillation (SO) [18], which involves removing and adding nodes in solutions with capacities that are very close to the limit. The work of Martí et al. [19] outperformed the aforementioned SO algorithm. This work proposes three phases to achieve a solution. First, the diversification generation method (DGM) aims to initialize a solution by picking nodes one by one until the capacity restriction is satisfied. The difference from the GRASP phase of the SO algorithm is that the node chosen depends on the distance and an α parameter, which shortens the candidate list to pick the node with the largest capacity. Then, the solution is improved using a method that differs from VND in the SO algorithm by taking into account more than one node to replace others. Finally, it employs a multiple path relinking approach [20]. More recently, Lu et al. [21] employed a solution based on a tabu search algorithm [22] using hash functions to identify the tabu status of the candidate solutions created by a greedy algorithm.
On the other part, reinforcement learning [23] is a machine learning paradigm that focuses on training agents (entities that interact with the environment) to make sequential decisions in an environment to maximize a cumulative reward. It is inspired by behavioral psychology and is commonly used in fields such as artificial intelligence, robotics, game playing, and autonomous systems. This technique plays a relevant role in the current paradigm of optimization problems as described in Mazyavkina et al. [24]. During the last years, some studies have successfully forged a synergy between reinforcement learning and heuristic algorithms to construct solutions for dynamic problems [25]. This symbiotic relationship not only enhances the robustness of heuristic algorithms when confronted with dynamic scenarios, but also allows them to continually improve over time and adapt to changes in the environment. An instance of this is detailed in reference [9]. To the best of our knowledge, our work is the first study to incorporate dynamic conditions into the CDP, which requires the combination of a heuristic algorithm with reinforcement learning.

3. Problem Definition

The problem addressed in this article is an extension of the CDP, thereby considering the possibility of selected nodes not meeting their capacity. This probability of not utilizing the capacity is determined by dynamic contextual conditions.
In a more formal way, the CDP can be defined on a complete, weighted, and undirected graph G ( V , E ) , in which V is the set of facilities, and E is the set of edges connecting these facilities. If  i , j V , with  i j , each edge ( i , j ) E has a distance d i j 0 that satisfies the triangle inequality. All distances are symmetric, i.e.,  d i j = d j i . Each facility i V has a known deterministic capacity c i > 0 . An aggregated servicing capacity b > 0 is required as a threshold. Then, the CDP consists of finding a subset O V of facilities to open, with the aggregated capacity exceeding b and such that the minimum distance between any pair of facilities i , j S is maximized. In this context, x i is a binary variable that takes the value one if facility i V is included in O, and it takes the value zero if otherwise. Then, an integer programming model for this problem can be formulated as follows:
max min i , j V , i < j { d i j : x i · x j = 1 }
s . t . i V c i · p i · x i B
x i { 0 , 1 } i V
The objective function (3) maximizes the minimum distance between any pair of open facilities. Constraint (4) guarantees that the minimum required capacity is met where the probability of obtaining a capacity for a node i also depends on the weather w, the congestion of the node ( c i ), and the type of nodes already opened, which are determined by the dynamic context condition, i.e.,  p i B e r n ( ϕ ( w , c , o ) ) for an unknown function ϕ , for a black box emulating reality. Finally, Constraint (5) defines the values that decision variables can take. Figure 1 shows an illustrative example of the dynamic CDP. It can be seen that the selection of the top right node has resulted in failure and a capacity of zero, but it still remains part of the solution, thus affecting the objective function and forcing the search for other nodes.

Black Box Definition

In order to define the aforementioned black box function, which emulates the behavior of real life, let us consider the following function:
ϕ ( w , c , o ) = 1 1 e β 0 + β 1 · w + β 2 · c + β 3 · o
where:
  • w represents the meteorological conditions directly affecting the networks.
  • c denotes the congestion level between nodes.
  • o indicates the ratio of operational facilities for each category—each facility belongs to a particular type.
To model the effect of these variables on the network, the black box function uses a set of β coefficients: β 0 , β 1 , β 2 , and  β 3 ; each coefficient is associated with a particular variable as in Equation (6):
  • β 0 serves as the intercept term, which represents the baseline log-odds when all other influencing factors are zero.
  • β 1 modulates the influence of the meteorological conditions (w) on the outcome.
  • β 2 influences the outcome based on the congestion (c) between nodes.
  • β 3 influences the outcome based on the ratio of the operation facilities (o) that are currently open.
We employed a rule based on the name of the node to determine its type. Specifically, the type of node with name n, where n > 1 is the number of nodes of the instance, is determined by calculating the remainder when n is divided by five, as displayed in Equation (7), since there are five different types of nodes.
node _ type ( n ) = n mod 5
Figure 2 provides a comprehensive visualization of the probabilities associated with selecting a new node, which are generated by the black box model.
The visualization takes into account three pivotal variables: node type, weather conditions, and congestion levels. This display is designed to facilitate an in-depth understanding of how these variables interact to determine the probabilities produced by the model. The figure consists of four distinct heatmaps, each of which corresponds to a unique combination of weather (W) and congestion (C) conditions. The first row of heatmaps represents scenarios where the weather condition is W = 0 , interpreted as ‘Good Weather’, while the second row corresponds to W = 1 , which is interpreted as ‘Bad Weather’. Similarly, the first column presents cases of C = 0 , meaning ‘Low Congestion’, and the second column represents C = 1 , indicating ‘High Congestion’.
In each heatmap, the x axis represents the node types, which are denoted by integers ranging from 0 to 4. These node types could, for instance, represent different geographical locations or functionalities within a network. The y axis lists various configurations for the open_type variable, which are labeled from A to C. These configurations represent the state of multiple facilities, and they play a crucial role in determining the probabilities in question. In the heatmap, every cell shows the likelihood of choosing the next node. This takes into consideration both the node type of the next node and the types of nodes that are already opened (open_type). It also considers the specific weather of the instance and the congestion conditions of the next node, which are indicated by the subplot. The color scale enhances visual comprehension: dark black represents lower probabilities, while dark red signifies higher probabilities.
A general legend at the bottom of the figure serves as a key to the open_type configurations. It clarifies that A corresponds to an open_type configuration where we have opened 30 facilities of type 3 and 4, B corresponds to opening 12 facilities of each type, and C is where we have opened 30 facilities of type 0 and 1. This legend serves as a useful guide for interpreting the complex interactions captured in the heatmaps. This intricate visualization serves as a valuable tool for comprehending the multifaceted relationships between weather conditions, congestion levels, node types, and open_type configurations, which are  modeled by the black box.

4. Solution Approaches for the Dynamic Capacitated Dispersion Problem

The approach proposed to solve the dynamic CDP is a heuristic algorithm that uses reinforcement learning to deal with uncertainly. The way to approach the dynamism of the problem is through a function that approximates the black box, which is designated from from now on as the white box. In this example, a multivariate logistic regression was used for prediction. It should be noted that another model, such as a neural network, could also have been selected.

A Constructive Heuristic

Algorithm 1 shows the main procedure of the constructive heuristic, which extends the algorithm proposed by Martí et al. [19], thus considering the predictions provided by the white box to deal with the uncertainty of the problem. It starts from a hypothetical scenario where all facilities are initially closed, i.e., the algorithm starts with an empty initial solution (line 1). Next, a list of potential candidates for the first two nodes is generated. This list includes the edges that connect each pair of facilities and is sorted in descending order based on a combination of the distances between them and their respective capacities. This combination is weighted by the likelihood of successfully covering the capacity of a node (line 2). Equation (8) shows the edge value, where π ( n i ) is the probability of n i covering the capacity.
e d g e ( n i , n j ) = δ · d i , j + ( 1 δ ) 2 · π ( n i ) · c i + π ( n j ) · c j
Next, the algorithm selects the element at the top position of the described e d g e _ l i s t (line 3). Subsequently, the pair of facilities that compose the selected edge are included in the solution (line 4). Then, the simulation of the black box with the environment condition is performed, and the capacity of the solution is updated (line 5). Later, the objective function (of) is computed for the first time, thus considering the distance between these two facilities (line 6). To construct a feasible solution, the heuristic generates a preference list of candidate facilities ( n o d e _ l i s t ) to be included in the solution in order to decide the most suitable ones to add in each iteration (line 8). To select the best nodes, a candidate list to evaluate them is implemented, with an evaluation function employed to sort the preference. The evaluation value associated with each element of the list is defined in Equation (9). It considers both the calculated distance from the new facility ( d i s t a n c e _ t o _ s o l u t i o n ), i.e., the minimum distance between the node and each of the facilities currently selected, and the capacity of the candidate facility ( c i ), which is weighted by the predicted probability of successfully covering the node capacity, π ( n i ) . Both of these factors are adjusted to fit within a common scale by dividing them by the maximum distance between nodes and the maximum capacity in the set of candidates V S , where S represents the current solution. To properly adjust the importance of each of the terms, a  δ ( 0 , 1 ) is added. Later on, the meaning of this parameter will be explained in more depth.
e v a l ( n i ) = δ · d i s t a n c e _ t o _ s o l u t i o n ( n i ) m a x _ d i s t a n c e ( V S ) + ( 1 δ ) · π ( n i ) · c i m a x _ c a p a c i t y ( V S )
The list of facilities is sorted in descending order, and then a modified version of the semigreedy algorithm introduced by Martí et al. [19] is performed. This variation involves selecting a subset of the candidate list based on the parameter 0 α 1 . The subset is described in (10):
C L : = { i V M : e v a l ( n i ) α · ( e v a l ( n 0 ) e v a l ( n m ) ) } ,
where n 0 and n m are the nodes with best evaluation value and the worst in the list, respectively. Finally, unlike the deterministic algorithm, the node n i with the highest π ( n i ) · c i in the subset is chosen. The parameter α enables the algorithm to make choices among various candidates and avoid those facilities with a low probability of successfully providing the required capacity.
Once the facility has been selected, the simulation under the environment condition is performed to determinate the success of the node (line 9). Finally, the of and the n o d e _ l i s t are updated to consider the new facility in the solution. This procedure is repeated until the capacity constraint is met (line 7). Notice that δ ( 0 , 1 ) is a tuning parameter, which depends on the heterogeneity of the facilities in terms of the capacity. In particular, in scenarios with heterogeneous facilities in terms of the capacity, δ will be close to zero. On the contrary, δ will be close to one in scenarios with homogeneous facilities in terms of the capacity. To choose the better δ parameter, the deterministic form of the problem is taken. Then, the heuristic algorithm is performed for different δ candidates (such as values ranging from 0.1 to 0.9 ). The  δ parameter that is ultimately chosen is the one that yields the best result. The parameter α is another tuning parameter, and it can be selected in a similar way.
To study the performance of the algorithm, a number of maximum iterations m a x _ i t e r is fixed, and the heuristic Algorithm 1 is performed for different environment conditions. In Algorithm 2, the procedure initiates by creating an empty solution set S and an empty set of objective function values o f _ l i s t . It also initializes the white box, which is going to generate random predictions in the first iteration. Until the iteration limit is reach, a new set of context conditions is created, and the constructive heuristic algorithm (referenced in Algorithm 1) is repeatedly invoked to construct solution S, and its performance in terms of the objective function value is recorded. Simultaneously, the white box object is subject to modifications ruled by the f i t _ w h i t e b o x method, which is executed based on a probability function that is dependent on the iteration. As the iteration goes on, the chance of using the fit operation on the white box decreases. This behavior is encapsulated in Equation (11).
E x p o n e n t i a l   F u n c t i o n : = P ( t ) = e ln ( 0.01 ) · i t e r max _ iteration
This cycle continues until the algorithm reaches the maximum iteration, thus culminating in the output of the refined solution set S.
Algorithm 1 Constructive heuristic ( e d g e _ l i s t , e n v i r o n m e n t _ c o n d i t i o n s , w h i t e b o x )
1:
S
2:
e d g e _ l i s t s o r t ( e d g e _ l i s t )
3:
Select an element i from e d g e _ l i s t
4:
S S N o d e s ( e d g e _ l i s t ( i ) )
5:
c a p a c i t y b l a c k b o x ( N o d e s ( m o d i f i e d _ e d g e _ l i s t ( i ) ) )
6:
o f d i s t ( e d g e _ l i s t ( i ) )
7:
while  not _ feasible _ solution ( S )  do
8:
      Select an element i from n o d e _ l i s t
9:
       c a p a c i t y c a p a c i t y + b l a c k b o x ( n o d e _ l i s t ( i ) )
10:
     S S n o d e _ l i s t ( i )
11:
     o f U p d a t e _ f ( S )
12:
     n o d e _ l i s t U p d a t e ( n o d e _ l i s t )
13:
end while
14:
return  S
Algorithm 2 Algorithm performance evaluation
1:
S
2:
o f _ l i s t
3:
i t e r 0
4:
w h i t e b o x i n i t i a l i z e ( )
5:
while  i t e r < m a x _ i t e r  do
6:
       e n v i r o n m e n t _ c o n d i t i o n s r a n d o m _ e n v i r o n m e n t _ c o n d i t i o n s
7:
       S C o n s t r u c t i v e _ h e u r i s t i c ( e d g e _ l i s t , e n v i r o n m e n t _ c o n d i t i o n s , w h i t e b o x )
8:
       i t e r i t e r + 1
9:
       o f _ l i s t o f _ l i s t o f ( S )
10:
     w h i t e b o x f i t _ w h i t e b o x ( i t e r , w h i t e b o x )
11:
end while
12:
return  m e a n ( o f _ l i s t )

5. Numerical Experiments and Results

This section outlines the computational trials we conducted to evaluate the efficacy and performance of the previously discussed algorithms.

5.1. Computational Environment

All of the algorithms were developed using Python version 3.11 and operated on a Google cloud platform (GCP) computing instance equipped with an Intel(R) Xeon(R) CPU @ 2.20 GHz, 8 GB of RAM, and the Debian 11 operating system. We established 1000 instances to run with different parameters of congestion and weather defined by the seed of the problem. We juxtaposed the performance of our proposed algorithms with the static method. In the static method, the same algorithm was used without utilizing the learned information from the white box. In other words, it assumes that π ( n i ) = 1 , i V . This approach allows for a thorough comparison. Our evaluations took into account three sets of instances as proposed by Martí et al. [19], i.e.:
  • GKD instances: This is a dataset formed using Euclidean distances with node coordinates generated within a uniform distribution ranging from 0 to 10. Initially proposed in Martí et al. [26], it is divided into two subsets: the GKD-b, which includes instances with 50 and 150 nodes, and the GKD-c subset, which includes instances harboring 500 nodes.
  • MDG instances: This dataset encompasses real numbers arbitrarily chosen between 0 and 1000 that originate from a uniform distribution. It was introduced by Duarte and Martí [27] and comprises instances with a sizable 500 nodes.
For each instance, it is important to include all the required data to conduct the experiments, as was done by Martí et al. [19] and Gomez et al. [28] using the parameter b to determine the percentage of the total capacity obtained by summing the capacities of all facilities. In this case, the parameter was set at either 0.2 or 0.3, which represents 20 % or 30 % of the total capacity, respectively. One of the main goals of this section is to conduct a comparison between the learnheuristic algorithm introduced in this study and its static counterpart. To achieve this, we compared the results of 1000 iterations of the algorithm using identical scenarios in each algorithm. In our experiments, we utilized three distinct seeds to create the environment of the instances.

5.2. Black Box Parameters

In the context of this study, the black box works as a mechanism that predicts whether a particular node will be operational or not. It is essential to include the black box in our solution in order to obtain these predictions. Utilizing specific beta coefficients, this function provides a probability of a node having capacity. Subsequently, we generate a random number from a uniform distribution to determine whether the node indeed possesses the predicted capacity. The black box is only utilized to predict a node’s operational status once that node has been specifically incorporated into our solution.
We have assumed that the black box follows a logistic regression for its predictions. We have defined three types of dynamism, i.e., we have built three sets of parameters, with one for each type of dynamism. The parameters used for the black box in the Formula (6) for each type of dynamism are the following:
Table 1 details the probability of a node being operational and offering capacity when it is opened, which is based on different scenarios combining weather, congestion, and the configuration of nodes already included in the solution. Notice that congestion has a great influence on this probability for all nodes.

5.3. Analysis of Results

Table 2 shows the results obtained with the static and learnheuristic approaches for all benchmark instances. Moreover, with the objective of comparing the quality of our methods, we also report three different types of dynamism. Since we used three different seeds, Table 2 shows two columns of results for each heuristic strategy: a column with the average result of the objective function and a column with the average nodes included in the solution (Nodes). The last four columns in Table 2 show the percentage gaps between some of the tested solution approaches. Hence, if we denote as a learnheuristic the average objective function obtained with our learnheuristic approach, then the percentage gap between the learnheuristic and static solutions is computed using the formula Gap = l e a r n h e u r i s t i c s t a t i c S t a t i c . In this context, a negative performance gap implies that the static algorithm yields a better solution, while a positive performance gap suggests that the learnheuristic algorithm outperforms its static counterpart.
As can be seen in Table 2, the average of our objective function obtained by our learnheuristic algorithm usually outperformed the static algorithm, sometimes by a noticeable percentage gap. For example, while the global average gaps between these approaches were from 7.30 to 16.07 % in a low and in a high dynamic environment, respectively, we can see that this gap grew by up to 40.13 % for the MDG instances, which are the largest ones. Notably, we observed a negative performance gap in one particular instance characterized by low dynamism. In this isolated case, the static algorithm outperformed the dynamic counterpart with a gap of 1.40 % . This anomaly can be attributed to the specific dynamic properties and topological structure inherent to the instance. Given that this negative gap is an isolated occurrence, it does not warrant undue concern. Additionally, notice that the average gap between the static and the learnheuristic approaches increased, as the dynamism was greater due to the static counterpart not taking into account the probability of a node, thus having fewer nodes in the solution. Based on the values of the aforementioned tables and Table 3, Figure 3 shows the gap between the different dynamism values of the problem comparing the static and learnheuristic approaches in all cited instances, where the upward trend that the gap has can be seen for each dynamic environment.
Nevertheless, Figure 4 shows the average nodes opened by the algorithm to have a solution. The learnheuristic algorithm, represented in green, opened 4.72 % fewer nodes in scenarios with low dynamism, 9.47 % fewer nodes for a scenario with medium dynamism, and up to 13.50 % fewer nodes in scenarios with high dynamism, thereby reducing the number of nodes that needed to be opened. This fact may be of interest for other variations of the CDP problem. For instance, in a study by Lozano-Osorio et al. [29], a cost is assigned to each node, along with a constraint that the total cost of all opened nodes must not exceed a constant K. This constraint is similar to the capacity constraint with the static algorithm, shown in orange, serving as a comparative baseline.

6. Conclusions and Future Work

This work introduces a learnheuristic algorithm that hybridizes a constructive heuristic with reinforcement learning to tackle a dynamic version of the capacitated dispersion problem, thus offering a robust solution that surpasses existing methods for deterministic static versions. The presented algorithm adopts a constructive approach, where it integrates elements using reinforcement learning techniques, thus continuously learning and adapting to the dynamism inherent in the problem space. The first part of our study explores the advantages of this constructive mechanism. Beginning with an empty set, the algorithm iteratively adds promising components, which are influenced by heuristic knowledge and guided by reinforcement learning feedback, until an efficient solution structure is achieved. This phase demonstrates a remarkable capacity to navigate and adapt to the problem’s complexities, thus showcasing superiority over established methods, particularly in deterministic scenarios. The results show a generalized decrease in the number of selected nodes compared to the static version of the algorithm. It does, however, raise the computational time due to its complex learning and adaptation processes. This minor increase in resource is usually made up for by the algorithm’s skill in quickly identifying and skipping nodes that probably will not give the best solution, thereby ensuring a more efficient problem-solving process. In the CDP, the objective function is logically connected to the number of selected nodes; this reduction, which was undoubtedly due to the predictive intelligence used by the algorithm to avoid nodes with a low chance of success, led to an improvement in the objective function.
Looking forward, we identify several directions for further research. Firstly, future studies could explore problem variants where the element selection is associated with specific costs that are possibly linked to individual element capacities. This expansion would bring a new layer of complexity and realism to the problem, thereby providing possibilities for further optimization and refinement. Secondly, considering scenarios with stochastic inputs, including fluctuating demands or capacities, would be a natural next step: this would extend our algorithm to include simheuristic components [30,31]. Thus, combining our learnheuristic with simulation could provide robust solutions in fluctuating environments, thus clearing the path for optimization algorithms that are flexible and capable of adapting to dynamic and stochastic scenarios. Additionally, another possible extension refers to the combined use of parallel computing [32] and biased randomization techniques [33] in order to promote ‘agile’ optimization algorithms.

Author Contributions

Conceptualization, A.A.J. and J.P.; methodology, J.F.G. and A.R.U.; software, J.F.G. and A.R.U.; validation, A.A.J. and J.P.; formal analysis, J.F.G. and A.R.U.; investigation, J.F.G. and A.R.U.; writing—original draft preparation, J.F.G. and A.R.U.; writing—review and editing, A.A.J. and J.P.; supervision, A.A.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been partially funded by the Spanish Ministry of Science and Innovation (PID2022-138860NB-I00, RED2022-134703-T), by the SUN project of the Horizon Europe program (HORIZON-CL4-2022-HUMAN-01-14-101092612), and by the i4OPT project of the Generalitat Valenciana (PROMETEO/2021/065).

Data Availability Statement

All data employed is publicly available in the given references.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Govindan, K.; Fattahi, M.; Keyvanshokooh, E. Supply chain network design under uncertainty: A comprehensive review and future research directions. Eur. J. Oper. Res. 2017, 263, 108–141. [Google Scholar] [CrossRef]
  2. Eskandarpour, M.; Dejax, P.; Miemczyk, J.; Péton, O. Sustainable supply chain network design: An optimization-oriented review. Omega 2015, 54, 11–32. [Google Scholar] [CrossRef]
  3. Nataraj, S.; Ferone, D.; Quintero-Araujo, C.; Juan, A.; Festa, P. Consolidation centers in city logistics: A cooperative approach based on the location routing problem. Int. J. Ind. Eng. Comput. 2019, 10, 393–404. [Google Scholar] [CrossRef]
  4. Martí, R.; Martínez-Gavara, A.; Pérez-Peló, S.; Sánchez-Oro, J. A review on discrete diversity and dispersion maximization from an OR perspective. Eur. J. Oper. Res. 2022, 299, 795–813. [Google Scholar] [CrossRef]
  5. Correia, I.; Melo, T.; Saldanha-da Gama, F. Comparing classical performance measures for a multi-period, two-echelon supply chain network design problem with sizing decisions. Comput. Ind. Eng. 2013, 64, 366–380. [Google Scholar] [CrossRef]
  6. Tordecilla, R.D.; Copado-Méndez, P.J.; Panadero, J.; Quintero-Araujo, C.L.; Montoya-Torres, J.R.; Juan, A.A. Combining heuristics with simulation and fuzzy logic to solve a flexible-size location routing problem under uncertainty. Algorithms 2021, 14, 45. [Google Scholar] [CrossRef]
  7. Osaba, E.; Villar-Rodriguez, E.; Del Ser, J.; Nebro, A.J.; Molina, D.; LaTorre, A.; Suganthan, P.N.; Coello, C.A.C.; Herrera, F. A tutorial on the design, experimentation and application of metaheuristic algorithms to real-world optimization problems. Swarm Evol. Comput. 2021, 64, 100888. [Google Scholar] [CrossRef]
  8. Szepesvári, C. Algorithms for Reinforcement Learning; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
  9. Juan, A.A.; Marugan, C.A.; Ahsini, Y.; Fornes, R.; Panadero, J.; Martin, X.A. Using Reinforcement Learning to Solve a Dynamic Orienteering Problem with Random Rewards Affected by the Battery Status. Batteries 2023, 9, 416. [Google Scholar] [CrossRef]
  10. Rosenkrantz, D.J.; Tayi, G.K.; Ravi, S.S. Facility Dispersion Problems under Capacity and Cost Constraints. J. Comb. Optim. 2000, 4, 7–33. [Google Scholar] [CrossRef]
  11. Bayliss, C. Machine learning based simulation optimisation for urban routing problems. Appl. Soft Comput. 2021, 105, 107269. [Google Scholar] [CrossRef]
  12. Mele, U.J.; Gambardella, L.M.; Montemanni, R. A new constructive heuristic driven by machine learning for the traveling salesman problem. Algorithms 2021, 14, 267. [Google Scholar] [CrossRef]
  13. Kuo, C.C.; Glover, F.; Dhir, K.S. Analyzing and modeling the maximum diversity problem by zero-one programming. Decis. Sci. 1993, 24, 1171–1185. [Google Scholar] [CrossRef]
  14. Sandoya, F.; Martínez-Gavara, A.; Aceves, R.; Duarte, A.; Martí, R. Diversity and equity models. In Handbook of Heuristics; Springer: Berlin/Heidelberg, Germany, 2018; Volume 2-2, pp. 979–998. [Google Scholar]
  15. Peiró, J.; Jiménez, I.; Laguardia, J.; Martí, R. Heuristics for the capacitated dispersion problem. Int. Trans. Oper. Res. 2021, 28, 119–141. [Google Scholar] [CrossRef]
  16. Resende, M.G.; Ribeiro, C.C. Optimization by GRASP; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  17. Duarte, A.; Mladenovic, N.; Sánchez-Oro, J.; Todosijević, R. Variable neighborhood descent. In Handbook of Heuristics; Springer: Cham, Switzerland, 2018. [Google Scholar]
  18. Glover, F.; Hao, J.K. The case for strategic oscillation. Ann. Oper. Res. 2011, 183, 163–173. [Google Scholar] [CrossRef]
  19. Martí, R.; Martínez-Gavara, A.; Sánchez-Oro, J. The capacitated dispersion problem: An optimization model and a memetic algorithm. Memetic Comput. 2021, 13, 131–146. [Google Scholar] [CrossRef]
  20. Laguna, M.; Martí, R.C. Scatter Search: Methodology and Implementations in C; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2003. [Google Scholar]
  21. Lu, Z.; Martínez-Gavara, A.; Hao, J.K.; Lai, X. Solution-based tabu search for the capacitated dispersion problem. Expert Syst. Appl. 2023, 223, 119856. [Google Scholar] [CrossRef]
  22. Gendreau, M. An introduction to tabu search. In Handbook of Metaheuristics; Springer: Berlin/Heidelberg, Germany, 2003; pp. 37–54. [Google Scholar]
  23. Lu, X.; Van Roy, B.; Dwaracherla, V.; Ibrahimi, M.; Osband, I.; Wen, Z. Reinforcement learning, bit by bit. Found. Trends Mach. Learn. 2023, 16, 733–865. [Google Scholar] [CrossRef]
  24. Mazyavkina, N.; Sviridov, S.; Ivanov, S.; Burnaev, E. Reinforcement learning for combinatorial optimization: A survey. Comput. Oper. Res. 2021, 134, 105400. [Google Scholar] [CrossRef]
  25. Calvet, L.; de Armas, J.; Masip, D.; Juan, A.A. Learnheuristics: Hybridizing metaheuristics with machine learning for optimization with dynamic inputs. Open Math. 2017, 15, 261–280. [Google Scholar] [CrossRef]
  26. Martí, R.; Gallego, M.; Duarte, A. A branch and bound algorithm for the maximum diversity problem. Eur. J. Oper. Res. 2010, 200, 36–44. [Google Scholar] [CrossRef]
  27. Duarte, A.; Martí, R. Tabu search and GRASP for the maximum diversity problem. Eur. J. Oper. Res. 2007, 178, 71–84. [Google Scholar] [CrossRef]
  28. Gomez, J.F.; Panadero, J.; Tordecilla, R.D.; Castaneda, J.; Juan, A.A. A multi-start biased-randomized algorithm for the capacitated dispersion problem. Mathematics 2022, 10, 2405. [Google Scholar] [CrossRef]
  29. Lozano-Osorio, I.; Martínez-Gavara, A.; Martí, R.; Duarte, A. Max–min dispersion with capacity and cost for a practical location problem. Expert Syst. Appl. 2022, 200, 116899. [Google Scholar] [CrossRef]
  30. Hatami, S.; Calvet, L.; Fernández-Viagas, V.; Framinan, J.M.; Juan, A.A. A simheuristic algorithm to set up starting times in the stochastic parallel flowshop problem. Simul. Model. Pract. Theory 2018, 86, 55–71. [Google Scholar] [CrossRef]
  31. Rabe, M.; Gonzalez-Feliu, J.; Chicaiza-Vaca, J.; Tordecilla, R.D. Simulation-optimization approach for multi-period facility location problems with forecasted and random demands in a last-mile logistics application. Algorithms 2021, 14, 41. [Google Scholar] [CrossRef]
  32. Essaid, M.; Idoumghar, L.; Lepagnot, J.; Brévilliers, M. GPU parallelization strategies for metaheuristics: A survey. Int. J. Parallel Emergent Distrib. Syst. 2019, 34, 497–522. [Google Scholar] [CrossRef]
  33. Dominguez, O.; Juan, A.A.; Faulin, J. A biased-randomized algorithm for the two-dimensional vehicle routing problem with and without item rotations. Int. Trans. Oper. Res. 2014, 21, 375–398. [Google Scholar] [CrossRef]
Figure 1. Simple example of a CDP solution under dynamic conditions.
Figure 1. Simple example of a CDP solution under dynamic conditions.
Algorithms 16 00532 g001
Figure 2. Different behaviors depending on the beta values.
Figure 2. Different behaviors depending on the beta values.
Algorithms 16 00532 g002
Figure 3. Gaps between static and dynamic approaches for different environments.
Figure 3. Gaps between static and dynamic approaches for different environments.
Algorithms 16 00532 g003
Figure 4. Average of nodes in the solution by type of instance.
Figure 4. Average of nodes in the solution by type of instance.
Algorithms 16 00532 g004
Table 1. Black box’s parameters for the Formula (6).
Table 1. Black box’s parameters for the Formula (6).
Node TypeLowMediumHigh
β 0 β 1 β 2 β 3 β 0 β 1 β 2 β 3 β 0 β 1 β 2 β 3
N10.7−0.3−2.5−0.80.5−0.6−3−1.85−0.4−1−3.2−2
N20.8−0.2−2−0.550.6−0.4−2.5−0.8−0.2−0.6−2.7−1.2
N30.9−0.15−0.501−0.25−0.7500.75−0.5−1.75−1
N41−0.1−0.250.41.2−0.15−0.450.61−0.25−0.450.3
N51.1−0.05−0.10.61.3−0.1−0.20.81.1−0.1−0.20.5
Table 2. Comparative results between learnheuristic and static algorithms for instances GKD_b.
Table 2. Comparative results between learnheuristic and static algorithms for instances GKD_b.
InstanceLearnheuristicStaticGap (%)
Low (1)Medium (2)High (3)Low (4)Medium (5)High (6)
TimeNodesOFTimeNodesOFTimeNodesOFTimeNodesOFTimeNodesOFTimeNodesOF(1)–(4)(2)–(5)(3)–(6)
GKD-b_11_n50_b02_m538.413.6117.934.114.4116.738.719.9110.41.913.0115.51.714.2113.62.019.7105.72.10%2.75%4.39%
GKD-b_11_n50_b03_m531.720.6107.734.821.6106.746.531.494.41.521.7107.51.623.5105.12.033.791.00.21%1.51%3.72%
GKD-b_12_n50_b02_m522.011.5146.322.211.6145.428.714.6140.41.013.4138.01.214.6135.31.420.2128.36.05%7.45%9.42%
GKD-b_12_n50_b03_m535.322.2134.637.223.1132.842.932.0122.91.421.6130.51.523.1128.41.932.7119.43.13%3.40%2.95%
GKD-b_13_n50_b02_m526.614.069.027.214.468.534.419.362.61.215.263.21.216.161.91.523.653.39.09%10.73%17.44%
GKD-b_13_n50_b03_m532.719.158.030.919.257.943.027.949.01.522.054.01.623.452.62.033.443.07.36%10.13%13.99%
GKD-b_14_n50_b02_m527.314.260.528.115.259.034.321.650.71.215.656.71.317.653.51.625.542.76.61%10.30%18.63%
GKD-b_14_n50_b03_m538.124.247.041.325.445.649.334.934.91.625.445.41.727.343.42.037.431.73.39%4.94%10.19%
GKD-b_15_n50_b02_m521.78.9115.421.08.9113.024.49.8104.30.911.6114.01.012.3110.11.115.399.31.16%2.61%4.99%
GKD-b_15_n50_b03_m539.321.2113.042.423.3109.855.635.293.71.723.6109.51.927.4104.92.239.588.63.15%4.68%5.75%
GKD-b_16_n50_b02_m1526.913.351.327.613.949.934.521.140.21.113.848.91.114.447.51.521.738.44.89%5.05%4.53%
GKD-b_16_n50_b03_m1530.314.828.929.614.629.535.319.226.61.317.926.21.419.025.21.726.820.910.31%17.30%27.34%
GKD-b_17_n50_b02_m1524.812.323.225.312.323.130.515.819.31.013.920.31.115.518.51.522.813.514.02%24.86%43.48%
GKD-b_17_n50_b03_m1535.521.613.536.422.912.742.731.38.81.522.513.01.724.411.91.933.28.14.27%6.36%8.96%
GKD-b_18_n50_b02_m1525.611.591.224.711.092.226.712.589.11.114.881.01.215.879.61.421.971.812.64%15.78%23.98%
GKD-b_18_n50_b03_m1531.317.474.333.217.473.934.422.966.81.318.971.21.419.969.41.727.160.54.34%6.55%10.38%
GKD-b_19_n50_b02_m1524.113.589.926.814.388.235.221.779.41.013.188.41.114.386.41.422.076.21.71%2.09%4.24%
GKD-b_19_n50_b03_m1536.322.677.937.323.676.644.933.365.51.624.076.01.626.473.22.036.960.32.52%4.62%8.64%
GKD-b_20_n50_b02_m1525.013.690.724.513.690.129.117.084.71.013.987.81.115.086.01.421.077.83.29%4.78%8.91%
GKD-b_20_n50_b03_m1530.519.779.733.120.978.344.530.664.41.320.277.91.521.975.61.932.760.32.30%3.58%6.82%
GKD-b_41_n150_b02_m15162.235.7141.2165.637.5139.9197.555.5132.77.839.4139.98.543.9137.711.767.6127.50.87%1.64%3.41%
GKD-b_41_n150_b03_m15198.656.9132.4205.860.1130.7254.395.7117.610.658.0130.011.464.0127.015.2105.6111.51.82%2.88%5.51%
GKD-b_42_n150_b02_m15166.938.462.5165.339.461.6194.953.955.58.241.961.18.845.859.411.162.153.72.40%3.71%3.34%
GKD-b_42_n150_b03_m15196.957.551.5195.458.750.7234.684.045.111.060.451.511.463.850.414.292.942.70.02%0.60%5.57%
GKD-b_43_n150_b02_m15165.639.046.0166.738.645.9182.450.541.58.845.742.89.549.241.611.665.437.67.33%10.44%10.30%
GKD-b_43_n150_b03_m15197.158.139.3197.461.738.2243.994.529.910.960.538.211.667.136.415.2105.326.42.86%4.75%13.00%
GKD-b_44_n150_b02_m15166.838.280.7166.439.380.0193.853.173.87.738.478.98.241.677.210.759.669.42.19%3.66%6.38%
GKD-b_44_n150_b03_m15191.456.569.8201.358.869.2249.191.360.611.262.668.212.371.365.515.4108.254.52.33%5.65%11.18%
GKD-b_45_n150_b02_m15159.637.188.6163.737.987.9184.349.782.07.537.586.17.839.884.79.953.678.02.89%3.67%5.13%
GKD-b_45_n150_b03_m15207.260.677.4209.464.175.9254.494.066.411.363.173.812.069.371.515.0102.162.34.92%6.22%6.57%
GKD-b_46_n150_b02_m45162.937.1101.3161.236.9101.2177.947.595.97.437.199.37.839.698.110.055.390.51.95%3.07%5.98%
GKD-b_46_n150_b03_m45199.459.090.5204.662.388.9248.393.478.310.758.689.411.464.087.414.8101.274.01.22%1.72%5.81%
GKD-b_47_n150_b02_m45163.737.4142.0154.538.3141.2174.949.1135.97.738.6139.78.141.6137.910.255.4130.71.68%2.42%4.01%
GKD-b_47_n150_b03_m45178.951.9133.0180.053.7132.3223.181.9122.210.053.8130.410.758.2129.114.091.5117.11.99%2.43%4.32%
GKD-b_48_n150_b02_m45151.837.079.3157.938.678.2182.655.671.37.638.078.58.342.276.411.865.867.70.93%2.33%5.19%
GKD-b_48_n150_b03_m45194.061.069.1194.166.267.7238.9102.757.111.060.768.011.867.865.915.4110.753.41.61%2.71%6.78%
GKD-b_49_n150_b02_m45140.030.3148.7137.030.2148.0146.336.1143.46.531.1146.16.632.5145.28.443.5139.51.78%1.93%2.78%
GKD-b_49_n150_b03_m45170.649.9134.2170.151.2133.9210.579.7123.29.952.9128.510.758.1126.214.091.2115.74.44%6.15%6.48%
GKD-b_50_n150_b02_m45140.335.388.8148.436.587.8172.850.082.17.637.786.48.040.985.011.060.878.02.74%3.27%5.26%
GKD-b_50_n150_b03_m45187.850.677.1190.951.076.5212.269.972.19.650.772.79.953.471.812.675.966.06.04%6.63%9.15%
Table 3. Comparative results between learnheuristic and static algorithms for instances GKD_c and MDG.
Table 3. Comparative results between learnheuristic and static algorithms for instances GKD_c and MDG.
InstanceLearnheuristicStaticGap (%)
Low (1)Medium (2)High (3)Low (4)Medium (5)High (6)
TimeNodesOFTimeNodesOFTimeNodesOFTimeNodesOFTimeNodesOFTimeNodesOF(1)–(4)(2)–(5)(3)–(6)
GKD-c_01_n500_b02_m501579.2119.97.61584.2122.97.51776.5160.87.076.8124.17.481.4133.07.3104.3183.56.82.11%3.13%3.25%
GKD-c_01_n500_b03_m501895.2170.06.71844.2173.76.72120.7258.56.0105.2179.16.6109.7191.96.5144.3296.35.72.12%2.95%5.44%
GKD-c_02_n500_b02_m501630.6114.27.71665.4116.47.61780.5150.67.173.2115.77.575.5121.57.498.2168.96.92.48%3.04%4.04%
GKD-c_02_n500_b03_m501956.9195.56.71964.5200.26.62183.0284.95.9114.9206.56.4123.4219.36.3151.9329.75.53.93%4.33%6.60%
GKD-c_03_n500_b02_m501621.8115.87.51596.2116.87.51745.8151.57.074.6120.67.378.7127.47.398.8171.16.83.04%3.39%2.68%
GKD-c_03_n500_b03_m501972.0199.96.62065.7210.36.42334.2329.05.5115.8210.36.4126.0230.86.2160.5364.45.21.56%3.33%6.71%
GKD-c_04_n500_b02_m501582.6100.67.31608.4101.67.21698.1132.06.768.5106.86.972.3115.56.9102.2171.46.15.35%5.10%9.34%
GKD-c_04_n500_b03_m501933.4193.16.41991.6202.56.32248.1288.15.6116.1207.96.2124.2227.76.0162.6341.75.02.89%5.45%11.33%
GKD-c_05_n500_b02_m501508.3110.37.41605.7111.67.31774.8146.26.973.3117.17.479.6128.67.2106.2183.76.50.76%2.14%6.31%
GKD-c_05_n500_b03_m501915.9200.16.61907.1208.66.52156.4307.85.7113.7201.86.4118.2217.56.3153.7343.75.32.88%3.21%8.22%
GKD-c_06_n500_b02_m501511.0117.07.51518.4120.27.51680.1161.76.977.4125.87.381.6134.97.2107.1192.16.63.36%4.53%5.05%
GKD-c_06_n500_b03_m501827.0184.66.51925.6190.66.42124.9275.35.7114.2201.16.3122.8222.16.1157.5342.15.23.38%4.59%10.45%
GKD-c_07_n500_b02_m501551.2120.17.41523.6122.17.41720.3156.47.080.8130.57.186.9140.57.1110.1191.46.74.34%4.19%4.78%
GKD-c_07_n500_b03_m501968.1200.56.61905.1210.36.52171.7317.95.7113.3204.16.5121.2222.06.4156.3347.75.30.44%1.05%5.75%
GKD-c_08_n500_b02_m501486.1116.17.61510.8119.47.61619.2163.66.972.7118.27.577.2127.37.4103.1182.96.81.65%2.03%1.61%
GKD-c_08_n500_b03_m501771.6197.26.61867.0204.26.52116.3296.45.8111.9204.16.5119.0220.66.4149.0324.15.61.47%1.48%3.94%
GKD-c_09_n500_b02_m501651.9121.97.41634.3125.57.41793.8167.16.878.2128.27.384.0138.77.2110.2197.46.51.23%2.45%5.39%
GKD-c_09_n500_b03_m501803.9178.66.41786.6185.46.32060.3272.65.7111.1196.76.3119.4218.76.1153.1333.25.41.08%3.65%6.37%
GKD-c_10_n500_b02_m501465.0113.47.71483.2115.97.71621.5152.87.172.4117.47.676.3124.77.499.5173.56.81.77%2.98%3.82%
GKD-c_10_n500_b03_m501730.9191.96.71738.6197.16.61969.5285.15.9108.5195.66.5114.9212.16.4147.4318.35.52.26%3.45%7.34%
MDG-b_01_n500_b02_m501386.4105.814.61390.5107.714.21565.6147.29.269.9110.412.273.8118.210.799.1170.55.920.24%33.02%56.78%
MDG-b_01_n500_b03_m501798.3214.44.41846.4223.24.02184.6322.91.9119.4221.53.8126.8241.13.3160.8368.51.114.63%21.87%70.62%
MDG-b_02_n500_b02_m501563.4142.613.21555.6148.812.21698.6199.47.491.6151.412.693.6160.811.4116.0216.26.95.21%7.45%7.28%
MDG-b_02_n500_b03_m501772.0211.75.41797.4221.74.82070.1325.22.0115.1212.14.8122.1230.74.0153.7344.91.412.08%20.98%44.87%
MDG-b_03_n500_b02_m501475.1134.212.01499.1140.211.11678.7190.26.688.6150.09.493.8159.98.6116.7217.65.328.66%29.06%25.48%
MDG-b_03_n500_b03_m501841.9239.54.71878.1255.64.12095.0362.81.5126.1241.54.5133.5264.43.7161.1383.31.24.66%9.04%21.97%
MDG-b_04_n500_b02_m501462.3122.412.91470.3127.611.91648.5173.87.081.9134.47.986.8144.57.3113.6206.95.563.56%62.27%26.19%
MDG-b_04_n500_b03_m501811.2210.54.91812.0225.24.32094.0331.81.6122.6227.84.3130.2252.63.4164.4365.51.012.89%25.96%51.06%
MDG-b_05_n500_b02_m501419.5118.513.61427.3123.112.71600.5167.97.677.3125.511.582.0134.510.4108.1192.65.718.35%22.27%33.91%
MDG-b_05_n500_b03_m501735.6197.05.01767.8207.54.42015.4306.81.8111.2198.24.0118.3217.13.6153.7337.51.424.19%20.74%32.24%
MDG-b_06_n500_b02_m501444.1116.513.71461.7120.013.11609.5163.88.275.6121.813.080.2132.211.5108.2194.76.05.57%13.57%36.91%
MDG-b_06_n500_b03_m501796.1222.54.91844.9237.94.22175.7347.61.5126.6239.63.8137.5271.83.1166.1401.90.928.53%37.25%72.18%
MDG-b_07_n500_b02_m501580.0122.013.71653.7126.412.81767.7169.87.882.4134.011.287.7144.69.3113.3203.45.622.55%38.32%39.54%
MDG-b_07_n500_b03_m501874.5210.05.51928.2221.55.02249.5336.12.0116.5213.25.0123.6230.44.3156.3347.61.711.24%16.10%17.93%
MDG-b_08_n500_b02_m501639.8115.213.91658.6117.413.01834.8162.48.073.8117.98.880.1130.08.4108.6193.04.857.79%54.24%68.93%
MDG-b_08_n500_b03_m501975.8220.35.72119.5232.15.12359.8338.22.1120.0219.05.3126.3237.04.5155.7344.31.77.72%13.94%23.52%
MDG-b_09_n500_b02_m501570.9104.915.71577.4108.215.11739.8150.89.372.1112.113.376.5121.011.8105.1181.96.318.81%27.92%47.93%
MDG-b_09_n500_b03_m501918.4198.15.51929.3207.25.02182.6305.42.1115.8203.05.6123.7223.44.7158.4341.01.4–1.41%5.68%47.44%
MDG-b_10_n500_b02_m501505.8117.014.81517.1119.713.91653.5159.48.778.4125.513.984.0133.911.7106.9185.47.15.93%18.97%23.02%
MDG-b_10_n500_b03_m501842.2203.75.11890.6216.84.52135.5327.61.7117.7212.74.3124.4235.83.6157.9359.41.120.19%27.95%54.85%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gomez, J.F.; Uguina, A.R.; Panadero, J.; Juan, A.A. A Learnheuristic Algorithm for the Capacitated Dispersion Problem under Dynamic Conditions. Algorithms 2023, 16, 532. https://doi.org/10.3390/a16120532

AMA Style

Gomez JF, Uguina AR, Panadero J, Juan AA. A Learnheuristic Algorithm for the Capacitated Dispersion Problem under Dynamic Conditions. Algorithms. 2023; 16(12):532. https://doi.org/10.3390/a16120532

Chicago/Turabian Style

Gomez, Juan F., Antonio R. Uguina, Javier Panadero, and Angel A. Juan. 2023. "A Learnheuristic Algorithm for the Capacitated Dispersion Problem under Dynamic Conditions" Algorithms 16, no. 12: 532. https://doi.org/10.3390/a16120532

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop