Next Article in Journal
Carrier Phase Residual Modeling and Fault Monitoring Using Short-Baseline Double Difference and Machine Learning
Next Article in Special Issue
A Novel Hybrid Algorithm Based on Jellyfish Search and Particle Swarm Optimization
Previous Article in Journal
A New Estimator: Median of the Distribution of the Mean in Robustness
Previous Article in Special Issue
Bio-Inspired Multi-UAV Path Planning Heuristics: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploring Initialization Strategies for Metaheuristic Optimization: Case Study of the Set-Union Knapsack Problem

1
Escuela de Ingeniería de Construcción y Transporte, Pontificia Universidad Católica de Valparaíso, Avenida Brasil 2241, Valparaíso 2362804, Chile
2
Facultad de Ingeniería, Centro de Transformación Digital, Universidad del Desarrollo, Santiago 7610658, Chile
3
Escuela de Ingeniería Informática, Pontificia Universidad Católica de Valparaíso, Avenida Brasil 2241, Valparaíso 2362807, Chile
*
Authors to whom correspondence should be addressed.
Mathematics 2023, 11(12), 2695; https://doi.org/10.3390/math11122695
Submission received: 25 May 2023 / Revised: 7 June 2023 / Accepted: 9 June 2023 / Published: 14 June 2023
(This article belongs to the Special Issue Metaheuristic Algorithms)

Abstract

:
In recent years, metaheuristic methods have shown remarkable efficacy in resolving complex combinatorial challenges across a broad spectrum of fields. Nevertheless, the escalating complexity of these problems necessitates the continuous development of innovative techniques to enhance the performance and reliability of these methods. This paper aims to contribute to this endeavor by examining the impact of solution initialization methods on the performance of a hybrid algorithm applied to the set union knapsack problem (SUKP). Three distinct solution initialization methods, random, greedy, and weighted, have been proposed and evaluated. These have been integrated within a sine cosine algorithm employing k-means as a binarization procedure. Through testing on medium- and large-sized SUKP instances, the study reveals that the solution initialization strategy influences the algorithm’s performance, with the weighted method consistently outperforming the other two. Additionally, the obtained results were benchmarked against various metaheuristics that have previously solved SUKP, showing favorable performance in this comparison.

1. Introduction

Over the past few years, metaheuristic methods have emerged as powerful tools in addressing complex problems, particularly those pertaining to the realm of combinatorial challenges. A wide array of applications across various fields, including biology [1,2], logistics [3], civil engineering [4,5], machine learning [6], and many more, serve as compelling evidence of their effectiveness in problem-solving.
These methods have garnered considerable attention owing to their capacity to effectively navigate extensive search spaces and identify near-optimal solutions within relatively brief timeframes. This characteristic has demonstrated its particular value in addressing the immense scale and complexity inherent in numerous combinatorial problems. The ability to find satisfactory solutions for complex problems in a time-efficient manner has solidified the importance of these methods in the field of optimization.
However, as the intricacy of such problems continues to grow, the challenges associated with efficiently solving them also intensify. Consequently, there is a pressing need for innovative techniques and strategies that can enhance the performance of these methods, ensuring that they remain relevant and effective when confronted with increasingly complex optimization tasks. The development of advanced algorithms and the integration of novel approaches in the initialization process can contribute to overcoming these obstacles, ultimately bolstering the performance and reliability of these methods in tackling large-scale, intricate combinatorial problems.
In response to these challenges, researchers have focused on the development and application of diverse approaches aimed at enhancing and fortifying metaheuristic algorithms. One notable example involves hybrid techniques, which amalgamate the strengths of various metaheuristic algorithms. By integrating complementary methods and harnessing their synergies, these hybrid techniques have demonstrated success in augmenting the overall performance, accuracy, and efficiency of the underlying metaheuristic algorithms. Another intriguing aspect to explore is the examination of the forms and parameters of solution initialization.
The significance of initialization operators lies in the fact that the diversity and nature of the initial population have an impact on the algorithm’s performance, potentially leading to enhanced solutions and improved convergence rates. Furthermore, the sensitivity of the algorithms is problem-dependent, signifying that the choice of an initialization method can have a substantial influence on the performance for certain problems. In addition, the population size and the number of iterations also play a role in the algorithms’ performance, necessitating an appropriate population size and a sufficient number of iterations to achieve optimal solutions. Lastly, the effectiveness of initialization methods varies depending on the specific metaheuristic optimizer employed, making it crucial to comprehend the relationship between initialization methods and optimizers in order to select the most suitable combination for specific problems. In summary, good population diversity and an adequate number of iterations, combined with an appropriate initialization method, are likely to lead to optimal solutions.
In the literature exploration, it has been observed that studies such as [7] compare the effects of population size, maximum iterations, and eleven initialization methods on the convergence and accuracy of metaheuristic optimizers. Results have indicated that sensitivity to initialization schemes varies among algorithms and is problem-dependent. Furthermore, performance has been found to rely on population size and the number of iterations, with greater diversity and a suitable quantity of iterations being more likely to produce optimal solutions. In a study by [8], a systematic comparison of 22 initialization methods was conducted, analyzing their impact on the convergence and accuracy of five optimizers: DE, PSO, CS, ABC, and GA. The findings revealed that 43.37% of DE functions and 73.68% of PSO and CS functions were significantly affected by initialization methods. Population size, number of iterations, and certain probability distributions also influenced performance.
In the study by [9], a reliability-analysis-based structural shape optimization formulation was proposed, incorporating Latin hypercube sampling (LHS) as the initialization scheme. The investigation focused on the relationship between geometry and fatigue life in structural component design. The extended finite element method (XFEM) and level set description were utilized, and nature-inspired optimization techniques were employed to solve the problems. Results indicated that proper shape changes can enhance the service life of structural components subjected to fatigue loads, with the location and orientation of initial imperfections significantly affecting optimal shapes. Finally, in [10], the authors provided an extensive review of diverse initialization strategies designed to improve the performance of metaheuristic optimization algorithms. Emphasizing the crucial role of initialization, various distribution schemes, population sizes, and iteration numbers have been investigated by researchers in pursuit of optimal solutions. Notable schemes encompass random numbers, quasirandom sequences, chaos theory, probability distributions, hybrid algorithms, Lévy flights, and more. Additionally, the paper evaluated the influence of population size, maximum iterations, and ten distinct initialization methods on three prominent population-based metaheuristic optimizers: bat algorithm (BA), grey wolf optimizer (GWO), and butterfly optimization algorithm (BOA).
Aligning with the process of solution initialization, this paper proposes various strategies for initializing solutions, incorporating these strategies into a discrete hybrid algorithm detailed in [11]. This algorithm merges the concept of k-means with metaheuristics and is applied to the set union knapsack problem (SUKP). The SUKP [12] is an extended version of the traditional knapsack problem and has attracted considerable research interest in recent years [13,14,15]. This attention is primarily due to its intriguing applications [16,17], coupled with the complexity and challenge involved in solving it efficiently.
In the context of the SUKP, an assortment of items is identified, each with a specific profit value attributed. Additionally, a correspondence is established between each item and a group of elements, with each carrying a weight that impacts the knapsack constraint. A scrutiny of the existing body of literature reveals that the SUKP is predominantly addressed using advanced metaheuristics, with outcomes provided within acceptable time limits. However, when conventional metaheuristics are utilized in the SUKP, issues including instability and diminished effectiveness are exposed as the instance size increases. For instance, a variety of transfer functions were deployed and evaluated within small to medium SUKP instances, as documented in [18]. A reduction in effectiveness was noted when these algorithms were applied to standard SUKP instances. Adding to the complexity, a new series of benchmark problems have been recently introduced, as noted in [19].
Given these considerations, it becomes imperative to explore solution initialization techniques to assess the algorithm’s performance. The following are the significant contributions of this study:
  • Three solution initialization strategies are proposed.
  • The initialization solutions are integrated with the SIN-COS metaheuristic and the k-means technique, following the strategy proposed in [20].
  • The three initialization strategies are evaluated using medium-sized and large-sized SUKP problems, with the latter being proposed in [19].
The structure of this paper is organized as follows: Section 2 delivers a comprehensive examination of the set-union knapsack problem and its related applications. In Section 3, the k-means sine cosine search algorithm and the initialization operator are thoroughly described. Section 4 expounds on the numerical experiments undertaken and the resulting comparisons. Finally, Section 6 presents concluding insights and explores potential directions for future research.

2. Advancements in Solving the Set-Union Knapsack Problem

The SUKP represents an extended model of a knapsack, which is defined as follows. A set of n elements, denoted as U, is assumed to exist, with each element j U possessing a positive weight w j . Assumed also is a set of m items, named V, where each item i V is a subset of elements U i U and holds a profit p i . Given the presence of a knapsack with a capacity C, the aim of SUKP is to identify a subset of items S V that allows the maximization of the total profit of S, while ensuring that the combined weight of the components belonging to S does not exceed the capacity C of the knapsack. The decision variables of the problem are identified as the elements belonging to the set S. It is noteworthy that an element’s weight is considered only once, even if it corresponds to multiple chosen items in S. The mathematical depiction of SUKP is presented subsequently:
Maximize P ( S ) = i S p i .
subject to:
W ( S ) = j i S U i w j C , S V .
In the research work, interesting applications of SUKP are identified, such as the one introduced in [16]. The aim of this application is the enhancement of robustness and scalability within cybernetic systems. The consideration is given to a centralized cyber system with a fixed memory capacity, which hosts an assortment of profit-generating services (or requests), each inclusive of a set of data objects. The activation of a data object consumes a particular amount of memory; however, recurring utilization of the same data object does not incur additional memory consumption (a pivotal condition of SUKP). The objective involves the selection of a service subset from the pool of available candidates, with the intention to maximize the total profit of these services, whilst keeping the total memory required by the linked data objects within the cyber system’s memory capacity. The SUKP model, in which an item symbolizes a service with its profit and an element represents a data object with its memory usage (element weight), fittingly structures this application. Consequently, the determination of the optimal solution to the resulting SUKP problem parallels the resolution of the data allocation problem.
An additional application worth mentioning relates to the real-time rendering of animated crowds, as noted in [21]. In this study, a method is introduced by the authors to hasten the visualization process for large gatherings of animated characters. A caching system is implemented that permits the reuse of skinned key-poses (elements) in multi-pass rendering, across multiple agents and frames, while endorsing an interpolative approach for key-pose blending. Within this context, each item symbolizes a member of the crowd. More applications are evident in data stream compression using bloom filters, as reported in [17].
SUKP is an NP -hard problem [12], and various methods have been employed to address it. Theoretical studies using greedy approaches or dynamic programming are presented in [12,22]. In [23], an integer linear programming model was developed and applied to small instances comprising 85 and 100 items, yielding optimal solutions.
Metaheuristic algorithms have been employed to tackle SUKP in various studies. In [24], the authors utilize an artificial bee colony technique to address SUKP, integrating a greedy operator to manage infeasible solutions. An enhanced moth search algorithm is developed in [25], incorporating a differential mutation operator to boost efficiency. The Jaya algorithm, along with a differential evolution technique, is applied in [26] to enhance exploration capability. A Cauchy mutation is used to improve exploitation ability, while an enhanced repair operator is designed to rectify infeasible solutions.
In [18], the efficacy of various transfer functions is examined for binarizing moth metaheuristics. A local search operator is devised in [27] and applied to large-scale instances of SUKP. The study proposes three strategies in line with the adaptive tabu search framework, enabling efficient solutions for new SUKP instances. In [28], the grey wolf optimizer (GWO) algorithm is adapted to tackle binary problems. Instead of traditional binarization methods, the study employs a multiple parent crossover with two distinct dominance tactics, replicating GWO’s leadership hierarchy technique. Furthermore, an adaptive mutation featuring an exponentially decreasing step size is employed, aiming to inhibit premature convergence and establish a balance between intensification and diversification.
In [29], the authors merge machine learning and metaheuristics to devise a Q-learning reinforcement strategy for binary optimization problems, using PSO, genetic algorithm, and gbPSO as optimizers. Enhanced optimizers incorporate initial solution generation and a random immigrants mechanism, while a mutation procedure fosters intensified search. This approach is applied to the set-union knapsack problem, producing promising outcomes. Meanwhile, [30] investigates the impact of integrating a backtracking strategy into population-based approaches for the same problem. The proposed method features swarm optimization, an iterative search operator, and a path-relinking strategy for obtaining high-quality solutions. Performance evaluation using benchmark instances demonstrates promising results when compared to existing methods.

3. Initialization, Metaheuristic, and Search Operators

This section outlines a comparison of three distinct initialization operators: a random operator, a greedy operator, and a weighted operator based on a specific indicator. The overall functioning of the proposed algorithm is depicted in Figure 1, with subsequent sections delving into detailed descriptions of each operator. These initialization operators are further elaborated in Section 3.1. Additionally, we discuss the hybrid operator, a unique blend of machine learning techniques and metaheuristics, which is tasked with executing movements. For this study, we implemented a hybrid SCA, as detailed in Section 3.2. Finally, we examine the local search operator, which is utilized to refine the results obtained. This operator is discussed in Section 3.3.

3.1. Initialization Operators

The aim of these operators is to construct the initial solutions that will initiate the search process. To achieve this, items are sorted according to the ratio defined in Equation (3). The operators take s o r t I t e m s as input, which consists of elements arranged in descending order based on their r values. The output is a valid solution, denoted as Sol.
r = item   profit sum   of   element   weights
In the case of the greedy operator, Algorithm 1, at line 4, S o l is initialized with a random element, and this element is removed from the s o r t I t e m s list. Subsequently, at line 6, the knapsack constraint is checked; if the condition is met, the while loop is entered. Within the loop, a new item is assigned based on Equation (3), and it is removed again from s o r t I t e m s . Once the knapsack is full, the solution requires cleaning in line 10, as its weight is greater than or equal to k n a p s a c k S i z e . If the weights are equal, no action is taken. However, if the weight is greater, the items in S o l must be sorted using the r value defined in Equation (3), and removed in ascending order while checking the constraint after each removal. Once the constraint is satisfied, the procedure halts, and the solution S o l is returned.
In the context of the random operator, Algorithm 2, at line 4, a random item is used to initialize S o l , and this item is subsequently removed from the s o r t I t e m s list. Following this, the knapsack constraint is evaluated at line 6. If the condition is fulfilled, the program enters the while loop. Within this loop, a new item is randomly selected without considering the greedy condition and is promptly removed from s o r t I t e m s . When the knapsack is filled, the solution requires adjustment at line 10 as its weight equals or exceeds k n a p s a c k S i z e . If the weights are identical, no action is required. However, if the weight of S o l surpasses k n a p s a c k S i z e , the items in G must be arranged in accordance with the r-value defined in Equation (3) and are removed in ascending order, with the constraint being evaluated after each removal. The procedure concludes once the constraint is met, returning the solution S o l as the output.
Algorithm 1 Greedy initialization operator.
  1:
Function initGreedySolutions( s o r t I t e m s )
  2:
Input  s o r t I t e m s
  3:
Output  S o l
  4:
S o l getRandom()
  5:
s o r t I t e m s r e m o v e F r o m S o r t I t e m s ( I t e m )
  6:
while (weightSol < knapsackSize) do
  7:
    S o l a d d S o r t I t e m ( s o r t I t e m s )
  8:
    s o r t I t e m s r e m o v e F r o m S o r t I t e m s ( I t e m )
  9:
end while
10:
S o l cleanSol( S o l )
11:
return  S o l =0
Algorithm 2 Random initial operator.
  1:
Function initRandomSolutions( s o r t I t e m s )
  2:
Input  s o r t I t e m s
  3:
Output  S o l
  4:
S o l getRandom()
  5:
s o r t I t e m s r e m o v e F r o m S o r t I t e m s ( I t e m )
  6:
while (weightSol < knapsackSize) do
  7:
    S o l a d d R a n d o m I t e m ( s o r t I t e m s )
  8:
    s o r t I t e m s r e m o v e F r o m S o r t I t e m s ( I t e m )
  9:
end while
10:
S o l cleanSol( S o l )
11:
return  S o l
In the case of the weighted operator, Algorithm 3, the selection process differs in that items are chosen randomly but with a probability governed by Equation (3). In this approach, a normalized probability is constructed for each item, where the sum of all probabilities equals 1. Subsequently, the random selection of items is performed, taking into account the probability assigned to each of them.
Algorithm 3 Weighted initial operator.
  1:
Function initWeightedSolutions( s o r t I t e m s )
  2:
Input  s o r t I t e m s
  3:
Output  S o l
  4:
S o l getRandom()
  5:
s o r t I t e m s r e m o v e F r o m S o r t I t e m s ( I t e m )
  6:
while (weightSol < knapsackSize) do
  7:
    S o l a d d W e i g h t e d I t e m ( s o r t I t e m s )
  8:
    s o r t I t e m s r e m o v e F r o m S o r t I t e m s ( I t e m )
  9:
end while
10:
S o l cleanSol( S o l )
11:
return  S o l

3.2. Machine Learning Binarization Operator

The binarization process relies heavily on the machine learning binarization algorithm (MLBA). This algorithm receives the list of solutions l S o l from the prior iteration, the metaheuristic ( M H )—in this scenario, SCA, the optimal solution achieved thus far ( b e s t S o l ), and the transition probability for each cluster, t r a n s P r o b s , as input. In line 4, the metaheuristic M H is utilized on the list l S o l ; in this specific situation, it corresponds to SCA. The absolute values of velocities v l S o l are extracted from the result of applying M H to l S o l . These velocities symbolize the transition vector obtained through the application of the metaheuristic to the solution list. In line 5, k-means is used to cluster the entire set of velocities (getKmeansClustering); in this particular instance, K is designated as 5. It should be emphasized that the Algorithm 4 in conjunction with Algorithm 5 were proposed in the context of [11].
Algorithm 4 Machine learning binarization operator (MLBA).
  1:
Function MLBA( l S o l , M H , t r a n s P r o b a b s , b e s t S o l )
  2:
Input  l S o l , M H , t r a n s P r o b a b s
  3:
Output  l S o l , b e s t S o l
  4:
v l S o l getAbsValueVelocities( l S o l , M H )
  5:
l S o l C l u s t getKmeansClustering( v l S o l , K)
  6:
for (each S o l i in l S o l C l u s t ) do
  7:
   for (each d i m S o l i , j l in S o l i ) do
  8:
      d i m S o l P r o b i , j = getClusterProbability( d i m S o l , t r a n s P r o b a b s )
  9:
     if  d i m S o l P r o b i , j > r 1  then
10:
        Update l S o l i , j considering the best.
11:
     else
12:
        Do not update the item in l S o l i , j
13:
     end if
14:
   end for
15:
    S o l i cleanSol( S o l i )
16:
end for
17:
t e m p B e s t getBest( l S o l )
18:
if  c o s t ( t e m p B e s t ) > c o s t ( b e s t S o l )  then
19:
    t e m p B e s t execLocalSearch( t e m p B e s t )
20:
    b e s t S o l t e m p B e s t
21:
end if
22:
return  l S o l , b e s t S o l
Algorithm 5 Local search.
  1:
Function LocalSearch( b e s t S o l )
  2:
Input  b e s t S o l
  3:
Output  b e s t S o l
  4:
l s o l I t e m s , l s o l N o I t e m s getItems( b e s t S o l )
  5:
i = 0
  6:
while (i < T) do
  7:
    t e m p S o l swap( l s o l I t e m s , l s o l N o I t e m s )
  8:
   if  p r o f i t ( t e m p S o l ) > p r o f i t ( b e s t S o l ) and k n a p s a c k ( t e m p S o l ) < = k n a p s a c k S i z e  then
  9:
      b e s t S o l t e m p S o l
10:
   end if
11:
   i += 1
12:
end while
13:
return  b e s t S o l =0
For each solution S o l i and dimension j, a cluster assignment is made, and every cluster is linked with a transition probability ( t r a n s P r o b a b s ), organized based on the value of the cluster centroid. In this situation, the transition probabilities employed are [0.1, 0.2, 0.4, 0.5, 0.9]. The set of points that belongs to the cluster with the smallest centroid, depicted by the green color in Figure 2, is connected with a transition probability of 0.1. For the collection of blue points with the highest centroid value, a transition probability of 0.9 is associated. The smaller the value of the centroid, the lower the related t r a n s P r o b s .
In line 8, for each l S o l i , j , a transition probability d i m S o l P r o b i , j is allocated and subsequently contrasted with a random number r 1 in line 9. If d i m S o l P r o b i , j > r 1 , the solution undergoes an update considering the best value in line 10; otherwise, no update occurs, as indicated in line 12. After all solutions have been refreshed, a cleanup process, explained in Section 3.1, is applied. If a new best value emerges, a local search operator is executed in line 19. The details of this local search operator are provided in the following section. Ultimately, the revised list of solutions l S o l and the optimal solution b e s t S o l are returned.

3.3. Local Search Operator

The local search operator is invoked whenever a new best value is uncovered by the metaheuristic. This operator accepts the new best values ( b e s t S o l ) as input and, for its initial step, leverages them to ascertain the items that are included and excluded from b e s t S o l , as displayed in line 4 of Algorithm 5. These two lists of items undergo an iteration T = 200 times, effectuating a nonrepetitive swap, as exhibited in line 7 of Algorithm 5. Upon the completion of the swap, two conditions are assessed: whether the profit has improved and if the weight of the knapsack is less than or equal to k n a p s a c k S i z e . If both conditions are met, b e s t S o l is updated with t e m p S o l , and, ultimately, the refreshed b e s t S o l is returned.

4. Results

This section introduces the experiments performed using MLBA in combination with the sine cosine metaheuristic, aiming to assess the efficacy and contribution of the proposed algorithms when deployed for an NP -hard combinatorial problem. This specific variant of MLBA, which employs the sine cosine algorithm, will be referred to as MLSCABA. The SUKP was elected as a benchmark problem due to its extensive addressal by numerous algorithms and its presentation of nontrivial challenges when it comes to resolving small, medium, and large instances. It should be emphasized that the MLBA binarization technique is highly adaptable to other optimization algorithms. The optimization algorithm of choice was SCA, given its absence of a requirement for parameter tuning and its wide use in solving a variety of optimization problems.
The algorithm was implemented using Python 3.8 and executed on a Windows 10 PC equipped with a Core i7 processor and 32 GB of RAM. To evaluate the statistical significance of the differences, the Wilcoxon signed-rank test was applied with a significance level of 0.05. This test was selected following the methodology delineated in [31]. The Shapiro–Wilk normality test is utilized first in this process. If one of the populations does not adhere to a normal distribution and both populations have an identical number of points, the Wilcoxon signed-rank test is suggested for identifying the difference. In the experiments, the Wilcoxon test was employed to contrast the MLSCABA results with other variants or algorithms used in pairs. A comprehensive list of results was consistently employed for comparisons. The tests were constructed using the statsmodels and scipy libraries in Python. Each instance was solved 30 times to gather the best value and average indicators. Moreover, the average time (in seconds) necessary for the algorithm to discover the optimal solution is documented for each instance.
The initial set of instances, employed during the first phase of this study, were introduced in [32]. These instances encompass between 85 and 500 items and elements. They are distinguished by two parameters. The first parameter, μ = ( i = 1 m j = 1 n R i j ) / ( m n ) , signifies the density in the matrix, where R i j = 1 denotes that item i is included in element j. The second parameter, ν = C / ( j = 1 n w j ) , denotes the capacity ratio C over the total weight of the elements. As a result, an SUKP instance is labeled as m _ n _ μ _ ν . The secondary group of instances was proposed in [19] and ranges between 585 and 1000 items and elements. These instances were assembled following the same framework as the preceding set.

4.1. Parameter Setting

The parameter selection was guided by the methodology delineated in [20,33]. This technique draws upon four metrics, encapsulated in Equations (4)–(7), to facilitate judicious parameter selection. We generated values through instances 100_85_0.10_0.75, 100_100_0.15_0.85, and 85_100_0.10_0.75, with each parameter combination undergoing a tenfold validation. The parameters examined and subsequently chosen are documented in Table 1. To identify the optimal configuration, the polygonal area derived from the four-metric radar chart was computed for each setting. The configuration yielding the most expansive area was subsequently selected. Regarding transition probabilities, variation was exclusively confined to the probability of the fourth cluster, assessed at values of [0.5, 0.6, 0.7], while maintaining the rest at a constant level.
  • The percentage difference between the best value achieved and the best known value:
    b S o l u t i o n = 1 K n o w n B e s t V a l u e B e s t V a l u e K n o w n B e s t V a l u e
  • The percentage difference between the worst value achieved and the best known value:
    w S o l = 1 K n o w n B e s t V a l u e W o r s t V a l u e K n o w n B e s t V a l u e
  • The percentage deviation of the obtained average value from the best known value:
    a S o l = 1 K n o w n B e s t V a l u e A v e r a g e V a l u e K n o w n B e s t V a l u e
  • The convergence time utilized during the execution:
    n T i m e = 1 c o n v e r g e n c e T i m e m i n T i m e m a x T i m e m i n T i m e

4.2. Insight into Binary Algorithm

The objective of this section is to have the performance of various initialization operators assessed and compared, specifically random, weighted, and greedy, when applied to two sets of data. In order to carry out this comparison, several key performance indicators were considered, including best value achieved, average value, average execution time, and standard deviation. Valuable insights into the efficiency, effectiveness, and consistency of each initialization operator can be offered by these metrics. To ensure the reliability of the findings, each instance was executed 30 times, providing a more comprehensive evaluation of the performance of the operators across multiple runs.
For the analysis of the results, comparative tables are generated to present quantitative data, box plots are created to facilitate visual comparisons of the performance distributions of the operators, and convergence graphs for selected instances are illustrated to depict the progress of each operator over time. By utilizing these various data visualization techniques, a thorough understanding of the strengths and weaknesses of each initialization operator can be provided, enabling more informed decisions when selecting the most suitable operator for a specific problem or dataset.
The findings from an experimental study on medium-sized instances of the set-union knapsack problem are delineated in Table 2, with accompanying visualizations depicted in Figure 3 and Figure 4. Three distinct initialization operators—random, greedy, and weighted—were examined in this study. We investigated three different initialization operators, namely, random, greedy, and weighted. The table is formatted in such a way that the first column provides the designation of the instance being analyzed, followed by the column depicting the optimum known solutions for the respective instances. The subsequent four columns illustrate the findings associated with the random initialization operator, including the best-found solution, the average solution, computational time measured in seconds, and the standard deviation of obtained values. In a similar format, the next four columns present the results derived from the weighted operator, while the final quartet of columns elaborate on the outcomes resulting from the application of the greedy operator.
From Table 2, it can be inferred that the best values obtained are associated with the weighted operator, both in terms of best value and average. Additionally, it can be observed that the average convergence times are quite similar among the three operators. The exact number of instances in which the weighted operator outperformed the other two operators in terms of best value was counted, and it was found to be superior in five instances, the random operator in one instance, and the greedy operator in three instances. When the average indicator was analyzed, the weighted initialization operator was observed to outperform the others in 16 instances, while the random operator excelled in seven instances and the greedy operator in seven instances. This further demonstrated the consistent superiority of the weighted operator compared to the others. In the significance analysis, the weighted operator was indicated to be significantly superior to the other two initialization operators, both in terms of best value and average.
In Figure 3, the % gap of average value, defined in Equation (8), is compared with respect to the best-known value for the different variants developed in this experiment. The comparison is made through box plots. In the figure, it can be observed that visually, the three box plots are similar; however, the median is relatively better for the weighted operator. On the other hand, it is observed that both random and weighted operators have one outlier each. In both cases, it occurred for the same instance, sukp-200-185-0.15-0.85. The dispersion in the case of the greedy operator was greater than that of the other two, even though it is not apparent in the average.
The convergence patterns for four instances are delineated in Figure 4. On initial observation, all plots exhibit analogous trends of convergence, yet certain distinctions arise when delving into the specifics. Notably, the greedy operator demonstrates slower convergence rates in the cases of graphs ’a’ and ’c’ compared to the other two operators. However, this pattern is intriguingly inverted in the context of graph ’d’. Such variations indicate that, on average, the performance disparities among the operators tend to counterbalance across different instances.
% G a p = 100 B e s t k n o w n V a l u e A v e r a g e V a l u e B e s t k n o w n V a l u e
Experimental findings derived from large-sized instances of the set-union knapsack problem are presented in Table 3 and further elucidated by the illustrative graphics in Figure 5 and Figure 6. From the data in Table 3, it can be seen that superior results for both average and best value metrics were achieved on average by ’weighted’. Upon detailed inspection of individual instances, it was found that 14 ’best values’ were obtained by the ’weighted’ initialization operator, followed by ’random’ and ’greedy’, each achieving three. In terms of ’average’ values, a similar pattern was seen, with 19 optimal averages being obtained by ’weighted’, followed by ’greedy’ with six, and ’random’ with five. These observations suggest that the final outcome of the optimization process can be influenced by the method used for initializing the solutions. This is further confirmed by significance tests, which indicate a statistically significant difference.
Upon examining the box plots, a favorable trend towards better values for the ’weighted’ initialization operator is also discernible within the interquartile range. The dispersion of results appears more controlled, and the median value also demonstrates superior performance. Furthermore, when analyzing convergence times in conjunction with the graphs, it is observed that there is generally no significant difference among them, with very similar timings being recorded across all three operators. The convergence graphs, likewise, exhibit a comparable pattern.
Based on the findings from the experimental study, it can be concluded that superior performance was exhibited by the weighted initialization operator when compared to the random and greedy operators in solving medium to large-sized instances of the set-union knapsack problem. Although the average convergence times were found to be similar across all three operators, the most optimal solutions were consistently achieved by the weighted operator, along with superior average performance. These assertions were further corroborated by the significance analysis, in which superior performance in terms of best value and average was shown by the weighted operator. However, it must be noted that in certain instances, superior performance was achieved by the random and greedy operators over the weighted operator, indicating that the choice of initialization operator may be contingent upon the unique characteristics of each problem instance.

4.3. Comparisons

This section endeavors to contrast the proposed algorithm with a variety of optimization strategies, particularly focusing on BABC [24], binDE [24], gPSO [14], intAgents [34], and DH-Jaya [26]. The gPSO algorithm commences by creating a randomly populated binary set. By employing crossover and gene selection mechanisms, it generates new particles based on the current, personal, and globally optimal solutions. After the solutions are evaluated and updated accordingly, the algorithm perpetuates iterations until a designated termination criterion is reached. Furthermore, an optional mutation procedure is integrated as needed to preclude premature convergence. The binary artificial bee colony (BABC) is an optimization algorithm that draws inspiration from the foraging practices of honeybees. It incorporates three categories of bees: employed, onlooker, and scout bees. The employed bees actively seek food sources and relay their discoveries to the onlooker bees. Subsequently, the onlookers choose a food source predicated upon its perceived quality. Meanwhile, scout bees traverse uncharted territories in search of superior food sources. The algorithm operates in an iterative manner, continuously refining the food sources until it satisfies a predetermined stopping criterion.
The binary differential evolution (BinDE) algorithm is a population-based optimization strategy. It employs the differential evolution operator to generate novel candidate solutions. The algorithm sustains a population of potential solutions, iteratively updating them by crafting new ones via the differential evolution operator. This operator constructs a new solution by amalgamating two existing solutions with a third, randomly chosen one. Subsequently, the algorithm opts for the superior solution from among the new and existing ones to revamp the population. The algorithm culminates upon reaching a termination criterion, such as attaining a maximum number of iterations or a minimal enhancement in the objective function. The IntAgents algorithm is a swarm-based optimization strategy tailored to solve binary optimization problems. It comprises artificial search agents, each possessing unique cognitive intelligence that enables individual learning within the problem space. While these agents display varied search characteristics, they periodically share information about promising regions. Guided by a central swarm intelligence, these independent agents make use of adaptive information-sharing techniques. These techniques allow the search agents to learn across generations, mitigating the issues of premature convergence and local optima as effectively as possible.
The results are presented in Table 4. The table indicates that MLSCABA uniquely achieved the optimal values in 19 instances, underscoring its robust performance. Nevertheless, in other situations, at least two algorithms jointly attained the best values, highlighting a competitive landscape. When considering average values, both DH-Jaya and MLSCABA exhibited superior results, with DH-Jaya outperforming in 7 instances and MLSCABA in 23. Notably, while MLSCABA generally yields commendable outcomes, its performance is influenced by the process of solution initialization, indicating that its robustness may be tempered by these initial conditions.

5. Discussion

The findings of this study indicate that the weighted initialization operator tends to outperform the random and greedy operators in most instances of the medium- and large-sized SUKP. This is evident both in terms of the best value obtained and the average value. In particular, the weighted operator achieved the best value in more instances than the other two operators in both datasets. Furthermore, the weighted operator also outperformed the other two operators in terms of the average value in most instances. These results suggest that the weighted operator is more efficient and effective in initializing solutions for this problem.
However, it is important to note that the average execution time was similar among the three operators. This suggests that, although the weighted operator may produce higher-quality solutions, it does not necessarily do so faster than the other operators. Additionally, the convergence plots show that, although all three operators tend to converge to a solution over time, there are some differences in their convergence patterns. For instance, the greedy operator showed a slower convergence rate in some cases, but in other cases, its convergence rate was faster. These differences may be due to the specific characteristics of the problem instances.
In summary, the results suggest that the weighted initialization operator is generally the most effective for the set-union knapsack problem. However, it is also important to consider other factors, such as execution time and the specific characteristics of the problem, when selecting an initialization operator. The results obtained from the comparison of the gPSO, BABC, binDE, intAgents, DH-Jaya, and MLSCABA algorithms in solving the SUKP reveal a variety of strengths and weaknesses in each approach.
The gPSO algorithm, which uses a swarm optimization approach and an iterative process of crossover and gene selection, proved capable of finding optimal solutions in several instances. However, its average performance was surpassed by MLSCABA and DH-Jaya, suggesting it may struggle to maintain consistent performance across multiple iterations. On the other hand, BABC, which draws inspiration from the behavior of bees to search for and improve solutions, achieved the best value in some instances, but its average performance was inferior to that of MLSCABA and DH-Jaya. This may indicate that, although BABC’s approach can be effective in finding optimal solutions, it may not be as effective in maintaining consistent performance across multiple iterations.
The intAgents algorithm, which uses a swarm-based optimization strategy with artificial search agents, achieved the best value in several instances, but its average results were inferior to those of MLSCABA and DH-Jaya. This may indicate that, although intAgents’ approach can be effective in finding optimal solutions, it may struggle to maintain consistent performance across multiple iterations.
DH-Jaya showed solid performance both in terms of the best values and average results. Although it did not achieve the best value in as many instances as MLSCABA, it surpassed MLSCABA in terms of average results in several instances. This suggests that DH-Jaya may be a viable option for this problem, especially in situations where the average quality of the solution is more important than achieving the best possible value. Finally, the proposed algorithm, MLSCABA, achieved the best value in most instances and also showed solid performance in terms of average results. However, its performance may be affected by the initial conditions of the solution, suggesting that the selection of a good initialization operator may be crucial for its performance.

6. Conclusions

In the context of this research, three solution initialization methods were developed and evaluated: random, greedy, and weighted. These methods were integrated and tested in relation to a sine cosine algorithm that uses k-means as a binarization procedure. Tests were conducted with medium- and large-sized instances, and the results show that the solution initialization method significantly impacts the performance of the algorithm.
Specifically, it was observed that the weighted method, which introduces some control over the weight of each item while also incorporating a random component, exhibits superior performance compared to the greedy method. The latter focuses its attention on the quality of items based on a specific heuristic, but does not introduce random elements into the process.
In addition, the weighted method was proven to perform better than the completely random method (random), which does not consider any heuristic for the construction of functions. Regarding convergence times, no significant differences were observed among the methods. However, an improvement in terms of the quality of the solutions obtained was noted, which reiterates the importance of the initialization strategy on the effectiveness of the algorithm.
This study, while concentrating on the set union knapsack problem, unveils initialization methods with potentially broader applicability to a diverse range of combinatorial problems. These might encompass challenges as varied as the traveling salesman problem (TSP), vehicle routing problem (VRP), job shop scheduling problem (JSSP), and quadratic assignment problem (QAP). Each of these problems presents its own unique set of challenges, thus creating a plethora of opportunities for more in-depth examination and investigation in future studies.
Moreover, we believe that there is an opportunity to enhance the solution initialization process further. In this regard, one promising avenue for future research involves the development of adaptive initialization methods. Such methods, with the ability to modulate their behavior based on either the specific characteristics of the problem instance or the algorithm’s progress, could potentially contribute to a more refined, efficient, and effective problem-solving approach.

Author Contributions

Conceptualization, J.G.; Methodology, J.G.; Software, J.G.; Validation, J.G.; Formal analysis, J.G.; Investigation, J.G.; Data curation, J.G.; Writing—original draft, J.G.; Writing—review & editing, J.G. and A.L.-A.; Funding acquisition, B.C., R.S. and H.P. All authors have read and agreed to the published version of the manuscript.

Funding

José Gracía was supported by PROYECTO DI Regular 039.300/2023. Broderick Crawford was supported by the grant Broderick Crawford is supported by Grant CONICYT/FONDECYT/REGULAR/1210810.

Data Availability Statement

The results obtained can be accessed at the following Google Drive link: https://drive.google.com/drive/folders/19D-vuV55a19MVjNZCEGCnFtKV2RiHC04?usp=sharing, accessed on 8 June 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ma, L.; Shao, Z.; Li, L.; Huang, J.; Wang, S.; Lin, Q.; Li, J.; Gong, M.; Nandi, A.K. Heuristics and metaheuristics for biological network alignment: A review. Neurocomputing 2022, 491, 426–441. [Google Scholar] [CrossRef]
  2. Guo, H.; Liu, B.; Cai, D.; Lu, T. Predicting protein–protein interaction sites using modified support vector machine. Int. J. Mach. Learn. Cybern. 2018, 9, 393–398. [Google Scholar] [CrossRef]
  3. Gholizadeh, H.; Goh, M.; Fazlollahtabar, H.; Mamashli, Z. Modelling uncertainty in sustainable-green integrated reverse logistics network using metaheuristics optimization. Comput. Ind. Eng. 2022, 163, 107828. [Google Scholar] [CrossRef]
  4. Martínez-Muñoz, D.; García, J.; Martí, J.; Yepes, V. Discrete swarm intelligence optimization algorithms applied to steel–concrete composite bridges. Eng. Struct. 2022, 266, 114607. [Google Scholar] [CrossRef]
  5. Martínez-Muñoz, D.; García, J.; Martí, J.V.; Yepes, V. Optimal design of steel–concrete composite bridge based on a transfer function discrete swarm intelligence algorithm. Struct. Multidiscip. Optim. 2022, 65, 312. [Google Scholar] [CrossRef]
  6. Dokeroglu, T.; Deniz, A.; Kiziloz, H.E. A comprehensive survey on recent metaheuristics for feature selection. Neurocomputing 2022, 494, 269–296. [Google Scholar] [CrossRef]
  7. Agushaka, J.O.; Ezugwu, A.E.; Abualigah, L.; Alharbi, S.K.; Khalifa, H.A.E.W. Efficient Initialization Methods for Population-Based Metaheuristic Algorithms: A Comparative Study. Arch. Comput. Methods Eng. 2023, 30, 1727–1787. [Google Scholar] [CrossRef]
  8. Li, Q.; Liu, S.Y.; Yang, X.S. Influence of initialization on the performance of metaheuristic optimizers. Appl. Soft Comput. 2020, 91, 106193. [Google Scholar] [CrossRef] [Green Version]
  9. Georgioudakis, M.; Lagaros, N.D.; Papadrakakis, M. Probabilistic shape design optimization of structural components under fatigue. Comput. Struct. 2017, 182, 252–266. [Google Scholar] [CrossRef]
  10. Agushaka, J.O.; Ezugwu, A.E. Initialisation approaches for population-based metaheuristic algorithms: A comprehensive review. Appl. Sci. 2022, 12, 896. [Google Scholar] [CrossRef]
  11. García, J.; Lemus-Romani, J.; Altimiras, F.; Crawford, B.; Soto, R.; Becerra-Rozas, M.; Moraga, P.; Becerra, A.P.; Fritz, A.P.; Rubio, J.M.; et al. A binary machine learning cuckoo search algorithm improved by a local search operator for the set-union knapsack problem. Mathematics 2021, 9, 2611. [Google Scholar] [CrossRef]
  12. Goldschmidt, O.; Nehme, D.; Yu, G. Note: On the set-union knapsack problem. Nav. Res. Logist. (NRL) 1994, 41, 833–842. [Google Scholar] [CrossRef]
  13. Wei, Z.; Hao, J.K. Multistart solution-based tabu search for the Set-Union Knapsack Problem. Appl. Soft Comput. 2021, 105, 107260. [Google Scholar] [CrossRef]
  14. Ozsoydan, F.B.; Baykasoglu, A. A swarm intelligence-based algorithm for the set-union knapsack problem. Future Gener. Comput. Syst. 2019, 93, 560–569. [Google Scholar] [CrossRef]
  15. Liu, X.J.; He, Y.C. Estimation of distribution algorithm based on Lévy flight for solving the set-union knapsack problem. IEEE Access 2019, 7, 132217–132227. [Google Scholar] [CrossRef]
  16. Tu, M.; Xiao, L. System resilience enhancement through modularization for large scale cyber systems. In Proceedings of the 2016 IEEE/CIC International Conference on Communications in China (ICCC Workshops), IEEE, Chengdu, China, 27–29 July 2016; pp. 1–6. [Google Scholar]
  17. Yang, X.; Vernitski, A.; Carrea, L. An approximate dynamic programming approach for improving accuracy of lossy data compression by Bloom filters. Eur. J. Oper. Res. 2016, 252, 985–994. [Google Scholar] [CrossRef] [Green Version]
  18. Feng, Y.; An, H.; Gao, X. The importance of transfer function in solving set-union knapsack problem based on discrete moth search algorithm. Mathematics 2019, 7, 17. [Google Scholar] [CrossRef] [Green Version]
  19. Wei, Z.; Hao, J.K. Kernel based tabu search for the Set-union Knapsack Problem. Expert Syst. Appl. 2021, 165, 113802. [Google Scholar] [CrossRef]
  20. García, J.; Crawford, B.; Soto, R.; Castro, C.; Paredes, F. A k-means binarization framework applied to multidimensional knapsack problem. Appl. Intell. 2018, 48, 357–380. [Google Scholar] [CrossRef]
  21. Lister, W.; Laycock, R.; Day, A. A Key-Pose Caching System for Rendering an Animated Crowd in Real-Time. Comput. Graph. Forum 2010, 29, 2304–2312. [Google Scholar] [CrossRef]
  22. Arulselvan, A. A note on the set union knapsack problem. Discret. Appl. Math. 2014, 169, 214–218. [Google Scholar] [CrossRef]
  23. Wei, Z.; Hao, J.K. Iterated two-phase local search for the Set-Union Knapsack Problem. Future Gener. Comput. Syst. 2019, 101, 1005–1017. [Google Scholar] [CrossRef] [Green Version]
  24. He, Y.; Xie, H.; Wong, T.L.; Wang, X. A novel binary artificial bee colony algorithm for the set-union knapsack problem. Future Gener. Comput. Syst. 2018, 78, 77–86. [Google Scholar] [CrossRef]
  25. Feng, Y.; Yi, J.H.; Wang, G.G. Enhanced moth search algorithm for the set-union knapsack problems. IEEE Access 2019, 7, 173774–173785. [Google Scholar] [CrossRef]
  26. Wu, C.; He, Y. Solving the set-union knapsack problem by a novel hybrid Jaya algorithm. Soft Comput. 2020, 24, 1883–1902. [Google Scholar] [CrossRef]
  27. Zhou, Y.; Zhao, M.; Fan, M.; Wang, Y.; Wang, J. An efficient local search for large-scale set-union knapsack problem. Data Technol. Appl. 2020, 55, 233–250. [Google Scholar] [CrossRef]
  28. Gölcük, İ.; Ozsoydan, F.B. Evolutionary and adaptive inheritance enhanced Grey Wolf Optimization algorithm for binary domains. Knowl.-Based Syst. 2020, 194, 105586. [Google Scholar] [CrossRef]
  29. Ozsoydan, F.B.; Gölcük, İ. A reinforcement learning based computational intelligence approach for binary optimization problems: The case of the set-union knapsack problem. Eng. Appl. Artif. Intell. 2023, 118, 105688. [Google Scholar] [CrossRef]
  30. Dahmani, I.; Ferroum, M.; Hifi, M. Effect of Backtracking Strategy in Population-Based Approach: The Case of the Set-Union Knapsack Problem. Cybern. Syst. 2022, 53, 168–185. [Google Scholar] [CrossRef]
  31. Martínez-Muñoz, D.; García, J.; Martí, J.V.; Yepes, V. Hybrid Swarm Intelligence Optimization Methods for Low-Embodied Energy Steel-Concrete Composite Bridges. Mathematics 2022, 11, 140. [Google Scholar] [CrossRef]
  32. He, Y.; Wang, X. Group theory-based optimization algorithm for solving knapsack problems. Knowl.-Based Syst. 2021, 219, 104445. [Google Scholar] [CrossRef]
  33. García, J.; Moraga, P.; Valenzuela, M.; Pinto, H. A db-scan hybrid algorithm: An application to the multidimensional knapsack problem. Mathematics 2020, 8, 507. [Google Scholar] [CrossRef] [Green Version]
  34. Ozsoydan, F.B. Artificial search agents with cognitive intelligence for binary optimization problems. Comput. Ind. Eng. 2019, 136, 18–30. [Google Scholar] [CrossRef]
Figure 1. MLSCABA flow chart.
Figure 1. MLSCABA flow chart.
Mathematics 11 02695 g001
Figure 2. K-means binarization procedure.
Figure 2. K-means binarization procedure.
Mathematics 11 02695 g002
Figure 3. Box plot of percentage gap for random, weighted, and greedy initialization operators for medium-sized instances.
Figure 3. Box plot of percentage gap for random, weighted, and greedy initialization operators for medium-sized instances.
Mathematics 11 02695 g003
Figure 4. Convergence chart for random, weighted, and greedy initialization operators for selected medium-sized instances.
Figure 4. Convergence chart for random, weighted, and greedy initialization operators for selected medium-sized instances.
Mathematics 11 02695 g004
Figure 5. Box plot of percentage gap for random, weighted, and greedy initialization operators for large instances.
Figure 5. Box plot of percentage gap for random, weighted, and greedy initialization operators for large instances.
Mathematics 11 02695 g005
Figure 6. Convergence chart for random, weighted, and greedy initialization operators for selected large-sized instances.
Figure 6. Convergence chart for random, weighted, and greedy initialization operators for selected large-sized instances.
Mathematics 11 02695 g006
Table 1. Parameter setting for the MLSCABA.
Table 1. Parameter setting for the MLSCABA.
ParametersDescriptionValueRange
NNumber of solutions10[5, 10, 20]
KClusters number5[4, 5]
TMaximum local search iterations200[100, 200, 400]
Transition probabilityTransition probability[0.1, 0.2, 0.4, 0.5, 0.9][0.1, 0.2, 0.4, [0.5, 0.6, 0.7], 0.9]
Table 2. Comparison of initialization operators: random, weighted, and greedy for the medium-sized instance set.
Table 2. Comparison of initialization operators: random, weighted, and greedy for the medium-sized instance set.
RandomWeightedGreedy
InstancesBest KnownBest ValueAverageTime (s)StdBest ValueAverageTime (s)StdBest ValueAverageTime (s)Std
85_100_0.10_0.7512,045.0012,045.0011,835.331.37193.7212,045.0011,866.071.22206.9312,045.0011,864.571.24177.94
85_100_0.15_0.8512,369.0012,369.0012,009.372.32392.5912,369.0012,043.532.52328.8612,369.0011,938.572.17361.75
100_100_0.10_0.7514,044.0014,044.0013,926.972.3992.8814,044.0013,882.302.34109.2814,044.0013,894.903.64134.65
100_100_0.15_0.8513,508.0013,508.0013,116.073.90239.4713,508.0013,174.175.46267.7513,498.0012,998.703.60429.24
100_85_0.10_0.7513,283.0013,283.0013,021.873.01154.0913,283.0013,029.032.42169.8413,283.0012,992.833.84138.88
100_85_0.15_0.8512,479.0012,273.0012,044.774.71203.2412,479.0012,015.005.96340.2412,348.0012,034.603.18228.98
185_200_0.10_0.7513,696.0013,696.0013,582.933.90101.9713,696.0013,559.104.64113.6813,696.0013,593.775.5482.97
185_200_0.15_0.8511,298.0011,298.0010,908.874.90248.2811,298.0010,923.505.64278.4711,298.0010,821.304.87255.41
200_185_0.10_0.7513,521.0013,502.0013,284.006.97176.3713,502.0013,287.406.56143.3813,502.0013,348.807.9299.57
200_185_0.15_0.8514,215.0013,993.0013,243.1710.40479.6014,215.0013,234.8312.84477.2113,993.0013,282.679.08387.88
200_200_0.10_0.7512,522.0012,522.0012,309.574.22183.4612,522.0012,262.774.37117.0012,522.0012,266.137.26150.87
200_200_0.15_0.8512,317.0012,238.0011,809.537.54230.2112,317.0011,842.537.00214.4212,167.0011,838.7311.66182.89
285_300_0.10_0.7511,568.0011,568.0011,468.432.9886.2811,568.0011,495.172.5469.0011,568.0011,479.573.3973.66
285_300_0.15_0.8511,802.0011,802.0011,372.005.77205.8311,802.0011,411.338.18243.8811,763.0011,342.506.26241.63
300_285_0.10_0.7511,563.0011,559.0011,314.1710.06143.5511,563.0011,346.009.05179.3111,563.0011,338.937.24170.56
300_285_0.15_0.8512,607.0012,380.0012,062.0710.79300.2112,402.0012,084.538.04258.4712,411.0012,034.7310.49216.47
300_300_0.10_0.7512,817.0012,817.0012,634.205.4193.4312,817.0012,652.274.8387.3112,817.0012,620.904.0897.47
300_300_0.15_0.8511,585.0011,425.0011,322.176.46140.2411,425.0011,288.097.88212.7611,448.0011,317.237.32133.51
385_400_0.10_0.7510,600.0010,490.0010,344.233.4974.4310,600.0010,357.734.6982.1010,483.0010,359.404.1567.64
385_400_0.15_0.8510,506.0010,506.0010,164.507.07153.8810,506.0010,175.876.72211.3610,506.0010,131.307.25205.12
400_385_0.10_0.7511,484.0011,484.0011,391.972.5789.4911,484.0011,409.572.8288.9711,484.0011,410.732.4987.33
400_385_0.15_0.8511,209.0011,209.0010,774.0010.52292.2111,209.0010,725.408.82292.5511,209.0010,805.5311.31299.81
400_400_0.10_0.7511,665.0011,665.0011,495.6710.54141.4911,665.0011,501.204.21157.2511,665.0011,451.634.84132.43
400_400_0.15_0.8511,325.0011,325.0010,907.2310.47422.5211,325.0010,976.209.24404.4711,325.0010,930.608.70441.57
485_500_0.10_0.7511,321.0011,115.0010,890.136.47106.0711,260.0010,885.777.73175.5311,186.0010,860.707.95112.53
485_500_0.15_0.8510,220.0010,208.009835.437.70234.4910,208.009871.239.36215.2910,208.009936.117.67220.54
500_485_0.10_0.7511,771.0011,729.0011,523.778.68151.0811,729.0011,504.108.64139.3111,698.0011,480.538.08159.68
500_485_0.15_0.8510,238.0010,059.009834.577.23139.1810,059.009856.207.54111.3210,086.009840.107.25141.35
500_500_0.10_0.7511,249.0011,217.0010,887.076.35120.3411,123.0010,854.376.08113.8611,078.0010,840.976.0897.11
500_500_0.15_0.8510,381.0010,203.009931.939.08185.6610,381.009953.778.68261.3810,381.009885.107.00235.07
Average 11,897.0711,621.476.56185.3611,928.2111,627.126.51197.6911,901.0711612.116.52186.60
Wilcoxon p-value 0.020.02 0.010.009
The values highlighted in bold represent the optimal outcomes obtained. It’s important to note that these are only marked when the best result is exclusive to a single algorithm. If there are two algorithms yielding identical optimal values, that instance will not be emphasized.
Table 3. Comparison of initialization operators: random, weighted, and greedy for the large-sized instance set.
Table 3. Comparison of initialization operators: random, weighted, and greedy for the large-sized instance set.
RandomWeightedGreedy
InstancesBest KnownBest ValueAverageTime (s)StdBest ValueAverageTime (s)StdBest ValueAverageTime (s)Std
1000_1000_0.10_0.7595448985.008429.477.15365.169084.008428.635.41259.409046.008347.775.75270.27
1000_1000_0.15_0.8584748129.007524.077.56369.288165.007616.438.34341.378079.007550.739.61338.77
1000_985_0.10_0.7596688776.008165.174.39336.539170.008318.835.46390.168788.008221.975.15320.91
1000_985_0.15_0.8584538159.007528.705.50379.778159.007528.705.50379.778159.007741.676.58351.04
585_600_0.10_0.7510,39310,334.009932.169.26146.2610,334.009959.278.15146.2610,233.009880.338.95156.26
585_600_0.15_0.8592569256.008802.116.57164.139256.008812.536.57164.139256.008827.277.69155.69
600_585_0.10_0.7599149741.009680.723.0441.509741.009681.613.1541.509914.009676.734.9165.38
600_585_0.15_0.8593579357.009052.1410.01137.279357.009054.079.86137.279258.009042.677.9695.32
600_600_0.10_0.751052410,469.0010,350.586.57101.4410,490.0010,384.706.7091.8710,469.0010,354.776.0074.25
600_600_0.15_0.8590629024.008847.175.4773.519024.008857.096.6780.378976.008834.935.9167.09
685_700_0.10_0.7510,1219827.009420.178.16235.369926.009548.4010.38182.619786.009472.538.89201.18
685_700_0.15_0.8592568990.008699.4010.20148.289110.008779.239.04124.609110.008727.578.12140.91
700_685_0.10_0.7598819635.009397.328.69196.049736.009452.778.07226.749695.009410.737.39147.66
700_685_0.15_0.8591639106.008811.938.24158.989135.008790.239.75171.429106.008718.5310.97176.12
700_700_0.10_0.7597869585.009138.309.88357.459467.009171.7810.02235.859637.009307.9310.76145.67
700_700_0.15_0.8592298886.008582.137.68147.988924.008602.878.70172.419143.008664.437.13219.65
785_800_0.10_0.7593849096.008640.279.71266.719077.008603.409.95353.379124.008627.978.71214.81
785_800_0.15_0.8587468355.008062.077.87238.448572.008159.539.64181.678366.008095.3710.11186.26
800_785_0.10_0.7598379525.009091.6711.18324.689577.009060.8312.39297.779343.008955.309.54314.30
800_785_0.15_0.8590248679.008456.637.12142.718907.008555.937.13145.528907.008498.407.31152.91
800_800_0.10_0.7599329607.009309.7310.26227.549786.009372.3710.36216.129661.009413.879.43171.85
800_800_0.15_0.8591018864.008389.178.97287.138841.008395.378.11275.128804.008381.908.17242.94
885_900_0.10_0.7593189058.008656.777.97304.859019.008596.407.31282.419030.008653.277.73228.28
885_900_0.15_0.8584258125.007669.239.88297.908027.007568.778.21318.228120.007620.409.08325.51
900_885_0.10_0.7597259177.008859.878.70346.509464.008953.639.16327.099305.008861.337.85315.26
900_885_0.15_0.8586208385.007814.577.69327.928427.007902.339.38324.418385.007872.337.71342.64
900_900_0.10_0.7597459467.008989.508.50323.269499.008870.777.41406.929465.008949.409.28437.68
900_900_0.15_0.8589908510.007747.235.13522.618647.007908.007.73438.398647.007888.376.44479.86
985_1000_0.10_0.7591938703.008106.176.42323.158931.008076.707.06436.638665.008062.875.70327.51
985_1000_0.15_0.8585288143.007546.136.16294.458143.007571.375.42281.918134.007506.235.61332.20
Average9354.979065.108658.007.80252.899133.178686.068.07247.719087.038672.257.85233.27
Wilcoxon p-value 0.0020.01 0.020.03
The values highlighted in bold represent the optimal outcomes obtained. It’s important to note that these are only marked when the best result is exclusive to a single algorithm. If there are two algorithms yielding identical optimal values, that instance will not be emphasized.
Table 4. Comparative analysis of MLSCABA and various algorithms in solving SUKP for medium-sized instances.
Table 4. Comparative analysis of MLSCABA and various algorithms in solving SUKP for medium-sized instances.
Best KnownBest ValueAverage Value
Instance BABCbinDEgPSOintAgentsDH-jayaMLSCABABABCbinDEgPSOintAgentsDH-jayaMLSCABA
85_100_0.10_0.751204511,66411,35212,04512,04512,04512,04511,182.711,07511,486.9511,419.7511,570.611,866.1
85_100_0.15_0.8512,36912,36912,36912,36912,36912,36912,36912,081.611,875.911,994.3611,885.2112,31812,043.5
100_100_0.10_0.7514,04413,86013,81414,04414,04414,04414,04413,734.913,675.913,854.7113,767.2313,912.513,882.3
100_100_0.15_0.8513,50813,50813,40713,50813,50813,50813,50813,352.413,212.813,347.5813,003.6213,439.113,174.2
100_85_0.10_0.7513,28313,25113,04413,28313,28313,28313,28313028.512,99113,050.5313,061.0213,07613,029.0
100_85_0.15_0.8512,47912,23812,27412,27412,27412,27412,47912,15512,123.912,084.8212,074.8412,192.512,015.0
185_200_0.10_0.7513,69613,04713,02413,69613,69613,69613,69612,522.812,277.513,204.2613,084.5213,350.213,559.1
185_200_0.15_0.8511,29810,60210,54711,29811,29811,29811,29810,150.610,085.410,801.4110,780.1410,828.910,923.5
200_185_0.10_0.7513,52113,24113,24113,40513,50213,40513,50213,064.412,940.713,286.5613,226.2813,306.613,287.4
200_185_0.15_0.8514,21513,82913,67114,04414,04414,21514,21513,359.213,11013,492.613,441.0613,660.213,234.8
200_200_0.10_0.7512,52211,84611,53512,52212,52212,52212,52211,194.310,969.411,898.7311,586.2612,171.612,262.8
200_200_0.15_0.8512,31711,52111,46912,31711,91112,18712,31710,94510,717.111,584.6411,288.2511,74611,842.5
285_300_0.10_0.7511,56811,15811,15211,56811,56811,56811,56810,775.910,661.311,317.9911,205.7211,327.711,495.2
285_300_0.15_0.8511,80210,52810,52811,51711,51711,40111,8029897.929832.3210,899.210,747.3311,025.911,411.3
300_285_0.10_0.7511,56310,42810,42011,33511,33510,93411,5639994.769899.2410,669.5110,576.110,703.211,346.0
300_285_0.15_0.8512,60712,01211,66112,24512,24712,24512,40210,902.910,499.411,607.111,490.2612,037.512,084.5
300_300_0.10_0.7512,81712,18612,30412,69512,69512,69512,81711,945.811,864.412,411.2712,310.1912,569.312,652.3
300_300_0.15_0.8511,58510,38210,38211,42511,42511,11311,4259859.699710.3710,568.4110,38410,701.911,288.1
385_400_0.10_0.7510,60010,085988310,48310,32610,41410,6009537.59314.5710,013.439892.1710,01710,357.7
385_400_0.15_0.8510,5069456935210,33810,13110,30210,5069090.038846.999524.989339.679565.7210,175.9
400_385_0.10_0.7511,48410,76610,57611,48411,48411,33711,48410,065.29681.4610,915.8710,734.6211,06211,409.6
400_385_0.15_0.8511,2099649964910,71010,71010,43111,2099135.989020.879864.55973510,017.910,725.4
400_400_0.10_0.7511,66510,62610,46211,53111,53111,31011,66510,101.19975.810,958.9610,756.9210,914.811,501.2
400_400_0.15_0.8511,3259541938810,92710,92710,91511,3259032.958768.429845.179608.079969.910,976.2
485_500_0.10_0.7511,32110,82310,72811,09411,09410,97111,26010,483.410,159.410,687.6210,603.5310,754.810,885.8
485_500_0.15_0.8510,2209333921810,10410,104971510,2089085.578919.649383.289259.369467.89871.2
500_485_0.10_0.7511,77110,78410,58611,72211,72211,72211,72910,452.210,363.811,184.5111,111.6311,269.411,504.1
500_485_0.15_0.8510,2389090919110,02210,059977010,0598857.898783.999299.569165.269354.289856.2
500_500_0.10_0.7511,24910,75510,54610,88810,96010,96011,12310,328.510,227.710,681.4610,610.5310,703.510,854.4
500_500_0.15_0.8510,3819318931210,19410,38110,17610,3819180.749096.139703.629578.899801.59953.8
Average11,956.911,209.411,120.111,809.811,796.411,729.011,946.810,794.110,633.211,290.811,157.911,391.011,649.0
The values highlighted in bold represent the optimal outcomes obtained. It’s important to note that these are only marked when the best result is exclusive to a single algorithm. If there are two algorithms yielding identical optimal values, that instance will not be emphasized.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

García, J.; Leiva-Araos, A.; Crawford, B.; Soto, R.; Pinto, H. Exploring Initialization Strategies for Metaheuristic Optimization: Case Study of the Set-Union Knapsack Problem. Mathematics 2023, 11, 2695. https://doi.org/10.3390/math11122695

AMA Style

García J, Leiva-Araos A, Crawford B, Soto R, Pinto H. Exploring Initialization Strategies for Metaheuristic Optimization: Case Study of the Set-Union Knapsack Problem. Mathematics. 2023; 11(12):2695. https://doi.org/10.3390/math11122695

Chicago/Turabian Style

García, José, Andres Leiva-Araos, Broderick Crawford, Ricardo Soto, and Hernan Pinto. 2023. "Exploring Initialization Strategies for Metaheuristic Optimization: Case Study of the Set-Union Knapsack Problem" Mathematics 11, no. 12: 2695. https://doi.org/10.3390/math11122695

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop