You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

11 April 2023

Hybrid Learning Moth Search Algorithm for Solving Multidimensional Knapsack Problems

,
,
,
and
1
School of Information Engineering, Hebei GEO University, Shijiazhuang 050031, China
2
Intelligent Sensor Network Engineering Research Center of Hebei Province, Shijiazhuang 050031, China
3
Xinjiang Institute of Engineering, Information Engineering College, Ürümqi 830023, China
4
Shanwei Institute of Technology, Shanwei 516600, China
This article belongs to the Special Issue Evolutionary Computation 2022

Abstract

The moth search algorithm (MS) is a relatively new metaheuristic optimization algorithm which mimics the phototaxis and Lévy flights of moths. Being an NP-hard problem, the 0–1 multidimensional knapsack problem (MKP) is a classical multi-constraint complicated combinatorial optimization problem with numerous applications. In this paper, we present a hybrid learning MS (HLMS) by incorporating two learning mechanisms, global-best harmony search (GHS) learning and Baldwinian learning for solving MKP. (1) GHS learning guides moth individuals to search for more valuable space and the potential dimensional learning uses the difference between two random dimensions to generate a large jump. (2) Baldwinian learning guides moth individuals to change the search space by making full use of the beneficial information of other individuals. Hence, GHS learning mainly provides global exploration and Baldwinian learning works for local exploitation. We demonstrate the competitiveness and effectiveness of the proposed HLMS by conducting extensive experiments on 87 benchmark instances. The experimental results show that the proposed HLMS has better or at least competitive performance against the original MS and some other state-of-the-art metaheuristic algorithms. In addition, the parameter sensitivity of Baldwinian learning is analyzed and two important components of HLMS are investigated to understand their impacts on the performance of the proposed algorithm.

1. Introduction

The multidimensional knapsack problem (MKP) [1,2] is a generalization of the 0–1 knapsack problem (0–1 KP) [3]. As a classical combinatorial optimization problem, numerous real-world applications can be modeled as MKP, such as capital budgeting [4,5], cutting stock [6], and loading problem [7,8].
Then the MKP is: given a set of n items and a set of m knapsacks (m < n), with cj = profit of item j, bi = capacity constraint of knapsack i,aij = resource consumption of item j in the ith knapsack. The goal of MKP is to select a subset of items so that the total profit of the selected items is a maximum while the total weights in each dimension i (i = 1, 2, …, m) do not exceed the corresponding capacity bi. Formally,
m a x z = j = 1 n c j x j
s . t . j = 1 n a i j x j b i , i { 1 , 2 , , m }
x j { 0 , 1 } , j { 1 , 2 , , n }
where xj (j = 1, 2, …, n) is a 0–1 decision variable, such that xj = 1 if item j is assigned to a knapsack, xj = 0 otherwise. When m = 1, MKP reduces to the 0–1 KP.
Known as the NP-hard combinatorial optimization problem of MKP, the conventional exact algorithms usually are unable to obtain a satisfactory solution in a reasonable time, especially for large-scale instances. It is because that combinatorial optimization problems often feature a combination of the explosion and then the search space grows exponentially with the expansion of the scale. In this case, metaheuristic algorithms are effective methods for MKP, which can obtain near-optimal solutions in a reasonable acceptable time. Representative metaheuristic algorithms for solving MKP include the two-phase tabu evolutionary algorithm (TPTEA) [9], binary particle swarm optimization (PSO) [10], quantum particle swarm optimization (QPSO) [11], diversity-preserving quantum particle swarm optimization (DQPSO*) [12], hybrid estimation of distribution algorithm (EDA) [13], memetic algorithm (MA) [14], binary grey wolf optimizer (GWO) [15], hybrid harmony search algorithm (HS) [16], binary multi-verse optimizer (MMVO) [17], cuckoo search (CS) [18], pigeon-inspired optimization algorithm (PIO) [19], binary moth search algorithm (MS) [20], sine cosine algorithm (SCA) [21], and binary slime mould algorithm (BSMA) [22]. For more information on evolutionary algorithms (EAs) [23] and exact methods for solving MKP problem, please refer to [24].
Recently, some novel metaheuristic algorithms have been proposed and used to solve various optimization problems, including continuous optimization problems and discrete optimization problems, such as differential evolution algorithm (DE) [25,26], cuckoo search algorithm (CS) [27], bat algorithm (BA) [28], krill herd (KH) [29,30], elephant herding optimization (EHO) [31], monarch butterfly optimization algorithm (MBO) [32,33], brain storm optimization algorithm (BSO) [34], fruit fly optimizer (FFO) [35], whale optimization algorithm (WOA) [36], wind driven optimization (WDO) [37], salp swarm algorithm (SSA) [38], Harris hawks optimization (HHO) [39], artificial Jellyfish Search (JS) optimizer [40], artificial rabbits optimization (ARO) [41], mountain gazelle optimizer (MGO) [42], moth-flame optimizer (MFO) [43], and many more.
The moth search (MS) [44,45] algorithm was recently developed by Wang and inspired by the phototaxis and Lévy flights of moths. Owing to its simplicity and high search efficiency, MS has been successfully applied to solve various optimization problems, such as the cloud task scheduling problem [46], drone placement problem [47], constrained optimization problems [48], discounted {0–1} knapsack problem [49], and set-union knapsack problem [50]. Previous studies have shown that MS is effective for solving various combinatorial optimization problems, especially for various variants of KP problems [49,51]. To the best of our knowledge, there are few literatures on applying MS to solve MKP. Hence, we concentrated on the MS algorithm for solving MKP.
In recent years, the learning-based metaheuristic algorithms have been extensively reported in literature because of their effectiveness and efficiency in solving various optimization problems. The core idea is that the metaheuristic algorithm combines specific learning operators or learning mechanisms to enhance itself some learning ability, and then owning better optimization behavior. There are many learning methods, for example, deep learning [52,53], reinforcement learning [54,55,56], transfer learning [57], Q-learning [58], information feedback [59], orthogonal learning [34], comprehensive learning [60], Baldwinian learning, and so on.
Inspired by Darwinian evolution, evolutionary computation mainly includes the random variation of individuals and fitness selection mechanism [61]. It seems time-consuming to search randomly for good genotypes without using phenotypes. One of the possible ways to surmount this deficiency is to integrate the learning mechanism into the evolutionary search. The learning mechanism can provide a more effective search path [62]. Hence, learning-based mechanisms have been extensive researched and applied in enhancing the search performances of evolution algorithms [14,31]. The Baldwin effect, sometimes called Baldwinian learning, was widely incorporated into a variety of evolutionary computing models to enhance their performance [61,63,64]. The above works motivate us to propose a Baldwinian learning MS to solve MKP.
Harmony search (HS) [65] is metaheuristic and imitates the musical improvization process to search for a perfect state of harmony. Since proposed, HS has always attracted much attention from researchers and has been successfully applied to deal with various optimization problems. An effective variant of HS, called global-best HS (GHS), was proposed by Omran [66]. Compared with HS and other variants, GHS alleviates the problem of tuning the parameter bw and can work efficiently on both continuous and discrete problems. The GHS algorithm, due to its easy implementation and quick convergence, has been applied to many fields. Xiang et al. [67] proposed a discrete GHS (DGHS) for solving 0–1 KP. Keshtegar et al. [68] proposed a Gaussian GHS (GGHS) algorithm for solving numerical optimization problems. EI-Abd et al. [69] proposed an improved GHS (IGHS) algorithm to solve continuous optimization. In essence, musical performances seek to find pleasing harmony, just like the process of finding the global optimal solution through continuous learning. Based on this, GHS, as another learning scheme, is integrated into the MS for a global search.
Aiming at resolving the above issues, we propose a hybrid learning MS (HLMS), by introducing the Baldwinian learning and GHS learning mechanism. In HLMS, a novel Baldwin learning strategy based on Cauchy distribution is proposed instead of Gaussian distribution in [70]. During the evolutionary process, each individual in the whole population performs GHS learning and Baldwinian learning successively with a certain probability to reduce the time-consumption. The beneficial combination and complementarity of these two mechanisms lead HLMS to evolve towards the global optimum. Intuitively, HLMS has better search performance than the original MS because of the good balance between the exploration capacity of GHS learning and the exploitation ability of the Baldwinian learning.
The novelty and main contributions of this work include.
  • A novel Baldwinian learning strategy based on Cauchy distribution is proposed, the analysis and experiment of which show this strategy is more effective than the Baldwin learning strategy based on Gaussian distribution.
  • Combined with Baldwin learning, GHS is an effective global search operator to enhance the exploration ability of HLMS. Meanwhile, the pitch adjustment process based on dimensional learning can generate a large jump in the search process.
  • To reduce computation costs, the proposed HLMS triggers Baldwin learning and GHS learning with a certain probability in each iteration.
  • Exploration and exploitation are two common and fundamental features of any optimization method. In the evolution of HLMS, Lévy flight and GHS learning are mainly responsible for exploration, whilst flight straightly and Baldwinian mainly implement exploitation.
The rest of this paper is organized as follows. Section 2 reviews the MS algorithm, the Baldwinian learning scheme, and the GHS algorithm. In Section 3, the proposed HLMS for the MKP is introduced in detail. Extensive experiments and comparisons are conducted in Section 4. Finally, conclusions and suggestions are provided in Section 5.

2. Preliminaries

In this section, we summarize the core idea of MS, Baldwinian learning, and Global-best HS algorithm, which form the basis of the proposed HLMS framework.

2.1. Moth Search Algorithm

The moth search algorithm (MS) [44] is a new metaheuristic algorithm developed by Wang and inspired by the phototaxis and Lévy flights of moths. Based on these two behaviors, the flight straightly operator and Lévy flight operator of MS are derived, which can achieve good balance between the exploration capability and exploitation ability. Meanwhile, the whole population is subdivided into two subpopulations (named subpopulation1 and subpopulation2) based on the fitness of moth individuals. To this end, the position of offspring in subpopulation1 and subpopulation2 is updated by Lévy flight and flight straightly, respectively.
In the Lévy flight stage, the core mathematical formulation is:
X i t + 1 = X i t + α L ( s )
α = S m a x / t 2
L ( s ) = ( β 1 ) Γ ( β 1 ) sin ( π ( β 1 ) 2 ) π s β
where Xit and Xit+1 denote the position vector of the ith moth at generation t and t + 1, respectively.  α is the scale factor and Smax is the max walk step. L(s) represents the step drawn from Lévy distribition with  β = 1.5 and  Γ ( x ) is the gamma function.
In the flight straightly stage, the ith individual in subpopulation2 is considered to fly in a straight line towards the light source. The mathematical model of the flight straightly operator is formulated as follows:
X i t + 1 = { λ × ( X i t + φ × ( X b e s t t X i t ) ) i f r a n d > 0.5 λ × ( X i t + 1 φ × ( X b e s t t X i t ) ) e l s e
where  λ is the scale factor, which is used to control the convergence speed of the algorithm and improve population diversity.  λ is set to a random number drawn by the standard uniform distribution. The acceleration factor  φ is set to golden ratio (0.618).  X b e s t t is the best moth individual at generation t. rand returns a random number that is uniformly distributed in (0, 1).
The pseudo code of MS is shown in Algorithm 1.
Algorithm 1. Moth search algorithm
Begin
Step 1: Initialization.
 Set the maximum iteration number MaxGen and iteration counter G = 1; Initialize the parameters max walk step Smax, the index  β , and acceleration factor  φ .
 According to uniform distribution, the population with NP individuals is randomly initialized.
Step 2: Fitness calculation.
 Compute the initial objective function values of each individual according to their position. Memory the best individual (denotes as Xbest).
Step 3: While (G < MaxGen) do
 Divide the whole population into two subpopulations with equal size: subpopulation1 and subpopulation2, based on their fitness.
 Update subpopulation1 by using Lévy flight operator (Equation (4)).
 Update subpopulation2 by using flight straightly operator (Equation (7)).
 Evaluate the objective function values of each individual and update Xbest.
G = G + 1.
 Sort the population by fitness.
Step 4: End while
Step 5: Output: the best results.
End.

2.2. Baldwin Effect and Baldwinian Learning

The interactive way of learning and evolution was first proposed by Baldwin, known as the Baldwin effect. Hence, different Baldwinian learning models have been proposed based on the Baldwin effect. Hinton and Nowlan [62] found that it was difficult to find the optimal solution of more complex problems only by an evolutionary algorithm. However, when combined with Baldwinian learning, the performance of the hybrid algorithm can be effectively improved.
Generally speaking, Baldwinian learning is a type of local search strategy in an evolutionary algorithm. The Baldwinian learning mechanism was first combined with the clonal selection algorithm (CSA) by Gong et al. [61] to improve the performance of BCSA. Based on this, Peng et al. [70] proposed four Baldwinian learning strategies inspired by the trial vector generating strategy of differential evolution (DE).
In evolutionary computation, Cauchy mutation and Gaussian mutation are two popular and effective mutation techniques [71]. The characteristic of Gaussian mutation is to speed up the local convergence, and Cauchy mutation is better at escaping from local optimum. However, compared with Gaussian mutation, Cauchy mutation is insensitive to mutation step size and can achieve the acceptable performance.
According to the above analysis, the newly designed Baldwinian learning operator based on Cauchy distribution is proposed in HLMS. The mathematical expression is as follows:
Y i = X r 1 + c ( X r 2 X r 3 )
where  Y i is the donor vector for each moth individual  X i from the current population after applying Baldwinian learning.  X r 1 X r 2 , and  X r 3 are sampled randomly from the current population and r1, r2, and r3 are mutually exclusive integers randomly chosen from the range [1, NP], which are also different from the selected individual index i. The parameter c is the strength of Baldwinian learning and is a random number based on Cauchy distribution.

2.3. Global-Best HS Algorithm

GHS is a novel variant of HS and inspired by the concept of swarm intelligence of PSO [66]. The difference from the original HS is that the new harmony can mimic the best harmony in the harmony memory HM. Meanwhile, the parameter bw in HS is replaced and a social dimension is added to the GHS. In addition, the GHS dynamically updates the pitch adjusting rate PAR according to the following equation [72]:
P A R ( t ) = P A R m i n + P A R m a x P A R m i n N I × t
where PAR(t) is the pitch adjusting rate for generation t,  P A R m i n is the minimum adjusting rate, and  P A R m a x is the maximum adjusting rate. NI is the number of improvisations and t is the generation number.
The procedure of GHS is given in Algorithm 2.
Algorithm 2. The Global-best HS algorithm (GHS)
Begin
For each  i [ 1 , n ]  do     /*n is the dimension of the problem*/
If U(0, 1) ≤ HMCR then   /* memory consideration*/
x i h = x i j , where  j U ( 1 , , H M S )
If U(0,1) ≤ PAR (t) then /*pitch adjustment*/
x i h = x k b e s t , where best is the best harmony in the HM and  k U ( 1 , n )
Else         /*random selection*/
x i h = L B i + r a n d * ( U B i L B i )
End.
Furthermore, it should be emphasized that, the concept of dimensional learning [73] is embodied in Algorithm 2, as shown below:
X i h = X k b e s t
In Equation (7) of MS, the same dimension j is selected in  X b e s t t X i t for conducting the new solution. Under this dimension, if the component value of the ith individual is similar to the best individual, the difference  X b e s t t X i t will be very small, especially in the later stage of evolution. This means that such a step size is not conductive to  X i jumping to a far position. If the best moth individual is local optimum, the solution hardly escapes from the local extremum. In Equation (10), the dimension index i of  X h is not equal to the dimension index k of  X b e s t . Generally, the difference between two different dimensions is large. Different dimensions can carry different information.
Based on the above analysis, dimension learning is embedding into Algorithm 2 and it should be an effective global search operator.

3. The Proposed HLMS for the MKP

The proposed HLMS algorithm for MKP is inspired from the studies [66,70] and distinguishes itself with two new features. First, GHS as a powerful global search operator is introduced to enhance the exploration ability of the algorithm. Second, a new Baldwinian learning strategy by replacing Gaussian distribution with Cauchy distribution is introduced on HLMS. The proposed algorithm framework and the main components of the MKP problem are described in the following subsection.

3.1. Population Initialization

In this stage, NP moth individuals are randomly generated in the search space. The swarm X = {X(1), X(2), …, X(NP)} is maintained and evolves, where each moth individual X(i) is a n-dimensional real-valued vector  X ( i ) = ( x 1 i , x 2 i , , x n i ) with  x j i { a , a } j { 1 , 2 , , n } and n is the number of objects or items. Here, a takes the value 3 or 5 in this paper. Then, each moth individual X(i) is transformed into an n-dimensional binary vector by a mapping method, which is called discrete moth  Y ( i ) = ( y 1 i , y 2 i , , y n i ) with  y j i { 0 , 1 } j { 1 , 2 , , n } .

3.2. Solution Representation

In HLMS, an n-bit binary string consisting of 0 and 1 is used to represent a candidate solution. If the item is selected, the bit is 1, otherwise it is 0. It should be noted that the MKP is a constrained optimization problem, so the solution generated in the evolution process may be infeasible.
In this paper, a simple and effective transfer function [50] is adopted and the function expression is as follows:
T ( x ) = x
The transfer method from a real-valued variable  x i to a binary variable  y i is calculated by:
y i = { 1 , i f T ( x i ) > 0 0 , else

3.3. Quick Repair Operator

Learning from previous research work [10,74], the HLMS algorithm also adopts a popular quick repair operator based on pseudo-utility which was proposed by Luo et al. [15]. In order to effectively apply the repair operator, the given MKP instance needs to be preprocessed. Specifically, all the items are renumbered in an ascending order based on their scaled pseudo-utility ratios  σ j [75] defined as follows:
σ j = c j i = 1 m a i j b i , j { 1 , 2 , , n }
More exactly, the index values of all sorted items are stored in array J [1…n], such that  σ J [ 1 ] σ J [ 2 ] σ J [ n ] . The vectors  ( c 1 , c 2 , , c n ) and  a i , j (i = 1, 2, …, m, j = 1, 2, …, n) should be adjusted as well. The main framework of the quick repair procedure can be summarized in Algorithm 3.
Algorithm 3. The quick repair operator base on the scaled pseudo-utility
Begin
Step 1: Input.  X = { x 1 , x 2 , , x n } { 0 , 1 } n .
Step 2: Calculation. Calculate the current total consumption of each resource.
r i = j = 1 n a i j * x j , i = 1 , 2 , , m .
Step 3: Repair process.
For j = n to 1 do
If  r i b i , i = 1 , 2 , , m then
 Break.
else if  x J [ j ] = 1 then
 Set  x J [ j ] = 0, and  r i r i a i J [ j ] , i = 1 , 2 , , m .
End
End
Step 4: Optimization process.
For j = 1 to n do
If  x J [ j ] = 0 and  r i + a i J [ j ] b i , i = 1 , 2 , , m  then
 Set  x J [ j ] = 1 and  r i r i + a i J [ j ] , i = 1 , 2 , , m .
End
End
Step 5: Output:  X = { x 1 , x 2 , , x n } { 0 , 1 } n .
End.
Obviously, the quick repair operator of Algorithm 3 mainly consists of two phases. In the first phase, called the repair process, according to the ascending order of the scaled pseudo-utility ratios, the items are removed from the knapsack one by one until the solution is feasible. In the second phase, called the optimization process, for all the feasible solutions, greedily packed the items to be loaded into the knapsack based on the descending order of the scaled pseudo-utility ratios one by one. In this process, the feasibility of the solution needs to be maintained all the time. In brief, the first phase makes all the infeasible solutions become feasible, and the second phase enables the quality of feasible solutions better.

3.4. Procedure of HLMS for MKP

Based on the analysis above, the proposed HLMS algorithm for MKP is outlined in Algorithm 4. The algorithm framework includes the following main steps. (1) After initialization, the repaired population is divided into two subpopulations based on the fitness. (2) Subpopulation1 and subpopulation2 apply the Lévy flight operator and flight straightly operator, respectively. (3) GHS learning and Baldwinian learning are implemented in sequence to the whole population with a probability of 0.5. (4) The mapping from the real-valued vector to the binary vector is realized with a transfer function and then the repair of infeasible solutions and the optimization of feasible solutions are performed. (5) Evaluating the solution is based on the objective function and then the whole population is divided into two subpopulations. Steps (2)(5) are repeated until the termination condition is reached.
Algorithm 4. Procedure of HLMS for MKP
Begin
  •  Step 1: Initialization.
  • Set the maximum iteration number MaxGen and iteration counter G = 1; Initialize the parameters max walk step Smax, the index  β φ , strength of Baldwinian learning c.
  • According to uniform distribution, the population with NP individuals is randomly initialized.
  • The transform function is used to discretize the real number vector to obtain the initial solution  X = { x 1 , x 2 , , x n } { 0 , 1 } n .
  • Repair the initial solution by Algorithm 3.
  • Step 2: Fitness evaluation.
  • Evaluate the initial solution using the objective function of MKP.
  • Step 3: Whileg < MaxGen do
 3.1 Divide the whole population into two subpopulations with equal size: subpopulation1 and subpopulation2, based on their fitness.
    3.2 Update subpopulation1 by Lévy flight operator.
    3.3 Update subpopulation2 by flight straightly operator.
    3.4 GHS search
     If U(0, 1) ≤ 0.5
  Apply GHS algorithm on each individual X and generate the trial individual Y.
       Choose the best one of X and Y to enter the next generation.
    3.5 Baldwinian Learning
     If U(0, 1) ≤ 0.5
  Apply Baldwinin learning strategy on each individual X and generate the trial individual Y.
       Choose the best one of X and Y to enter the next generation.
    3.6 Apply transform function to obtain the potential solution of MKP.
    3.7 Repair the potential solution by Algorithm 3.
    3.8 Evaluate the fitness of the population and record the global best fitness.
G = G + 1.
    3.9 Sort the population by fitness.
   Step 4: End while
   Step 5: Output: the best results.
End.
The algorithm framework is shown in Figure 1.
Figure 1. The framework of HLMS for MKP.

3.5. Computational Complexity of One Iteration of HLMS

The computational complexity of one iteration of HLMS based on Algorithm 4 is described as follows.
(1)
The initialization of HLMS requires O(NP × n) time, where NP denotes the population size, and n is the dimension of MKP (the number of the items).
(2)
The discretization process of NP moth individual costs O(NP × n) time.
(3)
The quick repair operator takes O(n × m) time, where m is the constraints of the MKP instance.
(4)
Fitness evaluation has O(NP) time.
(5)
Lévy flight operator has O(NP1 × n) time, where NP1 is the number of individuals of subpopulation1.
(6)
Flight straightly operator has O(NP2 × n) time, where NP2 is the number of individuals of subpopulation2.
(7)
GHS learning requires O(NP × n) time.
(8)
Baldwinian learning requires O(NP × n) time.
(9)
Sort the population based on Quick sort algorithm and it takes time O(NPlogNP).
In summary, the total computational complexity is O(NP × n) per generation for fixed m.

4. Experimental Studies

To comprehensively evaluate the performance of the proposed HLMS, large numbers of experiments are implemented on the benchmark instances commonly used in the literature and a comparative study is conducted between HLMS and several populations based on optimization algorithms.

4.1. Benchmark Test Functions

In this paper, four sets of well-known benchmark data for MKP are used to test the performance of HLMS. These instances are described in [74] and available at OR-Library (http://people.brunel.ac.uk/~mastjjb/jeb/orlib/mknapinfo.html, accessed on 1 July 2021).
Test set I contains 18 small-scale instances with m = 2 to 30 and n = 20 to 105, which are denoted as Sento [76], HP [77], PB [77], and Weing [4].
Test set II contains 30 medium-scale instances with m = 5 and n = 30 to 90, which are marked as Weish [7].
Test set III contains 30 large-scale instances. These instances are divided into two subsets. Subset I includes 15 instances with m = 15 and n  {100, 250, 500}, which are labeled as cb1, cb2, and cb3, respectively. Subset II also includes 15 instances with m = 15 and n  {100, 250, 500}, which are called as cb4, cb5, and cb6, respectively.
Test set IV contains 9 instances with m {15, 25, 50} and n  {100, 200, 500, 1000, 1500}, which were created by Glover and Kochenberger and then are marked as GK.

4.2. Experimental Environment and Parameters Setting

The proposed HLMS algorithm includes several important parameters, whose values are empirically set based on the preliminary experiments and the details are recorded in Table 1.
Table 1. Settings of parameters of HLMS.
The HLMS algorithm was implemented in C language and compiled using the GNU GCC compiler. All the experiments were carried out on a computer with Intel (R) Core (TM) i7-7500 CPU (2.90 GHz and 8.00 GB RAM), running the Windows 10 operating system. HLMS was independently run 30 times for each instance to eliminate the unfairness brought by the stochastic characteristic.
To comprehensively evaluate the performance of the HLMS algorithm, fourteen MKP algorithms in the literature are selected as our comparative algorithms, which are listed as follows.
  • Modified binary particle swarm optimization (MBPSO) [78];
  • Chaotic binary particle swarm optimization with time-varying acceleration coefficients (CBPSOTVAC) [79];
  • Binary PSO with time-varying acceleration coefficients (BPSOTVAC) [79];
  • Modified multi-verse optimization (MMVO) algorithm [17];
  • New binary particle swarm optimization with immunity-clonal algorithm (NPSOCLA) [80];
  • Binary gravitational search algorithm (BGSA) [81];
  • Binary hybrid topology particle swarm optimization (BHTPSO) [81];
  • Binary hybrid topology particle swarm optimization quadratic interpolation (BHTPSO-QI) [81];
  • New binary particle swarm optimization (NBPSO) [81];
  • Binary version of PSO (BPSO) [82];
  • Binary version of the Harris hawks algorithm (BHHA) [22,83];
  • Binary version of the salp swarm algorithm (BSSA) [84];
  • Binary version of the modified whale optimization algorithm (BIWOA) [85];
  • Binary version of the sin-cosine algorithm (BSCA) [86].
It should be noted that the results of the comparative algorithms are compiled from the related papers. If the result of an algorithm for a MKP instance is not available, the result of the instance is ignored. In addition, considering that different comparative algorithms are written in different programming languages, or run on different computing platforms based on different termination conditions and algorithm parameters, we focus on comparing solution quality.
For this purpose, in this paper, eight typical statistical evaluation criteria are selected to evaluate the performance of all the comparative algorithms.
  • Best value (Best):
B e s t = max ( f i ) , for i [ 1 , t ]
where  f i is the fitness value for  i t h time. t is the total number of independent experiments.
  • Worst value (Worst):
W o r s t = min ( f i ) , for i [ 1 , t ]
  • Mean value (Mean):
M e a n = 1 t i = 1 t f i
The mean value characterizes the centralized trend of the values of random variables. The larger the mean value, the more concentrated the results of multiple runs of the algorithm will be.
  • Standard deviation (Std):
S t d = 1 t i = 1 t ( f i m e a n )
Standard deviation describes the degree of dispersion of random variable values relative to the mean value. Meanwhile, standard deviation reflects the fluctuation in the value of a random variable. In other words, stability is an important evaluation criterion of a stochastic algorithm. If the Std value is high, the stability of the algorithm is poor, otherwise, its performance is good.
  • Success rate (SR):
S R = s t t
where st denotes the success times, that is, the number of the known theoretical optimal solution is obtained. The high success rate indicates that the algorithm has good stability and optimization performance.
  • Percent deviation (PDev):
P D e v = O p t M e a n O p t 100
where Opt represents the optimal or the best-known solution. PDev reflects the degree to which the mean value deviates from the known theoretical optimal solution when the algorithm solves a single instance.
  • Average error (AE):
A E = 1 N i = 1 N ( O p t p r o f i t ) O p t 100
Average error is an indicator that reflects the general level of error between random variables and Opt. Here, profit can be Best, Worst, or Mean. N is the number of benchmark instances. Clearly, the smaller AE value indicates that the algorithm has better performance. AE indicates the overall performance of the algorithm for solving a set of MKP instances.
  • Percentage gap (Gap):
G a p = O p t B e s t O p t 100
Similar to AE, for the maximization problem, the smaller the Gap is, the better the performance of the algorithm is. Gap investigates the performance of the algorithm to solve a single MKP instance.
Moreover, to determine whether there are significance differences between HLMS and other algorithms, the p-value based on the nonparametric Wilcoxon signed ranks test at the 95% confidence level is reported as well. Note that a p-value less than 0.05 represents that there exists a significant difference between the paired compared results. All the statistical results have been performed by the statistical software R language.

4.3. Comparisons on the Small-Scale Test Set

The proposed HLMS is first substantiated based on the 18 small-scale instances of Test set I and the experimental result is listed in Table 2, along with the available results of the comparative algorithms. In Table 2, the first three columns record the names of instances, the dimension information (n is the number of items, and m is the number of knapsacks), and the known optimum results (Opt), respectively. In addition, aggregate data are recorded at the bottom of the table. #Opt shows the number of the best-known solution obtained by the corresponding algorithm. #SR and #Std represent the number of instances for which the corresponding algorithm obtained a better result in terms of SR and Std among the comparative algorithms. MSR is the mean value of success rate and the ranks in descending order on the MSR are provided. Besides, the best findings among the comparison results are indicated in bold.
Table 2. The results of HLMS with 3 comparative algorithms for Test set I.
From Table 2, the proposed HLMS is able to obtain the known optimum solution for almost all the 18 instances except for Weing7. However, MS can only reach the known optimum solution for 13 instances. Considering the #SR, HLMS performs much better than the competing algorithms. In terms of MSR, HLMS is slightly worse than CBPSOTVAC and ranks two. Moreover, the clear superiority of HLMS is established in comparison with MS in terms of all evaluation criteria. Therefore, we can conclude that it is beneficial to use the GHS global search algorithm combined with the Baldwinian learning strategy. In terms of the p-value of SR, the difference between HLMS and MS, HLMS and MBPSO is statistically significant (p-value < 0.05). However, there are no significant differences in Std among the latter two groups (p-value > 0.05).
The related box plots are given in Figure 2 in terms of SR. As can be seen from Figure 2, the difference of SR among four algorithms is very obvious. Although the distributions of the SR value of MS and MBPSO are more uniform than that of HLMS and CBPSOTVAC, the interquartile ranges of the former are worse than that of the latter. Moreover, outliers exist in MS and MBPSO on the SR value. In addition, we also observe the maximum, upper quartile, the mean value of HLMS is equal to or close to 1.0, 0.8, and 0.6, respectively.
Figure 2. Boxplot for 4 comparative algorithms on SR for Test set I.
Based on the above analysis, we can draw a conclusion that HLMS can obtain the best-known solution of 18 small-scale instances with a high success rate.

4.4. Comparisons on the Medium-Scale Test Set

In the second experiment, HLMS is used to solve medium-scale test instance (Test set II) to verify the performance of algorithms. The results are reported in Table 3, together with the results of other six state-of-the-art MKP algorithms, including MS, BIWOA, BMMVO, BSCA, BHHA, and BSSA. Note that these six algorithms are all novel swarm intelligence algorithms proposed in recent years and it is meaningful to select them for comparative study of MKP.
Table 3. The results of HLMS with 6 comparative algorithms for Test set II.
From Table 3, the results demonstrate that HLMS reaches the optimum solutions on 27 out of 30 instances, while the six comparative algorithms obtain the optimum solutions only on 20, 16, 14, 18, 19, and 19, respectively. In terms of #Worst, BSSA outperforms the other six algorithms on 23 instances. PDev measures the deviation between the mean and the best-known solution. The small mean value of PDev also confirms the superiority of HLMS. Comprehensively speaking, MS has the worst performance among all the comparative algorithms. Moreover, there are significant differences (p-value < 0.05) between the comparisons of the first four groups concerning PDev.
Figure 3 presents the box plots of the PDev values for all the comparative algorithms. The span of each box implicitly reflects the stability of the algorithm. The smaller the span is, the better the stability of the algorithm is. As can be seen from Figure 3, HLMS has a significant advantage over the other six algorithms since the span of the box for HLMS is obviously smaller than that of the other comparative algorithms. It should be noted that the dot in Figure 3. represent outliers.
Figure 3. Boxplot for 7 comparative algorithms on PDev for Test set II.
In summary, the results in Table 3 and Figure 3 indicate that our HLMS algorithm is very competitive compared to the other six MKP algorithms. The findings are based on the fact that the GHS learning scheme enhances the global search ability of HLMS. On this basis, Baldwinian learning can effectively adjust the shape of search space and thereby provides good search paths towards the best solutions.

4.5. Comparisons on the Large-Scale Test Set

In the third experiment, the performance of HLMS is verified by solving large-scale problems, and the comparative results on Test set III and Test set IV are reported in Table 4 and Table 5. In order to make a fair comparison with different classical algorithms using appropriate evaluation criteria, the experiment is divided into three groups on different scale instances.
Table 4. The results of HLMS with 5 comparative algorithms for Test set III (cb1-cb3).
Table 5. The results of HLMS with 5 comparative algorithms for Test set III (cb4–cb5).

4.5.1. Performance Comparison on Test Set III (cb1–cb3)

Table 4 summarizes the experimental results of the first group large-scale benchmarks. From Table 4, it can be seen clearly that the proposed HLMS still keeps the best performance in terms of all six evaluation criteria. Specifically speaking, HLMS outperforms the other comparative algorithms. In addition, the p-value indicates that there is significant difference between HLMS and MS, BGSA, BHTPSO, and BHTPSO-QI in terms of mean. However, the p-value is 0.201 for MMVO (p-value > 0.05) and then reject the null hypothesis. Hence, insignificant difference can be detected between HLMS and MMVO.
The performance comparison of six methods based on AE is plotted in Figure 4. It is evident that the axis of HLMS on radar charts has a point nearer to the center in comparison with the other five algorithms when considering AE of Best, AE of Mean, and AE of Worst, which indicates that it is more effective with respect to quality of solutions. It can be considered that HLMS obtained optimal or near-optimal for most of the instances in terms of Best, Mean, and Worst, and could beat all the other competing algorithms.
Figure 4. The performance comparison of 6 methods based on AE for Test set III (cb1–cb3).
To observe the stability of MS and HLMS more intuitively, the error bar based on variance and the trends plot based on Std are shown in Figure 5 and Figure 6, respectively. It can be observed clearly from Figure 5 that the variances of HLMS are apparently smaller than that of MS for all benchmarks. Moreover, the variances will increase with the growth of the scale of instances. It is clear from Figure 6 that the trend lines of HLMS are located in the relatively low area for ten instances of cb1 and cb2. However, the curve of cb3 has an upward trend. In brief, the curve of MS is higher than that of HLMS, which indicates that HLMS has more stable performance than MS.
Figure 5. The error bars (Variance) for MS and HLMS on Test set III (cb1–cb3).
Figure 6. The trends of Std for MS and HLMS on Test set III (cb1–cb3).

4.5.2. Performance Comparison on Test Set III (cb4–cb6)

Table 5 summarizes the experimental results of the second group large-scale benchmarks. Table 5 shows that HLMS is also very efficient for 15 large instances with m = 10. Moreover, HLMS is superior to other five algorithms in absolute advantage, which is confirmed by the small p-values (0.000 ≤ 0.05).
Similarly, radar charts are plotted to visualize three evaluation criteria, AE of Best, AE of Mean, and AE of Worst in Figure 7. From Figure 7, the phenomenon is almost consistent with Figure 4. The point on the HLMS axis is very close to the center point. By implication, HLMS has smaller AE value compared with other algorithms.
Figure 7. The performance comparison of 5 methods based on AE for Test set III (cb4–cb6).
The error bar based on variance for HLMS and MS is illustrated in Figure 8, which is to assess the stability of algorithms. As can be seen from Figure 8, the variance of HLMS is almost unaffected by the scale of MKP. However, with the expansion of the scale, the variance of MS is increasing gradually.
Figure 8. Error bars (Variance) for MS and HLMS on Test set III (cb4–cb6).
The trend plot of Std for HLMS and MS is given in Figure 9. It can be seen from Figure 9 that the trend curve of MS is significantly higher than that of HLMS, which further indicates that HLMS has better stability than MS.
Figure 9. The trends of Std for MS and HLMS on Test set III (cb4–cb6).

4.5.3. Performance Comparison on Test Set IV

Table 6 summarizes the experimental results of the third group large-scale benchmarks. Overall, HLMS still outperforms all other comparative algorithms. In terms of #Best, #Worst, and #Mean, HLMS obtains a better result respectively on 5, 1, and 3 out of 9 instances. The results of BIWOA are respectively on 2, 2, and 2 out of 9 instances. BSSA with better performance obtains 2, 6, and 5 out of 9 instances, respectively. For the significance, the p-values for BMMVO and BSCA are both smaller than 0.05 concerned with Mean, except for MS, BIWOA, BHHA, and BSSA, which indicates that the difference between HLMS and most comparative algorithms is not significant when facing Test set IV.
Table 6. The results of HLMS with 6 comparative algorithms for Test set IV.
Graphically, Figure 10 shows the AE of Best, AE of Mean, and AE of Worst obtained by seven methods. In terms of Best, HLMS outperforms the other comparative algorithms in absolute small AE value. However, HLMS is slightly worse than BIWOA in terms of AE of Mean.
Figure 10. The performance comparison of 7 methods based on AE for Test set IV.
In summary, the above experimental results and comparisons show that the proposed HLMS also has excellent optimization performance in solving large-scale MKP instances, in terms of solution quality, convergence, and stability of algorithms. This is mainly due to GHS learning and Baldwinian learning being able to effectively balance exploration and exploitation in the evolution process. Compared with the original MS, the hybrid learning strategy focuses more on discovering and utilizing useful information from the whole population and whole search experience, rather than the experience of some random local individuals.

4.6. Sensitivity Analysis on the Positional Parameter and the Scale Parameter

As is known to all, Gaussian distribution and Cauchy distribution are two important distributions, which have been integrated into many algorithms and proved an effective strategy to enhance the ability of an elaborate search. The strength of Baldwinian learning in [47] is random real number obeying Gaussian distribution. However, previous studies have revealed that Cauchy mutation possesses more power in escaping local optima and converging to the global optimum. Hence, the parameters used in [47] for Gaussian distribution and the positional parameter x and the scale parameter y for Cauchy distribution are investigated in this subsection. It is noted that, to eliminate the influence of GHS, HLMS only adopted the Baldwinian learning strategy in this experiment. We tested HLMS for Gaussian distribution and different combinations for (x, y) of Cauchy distribution: (0, 1), (0, 0.5), (0, 2), and (−2, 1). Table 7 summaries the Mean and SR for 18 instances of Test set I. In addition, boxplot for five parameter combinations on SR is plotted in Figure 11.
Table 7. The results of HLMS using different parameters of two distributions for Test set I.
Figure 11. Boxplot for 5 parameter combinations on SR for Test set I.
From Table 7, all HLMS with four different combinations of x and y find better results than that of HLMS combined with Gaussian distribution. In terms of #Mean, #SR, and MSR, HLMS with four groups (x, y) shows similar results. It can be observed from Figure 11 that HLMS-C2 shows excellent comprehensive performance. HLMS-C2 has the best maximum and three-quarter quantile. Hence, considering all of the parameter combinations, we concluded that the setting x = 0 and y = 0.5 for the HLMS is an appropriate choice.
We can draw a conclusion from this experiment that it is better to use a random real obeying Cauchy distribution than Gaussian distribution as the Baldwinian learning strength. The reason may be that Cauchy mutation has stronger ability to jump from local optimum than Gaussian mutation.

4.7. The Effectiveness of the Two Components in HLMS

As mentioned above, HLMS includes two learning strategies: Baldwinian learning strategy and GHS learning strategy. The aim of this subsection is to investigate the effectiveness of these two learning strategies. Therefore, one additional experiment is conducted on Test set I and the results are summarized in Table 8. HLMS, which only adopted the Baldwinian learning strategy, is denoted as HLMS-B. HLMS, which only adopted the GHS learning strategy, is denoted as HLMS-H.
Table 8. Comparisons of MS, HLMS-B, HLMS-H, and HLMS on Test set I.
As can be seen from Table 8, compared with HLMS-B, HLMS-H, and HLMS, MS shows the worst performance in terms of #Mean, #SR, and MSR. The results further reveal that two learning strategies are effective in the search process. Additionally, it is noted that HLMS-B shows similar performance with MS while the performance difference between HLMS-H and MS is significant. However, the performance of HLMS integrated with two learning strategies is obviously better than that of any one.
Moreover, the convergence graphs of the average objective function values obtained by four algorithms are plotted in Figure 12 and Figure 13 for four representative instances: Sento1, Sento2, Weing7, and Weing8. As seen from Figure 12 and Figure 13, MS has the slowest convergence speed, while HLMS-H and HLMS have similar convergence speed and converge both faster than MS and HLMS-B.
Figure 12. Convergence graphs of MS, HLMS-B, HLMS-H, and HLMS on Sento1 and Sento2.
Figure 13. Convergence graphs of MS, HLMS-B, HLMS-H, and HLMS on Weing7 and Weing8.
In summary, we can draw a conclusion that the above two learning strategies can realize complementary advantages to enhance the performance of MS. Indeed, only using Baldwinian learning is not sufficient for exploitation. The conclusion is that the beneficial combination of the two strategies is significant for improving the performance of the algorithm.

4.8. Discussion

Comparison results demonstrate that the Baldwinian learning and GHS learning strategies can really improve the performance of HLMS, thereby making it better than the original MS and most other comparative algorithms on the majority of MKP instances, in terms of solution accuracy, convergence speed, and algorithm stability. The advantage of HLMS mainly lies in that GHS is used to guide the global search and dimensional learning can achieve a large jump to help solutions escape local extremum. The other reason is that Baldwinian learning as a local search strategy has the effect of changing the fitness landscape. This interaction between learning and evolution is very beneficial. Accordingly, both Baldwinian learning and GHS learning are more efficient and effective than MS alone.
In fact, based on the previous experimental results, we can find that considering small-scale MKP instances, medium-sized MKP instances, and large-scale MKP instances, the HLMS algorithm combining Baldwinian learning and GHS learning is an effective algorithm for solving MKP problems. Besides that, balancing exploration and exploitation is an important factor in metaheuristic algorithms by maintaining adequate diversity in swarm individuals so that reducing the probability of trapping in local optimal locations. In HLMS, GHS learning and Baldwinian learning respectively play the roles of exploration and exploitation, making the optimization performance of the algorithm better.

5. Conclusions and Future Work

This paper proposed a hybrid learning moth search algorithm (HLMS) inspired by the idea that the learning strategy could direct the evolutionary process. The framework proposed in this work includes two learning strategies: Baldwinian learning and GHS learning. In the search process, the two learning strategies play the role of local exploitation and global exploration, respectively.
HLMS is verified by solving the NP-hard 0–1 multidimensional knapsack problems. The experimental results on the 87 instances commonly used in literature showed that HLMS performs competitively in comparison with MS and other state-of-the-art meta-heuristics algorithms. Sensitivity analysis of Gaussian distribution and Cauchy distribution on Baldwinian learning is provided. The results proved that Cauchy mutation is more effective than Gaussian mutation as learning length. The effectiveness of two important learning strategies of HLMS is investigated. The results demonstrated that Baldwinian learning and GHS learning both play a major role in improving the performance of HLMS. MS with two learning strategies surpasses MS and MS with a single strategy. It confirms the effectiveness of our proposed learning strategies.
Future research on MS can be divided into two main directions: research on more real-world applications and research on improvements of the algorithms. Concerning the application of MS, it has the potential to solve more combinatorial optimization problems, such as maximum diversity problems (MDP), multi-objective knapsack problems (MOKP), and multi-demand multidimensional knapsack problems (MDMKP). In terms of the improvements of algorithms, more effective learning based on strategies can be adopted to enhance the search ability of the algorithm, such as orthogonal learning, reinforcement learning, and adaptive learning.

Author Contributions

Y.F. performed the methodology, investigation and wrote the draft. H.W. and Z.C. supervised the research and edited the final draft. X.L. performed the experiments. M.L. reviewed the final draft. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (No. 61806069), Key projects of science and technology research in Colleges of Hebei Province, (No. ZD2022083), Key R & D plan project of Hebei Province (No. 22375415D), Major Projects of Guangdong Education Department for Foundation Research and Applied Research (No. 2017KZDXM081, 2018KZDXM066), Guangdong Provincial University Innovation Team Project (No. 2020KCXTD045).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Toyoda, Y. A simplified algorithm for obtaining approximate solutions to zero-one programming problems. Manag. Sci. 1975, 21, 1417–1427. [Google Scholar] [CrossRef]
  2. Vasquez, M.; Vimont, Y. Improved results on the 0–1 multidimensional knapsack problem. Eur. J. Oper. Res. 2005, 165, 70–81. [Google Scholar] [CrossRef]
  3. Feng, Y.; Yang, J.; Wu, C.; Lu, M.; Zhao, X.-J. Solving 0-1 knapsack problems by chaotic monarch butterfly optimization algorithm with Gaussian mutation. Memetic Comput. 2018, 10, 135–150. [Google Scholar] [CrossRef]
  4. Weingartner, H.M.; Ness, D.N. Methods for the solution of the multidimensional 0/1 knapsack problem. Oper. Res. 1967, 15, 83–103. [Google Scholar] [CrossRef]
  5. Drake, J.H.; Özcan, E.; Burke, E.K. A case study of controlling crossover in a selection hyper-heuristic framework using the multidimensional knapsack problem. Evol. Comput. 2016, 24, 113–141. [Google Scholar] [CrossRef]
  6. Gilmore, P.; Gomory, R.E. The theory and computation of knapsack functions. Oper. Res. 1966, 14, 1045–1074. [Google Scholar] [CrossRef]
  7. Shih, W. A branch and bound method for the multiconstraint zero-one knapsack problem. J. Oper. Res. Soc. 1979, 30, 369–378. [Google Scholar] [CrossRef]
  8. Boussier, S.; Vasquez, M.; Vimont, Y.; Hanafi, S.; Michelon, P. A multi-level search strategy for the 0-1 Multidimensional Knapsack Problem. Discret. Appl. Math. 2010, 158, 97–109. [Google Scholar] [CrossRef]
  9. Lai, X.; Hao, J.-K.; Glover, F.; Lü, Z. A two-phase tabu-evolutionary algorithm for the 0-1 multidimensional knapsack problem. Inf. Sci. 2018, 436, 282–301. [Google Scholar] [CrossRef]
  10. Chih, M. Three pseudo-utility ratio-inspired particle swarm optimization with local search for multidimensional knapsack problem. Swarm Evol. Comput. 2018, 39, 279–296. [Google Scholar] [CrossRef]
  11. Haddar, B.; Khemakhem, M.; Hanafi, S.; Wilbaut, C. A hybrid quantum particle swarm optimization for the multidimensional knapsack problem. Eng. Appl. Artif. Intell. 2016, 55, 1–13. [Google Scholar] [CrossRef]
  12. Lai, X.; Hao, J.-K.; Fu, Z.-H.; Yue, D. Diversity-preserving quantum particle swarm optimization for the multidimensional knapsack problem. Expert Syst. Appl. 2020, 149, 113310. [Google Scholar] [CrossRef]
  13. Wang, L.; Wang, S.-y.; Xu, Y. An effective hybrid EDA-based algorithm for solving multidimensional knapsack problem. Expert Syst. Appl. 2012, 39, 5593–5599. [Google Scholar] [CrossRef]
  14. Li, Z.; Tang, L.; Liu, J. A memetic algorithm based on probability learning for solving the multidimensional knapsack problem. IEEE Trans. Cybern. 2020, 54, 2284–2299. [Google Scholar] [CrossRef] [PubMed]
  15. Luo, K.; Zhao, Q. A binary grey wolf optimizer for the multidimensional knapsack problem. Appl. Soft Comput. 2019, 83, 105645. [Google Scholar] [CrossRef]
  16. Zhang, B.; Pan, Q.K.; Zhang, X.L.; Duan, P.Y. An effective hybrid harmony search-based algorithm for solving multidimensional knapsack problems. Appl. Soft Comput. 2015, 29, 288–297. [Google Scholar] [CrossRef]
  17. Abdel-Basset, M.; El-Shahat, D.; Faris, H.; Mirjalili, S. A binary multi-verse optimizer for 0-1 multidimensional knapsack problems with application in interactive multimedia systems. Comput. Ind. Eng. 2019, 132, 187–206. [Google Scholar] [CrossRef]
  18. García, J.; Maureira, C. A KNN quantum cuckoo search algorithm applied to the multidimensional knapsack problem. Appl. Soft Comput. 2021, 102, 107077. [Google Scholar] [CrossRef]
  19. Bolaji, A.L.a.; Okwonu, F.Z.; Shola, P.B.; Balogun, B.S.; Adubisi, O.D. A modified binary pigeon-inspired algorithm for solving the multi-dimensional knapsack problem. J. Intell. Syst. 2021, 30, 90–103. [Google Scholar] [CrossRef]
  20. Feng, Y.; Wang, G.-G. A binary moth search algorithm based on self-learning for multidimensional knapsack problems. Future Gener. Comput. Syst. 2022, 126, 48–64. [Google Scholar] [CrossRef]
  21. Gupta, S.; Su, R.; Singh, S. Diversified sine-cosine algorithm based on differential evolution for multidimensional knapsack problem. Appl. Soft Comput. 2022, 130, 109682. [Google Scholar] [CrossRef]
  22. Abdel-Basset, M.; Mohamed, R.; Sallam, K.M.; Chakrabortty, R.K.; Ryan, M.J. BSMA: A novel metaheuristic algorithm for Multi-dimensional knapsack problems: Method and comprehensive analysis. Comput. Ind. Eng. 2021, 159, 107469. [Google Scholar] [CrossRef]
  23. Li, W.; Yao, X.; Zhang, T.; Wang, R.; Wang, L. Hierarchy ranking method for multimodal multi-objective optimization with local pareto fronts. IEEE Trans. Evol. Comput. 2022, 27, 98–110. [Google Scholar] [CrossRef]
  24. Cacchiani, V.; Iori, M.; Locatelli, A.; Martello, S. Knapsack problems-An overview of recent advances. Part II: Multiple, multidimensional, and quadratic knapsack problems. Comput. Oper. Res. 2022, 143, 105693. [Google Scholar] [CrossRef]
  25. Gao, D.; Wang, G.-G.; Pedrycz, W. Solving fuzzy job-shop scheduling problem using DE algorithm improved by a selection mechanism. IEEE Trans. Fuzzy Syst. 2020, 28, 3265–3275. [Google Scholar] [CrossRef]
  26. Wang, G.-G.; Gao, D.; Pedrycz, W. Solving multi-objective fuzzy job-shop scheduling problem by a hybrid adaptive differential evolution algorithm. IEEE Trans. Ind. Inform. 2022, 18, 8519–8528. [Google Scholar] [CrossRef]
  27. Shadkam, E.; Bijari, M. A novel improved cuckoo optimisation algorithm for engineering optimisation. Int. J. Artif. Intell. Soft Comput. 2020, 7, 164–177. [Google Scholar] [CrossRef]
  28. Parashar, S.; Senthilnath, J.; Yang, X.-S. A novel bat algorithm fuzzy classifier approach for classification problems. Int. J. Artif. Intell. Soft Comput. 2017, 6, 108–128. [Google Scholar] [CrossRef]
  29. Wang, G.-G.; Deb, S.; Gandomi, A.H.; Alavi, A.H. Opposition-based krill herd algorithm with cauchy mutation and position clamping. Neurocomputing 2016, 177, 147–157. [Google Scholar] [CrossRef]
  30. Wang, G.-G.; Guo, L.; Gandomi, A.H.; Hao, G.S.; Wang, H. Chaotic krill herd algorithm. Inf. Sci. 2014, 274, 17–34. [Google Scholar] [CrossRef]
  31. Li, W.; Wang, G.-G.; Alavi, A.H. Learning-based elephant herding optimization algorithm for solving numerical optimization problems. Knowl.-Based Syst. 2020, 195, 105675. [Google Scholar] [CrossRef]
  32. Wang, G.-G.; Deb, S.; Cui, Z. Monarch butterfly optimization. Neural Comput. Appl. 2019, 31, 1995–2014. [Google Scholar] [CrossRef]
  33. Feng, Y.; Deb, S.; Wang, G.-G.; Alavi, A.H. Monarch butterfly optimization: A comprehensive review. Expert Syst. Appl. 2021, 168, 114418. [Google Scholar] [CrossRef]
  34. Ma, L.; Cheng, S.; Shi, Y. Enhancing learning efficiency of brain storm optimization via orthogonal learning design. IEEE Trans. Syst. Man Cybern. Syst. 2020, 51, 6723–6742. [Google Scholar] [CrossRef]
  35. Yu, H.; Li, W.; Chen, C.; Liang, J.; Gui, W.; Wang, M.; Chen, H. Dynamic Gaussian bare-bones fruit fly optimizers with abandonment mechanism: Method and analysis. Eng. Comput. 2020, 38, 743–771. [Google Scholar] [CrossRef]
  36. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  37. Zhou, Y.; Bao, Z.; Luo, Q.; Zhang, S. A complex-valued encoding wind driven optimization for the 0-1 knapsack problem. Appl. Intell. 2017, 46, 684–702. [Google Scholar] [CrossRef]
  38. Song, J.; Chen, C.; Heidari, A.A.; Liu, J.; Yu, H.; Chen, H. Performance optimization of annealing salp swarm algorithm: Frameworks and applications for engineering design. J. Comput. Des. Eng. 2022, 9, 633–669. [Google Scholar] [CrossRef]
  39. Gezici, H.; Livatyalı, H. Chaotic Harris hawks optimization algorithm. J. Comput. Des. Eng. 2022, 9, 216–245. [Google Scholar] [CrossRef]
  40. Chou, J.S.; Truong, D.N. A novel metaheuristic optimizer inspired by behavior of jellyfish in ocean. Appl. Math. Comput. 2021, 389, 125535. [Google Scholar] [CrossRef]
  41. Wang, L.Y.; Cao, Q.J.; Zhang, Z.X.; Mirjalili, S.; Zhao, W.G. Artificial rabbits optimization: A new bio-inspired meta-heuristic algorithm for solving engineering optimization problems. Eng. Appl. Artif. Intell. 2022, 114, 31. [Google Scholar] [CrossRef]
  42. Abdollahzadeh, B.; Gharehchopogh, F.S.; Khodadadi, N.; Mirjalili, S. Mountain Gazelle Optimizer: A new Nature-inspired Metaheuristic Algorithm for Global Optimization Problems. Adv. Eng. Softw. 2022, 174, 34. [Google Scholar] [CrossRef]
  43. Chen, C.; Wang, X.; Yu, H.; Wang, M.; Chen, H. Dealing with multi-modality using synthesis of Moth-flame optimizer with sine cosine mechanisms. Math. Comput. Simul. 2021, 188, 291–318. [Google Scholar] [CrossRef]
  44. Wang, G.-G. Moth search algorithm: A bio-inspired metaheuristic algorithm for global optimization problems. Memetic Comput. 2018, 10, 151–164. [Google Scholar] [CrossRef]
  45. Li, J.; Yang, Y.H.; An, Q.; Lei, H.; Deng, Q.; Wang, G.G. Moth Search: Variants, Hybrids, and Applications. Mathematics 2022, 10, 4162. [Google Scholar] [CrossRef]
  46. Elaziz, M.A.; Xiong, S.; Jayasena, K.; Li, L. Task scheduling in cloud computing based on hybrid moth search algorithm and differential evolution. Knowl.-Based Syst. 2019, 169, 39–52. [Google Scholar] [CrossRef]
  47. Strumberger, I.; Sarac, M.; Markovic, D.; Bacanin, N. Moth search algorithm for drone placement problem. Int. J. Comput. 2018, 3, 75–80. [Google Scholar]
  48. Strumberger, I.; Tuba, E.; Bacanin, N.; Beko, M.; Tuba, M. Hybridized moth search algorithm for constrained optimization problems. In Proceedings of the 2018 International Young Engineers Forum (YEF-ECE), Monte de Caparica, Portugal, 4 May 2018; pp. 1–5. [Google Scholar]
  49. Feng, Y.; Wang, G.-G. Binary moth search algorithm for discounted {0-1} knapsack problem. IEEE Access 2018, 6, 10708–10719. [Google Scholar] [CrossRef]
  50. Feng, Y.; An, H.; Gao, X. The importance of transfer function in solving set-union knapsack problem based on discrete moth search algorithm. Mathematics 2019, 7, 17. [Google Scholar] [CrossRef]
  51. Feng, Y.; Yi, J.-H.; Wang, G.-G. Enhanced moth search algorithm for the set-union knapsack problems. IEEE Access 2019, 7, 173774–173785. [Google Scholar] [CrossRef]
  52. Cui, Z.; Xue, F.; Cai, X.; Cao, Y.; Wang, G.-G.; Chen, J. Detection of malicious code variants based on deep learning. IEEE Trans. Ind. Inform. 2018, 14, 3187–3196. [Google Scholar] [CrossRef]
  53. Zhang, J.; Pan, L.; Han, Q.-L.; Chen, C.; Wen, S.; Xiang, Y. Deep Learning Based Attack Detection for Cyber-Physical System Cybersecurity: A Survey. IEEE/CAA J. Autom. Sin. 2021, 9, 377–391. [Google Scholar] [CrossRef]
  54. Wang, F.; Wang, X.; Sun, S. A reinforcement learning level-based particle swarm optimization algorithm for large-scale optimization. Inf. Sci. 2022, 602, 298–312. [Google Scholar] [CrossRef]
  55. Wang, L.; Pan, Z.; Wang, J. A review of reinforcement learning based intelligent optimization for manufacturing scheduling. Complex Syst. Model. Simul. 2021, 1, 257–270. [Google Scholar] [CrossRef]
  56. Zhou, T.; Chen, M.; Zou, J. Reinforcement learning based data fusion method for multi-sensors. IEEE/CAA J. Autom. Sin. 2020, 7, 1489–1497. [Google Scholar] [CrossRef]
  57. Zheng, R.-z.; Zhang, Y.; Yang, K. A transfer learning-based particle swarm optimization algorithm for travelling salesman problem. J. Comput. Des. Eng. 2022, 9, 933–948. [Google Scholar] [CrossRef]
  58. Bingjie, X.; Deming, L. Q-Learning-based teaching-learning optimization for distributed two-stage hybrid flow shop scheduling with fuzzy processing time. Complex Syst. Model. Simul. 2022, 2, 113–129. [Google Scholar] [CrossRef]
  59. Wang, G.-G.; Tan, Y. Improving metaheuristic algorithms with information feedback models. IEEE Trans. Cybern. 2019, 49, 542–555. [Google Scholar] [CrossRef]
  60. Chen, C.; Wang, X.; Yu, H.; Zhao, N.; Wang, M.; Chen, H.; Kumarappan, N. An Enhanced Comprehensive Learning Particle Swarm Optimizer with the Elite-Based Dominance Scheme. Complexity 2020, 2020, 1–24. [Google Scholar] [CrossRef]
  61. Gong, M.; Jiao, L.; Zhang, L. Baldwinian learning in clonal selection algorithm for optimization. Inf. Sci. 2010, 180, 1218–1236. [Google Scholar] [CrossRef]
  62. Hinton, G.E.; Nowlan, S.J. How learning can guide evolution. In Adaptive Individuals in Evolving Populations: Models and Algorithms; Addison-Wesley Longman Publishing Co., Inc.: Boston, MA, USA, 1996; Volume 26, pp. 447–454. [Google Scholar]
  63. Qi, Y.; Liu, F.; Liu, M.; Gong, M.; Jiao, L. Multi-objective immune algorithm with Baldwinian learning. Appl. Soft Comput. 2012, 12, 2654–2674. [Google Scholar] [CrossRef]
  64. Zhang, C.; Chen, J.; Xin, B. Distributed memetic differential evolution with the synergy of Lamarckian and Baldwinian learning. Appl. Soft Comput. 2013, 13, 2947–2959. [Google Scholar] [CrossRef]
  65. Geem, Z.W.; Kim, J.H.; Loganathan, G.V. A new heuristic optimization algorithm: Harmony search. Simulation 2001, 76, 60–68. [Google Scholar] [CrossRef]
  66. Omran, M.G.H.; Mahdavi, M. Global-best harmony search. Appl. Math. Comput. 2008, 198, 643–656. [Google Scholar] [CrossRef]
  67. Xiang, W.L.; An, M.Q.; Li, Y.Z.; He, R.C.; Zhang, J.F. A novel discrete global-best harmony search algorithm for solving 0-1 knapsack problems. Discret. Dyn. Nat. Soc. 2014, 2014, 1–12. [Google Scholar] [CrossRef]
  68. Keshtegar, B.; Sadeq, M.O. Gaussian global-best harmony search algorithm for optimization problems. Soft Comput. 2017, 21, 7337–7349. [Google Scholar] [CrossRef]
  69. El-Abd, M. An improved global-best harmony search algorithm. Appl. Math. Comput. 2013, 222, 94–106. [Google Scholar] [CrossRef]
  70. Peng, Y.; Lu, B.-L. Hybrid learning clonal selection algorithm. Inf. Sci. 2015, 296, 128–146. [Google Scholar] [CrossRef]
  71. Lan, K.-T.; Lan, C.-H. Notes on the distinction of Gaussian and Cauchy mutations. In Proceedings of the 2008 Eighth International Conference on Intelligent Systems Design and Applications, Kaohsuing, Taiwan, 26–28 November 2008; pp. 272–277. [Google Scholar]
  72. Mahdavi, M.; Fesanghary, M.; Damangir, E. An improved harmony search algorithm for solving optimization problems. Appl. Math. Comput. 2007, 188, 1567–1579. [Google Scholar] [CrossRef]
  73. Xiao, S.; Wang, W.; Wang, H.; Tan, D.; Wang, Y.; Yu, X.; Wu, R. An Improved Artificial Bee Colony Algorithm Based on Elite Strategy and Dimension Learning. Mathematics 2019, 7, 289. [Google Scholar] [CrossRef]
  74. Chu, P.C.; Beasley, J.E. A genetic algorithm for the multidimensional knapsack problem. J. Heuristics 1998, 4, 63–86. [Google Scholar] [CrossRef]
  75. Puchinger, J.; Raidl, G.R.; Pferschy, U. The multidimensional knapsack problem: Structure and algorithms. Inf. J. Comput. 2010, 22, 250–265. [Google Scholar] [CrossRef]
  76. Senju, S.; Toyoda, Y. An approach to linear programming with 0-1 variables. Manag. Sci. 1968, 15, B196–B207. [Google Scholar] [CrossRef]
  77. Freville, A. Hard 0-1 multiknapsack test problems for size reduction methods. Investig. Oper. 1990, 1, 251–270. [Google Scholar]
  78. Bansal, J.C.; Deep, K. A modified binary particle swarm optimization for knapsack problems. Appl. Math. Comput. 2012, 218, 11042–11061. [Google Scholar] [CrossRef]
  79. Chih, M.; Lin, C.-J.; Chern, M.-S.; Ou, T.-Y. Particle swarm optimization with time-varying acceleration coefficients for the multidimensional knapsack problem. Appl. Math. Model. 2014, 38, 1338–1350. [Google Scholar] [CrossRef]
  80. Dina, E.-G.; Badr, A.; Abd El Azeim, M. New binary particle swarm optimization with immunity-clonal algorithm. J. Comput. Sci. 2013, 9, 1534. [Google Scholar]
  81. Beheshti, Z.; Shamsuddin, S.M.; Hasan, S. Memetic binary particle swarm optimization for discrete optimization problems. Inf. Sci. 2015, 299, 58–84. [Google Scholar] [CrossRef]
  82. Kennedy, J.; Eberhart, R.C. A discrete binary version of the particle swarm algorithm. In Proceedings of the 1997 IEEE International Conference on Systems, Man, and Cybernetics, Orlando, FL, USA, 12–15 October 1997; pp. 4104–4108. [Google Scholar]
  83. Aaha, B.; Sm, C.; Hf, D.; Ia, D.; Mm, E.; Hc, F. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar]
  84. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp swarm algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  85. Abdel-Basset, M.; El-Shahat, D.; Sangaiah, A.K. A modified nature inspired meta-heuristic whale optimization algorithm for solving 0-1 knapsack problem. Int. J. Mach. Learn. Cybern. 2017, 10, 1–20. [Google Scholar] [CrossRef]
  86. Pinto, H.; Pea, A.; Valenzuela, M.; Fernández, A. A binary sine-cosine algorithm applied to the knapsack problem. In Proceedings of the Computer Science Online Conference, Zlin, Czech Republic, 24–27 April 2019; pp. 128–138. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.