Next Article in Journal
Leveraging Machine Learning for Optimized Mechanical Properties and 3D Printing of PLA/cHAP for Bone Implant
Next Article in Special Issue
MSAO-EDA: A Modified Snow Ablation Optimizer by Hybridizing with Estimation of Distribution Algorithm
Previous Article in Journal
Active Vibration Control and Parameter Optimization of Genetic Algorithm for Partially Damped Composites Beams
Previous Article in Special Issue
MICFOA: A Novel Improved Catch Fish Optimization Algorithm with Multi-Strategy for Solving Global Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Set Packing Optimization by Evolutionary Algorithms with Theoretical Guarantees

1
School of Information Science and Engineering, Jiaxing University, Jiaxing 314001, China
2
School of Computer Science and Cyber Engineering, Guangzhou University, Guangzhou 510006, China
3
School of Mathematics and Systems Science, Guangdong Polytechnic Normal University, Guangzhou 510665, China
4
College of Artificial Intelligence, Nankai University, Tianjin 300350, China
5
Department of Electrical and Electronic Engineering, Hanyang University, Ansan 15588, Republic of Korea
*
Authors to whom correspondence should be addressed.
Biomimetics 2024, 9(10), 586; https://doi.org/10.3390/biomimetics9100586
Submission received: 14 September 2024 / Revised: 23 September 2024 / Accepted: 25 September 2024 / Published: 27 September 2024

Abstract

:
The set packing problem is a core NP-complete combinatorial optimization problem which aims to find the maximum collection of disjoint sets from a given collection of sets, S, over a ground set, U. Evolutionary algorithms (EAs) have been widely used as general-purpose global optimization methods and have shown promising performance for the set packing problem. While most previous studies are mainly based on experimentation, there is little theoretical investigation available in this area. In this study, we analyze the approximation performance of simplified versions of EAs, specifically the (1+1) EA, for the set packing problem from a theoretical perspective. Our analysis demonstrates that the (1+1) EA can provide an approximation guarantee in solving the k-set packing problem. Additionally, we construct a problem instance and prove that the (1+1) EA beats the local search algorithm on this specific instance. This proof reveals that evolutionary algorithms can have theoretical guarantees for solving NP-hard optimization problems.

1. Introduction

The set packing problem is a classical and significant NP-complete combinatorial optimization problem, which is included in 21 NP-complete problems famously established by Karp [1]. In the set packing problem, we are given a base set containing m elements, along with a collection of n subsets derived from the base set. The objective is to maximize the count of disjoint sets within this collection [2]. The set packing problem has attracted widespread attention from scholars in academic research and engineering fields, and has been investigated for decades. This problem has extensive practical applications, such as scheduling [3], combinatorial auctions [4], communication networks [5], product management [6], and computer vision [7], just to name a few. The significance of set packing optimization lies in its ability to efficiently allocate resources, schedule tasks, and optimize operations by selecting a subset of items while maximizing the overall value or minimizing the costs involved. Therefore, the research on the set packing problem has important application value and practical significance.
Considering the inherent complexity of the set packing problem, researchers have devoted their efforts to finding efficient approaches to address this challenging task. However, unless P = N P , it is unlikely that we can discover an algorithm for solving the set packing problem in polynomial time. Therefore, we usually only find satisfactory or approximate solutions to this NP-hard problem in practice. Hurkens and Schrijver [8] introduced a local search method that incorporates a p-neighborhood for solving the set packing problem. The results indicated that the local search technique can obtain k 2 + ϵ approximate solutions on this problem for any ϵ > 0 . Sviridenko and Ward [9] proposed a local improvement technique that incorporates the color-coding technique, resulting in an improved approximation ratio from k 2 + ϵ [8] to k + 2 3 . Cygan [10] considered local search techniques with swaps of bounded pathwidth, which resulted in a polynomial time approximation algorithm with an approximation ratio of k + 1 + ϵ 3 . Later, Fürer and Yu [11] achieved the same approximation ratio k + 1 + ϵ 3 as [10]. However, the runtime of [11] is singly exponential in ϵ 2 while it is doubly exponential in ϵ 1 for [10]; the time complexity is decreased.
In the context of the set packing problem, when assigning weights to each subset, it is called the weighted set packing problem. Bafna et al. [12] studied a local improvement heuristic for the weighted set packing problem, resulting in a tight approximation ratio of k 1 + 1 k . Arkin and Hassin [13] obtained the same approximation guarantee for the problem by using a local improvement method, and extended to a general bound k 1 + ϵ for any fixed ϵ . Furthermore, Chandra and Halldórsson [14] proposed a greedy strategy called B E S T I M P for the k-set packing problem, which combines greedy with local improvement and achieved an approximation guarantee of 2 ( k + 1 ) 3 . They also proposed another greedy strategy called A N Y I M P which can find a bigger improvement than a specified threshold rather than seeking the best improvement. The A N Y I M P algorithm can obtain an approximation ratio 4 k + 2 5 for the optimal threshold selection. Recently, based on the local search procedure of Berman’s algorithm [15], Thiery and Ward [16] proposed an improved squared-weight local search algorithm with a large exchange for the weighted k-set packing problem. It is shown that, for the weighted k-set packing problem, when performing exchanges of size reaching to 2 k 2 ( k 1 ) + k , the algorithm can attain an approximation ratio of k + 1 τ k 2 for τ k τ 3 = 0.214 . The authors presented the analysis and design of a local search algorithm that incorporates extensive exchanges within a large neighborhood for this problem. Recently, Neuwohner [17] also studied the weighted k-set packing problem; some approximation guarantees are obtained by combining local search with the application of a black box algorithm. Gadekar [18] investigated the parameterized complexity on the set packing problem; he constructed a gadget called the compatible intersecting set system pair to obtain some running time bounds for the parameterized set packing problem. Moreover, Duppala et al. [19] proved that the randomized algorithms can achieve a probabilistically fair solution with provable guarantees for a fair k-set packing problem.
Evolutionary algorithms (EAs) [20,21,22] belong to a broad category of stochastic heuristic search methods that draw inspiration from natural processes, which can effectively solve optimization problems without any prior knowledge. Due to their simplicity and ease of implementation, EAs have been widely used to solve various complex problems, which show superior search capabilities compared to local search and greedy algorithms [23,24,25,26,27,28,29]. Generally, EAs solve optimization problems by updating the population via selection (to select the better individual to enter the next generation), crossover mutation (to generate the new offspring), and fitness evaluation (to evaluate the performance of different individuals) operators [30,31,32,33]. Since EAs have shown remarkable success in practice, the need for theoretical underpinnings is crucial to help understand the algorithm’s behavior, convergence properties, and performance guarantees. In order to further investigate the execution mechanism and essence of EAs, many scholars have carried out some theoretical research for evolutionary algorithms and obtained some achievements in the past few years. Especially He and Yao [34,35,36,37], Neumann [38,39], Qian, Yu and Zhou [40,41], and Xia et al. [42,43,44] who obtained a series of theoretical research results of evolutionary algorithms, including average convergence rate, noisy evolutionary optimization, evolutionary Pareto optimization, evolutionary discrete optimization, etc. EAs have also been widely used for solving the set packing problem. From an experimental perspective, the research findings indicate that EAs demonstrate favorable performance when applied to the set packing problem [45,46,47].
However, there has been limited research conducted on the theoretical analysis of EAs’ performance for the set packing problem. Therefore, it is natural to question whether evolutionary algorithms can obtain some performance guarantees on the set packing problem. We concentrate on the worst-case approximation guarantee achieved by some stochastic algorithms for any problem instance. That is, what the quality of solutions generated by EAs is in what expected runtime, and how the relationships between the solution quality and runtime are. This will help strengthen the theoretical foundation of EAs, and provide guidance to the design of the algorithms.
In this study, we address this question by presenting an analysis of the approximation performance of evolutionary algorithms on the set packing problem. Our primary emphasis lies in the approximation ratio achieved by the algorithms, as well as when the algorithms can obtain this approximation ratio. Specifically, we focus on a simplified variant of an EA known as (1+1) EA, which employs only the mutation operator to iteratively generate improved solutions, and the population size is one. Our investigation focuses on analyzing the performance guarantees of the (1+1) EA when it is applied to the set packing problem. Our analysis reveals that the (1+1) EA has the capability to achieve an approximate ratio of k 2 + ϵ on the k-set packing problem within an expected runtime of O ( n 2 p ) , where ϵ > 0 and p = O ( log k 1 ϵ ) is an integer. Moreover, we generate a specific instance of the k-set packing problem and demonstrate that the (1+1) EA can effectively discover the global optima within an expected polynomial runtime. In contrast, local search algorithms tend to become trapped in local optima when dealing with this particular instance.
The rest of this paper is structured as follows. In the subsequent section, we provide an introduction to the set packing problem, along with an overview of the relevant algorithms such as local search and greedy approaches. Additionally, we discuss the analysis tools that are considered in this study. In Section 3, we conduct a theoretical analysis of the approximation guarantees of the (1+1) EA for the k-set packing problem. Additionally, Section 4 provides a performance analysis for the (1+1) EA applied to a problem instance that we construct. Finally, the conclusions are presented in Section 5.

2. Preliminaries

This section begins by providing a comprehensive description of set packing problems and the corresponding heuristic algorithms. Moreover, we introduce pertinent concepts and analysis tools that will be utilized throughout this paper.

2.1. Set Packing Problem

To begin, we present the concept of an independent set before introducing the specific set packing problem addressed within this work.
Definition 1. 
(Independent set) For an undirected graph G = ( V , E ) , where V represents the set of vertices, an independent set I V is defined as a subset of vertices in which no two vertices are adjacent to each other. In other words, there are no edges in E connecting any two vertices within the set I.
The set packing problem can be considered as an extension or generalization of the independent set problem [48].
In the set packing problem, we are given a universal set of n elements denoted as U = { 1 , 2 , , N } , along with a collection C = { C 1 , C 2 , , C n } consisting of subsets of U. The problem is to identify a collection C C of sets from C that are pairwise disjoint. The goal is to maximize the cardinality of C , represented as | C | .
A set packing corresponds to an independent set in the intersection graph. Hence, we give the formal definitions of the set packing problem in the context of an undirected graph.
Definition 2. 
(Set packing problem) For an undirected graph G = ( V , E ) with a vertices set V = { 1 , 2 , , N } , and a collection C = { C 1 , C 2 , , C n } consisting of subsets of V, we aim to discover a collection C C of sets from C that are pairwise disjoint. The goal is to maximize the cardinality of C , denoted as | C | .
Definition 3. 
(k-Set packing problem) In the set packing problem, when there is a constraint on the size of C i such that | C i |     k for every i [ n ] , it is referred to as the k-set packing problem.

2.2. Local Search Algorithm

The local search algorithm is a commonly used heuristic approach for solving the set packing problem. It operates by iteratively modifying the current solution in a local manner to improve its quality. A local search approach for the k-set packing problem was introduced by Hurkens and Schrijver [8]. The algorithm begins with an empty set I and examines whether there exist t p disjoint sets that are not part of the current solution set. These sets should intersect at most t 1 sets from the current collection. If such a collection S of t sets is found, the algorithm improves the current solution by adding S while simultaneously removing all sets from the current solution that intersect any set in S (up to t 1 sets). This local search algorithm is given in  Algorithm 1 below. The search continues until no further improvement can be made. The resulting solution is referred to as p-optimal (p-opt).
Algorithm 1: p-local search algorithm for the k-set packing problem
Input: Graph G = ( V , E ) ;
Output: Maximum Independent Set I;
1: Initialize: I ;
2: While termination condition does not hold do
3:    Find a set I g o o d C I such that | I g o o d | p and the sets in I g o o d
         are disjoint.
4:     I b a d = { x I | x ( I g o o d ) } ;
5:    if  | I g o o d | > | I b a d |  then
6:        I = I I g o o d I b a d ;
7:  end if
8: end while  
As shown in Algorithm 1, we can see that the p-local search algorithm aims to iteratively enhance the current solution in order to achieve an improved solution.

2.3. Analysis Method

This section presents a fitness-level method (also known as fitness-based partitions) [49], a theoretical analysis technique. The fitness-level method is a straightforward yet useful approach for estimating the expected runtime of evolutionary algorithms and other randomized search heuristics applied to combinatorial optimization problems. Next, we provide the formal definition of the fitness-based partitions method.
Definition 4. 
(Fitness-level method). Let S = { 0 , 1 } n be a finite search space, and let f : S R be a fitness function that needs to be maximized. For any two sets A and B that are subsets of S, we define A < f B if f ( a ) < f ( b ) holds for all a in A and all b in B. To analyze the search process, we partition the search space S into disjoint and non-empty sets A 1 , A 2 , , A m , satisfying the ordering
A 1 < f A 2 < f < f A m .
The set A m exclusively comprises optimal search points. We denote p i as the lower bound probability that, in the subsequent iteration, a new search point x will be generated from the current search point x, where x belongs to A i + 1 A i + 2 A m . Additionally, we define T f as the expected number of generations required by the (1+1) EA to discover an optimal search point for the fitness function f, then we have
E ( T f ) i = 1 m 1 1 p i .
The fitness-level method is a simple yet efficient approach for runtime analysis for EAs, also called fitness-based partitions. However, this technique only presents the upper bound estimation, and a general lower bound estimation method is difficult to establish. It should be pointed out that if you use this method, you have to give an appropriate partition for the search space S. The number of partitions cannot be too large, i.e., exponential. Furthermore, it is relatively straightforward to estimate the probability of transitioning from the current partition to a better partition and generating a new offspring.

3. Theoretical Analysis of a Simple Evolutionary Algorithm for k -Set Packing Problem

In this section, we focus on analyzing the approximation guarantee of evolutionary algorithms applied to the set packing problem. An approximation algorithm is designed to provide a solution for a combinatorial optimization problem, such that its objective value is guaranteed to be within a bounded factor of the optimal solution’s value. The approximation ratio serves as a fundamental metric for evaluating the quality of an approximation algorithm.
Let us begin by introducing the notion of approximation ratio. Without a loss of generality, we consider a maximization problem with an objective function denoted as f. For any given instance I, the value of an optimal solution is represented by O P T ( I ) . An algorithm is considered a ρ -approximation algorithm if the best solution x obtained by the algorithm satisfies f ( x ) 1 ρ · O P T ( I ) , where ρ 1 . In other words, the algorithm achieves a solution with a value that is at least 1 ρ times the value of the optimal solution. We refer to this as the ρ -approximation ratio for the problem. If the value of ρ is equal to 1, it indicates that the algorithm is considered optimal. It is worth noting that a higher approximation ratio indicates a worse solution in terms of quality.
Next, we present the (1+1) EA used in this study as shown in Algorithm 2.
Algorithm 2: The (1+1) EA for the set packing problem
Input: A Graph G = ( V , E ) ;
Output: A vertices set C with the cardinality of C as maximized as possible.
1: Initialization: Randomly select an initial bit string x { 0 , 1 } m ;
2: While termination condition is not met do
3:    Generate an offspring x by independently flipping each bit of x with
         a probability of 1 / n ;
4:    if  f ( x ) > f ( x )  then
5:       x : = x .
6:   end if
7: end while  
The (1+1) EA begins with an arbitrary solution and proceeds iteratively by generating an offspring solution using the mutation operator but without crossover. The current solution is replaced by the new solution only if it is strictly superior to the current one.
To utilize the (1+1) EA for finding the optimal or approximate optimal solution to the k-set packing problem, we employ a bit string x = ( x 1 , x 2 , , x n ) { 0 , 1 } n to encode the problem’s solution. Each bit x i corresponds to a subset C i , where i [ n ] . If x i = 1 , it indicates that the subset C i is selected. Conversely, if x i = 0 , it implies that C i is not selected. Consequently, a solution or bit string x represents a collection of subsets, and | x | = i = 1 n x i represents the count of subsets contained in x.
In the following, we define the fitness function for the k-set packing problem, denoted as
f ( x ) = | x | n 2 × u ( x ) .
We aim to maximize the fitness function f ( x ) . In the above Equation (1), the item u ( x ) represents the number of subsets with an intersection. The penalty term in the fitness function, n 2 × u ( x ) , serves to encourage the algorithm to minimize the value of u ( x ) . This, in turn, leads to a reduction in the number of intersecting subsets within the current solution. The primary goal is to obtain a solution where all subsets C i are disjoint. The first part of Equation (1) ensures that the number of disjoint subsets is maximized if all subsets are disjoint.
Hurkens and Schrijver [8] demonstrated that a simple local search algorithm can attain a polynomial time approximation ratio on the k-set packing problem.
Lemma 1 
([8]). For any ϵ > 0 , let p be an integer and p = O ( log k 1 ϵ ) . The local search algorithm with p local improvement can achieve an approximation ratio k 2 + ϵ when applied to any instance of the k-set packing problem.
Next, we demonstrate that the (1+1) EA can effectively discover a 2 k + 2 ϵ -approximation solution for the unweighted k-set packing problem, while simulating the aforementioned result established by Hurkens and Schrijver [8]. Additionally, the expected runtime for this achievement is bounded by O ( n 2 p ) .
The (1+1) EA has the capability to discover a packing that includes a minimum number of sets, reaching at least 2 k + 2 ϵ O P T , for the k-set packing problem. This accomplishment is expected to occur within a runtime of O ( n 2 p ) , where O P T represents the global optimum.
Theorem 1. 
The (1+1) EA has the capability to discover a packing with a number of sets that is at least 2 k + 2 ϵ O P T for the k-set packing problem. This achievement can be accomplished within an expected runtime of O ( n 2 p ) , where O P T represents the global optimum.
Proof. 
During the optimization process of the (1+1) EA, it is important to note that the value of the fitness function is never reduced. To facilitate this, the search space { 0 , 1 } n is divided into three distinct sets, namely P 1 , P 2 , and P 3 , based on their respective fitness values.
P 1 = { x | x { 0 , 1 } n , f ( x ) < 0 } ,
P 2 = { x | x { 0 , 1 } n , 0 f ( x ) < 2 k + 2 ϵ O P T } ,
P 3 = { x | x { 0 , 1 } n , f ( x ) 2 k + 2 ϵ O P T } .
It is obvious that P 1 represents the set of infeasible solutions, and P 2 and P 3 represent the set of feasible solutions.
In accordance with the fitness function (1), the fitness values in the evolutionary process of the (1+1) EA are guaranteed to never decrease. The (1+1) EA initiates by randomly selecting an initial solution from the search space. Suppose the current solution x belongs to P 1 , then x is a collection of subsets containing intersection, i.e., an infeasible solution. At least two subsets in the current solution are intersecting. The algorithm only accepts the event that reduces the number of subsets with an intersection. In this scenario, the (1+1) EA has the capability to enhance the fitness value by at least one by removing such a subset with an intersection from the solution. In this case, the probability of removing such a specific subset is 1 n ( 1 1 n ) n 1 1 e n = Ω ( 1 n ) ; this suggests that the fitness value is expected to increase by at least one within a runtime of O ( n ) . Note that the maximum fitness value is n. Consequently, the solutions from set P 1 will be transformed into set P 2 by the (1+1) EA, and this transformation will be accomplished within an expected runtime of O ( n 2 ) .
Let us assume that the current solution x belongs to set P 2 . On the basis of Lemma 1, if some solution satisfies 0 f ( x ) < 2 k + 2 ϵ O P T , then the algorithm evolves towards increasing the number of disjoint subsets. For this case, the following possibilities exist. There exist p disjoint sets that are not included in the current solution, and these p subsets intersect at most p 1 subsets in the current solution. Consequently, by performing a p-local improvement operation, the fitness value is increased by at least 1 by the (1+1) EA. The probability of performing such operation for the (1+1) EA is ( 1 n ) 2 p 1 ( 1 1 n ) n 2 p + 1 1 e n 2 p 1 = Ω ( 1 e n 2 p 1 ) . Since there are most n elements in the solution, according to the fitness-level method, performing such n operations, the (1+1) EA will transform the solutions into set P 3 within an expected runtime of O ( 1 n 2 p ) .
Combining the above analysis for different phases, we complete the proof. □
As shown in Theorem 1, the (1+1) EA can achieve a performance guarantee equivalent to that of the local search algorithm in the worst case. Understanding the theoretical properties of different algorithms can guide researchers and practitioners in choosing the most suitable approach based on the problem characteristics. Applying theoretical guarantees to real-world optimization problems can help in predicting algorithm behavior and performance outcomes. By leveraging theoretical insights, practitioners can tailor optimization algorithms to specific problem domains and optimize them for practical applications.

4. The Power of Evolutionary Algorithms on Some Instances of the k -Set Packing Problem

Local search and greedy algorithms are widely used and are practical approaches for efficiently searching approximate solutions in various optimization problems. However, both of these techniques are prone to becoming stuck in local optima. In this section, we present a constructed instance of the k-set packing problem. We demonstrate that the simple evolutionary algorithm has the capability to discover global optimal solutions within an expected polynomial runtime. On the other hand, the 2-local search algorithm may be apt to fall into the local optimum.
First, we give two sets with base elements as follows.
A = { v 1 , v 2 , , v s } ,
and
B = { u 1 , u 2 , , u s , u 1 , 2 , , u 1 , s , u 2 , 3 , , u 2 , s , , u i , i + 1 , u i , i + 2 , , u i , s , , u s 1 , s } ,
where s is an integer.
The instance I s p is constructed using the following steps. First, we construct three groups of sets C i , 0 , C 0 , i and C i , j . Let
C i , 0 = { { v i , u i } , { v i , u 1 , i } , , { v i , u i 1 , i } , { v i , u i , i + 1 } , , { v i , u i , s } } , w h e r e 2 i s 1 .
Here, C 1 , 0 = { { v 1 , u 1 } , { v 1 , u 1 , 2 } , , { v 1 , u 1 , s } } , and C s , 0 = { { v s , u s } , { v s , u 1 , s } , { v s , u 2 , s } , , { v s , u s 1 , s } } .
If the subscript of u is 0, it means that the element does not exist.
For 1 i s , we set
C 0 , i = { { v i , u i } }
and
C i , j = { { v i , u i , j } , { v j , u i , j } } .
After that, we define two sets X and Y. For any 1 i s , let X = { C 1 , 0 , C 2 , 0 , , C s , 0 } , Y 1 = { C 0 , 1 , C 0 , 2 , , C 0 , s } and Y 2 = { C 1 , 2 , , C 1 , s , C 2 , 3 , , C 2 , s , , C s 1 , s } . The set Y = Y 1 Y 2 . We can see that the size of set X is s, i.e., | X | = s , and | Y | = | Y 1 | + | Y 2 | = s + s ( s 1 ) 2 . The number of all elements contained in sets X and Y is n. The constructed instance I s p is shown in Figure 1.
In the following analysis, we demonstrate that the (1+1) EA can achieve a global optima for instance I s p within an expected polynomial runtime.
Theorem 2. 
For instance I s p , the (1+1) EA can efficiently obtain the global optimum starting from any initial solution within an expected runtime of O ( n 5 ) .
Proof. 
Obviously, the global optimum of I s p is { C 0 , 1 , , C 0 , s , C 1 , 2 , , C 1 , s , C 2 , 3 , , C 2 , s , , C s 1 , s } as shown in Figure 2. We can divide the optimization process of the algorithm on I s p into two stages. In the first stage, the fitness value is strictly negative, while in the second stage, the fitness value becomes positive.
In the first stage, since the fitness value is strictly less than 0, the intersection of some subsets exists in the current solution. In this case, the (1+1) EA will only accept operations that reduce the number of intersections in the subset. The probability that the algorithm reduces a particular subset with intersection is 1 n ( 1 1 n ) n 1 = Ω ( 1 n ) . It is noticed that the total set contains n elements. By employing the fitness-level method, the (1+1) EA is transitioned from the first stage to the second stage within an expected runtime of O ( n 2 ) .
During the second stage, the subsets within the current solution are mutually disjoint. It is necessary to consider four distinct cases in this context.
Case 1: All the elements from X are included in the current solution, i.e., a 2-optimal solution, as shown in Figure 3. In this case, any subset in Y can not be included into the current solution. Because the algorithm only accepts better solutions, this case occurs only once during the optimization process. We claim that the fitness value can be increased by at least one by the (1+1) EA through deleting two subsets C i , 0 and C j , 0 in X and simultaneously adding three subsets C 0 , i , C 0 , j and C i , j in Y. In the following, we estimate the probability that the algorithm performs this operation. We first choose one element from the set Y 2 , where the probability for this operation is s ( s 1 ) 2 n . We assume that we choose the element C i , j , then we need to remove the elements C i , 0 and C j , 0 from set X while adding the elements C 0 , i and C 0 , j from set Y simultaneously. Such an improvement will occur with probability s ( s 1 ) 2 n 1 n 4 ( 1 1 n ) n 5 = Ω ( 1 n 4 ) , which implies that the fitness value can be increased by at least one in the expected runtime O ( n 4 ) .
Case 2: The current solution contains, at most, s 1 elements selected from X, but there are no elements from Y contained in the current solution. In this scenario, by either adding the elements from set X or removing the elements from set Y that are disconnected from the elements in set X, the fitness value can be increased. Such an operation is executed by the (1+1) EA with probability 1 n ( 1 1 n ) n 1 = Ω ( 1 n ) , which implies that the fitness value can be increased by at least one in expected runtime O ( n ) .
Case 3: The current solution contains at most s 1 elements selected from X, also there exist some elements from Y contained in the current solution. For this case, we can divide it into two subcases for further consideration.
Subcase 3.1: The current solution contains less than s 1 elements selected from X. Let us assume that the subset C 0 , k is included in the current solution but without C i , 0 and C j , 0 for the sake of simplicity. Based on the aforementioned analysis, we can conclude that the current solution forms an independent set. Therefore, neither subset C i , k nor C j , k is present in the current solution. Through adding subsets C i , k and C j , k while simultaneously removing subset C 0 , k from set X by the (1+1) EA, the fitness value will be increased. Such an operation is executed by the (1+1) EA with probability 1 n 3 ( 1 1 n ) n 3 = Ω ( 1 n 3 ) , which implies that the fitness value can be increased by at least one in the expected runtime O ( n 3 ) .
Subcase 3.2: The current solution contains exactly s 1 elements from X, meaning that only a single element in X is excluded form the current solution. For the sake of simplicity, let us assume that the element C i , 0 from X is not included in the current solution. Furthermore, since the current solution is an independent set, we can know that only the element C 0 , i in Y is contained in the current solution. Through adding two subsets C 0 , j and C i , j and simultaneously deleting the subset C j , 0 from X by the (1+1) EA, the fitness value will be increased. The probability of the (1+1) EA executing such an operation is 1 n 3 ( 1 1 n ) n 3 = Ω ( 1 n 3 ) , which implies that the fitness value can be increased by at least one in the expected runtime O ( n 3 ) .
Case 4: The current solution exclusively consists of the elements from set Y. For this case, through adding any other element from Y into the current solution by the (1+1) EA, the fitness value can be increased by at least one. Such an operation is executed by the (1+1) EA with probability 1 n ( 1 1 n ) n 1 = Ω ( 1 n ) , which implies that the fitness value can be increased by at least one in the expected runtime O ( n ) .
Note that there are mostly n elements in the solution, i.e., the maximum fitness value is n. Therefore, according to the above analysis, starting with an arbitrary solution, the (1+1) EA can obtain the global optimal solution of instance I s p in an expected runtime of O ( n 5 ) . □
As we can see from Theorem 2, as a global optimization method, the evolutionary algorithm can avoid becoming stuck in local optima, but it is easy for the local search algorithm to become trapped in local optima.

5. Conclusions and Future Work

EAs have been commonly used for tackling the set packing problem, an NP-complete combinatorial optimization problem, and some experimental investigations have demonstrated that EAs exhibit efficiency in solving this problem. However, the understanding of the performance of EAs in the set packing problem remains limited in theory. This work focuses on the performance analysis of a simple but effective (1+1) EA on the set packing problem. We have demonstrated that the simple algorithm can obtain some approximation guarantees on the set packing problem. Furthermore, we theoretically reveal that the (1+1) EA outperforms the local search algorithm on a set packing instance. This study contributes to bridging the gap between theory and practice in the context of evolutionary algorithms. Theoretical investigation can give insight in how algorithms work and help select the most appropriate algorithm for a given problem. By understanding these theoretical insights into practical scenarios, researchers and practitioners can make decisions, optimize algorithm performance, and enhance the reliability of algorithms across practical applications.
In the future, we will further conduct other theoretical analysis for the more complex EAs on the k-set packing problem, and will also consider other theoretical analysis methods, such as parameter analysis, in some practical intelligence optimization algorithms [50,51,52,53] to see whether EAs can still maintain the approximation guarantees, or potentially improve the existing approximation guarantees established in the set packing problem. We can study the impact of crossover operators on algorithm performance. It is also interesting to theoretically analyze the approximation performance on the weighted k-set packing problem [54]. This will help to provide some valuable insights into the correlation between problem features and algorithm performance, facilitating a deeper understanding of their relationship. Evolutionary algorithms are commonly used to solve complex optimization problems in various fields, they are expected to provide theoretical evidence for more applications, e.g., machine learning and multi-objective optimization.

Author Contributions

Y.J.: conceptualization, methodology, formal analysis, investigation, writing—original draft. X.X.: conceptualization, methodology, writing—review and editing. Z.W.: conceptualization, methodology, supervision, writing—review and editing. X.P.: validation, writing—review and editing. J.Z.: writing—review and editing, project administration. W.L.: validation, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (62106055, 61703183), National Key Research and Development Program of China (2023YFC3305900, 2023YFC3305903), Natural Science Foundation of Zhejiang Province (LGG19F030010), Guangdong Natural Science Foundation (2022A1515011825), Guangzhou Science and Technology Planning Project (2023A04J0388, 2023A03J0662), and the Qin Shen Scholar Program of Jiaxing University.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon request. There are no restrictions on data availability.

Conflicts of Interest

The authors declare no competing interests.

References

  1. Karp, R. Reducibility among Combinatorial Problems. In 50 Years of Integer Programming 1958–2008; Jünger, M., Liebling, T.M., Naddef, D., Nemhauser, G.L., Pulleyblank, W.R., Reinelt, G., Rinaldi, G., Wolseyet, L.A., Eds.; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  2. Delorme, X.; Gandibleux, X.; Rodriguez, J. GRASP for set packing problems. Eur. J. Oper. Res. 2004, 153, 564–580. [Google Scholar] [CrossRef]
  3. Velásquez, R.; Melo, M.T. A set packing approach for scheduling elective surgical procedures. In Operations Research Proceedings 2005: Selected Papers of the Annual International Conference of the German Operations Research Society (GOR), Bremen, Germany, 7–9 September 2005; Springer: Berlin/Heidelberg, Germany, 2006; pp. 425–430. [Google Scholar]
  4. De Vries, S.; Vohra, R.V. Combinatorial auctions: A survey. Informs J. Comput. 2003, 15, 284–309. [Google Scholar] [CrossRef]
  5. Emek, Y.; Halldórsson, M.M.; Mansour, Y.; Patt-Shamir, B.; Radhakrishnan, J.; Rawitz, D. Online set packing. SIAM J. Comput. 2012, 41, 728–746. [Google Scholar] [CrossRef]
  6. Ray, A.K.; Jenamani, M.; Mohapatra, P.K.J. Supplier behavior modeling and winner determination using parallel MDP. Expert Syst. Appl. 2011, 38, 4689–4697. [Google Scholar] [CrossRef]
  7. Yarkony, J.; Adulyasak, Y.; Singh, M.; Desaulniers, G. Data association via set packing for computer vision applications. Informs J. Optim. 2020, 2, 167–191. [Google Scholar] [CrossRef]
  8. Hurkens, C.A.J.; Schrijver, A. On the size of systems of sets every t of which have an SDR, with an application to the worst-case ratio of heuristics for packing problems. SIAM J. Discret. Math. 1989, 2, 68–72. [Google Scholar] [CrossRef]
  9. Sviridenko, M.; Ward, J. Large neighborhood local search for the maximum set packing problem. In Automata, Languages, and Programming: 40th International Colloquium, ICALP 2013, Riga, Latvia, 8–12 July 2013; Proceedings, Part I 40; Springer: Berlin/Heidelberg, Germany, 2013; pp. 792–803. [Google Scholar]
  10. Cygan, M. Improved approximation for 3-dimensional matching via bounded pathwidth local search. In Proceedings of the 2013 IEEE 54th Annual Symposium on Foundations of Computer Science, Berkeley, CA, USA, 26–29 October 2013; pp. 509–518. [Google Scholar]
  11. Fürer, M.; Yu, H. Approximating the-set packing problem by local improvements. In Combinatorial Optimization: Third International Symposium, ISCO 2014, Lisbon, Portugal, 5–7 March 2014; Springer International Publishing: Cham, Switzerland, 2014; pp. 408–420. [Google Scholar]
  12. Bafna, V.; Narayanan, B.; Ravi, R. Nonoverlapping local alignments (weighted independent sets of axis-parallel rectangles). Discret. Appl. Math. 1996, 71, 41–53. [Google Scholar] [CrossRef]
  13. Arkin, E.M.; Hassin, R. On local search for weighted k-set packing. Math. Oper. Res. 1998, 23, 640–648. [Google Scholar] [CrossRef]
  14. Chandra, B.; Halldórsson, M.M. Greedy local improvement and weighted set packing approximation. J. Algorithms 2001, 39, 223–240. [Google Scholar] [CrossRef]
  15. Berman, P. A d/2 approximation for maximum weight independent set in d-claw free graphs. In Algorithm Theory-SWAT 2000: 7th Scandinavian Workshop on Algorithm Theory Bergen, Norway, 5–7 July 2000 Proceedings; Springer: Berlin/Heidelberg, Germany, 2002; pp. 214–219. [Google Scholar]
  16. Thiery, T.; Ward, J. An Improved Approximation for Maximum Weighted k-Set Packing. In Proceedings of the 2023 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), Florence, Italy, 22–25 January 2023; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2023; pp. 1138–1162. [Google Scholar]
  17. Neuwohner, M. Passing the limits of pure local search for weighted k-set packing. In Proceedings of the 2023 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), Florence, Italy, 22–25 January 2023; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 2023; pp. 1090–1137. [Google Scholar]
  18. Gadekar, A. On the parameterized complexity of compact set packing. Algorithmica 2024, 1–19. [Google Scholar] [CrossRef]
  19. Duppala, S.; Luque, J.; Dickerson, J.P.; Srinivasan, A. Group Fairness in Set Packing Problems. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence (IJCAI), Macao, China, 19–25 August 2023; pp. 391–399. [Google Scholar]
  20. Wang, Z.J.; Jian, J.R.; Zhan, Z.H.; Li, Y.; Kwong, S.; Zhang, J. Gene targeting differential evolution: A simple and efficient method for large-scale optimization. IEEE Trans. Evol. Comput. 2022, 27, 964–979. [Google Scholar] [CrossRef]
  21. Liu, S.; Wang, Z.J.; Wang, Y.G.; Kwong, S.; Zhang, J. Bi-directional learning particle swarm optimization for large-scale optimization. Appl. Soft Comput. 2023, 149, 110990. [Google Scholar] [CrossRef]
  22. Wang, Z.J.; Zhan, Z.H.; Yu, W.J.; Lin, Y.; Zhang, J.; Gu, T.; Zhang, J. Dynamic group learning distributed particle swarm optimization for large-scale optimization and its application in cloud workflow scheduling. IEEE Trans. Cybern. 2019, 50, 2715–2729. [Google Scholar] [CrossRef] [PubMed]
  23. Wang, Z.J.; Zhan, Z.H.; Lin, Y.; Yu, W.; Wang, H.; Kwong, S.; Zhang, J. Automatic niching differential evolution with contour prediction approach for multimodal optimization problems. IEEE Trans. Evol. Comput. 2020, 24, 114–128. [Google Scholar] [CrossRef]
  24. Wang, Z.J.; Zhou, Y.R.; Zhang, J. Adaptive estimation distribution distributed differential evolution for multimodal optimization problems. IEEE Trans. Cybern. 2020, 52, 6059–6070. [Google Scholar] [CrossRef]
  25. Li, H.; Liao, B.; Li, J.; Li, S. A Survey on Biomimetic and Intelligent Algorithms with Applications. Biomimetics 2024, 9, 453. [Google Scholar] [CrossRef]
  26. Zhu, Q.; Wu, X.; Lin, Q.; Ma, L.; Li, J.; Ming, Z.; Chen, J. A survey on evolutionary reinforcement learning algorithms. Neurocomputing 2023, 556, 126628. [Google Scholar] [CrossRef]
  27. Tian, Y.; Zhu, W.; Zhang, X.; Jin, Y. A practical tutorial on solving optimization problems via PlatEMO. Neurocomputing 2023, 518, 190–205. [Google Scholar] [CrossRef]
  28. Wang, Z.J.; Zhan, Z.H.; Kwong, S.; Jin, H.; Zhang, J. Adaptive granularity learning distributed particle swarm optimization for large-scale optimization. IEEE Trans. Cybern. 2020, 51, 1175–1188. [Google Scholar] [CrossRef]
  29. Jiang, H.; Lu, M.; Tian, Y.; Qiu, J.; Zhang, X. An evolutionary algorithm for solving capacitated vehicle routing problems by using local information. Appl. Soft Comput. 2022, 117, 108431. [Google Scholar] [CrossRef]
  30. Wang, Z.J.; Zhan, Z.H.; Lin, Y.; Yu, W.; Yuan, H.; Gu, T.; Kwong, S.; Zhang, J. Dual-strategy differential evolution with affinity propagation clustering for multimodal optimization problems. IEEE Trans. Evol. Comput. 2017, 22, 894–908. [Google Scholar] [CrossRef]
  31. Huang, Y.B.; Wang, Z.J.; Zhang, Y.H.; Wang, Y.; Kwong, S.; Zhang, J. Wireless sensor networks-based adaptive differential evolution for multimodal optimization problems. Appl. Soft Comput. 2024, 158, 111541. [Google Scholar] [CrossRef]
  32. Liang, S.M.; Wang, Z.J.; Huang, Y.B.; Zhan, Z.H.; Kwong, S.; Zhang, J. Niche Center Identification Differential Evolution for Multimodal Optimization Problems. Inf. Sci. 2024, 678, 121009. [Google Scholar] [CrossRef]
  33. Wang, Z.J.; Yang, Q.; Zhang, Y.H.; Cheng, S.H.; Wang, Y.G. Superiority combination learning distributed particle swarm optimization for large-scale optimization. Appl. Soft Comput. 2023, 136, 110101. [Google Scholar] [CrossRef]
  34. He, J.; Yao, X. Average drift analysis and population scalability. IEEE Trans. Evol. Comput. 2016, 21, 426–439. [Google Scholar] [CrossRef]
  35. He, J.; Lin, G. Average convergence rate of evolutionary algorithms. IEEE Trans. Evol. Comput. 2015, 20, 316–321. [Google Scholar] [CrossRef]
  36. He, J.; Chen, T.; Yao, X. On the easiest and hardest fitness functions. IEEE Trans. Evol. Comput. 2014, 19, 295–305. [Google Scholar] [CrossRef]
  37. He, J.; Zhou, Y. Drift analysis with fitness levels for elitist evolutionary algorithms. Evol. Comput. 2024. [Google Scholar] [CrossRef] [PubMed]
  38. Roostapour, V.; Neumann, A.; Neumann, F.; Friedrich, T. Pareto optimization for subset selection with dynamic cost constraints. Artif. Intell. 2022, 302, 103597. [Google Scholar] [CrossRef]
  39. Doerr, B.; Neumann, F. A survey on recent progress in the theory of evolutionary algorithms for discrete optimization. ACM Trans. Evol. Learn. Optim. 2021, 1, 1–43. [Google Scholar] [CrossRef]
  40. Zhou, Z.-H.; Yu, Y.; Qian, C. Evolutionary Learning: Advances in Theories and Algorithms; Springer: Singapore, 2019. [Google Scholar]
  41. Qian, C. Can Evolutionary Clustering Have Theoretical Guarantees. IEEE Trans. Evol. Comput. 2023. [Google Scholar] [CrossRef]
  42. Xia, X.; Zhou, Y. On the effectiveness of immune inspired mutation operators in some discrete optimization problems. Inf. Sci. 2018, 426, 87–100. [Google Scholar] [CrossRef]
  43. Xia, X.; Zhou, Y. Performance analysis of ACO on the quadratic assignment problem. Chin. J. Electron. 2018, 27, 26–34. [Google Scholar] [CrossRef]
  44. Xia, X.; Peng, X.; Liao, W. On the analysis of ant colony optimization for the maximum independent set problem. Front. Comput. Sci. 2021, 15, 154329. [Google Scholar] [CrossRef]
  45. Delorme, X.; Gandibleux, X.; Degoutin, F. Evolutionary, constructive and hybrid procedures for the bi-objective set packing problem. Eur. J. Oper. Res. 2010, 204, 206–217. [Google Scholar] [CrossRef]
  46. Chaurasia, S.N.; Sundar, S.; Singh, A. A hybrid evolutionary approach for set packing problem. Opsearch 2015, 52, 271–284. [Google Scholar] [CrossRef]
  47. Chaurasia, S.N.; Kim, J.H. An evolutionary algorithm based hyper-heuristic framework for the set packing problem. Inf. Sci. 2019, 505, 1–31. [Google Scholar] [CrossRef]
  48. Peng, X. Performance analysis of (1+1) EA on the maximum independent set problem. In Proceedings of the International Conference on Cloud Computing and Security, Nanjing, China, 13–15 August 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 448–456. [Google Scholar]
  49. Wegener, I. Methods for the analysis of evolutionary algorithms on pseudo-Boolean functions. In Evolutionary Optimization; Sarker, R., Mohammadian, M., Yao, X., Eds.; Kluwer, Springer: New York, NY, USA, 2002; pp. 349–369. [Google Scholar]
  50. Wang, Z.J.; Zhan, Z.H.; Li, Y.; Kwong, S.; Jeon, S.W.; Zhang, J. Fitness and distance based local search with adaptive differential evolution for multimodal optimization problems. IEEE Trans. Emerg. Top. Comput. Intell. 2023, 7, 684–699. [Google Scholar] [CrossRef]
  51. Xia, X.; Zhuang, H.; Wang, Z.; Chen, Z. Two-stage heuristic algorithm with pseudo node-based model for electric vehicle routing problem. Appl. Soft Comput. 2024, 165, 112102. [Google Scholar] [CrossRef]
  52. Li, A.D.; Xue, B.; Zhang, M. Improved binary particle swarm optimization for feature selection with new initialization and search space reduction strategies. Appl. Soft Comput. 2021, 106, 107302. [Google Scholar] [CrossRef]
  53. Zhan, Z.H.; Wang, Z.J.; Jin, H.; Zhang, J. Adaptive distributed differential evolution. IEEE Trans. Cybern. 2020, 50, 4633–4647. [Google Scholar] [CrossRef] [PubMed]
  54. Neuwohner, M. The limits of local search for weighted k-set packing. Math. Program. 2024, 206, 389–427. [Google Scholar] [CrossRef]
Figure 1. Instance I s p with s = 3 .
Figure 1. Instance I s p with s = 3 .
Biomimetics 09 00586 g001
Figure 2. The global optimum of I s p with s = 3 .
Figure 2. The global optimum of I s p with s = 3 .
Biomimetics 09 00586 g002
Figure 3. A 2-optimum of I s p with s = 3 .
Figure 3. A 2-optimum of I s p with s = 3 .
Biomimetics 09 00586 g003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, Y.; Xia, X.; Wang, Z.; Peng, X.; Zhang, J.; Liao, W. Set Packing Optimization by Evolutionary Algorithms with Theoretical Guarantees. Biomimetics 2024, 9, 586. https://doi.org/10.3390/biomimetics9100586

AMA Style

Jin Y, Xia X, Wang Z, Peng X, Zhang J, Liao W. Set Packing Optimization by Evolutionary Algorithms with Theoretical Guarantees. Biomimetics. 2024; 9(10):586. https://doi.org/10.3390/biomimetics9100586

Chicago/Turabian Style

Jin, Youzhen, Xiaoyun Xia, Zijia Wang, Xue Peng, Jun Zhang, and Weizhi Liao. 2024. "Set Packing Optimization by Evolutionary Algorithms with Theoretical Guarantees" Biomimetics 9, no. 10: 586. https://doi.org/10.3390/biomimetics9100586

APA Style

Jin, Y., Xia, X., Wang, Z., Peng, X., Zhang, J., & Liao, W. (2024). Set Packing Optimization by Evolutionary Algorithms with Theoretical Guarantees. Biomimetics, 9(10), 586. https://doi.org/10.3390/biomimetics9100586

Article Metrics

Back to TopTop