Next Article in Journal
Natural Disaster Information System (NDIS) for RPAS Mission Planning
Previous Article in Journal
Study of the Optimal YOLO Visual Detector Model for Enhancing UAV Detection and Classification in Optoelectronic Channels of Sensor Fusion Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cooperative Task Allocation for Unmanned Aerial Vehicle Swarm Using Multi-Objective Multi-Population Self-Adaptive Ant Lion Optimizer

1
National Innovation Institute of Defense Technology, Beijing 100071, China
2
Academy of Military Sciences, Beijing 100091, China
*
Author to whom correspondence should be addressed.
Drones 2025, 9(11), 733; https://doi.org/10.3390/drones9110733
Submission received: 11 September 2025 / Revised: 20 October 2025 / Accepted: 20 October 2025 / Published: 23 October 2025
(This article belongs to the Section Artificial Intelligence in Drones (AID))

Highlights

What are the main findings?
  • A double-layer encoding mechanism and adaptive penalty function were designed for constraint handling.
  • A multi-objective multi-population self-adaptive ant lion optimizer is proposed.
What are the implications of the main findings?
  • The coupling constraints imposed by heterogeneous tasks, such as task sequencing and time-window requirements, are addressed.
  • The diversity of task-allocation outcomes is enhanced.

Abstract

The rational allocation of tasks is a critical issue in enhancing the mission execution capability of unmanned aerial vehicle (UAV) swarms, which is difficult to solve exactly in polynomial time. Evolutionary-algorithm-based approaches are among the popular methods for addressing this problem. However, existing methods often suffer from insufficiently rigorous constraint settings and a focus on single-objective optimization. To address these limitations, this paper considers multiple types of constraints—including temporal constraints, time window constraints, and task integrity constraints—and establishes a model with optimization objectives comprising task reward, task execution cost, and task execution time. A multi-objective multi-population self-adaptive ant lion optimizer (MMSALO) is proposed to solve the problem. In MMSALO, a sparsity-based selection mechanism replaces roulette wheel selection, effectively enhancing the global search capability. A random boundary strategy is adopted to increase the randomness and diversity of ant movement around antlions, thereby improving population diversity. An adaptive position update strategy is employed to strengthen exploration in the early stages and exploitation in the later stages of the algorithm. Additionally, a preference-based elite selection mechanism is introduced to enhance optimization performance and improve the distribution of solutions. Finally, to handle complex multiple constraints, a double-layer encoding mechanism and an adaptive penalty strategy are implemented. Simulation experiments were conducted to validate the proposed algorithm. The results demonstrate that MMSALO exhibits superior performance in solving multi-task, multi-constraint task-allocation problems for UAV swarms.

1. Introduction

The rapid development of unmanned aerial vehicle (UAV) technology has garnered widespread global attention. As a technology with immense potential, UAVs have demonstrated significant application prospects in military, civilian, and commercial domains [1,2,3,4]. With advancements in technology and reductions in cost, UAVs are gradually transitioning from single-platform applications to swarm-based applications [5]. A UAV swarm system refers to a system where multiple UAVs work collaboratively, sharing information and distributing tasks to accomplish complex missions [6]. This system offers numerous advantages, such as high task-execution efficiency, strong reliability, and high flexibility. However, a key challenge facing UAV swarm systems is the rational allocation of tasks to achieve optimal system performance. The task-allocation problem for UAV swarms involves multiple factors, including task priority, UAV capabilities, and constraints, making it complex and diverse. It is a typical NP-hard problem, and the main difficulties in addressing it are problem modeling and model solving [7].
In terms of problem modeling, the research on multi-task multi-constraint task-allocation problems is relatively limited, and there has been insufficient attention paid to multi-objective optimization algorithms. One of the significant advantages of UAV swarms over individual UAVs is their ability to perform diverse and complex tasks. Therefore, multi-task types are a key focus in the study of UAV swarm task allocation, and the task sequence constraints brought by multi-task types must also be taken into account. Some studies on this subject did not consider multiple task types and temporal constraints of tasks, lacking rigor [8,9]. Due to the complexity and diversity of UAV swarm task allocation, evaluation criteria for task-allocation schemes should not be overly singular. Thus, traditional single-objective optimization strategies are not applicable, and a multi-objective optimization strategy is more appropriate. Other studies converted multi-objective optimization problems into single-objective optimization problems by weighting, which leads to the solution schemes being subjectively influenced by the weights. Therefore, it is not conducive to comprehensive decision-making [10,11].
In response to the aforementioned issues, this paper describes the collaborative multi-task-allocation model as a multiple traveling salesman problem. The model considers multiple task types, temporal constraints, and multi-objective optimization. First, UAVs can perform reconnaissance, supplies delivery, and assessment tasks, and a UAV swarm is required to execute all three types of tasks for all targets. Second, the three tasks on each target must be strictly carried out in the order from reconnaissance, supplies delivery, to assessment, and some tasks have specific requirements for the time period of execution. Finally, three optimization objectives need to be considered: task reward, task execution cost, and task execution time.
Generally, solution approaches can be categorized into two main types: exact methods and approximate methods. Exact methods yield optimal solutions at the expense of higher computational time, whereas approximate methods obtain satisfactory solutions within limited time.
Exact methods, such as dynamic programming [12] and branch-and-bound algorithms [13], can provide accurate references for the analysis of task-allocation problems. Dynamic programming is a global optimization method for solving multi-step optimization problems. Task allocation, which involves optimizing the execution sequence of multiple tasks by UAVs, is well-suited for dynamic programming. To address the multi-area reconnaissance coverage problem, Xie et al. [14] employed dynamic programming to recursively determine the optimal visiting sequence and entry–exit positions for multiple regions. In scenarios involving cooperative multi-UAV operations in adversarial environments, Alighanbari et al. [15] applied a two-step look-ahead strategy to derive action plans for UAVs, achieving optimal solutions with reduced computation time. Branching methods, a class of tree search strategies, primarily include branch and bound, branch and price, and branch and cut. The branch-and-bound method prunes suboptimal branches during the solution of mixed-integer linear programming models, thereby avoiding extensive unnecessary searches. Casbeer et al. [16] utilized a branch and price approach to solve temporally coupled problems. By leveraging the sparsity of constraint distributions, they decomposed the problem into smaller subproblems via column generation, progressively converging to the optimal solution while significantly reducing computational effort. Subramanyam et al. [17] formulated a robust linear programming model for heterogeneous vehicle routing under worst-case scenarios and employed a branch and cut algorithm to dynamically introduce constraints, thereby approaching the optimal solution and obtaining a tight lower bound for the problem.
In contrast, approximate methods do not aim to obtain the optimal solution. Instead, they utilize heuristic information to guide the search process in order to achieve satisfactory solutions within polynomial time. In the field of task allocation for UAV swarms, evolutionary algorithms have emerged as popular and efficient approximate methods. Commonly used algorithms include particle swarm optimization (PSO) [18], genetic algorithms (GA) [19], and ant colony optimization (ACO) [20]. Researchers have tailored these algorithms for specific problem environments through various improvements, such as strategy optimization [21], algorithm hybridization [22], and enhanced encoding schemes [23]. These modifications have effectively advanced the development of UAV swarm task allocation. Based on the context of multi-UAV cooperative reconnaissance, Wang et al. [21] further enhanced the population diversity and convergence capability of the algorithm by introducing a constraint-based particle initialization strategy, an experience-pool-based particle position reconstruction strategy, and a time-varying parameter adjustment strategy into the PSO. Liu et al. [7] employed a greedy selection operator and introduced elite opposition-based learning, constructing opposition solutions from feasible solutions of the current problem to increase population diversity, thereby improving the convergence speed of the GA. He et al. [24] first performed a preliminary search on the data using a simplified artificial fish swarm algorithm to generate an initial pheromone distribution, which was then transferred to an improved ACO for solution. This approach reduced the initial search blindness of the ACO and accelerated the algorithm’s convergence.
Ant lion optimizer (ALO) [25] is a swarm intelligence algorithm proposed by Seyedali Mirjalili in 2015. It simulates the foraging behavior of antlions trapping ants. The main steps include random walking of ants, setting traps by antlions, ants falling into traps, antlions preying on ants, and antlions rebuilding traps. It has been widely applied in various fields, including computer science, engineering, energy, and operations management due to its ease of implementation, strong scalability, and high flexibility [26].
This paper proposes a multi-objective multi-population self-adaptive ant lion optimizer (MMSALO) to efficiently solve the model. To enhance global exploration, MMSALO replaces the roulette-wheel operator with a sparsity selection mechanism. A random boundary strategy redefines the margins of the random walk that each ant performs around an antlion, thereby increasing the population diversity. An adaptive position update schedule then biases early iterations toward sparsity-selected antlions to favor exploration, whereas later iterations pivot around the elite antlion to intensify exploitation. Finally, a preference-based elite selection scheme drives each ant to track the best individual discovered thus far along distinct objective dimensions, which strengthens the algorithm’s convergence behavior and improves the spread of the resulting Pareto front. This study implemented a comprehensive experiment comparing MMSALO with four representative algorithms on the DTLZ test function suite, multi-objective 0-1 knapsack problem (MOKP), and two task-allocation models. To address the complex, multiple constraints inherent in the task-allocation models, a dual-layer encoding scheme, and an adaptive penalty strategy are proposed. The results demonstrate the effectiveness and superiority of the proposed approach.
The remainder of this paper is organized as follows. Section 2 formulates the task-allocation model for UAV swarms. Section 3 presents the MMSALO algorithm in detail. Section 4 describes the simulations and results. Finally, the conclusions of this study are presented in Section 5.

2. Task-Allocation Model for UAV Swarms

2.1. Scenario Description

In the scenario, multiple UAVs and multiple targets are present, with each target involving individuals in need of supplies assistance. To fulfill rescue requirements, the UAV swarm must depart from their initial positions and, based on task-allocation results, sequentially conduct reconnaissance, supplies delivery, and assessment for all targets. Effective collaboration among the UAVs is essential to executing these complex missions. Accordingly, the following assumptions are established to analyze the cooperative task-allocation problem:
1.
All targets are stationary and their locations are known. The operational area is simplified to a two-dimensional region.
2.
All UAVs have limited range and onboard resources. Tasks such as supplies delivery consume these onboard resources. Since the drones operate at different altitudes, the possibility of collision between them is not considered.
3.
During task execution, temporal sequence constraints and time-window constraints must be considered.
Current methods for solving task-allocation problems are primarily categorized into centralized, distributed, and mixed approaches [27]. Centralized methods utilize a central node to collect information and assign tasks, thereby enabling the full coordination of various resources and reflecting global characteristics, suitable for scheduling based on the integration of known information prior to task execution. Compared with distributed methods, centralized methods can effectively avoid task competition among UAVs, eliminating the need for additional resources to coordinate conflicts. In contrast to hybrid methods, centralized methods offer more concentrated resources, and the unified scheduling by the central node can achieve better global scheduling outcomes. Therefore, this study adopted a centralized allocation strategy.

2.2. Basic Model Definition

In this section, three fundamental models are defined: the target model, the task model, and the UAV model.

2.2.1. Modeling of Target and Task

This study primarily investigated static targets. Figure 1 illustrates a schematic diagram of UAV swarm task allocation, where “RE” denotes reconnaissance tasks, “AT” represents supplies-delivery tasks, and “AS” indicates assessment tasks. There are N T targets with known location information, and each target T j ( j = 1 , 2 , , N T ) comprises three types of tasks ( N t y p e = 3 ) : reconnaissance, supplies delivery, and assessment. The three tasks on each target must be executed strictly in order from reconnaissance, to supplies delivery, to assessment. Let M k ( k = 1 , 2 , , N M ) denote the set of tasks for the k-th task and N M = N T N t y p e represent the total number of tasks.
Table 1 provides detailed attributes of targets and tasks. For targets, X j T denotes the location of the target. For tasks, V k M represents the value of the task; R k represents the demand of the supplies-delivery task for onboard resources; [ t k s , t k e ] indicates the task schedule, where t k s and t k e represent the start and completion times, respectively; t k d o denotes the task execution time; [ E T k , L T k ] represents the task time window, where E T k is the earliest allowable start time, and L T k is the latest allowable completion time; S k denotes the probability of task failure.

2.2.2. Modeling of UAVs

Table 2 presents the detailed attributes of UAVs. Specifically, X i U denotes the location of the UAV, v i represents the speed of the UAV, P i k is the task capability parameter of the UAV, D i U represents the maximum range of the UAV, V i U refers to the value of the UAV, and L i signifies the amount of onboard resources of the UAV.
Based on the target model, task model, and UAV model, the relationship between the UAV and the task is described using x i k as the decision variable, which is defined as follows:
x i k = 1 , allocate task M k to UAV U i 0 , else

2.3. Constraints

Based on the three fundamental models, this section establishes the model constraints. All constraints can be categorized into two parts: UAV constraints and task constraints. UAV constraints are derived from the limitations of UAV capabilities, including collaboration constraints, range constraints, and supplies-delivery-capability constraints. Task constraints are related to the requirements of the task, such as integrity constraints, temporal sequence constraints, and time-window constraints.
Collaborative constraints: A UAV can perform only one task for a single target at any given moment.
i = 1 N V x i k = 1 , k N M
Range constraints: The flight distance of the UAV must not exceed its maximum range.
m a x x i k t k e k = 1 N M x i k t k d o v i D i U , i N V
where m a x x i k t k e denotes the completion time of the last task assigned to the UAV i, k = 1 N M x i k t k d o represents the total time taken by the UAV i to complete all its tasks, and m a x x i k t k e k = 1 N M x i k t k d o corresponds to the actual flight duration of the UAV i. The actual flight distance of the UAV i is then calculated as m a x x i k t k e k = 1 N M x i k t k d o v i .
Supplies-delivery-capability constraints: The amount of supplies delivered by each UAV shall not exceed its onboard resources.
k = 1 N M x i k R k L i , i N V
Integrity constraints: Each task on every target must be executed and only executed once.
i = 1 N V k = 1 N M x i k = N M
Temporal sequence constraints: Tasks on each target must be executed in the sequence of reconnaissance, supplies delivery, and assessment.
t r e c o n n a i s s a n c e e t s u p p l i e s d e l i v e r y s t s u p p l i e s d e l i v e r y e t a s s e s s m e n t s
Time-window constraints: Some tasks must be executed within specific time windows, which restrict the earliest start time and latest end time of the tasks.
E T k t k s < t k e L T k , k N M

2.4. Optimization Objectives

In the task-allocation model, three optimization objectives are defined to evaluate various allocation schemes: task rewards, task execution costs, and task execution time.
Task rewards: Task rewards refer to the value obtained upon successful execution of a task. Rewards are categorized into three types based on task types: reconnaissance rewards, supplies-delivery rewards, and assessment rewards. Reconnaissance rewards refer to the amount of information acquired by the UAV about the target. The purpose of reconnaissance is to verify the status of the target and environmental information, thereby reducing uncertainty. The supplies-delivery reward is associated with the accurate transportation of required supplies to designated locations, ensuring effective resource replenishment. The assessment reward pertains to the evaluation of task effectiveness after supplies delivery, examining both delivery accuracy and execution performance. These three types of rewards are associated with task value, UAV task capability parameters, and the probability of task failure. To minimize all optimization objectives, the reward optimization function was set as the loss value of the objective, with M R representing the maximum value of the objective. A lower reward objective value indicates better task completion.
F r e w a r d = M R i = 1 N V k = 1 N M 1 i = 1 N V 1 x i k P i k 1 S k V k M
In the equation, 1 i = 1 N V 1 x i k P i k 1 S k represents the probability of successfully performing task k.
Task execution costs: Costs incurred by UAVs for reconnaissance, supplies delivery, and assessment tasks are defined as task execution costs. Task execution costs are related to the UAVs’ value and the probabilities of task failure.
F c o s t = i = 1 N V k = 1 N M x i k S k V i U
Task execution time: The task execution time is defined as the total time required for the UAV swarm to complete all tasks, that is, the latest task completion time in the UAV task timetable.
F t i m e = m a x x i k t k e , i , k N V , N M

3. Method for Multi-Objective Optimization

This section first presents a concise overview of both ALO and multi-objective ant lion optimizer (MOALO) and then elaborates on the improvement strategies implemented in MMSALO.

3.1. ALO

ALO mimics the process of antlions preying on ants in nature. The main steps of the algorithm include the random walking of ants, trap setting by antlions, ants falling into the traps, antlions preying on ants, and trap reconstruction by antlions.
The random walk of ants is demonstrated as follows:
X t = 0 , c s r t 1 , c s r t 2 , , c s r t T
where t denotes the current iteration number, T represents the maximum number of iterations, X t indicates the position of the random walk, c s represents the cumulative sum of the random walk step lengths, and r t takes the value of 1 or −1 with equal probability.
During the iterative process, antlions set traps in the search space to prey on ants, and the random walk of ants is influenced by these traps.
c i t = Antlion j t + c t
d i t = Antlion j t + d t
where c t denotes the lower bounds of the ant colony’s individual values across all dimensions at the t-th iteration, d t represents the upper bounds of the ant colony’s individual values across all dimensions at the t-th iteration, and Antlion j t indicates the position of the j-th antlion at the t-th iteration.
After constructing pits, antlions preys on ants that wander randomly. Once an ant enters an antlion’s pit, the antlion will eject sand towards the center of the pit to reduce the ant’s range of movement. The specific process is as follows:
c t = c I
d t = d I
I = 10 w × t T
where c represents the lower bounds of the values of each dimension for individual ants, and d represents the upper bounds of the values of each dimension for individual ants. The value of w is as follows:
w = 0 , t 0.1 T 2 , 0.1 T < t 0.5 T 3 , 0.5 T < t 0.75 T 4 , 0.75 T < t 0.9 T 5 , 0.9 T < t 0.95 T 6 , t > 0.95 T
Antlions that are better adapted to the environment possess stronger predatory capabilities. To simulate this scenario, ALO retains the antlion with the optimal fitness obtained in each iteration as the elite antlion. Each ant then performs a random walk around the elite antlion. Additionally, each ant selects another antlion based on its fitness. This selection is done using the roulette-wheel method. Once selected, the ant performs a random walk around this chosen antlion.
After the ants update their positions through random walks, they are considered to be preyed upon by the selected antlion. The position update is described as follows:
Ant i t = R e t + R r t 2
where R e t represents the vector obtained by the ant performing random walks around the elite antlion during the t-th iteration, while R r t denotes the vector obtained by the ant performing random walks around the roulette-wheel-selected antlion during the t-th iteration.
When an ant slides into the bottom of a pit and is captured by an antlion, the antlion assesses whether this location is more suitable for capturing ants than its original position. If it is deemed more advantageous, the antlion relocates to this position and reconstructs its pit. The specific process is as follows:
Antlion j t = Ant i t , f Ant i t > f Antlion j t
where Ant i t denotes the position of the i-th ant at the t-th iteration.
At the end of the iteration, the antlion population is updated. After which, the antlion with the optimal fitness value is designated as the elite.

3.2. MOALO

Pareto optimality is a core concept in multi-objective optimization. In multi-objective optimization, given two solutions x and y in the decision space, x is said to dominate y if x is superior to y in at least one objective and not inferior to y in the remaining objectives. A solution x is defined as a non-dominated solution or a Pareto optimal solution when no other solution in the decision space dominates it. The set of all Pareto optimal solutions in the decision space is the Pareto set (PS), and the corresponding set of objective vectors is the Pareto front (PF).
The specific algorithm for maintaining the PS is presented in Algorithm 1. In each iteration, after updating all individuals in the population, the algorithm merges the current population with the PS. Subsequently, it employs the Pareto dominance strategy to select non-dominated solutions, which are then stored in the PS.   
Algorithm 1 PS Maintenance algorithm
Input: 
Original Pareto set P S ; Population P
Output: 
New Pareto set P S n e w
1:
Combine P S and P into P S t m p ;
2:
for  x in P S  do
3:
    if  x is not dominated by any solution in P S t m p  then
4:
        Designate x as a non-dominated solution;
5:
    end if
6:
end for
7:
Preserve all non-dominated solutions in P S n e w ;
   MOALO [28] incorporates the concept of sparsity to determine the density of the location occupied by a non-dominated solution. The specific algorithm for calculating sparsity is presented in Algorithm 2, where d i s t a n c e ( x , y ) denotes the Euclidean distance between x and y . In brief, the sparsity of a non-dominated solution is defined as the number of other non-dominated solutions within its local neighborhood.
Algorithm 2 Sparsity calculation
Input: 
Pareto set P S ; The number of optimization objectives O b j
Output: 
The sparsity S of all solutions in P S
1:
Calculate the neighborhood radius r for all objectives;
2:
S is initialized to 0;
3:
for  x in P S  do
4:
    for  y in P S  do
5:
        if  d i s t a n c e ( x , y ) < r in all objectives then
6:
            S [ x ] = S [ x ] + 1 ;
7:
        end if
8:
    end for
9:
end for
After obtaining the sparsity of all solutions in PS, MOALO employs a sparsity-based roulette wheel to select solutions from the PS as roulette-wheel antlions. The specific algorithm for the sparsity-based roulette-wheel selection is presented in Algorithm 3.
Algorithm 3 Sparsity roulette algorithm
Input: 
The sparsity S of all solutions in P S
Output: 
The antlion index selected by the roulette wheel based on sparsity
  1:
The reciprocal of the sparsity is calculated and stored in X;
  2:
for x in X do
  3:
    The cumulative sum from X [ 0 ] to x is calculated and stored in W;
  4:
end for
  5:
The sum of the reciprocals of all sparsities is calculated and then multiplied by a random number between 0 and 1 to obtain the selection probability p;
  6:
for w in W do
  7:
    if  w > p  then
  8:
        The index value of w is selected as the antlion index;
  9:
        Terminate the loop;
10:
    end if
11:
end for
   When the number of non-dominated solutions in PS exceeds its maximum capacity, MOALO also employs the sparsity-based roulette-wheel selection to remove the surplus solutions. The method follows the same procedure as Algorithm 3, except that the data stored in X are replaced by the sparsity itself.

3.3. MMSALO

3.3.1. Sparsity Selection Mechanism

The conventional MOALO employs a sparsity-based roulette wheel to maintain the Pareto set and to select ant lions, thereby injecting stochasticity into the search. However, this roulette mechanism occasionally retains inferior solutions, which hinders global exploration. Consequently, the present study introduces a deterministic selection to replace the original roulette wheel. This method is guided strictly by sparsity magnitude, which significantly enhances the algorithm’s global search capability. The modified PS maintenance algorithm and antlion selection algorithm are presented in Algorithm 4 and Algorithm 5, respectively.   
Algorithm 4 Modified PS maintenance algorithm
Input: 
Original Pareto set P S ; Population P; P S capacity c; The sparsity S of all solutions in P S
Output: 
New Pareto set P S n e w
  1:
Combine P S and P into P S t m p ;
  2:
for  x in P S  do
  3:
    if  x is not dominated by any solution in P S t m p  then
  4:
        Designate x as a non-dominated solution;
  5:
    end if
  6:
end for
  7:
Preserve all non-dominated solutions in P S n e w ;
  8:
while the number of solutions in P S n e w > c do
  9:
    Delete the solution with the highest sparsity;
10:
    Delete the corresponding sparsity from S;
11:
end while
Algorithm 5 Sparsity selection algorithm
Input: 
The sparsity S of all solutions in P S
Output: 
The antlion index selected based on sparsity
1:
Find the minimum value s in S;
2:
The index value of s is selected as the antlion index;

3.3.2. Random Boundary Strategy

In the classical MOALO, the value of I decreases with increasing iteration counts, exhibiting a piecewise linear trend. Moreover, in each iteration, all ants traverse the same boundaries, which reduces the diversity of the algorithm and is unfavorable for locating the global optimum. To address this issue, this study proposes a stochastic boundary strategy that modifies the calculation of I as follows to enhance the diversity of ants’ movement around antlions:
I = 10 w t T 1 + sin π 2 t T 6 T · r a n d
where r a n d denotes a random number between 0 and 1. Because sin π 2 t T 6 T exhibits a nonlinear increasing trend within the range of [ 0.5 , 0.5 ] as t increases, 1 + sin π 2 t T 6 T · r a n d displays a stochastic nonlinear increasing trend within the range of [ 0.5 , 1.5 ] . This introduces a degree of randomness into the variation process of I.
Overall, the value of I exhibits a nonlinear increasing trend with a certain degree of randomness as the number of evolutionary generations increases. Consequently, the boundary size decreases nonlinearly with a certain degree of randomness as the number of evolutionary generations progresses. This phenomenon enhances the randomness and diversity of the ants’ movement around antlions, thereby improving the diversity of the ant population to some extent.

3.3.3. Adaptive Position Update Strategy

The standard MOALO employs the method of calculating the average values of two types of random walks, R e t and R r t , to balance exploration and exploitation capabilities. However, it overlooks the issue that the proportions of exploration and exploitation should vary at different stages. In this study, we improved the proportionality coefficients of R e t and R r t , allowing the two types of walks to have different weights at various stages. In the early stages of the algorithm, the random walk around the sparsity-selected antlion is dominant, while in the later stages, the random walk around the elite antlion prevails. The improved position update strategy is described as follows:
Ant i t = 1 + t T · r a n d R e t + 1 t T · r a n d R r t 2
Through these improvements, the ants primarily wandered around the sparsity-selected antlion in the early stages. At the same time, the algorithm gradually enhanced the directional influence of the elite antlion. In the later stages, the ants mainly wandered around the elite antlion, with a certain degree of randomness retained. This approach effectively enhanced the exploration capability of the algorithm in the early stages and the exploitation capability in the later stages.

3.3.4. Preference-Based Elite Selection Mechanism

As shown in Figure 2, MMSALO divides the original population into multiple subpopulations based on the number of optimization objectives. Each subpopulation focuses on a single optimization objective and selects the solution with the best performance on that objective as its elite antlion. During the iterative process, each ant first selects an antlion using a sparsity selection mechanism. Then, it performs random walks around two key targets: the elite antlion of its own subpopulation and the sparsity-selected antlion. This process ultimately generates new individual solutions.    
By using a preference-based elite selection mechanism, each ant iteratively updates its search around the best individual for a distinct objective. This enhances the algorithm’s optimization capability and produces a more widely distributed set of solutions.

3.3.5. Complete Algorithm Framework

The MMSALO procedure is illustrated in Figure 3. The algorithm first initializes the ant population. The fitness of each ant is then calculated, and Pareto solutions are selected from the population based on Pareto dominance relationships and preserved in the PS. The sparsity of the antlions is subsequently computed, and elite antlions are selected for each optimization objective according to a preference-based elite selection mechanism. During the iterative process, each initialized ant performs a random walk around the elite antlion of its own population and an antlion selected by sparsity, employing a random boundary strategy. New ants are generated using an adaptive position update strategy. The fitness of the new ant population is calculated, and these ants are incorporated into the archive, followed by an update of the archive and elite antlions. Finally, upon reaching the maximum number of iterations, the task-allocation scheme constructed by the antlions is output.    
An analysis of the time complexity of MMSALO is presented as follows. Let D denote the dimensionality of the decision variables, N the population size, and T the maximum number of iterations. The time complexity of the ant random walk iteration is O ( N × D × T ) , while the time complexity of computing the dominance relations and sparsity of solutions is O ( N 2 × T ) . In summary, the overall time complexity of MMSALO is O ( N × T × ( N + D ) ) .

4. Simulations and Results

Two experiments were designed to verify the effectiveness of the proposed MMSALO in solving the multi-task and multi-constraint task-allocation problem for UAV swarm. In experiment 1, the DTLZ test function suite and MOKP were employed to evaluate the distribution and convergence performance of MMSALO on both continuous and discrete problems, thereby validating its capability for solving multi-objective optimization problems. Experiment 2 employed a battlefield model to create a simulation environment and solved the multi-task and multi-constraint task-allocation model to verify the algorithm’s effectiveness. In the comparison, four algorithms were employed: MOALO, improved multi-objective particle swarm optimization (IMOPSO) [29], improved non-dominated sorting genetic algorithm-II (INSGA-II) [30], and improved non-dominated sorting genetic algorithm-III (INSGA-III) [31]. These algorithms are specifically devised for the task-allocation problem of UAV swarms. IMOPSO balances diversity and convergence by adaptive angular region partitioning and a dedicated reuse rule for infeasible solutions, which effectively mitigates premature convergence. INSGA-II reshapes the non-dominated sorting criterion through a “constraint-tolerance” mechanism, significantly improving both convergence and feasibility under intricate constraints. INSGA-III introduces an enhanced reference-point association strategy to promote uniform solution distribution and employs a dynamic elite-selection scheme to balance exploration and exploitation, enabling rapid and evenly distributed convergence to optimal solutions in high-dimensional objective spaces with complex constraints. The detailed parameters of these algorithms are listed in Table 3. Among these parameters, S e and S r denote the weighting coefficients for the ant’s random walk around the elite antlion and the roulette-wheel-selected antlion, respectively; w denotes the weight vector; P c and P m represent the crossover probability and mutation probability, respectively. Both experiments were conducted on a computer equipped with an Intel i7-12650H CPU (2.30 GHz) and 16.0 GB of memory.
The performance of all algorithms was evaluated using hyper volume (HV) [32]. The formulas for calculating HV is shown below:
H V = v o l u m e ( v P v )
where P represents the set of non-dominated solutions within the target vector set calculated by the algorithm, while volume refers to the hypervolume calculated from the vector to the reference point.

4.1. Simulation on the Test Functions

In the experiment, the test functions were employed to evaluate the performance of MMSALO in solving multi-objective optimization problems. Given that the problem model in this study was a three-objective model, the number of objective functions, M, was set to 3.
The algorithms were assessed on DTLZ1 and DTLZ3, while | x M | was set to 2 and 8. Because the dimensionality of the decision variables is n = M + | x M | 1 , the resulting problem dimensions are 4 and 10. The analytical expressions of the three-objective DTLZ1 and DTLZ3 are listed in Table 4.
   Given a set of n items and a set of m knapsacks, MOKP can be stated as Equation (23):
maximize f i ( x ) = j = 1 n p i j x j , i = 1 , , m subject to j = 1 n w i j x j c i , i = 1 , , m x = ( x 1 , , x n ) T 0 , 1 n
where p i j 0 is the profit of item j in knapsack i, w i j 0 is the weight of item j in knapsack i, and c i is the capacity of knapsack i. x j = 1 means that item j is selected and put in all the knapsacks. Since this study addressed a three-objective optimization problem, the parameter m was set to 3 to evaluate the performance of all algorithms when n was set to 200 and 800, respectively.
Each algorithm-MMSALO, MOALO, IMOPSO, INSGA-II, and INSGA-III-was independently executed 20 times with a population size of 100 and a PS size of 100. For DTLZ1 and DTLZ3, the maximum number of iterations was set to 1000, and the reference point was defined as (5, 5, 5). For MOKP, the maximum number of iterations was set to 100, with a reference point at (1, 1, 1).
The comparison of HV results obtained by five algorithms on DTLZ1, DTLZ3, and MOKP is presented in Table 5 and Figure 4. In the comparison of the mean values of the HV results, it can be observed that among the six test functions, MMSALO outperforms the other four algorithms on DTLZ1 (n = 10) and DTLZ3 and achieves the second-best performance on DTLZ1 (n = 4) and MOKP. This demonstrates the overall superiority of MMSALO on these test functions. Compared to MOALO, MMSALO yields better results across all test functions, indicating that the proposed improvements enhance the algorithm’s exploitation capability and global search ability, thereby ensuring solution diversity. Furthermore, in the comparison of the standard deviations of the HV results, it can be seen that the standard deviations of MMSALO and MOALO are comparable, with each performing better on certain test functions. This confirms that the proposed modifications do not compromise the stability of the algorithm.
Table 6 reports the p-values obtained from the Wilcoxon rank-sum test on HV. At α = 0.05, most p-values for the two MMSALO-optimal and two MMSALO-second-best metrics were significantly lower than α , demonstrating the statistically significant superiority of MMSALO over the comparison algorithms.
Table 7 presents the average time required for all algorithms to run the test functions once. It can be observed that the computation times of MMSALO and MOALO are within the same order of magnitude. Compared to MOALO, MMSALO requires a longer computation time because its PS reaches the capacity limit more rapidly, owing to the sparsity selection mechanism, thus necessitating additional PS update operations. However, on the MOKP (n = 200) instance, MMSALO exhibits a shorter runtime than does MOALO. This occurs because the test problem is relatively simple, leading both algorithms to approach the capacity limit almost simultaneously. Since MMSALO employs a direct sparsity selection mechanism while MOALO utilizes a sparsity-based roulette wheel selection, MMSALO achieves faster updates once the PS capacity is reached, resulting in a shorter overall computation time for MMSALO in this scenario.
Table 8 presents the HV and runtime results of the ablation studies conducted for MMSALO on the DTLZ test suite. In the table, “SS” denotes the sparsity selection mechanism component, “RB” represents the random boundary strategy component, “APU” indicates the adaptive position update strategy component, and “PBES” refers to the preference-based elite selection mechanism component. A comparative analysis between MMSALO and MOALO on MOKP is omitted here due to the negligible differences observed in their HV results. As can be observed, the performance contributions of the different components are relatively balanced, with each component demonstrating distinct advantages across different problems. Collectively, these components contribute to the enhanced overall performance of MMSALO.

4.2. Simulation on the Task Allocation Model

To address the intricate and multiple constraints introduced by the task-allocation model, we introduce a double-layer encoding scheme and an adaptive penalty strategy.

4.2.1. Double-Layer Encoding

The present study employs a real-number encoding method, representing individual positions as real-valued vectors. These vectors are then mapped to the discrete solution space of task allocation through a specific approach. This method allows individuals to search in the continuous domain space and converts the continuous encoding into practical task-allocation schemes through a particular decoding process.
The task-allocation scheme is represented by a task-allocation vector, which comprises two rows. The length of each row corresponds to the number of targets multiplied by the number of task types. Each element in the first row corresponds to a task assigned to a target and is encoded as a real number. The integer part of this value indicates the UAV assigned to execute the task, while the fractional part denotes the task’s priority, with a smaller value representing higher priority. The elements in the second row are also real numbers and are used to determine whether the corresponding task is reconnaissance, supplies delivery, or assessment. For tasks associated with the same target, they are classified as reconnaissance, supplies delivery, and assessment in ascending order of the values in the second row. Figure 5 illustrates the mapping between the algorithm encoding and the task-allocation result in a scenario where two UAVs are assigned to two targets. Specifically, given the task-allocation result [[1.2837, 2.8449, 2.5364, 1.0482, 2.4619, 1.2984], [1.3283, 2.2581, 1.9564, 1.1012]], the following interpretation applies: tasks 1, 4, and 6 on target 2 are assigned to UAV 1, with task priorities of 0.2837, 0.0482, and 0.2984, respectively. The execution sequence is task 4, followed by task 1 and then task 6. Similarly, tasks assigned to UAV 2 can be derived accordingly. For target 1, the second-row values for tasks 1, 2, and 3 are 1.3283, 2.2581, and 1.9564, respectively. Sorting these values in ascending order yields the following task sequence: task 1 (reconnaissance), task 3 (supplies delivery), and task 2 (assessment). The same logic applies to target 2.

4.2.2. Adaptive Penalty Function

The introduction of appropriate constraint-handling methods can effectively enhance the search efficiency of algorithms for feasible solutions. These methods achieve a balance between convergence, diversity, and feasibility by weighting constraint violations and coordinating the exploration of feasible and infeasible regions. Common constraint-handling methods include penalty function methods, constraint dominance methods, and learning-based methods [33]. Each of these methods has its own focus and can leverage its strengths in different aspects. Penalty function methods handle constraints by adding penalty terms for constraint violations to the objective function, ensuring that the solution set meets the constraints. Constraint dominance methods prioritize feasible solutions by comparing the feasibility of solutions with their dominance relationships. Learning-based methods utilize machine learning techniques to predict and improve the search process of the algorithm. In this study, an adaptive penalty function method was adopted, which dynamically adjusts the penalty value based on the degree of constraint conflicts during the search process. This approach avoids the need for parameter tuning and enhances the convergence speed of the algorithm.
The adaptive penalty function method converts the original constrained optimization problem into an unconstrained one by incorporating a penalty term into the objective function. This penalty term adjusts the optimization target value of the model based on the degree of constraint violation of the solution. In this study, an adaptive penalty function was designed to calculate the total number of constraint violations for a given solution. The total number of violations was then multiplied by a fixed value to serve as the penalty term. The modified objective function incorporating the penalty term is as follows:
F = F x , x X F x + r N , x X
where F denotes the objective function, x represents the solution found by the algorithm during the optimization process, X indicates the feasible region, r is the total number of constraint violations, and N is an adaptive value set here to the number of tasks. This allows the algorithm to increase the penalty for infeasible solutions in more complex problems, thereby driving the algorithm to find feasible solutions more rapidly.
Algorithm 6 shows the pseudocode of adaptive penalty function for calculating the total number of violations, which considers the UAV range constraint, UAV supplies-delivery capability constraint, task temporal sequence constraint, and task time window constraint.
Algorithm 6 Adaptive penalty function
Input: 
Task allocation solution A
Output: 
The total number of constraint violations r
  • r = 0 ;
  • for i in N V  do
  •     Obtain the task list TL of UAV i through A ;
  •     Calculate the flight distance d of UAV i using TL ;
  •     if  d > D i U  then
  •          r = r + 1 ;
  •     end if
  •     Obtain the number of supplies-delivery tasks l executed by UAV i through TL ;
  •     if  l > L i  then
  •          r = r + 1 ;
  •     end if
  • end for
  • for j in N T  do
  •     Obtain the task list TL on target j through A ;
  •     if  t r e c o n n a i s s a n c e e > t s u p p l i e s d e l i v e r y s  then
  •          r = r + 1 ;
  •     end if
  •     if  t s u p p l i e s d e l i v e r y e > t a s s e s s m e n t s  then
  •          r = r + 1 ;
  •     end if
  •     for task k in TL  do
  •         if  E T k > t k s  then
  •             r = r + 1 ;
  •         end if
  •         if  t k e > L T k  then
  •             r = r + 1 ;
  •         end if
  •     end for
  • end for

4.2.3. Task-Allocation Model Configuration and Experimental Results

In the simulation, two scenarios were established. The first scenario simulated the process of six UAVs performing tasks at 18 targets, while the second scenario simulated the process of eight UAVs performing tasks at 24 targets. The positions of the targets and UAVs were simplified as points. The total coverage area of the operational zone was 90,000 km2, represented by the x-axis coordinates (0, 300) and y-axis coordinates (0, 300). To reflect realistic conditions, UAVs with lower flight speeds were assigned a greater maximum range. The task capability parameter of a UAV was positively correlated with its value, and the probability of task failure was positively correlated with its value. The reconnaissance task was required to be completed within the first 4000 s to ensure the UAV swarm’s grasp of global information. The supplies-delivery task was required to be completed within the first 8000 s to avoid the issue of reconnaissance information becoming obsolete due to excessive task intervals. Additionally, a minimum time interval of 300 s was required to be maintained between the assessment and supplies-delivery tasks to ensure that personnel at the target location have received the supplies. Reconnaissance and assessment tasks did not consume any onboard resources, whereas every supplies-delivery task consumed exactly one unit. The speed, task capability parameter, and UAV value varied among individuals. Limited by their range and onboard resources, the UAVs were restricted in the number of tasks they can perform.
Table 9 and Table 10 list the detailed information of the UAVs, targets, and the tasks on these targets under Scenario 1.
Table 11 and Table 12 list the detailed information of the UAVs, targets, and the tasks on these targets under Scenario 2.
The population size was 100, the maximum number of iterations was 100, and the PS archive capacity was 100. MMSALO, MOALO, IMOPSO, INSGA-II, and INSGA-III were executed independently 20 times. The reference point was set to (108, 108, 108) for Scenario 1 and (192, 192, 192) for Scenario 2. The resulting HV values for the five algorithms are compared in Table 13. To facilitate HV computation, the objective values of task execution time were linearly scaled to [0, 108] in Scenario 1 and to [0, 192] in Scenario 2.
Table 13 and Figure 6 show that MMSALO achieves the highest average HV across all scenarios, demonstrating its superior capability in tackling the multi-task, multi-constraint assignment problem for UAV swarms. Moreover, MMSALO exhibits a significantly smaller HV standard deviation than does ƒMOALO, indicating enhanced stability. Because operational environments preclude repeated algorithmic runs to cherry-pick favorable solutions, this increased robustness renders MMSALO more valuable in practice.
Table 14 presents the p-values of the Wilcoxon rank-sum test. It can be observed that when α is set to 0.05, all p-values are significantly lower than α . This indicates that the performance of MMSALO is statistically significantly better than that of the comparison algorithms.
Table 15 presents the average computation time required for each algorithm to solve the task-allocation problem in a single run. As can be observed, the computation times of MMSALO and MOALO remain within the same order of magnitude. Due to the significantly higher complexity of the task-allocation problem compared to the test functions, the computation times of IMOPSO, INSGA-II, and INSGA-III increase by one order of magnitude. In contrast, the computation times of MMSALO and MOALO show no significant change. This demonstrates the favorable scalability of MMSALO and its efficient performance in addressing complex problems.
Table 16 presents the HV and runtime results obtained from the ablation experiments conducted by MMSALO on the task-allocation problem. It can be observed that “SS” and “APU” contribute significantly when addressing complex problems.
Collectively, MMSALO attains the highest HV value and outperforms the competing algorithms by a statistically significant margin. These results demonstrate that the sparsity selection mechanism, random boundary strategy, adaptive position update rule, and preference-based elite selection mechanism collectively enhance algorithmic performance, enabling MMSALO to provide an effective reference for analyzing multi-task, multi-constraint task-allocation problems in UAV swarms.

5. Conclusions

The UAV swarm collaborative task-allocation problem considered in this study is a multi-objective, multi-constrained optimization problem. The main conclusions drawn are as follows:
A multi-task, multi-constraint task-allocation model for UAV swarms was established. To meet the demands of practical applications, more detailed constraints need to be considered. Three optimization objectives were introduced to effectively evaluate various allocation schemes. A double-layer encoding mechanism and an adaptive penalty function method were designed to handle the key constraints. Compared with traditional penalty methods, this approach can dynamically adjust the penalty values based on the degree of constraint violations during the search process, thereby avoiding the need for parameter tuning and accelerating the convergence speed of the algorithm.
The MMSALO algorithm is proposed to effectively solve the model. MMSALO integrates a sparsity selection mechanism that significantly improves global exploration. A random boundary strategy increases the stochasticity and diversity of the ant walks around antlions, thereby enhancing population diversity. An adaptive position update strategy balances exploration in the early phase and exploitation in the latter phase. A preference-based elitist selection mechanism further strengthens the search capability and ensures better solution distribution. Simulation results demonstrate the feasibility of the allocation model and confirm that the resulting schedules accurately capture the characteristics of both UAVs and tasks. Comparative experiments against state-of-the-art algorithms verify the superior performance of the proposed MMSALO.
This study was limited to the task allocation for stationary targets using a centralized strategy and did not account for potential collisions among UAVs. In future work, we plan to design larger-scale scenarios and account for uncertainties in real-world environments. A distributed task0allocation approach will be adopted and dynamic task-reallocation strategies developed. Corresponding models and algorithms will also be investigated.

Author Contributions

Conceptualization, C.L. and Y.L.; methodology, G.L. and Y.L.; software, Q.Z. and Y.L.; validation, K.L. and Y.L.; formal analysis, Q.Z. and G.Y.; investigation, Q.Z. and Y.L.; resources, K.L.; data curation, Y.L.; writing—original draft preparation, C.L. and G.L.; writing—review and editing, C.L. and Y.L.; supervision, X.D.; project administration, G.Y. and X.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data will be made available on request.

DURC Statement

Current research is limited to intelligent decision-making, which is beneficial for enhancing the task-execution capability of UAV swarms and does not pose a threat to public health or national security. Authors acknowledge the dual-use potential of the research involving UAV task allocation and confirm that all necessary precautions have been taken to prevent potential misuse. As an ethical responsibility, authors strictly adhere to relevant national and international laws about DURC. Authors advocate for responsible deployment, ethical considerations, regulatory compliance, and transparent reporting to mitigate misuse risks and foster beneficial outcomes.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, X.J.; Zheng, J.B.; Su, T.; Ding, M.H.; Liu, H.W. An Effective Dynamic Constrained Two-Archive Evolutionary Algorithm for Cooperative Search-Track Mission Planning by UAV Swarms in Air Intelligent Transportation. IEEE Trans. Intell. Transp. Syst. 2024, 25, 944–958. [Google Scholar] [CrossRef]
  2. Cui, Y. An Improved Chaotic Self-Adapting Monkey Algorithm for Multi-UAV Task Assignment. IEEE J. Miniaturization Air Space Syst. 2024, 5, 9–15. [Google Scholar] [CrossRef]
  3. Han, D.; Jiang, H.; Wang, L.F.; Zhu, X.Y.; Chen, Y.Q.; Yu, Q.Z. Collaborative Task Allocation and Optimization Solution for Unmanned Aerial Vehicles in Search and Rescue. Drones 2024, 8, 138. [Google Scholar] [CrossRef]
  4. Lou, J.B.; Ding, R.; Wu, W.J. HHPSO: A Heuristic Hybrid Particle Swarm Optimization Path Planner for Quadcopters. Drones 2024, 8, 221. [Google Scholar] [CrossRef]
  5. Deng, M.; Yao, Z.Q.; Li, X.W.; Wang, H.; Nallanathan, A.; Zhang, Z.Y. Dynamic Multi-Objective AWPSO in DT-assisted UAV Cooperative Task Assignment. IEEE J. Sel. Areas Commun. 2023, 41, 3444–3460. [Google Scholar] [CrossRef]
  6. You, W.J.; Dong, C.; Wu, Q.H.; Qu, Y.B.; Wu, Y.L.; He, R. Joint Task Scheduling, Resource Allocation, and UAV Trajectory under Clustering for FANETs. China Commun. 2022, 19, 104–118. [Google Scholar] [CrossRef]
  7. Liu, X.L.; Jing, T.; Hou, L.Y. An FW–GA Hybrid Algorithm Combined with Clustering for UAV Forest Fire Reconnaissance Task Assignment. Mathematics 2023, 11, 2400. [Google Scholar] [CrossRef]
  8. Wang, X.F.; Yin, S.; Luo, L.Y.; Qiao, X. Research on Multi-UAV Task Assignment Based on a Multi-Objective, Improved Brainstorming Optimization Algorithm. Appl. Sci. 2024, 14, 2365. [Google Scholar] [CrossRef]
  9. Huang, G.; Hu, M.; Yang, X.Y. Multi-UAV Collaborative Target Allocation Based on Two-Way Knowledge Transfer in Complex Environments. In Proceedings of the 2024 36th Chinese Control and Decision Conference, Xi’an, China, 25–27 May 2024; pp. 2755–2762. [Google Scholar] [CrossRef]
  10. Dong, X.L.; Shi, C.G.; Wen, W.; Zhou, J.J. Multi-Mission Oriented Joint Optimization of Task Assignment and Flight Path Planning for Heterogeneous UAV Cluster. Remote Sens. 2023, 15, 5315. [Google Scholar] [CrossRef]
  11. Li, K.; Yan, X.X.; Han, Y. Multi-Mechanism Swarm Optimization for Multi-UAV Task Assignment and Path Planning in Transmission Line Inspection under Multi-Wind Field. Appl. Soft Comput. 2024, 150, 111033. [Google Scholar] [CrossRef]
  12. Choudhury, S.; Gupta, J.K.; Kochenderfer, M.J.; Sadigh, D.; Bohg, J. Dynamic Multi-Robot Task Allocation under Uncertainty and Temporal Constraints. Auton. Rob. 2022, 46, 231–247. [Google Scholar] [CrossRef]
  13. Martin, J.G.; Frejo, J.R.D.; García, R.A.; Camacho, E.F. Multi-Robot Task Allocation Problem with Multiple Nonlinear Criteria Using Branch and Bound and Genetic Algorithms. Intell. Serv. Robot. 2021, 14, 707–727. [Google Scholar] [CrossRef]
  14. Xie, J.F.; Carrillo, L.R.G.; Jin, L. An Integrated Traveling Salesman and Coverage Path Planning Problem for Unmanned Aircraft Systems. IEEE Control Syst. Lett. 2019, 3, 67–72. [Google Scholar] [CrossRef]
  15. Alighanbari, M.; How, J. Cooperative Task Assignment of Unmanned Aerial Vehicles in Adversarial Environments. In Proceedings of the 2005, American Control Conference, 2005, Portland, OR, USA, 8–10 June 2005; Volume 7. pp. 4661–4666. [Google Scholar] [CrossRef]
  16. Casbeer, D.W.; Holsapple, R.W. Column Generation for a UAV Assignment Problem with Precedence Constraints. Int. J. Robust Nonlinear Control 2011, 21, 1421–1433. [Google Scholar] [CrossRef]
  17. Subramanyam, A.; Repoussis, P.P.; Gounaris, C.E. Robust Optimization of a Broad Class of Heterogeneous Vehicle Routing Problems under Demand Uncertainty. INFORMS J. Comput. 2020, 32, 661–681. [Google Scholar] [CrossRef]
  18. Jia, Z.S.; Xiao, B.; Qian, H.Y. Improved Mixed Discrete Particle Swarms Based Multi-Task Assignment for UAVs. In Proceedings of the 2023 IEEE 12th Data Driven Control and Learning Systems Conference, Xiangtan, China, 12–14 May 2023; pp. 442–448. [Google Scholar] [CrossRef]
  19. Tang, G.K.; Chen, Y.Y.; Yu, R.; Dong, Y.; Yang, Y.Z. An Improved Genetic Algorithm for Cooperative Allocation with Orbit-like Tasks. In Proceedings of the 2023 42nd Chinese Control Conference, Tianjin, China, 24–26 July 2023; pp. 5664–5669. [Google Scholar] [CrossRef]
  20. Yan, X.H.; Chen, R.W.; Jiang, Z.H. UAV Cluster Mission Planning Strategy for Area Coverage Tasks. Sensors 2023, 23, 9122. [Google Scholar] [CrossRef]
  21. Wang, G.; Lv, X.; Cui, L.Z.; Yan, X.H. The Methods of Task Pre-Allocation and Reallocation for Multi-UAV Cooperative Reconnaissance Mission. IET Collab. Intell. Manuf. 2023, 5, e12090. [Google Scholar] [CrossRef]
  22. Si, J.S.; Hao, M.R. Online Weapon-Target Assignment Based on Distributed Auction Mechanism. J. Phys. Conf. Ser. 2023, 2456, 12044. [Google Scholar] [CrossRef]
  23. Tian, X.R.; Xu, T.; Luo, X.; Jia, Y.J.; Yin, J. Multi-UAV Reconnaissance Task Allocation in 3D Urban Environments. IEEE Access 2024, 12, 30989–30999. [Google Scholar] [CrossRef]
  24. He, W.B.; Li, W.G.; Hu, Y.J. UAV Mission Planning Based on Fish Swarm-ant Colony Optimization. In Proceedings of the 2023 5th International Conference on Intelligent Control, Measurement and Signal Processing (ICMSP), Chengdu, China, 19–21 May 2023; pp. 428–432. [Google Scholar] [CrossRef]
  25. Mirjalili, S. The Ant Lion Optimizer. Adv. Eng. Softw. 2015, 83, 80–98. [Google Scholar] [CrossRef]
  26. Assiri, A.S.; Hussien, A.G.; Amin, M. Ant Lion Optimization: Variants, Hybrids, and Applications. IEEE Access 2020, 8, 77746–77764. [Google Scholar] [CrossRef]
  27. Song, J.; Zhao, K.; Liu, Y. Survey on Mission Planning of Multiple Unmanned Aerial Vehicles. Aerospace 2023, 10, 208. [Google Scholar] [CrossRef]
  28. Mirjalili, S.; Jangir, P.; Saremi, S. Multi-Objective Ant Lion Optimizer: A Multi-Objective Optimization Algorithm for Solving Engineering Problems. Appl. Intell. 2017, 46, 79–95. [Google Scholar] [CrossRef]
  29. Wang, Y.F.; Zhang, L. Improved Multi-Objective Particle Swarm Optimization Algorithm Based on Area Division with Application in Multi-UAV Task Assignment. IEEE Access 2023, 11, 123519–123530. [Google Scholar] [CrossRef]
  30. Zhu, Y.C.; Liang, Y.G.; Jiao, Y.J.; Ren, H.P.; Li, K.B. Multi-Type Task Assignment Algorithm for Heterogeneous UAV Cluster Based on Improved NSGA-II. Drones 2024, 8, 384. [Google Scholar] [CrossRef]
  31. Yang, M.; Zhang, B.C.; Shi, Z.F.; Li, J.G. A Hybrid Task Allocation Approach for Multi-UAV Systems with Complex Constraints: A Market-Based Bidding Strategy and Improved NSGA-III Optimization. J. Supercomput. 2025, 81, 546. [Google Scholar] [CrossRef]
  32. Chen, J.M.; Zhang, K.; Zeng, H.; Yan, J.; Dai, J.; Dai, Z.D. Adaptive Constraint Relaxation-Based Evolutionary Algorithm for Constrained Multi-Objective Optimization. Mathematics 2024, 12, 3075. [Google Scholar] [CrossRef]
  33. Huang, G.; Hu, M.; Yang, X.Y.; Wang, X.; Wang, Y.J.; Huang, F.Y. A Review of Constrained Multi-Objective Evolutionary Algorithm-Based Unmanned Aerial Vehicle Mission Planning: Key Techniques and Challenges. Drones 2024, 8, 316. [Google Scholar] [CrossRef]
Figure 1. Schematic of targets and tasks.
Figure 1. Schematic of targets and tasks.
Drones 09 00733 g001
Figure 2. Preference-based elite selection mechanism.
Figure 2. Preference-based elite selection mechanism.
Drones 09 00733 g002
Figure 3. Process of MMSALO.
Figure 3. Process of MMSALO.
Drones 09 00733 g003
Figure 4. HV results on the test functions.
Figure 4. HV results on the test functions.
Drones 09 00733 g004
Figure 5. Individual decoding schematic.
Figure 5. Individual decoding schematic.
Drones 09 00733 g005
Figure 6. HV results on the task-allocation model.
Figure 6. HV results on the task-allocation model.
Drones 09 00733 g006
Table 1. Attributes of targets and tasks.
Table 1. Attributes of targets and tasks.
ModelAttributeParameter
Target, T j Target number N T
Task type number N t y p e
Target position X j T
Task, M k Task number N M
Task value V k M
Onboard resources requirement R k
Task timetable [ t k s , t k e ]
Task execution time t k d o
Task time window [ E T k , L T k ]
Probability of task failure S k
Table 2. Attributes of UAVs.
Table 2. Attributes of UAVs.
ModelAttributeParameter
UAV, U i UAV number N V
UAV position X i U
UAV speed v i
Task capability parameter P i k
UAV maximum range D i U
UAV value V i U
Amount of UAV onboard resources L i
Table 3. Basic parameter for algorithms.
Table 3. Basic parameter for algorithms.
AlgorithmParameter Setting
MOALO S e = 0.5 , S r = 0.5
IMOPSO w = 0.4 , c 1 = 1 , c 2 = 1
INSGA-II P c = 0.7 , P m = 0.3
INSGA-III P c = 0.7 , P m = 0.3
Table 4. Expressions for the three-objective DTLZ1 and DTLZ3.
Table 4. Expressions for the three-objective DTLZ1 and DTLZ3.
InstanceSearch SpaceObjectives, PS and PF
DTLZ1 [ 0 , 1 ] n f 1 ( x ) = 1 2 x 1 x 2 ( 1 + g ( x M ) )
f 2 ( x ) = 1 2 x 1 ( 1 x 2 ) ( 1 + g ( x M ) )
f 3 ( x ) = 1 2 ( 1 x 1 ) ( 1 + g ( x M ) )
g ( x M ) = 100 ( | x M | + x i x M ( ( x i 0.5 ) 2 cos ( 20 π ( x i 0.5 ) ) ) )
PS: 0 x 1 , x 2 1 , x i = 0.5 , for i = 3 , , n
PF: f 1 = 1 2 u v , f 2 = 1 2 u ( 1 v ) , f 3 = 1 2 ( 1 u ) , 0 u , v 1
DTLZ3 [ 0 , 1 ] n f 1 ( x ) = cos ( π 2 x 1 ) cos ( π 2 x 2 ) ( 1 + g ( x M ) )
f 2 ( x ) = cos ( π 2 x 1 ) sin ( π 2 x 2 ) ( 1 + g ( x M ) )
f 3 ( x ) = sin ( π 2 x 1 ) ( 1 + g ( x M ) )
g ( x M ) = 100 ( | x M | + x i x M ( ( x i 0.5 ) 2 cos ( 20 π ( x i 0.5 ) ) ) )
PS: 0 x 1 , x 2 1 , x i = 0.5 , for i = 3 , , n
PF: f 1 = cos ( u ) cos ( v ) , f 2 = cos ( u ) sin ( v ) , f 3 = sin ( u ) , 0 u , v π 2
Table 5. Hyper volume index of algorithms on the test functions.
Table 5. Hyper volume index of algorithms on the test functions.
Test FunctionnHVMMSALOMOALOIMOPSOINSGA-IIINSGA-III
DTLZ14Avg. 1.24 × 10 2 1.23 × 10 2 8.42 × 10 1 1.20 × 10 2 1 . 24 × 10 2
Std. 8.48 × 10 1 6.02 × 10 1 2.43 × 10 1 6.19 × 10 0 9.39 × 10 1
10Avg. 3 . 85 × 10 1 1.92 × 10 1 0.00 × 10 0 3.16 × 10 1 1.71 × 10 1
Std. 3.13 × 10 1 3.44 × 10 1 0.00 × 10 0 2.78 × 10 1 2.23 × 10 1
DTLZ34Avg. 1 . 14 × 10 2 1.10 × 10 2 4.62 × 10 1 9.59 × 10 1 9.05 × 10 1
Std. 4.81 × 10 0 4.39 × 10 0 3.06 × 10 1 3.32 × 10 1 3.64 × 10 1
10Avg. 1 . 82 × 10 0 0.00 × 10 0 0.00 × 10 0 1.93 × 10 1 6.24 × 10 1
Std. 7.63 × 10 0 0.00 × 10 0 0.00 × 10 0 8.43 × 10 1 2.23 × 10 0
MOKP200Avg. 9.99502 × 10 1 9.99499 × 10 1 9 . 99532 × 10 1 9.99485 × 10 1 9.99481 × 10 1
Std. 8.68 × 10 6 8.54 × 10 6 7.58 × 10 6 8.59 × 10 6 7.15 × 10 6
800Avg. 9.99869 × 10 1 9.99869 × 10 1 9 . 99874 × 10 1 9.99867 × 10 1 9.99866 × 10 1
Std. 1.08 × 10 6 1.46 × 10 6 1.32 × 10 6 9.27 × 10 7 1.19 × 10 6
Bold denotes the optimal result.
Table 6. Wilcoxon rank-sum test analysis of HV results on the test functions.
Table 6. Wilcoxon rank-sum test analysis of HV results on the test functions.
Test FunctionnMOALOIMOPSOINSGA-IIINSGA-III
DTLZ14 1.85 × 10 1 6 . 30 × 10 8 1.17 × 10 1 2.04 × 10 1
10 1 . 49 × 10 2 4 . 26 × 10 6 5.89 × 10 1 2 . 56 × 10 2
DTLZ34 3 . 26 × 10 2 6 . 30 × 10 8 3 . 49 × 10 2 2 . 00 × 10 2
10 5.89 × 10 1 5.89 × 10 1 7.87 × 10 1 1.00 × 10 0
MOKP200 4.33 × 10 1 1 . 33 × 10 7 3 . 74 × 10 6 6 . 01 × 10 7
800 6.75 × 10 1 6 . 80 × 10 8 2 . 87 × 10 6 6 . 45 × 10 7
Bold denotes statistical significance.
Table 7. Runtime performance comparison on the test functions.
Table 7. Runtime performance comparison on the test functions.
Test FunctionnMMSALOMOALOIMOPSOINSGA-IIINSGA-III
DTLZ14 6.23 × 10 1 2.68 × 10 1 1.00 × 10 1 1 . 20 × 10 0 4.86 × 10 0
10 9.08 × 10 1 7.06 × 10 1 9.62 × 10 0 1 . 01 × 10 0 3.84 × 10 0
DTLZ34 6.73 × 10 1 5.18 × 10 1 8.71 × 10 0 1 . 16 × 10 0 4.06 × 10 0
10 1.31 × 10 2 9.12 × 10 1 1.38 × 10 1 1 . 09 × 10 0 3.84 × 10 0
MOKP200 3.63 × 10 1 4.12 × 10 1 2.06 × 10 0 5 . 97 × 10 1 8.17 × 10 1
800 1.52 × 10 2 1.32 × 10 2 3.41 × 10 0 1 . 94 × 10 0 2.10 × 10 0
Bold denotes the shortest runtime.
Table 8. Results from ablation studies on the test functions.
Table 8. Results from ablation studies on the test functions.
Test FunctionnMMSALOSSRBAPUPBESMOALO
DTLZ14 1.2359 × 10 2 1.2331 × 10 2 1.2358 × 10 2 1.2329 × 10 2 1.2342 × 10 2 1.2326 × 10 2
Improvement 1.00271.00041.00251.00021.00131.0000
Runtime 6.2275 × 10 1 6.0715 × 10 1 2.6168 × 10 1 2.5005 × 10 1 2.5745 × 10 1 2.6802 × 10 1
DTLZ34 1.1412 × 10 2 1.1145 × 10 2 1.1084 × 10 2 1.1133 × 10 2 1.1072 × 10 2 1.1048 × 10 2
Improvement 1.03291.00881.00321.00771.00221.0000
Runtime 6.7328 × 10 1 6.9311 × 10 1 2.4247 × 10 1 2.2199 × 10 1 2.3500 × 10 1 5.1819 × 10 1
Table 9. Details of UAVs (Scenario 1).
Table 9. Details of UAVs (Scenario 1).
UAVPosition/kmSpeed/(km/s)Task Capability ParameterMax Range/kmValueOnboard Resources
1(34.45, 34.43)0.24[0.90, 0.87, 0.79]32000.8515
2(9.72, 50.85)0.28[0.77, 0.82, 0.75]30000.7813
3(50.30, 95.56)0.28[0.78, 0.83, 0.87]30000.8315
4(60.83, 46.54)0.28[0.80, 0.83, 0.75]30000.7913
5(38.90, 7.09)0.18[0.76, 0.82, 0.81]34000.8010
6(29.80, 1.13)0.16[0.87, 0.84, 0.83]34000.8413
Table 10. Details of targets and tasks (Scenario 1).
Table 10. Details of targets and tasks (Scenario 1).
TargetPosition/kmTask ValuesTask Execution Time/sProbability of Task Failure
1(273.22, 131.11)[0.76, 0.76, 0.84][16.79, 11.07, 20.94][0.38, 0.38, 0.42]
2(136.10, 177.79)[0.78, 0.76, 0.89][19.36, 21.43, 28.78][0.39, 0.38, 0.44]
3(252.56, 224.05)[0.83, 0.90, 0.76][27.06, 29.83, 13.00][0.41, 0.45, 0.38]
4(245.53, 164.42)[0.81, 0.79, 0.85][27.90, 12.99, 11.29][0.40, 0.39, 0.43]
5(205.65, 290.38)[0.78, 0.77, 0.84][25.76, 13.22, 27.95][0.39, 0.38, 0.42]
6(175.38, 175.20)[0.89, 0.90, 0.84][29.07, 27.94, 12.55][0.45, 0.45, 0.42]
7(154.83, 245.09)[0.81, 0.80, 0.85][15.41, 11.67, 20.64][0.41, 0.40, 0.43]
8(256.92, 235.26)[0.84, 0.80, 0.87][12.90, 11.02, 23.62][0.42, 0.40, 0.44]
9(295.72, 171.83)[0.84, 0.89, 0.75][11.85, 14.65, 17.31][0.42, 0.45, 0.38]
10(218.45, 238.61)[0.82, 0.76, 0.87][25.45, 12.67, 18.29][0.41, 0.38, 0.43]
11(233.40, 230.03)[0.85, 0.90, 0.84][19.09, 22.49, 18.54][0.43, 0.45, 0.42]
12(185.68, 281.01)[0.83, 0.79, 0.85][16.06, 19.26, 12.87][0.41, 0.39, 0.43]
13(127.31, 128.02)[0.83, 0.89, 0.84][24.77, 20.16, 16.84][0.42, 0.44, 0.42]
14(192.93, 117.56)[0.87, 0.86, 0.85][24.34, 16.50, 18.01][0.44, 0.43, 0.43]
15(233.82, 281.93)[0.88, 0.81, 0.76][24.68, 27.43, 28.66][0.44, 0.40, 0.38]
16(154.15, 263.87)[0.80, 0.77, 0.78][11.83, 10.10, 23.88][0.40, 0.39, 0.39]
17(263.45, 158.86)[0.85, 0.87, 0.77][12.50, 15.12, 13.35][0.43, 0.43, 0.38]
18(264.43, 146.68)[0.79, 0.77, 0.89][12.31, 19.44, 23.24][0.39, 0.38, 0.45]
Table 11. Details of UAVs (Scenario 2).
Table 11. Details of UAVs (Scenario 2).
UAVPosition/kmSpeed/(km/s)Task Capability ParameterMax Range/kmValueOnboard Resources
1(34.45, 34.43)0.24[0.81, 0.79, 0.79]32000.8024
2(87.00, 36.35)0.28[0.79, 0.82, 0.83]30000.8114
3(6.08, 19.43)0.26[0.90, 0.81, 0.77]30000.8313
4(43.15, 51.61)0.29[0.88, 0.85, 0.80]30000.8417
5(52.12, 1.67)0.21[0.78, 0.84, 0.76]32000.7913
6(46.45, 41.52)0.19[0.89, 0.88, 0.87]34000.8812
7(56.76, 50.16)0.28[0.87, 0.82, 0.75]30000.8114
8(26.68, 82.16)0.27[0.83, 0.89, 0.81]30000.8419
Table 12. Details of targets and tasks (Scenario 2).
Table 12. Details of targets and tasks (Scenario 2).
TargetPosition/kmTask ValuesTask Execution Time/sProbability of Task Failure
1(175.20, 154.83)[0.87, 0.77, 0.88][23.53, 13.54, 12.03][0.43, 0.39, 0.44]
2(245.09, 256.92)[0.89, 0.88, 0.77][22.01, 29.15, 29.46][0.45, 0.44, 0.38]
3(235.26, 295.72)[0.79, 0.76, 0.83][21.78, 18.60, 17.09][0.40, 0.38, 0.41]
4(171.83, 218.45)[0.77, 0.76, 0.85][23.66, 21.47, 17.15][0.39, 0.38, 0.43]
5(238.61, 233.40)[0.76, 0.78, 0.80][26.23, 21.64, 29.05][0.38, 0.39, 0.40]
6(230.03, 185.68)[0.87, 0.77, 0.81][10.30, 19.37, 10.98][0.43, 0.39, 0.41]
7(281.01, 127.31)[0.82, 0.84, 0.81][25.36, 23.86, 29.54][0.41, 0.42, 0.41]
8(128.02, 192.93)[0.80, 0.82, 0.77][22.64, 20.55, 14.92][0.40, 0.41, 0.39]
9(117.56, 233.82)[0.86, 0.83, 0.80][23.37, 21.02, 28.38][0.43, 0.41, 0.40]
10(281.93, 154.15)[0.86, 0.80, 0.81][22.46, 26.24, 24.51][0.43, 0.40, 0.41]
11(263.87, 263.45)[0.86, 0.88, 0.89][23.64, 27.72, 17.99][0.43, 0.44, 0.44]
12(158.86, 264.43)[0.76, 0.75, 0.85][11.43, 16.89, 13.13][0.38, 0.38, 0.43]
13(146.68, 167.94)[0.77, 0.79, 0.78][14.34, 23.84, 25.53][0.38, 0.39, 0.39]
14(108.40, 110.73)[0.77, 0.82, 0.85][12.29, 15.27, 12.53][0.38, 0.41, 0.42]
15(116.92, 209.40)[0.79, 0.90, 0.88][28.76, 10.79, 22.95][0.40, 0.45, 0.44]
16(225.76, 193.61)[0.79, 0.85, 0.83][18.21, 22.86, 11.13][0.39, 0.42, 0.42]
17(139.28, 214.32)[0.88, 0.76, 0.81][24.36, 12.23, 19.57][0.44, 0.38, 0.40]
18(108.47, 287.79)[0.90, 0.80, 0.79][23.63, 15.54, 26.86][0.45, 0.40, 0.39]
19(282.10, 270.60)[0.77, 0.85, 0.80][20.78, 11.26, 25.89][0.39, 0.43, 0.40]
20(203.66, 298.34)[0.89, 0.87, 0.84][25.67, 12.93, 22.59][0.44, 0.44, 0.42]
21(299.17, 130.00)[0.79, 0.77, 0.82][10.16, 16.60, 11.15][0.39, 0.39, 0.41]
22(107.05, 279.01)[0.88, 0.79, 0.79][15.89, 25.08, 18.77][0.44, 0.40, 0.39]
23(175.96, 129.86)[0.75, 0.77, 0.84][21.82, 19.09, 27.25][0.38, 0.39, 0.42]
24(151.47, 112.94)[0.90, 0.77, 0.83][16.10, 19.74, 27.79][0.45, 0.38, 0.41]
Table 13. HV values of algorithms on the task-allocation model.
Table 13. HV values of algorithms on the task-allocation model.
ScenarioHV ValueMMSALOMOALOIMOPSOINSGA-IIINSGA-III
1Avg. 6 . 22 × 10 5 6.14 × 10 5 6.20 × 10 5 6.19 × 10 5 6.19 × 10 5
Std. 3.71 × 10 3 6.68 × 10 3 1.83 × 10 3 5.86 × 10 3 4.38 × 10 3
2Avg. 4 . 18 × 10 6 4.11 × 10 6 4.16 × 10 6 4.16 × 10 6 4.16 × 10 6
Std. 1.96 × 10 4 3.61 × 10 4 1.49 × 10 4 2.54 × 10 4 1.59 × 10 4
Bold denotes the optimal result.
Table 14. Wilcoxon rank-sum test analysis of HV results on the task-allocation model.
Table 14. Wilcoxon rank-sum test analysis of HV results on the task-allocation model.
ScenarioMOALOIMOPSOINSGA-IIINSGA-III
1 2 . 34 × 10 4 3 . 98 × 10 2 3 . 26 × 10 2 2 . 65 × 10 2
2 3 . 67 × 10 7 2 . 65 × 10 2 1 . 28 × 10 2 2 . 48 × 10 2
Bold denotes statistical significance.
Table 15. Runtime performance comparison on the task-allocation model.
Table 15. Runtime performance comparison on the task-allocation model.
ScenarioMMSALOMOALOIMOPSOINSGA-IIINSGA-III
1 6.58 × 10 1 6.62 × 10 1 1.44 × 10 2 1.81 × 10 1 1 . 67 × 10 1
2 1.05 × 10 2 8.66 × 10 1 2.02 × 10 2 2.89 × 10 1 2 . 66 × 10 1
Bold denotes the shortest runtime.
Table 16. Results from ablation studies on the task-allocation model.
Table 16. Results from ablation studies on the task-allocation model.
ScenarioMMSALOSSRBAPUPBESMOALO
1 6.2235 × 10 5 6.1626 × 10 5 6.1409 × 10 5 6.1517 × 10 5 6.1415 × 10 5 6.1408 × 10 5
Improvement1.01351.00361.00001.00181.00011.0000
Runtime 6.5766 × 10 1 6.1644 × 10 1 7.9269 × 10 1 5.9756 × 10 1 6.8892 × 10 1 6.6172 × 10 1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, C.; Li, G.; Liu, Y.; Zheng, Q.; Yang, G.; Liu, K.; Diao, X. Cooperative Task Allocation for Unmanned Aerial Vehicle Swarm Using Multi-Objective Multi-Population Self-Adaptive Ant Lion Optimizer. Drones 2025, 9, 733. https://doi.org/10.3390/drones9110733

AMA Style

Li C, Li G, Liu Y, Zheng Q, Yang G, Liu K, Diao X. Cooperative Task Allocation for Unmanned Aerial Vehicle Swarm Using Multi-Objective Multi-Population Self-Adaptive Ant Lion Optimizer. Drones. 2025; 9(11):733. https://doi.org/10.3390/drones9110733

Chicago/Turabian Style

Li, Chengze, Gengsong Li, Yi Liu, Qibin Zheng, Guoli Yang, Kun Liu, and Xingchun Diao. 2025. "Cooperative Task Allocation for Unmanned Aerial Vehicle Swarm Using Multi-Objective Multi-Population Self-Adaptive Ant Lion Optimizer" Drones 9, no. 11: 733. https://doi.org/10.3390/drones9110733

APA Style

Li, C., Li, G., Liu, Y., Zheng, Q., Yang, G., Liu, K., & Diao, X. (2025). Cooperative Task Allocation for Unmanned Aerial Vehicle Swarm Using Multi-Objective Multi-Population Self-Adaptive Ant Lion Optimizer. Drones, 9(11), 733. https://doi.org/10.3390/drones9110733

Article Metrics

Back to TopTop