Next Article in Journal
Multi-Stimuli Responsive Viologen-Imprinted Polyvinyl Alcohol and Tricarboxy Cellulose Nanocomposite Hydrogels
Previous Article in Journal
Multi-Sensor Soil Probe and Machine Learning Modeling for Predicting Soil Properties
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-UAV Escape Target Search: A Multi-Agent Reinforcement Learning Method

1
College of Electronic Engineering, National University of Defense Technology, Hefei 230037, China
2
Anhui Province Key Laboratory of Electronic Restriction, Hefei 230037, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(21), 6859; https://doi.org/10.3390/s24216859
Submission received: 13 September 2024 / Revised: 18 October 2024 / Accepted: 23 October 2024 / Published: 25 October 2024
(This article belongs to the Section Remote Sensors)

Abstract

:
The multi-UAV target search problem is crucial in the field of autonomous Unmanned Aerial Vehicle (UAV) decision-making. The algorithm design of Multi-Agent Reinforcement Learning (MARL) methods has become integral to research on multi-UAV target search owing to its adaptability to the rapid online decision-making required by UAVs in complex, uncertain environments. In non-cooperative target search scenarios, targets may have the ability to escape. Target probability maps are used in many studies to characterize the likelihood of a target’s existence, guiding the UAV to efficiently explore the task area and locate the target more quickly. However, the escape behavior of the target causes the target probability map to deviate from the actual target’s position, thereby reducing its effectiveness in measuring the target’s probability of existence and diminishing the efficiency of the UAV search. This paper investigates the multi-UAV target search problem in scenarios involving static obstacles and dynamic escape targets, modeling the problem within the framework of decentralized partially observable Markov decision process. Based on this model, a spatio-temporal efficient exploration network and a global convolutional local ascent mechanism are proposed. Subsequently, we introduce a multi-UAV Escape Target Search algorithm based on MAPPO (ETS–MAPPO) for addressing the escape target search difficulty problem. Simulation results demonstrate that the ETS–MAPPO algorithm outperforms five classic MARL algorithms in terms of the number of target searches, area coverage rate, and other metrics.

1. Introduction

In recent years, unmanned aerial vehicles (UAVs) have found applications in various military and civilian domains due to their advantages, such as high mobility, accessibility, convenient deployment, and low cost. They have gradually become indispensable in modern society, with roles in civil sectors, such as agriculture [1,2], mineral exploration [3], and forest rescue [4], as well as in military reconnaissance [5] and strikes [6]. Multi-UAV target search problems is a significant issue in autonomous UAV decision-making and has garnered extensive academic attention recently. Multi-UAV target search involves UAVs using on-board detection equipment to reconnoiter designated areas. They share information via a communication network, thereby jointly capturing targets. Currently, three primary methods are used for multi-UAV target search. The first category is pre-planning methods, such as partition search [7] and formation search [8]. These methods, essentially, transform the target search problem into a planning problem with area coverage, offering high reliability and easy evaluation of the solution results. However, they require a known mission area model in advance, involve longer planning times, and are not highly adaptive to dynamic environmental changes. The second category is online optimization methods, which approximate the search problem as a real-time objective function optimization problem. These methods typically employ traditional or heuristic algorithms, such as ant colony algorithms [9] and genetic algorithms [10]. They are better adapted to environmental dynamics than to pre-planning approaches. However, they depend on a central node for decision-making and exhibit low adaptability in distributed environments. The third category entails Multi-Agent Reinforcement Learning (MARL) methods, which model the problem as a Partially Observable Markov Decision Process (POMDP) and utilize algorithms based on the MARL framework. These methods enable agents to learn and optimize their behavior through interactions with the environment and other agents, allowing them to adapt to dynamic changes and make rapid decisions [11,12]. The primary challenge of these methods lies in designing the algorithm training architecture, agent exploration mechanism, and reward function tailored to specific task requirements.
Recently, the design of MARL methods has become a prominent area of research in the field of artificial intelligence. It has found applications in areas such as multi-UAV target search [13], autonomous vehicle path planning [14], and other related applications. Within the MARL framework, Shen et al. [15] proposed the DNQMIX algorithm, which enhances search rate and coverage. Lu et al. [16] proposed the MUICTSTP algorithm, demonstrating superior performance in terms of anti-interference and collision avoidance. Yu et al. [17] proposed the Multi-Agent Proximal Policy Optimization (MAPPO) algorithm, which has exhibited excellent performance in multi-agent testing environments and is regarded as one of the most advanced algorithms available. Wei et al. [18] combined the MAPPO algorithm with optimal control (OC) and GPU parallelization to propose the OC–MAPPO algorithm, which accelerates UAV learning.
To better describe environmental uncertainty, Bertuccelli et al. [19] proposed a probabilistic approach. This method divides the task area into units, each associated with the probability of target presence, establishing a target probability graph. The method has achieved good results in target search and is widely recognized. Building on the MARL framework and the target probability graph, Zhang et al. [20] designed a confidence probability graph using evidence theory and proposed a double critic DDPG algorithm. This approach effectively balances the bias in action–value function estimation and the variance in strategy updates. Hou et al. [21] converted the probability function into a grid-based goal probability graph and proposed a MADDPG-based search algorithm, improving search speed and avoiding collisions and duplications.
Multi-UAV target search has made some progress over time, but two challenges remain. Firstly, the utilization of sample data remains inefficient, and balancing utilization and exploration presents a challenge. Existing MARL algorithms primarily employ neural networks, such as fully connected networks and convolutional networks. These networks often fail to simultaneously achieve efficient utilization of temporal and spatial information in the sample data, and also lack effective environmental exploration. Secondly, the behavioral modeling of dynamic target is relatively simple. Existing work primarily considers changes in the target’s position over time, often transforming the target search problem into a target tracking problem. In actual non-cooperative target search scenarios, targets may exhibit escape behavior. They actively change their position to evade detection and, potentially, use environmental blind spots to hide, preventing real-time tracking by UAVs. Addressing the challenges identified above, this paper investigates the Multi-UAV Escape Target Search (METS) problem in complex environments. The contributions of this paper are summarized as follows:
  • The simulation environment for the METS problem is constructed, introducing a Target Existence Probability Map (TEPM), with an appropriate probability update method employed for the escaping target. Based on the TEPM, a local state field of view is designed, with local state values obtained through entropy calculation. Additionally, a new state space, action space, and reward function are devised within the framework of Decentralized Partially Observable Markov Decision process (DEC-POMDP). Ultimately, a model that addresses the METS problem is established.
  • To enhance the MARL algorithm’s ability to process spatio-temporal sequence information and improve environmental exploration, this paper proposes the Spatio-Temporal Efficient Exploration (STEE) network, constructed using a convolutional long short-term memory network and a noise network. This network is integrated into the MAPPO algorithm, and its impact on the overall performance of the MAPPO algorithm is validated.
  • To search the escaping target in the METS problem, the Global Convolutional Local Ascent (GCLA) mechanism is proposed. A Multi-UAV Escape Target Search algorithm based on MAPPO (ETS–MAPPO) is introduced by combining MAPPO with the STEE network. This algorithm effectively addresses the challenges of searching for escape targets. Experimental comparisons with five classic MARL algorithms show significant improvements in the number of target searches, area coverage rate, and other metrics.
The remaining chapters of this paper are organized as follows: Section 2 defines the system model and provides a mathematical formulation of the METS problem. Section 3 introduces the ETS–MAPPO algorithm within the MARL framework and describes it in detail. In Section 4, experiment results are presented to validate the effectiveness of ETS–MAPPO. Section 5 concludes the paper and discusses future research.

2. System Model and Problem Formulation

2.1. System Model

A typical multi-UAV escape target search mission scenario is shown in Figure 1. Assuming that several escape targets and fixed obstacles are distributed within an unknown mission area Ω , N u UAVs depart from the base and cooperatively search the area. The mission requires UAVs to capture as many escape targets as possible, maximize coverage of mission area Ω , and avoid collisions with obstacles to the greatest extent possible.

2.1.1. Mission Area Model

Considering the need for relative stability in the operation of the UAV detection payload, the UAV was set to maintain a constant flight altitude during target search operations in the mission area Ω [22]. In this scenario, the mission area Ω was conceptualized as a two-dimensional plane with two layers, where the UAV operated in the upper layer. Escape targets were positioned on the ground in the lower layer of the mission area. Obstacles were distributed across both the upper and the lower layers.
In Figure 2, each layer is a rectangular region L x × L y , which is rasterized into a cellular grid N x × N y , with each cell having an area of S = Δ x × Δ x . The positional coordinates of any cell C x , y in the Cartesian coordinate system are denoted by Equation (1):
C ( x , y ) = { ( x , y ) | x [ 1 ,   N x ] , y [ 1 ,   N y ] }

2.1.2. Unmanned Aerial Vehicle Model

To facilitate the study, the motions of N u identical quadrotor UAVs were modeled as depicted in Figure 3. The UAVs were capable of discrete movements with a total of 9 degrees of freedom: left up, up, right up, left down, down, right down, left, right, and stationary.
The UAVs were equipped with a detection payload for target identification, with a detection range of L t a r g e t . In the rasterized mission area model, this detection range was mapped to n u = L t a r g e t / S cells. Assuming no omissions or false detections by the detection payload, it was considered that the UAVs could correctly identify a target within a cell at time t with a probability p . The UAVs’ detection range for stationary obstacles was L o b s t a c l e , which was mapped to n o = L o b s t a c l e / S cells.

2.1.3. Escape Target Model

In the METS mission, the mission area contained N t escape targets, each of which could be in one of two states: stationary or escaping. These targets could detect UAVs within a detectable range of L u a v ,which was mapped to cells n t = L u a v / S . Initially, each escape target remained stationary, meaning that it maintained its current position while detecting UAVs. When a UAV entered the detection range of an escape target, the escape target transitioned from the stationary state to the escaping state. In the escaping state, the escape target engaged in active concealment, released false information to the UAV, and chose a movement distance L c h a n g in a random direction with probability P c , mapped to n c = L c h a n g e / S cells, or remained stationary with probability 1 P c , within the escape time step T c h a n g e . The escape target returned to the stationary state at the end of the escaping state.

2.1.4. Target Existence Probability Map Model

In the METS problem, at the start of each search cycle, a probability distribution function was employed to model the target’s location information. In the rasterized mission area, the probability function was transformed into a cell-based Target Existence Probability Map (TEPM), with each cell having an associated target existence probability b x , y [ 0 , 1 ] . The initial target existence probability for all cells in the TEPM was set to b x , y ( 0 ) = 0.5 , indicating that the UAVs lacked any a priori information about target presence in the mission area.
As the UAVs continuously searched the mission area, their detection payload scanned for target information along the path, and the TEPM was probabilistically updated based on this detection information. We used a Bayesian update model to update the TEPM. When the UAV scanned cell C x , y at time t , the probability update for the target appearing in cell C x , y was given by Equation (2) [23]. If multiple UAVs scanned cell b x , y simultaneously, the probability value was updated correspondingly for each scan.
b x , y ( t ) = p s b x , y ( t 1 ) p s b x , y ( t 1 ) + ( 1 p s ) ( 1 b x , y ( t 1 ) )
where p s is the correct probability p of the UAV’s detection payload in detecting a target in cell C x , y . If no targets existed, p s was substituted with 1 p .

2.2. Formulation of the Problem

In the METS mission, the UAVs began at a base position and made operational decisions based on state information, aiming to explore the unknown mission area and search for the escape targets as extensively as possible. Consequently, the objective function was defined to maximize the exploration degree f 1 of the entire region while also maximizing the number of target searches f 2 , as represented in Equation (3):
{ f 1 = x = 1 N x y = 1 N y | b x , y ( t ) p max | f 2 = x = 1 N x y = 1 N y 1 , b x , y ( t ) > ε
where ε is the confidence level of the target’s existence in a cell and p max is the maximum uncertainty value.
Considering that the UAV motion was constrained by the mission area’s boundaries, the boundary condition for the UAV at position u i , t ( x , y ) was expressed in Equation (4):
{ 1 u i , t ( x ) N x , u U 1 u i , t ( y ) N y , u U
Additionally, it is essential to consider that the UAV should avoid collisions with obstacles, with the collision constraint expressed in Equation (5):
B k ( x , y ) u i , t ( x , y ) > d s a f e , k { 1 , 2 , , N B }
where B k ( x , y ) is the position of obstacles k , d s a f e is the safety distance, and N B is the number of obstacles.
The collision constraint was subsequently transformed into the optimization objective of the objective function f 3 , as expressed in Equation (6):
f 3 = i = 1 N u k = 1 N B 1 , B k ( x , y ) u i , t ( x , y ) d s a f e
Finally, the METS problem was formulated as an optimization problem with an objective function, as represented in Equation (7):
max     f 1 + f 2 f 3 s . t .       1 u i , x ( t ) N x , u U                   1 u i , y ( t ) N y , u U

3. Multi-UAV Escape Target Search: MARL Approach

In this section, the optimization problem presented in Equation (7) is first reformulated within the framework of DEC-POMDP based on MARL methods. The state space of UAVs and the reward function are then defined according to the TEPM. Finally, the ETS–MAPPO algorithm is proposed, and its framework is detailed.

3.1. Decentralized Partially Observable Markov Decision Process

In the METS, an extended form of the POMDP was presented as the DEC-POMDP to model the optimization problem [24], considering multiple UAVs and decentralized decision-making. The DEC-POMDP is defined by the tuple ( N , O , A , S , F s , F o , R , γ ) .
  • N is the number of UAVs. γ ( 0 , 1 ] is the reward discount factor.
  • O = { O 1 , , O N } is the joint observation space, which consists of the observation state space O i of all UAVs.
Based on the UAV modeling in Section 2, it was assumed that each UAV had a Local State Field of View (LSFV) that corresponded to the obstacle detection range, i.e., n o cells. By extracting the target probability information of n o cells near the UAV’s location on the TEPM, the local state value of the LSFV was calculated at each time step t. The extracted cell uncertainty u x , y ( t ) was calculated as the Shannon entropy, serving as the local state value of the LSFV, as defined in Equation (8). Additionally, the local state value of the cell where an obstacle was detected was set to −1.
u x , y ( t ) = H [ b x , y ( t ) ] = b x , y ( t ) log 2 b x , y ( t ) ( 1 b x , y ( t ) ) log 2 ( 1 b x , y ( t ) )
In the METS mission, it was considered that the communication between individual UAVs would be unrestricted. Each UAV could obtain the current position information of other UAVs and the TEPM, aiding decision-making. Therefore, the observation state space o i , t of UAV i consisted of four parts, as defined in Equation (9):
o i , t = { P t , D t , l o c t , u , l o c t , u ¯ }
where P t is the TEPM at time t , D t is the LSFV of the UAV, l o c t , u is the position of the UAV, and l o c t , u ¯ is the position of the other UAVs.
3.
A = { A 1 , , A N } is the joint action space, which consists of each UAV i choosing action a i , t A i based on its own observed state space o i , t .
4.
S is the state space of the environment, s t S is the state of time slot t .
5.
F o ( s t , a t 1 ) = P ( o t = o | s t , a t 1 ) is the observation probability function.
6.
F s ( s t , a t ) = P ( s t + 1 = s | s t , a t ) is the state transition probability function.
7.
R is the reward function, consisting of three components: an escape target search reward J T ( t ) , an environment search reward J E ( t ) , and a collision prevention reward J C ( t ) . Weighting coefficients ω 1 , ω 2 , and ω 3 were applied to balance these three components, as defined in Equation (10):
R ( t ) = ω 1 J T ( t ) + ω 2 J E ( t ) + ω 3 J C ( t )
The escape target search reward J T ( t ) was rewarded based on both the initial discovery of targets and the subsequent rediscovery of these targets. The search for an escape target in cell C x , y at time step t was evaluated as described in Equation (11). If the probability of a target’s existence in cell C x , y exceeded a threshold ε , the target was considered to be present in that cell.
β x , y ( t ) = { 1   , b x , y ( t ) > ε 0   ,   e l s e
The evaluation process for determining the existence of an escape target in a cell, initially and in subsequent evaluations, is represented by β x , y 1 ( t ) and β x , y 2 ( t ) . Therefore, the escape target search reward was calculated as shown in Equation (12):
J T ( t ) = x , y T 1 β x , y 1 ( t ) r t 1 + x , y T 2 β x , y 2 ( t ) r t 2
where T 1 and T 2 are the sets of locations where the escape target was initially discovered and subsequently rediscovered, respectively, and r t 1 and r t 2 are the reward factors for the initial discovery and subsequent rediscovery of the escape target, respectively.
The environment search reward J E ( t ) quantifies the extent of environmental exploration and is determined by the change in cell uncertainty u x , y ( t ) over time, as represented in Equation (13):
J E ( t ) = x = 1 N x y = 1 N y u x , y ( t ) u x , y ( t + 1 )
The collision prevention reward J C ( t ) was defined as the condition in which the distance between the UAV and an obstacle falls below a safe threshold, as represented in Equation (14):
J C ( t ) = r c * i = 1 N u k = 1 N B 1 , B k u i , t ( x , y ) d s a f e
where r c is the collision penalty factor.

3.2. Multi-UAV Escape Target Search Algorithm Based on MAPPO

In a multi-agent environment, the MAPPO algorithm [17] demonstrates superior adaptability and stability in strategy optimization compared to other MARL algorithms. To address the METS problem, we propose the multi-UAV Escape Target Search algorithm based on MAPPO (ETS–MAPPO), which includes two key components: first, the Spatio-Temporal Efficient Exploration (STEE) network to enhance the MAPPO algorithm’s capability to process spatio-temporal sequence information and explore the environment. Second, the Global Convolution and Local Ascent (GCLA) mechanism to overcome challenges posed by the variability of the escape target and the weak directionality of state inputs.

3.2.1. Spatio-Temporal Efficient Exploration Network

The MARL algorithm focuses on both data utilization and environmental exploration when solving search problems. Utilization refers to the agent selecting the action that maximizes the reward from previous actions, while exploration involves the agent choosing a new action in anticipation of a higher reward. Balancing utilization and exploration poses a significant challenge in reinforcement learning [25], particularly in complex environments with multiple state inputs and sparse rewards.
In the MARL, empirical data contain both temporal and spatial information. Temporal information includes state dynamics, action continuity, and reward latency. Spatial information encompasses environmental complexity, relative UAV positions, and target distribution. Current approaches for data utilization commonly employ recurrent neural networks, such as LSTM [26] and GRU [27], for temporal data processing. Spatial data are primarily processed using convolutional neural networks, such as GCN [28] and HGCN [29]. These networks face challenges in relation to the efficient processing of both temporal and spatial information. As a solution, the Convolutional Long Short-Term Memory (ConvLSTM) [30] network, which combines convolutional operations with memory networks, has been introduced. The ConvLSTM network can capture both spatial and temporal features, thereby enhancing the performance of MARL algorithms. However, the ConvLSTM network lacks strong environmental exploration capabilities, necessitating improvements in this area. Currently, MARL algorithms employ epsilon-greedy and entropy regularization strategies for exploration, which introduce noise to the agents’ actions, but often result in low exploration efficiency. This paper proposes the use of the noise network, which adds parameterized noise to the neural network, thereby improving the exploration efficiency of MARL algorithms and preventing them from converging into local optima.
This paper proposes the STEE network architecture, which integrates the ConvLSTM network’s efficient processing of spatio-temporal sequential information with the noise network’s effective capability to conduct environmental exploration. The STEE network architecture is illustrated in Figure 4. Data were first normalized with features and then fed into the noise network to process state data features. The output from the noise network was then used as the input for the ConvLSTM network unit, and the final output of the STEE network was obtained through a multi-layer ConvLSTM network unit.
The ConvLSTM network unit was used to capture spatio-temporal data, enabling the modeling of dynamic environmental changes and long-term dependencies on historical information. The ConvLSTM network unit structure primarily comprises a forget gate, an input gate, a memory cell, and an output gate. The noise network is a fully connected neural network that incorporates parameterized noise, as depicted in Figure 5. It allows the STEE network to continuously learn the mean μ and variance σ of w , b during the training process.

3.2.2. Global Convolution Local Ascent Mechanism

In the METS problem, the escape target initially remained stationary and only entered the escaping state after detecting the UAV. During the escaping state, the target took random actions, leading to positional changes and increased environmental instability. Under the existing mechanism, each UAV observes a larger number of state space parameters than the TEPM’s N x × N y . However, these state parameters are sparse and lack clear directionality, offering no advantage when searching for the escape target. The UAV’s limited ability to capture changes in the escape target’s location hinders its ability to relocate the target, resulting in poor network convergence.
Currently, there is insufficient research on the poor convergence of networks due to target escape. This paper proposes the GCLA mechanism, which employs a global TEPM convolution operation and a target local area uncertain probability ascent operation. This mechanism enhances both the capability to capture escaping targets and the convergence of the algorithmic network.
The global TEPM convolution operation optimizes the state parameters of the TEPM’s N x × N y into a set of values 1 × 9 through convolution. This operation aims to divide the entire task area into nine orientation zones to capture global direction guidance information. The TEPM convolution operation for the state parameters 10 × 10 is illustrated in Figure 6.
The target local area uncertain probability ascent operation involves increasing the uncertain probability of the target local area, after the initial search of the escape target. To enhance the UAV’s ability to re-search the area from which the target may have escaped, the operation increases the uncertain probability p ( t ) at a specific rate within the escaped area. This area is defined as the area spanning n c cells around cell C x , y , as calculated in Equation (15). Subsequently, the TEPM is updated based on this revised uncertain probability within the escaped area.
{ C ( x , y ) = C ( x + i , y + i ) , i = [ n c , n c ] p x , y ( t ) = p max + p x , y ( t 1 ) e τ t
where p x , y ( t ) is the uncertain probability of target existence at C x , y at time t and τ is the ascent rate.
After processing with the GCLA mechanism, the UAV’s observation state space o i , t is reformulated as shown in Equation (16):
o i , t = { P t , D t , C t , l o c t , u , l o c t , u ¯ }
where C t is the result following the convolution operation on the TEPM and P t is the updated TEPM.

3.3. ETS–MAPPO Algorithmic Framework

The ETS–MAPPO algorithm builds upon the foundational architecture of MAPPO, incorporating the STEE network and the GCLA mechanism. As shown in Figure 7, the algorithm adopts the actor–critic network architecture and applies the centralized training and decentralized execution framework to solve the DEC-POMDP. Through interaction between the UAV and the simulation environment, empirical data are generated and stored in experience pools, from which the UAV extracts data to calculate the actor and critic network losses. The backpropagation algorithm is then employed to update these networks, enabling centralized training. Each UAV obtains observation state data from the environment and processes them through the GCLA mechanism. Based on the current actor network, these UAVs then select actions to interact with the simulation environment, thereby enabling decentralized execution.
The application paradigm of the ETS–MAPPO algorithm involves offline training and online execution. The pseudocode for the ETS–MAPPO algorithm is provided in Algorithm 1, below. The training phase focused on the optimization of both the actor and the critic networks. The actor network is designed to learn the mapping function π φ from the observed state space O t to the action a t , and is constructed using a fully connected neural network. To avoid overmodification of the objective values, the actor network is implemented by optimizing a cropped alternative objective function L t ( θ ) , as shown in Equation (17):
L t ( θ ) = E { min [ r t ( θ ) A t , c l i p ( r t ( θ ) , 1 ε , 1 + ε ) A t ] + σ S [ π θ ( O t ) ] }
where θ is the parameter of the actor network, r t ( θ ) is the ratio of the new strategy to the old one, A t is the General Advantage Estimation (GAE), c l i p ( ) is the cropping function, ε [ 0 , 1 ] is the truncation factor, S [ π θ ( O t ) ] is the entropy of the strategy π θ ( O t ) , and σ is the hyper-parameter controlling the entropy coefficient.
The calculation of GAE is based on the Temporal Difference (TD) error δ t , as defined in Equation (18):
δ t = r t + γ V ϕ ( s t + 1 ) V ϕ ( s t )
where r t is the reward, γ is the discount factor, and V ϕ ( s t ) is the value function of the critic network at time step t .
The GAE estimates the dominance function A t G A E ( λ ) by considering the TD errors for k time steps forward from the current time step t and by performing a weighted summation of them. The specific weighted summation expression is provided in Equation (19):
A t G A E ( λ ) = ( 1 λ ) ( δ t + λ δ t + 1 + + λ k 1 δ t + k 1 )
where λ is the hyperparameter of TD error weights.
The purpose of the critic network, which is constructed using the STEE network, is to learn the mapping function V ϕ from the state space S t to the real values. The critic network optimizes the network parameters using the mean square error loss function L t ( ϕ ) , with its expression provided in Equation (20):
L t ( ϕ ) = E { max [ ( V ϕ ( s t ) R ^ t ) 2 , ( c l i p ( V ϕ ( s t ) , V ϕ o l d ( s t ) ε , V ϕ o l d ( s t ) + ε , ) R ^ t ) 2 ] }
where ϕ is the critic network parameters and R ^ t is the discount reward.
Algorithm 1: ETS–MAPPO
Initialization: The actor network parameters and the critic network parameters for each UAV and other relevant parameters.
1. For episode = 1,2, …, M do.
2. Reset environment state: x ← env.reset()
3.      For time step = 1,2, …, T do
4.           For UAV i = 1,2, …, N do
5.                  Receive P t , D t , l o c t , u , l o c t , u ¯
6.                  Obtain o i , t = { P t , D t , C t , l o c t , u , l o c t , u ¯ } by the GCLA mechanism
7.                  Obtain the UAV action a i , t through π i , θ ( a i , t | o i , t )
8.           End
9.           Execute actions a t and update the environment
10.         Receive the environment reward r t and next state o t + 1
11.         Store trajectory τ + = [ o t , r t , a t , o t + 1 ]
12.      End
13.      Compute advantage estimate A t via GAE on τ
14.      Compute the discount reward R ^ t and normalize
15.      Split trajectory τ into chunks of length Y and store in experience pools
16.      For mini-batch = 1,2,…,H do
17.         b ← random sample mini batch from experience pools with all UAV data
18.         Compute loss functions L t ( θ ) , L t ( ϕ ) with data b
19.         Adam to update θ on L t ( θ )
20.         Adam to update ϕ on L t ( ϕ )
21.      End
22. End

4. Experiments

To verify the effectiveness of the proposed ETS–MAPPO algorithm, it was compared against five classic MARL algorithms: MAPPO [17], MADDPG [31], MATD3 [32], QMIX [33], and IQL [34]. Ablation experiments were also conducted to demonstrate the contributions of the STEE network and the GCLA mechanism.

4.1. Simulation Environment and Parameter Settings

In the multi-UAV escape target search simulation scenario established, the environment was divided into two layers with a size of 2000 m × 2000 m. There were three UAVs, with a detection range for targets of 200 m and a detection range for stationary obstacles of 400 m. There were 10 escape targets, with a detection range for UAVs of 200 m, an escape range of 400 m, and a single escape attempt. The initial position of the target was taken to be randomly generated. The number of obstacles was 15. The simulation environment and network parameters are detailed in Table 1.
The simulation experiments were conducted using the following computer hardware and software: an Intel i5-12400F CPU manufactured by Intel Corporation and sourced from Hefei, China, 32 GB RAM, an NVIDIA GeForce RTX 4060Ti GPU manufactured by PC Partner and sourced from Dongguan, China, Python 3.11.5, and Pytorch 2.1.0.

4.2. Model Performance Analysis

The analysis of model performance began with the evaluation of the training results of each model, focusing on the convergence and performance metrics of six algorithm models: ETS–MAPPO, MAPPO, MADDPG, MATD3, QMIX, and IQL. Subsequently, the test results were analyzed to assess the generalization performance and real-time performance of the ETS–MAPPO algorithm model. Finally, the operational state of the ETS–MAPPO algorithm in the simulation environment at different time step was obtained. To ensure the reliability of the experiments, the number of network layers and neurons in each algorithm was kept consistent, and the algorithm parameters were appropriately tuned. All experiments were conducted three times under random seeds of 1, 2, and 3, and the average value of the three experiments was taken as the final experimental result.

4.2.1. Analysis of Model Training Results

Model training was conducted according to the experimental setup described above, with the convergence curves of the six algorithms—ETS–MAPPO, MAPPO, MAD-DPG, MATD3, QMIX, and IQL—presented in Figure 8. As the number of training rounds increased, the UAV average reward value gradually increased and converged. The average reward value curve of the ETS–MAPPO algorithm demonstrates a significant advantage over the other five classic algorithms.
Besides the average reward results representing the combined effect of multiple metrics, the specific parameter metrics that measure the actual performance in the METS problem include the area coverage rate, the number of collisions, the number of initial target searches, and the number of target re-searches. As shown in Figure 9a–c, the ETS–MAPPO algorithm was compared with the other five algorithms in terms of area coverage, the number of initial target searches, and the number of target re-searches. Although the convergence curves are intertwined in the early stages, the results after convergence display clear advantages.
With regard to the indicator of the number of collisions shown in Figure 9d, the convergence trend of MADDPG and MATD3 differed from the downward convergence observed in the ETS–MAPPO, MAPPO, QMIX, and IQL algorithms by exhibiting upward convergence, leading to poorer results. Additionally, the results of the ETS–MAPPO, MAPPO, and IQL algorithms were similar.
The observed results can be attributed to the adoption of the STEE network and the GCLA mechanism within the ETS–MAPPO algorithm. These components enhance the processing of spatio-temporal sequence data and environmental exploration, strengthen the network state parameters, improve the algorithm’s learning efficiency, accelerate network convergence, and optimize the search for the escape target, ultimately yielding higher reward values, as reflected in the specific parameter metrics.

4.2.2. Analysis of Model Testing Results

In order to further validate the experimental results, six algorithms, ETS–MAPPO, MAPPO, MADDPG, MATD3, QMIX and IQL, were tested. The model with the random seed of 1 was chosen for model testing, and each algorithm was tested over 20 rounds so as to take the average value as the test result; other environmental parameters were consistent with the model training above. Then, the number of targets and time steps were changed to verify the generalization performance of the ETS–MAPPO algorithm.
The model test results are shown in Table 2, which compares the performance of the six algorithms across five metrics. The ETS–MAPPO algorithm outperformed the others on four of these metrics: the average reward, the area coverage rate, the number of initial target searches, and the number of target re-searches. In terms of obstacle collisions, comparison of the training curves and other indicators revealed that the IQL algorithm experienced fewer obstacle collisions, likely due to its limited UAV exploration of the environment, rendering it less comparable. Therefore, excluding the IQL algorithm, the ETS–MAPPO algorithm surpassed the remaining four algorithms in this metric. To summarize, the ETS–MAPPO algorithm outperformed the other five classic algorithms across all five performance metrics and demonstrated superior performance overall.
Figure 10 gives the results of the environmental operation under 50 consecutive time steps in the test phase, at time step 1, time step 10, time step 30, and time step 50. At the beginning of the simulation, the UAV departed from the base and followed two paths for searching. From time step 1 to time step 10, it can be found that the UAV was trying to avoid collision with obstacles. From time step 10 to time step 30, it can be observed that the UAVs undertook a decentralized search and started searching for targets in full range, with many targets successfully searched for the first time. It can also be observed that some of the escaped targets had already escaped and there was a change in their position. From the 30th to the 50th time step, some targets were successfully searched for again. It can be observed that the UAV covered and searched most of the mission area. The above test results show that the ETS–MAPPO algorithm can reasonably and efficiently complete the area reconnaissance and escape target search.
Figure 11 presents the results of the algorithm’s real-time performance test. From the results it can be seen that, as the number of UAVs increased and the range of the explored area expanded, the average decision time per UAV also increased. The variation in average decision time per UAV remained within the millisecond range (0.8–2.4 ms). In practice, this variation did not noticeably affect UAV operation.
To verify the generalization performance of the ETS–MAPPO algorithm, we varied the number of time steps and targets during testing. From Table 3 and Table 4, it can be seen that the average reward, the number of initial target searches, and the number of target re-searches increased with the increase in time steps and target count. The area coverage ratio remained near 0.9, and the percentage of initial repeat target searches ratio stayed around 0.5. This demonstrates that the ETS–MAPPO algorithm remained effective across different number of time steps and targets.

4.3. Ablation Experiment

To investigate the impact of the STEE network and the GCLA mechanism on the ETS–MAPPO algorithm’s performance, two variants were constructed: the GCLA–MAPPO, by removing the STEE network, and the STEE–MAPPO, by removing the GCLA mechanism. The test results for these algorithms across five performance metrics were analyzed. The specific configurations of the four algorithms in the ablation experiments are presented in Table 5, where “√” indicates inclusion and “×” indicates exclusion.
Figure 12 presents the training results of each ablation experiment algorithm in terms of average reward. From these results, it can be seen that the GCLA–MAPPO and MAPPO algorithms had similar reward values, while the ETS–MAPPO and STEE–MAPPO algorithms demonstrated higher reward values, with the ETS–MAPPO algorithm being slightly lower than the STEE–MAPPO algorithm.
Table 6 shows the test results of the four ablation experimental algorithms across five performance metrics. The results demonstrate that the ETS–MAPPO, STEE–MAPPO, and GCLA–MAPPO algorithms showed improvements over the MAPPO algorithm in four metrics: the average reward, the number of collisions, the number of initial target searches, and the number of target re-searches. However, on the area coverage rate metric, the GCLA–MAPPO algorithm was associated with a slightly lower value than that associated with the MAPPO algorithm.
These results can be attributed to the STEE network’s ability to efficiently process spatio-temporal sequence data and enhance environmental exploration, thereby improving the overall performance of the algorithms. Consequently, algorithms utilizing the STEE network achieved higher reward values. The GCLA mechanism caused the UAV to focus its search near the initial target discovery area, which often includes regions that have already been explored. Given the limited number of time steps, the UAV lacked sufficient time to search unexplored areas after repeatedly scanning near the initial target area. This led to a decrease in the area coverage rate and an increase in the number of target re-searches. These findings suggest that the STEE network and the GCLA mechanism effectively enhance the performance of the ETS–MAPPO algorithm in the METS problem.

5. Conclusions

With regard to the multi-UAV escape target search task, this paper addresses the challenges associated with enabling escape target search and efficiently utilizing the sample data from the MARL algorithm, particularly in relation to maintaining a balance between utilization and exploration. This paper proposes a multi-UAV escape target search algorithm that combines the STEE network with the GCLA mechanism, built upon the MAPPO algorithm and applied to the multi-UAV escape target search task. Experimental results demonstrate that the ETS–MAPPO algorithm excels in addressing the METS problem, outperforming the other five MARL algorithms across all metrics. Ablation experiments confirm that the STEE network enhances the utilization of spatio-temporal sequence data while effectively balancing environmental exploration, thereby improving the algorithm’s overall performance. Additionally, the GCLA mechanism significantly improves performance in the escape target search problem.
Future work will investigate the performance of our proposed ETS–MAPPO algorithm in larger-scale scenarios involving more UAVs, while continually optimizing the training efficiency of the network model.

Author Contributions

G.L. completed the main experiment and wrote the initial paper; J.W. gave the experimental guidance and revised the initial article; D.Y. and J.Y. guided the revision of the article. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Francesco, T.; Costanza, F.; Nicola, C. Unmanned Aerial Vehicle for Precision Agriculture: A Review. IEEE Access 2024, 12, 69188–69205. [Google Scholar]
  2. Dimosthenis, C.T.; Stamatia, B.; Panagiotis, G.S. A Review on UAV-Based Applications for Precision Agriculture. Information 2019, 10, 349. [Google Scholar] [CrossRef]
  3. Su, Z.N.; Jiao, J.; Lin, T. Integrated development of aeromagnetic system based on UAV platform and its application in mineral resources exploration. IOP Conf. Ser. Earth Environ. Sci. 2021, 660, 012103. [Google Scholar] [CrossRef]
  4. Yang, H.Y.; Wang, J.; Wang, J.C. Efficient Detection of Forest Fire Smoke in UAV Aerial Imagery Based on an Improved Yolov5 Model and Transfer Learning. Remote Sens. 2023, 15, 5527. [Google Scholar] [CrossRef]
  5. Tang, J.Z.; Liu, D.; Wang, Q.S. Probabilistic Chain-Enhanced Parallel Genetic Algorithm for UAV Reconnaissance Task Assignment. Drones 2024, 8, 213. [Google Scholar] [CrossRef]
  6. Hu, J.Q.; Wu, H.S.; Zhan, R.J. Self-organized search-attack mission planning for UAV swarm based on wolf pack hunting behavior. J. Syst. Eng. Electron. 2021, 32, 1463–1476. [Google Scholar]
  7. Vitaly, A.; Magnus, S. Optimal search for a moving target—A geometric approach. In Proceedings of the AIAA Guidance, Navigation, and Control Conference and Exhibit, Dever, CO, USA, 14–17 August 2000. [Google Scholar]
  8. Howie, C. Coverage for robotics—A survey of recent results. Ann. Math. Artif. Intell. 2001, 31, 113–126. [Google Scholar]
  9. Liang, Z.B.; Li, Q.; Fu, G.D. Multi-UAV Collaborative Search and Attack Mission Decision-Making in Unknown Environments. Sensors 2023, 23, 7398. [Google Scholar] [CrossRef]
  10. Mohammed, A.; Houssem, B.; Tijani, A. Dynamic Target Search Using Multi-UAVs Based on Motion-Encoded Genetic Algorithm With Multiple Parents. IEEE Access 2022, 10, 77922–77939. [Google Scholar]
  11. Chen, R.L.; Chen, J.L.; Li, S.Q. A review of multi-agent reinforcement learning methods. Inf. Countermeas. Technol. 2024, 3, 18–32. [Google Scholar]
  12. Littman, M.L. Markov Games as a Framework for Multi-Agent Reinforcement Learning. In Machine Learning Proceedings; Elsevier: Amsterdam, The Netherlands, 1994; pp. 157–163. [Google Scholar]
  13. Su, K.; Qian, F. Multi-UAV Cooperative Searching and Tracking for Moving Targets Based on Multi-Agent Reinforcement Learning. Appl. Sci. 2023, 13, 11905. [Google Scholar] [CrossRef]
  14. Cheng, C.X.; Zhang, H.; Sun, Y. A cross-platform deep reinforcement learning model for autonomous navigation without global information in different scenes. Control Eng. Pract. 2024, 150, 105991. [Google Scholar] [CrossRef]
  15. Shen, G.Q.; Lei, L.; Zhang, X.T. Multi-UAV Cooperative Search Based on Reinforcement Learning with a Digital Twin Driven Training Framework. IEEE Trans. Veh. Technol. 2023, 72, 8354–8368. [Google Scholar] [CrossRef]
  16. Lu, Z.A.; Wu, Q.H.; Zhou, F.H. Algorithm for intelligent collaborative target search and trajectory planning of MAv/UAv. J. Commun. 2024, 45, 31–40. [Google Scholar]
  17. Yu, C.; Velu, A.; Vinitsky, E. The Surprising Effectiveness of PPO in Cooperative Multi-Agent Games. Adv. Neural Inf. Process. Syst. 2022, 35, 24611–24624. [Google Scholar]
  18. Wei, D.X.; Zhang, L.; Liu, Q. UAV Swarm Cooperative Dynamic Target Search: A MAPPO-Based Discrete Optimal Control Method. Drones 2024, 8, 214. [Google Scholar] [CrossRef]
  19. Bertuccelli, L.F.; How, J.P. Robust UAV search for environments with imprecise probability maps. In Proceedings of the 44th IEEE Conference on CCD-ECC, Seville, Spain, 12–15 December 2005. [Google Scholar]
  20. Zhang, B.Q.; Lin, X.; Zhu, Y.F.; Tian, J. Enhancing Multi-UAV Reconnaissance and Search Through Double Critic DDPG With Belief Probability Maps. IEEE Trans. Intell. Veh. 2024, 9, 3827–3842. [Google Scholar] [CrossRef]
  21. Hou, Y.K.; Zhao, J.; Zhang, R.Q. UAV Swarm Cooperative Target Search: A Multi-Agent Reinforcement Learning Approach. J. Cloud Comput. 2024, 9, 568–578. [Google Scholar] [CrossRef]
  22. Vinh, K.; Gebreyohannes, S.; Karimoddini, A. An Area-Decomposition Based Approach for Cooperative Tasking and Coordination of UAVs in a Search and Coverage Mission. In Proceedings of the 2019 IEEE Aerospace Conference, Big Sky, MT, USA, 2–9 March 2019. [Google Scholar]
  23. Chung, T.H.; Burdick, J.W. Analysis of search decision making using probabilistic search strategies. IEEE Trans. Robot. 2012, 28, 132–144. [Google Scholar] [CrossRef]
  24. Frans, A.O.; Matthijs, T.S.; Nikos, V. Optimal and Approximate Q-value Functions for Decentralized POMDPs. J. Artif. Intell. Res. 2011, 32, 289–353. [Google Scholar]
  25. Mahajan, A.; Rashid, T.; Samvelyan, M. Maven: Multi-agent variational exploration. In Proceedings of the 33rd Conference on NeurIPS, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  26. Zhang, B.; Hu, W.H.; Cao, D. Novel Data-Driven decentralized coordination model for electric vehicle aggregator and energy hub entities in multi-energy system using an improved multi-agent DRL approach. Appl. Energy 2023, 339, 120902. [Google Scholar] [CrossRef]
  27. Shi, D.X.; Peng, Y.X.; Yang, H. A DQN-based deep reinforcement learning motion planning method for multiple agents. Comput. Sci. 2024, 51, 268–277. [Google Scholar]
  28. Sun, D.M.; Chen, Y.M.; Li, H. Intelligent Vehicle Computation Offloading in Vehicular Ad Hoc Networks: A Multi-Agent LSTM Approach with Deep Reinforcement Learning. Mathematics 2024, 12, 424. [Google Scholar] [CrossRef]
  29. He, H.R.; Zhou, F.Q.; Zhao, Y.K. Hypergraph convolution mix DDPG for multi-aerial base station deployment. J. Cloud Comput. 2023, 12, 172. [Google Scholar] [CrossRef]
  30. Shi, X.J.; Chen, Z.R.; Wang, H. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. Computer Science. In Proceedings of the 29th Annual Conference on NIPS, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  31. Lowe, R.; Wu, Y.; Tamar, A. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. In Proceedings of the 31st Annual Conference on NIPS, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  32. Fujimoto, S.; van Hoof, H.; Meger, D. Addressing Function Approximation Error in Actor-Critic Methods. In Proceedings of the 35th ICML, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  33. Rashid, T.; Samvelyan, M.; Farquhar, G. Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning. J. Mach. Learn. Res. 2020, 21, 178. [Google Scholar]
  34. Tampuu, A.; Matiisen, T.; Kodelja, D. Multiagent cooperation and competition with deep reinforcement learning. PLoS ONE 2017, 12, e0172395. [Google Scholar] [CrossRef]
Figure 1. Typical scenario diagram for a multi-UAV escape target search mission.
Figure 1. Typical scenario diagram for a multi-UAV escape target search mission.
Sensors 24 06859 g001
Figure 2. Rasterized model map of the task area.
Figure 2. Rasterized model map of the task area.
Sensors 24 06859 g002
Figure 3. UAV action space model.
Figure 3. UAV action space model.
Sensors 24 06859 g003
Figure 4. The STEE network architecture.
Figure 4. The STEE network architecture.
Sensors 24 06859 g004
Figure 5. Network parameterized noise.
Figure 5. Network parameterized noise.
Sensors 24 06859 g005
Figure 6. Processing of TEPM convolutional operations for 10 × 10 state parameters.
Figure 6. Processing of TEPM convolutional operations for 10 × 10 state parameters.
Sensors 24 06859 g006
Figure 7. The ETS–MAPPO algorithmic framework.
Figure 7. The ETS–MAPPO algorithmic framework.
Sensors 24 06859 g007
Figure 8. Comparison of the average reward values of the six algorithms.
Figure 8. Comparison of the average reward values of the six algorithms.
Sensors 24 06859 g008
Figure 9. The specific parameter metrics train results. (a) Area coverage rate, (b) the number of initial target searches, (c) the number of target re-searches, (d) the number of collisions.
Figure 9. The specific parameter metrics train results. (a) Area coverage rate, (b) the number of initial target searches, (c) the number of target re-searches, (d) the number of collisions.
Sensors 24 06859 g009
Figure 10. Results of UAV operation under 50 consecutive time steps in the simulation testing phase, where yellow pentacles represent targets, green diamonds represent obstacles, and black triangles represent UAVs.
Figure 10. Results of UAV operation under 50 consecutive time steps in the simulation testing phase, where yellow pentacles represent targets, green diamonds represent obstacles, and black triangles represent UAVs.
Sensors 24 06859 g010
Figure 11. Results of the algorithm’s real-time performance test.
Figure 11. Results of the algorithm’s real-time performance test.
Sensors 24 06859 g011
Figure 12. Comparison of average reward results from ablation experiment.
Figure 12. Comparison of average reward results from ablation experiment.
Sensors 24 06859 g012
Table 1. The simulation environment and network parameter settings.
Table 1. The simulation environment and network parameter settings.
ParameterValue
Total running steps2,000,000
Time step per round50
Actor network learning rate5 × 10−4
Critic network learning rate5 × 10−4
Hidden layer size64
Discount factor0.99
Detection payload correct rate0.90
Table 2. Six algorithms test results.
Table 2. Six algorithms test results.
AlgorithmsAverage RewardArea Coverage RateNumber of CollisionsNumber of Target Initial SearchesNumber of Target Re-Searches
IQL24.850.152.68 *0.550.09
QMIX41.950.6318.115.252.43
MADDPG79.730.8018.956.932.18
MATD371.110.7812.677.871.80
MAPPO88.860.825.127.452.75
ETS–MAPPO106.210.893.189.095.11
* indicates that the data is not compared with other data.
Table 3. Experimental test results for changing the number of time steps.
Table 3. Experimental test results for changing the number of time steps.
Time StepsAverage RewardArea Coverage ReteNumber of CollisionsNumber of Target Initial SearchesNumber of Target Re-Searches
50106.210.893.189.095.11
60111.650.969.709.555.65
80124.860.9916.209.737.05
100129.761.0018.49.827.10
Table 4. Experimental test results of changing the number of targets.
Table 4. Experimental test results of changing the number of targets.
Target NumbersAverage RewardArea Coverage RatePercentage of Initial Repeat Target SearchesNumber of Target Initial SearchesNumber of Target Re-Searches
5102.260.940.514.552.33
10106.210.890.579.095.11
15107.510.920.4813.056.35
Table 5. Ablation experimental setup.
Table 5. Ablation experimental setup.
AlgorithmsSTEE NetworkGCLA Mechanism
ETS–MAPPO
STEE–MAPPO×
GCLA–MAPPO×
MAPPO××
√ indicates that the network or mechanism is used. × indicates that the network or mechanism is not used.
Table 6. Results of ablation experiment.
Table 6. Results of ablation experiment.
AlgorithmsAverage RewardArea Coverage ReteNumber of CollisionsNumber of Target Initial SearchesNumber of Target Re-Searches
MAPPO88.860.825.127.452.75
STEE–MAPPO109.280.930.369.853.05
GCLA–MAPPO89.980.770.848.894.23
ETS–MAPPO106.210.893.189.095.11
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liao, G.; Wang, J.; Yang, D.; Yang, J. Multi-UAV Escape Target Search: A Multi-Agent Reinforcement Learning Method. Sensors 2024, 24, 6859. https://doi.org/10.3390/s24216859

AMA Style

Liao G, Wang J, Yang D, Yang J. Multi-UAV Escape Target Search: A Multi-Agent Reinforcement Learning Method. Sensors. 2024; 24(21):6859. https://doi.org/10.3390/s24216859

Chicago/Turabian Style

Liao, Guang, Jian Wang, Dujia Yang, and Junan Yang. 2024. "Multi-UAV Escape Target Search: A Multi-Agent Reinforcement Learning Method" Sensors 24, no. 21: 6859. https://doi.org/10.3390/s24216859

APA Style

Liao, G., Wang, J., Yang, D., & Yang, J. (2024). Multi-UAV Escape Target Search: A Multi-Agent Reinforcement Learning Method. Sensors, 24(21), 6859. https://doi.org/10.3390/s24216859

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop