Next Article in Journal
Edge AI for Early Detection of Chronic Diseases and the Spread of Infectious Diseases: Opportunities, Challenges, and Future Directions
Previous Article in Journal
GRAPH4: A Security Monitoring Architecture Based on Data Plane Anomaly Detection Metrics Calculated over Attack Graphs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Maximizing UAV Coverage in Maritime Wireless Networks: A Multiagent Reinforcement Learning Approach

1
Beijing Key Laboratory of Transportation Data Analysis and Mining, School of Computer and Information Technology, Beijing Jiaotong University, Beijing 100044, China
2
School of Information Communication and Engineering, Communication University of China, Beijing 100024, China
*
Author to whom correspondence should be addressed.
Future Internet 2023, 15(11), 369; https://doi.org/10.3390/fi15110369
Submission received: 15 October 2023 / Revised: 7 November 2023 / Accepted: 10 November 2023 / Published: 16 November 2023
(This article belongs to the Section Smart System Infrastructure and Applications)

Abstract

:
In the field of ocean data monitoring, collaborative control and path planning of unmanned aerial vehicles (UAVs) are essential for improving data collection efficiency and quality. In this study, we focus on how to utilize multiple UAVs to efficiently cover the target area in ocean data monitoring tasks. First, we propose a multiagent deep reinforcement learning (DRL)-based path-planning method for multiple UAVs to perform efficient coverage tasks in a target area in the field of ocean data monitoring. Additionally, the traditional Multi-Agent Twin Delayed Deep Deterministic policy gradient (MATD3) algorithm only considers the current state of the agents, leading to poor performance in path planning. To address this issue, we introduce an improved MATD3 algorithm with the integration of a stacked long short-term memory (S-LSTM) network to incorporate the historical interaction information and environmental changes among agents. Finally, the experimental results demonstrate that the proposed MATD3-Stacked_LSTM algorithm can effectively improve the efficiency and practicality of UAV path planning by achieving a high coverage rate of the target area and reducing the redundant coverage rate among UAVs compared with two other advanced DRL algorithms.

1. Introduction

Traditional ocean monitoring methods are limited by constraints on human resources, time, and space, making it difficult to meet the demands of large-scale, high-spatiotemporal-resolution ocean monitoring. Therefore, an efficient, flexible, and cost-effective ocean monitoring approach is needed to achieve high-quality data collection and information acquisition in complex and dynamic ocean environments. Unmanned aerial vehicles (UAVs), due to their small size, low cost, and high degree of flexibility, have been widely used in military and civilian fields, such as enemy surveillance, data collection, and disaster relief, among others [1,2,3]. As shown in Figure 1, in ocean monitoring tasks, drones, with their flexible high-altitude perspective, can rapidly locate and monitor areas using various sensors and equipment in three-dimensional space, acquiring oceanic environmental data at different heights and speeds. This provides powerful support for oceanographic research and environmental protection [4].
Single UAVs have relatively high levels of controllability and operability, but there are also some drawbacks. Firstly, because a single UAV can only perform one task at a time during monitoring of marine areas, a malfunction of the UAV can lead to the entire mission failing. Secondly, due to the limited endurance of UAVs and the additional equipment they carry, they often face battery-life issues during missions. In contrast, when multiple UAVs are used for marine monitoring, tasks can be allocated reasonably to solve the endurance problem [5,6]. Therefore, the coordination of multiple UAVs for area coverage has become the main research direction with respect to improving monitoring, positioning, and perception capabilities in marine environments [7].
Nowadays, for the problem of cooperative coverage path planning of UAV swarms, most studies adopt intelligent optimization algorithms such as particle swarm optimization and the genetic algorithm for stochastic search. However, the marine environment is complex and ever-changing, and the above algorithms can only obtain optimal and suboptimal solutions through multiple iterations, which cannot be generalized to dynamic, unknown environments. Furthermore, the formulated problem is a typical non-convex problem, which are extremely difficult to solve, and its complexity increases sharply with the number of parameters to be optimized [8,9]. Therefore, how to plan the path of UAVs reasonably in marine monitoring tasks to achieve efficient coverage of the target area is still subject to a series of challenges in academia.
Fortunately, with the rapid development of artificial intelligence, the combination of deep reinforcement learning (DRL) algorithms and UAVs enables more intelligent and autonomous execution of tasks [10]. Traditional reinforcement learning algorithms only consider the interaction between a single agent and the environment, making it difficult to effectively handle the problem of multiagent cooperative decision making [11]. Nguyen et al. [12] conducted a comprehensive survey of multiagent DRL. In multiagent reinforcement learning, multiple agents learn and influence each other to optimize their cooperative decision making. In the context of ocean monitoring applications, multiple UAVs need to work together to achieve efficient target coverage, and using multiagent reinforcement learning can better address this issue. Therefore, compared to traditional reinforcement learning algorithms, the use of multiagent reinforcement learning has a broader and more significant application prospect in multi-UAV path-planning research based on ocean monitoring. Compared with other multiagent algorithms, the MATD3 algorithm has several advantages. (i) The MATD3 algorithm ensures stability and improves the robustness of the algorithm in multiagent decision making. (ii) The algorithm also uses double Q networks, which can more accurately evaluate the value of actions. MATD3 not only coordinates the actions of multiple UAVs but also handles uncertainty caused by environmental and UAV behavioral changes. Therefore, using the MATD3 algorithm to solve multi-UAV path-planning problems in ocean monitoring is a promising research direction.
In this paper, considering the energy limitation and obstacle-avoidance ability of UAVs, we investigate how to achieve maximum coverage of the target area in ocean monitoring scenarios. Specifically, we propose a path-planning method based on a multiagent DRL framework to achieve maximum coverage of the target area. The traditional MATD3 algorithm can only consider the current state of the agents, ignoring the historical interaction information among the agents and environmental changes, which leads to poor path-planning results. Our method uses long short-term memory (LSTM) to store and utilize previous states, which helps to solve the path-planning problem in cases of incomplete information. By leveraging multiagent collaboration and experience sharing in MATD3, we address the non-convex problem of optimizing the objective function. Experimental results demonstrate that our method achieves good performance in terms of coverage paths in ocean monitoring scenarios.
The main contributions of this work can be summarized as follows:
(1)
A path-planning method based on a multi-intelligence deep reinforcement learning framework is proposed to maximize the coverage of the target area.
(2)
To address the path-planning problem under the condition of partially incomplete information, we utilized an S-LSTM to store and utilize previous states. Furthermore, to address the non-convex problem in optimizing the objective function, we employed multiagent collaboration and experience sharing within the MATD3 algorithm.
(3)
The simulation results demonstrate that the proposed MATD3-Stacked_LSTM algorithm has better convergence and significantly improves the coverage of the target area compared to the other two algorithms.

2. Related Work

Nowadays, heuristic algorithms are widely used for UAV path planning. For instance, Recep Ozdag [13] developed an optimal fitness value search method for continuous and discrete data ranges based on particle swarm optimization (PSO) and electromagnetic-like (EML) algorithms, aiming to ensure the best coverage rate of user devices positioned at certain distance intervals. However, heuristic algorithms require reoptimization and solving when ocean environment data change, making them less adaptive. Additionally, due to inherent algorithmic limitations, they often cannot provide corresponding solutions quickly, making it difficult to adapt to scenarios with high real-time requirements. Therefore, combining cooperative exploration among drones and DRL algorithms can address the challenges posed by the real-world environment. Many existing DRL methods focus only on a single-drone scenario. Bouhamed et al. [14] proposed an autonomous drone path-planning framework using the DDPG method, aimed at training drones to avoid obstacles and reach target points. Li et al. [15] proposed a real-time path-planning approach based on a deep Q network (DQN) to solve the problem of an unknown environment. Unlike quadrotor UAVs, this method provides real-time action planning for fixed-wing UAVs on a feasible path while satisfying the kinematic constraints.
Recently, several works have been proposed to address the problem of path planning using multiple UAVs. Wang et al. [16] proposed a D3QN-based learning method for decision-making policies of typical UAVs, which requires no prior knowledge of the environment (such as channel propagation models and obstacle positions) and other UAVs (such as their tasks, movements, and policies). Numerical results show that real-time navigation can be performed efficiently with high success rates and low collision rates. Liu et al. [17] studied a mobile edge computing network installed on UAVs. They formulated the trajectory optimization problem as a Markov decision process and developed a QoS-based action selection policy based on the DDQN algorithm. DTPC [18] was designed in a multiagent reinforcement learning (MARL) framework based on MADDPG, which solves the problem of decentralized joint trajectory and transmission power control for multi-UAV ABS networks. Harald Bayerlein et al. [19] proposed a novel end-to-end reinforcement learning approach for autonomous trajectory planning of UAVs to collect data from IoT devices in urban environments in support of next-generation communication networks. HAS-DQN [20] is an algorithm that combines the hexagonal area search (HAS) approach with a multiagent DQN. It effectively addresses collision issues among drones by constraining the overall coverage area of the UAVs. Xie et al. [9]. designed a locally optimal UAV path-planning algorithm by leveraging 3D antenna radiation patterns and multistage D3QN. Numerical results substantiate the algorithm’s effectiveness in the context of connectivity-aware UAV path planning.
However, in MARL-based path planning algorithms, each intelligent agent can only observe partial environmental information, overlooking the historical interaction among agents and environmental changes [21]. This may lead to increased difficulties in information propagation and coordination, especially in large-scale multiagent systems. Recurrent Neural networks (RNNs) can memorize and process sequential data [22]. However, RNNs tend to suffer from the vanishing gradient problem, causing them to forget long-term information. LSTM, as a sequence-modeling method, addresses this issue by introducing “forget gates” to improve the RNN structure. Therefore, using MATD3 combined with LSTM can consider the temporal correlations in agent path planning, enabling agents to make decisions with a more long-term planning capability rather than solely focusing on the current state.

3. System Model and Problem Statement

In this study, we propose a multidrone path coverage planning method based on MDRL for data collection in unknown marine environments. In this section, we first introduce the various system models in detail, then formulate the optimization problem for maximization of the coverage rate. We summarize the parameters used in this section in Table 1.

3.1. System Model

3.1.1. UAV Model

Let M = 1 , 2 , 3 , , M denote the set of UAVs. At each time step (t), the flight direction of UAV m M is determined by the turn angle ( θ m , t [ 0 , 2 π ) ). The flight distance is determined by d m , t [ 0 , d max ] , where d max is the maximum allowed flight distance, and UAVs cannot exceed the boundaries of the target area.
The trajectory of each drone can be represented as:
P m = ( x m , 1 , y m , 1 , z m , 1 ) , ( x m , 2 , y m , 2 , z m , 2 ) , , ( x m , t , y m , t , z m , t )
where P m is the trajectory of UAV m, and ( x m , t , y m , t , z m , t ) represents the position of UAV m at time step t.
When performing coverage tasks, UAVs are not allowed to enter no-fly zones and pass through obstacles. Let F = { f 1 , f 2 , , f n } represent the set of no-fly zones and O = { o 1 , o 2 , , o k } represent the set of obstacles. The UAV trajectories should avoid crossing F and O.
The constraints can be formulated as follows: The UAV trajectory should not intersect any no-fly zone:
( x m , i , y m , i , z m , i ) , ( x m , i + 1 , y m , i + 1 , z m , i + 1 ) F , m M , i = 1 , 2 , , t 1
The UAV trajectory should not intersect any obstacle:
( x m , i , y m , i , z m , i ) , ( x m , i + 1 , y m , i + 1 , z m , i + 1 ) O , m M , i = 1 , 2 , , t 1
where t is the number of time steps taken by UAV m in its trajectory. The notation ( x m , i , y m , i , z m , i ) represents the position of UAV m at time step i in the 3D space. The constraints ensure that each UAV’s trajectory avoids both the no-fly zones and obstacles during the coverage tasks.
Subsequently, the trajectory of the UAV must not exceed the specified airspace boundary, which is denoted as
B = { ( x m , 1 , y m , 1 , z m , 1 ) , ( x m , 2 , y m , 2 , z m , 2 ) , , ( x m , b , y m , b , z m , b ) } .
Each point ( r i ) in B represents a 3D coordinate in the airspace. To ensure that the UAV remains within the defined airspace at all times, the position of the UAV at each time step (i) should satisfy the following constraint:
( x i , y i , z i ) B , i = 1 , 2 , , t
This constraint ensures that at each time step, the position of the UAV is within the specified airspace boundary defined by the B set. It guarantees that the UAV remains within the specified airspace during its entire trajectory.
A UAV flying at a given height (H) from the ground acquires an image corresponding to a specific portion of the area (projected area). The size of the projected area depends on the height (H) and the angle of view ( f O A ). The size of the projected area ( L x , L y ) can be calculated as
L x = 2 H · tan f O A x 2
L y = 2 H · tan f O A y 2 .
Therefore, the spacial resolution ( s r ) obtained by taking a picture at height H is
s r = l x L x = l x 2 H · tan f O A x 2 ,
where the pixel width of the photo is denoted as l x . Similarly, the formula for calculating the vertical spatial resolution ( R y ) is given as l x / ( 2 × H × tan ( f O A y / 2 ) ) .
Therefore,
H min H l x 2 s r · tan f O A x 2
where H min is the lower bound for the height of UAVs for safety. Hence, the mission requirement imposes a constraint on the maximum height the UAV can fly, which is
H max = l x 2 s r · tan f O A 2
Furthermore, we denote the distance between UAV m and UAV m in t as D m , m , t , which can be expressed as
D m , m , t = X m , t X m , t 2 + Y m , t Y m , t 2
We assume that the UAVs should keep a minimal distance of D ^ min to avoiding their collision in each T . Then, we have
D m , m , t D min , m , m M , m m .
To avoid interference beyond the coverage area, we consider ensuring that the minimum distance between drones is greater than their communication coverage range. Therefore, we introduced communication coverage constraints into path optimization, which can be expressed as:
d m m > = R cov
where d m m is the communication distance between UAVs m and m , and  R c o v is the communication coverage radius.
In addition, we can obtain the maximum velocity and maximum acceleration of the UAV.
V m , t V m , t 1 Δ t < m a x ( A m )
V m , t V max
During the execution of a mission, constraints on the turning angles of multiple drones need to be considered. Turning a drone requires both time and space, and excessively large or small turning angles can affect the efficiency of the drone’s path planning and execution. If the turning angle is too large, the drone requires more time and space to complete the turn, which can reduce its execution efficiency. Conversely, if the turning angle is too small, the drone may deviate from the planned path during execution, resulting in flight errors.
| θ m ( t ) θ m ( t 1 ) | Δ θ , m M , t 2
where θ m ( t ) represents the heading of the m-th drone at time t, while Δ θ denotes the maximum turning angle.
In addition, to ensure that drones do not leave the coverage area when turning and to control the path-planning trajectory of the drones, avoiding excessively complex path planning, constraints on the turning radius need to be considered in multidrone path planning.
cos ( θ m ( t ) ) sin ( θ m ( t ) ) sin ( θ m ( t ) ) cos ( θ m ( t ) ) Δ x Δ y R , m , t
where Δ x and Δ y refer to the difference between the x and y coordinates of the m-th drone between time steps t and t 1 , θ m ( t ) is the turn angle of the m-th drone at time, R represents the minimum turn radius in drone path planning, and · is the Euclidean paradigm (L2 paradigm) used to compute the length of a vector obtained by matrix–vector multiplication. For vector ( x , y ) , the Euclidean norm is given by ( x , y ) = x 2 + y 2 .

3.1.2. Energy Model

During mission execution, each drone has a different battery-life limit. Failure to consider battery-life limits may result in certain drones losing control due to battery depletion during task execution and thus being unable to complete the task, affecting coverage and task performance. Therefore, to ensure that each drone can complete the task and return, constraints on battery-life limits need to be considered.
Taking into account factors such as flight energy consumption and turning energy consumption, we define the energy constraint of each drone as its total energy consumption throughout the flight mission, which cannot exceed its energy capacity. Let the energy capacity of each drone be denoted as C, the energy consumption rate be denoted as r , the total time denoted as T, and the initial energy level denoted as E 0 . The energy level ( E t ) of a drone at time t can be calculated using the following equation:
E t = E 0 0 t r d x d t d t
where x is the position vector of the drone, d x d t is the magnitude of the drone’s velocity, and t is the current time.
Therefore, for the power constraint of each UAV can be described using the following formula:
0 T r d x d t d t C E 0
Considering the flight energy consumption, turning energy consumption, and communication energy consumption, let the speed of UAV i be V i and its flight energy consumption at moment t be E i ( t ) ; then:
0 t f E m t d t E m , m a x
0 t f E m t u r n t d t E m , m a x t u r n
0 t f E m m c o m m t d t E m m , m a x c o m m
where t f denotes the moment when the UAV ends its flight, E m , m a x is the maximum energy consumption of UAV m, E m , m a x t u r n is the maximum energy consumption of UAV m for a turn, and E m m , m a x c o m m is the maximum communication energy consumption between UAV m and UAV m .
E m , m a x t u r n + E m m , m a x c o m m + E m , m a x 0 T r d x d t d t
Therefore, our goal is to find a set of paths ( ( x m , 1 , y m , 1 , z m , 1 ) , ( x m , 2 , y m , 2 , z m , 2 ) , , ( x m , t , y m , t , z m , t ) ) that maximizes the covered area. Specifically, our objective is to maximize the following expression, considering the energy consumption and turn constraints:
1 N i = 1 N ( x i , y i , z i ) A
where N represents the total number of grid cells in the grid environment. The formula for maximizing the coverage rate describes how to evaluate the quality of a set of paths. A denotes the set of all non-obstacle and non-restricted areas, where x i A is an indicator function that determines whether a path ( x i ) can cover a non-obstacle and non-restricted area. i = 1 N x i A indicates the number of cells that the set of paths ( x 1 , x 2 , , x m ) can cover.

3.2. Problem Formulation

The objective of the optimization problem is to maximize this coverage rate. The coverage path-planning problem consists of one objective function and ten constraint conditions, which represent different aspects of task performance. We formalize the optimization problem as follows:
max x 1 , x 2 , , x m 1 N i = 1 N I ( ( x i , y i , z i ) A )
where ( x i , y i , z i ) represents the path of the i-th UAV, A denotes the coverable area, and I ( x i A ) indicates whether x i falls within A .
s. t.:
(1)
d m m R cov m , m M , m m
(2)
0 d m , t d max , m M , t T
(3)
H min H l x 2 r · tan f O A x 2
(4)
D m , m , t D min , m , m M , m m
(5)
V m , t V m , t 1 Δ t < m a x ( A m ) , m M
(6)
0 t r d x d t d t C E 0
(7)
E m , m a x t u r n + E m m , m a x c o m m + E m , m a x 0 T r d x d t d t
(8)
| θ m t θ m t 1 | Δ θ , m M , t T

4. Design

4.1. MATD3

The MADDPG algorithm tends to overestimate Q value, especially during early stages of training [23]. It should be noted that most real-world applications are stochastic, which has been shown to lead to higher value-function overestimation due to added noise from function approximator errors.
The MATD3 algorithm extends TD3 to the multiagent setting in a way similar to how DDPG is extended to MADDPG. This algorithm addresses the stability problem and non-convex optimization objective in collaborative multiagent decision making. It improves upon the DDPG algorithm by optimizing policies, introducing interactions among multiple agents and using double Q networks to evaluate action values. In Figure 2, ⑤ and ⑥ illustrate the update process of the actor and critic networks, respectively.
The Q-function update formula in the MATD3 algorithm is expressed as:
L i Q θ = E s , a , r , s U D y i Q s , a ; θ 2
where L i Q ( θ ) is the loss function of the critic network for agent i with its parameters ( θ ); E ( s , a , r , s ) U ( D ) denotes the expectation over the replay buffer (D) containing samples of state–action–reward–next state tuples; and y i is the target value for updating the critic, representing the TD target for the MATD3 algorithm.
The TD target ( y t = r t + γ Q ( s t + 1 , μ ( s t + 1 | θ μ ) | θ Q ) ) in the MATD3 algorithm is computed by the target critic network. To enhance exploration and accelerate convergence, a clipped Gaussian noise is introduced for action exploration. Specifically, a noise vector ( ϵ ) is generated based on the deterministic policy ( μ ( s | θ μ ) ), then multiplied by a learnable parameter ( σ ) and added to the selected action; that is:
a = μ ( s | θ μ ) + ϵ , ϵ N ( 0 , σ 2 )
Therefore, the noisy Q value can be written as Q ( s t , a t | θ Q ) + ϵ , where θ Q represents the parameters of the critic network. The clipped Gaussian noise can be bounded using the clip function, restricting it to a certain range, such as [ c , c ] . Thus, the noisy target ( y t ) can be expressed as:
y t = r t + γ Q ( s t + 1 , μ ( s t + 1 | θ μ ) | θ Q ) + c l i p ( ϵ , c , c )
The actor network update in MATD3 is as follows:
θ i J 1 N j θ i Q i s , a i ; θ i | s = s j , a i = f i o j ; θ f i
where f i represents the actor network of the i-th agent, o j represents the observation vector of the j-th agent, and  N denotes the number of training samples. θ i J is the gradient of the actor network’s objective function for agent i with respect to its parameters ( θ i ). The update is approximated using samples from the replay buffer, where s j is the state observation of the j-th sample, and  f i is the actor network function for agent i with its parameters ( θ f i ).
The MATD3 algorithm exhibits excellent performance in multiagent cooperation problems, but it has certain limitations when dealing with high-dimensional state spaces, as it cannot fully exploit the dynamic relationships and patterns between state sequences. LSTM neural networks, on the other hand, can capture the dynamic relationships and patterns within state sequences, providing better state representations. Combining MATD3 and LSTM can enable agents to better understand and adapt to changes in the environment, further improving the performance of multiagent systems.
A single LSTM layer consistently underperforms the stacked LSTM structure. To capture the experience gained by UAVs during environmental interactions, we employed a two-layer stacked LSTM network. As illustrated in Figure 3, in the first LSTM layer, the UAV’s position, velocity, and direction serve as inputs. The outputs from this layer, along with the rewards and actions obtained by the UAV in the previous time step, are then fed into the second LSTM layer.
Moreover, LSTM can store and utilize previous states through its memory function, which is beneficial for solving path-planning problems under partially observable conditions. The output of the LSTM is given by:
h t = L S T M s t , h t 1
where s t represents the state at time t, and  h t denotes the output state of the LSTM.
Finally, the output of the LSTM is combined with the observation vector in MATD3 to obtain the agent’s state representation:
s t i = o t i h t i
where o t i is the observation vector of agent i at time t, and  h t i is the LSTM output state of agent i at time t.
Thus, as shown in Figure 2, the MATD3-Staked_LSTM algorithm leverages LSTM networks to model state representations, capturing temporal information within time series data, as well as sequential dependencies between UAVs. The agent selects actions through an actor network and incorporates noise upon execution. Subsequently, a batch of data is randomly sampled from the experience replay buffer comprising states, actions, rewards, and next states. A critic network then computes target Q values based on this sampled batch. The TD target is calculated using the target Q values to minimize the loss function, thereby updating the critic network. Concurrently, the actor network parameters are updated through policy gradient estimation based on the sampled policies. Finally, the target networks are softly updated. As presented in Algorithm 1, we can obtain the MATD3-Stacked_LSTM-based path-planning algorithm for multiple UAVs.
Algorithm 1 Multi-UAV coverage path planning based on MATD3-Stacked_LSTM algorithm
1:
Initialize critic network parameters θ Q , actor network parameters θ μ , set discount factor γ , experience replay buffer D
2:
for t = 1 to T do
3:
   Get the current state s t representation from LSTM;
4:
   Select action a i
5:
   According to the noise ϵ i N ( 0 , σ ) get the execution action a i μ s t + ϵ , observe the r value
6:
   for agent i = 1 to N do
7:
       Randomly fetch a batch of data ( s i , a i , r i , s i + 1 ) from experience playback cache D
8:
       Process via LSTM network;
9:
       Calculate the target Q value: Q ( s i + 1 , u )
10:
     Calculate TD target: y i = r i + γ Q ( s i + 1 , u + ϵ ) ,where ϵ i c l i p ( N ( 0 , σ ) )
11:
     Update critic network by minimizing the loss L = ( y i Q ( s i , a i ) ) 2
12:
     Update actor network using the sampled policy gradient:
      θ μ J 1 N i = 1 N a Q ( s i , a ; θ Q ) | a = μ ( s i ; θ μ ) θ μ μ ( s i ; θ μ )
13:
     Update target networks:
      θ Q θ Q τ + θ Q ( 1 τ )
      θ μ θ μ τ + θ μ ( 1 τ )
14:
   end for
15:
end for

4.2. Markov Decision Process

In the MATD3 algorithm, each UAV has its own agent, and they collaborate to accomplish the path-planning task. As shown in Figure 4, each agent receives state information and, based on this information, decides on the next action. We model the UAV path-planning problem as a multiagent reinforcement learning problem, with reinforcement learning model < S , A , π , R , P >. Suppose there are m UAVs that need path planning; the state ( s i ) of each UAV can be represented as its position. Each UAV needs to take an action ( a i ) to move to the next position. The policy ( π i ) of each UAV can be represented as a probability distribution mapping the current state to the action space.

4.2.1. Environment

e t represents the spatial environment where the UAVs are located at time t, including no-fly zones and other obstacles.

4.2.2. State

At each time step (t), the observation state is given as o t = [ p t , v t , d t ] , where p t , v t , and d t represent the position, velocity, and direction of the m-th UAV, respectively. The state space is defined as
s t = [ o t L ^ + 1 , , o t 1 , o t ]
where L ^ represents the length of the historical state sequence, encompassing the past L ^ states of the i-th drone, leading up to its current state at time step t.

4.2.3. Action

The available actions for all UAVs are a t = ( a 1 , t , a 2 , t , , a n , t ) , where a i , t represents the actions that the i-th UAV can choose at time step t. a t A represents the actions that the UAVs can perform at time t, including moving east, south, west, and north, as well as and descending.

4.2.4. Reward

In this task, we want agents to cover as much of the map as possible. Therefore, we can use the coverage area of the agents as part of the reward. Specifically, we can evaluate the performance of the agents by calculating the number of grid cells covered by the agents on the map. Then, we can divide this number by the total number of grid cells to obtain a coverage rate. This coverage rate is the value of the reward and can be represented as:
r r e w = 1 N i = 1 N ρ i = = 1
where N is the total number of grid cells on the map, γ i = = 1 is the i-th grid cell covered by the agents, and [ ρ i = = 1 ] is an indicator function, taking the value of 1 when ρ i = = 1 and 0 otherwise. If an agent collides with another agent or an obstacle or flies into a no-fly zone, it is penalized. Thus, we can add the penalty value for such situations to the reward value to obtain the final reward value. The penalty value can be represented as:
r p u n = 1 , collide or in no flyzone 0 , otherwise
To ensure that UAVs can work continuously during the mission, the UAV’s energy consumption needs to be considered. We can use the energy consumption of the UAV as part of the reward and encourage agents to plan their paths efficiently by limiting the UAV’s energy consumption.
r p u n = k e n e r g y · E t , E t > E c c e n e r g y , E t E c
where E t represents the current energy consumption value of the UAV, and E c represents the upper limit of the UAV’s energy consumption. When the current energy consumption value ( E t ) exceeds the energy consumption limit ( E c ), the reward value corresponding to the reward function is k e n e r g y · E t , where k e n e r g y is a positive number representing the penalty coefficient for energy consumption. When the energy consumption of the UAV is less than or equal to the energy consumption limit ( E c ), the reward value corresponding to the reward function is c e n e r g y , where c e n e r g y is a positive number representing the additional reward value after the energy consumption reaches the limit, aiming to encourage agents to better control the energy consumption of the UAV.
The final reward value is the sum of the reward part and the penalty part:
r e w a r d n = r r e w + r p u n

5. Numerical Result

The task studied in this paper is multi-UAV path planning based on ocean monitoring. In this scenario, the UAVs need to detect and cover targets within a rectangular area in the ocean. There exist unknown obstacles in the area (e.g., sudden emergence of a ship or other UAVs used for detection). Therefore, the UAVs need to autonomously avoid obstacles while detecting the target points. As the target points are all in the ocean area, the scenario can be simplified to a two-dimensional plane.

5.1. Simulation Settings

We constructed a simulated scenario involving multiple UAVs performing coverage tasks. The map size was assumed to be 10 × 10 , with each grid cell having a size of 1 × 1 , amounting to a total of 100 grid cells. The simulation environment was set up using Python 3.6 and PyCharm, with the deep learning library implemented in PyTorch. The initial UAV positions were initialized randomly, and initial actions were randomly selected. The training loop was then executed using the hyperparameters listed in Table 2. At each iteration, experiences were sampled from memory to update the networks. The current policy was evaluated after completing each round of updates. The scenario simulation, algorithm implementation, initialization of parameters, and training procedure helped validate the effectiveness of the proposed MATD3 approach for multi-UAV coverage path planning in a systematic manner.

5.2. Convergence Analysis

We compared the performance of three algorithms: MATD3, MADDPG, and our proposed MATD3-Stacked_LSTM. A scenario with four UAVs was used for evaluation. As shown in Figure 5, the reward curves of MATD3 and MADDPG gradually increased and stabilized over training, but MADDPG required about 10,000 episodes to converge, while MATD3-Stacked_LSTM achieved convergence within around 5000 episodes and attained higher final rewards. This suggests that MATD3-Stacked_LSTM outperforms the other two algorithms in terms of faster convergence and superior performance. This superior performance of MATD3-Stacked_LSTM can be attributed to its use of an RNN-based LSTM network for the modeling of joint state representations. LSTM excels at capturing temporal dependencies, and in this multiagent task with long-term dependencies, the LSTM network can more comprehensively learn complex interactions among agents, leading to enhanced state representations.
As shown in Figure 6, the loss values of both the actor and critic networks decrease significantly at the beginning of the training, and the decrease rate gradually slows down with the increase in training steps, eventually approaching stability. This indicates that the MATD3-Stacked_LSTM model gradually learns the optimal policy and value function during the training process and converges to the optimal solution. This also proves the effectiveness and reliability of the model.
The primary requirement of this coverage task is to cover all areas except for obstacles and no-fly zones; hence, the coverage rate is the most important metric. The coverage rate is calculated as the number of covered cells divided by the total number of cells, that is:
C o v e r = i = 1 m j = 1 n A r e a i , j m × n
where m and n are the numbers of rows and columns of the map, respectively; A r e a ( i , j ) is the coverage status at the ith row and jth column and takes a value of 0 or 1; 1 means covered; and 0 means not covered. The range of the summation is all positions of the whole map. In this experiment, we found:
i , j [ A r e a i , j = = 1 ] 100
where 100 represents the total number of grids.
As shown in Figure 7, the coverage rates obtained by the three algorithms increase rapidly with the increase in the number of episodes, indicating that the three algorithms can quickly adapt to the task environment and gradually improve their exploration and exploitation abilities. MATD3-Stacked_LSTM models temporal dependencies in states using an LSTM network. At each time step, it aggregates observations across agents to learn coordinated joint actions. MATD3 directly operates on raw states without modeling interdependencies. MADDPG extends the single-agent DDPG to a multiagent setting but does not explicitly consider interactions between agents. The coverage rates based on the MATD3-Stacked_LSTM and MATD3 algorithms reach around 80% and 70%, respectively, while the MADDPG algorithm can only reach 60%. This underscores the benefits of explicitly incorporating temporal modeling capabilities via LSTM. It should be noted that since the no-fly zone and obstacles are not coverable, these areas need to be excluded when calculating the coverage rate, so the coverage rate cannot reach 100%.
In order to achieve collaborative data monitoring in a target area using multiple UAVs, we conducted trajectory planning for the UAVs. Figure 8 and Figure 9 show the trajectory maps for two and four UAVs, respectively, completing the coverage task in a 10 × 10 discrete environment. Each UAV cooperatively covers a specific area and avoids obstacles and no-fly zones to maximize the defined reward. As depicted in Figure 9, for the scenario with two drones, the coverage rate of the MATD3-Stacked LSTM algorithm is approximately 65%, which is significantly higher than the other two algorithms. All three algorithms effectively mitigate path overlap. Furthermore, as shown in Figure 9, when deploying four UAVs, their paths overlap, but they can still avoid obstacles and no-fly zones well. Obviously, the MATD3-Stacked_LSTM algorithm covers more cells with lower overlap rates, as shown in Figure 9a. Although the other two algorithms also cover most of the areas, there are higher coverage overlap rates. This indicates that the MATD3-Stacked_LSTM algorithm can better coordinate the decision making of multiple UAVs, thereby achieving more efficient path planning.
The coverage of certain areas may have some degree of overlap, as shown in Figure 10. Therefore, we compared the coverage overlap rate of the three algorithms in multi-UAV coverage tasks. From Figure 10, it can be seen that as the number of UAVs increases, the coverage overlap rate based on the MATD3-Stacked_LSTM algorithm becomes lower than that of the other two algorithms. For example, when the number of UAVs is 4, the coverage redundancy of MATD3-Stacked_LSTM is approximately 9.563, while the other two algorithms have redundancy rates of about 11.754 and 15.654, respectively. This represents a reduction of 18.61% compared to MATD3 and 38.91% compared to MADDPG. The results demonstrate that for multi-UAV coverage tasks, the MATD3-Stacked LSTM algorithm is more suitable for reducing overlap between UAV paths, thereby improving the efficiency and effectiveness of the overall coverage mission compared to MADDPG and MATD3. By modeling temporal dependencies, the LSTM network enables the MATD3-Stacked_LSTM approach to learn coordinated policies that achieve higher coverage rates with less redundancy during multi-UAV path planning.

6. Conclusions

In this paper, we propose an improved MATD3 algorithm with an LSTM model incorporated into the MATD3 model to address the coverage problem of UAVs in the context of ocean data monitoring. The LSTM model is used to consider time series information during the UAV path-planning process. Through comparative experiments between the three algorithms, we have demonstrated that the improved MATD3 algorithm can effectively reduce the redundancy of coverage while ensuring high coverage rates of the target area. Future work can explore online reinforcement learning methods and multiagent communication mechanisms to continuously adapt to dynamic environments during execution. This will advance the applications of drone technology in various real-world settings, making a substantial contribution to the broader field of autonomous navigation systems.

Author Contributions

Q.W. wrote the main manuscript text, including the study design, experiments, and data analysis. Z.W. prepared the figures. J.Z. reviewed and edited the English grammar, and Q.L. revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Fundamental Research Funds for the Central Universities 2023YJS104.

Data Availability Statement

The data supporting this article are from previously reported studies and datasets, which have been cited.

Acknowledgments

The authors express their gratitude to the editor and reviewers for their work and valuable comments on the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, F.; Wang, P.; Zhang, Y.; Zheng, L.; Lu, J. Survey of swarm intelligence optimization algorithms. In Proceedings of the 2017 IEEE International Conference on Unmanned Systems (ICUS), Beijing, China, 27–29 October 2017; pp. 544–549. [Google Scholar] [CrossRef]
  2. Chang, S.Y.; Park, K.; Kim, J.; Kim, J. Securing UAV Flying Base Station for Mobile Networking: A Review. Future Internet 2023, 15, 176. [Google Scholar] [CrossRef]
  3. Li, H.; Wu, S.; Jiao, J.; Lin, X.H.; Zhang, N.; Zhang, Q. Energy-Efficient Task Offloading of Edge-Aided Maritime UAV Systems. IEEE Trans. Veh. Technol. 2022, 72, 1116–1126. [Google Scholar] [CrossRef]
  4. Ma, M.; Wang, Z. Distributed Offloading for Multi-UAV Swarms in MEC-Assisted 5G Heterogeneous Networks. Drones 2023, 7, 226. [Google Scholar] [CrossRef]
  5. Dai, Z.; Xu, G.; Liu, Z.; Ge, J.; Wang, W. Energy saving strategy of uav in mec based on deep reinforcement learning. Future Internet 2022, 14, 226. [Google Scholar] [CrossRef]
  6. Jiang, Y.; Ma, Y.; Liu, J.; Hu, L.; Chen, M.; Humar, I. MER-WearNet: Medical-Emergency Response Wearable Networking Powered by UAV-Assisted Computing Offloading and WPT. IEEE Trans. Netw. Sci. Eng. 2022, 9, 299–309. [Google Scholar] [CrossRef]
  7. Savkin, A.V.; Huang, C.; Ni, W. Joint multi-UAV path planning and LoS communication for mobile-edge computing in IoT networks with RISs. IEEE Internet Things J. 2022, 10, 2720–2727. [Google Scholar] [CrossRef]
  8. Aggarwal, S.; Kumar, N. Path planning techniques for unmanned aerial vehicles: A review, solutions, and challenges. Comput. Commun. 2020, 149, 270–299. [Google Scholar] [CrossRef]
  9. Xie, H.; Yang, D.; Xiao, L.; Lyu, J. Connectivity-Aware 3D UAV Path Design with Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2021, 70, 13022–13034. [Google Scholar] [CrossRef]
  10. Qiming, Z.; Husheng, W.; Zhaowang, F. A review of intelligent optimization algorithm applied to unmanned aerial vehicle swarm search task. In Proceedings of the 2021 11th International Conference on Information Science and Technology (ICIST), Chengdu, China, 21–23 May 2021; pp. 383–393. [Google Scholar] [CrossRef]
  11. Gao, S.; Wang, Y.; Feng, N.; Wei, Z.; Zhao, J. Deep Reinforcement Learning-Based Video Offloading and Resource Allocation in NOMA-Enabled Networks. Future Internet 2023, 15, 184. [Google Scholar] [CrossRef]
  12. Nguyen, T.T.; Nguyen, N.D.; Nahavandi, S. Deep reinforcement learning for multiagent systems: A review of challenges, solutions, and applications. IEEE Trans. Cybern. 2020, 50, 3826–3839. [Google Scholar] [CrossRef] [PubMed]
  13. Ozdag, R. Multi-metric optimization with a new metaheuristic approach developed for 3D deployment of multiple drone-BSs. Peer-Peer Netw. Appl. 2022, 15, 1535–1561. [Google Scholar] [CrossRef]
  14. Bouhamed, O.; Ghazzai, H.; Besbes, H.; Massoud, Y. Autonomous UAV Navigation: A DDPG-Based Deep Reinforcement Learning Approach. In Proceedings of the 2020 IEEE International Symposium on Circuits and Systems (ISCAS), Sevilla, Spain, 10–21 October 2020; pp. 1–5. [Google Scholar] [CrossRef]
  15. Li, J.; Liu, Y. Deep Reinforcement Learning based Adaptive Real-Time Path Planning for UAV. In Proceedings of the 2021 8th International Conference on Dependable Systems and Their Applications (DSA), Yinchuan, China, 11–12 September 2021; pp. 522–530. [Google Scholar]
  16. Wang, X.; Gursoy, M.C.; Erpek, T.; Sagduyu, Y.E. Learning-Based UAV Path Planning for Data Collection with Integrated Collision Avoidance. IEEE Internet Things J. 2022, 9, 16663–16676. [Google Scholar] [CrossRef]
  17. Liu, Q.; Shi, L.; Sun, L.; Li, J.; Ding, M.; Shu, F. Path planning for UAV-mounted mobile edge computing with deep reinforcement learning. IEEE Trans. Veh. Technol. 2020, 69, 5723–5728. [Google Scholar] [CrossRef]
  18. Chen, B.; Liu, D.; Hanzo, L. Decentralized Trajectory and Power Control Based on Multi-Agent Deep Reinforcement Learning in UAV Networks. In Proceedings of the ICC 2022—IEEE International Conference on Communications, Seoul, Republic of Korea, 16–20 May 2022; pp. 3983–3988. [Google Scholar] [CrossRef]
  19. Bayerlein, H.; Theile, M.; Caccamo, M.; Gesbert, D. UAV Path Planning for Wireless Data Harvesting: A Deep Reinforcement Learning Approach. In Proceedings of the GLOBECOM 2020—2020 IEEE Global Communications Conference, Taipei, China, 7–11 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
  20. Zhu, X.; Wang, L.; Li, Y.; Song, S.; Ma, S.; Yang, F.; Zhai, L. Path planning of multi-UAVs based on deep Q-network for energy-efficient data collection in UAVs-assisted IoT. Veh. Commun. 2022, 36, 100491. [Google Scholar] [CrossRef]
  21. Prédhumeau, M.; Mancheva, L.; Dugdale, J.; Spalanzani, A. Agent-based modeling for predicting pedestrian trajectories around an autonomous vehicle. J. Artif. Intell. Res. 2022, 73, 1385–1433. [Google Scholar] [CrossRef]
  22. Wen, G.; Qin, J.; Fu, X.; Yu, W. DLSTM: Distributed Long Short-Term Memory Neural Networks for the Internet of Things. IEEE Trans. Netw. Sci. Eng. 2022, 9, 111–120. [Google Scholar] [CrossRef]
  23. Ackermann, J.; Gabler, V.; Osa, T.; Sugiyama, M. Reducing Overestimation Bias in Multi-Agent Domains Using Double Centralized Critics. arXiv 2019, arXiv:1910.01465. [Google Scholar]
Figure 1. Multi-UAVs perform marine data-monitoring missions in target areas.
Figure 1. Multi-UAVs perform marine data-monitoring missions in target areas.
Futureinternet 15 00369 g001
Figure 2. A framework diagram based on MATD3-Stacked_LSTM for the implementation of multi-UAV coverage planning.
Figure 2. A framework diagram based on MATD3-Stacked_LSTM for the implementation of multi-UAV coverage planning.
Futureinternet 15 00369 g002
Figure 3. Stacked LSTM structure.
Figure 3. Stacked LSTM structure.
Futureinternet 15 00369 g003
Figure 4. Autonomous decision making for UAVs based on the MATD3 algorithm.
Figure 4. Autonomous decision making for UAVs based on the MATD3 algorithm.
Futureinternet 15 00369 g004
Figure 5. Average reward trends with training episodes (smooth = 0.85).
Figure 5. Average reward trends with training episodes (smooth = 0.85).
Futureinternet 15 00369 g005
Figure 6. Average loss trends with training steps (smooth = 0.85).
Figure 6. Average loss trends with training steps (smooth = 0.85).
Futureinternet 15 00369 g006
Figure 7. The changing trend of the coverage of the three algorithms (four UAVs).
Figure 7. The changing trend of the coverage of the three algorithms (four UAVs).
Futureinternet 15 00369 g007
Figure 8. The coverage area of two UAVs under different algorithms. The circles are obstacles, and the rectangles are no-fly zones. Both are randomly generated. (a) Coverage area of two UAVs based on the MATD3-Stacked_LSTM algorithm. (b) Coverage area of two drones based on the MATD3 algorithm. (c) Coverage area of two drones based on the MADDPG algorithm.
Figure 8. The coverage area of two UAVs under different algorithms. The circles are obstacles, and the rectangles are no-fly zones. Both are randomly generated. (a) Coverage area of two UAVs based on the MATD3-Stacked_LSTM algorithm. (b) Coverage area of two drones based on the MATD3 algorithm. (c) Coverage area of two drones based on the MADDPG algorithm.
Futureinternet 15 00369 g008
Figure 9. The coverage area of four UAVs under different algorithms. The circles are obstacles, and the rectangles are no-fly zones. Both are randomly generated. (a) Coverage area of four UAVs based on the MATD3-Stacked_LSTM algorithm. (b) Coverage area of four drones based on the MATD3 algorithm. (c) Coverage area of four drones based on the MADDPG algorithm.
Figure 9. The coverage area of four UAVs under different algorithms. The circles are obstacles, and the rectangles are no-fly zones. Both are randomly generated. (a) Coverage area of four UAVs based on the MATD3-Stacked_LSTM algorithm. (b) Coverage area of four drones based on the MATD3 algorithm. (c) Coverage area of four drones based on the MADDPG algorithm.
Futureinternet 15 00369 g009
Figure 10. Repeated coverage of three algorithms with different numbers of UAVs.
Figure 10. Repeated coverage of three algorithms with different numbers of UAVs.
Futureinternet 15 00369 g010
Table 1. Summary of symbols used in this chapter.
Table 1. Summary of symbols used in this chapter.
3.1 System Model
m, M Index, set of UAVs
t , T Time slot
( x , y , z ) , ( X , Y , Z ) Drone coordinates
P UAV path collection
F Indicator for UAV trajectory avoiding no-fly zones
O Indicator for UAV trajectory avoiding obstacles
θ Flight steering angle
L x , L y Shadow area
B Specified airspace boundary
HFlying height
s r Spatial resolution
R Turning radius
D ^ , d ^ Distance between UAVs
V, A Speed and acceleration
EEnergy consumption of UAVs
CBattery capacity
r Battery consumption rate
TTotal time
A Set of non-obstacle and non-restricted areas
3.2 Problem Formulation
NThe number of grid cells
A The set of all non-obstacles (non-restricted)
Note: The definitions of notations above apply only to this section.
Table 2. Summary of notations used in the paper.
Table 2. Summary of notations used in the paper.
NotationDefinition
Actor learning rate1 × 10 6
Critic learning rate1 × 10 3
γ 0.95
τ 1 × 10 5
Memory capacity100,000
Batch size512
Policy noise0.6
Noise clip0.5
Policy delay2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, Q.; Liu, Q.; Wu, Z.; Zhang, J. Maximizing UAV Coverage in Maritime Wireless Networks: A Multiagent Reinforcement Learning Approach. Future Internet 2023, 15, 369. https://doi.org/10.3390/fi15110369

AMA Style

Wu Q, Liu Q, Wu Z, Zhang J. Maximizing UAV Coverage in Maritime Wireless Networks: A Multiagent Reinforcement Learning Approach. Future Internet. 2023; 15(11):369. https://doi.org/10.3390/fi15110369

Chicago/Turabian Style

Wu, Qianqian, Qiang Liu, Zefan Wu, and Jiye Zhang. 2023. "Maximizing UAV Coverage in Maritime Wireless Networks: A Multiagent Reinforcement Learning Approach" Future Internet 15, no. 11: 369. https://doi.org/10.3390/fi15110369

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop