Next Article in Journal
Hardware-Assisted Low-Latency NPU Virtualization Method for Multi-Sensor AI Systems
Previous Article in Journal
MIRA-CAP: Memory-Integrated Retrieval-Augmented Captioning for State-of-the-Art Image and Video Captioning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Agent Reinforcement Learning-Based Computation Offloading for Unmanned Aerial Vehicle Post-Disaster Rescue

School of Computer Science and Engineering, Northeastern University, Shenyang 110000, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(24), 8014; https://doi.org/10.3390/s24248014
Submission received: 30 October 2024 / Revised: 9 December 2024 / Accepted: 13 December 2024 / Published: 15 December 2024
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Natural disasters cause significant losses. Unmanned aerial vehicles (UAVs) are valuable in rescue missions but need to offload tasks to edge servers due to their limited computing power and battery life. This study proposes a task offloading decision algorithm called the multi-agent deep deterministic policy gradient with cooperation and experience replay (CER-MADDPG), which is based on multi-agent reinforcement learning for UAV computation offloading. CER-MADDPG emphasizes collaboration between UAVs and uses historical UAV experiences to classify and obtain optimal strategies. It enables collaboration among edge devices through the design of the ’critic’ network. Additionally, by defining good and bad experiences for UAVs, experiences are classified into two separate buffers, allowing UAVs to learn from them, seek benefits, avoid harm, and reduce system overhead. The performance of CER-MADDPG was verified through simulations in two aspects. First, the influence of key hyperparameters on performance was examined, and the optimal values were determined. Second, CER-MADDPG was compared with other baseline algorithms. The results show that compared with MADDPG and stochastic game-based resource allocation with prioritized experience replay, CER-MADDPG achieves the lowest system overhead and superior stability and scalability.

1. Introduction

Unmanned aerial vehicles (UAVs) are widely used in various fields, such as crime scene surveillance [1], habitat destruction assessment [2], crop monitoring [3], and vegetation mapping [4]. Recently, UAVs have been used in humanitarian rescue operations [5]. UAVs can be combined with artificial intelligence technology to improve the efficiency of natural disaster rescue efforts. UAVs are characterized by their low cost, high maneuverability, and flexible deployment, allowing them to easily enter disaster-stricken areas that are difficult to reach otherwise [6]. UAVs use onboard sensors and wireless communications to create digital maps of post-disaster areas and can be used as communication relays to effectively forward scheduling and commands of post-disaster rescue to ground rescue equipment [7]. Recent work has investigated the age of information (AoI)-minimal clustering, transmission, and trajectory co-design in UAV-assisted wireless-powered communication networks (WPCNs). In these networks, UAVs with limited battery capacity must manage energy-efficient clustering of islands and optimize their flight trajectories to ensure effective data transmission and long-term performance [8].
The use of UAVs in post-disaster rescue operations is subject to fundamental engineering challenges. On the one hand, UAVs have strict weight requirements, limiting the capacity of their onboard batteries [9]. On the other hand, some tasks have deadlines, and because of the limited computational capacity of UAVs, real-time applications, such as image processing and video streaming, often exceed the local data-processing capacity of UAVs. Therefore, UAVs must offload a large number of intensive computing tasks to nearby edge servers during post-disaster rescue to respond to rescue efforts in a timely manner and extend their lifecycle. These studies highlight the importance of designing intelligent offloading strategies to balance energy consumption and task performance in UAV-assisted systems. UAVs must consider the weighted sum of the task completion time and system energy consumption as the system overhead, and the goal of the joint offloading strategy is to minimize system utilization. Our motivation is to optimize the UAV’s computation offloading decisions to reduce system overhead. Minimizing system overhead can improve the overall task completion efficiency. System overhead includes both time and communication overheads. Therefore, the means to improve the overall task completion efficiency is to minimize time consumption and communication consumption, achieving the highest efficiency for completing computation offloading tasks. The use of multi-agent reinforcement learning methods to determine optimal offloading strategies for UAV computation offloading scenarios has emerged as a topic of significant research interest [10].
In this study, a mobile edge computing offloading system called the multi-agent deep deterministic policy gradient with cooperation and experience replay (CER-MADDPG) with multiple UAVs and servers is proposed, in which UAVs generate tasks with delay constraints and continuously send offloading requests to edge servers during their movement. We define the trade-off between the delay and energy consumption as the system overhead. To minimize the system overhead caused by the decisions made by the UAVs, a swarm of UAVs makes joint offloading decisions using CER-MADDPG. The main contributions of this study are summarized as follows:
  • The UAV swarm makes joint offloading decisions by modeling the offloading environment, delegating offloading decision-making to UAVs, and considering the random mobility of edge server clusters, time-varying nature of channels, and signal blockage of UAVs. To avoid wasting server computing resources while ensuring the successful execution of user tasks, the system is modeled as an optimization problem that minimizes the system overhead using the weighted average of the task execution time and server energy consumption as the system overhead.
  • The CER-MADDPG algorithm based on multi-agent reinforcement learning is proposed. The UAVs use this algorithm to obtain their next coordinates, select edge servers, and determine the task offloading ratios. The algorithm enhances the collaborative decisions between agents by considering both global and individual information. Additionally, the algorithm classifies historical experiences and learns according to different categories of experiences, allowing UAVs to fully utilize historical experiences in continuous decision-making, cooperation, and optimal joint decisions.
  • Experiments were conducted using the PyTorch platform to simulate the proposed algorithm and compare it with MADDPG and stochastic game-based resource allocation with prioritized experience replays (SGRA-PERs), thereby verifying that the proposed algorithm is superior to other algorithms in terms of optimality, stability, and scalability.
The remainder of the article is organized as follows: Section 2 reviews relevant related work. Section 3 describes the UAV edge server problem. Section 4 introduces an offloading decision model based on the CER-MADDPG algorithm. Section 5 presents an analysis of the performance of the CER-MADDPG algorithm using simulations, and Section 6 presents the conclusions.

2. Related Work

Computation offloading is a core issue in edge computing. Zhang et al. [11] argued that the main solutions to computation offloading are tasks to be offloaded and offloading locations. Current offloading decisions often involve comprehensive delay, energy consumption, or user benefits as offloading objectives to meet the needs of real-time and low energy consumption.
Because finding an optimal solution to the computation offloading problem is difficult, several studies have proposed algorithms based on searching for suboptimal solutions, such as metaheuristic-based algorithms. Among these, the genetic algorithm is most widely used, which has low computational cost and can obtain solutions very close to the optimal solution in some cases. Li et al. [12] used the genetic algorithm to solve a scenario with multiple mobile devices and one edge server, dividing the tasks of edge devices proportionally to minimize the overall task completion time. Al-habob et al. [13] modeled the computation offloading scenario as an optimization problem with binary scheduling decision variables and designed a genetic algorithm to solve the optimization problem.
Another approach for solving optimization problems is based on deep reinforcement learning algorithms. Hu et al. [14] modeled the offloading problem as a Markov decision process and designed an offloading strategy based on deep deterministic policy gradient to dynamically adjust the offloading ratio, maximizing the system overhead considering server energy consumption and task duration. Huang et al. [15] proposed an online offloading framework based on deep reinforcement learning, optimizing the task offloading decision of a single wireless device according to the time-varying wireless channel. Yang et al. [16] divided the offloading decision into discrete actions, jointly deciding on task offloading, wireless channel allocation, and image compression rate selection to achieve higher average image recognition accuracy and lower average processing latency.
In some scenarios, single-agent reinforcement learning methods are not applicable. For example, in large-scale disaster relief, UAVs have different sensors and capabilities. Multi-agent reinforcement learning allows multiple agents to make independent decisions, with each agent responsible for solving specific sub-problems, thus better adapting to the decomposition and decision-making of complex tasks. Nguyen et al. [17] used MADDPG for joint data offloading in multiple independent edge clouds. Peng et al. [18] used the MADDPG method to solve the decision-making problem of UAV computation offloading with the goal of maximum reduction in the number of tasks. Lu et al. [19] considered minimizing the task failure rate and improving system utility as objectives to solve the task offloading problem of edge devices, based on the MADDPG algorithm. Huang et al. [20] proposed the cost-aware collaborative task-execution model, which specifically considers long-term reward indicators such as the number of lost tasks and uses the MADDPG method to coordinate multiple energy harvesting mobile devices in an EH-D2D network to execute computing tasks. Kumar et al. [21] treated each vehicle as an intelligent agent with decision-making capabilities and classified tasks generated by vehicles into local execution queues or task offloading queues using a method known as the Lyapunov-based MADDPG, which can minimize energy consumption while maintaining queue stability. However, these studies have shortcomings considering the cooperation between edge devices, which can lead to suboptimal joint decisions made by a group of edge devices.
Insufficient experience can lead to a system that does not achieve an optimal solution. Generally, reinforcement learning algorithms require a series of iterations to achieve the optimal solution. Argerich et al. [22] proposed using external knowledge to guide agent decisions, which improves performance during training using expert knowledge from a database in the form of a programmable function. Chen et al. [23] stored the historical experiences of edge devices in a priority experience buffer, repeatedly learning from key historical experiences to improve the learning efficiency and obtain higher-quality solutions. Wu et al. [24] prioritized the historical experiences of each UAV in a tree-like manner, repeatedly learning from less frequent but important experiences to improve the system performance. However, these methods have shortcomings in the utilization of historical experiences of edge devices, which is also one of the reasons why the joint decisions made by a group of edge devices fall into local optima.
In recent years, with the widespread application of unmanned aerial vehicles (UAVs) in fields such as post-disaster rescue and environmental monitoring, multi-UAV and multi-MEC (mobile edge computing) collaborative systems have become a hot research topic. Multi-UAV systems are capable of providing flexible computational and communication resources, particularly in scenarios where traditional cloud and edge computing capabilities are limited, effectively compensating for the lack of computational resources. Asif et al. [25] proposed a Joint Data Aggregation and Computation Offloading (JDACO) scheme, utilizing multiple UAVs as MEC servers to optimize data aggregation and task offloading processes, minimizing energy consumption and latency, showcasing UAVs’ potential in providing computational support for IoT devices in post-disaster scenarios. Zhang et al. [26] introduced a device-edge-cloud collaborative computing model that optimizes system delay and energy consumption, addressing the limited computational power of UAVs. The study employs a SAC-based reinforcement learning algorithm (STS-UDCO) to improve system convergence speed and stability, offering an effective solution for multi-UAV computation offloading. Wang et al. [27] presented a multi-UAV-assisted MEC system, incorporating Reconfigurable Intelligent Surfaces (RIS) to enhance communication performance and reduce system delay. By optimizing the computation offloading strategy and UAV trajectory, and employing a multi-agent deep reinforcement learning algorithm (MATD3), the study significantly improves system fairness and delay performance. Although these studies provide valuable insights into multi-UAV and multi-MEC systems for post-disaster rescue and resource allocation, they share a common limitation. While collaboration mechanisms and task offloading optimization are addressed in each study, they do not fully consider the integration of collaboration with the effective utilization of historical experiences. Specifically, existing approaches typically focus on optimizing a single objective, neglecting the potential of using intelligent experience replay mechanisms and collaborative decision-making to enhance overall system performance.
Although several studies have attempted to use multi-agent reinforcement learning methods to solve the UAV computation offloading problem, we found the following:
1.
Most current works consider scenarios with a single server and multiple UAVs, whereas real scenarios often involve multiple servers and multiple UAVs.
2.
Existing research does not fully consider cooperative decision-making between UAVs. For example, when multiple UAVs observe the same environmental state in the same time slot, they make the same decisions. This homogenized decision-making is not necessarily a joint optimal decision, which may lead to joint offloading decisions made by a group of UAVs to fall into local optima.
3.
Existing studies have not sufficiently considered the utilization of historical UAV decision-making experiences. Current research focuses on prioritizing historical UAV experiences or introducing external knowledge bases. The UAV group cannot learn quickly from high-quality historical decision-making experiences and avoid poor experiences. This can also lead to joint offloading decisions made by a group of UAVs to fall into local optima.
As discussed in this section, several architectures and technologies exist to solve the computation offloading problem, each with its own advantages and disadvantages. Among these, technology based on multi-agent reinforcement learning is the most promising. Therefore, we propose a UAV edge-computing offloading method based on multi-agent reinforcement learning, aiming to jointly optimize the task completion time and server energy consumption. Existing studies do not fully consider cooperative decision-making between UAVs or the full utilization of historical UAV decision-making experiences, resulting in suboptimal system overhead. To address these shortcomings, we propose the CER-MADDPG algorithm, which uses a neural network design to achieve cooperative decision-making between UAVs and guides the joint decision-making of a UAV group toward an optimal solution through the classification of historical UAV decision-making experiences.

3. Problem Description

In this section, we introduce the UAV computation offloading system. The system is divided into the UAV and edge layers, and the movement of devices in both layers is modeled with the devices able to move in their respective layers. Information exchange can occur between UAVs and edge devices. The system comprises three sub-models of mobile edge computing: network communication, task computation, and task transfer.

3.1. MEC Environment Model

The system comprises two layers of devices, as shown in Figure 1, namely the UAV layer and edge layer. In this structure, the devices in the edge layer provide computing resources for the UAV group, and the UAVs can fly within the system architecture. Our work focuses on effective offloading decisions of UAVs. Let the set of UAVs be N = { 1 , 2 , , n } and the set of edge devices be M = { 1 , 2 , , m } . After each UAV randomly generates a task, it determines the distance and angle of flight and offloading ratio of the task and selects an edge device to send the offloading request in each time interval. The corresponding device performs the offloading computations. The main components of the MEC environment model include the UAV mobility, edge server mobility, and UAV task generation models, as described below.

3.1.1. UAV Mobility Model

Let the flight angle and flight distance of the UAV be θ and d i s t , respectively, and the current coordinates of the UAV be u a v x , u a v y . The position of the UAV in the next time slot is given by (1):
u a v x = u a v x + d i s t · cos ( θ ) u a v y = u a v y + d i s t · sin ( θ )
The energy consumption b a t t e r y of the UAV flight to the next destination is given by (2).
b a t t e r y = b a t t e r y r · f u a v 2 · x · t a s k S i z e · s
where f u a v represents the computation frequency of the UAV, x represents the computation offloading ratio, t a s k S i z e represents the size of the UAV task, s represents the number of CPU cycles required to process one bit, and r represents the impact factor of the chip structure on the CPU processing.

3.1.2. Edge Server Mobility Model

The current coordinates of the edge server are u e x , u e y . At each time step, it randomly moves one step. The possible coordinates of the edge server in the next time step are as follows:
u e x 1 , u e y u e x + 1 , u e y u e x , u e y 1 u e x , u e y + 1 u e x , u e y

3.1.3. UAV Task Generation Model

Each UAV generates task T n in the initial state, including the task generation time, expiration time, and task size. In each time slot, each UAV selects an angle and distance to move and an edge server for the partial offloading of tasks.

3.2. Communication Model

We assume that the wireless channel bandwidth for transmission from the UAV to the edge server is B Hz. Additionally, the transmission loss power is defined as p n o i s y , and the reference channel gain when the distance between UAV and edge server is 1 m is defined as α 0 . We denote the uplink transmission power as p u p l i n k . The distance between the UAV and selected edge server was calculated based on the positions of the UAV and edge server as dist ( u a v , u e ) .
The channel gain g ( u a v , u e ) and uplink transmission rate t r a n s R a t e are given by Formulas (3) and (4), respectively.
g ( u a v , u e ) = α 0 dist ( u a v , u e ) 2
t r a n s R a t e = B · log 2 1 + p u p l i n k · g ( u a v , u e ) p n o i s y
The communication delay t t r and communication energy consumption e t r are given by Formulas (5) and (6), respectively, where t a s k S i z e represents the task size of UAV n, x represents the task offloading ratio, t r a n s R a t e is the uplink transmission rate obtained from (4), and P 1 represents the power consumed for communication computation.
t t r = x · t a s k S i z e t r a n s R a t e
e t r = x · t a s k S i z e t r a n s R a t e · P 1

3.3. Computation Offloading Model

3.3.1. Local Model

The tasks selected for local computation are provided with computing resources by the UAV. The local computation delay t l o c a l and local energy consumption e l o c a l are given by Formulas (7) and (8), respectively.
t l o c a l = ( 1 x ) · t a s k S i z e f u a v · s
e l o c a l = ( 1 x ) · t a s k S i z e f u a v · s · P 2
where x is the offloading ratio, t a s k S i z e represents the task size of UAV n, f u a v is the computing frequency of the UAV, s is the number of CPU cycles required to process one bit of the task, and P 2 represents the power consumed for local computation.

3.3.2. Server Model

The tasks selected for server computation are provided with computing resources using an edge server. The server computation delay t e d g e and server energy consumption e e d g e are given by Formulas (9) and (10), respectively.
t e d g e = x · t a s k S i z e f u a v · s
e e d g e = x · t a s k S i z e f u a v · s · P 3
Therefore, the total time consumption t c o m is given by Formula (11), and the total energy consumption e c o m is given by Formula (12), where x is the offloading ratio, t a s k S i z e represents the task size of UAV n, f u e is the computing frequency of the edge server, s is the number of CPU cycles required to process one bit of the task, and P 3 represents the power consumed for the server computation.
t c o m = max t t r + t e d g e , t l o c a l
e c o m = e t r + e e d g e + e l o c a l
The objective of the optimization problem for the optimal offloading decision is to minimize the system cost through rational decision-making of the UAV fleet. The system cost comprises the total time consumption t c o m and total energy consumption e c o m . The specific optimization problem is shown in Formula (13).
Minimize D = i = 1 N [ λ 1 · max ( t t r i + t e d g e i , t l o c a l i )
+ λ 2 · max ( e t r i + e e d g e i , e l o c a l i ) ]
subject to max ( t t r i + t e d g e i , t l o c a l i ) < T max i
x ( i ) [ 0 , 1 ]
Here, λ 1 represents the hyperparameter related to the total time consumption, and λ 2 represents the hyperparameter related to the total energy consumption. Formula (13a) indicates that the tasks processed by the UAV should not exceed the deadline, and Formula (13b) indicates that the offloading ratio should be between 0 and 1.

4. Multi-Agent Deep Deterministic Policy Gradient Considering Collaboration and Experience Utilization

In this section, we will define the UAV computation offloading system designed in Section 3 using the partially observable Markov decision process (POMDP). In addition, we provide a detailed introduction to the improved MADDPG (CER-MADDPG) model and explain its training process.

4.1. Partially Observable Markov Decision Process

Each UAV can be considered an intelligent agent with decision-making capabilities, and we train an offloading strategy for each UAV. The POMDP is defined as ( S , A , R , P , O ) , where S represents the set of states, A the set of actions, R the reward function, P the state transition probability, and O the set of all joint observation values of the agents.
State: At time slot t, the state s t i of UAV i includes the characteristics of the UAV N n and the characteristics of the base station D m . N n includes the remaining battery level b a t t e r y , the coordinates of the UAV ( u a v x , u a v y ) , and the flight speed of the UAV s p e e d . D m includes the current CPU usage c p u , number of tasks currently executed t r , maximum computing capability of the processor m i p s , energy consumption per bit for the transmission e e , and bandwidth of the edge device e b . The specific content is shown in Formula (14).
s t i = N i , D j i = 1 , j = 1 N , M
The state of the entire system contains the states of all UAVs, denoted by S t , as shown in (15).
S t = s t i i = 1 N
Action: The decision a t i made by UAV i at time t includes the selection of the edge server m, next flight angle a n g l e , distance d i s t , and offloading ratio for computation x, as shown in Formula (16).
a t i = m , a n g l e i , d i s t i , x
The action set of each agent forms the action space A t of the entire system, as expressed in Formula (17).
A t = a t i i = 1 N
Reward: Referring to Formula (13), the reward r e w a r d considers both the completion time and energy consumption of the same batch of tasks. The reward is defined as shown in Formula (18).
r e w a r d = D
Transition Probability: In a continuous action space, state transition is directly caused by the actions of the UAVs. As shown in Formula (19), the deterministic policy μ is used in this paper because of its efficiency, which is tens of times higher than that of a random policy, significantly reducing the training time.
a t i = ( s t i )
Observation: The information that UAV i can perceive is represented by observation o t i of UAV i, as shown in Formula (20). UAV i can only observe information from its local environment and partial base stations.
o t i = N i , D j j = 1 k
The observation set O t is shown in Formula (21).
O t = { o t i } i = 1 N
Cumulative Return: The return r sum , as shown in Formula (22), is the accumulation of rewards over time, which is the sum of all rewards on the time axis. The goal of this study was to maximize the returns within each time period.
r s u m = i = 1 N γ N i r i

4.2. Multi-Agent Deep Deterministic Policy Gradient Considering Collaboration and Experience Utilization

4.2.1. General Description

Based on the model described in the previous section, this study proposes an improved MADDPG algorithm (CER-MADDPG) to solve the optimization problem described in (13). First, the CER-MADDPG algorithm considers cooperative decision making among multiple UAVs. Second, the CER-MADDPG algorithm focuses on guiding the UAVs’ decisions toward good experiences and away from bad ones. When the UAVs are in the same state, they may perform homogeneous actions, which reduces the overall efficiency of the system. Therefore, collaborative decision-making among UAVs is necessary. In addition, the traditional priority-experience buffer does not fully utilize these experiences. Therefore, this study considers categorizing experiences to make the decisions of the UAV group closer to good experiences and further from bad ones.
In this study, the proposed CER-MADDPG algorithm is based on the partial offloading model. Compared to the full offloading model, partial offloading offers greater flexibility and efficiency in task processing. While full offloading alleviates the UAV’s computational burden by offloading all tasks to the edge server, it fails to fully utilize the local computation resources of the UAV, potentially leading to unnecessary network transmission delays and bandwidth wastage. In contrast, partial offloading allows UAVs to flexibly allocate tasks between local computation and edge servers based on the specific requirements of the tasks, thereby optimizing the balance between computational capacity and energy consumption. By dynamically adjusting the ratio of local computation and offloading, partial offloading not only reduces network transmission load but also minimizes system overhead, improving task processing efficiency and real-time performance. Additionally, partial offloading can extend the UAV’s flight time to some extent, as UAVs can decide whether to keep part of the computation locally based on task complexity, thus reducing reliance on battery power.
CER-MADDPG comprises multiple collaborating UAVs, each comprising an evaluation network, target network, and replay buffer R B i . The evaluation network includes actor evaluation networks μ i θ μ and critic evaluation networks Q i θ Q . The target network includes actor target networks μ i θ μ and critic target networks Q i θ Q . Here, θ μ , θ Q , θ μ and θ Q represent the parameters of the networks. As shown in Figure 2, at time t, all UAVs observe the state S t and input it into the action network. By calculating the policy function, the action vector A t is obtained. Subsequently, the state is updated to S t + 1 , and a reward is generated. The observed state of the UAV, action vector of the UAV, reward feedback from the environment, and next state of the environment are stored in the UAV’s respective experience replay buffer R B i in the form of ( s , a , r , s ) tuples. A batch of experiences ( s , a , r , s ) is extracted from the experience replay buffer, and state s and action a are input into the critic network. The Q value is calculated through Q i θ Q to describe whether the action is appropriate in the current state.
Joint actions a determine the next state and reward. The goal of the system is to find the optimal policy π * ( s ) = arg max π Q n π ( s , a ) that can select the optimal action in the current state to maximize the expected total discounted reward in the future. We use Q i s , a θ Q and Q i s , a θ Q to represent the Q-values at the current action state and Q-values at the next action and state, defined as (23):
Q μ ( s , a ) = t = 0 γ k R t + 1 s , a

4.2.2. Critic Networks That Consider Both Global and Local Information

Here, we design the Q i θ Q network structure in a parallel form, as shown in Figure 3. To avoid the UAV group taking the same actions when facing the same environmental state, which may lead to suboptimal system utility, we integrated the individual information of the UAVs into the neural network. The network was designed using shared and individual layers. The shared part of the neural network is used to process the global information and obtain Q s h a r e . In addition, a separate neural network was designed for each agent to process the individual information, and the individual information of the agent was concatenated to obtain Q l o c a l . This approach helps balance the processing of global and individual information to some extent. Finally, Q s h a r e and Q l o c a l are concatenated to obtain the final value Q t o t a l , which guides the update of the actor network.
Here, γ is the discount factor, which represents the importance of future states for the current state. The parameter update of the evaluation network depends on the TD error, and θ Q is updated by minimizing the loss function. The loss function can be expressed as (24):
L ( θ Q ) = 1 sum i [ ( R i + γ Q ( S i + 1 , μ ( S i + 1 θ μ ) θ Q ) ) Q ( S i , a i θ Q ) ] 2
Here, s u m i represents the total number of UAVs, and the critic network must be continuously optimized during each iteration of the training process to minimize the loss function. To ensure computational efficiency, a batch gradient descent was used to optimize the loss function and update the weight parameters. The parameter update of the actor network depends on the output of the critic network, which trains the policy network by maximizing the Q-value estimate of the critic network for the action output by the actor network. The update of the actor network parameters according to the deterministic policy gradient ascent strategy can be expressed as (25):
θ μ J ( μ ) 1 sum i a Q S i , a i θ Q θ μ μ S i θ μ
During the training process, exponential smoothing is used to update the two target networks. The updating method for the target network parameters can be expressed as (26):
θ Q t a u θ Q + ( 1 t a u ) θ Q θ μ t a u θ μ + ( 1 t a u ) θ μ
Here, the parameter 0 < θ 1 is used to ensure that the target networks update slowly and steadily, improving the stability of learning.

4.2.3. Utilization of Categorized Experience

The UAVs’ experiences generated in this process are stored in their respective experience replay buffers (RB). To better utilize the UAVs’ experiences, Figure 4 describes the categorization of the UAV group’s experiences, guiding them toward joint actions that are beneficial and avoiding harmful ones.
I ( S t , A t ) represents the mutual information between the global state and the joint decision, which can describe the relevance between the global state and decisions. The larger I ( S t , A t ) is, the higher the correlation between the global state and joint decision, indicating that both high- and low-quality joint actions may have a high correlation with the environment. In this study, a bidirectional guidance module was designed to increase the correlation between high-quality joint actions and their corresponding global states while reducing the correlation between low-quality joint actions and their corresponding global states. The mutual information neural estimation (MINE) and contrastive log-ratio upper bound (CLUB) networks are used to estimate the correlation between two random variables. The MINE network [28] is used to estimate the lower bound of I ( S t , A t ) , and the CLUB network [29] is used to estimate the upper bound of I ( S t , A t ) . Let M I l o w e r be the value estimated by the MINE network, and M I u p p e r be the value estimated value provided by the CLUB network. The values from both neural networks were integrated into the reward to train the policy selection of the UAV network. Increasing the lower bound of the high-quality experience buffer and decreasing that of the low-quality experience buffer can promote collaboration among UAVs, thereby achieving the goal of reducing system utility. The reward was supplemented as shown in the following Formula (27).
r e w a r d = D α 1 · M I u p p e r + α 2 · M I l o w e r
The method proposed in this paper for classifying good and bad experiences is as follows. An experience is considered a high-quality experience and enters the high-quality experience buffer P B if its reward is greater than the worst experience in the buffer P B l o w and its reward exceeds the average reward r ¯ over a period of time. Otherwise, the experience enters the buffer N B . Regardless of whether the experience is good or bad, it enters the priority buffer R B for extraction and learning during UAV model training. The determination of a high-quality experience is shown in the following Formula (28).
r s u m > max r ¯ , P B l o w
The experiences of the UAV swarm are classified by the classifier, with high-quality experiences entering P B and poor-quality experiences entering N B . Both high- and poor-quality experiences enter the general experience buffer, R B . Experiences in the high-quality experience buffer P B were used to train the MINE network, whereas experiences in the poor-quality experience buffer N B were used to train the CLUB network. The overall reward for training the critic network comprises both the environmental reward and estimated mutual information value.

4.3. Training Process

The training process using CER-MADDPG to solve the UAV computation offloading problem proposed in this study is as follows:
The Algorithm 1 initialization is described in lines 1–3. First, the action networks, value networks, MINE, and CLUB were initialized for each UAV. Additionally, the experience buffers P B , R B , and N B , and finally, the positions of the UAVs and edge devices are initialized. Training starts in line 4, where each UAV observes its state s n from the environment. Each UAV obtains an action from the action network based on its current state. The UAVs then execute joint actions and receive the current reward r n and the next state s n . The ( s , a , r , s ) tuples of the UAVs are stored in their respective experience replay buffers, R B . The experiences of the UAV swarm are classified by the classifier to determine whether to store them in the high-quality experience buffer P B or the low-quality experience buffer N B . When there are sufficient experiences with R B , a small batch is sampled from the replay buffer. Based on this batch, the critic and actor networks are updated. When there are sufficient experiences in both P B and N B , experiences are extracted from both to update the MINE and CLUB networks.
Algorithm 1 The CER-MADDPG Training Procedure in the UAV Computation Offloading System
Require: Replay buffer R B , positive buffer P B , negative buffer N B , time budget T, exploration probability ε , discount factor γ , update step ξ
Ensure: The optimal policy π θ n *
1:
Initialize actor and critic networks for each UAV, MINE and CLUB.
2:
Initialize experience replay buffer R B , P B , and N B .
3:
Initialize the location of UAVs and edge servers.
4:
for  e p i s o d e = 1 to M do
5:
       Initialize the state s 0 N i , D j i = 1 , j = 1 N , M , t 0
6:
   for  t = 1 to T do
7:
            Select joint actions u t of UAVs based on policy π θ i ( o t i )
8:
            Execute joint actions u t
9:
            Receive the team reward r t and the new state o
10:
         Store trajectory v = v = { u t , o t , u t + 1 , r t } to RB.
11:
   end for
12:
   if  r sum > max P B low , r ¯  then
13:
         Add v to P B .
14:
   else
15:
         Add v to N B .
16:
   end if
17:
     Update the actor and critic networks.
18:
     Update MINE and CLUB with every k episodes.
19:
end for

5. Simulation, Results and Analysis

In this section, we describe extensive simulations conducted to validate the performance of the proposed CER-MADDPG scheme in UAV computation offloading and edge–server resource allocation problems.

5.1. Simulation Environment and Parameter Setting

We simulated an MEC environment using Python, where MEC servers were allocated alongside roads and some edge servers moved randomly in the communication area. In the simulation, all UAVs selected angles and distances to move within the communication range and edge servers for offloading requests based on the CER-MADDPG method. The simulation was performed using Python on a PyTorch platform. All simulation parameters used are shown in Table 1 and Table 2, which were used in our simulation experiments. Referring to the works [14,30,31,32,33,34], similar values are set in our simulations.
The UAV’s computation frequency, f u a v is set to 1.2 GHz. The task size, t a s k S i z e is randomly generated when adjusting parameters, with values ranging from 30 to 40 Mbits. The number of CPU cycles required to process one bit, denoted as s, is set to 1000 cycles/bit. The UAV’s initial battery level is set to 500,000 J. These values are based on typical UAV parameters.
In our model, the location of the edge servers changes dynamically, simulating the potential movement of servers over time in a post-disaster environment. We use a simple linear mobility model to simulate server movement, where the servers can randomly select paths within a predefined area. The server’s movement speed is set to 1 m/s, and the server’s computation frequency, f u e is set to 5 GHz. This model can be further improved in the future with more complex random walk models to adapt to more complicated environmental changes.

5.2. Results and Analysis

5.2.1. Selection of Hyperparameters

Setting the hyperparameters is crucial in the CER-MADDPG algorithm, because different hyperparameters can affect the optimization, convergence, and stability of the algorithm. Experiments were conducted to determine the optimal hyperparameters for the proposed algorithm.
In the CER-MADDPG algorithm, the learning rates of the actor and critic networks influence the training and updating of the neural networks. A high learning rate can lead to poor optimization and stability, whereas a low learning rate can result in poor optimization and slow convergence. In general, because the critic guides the update of the actor, the critic’s learning rate should be set slightly higher than that of the actor to ensure faster convergence. As shown in Figure 5, when the critic network’s learning rate is 1.2 × 10 5 and that of the actor network is 6 × 10 6 , the reward increases with the number of iterations. However, the optimal solution here is not as good as that with a critic network’s learning rate of 1.2 × 10 6 and the actor network’s learning rate is 6 × 10 7 . This is because a high learning rate causes both the critic and actor networks to update significantly, whereas the optimal solution requires smaller updates. When the learning rate for the critic network is 1.2 × 10 6 and that for the actor network is 6 × 10 7 , the system’s cost can achieve a converged and stable optimal solution. Although a locally optimal solution was obtained after 170 rounds of training, a globally optimal solution could still be achieved by increasing the number of iterations. When the learning rate of the critic network is 1.2 × 10 7 and that for the actor network’s learning rate is 6 × 10 8 , the system reward cannot reach the optimal solution, as the lower learning rates result in slower neural network updates, requiring more iterations to converge. Therefore, in this study, the optimal learning rate for the critic network is 1.2 × 10 6 , and that for the actor network is 6 × 10 7 . Additionally, it can be observed from the figure that after 100 rounds there was a significant change in the system cost. This is because at this point, the experience replay buffer is full, and the neural networks have sufficient useful information for training.
The experience replay buffer affected the training time, optimality, and convergence of the algorithm. As shown in Figure 6, a local optimal solution was obtained with an experience replay buffer size of 1000 and 320 training rounds; however, a global optimal solution could not be achieved. This is because a small experience replay buffer can affect the extraction of data features, leading to an inability to learn the optimal policy. When the experience replay buffer was increased to 10,000, a globally optimal solution was achieved after 480 rounds. However, with an experience replay buffer of 100,000, the optimal solution could not be obtained, because the buffer was too large and the data could not be updated, thus preventing the attainment of the best solution. Additionally, the algorithm with an experience replay buffer size of 10,000 trains was three times faster than that with an experience replay buffer size of 100,000. Therefore, we selected an optimal experience replay buffer size of 10,000.
The selection of the mutual information parameters α 1 for the lower bound of the high-quality experience buffer and α 2 for the upper bound of the poor-quality experience buffer is shown in Figure 7. The choice of mutual information parameters affects the optimality and stability of the algorithm. When α 1 is chosen as 0.5 and α 2 is chosen as 0.5, the system reward converges steadily to the optimal solution. When α 1 is chosen as 0.9 and α 2 is chosen as 0.5, the system converges to a local optimal solution after 220 rounds. When α 1 is chosen as 0.5 and α 2 is chosen as 0.9, the system reward first decreases and then increases, and there is no stable convergence to the optimal solution. This is because neither a dominant weight for the upper bound of the poor-quality experience buffer nor a dominant weight for the lower bound of the high-quality experience buffer leads to convergence to the optimal solution. We summarize the experimental results in Table 3.
Through comparative experiments, it was verified that the experience buffer size should be set to 10 4 , learning rate of the actor network should be set to 10 7 , learning rate of the critic network should be set to 10 6 , α 1 should be set to 0.5, and α 2 should be set to 0.5.

5.2.2. Performance Comparison and Analysis

We compared the CER-MADDPG method with MADDPG and SGRA-PERs to demonstrate the effectiveness of our approach. Brief introductions to these two methods are presented below.
  • MADDPG: A policy gradient algorithm in which each agent has actor and critic networks. The critic network can access the states and actions of other agents during training, whereas the actor network requires only its own information. Consequently, the critic network is trained centrally, whereas the actor network is executed in a distributed manner.
  • SGRA-PERs: Utilizes a prioritized experience replay mechanism for each UAV’s experience buffer, allowing important but infrequent experiences to be repeatedly utilized for learning, thereby enhancing system utility.
As shown in Figure 8, with an increase in iterations, the proposed CER-MADDPG algorithm achieved the lowest system cost and optimal stability after 480 rounds. In comparison, the MADDPG algorithm achieved stable system utility after 460 rounds, and the SGRA-PERs achieved stable system utility after 520 rounds of training. However, the solutions obtained by these algorithms are locally optimal and do not achieve the best results. This is because the proposed CER-MADDPG considers the cooperation between agents more deeply and utilizes the classification of good and bad decision-making experiences of UAVs. The bidirectional guidance of joint decision-making by UAVs tends to exhibit good behavior and avoid bad behavior, thus achieving the best system utility. SGRA-PERs prioritizes historical experiences for each UAV, which leads to a lower system utility compared with that of MADDPG. It can guide UAVs to learn better strategies to some extent but still has limitations. Therefore, compared with other algorithms, the CER-MADDPG proposed in this study has advantages in terms of system cost and can obtain solutions with lower system costs.
To further validate the performance of the CER-MADDPG algorithm, Figure 9 illustrates the relationship between the initial task size of the UAVs and time required to complete the tasks. As the initial task size of the UAVs increases, the time required to complete the tasks also increases. Regardless of the initial task size, the CER-MADDPG required the shortest time to complete the tasks compared to those of the MADDPG and SGRA-PERs. Additionally, as the initial task size of the UAVs increased, CER-MADDPG exhibited the smallest increase in completion time, indicating higher stability than that of the MADDPG algorithm. This is because the CER-MADDPG algorithm proposed in this study fully utilizes good and bad experiences, enabling the UAV group to make more reasonable joint decisions and demonstrating advantages in stability and effectiveness.
We compared the system energy consumption of the four algorithms for different numbers of UAVs. We set the number of UAVs N to 2, 3, 5, 10, and 20. As shown in Figure 10, with an increase in the number of UAVs, the system costs of the different algorithms increase to some extent. The CER-MADDPG algorithm outperformed the other two algorithms in controlling system utility. This is because, as the number of UAVs increases, the total task processing time and energy consumption of the UAV group increase, resulting in an overall increase in system costs. The growth rate of CER-MADDPG was approximately constant, whereas those of MADDPG and SGRA-PERs were higher than that of CER-MADDPG. This is because CER-MADDPG utilizes an experience classification learning mechanism and considers both global and individual critic networks, enabling a more comprehensive consideration of the collaborative nature of joint decisions, thereby improving the system utility. In summary, the CER-MADDPG algorithm has significantly lower system costs than those of the other offloading methods, and this advantage increases with the number of UAVs. We summarize the experimental results in Table 4.
As shown in Table 4, irrespective of the number of UAVs, the system overhead of CER-MADDPG is smaller than those of SGRA-PERs and MADDPG. Additionally, from the changes in the system overhead, it can be observed that as the number of UAVs increases, the system overhead of the CER-MADDPG becomes more stable than those of the SGRA-PERs and MADDPG. Irrespective of the initial task of the UAVs (30 Mbits or 1000 Mbits), CER-MADDPG has the shortest average completion time for UAVs compared to those of the SGRA-PERs and MADDPG.

6. Conclusions

To address the limitations of UAV battery life and computing capabilities, this study considers a scenario with multiple UAVs and servers and proposes a UAV-assisted post-disaster rescue computation offloading scheme based on multi-agent reinforcement learning. The algorithm proposed herein delegates decision-making to UAVs, and is designed to consider cooperative decision-making among multiple UAVs. Additionally, the algorithm focuses on guiding UAVs to make decisions based on good experiences and avoid bad experiences, enabling them to make optimal joint offloading decisions and ultimately improving system performance. The simulation results demonstrate that the proposed CER-MADDPG algorithm achieves better effectiveness and stability in joint decision-making compared with SGRA-PERs and MADDPG, effectively reducing the system overhead. The CER-MADDPG can generate optimal strategies based on time-varying channel conditions and edge server states.
Fault tolerance in UAV edge computing systems [35] is a new field that must be explored in the future. In this study, the UAVs, channels, and edge devices were assumed to be honest. In future research, we will consider security issues, improve the existing models, and design a model that is more suitable for practical applications.

Author Contributions

L.W. conceived the ideas; L.W. and H.J. designed the methodology; H.J. implemented the software and conducted the validation; L.W. and H.J. performed the formal analysis; H.J. prepared the original draft; L.W. reviewed and edited the manuscript; L.W. acquired the funding. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Liaoning of funder grant number 2022-MS-115.

Institutional Review Board Statement

Not applicable for studies not involving humans or animals.

Informed Consent Statement

Not applicable for studies not involving humans.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bucknell, A.; Bassindale, T. An investigation into the effect of surveillance drones on textile evidence at crime scenes. Sci. Justice 2017, 57, 373–375. [Google Scholar] [CrossRef] [PubMed]
  2. Barnas, A.F.; Darby, B.J.; Vandeberg, G.S.; Rockwell, R.F.; Ellis-Felege, S.N. A comparison of drone imagery and ground-based methods for estimating the extent of habitat destruction by lesser snow geese (Anser caerulescens caerulescens) in La Pérouse Bay. PLoS ONE 2019, 14, e0217049. [Google Scholar] [CrossRef] [PubMed]
  3. Bendig, J.; Bolten, A.; Bennertz, S.; Broscheit, J.; Eichfuss, S.; Bareth, G. Estimating biomass of barley using crop surface models (CSMs) derived from UAV-based RGB imaging. Remote Sens. 2014, 6, 10395–10412. [Google Scholar] [CrossRef]
  4. Díaz-Varela, R.A.; De la Rosa, R.; León, L.; Zarco-Tejada, P.J. High-resolution airborne UAV imagery to assess olive tree crown parameters using 3D photo reconstruction: Application in breeding trials. Remote Sens. 2015, 7, 4213–4232. [Google Scholar] [CrossRef]
  5. Daud, S.M.S.M.; Yusof, M.Y.P.M.; Heo, C.C.; Khoo, L.S.; Singh, M.K.C.; Mahmood, M.S.; Nawawi, H. Applications of drone in disaster management: A scoping review. Sci. Justice 2022, 62, 30–42. [Google Scholar] [CrossRef]
  6. Yin, S.; Zhao, Y.; Li, L.; Yu, R.F. UAV-assisted cooperative communications with time-sharing information and power transfer. IEEE Trans. Veh. Technol. 2019, 69, 1554–1567. [Google Scholar] [CrossRef]
  7. Su, Z.; Wang, Y.; Xu, Q.; Zhang, N. LVBS: Lightweight vehicular blockchain for secure data sharing in disaster rescue. IEEE Trans. Dependable Secure Comput. 2020, 19, 19–32. [Google Scholar] [CrossRef]
  8. Liu, X.; Liu, H.; Zheng, K.; Liu, J.; Taleb, T.; Shiratori, N. AoI-minimal clustering, transmission and trajectory co-design for UAV-assisted WPCNs. IEEE Trans. Veh. Technol. 2024, 1–16. [Google Scholar] [CrossRef]
  9. Zeng, Y.; Zhang, R. Energy-efficient UAV communication with trajectory optimization. IEEE Trans. Wireless Commun. 2017, 16, 3747–3760. [Google Scholar] [CrossRef]
  10. Yadav, P.; Mishra, A.; Kim, S. A Comprehensive Survey on Multi-Agent Reinforcement Learning for Connected and Automated Vehicles. Sensors 2023, 23, 4710. [Google Scholar] [CrossRef]
  11. Zhang, J.; Letaief, K.B. Mobile edge intelligence and computing for the internet of vehicles. Proc. IEEE 2019, 108, 246–261. [Google Scholar] [CrossRef]
  12. Li, Z.; Zhu, Q. Genetic algorithm-based optimization of offloading and resource allocation in mobile-edge computing. Information 2020, 11, 83. [Google Scholar] [CrossRef]
  13. Al-Habob, A.A.; Dobre, O.A.; Armada, A.G. Sequential task scheduling for mobile edge computing using genetic algorithm. In Proceedings of the IEEE Globecom Workshops (GC Wkshps), Waikoloa, HI, USA, 9–13 December 2019; pp. 1–6. [Google Scholar]
  14. Hu, X.; Huang, Y. Deep reinforcement learning based offloading decision algorithm for vehicular edge computing. PeerJ Comput. Sci. 2022, 8, e1126. [Google Scholar] [CrossRef] [PubMed]
  15. Huang, L.; Bi, S.; Zhang, Y.J.A. Deep reinforcement learning for online computation offloading in wireless powered mobile-edge computing networks. IEEE Trans. Mob. Comput. 2019, 19, 2581–2593. [Google Scholar] [CrossRef]
  16. Yan, K.; Shan, H.; Sun, T.; Hu, H.; Wu, Y.; Yu, L. Reinforcement learning-based mobile edge computing and transmission scheduling for video surveillance. IEEE Trans. Emerg. Top. Comput. 2021, 10, 1142–1156. [Google Scholar]
  17. Nguyen, D.; Ding, M.; Pathirana, P.; Seneviratne, A.; Li, J.; Vincent Poor, H. Cooperative task offloading and block mining in blockchain-based edge computing with multi-agent deep reinforcement learning. IEEE Trans. Mob. Comput. 2021, 22, 2021–2037. [Google Scholar] [CrossRef]
  18. Peng, H.; Shen, X. Multi-agent reinforcement learning based resource management in MEC-and UAV-assisted vehicular networks. IEEE J. Sel. Areas Commun. 2020, 39, 131–141. [Google Scholar] [CrossRef]
  19. Lu, K.; Li, R.D.; Li, M.C.; Xu, G.R. MADDPG-based joint optimization of task partitioning and computation resource allocation in mobile edge computing. Neural Comput. Appl. 2023, 35, 16559–16576. [Google Scholar] [CrossRef]
  20. Huang, B.; Liu, X.; Wang, S.; Pan, L.; Chang, V. Multi-agent reinforcement learning for cost-aware collaborative task execution in energy-harvesting D2D networks. Comput. Netw. 2021, 195, 108176. [Google Scholar] [CrossRef]
  21. Kumar, A.S.; Zhao, L.; Fernando, X. Task Offloading and Resource Allocation in Vehicular Networks: A Lyapunov-based Deep Reinforcement Learning Approach. IEEE Trans. Veh. Technol. 2023, 72, 13360–13373. [Google Scholar] [CrossRef]
  22. Argerich, M.F.; Fürst, J.; Cheng, B. Tutor4RL: Guiding Reinforcement Learning with External Knowledge. In Proceedings of the AAAI Spring Symposium on Combining Machine Learning and Knowledge Engineering, Palo Alto, CA, USA, 23–25 March 2020. [Google Scholar]
  23. Chen, N.; Zhang, S.; Qian, Z.; Wu, J.; Lu, S. When learning joins edge: Real-time proportional computation offloading via deep reinforcement learning. In Proceedings of the IEEE 25th International Conference on Parallel and Distributed Systems (ICPADS), Tianjin, China, 4–6 December 2019; pp. 414–421. [Google Scholar]
  24. Wu, G.; Wang, H.; Zhang, H.; Zhao, Y.; Yu, S.; Shen, S. Computation Offloading Method Using Stochastic Games for Software Defined Network-based Multi-Agent Mobile Edge Computing. IEEE Internet Things J. 2023, 10, 17620–17634. [Google Scholar] [CrossRef]
  25. Raivi, A.M.; Moh, S. JDACO: Joint Data Aggregation and Computation Offloading in UAV-Enabled Internet of Things for Post-Disaster Scenarios. IEEE Internet Things J. 2024, 11, 16529–16544. [Google Scholar] [CrossRef]
  26. Zhang, L.; Tan, R.; Zhang, Y.; Peng, J.; Liu, J.; Li, K. UAV-assisted Dependency-aware Computation Offloading in Device–Edge–Cloud Collaborative Computing Based on Improved Actor–Critic DRL. J. Syst. Archit. 2024, 154, 103215. [Google Scholar] [CrossRef]
  27. Wang, S.; Song, X.; Song, T.; Yang, Y. Fairness-aware Computation Offloading with Trajectory Optimization and Phase-shift Design in RIS-assisted Multi-UAV MEC Network. IEEE Internet Things J. 2024, 11, 20547–20561. [Google Scholar] [CrossRef]
  28. Belghazi, M.I.; Baratin, A.; Rajeshwar, S.; Ozair, S.; Bengio, Y.; Courville, A.; Hjelm, D. Mutual information neural estimation. In Proceedings of the International Conference on Machine Learning (ICML), Stockholm, Sweden, 10–15 July 2018; pp. 531–540. [Google Scholar]
  29. Cheng, P.; Hao, W.; Dai, S.; Liu, J.; Gan, Z.; Carin, L. Club: A contrastive log-ratio upper bound of mutual information. In Proceedings of the International Conference on Machine Learning (ICML), Vienna, Austria, 12–18 July 2020; pp. 1779–1788. [Google Scholar]
  30. Li, J.; Gao, H.; Lv, T.; Lu, Y. Deep reinforcement learning based computation offloading and resource allocation for MEC. In Proceedings of the 2018 IEEE Wireless Communications and Networking Conference (WCNC), Barcelona, Spain, 15–18 April 2018; pp. 1–6. [Google Scholar]
  31. Tran, T.X.; Pompili, D. Joint task offloading and resource allocation for multi-server mobile-edge computing networks. IEEE Trans. Veh. Technol. 2018, 68, 856–868. [Google Scholar] [CrossRef]
  32. Wang, J.; Feng, D.; Zhang, S.; Tang, J. Computation offloading for mobile edge computing enabled vehicular networks. IEEE Access 2019, 7, 62624–62632. [Google Scholar] [CrossRef]
  33. Wang, K.; Wang, X.; Liu, X. A high reliable computing offloading strategy using deep reinforcement learning for IOVs in edge computing. J. Grid Comput. 2021, 19, 15. [Google Scholar] [CrossRef]
  34. Cai, T.; Yang, Z.; Chen, Y.; Zhang, Y.; Wang, Z. Cooperative Data Sensing and Computation Offloading in UAV-assisted Crowdsensing with Multi-agent Deep Reinforcement Learning. IEEE Trans. Netw. Sci. Eng. 2021, 9, 3197–3211. [Google Scholar] [CrossRef]
  35. Bosilca, G.; Bouteiller, A.; Guermouche, A.; Herault, T.; Robert, Y.; Sens, P. Failure detection and propagation in HPC systems. In Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis (SC’16), Salt Lake City, UT, USA, 13–18 November 2016; pp. 312–322. [Google Scholar]
Figure 1. Edge computing architecture.
Figure 1. Edge computing architecture.
Sensors 24 08014 g001
Figure 2. CER-MADDPG algorithm structure.
Figure 2. CER-MADDPG algorithm structure.
Sensors 24 08014 g002
Figure 3. Improved critic network structure.
Figure 3. Improved critic network structure.
Sensors 24 08014 g003
Figure 4. Good and bad behavior guidance model.
Figure 4. Good and bad behavior guidance model.
Sensors 24 08014 g004
Figure 5. Selection of learning rates for the critic and actor networks.
Figure 5. Selection of learning rates for the critic and actor networks.
Sensors 24 08014 g005
Figure 6. Selection of replay buffer size.
Figure 6. Selection of replay buffer size.
Sensors 24 08014 g006
Figure 7. Selection of α 1 and α 2 .
Figure 7. Selection of α 1 and α 2 .
Sensors 24 08014 g007
Figure 8. Comparison of system overhead of different algorithms.
Figure 8. Comparison of system overhead of different algorithms.
Sensors 24 08014 g008
Figure 9. Comparison of task completion time for different UAV mission sizes.
Figure 9. Comparison of task completion time for different UAV mission sizes.
Sensors 24 08014 g009
Figure 10. System consumption comparison as the number of UAVs increases.
Figure 10. System consumption comparison as the number of UAVs increases.
Sensors 24 08014 g010
Table 1. Related parameters of environment.
Table 1. Related parameters of environment.
ParametersValues
groundHeight100 m
groundLength100 m
groundWidth100 m
f u e 5 GHz
f u a v 1.2 GHz
s1000 cycles/bit
v u e 1 m/s
battery500,000
t a s k S i z e 30–40 Mbits
Table 2. Related parameters of CER-MADDPG.
Table 2. Related parameters of CER-MADDPG.
ParametersValues
Critical neural network structureGRU, MLP
GRU network layers1
MLP network layersThe number of UAVs
Batch size128
Positive buffer size 10 4
Negative buffer size 10 4
Soft update factor0.1
Epsilon0.1
λ 1 1
λ 2 1
Table 3. Hyperparameter settings.
Table 3. Hyperparameter settings.
ParametersValues
Replay buffer size 10 4
Actor network learning rate 6 × 10 7
Critic network learning rate 1.2 × 10 6
α 1 0.5
α 2 0.5
Table 4. Simulation results.
Table 4. Simulation results.
NameCER-MADDPGSGRA-PERsMADDPG
System overhead (2 UAVs)232257282
System overhead (20 UAVs)255428473250
Average task completion time of UAV (task size 30 Mbits)333436
Average task completion time of UAV (task size 1000 Mbits)354386398
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, L.; Jiao, H. Multi-Agent Reinforcement Learning-Based Computation Offloading for Unmanned Aerial Vehicle Post-Disaster Rescue. Sensors 2024, 24, 8014. https://doi.org/10.3390/s24248014

AMA Style

Wang L, Jiao H. Multi-Agent Reinforcement Learning-Based Computation Offloading for Unmanned Aerial Vehicle Post-Disaster Rescue. Sensors. 2024; 24(24):8014. https://doi.org/10.3390/s24248014

Chicago/Turabian Style

Wang, Lixing, and Huirong Jiao. 2024. "Multi-Agent Reinforcement Learning-Based Computation Offloading for Unmanned Aerial Vehicle Post-Disaster Rescue" Sensors 24, no. 24: 8014. https://doi.org/10.3390/s24248014

APA Style

Wang, L., & Jiao, H. (2024). Multi-Agent Reinforcement Learning-Based Computation Offloading for Unmanned Aerial Vehicle Post-Disaster Rescue. Sensors, 24(24), 8014. https://doi.org/10.3390/s24248014

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop