Next Article in Journal
Time-Frequency Feature Extraction and Analysis of Inland Waterway Buoy Motion Based on Massive Monitoring Data
Previous Article in Journal
SOAR-RL: Safe and Open-Space Aware Reinforcement Learning for Mobile Robot Navigation in Narrow Spaces
Previous Article in Special Issue
Minimization of Resource Consumption with URLLC Constraints for Relay-Assisted IIoT
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cooperative Schemes for Joint Latency and Energy Consumption Minimization in UAV-MEC Networks †

by
Ming Cheng
1,*,
Saifei He
1,
Yijin Pan
2,
Min Lin
1 and
Wei-Ping Zhu
3
1
School of Communications and Information Engineering, Nanjing University of Posts and Telecommunications, Nanjing 210003, China
2
National Mobile Communications Research Laboratory, Southeast University, Nanjing 210096, China
3
Department of Electrical and Computer Engineering, Concordia University, Montreal, QC H3G 1M8, Canada
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in He, S.; Cheng, M.; Pan, Y.; Lin, M.; Zhu, W.-P. Distributed access and offloading scheme for multiple UAVs assisted MEC network. In Proceedings of the IEEE 98th Vehicular Technology Conference (VTC2023-Fall), Hong Kong, China, 10–13 October 2023.
Sensors 2025, 25(17), 5234; https://doi.org/10.3390/s25175234
Submission received: 8 July 2025 / Revised: 16 August 2025 / Accepted: 18 August 2025 / Published: 22 August 2025

Abstract

The Internet of Things (IoT) has promoted emerging applications that require massive device collaboration, heavy computation, and stringent latency. Unmanned aerial vehicle (UAV)-assisted mobile edge computing (MEC) systems can provide flexible services for user devices (UDs) with wide coverage. The optimization of both latency and energy consumption remains a critical yet challenging task due to the inherent trade-off between them. Joint association, offloading, and computing resource allocation are essential to achieving satisfying system performance. However, these processes are difficult due to the highly dynamic environment and the exponentially increasing complexity of large-scale networks. To address these challenges, we introduce a carefully designed cost function to balance the latency and the energy consumption, formulate the joint problem into a partially observable Markov decision process, and propose two multi-agent deep-reinforcement-learning-based schemes to tackle the long-term problem. Specifically, the multi-agent proximal policy optimization (MAPPO)-based scheme uses centralized learning and decentralized execution, while the closed-form enhanced multi-armed bandit (CF-MAB)-based scheme decouples association from offloading and computing resource allocation. In both schemes, UDs act as independent agents that learn from environmental interactions and historic decisions, make decision to maximize its individual reward function, and achieve implicit collaboration through the reward mechanism. The numerical results validate the effectiveness and show the superiority of our proposed schemes. The MAPPO-based scheme enables collaborative agent decisions for high performance in complex dynamic environments, while the CF-MAB-based scheme supports independent rapid response decisions.

1. Introduction

Wireless networks have evolved over time into the Internet of Things (IoT), which has promoted emerging applications including autonomous driving, smart cities, virtual reality, and smart healthcare [1,2]. These applications generally rely on the cooperation of a large number of mobile devices, bring exponentially increasing computation load, and require stringent latency standards. However, user devices (UDs) are equipped with limited computing capability and battery capacity, meaning that they struggle to achieve satisfactory performance in these emerging scenarios. Mobile edge computing (MEC) brings high-performance servers to the network edge and provides efficient computing services for UDs [3,4]. MEC is thus a promising technology to address the latency-sensitive and localized processing requirements of future networks.
The terrestrial network infrastructure has a limited coverage, with some dead zones, meaning that it cannot provide sufficient computing and communications services, especially in remote, rural, or disaster-prone areas. Non-terrestrial networks are considered as enablers to provide a reliable and cost-effective solution for continuous and ubiquitous wireless coverage [5,6]. Unmanned aerial vehicles (UAVs) are widely used in current communication networks due to their flexible movement, easy deployment, and line-of-sight (LoS) connections [7,8]. It is effective and economical to deploy swarms of UAVs to form scalable networks. UAVs act as flying edge nodes that can provide computing, caching, and analytic services for UDs on demand. Consequently, UAV-assisted MEC has become an ideal solution to address scenarios where terrestrial infrastructure is unavailable, insufficient, or impractical [9,10].

1.1. Related Works

UAV-assisted MEC mainly aims to reduce energy consumption and latency. Effective and efficient resource allocation is key to achieving high performance. Most studies in this area concern energy consumption [11,12,13,14,15,16], latency [17,18,19,20,21,22], or both [23,24,25,26,27,28,29]. Latency and the energy consumption are inherently conflicting, meaning that cost functions are generally introduced as a trade-off. The cost can be defined as the weighted sum of latency and energy consumption [24,27], the minimum value of biased metrics [28], or the specific objective [29]. High system performance can be achieved through the joint optimization of task offloading, communication and computing resource allocation, UAVs’ trajectories, and device association.
Convex optimization methods, such as block coordinate descent (BCD) and successive convex approximation (SCA), are used to solve these non-trivial joint problems. To reduce the energy consumption of a UAV-assisted MEC system, Wang et al. [15] divided the joint optimization of device association, UAVs’ trajectories, task offloading, and resource allocation into three sub-problems. The non-convex trajectory optimization problem is transformed into a convex one through SCA. Then, the BCD method is used to optimize each set of control variables iteratively. To minimize the average response delay in a UAV-assisted MEC system, Zhang et al. [22] also divided the joint optimization problem into three sub-problems of UAV deployment, user access, and task offloading. Then, SCA is used to transform these sub-problems into convex problems and BCD is used to solve them to obtain a near-optimal solution. To minimize the delay, reduce the energy consumption, and maximize the offloaded tasks simultaneously, Sun et al. [28] formulated a multi-objective joint problem, divided it into three sub-problems, and then solved them by using the BCD method. The above studies usually assume that the task arrival rate and channel state are known or can be predicted from historical data. These optimization schemes only aim to obtain short-term optimal solutions based on deterministic conditions.
It is difficult to predict the task arrival and the channel state accurately in many practical scenarios. Buffers are generally used to store the stochastic tasks. Then, the system stability, the queue backlog, and the task completion rate become important and the UAV-assisted MEC system should be investigated from a long-term perspective. The Lyapunov method is one of the most effective solutions to ensure the long-term stability and to maximize the utility [30,31,32,33,34]. Specifically, the Lyapunov method introduces a drift-plus-penalty function that balances the utility and the queue backlog. Then, the original long-term stochastic problem is converted into an online sequence of deterministic problems in each timeslot. These short-term problems minimize the upper bound of the drift-plus-penalty function instead and can be solved by using convex optimization. Zhang et al. [30] minimized the average long-term energy consumption in a UAV-assisted MEC system by jointly optimizing the task offloading, the resource allocation, and the trajectory. Liu et al. [32] investigated the long-term problem in multiple-UAV-assisted MEC systems and took the timeslot scheduling into consideration. Zhao et al. [34] further considered the task caching and migration in multiple-UAV-assisted MEC systems and maximized the long-term average service quantity.
In conventional convex optimization-based and Lyapunov-based methods, the original highly coupled problem is transformed into several approximate convex problems and the local optimal solution is obtained by means of an iterative process and alternative optimization. These methods require well-known information, heavy interaction, and exponentially increasing complexity, meaning that they are unsuitable in practical scenarios with high dynamics or a large scale. Fortunately, deep reinforcement learning (DRL)-based schemes have shown significant advantages in solving these problems with high-dimensional dynamic state space, continuous or discrete action space, and non-linear relationships. DRL-based schemes formulate the joint long-term optimization problem in dynamic environments as a Markov decision process (MDP) [35,36,37,38,39] or partially observable MDP (POMDP) [40,41,42,43,44,45]. Then, one or multiple agents learn from environmental and historic information, update the strategies, and make decisions.
DRL has been applied in UAV-assisted MEC networks [10]. Specifically, to minimize the latency in UAV-assisted MEC systems with caching, Zhang et al. [35] used the actor–critic (AC) architecture to optimize the UAVs’ motion, communication and computing resource allocation, and caching decisions simultaneously. To balance the energy consumption and the latency in multi-UAV-assisted vehicular edge computing systems, Li et al. [36] proposed a double-depth Q-network-based algorithm to optimize the vehicle selection and the resource allocation jointly. Seid et al. [37] proposed a deep deterministic policy gradient (DDPG)-based algorithm to optimize the association, offloading, and computing resource jointly. He et al. [39] further proposed a multi-agent DDPG (MADDPG)-based algorithm to optimize the trajectories of UAVs while power and offloading are determined by conventional optimization. Each agent can only obtain its local observation instead of the global state in most practical scenarios. Yan et al. [40], Liu et al. [41], and Wang et al. [42] formulated the long-term optimization to a POMDP and applied the MADDPG scheme to decide the trajectories of UAVs. Though the above schemes excel in specific scenarios, they show inferior performance regarding adaptability and stability. DQN is limited to discrete actions apace and DDPG is designed for continuous actions space. AC is adaptable to both action spaces but suffers from instable training and poor sample efficiency. Proximal policy optimization (PPO) [46,47,48], which is particularly well-suited for high-dimensional actions space with both continuous and discrete actions, has been successfully applied in UAV-assisted MEC networks [38,43,44,45]. Specifically, Wang et al. [38] proposed a PPO-based algorithm to optimize the continuous movement of UAVs and offloading proportions jointly. Kang et al. [43] further applied the PPO into scenarios with partial observation and proposed a multi-agent PPO (MAPPO)-based scheme to optimize offloading proportion, power control, and computing frequency allocation simultaneously.
The related works are summarized in Table 1.

1.2. Motivations and Contributions

UAV-assisted MEC can significantly improve the quality of experience in the IoT. Multi-agent DRL (MADRL)-based methods have shown great potential to improve the efficiency and effectiveness of joint resource allocation. Though MADRL-based methods are generally superior to convex optimization-based and reinforcement learning (RL)-based methods in the performance of their target metrics, their detailed superiority should be further investigated.
This paper investigates a multi-UAV-assisted MEC system in which each UAV is equipped with an MEC server to offer computing services and each UD decides its serving UAV, task offloading proportion, and required computing resources. The network is highly dynamic due to the movement of UAVs and time-varying channel state. We aim to minimize the latency and energy consumption by optimizing the association, task offloading, and computing resource allocation simultaneously. Our previous work [49] proposed an MAB-based distributed scheme to tackle the joint long-term problem. This work significantly extends our previous research with the following contributions:
  • We introduce cost functions to balance the latency and energy consumption for both the whole system and individual UDs. Then, we formulate a long-term cost minimization problem that has discrete association constraints and continuous offloading and computing frequency constraints. We further formulate this long-term problem into a POMDP and propose a cooperative multi-agent DRL framework. All UDs are agents and each of them makes its own decision based on its local observation and individual reward function.
  • We propose a MAPPO-based scheme that adopts centralized training and decentralized execution to tackle the POMDP. In the training stage, the global observations and the system reward function are used to train the actor and the critic networks for all UDs. In the execution stage, each UD agent uses its local observation and individual reward function for decision-making and network updates. Simulation results validate the MAPPO-based scheme and show its superiority in reducing system cost.
  • We decouple the association from the task offloading and the computing resource allocation and propose a lightweight scheme based on closed-form enhanced multi-armed bandit (CF-MAB). In the CF-MAB-based scheme, each UD agent selects its association to maximize its long-term achievable rate and then the optimal offloading and computing resource allocation can be obtained in closed form given the association. Simulation results validate the CF-MAB-based scheme and show its superiority regarding its complexity and task completion rate.
The remainder of this paper is organized as follows. Section 2 describes the system model and formulates the long-term problem. Section 3 presents the MADRL framework, the MAPPO-based scheme, and the MAB-based scheme in detail. Simulation results and discussions are presented in Section 4. Finally, Section 5 provides some concluding remarks.
Some key notations are listed in Table 2.

2. System Model

Consider a multiple-UAV-assisted MEC system that consists of N UAVs and M ( M > N ) UDs as shown in Figure 1. We denote the sets of UDs and UAVs by M = { 1 , 2 , , M } and N = { 1 , 2 , , N } , respectively. The UDs are randomly located and the UAVs fly along pre-determined trajectories. Each UAV is equipped with an MEC server and provides computing service for UDs. Each UD decides its own association and offloading strategy.

2.1. Association

At each timeslot, each UD associates with a suitable UAV. If multiple UDs associate with the same UAV, frequency division multiple access (FDMA) is adopted to avoid interference. The association between UDs and UAVs is indicated by a m , n , t . If UD m associates with UAV n in timeslot t, a m , n , t = 1 ; otherwise, a m , n , t = 0 . Each UD can only associate with one UAV simultaneously, meaning that we have n N a m , n , t = 1 .

2.2. Communication Model

The UAVs move continuously and the transmission links between UDs and UAVs can be blocked by buildings or trees. We assume that the state of each transmission link is determined by the environment and the angle of elevation between UDs and UAVs [50]. The transmission link between UD m and UAV n is line-of-sight (LOS) with probability p m , n L O S and non-line-of-sight (NLOS) with probability p m , n N L O S ( t ) = 1 p m , n L O S .
The channel gain of transmission links is determined by large-scale fading and small-scale fading. The channel gain between UAV n and UD m is obtained as
g m , n = ( 4 π f c c ) 2 μ L d m , n β L h m , n R i c e 2 , L O S ( 4 π f c c ) 2 μ N d m , n β N h m , n R a y l e i g h 2 , N L O S
where f c is the carrier frequency; c is the speed of light; β L ( β N ) and μ L ( μ N ) are the path loss exponent and the attenuation factor for LOS (NLOS) links, respectively; d m , n is the distance between UD m and UAV n; and h i , j R i c e and h i , j R a y l e i g h are the Rice and Rayleigh fading coefficients, respectively.
Assume that the channel gains are constant during the channel coherent time interval. The signal-to-noise ratio (SNR) between UD m and UAV-BS n in timeslot t is
γ m , n , t = P g m , n , t σ 2
where P is the transmission power of the UD, g i , j , t is the channel gain, and σ 2 is the noise power.
M n , t denotes the number of UDs that access UAV n in timeslot t. Then, the maximum achievable transmission rate between UD m and UAV n in timeslot t is
r m , n , t = B M n , t log 2 ( 1 + γ m , n , t )
where B is the total bandwidth.

2.3. Computation Model

At each timeslot, each UD generates its computing tasks with a random size. D m , t denotes the size of tasks generated by UD m at timeslot t. All UDs adopt the partial offloading policy and determine the offloading proportion. The proportion of tasks that are offloaded to the MEC server is denoted by α m , t [ 0 , 1 ] and ( 1 α m , t ) is the proportion of tasks that are locally executed.

2.3.1. Edge Computing

The UAV-assisted computing can be divided into three steps: task offloading, edge computing, and result downloading. The UDs first offload their tasks to the serving UAVs and the UAVs compute the received tasks and download the computation results to UDs. The computing results obtained by the MEC server are usually small. The result downloading time is short and can be ignored [17]. The delay in MEC consists of two parts: transmission delay and computation delay.
We neglect the result downloading delay, meaning that the transmission delay equals the task offloading time. The task offloading time from UD m to UAV n at timeslot t can be expressed as
τ m , n , t trans = α m , t D m , t r m , n , t
The energy consumption observed when UD m offloads its task to UAV n is given by
E m , n , t trans = P τ m , n , t t r a n s .
The computing time at the MEC server can be expressed as
τ m , n , t comp = α m , t D m , t s f m , n , t
where s is the number of CPU cycles required to process each unit byte and f m , n , t denotes the computing capacity of the UAV n allocated to UD m. The energy consumption of the UAV n to compute the offloading task is given by
E m , n , t comp = κ 1 f m , n , t 2 α m , t D m , t s
where κ 1 is a constant that depends on the CPU of the UAV-assisted MEC server.

2.3.2. Local Computing

The local computation delay of UD m at timeslot t can be expressed as
τ m , t local = ( 1 α m , t ) D m , t s f u
where f u denotes the computing capability of the UD. We assume that all UDs have a predetermined and fixed computing capability.
The energy consumption of the UD computing the local task can be obtained as
E m , t local = κ 2 f u 2 ( 1 α m , t ) D m , t s
where κ 2 is a constant which depends on the CPU of UD.

2.4. System Energy Consumption and Latency

The energy consumption to complete UD m’s task in timeslot t comes from the local computing and the edge computing, which can be obtained as
E m , t = E m , t local Local   computing UD + n N a m , n , t ( E m , n , t trans UD + E m , n , t comp UAV ) Edge   computing .
The total energy consumption for tasks computing in timeslot t is
E t = m M E m , t
The latency to complete UD m’s task in timeslot t can be obtained as
τ m , t = max { τ m , t local , n N a m , n , t ( τ m , n , t trans + τ m , n , t comp ) } .
The system latency is defined as the maximum UD’s latency, which can be obtained as
τ t = max m M { τ m , t } .

2.5. Problem Formulation

This paper aims to minimize the energy consumption and the latency simultaneously. We define the system cost in timeslot t as the weighted sum of energy consumption and latency [23,24]
C t = λ E t + ( 1 λ ) τ t
where λ [ 0 , 1 ] adjusts the weights. The cost degenerates into energy consumption with λ = 1 and latency with λ = 0 .
Due to the time-varying channel state, UDs should select a suitable serving UAV, make a reasonable offloading decision, and request sufficient computing resources based on the current environment at each timeslot. We jointly optimize the association, offloading proportion, and computing resources allocation to minimize the long-term system cost. The joint optimization problem can be formulated as
P 1 : min { a m , n , t } , { α m , t } , { f m , n , t } t = 1 C t
s . t . a m , n , t = { 0 , 1 } , m M , n N , t 1 ,
n N a m , n , t = 1 , m M , t 1 ,
α m , t [ 0 , 1 ] , m M , t 1 ,
f m , n , t [ 0 , F max ] , m M , t 1
τ t τ max , t 1
where constraints (15b) and (15c) denote that each UD can only associate with one UAV in a timeslot; constraint (15d) is the range of the offloading proportion; constraint (15e) limits the maximum computing resources of the MEC; and constraint (15f) indicates that the latency requirement that cannot exceed the maximum value τ max . The difficulty of problem P 1 lies in the long-term objective and the coupling of association, offloading, and computing resource allocation.

3. Multi-Agent DRL-Based Association, Offloading, and Resource Allocation Schemes

Although the trajectories of UAVs can be predetermined, the variation in channel state and the variety in resources allocation make the system highly dynamic. It is challenging to tackle the long-term optimization problem in highly dynamic scenarios, especially for large networks. Fortunately, RL and DRL methods have been applied to solve these sequential decision problems successfully. The multi-agent scheme can further reduce the overheads of information collection and exchange in large networks. Therefore, we adopt the multi-agent RL framework and propose an RL- and a DRL-based scheme. We first introduce the multi-agent RL framework and formulate the long-term problem as a POMDP. Then, we elaborate the MAPPO-based and the CF-MAB-based schemes.

3.1. Multi-Agent RL Framework

In the multi-agent RL framework, all UDs are agents. Each agent maintains a stochastic policy to make decisions based on its local observations and then receives its corresponding reward independently. The long-term joint allocation problem can be treated as a POMDP. The policy of agent m is denoted by π ( m ) a t ( m ) | o t ( m ) , where a t ( m ) and o t ( m ) are the action and observation, respectively. We aim to find a joint policy for all UDs to minimize the long-term cost. The joint policy is the product of individual policies π = m = 1 M π ( m ) .
The observation contains the channel gains and the volume of current tasks. The observation of UD m at timeslot t can be denoted by
o t m = { { g m , n , t } n N , D m , t } .
We define the global state as the ensemble of observations of all UDs:
s t = { { o t m } m M } .
Each UD agent makes an action to decide its association, offloading proportion, and computing resource demand based on its policy. The action of UD m at timeslot t can be denoted by
a t m = { { a m , n , t } n N , α m , t , f m , n 0 , t | a m , n 0 , t = 1 } .
The joint action of all UDs is denoted by
a t = { { a t m } m M } .
After its action, each agent will receive an instantaneous reward from the environment. In our multi-agent framework, there is no global information exchange among UD agents, meaning that each UD can only obtain a direct reward based on its own local information, such as the transmission rate r m , n , t , energy consumption E m , t , and latency τ m , t . The reward of UD m after timeslot t can be denoted by
r t m = { { r m , n , t } n N , E m , t , τ m , t } .
Each UD agent can only obtain its local information, meaning that it cannot obtain the system cost. We introduce the individual cost based on its own reward for each UD as
C m , t = λ E m , t + ( 1 λ ) τ m , t .
Then, each UD agent updates its policy to minimize its individual cost instead.
There is a gap between minimizing the individual cost and minimizing the system cost. We use the substitution based on the fact that the system cost achieves the minimum when all UDs achieve their minimum individual cost synchronously.
All UD agents interact with the environment and other agents through the observation–action–reward process. A good policy is the key to achieving effective learning and efficient cooperation. We propose a MAPPO-based scheme and an MAB-based scheme to find the policies that can minimize the long-term cost.

3.2. MAPPO-Based Scheme for Long-Term Consumption Minimization

The MAPPO-based scheme is a multi-agent scheme that uses the AC architecture and the PPO. Specifically, each agent has an actor network and a critic network. The former makes actions and the later evaluates the actions. We first introduce the general MAPPO procedure. Then, we explain the reward functions in centralized training and decentralized execution, respectively. Finally, we summarize the MAPPO-based scheme.

3.2.1. Typical MAPPO Procedure

The actor network and the critic network are parameterized by θ and ϕ , respectively. The expected discounted accumulated reward can be denoted by
J ( θ ) = E a t , s t [ t γ t R t ] ,
where R t is the instantaneous reward in timeslot t and γ is the discount factor. Policy gradient methods maximize the expected reward by repeatedly estimating the gradient on θ . The most commonly used policy gradient has the form
g ^ = E ^ t θ log π θ a t | s t A ^ t
where π θ a t | s t is a stochastic policy, the expectation E ^ · indicates the empirical average over a finite batch of samples, and A ^ t is an estimator of the advantage function at timeslot t. We adopt the generalized advantage estimation (GAE) method and use the truncated estimator [51]
A ^ t = δ t + ( γ λ 0 ) δ t + 1 + + ( γ λ 0 ) T t + 1 δ T 1
where T is the length of trajectory segment, parameter λ 0 balances the bias and variance, δ t = R t + γ V ( s t + 1 ) V ( s t ) , and V ( s t ) is the state-value function.
The actor is trained to maximize the objective function F ( θ ) as [47]
F ( θ ) = 1 N b M i = 1 N b m = 1 M min { κ θ , i ( m ) A ^ i ( m ) , clip ( κ θ , i ( m ) , 1 ϵ , 1 + ϵ ) A ^ i ( m ) } + σ ν [ π θ ( s i ( n ) ) ] ,
where N b is the size of the mini-batch; M is the number of agents; κ θ , i ( m ) = π θ , i ( m ) ( a i | s i ) π θ o l d , i ( m ) ( a i | s i ) is the probability ratio between the current policy and the updated policy of agent m; A ^ i ( n ) is the advantage that can be estimated with (24); clip   ( κ θ , i ( n ) , 1 ϵ , 1 + ϵ ) is a clip function that restricts κ θ , i ( m ) into the interval [ 1 ϵ , 1 + ϵ ] ; σ is the entropy coefficient hyper parameter; and ν is the policy entropy to increase the exploration rate.
The critic network is trained to minimize the objective function [47]
L ( ϕ ) = 1 N b M i = 1 N b m = 1 M max { ( V ϕ ( s i ( n ) ) R i ) 2 , ( clip ( V ϕ ( s i ( n ) ) , V ϕ o l d ( o i ( n ) ) ϵ , V ϕ o l d ( o i ( n ) ) + ϵ ) R i ) 2 } ,
where R i is the cumulative discounted reward.
The objective functions in (25) and (26) can be optimized through the gradient ascent method. The objective functions are also propagated back to update parameters of the actor and critic by means of the Adam method.

3.2.2. Centralized Training and Decentralized Execution

The MAPPO-based scheme adopts centralized training and decentralized execution. In the training stage, all agents share a centralized critic network that evaluates the actions of all agents. This critic network should know all the global states and all agents’ observations and actions. Therefore, all agents share their local observations and receive the same critic network feedback. However, in the execution stage, each agent only has local observations and actions without coordination. To solve this problem, we use different rewards and gradients for the training and execution stages as follows.
To bridge the system cost and the individual cost, we set two kinds of reward functions. Each agent generates its individual reward function based its own cost as
R m , t = 1 C m , t .
Each UD should complete its task in the current timeslot, otherwise it will fail to meet the latency demand. Therefore, the reward function of the whole system should consider the cost of all UDs and the number of UDs that fail. We define the system reward function as
R M , t = 1 K t M m M R m , t
where K t is the number of UDs that fail to complete tasks in timeslot t and the term 1 K t M can be a penalty. The overall reward function R M , t is used in training for all agents and the individual reward function R m , t is used in execution for network updates.

3.2.3. MAPPO-Based Algorithm

The MAPPO-based algorithm for UAV-assisted MEC system is summarized in Algorithm 1. θ = { θ ( m ) } m M and ϕ = { ϕ ( m ) } m M are the combinations of parameters of all UD agents, where θ ( m ) and ϕ ( m ) are the parameters of the actor and the critic of UD agent m, respectively. There are some differences in the training stage and the execution stage. In the centralized training stage, the computations in line 11 use the overall reward function R M , t , while those in line 15 compute the gradients of objective functions in (25) and (26). In the decentralized execution stage, the computations in line 11 use the individual reward function R M , t , while those in line 15 compute the gradients of individual objective functions as
F ( m ) ( θ ) = 1 N b i = 1 N b min { κ θ , i ( m ) A ^ i ( m ) , clip ( κ θ , i ( m ) , 1 ϵ , 1 + ϵ ) A ^ i ( m ) } + σ ν [ π θ ( s i ( n ) ) ]
and
L ( m ) ( ϕ ) = 1 N b i = 1 N b max { ( V ϕ ( s i ( n ) ) R i ) 2 , ( clip ( V ϕ ( s i ( n ) ) , V ϕ o l d ( o i ( n ) ) ϵ , V ϕ o l d ( o i ( n ) ) + ϵ ) R i ) 2 } .
Algorithm 1 MAPPO-Based algorithm
1:Initialize parameter θ of the actor network and parameter ϕ of the critic network.
2:for each episode i = 1 , , E p  do
3:   Initialize experience reply buffer U;
4:   for each timeslot t = 1 , , T  do
5:     for each UD m M  do
6:        Execute action according to π θ o l d ( a t ( m ) | o t ( m ) ) ;
7:        Get the reward r t ( m ) and the next observation o t + 1 ( m ) ;
8:     end for
9:   end for
10:   Get trajectory τ = { o t , a t , r t } t = 1 T ;
11:   Compute the cumulative discounted reward R t , the state-value function V ϕ o l d , and the advantage { A t ( m ) } t = 1 T (According to (24)).
12:   Split τ , R t , and A t into chunks of length N b , and store them in U.
13:   for mini-batch j = 1 , . . . , B c h  do
14:     Randomly select a chunk from U as mini-batch b;
15:     Compute gradients of (25) and (26) on θ and ϕ , respectively, using mini-batch b;
16:     Update θ and ϕ using by Adam.
17:   end for
18:end for

3.3. CF-MAB-Based Scheme for Long-Term Consumption Minimization

The deployment and training of neural networks as agents can introduce high requirements for hardware and software in UDs. To reduce the overheads and complexity further, we propose a lightweight scheme that separates the association from the offloading decision and the computing resource allocation. The association directly determines the achievable rates between UDs and UAVs. Then, we find the closed-form offloading proportion and computing resource allocation for UDs given achievable rates. We first explain how to obtain the closed-form solution for the given association. Then, we explain the details of the MAB-based association. Finally, we summarize the CF-MAB-based scheme.

3.3.1. Closed-Form Offloading and Computing Resource Allocation for a Given Association

For a given association, each UD decides its offloading ratio and computation resources to minimize its individual cost. Assuming that UD m is associated with UAV n in timeslot t, the problem at UD m can be formulated into
P 2 : min α m , t , f m , n , t C m , t
s . t . α m , t [ 0 , 1 ] ,
f m , n , t [ 0 , F max ] ,
τ m , t τ max .
Problem P 2 is a short-term decision problem and can be solved at each UD independently.
From (31d), we can obtain
τ m , n , t trans + τ m , n , t comp τ max
and
τ m , t local τ max .
Substituting (4), (6), and (8) into (32) and (33), respectively, and combining (31c), we can obtain
1 f u τ max D m , t s α m , t τ max D m , t r m , n , t + D m , t s F max .
Usually, the system should satisfy τ max D m , t r m , n , t + D m , t s F max 1 so that all tasks can be executed at the MEC server. We can also obtain
α m , t D m , t s τ max α m , t D m , t r m , n , t f m , n , t F max .
We further analyze the individual cost function and obtain
C m , t = λ ( E m , t local + E m , n , t trans + E m , n , t comp ) + ( 1 λ ) ( τ m , n , t trans + τ m , n , t comp ) , α m , t α m , t λ ( E m , t local + E m , n , t trans + E m , n , t comp ) + ( 1 λ ) τ m , t local , α m , t < α m , t
where α m , t = s f u 1 α m , t f u s α m , t r m , n , t is obtained by setting τ m , n , t trans + τ m , n , t comp = τ m , t local . We check the gradient of C m , t on { α m , t , f m , n , t } , i.e., α m , t , f m , n , t C m , t , and find no stationary points in the feasible region. Consequently, the minimum value of C m , t is located at the boundaries with (1) α m , t = 1 , (2) f m , n , t = F m a x , and (3) α m , t = α m , t . A typical case of C m , t is shown in Figure 2.
When α m , t = 1 , UD m offloads all its tasks to the MEC server. The cost function becomes
C m , t = λ κ 1 D m , t s f m , n , t 2 + ( 1 λ ) D m , t s f m , n , t + λ P D m , t + ( 1 λ ) D m , t r m , n , t .
To achieve the minimum cost, we set d C m , t d f m , n , t = 0 and obtain the optimal solution f m , n , t * = 1 λ 2 λ κ 1 3 . We also get the lower bound f m , n , t lb = D m , t s τ max D m , t r m , n , t from (35). Then, we obtain a optimal solution candidate 1 , min { f m , n , t * , f m , n , t l b } .
When f m , n , t = F max , C m , t is a piecewise linear function with respect to α m , t . It is monotonically increasing with α m , t α m , t . Moreover, if λ < 1 κ 1 f u F max 2 κ 2 f u 3 + 1 and r m , n , t λ P f u 10 λ f u 3 + ( 1 λ ) λ F max 2 f u , C m , t is monotonically decreasing with α m , t α m , t . Then, the optimal candidate is α m , t , F max . Otherwise, C m , t is monotonically increasing with α i α m , t . Then, we get the lower bound α m , t l b = 1 f u τ max D m , t s from (34) and the optimal candidate is α m , t l b , F max .
When α m , t = α m , t , the cost function C m , t becomes
C m , t = f m , n , t ( f u r m , n , t s + 1 ) f m , n , t + f u λ κ 1 f m , n , t 2 + λ P r m , n , t s λ κ 2 f u 2 1 λ f u D m , t s + ( λ κ 2 f u 2 + 1 λ f u ) D m , t s .
We set d C m , t d f m , n , t = 0 and obtain
f m , n , t = X 3 2 3 κ 1 f u 6 2 3 κ 1 ( f u r m , n , t s + 1 ) + 3 f u 2 κ 1 2 2 2 3 κ 1 ( f u r m , n , t s + 1 ) X
where X = Y + Y 2 2916 f u 6 κ 1 6 1 3 , Y = f u 3 κ 2 Z f u P Z r m , n , t s + Z 54 f u 3 κ 1 3 , and Z = 108 κ 1 f u r m , n , t s + 1 2 . Then, we get a optimal candidate α m , t , f m , n , t .
Finally, we can obtain the optimal offloading proportion α m , t opt and computing resource allocation f m , n , t opt for each UD from the three candidates.

3.3.2. MAB-Based Long-Term Association

We tackle the long-term association problem using a distributed MAB scheme. Each UD is an agent that selects its association to maximize its long-term achievable rate. The achievable rate of each UD depends on its channel gain and association with other UDs. Then, it varies over time due to the UAVs moving and associations changing. On the other hand, a slight variety will incur frequent handover, especially when a UD can associate with different UAVs to achieve similar rates. Therefore, the MAB-based scheme should obtain the channel conditions in a timely manner and avoid frequent handover.
Each UD agent decides its association and receives a simultaneous reward at each timeslot. Then, it updates its exploration and exploitation strategy using the cumulative rewards. The weight, the association probability, and the reward function of UAV n for UD m in timeslot t are denoted by w m , n , t , p m , n , t , and R m , n , t , respectively.
Each UD selects its association according to the probability
p m , n , t = ( 1 γ 0 ) w m , n , t n N w i , n , t + γ 0 N
where γ 0 0 , 1 adjusts the proportion of exploitation and exploration.
After its action, each UD receives its reward and the reward function is defined as
R m , n , t = ψ m , n , t tanh r m , n , t r ¯ m , n , t r ˜ m , n , t , r i , j , t r ¯ m , n , t 0 , otherwise
where tanh ( · ) is the hyperbolic tangent activation function; ψ m , n , t is a reward discount rate to reduce frequent handover expressed as
ψ m , n , t = 1 , a m , n , t = a m , n , t 1 ψ , a m , n , t a m , n , t 1
with the discount factor ψ [ 0 , 1 ] ; r ¯ m , n , t is the average achieved rate in the past L timeslots, expressed as
r ¯ m , n , t = 1 L τ = t L τ = t 1 r m , n , τ ;
and r ˜ m , n , t is the bias to balance the load of UAVs, expressed as
r ˜ m , n , t = 1 M n , t k M n , t r k , n , t 1 M n , t 1 k M n , t , k m r k , n , t
where M n , t is the set of UDs that associate with UAV n at timeslot t.
Then, each UD can obtain its cumulative reward and update the weight as
G m , n , t = τ = 1 τ = t R m , n , τ p m , n , τ ,
and
w m , n , t + 1 = exp ( η G m , n , t ) = w m , n , t exp ( η R m , n , t p m , n , t ) ,
respectively, where η > 0 determines how aggressively the algorithm to learn and update the distribution.
It should be noted that the historical information becomes outdated and harmful for the current decision when the UAVs pass over a long distance. To improve the learning performance, we set a slide window with length D as shown in Figure 3, and the accumulative reward becomes the accumulation from the latest D timeslots as
G m , n , t = τ = t D + 1 τ = t R m , n , τ p m , n , τ .

3.3.3. CF-MAB-Based Algorithm

Finally, the CF-MAB-based algorithm at each UD is summarized in Algorithm 2. We focus on one UD and can find that each UD’s behavior is similar to the EXP3 algorithm [52]. The convergence proof of the algorithm is given in our previous work [53].
Algorithm 2 CF-MAB-based algorithm
Require:  η , γ 0
1:Initialize G m , n , t 0 = 0 , w m , n , t 0 = 1 for m M and n N ;
2:for each timeslot t = 1 , , T  do
3:   for each UD m M  do
4:     Calculate probability distribution p m , n , t with (40);
5:     Make action a m , n , t randomly according to the probabilities { p m , 1 , t , , p i , N , t } ;
6:     Calculate the achieved rate r m , n , t ;
7:     Obtain the optimal offloading proportion α m , t opt , computing resource allocation and f m , n , t opt , and the minimum cost C m , t ( α m , t opt , f m , n , t opt ) ;
8:     Calculate rewards R m , n , t with (41) and cumulative rewards G m , n , t with (45) or (47);
9:     Update the weights w m , n , t + 1 ;
10:   end for
11:end for

4. Performance Evaluation

This section evaluates the proposed algorithms. As shown in Figure 4, we assume that M = 10 UDs are randomly distributed in a 5 × 5 km2 square area with fixed locations and N = 4 UAV-BSs are deployed above the square area with a fixed height of 300 m. The square area is divided into N = 4 parts and each part shares its center with a UAV’s circular trajectory. The UAVs fly with a constant speed of 12.69 m/s and the trajectory has a radius of 0.75 km.
Some key parameters are given in the following. The channel coherence time is 5 ms and the channel state is constant during coherence time. The duration of each timeslot is 1 ms. The maximum latency is set to be the timeslot duration, i.e., τ max = 1 ms. The bandwidth is B = 10 MHz, the noise power density is 174 dBm/Hz, and the transmit power of UD is P = 0.35 W. The number of CPU cycles required to compute each bit is s = 10 3 cycles/bit. The computing capability of each UD is f u = 0.6 GHz. The parameters that adjust energy consumption at the UAV server and at the UD are set to be κ 1 = 10 27 and κ 2 = 10 26 , respectively. The parameter that balances latency and energy consumption is set to be λ = 0.5 . Unless otherwise stated, the total observed duration spans 5 × 10 4 timeslots, the maximum computing resource is F max = 1.5 GHz, and the volume of tasks that should be computed in one timeslot is D t = m = 1 M D m , t = 10 kbits.
To compare the performance, we employ MAPPO, received signal strength (RSS)-based, and random association strategies, and local-only, remote-only, closed-form, and random offloading strategies as benchmarks. Specifically,
  • CF-MAPPO scheme: The UDs’ association is determined via a simplified MAPPO-based scheme to maximize the total achievable rate. Then, the offloading and computing resource allocation are obtained through a closed-form solution;
  • RSS + Remote-only scheme: Each UD associates to the UAV with maximum RSS and all its tasks are executed at the UAV server;
  • Local-only scheme: Each UD computes its tasks locally;
  • Random + Random means: Each UD associates with a random UAV and a random part of its tasks are executed at the UAV server.
Figure 5 compares the convergence performance in the training stage of the proposed MAPPO-based and CF-MAB-based schemes with benchmarks. It is obvious that the RSS + Remote-only scheme, the local-only scheme, and the Random + Random scheme are too simple to achieve satisfying performance. It can be found that the MAPPO-based scheme reduces the system cost significantly and converges to the lowest cost. The reason for this is that the MAPPO-based scheme can learn knowledge from the environment and the historic strategies effectively and efficiently. Then, all UDs can make satisfactory decisions to obtain the best performance. It can also be found that the CF-MAB-based scheme converges quickly and achieves sufficiently low system cost. The reason for this is that the CF-MAB-based scheme reduces the dimensions of the action space significantly by separating the association from the offloading proportion and computing resource allocation. The MAPPO-based scheme adopts centralized training using global information and can learn knowledge more sufficiently and effectively than CF-MAB-based scheme in which each UD only learns from individual historic information. The CF-MAPPO-based scheme achieves a slightly weaker performance than the CF-MAB-based scheme since they both decouple the joint optimization and use the closed-form resource allocation. Moreover, the CF-MAPPO-based scheme uses more effective centralized learning than the CF-MAB-based scheme, meaning that it converges faster.
Figure 6 shows the cumulative system cost with different task computing demands. The tasks may not be completed under the maximum constraints, in which case a failure happens. To show the performance of all schemes completely, we still count the system cost in failures. Both the latency and the energy consumption increase with the task volume, meaning that the cumulative system cost increases in all schemes. The MAPPO-based scheme outperforms the others under all task demands. The CF-MAB-based scheme and the CF-MAPPO-based scheme achieve a similar performance, while the CF-MAB-based scheme demonstrates marginally better performance. It can also be noted that the RSS + Remote-only scheme obtains lower cumulative system cost than the CF-MAB- and CF-MAPPO-based schemes with tasks ( D t < 8 kbits). This is because the CF-MAB- and CF-MAPPO-based schemes take the latency constraints into consideration, meaning that they consume more energy during computing.
Figure 7 shows the cumulative system cost under different computing resource constraints. The volume of tasks in each timeslot is D t = 5 kbits. All schemes exhibit a monotonically decreasing cumulative system cost with respect to the computing resources below saturation thresholds (about 1 GHz). By optimally leveraging local and remote computation, these schemes achieve energy–latency trade-off optimization, resulting in a low cumulative system cost. The CF-MAB- and CF-MAPPO-based schemes decouple the transmission and the computation so that they cannot further reduce the cost when the computing resources are sufficiently large. The MAPPO-based scheme can use computation resources to compensate for communication in the joint optimization, meaning that it can further lower the cost.
The computation is completed successfully if the tasks are executed in the required time. Figure 8 shows the task completion rates of different schemes. The MAPPO-based scheme, the CF-MAB-based scheme, and the CF-MAPPO-based scheme outperform the other benchmarks significantly. Since the closed-form-based method takes the latency constraint into consideration, the CF-MAB and CF-MAPPO-based scheme can avoid failures and achieve slightly higher completion rates. The MAPPO-based scheme considers the latency constraint in the reward function and may select actions that result in exceeding latency, meaning that this scheme has a slightly lower completion rate. We further check Figure 6 and Figure 8 and can find that the RSS + Remote-only scheme has a significantly lower completion rate, though it achieves a slightly lower system cost.

5. Conclusions

This paper investigated the joint association, offloading, and computing resource allocation in multiple-UAV-assisted MEC networks. We introduce the cost function as a metric to balance the latency and the energy consumption. The joint latency and energy consumption minimization is transferred into a long-term cost minimization problem. Then, we adopt a multi-agent DRL framework and propose a MAPPO-based scheme and a CF-MAB-based scheme to solve this problem. The MAPPO-based scheme uses global observation for centralized learning and partial observation for decentralized execution. The CF-MAB-based scheme obtains associations to maximize the long-term transmission rates and achieve closed-form offloading and computing resource allocation. The numerical results validate the proposed schemes and show their superiority. Moreover, the MAPPO-based scheme can provide collaborative decisions for agents to obtain high performance in complex and dynamic environments. The CF-MAB-based scheme provides independent decisions and is suitable for scenarios needing rapid responses.

Author Contributions

Conceptualization, M.C., S.H., Y.P., M.L. and W.-P.Z.; methodology, M.C., S.H. and Y.P.; software, M.C. and S.H.; validation, M.C. and S.H.; writing-original draft preparation, M.C. and S.H.; writing-review and editing, M.C., M.L. and W.-P.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 62301282.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Nguyen, D.C.; Ding, M.; Pathirana, P.N.; Seneviratne, A.; Li, J.; Niyato, D.; Dobre, O.; Poor, H.V. 6G Internet of Things: A Comprehensive Survey. IEEE Internet Things J. 2022, 9, 359–383. [Google Scholar] [CrossRef]
  2. Aouedi, O.; Vu, T.H.; Sacco, A.; Nguyen, D.C.; Piamrat, K.; Marchetto, G.; Pham, Q.V. A Survey on Intelligent Internet of Things: Applications, Security, Privacy, and Future Directions. IEEE Commun. Surv. Tutor. 2025, 27, 1238–1292. [Google Scholar] [CrossRef]
  3. McEnroe, P.; Wang, S.; Liyanage, M. A Survey on the Convergence of Edge Computing and AI for UAVs: Opportunities and Challenges. IEEE Internet Things J. 2022, 9, 15435–15459. [Google Scholar] [CrossRef]
  4. Spinelli, F.; Mancuso, V. Toward Enabled Industrial Verticals in 5G: A Survey on MEC-Based Approaches to Provisioning and Flexibility. IEEE Commun. Surv. Tutor. 2021, 23, 596–630. [Google Scholar] [CrossRef]
  5. Azari, M.M.; Solanki, S.; Chatzinotas, S.; Kodheli, O.; Sallouha, H.; Colpaert, A.; Mendoza Montoya, J.F.; Pollin, S.; Haqiqatnejad, A.; Mostaani, A.; et al. Evolution of Non-Terrestrial Networks From 5G to 6G: A Survey. IEEE Commun. Surv. Tutor. 2022, 24, 2633–2672. [Google Scholar] [CrossRef]
  6. Mahboob, S.; Liu, L. Revolutionizing Future Connectivity: A Contemporary Survey on AI-Empowered Satellite-Based Non-Terrestrial Networks in 6G. IEEE Commun. Surv. Tutor. 2024, 26, 1279–1321. [Google Scholar] [CrossRef]
  7. Panahi, F.H.; Panahi, F.H. Reliable and Energy-Efficient UAV Communications: A Cost-Aware Perspective. IEEE Trans. Mobile Comput. 2024, 23, 4038–4049. [Google Scholar] [CrossRef]
  8. Kirubakaran, B.; Vikhrova, O.; Andreev, S.; Hosek, J. UAV-BS Integration with Urban Infrastructure: An Energy Efficiency Perspective. IEEE Commun. Mag. 2025, 63, 100–106. [Google Scholar] [CrossRef]
  9. Peng, H.; Shen, X. Multi-Agent Reinforcement Learning Based Resource Management in MEC- and UAV-Assisted Vehicular Networks. IEEE J. Sel. Areas Commun. 2021, 39, 131–141. [Google Scholar] [CrossRef]
  10. Ullah, S.A.; Hassan, S.A.; Abou-Zeid, H.; Qureshi, H.K.; Jung, H.; Mahmood, A.; Gidlund, M.; Imran, M.A.; Hossain, E. Convergence of MEC and DRL in Non-Terrestrial Wireless Networks: Key Innovations, Challenges, and Future Pathways. IEEE Commun. Surv. Tutor. 2025, 1–39, early access. [Google Scholar]
  11. Tun, Y.K.; Park, Y.M.; Tran, N.H.; Saad, W.; Pandey, S.R.; Hong, C.S. Energy-Efficient Resource Management in UAV-Assisted Mobile Edge Computing. IEEE Commun. Lett. 2021, 25, 249–253. [Google Scholar] [CrossRef]
  12. Liu, B.; Wan, Y.; Zhou, F.; Wu, Q.; Hu, R.Q. Resource Allocation and Trajectory Design for MISO UAV-Assisted MEC Networks. IEEE Trans. Veh. Technol. 2022, 71, 4933–4948. [Google Scholar] [CrossRef]
  13. Xu, B.; Kuang, Z.; Gao, J.; Zhao, L.; Wu, C. Joint Offloading Decision and Trajectory Design for UAV-Enabled Edge Computing With Task Dependency. IEEE Trans. Wireless Commun. 2023, 22, 5043–5055. [Google Scholar] [CrossRef]
  14. Li, Y.; Gao, X.; Shi, M.; Kang, J.; Niyato, D.; Yang, K. Dynamic Weighted Energy Minimization for Aerial Edge Computing Networks. IEEE Internet Things J. 2025, 12, 683–697. [Google Scholar] [CrossRef]
  15. Wang, C.; Zhai, D.; Zhang, R.; Cai, L.; Liu, L.; Dong, M. Joint Association, Trajectory, Offloading, and Resource Optimization in Air and Ground Cooperative MEC Systems. IEEE Trans. Veh. Technol. 2024, 73, 13076–13089. [Google Scholar] [CrossRef]
  16. Li, C.; Gan, Y.; Zhang, Y.; Luo, Y. A Cooperative Computation Offloading Strategy With On-Demand Deployment of Multi-UAVs in UAV-Aided Mobile Edge Computing. IEEE Trans. Network Serv. Manag. 2024, 21, 2095–2110. [Google Scholar] [CrossRef]
  17. Hu, Q.; Cai, Y.; Yu, G.; Qin, Z.; Zhao, M.; Li, G.Y. Joint Offloading and Trajectory Design for UAV-Enabled Mobile Edge Computing Systems. IEEE Internet Things J. 2019, 6, 1879–1892. [Google Scholar] [CrossRef]
  18. Zhang, L.; Ansari, N. Latency-Aware IoT Service Provisioning in UAV-Aided Mobile-Edge Computing Networks. IEEE Internet Things J. 2020, 7, 10573–10580. [Google Scholar] [CrossRef]
  19. Nasir, A.A. Latency Optimization of UAV-Enabled MEC System for Virtual Reality Applications Under Rician Fading Channels. IEEE Wireless Commun. Lett. 2021, 10, 1633–1637. [Google Scholar] [CrossRef]
  20. Sabuj, S.R.; Asiedu, D.K.P.; Lee, K.-J.; Jo, H.-S. Delay Optimization in Mobile Edge Computing: Cognitive UAV-Assisted eMBB and mMTC Services. IEEE Trans. Cogn. Commun. Netw. 2022, 8, 1019–1033. [Google Scholar] [CrossRef]
  21. Han, Z.; Zhou, T.; Xu, T.; Hu, H. Joint User Association and Deployment Optimization for Delay-Minimized UAV-Aided MEC Networks. IEEE Wirel. Commun. Lett. 2023, 12, 1791–1795. [Google Scholar] [CrossRef]
  22. Zhang, J.; Luo, H.; Chen, X.; Shen, H.; Guo, L. Minimizing Response Delay in UAV-Assisted Mobile Edge Computing by Joint UAV Deployment and Computation Offloading. IEEE Trans. Cloud Comput. 2024, 12, 1372–1386. [Google Scholar] [CrossRef]
  23. Yu, Z.; Gong, Y.; Gong, S.; Guo, Y. Joint Task Offloading and Resource Allocation in UAV-Enabled Mobile Edge Computing. IEEE Internet Things J. 2020, 7, 3147–3159. [Google Scholar] [CrossRef]
  24. Zhao, L.; Yang, K.; Tan, Z.; Li, X.; Sharma, S.; Liu, Z. A Novel Cost Optimization Strategy for SDN-Enabled UAV-Assisted Vehicular Computation Offloading. IEEE Trans. Intell. Transp. Syst. 2021, 22, 3664–3674. [Google Scholar] [CrossRef]
  25. Pervez, F.; Sultana, A.; Yang, C.; Zhao, L. Energy and Latency Efficient Joint Communication and Computation Optimization in a Multi-UAV-Assisted MEC Network. IEEE Trans. Wirel. Commun. 2024, 23, 1728–1741. [Google Scholar] [CrossRef]
  26. Kuang, Z.; Pan, Y.; Yang, F.; Zhang, Y. Joint Task Offloading Scheduling and Resource Allocation in Air–Ground Cooperation UAV-Enabled Mobile Edge Computing. IEEE Trans. Veh. Technol. 2024, 73, 5796–5807. [Google Scholar] [CrossRef]
  27. Yuan, H.; Wang, M.; Bi, J.; Shi, S.; Yang, J.; Zhang, J.; Zhou, M.; Buyya, R. Cost-Efficient Task Offloading in Mobile Edge Computing With Layered Unmanned Aerial Vehicles. IEEE Internet Things J. 2024, 11, 30496–30509. [Google Scholar] [CrossRef]
  28. Sun, G.; Wang, Y.; Sun, Z.; Wu, Q.; Kang, J.; Niyato, D.; Leung, V.C.M. Multi-Objective Optimization for Multi-UAV-Assisted Mobile Edge Computing. IEEE Trans. Mobile Comput. 2024, 23, 14803–14820. [Google Scholar] [CrossRef]
  29. Zhang, L.; Wen, F.; Zhang, Q.; Gui, G.; Sari, H.; Adachi, F. Constrained Multiobjective Decomposition Evolutionary Algorithm for UAV-Assisted Mobile Edge Computing Networks. IEEE Internet Things J. 2024, 11, 36673–36687. [Google Scholar] [CrossRef]
  30. Zhang, J.; Zhou, L.; Tang, Q.; Ngai, E.C.H.; Hu, X.; Zhao, H.; Wei, J. Stochastic Computation Offloading and Trajectory Scheduling for UAV-Assisted Mobile Edge Computing. IEEE Internet Things J. 2019, 6, 3688–3699. [Google Scholar] [CrossRef]
  31. Zeng, Y.; Chen, S.; Li, J.; Cui, Y.; Du, J. Online Optimization in UAV-Enabled MEC System: Minimizing Long-Term Energy Consumption Under Adapting to Heterogeneous Demands. IEEE Internet Things J. 2024, 11, 32143–32159. [Google Scholar] [CrossRef]
  32. Liu, B.; Peng, M. Online Offloading for Energy-Efficient and Delay-Aware MEC Systems with Cellular-Connected UAVs. IEEE Internet Things J. 2024, 11, 22321–22336. [Google Scholar] [CrossRef]
  33. Wang, J.; Wang, L.; Zhu, K.; Dai, P. Lyapunov-Based Joint Flight Trajectory and Computation Offloading Optimization for UAV-Assisted Vehicular Networks. IEEE Internet Things J. 2024, 11, 22243–22256. [Google Scholar] [CrossRef]
  34. Zhao, M.; Zhang, R.; He, Z.; Li, K. Joint Optimization of Trajectory, Offloading, Caching, and Migration for UAV-Assisted MEC. IEEE Trans. Mob. Comput. 2025, 24, 1981–1998. [Google Scholar] [CrossRef]
  35. Zhang, H.; Sun, Z.; Yang, C.; Cao, X. Latency Optimization in UAV-Assisted Mobile Edge Computing Empowered by Caching Mechanisms. IEEE J. Miniat. Air Space Syst. 2024, 5, 228–236. [Google Scholar] [CrossRef]
  36. Li, C.; Wu, J.; Zhang, Y.; Wan, S. Energy-Latency Tradeoff for Joint Optimization of Vehicle Selection and Resource Allocation in UAV-Assisted Vehicular Edge Computing. IEEE Trans. Green Commun. Netw. 2025, 9, 445–458. [Google Scholar] [CrossRef]
  37. Seid, A.M.; Boateng, G.O.; Anokye, S.; Kwantwi, T.; Sun, G.; Liu, G. Collaborative Computation Offloading and Resource Allocation in Multi-UAV-Assisted IoT Networks: A Deep Reinforcement Learning Approach. IEEE Internet Things J. 2021, 8, 12203–12218. [Google Scholar] [CrossRef]
  38. Wang, Y.; Farooq, J.; Ghazzai, H.; Setti, G. Joint Positioning and Computation Offloading in Multi-UAV MEC for Low Latency Applications: A Proximal Policy Optimization Approach. IEEE Trans. Mobile Comput. 2025, 1–15, early access. [Google Scholar]
  39. He, Y.; Xiang, K.; Cao, X.; Guizani, M. Task Scheduling and Trajectory Optimization-based on Fairness and Communication Security for Multi-UAV-MEC System. IEEE Internet Things J. 2021, 11, 30510–30523. [Google Scholar] [CrossRef]
  40. Yan, M.; Zhang, L.; Jiang, W.; Chan, C.A.; Gygax, A.F.; Nirmalathas, A. Energy Consumption Modeling and Optimization of UAV-Assisted MEC Networks Using Deep Reinforcement Learning. IEEE Sens. J. 2024, 24, 13629–13639. [Google Scholar] [CrossRef]
  41. Liu, Y.; Lin, P.; Zhang, M.; Zhang, Z.; Yu, F.R. Mobile-Aware Service Offloading for UAV-Assisted IoV: A Multiagent Tiny Distributed Learning Approach. IEEE Internet Things J. 2024, 11, 21191–21201. [Google Scholar] [CrossRef]
  42. Wang, Z.; Wang, H.; Liu, L.; Sun, E.; Zhang, H.; Li, Z.; Fang, C.; Li, M. Dynamic Trajectory Design for Multi-UAV-Assisted Mobile Edge Computing. IEEE Trans. Veh. Technol. 2025, 74, 4684–4697. [Google Scholar] [CrossRef]
  43. Kang, H.; Chang, X.; Mišić, J.; Mišić, V.B.; Fan, J.; Liu, Y. Cooperative UAV resource allocation and task offloading in hierarchical aerial computing systems: A MAPPO based approach. IEEE Internet Things J. 2023, 10, 10497–10509. [Google Scholar] [CrossRef]
  44. Cheng, M.; Zhu, C.; Lin, M.; Wang, J.B.; Zhu, W.P. An O-MAPPO scheme for joint computation offloading and resources allocation in UAV assisted MEC systems. Comp. Commun. 2023, 208, 190–199. [Google Scholar] [CrossRef]
  45. Cheng, M.; Zhu, C.; Lin, M.; Zhu, W.-P. A MAPPO Based Scheme for Joint Resource Allocation in UAV Assisted MEC Networks. In Proceedings of the IEEE/CIC International Conference on Communications in China (ICCC), Hangzhou, China, 7–9 August 2024. [Google Scholar]
  46. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal Policy Optimization Algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar] [CrossRef]
  47. Yu, C.; Velu, A.; Vinitsky, E.; Gao, J.; Wang, Y.; Bayen, A.; Wu, Y. The surprising effectiveness of PPO in cooperative multi-agent games. In Proceedings of the 36th International Conference on Neural Information Processing Systems (NIPS ’22), New Orleans, LA, USA, 28 November–9 December 2022. [Google Scholar]
  48. Guo, D.; Tang, L.; Zhang, X.; Liang, Y.-C. Joint Optimization of Handover Control and Power Allocation Based on Multi-Agent Deep Reinforcement Learning. IEEE Trans. Veh. Tech. 2020, 69, 13124–13138. [Google Scholar] [CrossRef]
  49. He, S.; Cheng, M.; Pan, Y.; Lin, M.; Zhu, W.-P. Distributed access and offloading scheme for multiple UAVs assisted MEC network. In Proceedings of the IEEE 98th Vehicular Technology Conference (VTC2023-Fall), Hong Kong, China, 10–13 October 2023. [Google Scholar]
  50. Pokkunuru, A.; Zhang, Q.; Wang, P. Capacity analysis of aerial small cells. In Proceedings of the IEEE International Conference on Communications (ICC), Paris, France, 21–25 May 2017. [Google Scholar]
  51. Schulman, J.; Moritz, P.; Levine, S.; Jordan, M.; Abbeel, P. High-Dimensional Continuous Control Using Generalized Advantage Estimation. arXiv 2018, arXiv:1506.02438. [Google Scholar] [CrossRef]
  52. Auer, P.; Cesa-Bianchi, N.; Freund, Y.; Schapire, R.E. Gambling in a rigged casino: The adversarial multi-armed bandit problem. In Proceedings of the IEEE 36th Annual Foundations of Computer Science, Milwaukee, WI, USA, 23–25 October 1995; pp. 322–331. [Google Scholar]
  53. Cheng, M.; He, S.; Lin, M.; Zhu, W.-P.; Wang, J. RL and DRL Based Distributed User Access Schemes in Multi-UAV Networks. IEEE Trans. Veh. Tech. 2025, 74, 5241–5246. [Google Scholar] [CrossRef]
Figure 1. The multi-UAV-assisted MEC system.
Figure 1. The multi-UAV-assisted MEC system.
Sensors 25 05234 g001
Figure 2. Plot of individual cost function C m , t with given association.
Figure 2. Plot of individual cost function C m , t with given association.
Sensors 25 05234 g002
Figure 3. The slide window with length D.
Figure 3. The slide window with length D.
Sensors 25 05234 g003
Figure 4. The locations of UAVs and UDs.
Figure 4. The locations of UAVs and UDs.
Sensors 25 05234 g004
Figure 5. Convergence performance in the training stage.
Figure 5. Convergence performance in the training stage.
Sensors 25 05234 g005
Figure 6. System’s cost with different task volumes.
Figure 6. System’s cost with different task volumes.
Sensors 25 05234 g006
Figure 7. System’s cost under different computing resource constraints.
Figure 7. System’s cost under different computing resource constraints.
Sensors 25 05234 g007
Figure 8. Task completion rate of different schemes.
Figure 8. Task completion rate of different schemes.
Sensors 25 05234 g008
Table 1. Comparison of related works.
Table 1. Comparison of related works.
WorkObjectiveMethod
EnergyLatency
[11,12,13,14,15,16] short-termConvex Optimization
[17,18,19,20,21,22]
[23,24,25,26,27,28,29]
[30,31] long-termLyapunov
[32,33]
[34]Service quantity
[35] ACSingle agent
[36]DDQN
[37]DDPG
[38] PPO
[39,42]MADDPGMultiple agents
[40]
[41]
[43]Task amountMAPPO
[44]Energy efficiency
[45]
Table 2. List of key notations.
Table 2. List of key notations.
NotationDescription
α m , t Offloading task proportion of UD m at timeslot t
β L , β N Path loss exponent for LOS and NLOS links
γ Discount factor for rewards
γ 0 Parameter that adjusts the exploitation and exploration
γ m , n , t SNR between UD m and UAV n in timeslot t
δ TD residual to calculate GAE in (24)
ϵ Determines the interval of clip function
η Parameter that determines how aggressively to learn and update
θ Parameter of actor network
κ 1 , κ 2 Weights of energy consumption by UAVs and UDs
κ θ , i ( m ) Probability ratio between the current and the updated policy
λ Adjusts the weights of energy consumption and latency
λ 0 Balances the bias and variance in GAE
μ L , μ N Attenuation factor for LOS and NLOS links
π ( m ) Policy of agent m
τ Latency
ϕ Parameter of critic network
a m , n , t Association between UD m and UAV n in timeslot t
a t ( m ) Action of agent m in timeslot t
C t , C m , t System’s cost and individual cost in timeslot t
D m , t Volume of tasks
EEnergy consumption
f u , f m , n , t Computing frequencies at UD and UAV
o t ( m ) Observation of UD m at timeslot t
r m , n , t Achievable rate between UD m and UAV n in timeslot t                  
r t ( m ) Reward of UD m at timeslot t
R m , t Individual reward function of UD m at timeslot t
R M , t System’s reward function at timeslot t
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cheng, M.; He, S.; Pan, Y.; Lin, M.; Zhu, W.-P. Cooperative Schemes for Joint Latency and Energy Consumption Minimization in UAV-MEC Networks. Sensors 2025, 25, 5234. https://doi.org/10.3390/s25175234

AMA Style

Cheng M, He S, Pan Y, Lin M, Zhu W-P. Cooperative Schemes for Joint Latency and Energy Consumption Minimization in UAV-MEC Networks. Sensors. 2025; 25(17):5234. https://doi.org/10.3390/s25175234

Chicago/Turabian Style

Cheng, Ming, Saifei He, Yijin Pan, Min Lin, and Wei-Ping Zhu. 2025. "Cooperative Schemes for Joint Latency and Energy Consumption Minimization in UAV-MEC Networks" Sensors 25, no. 17: 5234. https://doi.org/10.3390/s25175234

APA Style

Cheng, M., He, S., Pan, Y., Lin, M., & Zhu, W.-P. (2025). Cooperative Schemes for Joint Latency and Energy Consumption Minimization in UAV-MEC Networks. Sensors, 25(17), 5234. https://doi.org/10.3390/s25175234

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop