Next Article in Journal
Airport Ground-Based Aerial Object Surveillance Technologies for Enhanced Safety: A Systematic Review
Previous Article in Journal
A Lightweight Multi-Module Collaborative Optimization Framework for Detecting Small Unmanned Aerial Vehicles in Anti-Unmanned Aerial Vehicle Systems
Previous Article in Special Issue
Towards Scalable Intelligence: A Low-Complexity Multi-Agent Soft Actor–Critic for Large-Model-Driven UAV Swarms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Edge Intelligent Joint Optimization of UAV Computation Offloading and Trajectory Under Time-Varying Channels

1
Zhejiang University Binjiang Research Institute, Hangzhou 310053, China
2
Zhejiang Zhonghao Applied Engineering Technology Research Institute, Hangzhou 310016, China
*
Author to whom correspondence should be addressed.
Drones 2026, 10(1), 21; https://doi.org/10.3390/drones10010021
Submission received: 23 November 2025 / Revised: 20 December 2025 / Accepted: 25 December 2025 / Published: 31 December 2025
(This article belongs to the Special Issue Advances in AI Large Models for Unmanned Aerial Vehicles)

Highlights

What are the main findings?
  • The proposed Adaptive UAV Edge Intelligence Framework (AUEIF) effectively decouples the complexities of UAV trajectory, computation offloading, and dynamic air-to-ground communication, significantly improving adaptability and efficiency in multi-UAV MEC systems.
  • A hierarchical reinforcement learning (HRL) approach optimizes large-scale action spaces by combining high-level trajectory control with fine-grained offloading and resource allocation, enabling scalable and adaptive decisionmaking across different time scales.
What are the implications of the main findings?
  • The AUEIF demonstrates superior performance in task latency reduction and energy efficiency compared to conventional methods, with enhanced robustness under severe channel fading conditions, supporting reliable real-world UAV deployment.
  • The integration of dynamic modeling, predictive LSTM-based channel forecasting, and hierarchical learning offers a modular and scalable blueprint for next-generation adaptive MEC systems capable of operating in highly dynamic environments.

Abstract

With the rapid development of mobile edge computing (MEC) and unmanned aerial vehicle (UAV) communication networks, UAV-assisted edge computing has emerged as a promising paradigm for low-latency and energy-efficient computation. However, the time-varying nature of air-to-ground channels and the coupling between UAV trajectories and computation offloading decisions significantly increase system complexity. To address these challenges, this paper proposes an Adaptive UAV Edge Intelligence Framework (AUEIF) for joint UAV computation offloading and trajectory optimization under dynamic channels. Specifically, a dynamic graph-based system model is constructed to characterize the spatio-temporal correlation between UAV motion and channel variations. A hierarchical reinforcement learning-based optimization framework is developed, in which a high-level actor–critic module is responsible for generating coarse-grained UAV flight trajectories, while a low-level deep Q-network performs fine-grained optimization of task offloading ratios and computational resource allocation in real time. In addition, an adaptive channel prediction module leveraging long short-term memory (LSTM) networks is integrated to model temporal channel state transitions and to assist policy learning and updates. Extensive simulation results demonstrate that the proposed AUEIF achieves significant improvements in end-to-end latency, energy efficiency, and overall system stability compared with conventional deep reinforcement learning approaches and heuristic-based schemes while exhibiting strong robustness against dynamic and fluctuating wireless channel conditions.

1. Introduction

With the rapid advancement of mobile edge computing (MEC) and unmanned aerial vehicle (UAV) communications, UAV-assisted MEC has emerged as a promising paradigm for delivering low-latency and energy-efficient computing services to mobile users (MUs), particularly in areas with limited terrestrial infrastructure or unreliable network coverage [1,2,3]. Unlike conventional fixed MEC servers, UAVs possess high mobility and flexible deployment capabilities, enabling dynamic trajectory adjustment in response to spatio-temporal variations in user distribution, task generation, and wireless channel conditions, thereby enhancing service adaptability and resilience [4]. Nevertheless, UAV-assisted MEC faces significant challenges due to the highly dynamic air-to-ground (A2G) channels, which induce strong coupling between UAV mobility, task offloading decisions, and resource allocation. Many existing solutions inadequately capture the spatio-temporal correlations among these tightly interdependent components. Moreover, the high-dimensional decision space and complex system dynamics often lead to excessive computational complexity and slow convergence, hindering real-time adaptability in multi-UAV environments [5].
To address the above challenges, extensive research efforts have investigated optimization-based approaches for UAV-assisted MEC systems, primarily focusing on joint UAV trajectory planning and task offloading design. Yang et al. [6] proposed a perturbation-based Lyapunov optimization (PLOT) framework for energy-harvesting UAV-MEC systems, which transforms long-term stochastic optimization problems into per-slot deterministic subproblems. Hao et al. [7] developed a reliability-aware offloading framework by integrating perceptual representation with TD3-based deep reinforcement learning to maximize long-term task success probability. Nguyen et al. [8] studied joint partial offloading, UAV trajectory control, edge–cloud computation, and radio resource allocation in space–air–ground integrated networks, demonstrating notable improvements in energy efficiency. Similarly, Zhang et al. [9] employed convex optimization techniques to minimize average response delay through joint UAV deployment, user association, and task offloading strategies. Xu et al. [10] formulated a weighted computation efficiency maximization problem with aerial–ground cooperation, jointly optimizing task allocation, bandwidth assignment, CPU frequency, and UAV trajectory to balance computation performance and energy consumption.
To cope with the high-dimensional state and action spaces as well as the dynamic nature of UAV-MEC systems, learning-based and multi-agent approaches have attracted increasing attention. Zhao et al. [11] proposed a multi-agent deep reinforcement learning (MADRL) framework to jointly optimize UAV trajectories, task allocation, and communication resources, enabling scalable control in continuous action spaces. Chen et al. [12] developed a hybrid nature-inspired optimization scheme for joint UAV positioning, task offloading, and resource allocation. Al-Bakhrani et al. [13] introduced a multi-objective adaptive learning framework that integrates reinforcement learning, model predictive control, and particle swarm optimization to improve UAV trajectory planning and dynamic resource management. Song et al. [14] proposed an evolutionary multi-objective reinforcement learning algorithm to jointly optimize UAV trajectories and task offloading, balancing task latency, energy consumption, and data collection performance. Furthermore, Zhou et al. [15] extended learning-based frameworks to support real-time radio map construction, caching, and online task offloading, significantly reducing service latency under UAV energy constraints.
Despite these efforts, several limitations persist. Most existing methods either overlook the coupling between UAV trajectories and computation offloading in dynamic channels, or fail to consider multiple timescales in decision-making [16,17]. Additionally, achieving low-latency, energy-efficient, and stable UAV-MEC operations under time-varying channels remains an open problem, particularly in large-scale, multi-UAV scenarios with high-dimensional state and action spaces. These challenges motivate the need for an adaptive framework that can jointly optimize UAV trajectory, task offloading, and resource allocation while capturing channel dynamics and ensuring real-time adaptability.
To address these challenges, this paper proposes an Adaptive UAV Edge Intelligence Framework (AUEIF). The key contributions are summarized as follows:
  • We develop a dynamic system model that characterizes the spatio-temporal correlations among UAV mobility, task arrivals, and time-varying wireless channel conditions using a graph-based representation.
  • We design a hierarchical reinforcement learning architecture in which a high-level actor–critic module plans coarse UAV trajectories, while a low-level deep Q-network performs fine-grained task offloading and resource allocation in real time.
  • We introduce an adaptive channel prediction module based on LSTM to anticipate channel state evolution and enhance decision-making efficiency with respect to latency and energy consumption.
  • We conduct extensive simulation studies showing that the proposed AUEIF framework outperforms conventional DRL-based and heuristic approaches in terms of latency, energy efficiency, and policy stability under dynamic channel environments.
The remainder of this paper is organized as follows. Section 2 reviews the related work on UAV-assisted mobile edge computing. Section 3 presents the system model and problem formulation. Section 4 details the proposed Adaptive UAV Edge Intelligence Framework (AUEIF) and the hierarchical reinforcement learning algorithm. Section 5 describes the experimental setup. Section 6 presents the simulation results and performance analysis. Finally, Section 7 concludes the paper and discusses potential future research directions.

2. Related Work

Research on UAV-assisted mobile edge computing (MEC) can be categorized into three main directions: (i) joint task offloading and UAV trajectory optimization, (ii) security, anti-jamming, and multi-vehicle or multi-platoon collaboration, and (iii) reinforcement learning-based resource management and UAV-MEC integrated frameworks. The following subsections provide a comprehensive overview.
Many studies have focused on enhancing the performance of mobile edge computing (MEC) systems through the joint optimization of task offloading, communication resources, and UAV trajectories. For instance, Yang et al. [6] investigated a UAV-assisted MEC system with energy harvesting devices, where dynamic task offloading and UAV trajectory control were jointly optimized to minimize the overall energy consumption. Zhang et al. [18] proposed a comprehensive energy-efficient optimization framework for IoT devices, considering three computation modes, namely local execution, direct offloading to UAVs, and UAV-assisted relaying to access points. Their approach jointly optimized task bit allocation, time-slot scheduling, transmit power, and UAV trajectory using Lagrange dual decomposition and successive convex approximation techniques to obtain near-optimal solutions.
In multi-antenna UAV scenarios, Liu et al. [19] studied MISO UAV-assisted MEC networks, jointly optimizing UAV beamforming vectors, CPU frequencies, UAV trajectories, and UE transmit powers to minimize system energy consumption while satisfying task and trajectory constraints. Hui et al. [20] investigated service coverage optimization for UAV-assisted MEC networks, jointly optimizing UAV altitude and task offloading probability to maximize the successful edge computing probability (SECP). Sun et al. [21] proposed a multi-UAV joint optimization framework (JTORATC), which minimizes task completion latency, reduces UAV energy consumption, and maximizes offloaded tasks using decomposition methods for MINLP problems.
Security and robustness in UAV-assisted MEC have been investigated from multiple perspectives. Kwon et al. [22] introduced a self-certified broadcast authentication protocol for intelligent transportation systems (ITS) within UAV-MEC environments, enabling secure communication without relying on centralized authorities while preserving user privacy. Xu et al. [23] developed a joint resource and trajectory optimization framework to maximize secure computation in dual-UAV MEC systems, where one UAV executes computational tasks and a secondary UAV functions as a jammer to suppress potential eavesdroppers. Shao et al. [24] proposed a multi-agent deep reinforcement learning-based resource management scheme to counteract jamming attacks, dynamically adjusting UAV CPU frequency, bandwidth allocation, and channel access to enhance interference resilience.
Energy efficiency and cooperative strategies in multi-platoon or multi-UAV MEC networks have also been extensively studied. Duan et al. [25] formulated a weighted energy-efficiency maximization approach for multi-platoon UAV-assisted MEC systems, leveraging a 2D path tracking model and sequential quadratic programming for trajectory optimization. Zhao et al. [26] presented a cooperative secure transmission and computation (CSTC) strategy to mitigate threats from mobile collusive eavesdroppers, jointly optimizing UAV trajectories, interference beam patterns, transmit power, and task offloading to enhance system-wide security and efficiency.
Recent works leverage deep reinforcement learning and multi-agent frameworks for adaptive UAV-MEC resource management. Apostolopoulos et al. [27] introduced a risk-aware data offloading framework for UAV-assisted MEC under resource uncertainty, modeling user task offloading as a non-cooperative game. Liu et al. [28] developed a joint communication and computation scheduling approach for platooning vehicles using UAV-assisted MEC, incorporating wireless power transfer, air-to-ground and ground-to-air channels, and vehicular computation. Wang et al. [29] proposed a bi-objective ant colony optimization framework (bi-ACO) for UAV trajectory planning and task offloading. Sun et al. [30] developed a two-timescale approach (TJCCT) combining task offloading, computation resource allocation, and UAV trajectory control. Miao et al. [31] presented a drone swarm path planning method for MEC in industrial IoT, integrating global and local UAV path planning with task offloading. Li et al. [32] designed a multi-UAV framework for VR rendering, using a multi-agent deep reinforcement learning algorithm (MATD3) to jointly optimize UAV flight trajectories and rendering task allocation. Chen et al. [33] introduced a game-theory-based task offloading and resource pricing scheme for autonomous UAV-assisted MEC. Peng et al. [34] proposed a multi-UAV sensing-communication-computation integrated framework, using attention-based multi-agent PPO (MAPPO) to minimize weighted energy consumption while satisfying communication and sensing constraints.
Despite extensive research, key challenges remain. Most prior work either separates UAV trajectory planning from task offloading or ignores time-varying air-to-ground channels. Multi-UAV and multi-task coordination under dynamic environments is challenging due to high-dimensional state-action spaces and real-time decision requirements. Moreover, integrating latency, energy efficiency, reliability, and security in a unified framework remains underexplored. These limitations motivate the design of an adaptive hierarchical UAV edge intelligence framework that jointly optimizes UAV trajectories, computation offloading, and resource allocation under time-varying channels, which is the focus of this paper.

3. System Model

As depicted in Figure 1, we consider a multi-UAV-assisted mobile edge computing (MEC) network consisting of U UAVs and M user equipment (UE) over a total task period T. The time horizon is discretized into N equal-duration slots, each of length δ t = T / N . The UAVs collaborate to serve UEs by jointly optimizing task offloading ratios, UAV computation resource allocation, three-dimensional trajectories, and adaptive weight scheduling to handle heterogeneous workloads and dynamic task arrivals. The primary goal is to minimize a weighted combination of system latency and energy consumption while ensuring UAV flight safety, communication, and computation constraints are satisfied. The main symbols in the system are shown in Table 1.

3.1. UAV Three-Dimensional Trajectory and Collision Avoidance

Let q u ( n ) = [ x u ( n ) , y u ( n ) , H u ( n ) ] T denote the three-dimensional (3D) position of UAV u at time slot n, where x u ( n ) and y u ( n ) are the horizontal coordinates and H u ( n ) is the altitude. Formally,
q u ( n ) = [ x u ( n ) , y u ( n ) , H u ( n ) ] T , u = 1 , 2 , , U ,
where x u ( n ) , y u ( n ) are in meters (m) and H u ( n ) satisfies the altitude limits H min H u ( n ) H max .
To avoid inter-UAV collisions, a minimum separation distance L col is enforced between UAV pairs:
q u ( n ) q u ( n ) L col , u u .
Static obstacles are modeled as spheres with centers o k = [ x k , y k , H k ] T and radius R k . UAVs must maintain a minimum safe distance L safe from obstacles:
q u ( n ) o k R k + L safe , k ,
where · denotes the Euclidean norm.
To avoid inter-UAV collisions, a minimum separation distance L col is enforced between UAV pairs:
| q u ( n ) q u ( n ) | L col , u u .
UAV velocities are constrained by both speed and heading angle:
v u ( n ) = q u ( n + 1 ) q u ( n ) δ t , 0 v u ( n ) v max , | θ u ( n ) | θ max ,
where v u ( n ) is the velocity vector of UAV u, v max is the maximum UAV speed, and θ u ( n ) is the heading angle relative to the x-axis.
Each time slot δ t is divided into flying and hovering durations:
δ t = δ t fly + δ t hov , δ t hov = max m M u ( n ) η m , u ( n ) D m ( n ) R u , m ( n ) ,
where δ t fly and δ t hov are the flight and hovering times, respectively. Here, M u ( n ) is the set of UEs served by UAV u, η m , u ( n ) is the fraction of UE m’s task offloaded to UAV u, D m ( n ) is the task data size (in bits), and R u , m ( n ) is the uplink transmission rate (in bits/s).
The energy consumption of UAV u in slot n is composed of flying and hovering components:
E u ( n ) = E fly ( n ) + E hov ( n ) , E fly ( n ) = 1 2 M u v u ( n ) 2 , E hov ( n ) = p hov δ t hov ,
where M u is the UAV mass (kg), p hov is hovering power (W), E fly ( n ) represents the kinetic energy associated with UAV motion, and E hov ( n ) is the energy consumed while hovering.

3.2. UAV–UE Communication Model

The air-to-ground channel gain between UAV u and UE m is modeled as
h u , m ( n ) = β 0 d u , m ( n ) α g u , m ( n ) , d u , m ( n ) = | q u ( n ) w m ( n ) | ,
where β 0 is the reference path gain at 1 m, α is the path loss exponent, g u , m ( n ) represents small-scale fading, and w m ( n ) is the UE’s position.
The achievable uplink rate is given by Shannon’s formula:
R u , m ( n ) = B m , n log 2 1 + p m ( n ) | h u , m ( n ) | 2 σ 2 , B m , n = W K n ,
where B m , n is the bandwidth allocated to UE m, p m ( n ) is the UE transmit power, σ 2 is noise power, W is total system bandwidth, and K n is the number of subchannels.
For multi-UAV cooperation, the total rate received by UE m is
R tot m ( n ) = u U m ( n ) R u , m ( n ) ,
where U m ( n ) is the set of UAVs serving UE m.

3.3. Computation Offloading Model

The local computation latency at UE m is
T loc m ( n ) = C m D m ( n ) 1 u η m , u ( n ) f m UE ,
where C m is the required CPU cycles per bit, and f m UE is the UE CPU frequency. The corresponding energy consumption is
E m loc ( n ) = k m ( f m UE ) 3 T m loc ( n ) ,
with k m being the UE computation energy coefficient.
For offloaded tasks, the UAV computation latency is:
T off m , u ( n ) = η m , u ( n ) D m ( n ) R u , m ( n ) + η m , u ( n ) D m ( n ) C m f UAV m , u ( n ) ,
and the energy consumption for offloading is:
E off m , u ( n ) = p m ( n ) η m , u ( n ) D m ( n ) R u , m ( n ) + k u ( f UAV m , u ( n ) ) 2 η m , u ( n ) D m ( n ) C m ,
where f m , u UAV ( n ) is the computation resource allocated to UE m by UAV u, and k u is the UAV computation energy coefficient.
Constraints on UAV computation resources and offloading fractions are
m M u ( n ) f UAV m , u ( n ) f u UAV , 0 η m , u ( n ) 1 , u U m ( n ) η m , u ( n ) 1 .

3.4. Weighted System Objective Function

The total latency and energy consumption of the system are
T total = n = 1 N m = 1 M max T loc m ( n ) , u U m ( n ) T off m , u ( n ) ,
E total = n = 1 N m = 1 M E loc m ( n ) + ϕ u = 1 U n = 1 N E u ( n ) + n = 1 N m = 1 M u U m ( n ) E off m , u ( n ) ,
where ϕ is a weighting factor to balance UAV energy in the system-level objective.
The weighted sum objective is
F = ω T T ^ total + ω E E ^ total , ω T + ω E = 1 ,
where T ^ total and E ^ total are normalized latency and energy metrics, and the dynamic weights are adapted as
ω T ( n ) = ω 0 + α η sys β E rem ( n ) u = 1 U E u max , ω E ( n ) = 1 ω T ( n ) ,
with ω 0 being the initial latency weight, η sys representing the system-level workload utilization, E rem ( n ) the remaining energy of all UAVs, and E u max the maximum UAV energy capacity. The parameters α and β are design coefficients that control the sensitivity of the objective weights to system workload utilization and remaining energy, respectively. Specifically, α regulates the emphasis on latency minimization under high workload conditions, while β determines the extent to which energy conservation is prioritized as the overall UAV energy budget decreases.

3.5. Problem Formulation

The joint optimization problem aims to minimize the weighted sum of total system latency and total energy consumption across all UAVs and UEs, by jointly optimizing UAV trajectories, UE offloading ratios, transmit powers, and UAV computation resource allocations. Let q u [ n ] denote the 3D position of UAV u at time slot n, η m , u [ n ] the task offloading fraction from UE m to UAV u, p m [ n ] the UE transmit power, and f m , u UAV [ n ] the computation resource allocated by UAV u to UE m. The optimization problem is formulated as
P : min q u [ n ] , η m , u [ n ] , p m [ n ] , f m , u UAV [ n ] F s . t . C 1 : 0 η m , u [ n ] 1 , m M ,   u U m [ n ] ,   n N , C 2 : u U m [ n ] η m , u [ n ] 1 , m M ,   n N , C 3 : q u [ 1 ] = q u [ N + 1 ] , u U , C 4 :   q u [ n + 1 ] q u [ n ] v max δ t , u U ,   n N , C 5 :   q u [ n ] q u [ n ] L col , u u ,   n N , C 6 :   q u [ n ] o k R k + L safe , k , u U ,   n N , C 7 : H min H u [ n ] H max , u U ,   n N , C 8 : 0 p m [ n ] P m , max , m M ,   n N , C 9 : m M u [ n ] f m , u UAV [ n ] f u UAV , u U ,   n N , C 10 : n N f m , u UAV [ n ] δ t C m n N η m , u [ n ] D m ( n ) , m M ,   u U m [ n ] , C 11 : E m loc ( n ) + u U m [ n ] E m , u off ( n ) E m , max , m M ,   n N .
Constraints C1–C2 ensure that the offloading ratios are valid and do not exceed the total task size. Constraints C3–C7 enforce UAV trajectory continuity, collision avoidance, obstacle clearance, and altitude limits. Constraint C8 restricts UE transmit power. Constraints C9–C10 enforce UAV computation resource limits and guarantee task processing completion. Constraint C11 ensures that UE energy budgets are not exceeded.

4. Approach Design

To address the coupled optimization of UAV trajectory planning and computation offloading under highly dynamic air-to-ground channels, we propose an Adaptive UAV Edge Intelligence Framework (AUEIF). This framework is designed to simultaneously capture the spatial and temporal correlations among UAV mobility, wireless channel fluctuations, and computational workload distributions, enabling adaptive, reliable, and efficient decision-making in complex aerial networks.
AUEIF comprises three principal components. First, a dynamic graph-based modeling approach is employed to represent the UAV-MEC network topology and its time-varying channel states, facilitating explicit reasoning over the spatial relationships among UAVs, ground users, and communication links. Second, an LSTM-based channel prediction module is incorporated to forecast short-term channel variations, mitigating uncertainty and enhancing the accuracy and reliability of trajectory and offloading decisions. Third, a hierarchical reinforcement learning (HRL) strategy is utilized for real-time joint optimization, where macro-level policies determine UAV mobility and navigation, while micro-level policies optimize task offloading ratios, computational resource allocation, and transmission power, thereby achieving coordinated and efficient UAV-MEC operation.

4.1. Dynamic Graph Construction

To accurately capture the spatio-temporal dynamics of the UAV-MEC network, we represent the system as a dynamic graph G [ n ] at each discrete time slot n:
G [ n ] = ( V [ n ] , E [ n ] ) ,
where V [ n ] denotes the set of nodes and E [ n ] represents the set of edges encoding communication links between nodes. The node set is defined as
V [ n ] = { q u [ n ] , w m [ n ] } , u U ,   m M ,
where q u [ n ] corresponds to the three-dimensional position of UAV u at time slot n, and w m [ n ] captures the state of user equipment (UE) m, encompassing its location and computational workload. By integrating both UAV spatial positions and UE computational attributes into the graph nodes, this formulation provides a unified representation of network mobility and task demand, serving as a rigorous basis for coordinated trajectory planning and adaptive offloading strategies.
Edges are used to represent the time-varying communication quality between UAVs and UEs:
E [ n ] = { h u , m [ n ] } , h u , m [ n ] = β 0 d u , m α [ n ] g u , m [ n ] ,
where d u , m [ n ] is the Euclidean distance between UAV u and UE m, β 0 is the reference channel gain in free-space conditions, α is the path-loss exponent, and g u , m [ n ] accounts for small-scale fading effects. This formulation captures both the deterministic path-loss attenuation and the stochastic variations induced by the wireless environment, thus providing a realistic representation of the air-to-ground channel for subsequent decision-making.
To support graph-based reasoning and learning, we construct an adjacency matrix:
A [ n ] = [ a u , m [ n ] ] , a u , m [ n ] = f ( h u , m [ n ] ) ,
where f ( · ) normalizes the raw channel gains to edge weights that reflect link quality. By emphasizing stronger links and attenuating weaker ones, this adjacency matrix not only facilitates effective message propagation in graph-based models but also informs UAV trajectory and offloading decisions with an explicit awareness of the current network topology and channel conditions.

4.2. LSTM-Based Channel Prediction

While the dynamic graph captures the instantaneous spatial structure and communication topology, UAV trajectory planning and task offloading decisions are also strongly affected by temporal channel variations. To model these dynamics, we adopt a long short-term memory (LSTM) network to forecast future channel states:
h ^ u , m [ n + Δ ] = LSTM h u , m [ n τ : n ] ,
where τ denotes the historical observation window and Δ is the prediction horizon. By leveraging temporal correlations in the observed channel sequences, the LSTM enables the system to anticipate future channel conditions, which is critical for robust and proactive trajectory and offloading optimization.
The LSTM updates its internal memory through gated mechanisms:
c [ n ] = f [ n ] c [ n 1 ] + i [ n ] c ˜ [ n ] ,
where f [ n ] and i [ n ] are the forget and input gates, respectively, c ˜ [ n ] is the candidate cell state, and ⊙ denotes element-wise multiplication. This structure ensures that relevant historical information is preserved while outdated or less informative data is filtered out, improving the accuracy of channel prediction.
The predicted channel states are subsequently integrated into the UAV decision-making framework:
η m , u [ n ] = softmax f θ ( h ^ u , m [ n + Δ ] , D m [ n ] , E rem , u [ n ] ) ,
where D m [ n ] is the task size and E rem , u [ n ] represents the remaining energy of UAV u. By fusing the structural insights from the dynamic graph with the temporally predicted channel states, the framework enables anticipatory and adaptive control, allowing UAVs to proactively adjust trajectories and offloading strategies to mitigate the impact of sudden channel fluctuations.

4.3. Hierarchical Reinforcement Learning Framework

By combining dynamic graph information with LSTM-predicted channels, we obtain a rich spatio-temporal representation, which enables the agent to make proactive decisions rather than purely reactive ones.
Based on this spatio-temporal input, the joint UAV trajectory and computation offloading problem is formulated as a Markov Decision Process (MDP) ( S , A , P , R , γ ) . The system state at time slot n is defined as
s [ n ] = { q u [ n ] , v u [ n ] , h ^ u , m [ n ] , E rem , u [ n ] , D m [ n ] , η m , u [ n 1 ] , f m , u UAV [ n 1 ] , p m [ n 1 ] } ,
where h ^ u , m [ n ] incorporates the LSTM-predicted channel gains. This state captures UAV positions, velocities, energy levels, task requirements, and prior resource allocation decisions, integrating both spatial and temporal information. Correspondingly, the action vector is defined as
a [ n ] = { Δ q u [ n ] , η m , u [ n ] , f m , u UAV [ n ] , p m [ n ] } ,
where the macro-scale component Δ q u [ n ] governs UAV trajectory adjustments, while micro-scale actions η m , u [ n ] , f m , u UAV [ n ] , p m [ n ] control task offloading, computation resource allocation, and transmit power.
The reward function is designed to balance latency and energy consumption:
r [ n ] = m T ^ m [ n ] + λ E E ^ m [ n ] ,
where T ^ m [ n ] and E ^ m [ n ] denote latency and energy consumption for UE m, and λ E is a weighting factor adjusting the trade-off between energy efficiency and latency reduction. By maximizing cumulative rewards, the agent implicitly seeks to optimize both trajectory planning and offloading decisions over the mission horizon.
To address the hierarchical structure of decision-making, we adopt a two-level reinforcement learning (RL) architecture. At the high level, the HL actor–critic agent observes aggregated state information, including UAV positions, velocities, predicted channel states, and residual energy:
s HL [ n ] = { q u [ n ] , v u [ n ] , h ^ u , m [ n ] , E rem , u [ n ] } ,
and generates coarse trajectory adjustments a HL [ n ] = Δ q u [ n ] . The HL agent is designed to capture long-term strategic behaviors by accumulating low-level rewards over a macro horizon K H , and updates its policy using an actor–critic formulation:
V HL ( s HL [ n ] ) = E k = 0 γ k r HL [ n + k ] |   s HL [ n ] ,
θ HL θ HL + α HL θ HL E [ A HL ( s HL [ n ] , a HL [ n ] ) ] ,
A HL ( s , a ) = Q HL ( s , a ) V HL ( s ) ,
r HL [ n ] = k = 0 K H γ k r LL [ n + k ] .
This hierarchical design enables the HL agent to effectively capture global trajectory strategies and long-term objectives, while coordinating with the low-level agent for fine-grained control over task offloading, resource allocation, and UAV mobility.
At the low level, the LL agent operates at a finer temporal resolution, observing detailed system states including UAV positions, predicted channel states, task sizes, and prior resource allocations:
s LL [ n ] = { q u [ n ] , h ^ u , m [ n ] , D m [ n ] , η m , u [ n 1 ] , f m , u UAV [ n 1 ] , p m [ n 1 ] } ,
and produces micro-scale actions:
a LL [ n ] = { η m , u [ n ] , f m , u UAV [ n ] , p m [ n ] } .
The LL agent is trained via Double Deep Q-Network (DDQN) to ensure stable and robust learning, updating its Q-function as follows:
Q ( s LL [ n ] , a LL [ n ] ; ϕ ) = E t = n N γ t n r LL [ t ] | s LL [ n ] , a LL [ n ] , ϕ ϕ α LL ϕ ( r LL [ n ] + γ max a Q ( s LL [ n + 1 ] , a ; ϕ )
Q ( s LL [ n ] , a LL [ n ] ; ϕ ) ) 2 .
The hierarchical framework operates through iterative interaction with a simulated environment that captures dynamic UAV network behaviors, including task arrivals, UAV mobility, and time-varying air-to-ground channels. At each time slot n, the high-level agent first observes the coarse state s HL [ n ] and outputs a trajectory adjustment Δ q u [ n ] , which defines a feasible spatial region constraining the low-level action space and ensuring safety and energy feasibility. Subsequently, the low-level agent observes the fine-grained state s LL [ n ] , encompassing UAV positions, predicted channel gains h ^ u , m [ n ] , pending task sizes D m [ n ] , and previous allocation decisions η m , u [ n 1 ] , f m , u UAV [ n 1 ] , p m [ n 1 ] . Utilizing the learned Q-function Q ( s LL [ n ] , a LL [ n ] ; ϕ ) , the LL agent determines the optimal micro-actions:
a LL [ n ] = arg   max a LL Q ( s LL [ n ] , a LL ; ϕ ) ,
which jointly specify the task offloading ratio η m , u [ n ] , the allocated UAV computation resources f m , u UAV [ n ] , and the transmission power p m [ n ] . These actions drive the environment to transition to the next state s [ n + 1 ] , yielding an immediate low-level reward r LL [ n ] . Aggregating these rewards over the macro horizon K H provides long-term feedback to update the high-level agent, enabling coordinated trajectory optimization while ensuring responsive micro-level control over computation and communication resources.
To stabilize training, both HL and LL agents maintain experience replay buffers. For the LL DDQN, mini-batches of transitions ( s LL [ n ] , a LL [ n ] , r LL [ n ] , s LL [ n + 1 ] ) are sampled to update the parameters ϕ via gradient descent. The HL actor–critic network updates its policy parameters θ HL using the aggregated rewards r HL [ n ] , ensuring that long-term objectives such as energy efficiency and mission completion are considered. During each iteration, the HL policy guides the feasible region of UAV trajectories, and the LL policy selects micro-scale actions within this region, creating a feedback loop where LL performance informs HL policy improvement, and HL guidance constrains LL exploration.
Formally, the training and decision-making process can be summarized as a sequential procedure: at each time slot, the HL agent observes s HL [ n ] and produces Δ q u [ n ] according to its policy π HL ( s HL [ n ] ; θ HL ) , defining the permissible movement for each UAV. Within this feasible region, the LL agent observes s LL [ n ] and computes the optimal action a LL [ n ] that maximizes its Q-function. The joint action ( Δ q u [ n ] , a LL [ n ] ) is then executed, the environment transitions to s [ n + 1 ] , and the low-level reward r LL [ n ] is obtained. Both agents update their respective networks using their experience buffers and reward signals, and the process continues iteratively until the cumulative reward converges.
Upon convergence, the learned hierarchical policy generates a sequence of coordinated outputs over the entire mission horizon:
{ q u [ n ] , η m , u [ n ] , f m , u UAV [ n ] , p m [ n ] } n = 1 N ,
which represent optimized UAV trajectories, task offloading ratios, computation resource allocations, and transmit powers for each UAV and UE. This framework allows the system to adaptively balance latency and energy consumption while responding to dynamic network topologies and predicted channel variations. By embedding predicted channel states into both HL and LL agents, the hierarchical reinforcement learning approach effectively anticipates future network conditions, enabling proactive and coordinated optimization of UAV trajectory and computation offloading. The pseudo-code implementation is presented in Algorithm 1.
Algorithm 1: Hierarchical UAV Trajectory and Computation Offloading via HRL
Drones 10 00021 i001
The computational complexity of the proposed AUEIF framework mainly stems from hierarchical policy inference, dynamic graph processing, and LSTM-based channel prediction. The high-level (HL) policy operates on a coarse time scale and performs a single forward inference to generate macro-level trajectory adjustments, resulting in constant per-decision complexity. The low-level (LL) DDQN policy executes fine-grained task offloading and resource allocation, with computational complexity scaling linearly with the number of UAV–UE associations, i.e., O ( U · M ) . The dynamic graph construction and LSTM-based channel prediction introduce only fixed or near-linear overhead due to sparse connectivity and short temporal observation windows. During online deployment, AUEIF relies exclusively on forward inference without iterative optimization, leading to millisecond-level per-slot execution time, which satisfies the real-time requirements of UAV-assisted MEC systems.

5. Experiment Setting

5.1. System Parameters

We consider a multi-UAV-assisted MEC network comprising U UAVs and M UEs, where the numbers of UAVs and UEs vary within the ranges U [ 2 , 6 ] and M [ 5 , 20 ] , respectively. The network operates in a three-dimensional airspace with dimensions of 1000– 1500   m × 1000 1500   m × 400 600   m . UEs are assumed to follow random mobility patterns with velocities uniformly distributed in the range of 1 to 5   m / s . In each time slot, UEs generate computational tasks with heterogeneous sizes D m [ n ] [ 5 × 10 6 , 2 × 10 7 ] bits, thereby capturing realistic variations in workload demand.
UAVs have a maximum flight speed v max [ 15 , 25 ]   m / s and a maximum heading angle θ max [ π / 6 , π / 3 ] rad, while their physical properties such as mass M u and hovering power p hov can vary in the ranges M u [ 4 , 6 ] kg and p hov [ 120 , 180 ] W, respectively. The onboard energy capacity is configurable within E max [ 4000 , 6000 ] J, and the total CPU computing power is set to f u UAV [ 0.8 , 1.2 ] × 10 9 cycles per second.
For the communication subsystem, the air-to-ground wireless links are modeled by jointly considering large-scale path loss and small-scale Rayleigh fading. The reference channel gain β 0 is set within the range of 35 to 25 dBm, with the path loss exponent α [ 2.5 , 3.0 ] and the noise power spectral density σ 2 [ 180 , 170 ] dBm/Hz. The total available system bandwidth is configured as W [ 15 , 25 ] MHz and is evenly partitioned into K [ 8 , 12 ] orthogonal subchannels. Each UE is subject to a maximum transmit power constraint P m , max [ 15 , 25 ] dBm. To emulate realistic time-varying wireless environments, the small-scale fading coefficients evolve dynamically across time slots with a variation rate in the range of [ 0.1 , 0.3 ] . Regarding the computation model, the CPU frequencies of UEs are set within f m UE [ 1.5 , 2.5 ] × 10 8 cycles/s, and the required number of CPU cycles per processed bit is C m [ 800 , 1200 ] . The energy consumption for computation follows a standard dynamic power model, where the energy coefficients for UEs and UAVs are given by k m [ 0.8 , 1.2 ] × 10 28 and k u [ 4 , 6 ] × 10 29 J/(cycle2), respectively.
Flight constraints for the UAVs are specified with minimum and maximum altitudes H min [ 40 , 60 ] m and H max [ 140 , 160 ] m, a minimum inter-UAV separation L col [ 40 , 60 ] m, and a minimum distance to obstacles L safe [ 25 , 35 ] m. Obstacles are modeled as spheres with radii R k [ 15 , 25 ] m, and the number of obstacles can be varied between one and five to study different levels of environmental complexity.
The total task execution horizon T is set within the range of [ 80 , 120 ] s and is discretized into N [ 40 , 60 ] equal time slots, resulting in a slot duration of δ t = T / N . The initial system workload utilization factor η sys is configured within [ 0.3 , 0.9 ] to reflect different traffic intensities. In the weighted optimization objective, the latency-related weight ω 0 is selected from [ 0.5 , 0.7 ] , while the corresponding energy weight is defined as ω E = 1 ω 0 , enabling a balanced trade-off between delay and energy consumption. For channel state prediction, the LSTM-based module leverages a sliding historical observation window of τ [ 8 , 12 ] time slots and produces channel forecasts over a prediction horizon of Δ [ 2 , 5 ] slots, thereby facilitating proactive adaptation to channel dynamics.

5.2. Baseline Methods and Evaluation Metrics

To evaluate the proposed AUEIF framework, we consider the following baselines:
  • Random Offloading (RO): UAVs randomly decide task offloading ratios to edge servers and cloud, without trajectory optimization or queue awareness.
  • Greedy Trajectory (GT): UAVs always move toward the UE with the largest task queue, without considering channel dynamics or offloading optimization.
  • Equal Resource Allocation (ERA): Computation and communication resources are evenly distributed among all UEs and UAVs, without adaptive optimization.
  • Time-Driven DRL (TDRL): Deep reinforcement learning algorithm optimizing UAV trajectories, offloading decisions, and transmission power based on time-driven strategy, without task-driven reliability awareness or hybrid action representation.
  • Alternating Optimization (AO): Joint optimization of partial offloading, UAV trajectory, user scheduling, edge-cloud computation, and resource allocation using alternating optimization and SCA techniques, without online or task-driven adaptation.
  • PLOT without Lyapunov Perturbation (PLOT-NS): Online UAV-MEC algorithm using standard Lyapunov optimization without perturbed control, which may suffer from coupling effects and suboptimal queue stability.

6. Experiment Result

6.1. Convergence Performance Analysis

Figure 2 compares the cumulative reward trajectories of the proposed AUEIF framework with four representative DRL baselines, namely DQN, PPO, TDRL, and A3C. Starting from the initial exploration stage, AUEIF demonstrates consistently faster convergence and higher steady-state rewards than all competing methods. Throughout the training process, AUEIF achieves an average cumulative reward improvement of approximately 15–20% over the strongest baseline. In addition, the reward curve of AUEIF exhibits noticeably reduced variance and smoother convergence behavior, indicating improved training stability in dynamic UAV-assisted MEC environments. These results suggest that the proposed framework more effectively captures the spatio-temporal coupling between UAV mobility and computation offloading decisions, thereby enabling more efficient policy learning and superior long-term performance compared with existing DRL-based approaches.
Figure 3 illustrates the convergence of the AUEIF framework under different learning rates. All curves start from the initial exploration phase, showing that the learning rate significantly affects both convergence speed and stability. Higher learning rates enable faster initial improvement but exhibit larger fluctuations, whereas moderate learning rates achieve a balanced trade-off between convergence speed and smoothness. Lower learning rates converge more gradually and steadily, maintaining stable performance throughout training. Overall, the proposed framework demonstrates substantial performance gains, achieving approximately fifteen to twenty percent higher cumulative reward than the least effective configuration. These results indicate that careful selection of the learning rate is critical for efficient and stable policy learning in dynamic UAV-assisted MEC environments.

6.2. Differential Performance Analysis Under Varying UE Density

Figure 4 depicts the average task completion delay as the number of UEs increases from light-load to high-density scenarios. As expected, all baseline methods exhibit a monotonic increase in delay due to aggravated channel contention and heightened computational demand. In contrast, AUEIF consistently achieves the lowest delay across the entire range of UE densities. Specifically, compared with heuristic baselines such as RO, GT, and ERA, AUEIF reduces the average delay by more than 40–60%, primarily because these methods lack adaptive trajectory control and coordinated computation allocation mechanisms. Furthermore, even when compared with advanced optimization-based approaches, including TDRL, AO, and PLOT-NS, the AUEIF still attains delay reductions of approximately 20–35%. This performance gain stems from its unified optimization of UAV trajectory planning, task offloading decisions, and resource scheduling. Notably, under heavy-load conditions, the delay curves of baseline methods exhibit sharp growth, whereas AUEIF demonstrates a markedly flatter increase, indicating superior scalability and robustness in high-density UAV-MEC scenarios.
Figure 5 compares the total system energy consumption of AUEIF with that of the benchmark methods. AUEIF consistently exhibits the lowest energy consumption, highlighting its effectiveness in energy-aware operation under dynamic air–ground channel conditions. Heuristic baselines such as RO, GT, and ERA incur more than 50–70% higher energy expenditure due to inefficient UAV trajectories, redundant task processing, and non-adaptive resource allocation. More advanced schemes, including TDRL, AO, and PLOT-NS, achieve partial energy savings; however, they still consume approximately 20–40% more energy than AUEIF, as they do not fully coordinate mobility control, communication, and computation decisions in an online manner. The superior energy efficiency of AUEIF is attributed to the synergistic integration of three mechanisms: energy-aware trajectory optimization that avoids unnecessary UAV maneuvers, task-driven computation offloading that reduces superfluous processing, and adaptive resource allocation that mitigates prolonged high-frequency operation. Overall, AUEIF reduces total energy consumption by over 30% relative to the highest-energy-consuming baselines, demonstrating its ability to achieve energy-efficient UAV-MEC operation without sacrificing task performance.
Figure 6 depicts the three-dimensional Pareto frontier among average task completion delay, total system energy consumption, and system load for all evaluated methods. AUEIF consistently lies on the Pareto-optimal surface, indicating its superior ability to jointly balance latency performance and energy efficiency across different load regimes. In contrast, heuristic baselines such as RO, GT, and ERA are located in dominated regions characterized by simultaneously high delay and excessive energy consumption. Under heavy-load conditions, the AUEIF reduces the average task completion delay by approximately 25% while achieving an energy saving of about 32% relative to these baselines. Even when compared with more advanced schemes, including TDRL, AO, and PLOT-NS, the AUEIF maintains a more favorable trade-off with lower latency and reduced energy usage. These results confirm that the integrated design of trajectory optimization, adaptive computation offloading, and resource allocation in the AUEIF enables more effective multi-objective optimization in UAV-assisted MEC systems.

6.3. Differential Performance Analysis Under Varying UAV Density

Figure 7 illustrates the variation in average task completion delay as the number of UAVs increases, where the horizontal axis denotes UAV density and the vertical axis represents the corresponding task completion delay. Each curve corresponds to a different scheme. As the UAV density grows, all methods benefit from the increased availability of computing and communication resources, leading to an overall reduction in delay. However, the AUEIF consistently achieves the lowest task completion delay across all UAV densities. In particular, the AUEIF reduces the average delay by approximately 25–40% compared with advanced baselines and by up to 50–60% relative to heuristic methods. This performance advantage is attributed to the AUEIF’s tightly coupled optimization of UAV trajectory planning, task offloading decisions, and resource allocation, which enables smooth delay reduction and robust adaptation to varying system capacities.
Figure 8 illustrates the variation in total system energy consumption as the number of UAVs increases for different methods. The horizontal axis represents the UAV population, while the vertical axis denotes the total system energy consumption. Each curve corresponds to a specific scheme. As the number of UAVs grows, the overall energy consumption increases for all methods due to higher coordination, communication, and mobility overhead. Nevertheless, the AUEIF consistently achieves the lowest energy consumption across all UAV densities, demonstrating its superior energy efficiency. In contrast, heuristic baselines such as RO, GT, and ERA exhibit the steepest growth in energy usage, consuming approximately 50–70% more energy than the AUEIF under high UAV density. More advanced approaches, including TDRL, AO, and PLOT-NS, provide moderate improvements over the heuristic schemes but still incur about 15–30% higher energy consumption compared with the AUEIF. These results highlight the effectiveness of AUEIF in jointly optimizing UAV trajectories, task-aware computation offloading, and adaptive resource allocation to achieve energy-efficient UAV-MEC operation.
Figure 9 depicts the three-dimensional Pareto frontier among average task completion delay, total system energy consumption, and UAV density for the evaluated methods. The horizontal axis indicates the number of deployed UAVs, the vertical axis represents the average task completion delay, and the depth axis corresponds to the total system energy consumption. Each curve denotes a specific method, with the AUEIF highlighted to emphasize its performance. As UAV density increases, both delay and energy consumption generally exhibit an upward trend, reflecting the additional communication, coordination, and mobility overhead introduced by denser UAV deployments. Notably, the AUEIF consistently occupies the region characterized by lower delay and lower energy consumption in the three-dimensional performance space, indicating a more favorable trade-off between computational efficiency and energy utilization. In contrast, heuristic baselines such as RO, GT, and ERA incur persistently high delay and energy costs across all UAV densities. More advanced schemes, including TDRL, AO, and PLOT-NS, achieve moderate improvements over the heuristic methods but still remain dominated by the AUEIF in terms of both latency and energy efficiency.

6.4. Ablation Experiment Analysis

Figure 10 presents the ablation study of the AUEIF under varying system scenarios, where the horizontal axis represents different combinations of UE numbers and UAV numbers, reflecting system load and available computing resources. The vertical axis denotes normalized performance in percentage, measured in terms of task success rate or system efficiency, where higher values indicate better performance. Each scenario includes results for four methods: Full AUEIF, representing the complete model with all modules; w/o Graph, which removes the dynamic graph module and thus cannot capture UAV network topology changes; w/o LSTM, which eliminates the LSTM-based channel prediction, reducing the ability to anticipate channel fluctuations; and w/o HRL, which discards the hierarchical reinforcement learning structure, leaving a single-layer policy that must handle the entire high-dimensional joint action space.
The results clearly demonstrate the contribution of each core component. Removing the dynamic graph module leads to a noticeable performance drop, particularly under high-load scenarios, indicating the importance of capturing inter-UAV collaboration. Eliminating the LSTM prediction module further degrades performance, highlighting its role in improving offloading decision accuracy. The most pronounced degradation occurs when the hierarchical reinforcement learning structure is removed, reflecting slower convergence and reduced overall efficiency. Across all scenarios, Full AUEIF consistently outperforms its ablated variants, and the relative performance decline of the ablated models ranges from approximately 15% to 25%, demonstrating that each component significantly contributes to robust and efficient UAV-MEC operations.

6.5. Robustness of AUEIF Under Varying Channel Fading Conditions

Figure 11 illustrates the distribution of task success rates achieved by the AUEIF and six baseline methods under four channel fading conditions, namely Low, Moderate, High, and Severe. The horizontal axis represents the fading severity, while the vertical axis denotes the overall task success rate, ranging from 0.3 to 1.0. For each fading level, boxplots corresponding to different algorithms are horizontally offset for clarity. Each box indicates the interquartile range (IQR), the central line denotes the median over multiple runs, and the whiskers represent the extrema within 1.5 × IQR. The median performance of the AUEIF is highlighted for visual emphasis.
Across all fading conditions, the AUEIF consistently achieves the highest median task success rate with the lowest performance variance, demonstrating strong robustness against channel degradation. Under Low and Moderate fading, the median success rate of the AUEIF remains above 0.94 , substantially outperforming all baseline approaches. As channel fading intensifies to High and Severe levels, the performance of all methods degrades due to increased signal attenuation and heightened air-to-ground channel instability. Nevertheless, the AUEIF maintains a median success rate exceeding 0.92 , indicating stable and reliable decision-making under adverse channel conditions. In contrast, learning-based baselines such as TDRL, AO, and PLOT-NS experience more pronounced performance degradation, with median success rates declining to the 0.72 0.80 range under Severe fading. Heuristic-based strategies (RO, GT, and ERA) exhibit the most severe degradation, with median performance dropping to the 0.50 0.60 range with substantially increased variability.
The superior robustness of the AUEIF can be attributed to two key design features. First, the LSTM-based temporal channel prediction module enables the framework to anticipate fading trends, allowing proactive adjustments in task offloading decisions and UAV trajectory planning before significant link deterioration occurs. Second, the unified hybrid action space facilitates fine-grained control over partial offloading ratios and transmission strategies, thereby stabilizing queue dynamics even under highly volatile channel conditions. By comparison, RO, GT, and ERA lack adaptive mechanisms, TDRL optimizes policies without explicit task reliability awareness, and PLOT-NS is more susceptible to queue–channel coupling effects when perturbed control mechanisms are absent.

7. Conclusions

This paper proposed an AUEIF framework for multi-UAV-assisted mobile edge computing networks to jointly address UAV trajectory planning, computation offloading, and time-varying air–ground channels. By integrating dynamic graph modeling, hierarchical reinforcement learning, and LSTM-based channel prediction, the AUEIF captures spatio-temporal correlations and enables coordinated optimization of mobility, communication, and computation. Simulation results demonstrate that the AUEIF consistently outperforms baseline methods in terms of task completion delay, energy efficiency, and robustness under diverse system loads and channel fading conditions, while ablation studies confirm the necessity of each core component.
Nevertheless, several limitations remain. The current evaluation relies on simulated mobility and channel models, which may not fully reflect complex urban environments. Moreover, the hierarchical learning framework may introduce additional computational overhead as system scale increases, and inaccurate channel prediction under highly unpredictable conditions could affect decision optimality. Future work will explore decentralized multi-agent extensions, more realistic environment modeling, and integration with emerging 5G/6G network slicing mechanisms to facilitate practical deployment.

Author Contributions

Conceptualization, J.X.; Methodology, J.X.; Software, J.X.; Validation, J.X.; Formal Analysis, D.X.; Investigation, D.X.; Data Curation, D.X.; Writing—Original Draft Preparation, D.X.; Visualization, D.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, F.; Luo, J.; Qiao, Y.; Li, Y. Joint UAV Deployment and Task Offloading Scheme for Multi-UAV-Assisted Edge Computing. Drones 2023, 7, 284. [Google Scholar] [CrossRef]
  2. Gao, C.; Wei, D.; Li, K.; Liu, W. UAV-Centric Privacy-Preserving Computation Offloading in Multi-UAV Mobile Edge Computing. Drones 2025, 9, 701. [Google Scholar] [CrossRef]
  3. Wei, J.; Guo, Y.; Wang, H.; Gu, J.; Liu, J.; Ding, G. Cognitive Jamming-aided UAV Multi-user Covert Communication. IEEE J. Sel. Areas Commun. 2025; early access. [Google Scholar] [CrossRef]
  4. Zhang, K.; Gui, X.; Ren, D.; Li, D. Energy–Latency Tradeoff for Computation Offloading in UAV-Assisted Multiaccess Edge Computing System. IEEE Internet Things J. 2021, 8, 6709–6719. [Google Scholar]
  5. Yang, X.; Feng, J.; Liu, L.; Pei, Q. Optimizing Resource Utilization in Consumer Electronics Networks Through an Enhanced Grey Wolf Optimization Algorithm With UAV Collaboration. IEEE Trans. Consum. Electron. 2025, 71, 7376–7386. [Google Scholar] [CrossRef]
  6. Yang, Z.; Bi, S.; Zhang, Y.-J.A. Dynamic Offloading and Trajectory Control for UAV-Enabled Mobile Edge Computing System With Energy Harvesting Devices. IEEE Trans. Wirel. Commun. 2022, 21, 10515–10528. [Google Scholar] [CrossRef]
  7. Hao, H.; Xu, C.; Zhang, W.; Chen, X.; Yang, S.; Muntean, G.-M. Reliability-Aware Optimization of Task Offloading for UAV-Assisted Edge Computing. IEEE Trans. Comput. 2025, 74, 3832–3844. [Google Scholar] [CrossRef]
  8. Nguyen, M.D.; Le, L.B.; Girard, A. Integrated Computation Offloading, UAV Trajectory Control, Edge-Cloud and Radio Resource Allocation in SAGIN. IEEE Trans. Cloud Comput. 2024, 12, 100–115. [Google Scholar] [CrossRef]
  9. Zhang, J.; Luo, H.; Chen, X.; Shen, H.; Guo, L. Minimizing Response Delay in UAV-Assisted Mobile Edge Computing by Joint UAV Deployment and Computation Offloading. IEEE Trans. Cloud Comput. 2024, 12, 1372–1386. [Google Scholar] [CrossRef]
  10. Xu, Y.; Zhang, T.; Liu, Y.; Yang, D.; Xiao, L.; Tao, M. UAV-Assisted MEC Networks with Aerial and Ground Cooperation. IEEE Trans. Wirel. Commun. 2021, 20, 7712–7727. [Google Scholar] [CrossRef]
  11. Zhao, N.; Ye, Z.; Pei, Y.; Liang, Y.-C.; Niyato, D. Multi-Agent Deep Reinforcement Learning for Task Offloading in UAV-Assisted Mobile Edge Computing. IEEE Trans. Wirel. Commun. 2022, 21, 6949–6960. [Google Scholar] [CrossRef]
  12. Chen, Y.; Pi, D.; Yang, S.; Xu, Y.; Chen, J.; Mohamed, A.W. HNIO: A Hybrid Nature-Inspired Optimization Algorithm for Energy Minimization in UAV-Assisted Mobile Edge Computing. IEEE Trans. Netw. Serv. Manag. 2022, 19, 3264–3275. [Google Scholar] [CrossRef]
  13. Al-Bakhrani, A.A.; Li, M.; Obaidat, M.S.; Amran, G.A. MOALF-UAV-MEC: Adaptive Multiobjective Optimization for UAV-Assisted Mobile Edge Computing in Dynamic IoT Environments. IEEE Internet Things J. 2025, 12, 20736–20756. [Google Scholar] [CrossRef]
  14. Song, F.; Xing, H.; Wang, X.; Luo, S.; Dai, P.; Xiao, Z.; Zhao, B. Evolutionary Multi-Objective Reinforcement Learning Based Trajectory Control and Task Offloading in UAV-Assisted Mobile Edge Computing. IEEE Trans. Mob. Comput. 2023, 22, 7387–7405. [Google Scholar] [CrossRef]
  15. Zhou, L.; Mao, H.; Deng, X.; Zhang, J.; Zhao, H.; Wei, J. Real-Time Radio Map Construction and Distribution for UAV-Assisted Mobile Edge Computing Networks. IEEE Internet Things J. 2024, 11, 21337–21346. [Google Scholar] [CrossRef]
  16. Chen, Y.; Pi, D.; Yang, S.; Xu, Y.; Wang, B.; Qin, S.; Wang, Y. A Dynamic Optimization Framework for Computation Rate Maximization in UAV-Assisted Mobile Edge Computing. IEEE Trans. Veh. Technol. 2025, 74, 11395–11409. [Google Scholar] [CrossRef]
  17. Zhou, R.; Huang, Y.; Wang, Y.; Jiao, L.; Tan, H.; Zhang, R.; Wu, L. User Preference Oriented Service Caching and Task Offloading for UAV-Assisted MEC Networks. IEEE Trans. Serv. Comput. 2025, 18, 1097–1109. [Google Scholar] [CrossRef]
  18. Zhang, T.; Xu, Y.; Loo, J.; Yang, D.; Xiao, L. Joint Computation and Communication Design for UAV-Assisted Mobile Edge Computing in IoT. IEEE Trans. Ind. Inform. 2020, 16, 5505–5516. [Google Scholar] [CrossRef]
  19. Liu, B.; Wan, Y.; Zhou, F.; Wu, Q.; Hu, R.Q. Resource Allocation and Trajectory Design for MISO UAV-Assisted MEC Networks. IEEE Trans. Veh. Technol. 2022, 71, 4933–4948. [Google Scholar] [CrossRef]
  20. Hui, M.; Chen, J.; Yang, L.; Lv, L.; Jiang, H.; Al-Dhahir, N. UAV-Assisted Mobile Edge Computing: Optimal Design of UAV Altitude and Task Offloading. IEEE Trans. Wirel. Commun. 2024, 23, 13633–13647. [Google Scholar] [CrossRef]
  21. Sun, G.; Wang, Y.; Sun, Z.; Wu, Q.; Kang, J.; Niyato, D.; Leung, V.C. Multi-Objective Optimization for Multi-UAV-Assisted Mobile Edge Computing. IEEE Trans. Mob. Comput. 2024, 23, 14803–14820. [Google Scholar] [CrossRef]
  22. Kwon, D.; Son, S.; Kim, M.; Lee, J.; Das, A.K.; Park, Y. A Secure Self-Certified Broadcast Authentication Protocol for Intelligent Transportation Systems in UAV-Assisted Mobile Edge Computing Environments. IEEE Trans. Intell. Transp. Syst. 2024, 25, 19004–19017. [Google Scholar] [CrossRef]
  23. Xu, Y.; Zhang, T.; Yang, D.; Liu, Y.; Tao, M. Joint Resource and Trajectory Optimization for Security in UAV-Assisted MEC Systems. IEEE Trans. Commun. 2021, 69, 573–588. [Google Scholar] [CrossRef]
  24. Shao, Z.; Yang, H.; Xiao, L.; Su, W.; Chen, Y.; Xiong, Z. Deep Reinforcement Learning-Based Resource Management for UAV-Assisted Mobile Edge Computing Against Jamming. IEEE Trans. Mob. Comput. 2024, 23, 13358–13374. [Google Scholar] [CrossRef]
  25. Duan, X.; Zhou, Y.; Tian, D.; Zhou, J.; Sheng, Z.; Shen, X. Weighted Energy-Efficiency Maximization for a UAV-Assisted Multiplatoon Mobile-Edge Computing System. IEEE Internet Things J. 2022, 9, 18208–18220. [Google Scholar] [CrossRef]
  26. Zhao, M.; Wang, Z.; Guo, K.; Zhang, R.; Quek, T.Q.S. Against Mobile Collusive Eavesdroppers: Cooperative Secure Transmission and Computation in UAV-Assisted MEC Networks. IEEE Trans. Mob. Comput. 2025, 24, 5280–5297. [Google Scholar] [CrossRef]
  27. Apostolopoulos, P.A.; Fragkos, G.; Tsiropoulou, E.E.; Papavassiliou, S. Data Offloading in UAV-Assisted Multi-Access Edge Computing Systems Under Resource Uncertainty. IEEE Trans. Mob. Comput. 2023, 22, 175–190. [Google Scholar] [CrossRef]
  28. Liu, Y.; Zhou, J.; Tian, D.; Sheng, Z.; Duan, X.; Qu, G.; Leung, V.C. Joint Communication and Computation Resource Scheduling of a UAV-Assisted Mobile Edge Computing System for Platooning Vehicles. IEEE Trans. Intell. Transp. Syst. 2022, 23, 8435–8450. [Google Scholar] [CrossRef]
  29. Wang, Y.; Zhu, J.; Huang, H.; Xiao, F. Bi-Objective Ant Colony Optimization for Trajectory Planning and Task Offloading in UAV-Assisted MEC Systems. IEEE Trans. Mob. Comput. 2024, 23, 12360–12377. [Google Scholar] [CrossRef]
  30. Sun, Z.; Sun, G.; Wu, Q.; He, L.; Liang, S.; Pan, H. TJCCT: A Two-Timescale Approach for UAV-Assisted Mobile Edge Computing. IEEE Trans. Mob. Comput. 2025, 24, 3130–3147. [Google Scholar] [CrossRef]
  31. Miao, Y.; Hwang, K.; Wu, D.; Hao, Y.; Chen, M. Drone Swarm Path Planning for Mobile Edge Computing in Industrial Internet of Things. IEEE Trans. Ind. Inform. 2023, 19, 6836–6848. [Google Scholar] [CrossRef]
  32. Li, Z.; Liang, X.; Liu, J.; He, X.; Xie, L.; Qu, L.; Feng, G. Optimizing Mobile-Edge Computing for Virtual Reality Rendering via UAVs: A Multiagent Deep Reinforcement Learning Approach. IEEE Internet Things J. 2025, 12, 35756–35772. [Google Scholar] [CrossRef]
  33. Chen, Z.; Yang, Y.; Xu, J.; Chen, Y.; Huang, J. Task Offloading and Resource Pricing Based on Game Theory in UAV-Assisted Edge Computing. IEEE Trans. Serv. Comput. 2025, 18, 440–452. [Google Scholar] [CrossRef]
  34. Peng, S.; Li, B.; Liu, L.; Fei, Z.; Niyato, D. Trajectory Design and Resource Allocation for Multi-UAV-Assisted Sensing, Communication, and Edge Computing Integration. IEEE Trans. Commun. 2025, 73, 2847–2861. [Google Scholar] [CrossRef]
Figure 1. System model of a multi-UAV-assisted mobile edge computing network.
Figure 1. System model of a multi-UAV-assisted mobile edge computing network.
Drones 10 00021 g001
Figure 2. Reward comparison of AUEIF and baseline DRL methods.
Figure 2. Reward comparison of AUEIF and baseline DRL methods.
Drones 10 00021 g002
Figure 3. Convergence of AUEIF under different learning rates.
Figure 3. Convergence of AUEIF under different learning rates.
Drones 10 00021 g003
Figure 4. Average task completion delay under varying UE density.
Figure 4. Average task completion delay under varying UE density.
Drones 10 00021 g004
Figure 5. Total system energy consumption under varying UE density.
Figure 5. Total system energy consumption under varying UE density.
Drones 10 00021 g005
Figure 6. Three-dimensional Pareto frontier of task completion delay and system energy under varying system load.
Figure 6. Three-dimensional Pareto frontier of task completion delay and system energy under varying system load.
Drones 10 00021 g006
Figure 7. Average task completion delay under varying UAV density.
Figure 7. Average task completion delay under varying UAV density.
Drones 10 00021 g007
Figure 8. Total system energy consumption under varying UAV density.
Figure 8. Total system energy consumption under varying UAV density.
Drones 10 00021 g008
Figure 9. Three-dimensional Pareto frontier of task completion delay and system energy under varying UAV density.
Figure 9. Three-dimensional Pareto frontier of task completion delay and system energy under varying UAV density.
Drones 10 00021 g009
Figure 10. Ablation study of AUEIF components under varying UE–UAV system configurations.
Figure 10. Ablation study of AUEIF components under varying UE–UAV system configurations.
Drones 10 00021 g010
Figure 11. Task success rate distribution under different channel fading conditions.
Figure 11. Task success rate distribution under different channel fading conditions.
Drones 10 00021 g011
Table 1. Key notations in multi-UAV MEC system.
Table 1. Key notations in multi-UAV MEC system.
SymbolDescriptionSymbolDescription
UNumber of UAVsMNumber of UEs
TTotal task period (s)NNumber of time slots
δ t Duration of each slot (s) q u ( n ) UAV u 3D position at slot n
H u ( n ) UAV altitude (m) x u ( n ) , y u ( n ) UAV horizontal coordinates (m)
o k Obstacle k center (m) R k Obstacle radius (m)
L safe Min distance to obstacle (m) L col Min inter-UAV distance (m)
v u ( n ) UAV velocity (m/s) v max Max UAV speed (m/s)
θ u ( n ) UAV heading angle (rad) θ max Max heading angle (rad)
δ t fly Flying duration (s) δ t hov Hovering duration (s)
M u ( n ) UEs served by UAV u η m , u ( n ) Offloading fraction of UE m
D m ( n ) Task data size (bits) R u , m ( n ) Transmission rate (bits/s)
E u ( n ) UAV energy (J) E fly ( n ) Flying energy (J)
E hov ( n ) Hovering energy (J) M u UAV mass (kg)
p hov Hovering power (W) w m ( n ) UE position (m)
h u , m ( n ) Channel gain β 0 Reference channel gain
α Path loss exponent g u , m ( n ) Small-scale fading
B m , n Bandwidth (Hz) p m ( n ) UE transmit power (W)
σ 2 Noise power (W) K n Number of subchannels
U m ( n ) UAVs serving UE m f m UE UE CPU frequency
C m CPU cycles per bit k m UE computation energy coefficient
f m , u UAV ( n ) UAV computation allocation k u UAV energy coefficient
f u UAV Total UAV CPU capacity T m loc ( n ) UE local computation latency
T m , u off ( n ) Offloaded latency E m loc ( n ) UE local computation energy
E m , u off ( n ) Offloading energy T total Total system latency
E total Total system energy ϕ Weight for UAV energy
FWeighted objective ω T , ω E Latency/energy weights
ω 0 Initial latency weight η sys System workload utilization
E rem ( n ) Remaining UAV energy E u max Max UAV energy capacity
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, J.; Xie, D. Adaptive Edge Intelligent Joint Optimization of UAV Computation Offloading and Trajectory Under Time-Varying Channels. Drones 2026, 10, 21. https://doi.org/10.3390/drones10010021

AMA Style

Xie J, Xie D. Adaptive Edge Intelligent Joint Optimization of UAV Computation Offloading and Trajectory Under Time-Varying Channels. Drones. 2026; 10(1):21. https://doi.org/10.3390/drones10010021

Chicago/Turabian Style

Xie, Jinwei, and Dimin Xie. 2026. "Adaptive Edge Intelligent Joint Optimization of UAV Computation Offloading and Trajectory Under Time-Varying Channels" Drones 10, no. 1: 21. https://doi.org/10.3390/drones10010021

APA Style

Xie, J., & Xie, D. (2026). Adaptive Edge Intelligent Joint Optimization of UAV Computation Offloading and Trajectory Under Time-Varying Channels. Drones, 10(1), 21. https://doi.org/10.3390/drones10010021

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop