Next Article in Journal
MITS: A Quantum Sorcerer’s Stone for Designing Surface Codes
Previous Article in Journal
Violations of Hyperscaling in Phase Transitions and Critical Phenomena—In Memory of Prof. Ralph Kenna
Previous Article in Special Issue
Age of Information Minimization in Multicarrier-Based Wireless Powered Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Reinforcement Learning-Based Resource Allocation for UAV-GAP Downlink Cooperative NOMA in IIoT Systems

by
Yuanyan Huang
1,2,†,
Jingjing Su
1,2,†,
Xuan Lu
1,2,
Shoulin Huang
1,2,
Hongyan Zhu
1,2 and
Haiyong Zeng
1,2,*
1
Guangxi Key Laboratory of Brain-Inspired Computing and Intelligent Chips, School of Electronic and Information Engineering, Guangxi Normal University, Guilin 541004, China
2
Key Laboratory of Integrated Circuits and Microsystems, Education Department of Guangxi Zhuang Autonomous Region, Guangxi Normal University, Guilin 541004, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2025, 27(8), 811; https://doi.org/10.3390/e27080811
Submission received: 27 June 2025 / Revised: 25 July 2025 / Accepted: 26 July 2025 / Published: 29 July 2025

Abstract

This paper studies deep reinforcement learning (DRL)-based joint resource allocation and three-dimensional (3D) trajectory optimization for unmanned aerial vehicle (UAV)–ground access point (GAP) cooperative non-orthogonal multiple access (NOMA) communication in Industrial Internet of Things (IIoT) systems. Cooperative and non-cooperative users adopt different signal transmission strategies to meet diverse, task-oriented, quality-of-service requirements. Specifically, the DRL framework based on the Soft Actor–Critic algorithm is proposed to jointly optimize user scheduling, power allocation, and UAV trajectory in continuous action spaces. Closed-form power allocation and maximum weight bipartite matching are integrated to enable efficient user pairing and resource management. Simulation results show that the proposed scheme significantly enhances system performance in terms of throughput, spectral efficiency, and interference management, while enabling robustness against channel uncertainties in dynamic IIoT environments. The findings indicate that combining model-free reinforcement learning with conventional optimization provides a viable solution for adaptive resource management in dynamic UAV-GAP cooperative communication scenarios.

1. Introduction

The rapid development of the Industrial Internet of Things (IIoT) and smart manufacturing has imposed increasingly stringent requirements on wireless communication networks, particularly in industrial automation, predictive maintenance, and remote monitoring applications [1,2]. These applications demand massive device connectivity, ultra-high reliability, and extensive coverage, creating significant challenges for traditional terrestrial infrastructures. To address these challenges, unmanned aerial vehicles (UAVs) have emerged as a promising solution due to their high mobility, flexible deployment capabilities, and ability to dynamically adapt to communication demands. Serving as aerial base stations, UAVs can establish high-quality air-to-ground (A2G) line-of-sight (LoS) links [3,4], effectively complementing ground access points (GAPs) in complex industrial environments such as clean rooms, large factory sites, emergency scenarios, disaster recovery, and remote areas [5,6,7].
Early UAV-assisted communication systems primarily employed orthogonal multiple access (OMA) schemes for resource allocation, such as TDMA-based UAV scheduling [8] and hybrid cellular networks with UAVs as edge aerial base stations [7]. While OMA-based schemes ensure simple implementation and interference-free transmission, they often fall short in meeting the growing demands for spectral efficiency and massive connectivity in IIoT applications. To address these limitations, power-domain non-orthogonal multiple access (NOMA) has been introduced in UAV systems [9,10,11], enabling simultaneous multi-user transmission with improved spectral and energy efficiency. Studies have demonstrated that UAV-assisted NOMA provides superior performance than traditional TDMA schemes in terms of throughput, spectral efficiency, and coverage [12,13]. The recent literature has explored AI techniques in various NOMA paradigms, including Sparse Code Multiple Access (SCMA), which offers grant-free access capabilities for dense IIoT applications through optimized sparse codebook design [14,15]. AI algorithms have also proven effective in dynamic resource allocation for NOMA systems [16,17,18], with reinforcement learning being applied to optimize resource allocation and minimize Age of Information [19], demonstrating the versatility of AI techniques across different NOMA paradigms.
In UAV-assisted IIoT systems, the spectrum sharing between UAV and GAP introduces significant cross-tier interference, degrading communication performance. To mitigate this issue, cooperative transmission has been proposed to enable joint service by UAV and GAP [20]. Recent works have integrated cooperative transmission with NOMA and successive interference cancellation (SIC) techniques [21] and designed NOMA precoding matrices on the GAP side to effectively suppress interference while optimizing UAV trajectories [22]. However, most of the existing literature assumes exclusive user association with either UAV or GAP, limiting the exploitation of interference channels. A generalized joint user scheduling and power allocation (G-USPA) algorithm with 3D UAV trajectory design based on successive convex approximation (SCA) was proposed for UAV-assisted cooperative NOMA systems [7]. Though this approach demonstrated superior performance compared to non-cooperative systems, the SCA-based trajectory optimization incurs high computational complexity. Similarly, other related studies employing SCA methods and traditional convex optimization techniques [23,24] suffer from computational intensity and convergence challenges, limiting their adaptability in dynamic IIoT environments and making real-time deployment difficult [25,26]. Furthermore, UAVs in industrial environments are subject to limited battery capacity and computing power [27,28], necessitating more efficient and low-complexity resource management algorithms.
Deep reinforcement learning (DRL) has recently attracted significant attention for UAV decision-making and trajectory optimization in dynamic environments, thanks to its strong capabilities in feature extraction, policy learning, and adaptive decision-making under uncertainty. Unlike traditional optimization methods that require precise models and may struggle with environmental changes, DRL can learn flexible policies that adapt to varying conditions through interaction with the environment. Various DRL algorithms have been applied to this domain. For example, Deep Q-Networks (DQN) model the problem as a Markov Decision Process (MDP) to improve throughput and fairness [28], but their discretized action spaces limit precision. Continuous control methods such as Deep Deterministic Policy Gradient (DDPG) enhance accuracy in continuous domains [29], while Proximal Policy Optimization (PPO) incorporates clipped policy updates to enhance stability; it has also been successfully applied to large-scale network resource scheduling [30]. More recently, the Soft Actor–Critic (SAC) algorithm, leveraging a maximum entropy framework, offers improved convergence stability and efficiency, making it well suited for resource-constrained scenarios [31].
Based on the above analysis, there remains a need for efficient joint optimization approaches that can handle the complexity of UAV-GAP cooperative NOMA systems while ensuring real-time adaptability in dynamic IIoT environments. Unlike previous works that rely on traditional convex optimization solvers with high computational complexity [32] and poor real-time performance, this paper proposes a hybrid framework that combines traditional optimization techniques with deep reinforcement learning. This approach significantly reduces computational complexity compared to traditional convex programming methods while maintaining superior performance and real-time responsiveness suitable for complex industrial scenarios.
The main contributions of this work are summarized as follows:
  • We propose a SAC-based joint optimization framework addressing UAV 3D trajectory planning and resource allocation challenges in dense and dynamic IIoT scenarios.
  • By integrating power allocation, user scheduling, and 3D UAV trajectory design, we develop a joint resource management scheme for UAV-GAP cooperative NOMA systems that exploits interference channels to improve spectral efficiency and system throughput under stringent reliability and latency constraints.
  • Simulation results validate the proposed approach’s performance improvements in IIoT downlink scenarios, including throughput gains, energy consumption reduction, and enhanced interference management, indicating its potential applicability in industrial contexts.
The remainder of this paper is organized as follows: Section 2 introduces the system model and problem formulation; Section 3 details the joint user scheduling and power allocation approach; Section 4 presents the trajectory optimization method based on DRL and SAC; Section 5 provides simulation results and discussions; Section 6 concludes this paper and discusses future research directions.

2. System Model and Problem Formulation

2.1. System Model

As illustrated in Figure 1, we consider a 3D downlink UAV-GAP cooperative communication system employing cooperative NOMA, serving a total of K terrestrial users. The UAV-BS flies above the coverage area and coordinates with the GAP to jointly serve these users. Users are divided into cooperative and non-cooperative groups based on their service modes, where different signal transmission strategies are employed to satisfy diverse, task-oriented, QoS requirements. Specifically, cooperative users are served jointly by both UAV and GAP to guarantee their QoS, while non-cooperative users are served only by the GAP. Both sets of users receive signals from the GAP under the NOMA scheme. We denote the sets of non-cooperative and cooperative users as K c and K e , respectively.
We assume that the UAV flies periodically over the target coverage area with a cycle flight time T, which is divided into N equal time slots indexed by the set N = { 1 , 2 , , N } . The UAV’s 3D position at time slot n is denoted by ( x n , y n , z n ) , where z n represents the UAV’s altitude. The horizontal coordinates of the GAP and the i-th terrestrial user are fixed at ( x b , y b , 0 ) and ( x i , y i , 0 ) , respectively.
The channel state information (CSI) from the GAP to user i at time slot n is denoted by h i , n , which is modeled as a Rayleigh fading channel with zero mean and unit variance. In contrast, the UAV-to-terrestrial user channel follows a probabilistic A2G line-of-sight/non-line-of-sight (LoS/NLoS) [33,34] propagation model. The Euclidean distance between the UAV and user i at time slot n is given by
d i , n = ( x n x i ) 2 + ( y n y i ) 2 + z n 2 .
The corresponding elevation angle θ i , n between the UAV and user i is
θ i , n = arccos z n d i , n .
The probability of LoS in the A2G channel between the UAV and user i is given by
P LoS ( θ i , n ) = 1 1 + a e b ( θ i , n a ) ,
where a and b are propagation constants that can be adjusted to different industrial environments [35,36,37]. The NLoS probability is P NLoS ( θ i , n ) = 1 P LoS ( θ i , n ) .
Accordingly, the channel power gain between the UAV and user i at time slot n is expressed as
v i , n = β u d i , n α P LoS ( θ i , n ) + β u d i , n α P NLoS ( θ i , n ) ,
where β u and β u denote the channel power gains at the reference distance d 0 = 1 m under LoS and NLoS conditions, respectively, and α and α represent the corresponding path loss exponents.
As introduced above, the GAP adopts the NOMA strategy to serve all users. The users served cooperatively by the UAV and GAP are additionally served by the UAV via TDMA to guarantee their QoS requirements. Let U n denote the set of users served by the GAP at time slot n and denote the index of a cooperative user as l. Then, the received signal at an arbitrary user k served only by the GAP can be written as
y k , n = k U n ( h ^ k , n + e k , n ) P k , n S k , n + v k , n P u S l , n + u k , n ,
where S k , n and S l , n denote the transmitted symbols from the GAP to single-service user k and the UAV to cooperative user l, respectively, both with unit power, i.e., | S k , n | 2 = | S l , n | 2 = 1 . P k , n and P u are transmit powers allocated by the GAP to user k and by the UAV, respectively. The noise term u k , n is additive white Gaussian noise with zero mean and variance σ 2 , i.e., u k , n CN ( 0 , σ 2 ) .
Define the true and estimated channel gains for user k at time slot n as H k , n = | h k , n | 2 and H ^ k , n = | h ^ k , n | 2 , respectively. Given that h ^ k , n and e k , n are statistically independent, the expectation of the true channel gain satisfies
E [ H k , n ] = E [ H ^ k , n ] + E [ E k , n ] ,
where E [ E k , n ] = σ error 2 represents the effect of imperfect CSI estimation.
At the receiver, SIC is performed to decode the composite signals [38]. For any user k, it sequentially decodes signals of users with higher decoding priorities before decoding its own signal by treating lower-priority users’ signals as noise [9]. To reduce SIC complexity and limit error propagation, we assume that each NOMA group contains exactly two users: one cooperative user and one user served solely by the GAP. Since the cooperative user typically experiences poorer channel conditions from the GAP due to longer distance or blockage, it is treated as the weak user relative to the other user in the same NOMA group.
For the user k served only by the GAP, the interference from the UAV’s signal must be accounted for. We denote the channel power gain from the UAV to this user as V k , n = | v k , n | 2 . Then, the signal-to-interference-plus-noise ratio (SINR) at user k after SIC is given by
γ k , n = H ^ k , n P k , n E k , n P k , n + V k , n P u + σ 2 .
The achievable data rate at user k in bits per second per Hz (bps/Hz) is
R k , n = log 2 ( 1 + γ k , n ) .
For the cooperative user l, which is served jointly by both UAV and GAP, the SINR after SIC can be expressed as
γ l , n = H ^ l , n P l , n + V l , n P u E l , n P l , n + σ 2 + ( H ^ l , n + E l , n ) P k , n ,
with the corresponding achievable rate
R l , n = log 2 ( 1 + γ l , n ) .

2.2. Problem Formulation

In this paper, we focus on jointly optimizing user scheduling, power allocation, and 3D UAV trajectory design to maximize the sum-rate of cooperative users while guaranteeing minimum rate requirements for all users to ensure fairness.
Let N = { 1 , 2 , , N } denote the set of time slots. We define the user scheduling indicator C i , n , which equals 1 if user i is scheduled at time slot n and 0 otherwise. The sum-rate of cooperative users over all time slots is
R sum e = n N l K e C l , n R l , n .
Define the UAV position at time slot n as ( x n , y n , z n ) . Its feasible 3D region is denoted by D , representing a bounded subset of the 3D space where the UAV is allowed to fly.
The optimization problem is formulated as
OP 1 : max { C i , n , P i , n , ( x n , y n , z n ) } R sum e
s . t . j U n P j , n P t , n N ; ( 11 a ) R k , n R min , n N , k U n ; ( 11 b ) l K e C l , n = 1 , n N ; ( 11 c ) k K c C k , n = 1 , n N ; ( 11 d ) m = 1 N C k , m L c , k K c ; ( 11 e ) m = 1 N C l , m L e , l K e ; ( 11 f ) ( x n , y n , z n ) D , n N ; ( 11 g ) | x n + 1 x n | v max , x T N , n = 1 , , N 1 ; ( 11 h ) | y n + 1 y n | v max , y T N , n = 1 , , N 1 ; ( 11 i ) | z n + 1 z n | v max , z T N , n = 1 , , N 1 . ( 11 j )
The constraints in problem OP1 have the following interpretations. Constraint (11a) ensures that the total transmit power allocated by the GAP at each time slot does not exceed its maximum power budget P t . Constraint (11b) guarantees that the achievable data rate of every scheduled user at each time slot meets or exceeds the minimum required rate R min , thus satisfying QoS requirements. Constraints (11c) and (11d) enforce that exactly one cooperative user and one non-cooperative user are scheduled in each time slot, ensuring fair and balanced resource allocation. Meanwhile, constraints (11e) and (11f) limit the total number of time slots that each non-cooperative user and cooperative user can be scheduled, respectively, through the upper bounds L c and L e , preventing excessive scheduling of individual users. Constraint (11g) restricts the UAV’s position in each time slot to lie within the predefined 3D spatial region D . Lastly, constraints (11h) through (11j) impose velocity limits on the UAV’s movement along the x-, y-, and z-axes, respectively, by limiting the maximum displacement between consecutive time slots according to the UAV’s maximum allowable velocities v max , x , v max , y , and v max , z .

3. Joint User Scheduling and Power Allocation

3.1. Closed-Form Power Allocation with Given User Scheduling

Given a fixed user scheduling and UAV trajectory, the optimal power allocation for UAV-GBS CoMP-NOMA can be derived to maximize the sum-rate of CoMP users while ensuring minimum rate requirements for all users.
For each time slot n, consider the non-CoMP user k 1 and CoMP user k 2 scheduled, with their estimated channel gains satisfying H ^ k 1 , n H ^ k 2 , n . Typically, due to poorer channel conditions, the CoMP user is regarded as the weak user in the NOMA pair, and the non-CoMP user as the strong one.
To determine the optimal power allocation, we analyze the relationship between the transmit power and rate of the non-CoMP user. According to the rate-to-power mapping, the transmit power allocated to user k 1 in time slot n is a function of its transmit rate R k 1 , n . Differentiating P k 1 , n with respect to R k 1 , n yields
P k 1 , n R k 1 , n 0 ,
which indicates that the transmit power P k 1 , n increases monotonically with R k 1 , n .
Since a lower power allocation to the non-CoMP user reduces interference imposed on the CoMP user, it effectively increases the leftover power budget P t P k 1 , n available for the CoMP user to improve its achievable rate R k 2 , n . Therefore, to maximize R k 2 , n , the transmit power to non-CoMP user k 1 should be minimized, which corresponds to setting its transmit rate to the minimum required rate:
R k 1 , n * = R min .
With this choice, the optimal transmit power allocated to the non-CoMP user is given by the closed-form expression:
P k 1 , n * = ( 2 R min 1 ) σ 2 + V k 1 , n P u H ^ k 1 , n + E k 1 , n .
Consequently, the remaining power P t P k 1 , n * is allocated to the CoMP user k 2 , whose achievable rate can be expressed as
R k 2 , n * = log 2 1 + H ^ k 2 , n P t P k 1 , n * + V k 2 , n P u σ 2 + E k 2 , n P t P k 1 , n * + H ^ k 2 , n + E k 2 , n P k 1 , n * .
Although the power allocation problem is originally non-convex with respect to the power variables, it has been shown that expressing the sum transmit power as a convex function of users’ rates enables an equivalent convex optimization formulation over rates. This transformation facilitates efficient and tractable solution techniques.
In summary, given the fixed user scheduling and UAV trajectory, the above closed-form expressions yield an optimal power allocation strategy that satisfies minimum rate requirements and maximizes CoMP user rates, providing a foundation for joint optimization schemes.

3.2. Joint User Scheduling and Power Allocation Based on Bipartite Matching

In each time slot n N , the GAP jointly schedules one non-cooperative user k 1 K c and one cooperative user k 2 K e while simultaneously allocating transmit power to maximize system performance. To capture the coupling between user scheduling and power allocation, this problem is formulated as a maximum weight bipartite matching between the two disjoint sets of users at each time slot.
Specifically, we first compute the achievable rate of the cooperative user k 2 when paired with non-cooperative user k 1 in time slot n based on the closed-form power allocation derived in Section III-B. This achievable rate is denoted as R k 2 , n * ( k 1 , k 2 ) . By evaluating all candidate user pairs, we construct a weight matrix W n R | K c | × | K e | , whose entries are given by
W n ( k 1 , k 2 ) = R k 2 , n * ( k 1 , k 2 ) .
The joint user scheduling and power allocation problem in time slot n thus reduces to selecting a matching M n between K c and K e that maximizes the sum of the corresponding achievable rates:
max M n ( k 1 , k 2 ) M n W n ( k 1 , k 2 ) ,
subject to the constraint that each user is scheduled at most once in the time slot, i.e.,
| M n { ( k 1 , · ) } | 1 , k 1 K c , and | M n { ( · , k 2 ) } | 1 , k 2 K e .
This maximum weight bipartite matching problem can be efficiently solved by the Hungarian algorithm, which guarantees finding the optimal user pairing in polynomial time. Once the matching M n is obtained, the closed-form power allocation expressions are applied to each scheduled user pair to ensure that both transmit power constraints and minimum rate requirements are satisfied while maximizing the cooperative users’ achievable rates.
By independently applying this procedure to each time slot, the system performs efficient real-time joint optimization of user scheduling and power allocation that accounts for instantaneous channel conditions. The proposed framework achieves significant performance gains compared to heuristic and decoupled methods by integrating power allocation into the scheduling decision through the weight matrix. The maximum weight bipartite matching ensures optimal user pairing with polynomial-time complexity, making it suitable for real-time implementation. Furthermore, the framework provides flexibility to incorporate fairness constraints by modifying the weight matrix design. This approach offers a tractable solution for joint resource optimization in UAV-assisted cooperative NOMA systems.

4. Trajectory Optimization Using Deep Reinforcement Learning

4.1. Markov Decision Process Formulation

We model the UAV trajectory optimization as an MDP defined by the tuple ( S , A , P , R , γ ) , capturing the interaction between the UAV agent and the wireless environment over discrete time slots indexed by n.

4.1.1. State Space S

At time slot n, the state S n observed by the UAV agent includes its 3D position, time index, and instantaneous channel gains to all users:
S n = x n , y n , z n , n , V 1 , n , V 2 , n , , V U , n ,
where ( x n , y n , z n ) is the UAV location, n is the current time slot, and V u , n is the UAV-to-user u channel gain.

4.1.2. Action Space A

The action a n = ( v x , n , v y , n , v z , n ) represents the UAV’s velocity vector at time n with bounded components:
v x , n , v y , n [ v max , v max ] , v z , n [ v max , z , v max , z ] .
This continuous action space enables smooth 3D trajectory control.

4.1.3. State Transition Probability P

The system state transitions according to the UAV’s mobility and wireless environment dynamics:
S n + 1 P S n + 1 S n , a n .
Given the current position x n , y n , z n and velocity command a n , the UAV position updates by
( x n + 1 , y n + 1 , z n + 1 ) = clip ( x n , y n , z n ) + a n , D ,
where clip ( · ) enforces spatial boundary constraints within the feasible flight domain D . The channel gains h u , n + 1 update as a function of the new UAV location and user positions following large-scale path loss models. The transition probability accounts implicitly for channel variations and system uncertainties. In our design, the transition model is unknown and learned implicitly by model-free reinforcement learning agents.

4.1.4. Reward Function R

At each time slot n, the agent receives a scalar reward r n = R ( S n , a n ) , designed to simultaneously encourage high edge user rates, promote user fairness, and drive the UAV to return to its initial position. The reward is structured as follows:
r n = w 1 R sum , n e + w 2 r fair , n + w 3 r return , n
where R sum , n e = l K e C l , n R l , n promotes maximizing edge user throughput, r fair , n penalizes unfair scheduling, and r return , n encourages the UAV’s return to its initial position. The weights w 1 , w 2 , w 3 are hyperparameters that control the relative importance of throughput maximization, fairness enforcement, and trajectory optimization, respectively.

4.1.5. Discount Factor γ

The discount factor γ ( 0 , 1 ) trades off the immediate and future rewards, with larger values emphasizing long-term system performance and trajectory planning.

4.2. Soft Actor–Critic Algorithm

To solve the formulated MDP, we employ the SAC algorithm, which is particularly suitable for continuous control tasks such as UAV trajectory optimization. SAC is an off-policy, model-free deep reinforcement learning method that simultaneously maximizes expected cumulative reward and policy entropy, thereby fostering robust and sample-efficient learning in complex and high-dimensional environments.
SAC trains three parameterized functions concurrently: a stochastic policy (actor) π ϕ ( a | s ) and two soft Q-functions (critics) Q θ 1 ( s , a ) and Q θ 2 ( s , a ) , parameterized by ϕ , θ 1 , and θ 2 , respectively. The actor outputs a probability distribution over continuous actions, enabling effective exploration, while the dual critics reduce overestimation bias through clipped double Q-learning.
The overall architecture of the Soft Actor–Critic algorithm for the UAV-GAP cooperative NOMA system is illustrated in Figure 2. This framework integrates the UAV’s continuous trajectory control, represented as velocity commands, with the joint user scheduling and power allocation coordinated between the UAV and the ground base station. The SAC agent observes the current system state, including UAV position and channel conditions, and outputs continuous control actions to optimize long-term performance.
Contrary to conventional RL algorithms that optimize expected return alone, SAC maximizes a maximum entropy objective, defined as
J ( π ) = n E ( s n , a n ) ρ π r ( s n , a n ) + α H π ( · | s n ) ,
where H ( π ( · | s n ) ) represents the Shannon entropy of the policy at state s n and α > 0 is the temperature parameter governing the trade-off between exploration and exploitation. This entropy regularization promotes stochasticity in the policy, encouraging diverse behaviors and avoiding premature convergence to suboptimal deterministic policies.
For stable training, SAC utilizes two Q-functions approximated by neural networks and employs clipped double Q-learning to mitigate positive bias in value estimates. Experience replay buffers and target networks further enhance sample efficiency and convergence stability.
During training, each critic network minimizes the soft Bellman residual:
L ( θ i ) = E ( s n , a n , r n , s n + 1 ) D Q θ i ( s n , a n ) y n 2 , i { 1 , 2 } ,
where the target y n is computed by
y n = r n + γ E a n + 1 π ϕ min j = 1 , 2 Q θ j ( s n + 1 , a n + 1 ) α log π ϕ ( a n + 1 | s n + 1 ) ,
and θ j denotes the parameters of the slowly updated target critic networks.
The actor updates its policy by minimizing the expected Kullback–Leibler divergence between π ϕ and the exponential of the Q-function, effectively maximizing
J ( ϕ ) = E s n D , a n π ϕ α log π ϕ ( a n | s n ) min j = 1 , 2 Q θ j ( s n , a n ) .
The temperature parameter α can be adjusted automatically to maintain a target entropy, further improving exploration and training stability.
In the context of UAV cooperative NOMA systems, SAC’s capabilities to handle continuous 3D velocity actions and noisy, high-dimensional observations such as UAV position and channel gains are crucial. The entropy-regularized policy encourages diverse trajectory exploration, which helps the UAV avoid suboptimal flight patterns caused by the highly non-convex wireless environment and dynamic user scheduling.
Overall, SAC combines rigorous theoretical foundations with practical efficacy, making it an ideal approach for the UAV trajectory optimization problem. The detailed procedure of the Soft Actor–Critic-based trajectory optimization is summarized in Algorithm 1, which outlines the main steps of network initialization, interaction, training updates, and policy improvement.
Algorithm 1 Soft Actor–Critic-Based Trajectory Optimization
  1:
Initialize actor network π ϕ ( a | s ) and two critic networks Q θ 1 ( s , a ) , Q θ 2 ( s , a ) with random weights ϕ , θ 1 , θ 2 .
  2:
Initialize target critic networks with weights θ 1 θ 1 , θ 2 θ 2 .
  3:
Initialize temperature parameter α and replay buffer D .
  4:
for episode = 1 to MaxEpisodes do
  5:
      Reset environment and obtain initial state s 0 .
  6:
      for time step n = 0 to N 1  do
  7:
            Sample action a n π ϕ ( · | s n ) .
  8:
            Execute a n ; observe reward r n and next state s n + 1 .
  9:
            Store ( s n , a n , r n , s n + 1 ) in D .
10:
            Sample random minibatch of size M from D .
11:
            for  i = 1 to M do
12:
                 Compute target value:
y i = r i + γ E a i + 1 π ϕ min j = 1 , 2 Q θ j ( s i + 1 , a i + 1 ) α log π ϕ ( a i + 1 | s i + 1 ) .
13:
            end for
14:
            Update critics θ 1 , θ 2 by minimizing:
L ( θ j ) = 1 M i = 1 M Q θ j ( s i , a i ) y i 2 , j = 1 , 2 .
15:
            Update actor ϕ by minimizing:
J ( ϕ ) = 1 M i = 1 M α log π ϕ ( a i | s i ) min j = 1 , 2 Q θ j ( s i , a i ) .
16:
            (Optional) Adjust temperature α to control entropy.
17:
            Soft update target networks:
θ j τ θ j + ( 1 τ ) θ j , j = 1 , 2 ; ϕ τ ϕ + ( 1 τ ) ϕ .
18:
      end for
19:
end for
20:
return Learned policy π ϕ ( a | s ) and optimized UAV trajectory.

4.3. Computational Complexity Analysis

This section analyzes the computational complexity of the proposed joint optimization framework to demonstrate its practical feasibility for real-time IIoT applications.
The complexity analysis considers our hybrid optimization design, which enables distributed implementation. In this design, resource allocation computations occur at the ground access point, while trajectory control runs on the UAV. The SAC policy network requires minimal computational resources for online execution through a single forward pass. The discrete time slot operation provides adequate time for optimization between transmission periods.

4.3.1. Joint User Scheduling and Power Allocation

The resource allocation algorithm operates in two sequential phases for each time slot. First, the algorithm computes the weight matrix by evaluating achievable rates for all possible user pairs. Given the closed-form power allocation solutions derived in Section 4.1, the rate computation for each cellular-edge user pair ( k 1 , k 2 ) requires constant time operations. Therefore, constructing the complete weight matrix has complexity O ( | K c | · | K e | ) .
Second, the maximum weight bipartite matching problem is solved using the Hungarian algorithm. For a bipartite graph with | K c | cellular users and | K e | edge users, the Hungarian algorithm requires O ( ( | K c | + | K e | ) 3 ) operations. Since the matching complexity typically dominates the weight computation for practical system sizes, the overall per-time-slot complexity is
O max { | K c | · | K e | , ( | K c | + | K e | ) 3 }

4.3.2. SAC-Based Trajectory Optimization

The SAC algorithm involves distinct training and inference phases with different computational requirements.
(1) Training Phase: The SAC framework maintains one actor network and two critic networks, along with corresponding target networks. During each training step, the algorithm performs forward and backward passes through all networks. With E training episodes, N time slots per episode, and neural networks having L layers with H hidden units each, the training complexity is
O E · N · 2 L H 2 Two Critics + L H 2 Actor + L H 2 Target Updates = O ( 4 E · N · L · H 2 )
(2) Inference Phase: During online operation, only the trained actor network executes policy decisions, requiring O ( L · H 2 ) operations per time slot.

4.3.3. Overall System Complexity

The total online computational complexity per time slot combines both resource allocation and trajectory optimization components:
O max { ( | K c | + | K e | ) 3 , L · H 2 }
The polynomial-time complexity ensures computational tractability for practical IIoT deployments. For typical deployment scenarios with moderate numbers of users and compact neural network architectures, the computational overhead remains within the processing capabilities of modern edge computing platforms, enabling real-time system operation.

5. Results and Discussion

In this section, we evaluate the performance of the proposed joint user scheduling and power allocation scheme in the UAV cooperative NOMA system through extensive simulations. The simulation settings and key system parameters [32] are summarized in Table 1.
For the SAC-based trajectory optimization, we employed a deep neural network architecture with appropriate hyperparameter settings to ensure stable learning and optimal convergence performance. The specific training parameters for the SAC algorithm are detailed in Table 2.
Figure 3 illustrates the evolution of the system reward during SAC algorithm training as a function of episodes. It can be observed that in the early stages of training, the reward curve exhibited significant fluctuations, reflecting the typical characteristics of random exploration and policy updates. As training progressed, the system reward gradually increased and the fluctuations decreased, eventually reaching a steady state, which indicated that the SAC agent’s policy gradually converged and stabilized around the optimal or near-optimal solution.
Figure 4 presents the convergence characteristics of the SAC algorithm under different hyperparameter configurations to demonstrate the robustness of our hybrid optimization framework against hyperparameter variations. The results reveal that the algorithm achieved relatively stable performance across various learning rates and discount factors, which validates the effectiveness of our architectural design, which combines closed-form power allocation solutions with efficient user scheduling to constrain the SAC search space.
Figure 5 compares cooperative user rate performance under different minimum rate requirements for cooperative NOMA and non-cooperative schemes, each trained with SAC or DDPG. The cooperative user rate decreased as minimum rate requirements increased. The proposed cooperative NOMA scheme combined with SAC consistently achieved the highest cooperative user rates, outperforming all other schemes. The DDPG-based cooperative scheme performed better than the corresponding non-cooperative scheme but was outperformed by SAC-based solutions. These results demonstrate the superiority of integrating cooperative transmission and SAC learning, especially under stringent rate constraints.
Figure 6 shows the effect of increasing GAP transmit power on cooperative user rates. As the GAP power increased, cooperative user rates rose, with the cooperative NOMA scheme demonstrating substantial gains over the non-cooperative scheme. This improvement was attributed to the proposed scheme’s effective interference management via UAV–GAP cooperative transmission, which mitigated interference from the GAP- to UAV-served users.
Figure 7 illustrates the overall system sum rate as a function of UAV transmit power, with the GAP power fixed at 1.5 W and the minimum rate requirement at 5 bps/Hz. For the cooperative NOMA scheme, while increasing UAV power improved the sum rate, the growth rate slowed as power continued to rise. This phenomenon occurred because although cooperative NOMA boosted cooperative user rates, higher UAV power also intensified interference for non-cooperative users within NOMA groups, constraining overall throughput gains. In contrast, for the non-cooperative scheme, the system sum rate decreased as UAV transmit power increased since the UAV acted purely as an interference source without providing beneficial cooperative transmission to users served by the GAP.
Figure 8 demonstrates system performance with varying numbers of users. The simulation results show a decreasing trend in average cooperative user rate as the number of users increased, which was expected due to the fundamental resource-sharing nature of the system. Specifically, while the SAC-based algorithm maintained robust resource allocation strategies that kept the total system throughput relatively stable, the increasing number of cooperative users led to more intensive resource competition. Consequently, when the total achievable rate was divided among a larger number of cooperative users, the per-user average rate naturally decreased. This trend validates the algorithm’s ability to maintain system-wide performance stability while fairly distributing resources among an expanding user population.
In Figure 9, we evaluate the energy efficiency (EE) performance of the proposed SAC and DDPG algorithms, considering both transmit power and propulsion power of the UAV. We set P t = 1 W and P u = 30 mW. The EE of the UAV-GAP cooperative NOMA systems could be obtained as E E = B N n = 1 N i L c S i , n R i , n + l L e S l , n R l , n P ( v n ) + P u + j n P j , n . Here, P ( v n ) represents the propulsion power, which can be regarded as a function of the UAV’s flight speed v n . As can be seen, there was a trade-off between system transmission rate and total power consumption, and the proposed SAC-based cooperative algorithm achieved superior EE performance compared to DDPG.
Figure 10 compares edge user rates under varying channel estimation errors across cooperative and non-cooperative modes trained with SAC and DDPG. Edge rates degraded as estimation error increased. The SAC-based cooperative scheme consistently achieved the highest rates, outperforming other schemes. Non-cooperative approaches suffered from a lack of coordination, resulting in lower performance and greater sensitivity to estimation errors. The results highlight that integrating cooperative transmission with high-performance learning algorithms improves robustness against channel uncertainties.
Figure 11 illustrates the UAV trajectory in both 2D and 3D views under varying flight durations T = 60 s, 75 s, and 120 s, with a minimum edge user rate requirement R min = 5 bps/Hz. All ground users were located at zero altitude. For a short flight time (T = 60 s), the UAV trajectory was compact and flight distance was limited due to constrained time and space, forcing the UAV to serve all users within a smaller region. As the flight duration was extended (T = 75 s), the UAV exploited its 3D mobility more flexibly, approaching users more closely and adjusting the altitude to optimize channel conditions. With sufficient flight time (T = 120 s), the UAV executed more diverse trajectories with wider coverage and dynamic service adaptation, fully leveraging its maneuverability to satisfy communication demands.

6. Conclusions

This paper presented an innovative framework for intelligent resource allocation and 3D trajectory optimization in UAV-GAP cooperative NOMA systems based on the SAC deep reinforcement learning algorithm. By integrating power allocation and user scheduling into the reinforcement learning decision-making process and utilizing the SAC algorithm to handle continuous action spaces, we successfully achieved adaptive optimization of UAV trajectories, thereby significantly improving system performance. Experimental results demonstrated that compared to traditional non-cooperative and DDPG algorithms, our proposed scheme exhibited superior performance in enhancing overall system throughput, improving spectral efficiency, effectively managing interference, and guaranteeing the quality of service for cooperative users. Particularly in dynamic and complex industrial IoT scenarios, this framework showed stronger adaptability and robustness, capable of effectively addressing changes in channel conditions and the impact of errors. Future research directions should include the further exploration of multi-UAV cooperation and integrated sensing and communication and the consideration of UAV energy efficiency constraints to meet more complex Industrial application requirements.

Author Contributions

Conceptualization, Y.H. and H.Z. (Haiyong Zeng); Methodology, Y.H., S.H., H.Z. (Hongyan Zhu) and H.Z. (Haiyong Zeng); Software, Y.H. and J.S.; Validation, Y.H.; Formal analysis, Y.H. and H.Z. (Haiyong Zeng); Investigation, Y.H., J.S., X.L. and H.Z. (Haiyong Zeng); Data curation, Y.H. and X.L.; Writing—original draft, Y.H. and J.S.; Writing—review and editing, Y.H., J.S. and H.Z. (Haiyong Zeng); Visualization, Y.H.; Supervision, S.H. and H.Z. (Hongyan Zhu); Project administration, H.Z. (Haiyong Zeng); Funding acquisition, H.Z. (Haiyong Zeng). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of China under grant 62301172, the Guangxi Natural Science Foundation under grant 2024GXNSFBA010246, the Guangxi Science and Technology Base and Special Talent Program under grant GuikeAD23026197, the Guangxi Young Talent Inclusive Support Program 2024, and the Guangxi Key Laboratory of Brain-inspired Computing and Intelligent Chips under grant BCIC-24-Z6.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author. The data are not publicly available due to the copyright.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, X.; Liu, Y.; Zhou, S.; Bian, J.; Huang, J.; Li, X.; Xin, Z. GAN-Based Channel Generation and Modeling for 6G Intelligent IIoT Communications. IEEE Internet Things J. 2025; in press. [Google Scholar] [CrossRef]
  2. Mostafa, S.; Mota, M.P.; Valcarce, A.; Bennis, M. Intent-Aware DRL-Based NOMA Uplink Dynamic Scheduler for IIoT. IEEE Trans. Cogn. Commun. Netw. 2025; in press. [Google Scholar] [CrossRef]
  3. Wu, W.; Zhou, F.; Wang, B.; Wu, Q.; Dong, C.; Hu, R.Q. Unmanned Aerial Vehicle Swarm-Enabled Edge Computing: Potentials, Promising Technologies, and Challenges. IEEE Wireless Commun. 2022, 29, 78–85. [Google Scholar] [CrossRef]
  4. Zeng, H.; Zhu, X.; Jiang, Y.; Wei, Z.; Sun, S.; Xiong, X. Toward UL-DL Rate Balancing: Joint Resource Allocation and Hybrid-Mode Multiple Access for UAV-BS-Assisted Communication Systems. IEEE Trans. Commun. 2022, 70, 2757–2771. [Google Scholar] [CrossRef]
  5. Cheng, F.; Gui, G.; Zhao, N.; Chen, Y.; Tang, J.; Sari, H. UAV-Relaying-Assisted Secure Transmission With Caching. IEEE Trans. Commun. 2019, 67, 3140–3153. [Google Scholar] [CrossRef]
  6. Zhao, N.; Lu, W.; Sheng, M.; Chen, Y.; Tang, J.; Yu, F.R.; Wong, K.-K. UAV-Assisted Emergency Networks in Disasters. IEEE Wireless Commun. 2019, 26, 45–51. [Google Scholar] [CrossRef]
  7. Zeng, H.; Zhang, R.; Zhu, X.; Wei, Z.; Jiang, Y.; Sun, S.; Zheng, F.-C.; Cao, B. Toward 3-D AAV-Ground BS CoMP-NOMA Transmission: Optimal Resource Allocation and Trajectory Design. IEEE Internet Things J. 2025, 12, 9671–9686. [Google Scholar] [CrossRef]
  8. Wu, Q.; Zhang, R. Common Throughput Maximization in UAV-Enabled OFDMA Systems With Delay Consideration. IEEE Trans. Commun. 2018, 66, 6614–6627. [Google Scholar] [CrossRef]
  9. Zhang, H.; Zhang, J.; Long, K. Energy Efficiency Optimization for NOMA UAV Network With Imperfect CSI. IEEE J. Sel. Areas Commun. 2020, 38, 2798–2809. [Google Scholar] [CrossRef]
  10. Lyu, J.; Zeng, Y.; Zhang, R. UAV-Aided Offloading for Cellular Hotspot. IEEE Trans. Wireless. Commun. 2018, 17, 3988–4001. [Google Scholar] [CrossRef]
  11. Dai, L.; Wang, B.; Ding, Z.; Wang, Z.; Chen, S.; Hanzo, L. A Survey of Non-Orthogonal Multiple Access for 5G. IEEE Commun. Surv. Tutorials. 2018, 20, 2294–2323. [Google Scholar] [CrossRef]
  12. Liu, Y.; Qin, Z.; Cai, Y.; Gao, Y.; Li, G.Y.; Nallanathan, A. UAV Communications Based on Non-Orthogonal Multiple Access. IEEE Wireless. Commun. 2019, 26, 52–57. [Google Scholar] [CrossRef]
  13. Hou, T.; Liu, Y.; Song, Z.; Sun, X.; Chen, Y. Multiple Antenna Aided NOMA in UAV Networks: A Stochastic Geometry Approach. IEEE Trans. Commun. 2019, 67, 1031–1044. [Google Scholar] [CrossRef]
  14. Miuccio, L.; Panno, D.; Riolo, S. A Flexible Encoding/Decoding Procedure for 6G SCMA Wireless Networks via Adversarial Machine Learning Techniques. IEEE Trans. Veh. Technol. 2022, 72, 3288–3303. [Google Scholar] [CrossRef]
  15. Ülgen, O.; Tufekci, T.K.; Sadi, Y.; Erkucuk, S.; Anpalagan, A.; Baykaş, T. Sparse Code Multiple Access with Time Spreading and Repetitive Transmissions. Int. J. Commun. Syst. 2025, 38, e6121. [Google Scholar] [CrossRef]
  16. Wu, G.; Chen, G.; Gu, X. NOMA-Based Rate Optimization for Multi-UAV-Assisted D2D Communication Networks. Drones 2025, 9, 62. [Google Scholar] [CrossRef]
  17. Wu, J.; Liu, C.; Wang, X.; Cheng, C.T.; Zhou, Q. Jointly Optimizing Resource Allocation, User Scheduling, and Grouping in SBMA Networks: A PSO Approach. Entropy 2025, 27, 691. [Google Scholar] [CrossRef]
  18. Li, N.; Wu, P.; Zhu, L.; Ng, D.W.K. Movable-Antenna Array Enhanced Downlink NOMA. arXiv 2025, arXiv:2506.11438. [Google Scholar]
  19. Pereira, F.M.; de Araújo Farhat, J.; Rebelatto, J.L.; Brante, G.; Souza, R.D. Reinforcement Learning-Aided NOMA Random Access: An AoI-Based Timeliness Perspective. IEEE Internet Things J. 2024, 12, 6058–6061. [Google Scholar] [CrossRef]
  20. Irmer, R.; Droste, H.; Marsch, P.; Grieger, M.; Fettweis, G.; Brueck, S.; Mayer, H.-P.; Thiele, L.; Jungnickel, V. Coordinated multipoint: Concepts, performance, and field trial results. IEEE Commun. Mag. 2011, 49, 102–111. [Google Scholar] [CrossRef]
  21. Nguyen, T.M.; Ajib, W.; Assi, C. A Novel Cooperative NOMA for Designing UAV-Assisted Wireless Backhaul Networks. IEEE J. Sel. Areas Commun. 2018, 36, 2497–2507. [Google Scholar] [CrossRef]
  22. Zhao, N.; Pang, X.; Li, Z.; Chen, Y.; Li, F.; Ding, Z.; Alouini, M.-S. Joint Trajectory and Precoding Optimization for UAV-Assisted NOMA Networks. IEEE Trans. Commun. 2019, 67, 3723–3735. [Google Scholar] [CrossRef]
  23. Ali, M.S.; Hossain, E.; Al-Dweik, A.; Kim, D.I. Downlink Power Allocation for CoMP-NOMA in Multi-Cell Networks. IEEE Trans. Commun. 2018, 66, 3982–3998. [Google Scholar] [CrossRef]
  24. Zeng, H.; Zhu, X.; Jiang, Y.; Wei, Z.; Wang, T. A Green Coordinated Multi-Cell NOMA System With Fuzzy Logic Based Multi-Criterion User Mode Selection and Resource Allocation. IEEE J. Sel. Top. Signal Process. 2019, 13, 480–495. [Google Scholar] [CrossRef]
  25. Qu, Y.; Dai, H.; Wang, H.; Dong, C.; Wu, F.; Guo, S.; Wu, Q. Service Provisioning for UAV-Enabled Mobile Edge Computing. IEEE J. Sel. Areas Commun. 2021, 39, 3287–3305. [Google Scholar] [CrossRef]
  26. Bayessa, G.A.; Chai, R.; Liang, C.; Jain, D.K.; Chen, Q. Joint UAV Deployment and Precoder Optimization for Multicasting and Target Sensing in UAV-Assisted ISAC Networks. IEEE Internet Things J. 2024, 11, 33392–33405. [Google Scholar] [CrossRef]
  27. Zhao, N.; Cheng, F.; Yu, F.R.; Tang, J.; Chen, Y.; Gui, G.; Sari, H. Caching UAV Assisted Secure Transmission in Hyper-Dense Networks Based on Interference Alignment. IEEE Trans. Commun. 2018, 66, 2281–2294. [Google Scholar] [CrossRef]
  28. Song, Q.; Zheng, F.-C.; Zeng, Y.; Zhang, J. Joint Beamforming and Power Allocation for UAV-Enabled Full-Duplex Relay. IEEE Trans. Veh. Technol. 2019, 68, 1657–1671. [Google Scholar] [CrossRef]
  29. Yin, S.; Yu, F.R. Resource Allocation and Trajectory Design in UAV-Aided Cellular Networks Based on Multiagent Reinforcement Learning. IEEE Internet Things J. 2022, 9, 2933–2943. [Google Scholar] [CrossRef]
  30. Yuan, X.; Hu, S.; Ni, W.; Wang, X.; Jamalipour, A. Deep Reinforcement Learning-Driven Reconfigurable Intelligent Surface-Assisted Radio Surveillance with a Fixed-Wing UAV. IEEE Trans. Inf. Forensics Secur. 2023, 18, 4546–4560. [Google Scholar] [CrossRef]
  31. Dong, R.; Wang, B.; Cao, K.; Tian, J.; Cheng, T. Secure Transmission Design of RIS Enabled UAV Communication Networks Exploiting Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2024, 73, 8404–8419. [Google Scholar] [CrossRef]
  32. Zeng, H.; Zhang, R.; Zhu, X.; Jiang, Y.; Wei, Z.; Zheng, F.-C. Interference-Aware AAV-TBS Coordinated NOMA: Joint User Scheduling, Power Allocation and Trajectory Design. IEEE Open J. Veh. Technol. 2025, 6, 812–828. [Google Scholar] [CrossRef]
  33. Khawaja, W.; Guvenc, I.; Matolak, D.W.; Fiebig, U.-C.; Schneckenburger, N. A Survey of Air-to-Ground Propagation Channel Modeling for Unmanned Aerial Vehicles. IEEE Commun. Surv. Tutor. 2019, 21, 2361–2391. [Google Scholar] [CrossRef]
  34. Al-Hourani, A.; Kandeepan, S.; Lardner, S. Optimal LAP Altitude for Maximum Coverage. IEEE Wireless Commun. Lett. 2014, 3, 569–572. [Google Scholar] [CrossRef]
  35. Ju, S.; Shakya, D.; Poddar, H.; Xing, Y.; Kanhere, O.; Rappaport, T.S. 142 GHz Sub-Terahertz Radio Propagation Measurements and Channel Characterization in Factory Buildings. IEEE Trans. Wireless Commun. 2023, 23, 7127–7143. [Google Scholar] [CrossRef]
  36. Qin, Y.; Tang, P.; Tian, L.; Lin, J.; Chang, Z.; Liu, P.; Zhang, J.; Jiang, T. Time-Varying Channel Measurement and Analysis at 105 GHz in an Indoor Factory. In Proceedings of the 2024 18th European Conference on Antennas and Propagation (EuCAP), Rome, Italy, 17–22 March 2024; pp. 1–5. [Google Scholar]
  37. Yusuf, T.A.O.; Petersen, S.S.; Li, P.; Ren, J.; Mursia, P.; Sciancalepore, V.; Pérez, X.C.; Berardinelli, G.; Shen, M. AI-Assisted NLOS Sensing for RIS-Based Indoor Localization in Smart Factories. arXiv 2025, arXiv:2505.15989. [Google Scholar]
  38. Fang, F.; Zhang, H.; Cheng, J.; Leung, V.C.M. Energy Efficient Resource Allocation for Downlink Nonorthogonal Multiple Access Network. IEEE Trans. Commun. 2016, 64, 3722–3732. [Google Scholar] [CrossRef]
Figure 1. System model of UAV-GAP cooperative NOMA for IIoT, where all ground users are served by the GAP via NOMA and the UAV coordinates with the GAP to cooperatively transmit signals to cooperative users.
Figure 1. System model of UAV-GAP cooperative NOMA for IIoT, where all ground users are served by the GAP via NOMA and the UAV coordinates with the GAP to cooperatively transmit signals to cooperative users.
Entropy 27 00811 g001
Figure 2. Soft Actor–Critic framework for UAV trajectory and joint scheduling optimization in a UAV-GAP cooperative NOMA system.
Figure 2. Soft Actor–Critic framework for UAV trajectory and joint scheduling optimization in a UAV-GAP cooperative NOMA system.
Entropy 27 00811 g002
Figure 3. Training reward curve of SAC algorithm.
Figure 3. Training reward curve of SAC algorithm.
Entropy 27 00811 g003
Figure 4. Convergence characteristics of the SAC algorithm with different hyperparameters: (a) impact of learning rate; (b) impact of discount factor.
Figure 4. Convergence characteristics of the SAC algorithm with different hyperparameters: (a) impact of learning rate; (b) impact of discount factor.
Entropy 27 00811 g004
Figure 5. Impact of minimum rate requirement on cooperative user performance.
Figure 5. Impact of minimum rate requirement on cooperative user performance.
Entropy 27 00811 g005
Figure 6. Impact of ground access point transmit power on cooperative user rate.
Figure 6. Impact of ground access point transmit power on cooperative user rate.
Entropy 27 00811 g006
Figure 7. System sum rate with varying UAV transmit power.
Figure 7. System sum rate with varying UAV transmit power.
Entropy 27 00811 g007
Figure 8. System performance with varying numbers of users.
Figure 8. System performance with varying numbers of users.
Entropy 27 00811 g008
Figure 9. Energy efficiency performance for UAV-GAP cooperative NOMA systems.
Figure 9. Energy efficiency performance for UAV-GAP cooperative NOMA systems.
Entropy 27 00811 g009
Figure 10. Effect of channel estimation error on cooperative user rate.
Figure 10. Effect of channel estimation error on cooperative user rate.
Entropy 27 00811 g010
Figure 11. UAV trajectory illustration with user locations. (a) UAV trajectory top-down view, (b) UAV trajectory 3D view.
Figure 11. UAV trajectory illustration with user locations. (a) UAV trajectory top-down view, (b) UAV trajectory 3D view.
Entropy 27 00811 g011
Table 1. Simulation parameters.
Table 1. Simulation parameters.
NotationDescriptionValue
BSystem bandwidth1 MHz
P t Ground base station transmit power1.5 W
P u UAV transmit power40 mW
K c Number of non-cooperative users3
K e Number of cooperative users4
HUAV initial flight altitude70 m
v max UAV maximum horizontal speed50 m/s
v max , z UAV maximum vertical speed20 m/s
NNumber of time slots80
D UAV flight area [ 400 , 400 ] × [ 400 , 400 ] × [ 50 , 100 ] m
R min Minimum rate requirement per user4 bits/s/Hz
σ 2 Noise power spectral density−110 dBm
a, bUrban LoS channel parameters4.88, 0.43
α Path loss exponent3
L c , L e Max scheduling counts for center and edge users27, 23
Table 2. SAC algorithm training parameters.
Table 2. SAC algorithm training parameters.
SymbolParameterValue
α a Actor learning rate0.0003
α c Critic learning rate0.0003
γ Discount factor0.98
| B | Replay buffer size 10 6
BBatch size64
τ Target network update rate0.005
EEpisodes1000
NTime slots per episode80
LNeural network layers3
HHidden units per layer128
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, Y.; Su, J.; Lu, X.; Huang, S.; Zhu, H.; Zeng, H. Deep Reinforcement Learning-Based Resource Allocation for UAV-GAP Downlink Cooperative NOMA in IIoT Systems. Entropy 2025, 27, 811. https://doi.org/10.3390/e27080811

AMA Style

Huang Y, Su J, Lu X, Huang S, Zhu H, Zeng H. Deep Reinforcement Learning-Based Resource Allocation for UAV-GAP Downlink Cooperative NOMA in IIoT Systems. Entropy. 2025; 27(8):811. https://doi.org/10.3390/e27080811

Chicago/Turabian Style

Huang, Yuanyan, Jingjing Su, Xuan Lu, Shoulin Huang, Hongyan Zhu, and Haiyong Zeng. 2025. "Deep Reinforcement Learning-Based Resource Allocation for UAV-GAP Downlink Cooperative NOMA in IIoT Systems" Entropy 27, no. 8: 811. https://doi.org/10.3390/e27080811

APA Style

Huang, Y., Su, J., Lu, X., Huang, S., Zhu, H., & Zeng, H. (2025). Deep Reinforcement Learning-Based Resource Allocation for UAV-GAP Downlink Cooperative NOMA in IIoT Systems. Entropy, 27(8), 811. https://doi.org/10.3390/e27080811

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop