Next Article in Journal
Biomechanical Analysis of an Elite Para Standing Cross-Country Skier Using Lower Limb Prostheses: A Case Study
Previous Article in Journal
Overcoming Scale Variations and Occlusions in Aerial Detection: A Context-Aware DEIM Framework
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FMA-MADDPG: Constrained Multi-Agent Resource Optimization with Channel Prediction in 6G Non-Terrestrial Networks

1
School of Astronautics, Harbin Institute of Technology, Harbin 150001, China
2
The 54th Research Institute, China Electronics Technology Group Corporation, Shijiazhuang 050081, China
3
School of Artificial Intelligence, Key Laboratory of Intelligent Perception and Image Understanding, Ministry of Education, Xidian University, Xi’an 710071, China
4
College of Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(1), 148; https://doi.org/10.3390/s26010148 (registering DOI)
Submission received: 21 October 2025 / Revised: 22 December 2025 / Accepted: 22 December 2025 / Published: 25 December 2025
(This article belongs to the Section Remote Sensors)

Abstract

Sixth-generation (6G) wireless systems aim to integrate terrestrial, aerial, and satellite networks to support large-scale remote sensing and service delivery. In such non-terrestrial networks (NTNs), channels change quickly and the multi-tier architecture is heterogeneous, which makes real-time channel state acquisition and cooperative resource scheduling difficult. This paper proposes an FMA-MADDPG framework that combines a channel prediction module with a constraint-based multi-agent deep deterministic policy gradient scheme. The Fusion of Mamba and Attention (FMA) predictor uses a Mamba state-space backbone and a multi-head self-attention block to learn both long-term channel evolution and short-term fluctuations, and forecasts future CSI. The predicted channel information is added to the agents’ observations so that scheduling decisions can take expected channel variations into account. A constraint-based reward is also designed, with explicit performance thresholds and anti-idle penalties, to encourage fairness, avoid free-riding, and promote cooperation among heterogeneous agents. In a representative NTN uplink scenario, the proposed method achieves higher total reward, efficiency, load balance, and cooperation than several DRL baselines, with relative gains around 10–20% on key metrics. These results indicate that prediction-aware cooperative reinforcement learning is a useful approach for resource optimization in future 6G NTN systems.

1. Introduction

Sixth-generation (6G) wireless communication is reshaping how networks are designed and operated [1]. Unlike fifth-generation (5G) wireless communication—which mainly targets terrestrial users and base stations—6G aims to build a global network that connects ground, air, and space. This development is closely linked to Earth observation and remote sensing. In these applications, large volumes of heterogeneous data are generated by distributed ground sensors, unmanned aerial vehicles (UAVs), and satellites. Typical missions such as environmental monitoring, disaster response, precision agriculture, and climate analysis require high data rates, low latency, and broad spatial coverage [2].
Current terrestrial networks are not well suited for large-scale remote sensing tasks [3]. In remote regions such as oceans, deserts, forests, and mountains, deploying dense ground infrastructure is often too costly or technically infeasible. Even in urban areas, multipath fading, building blockage, and congestion frequently degrade communication quality. These limitations create a mismatch between front-end data acquisition and back-end data processing. As a result, the timeliness and reliability of geospatial services are reduced.
Non-terrestrial networks (NTNs) have been introduced to bridge this gap. NTNs integrate low-Earth-orbit (LEO) satellites, high-altitude platforms (HAPs), and terrestrial networks into a three-dimensional architecture [4]. In this structure, user equipment (UE) represents ground sensors, UAVs, or mobile terminals that generate data. HAPs operate as relay and edge computing nodes that support local processing, data fusion, and temporary storage. LEO satellites provide wide-area backhaul and global coordination so that data from distributed sensors can be efficiently aggregated and delivered [5]. Together, these three layers form a cooperative communication and computing system capable of supporting data-intensive sensing tasks across large geographic regions.
Deploying and operating such an NTN architecture introduces several technical challenges. First, wireless channels between UEs, HAPs, and LEO satellites vary rapidly. Atmospheric turbulence, rain, blockages, and Doppler shifts cause fast and sometimes correlated changes in signal quality. The UE–HAP and HAP–LEO links also differ in propagation characteristics, which increases uncertainty. Under these conditions, obtaining accurate channel state information (CSI) in real time becomes difficult. Without reliable CSI, network agents must make scheduling decisions based on partial observations, often resulting in poor resource utilization. Second, LEO satellites move quickly along their orbits, causing continuous changes in network topology and connectivity [6]. The available bandwidth, computing resources, and relay paths in a given area depend on which satellites are visible at that moment. This makes multi-agent resource coordination a time-varying problem. Traditional optimization methods often assume fixed topologies, convex formulations, or full system knowledge, and therefore struggle under such dynamics. Third, reinforcement learning (RL) and multi-agent RL (MARL) methods face their own limitations. Many RL algorithms assume stationary environments or full observability, which is unrealistic in NTNs. Even with partial observability, learning can be unstable when state representations are weak or rewards are ambiguous. Without domain-specific prediction or augmentation, policies may become short-sighted and fail to generalize across agents, locations, and time scales [7].
In this work, we propose a unified framework, termed Fusion-of-Mamba-and-Attention-enhanced Multi-Agent Deep Deterministic Policy Gradient (FMA-MADDPG), that combines predictive modeling, constraint-aware learning, and coordinated control across the three-layer UE–HAP–LEO NTN. To improve channel observability, we design a prediction module named FMA (Fusion of Mamba and Attention). It employs Mamba blocks to capture long-term channel evolution and multi-head self-attention to model short-term variations. Through this design, FMA can forecast CSI several steps ahead. The predicted CSI is incorporated into the observations of HAP and LEO agents, enabling their decisions to reflect both current and near-future channel conditions. In parallel, we develop a constrained multi-agent reinforcement learning scheme based on multi-agent deep deterministic policy gradient (MADDPG) within a centralized training and decentralized execution (CTDE) framework. The goal is not only to increase throughput but also to meet constraints related to energy usage, task completion, and cooperation. A structured reward is introduced to connect local actions to global stability and to reduce issues such as reward exploitation and policy collapse.
We evaluate the proposed framework in a simulation environment that reflects key operational characteristics of NTNs. The environment includes dynamic user traffic, fading channels, HAP mobility, LEO orbital motion, and interactions among multiple agents. Several channel prediction models and MARL baselines are incorporated for comparison. Experimental results show that the proposed FMA-MADDPG method improves throughput, reduces delay, and enhances coordination efficiency relative to these baselines.
In summary, this work contributes to NTN research by integrating channel prediction, constraint-aware multi-agent learning, and multi-layer cooperation within a single framework. Existing studies typically focus on either channel prediction or RL-based resource allocation in isolation, or combine standard predictors (such as RNNs, CNNs, or Transformers) with RL in terrestrial or simplified satellite scenarios. In contrast, our framework advances NTN design in three main aspects. First, at the architectural level, we introduce the FMA model, which integrates a Mamba state-space backbone, multi-head attention, and attention-based temporal pooling to capture multi-scale channel dynamics and to improve decision robustness. Second, at the learning-strategy level, we embed this long-context predictor into a constraint-aware MARL scheme based on MADDPG with CTDE. Predicted CSI augments agent observations, while fairness and delay are incorporated through a shaped reward. Third, at the system-design level, we construct and evaluate a joint prediction-and-scheduling solution for a realistic three-layer UE–HAP–LEO NTN, rather than optimizing prediction and resource allocation separately. By addressing channel uncertainty, partial observability, and constrained optimization within this combined prediction–MARL setting, the proposed system provides a practical approach for improving resource allocation in future 6G NTNs.

2. Related Works

The rapid development of NTNs has attracted growing interest in intelligent and adaptive communication under dynamic and resource-limited conditions. In such systems, two related problems are central. The first is how to design distributed but coordinated decision policies for resource scheduling [8]. The second is how to sense or predict time-varying channels, since CSI quality directly affects what each agent can observe and how well it can learn [9]. Most existing studies address only one of these problems. They either apply deep reinforcement learning (DRL) to optimize resource allocation, or use sequential models for channel estimation, but they do not treat prediction and control together in the same NTN framework. In this section, we briefly review DRL-based scheduling and channel prediction work that is relevant to our setting.

2.1. Deep Reinforcement Learning for Resource Scheduling

Single-agent DRL methods such as DQN, DDPG [10], and PPO have been applied to power control, spectrum sharing, and task offloading. However, these methods face notable limitations in systems with many users or distributed controllers. As the number of agents grows, the environment becomes highly non-stationary and the state information becomes only partially observable, making single-agent DRL difficult to scale.
Multi-agent reinforcement learning (MARL) addresses these issues by letting multiple agents learn policies jointly. Algorithms such as MADDPG, QMIX, and MAPPO support decentralized execution while using centralized critics or shared information during training [11]. In wireless communication, MARL has been applied to mobile edge computing (MEC), UAV-assisted networks, and vehicular networks. In NTNs, however, MARL still faces several difficulties. The network topology changes due to satellite motion, channel quality is strongly affected by fading and blockage, and reward signals may be delayed or sparse [12]. In addition, many MARL schemes focus on a single metric, for example throughput or delay. They often do not model constraints that are important in NTNs, such as fairness among users, energy usage, or explicit cooperation between different layers. This limits their applicability in hierarchical UE–HAP–LEO systems.

2.2. Channel Prediction in Dynamic Wireless Environments

Accurate and timely CSI is important for good scheduling decisions, especially in mobile systems with fast channel variations. In NTNs, Doppler shifts and moving platforms make this problem even more pronounced. Traditional model-based estimators, such as autoregressive (AR) models, Kalman filters, and least-squares interpolation [13], can work well in slowly changing channels. They often perform poorly, however, when fading is fast, nonlinear, or strongly frequency-selective.
Data-driven methods based on deep learning have improved channel prediction in many wireless settings. Recurrent neural networks (RNNs), including long short-term memory (LSTM) and gated recurrent unit (GRU) models, are widely used to capture temporal correlations [14]. Later, hybrid architectures and temporal convolutional networks (TCNs) were introduced to extend the receptive field and reduce training cost. Attention-based models, such as the Transformer, further improved long-range sequence modeling, but they often require high computational resources when the sequence is long.
More recently, state-space sequence models (SSMs) such as S4 and its efficient variant Mamba [15] have been proposed. Mamba can model very long sequences with structured state spaces and linear complexity in the sequence length. This makes it a promising tool for NTN channel prediction, where long time windows are needed to capture slow orbital dynamics together with fast fading effects. However, the use of such long-context state-space models for NTN channel prediction is still limited in the current literature.
A common limitation in most channel prediction work is that prediction is treated as an isolated task, separate from resource allocation or control [16]. In NTNs, where agents’ capacity to perceive the environment is constrained by channel conditions, this separation is problematic. It is more natural to feed predicted CSI directly into the learning loop, so that agents can use both current and anticipated future states when choosing actions. Only a few studies attempt such a tight integration [17]. In addition, NTN channels show multi-scale behavior (fast fading versus slower orbital motion) and irregular degradation (such as sudden blockage). This calls for predictors that can extract fine-grained features and at the same time remain stable over long time horizons. Existing single-structure models do not fully address this need and motivate hybrid designs. To the best of our knowledge, no prior work for NTNs combines a state-space-based channel predictor, prediction-augmented multi-agent RL, and explicit constraint-shaped rewards in a unified three-layer UE–HAP–LEO architecture.

3. Methodology

This section introduces the proposed NTN uplink optimization framework. We first describe the system architecture and channel models. We then give the problem formulation. After that, we explain how the Fusion of Mamba and Attention (FMA) channel prediction model is integrated with the MADDPG-based multi-agent scheduling scheme.

3.1. System Architecture and Network Model

As shown in Figure 1, we consider an NTN uplink with two types of links: UE–HAP (uplink 1) and HAP–LEO (uplink 2). UEs send data to HAPs using OFDM. HAPs then forward the aggregated traffic to LEO satellites using NOMA. This layered design reflects the different roles of the ground, air, and space segments, and allows resource management to be coordinated across tiers. The right-hand side of Figure 1 shows how the FMA-based channel predictor and the multi-agent actor–critic network are combined, with predicted CSI fed into the decision module for resource allocation.
The system includes a set H = { 1 , , H } of HAPs deployed at identical altitudes. Each HAP covers a sub-area and serves a subset of UEs denoted by Z = { 1 , , Z } . This assumption simplifies the geometry and reflects typical HAP deployment where platforms operate in a narrow altitude band. The sub-areas are represented by A = { 1 , , A } . The 3D position of HAP h is d HAP = { x h , y h , t h } . Within each sub-area, UEs are randomly distributed and transmit data to their associated HAPs using the available sub-channel resources [18].
Each HAP is allocated an equal share of orthogonal sub-channels under OFDM. The total bandwidth is divided into | C Z | orthogonal sub-channels. At each time slot k K , where K = { 1 , , K } , the bandwidth assigned to the link between UE z and HAP h is
B z , h ( k ) = B Tot UE HAP | C Z | ,
subject to
z = 1 Z B z , h ( k ) B Tot UE HAP ,
where B Tot UE HAP is the total bandwidth for all UE–HAP links and C Z = { 1 , , C Tot } is the set of sub-channels. In most cases the number of UEs exceeds the number of sub-channels, i.e., | Z | > | C Z | , so an efficient allocation strategy is needed to avoid congestion.
The space segment is modeled by a set L = { 1 , , L } of LEO satellites that form a constellation. Each satellite has a sub-channel index set Y = { 1 , , Y tot } , where Y tot is the number of sub-channels in uplink 2. The resources assigned to each HAP are denoted by J R = { J 1 , , J h , , J H } . Each LEO satellite also supports a set ϑ l = { 1 , , r tot l } of sub-carriers, where r tot l is the total number of sub-carriers for satellite l. UEs first send data to HAPs, and HAPs then forward the data to the LEO satellites using NOMA. The satellites apply successive interference cancellation (SIC) to decode the superimposed signals from different HAPs. For simplicity, each UE is assumed to have a single antenna.
The UE–HAP link uses OFDM. OFDM is robust to multipath fading and is widely used in terrestrial links, which makes it suitable for dense terrestrial–aerial communication [18]. For the HAP–LEO link, we use NOMA. NOMA can handle heterogeneous traffic and varying user densities and can increase spectral efficiency by power-domain multiplexing. LEO satellites are assumed to have sufficient processing capability to run SIC and separate the received signals from multiple HAPs. Although NOMA increases receiver complexity, it is a reasonable choice for HAP–LEO links, since the processing can be done on the satellite side [19]. By adjusting transmit power levels of different HAPs, NOMA can also improve fairness, because nodes with better channels can use less power, while others can be assigned more power when needed, without reducing overall system efficiency.
In our model, multiple HAPs sharing the same HAP–LEO sub-channel are multiplexed in the power domain. At the LEO receiver, their superimposed signals are ordered according to the effective channel gains and decoded successively using SIC, assuming that the resulting signal-to-interference-plus-noise ratios are sufficient for reliable detection. After decoding, the contribution of each recovered HAP signal is subtracted from the composite received signal before decoding the next one, and any residual interference is treated as noise in the achievable-rate calculations. The channel prediction module focuses on forecasting the large-scale and small-scale gains of the individual HAP–LEO links; the aggregate interference level on a shared sub-channel is implicitly determined by the simultaneous transmissions from the scheduled HAPs.

3.2. Channel Models

The position of the z-th UE associated with the h-th HAP at time slot k is
d UE ( k ) = { x z ( k ) , y z ( k ) , t z ( k ) } .
Each UE has an initial traffic demand D z to be transmitted. We use φ h , z ( k ) { 0 , 1 } to denote the association indicator between UE z and HAP h at time k:
φ h , z ( k ) = 1 , if UE z is associated with HAP h , 0 , otherwise .
For simplicity, we assume that UEs are static during an episode. This matches many remote sensing scenarios with fixed ground sensors and allows us to focus on NTN-specific channel dynamics. Each UE can be associated with at most one HAP in a time slot, so we have
h H φ h , z ( k ) 1 , z Z , k K .
Power allocation is another key factor in the UE–HAP link. Let P = [ P 1 , , P Z ] T be the vector of transmit powers for all UEs. We write the allocation compactly as
P = P 1 , , P Z T , z Z ,
and constrain the link-level power by
0 P ( z , h ) ( k ) P z max .
This means that the transmit power from UE z to HAP h at time k cannot exceed the hardware limit P z max and cannot be negative.
The instantaneous SNR at HAP h from UE z at time slot k is modeled as
θ ( z , h ) ( k ) = φ h , z ( k ) R c f P ( z , h ) ( k ) G ( z , h ) ( k ) T h ( k ) F h C B B z , h ( k ) T s n ,
where G ( z , h ) ( k ) and T h ( k ) are the transmit and receive antenna gains, F h represents the path loss term (including free-space loss and large-scale attenuation), and R c f captures small-scale fading on the UE–HAP link. The factor C B T s n combines the noise spectral density and receiver temperature, and B z , h ( k ) is the bandwidth assigned to this link. Equation (5) thus follows a standard link-budget form for the air-to-ground channel. In this expression, φ h , z ( k ) { 0 , 1 } is an association indicator that guarantees that the SNR is zero whenever UE z is not served by HAP h at slot k. The coefficient R c f models small-scale fading on the UE–HAP link and is drawn from a Rayleigh block-fading process with unit average power. The product G ( z , h ) ( k ) T h ( k ) F h aggregates the transmit and receive antenna gains and the large-scale path loss, which is computed from a standard air-to-ground link budget including free-space loss and log-normal shadowing. The denominator C B T s n B z , h ( k ) corresponds to the thermal noise power over the allocated bandwidth, given a receiver temperature T and noise spectral density s n .
Given the SNR, the instantaneous rate on the UE–HAP link is
Φ ( z , h ) ( k ) = B z , h ( k ) log 2 1 + θ ( z , h ) ( k ) ,
where the Shannon formula is used as an upper bound on achievable rate. Under the equal-bandwidth allocation in (1), this expression reduces to
Φ ( z , h ) ( k ) = B Tot UE HAP | C Z | log 2 1 + θ ( z , h ) ( k ) .
Here B Tot UE HAP is the total bandwidth for all UE–HAP links and | C Z | is the number of sub-channels.
To evaluate fairness among users, we adopt Jain’s fairness index:
I f ( k ) = z = 1 Z Γ ^ z ( k ) 2 | Z | z = 1 Z Γ ^ z ( k ) 2 ,
where Γ ^ z ( k ) denotes the cumulative throughput of UE z up to time slot k. If all UEs obtain similar cumulative throughput, the index I f ( k ) is close to 1; if a few UEs dominate, the index drops towards 0. This index is used later in the reward function to encourage balanced service across UEs.
We also include end-to-end latency in the channel model. The delay for data from UE z to HAP h at time k is
ϖ ( z , h ) ( k ) = D z Φ ( z , h ) ( k ) + Δ ( μ 1 ) + ζ ( z , h ) ( k ) C l ,
where D z / Φ ( z , h ) ( k ) is the transmission time for payload D z , Δ ( μ 1 ) models queuing delay, and ζ ( z , h ) ( k ) / C l is the propagation delay on the UE–HAP path. This expression exposes the trade-off between rate, queuing delay, and propagation delay. It also shows that throughput, fairness, and latency must be considered together when designing the NTN resource allocation policy [20]. The above channel and delay models deliberately abstract away several propagation effects that may be present in real NTN deployments. Examples include detailed atmospheric absorption and scintillation, rain attenuation at higher frequency bands, Earth curvature over very long links, and spatial correlation among neighboring links and beams. We adopt this intermediate level of fidelity in order to preserve the dominant large-scale geometry, mobility-induced path loss variation, and small-scale fading dynamics that are critical for evaluating learning-based schedulers, while keeping the state dimension and parameter space manageable for extensive multi-agent training and ablation studies. Consequently, the absolute performance values reported in Section 4 should not be interpreted as predictions for a particular commercial system, but rather as relative comparisons among different prediction and control schemes under the same controlled conditions. Extending the framework to standardized NTN channel models and richer propagation phenomena is left for future work.
Based on the above network and channel models, we formulate the uplink optimization problem as a joint resource allocation task. At each time slot k, we define a scalar system reward R tot ( k ) that aggregates several performance aspects, including total capacity, average end-to-end delay, Jain’s fairness index, and the number of cooperation events between HAPs and LEOs. The detailed form of this composite reward and the corresponding normalization are given in Section 3.3, see (13). Formally, the optimization problem is to maximize the cumulative system reward over a finite horizon under UE–HAP association constraints, power limits, sub-channel availability, fairness requirements, and latency thresholds. The constraints are given by the association rules in (2)–(3), the power constraints in (4), the capacity relations in (6)–(7), the fairness index in (8), and the latency bound in (9). Because the problem mixes binary association variables with continuous power variables and includes nonlinear rate and delay terms, it falls into the class of mixed-integer nonlinear programming (MINLP), which is NP-hard and difficult to solve for large NTN systems. Over a finite horizon of K time slots, the corresponding optimization problem can be written compactly as
max { φ h , z ( k ) } , { P ( z , h ) ( k ) } k = 1 K R tot ( k ) s . t . φ h , z ( k ) { 0 , 1 } , h H , z Z , k K , h H φ h , z ( k ) 1 , z Z , k K , 0 P ( z , h ) ( k ) P z max , h H , z Z , k K , z Z B z , h ( k ) B Tot UE HAP , h H , k K , Φ ( z , h ) ( k ) given by ( 6 ) ( 7 ) , h H , z Z , k K , I f ( k ) I f min , k K , ϖ ( z , h ) ( k ) ϖ max , h H , z Z , k K ,
where R tot ( k ) denotes the composite system reward at slot k, I f min is a target fairness level, and ϖ max is the maximum tolerable end-to-end delay. The association constraints ensure that each UE is connected to at most one HAP per slot, the power and bandwidth constraints enforce hardware and spectral limits, and the fairness and delay constraints reflect service quality requirements. Together with the nonlinear rate and delay relations in 6–9, this yields a mixed-integer nonlinear program that is difficult to solve optimally in real time for large NTN systems.
To handle these difficulties, we do not solve this mixed-integer nonlinear program directly at run time. Instead, we adopt a predictive and learning-based approach. Channel state information is first estimated and predicted by the FMA module, and the predicted channel evolution is then used to guide decisions in a multi-agent deep reinforcement learning framework [21]. In this framework, the original optimization problem is cast as a Markov decision process: the state contains the (predicted) channel features and queue information, the actions represent association and power control decisions, and the reward reflects the composite objective R tot ( k ) .

3.3. DRL and Training Model for Multi-Agents

In the proposed three-layer space–air–ground network, the resource allocation problem involves many UEs, HAPs, and LEO satellites. The links have different characteristics, the traffic is time-varying, and the wireless channels change over time. Each network node needs to make sequential decisions on UE association, bandwidth allocation, and power control in a distributed but still coordinated way [22]. This naturally leads to a multi-agent decision-making problem. We model it as a partially observable Markov game (POMG) defined by
G = N , 𝒮 , { 𝒪 i } i N , { A i } i N , P , { r i } i N , γ ,
where N = { 1 , 2 , , N } is the set of agents, 𝒮 is the global state space, 𝒪 i is the local observation of agent i, A i is its action space, P is the state transition function, r i is the reward for agent i, and γ ( 0 , 1 ) is the discount factor.
At time slot t, the global state is
s t = H t UE HAP , H t HAP LEO , B t , P t , Q t ,
where H t UE HAP and H t HAP LEO are the CSI matrices of the UE–HAP and HAP–LEO links, B t is the vector of available bandwidths, P t is the available power, and Q t is the vector of queue lengths at the HAPs. In practice, each agent i only has access to a local observation o t i of the global state. This observation contains its local CSI h t i , its allocated bandwidth b t i , its transmit power p t i , its buffer state q t i , and the predicted CSI H ^ t pred from the channel prediction module [23]. Using predicted CSI in o t i is important in NTNs, because it lets agents base decisions on both current and expected future channel conditions instead of only reacting to the latest measurement. In (10), r i denotes the reward received by agent i; in our design we use a shared global reward so that all agents receive the same scalar signal r i ( t ) = R t defined below.
The action of agent i at time t is a vector
a t i = α u , h i , β h i , p h i ,
where α u , h i { 0 , 1 } indicates UE–HAP association, β h i is the bandwidth fraction assigned to HAP h, and p h i is the transmit power used by HAP h towards the LEO. The joint action a t = ( a t 1 , , a t N ) affects both local link performance and the global behavior of the network.
To steer learning towards cooperative and fair use of resources, we design the reward as a weighted sum of four terms:
R t = λ 1 C t C max λ 2 D t D max + λ 3 J t + λ 4 E t collab E max ,
where C t is the total capacity, D t is the average end-to-end delay, J t is Jain’s fairness index, and E t collab is the number of successful HAP–LEO collaboration events at time t. The coefficients λ 1 , λ 2 , λ 3 , λ 4 control the relative weight of each term. In our implementation, when these metrics violate preset thresholds, we convert them into penalty terms. This reduces free-riding behavior that can occur when agents only receive a shared global reward [24]. Here C max , D max , and E max are normalization constants that scale the corresponding terms to comparable numerical ranges in the canonical scenario. In our experiments, we choose λ 1 = 1.0 , λ 2 = 0.5 , λ 3 = 0.5 , and λ 4 = 0.8 . We first fixed λ 1 as the reference scale for the capacity term and then varied λ 2 , λ 3 , and λ 4 over a small set of candidate values in preliminary runs. Very large delay and fairness weights (for example, values on the order of 1.0) made the policies overly conservative and reduced the total reward, whereas very small weights led to unstable delay and degraded fairness. The reported choice provides a balance in which none of the components dominates the reward and all four metrics improve jointly; we did not attempt an exhaustive optimization over all weight combinations. In summary, the per-slot association, power, and bandwidth limits in (2)–(4) are enforced as hard constraints through the action parameterization, while the fairness and delay requirements from (8)–(9) are treated as soft constraints via the shaped reward in (13).
We adopt the Multi-Agent Deep Deterministic Policy Gradient (MADDPG) algorithm under the CTDE paradigm. The overall MADDPG-based multi-agent DRL framework under CTDE is illustrated in Figure 2. Each agent i maintains a deterministic policy
μ θ i : 𝒪 i A i ,
with parameters θ i , and a centralized critic
Q ϕ i ( s , a ) ,
with parameters ϕ i , that takes as input the global state s and the joint action a . The critic is trained by minimizing the temporal-difference (TD) error:
L ( ϕ i ) = E y i Q ϕ i ( s t , a t ) 2 ,
where
y i = r i + γ Q ϕ i s t + 1 , a t + 1 , a t + 1 j = μ θ j ( o t + 1 j ) ,
and θ j , ϕ i are the parameters of the target actor and target critic networks.
The actor parameters θ i are updated using the deterministic policy gradient:
θ i J E θ i μ θ i ( o i ) a i Q ϕ i ( s , a ) | a i = μ θ i ( o i ) .
Training is carried out in the simulated SAGIN environment. At each time step, each agent chooses an action from its policy, receives a reward and the next observation, and stores the transition ( o t , a t , r t , o t + 1 ) in a replay buffer. We then sample mini-batches from this buffer to update the actor and critic networks by gradient descent [25]. In CTDE, the critic has access to the global state and joint action during training, which improves value estimation, while at test time each agent uses only its local observation, which keeps execution decentralized and scalable.
A key feature of our framework is the integration of the FMA channel prediction module into the DRL loop. Instead of using only instantaneous CSI, each agent augments its local observation with predicted CSI generated by the FMA model described in Section 3.4. This allows the policies to exploit temporal correlations in the channels and to adjust allocations in advance when a link is expected to improve or degrade. In our experiments, this integration reduces policy oscillations, speeds up convergence, and improves robustness under the rapid channel changes that are typical of LEO-based NTNs, leading to higher throughput and system efficiency than the tested baseline methods.

3.4. Fusion of Mamba and Attention Model for Channel Prediction

As discussed in Section 1, fast-varying and partially observed channels make it desirable to use predicted CSI in NTN scheduling. To this end, we build an FMA model as a sequence predictor. It is a hybrid architecture that combines a state-space backbone with an attention block. The FMA module serves as a prediction front-end for the multi-agent DRL scheduler and provides each agent with an estimate of near-future CSI. This extra information helps to reduce policy oscillations and performance drops when channels vary fast. We choose Mamba because it can model long CSI sequences with linear complexity in the sequence length and matches well the continuous-time dynamics of NTN channels.
As shown in Figure 3, the FMA architecture has three main parts. The first part is a Mamba state-space backbone that extracts long-term temporal features from CSI histories with linear-time complexity. The second part is a multi-head self-attention (MHSA) block that focuses on short-term variations and local structures. The third part is a fusion layer with residual connections, normalization, and a feed-forward network, which combines the two feature sets into a single predictive embedding. A final output head maps this embedding to the predicted CSI.
Formally, let h t R d be the CSI vector at time slot t. It can include large-scale path loss, small-scale fading, Doppler shift, and shadowing. Given a history window
X t = [ h t L + 1 , , h t ] ,
the FMA model learns a mapping
h ^ t + Δ = F Θ ( X t ) ,
where Δ is the prediction horizon. In our implementation, we use a history length of L = 10 time slots and predict Δ = 3 slots ahead, which we found to offer a good trade-off between prediction accuracy and computational cost.
The Mamba part is a state-space model (SSM) with time-varying and input-dependent transition matrices:
s τ + 1 = A τ s τ + B τ x τ , z τ = C τ s τ + D τ x τ .
Compared with standard RNNs or CNNs, Mamba updates its state with a selective memory mechanism and has a dynamic receptive field. This produces a sequence of hidden states
Z t = [ z t L + 1 , , z t ] ,
which encodes long-range channel behavior such as orbital periodicity, slow fading trends, and repeated blockage events.
The MHSA block is then applied to Z t . While Mamba captures global trends, it may smooth out some abrupt changes. The MHSA module computes attention weights across all time positions:
Att ( Q , K , V ) = softmax Q K d k V ,
where queries, keys, and values are linear projections of Z t . Using multiple heads allows the model to focus on different temporal regions at the same time, for example recent Doppler spikes or fading dips. The output of this block is a feature sequence U t that preserves fine-grained CSI changes.
We then fuse the outputs of Mamba and MHSA:
F t = LayerNorm ( Z t + U t ) ,
followed by a position-wise feed-forward network:
F t = GELU ( F t W 1 + b 1 ) W 2 + b 2 .
An attention-based temporal pooling layer then aggregates the sequence into a compact vector f t , and the output layer predicts the future CSI as
h ^ t + Δ = W pred O f t + b pred O .
The FMA predictor is trained with a mean squared error loss:
L FMA = 1 B b = 1 B h ^ t + Δ ( b ) h t + Δ ( b ) 2 2 ,
where B is the batch size. In some experiments, we further fine-tune FMA together with the DRL scheduler using a combined loss
L total = L DRL + η L FMA ,
where η is a weighting factor.
Compared with a Transformer predictor, whose self-attention cost is O ( L 2 ) in the sequence length L, FMA keeps the Mamba backbone at O ( L h ) and adds O ( H L d k ) for the MHSA block. More specifically, for batch size B the overall time complexity of FMA is 𝒪 ( B L h + B H L d k ) with memory cost 𝒪 ( B L h ) , since no L × L attention matrices are formed. In contrast, a vanilla Transformer-based predictor has time complexity 𝒪 ( B L 2 d k ) and memory complexity 𝒪 ( B L 2 ) due to the quadratic self-attention. This linear dependence on L makes it practical to use longer CSI histories on resource-constrained HAP or LEO platforms. With the configuration in Table 1 (hidden size h = 128 , four state-space blocks, four attention heads, and a short history window), the resulting model has a few hundred thousand parameters and a per-step computational cost that is comparable to lightweight sequence models commonly used on embedded devices, making real-time inference on HAP or LEO payload processors feasible in principle. Long histories are useful in NTN because they contain both slow orbital trends and fast fading effects. In our design, the Mamba part mainly stabilizes predictions by modeling global patterns, while the MHSA part makes the predictor more sensitive to sudden changes. Compared with generic hybrid state-space–attention architectures, our FMA design is tailored to NTN channel prediction in three ways. First, multi-head self-attention is applied only on top of the Mamba-generated hidden sequence, rather than on the raw CSI input, so that the overall time and memory complexity remain linear in the history length. Second, an attention-based temporal pooling layer is introduced to emphasize the most relevant time steps when forecasting h t + Δ under multi-scale NTN dynamics. Third, the predictor is not used in isolation but is tightly coupled with a constraint-aware MADDPG scheduler under the CTDE paradigm, so that its capacity and receptive field are chosen with the joint prediction-and-control task in mind.
When FMA is used in front of the MARL scheduler, each agent receives both current CSI and its predicted continuation in its observation vector. For example, a HAP can reduce bandwidth assigned to a link that is predicted to degrade, or postpone a high-rate transmission until a predicted Doppler peak has passed. In the experiments, such prediction-aware decisions reduce the number of times that constraints are violated in the reward function and lead to higher total reward, better system efficiency, and more frequent collaboration events compared with using instantaneous CSI alone [26].

3.5. FMA-MADDPG for Cooperative Scheduling

To make full use of the predicted CSI from the FMA model and to cope with the dynamic and heterogeneous nature of NTNs, we build an integrated FMA-MADDPG framework for cooperative uplink scheduling [27]. Unlike MARL schemes that rely only on instantaneous CSI, our approach feeds predicted CSI into the learning loop. In this way, agents can make scheduling decisions that take expected channel changes into account instead of reacting only to the latest measurements.
At each time slot t, the network collects the current system state, including CSI on the UE–HAP and HAP–LEO links H t UE HAP and H t HAP LEO , the available bandwidth B t , the available transmission power P t , and the queue lengths Q t at the HAPs. For agent i, the initial local observation at time t is
o t i = h t i , b t i , p t i , q t i ,
where h t i is the instantaneous CSI relevant to agent i, b t i is its available bandwidth, p t i is its available power, and q t i is its local queue length.
At the same time, the CSI history
X t = [ h t L + 1 , , h t ]
is sent to the FMA predictor, which outputs predicted CSI for the next Δ time steps:
h ^ t + Δ = F Θ ( X t ) .
We then build an enhanced observation by concatenating current and predicted information:
o ˜ t i = o t i , h ^ t + Δ .
This ensures that each agent sees both the present channel state and an estimate of its near-future evolution.
Each agent feeds the enhanced observation o ˜ t i into its actor network μ θ i to obtain a composite action:
a t i = μ θ i ( o ˜ t i ) = α u , h i , β h i , p h i ,
where α u , h i is the UE–HAP association decision, β h i is the bandwidth allocation ratio, and p h i is the power level for HAP–LEO transmission [28]. The joint action of all agents, a t = ( a t 1 , , a t N ) , controls link usage and therefore affects system throughput, delay, and fairness. Under the CTDE setting, the critic network receives the global state s t and the joint action a t and evaluates their quality during training.
Once the actions are chosen, they are executed in the three-layer network: each UE is bound to a HAP according to α u , h i , bandwidth fractions are assigned to UE–HAP links according to β h i , and power levels on HAP–LEO links follow p h i . This changes the system state, queue lengths, and load for the next time slot. After execution, we compute the system-level feedback and give each agent a shared reward:
R t = λ 1 R t tot R max λ 2 E t E max λ 3 D t D max + λ 4 C t collab C max ,
where R t tot is the total throughput at time t, E t is the energy consumption, D t is the average transmission delay, and C t collab is the number of successful cooperation events. The weights λ 1 , , λ 4 balance throughput, energy–delay trade-off, load balancing, and cooperation. The terms R max , E max , D max , and C max are corresponding normalization constants chosen according to typical values observed in the canonical scenario so that the different components of the reward have comparable scales.
The FMA-MADDPG framework can be viewed as a data flow that includes CSI measurement, prediction, observation construction, action selection, and reward feedback. Predicted CSI extends the agents’ temporal horizon, the reward design promotes cooperation, and CTDE keeps training stable even when channel conditions and topology change quickly. The reward in (30) follows the same constraint-shaped design as (13), combining throughput, energy, delay, fairness, and cooperation into a single scalar signal. Overall, we first formulate the uplink resource allocation as a mixed-integer nonlinear program with explicit association, power, fairness, and delay constraints, and then approximate it by a prediction-augmented multi-agent RL scheme where hard per-slot limits are enforced at the action level and long-term fairness and delay are encoded through a shaped reward.

4. Experiments and Results

This section evaluates the proposed FMA-MADDPG framework in a simulated three-tier NTN scenario. We study both the FMA channel prediction module and the cooperative scheduling performance of the multi-agent RL scheme. Each simulation run includes multiple episodes with different traffic levels and mobility patterns. We focus on several key metrics [29,30,31]: total reward, system efficiency, number of collaboration events, and load balance. Together, these metrics reflect throughput, fairness, and cooperation across the UE–HAP–LEO layers.

4.1. Experimental Settings

To evaluate the proposed framework, we build a simulation dataset that emulates channel dynamics and traffic patterns in a three-layer SAGIN [32,33]. The dataset is generated using a Python 3.8 simulator [34]. In this environment, low-Earth-orbit (LEO) satellites follow predefined orbits and interact with HAPs and distributed UEs. The channel traces for UE–HAP links include large-scale path loss, log-normal shadowing, and Rayleigh fading. The HAP–LEO channels additionally include Doppler shifts and slowly varying shadowing caused by satellite motion. Each simulation episode produces a time series of CSI and traffic arrivals, which we use to train and test both the prediction and the scheduling modules. This setup allows controlled but still realistic tests under different traffic and mobility conditions. All results in Figure 4, Figure 5, Figure 6 and Figure 7 and Table 2 are obtained under one canonical NTN uplink configuration defined in Section 3.1 and this subsection. For each method, we train in this canonical scenario and then evaluate the learned policies over multiple test episodes with independently generated traffic and channel realizations. Due to computational constraints, we train one policy realization (i.e., one random initialization and one full training run) per learning-based method. The reported curves and scalar values are therefore averages over multiple evaluation episodes for a single trained policy per method, and we do not provide confidence intervals or statistical hypothesis tests; the results should be viewed as representative for this configuration rather than as statistically guaranteed performance bounds.
In the canonical UE–HAP–LEO scenario, each high-altitude platform (HAP) is deployed at an altitude of 20 , 000 m and each low-Earth-orbit (LEO) satellite orbits at 600 , 000 m. Every HAP has a computing capability of 8.0 GHz and can handle up to 20 concurrent tasks, while each LEO satellite has a computing capability of 15.0 GHz and a maximum load of 50 tasks. To reflect the different propagation characteristics of the links, the large-scale channel gains are initialized within predefined ranges in dB: 5–25 dB for UE–HAP links (short-range, relatively high gain), 10–30 dB for HAP–LEO links (more stable, high gain), and 5 –15 dB for UE–LEO links when such links are considered [35].
For the learning setup, we use N agents = 4 cooperating agents. The local observation (state) dimension is 15 and the action dimension is 4, consistent with the UE–HAP–LEO resource control problem. The actor networks follow a fully connected architecture of 15 256 128 64 4 , and the critic networks use a 76 512 256 128 64 1 architecture. These sizes are chosen to provide sufficient representational power for the joint prediction-and-scheduling task while keeping the computational cost moderate for potential deployment on HAP or satellite platforms. At deployment time, only the actor networks are used for inference, so the per-step cost of the MARL policy reduces to a few fully connected layers per agent. All reported metrics in Section 4 are obtained by averaging over multiple training and evaluation episodes with independently generated channel and traffic realizations for each trained policy.
In the canonical configuration, we simulate 50 ground UEs and 3 HAPs that forward traffic to a small constellation of 4 LEO satellites. The three HAPs are deployed at the same altitude with identical hardware capabilities and roughly symmetric coverage regions. The carrier frequency used in the path-loss calculations is 2.4 GHz. The total bandwidth available for the UE–HAP links is 100 MHz, which is partitioned into multiple OFDM sub-channels. Each LEO satellite has a per-satellite bandwidth of 50 MHz for HAP–LEO transmission. User transmit powers are limited to 16 dBm, while each HAP can transmit up to 30 dBm towards the LEO layer. Traffic arrivals at UEs follow independent stochastic processes with region-dependent mean arrival rates between 0.3 and 0.6 tasks per time slot. The regional mean loads are chosen to be similar, so that the canonical scenario is approximately symmetric in both HAP capabilities and long-term traffic intensity.
For all MADDPG-based schedulers and their baselines, we use a discount factor of 0.99, an actor learning rate of 1 × 10 4 , a critic learning rate of 5 × 10 5 , a replay buffer with capacity 100,000 transitions, and a mini-batch size of 256. The target actor and critic networks are updated using a soft update coefficient of 0.001. Each method is trained for a fixed number of episodes and then evaluated over a separate set of episodes with independently generated channel and traffic realizations; all curves and scalar metrics reported in Section 4 are averages over these evaluation episodes.

4.2. Performance Comparison of Channel Prediction Models

We first study how different channel prediction models affect the NTN scheduling performance. We compare five predictors under the same MADDPG scheduler: an LSTM, a TCN, a probabilistic TPM, a transformer-based predictor, and the proposed FMA model. Figure 4a,b show total reward and system efficiency. Figure 5a,b show the number of collaboration events and the prediction accuracy.
From the results, the FMA predictor achieves a total reward of about 3200, higher than TCN (2867), LSTM (2574), TPM (2431), and the transformer-based predictor (about 3050). The total reward summarizes throughput, latency, and energy terms in the reward function, so this gain suggests that the use of predicted CSI helps agents strike a better balance between data rate and power cost.
The system efficiency for a given scheduling method m is defined as the average throughput per unit total bandwidth and power over an evaluation episode of length T:
η sys ( m ) = 1 T t = 1 T C t ( m ) B Tot P Tot ,
where C t ( m ) is the total system throughput at time slot t under method m, and B Tot and P Tot denote the total available bandwidth and the total maximum transmit power in the canonical UE–HAP–LEO scenario, respectively. Since η sys ( m ) is a throughput-per-resource ratio rather than a probability or a normalized cumulative distribution, it is not mathematically restricted to the interval [ 0 , 1 ] . Values smaller than, equal to, or larger than one simply correspond to lower, equal, or higher levels of achieved bits per second per unit of aggregated bandwidth and power in the considered setting. The transformer-based model attains an efficiency close to 1.0 in the canonical scenario, while FMA reaches 0.95 and the LSTM and TPM baselines remain below 0.85. Since this metric is a throughput-per-resource ratio rather than a quantity normalized to a fixed upper bound, it is not restricted to the interval [ 0 , 1 ] ; values slightly below or above one correspond to lower or higher levels of bits per second achieved per unit of ( Hz · W ) . This means that prediction-aware schedulers use the same resources more effectively when longer temporal dependencies are exploited. The improvement is consistent across episodes and indicates that temporal foresight stabilizes scheduling decisions under fast channel variations.
The number of successful collaboration events between HAPs and LEOs reaches 956 under FMA, which is higher than the transformer-based predictor (about 915) and about 13.5% higher than the best non-transformer baseline. More collaboration events show that predicted CSI gives agents better awareness of cross-layer link conditions, so they can coordinate power control and association more often. FMA also attains a prediction accuracy of 0.78, compared with about 0.67 for the Transformer-based predictor, 0.64 for LSTM, 0.71 for TCN, and 0.68 for TPM. Since the Transformer serves as an attention-only baseline, this comparison supports the idea that combining Mamba’s state-space modeling with multi-head attention helps capture both long-term orbital patterns and short-term fading changes better than attention-only or purely recurrent/convnet-based models. A more fine-grained architectural ablation that explicitly isolates Mamba-only and attention-only variants of FMA is left for future work.
All five predictors in this subsection are paired with the same MADDPG-based scheduler and evaluated under the same canonical scenario, so the absolute total reward of the FMA-based configuration here matches the FMA-MADDPG result reported in Section 4.3. Overall, these results show that accurate and stable channel prediction is an important component for intelligent scheduling. By adding predicted CSI to the observations, FMA not only improves steady-state performance, but also makes the training process more stable, leading to smoother and more reliable learning curves.

4.3. Comparison with Other Deep Reinforcement Learning Frameworks

We next compare FMA-MADDPG with several multi-agent RL baselines (PPO, MAPPO, HAPPO, QMIX, and standard MADDPG) and with two non-RL baselines and a random scheduler. The first non-RL baseline, denoted Convex-CSI, approximates the per-slot joint association and power allocation under perfect instantaneous CSI by solving a convex-relaxed version of the uplink resource allocation problem. The second, denoted Greedy, is a myopic heuristic that associates each UE to the HAP with the best instantaneous channel quality, allocates bandwidth in proportion to instantaneous rates, and distributes transmit power according to a simple rule, without learning. We also include a purely random scheduler that selects feasible associations and power levels uniformly at random. All methods use the same network configuration and traffic settings for a fair comparison. The results in Figure 6, Figure 7, Figure 8 and Figure 9 and Table 3 show that FMA-MADDPG achieves higher performance than these baselines on most metrics in our experiments under the canonical scenario. All schedulers compared in this subsection share the same canonical NTN configuration as in Section 4.2; therefore, the FMA-MADDPG score refers to the same setting and should be interpreted in terms of relative differences to the other schedulers.
In terms of total reward, FMA-MADDPG achieves the highest average value of 3199.4, which is higher than HAPPO (2789.3), MADDPG (2749.1), MAPPO (2519.3), QMIX (2465.4), and PPO (1331.1). Among the non-RL baselines, Convex-CSI reaches an average reward of 2812.7, which is clearly better than the greedy heuristic (1974.6) and the random scheduler (1160.8) but still below the best RL-based methods. This shows that combining predicted CSI with a constraint-based cooperative scheduler can improve the overall return beyond both heuristic and convex-relaxation approaches. The number of collaboration events between the HAP and LEO layers is also highest for FMA-MADDPG (942 on average), followed by MADDPG (897) and HAPPO (858), while Convex-CSI (282), Greedy (100), and Random (36) exhibit much less frequent cooperation. This indicates that the prediction-aware MARL schemes are better able to exploit cross-layer links for coordination than the non-learning baselines in the dynamic NTN setting.
For load balance, FMA-MADDPG and MADDPG both achieve Jain’s indices close to one (0.997), while Convex-CSI and Greedy produce lower fairness values (0.920 and 0.825, respectively). These near-unity indices indicate very even traffic distribution among HAPs in the canonical scenario and are partly a consequence of the deliberately symmetric configuration, where HAPs have identical capabilities and regions have similar average traffic. Under such symmetric conditions and with a fairness-aware reward, Jain’s indices close to 1.0 are expected and should not be interpreted as typical for arbitrary NTN deployments. In more heterogeneous or larger-scale settings with asymmetric HAP capacities, more uneven user distributions, or different link budgets, we would expect lower fairness indices in absolute value, although we still anticipate that FMA-MADDPG would retain advantages relative to the baselines. In terms of system efficiency, FMA-MADDPG and MADDPG both attain values slightly above one (1.011 and 1.012, respectively), while Convex-CSI and HAPPO reach around 0.88–0.89 and the Greedy and Random baselines achieve much lower efficiencies (0.718 and 0.457). As discussed in (31), this metric is a linear throughput-per-resource measure and is therefore not mathematically bounded by one; values greater than one simply indicate that these schedulers achieve more than one unit of normalized throughput per unit of aggregated bandwidth and power in the canonical scenario. These results confirm that FMA-MADDPG delivers high throughput and efficient use of resources while maintaining fairness and cooperation advantages over both RL and non-RL alternatives in this canonical scenario.

4.4. Ablation Study on Prediction and Constraint Design

To further isolate the contributions of the channel prediction module and the constraint-aware reward, we conduct an ablation study under the same canonical UE–HAP–LEO configuration. We compare four variants built on the same MADDPG backbone: (A) a constrained scheduler without prediction, which uses the full structured reward but only instantaneous CSI; (B) an unconstrained scheduler without prediction, which removes the constraint penalties and fairness/cooperation terms; (C) an unconstrained scheduler with prediction, which removes the constraint penalties but augments the agents’ observations with predicted CSI; and (D) the full FMA-MADDPG scheme with both prediction and constraints.
Figure 10, Figure 11 and Figure 12 illustrate the learning curves of these four variants in terms of total reward, system efficiency, and collaboration events. The unconstrained and non-predictive variant (B) yields the lowest average total reward, around 2100, together with a degraded system efficiency of about 0.78 and roughly 450 collaboration events per episode. This behavior indicates that removing the constraint terms leads to unstable policies that can exploit the reward in the short term but eventually drift towards inefficient and poorly coordinated operation. Adding only the constraint-aware reward while still relying on instantaneous CSI (variant A) increases the average total reward to about 2400, maintains system efficiency around 1.0, and nearly doubles the number of collaboration events to about 900, showing that the constraint mechanism plays a key role in stabilizing learning and promoting cooperation even without prediction.
Introducing only the prediction module (variant C) produces a much larger improvement. In this case, the total reward rises to about 3000, the system efficiency recovers to approximately 1.0 from the degraded level seen in the unconstrained baseline, and the number of collaboration events increases to around 700. This demonstrates that predicted CSI by itself brings a substantial gain in both performance and coordination, as it allows agents to anticipate channel variations and to allocate resources more proactively. Finally, combining both prediction and constraints in the full FMA-MADDPG configuration (variant D) yields the best results: the total reward reaches about 3200, the convergence is the fastest among all variants, the system efficiency stabilizes slightly above 1.0, and the number of collaboration events approaches 950. Overall, these ablation results indicate that the constraint-aware reward is essential for stability and fairness, the prediction module provides a large boost in reward and efficiency, and their combination in FMA-MADDPG delivers the most stable and effective behavior. Together with the comparison to the Transformer-based predictor in Section 4.2, they suggest that both the presence of prediction and the specific Mamba + attention design matter for performance. A more detailed architectural ablation that separates Mamba-only and attention-only versions of FMA will be considered in future work.

5. Discussion

This section interprets the experimental results and explains how the FMA-MADDPG framework helps address key challenges in NTNs. This section interprets the experimental results and explains how the main components of FMA-MADDPG contribute to the observed improvements in total reward, collaboration frequency, load balance, and system efficiency. We discuss the roles of the prediction module, the reinforcement learning design, and their joint effect at the system level.

5.1. Impact of Channel Prediction on System Performance

Accurate channel prediction is important for link stability and resource allocation in NTNs, where channels change with user positions, weather, and satellite motion. In our experiments, the FMA predictor attains a prediction accuracy of 0.78, higher than the LSTM, TCN, and TPM baselines. This gain in prediction accuracy is reflected in the higher total reward and system efficiency, because agents with predicted CSI can anticipate link degradation and adjust power and bandwidth before performance drops.
The FMA module uses a Mamba state-space backbone together with multi-head self-attention. The state-space part captures long-term patterns, such as orbital periodicity and slow fading trends. The attention part focuses on short-term changes, such as Doppler fluctuations and sudden interference. In combination, these two parts produce predictions that are neither too smooth nor too noisy. This helps to reduce oscillations in the learned policies and improves convergence during training.
From a network operation point of view, better prediction allows UEs and HAPs to schedule more proactively. For example, a HAP can reduce bandwidth to a link that is predicted to experience deep fading in the next slots, or it can delay non-urgent transmissions when a Doppler peak is expected. Such decisions lead to smoother use of links, fewer retransmissions, and more balanced service, which is consistent with the observed load-balance index of 1.00 across HAPs. In this sense, channel prediction acts both as a performance booster and as a stabilizing factor for multi-agent coordination under time-varying conditions. For remote-sensing and Earth observation applications, higher throughput and lower delay directly improve the timeliness of sensing data delivery. They also enhance the reliability of coverage for missions such as disaster monitoring and environmental surveillance. These tasks represent typical service scenarios targeted by large-scale NTN deployments.

5.2. Effectiveness of Reinforcement Learning and Reward Design

The reinforcement learning part of FMA-MADDPG is designed to deal with two issues that often appear in MARL for NTNs: unstable convergence and free-riding. To this end, we use a reward that combines throughput, energy, delay, and cooperation, and that turns some metrics into penalties when they violate thresholds. This guides agents towards global objectives while still keeping local fairness in mind.
Concretely, the reward includes terms for capacity, average end-to-end delay, Jain’s fairness index, and the number of collaboration events. When delay is too high or fairness drops below a target, penalty terms are added. This reduces the misalignment that can appear in simple shared-reward schemes, where some agents might benefit from others’ actions without contributing much themselves.
The experimental results summarized in Table 3 support this design. FMA-MADDPG achieves the highest total reward and the largest number of collaboration events among all compared methods, indicating a high level of inter-agent coordination, especially between the HAP and LEO layers. At the same time, it maintains near-perfect load balance and high system efficiency, which means that resources are well used and traffic is spread evenly across HAPs. Taken together, these observations suggest that the constraint-aware reward formulation improves both stability and fairness under decentralized execution.
In preliminary sensitivity tests, we also varied the weights λ 2 λ 4 while keeping λ 1 fixed, to gain intuition about the effect of the constraint terms. Reducing the fairness weight λ 3 shifted the learned policies towards slightly higher total reward but visibly lower Jain’s fairness indices, indicating a more throughput-oriented behavior. Increasing λ 3 had the opposite effect, improving fairness at the cost of a small reduction in total reward and fewer collaboration events. Similarly, strengthening the delay penalty λ 2 led to lower average end-to-end delay but a modest loss in throughput, while too small a value of λ 2 made the delay distributions broader. These qualitative trends are consistent with the intended role of the weights as knobs that trade off throughput against delay, fairness, and cooperation; a full quantitative study of this trade-off is left for future work.
The CTDE setting also plays an important role. During training, the critics can use the full global state and joint action, which leads to better value estimates. At test time, each actor only needs its own local observation, which keeps the solution scalable and practical in large NTN systems where full central control is not realistic.

5.3. Overall Comparison and System-Level Insights

When compared with other MARL methods such as PPO, MAPPO, HAPPO, QMIX, and standard MADDPG, as well as with the Convex-CSI and Greedy non-learning baselines, the FMA-MADDPG framework performs better on most of the evaluated metrics, as summarized in Table 3. PPO and MAPPO achieve only moderate rewards, HAPPO and standard MADDPG come closer but still fall short in terms of total reward and collaboration frequency, and QMIX attains good load balance but a lower total reward. The non-RL baselines, in particular the greedy heuristic and the random scheduler, exhibit significantly lower reward, efficiency, and cooperation levels, which underlines the advantage of prediction-augmented multi-agent RL in this dynamic NTN setting.
In contrast, FMA-MADDPG consistently improves total reward, load balance, and collaboration events. The prediction module helps bridge the gap between purely reactive control and proactive scheduling. The constraint-based reward ensures that the learned policies remain cooperative and fair, not just throughput-oriented. Together, these elements allow the framework to handle the complex and non-stationary conditions of the three-layer UE–HAP–LEO system.
Another observation is that FMA-MADDPG converges faster than the baselines. The availability of predicted CSI reduces uncertainty in the state transitions seen by the RL agents. As a result, the value function and policies stabilize more quickly, which shortens training time and can make deployment easier in practice.
We also note that, for each learning-based method, only a single policy realization is trained in the canonical NTN scenario due to the computational cost of joint prediction-and-scheduling experiments with multiple baselines. As a result, we do not report confidence intervals or statistical significance tests across random seeds. The curves and scalar metrics in Section 4 should be interpreted as representative examples under the given configuration, rather than as statistically guaranteed performance bounds. Extending the evaluation to multiple seeds with uncertainty quantification is left for future work.
From a computational perspective, the FMA predictor and the MADDPG policies were configured to be lightweight. The FMA model uses a hidden size of 128 with four state-space blocks and four attention heads, resulting in a few hundred thousand parameters and linear-time complexity in the history length, while each actor is a small fully connected network with four layers. Training is performed offline on the ground, and only inference of the FMA predictor and the actor networks would run on HAP or LEO hardware. Given these model sizes and the linear-time design, real-time inference at typical scheduling rates appears feasible for current HAP/LEO payload processors in principle, especially when using standard optimizations such as mixed-precision or fixed-point implementations. However, we have not yet measured end-to-end latency on specific hardware platforms, and a detailed hardware-in-the-loop evaluation is left for future work.
Beyond computational issues, several modeling assumptions may also affect how the framework behaves when they are relaxed. The assumption of static UEs removes Doppler effects due to user motion and keeps the UE–HAP geometry fixed within an episode; with mobile users we would expect faster variations in link quality and queue lengths, more frequent handovers, and a tighter requirement on the predictor to track these changes. The symmetric HAP deployment and approximately balanced regional traffic mainly influence the absolute level of the fairness indices: in more heterogeneous deployments with different HAP altitudes or capacities we would expect lower Jain’s indices, although the relative ordering between methods is likely to persist. Finally, the assumption of accurate historical CSI is probably the most critical for prediction quality; if the CSI fed to the FMA module is noisy or delayed, prediction accuracy and, in turn, scheduling performance can degrade. Studying robustness to user mobility, imperfect and delayed CSI, and heterogeneous HAP deployments is therefore an important part of our future work.
In future work, several promising directions will be pursued to further enhance the applicability and robustness of the proposed framework. First, real-world deployment and hardware-in-the-loop experiments will be conducted to validate the system performance under practical NTN conditions, including hardware impairments, synchronization latency, and imperfect CSI feedback. Second, the current FMA predictor can be extended toward adaptive or hierarchical structures, enabling dynamic adjustment of prediction horizons based on mobility levels or environmental variations. Third, lightweight model compression, pruning, or knowledge distillation techniques will be explored to reduce computational overhead, facilitating deployment in resource-constrained HAP or satellite platforms. Finally, integration with cross-layer optimization, network slicing, and semantic communication frameworks may further improve the intelligence and flexibility of resource management in large-scale 6G NTN ecosystems.

6. Conclusions

This paper proposed an FMA-MADDPG framework for uplink resource scheduling in NTNs. The framework combines channel prediction with constraint-based cooperative reinforcement learning. It is designed for a three-layer space–air–ground architecture and targets practical issues in large NTN systems, such as time-varying channels, heterogeneous traffic, and partial observability.
The core component is the FMA predictor. It uses a Mamba state-space backbone to model long-term temporal dependencies and a multi-head self-attention block to capture short-term fluctuations. This hybrid design provides accurate and stable channel forecasts and gives agents advance knowledge of link variations. By embedding predicted CSI into the observation space, the framework allows agents to make scheduling decisions that are not only based on the current state but also on expected future conditions.
On top of this prediction module, we design a constraint-based MADDPG strategy. The reward function jointly considers throughput, latency, energy efficiency, and fairness. Unlike standard multi-agent RL approaches that only share a global reward, our formulation introduces explicit performance thresholds and anti-idle penalties. This reduces free-riding behavior and encourages more balanced resource use across agents in different layers.
Simulations in a customized NTN environment show that, in our experiments under the canonical scenario, the proposed FMA-MADDPG method achieves higher performance than several baseline algorithms, including PPO, MAPPO, HAPPO, QMIX, and standard MADDPG. Under this canonical scenario, the framework attains the highest total reward, system efficiency, and number of HAP–LEO collaboration events among all compared methods, while maintaining a load-balance index very close to one. These results support the effectiveness of combining prediction with constraint-aware cooperative learning to obtain stable, fair, and efficient scheduling under dynamic network conditions. We note that the current evaluation is limited to one canonical NTN uplink scenario with static UEs, symmetric HAP deployment, and simplified channel models, and assumes accurate historical CSI; it does not yet examine scalability with respect to the number of UEs, HAPs, and LEO satellites, or robustness under blocked links, noisy CSI, and different Doppler regimes. Studying more diverse network settings, mobile users, standardized NTN channels, imperfect CSI feedback, and these scalability and robustness aspects will be part of our future work.
Overall, the FMA-MADDPG framework offers a scalable approach to resource management in future 6G space–air–ground integrated networks. Its use of prediction and cooperative decision making makes it suitable for mission-critical remote sensing applications, such as disaster monitoring, environmental observation, and intelligent navigation. In future work, we plan to extend this framework to real-world NTN testbeds, include more explicit multi-objective optimization for cross-layer control, and explore lightweight implementations for edge intelligence on HAP and satellite platforms.

Author Contributions

Conceptualization, C.Y., K.S. and J.B.; methodology, C.Y. and K.S.; software, C.Y. and K.S.; validation, C.Y., K.S. and C.L.; formal analysis, C.Y.; investigation, C.Y. and K.S.; resources, J.B., Y.Z. and Z.X.; data curation, C.Y. and C.L.; writing—original draft preparation, C.Y.; writing—review and editing, J.B., K.S., Y.Z., Z.X. and Y.S.; visualization, C.Y. and K.S.; supervision, J.B. and Z.X.; project administration, J.B.; funding acquisition, J.B. and Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 62276206.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

Authors Chunyu Yang and Yanhong Sun were employed by the company China Electronics Technology Group Corporation. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Tang, C.; Zhou, X.; Zhang, L.; Liu, Y.; Dan, Z. LEO Satellite Navigation Signal Multi-Dimensional Interference Optimisation Method Based on Hybrid Game Theory. Remote Sens. 2025, 17, 1444. [Google Scholar] [CrossRef]
  2. Guo, N.; Liu, L.; Zhong, X. Task-Aware Distributed Inter-Layer Topology Optimization Method in Resource-Limited LEO-LEO Satellite Networks. IEEE Trans. Wirel. Commun. 2024, 23, 3572–3585. [Google Scholar] [CrossRef]
  3. Wang, G.; Yang, F.; Song, J.; Han, Z. Resource Allocation and Load Balancing for Beam Hopping Scheduling in Satellite-Terrestrial Communications: A Cooperative Satellite Approach. IEEE Trans. Wirel. Commun. 2025, 24, 1339–1354. [Google Scholar] [CrossRef]
  4. Kawamoto, Y.; Kamei, T.; Takahashi, M.; Kato, N.; Miura, A.; Toyoshima, M. Flexible Resource Allocation with Inter-Beam Interference in Satellite Communication Systems with a Digital Channelizer. IEEE Trans. Wirel. Commun. 2020, 19, 2934–2945. [Google Scholar] [CrossRef]
  5. Zhou, G.; Jin, Z.; Liu, D. Deep Reinforcement Learning-Based Two-Phase Hybrid Optimization for Scheduling Agile Earth Observation Satellites. Remote Sens. 2025, 17, 1972. [Google Scholar] [CrossRef]
  6. Zhou, D.; Sheng, M.; Wang, Y.; Li, J.; Han, Z. Machine Learning-Based Resource Allocation in Satellite Networks Supporting Internet of Remote Things. IEEE Trans. Wirel. Commun. 2021, 20, 6606–6621. [Google Scholar] [CrossRef]
  7. Gao, H.; Wang, J.; Xu, X.; Yuan, Q.; Wang, P.; Zhou, D. Research on Distributed Collaborative Task Planning and Countermeasure Strategies for Satellites Based on Game Theory Driven Approach. Remote Sens. 2025, 17, 2640. [Google Scholar] [CrossRef]
  8. Wang, F.; Zhang, S.; Hong, E.-K.; Quek, T.Q.S. Constellation as a Service: Tailored Connectivity Management in Direct-Satellite-to-Device Networks. IEEE Commun. Mag. 2025, 63, 30–36. [Google Scholar] [CrossRef]
  9. Shang, B.; Li, C.; Ma, J.; Fan, P. Inter-Satellite Links-Enabled Cooperative Edge Computing in Satellite-Terrestrial Networks. IEEE Trans. Wirel. Commun. 2025. [Google Scholar] [CrossRef]
  10. Zhu, Y.; Chen, M.; Wang, S.; Hu, Y.; Liu, Y.; Yin, C.; Quek, T.Q.S. Passive Inter-Satellite Localization Accuracy Optimization in Low Earth Orbit Satellite Networks. IEEE Trans. Wirel. Commun. 2025, 24, 2894–2909. [Google Scholar] [CrossRef]
  11. Bao, C.; Sheng, M.; Zhou, D.; Shi, Y.; Li, J. Toward Intelligent Cross-Domain Resource Coordinate Scheduling for Satellite Networks. IEEE Trans. Wirel. Commun. 2023, 22, 9610–9625. [Google Scholar] [CrossRef]
  12. Wen, X.; Ruan, Y.; Li, Y.; Pan, C.; Elkashlan, M.; Zhang, R.; Li, T. A Hierarchical Game Framework for Win-Win Resource Trading in Cognitive Satellite Terrestrial Networks. IEEE Trans. Wirel. Commun. 2024, 23, 13530–13544. [Google Scholar] [CrossRef]
  13. Fu, S.; Gao, J.; Zhao, L. Collaborative Multi-Resource Allocation in Terrestrial-Satellite Network Towards 6G. IEEE Trans. Wirel. Commun. 2021, 20, 7057–7071. [Google Scholar] [CrossRef]
  14. Ding, C.; Wang, J.-B.; Chen, Y.; Yu, H.; Cheng, M.; Lin, M.; Wang, J. Satellite-Terrestrial Assisted Multi-Tier Computing Networks with MIMO Precoding and Computation Optimization. IEEE Trans. Wirel. Commun. 2024, 23, 3763–3779. [Google Scholar] [CrossRef]
  15. Zhang, H.; Xu, J.; Liu, X.; Long, K.; Leung, V.C.M. Joint Optimization of Caching Placement and Power Allocation in Virtualized Satellite-Terrestrial Network. IEEE Trans. Wirel. Commun. 2023, 22, 7932–7943. [Google Scholar] [CrossRef]
  16. Zhang, B.; Gao, R.; Gao, P.; Wang, Y.; Yang, Z. Interference-Suppressed Joint Channel and Power Allocation for Downlinks in Large-Scale Satellite Networks: A Dynamic Hypergraph Neural Network Approach. IEEE Trans. Wirel. Commun. 2025. [Google Scholar] [CrossRef]
  17. Qin, X.; Ma, T.; Tang, Z.; Zhang, X.; Zhou, H.; Zhao, L. Service-Aware Resource Orchestration in Ultra-Dense LEO Satellite-Terrestrial Integrated 6G: A Service Function Chain Approach. IEEE Trans. Wirel. Commun. 2023, 22, 6003–6017. [Google Scholar] [CrossRef]
  18. Di, B.; Zhang, H.; Song, L.; Li, Y.; Li, G.Y. Ultra-Dense LEO: Integrating Terrestrial-Satellite Networks Into 5G and Beyond for Data Offloading. IEEE Trans. Wirel. Commun. 2019, 18, 47–62. [Google Scholar] [CrossRef]
  19. Ding, C.; Wang, J.-B.; Zhang, H.; Lin, M.; Li, G.Y. Joint Optimization of Transmission and Computation Resources for Satellite and High Altitude Platform Assisted Edge Computing. IEEE Trans. Wirel. Commun. 2022, 21, 1362–1377. [Google Scholar] [CrossRef]
  20. Liu, W.; Zhou, Q.; Zhang, H.; Li, W.; Gao, Z.; Tao, Y. Satellite Beam Slicing Resource Allocation Strategy Based on Business Recognition and Prediction. In Proceedings of the 2024 5th International Conference on Computers and Artificial Intelligence Technology (CAIT), Hangzhou, China, 20–22 December 2024; pp. 285–289. [Google Scholar] [CrossRef]
  21. Bai, J.; Yuan, A.; Xiao, Z.; Zhou, H.; Wang, D.; Jiang, H.; Jiao, L. Class Incremental Learning with Few-Shots Based on Linear Programming for Hyperspectral Image Classification. IEEE Trans. Cybern. 2022, 52, 5474–5485. [Google Scholar] [CrossRef]
  22. Zhuang, C.; Qian, C.; Wang, W.; Shi, X.; Wang, J.; Yue, C. Research on Intelligent Matching Technology for Natural Disaster Monitoring Needs Based on Multi Satellite and Multi Payload. In Proceedings of the 2024 5th International Conference on Geology, Mapping and Remote Sensing (ICGMRS), Wuhan, China, 12–14 April 2024; pp. 68–73. [Google Scholar] [CrossRef]
  23. Jo, G.; Kim, S.; Kim, H.W.; Oh, D. Comparative Analysis on the Resource Allocations for Interference-Limited LEO Satellite Systems. In Proceedings of the 2023 14th International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 11–13 October 2023; pp. 1645–1649. [Google Scholar] [CrossRef]
  24. Zhou, B.; Zhu, W.; Zhai, Y.; Yang, M.; Li, H. A Multi-Objective Optimization Based LEO Satellite Beam Hopping Resource Allocation Algorithm. In Proceedings of the 2024 10th International Conference on Computer and Communications (ICCC), Chengdu, China, 13–16 December 2024; pp. 1657–1662. [Google Scholar] [CrossRef]
  25. Sun, Z.; Wang, C.; Zhang, C.; Han, Z.; Qin, Z.; Lv, Y. User Position-Driven Resource Allocation for Beam Hopping in Low Earth Orbit Satellite. In Proceedings of the 2024 5th International Conference on Electronic Engineering and Signal Processing (EESP), Derby, UK, 2–4 December 2024; pp. 15–20. [Google Scholar] [CrossRef]
  26. Hao, Y.; Song, Z.; Zheng, Z.; Zhang, Q.; Miao, Z. Joint Communication, Computing, and Caching Resource Allocation in LEO Satellite MEC Networks. IEEE Access 2023, 11, 6708–6716. [Google Scholar] [CrossRef]
  27. Bai, J.; Ding, B.; Xiao, Z.; Jiao, L.; Chen, H.; Regan, A.C. Hyperspectral Image Classification Based on Deep Attention Graph Convolutional Network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5504316. [Google Scholar] [CrossRef]
  28. Li, J.; Sun, G.; Duan, L.; Wu, Q. Multi-Objective Optimization for UAV Swarm-Assisted IoT with Virtual Antenna Arrays. IEEE Trans. Mob. Comput. 2024, 23, 4890–4907. [Google Scholar] [CrossRef]
  29. Jia, G.; Sun, C. Research on the Application of Artificial Intelligence Technology in Beam Resource Allocation for Multi-Beam Satellite Communication Systems. In Proceedings of the 2025 6th International Conference on Computer Engineering and Application (ICCEA), Hangzhou, China, 25–27 April 2025; pp. 1170–1175. [Google Scholar] [CrossRef]
  30. Shan, Z.; Wei, L.; Li, J.; Wang, Y.; Zhu, K. Edge Computing-Enabled Real-Time Resource Allocation System Based on Satellite Cluster. In Proceedings of the 2024 10th International Conference on Computer and Communications (ICCC), Chengdu, China, 13–16 December 2024; pp. 2128–2132. [Google Scholar] [CrossRef]
  31. Tao, T.; Huang, W.; Sun, L.; Cao, H. Resource Optimization in Low Earth Orbit Satellites with Edge Computing: An ADMM-Based Resource Management Strategy. In Proceedings of the 2025 5th International Conference on Neural Networks, Information and Communication Engineering (NNICE), Guangzhou, China, 10–12 January 2025; pp. 787–792. [Google Scholar] [CrossRef]
  32. Cheng, L.; Wang, X.; Sun, X. SCODA: Joint Optimization of Task Offloading and Resource Allocation in LEO-Assisted MEC System with Satellite Cooperation. IEEE Wirel. Commun. Lett. 2025, 14, 3304–3308. [Google Scholar] [CrossRef]
  33. Fan, W.; Meng, Q.; Wang, G.; Bian, H.; Liu, Y.; Liu, Y. Satellite Edge Intelligence: DRL-Based Resource Management for Task Inference in LEO-Based Satellite-Ground Collaborative Networks. IEEE Trans. Mob. Comput. 2025, 24, 10710–10728. [Google Scholar] [CrossRef]
  34. Chai, R.; Liu, J.; Tang, X.; Gui, K.; Chen, Q. A Hybrid Offline and Online Resource Allocation Algorithm for Multibeam Satellite Communication Systems. IEEE Trans. Netw. Serv. Manag. 2024, 21, 3711–3726. [Google Scholar] [CrossRef]
  35. Peng, P.; Xu, T.; Chen, X.; Zarakovitis, C.C.; Wu, C. Blocked-Job-Offloading-Based Computing Resources Sharing in LEO Satellite Networks. IEEE Internet Things J. 2025, 12, 2287–2290. [Google Scholar] [CrossRef]
Figure 1. Illustration of the proposed NTN uplink architecture and the integration of the FMA-enhanced multi-agent decision framework. The left side shows the three-layer UE–HAP–LEO communication scenario across urban, industrial, and remote areas, while the right side highlights the actor–critic network design, where predicted CSI from the FMA module is embedded into multi-agent decision making for resource allocation.
Figure 1. Illustration of the proposed NTN uplink architecture and the integration of the FMA-enhanced multi-agent decision framework. The left side shows the three-layer UE–HAP–LEO communication scenario across urban, industrial, and remote areas, while the right side highlights the actor–critic network design, where predicted CSI from the FMA module is embedded into multi-agent decision making for resource allocation.
Sensors 26 00148 g001
Figure 2. MADDPG-based multi-agent DRL framework with centralized training and decentralized execution.
Figure 2. MADDPG-based multi-agent DRL framework with centralized training and decentralized execution.
Sensors 26 00148 g002
Figure 3. FMA model architecture combining state-space modeling and self-attention for NTN channel prediction.
Figure 3. FMA model architecture combining state-space modeling and self-attention for NTN channel prediction.
Sensors 26 00148 g003
Figure 4. Comparison of channel prediction models under the canonical UE–HAP–LEO NTN configuration. (a) Average total reward per episode versus training episodes. (b) Average normalized system efficiency versus training episodes.
Figure 4. Comparison of channel prediction models under the canonical UE–HAP–LEO NTN configuration. (a) Average total reward per episode versus training episodes. (b) Average normalized system efficiency versus training episodes.
Sensors 26 00148 g004
Figure 5. Comparison of channel prediction models under the canonical UE–HAP–LEO NTN configuration. (a) Average number of HAP–LEO collaboration events per episode versus training episodes. (b) Prediction accuracy versus training episodes.
Figure 5. Comparison of channel prediction models under the canonical UE–HAP–LEO NTN configuration. (a) Average number of HAP–LEO collaboration events per episode versus training episodes. (b) Prediction accuracy versus training episodes.
Sensors 26 00148 g005
Figure 6. Average total reward per episode versus training episodes for different scheduling methods under the canonical UE–HAP–LEO NTN configuration.
Figure 6. Average total reward per episode versus training episodes for different scheduling methods under the canonical UE–HAP–LEO NTN configuration.
Sensors 26 00148 g006
Figure 7. Jain’s fairness index across HAPs versus training episodes for different scheduling methods under the canonical UE–HAP–LEO NTN configuration.
Figure 7. Jain’s fairness index across HAPs versus training episodes for different scheduling methods under the canonical UE–HAP–LEO NTN configuration.
Sensors 26 00148 g007
Figure 8. Average number of HAP–LEO collaboration events per episode versus training episodes for different scheduling methods under the canonical UE–HAP–LEO NTN configuration.
Figure 8. Average number of HAP–LEO collaboration events per episode versus training episodes for different scheduling methods under the canonical UE–HAP–LEO NTN configuration.
Sensors 26 00148 g008
Figure 9. Average normalized system efficiency versus training episodes for different scheduling methods under the canonical UE–HAP–LEO NTN configuration.
Figure 9. Average normalized system efficiency versus training episodes for different scheduling methods under the canonical UE–HAP–LEO NTN configuration.
Sensors 26 00148 g009
Figure 10. Ablation study of prediction and constraint design under the canonical UE–HAP–LEO NTN configuration: average total reward per episode versus training episodes for different variants.
Figure 10. Ablation study of prediction and constraint design under the canonical UE–HAP–LEO NTN configuration: average total reward per episode versus training episodes for different variants.
Sensors 26 00148 g010
Figure 11. Ablation study of prediction and constraint design under the canonical UE–HAP–LEO NTN configuration: average normalized system efficiency per episode versus training episodes for different variants.
Figure 11. Ablation study of prediction and constraint design under the canonical UE–HAP–LEO NTN configuration: average normalized system efficiency per episode versus training episodes for different variants.
Sensors 26 00148 g011
Figure 12. Ablation study of prediction and constraint design under the canonical UE–HAP–LEO NTN configuration: average number of HAP–LEO collaboration events per episode versus training episodes for different variants.
Figure 12. Ablation study of prediction and constraint design under the canonical UE–HAP–LEO NTN configuration: average number of HAP–LEO collaboration events per episode versus training episodes for different variants.
Sensors 26 00148 g012
Table 1. Layer configuration of the FMA channel prediction model.
Table 1. Layer configuration of the FMA channel prediction model.
ModuleLayer TypeConfigurationOutput
Input EncodingLinear + Positional Embedding d h L × h
Mamba BackboneState-Space Block ( × 4 )Hidden size h = 128 L × h
MHSAMulti-Head Self-Attention H = 4 , d k = 32 L × h
FusionResidual + LayerNormSkip connection Z t + U t L × h
Feed-ForwardTwo-layer MLP h 2 h h , GELU L × h
Temporal PoolingAttention PoolingWeight vector w p 1 × h
Output HeadLinear Projection h d 1 × d
Table 2. Main notations used in the system model and optimization framework.
Table 2. Main notations used in the system model and optimization framework.
SymbolDescription
Z , ZSet and number of ground UEs
H , HSet and number of HAPs
L , LSet and number of LEO satellites
A , ASet and number of sub-areas
k, tTime-slot index in an episode
d UE ( k ) 3D position of a UE at slot k
d HAP 3D position of a HAP
B Tot UE HAP Total bandwidth of the UE–HAP uplink
C Z , | C Z | Set and number of UE–HAP sub-channels
B z , h ( k ) Bandwidth allocated to UE z–HAP h link at slot k
P z max Maximum transmit power of UE z
P ( z , h ) ( k ) Transmit power from UE z to HAP h at slot k
φ h , z ( k ) Binary UE–HAP association indicator at slot k
θ ( z , h ) ( k ) Instantaneous SNR on UE z–HAP h link at slot k
Φ ( z , h ) ( k ) Instantaneous rate on UE z–HAP h link at slot k
Γ ^ z ( k ) Cumulative throughput of UE z up to slot k
I f ( k ) Jain’s fairness index over all UEs at slot k
ϖ ( z , h ) ( k ) End-to-end delay of UE z–HAP h transmission at slot k
R tot ( k ) Composite system reward at slot k
H t UE HAP CSI matrix of UE–HAP links at time t
H t HAP LEO CSI matrix of HAP–LEO links at time t
B t Vector of available bandwidths at time t
P t Vector of available transmit powers at time t
Q t Vector of HAP queue lengths at time t
s t Global system state at time t
o t i Local observation of agent i at time t
a t i Action of agent i at time t
α u , h i UE–HAP association decision in agent i’s action
β h i Bandwidth allocation ratio to HAP h in agent i’s action
p h i Transmit power of HAP h in agent i’s action
C t Total capacity (throughput) at time t
D t Average end-to-end delay at time t
J t Jain’s fairness index at time t
E t collab Number of collaboration events at time t
λ 1 , λ 2 , λ 3 , λ 4 Weights of reward components in (13) and (30)
h t CSI feature vector at time slot t (FMA input)
X t History window of CSI features used by FMA
LLength of the CSI history window
Δ Prediction horizon (number of slots ahead)
Z t Hidden sequence produced by the Mamba backbone
U t Feature sequence output by the MHSA block
F t , F t Fused feature sequences before and after feed-forward network
f t Pooled feature vector used for CSI prediction
h ^ t + Δ Predicted future CSI vector at time t + Δ
Table 3. Summary of key performance metrics under different frameworks.
Table 3. Summary of key performance metrics under different frameworks.
Deep Reinforcement
Learning/Baseline Method
Total RewardCollaboration EventsSystem EfficiencyLoad Balance
PPO1331.13440.7450.975
MAPPO2519.37840.9000.981
HAPPO2789.38580.8910.985
QMIX2465.47010.9990.999
MADDPG2749.18971.0120.997
Convex-CSI2812.72820.8790.920
Greedy1974.61000.7180.825
Random1160.8360.4570.647
FMA-MADDPG (proposed)3199.49421.0110.997
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, C.; Song, K.; Bai, J.; Li, C.; Zhao, Y.; Xiao, Z.; Sun, Y. FMA-MADDPG: Constrained Multi-Agent Resource Optimization with Channel Prediction in 6G Non-Terrestrial Networks. Sensors 2026, 26, 148. https://doi.org/10.3390/s26010148

AMA Style

Yang C, Song K, Bai J, Li C, Zhao Y, Xiao Z, Sun Y. FMA-MADDPG: Constrained Multi-Agent Resource Optimization with Channel Prediction in 6G Non-Terrestrial Networks. Sensors. 2026; 26(1):148. https://doi.org/10.3390/s26010148

Chicago/Turabian Style

Yang, Chunyu, Kejian Song, Jing Bai, Cuixing Li, Yang Zhao, Zhu Xiao, and Yanhong Sun. 2026. "FMA-MADDPG: Constrained Multi-Agent Resource Optimization with Channel Prediction in 6G Non-Terrestrial Networks" Sensors 26, no. 1: 148. https://doi.org/10.3390/s26010148

APA Style

Yang, C., Song, K., Bai, J., Li, C., Zhao, Y., Xiao, Z., & Sun, Y. (2026). FMA-MADDPG: Constrained Multi-Agent Resource Optimization with Channel Prediction in 6G Non-Terrestrial Networks. Sensors, 26(1), 148. https://doi.org/10.3390/s26010148

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop