Next Article in Journal
Optimization Design of Microwave Filters Based on Deep Learning and Metaheuristic Algorithms
Previous Article in Journal
A Data-Centric Framework for Implementing Artificial Intelligence in Smart Manufacturing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Agent Deep Reinforcement Learning Anti-Jamming Spectrum-Access Method in LEO Satellites

1
The School of Space Information, Space Engineering University, Beijing 101416, China
2
National Key Laboratory of Space Target Awareness, School of Space Information, Space Engineering University, Beijing 101416, China
*
Authors to whom correspondence should be addressed.
Electronics 2025, 14(16), 3307; https://doi.org/10.3390/electronics14163307
Submission received: 19 July 2025 / Revised: 11 August 2025 / Accepted: 13 August 2025 / Published: 20 August 2025

Abstract

Low-Earth-orbit (LEO) satellite networks face significant vulnerabilities to malicious jamming and co-channel interference, compounded by dynamic topologies, resource constraints, and complex electromagnetic environments. Traditional anti-jamming approaches lack adaptability, centralized intelligent methods incur high overhead, and distributed intelligent methods fail to achieve global optimization. To address these limitations, this paper proposed a value decomposition network (VDN)-based multi-agent deep reinforcement learning (DRL) anti-jamming spectrum access approach with a centralized training and distributed execution architecture. Following offline centralized ground-based training, the model was deployed distributedly on satellites for real-time spectrum-access decision-making. The simulation results demonstrate that the proposed method effectively balances training costs with anti-jamming performance. The method achieved near-optimal user satisfaction (approximately 97%) with minimal link overhead, confirming its effectiveness for resource-constrained LEO satellite networks.

1. Introduction

Low-Earth-orbit (LEO) satellites offer significant advantages, including low-latency communication, high data rates, and wide geographical coverage capabilities [1,2]. These attributes make their application a powerful supplement to terrestrial networks [3,4], and they have broad application prospects [5]. In recent years, the global deployment of LEO constellations (e.g., Starlink, OneWeb, and Kuiper) has been rapidly expanding in both speed and scale [6]. However, the inherent exposure of satellites and their reliance on open wireless channels make them susceptible to external malicious jamming and internal co-channel interference. Consequently, developing effective anti-jamming techniques is critical to ensure communication security. However, dynamic topologies [7], limited resources [8], and complex electromagnetic conditions [9] in LEO networks pose significant challenges to achieving this goal.
Extensive research has been dedicated to anti-jamming technologies. Overall, the existing approaches can be divided into two categories: traditional anti-jamming technologies and intelligent anti-jamming technologies.

1.1. Traditional Anti-Jamming Techniques

Direct-sequence spread spectrum (DSSS) and frequency-hopping spread spectrum (FHSS) constitute the foundational techniques for anti-jamming communications [10]. The authors of [11] proposed the Turbo iterative acquisition algorithm based on factor graph modeling of time-varying Doppler rate to address the challenges of DSSS signal acquisition in highly mobile satellite communications. Ref. [12] proposed random differential DSSS based on the traditional DSSS technology, which can effectively alleviate reactive jamming attacks. FHSS uses pre-shared pseudorandom sequences to switch communication frequencies. Ref. [13] proved that the performance of FHSS signals is superior to that of DSSS signals, and FHSS signals are suitable for application in military satellite communication. The authors of [14] achieved a stable connection and good anti-jamming ability by applying adaptive frequency hopping technology. Ref. [15] proposed dynamic spectrum-access technology (DSA) based on cognitive database control, and it achieved co-channel sharing between non-terrestrial networks and terrestrial networks. The authors of [16] presented anti-jamming technologies in the time and transform domains and further investigated spatial-domain techniques, space–time processing, and adaptive beamforming based on array antennas. However, these methods typically rely on predefined rules or static configurations. Although effective in stable environments, their fixed strategies lack adaptability in high-mobility LEO satellite networks with dynamic uncertainties.

1.2. Intelligent Anti-Jamming Techniques

Intelligent anti-jamming techniques can explore the environment through trial-and-error, enabling them to learn to make robust and adaptive anti-jamming decisions under highly dynamic and uncertain conditions [17]. The adaptive learning capabilities, ability to handle complex high-dimensional state spaces, and innovative distributed/collaborative learning paradigms of these intelligent technologies overcome traditional bottlenecks and address anti-jamming challenges in satellite communications [18,19]. The authors of [20] proposed a decentralized federated learning-assisted deep Q-network (DQN) algorithm to determine collaborative anti-jamming strategies for UAV pairs. In  [21], an energy-saving and anti-jamming communication framework of a UAV cluster based on multi-agent cooperation was constructed. Ref. [22] introduced a deep reinforcement learning (DRL)-based dynamic spectrum-access method with accelerated convergence where soft labels replace conventional reward signals; it reduced iteration counts. Ref. [23] proposed an intelligent dynamic-spectrum anti-jamming communication approach based on DRL, and it improved in a trial-and-error manner to adapt to the dynamic nature of jamming. Ref. [24] proposed a collaborative multi-agent layered Q learning (MALQL)-based anti-jamming communication algorithm to reduce the high dimensionality of the action space. Ref. [25] proposed a multi-agent proximal policy optimization (MAPPO) algorithm to optimize a collaborative strategy for multiple UAVs regarding transmission power and flight trajectory. The authors of [26] designed a deep recurrent Q-network (DRQN) algorithm and an intelligent anti-jamming decision algorithm, which improved throughput and convergence performance. However, centralized intelligent anti-jamming architectures incur high computational overhead and complexity, while distributed intelligent anti-jamming architectures fail to converge to globally optimal solutions. Consequently, neither approach is suitable for resource-constrained LEO satellite networks.
To address the aforementioned issues, this paper proposes a value decomposition network (VDN)-based multi-agent DRL method that applies a centralized training and distributed execution architecture. This method achieves global optimization during training and local decision-making during execution, which has certain advantages in terms of performance improvement and cost reduction. The main contributions of this paper are as follows:
  • A VDN-based multi-agent DRL method is proposed to solve the anti-jamming spectrum-access problem. This method adopts an “offline centralized training–online distributed execution” architecture. After offline training on the ground, the model is deployed onto satellites to enable real-time anti-jamming spectrum-access decisions based on local observations.
  • During training, the parameter-sharing mechanism is employed to reduce communication overhead significantly. During execution, the incremental update mechanism is employed to enhance model adaptability.
  • The simulation results prove the proposed method’s effectiveness. It balances performance and cost better than fully centralized training and independent distributed training.
The rest of the paper is organized as follows. The system model and problem construction are outlined in Section 2. Section 3 proposes a multi-agent deep reinforcement learning algorithm. Section 4 presents the simulation results and discussion. Section 5 concludes this paper. The main notations employed in the paper are shown in Table 1.

2. System Model and Problem Construction

2.1. System Model

As shown in Figure 1, there is a jamming satellite and an LEO constellation. Region S is the area jointly covered by jamming satellite and LEO satellites, and users are randomly distributed within region S. The jamming satellite applies randomized frequency suppression jamming to the satellite downlinks. The frequency variations follow a Markov probability matrix. Jamming strategies are formulated by the ground control center and distributed to jamming satellites. The LEO satellites are equipped with multibeam phased array antennas that can dynamically adjust the beam frequency for anti-jamming, and user communication quality is maintained uninterrupted.
The set of LEO satellites is denoted as L E O = { L E O 1 , L E O 2 , , L E O K } ; the set of users is denoted as U = { U 1 , U 2 , , U N } . Each satellite has B beams, and each beam serves one user. The set of available channels for each beam is denoted as F L = { f 1 , f 2 , , f M } , and each channel bandwidth is B f . The jamming channel set is F J = F L , and the beam power of the jamming satellite and LEO satellites are denoted as p j and p l , respectively.
At each time slot, jamming satellite switches channels based on Markov probability matrix, and LEO satellites decide the optimal spectrum-access scheme for the downlinks based on real-time environment sensing and historical data analysis. This paper focuses on dynamic spectrum-access strategies for LEO satellite downlinks in malicious jamming environments, aiming to reduce interference and improve user satisfaction.

2.2. Propagation Model

According to ITU-R S.672 [27], the gain of jamming satellite transmitting antenna is
G J θ n j ( t ) = G J m a x , θ n j ( t ) θ J b , G J m a x 3 θ n j ( t ) θ J b 2 , θ J b < θ n j ( t ) 2.58 θ J b , G J m a x 20 , 2.58 θ J b < θ n j ( t ) 6.32 θ J b G J m a x 25 log ( θ n j ( t ) θ J b ) , 6.32 θ J b < θ n j ( t ) θ 1 , 0 , θ 1 < θ n j ( t ) ,
where G J m a x is the maximum gain of jamming satellite transmitting antenna, and θ n j ( t ) is the off-axis angle of the center of the jamming beam in the U n direction. θ J b is half the 3 dB beamwidth θ J 3 d B , and θ 1 denotes the value obtained when the fourth line of Equation (1) is equal to 0.
According to ITU-R S.1528 [28], the gain of LEO satellite transmitting antenna is
G L θ n m ( t ) = G L m a x , θ n m ( t ) θ L b , G L m a x 3 θ n m ( t ) θ L b 2 , θ L b < θ n m ( t ) Y , G L m a x + L s 25 log ( θ n m ( t ) Y ) , Y < θ n m ( t ) Z , L F , Z < θ n m ( t ) ,
where G L m a x is the maximum gain of LEO satellite transmitting antenna, and θ n m ( t ) is the off-axis angle of the link l m ( t ) in the U n direction. θ L b is half the 3 dB beamwidth θ L 3 d B , L s = 6.75   d B is main beam and proximal paraflap mask intersection, and L F = 0 dBi is near-axis paravalve level. Parameters Y and Z are denoted as
Y = 1.5 θ L b , Z = Y × 10 0.04 ( G L m a x + L s L F ) .
According to ITU-R S.465 [29], the gain of user receiving antenna is
G U θ m n ( t ) = G U m a x , θ m n ( t ) θ m i n , 32 25 log ( θ m n ( t ) ) , θ m i n < θ m n ( t ) 48 , 10 , 48 < θ m n ( t ) ,
where G U m a x is the maximum gain of user receiving antenna, and θ m n ( t ) is the off-axis angle of the link l n ( t ) in the U m direction. The parameter θ m i n is denoted as
θ m i n = max ( 1 , 100 λ / D ) , D / λ 50 , max ( 2 , 144 D / λ 1.09 ) , else ,
where D is the circular equivalent diameter of antenna, and λ is the wavelength.
As shown in Figure 2, users are disrupted in two ways:
➀ Malicious jamming from jamming satellite:
I n j ( t ) = p j δ n j ( t ) h n j ( t ) .
where the channel gain h n j ( t ) is
h n j ( t ) = G J θ n j ( t ) G U θ j n ( t ) 4 π d n j ( t ) λ 2 ,
and the function δ n j ( t ) is
δ n j ( t ) = 1 , a j ( t ) = a n ( t ) , 0 , a j ( t ) a n ( t ) .
where G J θ n j ( t ) is antenna gain of the jamming satellite in the U n direction, G U θ j n ( t ) is antenna gain of user in the jamming satellite direction, and d n j ( t ) is distance of the link  i n j ( t ) .
➁ Co-channel interference from other beams of the LEO satellites:
I n m ( t ) = m = 1 , m n N p l δ n m ( t ) h n m ( t ) .
where the channel gain h n m ( t ) is
h n m ( t ) = G L θ n m ( t ) G U θ m n ( t ) 4 π d n m ( t ) λ 2 ,
and G L θ n m ( t ) is antenna gain of the link l m ( t ) in the U n direction, G U θ m n ( t ) is antenna gain of the link l n ( t ) in the U m direction, and d n m ( t ) is distance of the link i n m ( t ) .
Consequently, the transmission rate of user U n is
r n ( a n t ) = B f log 2 ( 1 + p l h n ( t ) I n j ( t ) + I n m ( t ) + N n ( t ) ) ,
where N n ( t ) is the system noise, h n ( t ) is the channel gain, and its expression is
h n ( t ) = G L m a x G U m a x 4 π d n ( t ) λ 2 ,
where d n ( t ) is distance of the link l n ( t ) .
In order to capture the user’s individualized needs and preferences, perceived satisfaction [30] is introduced with the following expression:
s n ( a n t ) = 1 1 + exp ( c ( r n ( a n t ) r n t h ) ) ,
where a n t is the spectrum selection of user U n at time slot t, r n t h is transmission rate threshold, and the parameter c modulates the slope of the demand utility curve, reflecting the sensitivity of the s n ( a n t ) to deviations from the rate threshold. Higher values of the s n ( a n t ) indicate superior user satisfaction with communication services, whereas lower values signify that the system fails to adequately meet service requirements.

2.3. Problem Construction

This paper aims to determine optimal spectrum-access strategies for LEO satellites to maximize average user satisfaction over T periods while maximizing anti-jamming performance. The optimization problem can be formulated as follows:
P : max a n t , n , t E t = 1 T n = 1 N s n ( a n t )
s . t . C 1 : r n ( a n t ) r n t h , t 1 , , T
C 2 : T N + { }                                      
C 3 : a n t F L                                                                
In this optimization problem, C 1 indicates that the transmission rate must be greater than a threshold to ensure normal communication, C 2 indicates that the time horizon is either finite or infinite, and C 3 indicates that the spectrum accessed by the user U n at time slot t must be within the orthogonal frequency band. The model not only takes into account the immediate user needs but also optimizes the overall user satisfaction over a longer period of time.
The optimization problem P is a non-convex nonlinear program [31]. According to [32,33], this problem is generally NP-hard, and it is difficult to find the optimal solution. Although convex optimization, heuristic, or game-theoretic approaches may attain suboptimal solutions, they typically require global information collection. This makes them impractical for resource-constrained LEO satellites with strict real-time requirements [34]. Since the dynamic spectrum-access problem satisfies the characteristics of the Markov decision process (MDP), the multi-agent DRL solution is introduced in detail subsequently.

3. The Proposed Multi-Agent Deep Reinforcement Learning Method

This section first presents the multi-agent MDP [35] formulation for multi-user anti-jamming spectrum access. Subsequently, the process of VDN-based multi-agent DRL is detailed.

3.1. Multi-Agent MDP Problem Formulation

In this paper, each satellite operates as an agent that observes local environmental states and selects actions per time slot. With all agents collaboratively maximizing the long-term cumulative reward, this constitutes a fully cooperative multi-agent task modeled as an MDP, mathematically expressed as ( S , A , F , R , γ ) , where S is the set of environment state, A is the the set of action, F is the state transition probability function, R is the reward function, and γ is the discount factor; its values range from 0 to 1. Specifically, key parameters for the multi-user scenario are formally defined as follows:
  • State: The state of the environment s t is the global information of all users, which can be formed as follows:
    s t = ( s 1 t , s 2 t , , s N t ) ,
    where each s n t contains self-satisfaction, current jamming, and current co-channel interference.
  • Action: The action is the spectrum-access strategy for all users at time slot t, which can be formed as follows:
    a t = ( a 1 t , a 2 t , , a N t ) ,
    where a n t is the spectrum-access strategy of user n.
  • Reward: The agents can obtain immediate rewards after taking actions, which can be formed as follows:
    r t = n = 1 N s n a n t .
  • State–action value function: The state–action value function is the discounted reward obtained by taking action a n t in state s n t , which can be formed as follows:
    Q n s n t , a n t = max a n E r t + γ r t + 1 ,
    where a n is the action set of agent n over period T, a n = a n 1 , a n 2 , , a n T , r t + 1 is long-term reward, and E · is the expectation. The agent’s goal is to learn the optimal strategy that maximizes the long-term reward.

3.2. Multi-Agent DRL Algorithm Design

Centralized multi-agent DRL methods require global information and incur high computational overhead. Conversely, distributed multi-agent DRL approaches relying on local observations fail to converge to optimal solutions. To overcome these limitations, the VDN [36] algorithm is adopted for multi-satellite anti-jamming spectrum access. This approach implements a centralized training and distributed execution framework. Global information is utilized during offline training, while local observations are required during online execution. The overall structure of the proposed algorithm is shown in Figure 3.

3.2.1. The Offline Centralized Training Phase

The core idea of the VDN algorithm is to decompose the joint action-value function Q t o t ( s t , a t ; θ ) of a multi-agent system into the sum of the local action-value functions Q n ( s n t , a n t ; θ n ) of the individual agent, and it is denoted as
Q t o t s t , a t ; θ = n = 1 N Q n s n t , a n t ; θ n ,
where s t is joint state, a t is joint action, and θ n is parameter of local Q-network of the agent n. This decomposition allows the agent to independently select actions based on their local observations during the execution phase while optimizing the strategy through a global temporal difference (TD) objective during the centralized training phase. During the training process, the agent n selects action according to the ϵ -greedy strategy, and it is denoted as
a n t = random a n t , with probability ϵ , arg max a n t Q n s n t , a n t ; θ ,             with probability 1 ϵ .
The VDN network architecture is shown in Figure 4. After approximating the joint Q-value Q t o t ( s t , a t ; θ ) , the VDN updates the network parameters by minimizing the global reward-based TD error, and the loss function is defined as
L θ , θ n = i B ( y i t o t Q t o t s i , a i ; θ ) 2 ,
where B denotes the mini-batch set sampled from the replay buffer, and y i t o t is target value denoted as
y i t o t = r i + γ arg max a i + 1 Q t o t s i + 1 , a i + 1 ; θ ,
where r i is global immediate reward, γ is discount factor, and s i + 1 and a i + 1 denote the next joint state and joint action, respectively.
The gradient of the loss function L θ , θ n is backpropagated through Q t o t to each local Q function Q n , which in turn updates its parameters θ n . This optimization process ensures that the local Q-network parameters of each agent are updated from the perspective of global returns. In order to improve learning efficiency, reduce computational and storage overheads, and facilitate collaboration among the agents, the algorithm employs a parameter-sharing mechanism.
After the centralized training phase is completed, the optimized model parameters θ n are uploaded to the satellite agent to support subsequent distributed autonomous operation. Algorithm 1 describes the centralized training process in detail.
Algorithm 1 Offline Centralized Training Algorithm
  1:
Initialize: network parameter θ n , θ , target network parameter θ n , θ , replay buffer D , mini-batch size B , and exploration probability ϵ
  2:
for episode = 1 to M do
  3:
      Initial environment, observe states s
  4:
      for t = 1 to T do
  5:
          for agent = 1 to N do
  6:
             Select action a n t according to local state s n t and ϵ -greedy strategy as shown in Equation (20)
  7:
          end for
  8:
          Central node calculates additive joint Q-value according to Equation (19)
  9:
          Execute joint action a t = ( a 1 t , a 2 t , , a N t ) ; observe reward r t and next state s t + 1
10:
          Store experience ( s t , a t , r t , s t + 1 ) in replay buffer D
11:
      end for
12:
      if  | D | > | B |  then
13:
          Central node sample a B mini-batch experience from replay buffer D
14:
          Compute loss according to Equations (21) and (22)
15:
      end if
16:
      Use gradient descent to minimize losses; update parameter θ , θ n
17:
      Update the target network parameters softly every C episode:
θ = τ θ + ( 1 τ ) θ , θ n = τ θ n + ( 1 τ ) θ n
18:
end for
19:
Upload trained parameters θ n to satellite agents

3.2.2. The Online Distributed Execution Phase

Each satellite agent loads the trained agent network θ n . Satellite agent determines the optimal spectrum access by independently selecting the action that maximizes their local Q value Q n ( s n t , a n t ; θ n ) based only on their local observations.
An incremental parameter update mechanism is introduced to adapt to dynamic satellite network environment. Each agent continuously collects execution experiences, storing them in a local replay buffer. Periodically, new experiences from this buffer are utilized to refine pretrained network parameters θ n through local loss computation and stochastic gradient descent. This online adaptation accommodates environmental dynamics while avoiding complete retraining. Algorithm 2 describes the distributed execution process in detail.
Algorithm 2 Online Distributed Execution Algorithm
  1:
Initialize: Load trained parameter θ n
  2:
for each time slot t do
  3:
      for each agent n do
  4:
            Observe current state s n t
  5:
            Select action a n t = arg max a n t Q n s n t , a n t ; θ n
  6:
            Execute action a n t ; observe local reward r n t and next state s n t + 1
  7:
            Store experience ( s n t , a n t , r n t , s n t + 1 ) in replay buffer
  8:
      end for
  9:
      Periodically incremental updates using new experience in replay buffer
10:
end for

3.2.3. Complexity Analysis

The computational complexity of the proposed method is analyzed from two key phases: centralized training and distributed execution.
Centralized Training Phase: Each agent’s local Q-network performs forward propagation with complexity O ( L f · D 2 ) , where L f is the number of layers in forward propagation, and D is the dimension of hidden layers in the neural network. For N agents, the total forward complexity per time slot is O ( N · L f · D 2 ) . For each mini-batch, backpropagation of the TD error involves computing gradients across all layers, with per-sample complexity O ( L b · D 2 ) , where L b is the number of layers in backpropagation. For a mini-batch of size | B | , the total backpropagation complexity per episode is O ( | B | · L b · D 2 ) . Combining forward and backward operations over T time slots, the training complexity is O ( T · [ N · L f · D 2 + | B | · L b · D 2 ] ) . The parameter-sharing mechanism reduces the parameter space from O ( N · D 2 ) to O ( D 2 ) , significantly lowering computational costs.
Distributed Execution Phase: Each satellite agent requires only O ( L f · D 2 ) per time slot to compute Q n ( s n t , a n t ) and select actions locally. The incremental update further adapts to environmental dynamics with minimal overhead O ( | L | · L b · D 2 ) , where | L | is the size of the local replay buffer. Therefore, the overall complexity of per time slot is O ( L f · D 2 + | L | · L b · D 2 ) .

4. Simulation Results and Performance Analysis

4.1. Simulation Parameters

In this section, the performance of the proposed multi-agent DRL algorithm is evaluated. The simulated scenario consists of one jamming satellite and an LEO Walker constellation, with ten users randomly distributed within the jamming satellite’s beam coverage. Detailed simulation parameters are documented in Table 2. All the simulations are executed on a computer equipped with a 2.7 GHz Intel Core Ultra 9 CPU, 32 GB RAM, and NVIDIA 5070Ti GPU. Algorithm framework parameters are summarized in Table 3.
In this paper, we compare the proposed method with the following three different methods:
  • Centralized Training Execution (CTE) [37]: The CTE method employs a fully centralized architecture during both training and execution phases. It requires continuous global information interaction and incurs high communication overhead.
  • Non-Cooperative Independent Learning (NIL) [38]: In the NIL method, each agent independently optimizes strategies without coordination mechanisms or information sharing.
  • Random Action Selection (RAS): The RAS method is a benchmark strategy without learning ability, and each agent randomly selects actions.

4.2. Convergence Analysis

Figure 5 illustrates the convergence of the proposed method through episode reward and loss. The reward values in Figure 5a increase during training and stabilize after 200 episodes, while the loss values in Figure 5b exhibit progressive reduction. These observations validate the proposed method’s convergence.
Figure 6 compares convergence among different methods, with the vertical axis denoting the satisfaction index. Evidently, the CTE method demonstrates superior convergence performance due to its global information awareness; its practical application is significantly constrained by satellite network communication and computational resources. The proposed method achieves approximately 97% of CTE’s convergence performance without excessive link overhead. It validates the effectiveness of the proposed method. The NIL method achieves a fully independent decision-making training mode, but the convergence performance is significantly reduced due to the lack of a cooperation mechanism. The RAS method has the worst performance due to its lack of learning ability.

4.3. Performance Analysis

4.3.1. Jamming Avoidance

Table 4 and Table 5, respectively, present the channel selection of users and the jammer during initial and convergence episodes. Each row corresponds to a distinct time slot. As shown in Table 4, the users exhibit random channel selection during the initial exploration, resulting in significant channel overlap with jammer transmissions. In contrast, Table 5 demonstrates users’ effective avoidance of jammed channels after convergence. This comparative analysis confirms the proposed algorithm’s efficacy in achieving near-optimal jamming avoidance.
Figure 7 illustrates the users’ channel selection over 12 time slots after algorithm convergence under sweep jamming. It is obvious that the proposed method remains effective under sweep jamming, and users can thus avoid the channel selected by the jammer.

4.3.2. User Satisfaction

Figure 8 illustrates the impact of number of agents on user satisfaction. As number of agents increases, average satisfaction decreases for all the methods. The CTE method achieves the highest satisfaction. However, its computational costs rise significantly with user scaling. The NIL method reduces costs but causes frequent spectrum collisions from insufficient information exchange, so it has lower satisfaction. The proposed method maintains moderate satisfaction levels. Malicious jamming and computational costs are reduced through collaborative decisions. Thus, the proposed method is more suitable for spectrum management in LEO dynamic networks.

4.3.3. Network Fairness

Inspired by [39], to better demonstrate the effectiveness of the proposed method, the user fairness index is introduced as follows:
J = n N s n 2 N n N s n 2 ,
where N is the number of users, s n is the satisfaction of user U n . It is evident that, the closer the fairness index is to 1, the better the network fairness.
Figure 9 illustrates the impact of number of agents on network fairness. The fairness index exhibits a declining trend across all the methods with an increasing number of agents. Among them, the CTE method demonstrates the best fairness index, while the proposed method’s fairness index is slightly inferior to CTE. In contrast, due to a lack of coordination among agents, NIL exhibits a relatively poor fairness index. RAS achieves the worst fairness index owing to its lack of learning ability.

5. Conclusions

A VDN-based multi-agent DRL method is proposed for anti-jamming spectrum access in LEO satellite networks. To enhance training efficiency and minimize communication overhead, a centralized training with distributed execution architecture is adopted. Following offline centralized training, the network is distributedly deployed on LEO satellites to enable real-time spectrum-access decision-making. Parameter sharing is employed to accelerate convergence and reduce complexity, while incremental updates enhance model adaptability. The simulation results demonstrate improvements in jamming avoidance and user satisfaction, along with reduced computational overhead. Therefore, this approach provides an effective solution for dynamic anti-jamming spectrum access in resource-constrained LEO satellite networks. Future work will explore multi-domain joint anti-jamming technologies and integrate them with physical layer security techniques to further enhance resilience against evolving jamming threats.

Author Contributions

Conceptualization, F.C. and L.J.; Methodology, L.J. and W.C.; Validation, W.C.; Writing—original draft preparation, W.C.; Writing—review and editing, H.Z. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 62471491.

Data Availability Statement

The original contributions presented in this study are included in the article material. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Al-Hraishawi, H.; Chougrani, H.; Kisseleff, S.; Lagunas, E.; Chatzinotas, S. A survey on nongeostationary satellite systems: The communication perspective. IEEE Commun. Surv. Tutorials 2022, 25, 101–132. [Google Scholar] [CrossRef]
  2. Xiao, Z.; Yang, J.; Mao, T.; Xu, C.; Zhang, R.; Han, Z.; Xia, X.G. LEO satellite access network (LEO-SAN) toward 6G: Challenges and approaches. IEEE Wirel. Commun. 2022, 31, 89–96. [Google Scholar] [CrossRef]
  3. Luo, X.; Chen, H.H.; Guo, Q. LEO/VLEO satellite communications in 6G and beyond networks–technologies, applications, and challenges. IEEE Netw. 2024, 38, 273–285. [Google Scholar] [CrossRef]
  4. Wang, Q.; Chen, X.; Qi, Q. Energy-efficient design of satellite-terrestrial computing in 6G wireless networks. IEEE Trans. Commun. 2023, 72, 1759–1772. [Google Scholar] [CrossRef]
  5. Kodheli, O.; Lagunas, E.; Maturo, N.; Sharma, S.K.; Shankar, B.; Montoya, J.F.M.; Duncan, J.C.M.; Spano, D.; Chatzinotas, S.; Kisseleff, S.; et al. Satellite communications in the new space era: A survey and future challenges. IEEE Commun. Surv. Tutor. 2020, 23, 70–109. [Google Scholar] [CrossRef]
  6. Al-Hraishawi, H.; Chatzinotas, S.; Ottersten, B. Broadband non-geostationary satellite communication systems: Research challenges and key opportunities. In Proceedings of the 2021 IEEE International Conference on Communications Workshops (ICC Workshops), Montreal, QC, Canada, 14–23 June 2021; pp. 1–6. [Google Scholar]
  7. Li, W.; Jia, L.; Chen, Y.; Chen, Q.; Yan, J.; Qi, N. A game-theoretic approach for satellites beam scheduling and power control in a mega hybrid constellation spectrum sharing scenario. IEEE Internet Things J. 2025, 12, 20626–20639. [Google Scholar] [CrossRef]
  8. Kim, D.; Jung, H.; Lee, I.H.; Niyato, D. Multi-Beam Management and Resource Allocation for LEO Satellite-Assisted IoT Networks. IEEE Internet Things J. 2025, 12, 19443–19458. [Google Scholar] [CrossRef]
  9. Yang, Q.; Laurenson, D.I.; Barria, J.A. On the use of LEO satellite constellation for active network management in power distribution networks. IEEE Trans. Smart Grid 2012, 3, 1371–1381. [Google Scholar] [CrossRef]
  10. Hasan, M.; Thakur, J.M.; Podder, P. Design and implementation of FHSS and DSSS for secure data transmission. Int. J. Signal Process. Syst. 2016, 4, 144–149. [Google Scholar] [CrossRef]
  11. Wang, J.; Jiang, C.; Kuang, L. Turbo iterative DSSS acquisition in satellite high-mobility communications. IEEE Trans. Veh. Technol. 2021, 70, 12998–13009. [Google Scholar] [CrossRef]
  12. Alagil, A.; Liu, Y. Randomized positioning dsss with message shuffling for anti-jamming wireless communications. In Proceedings of the 2019 IEEE conference on dependable and secure computing (DSC), Hangzhou, China, 18–20 November 2019; pp. 1–8. [Google Scholar]
  13. Lu, R.; Ye, G.; Ma, J.; Li, Y.; Huang, W. A numerical comparison between FHSS and DSSS in satellite communication systems with on-board processing. In Proceedings of the 2009 2nd International Congress on Image and Signal Processing, Tianjin, China, 17–19 October 2009; pp. 1–4. [Google Scholar]
  14. Mast, J.; Hänel, T.; Aschenbruck, N. Enhancing adaptive frequency hopping for bluetooth low energy. In Proceedings of the 2021 IEEE 46th Conference on Local Computer Networks (LCN), Edmonton, AB, Canada, 4–7 October 2021; pp. 447–454. [Google Scholar]
  15. Kokkinen, H.; Piemontese, A.; Kulacz, L.; Arnal, F.; Amatetti, C. Coverage and interference in co-channel spectrum sharing between terrestrial and satellite networks. In Proceedings of the 2023 IEEE Aerospace Conference, Big Sky, MT, USA, 4–11 March 2023; pp. 1–9. [Google Scholar]
  16. Yan, D.; Ni, S. Overview of anti-jamming technologies for satellite navigation systems. In Proceedings of the 2022 IEEE 6th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 4–6 March 2022; Volume 6, pp. 118–124. [Google Scholar]
  17. Wang, X.; Wang, S.; Liang, X.; Zhao, D.; Huang, J.; Xu, X.; Dai, B.; Miao, Q. Deep Reinforcement Learning: A Survey. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 5064–5078. [Google Scholar] [CrossRef]
  18. Nguyen, T.T.; Nguyen, N.D.; Nahavandi, S. Deep Reinforcement Learning for Multiagent Systems: A Review of Challenges, Solutions, and Applications. IEEE Trans. Cybern. 2020, 50, 3826–3839. [Google Scholar] [CrossRef]
  19. Jia, L.; Qi, N.; Su, Z.; Chu, F.; Fang, S.; Wong, K.K.; Chae, C.B. Game theory and reinforcement learning for anti-jamming defense in wireless communications: Current research, challenges, and solutions. IEEE Commun. Surv. Tutorials 2024, 27, 1798–1838. [Google Scholar] [CrossRef]
  20. Yin, Z.; Li, J.; Wang, Z.; Qian, Y.; Lin, Y.; Shu, F.; Chen, W. UAV Communication Against Intelligent Jamming: A Stackelberg Game Approach With Federated Reinforcement Learning. IEEE Trans. Green Commun. Netw. 2024, 8, 1796–1808. [Google Scholar] [CrossRef]
  21. Wu, Z.; Lin, Y.; Zhang, Y.; Shu, F.; Li, J. Multi-agent collaboration based UAV clusters multi-domain energy-saving anti-jamming communication. Sci. Sin. Inf. 2023, 53, 2511. [Google Scholar] [CrossRef]
  22. Li, Y.; Xu, Y.; Li, G.; Gong, Y.; Liu, X.; Wang, H.; Li, W. Dynamic spectrum anti-jamming access with fast convergence: A labeled deep reinforcement learning approach. IEEE Trans. Inf. Forensics Secur. 2023, 18, 5447–5458. [Google Scholar] [CrossRef]
  23. Li, W.; Chen, J.; Liu, X.; Wang, X.; Li, Y.; Liu, D.; Xu, Y. Intelligent dynamic spectrum anti-jamming communications: A deep reinforcement learning perspective. IEEE Wirel. Commun. 2022, 29, 60–67. [Google Scholar] [CrossRef]
  24. Yin, Z.; Lin, Y.; Zhang, Y.; Qian, Y.; Shu, F.; Li, J. Collaborative Multiagent Reinforcement Learning Aided Resource Allocation for UAV Anti-Jamming Communication. IEEE Internet Things J. 2022, 9, 23995–24008. [Google Scholar] [CrossRef]
  25. Bai, H.; Wang, H.; Du, J.; He, R.; Li, G.; Xu, Y. Multi-Hop UAV Relay Covert Communication: A Multi-Agent Reinforcement Learning Approach. In Proceedings of the 2024 International Conference on Ubiquitous Communication (Ucom), Xi’an, China, 5–7 July 2024; pp. 356–360. [Google Scholar]
  26. Zhang, F.; Niu, Y.; Zhou, Q.; Chen, Q. Intelligent anti-jamming decision algorithm for wireless communication under limited channel state information conditions. Sci. Rep. 2025, 15, 6271. [Google Scholar] [CrossRef]
  27. ITU-R S.672; Satellite Antenna Radiation Patterns for Geostationary Orbit Satellite Antennas Operating in the Fixed-Satellite Service. International Telecommunication Union (ITU): Geneva, Switzerland, 1997.
  28. ITU-R S.1528; Satellite Antenna Radiation Patterns for Non-Geostationary Orbit Satellite Antennas Operating in the Fixed-Satellite Service Below 30 GHz. International Telecommunication Union (ITU): Geneva, Switzerland, 2001.
  29. ITU-R S.465; Reference radiation pattern for earth station antennas in the fixed-satellite service for use in coordination and interference assessment in the frequency range from 2 to 31 ghz. International Telecommunication Union (ITU): Geneva, Switzerland, 2010.
  30. Li, W.; Jia, L.; Chen, Q.; Chen, Y. A game theory-based distributed downlink spectrum sharing method in large-scale hybrid satellite constellations. IEEE Trans. Commun. 2024, 72, 4620–4632. [Google Scholar] [CrossRef]
  31. Islam, M.; Sharmin, S.; Nur, F.N.; Razzaque, M.A.; Hassan, M.M.; Alelaiwi, A. High-throughput link-channel selection and power allocation in wireless mesh networks. IEEE Access 2019, 7, 161040–161051. [Google Scholar] [CrossRef]
  32. Lin, Z.; Ni, Z.; Kuang, L.; Jiang, C.; Huang, Z. Dynamic beam pattern and bandwidth allocation based on multi-agent deep reinforcement learning for beam hopping satellite systems. IEEE Trans. Veh. Technol. 2022, 71, 3917–3930. [Google Scholar] [CrossRef]
  33. Luo, Z.Q.; Zhang, S. Dynamic spectrum management: Complexity and duality. IEEE J. Sel. Top. Signal Process. 2008, 2, 57–73. [Google Scholar] [CrossRef]
  34. Lin, Z.; Ni, Z.; Kuang, L.; Jiang, C.; Huang, Z. Satellite-terrestrial coordinated multi-satellite beam hopping scheduling based on multi-agent deep reinforcement learning. IEEE Trans. Wirel. Commun. 2024, 23, 10091–10103. [Google Scholar] [CrossRef]
  35. Busoniu, L.; Babuska, R.; De Schutter, B. A comprehensive survey of multiagent reinforcement learning. IEEE Trans. Syst. Man, Cybern. Part Appl. Rev. 2008, 38, 156–172. [Google Scholar] [CrossRef]
  36. Sunehag, P.; Lever, G.; Gruslys, A.; Czarnecki, W.M.; Zambaldi, V.; Jaderberg, M.; Lanctot, M.; Sonnerat, N.; Leibo, J.Z.; Tuyls, K.; et al. Value-decomposition networks for cooperative multi-agent learning. arXiv 2017, arXiv:1706.05296. [Google Scholar]
  37. Tampuu, A.; Matiisen, T.; Kodelja, D.; Kuzovkin, I.; Korjus, K.; Aru, J.; Aru, J.; Vicente, R. Multiagent cooperation and competition with deep reinforcement learning. PloS ONE 2017, 12, e0172395. [Google Scholar] [CrossRef]
  38. Aref, M.A.; Jayaweera, S.K.; Machuzak, S. Multi-agent reinforcement learning based cognitive anti-jamming. In Proceedings of the 2017 IEEE wireless communications and networking conference (WCNC), San Francisco, CA, USA, 19–22 March 2017; pp. 1–6. [Google Scholar]
  39. Zhang, Y.; Jia, L.; Qi, N.; Xu, Y.; Wang, M. Anti-jamming channel access in 5G ultra-dense networks: A game-theoretic learning approach. Digit. Commun. Netw. 2023, 9, 523–533. [Google Scholar] [CrossRef]
Figure 1. Anti-jamming communication scenario.
Figure 1. Anti-jamming communication scenario.
Electronics 14 03307 g001
Figure 2. Interference model.
Figure 2. Interference model.
Electronics 14 03307 g002
Figure 3. The overall architecture of the proposed algorithm.
Figure 3. The overall architecture of the proposed algorithm.
Electronics 14 03307 g003
Figure 4. The structure of the VDN.
Figure 4. The structure of the VDN.
Electronics 14 03307 g004
Figure 5. An illustration of (a) reward and (b) loss during the training process.
Figure 5. An illustration of (a) reward and (b) loss during the training process.
Electronics 14 03307 g005
Figure 6. Convergence comparison of different methods.
Figure 6. Convergence comparison of different methods.
Electronics 14 03307 g006
Figure 7. The channel selection under sweep jamming.
Figure 7. The channel selection under sweep jamming.
Electronics 14 03307 g007
Figure 8. Impact of number of agents on user satisfaction.
Figure 8. Impact of number of agents on user satisfaction.
Electronics 14 03307 g008
Figure 9. Impact of number of agents on network fairness.
Figure 9. Impact of number of agents on network fairness.
Electronics 14 03307 g009
Table 1. Main notations.
Table 1. Main notations.
NotationDescription
L E O k The kth LEO satellite
U n The nth user of LEO satellites
l n ( t ) The downlink accessed by U n at time slot t
l m ( t ) The downlink accessed by U m at time slot t
i n m ( t ) The interference link from l m ( t ) to U n at time slot t
i n j ( t ) The jamming link from the jamming satellite to U n at time slot t
I n m ( t ) The co-channel interference from l m ( t ) to U n at time slot t
I n j ( t ) Jamming from the jamming satellite to U n at time slot t
θ n m ( t ) The off-axis angle from the l m ( t ) to U n at time slot t
θ n j ( t ) The off-axis angle from the center of the jamming beam to U n at time slot t
B f Channel bandwidth
p j The beam power of jamming satellite
p l The beam power of LEO satellites
h n j ( t ) The channel gain of link i n j ( t )
h n m ( t ) The channel gain of link i n m ( t )
h n ( t ) The channel gain of link l n ( t )
r n ( t ) The transmission rate of user U n at time slot t
s n ( t ) The satisfaction of user U n at time slot t
Table 2. Parameters of constellation.
Table 2. Parameters of constellation.
ParameterJamming SatelliteLEO Satellite
Orbital altitude35,786 km550 km
Beam power45 dBw15 dBw
Beam bandwidth200 MHz200 MHz
Beam radius200 km50 km
Frequency range10.7–12.7 GHz10.7–12.7 GHz
Frequency channels44
Channel noise−100 dBm−100 dBm
Spectrum switchingMarkov probability matrixintelligent
Table 3. Parameters of algorithm framework.
Table 3. Parameters of algorithm framework.
ParameterValue
Learning rate0.001
Replay buffer size D 10,000
Mini-batch size B 64
Initial exploration rate0.1
Final exploration rate0.01
Discount factor γ 0.99
OptimizerAdam
Table 4. Channel selection in the initial episode.
Table 4. Channel selection in the initial episode.
User 1User 2User 3User 4User 5User 6User 7User 8User 9User 10Jammer
t = 1033343443434
t = 2033434343434
t = 3043333334334
t = 4032333333333
t = 5034343433342
Table 5. Channel selection in the convergence episode.
Table 5. Channel selection in the convergence episode.
User 1User 2User 3User 4User 5User 6User 7User 8User 9User 10Jammer
t = 1011424142423
t = 2022123123224
t = 3034414334342
t = 4043313343412
t = 5032331231134
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, W.; Chu, F.; Jia, L.; Zhou, H.; Zhang, Y. A Multi-Agent Deep Reinforcement Learning Anti-Jamming Spectrum-Access Method in LEO Satellites. Electronics 2025, 14, 3307. https://doi.org/10.3390/electronics14163307

AMA Style

Cao W, Chu F, Jia L, Zhou H, Zhang Y. A Multi-Agent Deep Reinforcement Learning Anti-Jamming Spectrum-Access Method in LEO Satellites. Electronics. 2025; 14(16):3307. https://doi.org/10.3390/electronics14163307

Chicago/Turabian Style

Cao, Wenting, Feihuang Chu, Luliang Jia, Hongyu Zhou, and Yunfan Zhang. 2025. "A Multi-Agent Deep Reinforcement Learning Anti-Jamming Spectrum-Access Method in LEO Satellites" Electronics 14, no. 16: 3307. https://doi.org/10.3390/electronics14163307

APA Style

Cao, W., Chu, F., Jia, L., Zhou, H., & Zhang, Y. (2025). A Multi-Agent Deep Reinforcement Learning Anti-Jamming Spectrum-Access Method in LEO Satellites. Electronics, 14(16), 3307. https://doi.org/10.3390/electronics14163307

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop