Next Article in Journal
Multispectral Portable Fibre-Optic Reflectometer for the Classification of the Origin of Chicken Eggshells in the Case of Mycoplasma synoviae Infections
Next Article in Special Issue
A Novel Method for Interferometric Phase Estimation in Dual-Channel Cancellation
Previous Article in Journal
A Quantum Ring Laser Gyroscope Based on Coherence de Broglie Waves
Previous Article in Special Issue
Multiple Mainlobe Interferences Suppression Based on Eigen-Subspace and Eigen-Oblique Projection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Airborne Radar Anti-Jamming Waveform Design Based on Deep Reinforcement Learning

Information and Navigation College, Air Force Engineering University, Xi’an 710077, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(22), 8689; https://doi.org/10.3390/s22228689
Submission received: 2 October 2022 / Revised: 29 October 2022 / Accepted: 7 November 2022 / Published: 10 November 2022

Abstract

:
Airborne radars are susceptible to a large number of clutter, noise and variable jamming signals in the real environment, especially when faced with active main lobe jamming, as the waveform shortcut technology in the traditional regime can no longer meet the actual battlefield radar anti-jamming requirements. Therefore, it is necessary to study anti-main-lobe jamming techniques for airborne radars in complex environments to improve their battlefield survivability. In this paper, we propose an airborne radar waveform design method based on a deep reinforcement learning (DRL) algorithm under clutter and jamming conditions, after previous research on reinforcement-learning (RL)-based airborne radar anti-jamming waveform design methods that have improved the anti-jamming performance of airborne radars. The method uses a Markov decision process (MDP) to describe the complex operating environment of airborne radars, calculates the value of the radar anti-jamming waveform strategy under various jamming states using deep neural networks and designs the optimal anti-jamming waveform strategy for airborne radars based on the duelling double deep Q network (D3QN) algorithm. In addition, the method uses an iterative transformation method (ITM) to generate the time domain signals of the optimal waveform strategy. Simulation results show that the airborne radar waveform designed based on the deep reinforcement learning algorithm proposed in this paper improves the signal-to-jamming plus noise ratio (SJNR) by 2.08 dB and 3.03 dB, and target detection probability by 26.79% and 44.25%, respectively, compared with the waveform designed based on the reinforcement learning algorithm and the conventional linear frequency modulation (LFM) signal at a radar transmit power of 5 W. The airborne radar waveform design method proposed in this paper helps airborne radars to enhance anti-jamming performance in complex environments while further improving target detection performance.

1. Introduction

1.1. Background and Motivation

The rapid development of airborne radar has eliminated the effects of the ground environment and the curvature of the earth on surface-to-air radar. Airborne radar has expanded the detection distance of various low-altitude and ultra-low-altitude targets on the ground, in the air and at sea, and has become one of the key factors of battlefield victory [1]. Airborne radar has the advantages of small blind spots, strong manoeuvrability and high resolution. It can detect long-range targets in all directions and around the clock, control and guide weapons and undertake tasks such as air alert, reconnaissance and flight safety. It plays an important role in land–air defence.
With the continuous development of communication technology and electronic warfare technology, there are many types of electronic equipment on the modern battlefield, and there are situations where the frequency bands used by electronic equipment overlap and interfere with each other. With the rapid development of manufacturing technology and the continuous upgrading of enemy jamming equipment, airborne radar must not only overcome the impact of terrain clutter caused by the complex geographical environment, but also the diversified, dexterous and intelligent jamming of aircraft and missiles that are rushing to attack. These include passive jamming techniques such as foil jamming and towed jamming; active jamming techniques such as deception jamming; and suppression jamming, for which there are a variety of implementation methods, such as long-range support, follow-on and self-defence jamming. Strong clutter and jamming signals in complex environments can reduce the probability of radar target detection by drowning out target echo information, as well as blurring the target signal and reducing the probability of target identification, or forming false targets and increasing the probability of false alarms. These jamming techniques all pose a challenge to airborne radars performing tasks such as search, identification and guidance [2].
Traditional anti-main-lobe jamming techniques include passive anti-jamming methods such as null-domain zeroing, frequency agility and phase encode technology. In particular, the airspace anti-jamming method [3] uses an antenna array to zero in a particular direction of the antenna direction diagram, thereby suppressing side lobe jamming from that direction, but when the target performs self-defence jamming, the jamming and target echoes come from the same angle, i.e., direction, and the radar is unable to detect the target while suppressing the main lobe self-defence jamming by zeroing at the same time. When an airborne radar is subjected to main lobe jamming, the detection probability can be improved by changing the detection method and detection threshold to reduce the probability of false alarms on the radar when the jamming signal has a specific distribution and the jamming power is low. However, this method has difficulty coping with high-power jamming and adapting to jamming from changing parameters such as signal distribution and energy spectra. It can be seen that anti-jamming means starting from airspace zeroing and detection techniques are hardly effective in suppressing main lobe jamming. Airborne radars using these means have low flexibility in transmitted waveforms, fixed received signal processing and a very limited anti-jamming capability, which make it difficult to meet the requirements of a non-uniform, non-smooth radar operating environment [4] and pose a huge threat to airborne radars performing target detection, tracking and guidance tasks. To ensure the good performance of airborne radars in complex electromagnetic environments, it is necessary to research radar anti-main-lobe jamming.
Designing and transmitting waveforms based on cognitive radar concepts [5] that match target characteristics and jamming spectral features can overcome the limitations of traditional methods, such as improving performance from the receiver side only or randomly shortcutting waveform parameters at the transmitter side, thereby radically improving radar detection performance in main lobe jamming conditions. Since then, waveform design based on cognitive radar thinking has gradually become a hot research topic in the field of waveform design.
This paper aims to design an optimal anti-jamming waveform in a complex and changing electromagnetic environment, based on the electromagnetic environment information acquired by the radar in advance and the mission requirements of the radar, and to lay the theoretical foundation for promoting the development of active anti-jamming technology for radars.

1.2. Related Works

Active radar jamming techniques seriously degrade the target detection, tracking and identification performance of radars and are the biggest threat to radar survival today. Therefore, it is necessary to study radar anti-active jamming techniques. With the improvement in hardware performance and in-depth research on signal processing technology, waveform design is favoured by radar researchers at home and abroad as one of the most effective ways to improve the overall performance of the radar.
Waveform shortcutting is valued for its anti-jamming properties in the face of main lobe jamming [6]. The main ways of dealing with jamming from a waveform perspective are currently frequency shortcuts and phase-coded signal shortcuts. The literature [7] proposed a dense dummy target jamming suppression algorithm based on the joint waveform entropy of shortcut frequency change and provided an outlook on the development trend of frequency shortcut radar waveform countermeasure technology [8]. The literature [9] also proposed an anti-main-lobe jamming method by sensing the jamming environment to achieve low intercept waveform scheduling for radar, which effectively improved the target tracking performance in multiple main lobe jamming environments. In the literature [10], a sparse step linear FM waveform optimization algorithm with frequency and pulse repetition frequency agility was proposed to study the distance and Doppler 2D autocorrelation functions of the shortcut waveform, paving the way for finding a frequency and PRF optimization strategy to suppress the distance and Doppler partials. However, the frequency shortcutting method mostly shortcuts the transmitting frequency in inter- or intra-vehicle according to the hopping pattern [11], and processing the hopping signal within the coherent processing time generates side lobes, resulting in spikes in the Doppler spectrum, and thus increasing the probability of false alarms and reducing the target detection performance of the radar [12]. At the same time, the airborne radar hopping pattern has been preset before working and cannot be changed in real time according to the target and jamming environment, resulting in sometimes jumping into the jamming signal coverage band, which in turn increases the possibility of jamming.
Phase-coded signals are also a common form of signal used for shortcuts [13]. Through rational design, the phase-coded signal can have a high autocorrelation function main-to-side-lobe ratio and also a high mutual quadrature performance, which provides good anti-jamming performance when the jammer cannot keep up with the transmitter changes in time or cannot forward the jamming signal in real time. For co-channel jamming, the literature [14] proposed a phase-coded anti-jamming method in conjunction with the constant false alarm technique. The literature [15] proposed a phase-coded signal system that effectively suppressed the main lobe jamming and improved the target detection probability. In the literature [16], an anti-retransmission jamming technique was proposed using phase- and frequency-coded waveform techniques. To generate effective agile waveforms, an improved Logistic-Map chaotic sequence generation and optimization algorithm was proposed in the literature, using phase and frequency coding techniques to obtain an orthogonal polyphase code set, which effectively suppresses retransmission jamming. To combat repeater jamming, the literature [17] proposes an adaptive transmitting scheme based on phase-coded signals by sensing jamming parameters and using genetic algorithms to optimize the waveform. To address the problem of extended target detection in cluttered and noisy environments, a phase-modulated waveform design method based on a maximum signal-to-noise ratio criterion is proposed in the literature [18]. However, there are significant difficulties in generating a sufficient number of coded signals that are orthogonal to each other. As the number of coded signals increases, the mutual orthogonality of the signals in the set decreases, which can lead to a certain level of jamming performance when forward jamming is based on the previous pulse or pulses.
The above research shows that radar anti-jamming technology has made great progress. However, the existing anti-jamming waveforms are not designed to change in real time according to the environment. The waveforms have limited variation and the echo signal processing is fixed, making it difficult to meet the time-varying, non-smooth and non-uniform operating environment. When airborne radars are confronted with new types of jamming, the practical anti-jamming performance of traditional anti-jamming techniques is limited [4], so there is an urgent need to investigate new waveform design methods to improve radar countermeasures.
Professor Simon Haykin has developed a cognitive radar concept [5] that draws on biological bionic knowledge to design and transmit waveforms that match the characteristics of the target and the jamming spectrum by sensing information about the environment and the target. The cognitive radar concept establishes closed-loop adaptive processing of radar from reception to transmission, overcoming the limitations of conventional radars that can only perform performance enhancements from the receiver side or randomly change transmission parameters at the transmitter side, radically improving airborne radar detection performance under clutter and jamming conditions.
Cognitive radar waveform design, as one of the main ways of active radar anti-jamming at present, can better design radar anti-jamming waveforms adapted to the current environment based on the electromagnetic spectrum characteristics of the signal and the characteristics of the detected target, to improve the detection and identification performance of the radar. Research on radar anti-jamming based on cognitive radar waveform design has yielded fruitful results. To address the extended clutter suppression problem, the literature [19] investigated radar waveform design using inter-pulse phase-encoded signals. For intermittently sampled forwarded jamming, the literature [20] demonstrates the design of constant-mode radar waveforms by reducing the side lobe level. In the context of inaccurate a priori information on both random target impulse response and signal-related clutter, the literature [21] proposed a robust algorithm for the joint design of cognitive radar low peak-to-average ratio transmit waveforms and receiver filters. The literature [22] introduces the basic framework of cognitive radar waveform optimization and systematically compares the main research contents and research progress of waveform optimization for detection, tracking, imaging and classification tasks as well as anti-jamming cognitive waveform optimization, providing a comparative vertical and horizontal perspective for the research of waveform optimization techniques for single tasks and joint multi-tasks. For the problem of adaptive waveform design for extended target recognition in cognitive radar networks, a minimum correlation algorithm is proposed in the literature [23] to adaptively design the transmit waveform for each radar in an amplitude fluctuation situation. However, waveform design based on cognitive radar thinking still requires the establishment of an optimized objective function. In a world where radars are faced with multi-functional and multi-mission requirements, the comprehensive enhancement of radar performance has become an issue that has to be addressed. However, most of the existing waveform design methods use fixed guidelines, which are difficult to adapt to complex electromagnetic environments. Deep reinforcement learning, with its strong ability to interact with the environment, can automate radar waveform design by sensing the input environmental information to design new radar-transmitted waveforms in real time.
Reinforcement learning can be found in many areas of artificial intelligence where decision making and control are required, such as in games and gaming. Reinforcement learning has the ability to interact with the environment in real time and to make optimal decisions by continuously learning information about the environment, and has been successfully applied in many areas, such as Go, gaming and medical diagnosis, even surpassing human levels. Reinforcement learning theory fits well with the idea of cognitive radar in that both involve exploring the environment, sensing information about the surroundings and interacting with it, using it to make the best possible decisions to achieve the ultimate goal. For example, the most important feature of reinforcement learning is that decisions are learned through rewards, and the reward function can correspond to rating criteria in cognitive radar systems, such as the signal-to-jamming-noise ratio, amount of mutual information and probability of detection, among other criteria. While reinforcement learning seeks optimal solutions for specific objectives, cognitive radar requires the design of optimal transmitting waveforms to meet specific mission requirements to achieve optimal radar performance. Reinforcement learning requires consideration of the long-term nature of the strategy, and the constant game of cognitive radar against jamming is also a process of finding winning strategies in terms of long-term goals. Therefore, the use of reinforcement learning theory, which is highly compatible with cognitive radar and game theory, can help radar achieve interactive sensing and adaptive firing with the environment, improve the ability of intelligent sensing, processing and feedback of radar and realize the overall performance of radar in complex adversarial environments [4].
With the continuous research on arbitrary waveform transmitters, the maturity of cognitive radar theory and the wide application of artificial intelligence technology, exploring waveform design methods for intelligent anti-jamming of airborne radars based on cognitive radar ideas and artificial intelligence technology has become a feasible new anti-jamming approach for radars. The method can greatly improve the active anti-jamming performance of radars and better adapt to the complex environment of the future intelligent battlefield. A joint adaptive frequency hopping and pulse width allocation scheme based on reinforcement learning was proposed in the literature [24], and the effectiveness of the scheme was verified by simulation. In the literature [25], an energy-effective power control scheme based on reinforcement learning was proposed for detecting spoofing jamming in split-frequency array MIMO radars, which effectively improved the detection accuracy of spoofing jamming and saved radar energy. The literature [26] improved the agent of radar anti-jamming by designing anti-jamming strategies using reinforcement learning algorithms. The literature [27] proposed a reinforcement learning-based algorithm for cognitive multi-target detection in the presence of unknown jamming statistics, and experiments show that the radar target detection performance is significantly improved under harsh environmental conditions, such as low signal-to-noise ratio and dragging jamming.
In recent years, the field of reinforcement learning has made significant breakthroughs in many artificial intelligence tasks based on a combination of traditional theory and deep learning, giving rise to a new research area of deep reinforcement learning, which provides a viable way to solve complex scenario problems in the real world.
Deep Q network (DQN), one of the classical algorithms for deep reinforcement learning, is highly regarded by the academic community. To address the situation in which cognitive radars do not know the exact jamming model, Li K. et al. [28] used reinforcement learning algorithms to design a frequency hopping strategy based on frequency agile radar. Based on this, Li K. et al. [29] also proposed a DQN-based frequency agile (FA) radar strategy design method, which achieves radar anti-jamming with a high target detection probability. Then, they [30] applied deep reinforcement learning to the design of an anti-main-lobe jamming strategy for frequency shortcut radar and achieved intelligent selection of pulse carrier frequency. The literature [31] used deep Q networks and long- and short-term memory networks to generate frequency hopping strategies under a partially observable Markov decision process model to improve cognitive radar anti-jamming performance. To address the problem of coexistence between radar and communication systems, the literature [32,33] modelled the radar environment as a Markov decision process to predict the frequency band with the lowest jamming energy and successfully avoided jamming from communication systems. In the literature [34], a deep deterministic policy gradient cognitive design based on a control framework was proposed for the radar power allocation problem in multi-manoeuvre target tracking.
As a value-based algorithm, DQN can only solve finite and discrete action spaces. A series of improvements have been proposed in academia for the DQN algorithm and many research results have been published [35], such as distributional DQN, double DQN, duelling DQN, noisy DQN, etc. However, almost all DRL algorithms have different characteristics and limitations in their applications, as shown in Table 1. Therefore, when selecting DRL algorithms, the need for realistic scenarios and problems is fully considered. The goal of the development of radar countermeasure technology is to achieve intelligent radars and jammers with an observing environment and independent decision making. To solve the problem of radar countermeasures with complex strategies, deep learning techniques with good perception characteristics can be incorporated into reinforcement learning algorithms to better realize environment observation, learning calculation, decision making and feedback, forming a closed-loop process of “observation-action-feedback”, which can effectively improve the overall performance of the radar. Figure 1 shows the steps in the implementation of a rational application of the DRL algorithm to a practical problem, where the red line indicates co-design between modules and the dotted line indicates optional.
As shown in Figure 1, a DRL algorithm that wants to solve a practical task must go through the following steps: The first step is to conduct a requirements analysis to objectively assess the adaptability and performance improvement of the DRL for the task at hand and to identify the agent and environment in the task. Next, the core elements of the state space, action space and reward function are designed separately to complete the complete definition of the reinforcement learning problem. Then, a suitable DRL algorithm is selected with the task characteristics, and the policy model is obtained after training and debugging. If the performance of the policy is not satisfactory, the above process and algorithm details need to be continuously improved based on the results of the experiments and additional performance enhancement measures taken if necessary until a satisfactory policy is achieved. Ultimately, the trained policies are deployed in real application environments.
In summary, to address the problem of intelligent anti-jamming of airborne radars in complex electromagnetic environments, we should introduce cognitive radar ideas and establish Markov decision process (MDP) models to describe radar and target jamming countermeasures from the perspective of waveform design. Interaction between airborne radar and environmental information is determined via analyzing the spectral characteristics of target response, jamming signals, clutter response and airborne radar transmit waveforms. Using deep reinforcement learning algorithms, we study waveform design methods for airborne radars under jamming conditions and explore feasible ways for airborne radars with predictive and anti-intelligent jamming capabilities, laying a solid foundation for the realization of intelligent radars.

1.3. Main Contributions

In the first stage, we studied the anti-jamming problem of the airborne radar in a complex environment and proposed a radar anti-jamming waveform design method based on the strategy iteration method [43]. However, this method can only solve the problem of low-dimensional discrete spatial reinforcement learning, and it applies to situations where there are many influencing factors, high data dimension and a large scale in the actual radar confrontation process. To solve the above problems, we propose a deep reinforcement-learning-algorithm-based airborne radar waveform design method for clutter and jamming conditions. Firstly, a Markov decision process-based game model between airborne radar and jammer is developed to describe the actual operating environment of airborne radar containing information on target characteristics, clutter, noise and jamming signals, etc., by combining cognitive radar ideas. Secondly, we build deep neural networks to sense, analyze and learn from complex electromagnetic environments. Then, the optimal transmitted waveform for airborne radar is designed based on the D3QN algorithm. Finally, an iterative transformation method is used to synthesize the time domain radar signal for the optimal waveform strategy.
The main contributions of this paper are as follows.
  • To address the problems of single and idealized airborne radar countermeasure scenario models, this paper investigates a Markov decision process (MDP)-based approach for modelling dynamic airborne radar countermeasure scenarios. The airborne radar electromagnetic environment under clutter and jamming conditions is investigated, and the dynamic countermeasure process between airborne radar and jamming is modelled using MDP. The MDP-based radar modelling approach allows for the analysis of the electromagnetic information present in realistic scenarios and the arbitrary design of countermeasure models containing a variety of influencing factors such as noise, clutter, the impulse response of the detected target, radar-transmitted signals and jamming signals. The method overcomes the limitations of traditional models and improves the flexibility, accuracy and predictability of the description of airborne radar operating scenarios, providing a viable way to realize the increasingly complex modelling of actual airborne radar operating scenarios.
  • This paper proposes an intelligent waveform design method based on DRL to address the problem of anti-jamming of radar under clutter and jamming conditions. Radar-environment information interaction is determined by modelling the MDP for radar and jamming countermeasures. A radar frequency domain optimal anti-jamming waveform strategy generation method based on the D3QN algorithm is proposed. We build two deep neural networks to fit the state value function and action value function of the radar. The “overestimation” problem of the value function is solved using the fixed Q target method and the preferred experience replay method. To improve radar learning efficiency, prevent training overfitting, etc., the action value function network adopts a pairwise structure. Once the optimal frequency domain waveform strategy has been computationally selected, the ITM is used to generate the corresponding time domain signal with a constant envelope. The simulation results show that the radar transmit waveform designed by this method achieves intelligent anti-jamming of airborne radar while further improving the radar target detection probability compared to the conventional LFM signal, based on the radar waveform strategy generated by the strategy iterative method.
  • Through extensive simulation experiments, the results show that the optimal waveform design method for airborne radar proposed in this paper significantly improves the anti-jamming performance and target detection performance of airborne radar, and outperforms traditional linear FM signals and previous research results based on reinforcement learning.

2. Airborne Radar Signal and MDP Model

2.1. Airborne Radar Signal Model

Figure 2 shows the airborne radar detection scene. To establish the airborne radar signal model in a complex electromagnetic space, we need to fully consider the influence of radar-transmitted signals, enemy jamming signals, target echo, noise and all kinds of environmental clutter and other factors. Figure 3 shows the airborne radar signal model, where s ( t ) is the radar-transmitted signal, the Fourier transform is S ( f ) , the signal bandwidth is W and the total power is P S ; j ( t ) is the jammer signal, the power spectral density is J ( f ) and the total power is P J . The Fourier transforms of the target impulse response h ( t ) and the receive filter impulse response r ( t ) are H ( f ) and R ( f ) , and h ( t ) is a time-limited random model. Information such as target impulse response and clutter response can be acquired during the airborne radar search phase. The clutter c ( t ) is a non-Gaussian random process, and the power spectral density S c ( f ) is not constant within W . The noise n ( t ) is a zero-mean Gaussian channel process with a power spectral density S n ( f ) that is not zero within W .
The expression of the signal y ( t ) at the output end of the radar receiver filter is as follows [44]:
y ( t ) = r ( t ) ( s ( t ) h ( t ) + s ( t ) c ( t ) + n ( t ) + j ( t ) )
where “*” is the convolution operator. The radar signal components and jamming components are expressed as:
y s ( t ) = r ( t ) ( s ( t ) h ( t ) )
y j ( t ) = r ( t ) ( s ( t ) c ( t ) + n ( t ) + j ( t ) )
At the time t 0 , the frequency domain expression of the signal-to-jamming-noise ratio (SJNR) is
( S J N R ) t 0 = y s ( t 0 ) 2 E ( y j ( t 0 ) 2 ) = + R ( f ) H ( f ) S ( f ) e j 2 π f t 0 d f 2 + R ( f ) 2 ( S c ( f ) S ( f ) 2 + J ( f ) + S n ( f ) ) d f
h ( t ) is a time-limited stochastic model, and the power spectral density can be replaced by energy spectrum variance [45].
σ h 2 ( f ) = E ( H ( f ) μ h ( f ) 2 )
Assuming that means is 0, substitute Equation (5) into Equation (4) and use Schwartz’s inequality to solve for Equation (6).
( S J N R ) t 0 + R ( f ) 2 ( S c ( f ) S ( f ) 2 J ( f ) + S n ( f ) ) d f + σ h 2 ( f ) S ( f ) 2 S c ( f ) S ( f ) 2 J ( f ) + S n ( f ) d f + R ( f ) 2 ( S c ( f ) S ( f ) 2 + J ( f ) + S n ( f ) ) d f
The equals sign holds under the condition that the SJNR takes the maximum value when and only when R ( f ) = [ k σ h ( f ) S ( f ) e j 2 π t 0 ] S c ( f ) S ( f ) 2 + J ( f ) + S n ( f ) ( k is an arbitrary constant). Assuming that the jammer can obtain the radar signal spectrum and adjust the jamming to the same band of the radar signal to achieve the maximum jamming effect, at this point, the following is obtained:
( S J N R ) t 0 = + σ h 2 ( f ) S ( f ) 2 S c ( f ) S ( f ) 2 + J ( f ) + S n ( f ) d f Δ f W k = 1 K σ h 2 ( f k ) S ( f k ) 2 S c ( f ) S ( f k ) 2 + J ( f k ) + S n ( f k )
where K is the number of frequency samples, and Δ f is the frequency sampling interval, K Δ f = W .

2.2. MDP Model

Agents are actionable entities for reinforcement learning. For self-driving cars, the environment is the current road conditions; for Go, the status is the current game. At each moment, the agent and environment have their state, such as the current position and speed of the car, the vehicle and pedestrian conditions on the road. The agent determines an action based on the current state and performs that action. It then moves on to the next state with the environment, and the system gives it a feedback value, rewarding or punishing the action, and forcing the agent to perform the desired action. As shown in Figure 4.
Reinforcement learning is a class of methods that solve such decision-making problems. The algorithm learns a mapping function called the strategy function through sample learning, whose input is the environment information at the current moment, and the output is the action to be performed, in this case a = π ( s ) , where s is the state, a is the action to be performed, and the state and action are from the state set and the action set, respectively. Actions and states can be discrete or continuous. For the former, the set of actions and states is finite, and for the latter, it is an infinite set. The goal of acting is to achieve a certain purpose, such as driving safely in an unmanned car, winning this Go game and modelling this with a reward function.
The problems to be solved by reinforcement learning can be abstracted into the MDP. Markov processes are characterized by the fact that the state of the system at the next moment is determined by the state of the current moment, regardless of the earlier moment. Unlike the Markov process, in MDP, the agent can perform actions that change the state of itself and the environment and receive a punishment or reward.
The MDP model can be described by a five-tuple S , A , P , R , γ , where S is the state space, A is the action space, P is the state transfer probability, R is the payoff function, and γ is the discount factor [46]. The state space S is the set of all the factors involved in the confrontation scenario; the action space A is the set of all possible behaviours. The state transfer probability function P ( s , a , s ' ) is the probability of the agent taking action a from the state s to transition to the state s ' . The reward function R ( s , a , s ' ) is the average value of the cumulative rewards for the agent performing action a transition to a state s ' in the state s . A positive reward value is rewarded for the selected action and a negative one is penalized. To ensure the convergence of the algorithm, a discount factor γ [ 0 , 1 ] is set. When γ tends to 0, the agent tends to obtain immediate rewards; when γ tends to 1, the agent prefers to obtain long-term gains.

3. MDP-Based Radar Countermeasure Environment Modelling

3.1. Radar Action, Status and Reward Design

The radar transmits signals to detect the target and the jammer performs jamming on the received radar signals. Radar for the electromagnetic space of jamming signals, environmental clutter and the characteristics of the target being detected integrates this information to design a new radar-transmitted signal to complete the anti-jamming and target detection tasks. The waveform changes in the process of radar-jamming confrontation are Markovian, and the radar confrontation environment can be modelled as an MDP model to achieve intelligent anti-jamming waveform design for airborne radars by interacting with environmental information, such as environmental noise, clutter and jamming signals. Figure 5 briefly describes the MDP-based radar countermeasure process, where blue represents the radar signal, the red waveform represents the jamming signal, S i is the current state, R i is the reward for the current state, and the number on the arrow indicates the transfer probability between states.
The radar signal s ( t ) and the jamming signal j ( t ) are equally divided into M sub-bands in the frequency domain, and the sub-band power is equally divided into N parts; i.e., the individual signal is modelled as an array consisting of M numbers, as shown in Equation (8).
s ( t ) , j ( t ) [ f 1 , f 2 , , f M ] , f i [ 0 , 1 , , N ]
All jamming signals form the state space S , which is defined as
S = ( s 1 , s 2 , , s ( N + 1 ) M )
where the subscript ( N + 1 ) M denotes the size of the state space S , and α i [ 0 , 1 , , N ] , i [ 1 , 2 , , M ] denotes the signal power allocation size of the sub-band i of a single state s t at the moment t , which can be expressed as
s t = [ α 1 , α 2 , , α M ]
Similarly, all radar signals form the action space A , which is defined as
A = ( a 1 , a 2 , , a ( N + 1 ) M )
where the subscript ( N + 1 ) M denotes the action space A size, and β i [ 0 , 1 , , N ] , i [ 1 , 2 , , M ] denotes the signal power allocation size of the sub-band i of action a t taken at the moment t , which can be expressed as
a t = [ β 1 , β 2 , , β M ]
The reward is the key factor that affects the good or bad decision. The SJNR of the radar signal is used as the action reward of the intelligent body decision, and the reward value size is proportional to the radar signal SJNR; the larger the SJNR, the larger the reward.
Reward S I N R Δ f W k = 1 K σ h 2 ( f k ) S ( f k ) 2 S c ( f ) S ( f k ) 2 + J ( f k ) + S n ( f k )

3.2. Confrontation Model Key Parameter Settings

Setting M = 5 ,   N = 5 , the airborne radar and jammer MDP game model uses an array of five digits to represent the frequency domain energy distribution state of the radar signal s ( t ) and jammer signal j ( t ) . The signal frequency domain is divided into five sub-bands, and the sub-band power is divided into five equal parts with a discount factor γ = 0.9 . The specific parameter settings are shown in Table 2.

4. D3QN-Based Strategy Generation for Optimal Radar Anti-Jamming

In this paper, the MDP model is established for the radar anti-jamming environment under clutter and jamming conditions. Firstly, the deep neural network is trained to realize the perceptual interaction with complex electromagnetic environment factors such as clutter, noise and jamming. Second, the SJNR of the radar signal is set as the reward function by combining the impulse response of the detected target acquired in advance. Then, the frequency domain energy optimal strategy of the radar anti-jamming signal is obtained based on the D3QN algorithm, and its time domain constant mode signal is synthesized to realize the intelligent anti-jamming waveform design for airborne radar.
The principle of D3QN-based radar intelligent anti-jamming is shown in Figure 6. The radar is regarded as the agent, the jamming signal in the electromagnetic space as the state information, the radar-transmitted signal as the action and the SJNR of the radar signal as the reward function. This is achieved by firstly continuously calculating the reward values for taking different actions in different states, and storing the corresponding state, action and reward information for training the neural network. Secondly, the neural network analyzes, calculates and selects the radar anti-jamming strategy with the highest Q value. Finally, the selected optimal anti-jamming strategy is output to the radar to synthesize the time domain signal and transmit it.

4.1. Fixed Q Targets

Due to the max operation, the value function has an overestimation problem at each point of the value, and an uneven amount of overestimation leads to a suboptimal solution, so a method for fixing the Q target is needed.
First, an evaluation network and a target network are created. A target network with fixed parameters ω t is used to estimate the time difference target. Second, a function is created to obtain the evaluation network parameters and replicate them to the target network. After performing N updates, the parameters ω e of the evaluation network are periodically copied to the target network to update the parameters ω t of the target network.
Δ ω = α [ ( R + γ max a Q ( s , a , ω t ) ) Q ( s , a , ω e ) ] w Q ( s , a , ω e )
At   every   N   step :   ω t ω e
Then, the evaluation network is used to obtain the action corresponding to the optimal action value in the state s t + 1 , and the target network is used to calculate the action value of this action to obtain the target value. The interaction of the two networks effectively avoids the problem of the “overestimation” of the algorithm.
y t = r t + 1 + γ q ( s t + 1 , arg max a q ( s t + 1 , a ; ω e ) ; ω t )

4.2. Prioritized Experience Replay

To help the radar fully learn and adapt to the current electromagnetic countermeasure environment and to improve the speed of environmental information acquisition, an empirical replay method is used. First, the data, such as the action, a , selected by the radar, the new jamming state received, s , the calculated reward value, R , and whether the interaction is in the terminated state, d o n e , in each jamming state, s , during the interaction between the radar and the complex electromagnetic environment, are recorded as { S , S , A , R ,   d o n e } , and all are stored in the replay buffer D . Second, when the number of data reaches a set batch size, a batch of data is evenly and randomly selected and put into the neural network for training. Then, after training, the radar continues to interact with the electromagnetic environment and repeats the operation. A batch size of data is sampled uniformly and randomly each time to train the neural network, which improves data utilization and reduces the overfitting problem arising from training. The experience replay process is shown in Figure 7.
However, the uniform random sampling of data during the neural network training process suffers from insufficient learning of important experiences. To avoid the possibility that more important radar countermeasure experiences may occur less frequently than others, a priority experience replay method is adopted so that important experiences have a high probability of being replayed, thus greatly improving the efficiency of the utilization and learning of radar countermeasure data [40].
The radar adversarial process is a model-free reinforcement learning, which requires the use of the concept of time difference error (TD error) to represent the difference between the current Q value and the target Q value as a measure of the importance of each set of radar training data. If the TD error is larger, it means that there is still much room for improvement in prediction accuracy, and the more intelligent the body updates, the more this sample needs to be learned, i.e., the higher the priority p is. Let the TD error at the sample i be σ i ; then, the sampling probability at the sample is [40].
P ( i ) = p i a k p k a
where p i denotes the priority level and both are greater than 0, α determines how many heavyweights to use, and if α = 0 , no importance sampling is used, that is, uniform random sampling. Priority p i is defined by the proportional prioritization method, as shown in Equation (18).
p i = δ i + e
where δ i is the magnitude of the TD error and e is a small normal number, ensuring that some special edge examples with a TD error of 0 can also be extracted.
With probability distribution sampling with priority playback, the action value estimate is biased because the sample distribution and the action value function distribution are two completely different distributions. To correct this deviation, it is necessary to multiply it by an importance sampling coefficient, ω i [40]:
ω i = ( 1 N 1 P ( i ) ) β
These weights are added to the Q network update using ω i δ i .

4.3. Value Function V and Dominance Function A

In the radar intelligent anti-jamming training, to improve the radar learning efficiency and prevent training overfitting and other situations, the action value function network outputs the state value function and the dominance function, respectively, in the hidden layer in the middle of the network by adopting a dyadic network structure using two fully connected sequences. This is shown in Figure 8 [41].
The advantage function A ( s , a ) represents the advantage of taking this action in the current state, and the advantage here is the advantage of the action value function over the value function of the current state, as shown in Equation (20).
A ( s , a ) = Q ( s , a ) V ( s )
where the action value function Q ( s , a ) is the value corresponding to a single action. V ( s ) is the action value function of all possible actions in the state s multiplied by the probability sum of taking the action; that is, the value function V ( s ) is the average of all the action value functions in that state concerning the probability of action, as shown in Equation (21).
V π ( s ) = a A π ( a | s ) q π ( s , a )
By calculating the advantage function, you can compare the average size of the action value of the current radar anti-jamming strategy relative to the action value function. If the advantage function is greater than 0, the action is better than the average action, and vice versa; the current action is not as good as the average action.
Thus, the Q value corresponding to each action can be calculated using Equation (22), and the output of the action value function network is:
Q ( s , a ; θ , α , β ) = V ( s ; θ , β ) + ( A ( s , a ; θ , α ) 1 A a A ( s , a ; θ , α ) )
Among them, θ is a commonly used network parameter, α is the advantage function network parameter, β is the state value function network parameter, and 1 A a A ( s , a ; θ , α ) is a centralized processing of the advantage function to ensure that there will be zero advantage under a certain action.

4.4. Design Flow of Optimal Anti-Jamming Strategy for Radar Based on the D3QN Algorithm

In summary, combined with the actual countermeasure scenario, the airborne radar anti-jamming problem is transformed into a reinforcement learning problem, the countermeasure model between radar and target is established, and the optimal anti-jamming strategy is obtained by describing the electromagnetic environment and training the neural network based on the D3QN algorithm with the following algorithm steps (Algorithm 1).
Algorithm 1: Airborne radar anti-jamming strategy design algorithm based on D3QN.
  • Initializes the current Q network parameter θ , and initializes the target Q network parameter θ . The Q network parameters are assigned to the Q network, θ θ . The total number of iterations T , discount factor γ , exploration rate ε , target Q network parameter update frequency P , and the number of samples sampled at a time m are determined.
  • Initialize the replay buffer D .
  • For t = 1 to T :
1)
Initialize the environment and obtain the state S (i.e., jamming signal), R = 0 , done = False .
2)
while True:
a)
Based on the state ϕ ( S ) acquisition, input the current Q network, calculate the Q value corresponding to each action (i.e., radar signal), and use the ε -greedy algorithm to select the action A corresponding to the current state S .
b)
Perform the action A , obtain new status S and rewards R , whether the anti-jamming process is in the end state done
c)
Save 5 elements { S , S , A , R ,   d o n e } into D .
d)
if done :
break
e)
Randomly sample m samples from D , { S , S j , R j , A j , done j } , j , 1 , 2 , 3 , 4 , , m . Calculate y j of the current Q network y j = R j + γ Q ( ϕ ( S j ) , arg max a Q ( ϕ ( S j ) , a ; θ ) ; θ ) .
f)
Use the mean squared loss function ( 1 m ) r = 1 n ( y j Q ( ϕ ( S j ) , A j , θ ) ) 2 , calculate loss, and backpropagate update parameter θ .
g)
if t % p = 0 : θ θ
h)
S = S

4.5. Performance Indicators

The radar detection problem in this paper can be defined as a hypothesis testing problem [47]. By solving the classical Neyman–Pearson theorem, the radar target detection probability P D is obtained as
P D = Q ( Q 1 ( P F A ) d 2 )
where P F A is the probability of a false alarm, P F A = Q ( τ ) = τ 1 2 π exp ( 1 2 t 2 ) d t , τ is the radar detection threshold, and d 2 is the offset coefficient, which is the SJNR in this model. The detection performance of such detectors is completely determined by the offset coefficient, so by calculating the SJNR, the relationship between radar waveform and target detection probability can be established.

5. Simulation Results and Analysis

The simulation was carried out based on foreign airborne radar setting parameters such as operating band, centre frequency and signal bandwidth, as well as target flight speed, target impulse response and environmental clutter information, as shown in Table 3. The environmental information is shown in Figure 9. Yellow indicates the target impulse response information and green indicates the environmental clutter information. The signal power of the sub-bands in all result maps is in per cent.

5.1. Radar and Jamming Game Strategy Generation

During the confrontation between radar and jamming, the radar generates the optimal anti-jamming waveform frequency domain strategy by interacting with the environment using the D3QN algorithm. To verify the effectiveness of the method in this paper, experimental comparisons were made between the conventional LFM signal of the radar and the frequency domain anti-jamming strategy generated by the team’s previous iterative method based on the classical reinforcement learning strategy under the same experimental conditions, respectively, and the results are shown in Figure 10.
As can be seen from Figure 10a, because the LFM signal has a large temporal bandwidth product, and the top undulation of the amplitude–frequency characteristics of the signal gradually decreases with the increase in the temporal bandwidth product and approaches rectangular shape, but the signal power is uniformly distributed in five sub-bands, which cannot be used for radar anti-jamming according to the current electromagnetic environment information and target characteristics. From Figure 10b,c, it can be seen that both the radar anti-jamming strategy generated based on the strategy iteration method and the radar anti-jamming strategy generated based on the D3QN algorithm in this paper can respond to the clutter, jamming and target characteristics to some extent to achieve the anti-jamming effect. As shown in Figure 10b, the radar anti-jamming strategy based on the strategy iterative method mainly distributes the radar signal power in sub-frequency bands 1 and 3 with no jamming signal and strong target impulse response, indicating that the strategy can respond to environmental information to a certain extent. In Figure 10c, under the same environmental conditions, the radar optimal anti-jamming strategy obtained by the proposed method is more sensitive to the clutter, jamming signal and the impulse response of the detected target, and the strategy distributes most of the radar signal power in the sub-frequency band with the low power of the clutter and jamming signal and the sub-frequency band 1, 2 and 4 where the target impulse response is strong, to obtain a higher radar SJNR and the detected target information.

5.2. Performance Comparison

The SJNR is an important parameter for measuring the performance of radar target detection. To verify the performance of the D3QN-based algorithm for airborne radar anti-jamming strategy design, the optimal anti-jamming waveform strategy designed in this paper is compared with the LFM signal and the anti-jamming signal designed by the classical strategy iteration method in reinforcement learning, and the target detection performance of the optimal waveform strategy is analyzed. The radar receiver SJNR and radar target detection probability are calculated by Equations (7) and (23), respectively, to compare and analyze the optimal strategy performance, and the simulation results are shown in Figure 11.

5.2.1. Performance Analysis When the Total Target Disturbance Power Pb Is a Constant Value

The simulation results in Figure 11 were achieved under experimental conditions where the jamming was a fixed total power of 10 W and the transmit power of the radar signal was P s = 1 30 W . The variation in the SJNR and target detection probability with the total radar signal power for the three radar waveform strategies is shown in Figure 11.
The simulation results in Figure 11 were achieved under the experimental conditions where the jamming was a fixed power of 10 W and the transmit power of the radar signal was increased from 1 W to 30 W. The blue line in the figure indicates the LFM signal, and the green and red lines indicate the waveform strategy generated by the strategy iterative method and the optimal strategy generated by the D3QN algorithm based on this paper, respectively. From Figure 11a, it can be seen that the SJNR of all three different signals keeps increasing as the radar signal power increases, and the anti-jamming strategy generated by the D3QN algorithm based on the D3QN algorithm is significantly better than the other two strategies, and the SJNR of the anti-jamming strategy generated by the strategy iteration method based on the reinforcement learning tradition is also higher than that of the radar LFM signal. At the radar signal power of 10 W, the SJNR of the optimal strategy generated by this paper is up to 15.8 dB, which is 1.8 dB and 3.0 dB higher than that of the signal generated by the strategy iterative method and the LFM signal, respectively. Similarly, it can be seen from Figure 11b that among the three different signals, the signal generated by this method has the highest target detection probability, the signal generated by the MDP-based strategy iterative method has the second highest probability, and the LFM signal has the lowest detection probability. The LFM signal has the lowest detection probability. At the radar signal power of 5 W, the target detection probability of the optimal strategy generated by this paper is up to 97%, which is 27% and 44% higher compared to the signal generated by the strategy iterative method and the LFM signal, respectively.

5.2.2. Performance Analysis When the Total Radar Signal Power Ps Is a Constant Value

The simulation results in Figure 12 were achieved under experimental conditions where the total power of the radar transmit signal was fixed at 10 W and the total target jamming power was P j = 1 30 W . The SJNR and target detection probability of the three radar waveform strategies with the variation in the total jamming power are shown in Figure 12.
From Figure 12a, it can be seen that the SJNR of all three radar transmit signals decreases continuously as the jamming signal power increases, but the SJNR of the anti-jamming waveform strategy generated by the D3QN algorithm decreases significantly less than the other two strategies, and the anti-jamming strategy based on the strategy iteration algorithm also significantly outperforms the LFM signal. Similarly, as can be seen from Figure 12b, the target detection probability of the three radar signals decreases as the jamming power increases, but the optimal radar strategy obtained in this chapter has a lower target detection probability than the other two strategies, being 2.01 dB and 2.68 dB higher, respectively. Until the jamming power increases to 30 W, the target detection probability of this optimal strategy is still 99.84%, which is 7.57% and 15.89% higher than the signals generated by the iterative strategy method and the LFM signal.

5.2.3. Performance Analysis When the Total Radar Signal Power Ps Is a Constant Value

For more visualization, a 3D plot reflecting the variation in the SJNR with P s = 1 30 W and P j = 1 30 W for the three radar waveform strategies in the clutter and jamming environment is shown in Figure 13.
Figure 13 shows the changes in the SJNR of the three signals, including the LFM signal, the radar anti-jam signal generated based on the strategy iterative method and the radar signal designed by the proposed method, when both the jamming power and the radar power increase from 1 W to 30 W. It can be seen that the SJNR of all three signals increases with the increase in radar power and decreases with the increase in jamming power. The simulation shows that the SJNR of the radar anti-jamming waveform generated by the method in this paper is the largest and has better anti-jamming performance when the jamming power is the largest (30 W) and the radar signal power is the smallest (1 W) under the same conditions.

5.3. Time Domain Waveform Synthesis

To better apply this method to practical radar detection systems and improve target detection performance, it is necessary to generate the time-domain-transmitted signals of the optimal waveform strategy in the radar frequency domain. There are many methods to obtain the time domain characteristics of the generated signal, and the simplest method is the direct inverse fast Fourier transform (IFFT) method, which performs an IFFT on the optimal amplitude spectrum and then normalizes the transformed signal in terms of amplitude. However, the time domain signal synthesized by this method differs significantly from the optimal strategy [48]. The fixed-phase technique is the conventional method for synthesizing nonlinear frequency-modulated signals and uses Newton’s method to calculate the numerical solution, which is complicated to derive. Jackson et al. [49] used the iterative transform method (ITM) to generate a constant envelope time domain signal with the best spectral fit. Therefore, in this section, ITM is used to fit the time domain signal with the optimal strategy in the frequency domain. The time domain form and spectral fitting results of the optimal radar waveform strategy designed based on the D3QN algorithm are as follows.
The radar time domain signal is synthesized according to the radar optimal anti-jamming frequency domain strategy in Figure 10c, and its frequency domain characteristics are verified, and the results are shown in Figure 14. Among them, Figure 14a depicts the real part, imaginary part, amplitude spectrum and phase spectrum of the synthesized time domain signal; Figure 14b shows the results of the frequency domain characteristics of the verified time domain waveform; the red dashed line is the optimal frequency domain strategy designed by the method in this paper, and the black solid line indicates the spectrum of the synthesized time domain signal by ITM. It can be seen that the synthesized time domain signal better achieves the frequency domain characteristics of the optimal strategy and has low interception performance, showing behaviours such as constant envelope and anti-jamming.

6. Conclusions

Airborne radars play a vital role in modern national defence by carrying out important tasks such as target search, detection, identification, tracking and imaging. With the rapid development of electronic countermeasure technology, the electromagnetic signals emitted by numerous electronic devices on the battlefield, enemy intelligent jamming signals and complex terrain clutter pose a serious threat to the survival and performance of airborne radars. Therefore, it is necessary to conduct research on anti-jamming technology for airborne radars in complex environments. In recent years, artificial intelligence algorithms have been widely used to provide a viable path for the implementation of cognitive radar. Cognitive radar waveform design is a main research area, involving modelling complex real-world operating environments. How to realize the interactive sensing of radar and electromagnetic environment information, make full use of target characteristics and jamming signal spectrum characteristics, design radar-transmitted waveforms in real time to adapt to complex environments and detected targets and achieve good anti-jamming while improving radar target detection performance is the focus of current research.

6.1. Article Work Summary

This paper outlines the current research status of airborne radar anti-jamming technology with research on deep-reinforcement-learning-based radar waveform design in complex environments and a focus on solving the dynamic confrontation problem between radar and jamming in electronic warfare, fundamentally improving radar’s active anti-jamming capability, enhancing radar adaptability in complex environments and laying the foundation for the development of radar confrontation. The main contributions of this work are as follows:
  • In response to the problems of single and idealized radar countermeasure scenario models in traditional anti-jamming technology research, we study a dynamic radar countermeasure scenario modelling method based on the Markov decision process. The MDP model overcomes the limitations of traditional models, improves the accuracy and predictability of the actual radar operating environment and provides a model basis for radar waveform design research. The MDP model overcomes the limitations of traditional models, improves the accuracy and prediction of the actual radar operating environment and lays the model foundation for post-radar waveform design research, further promoting the development and application of cognitive radar.
  • To address the problem of dynamic countermeasures between airborne radar and jamming in complex environments, we propose a deep-reinforcement-learning-based anti-jamming waveform design method for airborne radar under clutter and jamming conditions from a waveform design perspective and build a dynamic countermeasure model for airborne radar and jamming in complex environments based on MDP. We construct two deep neural networks for training learning of the neural networks by iteratively computing the reward values to the adversarial data using different radar waveforms in different jamming states. We recommend a design technique for the optimal anti-jamming waveform strategy of airborne radar based on the D3QN algorithm and synthesize the optimal time domain transmitted signal by iterative transformation method. It has certain reference significance for the realization of radar intelligent confrontation technology in electronic warfare and has laid a theoretical foundation for the realization of cognitive radar.

6.2. Future Research Perspectives

This paper mainly aims to investigate the design of anti-jamming waveforms for airborne radars in complex environments based on deep reinforcement learning, drawing on cognitive radar ideas. The paper has carried out some work at the basic theoretical level and obtained some preliminary research results, which should be followed up by research on the following aspects and shortcomings. The main areas include the following:
  • The waveform design problem for complex environments has focused on the waveform design problem for electromagnetic information environments containing clutter, noise and jamming, while the waveform design for the non-smooth and non-uniform actual combat environments faced by airborne radars has still not been considered. Therefore, subsequent work should fully investigate waveform design methods for airborne radars in various non-smooth, non-uniform electromagnetic environments, such as clutter and new types of jamming signals, and discuss the performance of waveform design technology methods in a variety of complex environments, which is of great significance for improving the anti-jamming performance of airborne radars in actual combat.
  • In response to traditional anti-principal flap jamming techniques’ difficulty coping with diverse and dexterous new types of jamming, we investigate a deep-reinforcement-learning-based approach to airborne radar waveform design. When building a dynamic countermeasure model between radar and jamming based on a Markov decision process, a discrete signal form and a reward function design based on a single performance indicator of the SJNR are mainly used, without considering continuous state and action representations, and a reward function setting that combines multiple performance indicators. The future development trend of radar must be multi-functional and intelligent, and in the subsequent research, attempts should be made to establish a continuous state and action space, which is conducive to the application of airborne radar intelligent waveform design technology in practice; the use of multiple radar performance indicators to combine the design of reward functions is of great significance to the development of multi-functional radar waveform design.
  • To address the difficulty of the practical application of theoretical waveform design methods, we focus on the design of frequency domain optimal anti-jamming waveform strategies based on artificial intelligence algorithms, using an iterative transformation method to synthesize the corresponding constant envelope time domain signal. However, only the real, imaginary, amplitude and phase information of the time domain signal is obtained, and no actual transmittable time domain waveform is found. Most of the current radar waveform design studies do not consider the actual emission of the designed waveform. On the one hand, this is because the modelling of the problem deviates significantly from the actual environment when considering waveform design methods, making it difficult to apply to real-world scenarios. On the other hand, the hardware is limited by the production process of electronic equipment, and current radar transmitters are not yet able to meet the demand for transmitting arbitrary waveforms. Therefore, to ensure that radar waveform design research results can be used in practical equipment and truly improve airborne radar performance, future radar waveform design research should be closely integrated with practical application scenarios and equipment, so as to lay a solid foundation for the successful implementation of airborne radar waveform design and the smooth development of new system radars.

Author Contributions

Conceptualization, Z.Z., W.L. and K.Z.; methodology, Z.Z., W.L. and K.Z.; software, Z.Z.; validation, Z.Z.; formal analysis, Z.Z. and W.L.; data curation, Z.Z.; writing—original draft preparation, Z.Z.; writing—review and editing, Z.Z. and W.L.; visualization, Z.Z.; supervision, W.L.; project administration, W.L. and K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Natural Science Foundation of China (Project No. 62271500) and the Natural Science Foundation of Shaanxi Province (Project 2020JM-343).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DQNdeep Q network
DRLdeep reinforcement learning
D3QNduelling double deep Q network
IFFTinverse fast Fourier transform
ITMiterative transform method
LFMlinear frequency modulation
MDPMarkov decision process
RLreinforcement learning
SJNRsignal-to-jamming plus noise ratio
TD errortime difference error

References

  1. Li, Z.; Tang, B.; Zhou, Q.; Shi, J.; Zhang, J. A review of research on the optimal design of new system airborne radar waveforms. Syst. Eng. Electron. Technol. 2022, 1–20. Available online: http://kns.cnki.net/kcms/detail/11.2422.TN.20220328.2210.037.html (accessed on 6 November 2022).
  2. Zhao, G. Principles of Radar Countermeasures, 2nd ed.; Xidian University Press: Xi’an, China, 2015. [Google Scholar]
  3. Skolnik, M. Translated by Nanjing Electronics Technology Research Institute. Radar Handbook; Electronic Industry Press: Beijing, China, 2010. [Google Scholar]
  4. Chen, X.; Xue, Y.; Zhang, L.; Huang, Y. Airborne Radar System and Information Processing; Electronic Industry Press: Beijing, China, 2021; pp. 333–347. [Google Scholar]
  5. Haykin, S. Cognitive radar networks. In Proceedings of the 2005 1st IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing, Puerta Vallarta, Mexico, 13–15 December 2005; pp. 1–3. [Google Scholar]
  6. Yan, Y.; Chen, H.; Su, J. Overview of anti-jamming technology in the main lobe of radar. In Proceedings of the 2021 IEEE 4th International Conference on Automation, Electronics and Electrical Engineering (AUTEEE), Shenyang, China, 19–21 November 2021; pp. 67–71. [Google Scholar]
  7. Fang, W.; Quan, Y.; Sha, M.; Liu, Z.; Gao, X.; Xing, M. Dense False Targets Jamming Suppression Algorithm Based on the Frequency Agility and Waveform Entropy. Syst. Eng. Electron. 2021, 43, 1506–1514. [Google Scholar]
  8. Quan, Y.; Fang, W.; Sha, M.; Chen, M.; Ruan, F.; Li, X.; Meng, F.; Wu, Y.; Xing, M. Status and Prospects of Frequency Agile Radar Waveform Countermeasure Technology. Syst. Eng. Electron. 2021, 43, 3126–3136. [Google Scholar]
  9. Yan, F.; Su, J.; Li, Y. Research and Test of Radar Low Intercept Waveform Anti-Main Lobe Jamming Technology. Fire Control. Radar Technol. 2021, 50, 31–35. [Google Scholar] [CrossRef]
  10. Wei, S.; Zhang, L.; Liu, H. Joint frequency and PRF agility waveform optimization for high-resolution ISAR imaging. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–23. [Google Scholar] [CrossRef]
  11. Ou, J.; Li, J.; Zhang, J.; Zhan, R. Frequency Agility Radar Signal Processing; Science Press: Beijing, China, 2020. [Google Scholar]
  12. Thornton, C.; Buehrer, R.M.; Martone, A. Constrained contextual andit learning for adaptive radar waveform selection. IEEE Trans. Aerosp. Electron. Syst. 2021. [Google Scholar]
  13. Cui, G.; De, M.; Farina, A.; Li, J. Radar Waveform Design Based on Optimization Theory; The Institution of Engineering and Technology: London, UK, 2020. [Google Scholar]
  14. Xia, D.; Zhang, K.; Ding, B.; Li, B. Anti-simultaneous jamming based on phase-coded waveform shortcuts and CFAR techniques. Syst. Eng. Electron. 2022, 44, 1210–1219. [Google Scholar]
  15. Hu, H.; Lui, R.; Zhang, J. An improved radar anti-jamming method. In Proceedings of the 2018 IEEE 3rd International Conference on Signal and Image Processing, Shenzhen, China, 13–15 July 2018; pp. 439–443. [Google Scholar]
  16. Wang, H.; Yang, X.; Li, Y. Radar anti-retransmitted jamming technology based on agility waveforms. In Proceedings of the 2019 IEEE International Conference on Signal, Information and Data Processing (ICSIDP), Chongqing, China, 11–13 December 2019; pp. 1–6. [Google Scholar]
  17. Zhou, C.; Liu, F.; Liu, Q. An Adaptive Transmitting Scheme for Interrupted Sampling Repeater Jamming Suppression. Sensors 2017, 17, 2480. [Google Scholar] [CrossRef] [Green Version]
  18. Gong, X.; Meng, H.; Wei, Y.; Wang, X. Phase-Modulated Waveform Design for Extended Target Detection in the Presence of Clutter. Sensors 2011, 11, 7162–7177. [Google Scholar] [CrossRef] [Green Version]
  19. Zhang, Y.; Wei, Y. A cognitive-based approach to the design of anti-folding extended clutter waveforms. Syst. Eng. Electron. 2018, 40, 2216–2222. [Google Scholar]
  20. He, J.; Cheng, Z.; He, Z. Cognitive constant-mode waveform design method for resisting intermittent sampling and forwarding jamming. Syst. Eng. Electron. 2021, 43, 2448–2456. [Google Scholar]
  21. Hao, T.; Hu, S.; Gao, W.; Li, J.; Cao, X.; Wang, P. Detection-oriented low peak-to-average ratio robust waveform design for cognitive radar. Electron. Inf. Warf. Technol. 2022, 37, 70–75. [Google Scholar]
  22. Yu, R.; Yang, W.; Fu, Y.; Zhang, W. A Review of Cognitive Waveform Optimisation for Different Radar Tasks. Acta Electron. Sin. 2022, 50, 726–752. [Google Scholar]
  23. Wei, Y.; Meng, H.; Liu, Y.; Wang, X. Extended Target Recognition in Cognitive Radar Networks. Sensors 2010, 10, 10181–10197. [Google Scholar] [CrossRef] [Green Version]
  24. Ailiya; Wei, Y.; Yuan, Y. Reinforcement learning-based joint adaptive frequency hopping and pulse-width allocation for radar anti-jamming. In Proceedings of the 2020 IEEE Radar Conference, Florence, Italy, 21–25 September 2020; pp. 1–6. [Google Scholar]
  25. Liu, K.; Lu, X.; Xiao, L.; Xu, L. Learning based energy efficient radar power control against deceptive jamming. In Proceedings of the 2020 IEEE Global Communications Conference, Taipei, Taiwan, 7–11 December 2020; pp. 1–6. [Google Scholar]
  26. Wang, H.; Wang, F. Application of a reinforcement learning algorithm in intelligent anti-jamming of radar. Mod. Radar 2020, 42, 40–44+48. [Google Scholar]
  27. Ahmed, A.; Ahmed, A.; Fortunati, S.; Sezgin, A.; Greco, M.; Gini, F. A reinforcement learning based approach for multitarget detection in massive MIMO radar. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 2622–2636. [Google Scholar] [CrossRef]
  28. Li, K.; Jiu, B.; Liu, H.; Liang, S. Reinforcement Learning based Anti-jamming Frequency Hopping Strategies Design for Cognitive Radar. In Proceedings of the 2018 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Qingdao, China, 14–16 September 2018; pp. 1–5. [Google Scholar] [CrossRef]
  29. Li, K.; Jiu, B.; Liu, H. Deep Q-network based anti-jamming strategy design for frequency agile radar. In Proceedings of the 2019 International Radar Conference, Toulon, France, 30 September–4 October 2019; pp. 1–5. [Google Scholar]
  30. Li, K.; Jiu, B.; Wang, P.; Liu, H.; Shi, Y. Radar Active Antagonism through Deep Reinforcement Learning: A Way to Address The Challenge of Mainlobe Jamming. Signal Process. 2021, 186, 108130. [Google Scholar] [CrossRef]
  31. Ak, S.; Brüggenwirth, S. Avoiding Jammers: A Reinforcement Learning Approach. In Proceedings of the 2020 IEEE International Radar Conferenc (RADAR), Washington, DC, USA, 28–30 April 2020; pp. 321–326. [Google Scholar] [CrossRef]
  32. Selvi, E.; Buehrer, R.; Martone, A.; Sherbondy, K. On The Use of Markov Decision Processes in Cognitive Radar: An Application to Target Tracking. In Proceedings of the 2018 IEEE Radar Conference (RadarConf18), Oklahoma City, OK, USA, 23–27 April 2018; pp. 537–542. [Google Scholar] [CrossRef]
  33. Kozy, M.; Yu, J.; Buehrer, R.; Martone, A.; Sherbondy, K. Applying Deep-Q Networks to Target Tracking to Improve Cognitive Radar. In Proceedings of the 2019 IEEE Radar Conference (RadarConf), Boston, MA, USA, 22–26 April 2019; pp. 1–6. [Google Scholar] [CrossRef]
  34. Shi, Y.; Jiu, B.; Yan, J.; Liu, H.; Li, K. Data-Driven Simultaneous Multibeam Power Allocation: When Multiple Targets Tracking Meets Deep Reinforcement Learning. IEEE Syst. J. 2020, 15, 1264–1274. [Google Scholar] [CrossRef]
  35. Hessel, M.; Modayil, J.; Hasselt, H.; Schaul, T.; Ostrovski, G.; Dabney, W.; Horgan, D.; Piot, B.; Azar, M.; Silver, D. Rainbow: Combining improvements in deep reinforcement learning. In Proceedings of the 32rd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; p. 32. [Google Scholar]
  36. Watkins, C.; Dayan, P. Q-learning. Mach. Learn 1992, 8, 279–292. [Google Scholar] [CrossRef]
  37. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M. Playing Atari with Deep Reinforcement Learning. arXiv 2013, arXiv:1312.5602. Available online: https://arxiv.org/pdf/1312.5602.pdf (accessed on 6 November 2022).
  38. Bellemare, M.G.; Dabney, W.; Munos, R. A Distributional Perspective on Reinforcement Learning. In Proceedings of the 34rd International Conference on Machine Learning (ICML 2017), Sydney, Australia, 6–11 August 2017; pp. 449–458. Available online: https://arxiv.org/pdf/1707.06887.pdf (accessed on 6 November 2022).
  39. Hasselt, V.H.; Guez, A.; Silver, D. Deep Reinforcement Learning with Double Q-Learning. The 30rd AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 2094–2100. Available online: https://arxiv.org/pdf/1509.06461v3.pdf (accessed on 6 November 2022).
  40. Schaul, T.; Quan, J.; Antonoglou, I.; Silver, D. Prioritized Experience Replay. arXiv 2015, arXiv:1511.05952. Available online: https://arxiv.org/pdf/1511.05952.pdf (accessed on 6 November 2022).
  41. Wang, Z.; Schaul, T.; Hessel, M.; Hasselt, H.; Lanctot, M.; Freitas, N. Dueling Network Architectures for Deep Reinforcement Learning. In Proceedings of the 33rd International Conference on Machine Learning (ICML 2016), New York, NY, USA, 19–24 June 2016; pp. 1995–2003. Available online: https://arxiv.org/pdf/1511.06581.pdf (accessed on 6 November 2022).
  42. Fortunato, M.; Azar, M.G.; Piot, B.; Menick, J.; Osband, I.; Graves, A.; Mnih, V.; Munos, R.; Hassabis, D.; Pietquin, O.; et al. Noisy Networks for Exploration. arXiv 2017, arXiv:1706.10295. Available online: https://arxiv.org/pdf/1706.10295.pdf (accessed on 6 November 2022).
  43. Zheng, Z.; Li, W.; Zou, K.; Li, Y. Anti-jamming Waveform Design of Ground-to-air Radar Based on Reinforcement Learning. Acta Armamentarii 2022, 1–9. Available online: http://kns.cnki.net/kcms/detail/11.2176.TJ.20220722.1003.002.html (accessed on 6 November 2022).
  44. Wang, H.; Li, W.; Wang, H.; Xu, J.; Zhao, J. Radar Waveform Strategy Based on Game Theory. Radio Eng. 2019, 28, 757–764. [Google Scholar] [CrossRef]
  45. Romero, R.; Bae, J.; Goodman, N. Theory and Application of SNR and Mutual Information Matched Illumination Waveforms. IEEE Trans. Aerosp. Electron. Syst. 2011, 47, 912–927. [Google Scholar] [CrossRef]
  46. Gagniuc, P. Markov Chains: From Theory to Implementation and Experimentation; Wiley: Hoboken, NJ, USA, 2017. [Google Scholar] [CrossRef]
  47. Steven, M.K. Fundamentals of Statistical Signal Processing: Estimation and Detection Theory; Electronic Industry Press: Beijing, China, 2014; pp. 425–445. [Google Scholar]
  48. Li, X.; Fan, M. Research Advance on Cognitive Radar and Its Key Technology. Acta Electron. Sin. 2012, 40, 1863–1870. Available online: https://www.ejournal.org.cn/EN/10.3969/j.issn.0372-2112.2012.09.025 (accessed on 6 November 2022).
  49. Jackson, L.; Kay, S.; Vankayalapati, N. Iterative Method for Nonlinear FM Synthesis of Radar Signals. IEEE Trans. Aerosp. Electron. Syst. 2010, 46, 910–917. [Google Scholar] [CrossRef]
Figure 1. DRL algorithm landing process.
Figure 1. DRL algorithm landing process.
Sensors 22 08689 g001
Figure 2. The airborne radar detection scene.
Figure 2. The airborne radar detection scene.
Sensors 22 08689 g002
Figure 3. Airborne radar signal model.
Figure 3. Airborne radar signal model.
Sensors 22 08689 g003
Figure 4. Reinforcement learning model.
Figure 4. Reinforcement learning model.
Sensors 22 08689 g004
Figure 5. The radar countermeasure process based on markov decision process (MDP).
Figure 5. The radar countermeasure process based on markov decision process (MDP).
Sensors 22 08689 g005
Figure 6. Block diagram of the radar intelligent anti-jamming based on the duelling double deep Q network (D3QN).
Figure 6. Block diagram of the radar intelligent anti-jamming based on the duelling double deep Q network (D3QN).
Sensors 22 08689 g006
Figure 7. Schematic diagram of the empirical replay process.
Figure 7. Schematic diagram of the empirical replay process.
Sensors 22 08689 g007
Figure 8. Paired network structure.
Figure 8. Paired network structure.
Sensors 22 08689 g008
Figure 9. Environmental clutter and target impulse response.
Figure 9. Environmental clutter and target impulse response.
Sensors 22 08689 g009
Figure 10. Three radar-transmitted signals under jamming conditions. (a) linear frequency modulation (LFM) signal; (b) anti-jamming signal generated based on policy iteration method; (c) radar optima anti-jamming signal generated based on D3QN.
Figure 10. Three radar-transmitted signals under jamming conditions. (a) linear frequency modulation (LFM) signal; (b) anti-jamming signal generated based on policy iteration method; (c) radar optima anti-jamming signal generated based on D3QN.
Sensors 22 08689 g010
Figure 11. Performance of different signals when radar power changes. (a) the signal-to-jamming plus noise ratio (SJNR) of different transmitted signals; (b) detection probability of different transmitted signals.
Figure 11. Performance of different signals when radar power changes. (a) the signal-to-jamming plus noise ratio (SJNR) of different transmitted signals; (b) detection probability of different transmitted signals.
Sensors 22 08689 g011
Figure 12. The performance of different signals when the jamming power changes. (a) SJNR of different transmitted signals; (b) detection probability of different transmitted signals.
Figure 12. The performance of different signals when the jamming power changes. (a) SJNR of different transmitted signals; (b) detection probability of different transmitted signals.
Sensors 22 08689 g012
Figure 13. SJNR of the three signals under the conditions of jamming power and radar power variation. (a) Variation in LFM signal SJNR; (b) SJNR change graph of the anti-jamming signal generated based on strategy iteration method; (c) variation diagram of optimal anti-jamming signal SJNR generated based on D3QN algorithm.
Figure 13. SJNR of the three signals under the conditions of jamming power and radar power variation. (a) Variation in LFM signal SJNR; (b) SJNR change graph of the anti-jamming signal generated based on strategy iteration method; (c) variation diagram of optimal anti-jamming signal SJNR generated based on D3QN algorithm.
Sensors 22 08689 g013
Figure 14. Time domain waveforms of the optimal radar anti-jamming strategy. (a) Magnitude and phase of the synthetic radar signal; (b) spectrum fitting results of the radar signal.
Figure 14. Time domain waveforms of the optimal radar anti-jamming strategy. (a) Magnitude and phase of the synthetic radar signal; (b) spectrum fitting results of the radar signal.
Sensors 22 08689 g014
Table 1. Comparison of methods and results of research in related work.
Table 1. Comparison of methods and results of research in related work.
Classification by DRL AlgorithmDRL AlgorithmBrief DescriptionReferences
Value-basedQ learning
(1992.05)
Q learning uses Q Table to store the value of each pair of state actions. When the state and action space are high-dimensional or continuous, it is not realistic to use Q Table.Watkins et al. [36]
Deep Q Network (DQN,2013.12)DQN can directly use raw high-dimensional sensor data for reinforcement learning, and training using stochastic gradient descent and the training process is more stable.Mnih et al. [37]
Distributional DQN
(2015.08)
In DQN, the expected estimate of the state-action value Q of the network output ignores a lot of information. A deep reinforcement learning model modelled from a distributed perspective can obtain more useful information and obtain better and more stable results.Bellemare et al. [38]
Double DQN
(2015.12)
In the original DQN, there will be a situation where the Q value is overestimated. To decouple the action selection and the value estimate, when calculating the actual value of Q in double DQN, the action selection is obtained by the evaluation network, while the value estimate is obtained by the target network.Hasselt et al. [39]
Prioritized Experience Replay
(2016.02)
In the experience pool of traditional DQNs, the data selected for batch training are random, without taking into account the priority relationship of the samples. This method gives each sample a priority and samples according to the priority of the sample.Schaul et al. [40]
Duelling DQN (2016.04)In the original DQN, the neural network directly outputs the Q value of each action, while the Q value of each action of the duelling DQN is determined by the state value V and the dominance function A, that is, Q = V + A.Wang et al. [41]
Noisy DQN
(2017.06)
This method increases the exploration ability of the model by adding noise to the parameters.Fortunato et al. [42]
Table 2. MDP model parameter settings.
Table 2. MDP model parameter settings.
Markov Decision ModelParameter Settings
Number of signal sub-band divisions M = 5
Number of sub-band energy divisions N = 5
State space S = ( s 1 , s 2 , s 3 , , s 6 5 )
Action space A = ( a 1 , a 2 , a 3 , , a 6 5 )
Single state s t = [ α 1 , α 2 , α 3 , α 4 , α 5 ] , α [ 0 , 5 ]
Single action a t = [ β 1 , β 2 , β 3 , β 4 , β 5 ] , β [ 0 , 5 ]
γ 0.9
Table 3. Simulation parameters.
Table 3. Simulation parameters.
A Certain Foreign RadarSimulation Parameter Settings
Working bandX-band
Centre frequency9.5 GHz
Signal bandwidth100 MHz
Sub-band bandwidth20 MHz
Target aircraft speed250 m/s
σ h 2 ( f ) 25 ,   4 ,   16 ,   64 ,   49
Clutter 0.8 ,   0.2 ,   1.0 ,   3.0 ,   4.5
Noise1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zheng, Z.; Li, W.; Zou, K. Airborne Radar Anti-Jamming Waveform Design Based on Deep Reinforcement Learning. Sensors 2022, 22, 8689. https://doi.org/10.3390/s22228689

AMA Style

Zheng Z, Li W, Zou K. Airborne Radar Anti-Jamming Waveform Design Based on Deep Reinforcement Learning. Sensors. 2022; 22(22):8689. https://doi.org/10.3390/s22228689

Chicago/Turabian Style

Zheng, Zexin, Wei Li, and Kun Zou. 2022. "Airborne Radar Anti-Jamming Waveform Design Based on Deep Reinforcement Learning" Sensors 22, no. 22: 8689. https://doi.org/10.3390/s22228689

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop