Next Article in Journal
Analysis of Bubble Flow in an Inclined Tube and Modeling of Flow Prediction
Previous Article in Journal
Time Optimal Altitude-Hold Flight Mode Transition Strategy for a Class of Ducted Fan Tail Sitter UAV
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Intelligent Bait Delivery Control Method for Flight Vehicle Evasion Based on Reinforcement Learning

by
Shuai Xue
1,
Zhaolei Wang
2,
Hongyang Bai
1,*,
Chunmei Yu
2,
Tianyu Deng
1 and
Ruisheng Sun
1
1
School of Energy and Power Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
2
National Key Laboratory of Science and Technology on Aerospace Intelligence Control, Beijing Aerospace Automatic Control Institute, Beijing 100854, China
*
Author to whom correspondence should be addressed.
Aerospace 2024, 11(8), 653; https://doi.org/10.3390/aerospace11080653
Submission received: 13 June 2024 / Revised: 5 August 2024 / Accepted: 7 August 2024 / Published: 11 August 2024
(This article belongs to the Section Aeronautics)

Abstract

:
During aerial combat, when an aircraft is facing an infrared air-to-air missile strike, infrared baiting technology is an important means of penetration, and the strategy of effective delivery of infrared bait is critical. To address this issue, this study proposes an improved deep deterministic policy gradient (DDPG) algorithm-based intelligent bait-dropping control method. Firstly, by modeling the relative motion between aircraft, bait, and incoming missiles, the Markov decision process of aircraft-bait-missile infrared effect was constructed with visual distance and line of sight angle as states. Then, the DDPG algorithm was improved by means of pre-training and classification sampling. Significantly, the infrared bait-dropping decision network was trained through interaction with the environment and iterative learning, which led to the development of the bait-dropping strategy. Finally, the corresponding environment was transferred to the Nvidia Jetson TX2 embedded platform for comparative testing. The simulation results showed that the convergence speed of this method was 46.3% faster than the traditional DDPG algorithm. More importantly, it was able to generate an effective bait-throwing strategy, enabling the aircraft to successfully evade the attack of the incoming missile. The strategy instruction generation time is only about 2.5 ms, giving it the ability to make online decisions.

1. Introduction

With the rapid development of modern military technology, infrared-guided missiles have become an important weapon in modern times. According to recent studies on aircraft damaged in wars abroad, approximately 90% of the aircraft were damaged by infrared-guided-missiles [1,2]. Therefore, studying infrared defense strategies is crucial for air combat.
Infrared baits are the most commonly used interference method against infrared-guided missiles. The deployment of infrared baits can interfere with the normal tracking of infrared-guided missiles and improve the survival rate of our aircraft. However, the current deployment of baits requires pilots to manually bind and control the bait deployment device based on the air condition, thus relying completely on the pilot’s experience. Improper deployment not only wastes valuable infrared bait resources but may also seriously endanger the survival and safety of the aircraft and pilots [3]. Therefore, making reasonable decisions in a battlefield environment, such as determining the launch speed, number of launches, interval between launches, and distance of infrared baits, has become a challenge that must be studied. Shen et al. [4] modeled and simulated the interference of infrared surface-source baits but did not analyze the interference strategy of infrared baits. Yang et al. [5] examined the infrared radiation characteristics of targets, infrared baits, and backgrounds and proposed a method for recognizing infrared baits; however, they did not study the interference characteristics of infrared baits. Chen et al. [6] provide a reference for the deployment strategy of infrared baits by studying the infrared radiation characteristics of an aircraft; however, they only obtained the deployment and use strategies of infrared baits through theoretical analyses and did not conduct simulation verifications. Zhang et al. [7] examined adversarial strategies from two perspectives: barrel roll maneuvers and unpowered point source bait. However, because they mainly focused on the target’s maneuver method and the bait’s deployment method, they were unable to generate the infrared bait projectile’s deployment plan through autonomous decision-making. In summary, most of the current research only focuses on the theoretical analysis of infrared modeling and scattering strategies for baits.
With the rapid development of artificial intelligence technology, countries are committed to researching more intelligent weapon systems to replace existing designs [8]. In recent years, deep reinforcement learning has become a research hotspot, attracting much attention from scholars [9]. Deep reinforcement learning technology is considered one of the most likely ways to achieve universal artificial intelligence computing, with strong universality. The deep reinforcement learning algorithm does not rely on accurate modeling and instead uses model-free methods to control the unknown dynamic environment [10], allowing it to handle the dynamics of unstable systems [11]. Shen et al. [12] proposed a data-driven method that uses Markov parameter sequences to replace the dynamics modeling of complex tension systems. At present, there is extensive and in-depth research on the use of deep reinforcement learning in fields such as guidance, control, and path planning. For example, Yang et al. [13] studied the impact time control guidance law of missiles under time-varying velocities caused by gravity and air resistance and proposed a collision time control guidance algorithm based on TD3. Aslan et al. [14] used deep reinforcement learning algorithms to solve the balance control problem of robots under external forces. Similarly, Lee et al. [15] used reinforcement learning to enable unmanned aerial vehicles to perform real-time autonomous path planning in unknown environments. Fan et al. [16] applied the DDPG algorithm to missile evasion decision training and simulated the effectiveness of escape strategies under four typical initial situations. Furthermore, Deng et al. [17] proposed a missile terminal guidance law based on an improved depth deterministic strategy gradient algorithm; however, it allowed an attacker to intercept maneuvering targets rather than scattering infrared baits, which would allow aircraft to successfully escape. Finally, Qiu et al. [18] proposed a recorded recurrent twin-delayed deep deterministic (RRTD3) strategy gradient algorithm for intercepting maneuvering missiles in the atmosphere, but this work did not study targets carrying baits.
In this study, an intelligent control method for aircraft penetration bait delivery based on an improved DDPG algorithm is proposed. Also, a dynamic model of incoming missiles, infrared bait, and aircraft is constructed in the air, with the state space, action space, and reward function being designed during the process. Meanwhile, the trained decision network is used to output the deployment strategy of baits. Significantly, to address the issue of the slow convergence speed of the algorithm, improvements were made through pre-training and classification sampling mechanisms. Importantly, the algorithm was transferred to embedded platforms for comparative testing to verify the online capability of the proposed method.
This paper is organized as follows: Section 2 describes the modeling of bait delivery strategies. Section 3 introduces the relationship between modules in aircraft penetration systems and the design of the Markov decision process. In Section 4, the deep reinforcement learning algorithm is introduced, and the algorithm is improved by pre-training and sampling mechanisms. Section 5 provides the algorithm training results and simulation verification examples of bait delivery in different scenarios. Section 6 summarizes the conclusions of this study.

2. Relative Motion Model of Aircraft-Bait-Incoming Missile

To simplify matters, assume that the incoming missile, bait, and aircraft are in a vacuum and within a uniform gravitational field. We only consider the motion of the center of mass in three dimensions. In addition, assume that the incoming missile and bait are rigid bodies with constant mass. The space coordinate systems [19] are defined as follows:
Inertial coordinate system S g ( O g X g Y g Z g ): The origin O g of the inertial coordinate system is taken at a point in the inertial space. The O g X g axis and O g Z g axis are in the horizontal plane, with the O g X g axis pointing due north and the O g Z g axis pointing due east. The O g Y g axis satisfies the right-hand rule of vertical upward.
Ballistic coordinate system St(OtXtYtZt): The origin Ot is the instantaneous center of mass of the missile. The OtXt axis coincides with the missile velocity vector. The OtYt axis is located in the plumb plane containing the velocity vector perpendicular to the OtXt axis. The OtYt axis points upward and is positive. The OtZt axis and OtXt, OtYt axis form the right-hand coordinate system.
Based on the above conditions, the dynamic models [20] were constructed as shown below.

2.1. Motion Modeling of Aircraft, Incoming Missile, and Bait

The dynamic equation of the aircraft is as follows:
x ˙ m = v x m y ˙ m = v y m z ˙ m = v z m v ˙ x m = P x m G M x m r m 3 v ˙ y m = P y m G M y m + R r m 3 v ˙ z m = P z m G M z m r m 3 m ˙ = | P | I s p r m = x m 2 + y m + R 2 + z m 2
where x m , y m , and z m are the position vectors of the aircraft; v x m , v y m , and v z m are the speed vectors of the aircraft; P x , P y , and P z are the thrust vectors generated by the aircraft’s maneuvering device; m is the mass of the aircraft; M is the mass of the earth; G is the constant of universal gravitation; R is the radius of the earth; r m is the distance of the aircraft relative to the origin; and I s p stands for fuel specific impulse.
The dynamic equation of the incoming missile is:
x ˙ t = v t cos θ t cos ϕ t y ˙ t = v t sin θ t z ˙ t = v t cos θ t sin ϕ t v ˙ t = g n t x sin θ t θ ˙ t = g v t n t y cos θ t ϕ ˙ t = g n t z / v t cos θ t
where x t , y t , z t represent the coordinates of the incoming missile in the inertial coordinate system; v t , θ t , and ϕ t respectively, represent the velocity, trajectory pitch angle, and trajectory deviation angle of the incoming missile; n t x represents the longitudinal control overload of the incoming missile; and n t y and n t z respectively, represent the turning control overload of the incoming missile in the yaw and pitch directions.
Assuming that the infrared bait is released, ignoring the influence of air resistance, and only maintains uniform acceleration motion under the action of gravity, the dynamic equation of the bait is written as:
x ˙ d = v d cos θ d cos ϕ d y ˙ d = v d sin θ d z ˙ d = v d cos θ d sin ϕ d v ˙ d = g sin θ d θ ˙ d = g v d cos θ d ϕ ˙ d = 0
where x d , y d , z d represent the coordinates of the bait in the inertial coordinate system, and v d , θ d , and ϕ d represent the velocity, trajectory pitch angle, and trajectory deflection angle of the bait, respectively.

2.2. Relative Motion Model between Incoming Missile and Target

Assuming the position vector of the target relative to the incoming missile is r and the velocity vector is V , it can be represented by ( r , V , q ε , q β ) in the inertial coordinate system as follows:
r = r x 2 + r y 2 + r z 2 V = V x 2 + V y 2 + V z 2 q β = arctan r z / r x q ε = arcsin r y / r x 2 + r z 2
where r x , r y , and r z are the components of the position of the target relative to the incoming missile in three directions, and r x = x t x m , r y = y t y m , and r z = z t z m ; V x , V y , and V z are the components of the velocity of the target relative to the incoming missile in three directions, and V x = v x _ t v x _ m , V y = v y _ t v y _ m , and V z = v z _ t v z _ m ; q β is the line of sight azimuth angle; and q ε is the height angle of the line of sight.
Taking the derivative of Equation (4) with respect to time yields:
r ˙ = r x v x + r y v y + r z v z r q ˙ β = v x r z r x v z r x 2 + r z 2 q ˙ ε = v y r x 2 + r z 2 r y r x v x + r z v z r 2 r x 2 + r z 2
where r ˙ represents the relative position change rate and q ˙ β and q ˙ ε represent the line of sight azimuth angle change rate and line of sight elevation angle change rate, respectively.
The infrared guided missile in this article adopts a proportional guidance rate, and its guidance instructions are as follows:
n t y = k y r ˙ q ˙ ε n t z = k z r ˙ q ˙ β
where k y and k z are the proportional guidance coefficients, respectively.

3. Design of Adversarial Defense Framework for Incoming Missile

3.1. Overall Framework Design

A defensive framework for incoming missiles was constructed to study the deployment strategy of using bait for interference defense in confrontation scenarios between our aircraft and incoming missiles. The entire framework consists of three modules: the bait autonomous deployment module (based on deep reinforcement learning), the target and scene real-time dynamic generation module, and the target detection and recognition module. The modules are connected and communicate via Ethernet. The general relationship between each module is shown in Figure 1.
The bait autonomous delivery module (based on deep reinforcement learning) sends the position, velocity, and angle information of incoming missiles, aircraft, and baits to the target and scene real-time dynamic generation module through a simulation. Based on that information, the module generates real-time infrared images and sends them to the target detection and recognition module. This module then sends the detection probability through the target detection and recognition module to the bait autonomous deployment module (based on deep reinforcement learning), where it serves as the input to simulate the recognition probability of the incoming missile seeker to the target.

3.2. Bait Delivery Model Based on Deep Reinforcement Learning

The Markov decision process (MDP) [21] model of autonomous delivery of baits can be described by a quintuple s , a , p , r , γ , where: s is the state space; a is the action space; p is the state transition probability; r is the reward function; and γ is a discount factor [22].

3.2.1. State Space and Action Space Design

When developing the autonomous launch of MDP by bait, the state space should be able to consider all states of the confrontation process. The state of the environment should be able to describe the dynamic characteristics of the environment, accurately simulate real-world conditions, and facilitate observation. Here we use the relative distance between the incoming missile and the aircraft, the elevation angle, and the azimuth angle of the line of sight to represent the state space of the environment as follows:
s = r , q ε , q β
where r is the relative distance between the incoming missile and the aircraft, q ε is the elevation angle of the line of sight, and q β is the azimuth angle of the line of sight.
The action space is:
a = v d x , v d y , v d z
where v d x , v d y , v d z represent the partial velocities of the bait projectile in the x , y , z direction.

3.2.2. Probability of State Transition

The state transition probability represents the probability of moving to the state set s t + 1 of time t + 1 when the incoming missile, aircraft, and infrared bait are in the state set s t and action set a t of the current time t , then the deterministic state transition probability of establishing the counter defense MDP of the incoming missile would be as follows:
p ( s t + 1 s t , a t ) = 1

3.2.3. Reward Function Design

The reasonable setting of a reward function is key to ensuring the convergence of the algorithm. To design the reward function we combined the ideas of sparse and process. The specific reward function design is as follows:
(1) Reward for successfully evading incoming missile attacks
The main purpose of the bait dropping strategy of our aircraft when it encounters the attack of the incoming infrared guided air-to-air missile is to successfully evade the attack of the incoming missile while deceiving the seeker. Therefore, its reward function was designed as follows:
r 1 = 100 ,    Successfully evade missile 0 ,    Failed to evade the missile
where the conditions for successful evasion are the distance between the incoming missile and the target: r 50   m .
(2) Reward the number of infrared baits consumed
The number of baits carried by the aircraft is limited when carrying out tasks. In order to improve the continuous defense capability of our aircraft against infrared guided missiles in subsequent operations, the number of infrared baits should be reduced as much as possible during the delivery process on the premise of ensuring the safety of our aircraft. Then the design reward function is presented as follows:
r 2 = 2 × ( n 1 n 2 )
where n 1 is the initial number of infrared baits, n 2 is the number of infrared baits deployed.
(3) Sight Angle reward
In the process of infrared bait interference on incoming missile, the greater the line of sight angle, the greater the probability our aircraft successfully can avoid an incoming missile attack, thus the design reward function is:
r 3 = 100 e q ε + q β 180
(4) Position reward of bait shot
In order to ensure that the aircraft can successfully evade an incoming missile under the action of the bait, a reward is given for its relative position during the evasion process (i.e., the minimum distance between the incoming missile and the aircraft). When the distance between the incoming missile and the aircraft is greater than 50 m, it indicates that the bait successfully caused the aircraft to escape, and a positive reward is provided. When the distance between the incoming missile and the aircraft is less than 50 m, it indicates that the aircraft failed to escape, and a negative reward is provided as punishment. Therefore, the reward function was designed as follows:
r 4 = 0.01 r m ,    Successfully evade missile 10 ,    Failed to evade the missile
where r m is the target distance at the time when the bait is deployed.
Considering the above four reward models, the total reward function was designed as follows:
r t o t a l = r 1 + r 2 + r 3 + r 4

3.3. Identification Model of Incoming Missile Seeker

The infrared seeker is the guided weapon’s direction control, which can automatically track its target through the continuous identification, detection, and tracking of the target’s radiation energy by the infrared detector. The infrared seeker, like other types of seekers, not only has the ability to continuously detect infrared radiation from the target, identify, detect, and track the target, but it can also eliminate the deviation angle distortion of the incoming missile, as well as automatically track the locked target and output information in real time [23].
In order to make the seeker enter the loop, a simulated incoming flight vehicle seeker model was designed. The relative motion model of incoming flight vehicle-bait-aircraft provides their respective particle motion trajectory. with the help of infrared target simulator, the infrared target characteristics of bait and aircraft are simulated. The tracking process of attack incoming flight vehicle seeker to infrared bait and aircraft is simulated.
The infrared target simulator in this study uses real-time dynamic generation of targets and scenes obtained from an image search in reference [24] and is able to generate real-time target infrared images using mathematical formulas. In addition, the target detection and recognition method uses the Yolov5 target detection algorithm to calculate the real-time recognition probability by detecting the target’s infrared image.

4. Deep Reinforcement Learning Algorithms

Deep reinforcement learning [25] is a decision algorithm based on deep learning models. Deep learning focuses on perception, while reinforcement learning differs from deep learning in that it focuses on expression. Deep reinforcement learning combines the perceptual ability of deep learning with the sequential decision-making ability of reinforcement learning, enabling intelligent agents to solve decision-making problems in complex state spaces with strong universality.

4.1. DDPG Algorithm

DDPG [26] is an important algorithm for deep reinforcement learning applied in the field of continuous control reinforcement learning.
A task-independent model is proposed by combining the deterministic strategy gradient algorithm with the actor-critic framework, which can solve numerous continuous control problems with different tasks using the same parameters. Figure 2 shows the basic structure diagram of the actor critic section of the DDPG algorithm.
The DDPG algorithm includes the current network and target network in both actor and critic. The critic value network estimates Q value and updates the current network parameter ω of the critic by minimizing loss function Loss , which is specifically expressed as follows:
Loss = 1 m i r i + γ Q s i + 1 , a i + 1 ω Q s i , a i ω 2
where m is the sampling number of empirical data; r i represents the reward at i time; γ is the reward discount factor; Q s i + 1 , a i + 1 ω represents the target network Q value of environment state s i + 1 at time i + 1 , and Q s i , a i ω represents the current network Q value obtained by inputting current state s i and current action a i at time i .
The network of actor strategy adopts the mode of deterministic strategy gradient, uses the gradient descent method to update, and then outputs a definite action a = μ s i | θ . The current network parameter θ of actor is updated as follows:
θ J ( θ ) = 1 m i a i Q s i , a i ω θ μ s i | θ
where θ μ s i | θ is the gradient of performance indicator J ( θ ) to actor online network parameter θ .
During the training process, parameters θ and ω of the target network of actor and critic are soft-updated through their current network at regular intervals. The specific updating method is as follows:
θ τ θ + ( 1 τ ) θ ω τ ω + ( 1 τ ) ω
where τ is used to control the update speed of target network parameters θ and ω .

4.2. Improved DDPG Algorithm

The traditional DDPG algorithm has strong randomness in the initial stages of policy exploration, and the gradient of the policy network declines slowly. Therefore, this study provides the initial network weight parameters through pre-training based on its random action selection, which decreases invalid searches during the early exploration stage and increases the algorithm’s speed of convergence. Specifically, for pre-training, the initial planning strategy is obtained by simply pre-training the agent using the reward function of the aircraft and successfully avoiding the incoming flight vehicle attack. Its network weight is then saved as the initial value of training. In this case, only the missed distance is taken into consideration.
In addition, the training effect is diminished by the fact that the traditional DDPG algorithm updates network parameters by random sampling during experience playback, and the samples extracted may have a large number of useless sample data. To solve this issue, this study classifies the samples in the experience pool during the training process and stores the aircraft data that can successfully evade incoming missiles in experience pool I. The data on aircraft hits by incoming missiles is stored in experience pool II. During the training process, the goal is that experience data of higher value can be sampled from experience pool I as much as possible. The method of proportional sampling from the two experience pools was designed with these factors in mind and is shown in Equation (18). In order to ensure randomness in the exploration and avoid falling into local optimality, random sampling is still adopted in the respective sampling of experience pools I and II.
N 1 = n N 2 N 2 = λ n
where, N 1 and N 2 are the sampling quantities from experience pools I and II, respectively. n is the total sampling quantity, λ is the sampling rate from experience pool II, and λ ( 0 , 1 ) .
The Figure 3 shows the structure of the infrared bait autonomous delivery framework designed by the improved DDPG algorithm in this paper.
The process of designing the algorithm according to the structure diagram in Figure 3 is shown in Table 1.

5. Simulation Verification and Analysis

5.1. Simulation Scenario Setting and Platform Construction

In the simulation scenario, the position, initial speed, and initial angle of the aircraft and incoming missile are mainly set, as shown in Table 2. Meanwhile, the mass of the aircraft is set to 600 kg, and the available thrust is set to 20,000 N. The mass of the incoming missile is 20 kg, the maximum usable overload is 4 g, and the minimum miss distance is 50 m. The incoming missile adopts the proportional guidance method to attack the aircraft, and the proportional guidance coefficient is 3. The four-order Runge-Kutta method was used in the integration process. The integration step was 0.005, and the longest integration time was 20 s.
Figure 4 shows the physical simulation system developed for this study. It shows the simulation interface diagram of the three modules in the simulation platform for data transmission through the Ethernet built for this study.

5.2. Setting Simulation Parameters

In this paper, the algorithm is programmed using Python, the agent environment is built by Gym, and the deep reinforcement learning training environment is built based on PyTorch. The test host is based on the Windows10 operating system, and the hardware configuration is an Intel Core i7 9700K, RTX 3080Ti, and 16 GB of RAM. The algorithm hyperparameter settings are shown in Table 3. In addition, the environment was ported to the Nvidia Jetson TX2 embedded development board, and the test was carried out on the embedded platform with other parameters being the same. The detailed parameters of the Nvidia Jetson TX2 are shown in Table 4.

5.3. Analysis of Simulation Results

In this paper, we use the traditional DDPG algorithm to train the autonomous delivery strategy network of infrared bait. Then, the DDPG algorithm is improved and trained on the basis of pre-training and classification sampling. Finally, the reward function convergence of the two algorithms is compared. The average reward of 100 events was selected as the result of multiple training sessions according to random initial conditions, as shown in Figure 5.
As can be seen in Figure 5, under the same simulation environment, the convergence steps of the traditional DDPG algorithm and the improved DDPG algorithm are 177 and 95, respectively, and the maximum rewards are 526 and 477, respectively. Thus, the convergence speed of the improved DDPG algorithm is 46.3% faster, and the maximum rewards can be explored. Therefore, the improved DDPG algorithm developed for this study enhances the exploration ability of the agent, and the convergence of the algorithm is significantly improved. In addition, the influence of Gaussian noise was added to the training process of the improved DDPG algorithm, and the training result is shown in Figure 5 as a blue curve. Figure 5 also shows that the algorithm is very robust; it can also reach the convergence state in the end, and the curve has little fluctuation compared with the conditions before the noise interference was introduced.
Following the training phase of the improved DDPG algorithm, the parameters of the strategy network were fixed. The interaction with the environment of the incoming missile, the aircraft, and the infrared bait was simulated and verified when the initial position and velocity parameters of the incoming missile and the aircraft differed by 5%. In addition, in order to verify the real-time performance and online capability of the algorithm on the embedded platform, the simulation environment built for this study was transplanted to the Nvidia Jetson TX2 embedded platform.
Figure 6 shows the three-dimensional motion trajectory diagram of the incoming missile when the aircraft only maneuvers without adopting the autonomous delivery strategy of infrared bait in this paper. Figure 6a shows the flight curve when the aircraft and the incoming missile are in the same direction (i.e., simulation situation I). Figure 6b shows the flight curve when the aircraft and the incoming missile are in relative motion (i.e., simulation situation II). Figure 6c shows the motion curve of the incoming missile attacking the aircraft from a high point at a certain height relative to the aircraft (i.e., simulation situation III).
Figure 7 shows the change in the relative distance between the incoming missile and the aircraft. Among them, Figure 7a shows the relative distance change curve during the simulation situation I. Figure 7b shows the relative distance change curve in simulation situation II. Figure 7c shows the relative distance change curve between the incoming missile and the aircraft in simulation situation III. In all three situations, the aircraft failed to evade the attack of the incoming missile.
Figure 8 shows the three-dimensional motion trajectory diagram of the infrared bait deployment by the delivery strategy network while the aircraft is maneuvering during an encounter with the incoming missile attack. Figure 8a shows the flight curve of the bait when the aircraft and the incoming missile are in the same direction (i.e., simulation situation I). Figure 8b shows the flight curve of the bait when the aircraft and the incoming missile are in relative motion (i.e., simulation situation II). Figure 8c shows the flight curve of the bait when the incoming missile is approaching from a high point over the aircraft (i.e., simulated situation III).
Figure 9 shows the variation curve of the relative distance between the incoming missile and the maneuvering aircraft under the interference of the bait. Figure 9a shows the relative distance change curve in simulation situation I. Figure 9b shows the relative distance change curve in simulation situation II. Figure 9c shows the relative distance change curve in simulation situation Ⅲ. In all three situations, the incoming missile is lured by the infrared baits, causing it to deviate from the target, so that the aircraft can successfully evade the attack and escape.
According to the curves in Figure 6, Figure 7, Figure 8 and Figure 9, the simulation results and calculation speed in the above cases were compared and as shown in Table 5.
Each simulation situation was repeated 2000 times. As Table 5 shows, in the case of flight situation I, when the aircraft uses the bait autonomous delivery strategy network, dropping five point-source infrared baits, the miss distance of the incoming missile is 670.2 m, and the successful escape rate of the aircraft was 96.8%. Conversely, when the bait strategy is not implemented, the miss distance is 0.5 m, and the successful escape rate is 0.3%. Similarly, in flight situation II, when the bait autonomous delivery strategy network is used and the aircraft adopts the strategy to deploy four point-source infrared baits, the miss distance of the incoming missile is 1000.0 m, and the successful escape rate of the aircraft is 97.2%. When the bait strategy is not used, the miss distance is 0.2 m, and the successful escape rate is 0.24%. In the case of flight situation Ⅲ, when the bait autonomous delivery strategy network is used and the aircraft deploys four point-source infrared baits, the average miss distance of the incoming missile is 750.5 m, and the successful escape rate of the aircraft is 96.0%. In contrast, when the bait strategy is not implemented, the average miss distance of the aircraft is 0.6 m, and the successful escape rate of the aircraft is 0.4%. Therefore, the bait autonomous delivery strategy in this study can improve the success rate of aircraft in evading an incoming missile. In addition, during the decision output, the time needed to generate control instructions is 2.5 ms, while the test time of the decision instruction output on the Nvidia Jetson TX2 embedded platform is 12.5 ms, which meets the real-time requirements. These results verify that the autonomous delivery method developed in this study can be deployed on the embedded platform and has online decision-making capability.

6. Conclusions

With the aim of assessing the success rate of infrared bait launching methodologies, this paper designed an intelligent control method for autonomous bait launching based on an improved DDPG algorithm. The proposed algorithm applies deep reinforcement learning to air operations to provide continuous deterministic action decisions for infrared bait. In addition, to further optimize the algorithm, this study also developed pre-training and classification sampling. The simulation results show that the intelligent control method can highly improve the convergence of the traditional DDPG algorithm, and the convergence speed is increased by about 46.3%. The decision network can also output an effective bait-spray strategy. Meanwhile, the speed of generating the decision control order is fast, about 2.5 ms, which can meet real-time requirements. Significantly, the control command output on the embedded platform can meet the required speed with online control ability and achieve rapid autonomous decision-making. Our results show that the probability of a successful escape is about 97%. The method proposed in this work highly improves the autonomous survivability of aircraft in the face of infrared-guided missile attacks.

Author Contributions

Conceptualization, S.X.; methodology, S.X. and H.B.; software, S.X., Z.W. and T.D.; validation, Z.W. and C.Y.; formal analysis, R.S.; investigation, H.B.; resources, H.B.; data curation, H.B.; writing—original draft preparation, S.X.; writing—review and editing, S.X. and Z.W.; visualization, S.X.; supervision, H.B.; project administration, S.X. and H.B.; funding acquisition, H.B. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China through Grant No. U21B2028.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Acknowledgments

Shuai Xue thanks Hongyang Bai for their helpful guidance.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nowak, J.; Achimowicz, J.; Ogonowski, K.; Biernacki, R. Protection of air transport against acts of unlawful interference: What’s next. Saf. Def. 2020, 2, 75–88. [Google Scholar]
  2. Wang, R.F.; Wu, W.D.; Zhang, Y.P. Summarizotion of defense measures for the IR guided flight vehicle. Laser Infrared 2006, 12, 1103–1105. [Google Scholar]
  3. Hu, Z.H.; Chen, K.; Yan, J. Research on control parameters of infrared bait delivery device. Infrared Laser Eng. 2008, 37, 396–399. [Google Scholar] [CrossRef]
  4. Shen, T.; Li, F.; Song, M.M. Modeling and simulation of infrared surface source interference based on SE-Work-Bench. Aerosp. Electron. Work. 2019, 39, 6–10. [Google Scholar]
  5. Yang, S.; Wang, B.; Yi, X.; Yu, H.; Li, J.; Zhou, H. Infrared baits recognition method based on dual-band information fusion. Infrared Phys. Technol. 2014, 67, 542–546. [Google Scholar] [CrossRef]
  6. Chen, S.H.; Zhu, N.Y.; Chen, N.; Ma, X.J.; Liu, J. Infrared radiation characteristics test of aircraft and research on infrared bait delivery. Infrared Technol. 2021, 43, 949–953. [Google Scholar]
  7. Zhang, N.; Chen, C.S.; Sun, J.G.; Liang, X.C. Research on infrared air-to-air flight vehicle based on barrel roll maneuver and bait projection. Infrared Technol. 2022, 44, 236–248. [Google Scholar]
  8. Huang, S.C.; Li, W.M.; Li, W. Two-sided optimal decision model used for ballistic flight vehicle attack-defense. J. Airf. Eng. Univ. Nat. Sci. Ed. 2007, 8, 23–25. [Google Scholar]
  9. Sutton, R.S.; Barto, A.G. Reinforcement learning: An introduction. IEEE Trans. Neural Netw. 1998, 9, 1054. [Google Scholar] [CrossRef]
  10. Shen, Y.; Chen, M.; Skelton, R.E. Markov data-based reference tracking control to tensegrity morphing airfoils. Eng. Struct. 2023, 291, 116430. [Google Scholar] [CrossRef]
  11. Shen, Y.; Chen, M.; Majji, M.; Skelton, R.E. Q-Markov covariance equivalent realizations for unstable and marginally stable systems. Mech. Syst. Signal Process. 2023, 196, 110343. [Google Scholar] [CrossRef]
  12. Shen, Y.; Chen, M.; Skelton, R.E. A Markov data-based approach to system identification and output error covariance analysis for tensegrity structures. Nonlinear Dyn. 2024, 112, 7215–7231. [Google Scholar] [CrossRef]
  13. Yang, Z.; Liu, X.; Liu, H. Impact time control guidance law with time-varying velocity based on deep reinforcement learning. Aerosp. Sci. Technol. 2023, 142, 108603. [Google Scholar] [CrossRef]
  14. Aslan, E.; Arserim, M.A.; Uçar, A. Development of push-recovery control system for humanoid robots using deep reinforcement learning. Ain Shams Eng. J. 2023, 14, 102167. [Google Scholar] [CrossRef]
  15. Lee, G.; Kim, K.; Jang, J. Real-time path planning of controllable UAV by subgoals using goal-conditioned reinforcement learning. Appl. Soft Comput. 2023, 146, 110660. [Google Scholar] [CrossRef]
  16. Fan, X.L.; Li, D.; Zhang, W.; Wang, J.Z.; Guo, J.W. Flight vehicle evasion decision training based on deep reinforcement learning. Electron. Opt. Control 2021, 28, 81–85. [Google Scholar]
  17. Deng, T.B.; Huang, H.; Fang, Y.W.; Yan, J.; Cheng, H.Y. Reinforcement learning-based flight vehicle terminal guidance of maneuvering targets with baits. Chin. J. Aeronaut. 2023, 36, 309–324. [Google Scholar] [CrossRef]
  18. Qiu, X.Q.; Lai, P.; Gao, C.S.; Jing, W.X. Recorded recurrent deep reinforcement learning guidance laws for intercepting endoatmospheric maneuvering flight vehicles. Def. Technol. 2024, 31, 457–470. [Google Scholar] [CrossRef]
  19. Qian, X.F.; Lin, R.X.; Zhao, Y.N. Aircraft Flight Mechanics; Beijing Institute of Technology Press: Beijing, China, 2012; pp. 102–110. [Google Scholar]
  20. Tang, S.J.; Yang, B.E.; Xu, L.F.; Zhang, Y. Research on anti-interference technology based on target and infrared bait projectile motion mode. Aerosp. Shanghai 2017, 34, 44–49. [Google Scholar]
  21. Sigaud, O.; Buffet, O. Markov decision processes in artificial intelligence. In Markov Processes & Controlled Markov Chains; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar] [CrossRef]
  22. Howard, M. Multi-Agent Machine Learning: A Reinforcement Approach; China Machine Press: Beijing, China, 2017. [Google Scholar]
  23. Ma, X.P.; Zhao, L.Y. Overview of research status of key technologies of infrared seeker at home and abroad. Aviat. Weapons 2018, 3, 3–10. [Google Scholar]
  24. Bai, H.Y.; Zhou, Y.X.; Zheng, P.; Guo, H.W.; Li, Z.M.; Hu, K. A Real-Time Dynamic Generation System and Method for Target and Scene Used in Image Seeker. CN202010103846.3, 28 July 2020. [Google Scholar]
  25. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
  26. Vosoogh, A.; Sorkherizi, M.S.; Zaman, A.U.; Yang, J.; Kishk, A.A. An integrated Ka-Band diplexer-antenna array module based on gap waveguide technology with simple mechanical assembly and no electrical contact requirements. IEEE Trans. Microw. Theory Tech. 2018, 66, 962–972. [Google Scholar] [CrossRef]
Figure 1. Overall structural framework diagram.
Figure 1. Overall structural framework diagram.
Aerospace 11 00653 g001
Figure 2. Basic structure block diagram of actor-critic.
Figure 2. Basic structure block diagram of actor-critic.
Aerospace 11 00653 g002
Figure 3. Structure diagram of an infrared bait autonomous delivery framework with improved DDPG algorithm.
Figure 3. Structure diagram of an infrared bait autonomous delivery framework with improved DDPG algorithm.
Aerospace 11 00653 g003
Figure 4. Physical simulation system.
Figure 4. Physical simulation system.
Aerospace 11 00653 g004
Figure 5. Algorithm reward convergence diagram.
Figure 5. Algorithm reward convergence diagram.
Aerospace 11 00653 g005
Figure 6. Trajectory of incoming missile and aircraft.
Figure 6. Trajectory of incoming missile and aircraft.
Aerospace 11 00653 g006aAerospace 11 00653 g006b
Figure 7. Relative distance between incoming missile and mobile aircraft.
Figure 7. Relative distance between incoming missile and mobile aircraft.
Aerospace 11 00653 g007aAerospace 11 00653 g007b
Figure 8. Trajectory diagram of incoming missile-bait-aircraft.
Figure 8. Trajectory diagram of incoming missile-bait-aircraft.
Aerospace 11 00653 g008aAerospace 11 00653 g008b
Figure 9. Relative distance between incoming missile and maneuvering aircraft under the bait effect.
Figure 9. Relative distance between incoming missile and maneuvering aircraft under the bait effect.
Aerospace 11 00653 g009aAerospace 11 00653 g009b
Table 1. Improved DDPG algorithm flow.
Table 1. Improved DDPG algorithm flow.
Improved DDPG Algorithm
1: 
Obtain initial network weight parameters ω and θ through neural network pre-training
2: 
The critic current network and actor current network are randomly initialized with pre-trained parameters ω and θ
3: 
The critic target network and the actor target network are initialized with parameters ω′←ω and θ′←θ
4: 
Initializing experience pools I and II
5: 
for episode = 1, M do
6:   
Initialization state s1
7:   
for t = 1, T do
8:     
According to the parameters of the pre-training model, an action at is selected from the current state st
9:     
Perform action at to get the reward value rt and the next state st+1
10:    
Experience s t , a t , r t , s t + 1 is stored in the experience pool. If our aircraft successfully evades the attack of incoming flight vehicle under the interference of infrared bait, its data is distributed in the experience pool I, otherwise it is distributed in the experience pool II.
11:    
The data of m samples were sampled, and the number of random samples from experience pools I and II was updated according to Equation (18)
12:    
Update the current network of critic according to Formula (15)
13:    
Update the current network of the actor according to Formula (16)
14:    
Update the actor target network and critic target network according to Formula (17).
15:  
end for
16: 
end for
Table 2. Initial motion parameters of an incoming missile and the aircraft.
Table 2. Initial motion parameters of an incoming missile and the aircraft.
NameParameterValue
Simulation Situation IInitial position of aircraft (m) ( x m , y m , z m ) = ( 1000 ,   1500 ,   20,000 )
Initial position of missile (m) x t , y t , z t = ( 0 ,   0 ,   0 )
Initial speed of aircraft (m·s−1) v m = 400
Initial velocity of missile (m·s−1) v t = 500
Initial azimuth angle of missile (rad) θ t , ϕ t = ( 0.2 ,   0 )
Simulation Situation IIInitial position of aircraft (m) ( x m , y m , z m ) = ( 1000 ,   0 ,   10,000 )
Initial position of missile (m) x t , y t , z t = ( 0 ,   0 ,   0 )
Initial speed of aircraft (m·s−1) v m = 210
Initial velocity of missile (m·s−1) v t = 500
Initial azimuth angle of missile (rad) θ t , ϕ t = ( 0 ,   0 )
Simulation Situation ⅢInitial position of aircraft (m) ( x m , y m , z m ) = ( 1000 ,   0 ,   12,000 )
Initial position of missile (m) x t , y t , z t = ( 0 ,   3000 ,   0 )
Initial speed of aircraft (m·s−1) v m = 210
Initial velocity of missile (m·s−1) v t = 400
Initial azimuth angle of missile (rad) θ t , ϕ t = ( 0 ,   0 )
Table 3. Algorithm hyperparameter setting.
Table 3. Algorithm hyperparameter setting.
NameValue
Actor and critic learning rate0.0003
Discount factor0.99
Experience pool1 × 107
Sample rate of experience pool II0.05
Number of batch samples256
Maximum steps400
OptimizerAdam
Table 4. Detailed parameters of the Nvidia Jetson TX2.
Table 4. Detailed parameters of the Nvidia Jetson TX2.
NameValue
GPUNVIDIA Pascal architecture with 256 NVIDIA CUDA cores
CPUDual core Denver 2, 64 bit CPU and quad core ARM A57 Complex
Memory8 GB 128 bit LPDDR4
Storage space32 GB eMMC 5.1
ConnectGigabit Ethernet
Size87 mm × 50 mm
Table 5. Comparison of results.
Table 5. Comparison of results.
Simulation SituationNameBait TypeThe Number of Baits DroppedBait Drop Position ( x d , y d , z d ) /mBait Drop Angle ( θ d , ϕ d ) /radAverage Miss Distance/mEscape Success Rate/%Decision Time for PC/msDecision Time for Nvidia Jetson TX2/ms
IRelease without bait____0.50.3__
Bait releasePoint source bait5(1025.1, 1500.0, 21,234.9)(0.8, −0.02)670.296.82.512.5
(1081.3, 1499.9, 24,013.3)(1.1, −0.02)
(1106.5, 1500.0, 25,248.2)(1.2, −0.01)
(1162.7, 1499.9, 28,026.7)(1.4, −0.01)
(1206.6, 1500.0, 30,187.8)(1.6, −0.01)
IIRelease without bait ____0.20.24__
Bait releasePoint source bait4(993.9, −0.1, 9969.1)(−0.7, −0.02)1000.097.22.512.5
(890.1, 0.0, 9573.2)(−1.1, −0.02)
(656.5, 0.0, 8988.7)(−1.6, −0.01)
(302.1, 0.0, 8394.1)(−2.1, −0.01)
IIIRelease without bait ____0.60.4__
Bait releasePoint source bait4(719.6, −0.01, 10,491.0)(−1.3, −0.01)750.596.02.512.5
(481.0, −0.08, 9977.2)(−1.7, −0.02)
(−281.2, −0.1, 8979.7)(−2.6, −0.02)
(301.8, −0.06, 9680.7)(−2.0, −0.02)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xue, S.; Wang, Z.; Bai, H.; Yu, C.; Deng, T.; Sun, R. An Intelligent Bait Delivery Control Method for Flight Vehicle Evasion Based on Reinforcement Learning. Aerospace 2024, 11, 653. https://doi.org/10.3390/aerospace11080653

AMA Style

Xue S, Wang Z, Bai H, Yu C, Deng T, Sun R. An Intelligent Bait Delivery Control Method for Flight Vehicle Evasion Based on Reinforcement Learning. Aerospace. 2024; 11(8):653. https://doi.org/10.3390/aerospace11080653

Chicago/Turabian Style

Xue, Shuai, Zhaolei Wang, Hongyang Bai, Chunmei Yu, Tianyu Deng, and Ruisheng Sun. 2024. "An Intelligent Bait Delivery Control Method for Flight Vehicle Evasion Based on Reinforcement Learning" Aerospace 11, no. 8: 653. https://doi.org/10.3390/aerospace11080653

APA Style

Xue, S., Wang, Z., Bai, H., Yu, C., Deng, T., & Sun, R. (2024). An Intelligent Bait Delivery Control Method for Flight Vehicle Evasion Based on Reinforcement Learning. Aerospace, 11(8), 653. https://doi.org/10.3390/aerospace11080653

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop