Next Article in Journal
A Hierarchical Blockchain-Based Trust Measurement Method for Drone Cluster Nodes
Previous Article in Journal
Integrating a UAV System Based on Pixhawk with a Laser Methane Mini Detector to Study Methane Emissions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Constraint Guidance and Maneuvering Penetration Strategy via Meta Deep Reinforcement Learning

1
School of Aerospace Science and Technology, Xidian University, Xi’an 710126, China
2
College of Missile Engineering, Rocket Force University of Engineering, Xi’an 710025, China
3
China Aerospace Science and Technology Corporation, Beijing 100048, China
*
Author to whom correspondence should be addressed.
Drones 2023, 7(10), 626; https://doi.org/10.3390/drones7100626
Submission received: 21 August 2023 / Revised: 24 September 2023 / Accepted: 28 September 2023 / Published: 8 October 2023

Abstract

:
In response to the issue of UAV escape guidance, this study proposed a unified intelligent control strategy synthesizing optimal guidance and meta deep reinforcement learning (DRL). Optimal control with minor energy consumption was introduced to meet terminal latitude, longitude, and altitude. Maneuvering escape was realized by adding longitudinal and lateral maneuver overloads. The Maneuver command decision model is calculated based on soft-actor–critic (SAC) networks. Meta-learning was introduced to enhance the autonomous escape capability, which improves the performance of applications in time-varying scenarios not encountered in the training process. In order to obtain training samples at a faster speed, this study used the prediction method to solve reward values, avoiding a large number of numerical integrations. The simulation results demonstrated that the proposed intelligent strategy can achieve highly precise guidance and effective escape.

1. Introduction

Hypersonic UAVs mainly glide in near-space [1]. In their early phase, a higher flight velocity is acquired, relying on the thin atmospheric environment, which is an advantage for effectively avoiding the interception of defense systems. At the end of a gliding flight, the velocity is mainly influenced by the aerodynamic force and suffers from the restrictions of heat flow, dynamic pressure, and overload [2]. The velocity advantage leads to penetration becoming more difficult, so orbital maneuvering is applied by UAVs to achieve penetration. The main flight mission is split into avoiding defense system interception and satisfying multiple terminal constraints [3]. The core of this manuscript is designing a penetration guidance strategy via orbital maneuvering capabilities, avoiding interception, and reducing the penetration’s impact on guidance accuracy.
The penetration strategy is summarized as a tactical penetration and a technical penetration strategy [4]. The technical penetration strategy changes the flight path through maneuvering, aiming to increase the missed distance in order to successfully penetrate. Common maneuver manners include the sine maneuver, step maneuver, square wave maneuver, and spiral maneuver [5]. There are some limitations and instability for technical penetration strategies, attributed to the UAV struggling to adopt an optimal penetration strategy according to the actual situation of the offensive and defensive confrontations. Compared with the traditional procedural maneuver strategy, the differential game guidance law has the characteristics of real-time operation and intelligence as a tactical penetration strategy [6]. Penetration problems are essentially regarded as the continuous dynamic conflict problem of multi-party participants, and this strategy is an essential solution for solving multi-party optimal control problems. Applying it to the problem of attack–defense confrontation can not only fully consider the relative information between UAVs and interceptors but also obtain a Nash equation strategy to reduce energy consumption. Many scholars have proposed differential game models of various maneuvering control strategies based on control indexes and motion models. Garcia [7] regarded the scenario of active target defense modeling as a zero-sum differential game, designed a complete differential game solution, and comprehensively considered the optimal strategy of closed-loop state feedback to obtain the value function. In Ref. [8], the optimal guidance problem was studied between an interceptor and an active defense ballistic UAV, and an optimal guidance scheme was proposed based on the linear quadratic differential game method and the numerical solution of Riccati differential equations. Liang [9] mainly analyzed the problem of pursuit and escape attacks of multiple players, inducted the three-body game confrontation into competition and cooperation problems, and solved the optimal solution of multiple players via differential game theory. The above methods are of great significance for analyzing and solving the confrontation process between UAVs and interceptors. Near-space UAVs have the characteristics of high velocity and short time in the phase of attack and defense confrontation terminal guidance [10], and the differential game guidance law struggles to show advantages in this phase. Moreover, the differential game method requires a large amount of calculation, and the bilateral performance indicators are difficult to model [11]; as a result, this theory is unable to be applied in practice.
DRL is a research hotspot in the field of artificial intelligence that has sprung up in recent years, and amazing results have been achieved in robot control, guidance, and control technologies [12]. DRL specifically refers to agents learning in the process of interaction with the environment to find the best strategy to maximize cumulative rewards [13]. With the advantages of dealing with high-dimensional abstract problems and making decisions quickly, DRL provides a new solution for the maneuvering penetration of high-velocity UAVs. In order to solve the problem of intercepting a high maneuvering target, an auxiliary DRL algorithm was proposed in Ref. [14] to optimize the frontal interception guidance control strategy based on a neural network. Simulation results showed that DRL had a higher hit rate and larger terminal interception angle than traditional methods and proximal policy optimization algorithms. Gong [15] proposed an Omni bearing attack guidance law for agile UAVs via DRL, which effectively dealt with aerodynamic uncertainty and strong nonlinearity at a high attack angle. DRL was used to generate a guidance law for the attack angle in an agile turning phase. Furfaro [16] proposed an adaptive guidance algorithm based on classical zero-effort velocity, and the limitations of this algorithm were overcome via RL. A closed-loop guidance algorithm was created that is lightweight and flexible enough to adapt to a given constrained scene.
Compared with differential game theory, DRL is convenient for establishing the performance index function, and it is a feasible method to solve the dynamic programming problem by utilizing the powerful numerical calculation ability of computers to skillfully avoid solving the function analytical solution. However, the traditional DRL has some limitations, such as high sample complexity, low sample utilization, a long training time, and so on. Once the mission changes, the original DRL parameters are hard to adapt to the new mission and need to be learned from scratch. A change in mission or environment will lead to the failure of the trained model and poor generalization ability of the model. In order to solve the existing problems in DRL, researchers introduced meta-learning into DRL and proposed Meta DRL [17]. Lu et al. [18] mainly solved the issues of maximizing the total data collected and avoiding collisions during the guidance flight and improved the adaptability to different missions via meta RL. Hu et al. [19] mainly studied a challenging trajectory design problem, optimized the DBS trajectory, and considered the uncertainty and dynamics of terrestrial users’ service requests. MAML was introduced for the purposes of solving raised POMDPS and enhancing adaptability to balance flight aggressiveness and safety [20]. For suspended payload transportation tasks, the paper [21] proposed a meta-learning approach to improve adaptability. The simulation demonstrated improvements in closed-loop performance compared to non-adaptive methods.
By learning useful meta knowledge from a group of related missions, agents acquire the ability to learn, and the learning efficiency on new missions is improved and the complexity of samples is reduced. When faced with new missions or environments, the network responds quickly based on the previously accumulated knowledge, so only a small number of samples are needed to quickly adapt to the new mission.
Based on the above analysis, this manuscript proposes Meta DRL to solve the UAV guidance penetration strategy, and the DRL is improved, resulting in enhancing the adaptability of UAVs in complex and changeable attack and defense confrontations. In addition, the idea of meta-learning is used to enable UAVs to learn and improve their ability for autonomous flight penetration. The core contributions of this manuscript are as follows:
  • By modeling the three-dimensional attack and defense scene between UAVs and interceptors, we analyze the terminal and process constraints of UAVs. A guidance penetration strategy based on DRL is proposed, aiming to provide the optimal solution for maneuvering penetration under a constant environment or mission.
  • Meta-learning is used to improve the UAV guidance penetration strategy. The improvement enables the UAV to learn and enhances the autonomous penetration ability.
  • Through the simulation analysis, the manuscript analyzes the penetration strategy, explores penetration timing and maneuvering overload, and summarizes the penetration tactics.

2. Modeling of the Penetration Guidance Problem

2.1. Modeling of UAV Motion

The three-degree-of-freedom motion equation is adopted to describe UAVs and the dynamic equation is established in the ballistic coordinating system:
v ˙ = ρ v 2 S m C D 2 m + g r sin θ + g ω e cos σ cos θ cos ϕ + sin θ sin ϕ               + ω e 2 r cos 2 ϕ sin θ cos ϕ sin ϕ cos σ cos θ θ ˙ = ρ v 2 S m C L cos υ 2 m v + g r cos θ v 2 ω e sin σ cos ϕ +       v cos θ r           + g ω e v cos θ sin ϕ cos σ sin θ cos ϕ + ω e 2 r v cos ϕ sin ϕ cos σ sin θ + cos 2 ϕ cos θ σ ˙ = ρ v 2 S m C L sin υ 2 m v cos θ g ω e sin σ cos ϕ v cos θ +     ω e 2 r cos ϕ sin ϕ sin σ v cos θ + v tan ϕ cos θ sin σ r           2 ω e sin ϕ cos σ tan θ cos ϕ ϕ ˙ = v cos θ cos σ r λ ˙ = v cos θ sin σ r cos ϕ r ˙ = v sin θ
where v is the velocity of UAV relative to the earth, θ is the velocity slope angle, σ is the velocity azimuth, and the positive direction is clockwise from the north. r represents the geocentric distance, and λ , ϕ is the longitude and latitude. The differential argument is flight time t . g ω e is the component of the Earth’s gravitational acceleration in the direction of the Earth’s rotational angular rate ω e , while g r is the component of the Earth’s gravitational acceleration in the direction of the geo-center. ρ is the density of the atmosphere, while m and S m are the mass and reference area of UAV. C D and C L are the drag and lift coefficient, respectively, relating to the Mach number and attack angle, so the control variable attack angle α is implicit in it, and the other control variable is the bank angle υ .
For a hypersonic UAV with a large lift–drag ratio (L/D), the heat flow, overload, and dynamic pressure are considered as flight process constraints, as follows:
k h ρ 1 / 2 v 3 Q s max ρ v 2 2 q max D 2 + L 2 m g 0 n max
where Q s max , q max , and n max are the maximum heat flow, dynamic pressure, and overload, respectively; the parameter k h is a constant coefficient, and g0 is the gravitational acceleration at sea level. In order to keep the gliding state steady and effectively prevent trajectory jumps, the balance of lift and gravity is required by UAV [22]. Stable gliding is generally considered as quasi-equilibrium gliding (QEGs). At present, the international definition of QEG can be summarized into two types [23]: the velocity slope angle is considered as constant, expressed by θ ˙ = 0 , or the flight altitude variance ratio is considered as constant, expressed by h ¨ = 0 . Moreover, θ ˙ = 0 was adopted by traditional QEG guidance, and h ¨ = 0 mainly appeared in the analytical prediction guidance of the Mars landing at the end of the last century [24].
For high L/D UAVs flying in the gliding phase, the Coriolis inertial force and centrifugal inertial force caused by the Earth’s rotation are relatively small compared to aerodynamic forces. Therefore, in the design of guidance laws, the Earth can be assumed to be a homogeneous sphere that does not rotate. We further decompose Equation (1) into longitudinal and lateral directions, where the longitudinal motion equation is as follows:
v ˙ = ρ v 2 S m C D 2 m g sin θ θ ˙ = ρ v 2 S m C L cos υ 2 m v g cos θ v + v cos θ r h ˙ = v sin θ L ˙ R = R e v cos θ / R e + h
where h = r R e represents the flight altitude, and L R is range. For a gliding UAV flying in close proximity to space, h R e . Therefore, range differentiation can be further described as follows:
L ˙ R = v cos θ
This manuscript adopts θ ˙ = 0 as the QEG condition. Referring to the second equation in Equation (3), θ ˙ = 0 is transferred into Equation (5).
θ ˙ = 0 = ρ v 2 S m C L cos υ 2 m v g cos θ v + v cos θ r ρ v 2 S m C L cos υ 2 m v = g cos θ v v cos θ r ρ v 3 S m C L cos υ 2 = m g cos θ v m v 2 cos θ r
where L = ρ v 3 S m C D 2 , hence
m ( g v 2 r ) cos θ L cos υ = 0

2.2. Description of Flight Missions

2.2.1. Guidance Mission

The physical meaning of gliding guidance is eliminating heading errors, satisfying complex process constraints, and minimizing energy loss. The UAV is guided to glide unpowered to the setting terminal target point ( h f , λ f , ϕ f ), satisfying the terminal altitude, longitude and latitude. Hence, the terminal constraints are expressed by Equation (7).
h ( L R f ) = h f ,   λ ( L R f ) = λ f ϕ ( L R f ) = ϕ f , Δ σ f Δ σ max
where the terminal range L R f is given, Δ σ represents the heading error, and Δ σ max is a pre-setting allowable value. The guidance problem is the process of determining α and υ .

2.2.2. Penetration Mission

The main indexes are used to judge penetration probability as follows:
(1)
The miss distance D m i s s with the interceptor at the encounter moment.
(2)
The overload N m of the interceptor in the last phase.
(3)
The line-of-sight (LOS) angular rates θ ˙ int l o s and σ ˙ int l o s with the interceptor at the encounter moment.
where D m i s s directly reflects the result of penetration—the larger the D m i s s , the greater of penetration probability. The N m and LOS angle indirectly reflect the result of penetration. The larger the θ ˙ int l o s and σ ˙ int l o s , the more difficult it is for the interceptor to successfully intercept, which is attributed to θ int l o s and σ int l o s not easily converging to a constant value at the encounter moment. In addition, a larger overload is required to adjust θ int l o s and σ int l o s at the end of interception. The larger the N m , the greater the control cost that the interceptor has to pay to complete the interception mission. Once N m exceeds the overload limit of the interceptor, it indicates that the control is saturated, which demonstrates that the interceptor struggles to intercept based on the current maximum overload constraint. Referring to the size and flight characteristics of UAVs [25], this manuscript assumes that the penetration mission is completed if D m i s s is greater than 2 m.
The maximum maneuvering overload of UAVs is constrained by the structure and QEG condition [24]. If the lateral maneuver overload is too large, it will cause the UAV to deviate from the course, eventually leading to the failure of the guidance mission. A large longitudinal maneuver amplitude will have a significant impact on the L/D of the UAV, affecting safety. According to the above analysis, the maximum lateral maneuvering overload is set as 2 g, and the maximum longitudinal maneuvering overload is set as 1 g.

2.3. The Guidance Law of Interceptor

The guidance law of the interceptor relies on the inertial navigation system to obtain information such as the position and velocity of the UAV. The relative motion is shown in Figure 1.
P, T, and M, respectively, represent the UAV, target, and interceptor. r is the relative position between the UAV and the interceptor, and r t is the target.
The LOS angular rate and the approach velocity to the UAV are obtained for the interceptor. Overload control command is derived via generalized proportional navigation guidance (GPNG).
n y 2 * = K D r ˙ θ ˙ int l o s / g 0 n z 2 * = K T r ˙ σ ˙ int l o s / g 0
As shown in Equation (8), K D and K T are navigation ratios in the longitudinal and lateral direction, respectively. r ˙ is the approach velocity. θ ˙ int l o s and σ ˙ int l o s represent the LOS angular rate in the longitudinal and lateral directions, respectively.
r = ( x x m ) 2 + ( y y m ) 2 + ( z z m ) 2 θ int l o s = arcsin ( ( y y m ) / r ) σ int l o s = arctan ( ( z z m ) / ( x x m ) )
Differentiating Equation (9) with respect to t , we obtain the following:
θ ˙ int l o s = 1 1 ( y y m r ) 2 ( y ˙ y ˙ m ) r ( y y m ) r ˙ r 2 σ ˙ int l o s = 1 1 + ( z z m x x m ) 2 ( z ˙ z ˙ m ) ( x x m ) ( z z m ) ( x ˙ x ˙ m ) ( x x m ) 2 r ˙ = ( x x m ) ( x ˙ x ˙ m ) + ( y y m ) ( y ˙ y ˙ m ) + ( z z m ) ( z ˙ z ˙ m ) r

3. Design of Penetration Strategy Considering Guidance

3.1. Guidance Penetration Strategy Analysis

Generally, the UAV achieves penetration through a velocity advantage or increased maneuvering overload. The former is used in pre-gliding flight, and compared with interceptors, the velocity of UAVs is relatively large, which is conducive to penetrating defenses. More threats to UAVs come from the defense systems of intended targets, resulting in an intercept threat focusing on the end of the glide flight. However, the flight velocity gradually decreases, which is not enough to penetrate escape. Based on the above analysis, the manuscript designs a penetration strategy by increasing the maneuvering overload.
Firstly, in previous research [26], the energy-optimized gliding guidance law was designed. In this study, avoiding the intercept by maneuvering becomes the key point. The real-time flight information of the interceptor cannot easily be accurately obtained by the UAV, while the real-time flight information of the UAV is obtained by the interceptor. Based on the UAV penetration mission, which only knew the initial launch position of the interceptor, this manuscript simulates the attack and defense environment between the UAV and the interceptor, applying the DRL method to solve the overload command of UAV maneuvering penetration. DNN parameters are trained offline, and a maneuvering penetration command is constructed online. The launch position of the interceptor is changeable, and a stable penetration command cannot adapt to the complex penetration environment. To improve the adaptability of DNN parameters, the original DRL is optimized by adopting the idea of meta-learning, and the ontology and environmental information are fully utilized. The optimization of meta-learning enhances flight capability, fast response to complex missions, and flight self-learning. Finally, the UAV guidance penetration strategy based on Meta DRL is proposed in this manuscript, as shown in Figure 2.
As for the gliding guidance mission, DRL and optimal control are used to achieve the guidance penetration mission of the UAV. An optimal guidance command was conducted in previous research [26], which was introduced to satisfy constraints regarding terminal position, altitude, and minimal energy loss. Maneuvering overloads are added between the longitudinal and lateral directions, aiming to achieve the penetration mission at the end of the gliding flight. Maneuvering overloads are solved via DRL. The DNN parameters are optimized by means of meta-learning.
Guidance command n y ,   n z is generated by the optimal guidance strategy, and maneuvering overload command Δ n y , Δ n z is generated by meta DRL. The flight total overload is shown in Equation (11).
N y = n y + Δ n y N z = n z + Δ n z
Maneuvering penetration command via meta DRL is the core of this manuscript. The penetration considering guidance is described as a Markov decision process (MDP), which consists of finite-dimensional continuous flight state space, longitudinal and lateral direction overload sets, and a reward function judging the penetration strategy. Flight data generated via numerical integration and SAC networks are introduced to train and learn MDP. Optimizing network parameters via meta-learning aims at adjusting network parameters with very little flight data when the UAV is faced with online mission changes, adapting to the new environment as soon as possible.

3.2. Energy Optimal Gliding Guidance Method

In previous research [26], based on the QEG condition and taking the required overload as the control amount, the performance index with minimum energy loss was established. The optimal longitudinal and lateral overload were designed, respectively, satisfying the constraints on terminal latitude, longitude, altitude, and velocity. The required overload command is shown in Equation (12).
u y * = k C h L R C θ + 1 u z * = σ l o s σ k L R f L R
where u y * = n y * and u z * = n z * are the optimal overload between the longitudinal and lateral dimensions. k = g 0 v 2 g v 2 , L R is the current range, and L R f is the total range of the gliding phase. C h and C θ are the guidance coefficients based on optimal control, represented as Equation (13).
C h = 6 L R L R f θ f + θ 2 h + 2 h f k 2 L R L R f 3 C θ = 2 L R L R f θ θ f L R f 2 2 θ + θ f + L R 2 2 θ f + θ + 3 L R f + L R h f h k 2 L R L R f 3
Based on Equation (14), the control variables α and υ are calculated as follows:
ρ v 2 S m C L ( M a , α ) 2 g 0 = n y * 2 + n z * 2 υ = arctan n z * n y *
where α is obtained by calculating the contrast value in Equation (14).

4. RL Model for Penetration Guidance

The problem of maneuvering penetration is modeled as a series of stationary MDPs [27] with unknown transition probabilities. The continuous flight state space, action set, and reward function for judging the command are determined in this section.

4.1. MDP of the Penetration Guidance Mission

Deterministic MDP with a continuous state and action is defined as ( S , A , T , R , γ ) via a quintuple. S is described as the continuous state space, A is described as the finite action set, and T is depicted as the state transition function. S × A T , reflecting deterministic state transition relationships. R is defined as the immediate reward. γ [ 0 , 1 ] is the discount factor, to balance immediate and forward reward.
The agent chooses the action a t A at the current state s t , while the state changes from s t to s t + 1 S , and the environment returns an immediate reward R t = f ( s t , a t ) for the agent. Cumulative rewards are obtained with a controlled action sequence τ , as shown in Equation (15).
G s 0 , τ = R 0 + γ R 1 + γ 2 R 2 + ... = t = 0 γ t R t
The goal of MDP is determining policy, and maximizing the expected accumulated rewards, as follows:
τ = arg max τ G s 0 , τ

4.1.1. State Space Design

The environment of attack–defense is abstracted as a state space of MDP, which applies guidance for action commands. UAVs struggle to acquire the flight information and guidance laws of interceptors. Hence, the information of the UAV and the target is only considered as state space.
S = x r , y r , z r , θ l o s , σ l o s
where x r , y r , z r represents the relative distance between the UAV and the target under the North East Down (NED) Coordinate System. θ l o s , σ l o s represents, respectively, the longitudinal and lateral LOS angles. In order to eliminate dimension differences and enhance compatibility among states, S is normalized using Equation (18).
S = x r = x r x r 0 ,   y r = y r y r 0 ,   z r = z r z r 0 ,   θ l o s = θ l o s 2 π ,   σ l o s = σ l o s 2 π

4.1.2. Action Space Design

Action is a decision selected by the UAV based on the current state, and action space A is the set of all possible decisions. Overload directly affects the velocity azimuth and slope angle and indirectly affects the gliding flight status; hence, this manuscript determines overload as an intermediate control variable.
On the basis of the optimal guidance law and flight process constraints, longitudinal overload Δ n y and lateral overload Δ n z are added to the action space of MDP. Considering the safety of the UAV and heading error, this manuscript assumes that the longitudinal and lateral maximum maneuvering overloads are 1 g and 2 g, respectively. A is defined as a continuous set, shown in Equation (19).
A = Δ n y [ g ,   g ] Δ n z [ 2 g ,   2 g ]

4.2. Multi-Missions Reward Function Designing

The reward function f r is an essential section for guiding and training the maneuvering penetration strategy. After execution of the action command, f r returns a reward value to the UAV, which reflects the fairness and rationality of the action judgment. The rationality of f r directly affects the training result and determines the efficiency of SAC training. In this manuscript, the aim of f r is to guide the UAV to achieve the guidance penetration mission while satisfying terminal multi-constraints. Given the requirements of the mission, f r consists of the miss distance with the interceptor and the terminal deviation from the target.
f r ( s , a ) = c 1 d m i s s c 2 d e r r o r
where d m i s s and d e r r o r are the miss distance and terminal deviation after execution the action command. The sufficient and necessary condition of satisfying terminal position deviation is eliminating the heading error, which is directly related to the LOS angular rate. Similarly, d m i s s can be reflected by the LOS angular rate at the encounter time between the UAV and the interceptor. Therefore, the normalized f r is expressed by Equation (21).
f r = c 1 θ ˙ ¯ int l o s 2 + σ ˙ ¯ int l o s 2 c 2 θ ˙ ¯ l o s 2 + σ ˙ ¯ l o s 2
where θ ˙ ¯ int l o s represents the normalized LOS angular rate in the longitudinal direction with the interceptor, σ ˙ ¯ int l o s represents the normalized LOS angular rate in the lateral direction with the interceptor. θ ˙ ¯ l o s represents the normalized LOS angular rate in the longitudinal direction with the target, and σ ˙ ¯ l o s represents the normalized LOS angular rate in the lateral direction with the target. In this manuscript, the LOS angular rates at the encounter time and terminal time are solved analytically via numerical calculation.

4.2.1. The Solution of LOS Angular Rate in the Lateral Direction

The attack–defense confrontation model in the lateral direction is shown in Figure 3.
P, T, and M, respectively represent UAV, target, and interceptor. LRgo and LMgo are represented as the remaining range among UAV, target and interceptor, which are calculated in Equations (22) and (23).
The lateral relative motion model between UAV and the target is shown in Equation (22).
L ˙ R g o = v cos Δ δ L R g o σ ˙ l o s = v sin Δ δ
where Δ δ = σ σ l o s , represents the heading error between UAV and the target point. Lateral relative motion model between UAV and interceptor is shown in Equation (23).
L ˙ M g o = v cos Δ σ v m cos Δ σ m L M g o σ ˙ int l o s = v sin Δ σ v m sin Δ σ m
where Δ σ = σ σ int l o s , represents heading error between UAV and the interceptor. In order to simplify the calculation, the relative motion equation Equation (23) is conducted as Equation (24):
L ˙ M g o = v r cos Δ σ m n L M g o σ ˙ int l o s = v r sin Δ σ m n
where v r and Δ σ m n are calculated as Equation (25).
v r = v cos σ v v m cos σ m 2 + v sin σ v v m sin σ m 2 Δ σ m n = atan v sin σ v v m sin σ m v cos σ v v m cos σ m
To facilitate the analysis and prediction, σ ˙ l o s is calculated as follows, taking the derivation of the second formula in Equation (22):
L ˙ R g o σ ˙ L O S + L R g o σ ¨ L O S = v ˙ sin Δ σ + v Δ σ ˙ cos Δ σ
Bring the heading error and first formula in Equation (22) into Equation (26), the rate of LOS angular rate is calculated by Equation (27).
L ˙ R g o σ ˙ l o s + L R g o σ ¨ l o s = v ˙ sin Δ σ + v Δ σ ˙ cos Δ σ L ˙ R g o σ ˙ l o s + L R g o σ ¨ l o s = v ˙ v L R g o σ ˙ l o s L ˙ R g o σ ˙ l o s + L ˙ R g o σ ˙ σ ¨ l o s = v ˙ v 2 L ˙ R g o L R g o σ ˙ l o s + L ˙ R g o L R g o σ ˙
Defining T g o c as the predicted remaining time of flight, T g o c is derived via the remaining flight range and variation in range, expressed in Equation (28).
T g o c = L R g o L ˙ R g o
Defining the value of state x = σ ˙ l o s and the value of control u = σ ˙ , the differential of LOS angular rate is obtained, which is shown in Equation (29).
x ˙ = v ˙ v + 2 T g o c x 1 T g o c u
In the latter phase of glide flight, v ˙ v is an order of magnitude smaller than 2 T g o c . Equation (29) is further simplified to Equation (30).
x ˙ = 2 T g o c x 1 T g o c u
The current remaining flight time Tgoc, as shown in Equation (31). a certain time t of future flight starting from the current time t, and the remaining flight time Tgo1 at time t satisfy the following relationship:
T g o c = T g o 1 + t
d T g o 1 = d t represents the derivation of remaining flight time. For a given control input u, the definite integral of Equation (30) is solved, and the calculated result is shown in Equation (32).
x t = e 2 T g o 1 d t u T g o 1 e 2 T g o 1 d t d t + C = e 2 T g o c t d t u T g o c t e 2 T g o c t d t d t + C = e 2 ln T g o c t u t T g o c e 2 ln t T g o c d t + C = 1 T g o c t 2 u t T g o c d t + C = 1 T g o c t 2 u 1 2 t 2 T g o c t + C
The LOS angular rate is σ ˙ l o s , at the current time t = 0, the constant C is expressed by Equation (33).
C = σ ˙ l o s T g o c 2
σ ˙ l o s is obtained by Equation (34).
σ ˙ l o s t = 1 T g o c t 2 u 1 2 t 2 T g o c t + σ ˙ l o s T g o c 2
σ ˙ l o s at the terminal time tf is shown in Equation (35).
σ ˙ l o s = σ ˙ l o s t f = 1 T g o c t f 2 u 1 2 t f 2 T g o c t f + σ ˙ l o s T g o c 2
Similarly, based on above analysis, σ ˙ int l o s is calculated via the analysis and prediction. The solution is shown in Equation (36).
σ ˙ int l o s = σ ˙ int l o s t int f = 1 T int g o c t int f 2 u 1 2 t int f 2 T int g o c t int f + σ ˙ int l o s T int g o c 2
where T int g o c represents the total encounter time based on the current moment, t int f represents encounter time with interceptor, and u is the input overload.

4.2.2. The Solution of LOS Angular Rate in the Longitudinal Direction

The attack–defense confrontation model in longitudinal direction is shown in Figure 4.
RRgo and RMgo represent the remaining range among the UAV, target and interceptor, which are calculated in Equations (37) and (38).
The longitudinal relative motion model between the UAV and the target is shown in Equation (22).
R ˙ R g o = v cos Δ θ R R g o θ ˙ l o s = v sin Δ θ
The longitudinal relative motion model between the UAV and the interceptor is shown in Equation (38).
R ˙ M g o = v cos Δ θ v m cos Δ θ m R ˙ M g o θ ˙ int l o s = v sin Δ θ v m sin Δ θ m
In order to simplify the calculation, the relative motion equation Equation (38) is expressed as Equation (39).
R ˙ M g o = v r cos Δ θ m n R M g o θ ˙ int l o s = v r sin Δ θ m n
where v r and Δ θ m n are calculated in Equation (40).
v r = v cos θ v v m cos θ m 2 + v sin θ v v m sin θ m 2 Δ θ m n = atan v sin θ v v m sin θ m v cos θ v v m cos θ m
To facilitate the analysis and prediction, θ ˙ l o s is calculated as follows. Taking the derivation of the second formula in Equation (37) results in Equation (41):
R ˙ R g o θ ˙ L O S + R R g o θ ¨ L O S = v ˙ sin Δ θ + v Δ θ ˙ cos Δ θ
Bring the heading error and first formula in Equation (37) results in Equation (41), the rate of the LOS angular rate is calculated using Equation (42).
R ˙ R g o θ ˙ l o s + R R g o θ ¨ l o s = v ˙ sin Δ θ + v Δ θ ˙ cos Δ θ R ˙ R g o θ ˙ l o s + R R g o θ ¨ l o s = v ˙ v R R g o θ ˙ l o s R ˙ R g o θ ˙ l o s + R ˙ R g o θ ˙ θ ¨ l o s = v ˙ v 2 R ˙ R g o R R g o θ ˙ l o s + R ˙ R g o R R g o θ ˙
Based on the predicted remaining time of flight T g o c in lateral prediction, the LOS angular rate is σ ˙ l o s , and at the current time t = 0, the constant C is expressed by Equation (43).
C = θ ˙ l o s T g o c 2
θ ˙ l o s is obtained using Equation (44).
θ ˙ l o s t = 1 T g o c t 2 u 1 2 t 2 T g o c t + θ ˙ l o s T g o c 2
θ ˙ l o s at the terminal time tf is shown in Equation (45).
θ ˙ l o s = θ ˙ l o s t f = 1 T g o c t f 2 u 1 2 t f 2 T g o c t f + θ ˙ l o s T g o c 2
Similarly, based on the above analysis, θ ˙ int l o s is calculated via the analysis and prediction. The solution is shown in Equation (46).
θ ˙ int l o s = θ ˙ int l o s t int f = 1 T int g o c t int f 2 u 1 2 t int f 2 T int g o c t int f + θ ˙ int l o s T int g o c 2
where T int g o c represents the total encounter time based on the current moment, t int f represents encounter time with interceptor, and u is the input overload.
According to the above analysis, σ ˙ l o s , θ ˙ l o s and σ ˙ int l o s , θ ˙ int l o s depend on the change in the LOS angle rate. For the guidance mission, σ ˙ l o s , θ ˙ l o s is related to the overload of the UAV, and the smaller the value, the closer the UAV approaches the target at the end of the gliding flight. For the penetration mission, σ ˙ int l o s , θ ˙ int l o s is related to the overloads of the UAV and the interceptor, and the greater the value, the higher the costs of the interceptor at the interception terminal phase, and the UAV will break through more easily if the control overload of the interceptor reaches saturation.

5. DRL Penetration Guidance Law

5.1. SAC Training Model

Standard DRL maximizes the sum of expected rewards t E ( s t , a t ) ρ π f r s t , a t . For the problem of multi-dimensional continuous state inputs and a continuous action output, SAC networks are introduced to solve the MDP model.
Compared with other policy learning algorithms [28], SAC augments the standard RL objective with expected policy entropy using Equation (47).
J π = t γ t E ( s t , a t ) ρ π f r s t , a t + τ H π   s t
The entropy term τ H π   s t is shown in Equation (48), which represents the stochastic feature of the strategy, balancing the exploration and learning of networks. The entropy parameter τ determines the relative importance of entropy against the immediate reward.
H π   s t = a A π a s t log π a s t d a = E a π   s t log π a s t
The optimal strategy of SAC is shown in the Equation (49), aiming at maximizing the cumulative reward and policy entropy.
π M a x E n t * = arg max π t γ t E ( s t , a t ) ρ π f r s t , a t + τ H π   s t
The framework of SAC networks is shown in Figure 5, consisting of an Actor network and a Critic network. The Actor network generates the action, and the environment returns the reward and the next state. All of the ballistics data are stored in the experience pool, including the state, action, reward, and next state.
The Critic network is used to judge the found strategies, which impartially guides the strategy of the Actor network. At the beginning, the Actor network and Critic network are given random parameters. The Actor network struggles to generate the optimal strategy, and Critic network struggles to scientifically judge the strategy of the Actor network. The parameters of the networks need to be updated based on continuously generating data and sampling ballistics data.
For updating the Critic network, it outputs the expected reward E a π   s t based on samples, and the Actor network outputs the action probability, which is depicted by the entropy term H π   s t . Combining E a π   s t with H π   s t , the value function is conducted and shown in Equation (50)
Q s o f t ( s t , a t ) = E ( s t , a t ) ρ π t = 0 γ t r ( s t , a t ) + τ t = 1 γ t H ( π ( | s t ) )
We further obtain the Bellman equation, as shown in Equation (51):
Q s o f t ( s t , a t ) = E s t + 1 ρ ( s t + 1 | s t , a t ) a t + 1 π r ( s t , a t ) + γ Q s o f t ( s t + 1 , a t + 1 ) + τ H ( π ( | s t + 1 ) )
Given by Equation (52), the loss function of the Critic network is acquired as follows:
J Q ( ψ ) = E ( s t , a t , s t + 1 ) D a t + 1 π 1 2 Q s o f t ( s t , a t ) r ( s t , a t ) + γ Q s o f t ( s t + 1 , a t + 1 ) τ log π a t + 1 | s t + 1 2
For updating the Actor network, the updating strategy is shown in Equation (53)
π n e w = arg min D K L π π ( | s t ) exp 1 τ Q s o f t π o l d ( s t , ) Z s o f t π o l d ( s t )
where Π represents the set of strategies, and Z is the partition function, used to normalize the distribution. D K L is the Kullback–Leibler (KL) divergence [29].
Combining the re-parameterization technique with Equation (54), the loss function of the Actor network is obtained as follows:
J π ( ϕ ) = E s t D , ε t N [ τ log π ( f ( ε t ; s t ) | s t ) Q s o f t ( s t , f ( ε t ; s t ) ) ]
in which a t = f ( ε t ; s t ) , and ε t is the input noise, obeying the distribution N .
The method of stochastic gradient descent is introduced to minimize the loss function of networks. The optimal parameters of the Actor–Critic networks are obtained by repeating the updating process and passing the parameters to the target networks via soft updating.

5.2. Meta SAC Optimization Algorithm

The learning algorithm in DRL relies on a lot of interaction between the agent and the environment and high training costs. Once the environment changes, the original strategy is no longer applicable and needs to be learned from scratch. The penetration guidance problem under the stable flight environment can be solved using SAC networks. For a changeable flight environment, such as where the initial position of the interceptor changes greatly or the interceptor guidance law deviates greatly from the preset value, the strategy solved via traditional SAC struggles to adapt, which thus requires to restudy and redesign. This manuscript introduces meta-learning to optimize and improve SAC performance. The training goal of Meta SAC is to obtain initial SAC model parameters. When the UAV penetration mission is changed, through a few scenes of learning, the UAV can adapt to the new environment and complete the corresponding guidance penetration mission, without relearning model parameters. Meta SAC can achieve “learn while flying” for UAVs and strengthen the adaptability of UAVs.
The Meta SAC algorithm is shown in Algorithm 1, which is divided into a meta-training and meta-testing phase. The meta-training phase seeks to determine the optimal meta-learning parameters based on multi-experience missions. In the meta-testing phase, the trained meta parameters are interactively learned in the new mission environment to fine-tune the meta parameters.
Algorithm 1 Meta SAC
1: Initialize the experience pool Ω , Storage space N
2: Meta training:
3: Inner loop
4: for iteration k do
5:     sample mission(k) from T p T
6:     update actor policy Θ to Θ using SAC based on mission(k):
7:      Θ S A C Θ , m i s s i o n k .
8: Outer loop
9:  Θ = m m s e i = 1 k Θ i
10: Generate D 1 from Θ and estimate the reward of Θ .
11: Add a hidden layer feature as a random noise.
12: Θ i = Θ + α Θ Θ E α t π α t s t ; Θ , z i , z i N μ i , σ i t R t s t
13: The meta-learning process of different missions is carried out through SGD.
14: for iteration mission(k) do
15:      min Θ T i p T L T i f Θ i = T i p T L T i f Θ α Θ L T i f Θ
16:      Θ = Θ β Θ T i p T L T i f Θ i
17: Meta testing
18: Initialize the experience pool Ω , Storage space N.
19: Load meta training network parameters Θ .
20: Set training parameters.
21: for iteration i do
22:   sample mission from T p T
23:    Θ = S A C Θ , m i s s i o n k
End for
The basic assumption of Meta SAC is that the experience mission for meta training and the new mission for meta testing obey the same mission distribution p T . Therefore, there are some common characteristics between different missions. In the DRL scenario, our goal is to learn a function f θ with parameter θ , which can minimize the loss function L T of a specific mission T . In the meta DRL scenario, our goal is to learn a learning process θ = μ ψ D T t r , θ , which can quickly adapt to the new mission T with a very small dataset D T t r . Meta SAC can be summarized as optimizing the parameters θ and ψ in the learning process, and the optimization equation is shown in Equation (55).
min E T P T θ , ψ L D T t e s t , θ         s . t .   θ = μ ψ D T t r , θ
where D T t e s t and D T t r , respectively, represent training and testing missions sampled from p T , and L D T t e s t , θ represents the testing loss function. In the meta-training phase, parameters are optimized via the inner loop and outer loop.
In the inner loop, Meta SAC updates the model parameters with a small amount of randomly selected data for the specific mission T as the training data, reducing the loss of the model in mission T . In this part, the updating of model parameters is the same as in the original SAC algorithm, and the agent learns several scenes from randomly selected missions.
The minimum mean square error of strategy parameters θ corresponding to different missions in the inner loop phase is solved to obtain the initial strategy parameters θ i n i of the outer loop. In this manuscript, a hidden layer feature is added to the input part of strategy θ i n i as a random noise. The random noise is sampled again in each episode, in order to provide a more continuous random exploration in time, which is helpful for agent to adjust their overall strategy exploration according to the current mission MDP. The goal of meta-learning is to enable the agent to learn how to quickly adapt to new missions by simultaneously updating a small amount of gradient of strategy parameters and hidden layer features. Therefore, the θ of θ = μ ψ D T t r , θ includes not only parameters of the neural network, but also the distribution parameters of hidden variables of each mission, namely the mean and variance of the Gaussian distribution, as shown in Equation (56).
μ i = μ i + α μ μ i E α t π α t s t ; θ , z i , z i N μ i , σ i t R t s t σ i = σ i + α σ σ i E α t π α t s t ; θ , z i , z i N μ i , σ i t R t s t θ i = θ + α θ θ E α t π α t s t ; θ , z i , z i N μ i , σ i t R t s t
The model is represented by a parameterized function f θ with parameter θ , and when it is transferred to a new mission T , model parameter θ is updated to θ through gradient rise, as shown in Equation (57).
θ i = θ α θ L T i f θ
We update step α is a fixed super parameter. Model parameter θ is updated to maximize the performance f θ i of different missions, as shown in Equation (58).
min θ T i p T L T i f θ i = T i p T L T i f θ α θ L T i f θ
The meta-learning process of different missions is carried out through SGD, and the principle of θ is as follows:
θ = θ β θ T i p T L T i f θ i
where β is the meta update step.
In the meta-testing phase, a small amount of experience in new missions is used to quickly learn strategies for solving new missions. A new mission may involve completing a new mission goal or achieving the same mission goal in a new environment. The updating process of the model in this phase is the same as the cycle part in the meta-training phase, and by calculating the loss function with the data collected in the new mission and adjusting the model through back propagation, the new mission is adapted by the agent.

6. Simulation Analysis

In this section, the manuscript analyzes and verifies the escape guidance strategy based on Meta SAC. SAC is used to solve a specific escape guidance mission. We conduct comprehensive experiments to verify whether the UAV can complete the guidance escape mission under satisfying terminal and process constraints. Once the UAV guidance escape mission changes, the original strategy based on SAC cannot be easily adapted to the changed mission and thus needs to be relearned and retrained. The manuscript proposes an optimization method via meta-learning that improves the learning ability of UAVs during the training process. This section focuses on verifying the validation of Meta SAC and demonstrating its performance in various new missions. In addition, the maneuvering overload commands under different pursuit evading distances are analyzed in order to explore the influence of different maneuvering timings and distances on the escape results. We use CAV-H to verify the escape guidance performance. The initial conditions, terminal altitude, and Meta SAC training parameters are given in Table 1.

6.1. Validity Verification on SAC

In order to verify the effectiveness of SAC, three different pursuit evading scenarios are constructed, and the terminal reward value, miss distance, and terminal position deviation are analyzed. As shown in Figure 6a, the terminal reward value is poor in the initial phase of training, which demonstrates that the optimal strategy is not found. After 500 episodes, the terminal reward value increases gradually, indicating that a better strategy has been explored and converged. In the last 100 episodes, the optimal strategy is trained and learned, while the network parameters are adjusted to the optimal strategy. As can be seen from Figure 6b, the miss distance is relatively divergent in the first 150 episodes of training, indicating that the Action network in SAC constantly explores new strategies, and the Critic network also learns scientific evaluation criteria. After 500 training episodes, the network gradually learns and trains in the direction of optimal solution. The miss distance at the encounter moment converges to about 20 m. As shown in Figure 6c, the terminal position of the UAV has a large deviation in the early training phase, which is attributed to the exploration of the escape strategy by the network. In the later training phase, the position deviation is less affected by exploration. These pursuit evading scenarios tested in the manuscript can achieve convergence, and the final convergence values are all within 1 m.
In order to verify whether the SAC algorithm can solve the escape guidance strategy that meets the mission requirements in different pursuit and evasion scenarios, the pursuing and evading distance is changed, and the training results are shown in Figure 7. In the medium-range scenario, the miss distance converges to about 2 m, and the terminal deviation converges to about 1 m.
As shown in Figure 8, in long-range attack and defense scenarios, the miss distance converges to about 5 m, and the terminal deviation converges to about 1 m.
Based on the above simulation analysis, SAC is a feasible method to solve the UAV guidance escape strategy. After limited episodes of learning and training, network parameters are converged, which is used to test the flight mission.

6.2. Validity Verification on Meta SAC

When the mission of the UAV changes, the original SAC parameters cannot meet the requirements of the new mission, and thus the parameters need to be retrained and relearned. The SAC proposed in the manuscript is improved via meta-learning. Strong adaptive network parameters are found using learning and training, and when the pursuit evading environment changes, the network parameters are fine-tuned to adapt to the new environment immediately.
Meta SAC is divided into a meta-training phase and a meta-testing phase, and initialization parameters for the SAC network are trained in the meta-training phase, which is fine-tuned by interacting with the new environment in the meta-testing phase. By changing the initial interceptor position, three different pursuit evading scenarios are constructed, which represent a short, medium, and long distance.
The training results of Meta SAC and SAC are compared, and terminal reward values are presented in Figure 9a. Meta SAC is an effective method to speed up the training process, and after 100 episodes, a better strategy is learned by the network and converged gradually. In contrast, the SAC network needs 500 episodes to find the optimal solution. Miss distance is shown in Figure 9b. The better strategy is quickly learned by Meta SAC, which is more effective than the SAC method. Figure 9c shows the terminal deviation between the UAV and the target.
To explore the optimal solution as much as possible, some strategies with large terminal position deviations appear in the training process. As shown in Figure 10b,c, in medium-range attack and defense scenarios, the miss distance converges to about 8 m based on Meta SAC, and the terminal deviation converges to about 1 m.
As shown in Figure 11b,c, in long-range attack and defense scenarios, the miss distance converges to about 10 m based on Meta SAC, and the terminal deviation converges to about 1 m.
According to the theoretical analysis, in the training process, new missions corresponding to the same distribution are used to execute micro-testing using Meta SAC, resulting in more gradient-descending directions of the optimal solution being learned by the network. Combined with the theory analysis and training results, the manuscript demonstrates that meta-learning is a feasible method to accelerate convergence and improve the efficiency of training.
In the previous analysis, when the pursuit evading scenario is changed, network parameters obtained in the meta-training phase are fine-tuned through a few interactions. The manuscript verifies meta testing performance by changing the initial interceptor position, and results compared with the SAC method are shown in Table 2. Based on the network parameters of the meta-training phase, the strategic solutions for the escape guidance missions are found through training within 10 episodes. On the contrary, network parameters based on SAC need more interaction to find solutions, and the episode of interactions comprises more than 50 episodes. According to the above simulation, the adaptability of Meta SAC is much greater than SAC, and once the escape mission changes, through very few episodes of learning, the new mission is completed by the UAV without re-learning and designing the strategy. The method provides the possibility of realizing UAV learning while flying.

6.3. Strategy Analysis Based on Meta SAC

This section tests the network parameters based on Meta SAC, and analyzes the escape strategy and flight state under different pursuit evading distances. As shown in Figure 12a, for the pursuit evading scene over a short distance, the longitudinal maneuvering overload is larger in the first half of the phase of escape, resulting in the velocity slope angle decreasing gradually. In the second half of the phase of escape, if the strategy is executed under the original maneuvering overload, the terminal altitude constraint cannot be satisfied, and therefore, the overload gradually decreases, and the velocity slope angle is slowly reduced. As shown in Figure 12b, at the beginning of escape, the lateral maneuvering overload is positive, and the velocity azimuth angle constantly increases. With the distance between the UAV and the interceptor reducing, the overload increases gradually in the opposite direction, and the velocity azimuth angle decreases. On the one hand, this can confuse the strategy of the interceptor, while on the other hand, the guidance course is corrected.
As shown in Figure 13a, compared with the pursuit evading scene of a short distance, the medium-distance escape process takes longer, the pursuing time left to the interceptor is longer, and the UAV flies in the direction of increasing the velocity slope angle. The timing of the maximum escape overload corresponding to the medium distance is also different. As shown in Figure 13b, in the first half of the phase of escape, lateral maneuvering overload corresponding to a medium distance is larger than that in a short distance, and in the second half of the phase of escape, the corresponding reverse maneuvering overload is smaller, resulting in the UAV being able to use the longer escape time to slowly correct the course.
As shown in Figure 14, under a long pursuit distance, the overload change of the UAV maneuver is similar to that for a medium range, and the escape timing is basically the same as the escape strategy.
According to the above analysis, the escape guidance strategy via Meta SAC can be used as a tactical escape strategy, and the timing of escape and the maneuvering overload are adjusted in a timely manner under different pursuit evading distances. On the one hand, the overload corresponding to this strategy can confuse the interceptor and cause some interference, and on the other hand, it can take into account the guidance mission, correcting the course deviation caused by escape.
Figure 15a shows the flight trajectory of the interceptor against UAV at the North East Down (NED) coordinate (10 km, 30 km, 30 km), the trajectory point at the encountering moment is shown in Figure 15b, and the miss distance is 19 m in this pursuit evading scene. To verify the scientific and applicability of Meta SAC, the initial position of interceptor is changed. Flight trajectories are shown in Figure 15c,e, and trajectory points at the encountering moment are shown in Figure 15d,f. The miss distances in these two pursuit evading scenarios are 3 m and 6 m, respectively. Based on the CAV-H structure, the miss distance between the UAV and the interceptor is greater than 2 m at the encountering moment, which means that the escape mission is achieved.
Based on the principles of Meta SAC and optimal guidance, flight states are shown in Figure 16. Longitude, latitude, and altitude during the flight of the UAV are shown in Figure 16a,b, under different pursuit evading scenarios, and the terminal position and altitude constraints are met. There is larger amplitude modification in the velocity slope and azimuth angle, which is attributed to the escape strategy via lateral and longitudinal maneuvering, as shown in Figure 16c,d. The total change in velocity slope and azimuth angle is within two degrees, which meets the flight process constraints. Through the analysis of flight states, this escape strategy is an effective measure for guidance and escape with high accuracy.
Flight process deviation mainly includes aerodynamic calculation deviation and output overload deviation. For the aerodynamic deviation, this manuscript uses the interpolation method to calculate it based on the flight Mach number and angle of attack, which may have some deviation. Therefore, when calculating the aerodynamic coefficient, random noise with an amplitude of 0.1 is added to verify whether the UAV can complete the guidance mission. As shown in Figure 17a, aerodynamic deviation noise causes certain disturbances to the angle of attack during the flight. At the 10th second and end of the flight, the maximum deviation of the angle of attack is 2°. However, overall, the impact of aerodynamic deviation on the entire flight is relatively small, and the change in angle of attack is still within the safe range of the UAV. As shown in Figure 17b, due to the constraints of UAV game confrontation and guidance missions, the bank angle during the entire flight process changes significantly, and aerodynamic deviation noise has a small impact on the bank angle. After increasing the aerodynamic deviation noise, the miss distance between the UAV and the interceptor at the time of encounter is 8.908 m, and the terminal position deviation is 0.52 m. Therefore, under the influence of aerodynamic deviation, the UAV can still complete the escape guidance mission.
For the output overload deviation, the total overload is composed of the guiding overload derived from the optimal guidance law and the maneuvering overload output derived from the neural network. Random maneuvering overload with an amplitude of 0.1 is added to verify whether the UAV can complete the maneuver guidance mission. As shown in Figure 18, random overloads are added in the longitudinal and lateral directions, respectively. Through simulation testing, the miss distance between the UAV and the interceptor at the encounter point is 10.51 m, and the terminal deviation of the UAV is 0.6 m. Under this deviation, the UAV can still achieve high-precision guidance and efficient penetration.

7. Conclusions

This manuscript proposes an escape guidance strategy satisfying multiple terminal constraints via SAC. The action space is designed under the UAV process constraint. Considering the fact that real-time interceptor information is hard to obtain in the actual escape process, the state space considering the target and the heading angle is designed. The reward function is an important index function that is used to guide and evaluate the training results. Based on the pursuit evading model, the terminal LOS angle rates in the lateral and longitudinal directions are derived to describe the deviation and miss distance. In order to improve the adaptability of the escape guidance strategy, the manuscript improves SAC via meta-learning and compares meta SAC with SAC. The strong adaptive escape strategy based on Meta SAC is analyzed. In view of the above theoretical numerical analysis, we have obtained the following conclusions:
(1)
The escape guidance strategy based on SAC is a feasible tactical escape strategy that can achieve high-precision guidance while meeting the escape requirements.
(2)
Meta SAC can significantly improve the adaptability of the escape strategy. When the escape mission changes, it can fine-tune network parameters to adapt to the mission through a small number of training epochs. This method provides the possibility for the UAV to learn while flying.
(3)
The strong adaptive escape strategy based on Meta SAC can adjust the escape timing and maneuver overload in real time according to the pursuit evading distance. On the one hand, the overload corresponding to this strategy can confuse the interceptor and cause some interference, while on the other hand, it can take into account the guidance mission, correcting the course deviation caused by escape.

Author Contributions

Conceptualization, S.Z. and J.Z.; methodology, S.Z.; software, S.Z; validation, S.Z. and J.Z.; formal analysis, S.Z; investigation, S.Z.; resources, X.L.; data curation, H.S.; writing—original draft preparation, H.S.; writing—review and editing, X.L.; visualization, S.Z.; supervision, W.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Z.B.; Huang, W.N.; Zhang, Z.J. Summary of the Hot Spots of Near Space Vehicles in 2018. Sci. Technol. Rev. 2019, 37, 44. [Google Scholar]
  2. Li, G.H.; Zhang, H.B.; Tang, G.J. Maneuver characteristics analysis for hypersonic glide vehicles. Aerosp. Sci. Technol. 2015, 43, 321–328. [Google Scholar] [CrossRef]
  3. Wang, L.L.; Lan, Y.B.; Zhang, Y.L.; Zhang, H.H.; Tahir, M.N.; Ou, S.C.; Liu, X.T.; Chen, P.C. Applications and prospects of agricultural unmanned aerial vehicle obstacle avoidance technology in China. J. Sens. 2019, 19, 642. [Google Scholar] [CrossRef]
  4. Wang, Y.F.; Zhou, T.; Chen, W.C.; He, T.J. Optimal maneuver penetration strategy based on power series solution of miss distance. J. Beijing Univ. Aeronaut. Astronaut. 2020, 46, 159. [Google Scholar]
  5. Rim, J.W.; Koh, I.S. Survivability simulation of airborne platform with expendable active decoy countering RF missile. IEEE Trans. Aerosp. Electron. Syst. 2019, 56, 196–207. [Google Scholar] [CrossRef]
  6. Liu, F.; Dong, X.W.; Li, Q.D.; Ren, Z. Robust multi-agent differential games with application to cooperative guidance. Aerosp. Sci. Technol. 2021, 111, 106568. [Google Scholar]
  7. Garcia, E.; Casbeer, D.W.; Pachter, M. Design and analysis of state-feedback optimal strategies for the differential game of active defense. IEEE Trans. Autom. Control. 2018, 64, 553–568. [Google Scholar] [CrossRef]
  8. Liang, H.Z.; Wang, J.Y.; Wang, Y.H.; Wang, L.L.; Liu, P. Optimal guidance against active defense ballistic missiles via differential game strategies. Chin. J. Aeronaut. 2020, 33, 978–989. [Google Scholar] [CrossRef]
  9. Liang, L.; Deng, F.; Peng, Z.H.; Li, X.X.; Zha, W.Z. A differential game for cooperative target defense. Automatica 2019, 102, 58–71. [Google Scholar]
  10. Liu, S.; Wang, Y.; Li, Y.; Yan, B.; Zhang, T. Cooperative guidance for active defence based on line-of-sight constraint under a low-speed ratio. Aeronaut. J. 2023, 127, 491–509. [Google Scholar] [CrossRef]
  11. Zhang, D.H.; Zhang, T.Y.; Lu, Y.P.; Zhu, Z.X.; Dong, B. You only propagate once: Accelerating adversarial training via maximal principle. In Proceedings of the Advances in Neural Information Processing Systems 32 (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; p. 32. [Google Scholar]
  12. Ruthotto, L.; Osher, S.J.; Li, W.C.; Nurbekyan, L.; Fung, S.W. A machine learning framework for solving high-dimensional mean field game and mean field control problems. Proc. Natl. Acad. Sci. USA 2020, 117, 9183–9193. [Google Scholar]
  13. Ullah, Z.; Al-Turjman, F.; Mostarda, L.; Gagliardi, R. Applications of artificial intelligence and machine learning in smart cities. Comput. Commun. 2020, 154, 313–323. [Google Scholar]
  14. Song, H.; Bai, J.N.; Yi, Y.; Wu, J.S.; Liu, L.J. Artificial intelligence enabled Internet of Things: Network architecture and spectrum access. IEEE Comput. Intell. Mag. 2020, 15, 44–51. [Google Scholar] [CrossRef]
  15. Gong, X.P.; Chen, W.C.; Chen, Z.Y. All-aspect attack guidance law for agile missiles based on deep reinforcement learning. Aerosp. Sci. Technol. 2022, 127, 107677. [Google Scholar]
  16. Furfaro, R.; Scorsoglio, A.; Linares, R.; Massari, M. Adaptive generalized ZEM-ZEV feedback guidance for planetary landing via a deep reinforcement learning approach. Acta Astronaut. 2020, 171, 156–171. [Google Scholar] [CrossRef]
  17. Yuan, Y.; Zheng, G.; Wong, K.K.; Letaief, K.B. Meta-reinforcement learning based resource allocation for dynamic V2X communications. IEEE Trans. Veh. Technol. 2021, 70, 8964–8977. [Google Scholar] [CrossRef]
  18. Lu, Z.Y.; Wang, X.Y.; Gursoy, M.C. Trajectory Design for Unmanned Aerial Vehicles via Meta-Reinforcement Learning. In Proceedings of the IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Hoboken, NJ, USA, 20 May 2023; pp. 1–6. [Google Scholar]
  19. Hu, Y.; Chen, M.Z.; Saad, W.; Poor, H.V.; Cui, S.G. Meta-reinforcement learning for trajectory design in wireless UAV networks. In Proceedings of the 2020 IEEE Global Communications Conference, Taipei, Taiwan, 7–11 December 2020; pp. 1–6. [Google Scholar]
  20. Yu, Q.Y.; Luo, L.K.; Liu, B.; Hu, S.Q. Re-planning of Quadrotors Under Disturbance Based on Meta Reinforcement Learning. J. Intell. Robot. Syst. 2023, 107, 13. [Google Scholar] [CrossRef]
  21. Belkhale, S.; Li, R.; Kahn, G.; McAllister, R.; Calandra, R.; Levine, S. Model-based meta-reinforcement learning for flight with suspended payloads. IEEE Robot. Autom. Lett. 2021, 6, 1471–1478. [Google Scholar] [CrossRef]
  22. Zhao, L.B.; Xu, W.; Dong, C.; Zhu, G.S.; Zhuang, L. Evasion guidance of re-entry vehicle satisfying no-fly zone constraints based on virtual goals. Sci. Sin. Phys. Mech. Astron. 2021, 51, 104706. [Google Scholar] [CrossRef]
  23. Guo, Y.H.; Li, X.; Zhang, H.J.; Wang, L.L.; Cai, M. Entry Guidance With Terminal Time Control Based on Quasi-Equilibrium Glide Condition. IEEE Trans. Aerosp. Electron. Syst. 2020, 56, 887–896. [Google Scholar] [CrossRef]
  24. Krasner, S.M.; Bruvold, K.; Call, J.A.; Fieseler, P.D.; Klesh, A.T.; Kobayashi, M.M.; Lay, N.E.; Lim, R.S.; Morabito, D.D.; Oudrhiri, K.; et al. Reconstruction of Entry, Descent, and Landing Communications for the InSight Mars Lander. J. Spacecr. Rocket. 2021, 58, 1569–1581. [Google Scholar] [CrossRef]
  25. Huang, Z.C.; Zhang, Y.S.; Liu, Y. Research on state estimation of hypersonic glide vehicle. J. Phys. Conf. Ser. 2018, 1060, 012088. [Google Scholar] [CrossRef]
  26. Zhu, J.; Su, D.; Xie, Y.; Sun, H. Impact time and angle control guidance independent of time-to-go prediction. Aerosp. Sci. Technol. 2019, 86, 818–825. [Google Scholar] [CrossRef]
  27. Ni, C.; Zhang, A.R.; Duan, Y.; Wang, M. Learning Good State and Action Representations via Tensor Decomposition. In Proceedings of the IEEE International Symposium on Information Theory, Melbourne, Australia, 12–20 July 2021; pp. 1682–1687. [Google Scholar]
  28. Ma, Y.; Wang, Z.; Castillo, I.; Rendall, R.; Bindlish, R.; Ashcraft, B.; Bentley, D.; Benton, M.G.; Romagnoli, J.A.; Chiang, L.H. Reinforcement Learning-Based Fed-Batch Optimization with Reaction Surrogate Model. In Proceedings of the 2021 American Control Conference (ACC), New Orleans, LA, USA, 25–28 May 2021; pp. 2581–2586. [Google Scholar]
  29. Yang, X.; Yang, X.; Yang, J.; Ming, Q.; Wang, W.; Tian, Q.; Yan, J. Learning high-precision bounding box for rotated object detection via kullback-leibler divergence. Adv. Neural Inf. Process. Syst. 2021, 34, 18381–18394. [Google Scholar]
Figure 1. Attack–defense model.
Figure 1. Attack–defense model.
Drones 07 00626 g001
Figure 2. Guidance penetration strategy.
Figure 2. Guidance penetration strategy.
Drones 07 00626 g002
Figure 3. The attack–defense confrontation model in the lateral direction.
Figure 3. The attack–defense confrontation model in the lateral direction.
Drones 07 00626 g003
Figure 4. The attack–defense confrontation model in longitudinal direction.
Figure 4. The attack–defense confrontation model in longitudinal direction.
Drones 07 00626 g004
Figure 5. Updating principle of SAC.
Figure 5. Updating principle of SAC.
Drones 07 00626 g005
Figure 6. Train results of SAC of the short-range scenario. (a) Reward value, (b) miss distance, (c) target deviation.
Figure 6. Train results of SAC of the short-range scenario. (a) Reward value, (b) miss distance, (c) target deviation.
Drones 07 00626 g006
Figure 7. Train results of SAC of the medium-range scenario. (a) reward value, (b) miss distance, (c) target deviation.
Figure 7. Train results of SAC of the medium-range scenario. (a) reward value, (b) miss distance, (c) target deviation.
Drones 07 00626 g007
Figure 8. Train results of SAC of the long-range scenario. (a) reward value, (b) miss distance, (c) target deviation.
Figure 8. Train results of SAC of the long-range scenario. (a) reward value, (b) miss distance, (c) target deviation.
Drones 07 00626 g008
Figure 9. Meta SAC training results of the short-range scenario. (a) Reward value, (b) miss distance, (c) target deviation.
Figure 9. Meta SAC training results of the short-range scenario. (a) Reward value, (b) miss distance, (c) target deviation.
Drones 07 00626 g009
Figure 10. Meta SAC training results of the medium-range scenario. (a) Reward value, (b) miss distance, (c) target deviation.
Figure 10. Meta SAC training results of the medium-range scenario. (a) Reward value, (b) miss distance, (c) target deviation.
Drones 07 00626 g010
Figure 11. Meta SAC training results of the long-range scenario. (a) Reward value, (b) miss distance, (c) target deviation.
Figure 11. Meta SAC training results of the long-range scenario. (a) Reward value, (b) miss distance, (c) target deviation.
Drones 07 00626 g011
Figure 12. The maneuvering overload. (a) Longitudinal direction under a short distance, (b) lateral direction under a short distance.
Figure 12. The maneuvering overload. (a) Longitudinal direction under a short distance, (b) lateral direction under a short distance.
Drones 07 00626 g012
Figure 13. The maneuvering overload. (a) Longitudinal direction under a medium distance, (b) lateral direction under a medium distance.
Figure 13. The maneuvering overload. (a) Longitudinal direction under a medium distance, (b) lateral direction under a medium distance.
Drones 07 00626 g013
Figure 14. The maneuvering overload. (a) Longitudinal direction under a long distance, (b) lateral direction under a long distance.
Figure 14. The maneuvering overload. (a) Longitudinal direction under a long distance, (b) lateral direction under a long distance.
Drones 07 00626 g014
Figure 15. The ballistic flight diagrams of the whole course under different pursuit evading distances. (b,d,f) represent trajectories at the encountering time under different pursuit evading distances.
Figure 15. The ballistic flight diagrams of the whole course under different pursuit evading distances. (b,d,f) represent trajectories at the encountering time under different pursuit evading distances.
Drones 07 00626 g015
Figure 16. Flight states of the UAV. (a) The longitude and latitude, (b) the height, (c) the velocity slope angle, (d) the velocity azimuth angle.
Figure 16. Flight states of the UAV. (a) The longitude and latitude, (b) the height, (c) the velocity slope angle, (d) the velocity azimuth angle.
Drones 07 00626 g016
Figure 17. (a) The attack angle, (b) the bank angle.
Figure 17. (a) The attack angle, (b) the bank angle.
Drones 07 00626 g017
Figure 18. (a) The maneuvering overload in the longitudinal direction, (b) the maneuvering overload in the lateral direction.
Figure 18. (a) The maneuvering overload in the longitudinal direction, (b) the maneuvering overload in the lateral direction.
Drones 07 00626 g018
Table 1. Simulation and Meta SAC training conditions.
Table 1. Simulation and Meta SAC training conditions.
SchemeMeta SAC Training Parameters
UAV initial velocity4000 m/sLearning episodes1000
Initial velocity Inclination Guidance period0.1 s
Initial velocity azimuthData sampling interval30 Km
Initial position(3° E, 1° N)Discount factor γ = 0.99
Initial altitude45 KmSoft update tau0.001
Terminal altitude 40 KmLearning rate0.005
Target position (0° E, 0° N)Sampling size for each train128
Interceptor Initial velocity1500 m/sNet layers 2
Initial velocity Inclination Longitudinal LOS angleNet nodes256
Initial velocity azimuthLateral LOS angleCapacity of experience pool20,000
Table 2. Results compared with the SAC method.
Table 2. Results compared with the SAC method.
Interceptor
Initial Position (Km)
Interaction EpisodesMiss Distance (m)Terminal Deviate (m)
SACMeta SACSACMeta SACSACMeta SAC
(0, 30, 0)7413.783.290.560.61
(2, 30, 6)7542.802.720.680.72
(4, 30, 12)5986.933.750.690.58
(6, 30, 18)5912.716.820.680.72
(8, 30, 24)2623.163.700.470.50
(10, 30, 30)5833.502.370.610.64
(12, 30, 36)6712.862.210.680.45
(14, 30, 42)5682.182.890.550.61
(16, 30, 48)6912.732.230.610.72
(18, 30, 54)10612.453.710.560.63
(20, 30, 60)9412.72.350.490.54
(22, 30, 66)5912.232.510.730.71
(24, 30, 72)6212.113.470.480.67
(26, 30, 78)6312.044.50.480.57
(28, 30, 84)6342.645.120.470.40
(30, 30, 90)6392.956.050.680.47
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, S.; Zhu, J.; Bao, W.; Li, X.; Sun, H. A Multi-Constraint Guidance and Maneuvering Penetration Strategy via Meta Deep Reinforcement Learning. Drones 2023, 7, 626. https://doi.org/10.3390/drones7100626

AMA Style

Zhao S, Zhu J, Bao W, Li X, Sun H. A Multi-Constraint Guidance and Maneuvering Penetration Strategy via Meta Deep Reinforcement Learning. Drones. 2023; 7(10):626. https://doi.org/10.3390/drones7100626

Chicago/Turabian Style

Zhao, Sibo, Jianwen Zhu, Weimin Bao, Xiaoping Li, and Haifeng Sun. 2023. "A Multi-Constraint Guidance and Maneuvering Penetration Strategy via Meta Deep Reinforcement Learning" Drones 7, no. 10: 626. https://doi.org/10.3390/drones7100626

APA Style

Zhao, S., Zhu, J., Bao, W., Li, X., & Sun, H. (2023). A Multi-Constraint Guidance and Maneuvering Penetration Strategy via Meta Deep Reinforcement Learning. Drones, 7(10), 626. https://doi.org/10.3390/drones7100626

Article Metrics

Back to TopTop