Next Article in Journal
Review of Alternative Sustainable Fuels for Hybrid Rocket Propulsion
Previous Article in Journal
Active Flow Control for Passage Vortex Reduction in a Linear Turbine Cascade with Various Tip Clearance Sizes Using a Dielectric Barrier Discharge Plasma Actuator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration

School of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Aerospace 2023, 10(7), 642; https://doi.org/10.3390/aerospace10070642
Submission received: 19 June 2023 / Revised: 12 July 2023 / Accepted: 14 July 2023 / Published: 16 July 2023
(This article belongs to the Section Aeronautics)

Abstract

:
The penetration of unmanned aerial vehicles (UAVs) is an important aspect of UAV games. In recent years, UAV penetration has generally been solved using artificial intelligence methods such as reinforcement learning. However, the high sample demand of the reinforcement learning method poses a significant challenge specifically in the context of UAV games. To improve the sample utilization in UAV penetration, this paper innovatively proposes an improved sampling mechanism called task completion division (TCD) and combines this method with the soft actor critic (SAC) algorithm to form the TCD-SAC algorithm. To compare the performance of the TCD-SAC algorithm with other related baseline algorithms, this study builds a dynamic environment, a UAV game, and conducts training and testing experiments in this environment. The results show that among all the algorithms, the TCD-SAC algorithm has the highest sample utilization rate and the best actual penetration results, and the algorithm has a good adaptability and robustness in dynamic environments.

1. Introduction

UAV penetration refers to the process in which UAVs change their original flight trajectory to avoid detection and interception by an enemy, thereby successfully infiltrating the enemy’s defense. The ability to penetrate enemy defenses is crucial for UAV operations. However, the current research on autonomous UAV penetration is still limited, and there is a need to enhance the effectiveness of UAV penetration. This paper is dedicated to studying the issue of autonomous UAV penetration.
Previous studies on UAV penetration have primarily focused on constraints such as the position, speed, and overload of the re-entry point [1,2,3]. They have also examined performance measures like the heading error of the UAV [4] and analyzed the miss distance of interceptors to discuss UAV penetration strategies [5]. However, these traditional methods are primarily intended for planned trajectories before a launch and cannot be applied to real-time autonomous penetration after a launch.
In recent years, artificial intelligence methods represented by reinforcement learning (RL) have gradually been applied to UAV penetration. RL offers the advantage of enabling online autonomous penetration, which is not feasible with conventional approaches. The RL method iteratively updates and improves a pre-designed reward function, using a large number of samples to obtain the optimal policy for UAV penetration [6]. However, the significant sample requirements of RL pose time and cost challenges in the context of UAV penetration [7,8].
In order to improve the sample utilization in UAV penetration, this paper proposes a reinforcement learning method based on an improved sampling mechanism. The structure of this paper is as follows. Firstly, UAV penetration is modeled as a reinforcement learning problem, and the corresponding action space, state space, and reward function are designed. Then, according to the decomposability of the task in UAV penetration, a task completion division (TCD) method is proposed to improve the sample utilization. Next, the TCD method is combined with the soft actor critic (SAC) algorithm, so as to obtain the TCD-SAC algorithm. Finally, the actual performance of the TCD-SAC algorithm in UAV penetration is analyzed through training and test experiments.

2. Related Works

The field studied in this paper belongs to UAV penetration, and algorithms in this field can be roughly divided into two categories. (1) Based on the swarm intelligence optimization algorithm. For example, Jing Luo [9] proposed a three-dimensional path planning algorithm based on improved holonic particle swarm optimization (IHPSO) to meet the requirements of safety, concealment, and timeliness of trajectory planning during the process of UAV penetration; Jinyu Fu [10] proposed a trajectory homotopy optimization framework for multiple UAVs, which solved the problem of dynamic penetration mission planning in complex environments; Zhe Zhang [11] proposed a penetration strategy based on an improved A-Star algorithm to solve the replanning problem of stealth UAVs in a three-dimensional complex dynamic environment. (2) Based on the reinforcement learning network. For example, Yuxie Luo [12] established an intelligent UAV model using the reinforcement learning network and built a reward function using the cooperative parameters of multiple UAVs to guide UAVs to conduct collaborative penetration; Yue Li [13] used reinforcement learning algorithms for training in four scenarios, frontal attack, escape, pursuit, and energy storage, thereby improving the intelligent decision-making level of air confrontation; Kaifang Wan [14] proposed a motion control method based on deep reinforcement learning (DRL), which provides additional flexibility for UAV penetration within the DRL framework; Liang Li [15] mainly focused on the winning region of three players in the reconnaissance penetration game, proposed an explicit policy method for analyzing and constructing barriers, and provided a complete solution by integrating the games of kind and degree.
In general, the swarm intelligence optimization method needs a long time period to update and iterate the final strategy, so it is not suitable for online real-time decision making. On the contrary, the reinforcement learning network has a strong data processing capability and the ability to make rapid decisions, which can meet the real-time online decision making of UAVs. Therefore, the reinforcement learning network is more suitable for the online autonomous penetration of UAVs.
Compared with the above references, the innovation of this paper lies in solving the problem of the high sample demand when RL is applied to UAV penetration. In addition, the related works often use the simplified three degrees of freedom particle model, while this study uses the six degrees of freedom aircraft model, so the results of the algorithm are closer to the application effect in the actual environment.
Specifically, this paper innovatively proposes an improved sampling mechanism called task completion division (TCD) and combines the TCD method with the soft actor critic (SAC) algorithm to form the TCD-SAC algorithm, designed to improve the sample utilization in UAV penetration. In addition, the feasibility of the algorithm is verified through a series of training and testing experiments.

3. Problem Description

3.1. UAV Penetration

UAV penetration refers to the penetration of drones into the opponent’s defense during the game process. The application scenario of this paper is shown in Figure 1. First, the UAV is launched and continues to fly according to the predetermined trajectory to hit the target. However, the other party finds and tracks the UAV during the flight, launching an interceptor to intercept the UAV. When the UAV finds the interceptor, it maneuvers to penetrate the defense, thereby flying according to the new trajectory and finally hitting the target. The interceptor is a high-speed antimissile missile with guidance capability.
In the above process, the method in this paper will play a role in the process of UAV penetration, and its task is to guide the UAV to penetrate the defense of the interceptor and successfully hit the target. This method of UAV penetration will output maneuvering commands, so as to control the UAV to complete the task of autonomous penetration and a target hit.

3.2. Reinforcement Learning Description of UAV Penetration

The next state of the UAV penetration environment is related to the current state and the current action, so the process of UAV penetration can be modeled as a Markov decision process (MDP) [16]. Therefore, the problem of UAV penetration can be solved using the reinforcement learning algorithm.
Before research on the reinforcement learning algorithm is carried out, a six degrees of freedom model of UAV should be built, so as to conduct subsequent training and test experiments. The UAV model built in this study is a hypersonic glider based on real parameters [17]. To facilitate calling, this study converts the model into a dynamic link library, which can also reduce the CPU usage and improve the computational efficiency to a certain extent.
For the control system of UAV, the design methods of the control system mainly include an attitude control method and overload control method [18]. The hypersonic glider built in this study is a complex aircraft with a large lift/drag ratio, which means that the attitude control method will have a better stability margin and reliability [19]. Therefore, this study adopts the attitude control method and selects the angles of attack and bank as the attitude control command to control the UAV.
The relationship between UAV penetration and reinforcement learning is shown in Figure 2. The reinforcement learning algorithm sends the attitude control command to the UAV, and the UAV receives and executes the command to penetrate the interceptor, which intercepts the UAV at the same time. The adversarial environment composed of the UAV and interceptor outputs the current state and reward. The reinforcement learning algorithm receives the current state and reward and updates the internal parameters of the algorithm accordingly to better output the next action command.
During the MDP process of UAV penetration, the action vector a t , state vector s t , and reward function r t of the penetration environment need to be determined. The specific design method is as follows.

3.2.1. State and Action Specification

The input of the control system is the attitude control command, so the angles of attack α and bank γ V are selected as the action vector of the reinforcement learning network. Therefore, the action vector a t can be expressed as Formula (1).
a t = α , γ V
In addition, to ensure flight stability, the ranges of attack angle α and bank angle γ V are 20 ° ,   20 ° and 60 ° ,   60 ° , respectively.
The variables related to the observation state are shown in Figure 3. The red icon represents our UAV, the blue icon represents the opponent’s interceptor, the yellow icon represents the target to be hit, and the red dashed line represents the trajectory of our UAV. This paper defines the relative distance between the UAV and interceptor as R 1 , the line-of-sight angle of the UAV relative to the interceptor as q 1 , and the relative distance between the UAV and target as R 2 .
In order to complete the task of autonomous penetration and hitting the target, the algorithm needs to obtain the relative distance and relative angle of the UAV relative to the interceptor and target. That is to say, the state information to be obtained includes the relative distance between the UAV and interceptor ( R 1 ), the line-of-sight angle of the UAV relative to the interceptor ( q 1 ), and the relative distance between the UAV and target ( R 2 ), so a three-dimensional state space is built, and its corresponding state vector s t is shown as Formula (2).
s t = R 1 , q 1 , R 2
In actual application, the state vector s t will be used as the input of the reinforcement learning network, and the network will output the corresponding action vector a t based on the state vector s t .

3.2.2. Reward Shaping

The reward function plays a vital role in reinforcement learning, and it will be used as the optimal target function for subsequent training. The environment for UAV penetration has high-dimensional and continuous characteristics, so the exploration space of the reinforcement learning algorithm is huge [20]. If there is only a main reward, it is difficult to obtain positive samples [21,22] (successful samples of autonomous penetration and a target hit). Therefore, it is considered to add a process reward in reward shaping to solve the problem of convergence difficulty.
The primary objective of this algorithm is to guide the UAV to penetrate the defense system and successfully strike the target. Consequently, the design of the reward function focuses on two key aspects:
(a).
Guiding the UAV through interception defenses
Guiding the UAV through interception defenses
When the UAV encounters interception measures, a greater relative distance R 1 between the two aircrafts increases the likelihood of evading the interceptor and achieving a higher success rate of penetration. Therefore, the reward function should assign a correspondingly higher value. The corresponding reward function is shown in Formula (3).
r t = a 1 e b 1 × R 1
where a 1 > 0 ,   b 1 > 0 , and r t is the increasing function of R 1 .
Additionally, a larger line-of-sight angle q 1 between the UAV and the interceptor also enhances the success rate of penetration. Thus, the reward function should give a more significant weight to such scenarios. The corresponding reward function is shown in Formula (4).
r t = a 2 arctan b 2 × q 1
where a 2 > 0 , b 2 > 0 , and r t is the increasing function of q 1 .
(b).
Guiding the UAV to hit the target after penetration
Once the successful penetration through the defense system is confirmed, the algorithm must direct the UAV towards the designated target to accomplish the mission of striking the target. During this phase, a shorter relative distance R 2 between the UAV and the target corresponds to a higher success rate of hitting the target. Therefore, the reward function should assign a greater value to scenarios where the UAV is in closer proximity to the target. The corresponding reward function is shown in Formula (5).
r t = a 3 e b 3 × R 2
where a 3 > 0 , b 3 < 0 , and r t is the decreasing function of R 2 .
According to the above analysis, the pseudo code of the reward function is outlined in Function 1.
Function 1: Reward Function of UAV Penetration.
r 1 = r 2 = r 3 = r 4 = r 5 = r 6 = r 7 = 0
r t = r 1 + r 2 + r 3 + r 4 + r 5 + r 6 + r 7
If the UAV is intercepted:
    r 1 = a 1 e b 1 × R 1 , where a 1 > 0 ,   b 1 > 0 , r 1 is the increasing function of R 1
    r 2 = a 2 arctan b 2 × q 1 , where a 2 > 0 , b 2 > 0 , r 2 is the increasing function of q 1
If the UAV penetrates successfully:
    r 3 = c 1 , where c 1 > 0 , r 3 is the instant reward for successful penetration
If the UAV fails to penetrate:
    r 4 = c 2 , where c 2 < 0 , r 4 is the immediate penalty for failed penetration
If the UAV successfully penetrates and before landing:
    r 5 = a 3 e b 3 × R 2 , where a 3 > 0 , b 3 < 0 , r 5 is the decreasing function of R 2
If the UAV lands and hits the target:
    r 6 = c 3 , where c 3 > 0 , r 6 is the instant reward for hitting the targetIf the UAV lands and misses the target:
    r 7 = c 4 , where c 4 < 0 , r 7 is the immediate penalty for missing the target
r t = r 1 + r 2 + r 3 + r 4 + r 5 + r 6 + r 7
To facilitate the convergence of the algorithm, this paper sets the value of the reward within the range of [−1, 1]. The specific values for the theoretical parameters mentioned above are assigned based on the actual performance parameters of the UAV model constructed.
This paper divides the UAV’s flight phase into several distinct phases. The phase in which the UAV is not intercepted by the interceptor is referred to as Phase 1. Phase 2 represents the phase in which the UAV is being intercepted. Phase 3 corresponds to the phase in which the UAV successfully penetrates the interceptor. Phase 4 denotes the phase in which the UAV fails to penetrate. Phase 5 represents the phase after successful penetration and before landing. Phase 6 represents the phase of landing and successfully hitting the target, while Phase 7 refers to the phase of landing but missing the target.
Based on these phases, the specific form of the reward function is defined as shown in Formula (6).
r t = { 0                     , p h a s e   1 0.23 × e 0.0023 R 1 + 0.17 × a r c t a n q 1   , p h a s e   2 1                     , p h a s e   3   1                     , p h a s e   4   0.35 × e 0.0012 R 2                     ,   p h a s e   5   1                     , p h a s e   6 1                     , p h a s e   7
Assuming that the warhead of the interceptor is an antipersonnel warhead, then the judgment condition for failed penetration is as follows: the relative distance between the UAV and interceptor R 1 is less than 300 m; otherwise, it is considered a successful penetration. The judgment condition for hitting the target is as follows: the relative distance between the UAV and target R 2 during the landing time is less than 10 m; otherwise, it is considered to have missed the target.

4. TCD-SAC Algorithm

4.1. SAC Algorithm

Both the action space and state space of the UAV are continuous, so the problem of UAV penetration should be solved with a continuous control algorithm. As shown in Figure 4, the continuous control algorithms in reinforcement learning are mainly divided into three categories: the proximal policy optimization (PPO) algorithm [23], the deterministic policy gradient (DDPG) algorithm [24], and the soft actor critic (SAC) algorithm [25]. The PPO algorithm is an on-policy algorithm, which faces the serious problem of sample efficiency, requiring a large number of samples to learn; the DDPG algorithm is an off-policy algorithm with a deterministic policy, which means that only the best action is considered in each state, so its exploration ability is poor; the SAC algorithm is an off-policy algorithm with a stochastic policy, which has a higher sampling efficiency than the PPO algorithm and has a stronger exploration ability than the DDPG algorithm.
There are many action combinations that can realize UAV penetration. Therefore, the algorithm for UAV penetration should use a stochastic policy to explore as many combinations as possible. At the same time, the cost of training for UAV penetration is high, so it is better to realize convergence with less sampling. Based on the above considerations, the SAC algorithm was initially selected for UAV penetration, with lower sample requirements and a stronger exploration ability.
The framework of the SAC algorithm is shown in Figure 5. The SAC algorithm consists of a Policy network, a Value network, and two Q networks. (1) The Policy network generates the optimal action by observing the current state of the environment. Therefore, the input of the Policy network is the state, and the output is the probability distribution of action, from which an action can be sampled. (2) The Value network is a value estimation network of state, which takes the current state as the input and outputs an evaluation value for the current state. (3) The Q network is a value estimation network of a state–action pair, which takes the state–action pair as the input and outputs an evaluation value for the current pair. The specific theoretical derivation of the SAC algorithm is provided in Appendix A.

4.2. Task Completion Division

Facing a huge exploration space of UAV penetration, the number of samples required for network convergence is high. Although the sample demand of the SAC algorithm is relatively small, the actual number of samples required is still huge [25]. To improve the sample utilization of UAV penetration, this paper proposes an improved sampling method, which will improve the sampling efficiency by dividing task completion. This paper defines the improved method as task completion division (TCD).
Specifically, the task of the UAV in this paper is autonomous penetration and a target hit, so its task can be preliminarily divided into two subtasks: a (1) penetration subtask and (2) hit subtask. For example, there is a sample that successfully penetrates but fails to hit the target; that is to say, the penetration subtask has been completed, but the hit subtask has not been completed. In the above example, the sample only completed part of the total task, so it will be regarded as worthless [26,27,28]. Although this sample did not hit the target, its successful action combination of penetration is of learning value. Therefore, if the successful part of this sample can be used for learning, it will be useful to reduce the sample demand.
During the training process of UAV penetration, the UAV will continuously interact with the environment to collect samples. In these samples, each episode corresponds to a trajectory S , which is based on the total task m . The trajectory S consists of all states from the start to the end of the corresponding episode, which is shown as Formula (7).
S = ( s 1 , , s T )
where s 1 is the state at the start time, s T is the state at the end time, and T is the total time spent in the episode.
Define the subtask set completed by S as M , and all subtasks included by M are shown in Formula (8).
M = m 1 , , m T
where m 1 is the subtask completed in state s 1 and m T is the subtask completed in state s T . If the total task m can be completed after all subtasks in M are combined, then the corresponding trajectory S is considered a positive sample; otherwise, S is a negative sample.
For a negative sample, the corresponding trajectory S does not provide any useful information. However, although S did not complete the total task m , it completed some subtasks. If these subtasks are used to replace the original total task m , and the value of reward in the trajectory S is recalculated, some useful information can be obtained. In the method of TCD, this trajectory S can select another task m t from M , and play back again, so as to learn from the successful part of the negative sample.
Based on the original task m , the original transition T o r i is expressed as Formula (9).
T o r i = s t m , a t , r t , s t + 1 m
where s t | | m represents a concatenation operation of the current state s t and original task m , s t + 1 | | m represents a concatenation operation of the next state s t + 1 and original task m , a t is the action at the current time, and r t is the reward of the original task m at the current time.
Based on the new task m t , the new transition T n e w is expressed as Formula (10).
T n e w = s t m t , a t , r t , s t + 1 m t
where m t ϵ M , s t | | m t represents a concatenation operation of the current state s t and new task m t , s t + 1 | | m represents a concatenation operation of the next state s t + 1 and new task m t , and r t is the reward of the new task m t at the current time.
As for the division of m t , it is judged according to the states of S . When UAV penetration is successful, it is considered to have completed penetration subtask m p e n , so the corresponding subtask m t is shown as Formula (11).
m t = m p e n
where the condition for completing penetration subtask m p e n is that the relative distance between the UAV and interceptor is always greater than 300 m and the UAV has passed the interceptor.
When the UAV hits the target successfully, it is considered to have completed hit subtask m h i t , so the corresponding subtask m t is shown as Formula (12).
m t = m h i t
where the condition for completing hit subtask m h i t is that the relative distance between the UAV and target is less than 10 m.
When both m p e n and m h i t are successfully completed, it indicates that the total task m has been completed, so the corresponding subtask m t is shown as Formula (13).
m t = m
In this case, many successful subtasks will be used as positive samples, and the number of positive samples will increase many times, thus increasing the sample utilization. A more formal description of TCD is shown in Algorithm 1.
Algorithm 1: Task completion division (TCD) algorithm
Given:
     a reinforcement learning algorithm A                     e . g . ,   SAC , DDPG , PPO
     a reward function r : S × A × M .
Initialize A
Initialize replay buffer D
for episode = 1, K do
     Set a task m and sample an initial state s 0 .
     For t = 0 , T 1 do
          Sample an action a t using the behavioral policy from A :
                                            a t π b ( s t | | m )
          Execute the action a t and observe the next state s t + 1
     end for
     for t = 0 , T 1 do
                                            r t : = r s t , a t , m
          Store the transition s t m , a t , r t , s t + 1 m in D                 standard   experience   replay
          for m t ϵ M do
                                            r t : = r s t , a t , m t
              Store the transition s t m i , a t , r t , s t + 1 m i in D           TCD
          end for
      end for
     for t = 1 , L do
          Sample a minibatch B from the replay buffer D
          Perform one step of optimization using A and minibatch B
     end for
end for

4.3. TCD-SAC Algorithm

This paper combines the above TCD algorithm with the SAC algorithm and finally obtains the TCD-SAC algorithm, whose framework is shown in Figure 6. The TCD-SAC algorithm samples from the replay buffer and outputs a corresponding action based on the current state. This action controls the UAV to engage in game confrontation with the interceptor, and the transition generated during the confrontation process is placed in the replay buffer. The above process is continuously cycled until the algorithm converges.
The pseudo code of the TCD-SAC algorithm is shown in Algorithm 2. First, input initial parameter vectors θ 1 , θ 2 , and ϕ . Then, initialize target network weights θ ¯ 1 and θ ¯ 2 , and initialize an empty replay pool D . Next, in each time step t , sample action a t from the policy π ϕ , sample the transition s t m , a t , r t , s t + 1 m from the environment, and store it in the replay pool D . For the subtask m t ϵ M , store the additional transition s t m t , a t , r t , s t + 1 m t in the replay pool D , so as to increase the number of valuable transitions [29,30,31]. Then, in each gradient step, update the Q-function parameters θ 1 and θ 2 , update policy weights ϕ , adjust temperature α , and update target network weights θ ¯ 1 and θ ¯ 2 , so as to output the optimized parameters θ 1 , θ 2 , and ϕ .
Because of the addition of the TCD algorithm, the TCD-SAC algorithm can learn from failure. For example, when the UAV fails to penetrate the defense in a certain episode because it is too close to the interceptor, a UAV with conventional algorithms cannot learn any useful information from this failed sample [32,33]. Instead, based on the experience of the previous failure, a UAV with the TCD-SAC algorithm will distance itself from the interceptor in the next episode, thereby increasing the probability of successful penetration [34,35,36].
Algorithm 2: Task completion division–soft actor critic (TCD-SAC) algorithm
Input:   θ 1 , θ 2 , ϕ
      θ ¯ 1 θ 1 , θ ¯ 2 θ 2
      D
   for each iteration do
     for each environment step do
          a t ~ π ϕ ( a t | s t )
          s t + 1 ~ p ( s t + 1 | s t , a t )
          r t : = r s t , a t , m
          D D s t m , a t , r t , s t + 1 m
      for m t ϵ M do
                  r : = r s t , a t , m t
                  D D s t m t , a t , r t , s t + 1 m t
      end for
    end for
    for each gradient step do
            θ i θ i λ Q ^ θ i J Q θ i   f o r   i 1 , 2  
            ϕ ϕ λ π ^ ϕ J π ϕ
            α α λ ^ α J α
            θ ¯ i τ θ i + 1 τ θ ¯ i   f o r   i 1 , 2
     end for
  end for
Output:   θ 1 , θ 2 , ϕ
Theoretically, the TCD-SAC algorithm is able to improve the efficiency of sampling and speed up the convergence of the network. However, the above is only a theoretical derivation. In order to verify the real performance of the algorithm, experiments should be carried out.
(1) To verify the performance of the TCD-SAC algorithm, it should be compared with the original SAC algorithm.
(2) To verify the superiority of the SAC algorithm in UAV penetration, it should be compared with the DDPG algorithm and PPO algorithm (the other two excellent algorithms mentioned above). Of course, the TCD-DDPG algorithm and TCD-PPO algorithm should be compared with the TCD-SAC algorithm, where the TCD-DDPG algorithm is a combination of the DDPG and TCD algorithms, and the TCD-PPO algorithm is a combination of the PPO and TCD algorithms.
Therefore, the TCD-SAC algorithm, TCD-DDPG algorithm, TCD-PPO algorithm, SAC algorithm, DDPG algorithm, and PPO algorithm need to be compared through subsequent experiments, and the experimental results will be presented and discussed in detail in Section 5.

5. Environmental Results and Discussion

5.1. Training and Testing Environment

For the convenience of observation, this study builds a complete visual simulation environment based on the 6-DOF UAV model. The parameters of the UAV model are based on real data, the visual part is written in the QT language, and the algorithm is written in the Python language; the algorithm calls the 6-DOF UAV model through the User Datagram Protocol.
As for the 6-DOF UAV model built in this study, its range is approximately 4000 to 6000 km, its maximum boost height is approximately 120~180 km in the boost section, its glide height is approximately 30~40 km in the gliding section, and its gliding speed is approximately 4~6 Ma. The specific physical parameters of the UAV are shown in Table 1.
As for the reinforcement learning algorithm for UAV penetration, it is in the software environment of Python 3.6 and the Pytorch framework. The actual training and test process is shown in Figure 7. The specific parameters of each algorithm are shown in Table 2.
The process of training: When the UAV detects the interceptor in the environment, the perception unit of the UAV will obtain the state vector s t = R 1 , q 1 , R 2 and transmit it to the algorithm. Inside the algorithm, the corresponding action vector a t = α , γ V will be given according to s t , and the reward function r t will be iteratively updated, whose value is obtained from Formula (3). With the progress of massive training, the reward function gradually converges to the optimal value, thus obtaining a policy π * , which can stably output a real-time action command α , γ V .
The process of testing: The weight parameters of the algorithm will be fixed after convergence. At this time, when the state vector s t is observed again by the perception unit, the stable policy π * will output a continuous action command α , γ V to control the flight. Under the guidance of policy π * , the UAV can achieve autonomous evasion and penetration and finally hit the target.

5.2. Performance in Training Experiment

The performances of all algorithms (SAC, DDPG, PPO, TCD-SAC, TCD-DDPG, and TCD-PPO) are trained in a dynamic environment. The operating system is Ubuntu 18.04, and the GPU is a GeForce RTX 3090. The maximum number of episodes is 10,000, and the maximum time step for each episode is 2000. Under these conditions, it takes 6~8 h to perform a training session with 10,000 episodes. The launch point and target point of the UAV are randomly initialized in each episode, and the performance of each algorithm is as shown in Figure 8. The abscissa in Figure 8 is the episodes, which means the number of training rounds, while the ordinate is the average reward, which means the average value of all groups of rewards. As for each algorithm, this study sets the value of the seed to vary from 0 to 9, corresponding to 10 rounds of training, so the average reward of an ordinate is the average value of 10 groups of rewards.
From Figure 8, the performance of each algorithm is analyzed as follows.
(1) From the convergence speed and convergence result: When the SAC algorithm is applied alone, the network cannot converge after 10,000 episodes, and the value of the average reward is low, indicating that the original SAC algorithm has poor performance in UAV penetration. However, when the TCD-SAC algorithm is applied, the algorithm converges at about 8000 episodes, and its average reward value corresponding to convergence is the highest of all algorithms, indicating that the TCD-SAC algorithm has the fastest convergence speed and the best convergence result.
(2) From the perspective of sample utilization: When the SAC, PPO, or DDPG algorithm is used alone, the network does not converge after 10,000 episodes, which is because their sample utilization is too low. However, when the above three algorithms are combined with the TCD method to obtain the TCD-SAC, TCD-PPO, and TCD-DDPG algorithms, they can all converge, which shows that the TCD method effectively improves the sample utilization. Moreover, the TCD-SAC algorithm with the fastest convergence speed shows that it has the highest sample utilization.
(3) From the superiority of the SAC algorithm in UAV penetration: When the TCD-DDPG algorithm is used, its network converges at about 4000 episodes, and its convergence result is lower than that of the other two algorithms (TCD-SAC, TCD-PPO). This is because the algorithm with a deterministic policy selects only one deterministic action output in each state, so its degree of exploration is low and its convergence result is not ideal. On the contrary, the TCD-SAC algorithm with a stochastic policy is indeed superior in UAV penetration. When the TCD-PPO algorithm is used, its convergence speed is similar to that of the TCD-SAC algorithm, but its convergence result is not as good as that of the TCD-SAC algorithm. This is because on-policy algorithms such as the TCD-PPO algorithm cannot be used while exploring, so it is easy to fall into local optimization. On the contrary, off-policy algorithms such as the TCD-SAC algorithm are superior in UAV penetration. Among the above three algorithms combined with the TCD method, the TCD-SAC algorithm performs the best, which proves the superiority of the SAC algorithm in UAV penetration.
In addition, the success rate of penetrations for each algorithm is shown in Figure 9, and the success rate of hits for each algorithm is shown in Figure 10. Their change with episodes is consistent with the average reward in Figure 8, which shows the rationality and correctness of reward shaping.
The above is the result of training experiments. To test the actual performance of each algorithm further, a test experiment should later be conducted, which is shown in Section 5.3.

5.3. Performance in Test Experiment

The maximum number of test experiments is 3000, and the launch point and target point are randomly initialized in each test to verify the general performance of all algorithms in a complex dynamic environment. Since it is impossible to show all test results here, the 600th, 1200th, 1800th, 2400th, and 3000th test results are selected, respectively, corresponding to Figure a–e in Figure 11. In Figure 11, the red line is the trajectory of the UAV, and the blue line is the trajectory of the interceptor. The performance of the UAV with different algorithms in the test environment is shown in Figure 11, which is from a two-dimensional perspective.
Figure 11a shows the result of the 600th test experiment. When the PPO algorithm, DDPG algorithm, or SAC algorithm is used, the test result is “penetration failed”; when the TCD-DDPG algorithm is used, the test result is “penetration succeeded but hit failed”; when the TCD-PPO algorithm or TCD-SAC algorithm is used, the test result is “successful penetration and hit”.
Figure 11b shows the result of the 1200th test experiment. When the PPO algorithm, DDPG algorithm, SAC algorithm, or TCD-PPO algorithm is used, the test result is “penetration failed”; when the TCD-DDPG algorithm is used, the test result is “penetration succeeded but hit failed”; when the TCD-SAC algorithm is used, the test result is “successful penetration and hit”.
Figure 11c shows the result of the 1800th test experiment. When the PPO algorithm, DDPG algorithm, or TCD-PPO algorithm is used, the test result is “penetration failed”; when the SAC algorithm is used, the test result is “penetration succeeded but hit failed”; when the TCD-DDPG algorithm or TCD-SAC algorithm is used, the test result is “successful penetration and hit”.
Figure 11d shows the result of the 2400th test experiment. When the PPO algorithm, DDPG algorithm, or TCD-DDPG algorithm is used, the test result is “penetration failed”; when the SAC algorithm, TCD-PPO algorithm, or TCD-SAC algorithm is used, the test result is “successful penetration and hit”.
Figure 11e shows the result of the 3000th test experiment. When the PPO algorithm, SAC algorithm, TCD-PPO algorithm, or TCD-DDPG algorithm is used, the test result is “UAV penetration failed”; when the DDPG algorithm or TCD-SAC algorithm is used, the test result is “penetration succeeded but hit failed”; when the TCD-DDPG algorithm or TCD-SAC algorithm is used, the test result is “successful penetration and hit”.
In the five test results above, the TCD-SAC algorithm has four perfect results of “successful penetration and hit” and only one bad result of “penetration succeeded but hit failed”, which is because the landing point of the UAV slightly deviated from the target point. On the contrary, other algorithms mostly have bad results of “penetration failed” or “penetration succeeded but hit failed”. Therefore, the TCD-SAC algorithm proposed in this paper has the best performance in the above test results.
The five results in Figure 11 are only a sample of 3000 test results. To objectively evaluate the performance of each algorithm, the 3000 test results should be comprehensively evaluated. Therefore, this paper draws the minimum distance between the UAV and interceptor in each of the 3000 tests, which is the index to evaluate whether the UAV has successfully penetrated. As mentioned earlier, once the minimum distance between the UAV and interceptor is less than 300 m, it is considered a failed penetration; otherwise, it is considered a successful penetration.
Figure 12 illustrates that when the PPO, SAC, or DDPG algorithm is tested, the minimum distance between the UAV and interceptor (hereinafter referred to as mini-distance) is less than 100 m in each test, which means failed penetration. When the TCD-DDPG algorithm is tested, the mini-distance increases to about 250 m, but its mini-distance is still less than 300 m, so most results are still failed penetration; when the TCD-PPO algorithm is tested, its mini-distance increases to about 280 m, which is still less than 300 m and most results are failed penetration; when the TCD-SAC algorithm is tested, the mini-distance increases to about 330 m. In this case, the mini-distance of most tests is greater than 300 m, so most results are successful penetration. Therefore, the TCD-SAC algorithm proposed in this paper has the best performance of penetration in 3000 tests.
For the convenience of description, this paper defines the number of tests as NT, the number of successful penetrations as NP, the number of successful hits as NH, the success rate of penetrations as RP, and the success rate of hits as RH. RP is equal to NP divided by NT, and RH is equal to NH divided by NT. Based on the results in Figure 12, the success rate of penetrations (RP) in 3000 tests is shown in Figure 13, and the success rate of hits (RH) in 3000 tests is shown in Figure 14. From Figure 13 and Figure 14, the success rates of penetrations (RP) of DDPG, PPO, and SAC algorithms are 0.63%, 3.27%, and 3.50%, respectively, which are all less than 4%, and their success rates of hits (RH) are 0.47%, 2.80%, and 2.87%, respectively, which are all less than 3%. The above results show that these algorithms have a poor performance in UAV penetration. In contrast, the RP and RH of the TCD-DDPG algorithm are 17.83% and 11.13%, respectively, and those of the TCD-PPO algorithm are 42.90% and 31.97%, respectively. Compared to the first three algorithms, both the TCD-DDPG and TCD-PPO algorithms exhibit significant performance improvements. Furthermore, the RP and RH of the TCD-SAC algorithm are 68.20% and 67.90%, respectively, whose performance is far ahead of the other algorithms. That is to say, when the TCD-SAC algorithm is tested, all subtasks are completed perfectly in most of the 3000 tests, indicating the reliability and universality of the algorithm.
The above test results show that the TCD-SAC algorithm obtained from training has a good feasibility and robustness in a complex dynamic environment. When the UAV encounters the interceptor, the penetration policy output by the TCD-SAC algorithm can play a significant role in time, quickly guiding the UAV to maneuver and avoid the interceptor to complete the task of autonomous penetration and hitting the target.

6. Conclusions and Future Work

Committed to improving the sample utilization of UAV penetration, this paper proposes the improved sampling method TCD and combines this TCD method with the SAC algorithm to obtain the TCD-SAC algorithm. To verify the real performance of the TCD-SAC algorithm, this study designs corresponding action space, state space, and a reward function, and conducts training experiments and test experiments for UAV penetration. All experiments are carried out in a complex dynamic environment. Compared with the original SAC algorithm and two other excellent algorithms (DDPG and PPO), the training experiments show that the improved TCD-SAC algorithm has the fastest convergence speed and the best convergence result in all algorithms. The test experiments show that the improved TCD-SAC algorithm has the highest success rate of penetrations and hits, illustrating the feasibility and robustness of the TCD-SAC algorithm. This paper presents a relatively complete study from model building to algorithm improvement to experimental validation, which provides potentially useful ideas for the future autonomous penetration of UAVs.
Of course, because the cost of training and testing a real physical UAV is too high, this paper only conducted experiments in a simulation environment. If the algorithm is to be applied to a real physical UAV, it will be necessary to eliminate the difference between the real model and simulation model as much as possible. Therefore, our next task is to try to bridge the reality gap between a real model and simulation model. For example, when an algorithm performs well in a simulation environment, methods such as transfer learning can be used to transfer the trained algorithm to a real environment quickly and cost effectively.

Author Contributions

Conceptualization, Y.W. and X.Z.; methodology, X.Z. and K.L.; software, X.L. and H.L.; validation, K.L.; formal analysis, K.L.; investigation, K.L.; resources, K.L. and Y.W.; data curation, K.L.; writing—original draft preparation, X.Z. and Y.W.; writing—review and editing, K.L., X.L., and H.L.; visualization, K.L.; supervision, X.L. and H.L.; project administration, Y.W.; funding acquisition, K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The SAC algorithm defines the policy in the Actor Network as π ϕ ( a t | s t ) and defines the output of the Q Critic Network as Q θ s t , a t . The update method of the Q Critic Network is shown in Formula (A1).
J Q θ = E s t , a t , s t + 1 ~ D 1 2 Q θ s t , a t r s t , a t + γ V θ ¯ s t + 1 2 = E s t , a t , s t + 1 ~ D , a t ~ π ϕ 1 2 Q θ s t , a t r s t , a t + γ ( Q θ ¯ s t + 1 , a t + 1 α log π ϕ s t + 1 , a t + 1 ) 2
where J Q θ is the cost function of the Q Critic Network, s t is the current state, a t is the current action, D is the replay pool, s t + 1 is the next state, r s t , a t is the corresponding reward, γ is the discount factor, θ and ϕ are parameter vectors of the Q Critic Network, and θ ¯ is obtained from θ through an exponentially moving average.
The update method of the Actor Network is to minimize the Kullback–Leibler Divergence, which is shown in Formula (A2).
J π ϕ = D K L ( π ϕ ( . | s t ) | | exp 1 α Q θ s t , . l o g Z s t ) = E s t ~ D , a t ~ π ϕ [ log ( π ϕ a t | s t exp 1 α Q θ s t , a t l o g Z s t ) ] = E s t ~ D , a t ~ π ϕ [ l o g π ϕ a t | s t 1 α Q θ s t , a t + l o g Z s t ]
where J π ϕ is the cost function of the Actor Network, and the generation method of a t is shown in Formula (A3).
a t = f ϕ ε t ; s t = f ϕ μ s t + ε t f ϕ σ s t
where the f function calculates the mean and variance of the sample, the sample follows a standard normal distribution, and ε is the noise of the standard normal distribution. Based on the reparameterization trick in Formula (A3), the whole process of Formula (A2) is completely differentiable. Therefore, Formula (A2) can be simplified into Formula (A4).
J π ϕ = E s t ~ D , ε ~ N α l o g π ϕ f ϕ ε t ; s t | s t Q θ s t , f ϕ ε t ; s t

References

  1. Kumar, G.N.; Sarkar, A.; Mangrulkar, K.; Talole, S. Atmospheric vehicle trajectory optimization with minimum dynamic pressure constraint. Proc. Inst. Mech. Eng. Part G J. Aerosp. Eng. 2018, 232, 837–846. [Google Scholar] [CrossRef]
  2. Chai, R.; Savvaris, A.; Tsourdos, A.; Chai, S.; Xia, Y.; Wang, S. Solving trajectory optimization problems in the presence of probabilistic constraints. IEEE Trans. Cybern. 2019, 50, 4332–4345. [Google Scholar] [CrossRef] [Green Version]
  3. Xu, J.; Dong, C.; Cheng, L. Deep neural network-based footprint prediction and attack intention inference of hypersonic glide vehicles. Mathematics 2022, 11, 185. [Google Scholar] [CrossRef]
  4. Gao, Y.; Cai, G.; Yang, X.; Hou, M. Improved tentacle-based guidance for reentry gliding hypersonic vehicle with no-fly zone constraint. IEEE Access 2019, 7, 119246-58. [Google Scholar] [CrossRef]
  5. Liao, Y.; Li, H. Trajectory optimization for terminal phase flight of hypersonic reentry vehicles with multi-constraints. In Proceedings of the 2013 25th Chinese Control and Decision Conference (CCDC), Guiyang, China, 25–27 May 2013; pp. 571–576. [Google Scholar]
  6. Sana, K.S.; Weiduo, H. Hypersonic reentry trajectory planning by using hybrid fractional-order particle swarm optimization and gravitational search algorithm. Chin. J. Aeronaut. 2021, 34, 50–67. [Google Scholar] [CrossRef]
  7. Chai, R.; Savvaris, A.; Tsourdos, A.; Chai, S.; Xia, Y. Improved gradient-based algorithm for solving aeroassisted vehicle trajectory optimization problems. J. Guid. Control Dyn. 2017, 40, 2093–2101. [Google Scholar] [CrossRef]
  8. Wan, K.; Gao, X.; Hu, Z.; Wu, G. Robust motion control for uav in dynamic uncertain environments using deep reinforcement learning. Remote Sens. 2020, 12, 640. [Google Scholar] [CrossRef] [Green Version]
  9. Luo, J.; Liang, Q.; Li, H. Uav penetration mission path planning based on improved holonic particle swarm optimization. J. Syst. Eng. Electron. 2023, 34, 197–213. [Google Scholar] [CrossRef]
  10. Fu, J.; Sun, G.; Yao, W.; Wu, L. On trajectory homotopy to explore and penetrate dynamically of multi-uav. IEEE Trans. Intell. Transp. Syst. 2022, 23, 24008–24019. [Google Scholar] [CrossRef]
  11. Zhang, Z.; Wu, J.; Dai, J.; He, C. A novel real-time penetration path planning algorithm for stealth uav in 3d complex dynamic environment. IEEE Access 2020, 8, 122757–122771. [Google Scholar] [CrossRef]
  12. Luo, Y.; Song, J.; Zhao, K.; Liu, Y. Uav-cooperative penetration dynamic-tracking interceptor method based on ddpg. Appl. Sci. 2022, 12, 1618. [Google Scholar] [CrossRef]
  13. Li, Y.; Han, W.; Wang, Y. Deep reinforcement learning with application to air confrontation intelligent decision-making of manned/unmanned aerial vehicle cooperative system. IEEE Access 2020, 8, 67887–67898. [Google Scholar] [CrossRef]
  14. Kaifang, W.; Bo, L.; Xiaoguang, G.; Zijian, H.; Zhipeng, Y. A learning-based flexible autonomous motion control method for uav in dynamic unknown environments. J. Syst. Eng. Electron. 2021, 32, 1490–1508. [Google Scholar] [CrossRef]
  15. Liang, L.; Deng, F.; Wang, J.; Lu, M.; Chen, J. A reconnaissance penetration game with territorial-constrained defender. IEEE Trans. Autom. Control 2022, 67, 6295–6302. [Google Scholar] [CrossRef]
  16. Bellman, R. A markovian decision process. J. Math. Mech. 1957, 6, 679–684. [Google Scholar] [CrossRef]
  17. Li, S.; Lei, H.; Shao, L.; Xiao, C. Multiple model tracking for hypersonic gliding vehicles with aerodynamic modeling and analysis. IEEE Access 2019, 7, 28011–28018. [Google Scholar] [CrossRef]
  18. Liu, X.; Zhang, Y.; Wang, S.; Huang, W. Backstepping attitude control for hypersonic gliding vehicle based on a robust dynamic inversion approach. Proc. Inst. Mech. Eng. Part I J. Syst. Control Eng. 2014, 228, 543–552. [Google Scholar] [CrossRef]
  19. Li, G.; Zhang, H.; Tang, G. Maneuver characteristics analysis for hypersonic glide vehicles. Aerosp. Sci. Technol. 2015, 43, 321–328. [Google Scholar] [CrossRef]
  20. Shen, Z.; Yu, J.; Dong, X.; Hua, Y.; Ren, Z. Penetration trajectory optimization for the hypersonic gliding vehicle encountering two interceptors. Aerosp. Sci. Technol. 2022, 121, 107363. [Google Scholar] [CrossRef]
  21. Yan, B.; Liu, R.; Dai, P.; Xing, M.; Liu, S. A rapid penetration trajectory optimization method for hypersonic vehicles. International J. Aerosp. Eng. 2019, 2019, 11. [Google Scholar] [CrossRef] [Green Version]
  22. Chai, R.; Tsourdos, A.; Savvaris, A.; Chai, S.; Xia, Y.; Chen, C.P. Six-dof spacecraft optimal trajectory planning and real-time attitude control: A deep neural network-based approach. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 5005–5013. [Google Scholar] [CrossRef] [PubMed]
  23. Xiang, J.; Li, Q.; Dong, X.; Ren, Z. Continuous control with deep reinforcement learning for mobile robot navigation. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 1501–1506. [Google Scholar]
  24. Wenjun, N.; Ying, B.; Di, W.; Xiaoping, M. Energy-optimal trajectory planning for solar-powered aircraft using soft actor-critic. Chin. J. Aeronaut. 2022, 35, 337–353. [Google Scholar]
  25. Eteke, C.; Kebüde, D.; Akgün, B. Reward learning from very few demonstrations. IEEE Trans. Robot. 2020, 37, 893–904. [Google Scholar] [CrossRef]
  26. Han, S.-C.; Bang, H.; Yoo, C.-S. Proportional navigation-based collision avoidance for uavs. Int. J. Control Autom. Syst. 2009, 7, 553. [Google Scholar] [CrossRef]
  27. Wang, Z.; Cheng, X.X.; Li, H. Hypersonic skipping trajectory planning for high l/d gliding vehicles. In Proceedings of the 21st AIAA International Space Planes and Hypersonics Technologies Conference, Xiamen, China, 6–9 March 2017; p. 2135. [Google Scholar]
  28. Tripathi, A.K.; Patel, V.V.; Padhi, R. Autonomous landing of fixed wing unmanned aerial vehicle with reactive collision avoidance. IFAC-PapersOnLine 2018, 51, 474–479. [Google Scholar] [CrossRef]
  29. Maeda, R.; Mimura, M. Automating post-exploitation with deep reinforcement learning. Comput. Secur. 2021, 100, 102108. [Google Scholar] [CrossRef]
  30. Sackmann, M.; Bey, H.; Hofmann, U.; Thielecke, J. Modeling driver behavior using adversarial inverse reinforcement learning. In Proceedings of the 2022 IEEE Intelligent Vehicles Symposium (IV), Aachen, Germany, 5–9 June 2022; pp. 1683–1690. [Google Scholar]
  31. Hu, Y.; Gao, C.; Li, J.; Jing, W.; Li, Z. Novel trajectory prediction algorithms for hypersonic gliding vehicles based on maneuver mode on-line identification and intent inference. Meas. Sci. Technol. 2021, 32, 115012. [Google Scholar] [CrossRef]
  32. Alzahrani, B.; Oubbati, O.S.; Barnawi, A.; Atiquzzaman, M.; Alghazzawi, D. Uav assistance paradigm: State-of-the-art in applications and challenges. J. Netw. Comput. Appl. 2020, 166, 102706. [Google Scholar] [CrossRef]
  33. Kontogiannis, S.G.; Ekaterinaris, J.A. Design, performance evaluation and optimization of a uav. Aerosp. Sci. Technol. 2013, 29, 339–350. [Google Scholar] [CrossRef]
  34. Zhang, S.; Zhang, H.; Di, B.; Song, L. Cellular uav-to-x communications: Design and optimization for multi-uav networks. IEEE Trans. Wirel. Commun. 2019, 18, 1346–1359. [Google Scholar] [CrossRef] [Green Version]
  35. Oubbati, O.S.; Atiquzzaman, M.; Ahanger, T.A.; Ibrahim, A. Softwarization of uav networks: A survey of applications and future trends. IEEE Access 2020, 8, 98073–98125. [Google Scholar] [CrossRef]
  36. Koch, W.; Mancuso, R.; West, R.; Bestavros, A. Reinforcement learning for uav attitude control. ACM Trans. Cyber-Phys. Syst. 2019, 3, 1–21. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The application scenario for UAV penetration.
Figure 1. The application scenario for UAV penetration.
Aerospace 10 00642 g001
Figure 2. The relationship between UAV penetration and reinforcement learning.
Figure 2. The relationship between UAV penetration and reinforcement learning.
Aerospace 10 00642 g002
Figure 3. The variables related to the observation state.
Figure 3. The variables related to the observation state.
Aerospace 10 00642 g003
Figure 4. Continuous control algorithm in reinforcement learning.
Figure 4. Continuous control algorithm in reinforcement learning.
Aerospace 10 00642 g004
Figure 5. Framework of the SAC algorithm.
Figure 5. Framework of the SAC algorithm.
Aerospace 10 00642 g005
Figure 6. Framework of TCD-SAC algorithm.
Figure 6. Framework of TCD-SAC algorithm.
Aerospace 10 00642 g006
Figure 7. The actual training and test process of the algorithm.
Figure 7. The actual training and test process of the algorithm.
Aerospace 10 00642 g007
Figure 8. The average reward for each algorithm.
Figure 8. The average reward for each algorithm.
Aerospace 10 00642 g008
Figure 9. The success rate of penetrations for each algorithm.
Figure 9. The success rate of penetrations for each algorithm.
Aerospace 10 00642 g009
Figure 10. The success rate of hits for each algorithm.
Figure 10. The success rate of hits for each algorithm.
Aerospace 10 00642 g010
Figure 11. The performance of different algorithms in the test environment. Among them, subfigure (a) represents the results of the 600th test, subfigure (b) represents the results of the 1200th test, subfigure (c) represents the results of the 1800th test, subfigure (d) represents the results of the 2400th test, and subfigure (e) represents the results of the 3000th test.
Figure 11. The performance of different algorithms in the test environment. Among them, subfigure (a) represents the results of the 600th test, subfigure (b) represents the results of the 1200th test, subfigure (c) represents the results of the 1800th test, subfigure (d) represents the results of the 2400th test, and subfigure (e) represents the results of the 3000th test.
Aerospace 10 00642 g011aAerospace 10 00642 g011bAerospace 10 00642 g011c
Figure 12. The minimum distance between the UAV and interceptor in 3000 tests.
Figure 12. The minimum distance between the UAV and interceptor in 3000 tests.
Aerospace 10 00642 g012
Figure 13. The success rate of penetrations in 3000 tests.
Figure 13. The success rate of penetrations in 3000 tests.
Aerospace 10 00642 g013
Figure 14. The success rate of hits in 3000 tests.
Figure 14. The success rate of hits in 3000 tests.
Aerospace 10 00642 g014
Table 1. Physical parameters of the UAV.
Table 1. Physical parameters of the UAV.
PerformanceValue
range4000~6000 km
maximum boost height120~180 km
glide height30~40 km
gliding speed4~6 Ma
Table 2. Parameters of each algorithm.
Table 2. Parameters of each algorithm.
AlgorithmNetwork Dimension NumberNetwork Layer NumberReplay BufferBatch SizeUpdate TimesDiscount Factor
SAC128421712810240.998
DDPG25642172565120.997
PPO128821712810240.994
TCD-SAC128421712810240.995
TCD-DDPG25642172565120.991
TCD-PPO128821712810240.997
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Li, K.; Zhuang, X.; Liu, X.; Li, H. A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration. Aerospace 2023, 10, 642. https://doi.org/10.3390/aerospace10070642

AMA Style

Wang Y, Li K, Zhuang X, Liu X, Li H. A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration. Aerospace. 2023; 10(7):642. https://doi.org/10.3390/aerospace10070642

Chicago/Turabian Style

Wang, Yue, Kexv Li, Xing Zhuang, Xinyu Liu, and Hanyu Li. 2023. "A Reinforcement Learning Method Based on an Improved Sampling Mechanism for Unmanned Aerial Vehicle Penetration" Aerospace 10, no. 7: 642. https://doi.org/10.3390/aerospace10070642

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop