Next Article in Journal
Robust Task Offloading and Trajectory Optimization for UAV-Mounted Mobile Edge Computing
Next Article in Special Issue
Utilizing the Finite Fourier Series to Generate Quadrotor Trajectories Through Multiple Waypoints
Previous Article in Journal
Rapid Integrated Design Verification of Vertical Take-Off and Landing UAVs Based on Modified Model-Based Systems Engineering
Previous Article in Special Issue
MPC-Based Dynamic Trajectory Spoofing for UAVs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Automated Maneuvering Decisions in UCAV Air Combat Games Using Homotopy-Based Reinforcement Learning

1
School of Aeronautics and Astronautics, Zhejiang University, Hangzhou 310058, China
2
Zhejiang Provincial Engineering Research Center of Advanced UAV Systems, Zhejiang University, Hangzhou 310027, China
3
School of Computer Science, Civil Aviation Flight University of China, Guanghan 618307, China
*
Author to whom correspondence should be addressed.
Drones 2024, 8(12), 756; https://doi.org/10.3390/drones8120756
Submission received: 12 November 2024 / Revised: 10 December 2024 / Accepted: 11 December 2024 / Published: 13 December 2024
(This article belongs to the Special Issue Path Planning, Trajectory Tracking and Guidance for UAVs: 2nd Edition)

Abstract

:
In the field of real-time autonomous decision-making for Unmanned Combat Aerial Vehicles (UCAVs), reinforcement learning is widely used to enhance their decision-making capabilities in high-dimensional spaces. These enhanced capabilities allow UCAVs to better respond to the maneuvers of various opponents, with the win rate often serving as the primary optimization metric. However, relying solely on the terminal outcome of victory or defeat as the optimization target, but without incorporating additional rewards throughout the process, poses significant challenges for reinforcement learning due to the sparse reward structure inherent in these scenarios. While algorithms enhanced with densely distributed artificial rewards show potential, they risk deviating from the primary objectives. To address these challenges, we introduce a novel approach: the homotopy-based soft actor–critic (HSAC) method. This technique gradually transitions from auxiliary tasks enriched with artificial rewards to the main task characterized by sparse rewards through homotopic paths. We demonstrate the consistent convergence of the HSAC method and its effectiveness through deployment in two distinct scenarios within a 3D air combat game simulation: attacking horizontally flying UCAVs and a combat scenario involving two UCAVs. Our experimental results reveal that HSAC significantly outperforms traditional algorithms, which rely solely on using sparse rewards or those supplemented with artificially aided rewards.

1. Introduction

The prevailing research in the field of Unmanned Combat Aerial Vehicles (UCAVs) is centered on enhancing autonomy, which necessitates advanced intelligent decision-making capabilities. To effectively navigate dynamic air combat game environments, a UCAV must execute swift and informed decisions. Traditional methods like the differential game method [1,2], influence diagram method [3,4], model predictive control [5], genetic algorithms [6], and Bayesian inference [7] have been deployed to tackle tactical decision-making challenges. These approaches analyze relationships in air combat games to derive optimal solutions but often falter in complex environments that require real-time calculations.
With the increasing complexity of air combat game scenarios, decision-making strategies must evolve. Recently, artificial intelligence (AI) has shown promise in addressing these challenges, particularly through its capability for real-time performance. Traditional AI methods include expert systems, which utilize comprehensive rule bases to generate empirical policies [8,9,10,11], and supervised learning, which employs neural networks to translate sensor data into tactical decisions based on expert demonstrations [12,13,14]. However, these methods are limited by their reliance on hard-to-obtain expert data, which may not adequately represent the complexities of real-world air combat games.
Reinforcement Learning (RL) addresses these limitations by enabling autonomous learning through environmental interactions, thus maximizing returns without prior expertise [15]. This approach has gained traction in autonomous air combat game applications, where UCAVs operate as agents [16,17,18,19,20]. Pioneers like McGrew [21] and Crumpacker [22] have applied approximate dynamic programming to develop maneuver decision models for 1 vs. 1 air combat game scenarios, with McGrew even testing these on micro-UCAVs. Their work underscores the viability of RL in autonomous tactical decision-making. Additional advancements include efforts by Xiaoteng Ma [23], who integrated speed control into the action space using Deep Q-Learning, and Zhuang Wang [24], who introduced a freeze game framework to mitigate nonstationarity in RL through a league system that adapts to varying opponent strategies. Furthermore, Qiming Yang [25] developed a motion model for a second-order UCAV and a close-range air combat game model in 3D, employing DQN and basic confrontation training techniques. Notably, Adrian’s use of Hierarchical RL [26] secured a second-place finish in DARPA’s AlphaDogfight Trials, showcasing the effectiveness of artificially aided rewards in challenging air combat game scenarios.
Despite these advancements, the sparse reward challenge in RL persists, especially in high-dimensional settings like air combat games, where agents must identify a sequence of correct actions to attain the sparse reward. Traditional random exploration often proves inadequate, which provides minimal feedback on action quality. In response, scholars have explored various strategies, including reward shaping [27,28,29], curriculum learning [30,31,32], learning from demonstrations [33,34,35], model-guided learning [36], and inverse RL [37]. However, these methods typically rely on artificial prior knowledge, potentially biasing policies toward suboptimal outcomes.
Furthermore, previous studies often oversimplify air combat game scenarios, assuming 2D movement or discretized basic flight maneuvers (BFM), thus hampering the realism and efficacy of learned policies. In contrast, this paper advances the following contributions:
(1)
We introduce the homotopy-based soft actor–critic (HSAC) algorithm to address the sparse reward problem in air combat games. We theoretically validate the feasibility and consistent convergence of this approach, offering a new perspective for RL-based methods.
(2)
We develop a more realistic air combat game environment, utilizing a refined 3D flight dynamics model and expanding the agent’s action space from discrete to continuous.
(3)
We implement the HSAC algorithm in this enhanced environment, leveraging self-play. Simulations demonstrate that HSAC significantly outperforms models trained with either sparse or artificially aided rewards.
The paper is structured as follows: Section 2 reviews existing RL-based methods and introduces the air combat game problem. Section 3 details our air combat game environment setup, incorporating a refined 3D flight dynamics model and a continuous action space. In Section 4, we describe the homotopy-based soft actor–critic algorithm and provide theoretical evidence of its convergence and effectiveness. An overview diagram of the HSAC algorithm is shown in Figure 1. Section 5 showcases the performance of the HSAC algorithm in various training and experimental setups. Links to video footage of the training process are available online (Supplementary Materials). Finally, Section 6 summarizes our contributions and discusses future research directions.

2. Preliminaries

In this section, we provide an overview of the dynamics model of a UCAV during short-range air combat game scenarios and introduce the soft actor–critic (SAC) method, a widely recognized reinforcement learning (RL) algorithm for training autonomous agents.

2.1. One-to-One Short-Range Air Combat Game Problem

Short-range air combat games, also known as dogfights, entail the objective of shooting down an enemy UCAV while avoiding return fire. To distinguish between the UCAVs of opposing sides, we use superscript b for blue-side UCAVs and superscript r for red-side UCAVs. The interaction between UCAV b and UCAV r is inherently adversarial. For simplicity, our focus will primarily be on the learning problem of UCAV b .
Figure 2 illustrates a typical one-to-one short-range air combat game scenario. In this model, four variables are crucial for defining the advantageous position of UCAV b : Aspect Angle (AA), Antenna Train Angle (ATA), Line of Sight (LOS), and Relative Distance. These metrics allow UCAV b to determine the position P r and velocity V r of the opposing UCAV r .
For UCAV b to successfully engage UCAV r , it must fulfill the following criteria:
(1)
The relative distance between the UCAVs must be within the maximum attack range and above the minimum collision range.
(2)
UCAV b must maintain an advantageous position, enabling the pursuit of UCAV r while evading potential counterattacks.
These requirements are governed by the following constraints [21]:
{ d min < d L O S < d max , A A b < 60 , A T A b < 60 .
If all constraints in Equation (1) are satisfied, UCAV b can engage UCAV r within its firing envelope [38]. The same operational constraints apply to UCAV r .
A two-target differential game model [39,40] aptly describes this air combat game scenario:
min u b max u r J ( u b , u r ) = f ( t f , x b ( t f ) , x r ( t f ) ) + t 0 t f L b ( t , x b ( t ) , x r ( t ) , u b ( t ) , u r ( t ) ) d t , min u r max u b J ( u r , u b ) = f ( t f , x r ( t f ) , x b ( t f ) ) + t 0 t f L r ( t , x r ( t ) , x b ( t ) , u r ( t ) , u b ( t ) ) d t .
Here, u b and u r denote the control policies of each UCAV, whereas x b and x r represent their respective state vectors, as defined in Equation (3). In Equation (2), L b ( · ) and L r ( · ) are the loss functions for UCAV b and UCAV r , respectively, with  f ( · ) representing the terminal punishment function:
x b ( t ) = [ x b ( t ) , y b ( t ) , z b ( t ) , v b ( t ) , γ b ( t ) , χ b ( t ) ] T , x r ( t ) = [ x r ( t ) , y r ( t ) , z r ( t ) , v r ( t ) , γ r ( t ) , χ r ( t ) ] T .
Here, v, γ , χ represents the velocity, the flight path angle, and the heading angle of UCAV.
The primary aim of this research is to enable UCAV b to rapidly attain an advantageous position. In reinforcement learning, the concept of self-play is often utilized to simplify the adversarial game, a strategy that has been successfully implemented in AlphaZero [41]. Consequently, we have simplified the two-target differential game into an optimization problem that satisfies the UCAV dynamics constraints. More details are provided in Section 4.4.

2.2. UCAV’s Dynamics Model

The dynamics model of the UCAV is crucial for accurately simulating air combat game scenarios and is constructed using the ground coordinate system. This model is described by a three-degree-of-freedom (3-DoF) particle motion model [3]:
{ x ˙ = v cos γ cos χ , y ˙ = v cos γ sin χ , z ˙ = v sin γ .
In Equation (4), x ˙ , y ˙ , and  z ˙ denote the UCAV’s rates of change in position along the X, Y, and Z axes, respectively. The variables γ , χ , and v represent the flight path angle, heading angle, and velocity, respectively. Figure 3 illustrates the associated flight dynamics.
Control variables for the UCAV, including the attack angle α , throttle setting η , and bank angle μ , interplay intricately via:
{ v ˙ = 1 m η T max cos α D ( α , v ) g sin γ , χ ˙ = 1 m v cos γ η T max sin α + L ( α , v ) sin μ , γ ˙ = 1 m v η T max sin α + L ( α , v ) cos μ m g cos γ .
Here, g, m, and  T max represent the acceleration due to gravity, the mass of the UCAV, and the maximum thrust force of the engine, respectively. Lift and drag forces are given as follows:
{ L ( α , v ) = 1 2 ρ v 2 S w C L ( α ) , D ( α , v ) = 1 2 ρ v 2 S w C D ( α ) .
Here, the reference wing area is denoted by S w . The air density ρ is assumed to be a constant in this study. C L and C D denote lift and drag coefficients, respectively, related through the attack angle α :
{ C L ( α ) = C L 0 + C L α α , C D ( α ) = C D 0 + B D P · C L ( α ) 2 .
Variables C L 0 , C L α , C D 0 , and  B D P in Equation (7) signify the zero-lift coefficient, the derivative of the lift coefficient with respect to attack angle, zero-drag coefficient, and the drag–lift coefficient, respectively.
Constraints on both the control variables and their rate of change are defined to ensure performance and maintain UCAV inertia within operational limits:
{ α [ α min , α max ] , α ˙ [ Δ α , Δ α ] , μ [ π , π ] , μ ˙ [ Δ μ , Δ μ ] , η = η max = 1 , η ˙ = 0 .
The throttle setting η is assumed constant at 1 to maximize pursuit capabilities.
The load factor n ( · ) and dynamic pressure q ( · ) are defined as follows:
n ( α , v ) = L ( α , v ) m g , q ( v ) = 1 2 ρ v 2 .
Constraints on these parameters prevent overloading of the UCAV:
{ n ( α , v ) n max 0 , h min h 0 , q ( v ) q max 0 .
where h min , n max , and  q max denote the minimum altitude, maximum load factor, and maximum dynamic pressure, respectively. Additionally, it is important to note that the relationship between altitude h and the functions n ( · ) and q ( · ) is not considered here, as the air density ρ is assumed constant.

2.3. Soft Actor–Critic Method

The Markov Decision Process (MDP) is fundamental to RL-based methods, facilitating the modeling of sequential decision-making problems using mathematical formalism. An MDP consists of a state space S, an action space A, a transition function T, and a reward function R, forming a tuple < S , A , T , R > . RL-based methods aim to optimize the agent’s policy through interaction with the environment. At each time step t, the agent in state s t chooses an action a t , and the environment provides feedback on the next state s t + 1 based on the transition probability T ( s t + 1 | s t , a t ) [ 0 , 1 ] , along with a reward r t + 1 = R ( s t , a t ) to evaluate the quality of the action. The policy π : S A directs the agent’s action selection in each state, with the probability of each action given by π ( a | s ) . The ultimate objective is to discover an optimal policy π * that maximizes the expected cumulative reward.
Finding the optimal policy π * can be challenging, especially if the agent has not extensively explored the state space. This challenge is addressed by employing maximum entropy reinforcement learning techniques, such as the soft actor–critic (SAC) algorithm [42]. SAC is a model-free, off-policy RL algorithm that integrates policy entropy into the objective function to encourage exploration across various states [43]. Recognized for its effectiveness, SAC has become a benchmark algorithm in RL implementations, outperforming other advanced methods like SQL [44] and TD3 [45] across multiple environments [42,46].
The optimal policy π * is described as follows:
π * = arg max π E s 0 ρ 0 ( s 0 ) a t π ( a t | s t ) s t + 1 T ( s t + 1 | s t , a t ) t = 0 T t e r R t + 1 ( s t , a t ) + α temp H [ π ( · | s t ) ] .
Here, H ( · | s t ) denotes the entropy of the action distribution at state s t , and  ρ 0 is the distribution of the agent’s initial state. The discount factor γ dis ( 0 , 1 ) balances the importance of immediate versus future rewards, and  α temp is the temperature parameter that influences the algorithm’s convergence.
Following the Bellman Equation [15], the soft Q-function is expressed as follows:
Q ( s t , a t ) = r t + 1 + γ dis E s t + 1 T ( s t + 1 | s t , π ( s t ) ) [ V ( s t + 1 ) ] .
The soft value function is derived from the soft Q-function:
V ( s t ) = E a t π ( s t ) [ Q ( s t , a t ) α temp log π ( a t | s t ) ] .
The parameters of the soft Q-function are trained to minimize the temporal difference (TD) error, whereas the policy parameters minimize the Kullback-Leibler divergence between the normalized soft Q-function and the policy distribution.
The critic network’s loss function is defined as follows:
J Q ( θ ) = E ( s t , a t ) D 1 2 ( Q θ ( s t , a t ) ( r + γ dis E s t + 1 T [ V θ ˜ ( s t + 1 ) ] ) ) 2 .
The actor network’s loss function, simplifying the actor’s learning objective, is given in Equation (15):
J π ( ϕ ) = E s t D [ E a t π ϕ [ α temp log π ( a t | s t ) Q θ ( s t , a t ) ] ] .
To manage the sensitivity of the hyperparameter α temp , an automatic adjustment technique is proposed, framing the problem within a maximum entropy RL paradigm with a minimum entropy constraint. This approach, outlined in Equation (16), employs a neural network to approximate α temp , targeting a desired minimum expected entropy H ¯ :
J ( α temp ) = E a t π t [ α temp log π t ( a t | s t ) α temp H ¯ ] .

3. Configuration of One-to-One Air Combat Game in RL

This section outlines the environment setup required for applying RL-based methods to the air combat game optimization problem specified in Equation (2). The environment encapsulates the state space, action space, transition function, and reward function. We also discuss the transformation of the global state space into a relative state space, which simplifies the representation of state information for UCAVs significantly.

3.1. Action Space

The control laws for UCAVs, denoted as u b and u r , are represented by the vector format in Equation (17):
u b = [ α b , μ b , η b ] T , u r = [ α r , μ r , η r ] T .
Constraints on these control variables are defined in Equation (8). The action vectors, which influence the control variables, are designed as shown in Equation (18):
a b = [ α b ˙ , μ b ˙ ] , a r = [ α r ˙ , μ r ˙ ] .
These action vectors modify the control laws at each timestep as outlined in Equation (19), where constraints ensure the actions remain within feasible limits:
{ α ˙ ( t ) [ Δ α , Δ α ] , α ( t ) = α 0 + t 0 t α ˙ ( u ) d u [ α min , α max ] , μ ˙ ( t ) [ Δ μ , Δ μ ] , μ ( t ) = μ 0 + t 0 t μ ˙ ( u ) d u [ π , π ] .
α 0 and μ 0 are the initial attack angle and initial bank angle of the UCAV, respectively.

3.2. State Space

A well-designed state space can facilitate faster convergence of learning algorithms. Referring to [47], the state space for each UCAV can initially be represented as follows:
s = [ x b , y b , z b , v b , γ b , χ b , x r , y r , z r , v r , γ r , χ r ] T .
To reduce the dimensionality and simplify this state information, we transform the global state into a relative state for each UCAV as follows:
s b = [ A T A X O Y b , A T A Y O Z b , A A X O Y b , A A Y O Z b , μ b , D L O S , v b , γ b , v r , α b , z b ] T .
s r = [ A T A X O Y r , A T A Y O Z r , A A X O Y r , A A Y O Z r , μ r , D L O S , v r , γ r , v b , α r , z r ] T .
The angle projection onto the X O Y and Y O Z planes, as illustrated in Figure 4, are denoted by the subscripts X O Y and Y O Z , respectively. D L O S represents the line of sight distance between the UCAVs. Additionally, we normalize the state variables to facilitate neural network processing, ensuring learning efficiency:
s b = [ A T A X O Y b 2 π , A T A Y O Z b 2 π , A A X O Y b 2 π , A A Y O Z b 2 π , μ b 2 π , D L O S D max , v b v max v min , γ b 2 π , v r v max v min , α b α max α min , z b h max h min ] T .
s r = [ A T A X O Y r 2 π , A T A Y O Z r 2 π , A A X O Y r 2 π , A A Y O Z r 2 π , μ r 2 π , D L O S D max , v r v max v min , γ r 2 π , v b v max v min , α r α max α min , z r h max h min ] T .

3.3. Transition Function

To improve the accuracy of our simulation results, we employ the fourth-order Runge-Kutta method [48] for solving the differential equation described in Section 2.2. The specifics of this method are presented in Equation (25), ensuring precise integration of the UCAV dynamics:
k 1 = f ( x ( t ) , t , a ( t ) ) , k 2 = f ( x ( t ) + k 1 Δ t 2 , t + Δ t 2 , a ( t ) ) , k 3 = f ( x ( t ) + k 2 Δ t 2 , t + Δ t 2 , a ( t ) ) , k 4 = f ( x ( t ) + k 3 Δ t , t + Δ t , a ( t ) ) , x ( t + Δ t ) = x ( t ) + 1 6 ( k 1 + 2 k 2 + 2 k 3 + k 4 ) Δ t .
where we consider the system of equations d x ( t , a ) d t = f ( x , t , a ) , with initial condition x ( t 0 ) = x 0 . In this context, the function f ( · ) denotes the rate of change of x when the agent chooses the action a at the time step t.
This method accounts for the nonlinear dynamics of the UCAVs, providing a robust framework for accurate state predictions over the simulation timeframe.

3.4. Reward

In this study, we categorize scenarios involving UCAVs based on Algorithm 1. This classification serves as the foundation for constructing the reward function for UCAV b , as detailed in Equation (26).
R = r 1 , if C b = win , r 2 , if C b = survival and C r = overloaded , r 3 , if C b = overloaded or C b = killed , r 4 , if C b = survival and C r = survival .
This reward structure incentivizes UCAV b to achieve a tactically advantageous position, granting a win reward of r 1 . Should UCAV r become overloaded while UCAV b remains operational, UCAV b is deemed the winner and receives a reward of r 2 . Conversely, if a UCAV is either killed or becomes overloaded, it incurs a penalty of r 3 . Additionally, to encourage expeditious resolution of engagements, a time penalty of r 4 is applied to each UCAV until the end of the game. Figure 5 visually represents the logical framework of the environment.
Algorithm 1 Air combat game criterion of UCAV b
Input: State: s b
Output: Situation of UCAV in air combat game: C b
  1:
D L O S : Euclidean distance between UCAV
  2:
if     h b [ h min , h max ] or n ( α , v ) > 10 g or v b [ v min , v m a x ]  then
  3:
      C b = o v e r l o a d e d
  4:
else if    | A A r |   < 60  and  | A T A r |   < 60  and  D L O S [ D min , D max ]  then
  5:
      C b = k i l l e d
  6:
else if    | A A b |   < 60  and  | A T A b |   < 60  and  D L O S [ D min , D max ]  then
  7:
      C b = w i n
  8:
else
  9:
      C b = s u r v i v a l
10:
end if

4. Homotopy-Based Soft Actor–Critic Method

Learning a policy for achieving an advantageous posture using only sparse rewards and random exploration is a formidable challenge for any agent. Alternatively, relying solely on artificial prior knowledge may prevent the discovery of the optimal policy. To navigate these challenges, we propose a novel homotopy-based soft actor–critic (HSAC) algorithm that leverages the benefits of both artificial prior knowledge and a sparse reward formulation. Initially, the algorithm employs artificial prior knowledge to establish a suboptimal policy. Subsequently, it follows a homotopic path in the solution space, guiding a sequence of suboptimal policies to converge toward the optimal policy. This method effectively balances exploration and exploitation, leading to a robust policy capable of securing an advantageous posture.

4.1. Homotopy-Based Reinforcement Learning

The challenge of reinforcement learning with sparse rewards, denoted as R, is modeled as a nonlinear programming (NLP) problem, referred to as NLP1, following the framework proposed by Bertsekas [49].
  • NLP1 (Original Problem)
max ϕ F 1 ( ϕ ) = E s 0 ρ 0 ( s 0 ) t = 0 T γ dis t [ R ( s t , a t ) ] , s . t . h j ( ϕ ) = 0 , j = 1 , , m , g j ( ϕ ) 0 , j = 1 , , s .
Here, a t π ( · | s t ) , and  ϕ denotes the parameters of the actor network. The functions h ( · ) = 0 and g ( · ) 0 correspond to the environment’s equality and inequality constraints, respectively.
To address the sparse reward challenge in reinforcement learning, researchers often introduce artificial priors that generate an additional reward signal, R e x t r a , to aid agents in policy discovery [23,24,25]. This enhancement improves the feedback on the agent’s actions at each step, facilitating learning. Thus, we define the total reward as the sum of R and R e x t r a , and formulate the enhanced reinforcement learning problem as NLP2.
  • NLP2 (Auxiliary Problem)
max ϕ F 2 ( ϕ ) = E s 0 ρ 0 ( s 0 ) t = 0 T γ dis t [ R ( s t , a t ) + R e x t r a ( s t , a t ) ] , s . t . h j ( ϕ ) = 0 , j = 1 , · · · , m g j ( ϕ ) 0 , j = 1 , · · · , s
The feasible policy derived from NLP2 is not the optimal solution to NLP1 due to distortions caused by the artificial priors. To achieve convergence to the optimal solution of the original problem, we recommend a gradual transition using homotopy. We introduce a function operator F 3 ( ϕ , q ) :
F 3 ( ϕ , q ) = ( 1 q ) F 2 ( ϕ ) + q F 1 ( ϕ ) ,
where q [ 0 , 1 ] . This function operator facilitates the formulation of a homotopy NLP problem as depicted in NLP3.
  • NLP3 (Homotopy Problem)
max ϕ F 3 ( ϕ , q ) = ( 1 q ) F 2 ( ϕ ) + q F 1 ( ϕ ) , s . t . h j ( ϕ ) = 0 , j = 1 , · · · , m g j ( ϕ ) 0 , j = 1 , · · · , s
It is evident that
F 3 [ ϕ , q ] | q = 0 = F 2 [ ϕ ] , F 3 [ ϕ , q ] | q = 1 = F 1 [ ϕ ] .
By adjusting q from 0 to 1, we can derive a continuous set of solutions transitioning from the solution to NLP2 to the solution to NLP1.
We propose a homotopic reward function, R H ( s t , a t , q ) , that integrates both sparse and extra rewards proportionally based on q:
R H ( s t , a t , q ) = q [ R ( s t , a t ) + R e x t r a ( s t , a t ) ] + ( 1 q ) R ( s t , a t ) .
The equivalence of maximizing the expected accumulated homotopic reward R H to solving the NLP3 problem is demonstrated in Theorem 1, with proof provided in Appendix A.
Theorem 1. 
F 3 ( · , · ) equals to the expectation of accumulated homotopic reward as shown in Equation (30).
F 3 ( ϕ , q ) = E s 0 ρ 0 ( s 0 ) t = 0 T γ dis t [ R H ( s t , a t , q ) ] .

4.2. Homotopic Path Following with Predictor–Corrector Method

The solution to the auxiliary problem (NLP2) can be seamlessly transformed into the solution for the original problem (NLP1) using the function F 3 ( · ) , as assured by Theorem 2.
Theorem 2. 
Under Assumption A1, a continuous homotopic path necessarily exists between the solutions to NLP2 and NLP1.
Here, a “path” denotes a piecewise differentiable curve within the solution space. The proof for the existence of this homotopic path is provided in Appendix B.
This section utilizes a classical method known as the horizontal corrector and elevator predictor [50] to navigate from a feasible solution toward the optimal solution for the original problem (NLP1), tracing the homotopic path in the solution space.
Initially, we identify the auxiliary problem (NLP2) as M 0 and label the original problem (NLP1) as M T . The goal of M 0 is to determine the optimal policy π M 0 * ( ϕ ) , maximizing cumulative reward. Theoretically, by following a feasible homotopic path, we can transition to obtaining the optimal policy π M T * ( ϕ ) for the original task.
Additionally, we introduce a parameter N, representing the number of steps required for q to complete the transition from the auxiliary to the original solution. This sequence is defined by q 0 , q 1 , q 2 , , q N , where q 0 = 0 < q 1 < q 2 < < q N = 1 . This definition allows for a sequence of corresponding sub-tasks M 0 , M 1 , , M N , where M N = M T . The pair consisting of the parameter of the actor network ϕ and the value of q at step n is denoted as y n = ( ϕ n , q n ) .

4.2.1. Predictor

If the optimal policy π M n * ( ϕ ) has been determined for task M n at step n, we can predict the parameter ϕ n + 1 * for the subsequent optimal policy π M n + 1 * ( ϕ ) on the homotopic path using the elevator predictor approach [50], as shown in Equation (31).
y ^ n + 1 = ( ϕ n * , q n + 1 ) .

4.2.2. Corrector

After predicting for the new task M n + 1 , the predicted parameter tuple y ^ n + 1 may not exactly align with the homotopic path. To re-align, the corrector step adjusts the predicted tuple y ^ n + 1 back to the true solution tuple y n + 1 using a technique known as the horizontal corrector. During this correction, q n + 1 remains constant.
The solution’s accuracy is the sole focus during this process, as RL-based methods inherently manage correction through gradient descent. The convergence criteria are established by the transformation of the homotopy problem (NLP3) into the condition H ( y n ) = 0 , with the precise conditions detailed in Appendix B.
H ( y n ) ε .
Here, ε denotes the convergence threshold for the criterion.

4.3. Algorithm

The convergence proof for the predictor–corrector path-following method is delineated in Appendix B, and Theorem 3 posits the following principle:
Theorem 3. 
As the parameter q n transitions from 0 to 1 using the horizontal corrector and elevator predictor methods, the solution π M n * ( ϕ ) for task M n may converge to the optimal solution π M T * ( ϕ ) for task M T .
Inspired by this theorem, we have developed the homotopy-based soft actor–critic (HSAC) algorithm, which is encapsulated in Algorithm 2. This approach utilizes the variance trend of the parameter ϕ within the actor network of the SAC algorithm, denoted as k π , to assess policy convergence. The slope k π of the policy gradient’s changes is computed by fitting the policy gradient data to a first-order function using the least squares method, subsequently utilizing this slope to evaluate the quality of convergence within task M n .
Algorithm 2 includes a buffer, P , to store the gradient of the policy π for computing k π . M denotes the sample size for the least squares method, while N represents the total number of iterations required for the homotopy method.
During the horizontal corrector step, the policy parameter ϕ within π M n ( ϕ ) is iteratively adjusted to approximate the optimal policy π M n * ( ϕ ) for task M n . If the slope k π meets the criterion | k π |   < ε , the  n t h corrector step is completed. Otherwise, the iteration of the policy parameter ϕ continues.
The elevator predictor step begins by clearing the policy gradient buffer P to accommodate new data for the subsequent task M n + 1 . The parameter ϕ from the optimal policy π M n * ( ϕ ) of the preceding task is directly transferred to the policy for the new task M n + 1 .
This structured approach ensures rapid convergence to a feasible policy for the initial task M 0 using the additional reward R e x t r a at the training’s commencement. The undesirable effects of this extra reward are progressively mitigated through iterations using the corrector-predictor method, leading to the attainment of the optimal policy for the target task M T as the weight q transitions from 0 to 1.
Algorithm 2 Homotopy-based soft actor–critic
Input: Initialized Parameters: { θ 1 , θ 2 , ϕ , M, N}
Output: Optimized parameters: { θ 1 , θ 2 , ϕ }
  1:
Initialize target network weights:
  2:
θ 1 ¯ θ 1 , θ 2 ¯ θ 2 , ε > 0
  3:
Initialize the auxiliary weight:
  4:
q 0 1 , n 0
  5:
Empty the Replay Buffer:
  6:
D
  7:
Empty the gradient of policy π Buffer:
  8:
P , X [ 1 , 2 , 3 , , M ] T
  9:
X Add A Column With Ones ( X )
10:
for each episode do
11:
      for each step in the episode do
12:
              a t π ϕ ( · | s t )
13:
              s t + 1 , R t + 1 H ( k n ) T ( · | s t , a t )
14:
              D D ( s t , a t , r t + 1 , s t + 1 )
15:
      end for
16:
      for each update step do
17:
             Horizontal Corrector:
18:
              θ i θ i λ Q ^ θ i J Q ( θ i ) for i { 1 , 2 }
19:
              ϕ ϕ λ π ^ ϕ J π ( ϕ , arg min θ i Q θ i )
20:
              α temp α temp λ α temp ^ α temp J α temp ( α temp )
21:
              θ i ¯ τ θ i + ( 1 τ ) θ i ¯ for i { 1 , 2 }
22:
              P P ^ ϕ J π
23:
             Elevator Predictor:
24:
             if  len ( P ) M  then
25:
                    k π ( X T X ) 1 X T P
26:
                   if  | k π | < ε  and  q n > 0  then
27:
                          q n + 1 q n 1 N
28:
                          n n + 1
29:
                          P
30:
                   end if
31:
             end if
32:
      end for
33:
end for

4.4. The Application of HSAC in Air Combat Game

In air combat game scenarios, developing a high-quality policy involves finding an equilibrium for the two-target differential game described in Equation (2). To streamline this complex game, we adopt the concept of self-play, where both UCAV b and UCAV r employ the same policy throughout the two-UCAV game task. This method aims to maximize the cumulative reward, aligning with the framework of NLP1, thereby transforming a complicated two-target differential game into a more manageable self-play reinforcement learning problem. This setup facilitates the application of the HSAC method within a self-play environment.
The HSAC approach unfolds in several phases. Initially, a simple reward function R is specified as shown in Equation (26), and the foundational problem is structured according to NLP1. The equality constraints from Equations (4)–(6) are reduced to h ( · ) = 0 . Meanwhile, the inequality constraints listed in Equations (8) and (10) are simplified to g ( · ) 0 .
Next, the additional reward, R e x t r a , is configured in accordance with Equation (33) to develop the auxiliary function F 2 ( · ) . This preparation makes it straightforward to construct the homotopic reward R H using Equation (29), aligning with the execution of the HSAC algorithm to address the original problem.
R e x t r a = Φ T Q Φ , if D L O S > D max , Φ T Q Φ k 2 D L O S D max + D min 1 2 + 1 , elif D L O S D max .
In this model, the vector Φ represents the relative angle, including A T A b and A A b . The matrix Q R 3 × 3 is a diagonal matrix filled with positive weights, and k is the penalty coefficient related to relative distance. The design of R e x t r a specifically accounts for the influence of relative distance only when it is less than the maximum attack range D max .

5. Simulation

This section demonstrates the effectiveness of our HSAC method in addressing sparse reward challenges in reinforcement learning, particularly in air combat game scenarios, through two distinct experiments.
In the first experiment, we focus on a simplified task—the attack on a horizontally flying UCAV—to highlight the advantages of the HSAC method. The second experiment extends the application of HSAC combined with self-play in a complex two-UCAVs game task, illustrating its potential in more dynamic and intricate scenarios.
We use two variants of the SAC algorithm for comparison: SAC-s, which utilizes only the original reward R, and SAC-r, which employs a combined reward of R + R e x t r a .

5.1. Simulation Platform

The simulation environment is built on the Gym framework [51], programmed in Python, and visualized through Unity3D. The data interaction between Gym and Unity3D is facilitated using socket technology. The platform interface is shown in Figure 6.
Training leverages an NVIDIA GeForce GTX 2080TI graphics card with PyTorch acceleration. The framework for the HSAC algorithm utilized by the agent is depicted in Figure 7. The model parameters for the air combat game, as discussed in Section 2.1 and Section 2.2, along with the hyperparameters for the HSAC algorithm, are detailed in Table A1 and Table A2, respectively.

5.2. Attacking Horizontal-Flight UCAV

5.2.1. Task Setting

This experiment is designed to compare the convergence rates of different methods in a straightforward task setup. In this task, UCAV r maintains a horizontal flight path at a random starting position and heading, while UCAV b , controlled by the trained agent, attempts to execute an attack.
The starting point for UCAV b is designated as point o in Figure 8, positioned at a height of 5 km, and aligned with the X axis with an initial heading angle of zero. The initial positioning for UCAV r is within a predefined area, as depicted in Figure 8, with the initial relative distance ranging from 4 km to 6 km and the relative height from −1 km to 1 km. The initial heading angle for the red UCAV varies randomly from 180 to 180 .

5.2.2. Training Process

To test the convergence of the methods and the performance in the original problem, We analyze from two perspectives: the training process and the evaluation process. During the training process, the episode reward comprises both the original reward R and additional rewards such as R e x t r a and R H . In contrast, during the evaluation process, only the original reward R is considered. We periodically evaluate the quality of the policy every ten training episodes.
The models are trained for 5 × 10 4 episodes as illustrated in Figure 9. Changes in episode rewards for different methods during the training process are displayed in Figure 9a, and those during the evaluation process are presented in Figure 9b.
As shown in Figure 9a, the episode rewards during the training process generally rise with iteration, indicating that agents trained by different methods are progressively mastering the task. Notably, the agent trained with SAC-r records the lowest episode rewards, which is attributable to the negative design of R e x t r a .
To gauge the training quality of each agent, attention must be directed to Figure 9b. Here, it is evident that the agent employing HSAC achieves the highest episode rewards compared to those trained by SAC-s and SAC-r.
Although R e x t r a in SAC-r is crafted using artificial priors, the agent trained with SAC-r still attains lower episode rewards than the agent trained with SAC-s. Conversely, our HSAC method, using the same artificial priors, outperforms both SAC-s and SAC-r. This demonstrates that HSAC is not adversely affected by the negative aspects of artificial priors and can effectively leverage them to address the challenges posed by sparse reward RL problems with the aid of R H .
Subsequent analyses will compare the performance of agents trained by these methods in various air combat game scenarios post-training.

5.2.3. Simulation Results

The initial state space of air combat game geometry can be divided into four categories [24] from the perspective of UCAV b : head-on, neutral, disadvantageous, and advantageous, as illustrated in Figure 10. Depending on these initial conditions, the win probability for UCAV b varies significantly. For instance, UCAV b is more likely to win in an advantageous position, while the odds are reversed in a disadvantageous scenario. In neutral and head-on scenarios, the win probabilities for both sides are similar. We evaluate the performance of each well-trained agent across these diverse initial scenarios.
We selected four representative initial states from each category, detailed in Table A3. The initial bank angle μ and path angle γ for both UCAV b and UCAV r are set to zero. The performance metrics, including the win rates and average time cost per episode for the agents trained by different methods, are captured in Figure 11 and Table 1.
As illustrated in Table 1, agents trained by SAC-s, SAC-r, and HSAC can successfully complete the task when starting from an advantageous position. The agent using SAC-r achieves this task faster than others in advantageous situations. However, in disadvantageous and head-on initial scenarios, the agent trained by HSAC significantly outperforms the others, with the highest success rates. In a disadvantageous scenario, the agent trained by SAC-s has only a 6.61 % success rate, while the agent trained by SAC-r fails to succeed at all. In contrast, in head-on scenarios, while the agent trained by SAC-s has a 73.32 % success rate, the one trained by SAC-r only achieves 7.43 % . Although both the SAC-s and HSAC-trained agents show equal winning rates in neutral scenarios, the HSAC-trained agent completes the task more quickly.
In summary, while SAC-s fairly represents the original task, it struggles with sparse reward issues. SAC-r performs well only in advantageous scenarios due to the bias introduced by the extra rewards. However, HSAC excels in all tested scenarios, outperforming SAC-s and SAC-r by a substantial margin. HSAC effectively harnesses artificial priors without being negatively impacted by them, demonstrating its robustness across varying initial air combat game scenarios.

5.3. Self-Play Training and Two-UCAVs Game Task

5.3.1. Two-UCAVs Game Task Setting

This experiment aims to demonstrate the effectiveness of the HSAC method combined with self-play in the two-UCAVs game task. Integrating self-play with HSAC, as described in Section 4.4, allows for the development of a more sophisticated UCAV policy under air combat game conditions, optimizing the policy through competitive interactions with itself. Figure 12 illustrates the self-play training setup, where both UCAV b and UCAV r share the same actor and critic networks. Furthermore, they share the experience replay buffer and the gradient of the policy buffer, which is essential for HSAC. The initial setup follows the parameters of the attacking horizontal-flight UCAV task outlined in Section 5.2.1. To enhance the realism of radar detection, we set the initial line-of-sight angle component for UCAV r in the X O Y plane between 60 and 60 . While this setup gives UCAV r an initial advantage, it serves to test if the policy can converge to an equilibrium point within the policy space, ideally demonstrating balanced offensive and defensive capabilities from any starting position.
Following the training, the performance of agents trained by SAC-s, SAC-r, and HSAC will be compared in direct confrontations.

5.3.2. Training Process

The training processes for SAC-s, SAC-r, and HSAC are depicted in Figure 13 after 10 4 training episodes. Variations in episode rewards for UCAV b and UCAV r during the self-play training are illustrated in Figure 13a. The difference in episode rewards between UCAV b and UCAV r throughout the training process is shown in Figure 13b. Additionally, we record the number of episodes required for method convergence, the rewards upon convergence of both UCAV b and UCAV r , and the difference in these rewards, as summarized in Table 2, using the criteria set forth in Algorithm 2 for convergence determination.
The data in Table 2 and Figure 13 highlight the minimal difference in rewards upon convergence, suggesting that the policy developed through HSAC aligns closely with an equilibrium point, in accordance with Nash’s theories [52], particularly when this difference approaches zero. This outcome underscores the efficacy of HSAC in rapidly achieving the policy that performs robustly in both offensive and defensive scenarios, as evidenced by the lowest convergence episode cost and the smallest reward discrepancies.
Post-training analysis reveals that HSAC not only converges to an equilibrium point more effectively than the other methods but also does so with superior performance consistency between the competing agents. We will next examine whether the policy honed through HSAC training outperforms those trained via other methods in subsequent head-to-head tests.

5.3.3. Simulation Results

We assess the performance of HSAC through 1 vs. 1 games played among agents trained by SAC-s, SAC-r, and HSAC. Each agent alternates between playing on the red side and the blue side over 10 5 games. The winning rates are detailed in Table 3, with DN representing the number of draws.
Table 3 shows that when UCAV b and UCAV r operate under the same policy, players trained by SAC-r seldom achieve draws, likely due to the disproportionate reward structures which favor aggressive tactics over balanced strategic play. This undermines their ability to perform defensively, rendering SAC-r less suited for self-play training.
In contrast, the HSAC-trained agents demonstrate almost equal win rates in head-to-head matchups, reflecting robust performance across both aggressive and defensive scenarios. Notably, HSAC consistently outperforms the other methods, showcasing superior adaptability and strategic depth.
Figure 14 depicts these interactions, demonstrating how HSAC-trained agents often reach equilibrium states, effectively balancing aggressive and defensive strategies.
In summary, while SAC-s often struggles with the sparse reward problem, leading to protracted exploratory phases and suboptimal local convergence, SAC-r tends to skew learning due to its reward design. HSAC, on the other hand, utilizes homotopic techniques to smoothly transition from exploratory to optimal strategies, effectively navigating the sparse reward landscape without deviating from the task’s objectives.

6. Conclusions

In this study, we introduced the HSAC method, a novel reinforcement learning approach specifically designed to tackle the challenges associated with sparse rewards in complex RL tasks such as those found in air combat game simulations. The HSAC method begins by integrating artificial reward priors to enhance exploration efficiency, allowing for the rapid development of an initial feasible policy. As training progresses, these artificial rewards are gradually reduced to ensure that the learning process remains focused on achieving the primary objectives, thus following a homotopic path toward the optimal policy.
The results of our experiments, conducted within a sophisticated 3D simulation environment that accurately models both offensive and defensive maneuvers, demonstrate that agents trained with HSAC significantly outperform those trained using the SAC-s and SAC-r methods. Specifically, HSAC-trained agents exhibit superior combat skills and achieve higher success rates, highlighting the algorithm’s ability to effectively balance exploration and exploitation in sparse reward settings.
HSAC not only accelerates the convergence process by efficiently navigating the sparse reward landscape but also maintains robustness across various initial conditions and scenarios. This adaptability is particularly evident in our self-play experiments, where HSAC consistently achieves near-equilibrium strategies, showcasing its potential for both competitive and cooperative multi-agent environments.
Looking forward, we plan to extend the application of HSAC to multi-agent systems, further enhancing its capabilities in complex, dynamic environments. By doing so, we aim to improve its proficiency in both competitive and cooperative contexts, paving the way for broader applications in real-world scenarios.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/drones8120756/s1, Video S1: HSAC.

Author Contributions

Conceptualization, Z.F., Y.Z. (Yiwen Zhu) and Y.Z. (Yuan Zheng); methodology, Y.Z. (Yiwen Zhu); software, Y.Z. (Yiwen Zhu); validation, Y.Z. (Yuan Zheng) and W.W.; formal analysis, Y.Z. (Yuan Zheng) and Y.Z. (Yiwen Zhu); investigation, Y.Z. (Yiwen Zhu); resources, Z.F.; data curation, Y.Z. (Yiwen Zhu); writing—original draft preparation, Y.Z. (Yiwen Zhu); writing—review and editing, Y.Z. (Yuan Zheng) and Z.F.; visualization, Y.Z. (Yiwen Zhu); supervision, Z.F.; project administration, Z.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (No. U2333214).

Data Availability Statement

Data available on request from the authors.

DURC Statement

Current research is limited to the Intelligent Decision Making for Drones, which is beneficial for the further enhancement of agent intelligence and does not pose a threat to public health or national security. The authors acknowledge the dual-use potential of the research involving Intelligent Decision Making for Drones and confirm that all necessary precautions have been taken to prevent potential misuse. As an ethical responsibility, authors strictly adhere to relevant national and international laws about DURC. Authors advocate for responsible deployment, ethical considerations, regulatory compliance, and transparent reporting to mitigate misuse risks and foster beneficial outcomes.

Acknowledgments

We would like to express our sincere gratitude to Zheng Chen from the Institute of Unmanned Aerial Vehicles at Zhejiang University for his invaluable guidance and support. We are also grateful for the technical and equipment support provided by the Institute. Furthermore, we would like to thank the China Scholarship Council for their financial support during the course of this work.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Proof of the Equation F3 (·)

The equation of F 3 ( · ) can be expanded as shown in Equation (A1).
F 3 ( ϕ , q ) = ( 1 q ) F 2 ( ϕ ) + q F 1 ( ϕ ) , = ( 1 q ) E s 0 ρ 0 ( s 0 ) t = 0 T γ dis t [ R ( s t , a t ) + R e x t r a ( s t , a t ) ] + q E s 0 ρ 0 ( s 0 ) t = 0 T γ dis t [ R ( s t , a t ) ) ] , = E s 0 ρ 0 ( s 0 ) t = 0 T γ dis t [ ( 1 q ) R ( s t , a t ) + R e x t r a ( s t , a t ) + q R ( s t , a t ) ] , = E s 0 ρ 0 ( s 0 ) t = 0 T γ dis t R H ( s t , a t , q ) .
The homotopy NLP problem mentioned in NLP3 can be described as follow:
max ϕ F 3 ( ϕ , q ) = E s 0 ρ 0 ( s 0 ) t = 0 T γ dis t [ R H ( s t , a t , q ) ] , s . t . h j ( ϕ ) = 0 , j = 1 , · · · , m g j ( ϕ ) 0 , j = 1 , · · · , s

Appendix B. Homotopic Path Existence and Convergence of the Path Following Method

Appendix B.1. Formalize the Problem as H(·) = 0

The s in the NLP problem of F 3 ( · ) , in Appendix A, can also be described by a state transition function T ( · ) :
s t = T ( s t 1 , π ϕ ( s t 1 ) ) , = T ( T ( s t 2 , π ϕ ( s t 2 ) ) , π ϕ ( s t 1 ) ) , = T ( T ( · · · T ( s 0 , π ϕ ( s 0 ) ) ) , π ϕ ( T ( · · · T ( s 0 , π ϕ ( s 0 ) ) ) ) ) , = T ( s 0 , t , ϕ ) .
which means that the state s only relates to the initial state s 0 , time step t, and the parameter of the policy ϕ . Here, we use an unknown operator T to describe this relationship.
Furthermore, we can also prove the function F 3 ( · ) only relates to parameter ϕ and q, as shown in Equation (A4)
F 3 ( ϕ , q ) = E s 0 ρ 0 ( s 0 ) t = 0 T γ dis t R H s t , π ϕ ( s t ) , q , = s 0 ρ 0 ( s 0 ) t = 0 T γ dis t R H T ( s 0 , t , ϕ ) , π ϕ ( T ( s 0 , t , ϕ ) ) , q , = s 0 ρ 0 ( s 0 ) t = 0 T γ dis t R s 0 , t , ϕ , q , = F ( ϕ , q ) .
Here, we use the operator F to describe the relationship between the expectation of accumulated homotopic reward with the parameter ϕ and q.
Then using the Karush–Kuhn–Tucker (KKT) equations [53] to formalize the problem described in Equation (A2) as Equation (A5)
H ( ϕ , β , ζ , q ) = ϕ F ( ϕ , q ) + j = 1 s β j + ϕ g j ( ϕ ) + j = 1 m ζ j ϕ h j ( ϕ ) β 1 g 1 ( ϕ ) · · · β s g s ( ϕ ) h 1 ( ϕ ) · · · h m ( ϕ ) .
Here, H ( · ) is called the homotopy function. And H : R n + m + s + 1 R n + m + s because ϕ R n , β R s , ζ R m , and q R 1 . We let β j + = m a x 0 , β j 3 and β j = m a x 0 , β j 3 , these and ζ here are only the auxiliary variables. Hence, the problem mentioned in NLP3 is equivalent to solving the KKT equation H ( ϕ , β , ζ , q ) = 0 .

Appendix B.2. Existence of Homotopy Path

The idea of the homotopy method is to follow a path to a solution, where the “path” indicates a piecewise differentiable curve in solution space.
Zangwill [50] defined the set of all solutions as H 1 .
Definition A1. 
Given a homotopy function H : R n + 1 R n , we must now be more explicit about solutions to
H ( ϕ , β , ζ , q ) = 0 .
In particular, define
H 1 = ( ϕ , β , ζ , q ) | H ( ϕ , β , ζ , q ) = 0 .
as the set of all solutions ( ϕ , β , ζ , q ) R n + 1
The implicit function theorem can ensure that H 1 consists solely of paths [50]. The Jacobian of homotopy function H ( ϕ , β , ζ , q ) can be written as an ( n + m + s ) × ( n + m + s + 1 ) matrix, as shown in Equation (A8).
H ( ϕ , β , ζ , q ) = H 1 ϕ 1 · · H 1 ϕ n H 1 β 1 · · H 1 β s H 1 ζ 1 · · H 1 ζ m H 1 q · · · · · · · · · · · · · · H n + m + s ϕ 1 · · H n + m + s ϕ n H n + m + s β 1 · · H n + m + s β s H n + m + s ζ 1 · · H n + m + s ζ m H n + m + s q .
Then, the existence of the path was given in [50]:
Theorem A1 
(Path Existence ([50])). Let H : R n + m + s + 1 R n + m + s be continuously differentiable and suppose that for every ( ϕ , β , ζ , q ) H 1 , the Jacobian H ( ϕ , β , ζ , q ) is of full rank. Then, H 1 consists only of continuously differentiable paths.
As the other scholars did in homotopy optimization [50], we give an assumption of the Jacobian matrix H ( · ) :
Assumption A1. 
H ( ϕ , β , ζ , q ) is of full rank for all ( ϕ , β , ζ , q ) H 1 .
Combining the Assumption A1 and Theorem A1, the existence of the homotopic path in this problem can be assured.

Appendix B.3. Convergence of Path Following Method

The homotopic path existence has been proved in Appendix B.2. And then we will give the proof of the convergence and the feasibility of the path following method. In this paper, the corrector-predictor method, with horizontal corrector and elevator predictor, is used to follow the homotopic path along with the iterations. To ensure the convergence and the feasibility of this corrector-predictor path-following method, we quote Theorem A2 proposed by Zangwill [50].
Theorem A2 
(Method Convergence). For a homotopy H C 2 at any y H 1 , let H ϕ ( ϕ , q ) be of full rank. Also, for ( ϕ 0 , q 0 ) H 1 , let the path be of finite length. Now, suppose that a predictor–corrector method uses the horizontal corrector and the elevator predictor.
Given ε 1 > ε 2 > 0 sufficiently small, if for all iterations k the predictor step length d k is sufficiently small, then the method will follow the entire path length to any degree of accuracy.
In the traditional predictor-corrector method [50], Equations (A9) and (A10) are used to decide if the solution satisfies the accuracy requirement in the predictor and corrector step.
H ( ϕ n , β n , ζ n , q n ) < ε 1 .
H ( ϕ n + 1 , β n + 1 , ζ n + 1 , q n + 1 ) < ε 2 .
In this paper, we suppose ε 1 = ε 2 = ε to be convenient to calculate.
Through Theorem A2, we can guarantee the convergence and the feasibility of using HSAC in the original problem (NLP1), with the help of the function F 3 ( · ) .

Appendix C. Parameters

The parameters of UCAV and the criterion of the air combat game scenario, mentioned in Section 2.1 and Section 2.2, are given in Table A1.
The hyperparameters of the HSAC algorithm, which we mentioned in Section 4.3, are given in Table A2.
In Table A3, we gave four typical initial states, which were used to evaluate the quality of the policy trained by different methods in the attacking horizontal-flight UCAV task.
Table A1. Air combat game simulation parameters.
Table A1. Air combat game simulation parameters.
Single UCAV ParameterValue
Mass of UCAV150 kg
Max Load Factor ( n m a x ) 10 g
Velocity Band [ 80 , 400 ] m/s
Height Range [ 2000 , 8000 ] m
Max Thrust ( T m a x ) 100 kg
Range of α ˙ [ 5 , 5 ] deg/s
Range of μ ˙ [ 50 , 50 ] deg/s
Range of α [ 15 , 15 ] deg
Game ParameterValue
Optimum Attack Range ( d m i n , d m a x ) (200, 2000) m
Maximum Iteration Step2000
Table A2. Hyperparameters of HSAC.
Table A2. Hyperparameters of HSAC.
ParameterValue
OptimizerAdam
Learning Rate 3 × 10 4
Discount ( γ dis )0.996
Number of Hidden Layers (All Networks)3
Number of Hidden Units per Layer256
Number of Samples per Minibatch256
NonlinearityReLU
Replay Buffer size 10 6
Entropy Target−dim(A)
Target Smoothing Coefficient ( τ )0.005
Policy Gradient Buffer (M) 10 4
Total Number of Homotopy Iterations (N)100
Threshold of k π ( ε ) 10 5
Table A3. Initial state settings for evaluating the quality of agents.
Table A3. Initial state settings for evaluating the quality of agents.
Initial State X ( m ) Y ( m ) H ( m ) V ( m / s ) χ ( )
AdvantageousBlue00500015045
Red50005000500015045
DisadvantageousBlue005000150−45
Red−500050005000150−45
Head-OnBlue00500015045
Red500050005000150−135
NeutralBlue00500015045
Red5000−50005000150−135

References

  1. Xu, G.; Wei, S.; Zhang, H. Application of situation function in air combat differential games. In Proceedings of the 2017 36th Chinese Control Conference (CCC), Dalian, China, 26–28 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 5865–5870. [Google Scholar]
  2. Park, H.; Lee, B.Y.; Tahk, M.J.; Yoo, D.W. Differential game based air combat maneuver generation using scoring function matrix. Int. J. Aeronaut. Space Sci. 2016, 17, 204–213. [Google Scholar] [CrossRef]
  3. Virtanen, K.; Karelahti, J.; Raivio, T. Modeling air combat by a moving horizon influence diagram game. J. Guid. Control Dyn. 2006, 29, 1080–1091. [Google Scholar] [CrossRef]
  4. Zhong, L.; Tong, M.; Zhong, W.; Zhang, S. Sequential maneuvering decisions based on multi-stage influence diagram in air combat. J. Syst. Eng. Electron. 2007, 18, 551–555. [Google Scholar]
  5. Ortiz, A.; Garcia-Nieto, S.; Simarro, R. Comparative Study of Optimal Multivariable LQR and MPC Controllers for Unmanned Combat Air Systems in Trajectory Tracking. Electronics 2021, 10, 331. [Google Scholar] [CrossRef]
  6. Smith, R.E.; Dike, B.; Mehra, R.; Ravichandran, B.; El-Fallah, A. Classifier systems in combat: Two-sided learning of maneuvers for advanced fighter aircraft. Comput. Methods Appl. Mech. Eng. 2000, 186, 421–437. [Google Scholar] [CrossRef]
  7. Changqiang, H.; Kangsheng, D.; Hanqiao, H.; Shangqin, T.; Zhuoran, Z. Autonomous air combat maneuver decision using Bayesian inference and moving horizon optimization. J. Syst. Eng. Electron. 2018, 29, 86–97. [Google Scholar]
  8. Shenyu, G. Research on Expert System and Decision Support System for Multiple Air Combat Tactical Maneuvering. Syst. Eng.-Theory Pract. 1999, 8, 76–80. [Google Scholar]
  9. Zhao, W.; Zhou, D.Y. Application of expert system in sequencing of air combat multi-target attacking. Electron. Opt. Control 2008, 2, 23–26. [Google Scholar]
  10. Bechtel, R.J. Air Combat Maneuvering Expert System Trainer; Technical Report; Merit Technology Inc.: Plano, TX, USA, 1992. [Google Scholar]
  11. Xu, J.; Zhang, J.; Yang, L.; Liu, C. Autonomous decision-making for dogfights based on a tactical pursuit point approach. Aerosp. Sci. Technol. 2022, 129, 107857. [Google Scholar] [CrossRef]
  12. Rodin, E.Y.; Amin, S.M. Maneuver prediction in air combat via artificial neural networks. Comput. Math. Appl. 1992, 24, 95–112. [Google Scholar] [CrossRef]
  13. Schvaneveldt, R.W.; Goldsmith, T.E.; Benson, A.E.; Waag, W.L. Neural Network Models of Air Combat Maneuvering; Technical Report; New Mexico State University: Las Cruces, NM, USA, 1992. [Google Scholar]
  14. Teng, T.H.; Tan, A.H.; Tan, Y.S.; Yeo, A. Self-organizing neural networks for learning air combat maneuvers. In Proceedings of the 2012 International Joint Conference on Neural Networks (IJCNN), Brisbane, Australia, 10–15 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 1–8. [Google Scholar]
  15. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  16. Ding, Y.; Kuang, M.; Shi, H.; Gao, J. Multi-UAV Cooperative Target Assignment Method Based on Reinforcement Learning. Drones 2024, 8, 562. [Google Scholar] [CrossRef]
  17. Yang, J.; Yang, X.; Yu, T. Multi-Unmanned Aerial Vehicle Confrontation in Intelligent Air Combat: A Multi-Agent Deep Reinforcement Learning Approach. Drones 2024, 8, 382. [Google Scholar] [CrossRef]
  18. Gao, X.; Zhang, Y.; Wang, B.; Leng, Z.; Hou, Z. The Optimal Strategies of Maneuver Decision in Air Combat of UCAV Based on the Improved TD3 Algorithm. Drones 2024, 8, 501. [Google Scholar] [CrossRef]
  19. Guo, J.; Zhang, J.; Wang, Z.; Liu, X.; Zhou, S.; Shi, G.; Shi, Z. Formation Cooperative Intelligent Tactical Decision Making Based on Bayesian Network Model. Drones 2024, 8, 427. [Google Scholar] [CrossRef]
  20. Chen, C.L.; Huang, Y.W.; Shen, T.J. Application of Deep Reinforcement Learning to Defense and Intrusion Strategies Using Unmanned Aerial Vehicles in a Versus Game. Drones 2024, 8, 365. [Google Scholar] [CrossRef]
  21. McGrew, J.S.; How, J.P.; Williams, B.; Roy, N. Air-combat strategy using approximate dynamic programming. J. Guid. Control Dyn. 2010, 33, 1641–1654. [Google Scholar] [CrossRef]
  22. Crumpacker, J.B.; Robbins, M.J.; Jenkins, P.R. An approximate dynamic programming approach for solving an air combat maneuvering problem. Expert Syst. Appl. 2022, 203, 117448. [Google Scholar] [CrossRef]
  23. Ma, X.; Xia, L.; Zhao, Q. Air-combat strategy using Deep Q-Learning. In Proceedings of the 2018 Chinese Automation Congress (CAC), Xi’an, China, 30 November–2 December 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 3952–3957. [Google Scholar]
  24. Wang, Z.; Li, H.; Wu, H.; Wu, Z. Improving maneuver strategy in air combat by alternate freeze games with a deep reinforcement learning algorithm. Math. Probl. Eng. 2020, 2020, 7180639. [Google Scholar] [CrossRef]
  25. Yang, Q.; Zhang, J.; Shi, G.; Hu, J.; Wu, Y. Maneuver decision of UAV in short-range air combat based on deep reinforcement learning. IEEE Access 2019, 8, 363–378. [Google Scholar] [CrossRef]
  26. Pope, A.P.; Ide, J.S.; Mićović, D.; Diaz, H.; Rosenbluth, D.; Ritholtz, L.; Twedt, J.C.; Walker, T.T.; Alcedo, K.; Javorsek, D. Hierarchical Reinforcement Learning for Air-to-Air Combat. In Proceedings of the 2021 International Conference on Unmanned Aircraft Systems (ICUAS), Athens, Greece, 15–18 June 2021; pp. 275–284. [Google Scholar] [CrossRef]
  27. Ng, A.Y.; Harada, D.; Russell, S. Policy invariance under reward transformations: Theory and application to reward shaping. In Proceedings of the Sixteenth International Conference on Machine Learning (ICML 1999), Bled, Slovenia, 27–30 June 1999; Volume 99, pp. 278–287. [Google Scholar]
  28. Randløv, J.; Alstrøm, P. Learning to Drive a Bicycle Using Reinforcement Learning and Shaping. In Proceedings of the ICML, San Francisco, CA, USA, 24–27 July 1998; Citeseer: Princeton, NJ, USA, 1998; Volume 98, pp. 463–471. [Google Scholar]
  29. Gu, S.; Holly, E.; Lillicrap, T.; Levine, S. Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 3389–3396. [Google Scholar]
  30. Heess, N.; TB, D.; Sriram, S.; Lemmon, J.; Merel, J.; Wayne, G.; Tassa, Y.; Erez, T.; Wang, Z.; Eslami, S.; et al. Emergence of locomotion behaviours in rich environments. arXiv 2017, arXiv:1707.02286. [Google Scholar]
  31. Ghosh, D.; Singh, A.; Rajeswaran, A.; Kumar, V.; Levine, S. Divide-and-conquer reinforcement learning. arXiv 2017, arXiv:1711.09874. [Google Scholar]
  32. Forestier, S.; Portelas, R.; Mollard, Y.; Oudeyer, P.Y. Intrinsically motivated goal exploration processes with automatic curriculum learning. arXiv 2017, arXiv:1708.02190. [Google Scholar]
  33. Ross, S.; Gordon, G.; Bagnell, D. A reduction of imitation learning and structured prediction to no-regret online learning. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 627–635. [Google Scholar]
  34. Vecerik, M.; Hester, T.; Scholz, J.; Wang, F.; Pietquin, O.; Piot, B.; Heess, N.; Rothörl, T.; Lampe, T.; Riedmiller, M. Leveraging demonstrations for deep reinforcement learning on robotics problems with sparse rewards. arXiv 2017, arXiv:1707.08817. [Google Scholar]
  35. Kober, J.; Peters, J. Policy search for motor primitives in robotics. Mach. Learn. 2011, 84, 171–203. [Google Scholar] [CrossRef]
  36. Montgomery, W.H.; Levine, S. Guided policy search via approximate mirror descent. Adv. Neural Inf. Process. Syst. 2016, 29, 4008–4016. [Google Scholar]
  37. Ziebart, B.D.; Maas, A.L.; Bagnell, J.A.; Dey, A.K. Maximum entropy inverse reinforcement learning. In Proceedings of the AAAI, Chicago, IL, USA, 13–17 July 2008; Volume 8, pp. 1433–1438. [Google Scholar]
  38. Shaw, R.L. Fighter Combat. Tactics and Maneuvering; Naval Institute Press: Annapolis, MD, USA, 1985. [Google Scholar]
  39. Grimm, W.; Well, K. Modelling air combat as differential game recent approaches and future requirements. In Differential Games—Developments in Modelling and Computation, Proceedings of the Fourth International Symposium on Differential Games and Applications, Otaniemi, Finland, 9–10 August 1990; Springer: Berlin/Heidelberg, Germany, 1991; pp. 1–13. [Google Scholar]
  40. Blaquière, A.; Gérard, F.; Leitmann, G. Quantitative and Qualitative Games by Austin Blaquiere, Francoise Gerard and George Leitmann; Academic Press: Cambridge, MA, USA, 1969. [Google Scholar]
  41. Silver, D.; Hubert, T.; Schrittwieser, J.; Antonoglou, I.; Lai, M.; Guez, A.; Lanctot, M.; Sifre, L.; Kumaran, D.; Graepel, T.; et al. A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play. Science 2018, 362, 1140–1144. [Google Scholar] [CrossRef]
  42. Haarnoja, T.; Zhou, A.; Abbeel, P.; Levine, S. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Proceedings of the International Conference on Machine Learning. PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 1861–1870. [Google Scholar]
  43. Ziebart, B.D. Modeling Purposeful Adaptive Behavior with the Principle of Maximum Causal Entropy; Carnegie Mellon University: Pittsburgh, PA, USA, 2010. [Google Scholar]
  44. Haarnoja, T.; Tang, H.; Abbeel, P.; Levine, S. Reinforcement learning with deep energy-based policies. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1352–1361. [Google Scholar]
  45. Fujimoto, S.; Hoof, H.; Meger, D. Addressing function approximation error in actor-critic methods. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 1587–1596. [Google Scholar]
  46. Haarnoja, T.; Zhou, A.; Hartikainen, K.; Tucker, G.; Ha, S.; Tan, J.; Kumar, V.; Zhu, H.; Gupta, A.; Abbeel, P.; et al. Soft actor-critic algorithms and applications. arXiv 2018, arXiv:1812.05905. [Google Scholar]
  47. Kong, W.; Zhou, D.; Yang, Z.; Zhao, Y.; Zhang, K. Uav autonomous aerial combat maneuver strategy generation with observation error based on state-adversarial deep deterministic policy gradient and inverse reinforcement learning. Electronics 2020, 9, 1121. [Google Scholar] [CrossRef]
  48. Forsythe, G.E. Computer Methods for Mathematical Computations; Prentice Hall: Upper Saddle River, NJ, USA, 1977. [Google Scholar]
  49. Bertsekas, D. Reinforcement Learning and Optimal Control; Athena Scientific: Nashua, NH, USA, 2019. [Google Scholar]
  50. Lemke, C. Pathways to solutions, fixed points, and equilibria (cb garcia and wj zangwill). Sch. J. 1984, 26, 445. [Google Scholar]
  51. Brockman, G.; Cheung, V.; Pettersson, L.; Schneider, J.; Schulman, J.; Tang, J.; Zaremba, W. OpenAI Gym. arXiv 2016, arXiv:1606.01540. [Google Scholar]
  52. Nash, J. Non-cooperative games. Ann. Math. 1951, 54, 286–295. [Google Scholar] [CrossRef]
  53. Fiacco, A.V.; McCormick, G.P. Nonlinear Programming: Sequential Unconstrained Minimization Techniques; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1990. [Google Scholar]
Figure 1. Overview diagram of HSAC in the air combat game environment.
Figure 1. Overview diagram of HSAC in the air combat game environment.
Drones 08 00756 g001
Figure 2. One-to-one short-range air combat game scenario.
Figure 2. One-to-one short-range air combat game scenario.
Drones 08 00756 g002
Figure 3. Three-degree-of-freedom particle model of the UCAV in the ground coordinate system with dotted lines as visual aids for attitude understanding.
Figure 3. Three-degree-of-freedom particle model of the UCAV in the ground coordinate system with dotted lines as visual aids for attitude understanding.
Drones 08 00756 g003
Figure 4. The diagrammatic sketch of three-dimensional air combat game.
Figure 4. The diagrammatic sketch of three-dimensional air combat game.
Drones 08 00756 g004
Figure 5. The framework of the air combat game environment.
Figure 5. The framework of the air combat game environment.
Drones 08 00756 g005
Figure 6. The simulation platform established by Unity3D and Gym.The blue and red lines in the figure correspond to the flight trajectories of the aircraft.
Figure 6. The simulation platform established by Unity3D and Gym.The blue and red lines in the figure correspond to the flight trajectories of the aircraft.
Drones 08 00756 g006
Figure 7. The construction of the HSAC method.
Figure 7. The construction of the HSAC method.
Drones 08 00756 g007
Figure 8. Initial state of blue and red UCAV.
Figure 8. Initial state of blue and red UCAV.
Drones 08 00756 g008
Figure 9. The episode reward of different methods in attacking horizontal-flight UCAV task, the results of the training process reflect the quality of convergence, and the results of the evaluating process reflect the performance of different methods in the original problem. (a) Illustrates the episode reward in the training process. (b) Denotes the episode reward in the evaluating process.
Figure 9. The episode reward of different methods in attacking horizontal-flight UCAV task, the results of the training process reflect the quality of convergence, and the results of the evaluating process reflect the performance of different methods in the original problem. (a) Illustrates the episode reward in the training process. (b) Denotes the episode reward in the evaluating process.
Drones 08 00756 g009
Figure 10. Four initial air combat game scenario categories.
Figure 10. Four initial air combat game scenario categories.
Drones 08 00756 g010
Figure 11. Performance of agents trained by different methods with different initial situations in the attacking horizontal-flight UCAV task, where red and blue are used to conveniently distinguish between different aircraft: (a) Agent trained by SAC-s in an advantageous situation; (b) Agent trained by SAC-s in a disadvantageous situation; (c) Agent trained by SAC-s in a head-on situation; (d) Agent trained by SAC-s in a neutral situation; (e) Agent trained by SAC-r in an advantageous situation; (f) Agent trained by SAC-r in a disadvantageous situation; (g) Agent trained by SAC-r in a head-on situation; (h) Agent trained by SAC-r in a neutral situation; (i) Agent trained by HSAC in an advantageous situation; (j) Agent trained by HSAC in a disadvantageous situation; (k) Agent trained by HSAC in a head-on situation; (l) Agent trained by HSAC in a neutral situation.
Figure 11. Performance of agents trained by different methods with different initial situations in the attacking horizontal-flight UCAV task, where red and blue are used to conveniently distinguish between different aircraft: (a) Agent trained by SAC-s in an advantageous situation; (b) Agent trained by SAC-s in a disadvantageous situation; (c) Agent trained by SAC-s in a head-on situation; (d) Agent trained by SAC-s in a neutral situation; (e) Agent trained by SAC-r in an advantageous situation; (f) Agent trained by SAC-r in a disadvantageous situation; (g) Agent trained by SAC-r in a head-on situation; (h) Agent trained by SAC-r in a neutral situation; (i) Agent trained by HSAC in an advantageous situation; (j) Agent trained by HSAC in a disadvantageous situation; (k) Agent trained by HSAC in a head-on situation; (l) Agent trained by HSAC in a neutral situation.
Drones 08 00756 g011
Figure 12. The structure of the two-UCAVs game task with the idea of self-play.
Figure 12. The structure of the two-UCAVs game task with the idea of self-play.
Drones 08 00756 g012
Figure 13. The self-play training process of SAC-s, HSAC, and SAC-r in the two-UCAVs game task: (a) Displays the episode rewards for UCAV b and UCAV r across different methods. (b) Shows the difference in episode rewards between UCAV b and UCAV r trained by different methods.
Figure 13. The self-play training process of SAC-s, HSAC, and SAC-r in the two-UCAVs game task: (a) Displays the episode rewards for UCAV b and UCAV r across different methods. (b) Shows the difference in episode rewards between UCAV b and UCAV r trained by different methods.
Drones 08 00756 g013
Figure 14. (a) HSAC-trained agent competing against SAC-s-trained agent; HSAC plays as UCAV b and SAC-s as UCAV r . (b) HSAC-trained agent competing against SAC-r-trained agent; HSAC plays as UCAV b and SAC-r as UCAV r . (c) HSAC-trained agent competes against another HSAC-trained agent, demonstrating the equilibrium strategy effectiveness. Red and blue are used to conveniently distinguish between different aircraft.
Figure 14. (a) HSAC-trained agent competing against SAC-s-trained agent; HSAC plays as UCAV b and SAC-s as UCAV r . (b) HSAC-trained agent competing against SAC-r-trained agent; HSAC plays as UCAV b and SAC-r as UCAV r . (c) HSAC-trained agent competes against another HSAC-trained agent, demonstrating the equilibrium strategy effectiveness. Red and blue are used to conveniently distinguish between different aircraft.
Drones 08 00756 g014
Table 1. Performance of agents trained by different methods in attacking horizontal-flight tasks with different initial states after 10 4 episode evaluations.
Table 1. Performance of agents trained by different methods in attacking horizontal-flight tasks with different initial states after 10 4 episode evaluations.
MethodSAC-sSAC-rHSAC
Initial State
AdvantageousWinning Rate100.00%100.00 %100.00%
Average Time Cost42.07 s37.90 s39.40 s
DisadvantageousWinning Rate 6.61 % 0.00 % 98.38%
Average Time Cost44.66 s26.23 s35.91 s
Head-OnWinning Rate 73.32 % 7.43 % 99.94%
Average Time Cost39.15 s31.02 s24.73 s
NeutralWinning Rate100.00% 3.66 % 100.00%
Average Time Cost40.05 s22.57 s33.20 s
Table 2. The convergence episode cost for three different methods, the episode reward of the converged policy, and the difference value of converged episode reward between UCAV b and UCAV r in the self-play training process.
Table 2. The convergence episode cost for three different methods, the episode reward of the converged policy, and the difference value of converged episode reward between UCAV b and UCAV r in the self-play training process.
SAC-rSAC-sHSAC
BlueRedBlueRedBlueRed
Quality
of Policy
Episode Reward−217.64111.72−1469.09−1339.13−387.52−382.57
Difference Value329.36129.964.95
Convergence Episode Cost (episodes)825066754455
Table 3. The relative winning rates of agents trained by different methods after 10 5 episodes’ evaluation in two-UCAVs game task, where DN means the number of draws.
Table 3. The relative winning rates of agents trained by different methods after 10 5 episodes’ evaluation in two-UCAVs game task, where DN means the number of draws.
RedSAC-sSAC-rHSAC
Blue
SAC-sBlue Win 34.9 % 42.1 % 18.7 %
Red Win 65 . 1 % 57 . 9 % 81 . 3 %
DN43,78537249,939
SAC-rBlue Win 61 . 0 % 30.5 % 39.4 %
Red Win 39.0 % 69 . 5 % 60 . 6 %
DN3281619
HSACBlue Win 66 . 8 % 60 . 8 % 45.4 %
Red Win 33.2 % 39.2 % 54 . 6 %
DN35,44884375,228
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, Y.; Zheng, Y.; Wei, W.; Fang, Z. Enhancing Automated Maneuvering Decisions in UCAV Air Combat Games Using Homotopy-Based Reinforcement Learning. Drones 2024, 8, 756. https://doi.org/10.3390/drones8120756

AMA Style

Zhu Y, Zheng Y, Wei W, Fang Z. Enhancing Automated Maneuvering Decisions in UCAV Air Combat Games Using Homotopy-Based Reinforcement Learning. Drones. 2024; 8(12):756. https://doi.org/10.3390/drones8120756

Chicago/Turabian Style

Zhu, Yiwen, Yuan Zheng, Wenya Wei, and Zhou Fang. 2024. "Enhancing Automated Maneuvering Decisions in UCAV Air Combat Games Using Homotopy-Based Reinforcement Learning" Drones 8, no. 12: 756. https://doi.org/10.3390/drones8120756

APA Style

Zhu, Y., Zheng, Y., Wei, W., & Fang, Z. (2024). Enhancing Automated Maneuvering Decisions in UCAV Air Combat Games Using Homotopy-Based Reinforcement Learning. Drones, 8(12), 756. https://doi.org/10.3390/drones8120756

Article Metrics

Back to TopTop