Next Article in Journal
AI-Driven Safety and Security for UAVs: From Machine Learning to Large Language Models
Previous Article in Journal
A Multi-UAV Formation Obstacle Avoidance Method Combined with Improved Simulated Annealing and an Adaptive Artificial Potential Field
Previous Article in Special Issue
Online Safe Flight Control Method Based on Constraint Reinforcement Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Composite Learning-Based Inverse Optimal Fault-Tolerant Control for Hierarchy-Structured Unmanned Helicopters

1
College of Automation Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing 211106, China
2
National Key Laboratory of Helicopter Dynamics, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China
3
School of Applied Mathematics, Nanjing University of Finance and Economics, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Drones 2025, 9(6), 391; https://doi.org/10.3390/drones9060391
Submission received: 25 March 2025 / Revised: 13 May 2025 / Accepted: 20 May 2025 / Published: 23 May 2025

Abstract

:
This article investigates the inverse optimal fault-tolerant formation-containment control problem for a group of unmanned helicopters, where the leaders form a desired formation pattern under the guidance of a virtual leader while the followers move toward the convex hull established by leaders. To facilitate control design and stability analysis, each helicopter’s dynamics are separated into an outer-loop (position) and an inner-loop (attitude) subsystem by exploiting their multi-time-scale characteristics. Next, the serial-parallel estimation model, designed to account for prediction error, is developed. On this foundation, the composite updating law for network weights is derived. Using these intelligent approximations, a fault estimation observer is constructed. The estimated fault information is further incorporated into the inverse optimal fault-tolerant control framework that avoids tackling either the Hamilton–Jacobi–Bellman or Hamilton–Jacobi–Issacs equation. Finally, simulation results are presented to demonstrate the superior control performance and accuracy of the proposed method.

1. Introduction

Unmanned helicopters (UHs) have been extensively applied in various surveillance and inspection tasks due to their exceptional abilities in hovering, vertical take-off and landing, and in accessing constrained environments where other aerial vehicles may be unable to operate. Nevertheless, UHs require complex, multivariable, nonlinear, and under-actuated systems, often characterized by strong coupling effects and uncertain or unidentifiable aerodynamic parameters [1,2]. To address these inherent limitations, control designs for UHs typically adopt a cascaded structure, consisting of an outer-loop subsystem and an inner-loop subsystem (i.e., translational and rotational). The authors of [3,4,5] discussed the trajectory tracking control methods for a single UH with unknown uncertainties and external disturbances. As mission complexity and environmental challenges escalate, employing helicopter formations rather than individual units has become crucial for meeting operational demands and enhancing overall system performance. The authors of [6] studied the intelligent formation control problem by utilizing recurrent neural networks, and the authors of [7,8,9] provided fast formation tracking control protocols for multiple UHs to improve convergence rate/time.
Compared to fixed-wing aircraft, helicopters are more prone to mechanical faults and operational challenges due to their unique structural characteristics, dynamic flight mechanics, and complex operational conditions. If a fault (e.g., performance degradation) occurring in an actuator, sensor, or other component is not detected and addressed, the helicopter may fail to perform tasks. To ensure that system performance remains within acceptable bounds when a fault occurs, the fault-tolerant control (FTC) scheme must be implemented to settle the presumed faults [10,11]. Depending on whether to address the faults proactively, the existing FTC methods can be generally categorized into passive FTC (PFTC) and active FTC (AFTC) [12]. PFTC does not modify the system structure and control unit. It only employs a resilient controller against the expected and considered fault categories without requiring accurate fault information [13,14,15]. However, this approach may have inherent limitations in fault accommodation, potentially compromising system performance. An AFTC is developed by reconstructing the related controller, or selecting a pre-devised controller through a fault diagnosis technique that can detect and identify faults [16,17]. In this manner, the authors of [18,19] developed an output feedback-based AFTC scheme for a single three-degree-of-freedom (3-DOF) helicopter suffering from angular position sensor faults. Next, the authors of [20] developed a decentralized fault estimation and a distributed fault-tolerant control hierarchy for a number of 3-DOF helicopters with nonlinearities, uncertainties, and both actuator and sensor faults. As the autonomy of the helicopter continues to improve, designing intelligent AFTC schemes to compensate for potential faults remains a crucial area for further investigation, serving as the primary motivation for this article.
It should be pointed out that the results above only require the helicopters to follow a desired trajectory under fault conditions, without imposing additional performance constraints such as minimizing tracking time, tracking distance, or energy consumption. Therefore, optimal control problems have received considerable attention in practice, which often need either the Hamilton–Jacobi–Bellman (HJB) or Hamilton–Jacobi–Issacs (HJI) equation to obtain an analytical solution [21,22,23]. Unlike the conventional optimal control method that initially defines the cost performance function and then minimizes it to design an optimized controller, the inverse optimal control method takes an alternative way. It starts by designing a stabilizing controller and then identifies a corresponding cost function for which the controller is optimal. Also, it no longer focuses on the solution of the HJB or HJI equation but searches inversely for a certain undetermined objective function. This capability makes the inverse optimal control an essential tool in improving system performance and achieving optimization without solving partial differential equations. The authors of [24,25,26,27] investigated the inverse optimal consensus, formation, and containment control problem for multi-agent systems. The authors of [28] devised a fuzzy-based inverse optimal output feedback strategy to resist actuator and sensor network attacks. The authors of [29,30,31] later incorporated disturbance estimation into an inverse optimal control. From the perspective of the remarkable ability, it is also necessary for UHs to equip the inverse optimal control with an FTC, which is the main motivation for investigating this problem in the subsequent sections.
Enlightened by the discussions above, the inverse optimal fault-tolerant formation-containment control strategy is proposed for multi-UHs with actuator faults. The primary innovations and contributions are outlined below:
(1)
Different from the previous control schemes in [6,7,8,9], multiple leaders and multiple followers exist in the formation system, and the leaders collaborate with each other to form a desired formation. Each UH modeling is equipped with nonlinearities and faults (simultaneous partial loss of effectiveness and bias), which are separated into outer-loop (position) and inner-loop (attitude) subsystems based on multiple-time-scale features.
(2)
The serial-parallel estimation model is constructed with intelligent approximation techniques. Accordingly, the composite learning (CL)-based updating law is designed for both the network weight and the fault estimation observer.
(3)
The estimated fault information is subsequently incorporated into the inverse optimal FTC to develop the controller that ensures both optimality and effective fault compensation. Owing to this, the formation of the leaders and reference tracking are obtained. Further, the followers are not only inside the convex hull established by leaders but maintain a specific formation based on the combination of the leaders.
The remaining sections of this article are structured as follows. Section 2 gives the problem formulation and related lemmas, definitions, and assumptions. Section 3 and Section 4 separately design the CL-based inverse optimal fault-tolerant formation and containment controller for leaders and followers. Finally, Section 5 presents the simulation flight results of the hierarchy-structured UHs, and Section 6 concludes this article.

2. The UH Flight Dynamics Model and Problem Formulation

Each UH can be recognized as a rigid body in executing formation flight tasks. Let O I = { x I , y I , z I } be the inertial reference frame positioned at the take-off location, and O ι = { x ι , y ι , z ι } be the body-fixed coordinate frame located at the center of gravity of the helicopter. Figure 1a presents the spatial configuration of the inertial and body-fixed frames. As depicted in Figure 1b, a flying system with N + M helicopters (including N leaders and M followers) is considered, where H = { 1 , 2 , N + M } is the set of serial numbers for the helicopters, and L = { 1 , 2 , N } and F = { N + 1 , N + 2 , , N + M } are the sets of serial numbers for the leaders and followers, respectively. Also, the virtual leader is labeled as 0. The 6-DOF rigid body model of the ι th UH is modeled as a position subsystem and an attitude subsystem as follows:
P ˙ ι = V ι , ι = 1 , 2 , N + M V ˙ ι = g e 3 + 1 m ι R ι f ι
Φ ˙ ι = Π ι Λ ι Λ ˙ ι = J ι 1 Λ ι × J ι Λ ι + J ι 1 τ ι
where P ι = [ x ι , y ι , z ι ] T and V ι = [ u ι , v ι , ω ι ] T are the position and the velocity vectors of the inertial reference frame O I , respectively. Φ ι = [ ϕ ι , θ ι , ψ ι ] T and Λ ι = [ p ι , q ι , r ι ] T denote the Euler angle and angular rate vectors of the body-fixed coordinate frame O ι , respectively. e 3 = [ 0 , 0 , 1 ] T is the unitary vector. g indicates the gravity acceleration, m ι denotes the mass of UH, and J ι = diag { J x ι , J y ι , J z ι } represents the diagonal inertia matrix. Further, the rotation matrix R ι , the anti-symmetric matrix Λ ι × , and the rotational kinematic matrix Π ι are defined below:
R ι = cos θ ι cos ψ ι sin ϕ ι sin θ ι cos ψ ι cos ϕ ι sin ψ ι cos ϕ ι sin θ ι cos ψ ι + sin ϕ ι sin ψ ι cos θ ι sin ψ ι sin ϕ ι sin θ ι sin ψ ι + cos ϕ ι cos ψ ι cos ϕ ι sin θ ι sin ψ ι sin ϕ ι cos ψ ι sin θ ι sin ϕ ι cos θ ι cos ϕ ι cos θ ι Λ ι × = 0 r ι q ι r ι 0 p ι q ι p ι 0 , Π ι = 1 sin ϕ ι tan θ ι cos ϕ ι tan θ ι 0 cos ϕ ι sin ϕ ι 0 sin ϕ ι / cos θ ι cos ϕ ι / cos θ ι
In translational dynamics, f ι = [ 0 , 0 , T m ι ] T implies the force vector, with T m ι m ι ( g + Z ω ι ω ι + Z c o ι δ c o ι ) as the main rotor thrust controlled by the collective pitch δ c o ι . Z ω ι and Z c o ι denote the constants associated with the main rotor rotation speed and the collective pitch δ c o ι , respectively.
For the rotational dynamics, τ ι is the moment vector applied to a helicopter, which is formed by
τ ι = J ι L a ι a ι + L b ι b ι M a ι a ι + M b ι b ι N r ι r ι + N t a ι δ t a ι + N c o ι δ c o ι
where L a ι , L b ι and M a ι , M b ι are the lateral and longitudinal wave motion coefficients of the main rotor. N r ι denotes the damping coefficient of the yaw angle. N t a ι and N c o ι are damping coefficients related to the tail rotor and main rotor (e.g., rotation speed and blade radius). δ t a ι is the collective pitch of the tail rotor. a ι and b ι denote the flapping angles of the main rotor along with the longitudinal and lateral axes, which are mainly controlled by longitudinal and lateral cyclic δ l o ι and δ l a ι . Their relationships can be approximated as
a ι = τ m ι q ι + A l a ι δ l a ι + A l o ι δ l o ι b ι = τ m ι p ι + B l a ι δ l a ι + B l o ι δ l o ι
Therefore, combining Equations (3) and (4), a modified moment vector can be rewritten as
τ ι = J ι τ m ι L a ι q ι τ m ι L b ι p ι τ m ι M a ι q ι τ m ι M b ι p ι N r ι r ι + J ι L l a ι δ l a ι + L l o ι δ l o ι M l a ι δ l a ι + M l o ι δ l o ι N t a ι δ t a ι + J ι 0 0 N c o ι δ c o ι = J ι A ι Ω ι + J ι B ι U ι + J ι N c o ι δ c o ι e 3
where U ι = [ δ l o ι , δ l a ι , δ t a ι ] T . τ m ι denotes the time constant of the main rotor. The parameters in the above matrices are L l a ι = L a ι A l a ι + L b ι B l a ι , M l a ι = M a ι A l a ι + M b ι B l a ι , L l o ι = L a ι A l o ι + L b ι B l o ι , and M l o ι = M a ι A l o ι + M b ι B l o ι , where the coefficients A l a ι , B l a ι and A l o ι , B l o ι can be obtained by the system model identification. The matrices A ι and B ι are given as
A ι = τ m ι L b ι τ m ι L a ι 0 τ m ι M b ι τ m ι M a ι 0 0 0 N r ι , B ι = L l o ι L l a ι 0 M l o ι M l a ι 0 0 0 N t a ι
Remark 1. 
Note that the representations of force vector f ι , moment vector τ ι , and blade flapping angles (both longitudinal a ι and lateral b ι ) are simplified in the above-mentioned model. The force and moment vectors generated by the main rotor, tail rotor, and fuselage can be highly complex due to unsteady aerodynamics and dynamic interactions. However, under steady or quasi-steady flight conditions (such as hover or low-speed cruise), these effects can be approximated using linearized or lumped-parameter models, significantly reducing computational complexity. Similarly, the blade flapping dynamics can be simplified under the small-angle assumption. In many control-oriented models, the flapping angles are approximated using first-order dynamics or even static gain relationships, assuming a linear response to cyclic pitch. These simplifications are reasonable in flight regimes, where the flapping angles remain small and high-frequency dynamics are negligible. Such simplifications keep a balance between model fidelity and computational efficiency, making them suitable for control design.
Remark 2. 
In this article, the virtual leader serves as a reference point that defines the desired trajectory and formation pattern for the actual leaders to follow and maintain. Although it does not physically exist, the virtual leader provides position or yaw angle information that coordinates the motions in the flying system. By tracking the virtual leader, each helicopter can maintain proper spacing and movement direction, thereby ensuring the overall geometric structure.
Let δ c o ι , δ l o ι , δ l a ι and δ t a ι be the actual input signal, which is combined with the collective pitch of the main and tail rotor, and the lateral and longitudinal cyclic. Each component plays a critical role in ensuring the helicopter’s stability and maneuverability. Due to the hydraulic or electrical malfunctions, and some environmental factors like extreme temperatures or corrosion, these control inputs may suffer from various types of faults. Among the most common are loss of effectiveness and bias faults, which allow the system to continue operating but with reduced performance. Such faults can result in sluggish or unresponsive control behavior, ultimately leading to degraded or even lost of control authority. In this article, the fault models are formulated as
δ υ f = ρ ι m δ υ + ζ ι m = δ υ + ( ρ ι m 1 ) δ υ + ζ ι m , m = 1 , 2 , 3 , 4
where υ = { c o ι , l o ι , l a ι , t a ι } , ρ ι m ( 0 , 1 ] , and ξ ι m denote the partial loss of actuator effectiveness and the additional bias faults, respectively.
Subsequently, the flight dynamics (1) and (2) of the ι th UH are separated into outer-loop (position) and inner-loop (attitude) subsystems for the control design. Since the design processes of three directions are similar, define [ P ι 1 , P ι 2 , P ι 3 ] T = [ x ι , y ι , z ι ] T and [ V ι 1 , V ι 2 , V ι 3 ] T = [ u ι , v ι , ω ι ] T , and the outer-loop (position) subsystem can be rewritten as
P ˙ ι n = V ι n V ˙ ι n = f ι 1 n + u ι 1 n + Θ ι 1 n , n = 1 , 2 , 3
where f ι 11 = ( cos ϕ ι sin θ ι cos ψ ι + sin ϕ ι sin ψ ι ) ( g + Z ω i ω ι ) , f ι 12 = ( cos ϕ ι sin θ ι sin ψ ι sin ϕ ι cos ψ ι ) ( g + Z ω ι ω ι ) and f ι 13 = cos ϕ ι cos θ ι ( g + Z ω ι ω ι ) are nonlinear functions. u ι 1 n and Θ ι 1 n imply the input signal to be designed and the lumped fault to be estimated, which are constructed as
u ι 11 = ( cos ϕ ι sin θ ι cos ψ ι + sin ϕ ι sin ψ ι ) Z c o ι δ c o ι u ι 12 = ( cos ϕ ι sin θ ι sin ψ ι sin ϕ ι sin ψ ι ) Z c o ι δ c o ι u ι 13 = g + cos ϕ ι cos θ ι Z c o ι δ c o ι
and
Θ ι 11 = ( cos ϕ ι sin θ ι cos ψ ι + sin ϕ ι sin ψ ι ) Z c o ι ( ( ρ ι 1 1 ) δ c o ι + ξ ι 1 ) Θ ι 12 = ( cos ϕ ι sin θ ι sin ψ ι sin ϕ ι sin ψ ι ) Z c o ι ( ( ρ ι 1 1 ) δ c o ι + ξ ι 1 ) Θ ι 13 = cos ϕ ι cos θ ι Z c o ι ( ( ρ ι 1 1 ) δ c o ι + ξ ι 1 )
Similarly, we denote [ Φ ι 1 , Φ ι 2 , Φ ι 3 ] T = [ ϕ ι , θ ι , ψ ι ] T and [ Λ ι 1 , Λ ι 2 , Λ ι 3 ] T = [ p ι , q ι , r ι ] T to reformulate the inner-loop (attitude) subsystem as
Φ ˙ ι n = f ι 2 n + g ι n Λ ι n Λ ˙ ι n = f ι 3 n + u ι 2 n + Θ ι 2 n , n = 1 , 2 , 3
where f ι 21 = sin ϕ ι tan θ ι q ι + cos ϕ ι tan θ ι r ι , f ι 22 = sin ϕ ι r ι , f ι 23 = q ι sin ϕ ι / cos θ ι , g i 1 = 1 , g ι 2 = cos ϕ ι , g ι 3 = cos ϕ ι / cos θ ι , f ι 31 = q ι r ι ( J z ι J y ι ) / J x ι τ m ι L b ι p ι τ m ι L a ι q ι , f ι 32 = p ι r ι ( J x ι J z ι ) / J y ι τ m ι M b ι p ι τ m ι M a ι q ι , and f ι 33 = p ι q ι ( J y ι J x ι ) / J z ι + N r ι r ι are nonlinear functions. u ι 2 n and Θ ι 2 n denote the input signal and the lumped fault as follows:
u ι 21 = L l o ι δ l o ι + L l a ι δ l a ι u ι 22 = M l o ι δ l o ι + M l a ι δ l a ι u ι 23 = N t a ι δ t a ι + N c o ι δ c o ι
and
Θ ι 21 = L l o ι ( ( ρ ι 2 1 ) δ l o ι + ξ ι 2 ) + L l a ι ( ( ρ ι 3 1 ) δ l a ι + ξ ι 3 ) Θ ι 22 = M l o ι ( ( ρ ι 2 1 ) δ l o ι + ξ ι 2 ) + M l a ι ( ( ρ ι 3 1 ) δ l a ι + ξ ι 3 ) Θ ι 23 = N t a ι ( ( ρ ι 4 1 ) δ t a ι + ξ ι 4 ) + N c o ι ( ( ρ ι 1 1 ) δ c o ι + ξ ι 1 )
Remark 3. 
It should be noted that the direct control inputs of the helicopter are δ c o ι , δ l o ι , δ l a ι , and δ t a ι . These inputs determine the control effectiveness in all flight directions and can be computed through the transformation relationships defined in (8) and (11). The above-mentioned (6) satisfies the following four cases: (1) When ρ ι m = 1 and ζ ι m = 0 , this indicates the fault-free (normal) case. (2) When ρ ι m ( 0 , 1 ) and ζ ι m = 0 , it implies the partial loss of actuator effectiveness only. (3) When ρ ι m = 1 and ζ ι m 0 , it indicates the bias fault only. (4) When ρ ι m ( 0 , 1 ) and ζ ι m 0 , it reveals both the partial loss of actuator effectiveness and bias faults. Note that when ζ ι m = 0 and ζ ι m 0 , this refers to the stuck fault, i.e., δ υ sticks at a bounded function ζ ι m . However, this case cannot be compensated for by designing the AFTC scheme; thus, we only consider the other four cases in the subsequent sections.
Before presenting the formation-containment control scheme, we make the following fundamental graph theory and some related lemmas, definitions, and assumptions.
The communication network among the N + M helicopters is represented by a directed graph G = ( V , E , A ) , where V is the node set and E V × V is the edge set. The vertex sets of the leaders and the followers are defined as V L and V F , respectively. Consequently, we have V = V L V F and V L V F = . The edge pair ( i , j ) E means that the ith UH can receive data from the jth UH. The adjacent matrix A = [ a i j ] R ( N + M ) × ( N + M ) is a position matrix that can be defined as a i j = 1 when ( i , j ) E , otherwise a i j = 0 . Denote the Laplacian matrix as L = D A = [ l i j ] R ( N + M ) × ( N + M ) , where l i j = a i j , i j , and l i i = j = 1 , j i N + M a i j . D = diag { d 1 , d 2 , , d N + M } represents the degree matrix, d i j = j = 1 N + M a i j . Here, the Laplacian matrix is described as
L = L l l 0 N × M L l f L f f
where L l l R N × N signifies the communication topology among the leader UHs, L l f R M × N denotes the communication interaction from the leaders to the followers, and L f f R M × M implies the communication topology among the follower UHs.
In addition, B = [ b 1 , b 2 , , b N ] , with b i being the connections between the ith leader UH and the virtual leader, where b i = 1 means that it can receive data from the virtual leader, or otherwise, b i = 0 .
Assumption 1 
([32]). As for the leader UHs, the virtual leader has at least one path to each leader UH. Also, there exists at least one path from the leader UHs to each follower UH.
Lemma 1 
([32]). All the eigenvalues of matrix L l l + B have positive real parts. Each entry of L f f 1 L l f is positive and each row sum of L f f 1 L l f is equal to 1.
Definition 1 
([32]). (Convex Hull) Suppose that C is a subset in the real vector space Z R ρ . The set C is said to be convex if the point ( 1 ρ ) α + ρ β C holds for any ρ [ 0 , 1 ] and arbitrary two elements α, β. The convex hull regarding a group of points X = { x 1 , x 2 , , x n } in Z is the minimal convex set consisting of all points in X, and then, it is represented as Co ( X ) = { k = 1 n ρ k | x k X , ρ k > 0 , k = 1 n ρ k = 1 } .
In this article, the control objective is to ensure the desired flight performance of N + M UHs. As shown in Figure 2, a two-layer structure framework is employed to settle the inverse optimal fault-tolerant formation-containment control problem: one is the formation layer of N leaders, and another is the containment layer of M followers. Specifically, the leaders in the first layer can simultaneously take action and form a desired reference formation pattern, while the followers in the second layer are contained in a convex hull constructed by the leaders. Thus, the inverse optimal fault-tolerant control of the hierarchy-structured UHs is completed if and only if both two layers have achieved their own control objectives. That is, for i , j L , there exists
lim t P i P 0 c i P d = 0 lim t P i P j ( c i P d c j P d ) = 0
and for k F , j L , there exists
lim t P k j L α k j P j = 0
where P 0 = [ x 0 , y 0 , z 0 ] T is the position of the virtual leader. c i P d = [ c i P 1 d , c i P 2 d , c i P 3 d ] T is the relative position of the desired configuration between the ith leader UH and the virtual leader.
Remark 4. 
Owing to the under-actuated mechanics of the ιth UH, the proposed control algorithm is separated into two cascaded control modules. The outer-loop (position) controller calculates the input signal u ι 1 n and appropriate attitude commands ϕ ι d and θ ι d to navigate the helicopter along the target trajectories. It can be calculated that ϕ ι d and θ ι d are
ϕ ι d = arcsin sin ψ ι d u ι 11 cos ψ ι d u ι 12 u ι 11 2 + u ι 12 2 + ( u ι 13 g ) 2 θ ι d = arctan cos ψ ι d u ι 11 + sin ψ ι d u ι 12 u ι 13 g , ι = 1 , 2 , , N + M
and ψ ι d = ψ 0 with ψ 0 denotes the yaw angle of the virtual leader. Then, the command Φ ι d = [ ϕ ι d , θ ι d , ψ ι d ] T is delivered to the inner-loop (attitude) controller that calculates u ι 2 n .
Assumption 2 
([33]). The position P 0 = [ x 0 , y 0 , z 0 ] T and the yaw angle reference ψ 0 from the virtual leader and their derivatives are bounded and continuous functions.
Assumption 3 
([7]). The Euler angle vectors (i.e., roll angle ϕ ι , pitch angle θ ι , and yaw angle ψ ι ) are bounded and satisfy ϕ ι ( π 2 , π 2 ) , θ ι ( π 2 , π 2 ) and ψ ι ( π , π ) for all t.
Lemma 2 
([30]). (Inverse Optimal Control) Consider a nonlinear system as follows:
x ˙ = f ( x ) + g ( x ) u
where x R n and u R m denote the state and control input vectors, while f ( x ) and g ( x ) are the smooth functions. If there exist a positive-definite control Lyapunov function V ( x ) , and a control strategy u = R ( x ) 1 ( L g V ) T stabilizing the whole system (17), then we have u * = β u , where β 2 is the inverse optimal control input for the system (17), and the minimized performance index function is
J = lim t { 0 ( l ( x ) + u T R ( x ) u ) d t }
where l ( x ) = 2 β ( L f V L g V R ( x ) 1 ( L g V ) T + β ( β 2 ) L g V R ( x ) 1 ( L g V ) T ) with L f V = V f ( x ) / x and L g V = V g ( x ) / x .
Lemma 3 
([7]). (Neural Network Approximation) For an unknown nonlinear function f ( x ) , there is a neural network w * T s ( x ) satisfying the following condition:
f ( x ) = w * T s ( x ) + ϵ x , | ϵ x | ϵ ¯ x
where w * represents the ideal neural network weight matrix, s ( x ) denotes the activation function, and ϵ ¯ x depicts the approximation error ϵ x bounded by | ϵ x | . In this article, s ( x ) is assumed as the Gaussian function, i.e., s ( x ) = e x p ( x c 2 / 2 ς 2 ) , where ς > 0 represents the width, and c = [ c 1 , c 2 , , c μ ] T , with μ as the node number.

3. CL-Based Inverse Optimal Fault-Tolerant Formation Controller

In this subsection, a CL-based inverse optimal fault-tolerant formation control scheme is put forward for N leader UHs to track the virtual leader and form a desired geometric pattern even though the actuator faults exist. Based on this, P i P j c i P d c j P d and P i P 0 c i P d can be achieved in the outer-loop (position) subsystem, and Φ i Φ i d can be obtained in the inner-loop (attitude) subsystem.

3.1. Outer-Loop (Position) Controller Design for the Leaders

For the formation layer, we define the tracking errors of each direction in the outer-loop (position) subsystem (7) as follows:
e i P n = j L a i j P i n P j n ( c i P n d c j P n d ) + b i ( P i n P 0 n c i P n d ) e i V n = V i n a ¯ i 1 n , n = 1 , 2 , 3
where i , j L . a i j is the connective relation between the ith helicopter and jth helicopter, and a ¯ i 1 n denotes the designed virtual controller in the following.
Subsequently, we elaborate an inverse optimal fault-tolerant control scheme for the ith leader UH via the CL method under the backstepping framework.
In light of Equation (20), the derivative of e i P n with respect of time can be attained as
e ˙ i P n = π i e i V n + π i a ¯ i 1 n j L a i j P ˙ j n b i P ˙ 0 n
where π i = j L a i j + b i .
Then, the virtual controller a ¯ i 1 n is defined as
a ¯ i 1 n = π i 1 k i 1 n e i P n + j L a i j P ˙ j n + b i P ˙ 0 n
where k i 1 n denotes a positive constant.
By substituting Equation (22) into Equation (21), we can obtain as e ˙ i P n = k i 1 n e i P n + π i e i V n .
For the unknown nonlinear function f i 1 n , we employ the neural network f i 1 n = w i 1 n * T s i 1 n + ϵ i 1 n to approximate it, where ϵ i 1 n denotes the neural network approximation error with | ϵ i 1 n | ϵ ¯ i 1 n . Then, the derivative of e i V n is achieved as
e ˙ i V n = w i 1 n * T s i 1 n + ϵ i 1 n + u i 1 n + Θ i 1 n a ¯ ˙ i 1 n
To compensate for the influence of lumped fault Θ i 1 n , the following CL-based fault estimation observer is utilized to estimate the exact information of Θ i 1 n :
Θ ^ i 1 n = L i 1 n ( V i n η i 1 n ) η ˙ i 1 n = w ^ i 1 n T s i 1 n + u i 1 n + Θ ^ i 1 n L i 1 n 1 ( r f i 1 n ϖ f i 1 n + e i V n )
where L i 1 n denotes a positive constant. ϖ f i 1 n = V i n V ^ i n indicates the prediction error, where V ^ i n is denoted with the serial-parallel estimation model as follows:
V ^ ˙ i n = w ^ i 1 n T s i 1 n + u i 1 n + Θ ^ i 1 n + γ i 1 n ϖ f i 1 n
where γ i 1 n is a user-defined positive constant.
For the neural network updating law, the prediction error ϖ f i 1 n is also applied to construct the learning law design:
w ^ ˙ i 1 n = e i V n s i 1 n + r f i 1 n ϖ f i 1 n s i 1 n κ i 1 n w ^ i 1 n
where r f i 1 n and κ i 1 n represent the positive constants to be designed.
With Equations (23)–(26), the time-derivative of Θ ^ i 1 n and e i V n are rewritten as follows:
Θ ^ ˙ i 1 n = L i 1 n ( w ˜ i 1 n T s i 1 n + ϵ i 1 n + Θ ˜ i 1 n ) + r f i 1 n ϖ f i 1 n + e i V n
e ˙ i V n = w ^ i 1 n T s i 1 n a ¯ ˙ i 1 n + w ˜ i 1 n T s i 1 n + ϵ i 1 n + u ¯ i 1 n + Θ ˜ i 1 n
where u ¯ i 1 n = u i 1 n + Θ ^ i 1 n . w ˜ i 1 n = w i 1 n * w ^ i 1 n and Θ ˜ i 1 n = Θ i 1 n Θ ^ i 1 n are the estimation errors of network weight and lumped faults.
Remark 5. 
In [6,7,34], only the tracking error is used to construct the neural network updating law and the fault estimation observer. Here, the prediction error ϖ f i 1 is employed to improve the approximation accuracy of network weight and fault estimation.
The design of the inverse optimal fault-tolerant controller u i 1 n * is summarized in the next theorem.
Theorem 1. 
Under the designed virtual controller (22), the CL-based fault estimation observer (24), and the updating law of network weight (26), the inverse optimal fault-tolerant controller u i 1 n * is designed with a positive constant k ¯ i 1 n :
u i 1 n * = 2 ( k ¯ i 1 n + 1 2 + π i 2 2 k i 1 n + Ψ i 1 n 2 2 k ¯ i 1 n ) e i V n Θ ^ i 1 n
which can make the tracking errors e i P n and e i V n as well as the errors Θ ˜ i 1 n , w ˜ i 1 n , and ϖ f i 1 n of the outer-loop (position) subsystem approach a small region of the origin, and achieve the minimum cost function as follows:
J 1 i = 1 N n = 1 3 J i 1 n = i = 1 N n = 1 3 lim t 0 ( l i 1 n + u ¯ i 1 n T R i 1 n u ¯ i 1 n ) d t
where u ¯ i 1 n = ( k ¯ i 1 n + 1 2 + π i 2 2 k i 1 n + Ψ i 1 n 2 2 k ¯ i 1 n ) e i V n , l i 1 n = 2 k i 1 n e i P n 2 + 2 k ¯ i 1 n e i V n 2 + ( 4 L i 1 n 2 2 L i 1 n ρ i 1 n s ¯ i 1 n 2 ) Θ ˜ i 1 n 2 + ( 2 κ i 1 n 2 ρ i 1 n L i 1 n ) w ˜ i 1 n T w ˜ i 1 n + ( 4 r f i 1 n γ i 1 n 2 ) ϖ f i 1 n 2 + 2 k i 1 n ( e i P n π i k i 1 n e i V n ) 2 + 2 k ¯ i 1 n ( e i V n Ψ i 1 n k ¯ i 1 n e i V n ) 2 4 ε i 1 n .
Proof of Theorem 1
The Lyapunov functional is chosen as follows:
V 1 i = 1 N n = 1 3 V i 1 n = i = 1 N n = 1 3 1 2 e i P n 2 + 1 2 e i V n 2 + 1 2 Θ ˜ i 1 n 2 + 1 2 w ˜ i 1 n T w ˜ i 1 n + 1 2 r f i 1 ϖ f i 1 n 2
Note that L g V i 1 n = V i 1 n / e i V n = e i V n . From Lemma 2, there exists a controller u ¯ i 1 n = R i 1 n 1 e i V n that can stabilize the whole subsystem. By calculating the time-derivative of V 1 and combining Equations (24)–(28), we have
V ˙ 1 = i = 1 N n = 1 3 { k i 1 n e i P n 2 + π i e i P n e i V n + ( Ψ i 1 n R i 1 n 1 ) e i V n 2 + e i V n ϵ i 1 n + Θ ˜ i 1 n Θ ˙ i 1 n Θ ˜ i 1 n L i 1 n w ˜ i 1 n T s i 1 n Θ ˜ i 1 n L i 1 n ϵ i 1 n L i 1 n Θ ˜ i 1 n 2 + κ i 1 w ˜ i 1 n T w ^ i 1 n + r f i 1 n ϖ f i 1 n ϵ i 1 n r f i 1 n γ i 1 n ϖ f i 1 n 2 }
where Ψ i 1 n is a smooth function satisfying w ^ i 1 n T s i 1 n a ¯ ˙ i 1 n = Ψ i 1 n e i V n .
The following facts exist by using Young’s inequality:
e i V n ϵ i 1 n 1 2 e i V n 2 + 1 2 ϵ ¯ i 1 n 2
Θ ˜ i 1 n Θ ˙ i 1 n 1 2 Θ ˜ i 1 n 2 + 1 2 χ i 1 n 2
r f i 1 n ϖ f i 1 n ϵ i 1 n 1 2 ϖ f i 1 n 2 + 1 2 r f i 1 n 2 ϵ ¯ i 1 n 2
Θ ˜ i 1 L i 1 n ϵ i 1 1 2 Θ ˜ i 1 2 + 1 2 L i 1 n 2 ϵ ¯ i n 2
κ i 1 n w ˜ i 1 n T w ^ i 1 n 1 2 κ i 1 n w ˜ i 1 n T w ˜ i 1 n + 1 2 κ i 1 n w i 1 n * 2
Θ ˜ i 1 n L i 1 n w ˜ i 1 n T s i 1 n 1 2 ρ i 1 n L i 1 n s ¯ i 1 n 2 Θ ˜ i 1 n 2 + 1 2 ρ i 1 n L i 1 n w ˜ i 1 n T w ˜ i 1 n
where | Θ ˙ i 1 n | χ i 1 n , s i 1 n s ¯ i 1 n , with χ i 1 n , s ¯ i 1 n , and ρ i 1 n as the positive constants.
Integrating the above-mentioned inequality (33)–(38) into Equation (32), we obtain
V ˙ 1 i = 1 N n = 1 3 { k i 1 n e i P n 2 ( R i 1 n 1 Ψ i 1 n 1 2 ) e i V n 2 ( L i 1 n 1 1 2 ρ i 1 n L i 1 n s ¯ i 1 n 2 ) Θ ˜ i 1 n 2 + π i e i P n e i V n + ε i 1 n ( 1 2 κ i 1 n 1 2 ρ i 1 n L i 1 n ) w ˜ i 1 n T w ˜ i 1 n ( r f i 1 n γ i 1 n 1 2 ) ϖ f i 1 n 2 }
where ε i 1 n = 1 2 χ i 1 n 2 + 1 2 κ i 1 n w i 1 n * T w i 1 n * + 1 2 r f i 1 n 2 ϵ ¯ i 1 n 2 + 1 2 ϵ ¯ i 1 n 2 + 1 2 L i 1 n 2 ϵ ¯ i 1 n 2 .
For a given R i 1 n 1 = k ¯ i 1 n + 1 2 + π i 2 2 k i 1 n + Ψ i 1 n 2 2 k ¯ i 1 n , V ˙ 1 can be further obtained as
V ˙ 1 i = 1 N n = 1 3 { 1 2 k i 1 n e i P n 2 1 2 k ¯ i 1 n e i V n 2 ( L i 1 n 1 1 2 ρ i 1 n L i 1 n s ¯ i 1 n 2 ) Θ ˜ i 1 n 2 ( 1 2 κ i 1 n 1 2 ρ i 1 n L i 1 n ) w ˜ i 1 n T w ˜ i 1 n ( r f i 1 n γ i 1 n 1 2 ) ϖ f i 1 n 2 1 2 k i 1 n ( e i P n π i k i 1 n e i V n ) 2 1 2 k ¯ i 1 n ( e i V n Ψ i 1 n k ¯ i 1 n e i V n ) 2 + ε i 1 n } i = 1 N n = 1 3 { 1 2 k i 1 n e i P n 2 1 2 k ¯ i 1 n e i V n 2 ( L i 1 n 1 1 2 ρ i 1 n L i 1 n s ¯ i 1 n 2 ) Θ ˜ i 1 n 2 ( 1 2 κ i 1 n 1 2 ρ i 1 n L i 1 n ) w ˜ i 1 n T w ˜ i 1 n ( r f i 1 n γ i 1 n 1 2 ) ϖ f i 1 n 2 + ε i 1 n }
this means that if the conditions L i 1 n 1 1 2 ρ i 1 n L i 1 n s ¯ i 1 n 2 > 0 , 1 2 κ i 1 n 1 2 ρ i 1 n L i 1 n > 0 and r f i 1 n γ i 1 n 1 2 > 0 hold, it follows that the controller u ¯ i 1 n = R i 1 n 1 e i V n can ensure that the following condition holds:
V ˙ 1 c 1 V 1 + ε 1
where c 1 = min 1 i N , 1 n 3 { k i 1 n , k ¯ i 1 n , 2 L i 1 n 2 L i 1 n ρ i 1 n s ¯ i 1 n 2 , κ i 1 n 1 ρ i 1 n L i 1 n , 2 r f i 1 n γ i 1 n 1 } , and ε 1 = i = 1 N n = 1 3 ε i 1 n .
Then, from Lemma 2, when β i = 2 , we can derive that the CL-based inverse optimal controller u ¯ i 1 n * = 2 u ¯ i 1 n = 2 ( k ¯ i 1 n + 1 2 + π i 2 2 k i 1 n + Ψ i 1 n 2 2 k ¯ i 1 n ) e i V n can minimize the cost function (30). Moreover, the tracking errors e i P n and e i V n , along with the errors Θ ˜ i 1 n , w ˜ i 1 n , and ϖ f i 1 n of the ith leader UH, converge to a small region around the origin. Accordingly, the optimal fault-tolerant controller is derived as (29).
Next, we will verify whether the control objective (14) is achieved for the N leaders’ formation system. Define P ˜ i n = P i n P 0 n + c i P n d , and from Equation (20), we have
e i P n = j L a i j ( P ˜ i n P ˜ j n ) + b i P ˜ i n
which can be rewritten as e i P = j L a i j ( P ˜ i P ˜ j ) + b i P ˜ i with P ˜ i = P i P 0 + c i P d . As for the N leader UHs, the following is derived:
e l P = ( ( L l l + B ) I 3 ) P ˜ l
where e l P = [ e 1 P , e 2 P , , e N P ] T and P ˜ l = [ P ˜ 1 , P ˜ 2 , , P ˜ N ] T . From Lemma 1, the matrix L l l + B is a non-singular matrix, and P ˜ l = ( ( L l l + B ) I 3 ) 1 e l P is derived, i.e., the formation control object is achieved. □

3.2. Inner-Loop (Attitude) Controller Design for the Leaders

Define the tracking errors of the inner-loop subsystem as follows:
e i Φ n = Φ i n Φ i d n e i Λ n = Λ i n a ¯ i 2 n
where [ Φ i d 1 , Φ i d 2 , Φ i d 3 ] T = [ ϕ i d , θ i d , ψ i d ] T , and a ¯ i 2 n is the virtual controller to be designed.
According to Lemma 3, the neural network is utilized to estimate the unknown function f i 2 n = w i 2 n * T s i 2 n + ϵ i 2 n , where ϵ i 2 n denotes the bounded approximation error satisfying | ϵ i 2 n | ϵ ¯ i 2 n . Similarly, the time-derivative of e i Φ n is derived as follows:
e ˙ i Φ n = w i 2 n * T s i 2 n + ϵ i 2 n Φ ˙ i d n + g i n ( e i Λ n + a ¯ i 2 n )
where the network weight updating law is designed as
w ^ ˙ i 2 n ¯ = e i Φ n s i 2 n + r f i 2 n ϖ f i 2 n s i 2 n κ i 2 n w ^ i 2 n
where r f i 2 n and κ i 2 n are positive constants. The prediction error is denoted as ϖ f i 2 n = Φ i n Φ ^ i n , where Φ ^ i n is estimated by the serial-parallel estimation model as follows:
Φ ^ ˙ i n = w ^ i 2 n s i 2 n + g i n ( e i Λ n + a ¯ i 2 n ) + γ i 2 n ϖ f i 2 n
where γ i 2 n represents a positive constant to be designed.
Furthermore, the virtual controller a ¯ i 2 n is designed as
a ¯ i 2 n = g i n 1 ( k i 2 n e i Φ n + Φ ˙ i d n w ^ i 2 n T s i 2 n )
where k i 2 n is a given positive constant.
Integrating Equation (48) into Equation (45), the time-derivative of e i Φ has the form
e ˙ i Φ n = k i 2 n e i Φ n + g i n e i Λ n + w ˜ i 2 n s i 2 n + ϵ i 2 n
where w ˜ i 2 n = w i 2 n * w ^ i 2 n is the approximation error of the network weight.
Similarly, with the help of the neural network approximation, e ˙ i Λ n is achieved as
e ˙ i Λ n = w i 3 n * T s i 3 n + ϵ i 3 n + u i 2 n + Θ i 2 n a ¯ ˙ i 2 n
where ϵ i 3 n is the bounded network approximation error with | ϵ i 3 n | ϵ ¯ i 3 n .
Correspondingly, the CL-based fault estimation observer is given below:
Θ ^ i 2 n = L i 2 n ( Λ i n η i 2 n ) η ˙ i 2 n = w ^ i 3 n T s i 3 n + u i 2 n + Θ ^ i 2 n L i 2 n 1 ( r f i 3 n ϖ f i 3 n + e i Λ n )
where L i 2 n denotes a positive constant, and the network weight training law of w ^ i 3 n with the prediction error ϖ f i 3 n = Λ i n Λ ^ i n is designed as follows:
w ^ ˙ i 3 n = e i Λ n s i 3 n + r f i 3 n ϖ f i 3 n s i 3 n κ i 3 n w ^ i 3 n
Λ ^ ˙ i n = w ^ i 3 n T s i 3 n + u i 2 n + Θ ^ i 2 n + γ i 3 n ϖ f i 3 n
where γ i 3 n , r f i 3 n , and κ i 3 n denote designed positive constants.
On this basis, we have
Θ ^ ˙ i 2 n = L i 3 n ( w ˜ i 3 n T s i 3 n + ϵ i 3 n + Θ ˜ i 2 n ) + r f i 3 n ϖ f i 3 n + e i Λ n
e ˙ i Λ n = w ^ i 3 n T s i 3 n a ¯ ˙ i 2 n + w ˜ i 3 n T s i 3 n + ϵ i 3 n + u ¯ i 2 n + Θ ˜ i 2 n
where u ¯ i 2 n = u i 2 n + Θ ^ i 2 n . Also, w ˜ i 3 n = w i 3 n * w ^ i 3 n and Θ ˜ i 2 n = Θ i 2 n Θ ^ i 2 n are approximation errors.
Theorem 2. 
Under the designed virtual controller (48), the fault estimation observer (51), and the composite updating law of network weight (52), the inverse optimal fault-tolerant controller u i 2 n * is designed with a positive constant k ¯ i 2 n :
u i 2 n * = 2 ( k ¯ i 2 n + g i n 2 2 k i 2 n 1 + Ψ i 2 n 2 2 k ¯ i 2 n ) e i Λ n Θ ^ i 2 n
which can make the tracking errors e i Φ n and e i Λ n as well as the estimation errors Θ ˜ i 2 n , w ˜ i 2 n , ϖ f i 2 n , w ˜ i 3 n , and ϖ f i 3 n of the inner-loop (attitude) subsystem approach a small region of the origin, and achieve the minimum cost function below:
J 2 = i = 1 N n = 1 3 J i 2 n = i = 1 N n = 1 3 lim t 0 ( l i 2 n + u ¯ i 2 n T R i 2 n u ¯ i 2 n ) d t
where  u ¯ i 2 n = ( k ¯ i 2 n + g i n 2 2 k i 2 n 1 + Ψ i 2 n 2 2 k ¯ i 2 n ) e i Λ n , l i 2 n = ( 2 k i 2 n 1 ) e i Φ n 2   + ( 4 r f i 2 n γ i 2 n 2 ) ϖ f i 2 n 2 + 2 κ i 2 n w ˜ i 2 n T w ˜ i 2 n   + 2 k ¯ i 2 n e i Λ n 2 + ( 4 L i 3 n 4 2 L i 3 n ρ i 3 n s ¯ i 3 n 2 ) Θ ˜ i 2   + ( 2 κ i 3 n 2 ρ i 3 n L i 3 n ) w ˜ i 3 n T w ˜ i 3 n +   ( 4 r f i 3 n γ i 3 n 2 ) ϖ f i 3 n 2 + ( 2 k i 2 n 1 ) ( e i Φ n g i n k i 2 n 1 2 e i Λ n ) 2   + 2 k ¯ i 2 n ( e i Λ n Ψ i 2 n k ¯ i 2 n e i Λ n ) 2 4 ε i 2 n .
Proof of Theorem 2. 
For Equations (45) and (50), we select the candidate Lyapunov functional as follows:
V 2 i = 1 N n = 1 3 V i 2 n = i = 1 N n = 1 3 { 1 2 e i Φ n 2 + 1 2 w ˜ i 2 n T w ˜ i 2 n + 1 2 r f i 2 n ϖ f i 2 n 2 + 1 2 e i Λ n 2 + 1 2 Θ ˜ i 2 n 2 + 1 2 w ˜ i 3 n T w ˜ i 3 n + 1 2 r f i 3 n ϖ f i 3 n 2 }
Note that L g V i 2 n = V i 2 n / e i Λ n = e i Λ n . From Lemma 2, there exists a controller u ¯ i 2 n = R i 2 n 1 e i Λ n that can stabilize the whole subsystem:
V 2 = i = 1 N n = 1 3 { k i 2 n e i Φ n 2 + g i n e i Φ n e i Λ n + e i Φ n ϵ i 2 n + κ i 2 n w ˜ i 2 n T w ^ i 2 n + r f i 2 n ϖ f i 2 n ϵ i 2 n r f i 2 n γ i 2 n ϖ f i 2 n 2 + ( Ψ i 2 n R i 2 n 1 ) e i Λ n 2 + e i Λ n ϵ i 3 n + Θ ˜ i 2 n Θ ˙ i 2 n L i 3 n Θ ˜ i 2 n w ˜ i 3 n T s i 3 n L i 3 n Θ ˜ i 2 n ϵ i 3 n L i 3 n Θ ˜ i 2 n 2 + κ i 3 n w ˜ i 3 n T w ^ i 3 n + r f i 3 n ϖ f i 3 n ϵ i 3 n r f i 3 n γ i 3 n ϖ f i 3 n 2 }
where w ^ i 3 n T s i 3 n a ¯ ˙ i 2 n = Ψ i 2 n e i Λ n , with Ψ i 2 n as a smooth function.
Using Young’s inequality, one determines that
V 2 i = 1 N n = 1 3 { ( k i 2 n 1 2 ) e i Φ n 2 + g i n e i Φ n e i Λ n ( r f i 2 n γ i 2 n 1 2 ) ϖ f i 2 n 2 1 2 κ i 2 n w ˜ i 2 n T w ˜ i 2 n ( R i 2 n 1 1 2 Ψ i 2 n ) e i Λ n 2 ( L i 3 n 1 1 2 L i 3 n ρ i 3 n s ¯ i 3 n 2 ) Θ ˜ i 2 n 2 ( 1 2 κ i 3 n 1 2 ρ i 3 n L i 3 n ) w ˜ i 3 n T w ˜ i 3 n ( r f i 3 n γ i 3 n 1 2 ) ϖ f i 3 n 2 + ε i 2 n }
where ε i 2 n = 1 2 ϵ ¯ i 2 n 2 + 1 2 κ i 2 n w i 2 n * T w i 2 n * + 1 2 r f i 2 n 2 ϵ ¯ i 2 n 2 + 1 2 ϵ ¯ i 3 n 2 + 1 2 χ i 2 n 2 + 1 2 L i 3 n 2 ϵ ¯ i 3 n 2 + 1 2 κ i 3 n w i 3 n * T w i 3 n * + 1 2 r f i 3 n 2 ϵ ¯ i 3 n 2 . ρ i 3 n , χ i 2 n , and s ¯ i 3 n are positive constants satisfying | Θ ˙ i 2 n | χ i 2 n and s i 3 n s ¯ i 3 n .
Defining R i 2 n 1 = k ¯ i 2 n + g i n 2 2 k i 2 n 1 + Ψ i 2 n 2 2 k ¯ i 2 n derives
V ˙ 2 i = 1 N n = 1 3 { 1 2 ( k i 2 n 1 2 ) e i Φ n 2 ( r f i 2 n γ i 2 n 1 2 ) ϖ f i 2 n 2 1 2 κ i 2 n w ˜ i 2 n T w ˜ i 2 n 1 2 k ¯ i 2 n e i Λ n 2 ( L i 3 n 1 1 2 L i 3 n ρ i 3 n s ¯ i 3 n 2 ) Θ ˜ i 2 n 2 ( 1 2 κ i 3 n 1 2 ρ i 3 n L i 3 n ) w ˜ i 3 n T w ˜ i 3 n ( r f i 3 n γ i 3 n 1 2 ) ϖ f i 3 n 2 + ε i 2 n 1 2 ( k i 2 n 1 2 ) ( e i Φ n g i n k i 2 n 1 2 e i Λ n ) 2 1 2 k ¯ i 2 n ( e i Λ n Ψ i 2 n k ¯ i 2 n e i Λ n ) 2 } i = 1 N n = 1 3 { 1 2 ( k i 2 n 1 2 ) e i Φ n 2 ( r f i 2 n γ i 2 n 1 2 ) ϖ f i 2 n 2 1 2 κ i 2 n w ˜ i 2 n T w ˜ i 2 n 1 2 k ¯ i 2 n e i Λ n 2 ( L i 3 n 1 1 2 L i 3 n ρ i 3 n s ¯ i 3 n 2 ) Θ ˜ i 2 n 2 ( 1 2 κ i 3 n 1 2 ρ i 3 n L i 3 n ) w ˜ i 3 n T w ˜ i 3 n ( r f i 3 n γ i 3 n 1 2 ) ϖ f i 3 n 2 + ε i 2 n }
which means that if the inequalities k i 2 n 1 2 > 0 , r f i 2 n γ i 2 n 1 2 > 0 , L i 3 n 1 1 2 L i 3 n ρ i 3 n s ¯ i 3 n 2 > 0 , 1 2 κ i 3 n 1 2 ρ i 3 n L i 3 n > 0 , and r f i 3 n γ i 3 n 1 2 > 0 hold, the controller u i 2 n = R i 2 n 1 e i Λ n can stabilize the whole subsystem, that is,
V ˙ 2 c 2 V 2 + ε 2
where c 2 = min 1 i N , 1 n 3 { k i 2 n 1 2 , 2 r f i 2 n γ i 2 n 1 , κ i 2 n , k ¯ i 2 n , 2 L i 3 n 2 L i 3 n ρ i 3 n s ¯ i 3 n 2 , κ i 3 n 1 ρ i 3 n L i 3 n , 2 r f i 3 n γ i 3 n 1 } and ε 2 = i = 1 N n = 1 3 ε i 2 n .
From Lemma 2, the optimized controller u ¯ i 2 n * = 2 u ¯ i 2 n = 2 ( k ¯ i 2 n + g i n 2 2 k i 2 n 1 + Ψ i 2 n 2 2 k ¯ i 2 n ) e i Λ n can minimize the performance index as (57). Further, the optimized fault-tolerant controller is calculated as (56), and the tracking errors e i Φ n and e i Λ n as well as the errors Θ ˜ i 2 n , w ˜ i 2 n , ϖ f i 2 n , w ˜ i 3 n , and ϖ f i 3 n of the ith leader UH approach a small region of the origin. □

4. CL-Based Inverse Optimal Fault-Tolerant Containment Controller

As for the second layer, the optimal position containment controller and optimal attitude tracking controller are separately designed for each follower UH in this subsection. Based on this, the followers will approach the convex region constructed by the leaders.
Firstly, the containment error e k n and the tracking error e k n ¯ of the kth helicopter are defined as
e k P n = l F a k l ( P k n P l n ) + j L a k j ( P k n P j n ) e k V n = V k n a ¯ k 1 n
and
e k Φ n = Φ k n Φ k d n e k Λ n = Λ k n a ¯ k 2 n
where k, l F , j L , a k l , and a k j are the connective relation between the kth follower and the lth follower or the jth leader. [ Φ k d 1 , Φ k d 2 , Φ k d 3 ] T = [ ϕ k d , θ k d , ψ k d ] T . a ¯ k 1 n and a ¯ k 2 n denote the virtual controller to be designed.
Next, the entirety of the CL-based inverse optimal control schemes with fault compensation design for the follower UHs are presented in the subsequent steps.
For the outer-loop (position) subsystem, differentiating e k P n yields
e ˙ k P n = π k e k V n + π k a ¯ k 1 n l F a k l P ˙ l n j L a k j P ˙ j n
where π k = l F a k l + j L a k j . Furthermore, we design the following feedforward virtual controller a ¯ k 1 n for target enclosing:
a ¯ k 1 n = π k 1 k k 1 n e k P n + l F a k l P ˙ l n + j L a k j P ˙ j n
where k k 1 n is a positive constant and has e ˙ k P n = k k 1 n e k P n + π k e k V n .
Similarly, the CL-based fault estimation observer Θ ^ k 1 n with the adaptive learning law of network weight w ^ k 1 n is constructed as
Θ ^ ˙ k 1 n = L k 1 n ( w ˜ k 1 n T s k 1 n + ϵ k 1 n + Θ ˜ k 1 n ) + r f k 1 n ϖ f k 1 n + e k V n
w ^ ˙ k 1 n = e k V n s k 1 n + r f k 1 n ϖ f k 1 n s k 1 n κ k 1 n w ^ k 1 n
where L k 1 n , r f k 1 n , and κ k 1 n are positive constants. ϖ f k 1 n = V k n V ^ k n is the prediction error with V ^ ˙ k n = w ^ k 1 n T s k 1 n + u k 1 n + Θ ^ k 1 n + γ k 1 n ϖ f k 1 n , with γ k 1 n as a positive constant. w ˜ k 1 n = w k 1 n * w ^ k 1 n and Θ ˜ k 1 n = Θ k 1 n Θ ^ k 1 n are the estimation errors of network weight and lumped faults.
Then, the derivative of e k V n is rewritten as
e ˙ k V n = w ^ k 1 n T s k 1 n a ¯ ˙ k 1 n + w ˜ k 1 n T s k 1 n + ϵ k 1 n + u ¯ k 1 n + Θ ˜ k 1 n
where u ¯ k 1 n = u k 1 n + Θ ^ k 1 n .
Similarly, for the inner-loop (attitude) subsystem, the time-derivation of e ˙ k Φ n is calculated as e ˙ k Φ n = w k 2 n * T s k 2 n + ϵ k 2 n Φ ˙ k d n + g k n ( e k Λ n + a ¯ k 2 n ) , and then, the virtual controller is
a ¯ k 2 n = g i n 1 ( k k 2 n e k Φ n + Φ ˙ k d n w ^ k 2 n s k 2 n )
where k k 2 n is a given positive constant, and the neural network weight w ^ k n is formed by
w ^ ˙ k 2 n ¯ = e k Φ n s k 2 n + r f k 2 n ϖ f k 2 n s k 2 n κ k 2 n w ^ k 2 n
where r f k 2 n and κ k 2 n are positive constants. ϖ f k 2 n = Φ k n Φ ^ k n denotes the prediction error with Φ ^ ˙ i n = w ^ k 2 n s k 2 n + g i n ( e k Λ n + a ¯ k 2 n ) + γ k 2 n ϖ f k 2 n , with γ k 2 n as a positive constant.
Combining (71) and (70), the time-derivative of e k Φ n has the form
e ˙ k Φ n = k k 2 n e k Φ n + g k n e k Λ n + w ˜ k 2 n s k 2 n + ϵ k 2 n
where w ˜ k 2 n = w k 2 n * w ^ k 2 n is the approximation error of the network weight.
By differentiating e k Λ n with respect to the time, we obtain e ˙ k Λ n = w k 3 n * T s k 3 n + ϵ k 3 n + u k 2 n + Θ k 2 n a ¯ ˙ k 2 n ; then, the CL-based fault estimation observer with the network weight training law is formulated as
Θ ^ ˙ k 2 n = L k 3 n ( w ˜ k 3 n T s k 3 n + ϵ k 3 n + Θ ˜ k 2 n ) + r f k 3 n ϖ f k 3 n + e k Λ n
w ^ ˙ k 3 n = e k Λ n s k 3 n + r f k 3 n ϖ f k 3 n s k 3 n κ k 3 n w ^ k 3 n
where L k 3 n , r f k 3 n and κ k 3 n denote the positive constants. ϖ f k 3 n = Λ k n Λ ^ k n is the prediction error with Λ ^ ˙ k n = w ^ k 3 n T s k 3 n + u k 2 n + Θ ^ k 2 n + γ k 3 n ϖ f k 3 n . Also, w ˜ k 3 n = w k 3 n * w ^ k 3 n and Θ ˜ k 2 n = Θ k 2 n Θ ^ k 2 n are the approximation errors.
Then, we have
e ˙ k Λ n = w ^ k 3 n T s k 3 n a ¯ ˙ k 2 n + w ˜ k 3 n T s k 3 n + ϵ k 3 n + u ¯ k 2 n + Θ ˜ k 2 n
where u ¯ k 2 n = u k 2 n + Θ ^ k 2 n .
Theorem 3. 
Under the designed virtual controllers (66) and (70), the fault estimation observers (67) and (73) with the updating laws of network weights (68), (71), and (74), the inverse optimal fault-tolerant controllers u k 1 n * and u k 2 n * are designed with the positive constants k ¯ k 1 n and k ¯ k 2 n :
u k 1 n * = 2 ( k ¯ k 1 n + 1 2 + π k 2 2 k k 1 n + Ψ k 1 n 2 2 k ¯ k 1 n ) e k V n Θ ^ k 1 n
u k 2 n * = 2 ( k ¯ k 2 n + g k n 2 2 k k 2 n 1 + Ψ k 2 n 2 2 k ¯ k 2 n ) e k Λ n Θ ^ k 2 n
which can make the tracking errors e k P n , e k V n , e k Φ n , and e k Λ n , as well as the errors Θ ˜ k 1 n , w ˜ k 1 n , ϖ f k 1 n , Θ ˜ k 2 n , w ˜ k 2 n , ϖ f k 2 n , w ˜ k 3 n , and ϖ f k 3 n approach a small region of the origin and achieve the minimum cost functions as follows:
J 3 k = N + 1 M n = 1 3 J k 1 n = k = N + 1 N n = 1 3 lim t { 0 ( l k 1 n + u ¯ k 1 n T R k 1 n u ¯ k 1 n ) d t }
J 4 k = N + 1 N n = 1 3 J k 2 n = k = N + 1 M n = 1 3 lim t { 0 ( l k 2 n + u ¯ k 2 n T R k 2 n u ¯ i 2 n ) d t }
where u ¯ k 1 n = ( k ¯ k 1 n + 1 2 + π i 2 2 k k 1 n + Ψ k 1 n 2 2 k ¯ k 1 n ) e k V n , u ¯ k 2 n = ( k ¯ k 2 n + g k n 2 2 k k 2 n 1 + Ψ k 2 n 2 2 k ¯ k 2 n ) e k Λ n . l k 1 n = 2 k k 1 n e k P n 2 + 2 k ¯ k 1 n e k V n 2 + ( 4 L k 1 n 2 2 L k 1 n ρ k 1 n s ¯ k 1 n 2 ) Θ ˜ k 1 n 2 + ( 2 κ k 1 n 2 ρ k 1 n L k 1 n ) w ˜ k 1 n T w ˜ k 1 n + ( 4 r f k 1 n γ k 1 n 2 ) ϖ f k 1 n 2 + 2 k k 1 n ( e k P n π k k k 1 n e i V n ) 2 + 2 k ¯ k 1 n ( e k V n Ψ k 1 n k ¯ k 1 n e k V n ) 2 4 ε k 1 n , and l k 2 n = ( 2 k k 2 n 1 ) e k Φ n 2 + ( 4 r f k 2 n γ k 2 n 2 ) ϖ f k 2 n 2 + 2 κ k 2 n w ˜ k 2 n T w ˜ k 2 n + ( 4 L k 3 n 4 2 L k 3 n ρ k 3 n s ¯ k 3 n 2 ) Θ ˜ i 2 + ( 2 κ k 3 n 2 ρ k 3 n L k 3 n ) w ˜ k 3 n T w ˜ k 3 n + 2 k ¯ k 2 n e k Λ n 2 + ( 2 k k 2 n 1 ) ( e k Φ n g i n k k 2 n 1 2 e k Λ n ) 2 + 2 k ¯ k 2 n ( e k Λ n Ψ k 2 n k ¯ k 2 n e k Λ n ) 2 + ( 4 r f k 3 n γ k 3 n 2 ) ϖ f k 3 n 2 4 ε k 2 n .
Proof of Theorem 2. 
For the outer-loop (position) subsystem, select the Lyapunov function as follows:
V 3 k = N + 1 M n = 1 3 V k 3 n = k = N + 1 N n = 1 3 { 1 2 e k P n 2 + 1 2 e k V n 2 + 1 2 Θ ˜ k 1 n 2 + 1 2 w ˜ k 1 n T w ˜ k 1 n + 1 2 r f k 1 ϖ f k 1 n 2 }
By employing the aforementioned procedures in (32)–(39), we can know that for a given R k 1 n 1 = k ¯ k 1 n + 1 2 + π k 2 2 k k 1 n + Ψ k 1 n 2 2 k ¯ k 1 n , V ˙ 3 becomes
V ˙ 3 k = N + 1 M n = 1 3 { 1 2 k k 1 n e k P n 2 1 2 k ¯ k 1 n e k V n 2 ( L k 1 n 1 1 2 ρ k 1 n L k 1 n s ¯ k 1 n 2 ) Θ ˜ k 1 n 2 ( 1 2 κ k 1 n 1 2 ρ k 1 n L k 1 n ) w ˜ k 1 n T w ˜ k 1 n ( r f k 1 n γ k 1 n 1 2 ) ϖ f k 1 n 2 + ε k 1 n }
where ε k 1 n = 1 2 χ k 1 n 2 + 1 2 κ k 1 n w k 1 n * T w k 1 n * + 1 2 r f k 1 n 2 ϵ ¯ k 1 n 2 + 1 2 ϵ ¯ k 1 n 2 + 1 2 L k 1 n 2 ϵ ¯ k 1 n 2 . ρ k 1 n , χ k 1 n , and s ¯ k 1 n are positive constants satisfying | Θ ˙ k 1 n | χ k 1 n and s k 1 n s ¯ k 1 n . Note that if the conditions L k 1 n 1 1 2 ρ k 1 n L k 1 n s ¯ k 1 n 2 > 0 , 1 2 κ k 1 n 1 1 2 ρ k 1 n L k 1 n > 0 and r f k 1 n γ k 1 n 1 2 > 0 hold, it follows that the controller u ¯ k 1 n = R k 1 n 1 e k V n can ensure that the following condition holds:
V ˙ 3 c 3 V 3 + ε 3
where c 3 = min N + 1 k M , 1 n 3 { k k 1 n , 2 L k 1 n 2 L k 1 n ρ k 1 n s ¯ k 1 n 2 , κ k 1 n 1 ρ k 1 n L k 1 n , 2 r f k 1 n γ k 1 n 1 , k ¯ k 1 n } and ε 3 = k = N + 1 M n = 1 3 ε k 1 n . From Lemma 2, when β k = 2 , the CL-based inverse optimal controller u ¯ k 1 n * = 2 u ¯ k 1 n = 2 ( k ¯ k 1 n + 1 2 + π i 2 2 k k 1 n + Ψ k 1 n 2 2 k ¯ k 1 n ) e k V n can minimize the cost function (78). Moreover, the tracking errors e k P n and e k V n along with the estimation errors Θ ˜ k 1 n , w ˜ k 1 n , and ϖ f k 1 n will converge to a small region around the origin. Then, the inverse optimal fault-tolerant controller is designed as (76).
For the inner-loop (attitude) subsystem, the Lyapunov functional is chosen as
V 4 k = N + 1 M n = 1 3 V k 2 n = i = 1 N n = 1 3 { 1 2 e k Φ n 2 + 1 2 w ˜ k 2 n T w ˜ k 2 n + 1 2 r f k 2 n ϖ f k 2 n 2 + 1 2 e k Λ n 2 + 1 2 Θ ˜ k 2 n 2 + 1 2 w ˜ k 3 n T w ˜ k 3 n + 1 2 r f k 3 n ϖ f k 3 n 2 }
and we can know that, if R k 2 n 1 = k ¯ k 2 n + g k n 2 2 k k 2 n 1 + Ψ k 2 n 2 2 k ¯ k 2 n , we have
V ˙ 4 k = N + 1 M n = 1 3 { 1 2 ( k k 2 n 1 2 ) e k Φ n 2 ( r f k 2 n γ k 2 n 1 2 ) ϖ f k 2 n 2 1 2 κ k 2 n w ˜ k 2 n T w ˜ k 2 n 1 2 k ¯ k 2 n e k Λ n 2 ( L k 3 n 1 1 2 L k 3 n ρ k 3 n s ¯ k 3 n 2 ) Θ ˜ k 2 n 2 ( 1 2 κ k 3 n 1 2 ρ k 3 n L k 3 n ) w ˜ k 3 n T w ˜ k 3 n ( r f k 3 n γ k 3 n 1 2 ) ϖ f k 3 n 2 + ε k 2 n }
where ε k 2 n = 1 2 ϵ ¯ k 2 n 2 + 1 2 κ k 2 n w k 2 n * T w k 2 n * + 1 2 r f k 2 n 2 ϵ ¯ k 2 n 2 + 1 2 ϵ ¯ k 3 n 2 + 1 2 L k 3 n 2 ϵ ¯ k 3 n 2 + 1 2 κ k 3 n w k 3 n * T w k 3 n * + 1 2 r f k 3 n 2 ϵ ¯ k 3 n 2 + 1 2 χ k 2 n 2 . ρ k 3 n , χ k 2 n , and s ¯ k 3 n are positive constants satisfying | Θ ˙ k 2 n | χ k 2 n and s k 3 n s ¯ k 3 n . This means that if the inequalities k k 2 n 1 2 > 0 , r f k 2 n γ k 2 n 1 2 > 0 , L k 3 n 1 1 2 L k 3 n ρ k 3 n s ¯ k 3 n 2 > 0 , 1 2 κ k 3 n 1 2 ρ k 3 n L k 3 n > 0 , and r f k 3 n γ k 3 n 1 2 > 0 hold, the controller u k 2 n = R k 2 n 1 e k Λ n can stabilize the whole subsystem such that
V ˙ 4 c 4 V 4 + ε 4
where c 4 = min N + 1 k M , 1 n 3 { k k 2 n 1 2 , 2 r f k 2 n γ k 2 n 1 , κ k 2 n , k ¯ k 2 n , 2 L k 3 n 2 L k 3 n ρ k 3 n s ¯ k 3 n 2 , κ k 3 n 1 ρ k 3 n L k 3 n , 2 r f k 3 n γ k 3 n 1 } and ε 4 = k = N + 1 M n = 1 3 ε k 2 n . Moreover, the tracking errors e k Φ n and e k Λ n alongside the errors Θ ˜ k 2 n , w ˜ k 2 n , ϖ f k 2 n , w ˜ k 3 n , and ϖ f k 3 n approach a small neighborhood of the origin.
The containment tracking objective is illustrated as follows. From (15), the containment error is described as e c P = P f ( L f f 1 L l f I 3 ) P l , where P l = [ P 1 , P 2 , , P N ] T , P f = [ P N + 1 , P N + 2 , , P N + M ] T . In terms of the condition (63), it has e k P = l F a k l ( P k P l ) + j L a k j ( P k P j ) , and the following equation can be derived:
e f P = ( L f f I 3 ) P f + ( L l f I 3 ) P l = ( L f f I 3 ) ( P f ( L f f 1 L l f I 3 ) P l ) = ( L f f I 3 ) e c P
where e f P = [ e ( N + 1 ) P , e ( N + 2 ) P , , e ( N + M ) P ] T , which can be written as e c P = ( L f f I 3 ) 1 e f P , i.e., the containment objective is achieved accordingly. □
Remark 6. 
For the flying system containing N leaders and M followers, if we choose the Lyapunov function as V = V 1 + V 2 + V 3 + V 4 , it derives that V ˙ < c V + ε with c = min { c 1 , c 2 , c 3 , c 4 } and ε = ε 1 + ε 2 + ε 3 + ε 4 , and the inverse optimal fault-tolerant controller u i 1 n * , u i 2 n * , and u k 1 n * , u k 2 n * can minimize the cost function J = J 1 + J 2 + J 3 + J 4 .
Remark 7. 
Among the above-mentioned parameters, larger values of k ι 1 n , k ¯ ι 1 n , k ι 2 n , k ¯ ι 2 n , γ i 1 n , γ i 2 n , and γ i 3 n along with smaller values of χ i 1 n , ϵ ¯ i 1 n , s ¯ i 1 n , χ i 2 n , ϵ ¯ i 2 n , ϵ ¯ i 3 n , and s ¯ i 3 n can be chosen such that L ι 1 n 1 1 2 ρ ι 1 n L ι 1 n s ¯ ι 1 n 2 > 0 , 1 2 κ ι 1 n 1 1 2 ρ ι 1 n L ι 1 n > 0 , r f ι 1 n γ ι 1 n 1 2 > 0 , k ι 2 n 1 2 > 0 , r f ι 2 n γ ι 2 n 1 2 > 0 , L ι 3 n 1 1 2 L ι 3 n ρ ι 3 n s ¯ ι 3 n 2 > 0 , 1 2 κ ι 3 n 1 2 ρ ι 3 n L ι 3 n > 0 and r f ι 3 n γ ι 3 n 1 2 > 0 hold, keeping a balance between the faster control convergence time/rate and the input characteristic.

5. Simulation Results

The flight performance of the inverse optimal fault-tolerant formation-containment control protocol was examined using a numerical example. A UH system composed of three leaders and two followers were considered. This leader-following platoon was supposed to accomplish a fault reconfiguration mission, changing from arbitrary positions in the air to an assigned pattern. The related physical parameters of each UH are given in Table 1.
The control objective of this theoretical case analysis is to enable three leaders (labeled as 1, 2, and 3) to maintain a triangle formation and encircle around a particular target or reference provided by a virtual leader. Afterward, the position trajectories of two followers (labeled as 4 and 5) need to be entered into the convex hull spanned by the leaders. The information transmission topology among members in the helicopter system is exhibited in Figure 3. The reference trajectory of the virtual leader can be expressed as x 0 = 15 sin ( 0.2 t ) , y 0 = 15 cos ( 0.2 t ) , z 0 = 0.1 t , and ψ 0 = 1 , and a desired formation pattern of leaders is set with the relative distances as c x 1 d = 0 , c x 2 d = a , c x 3 d = a , c y 1 d = 0 , c y 2 d = a , c y 3 d = a and c z 1 d = c z 2 d = c z 3 d = 0 with a = 5 .
The other related simulation parameters are given as k i 1 n = k ¯ i 1 n = 15 , k i 2 n = k ¯ i 2 n = 10 , k k 1 n = k ¯ k 1 n = 15 , k k 2 n = k ¯ k 2 n = 8 , L i 1 n = L i 1 n = 50 , L k 1 n = L k 1 n = 50 , γ i 1 n = γ i 2 n = γ i 3 n = 20 , γ k 1 n = γ k 2 n = γ k 3 n = 20 , r f i 1 n = r f i 2 n = r f i 3 n = 1 , r f k 1 n = r f k 2 n = r f k 3 n = 1 , κ i 1 n = κ i 2 n = κ i 3 n = 1 , and κ k 1 n = κ k 2 n = κ k 3 n = 1 . The faults are assumed to occur in leaders 2 and 3 and follower 5, while the others operate under normal conditions. The detailed fault parameters are set as follows:
ρ 21 = 1 , 0 < t < 5 s 0.9 t 5 s , ζ 21 = 0 , 0 < t < 5 s 0.1 cos ( t ) t 5 s , ρ 22 = 1 , 0 < t < 5 s 0.8 t 5 s ζ 22 = 0 , 0 < t < 5 s 0.1 cos ( t ) t 5 s , ρ 23 = 1 , 0 < t < 5 s 1 t 5 s , ζ 23 = 0 , 0 < t < 5 s 0.2 cos ( 1.5 t ) t 5 s ρ 24 = 1 , 0 < t < 5 s 0.6 t 5 s , ζ 24 = 0 , 0 < t < 5 s 0.3 cos ( 2.5 t ) t 5 s , ρ 31 = 1 , 0 < t < 5 s 0.9 t 5 s ζ 31 = 0 , 0 < t < 5 s 0.1 sin ( 2 t ) t 5 s , ρ 32 = 1 , 0 < t < 5 s 1 t 5 s , ζ 32 = 0 , 0 < t < 5 s 0.2 cos ( 0.5 t ) t 5 s ρ 33 = 1 , 0 < t < 5 s 0.6 t 5 s , ζ 33 = 0 , 0 < t < 5 s 0.3 cos ( t ) t 5 s , ζ 34 = 0 , 0 < t < 5 s 0.4 cos ( 1.5 t ) t 5 s
ρ 34 = 1 , 0 < t < 5 s 0.7 t 5 s , ρ 51 = 1 , 0 < t < 10 s 0.9 t 10 s , ζ 51 = 0 , 0 < t < 10 s 0.1 sin ( 1.5 t ) t 10 s ρ 52 = 1 , 0 < t < 10 s 1 , t 10 s , ζ 52 = 0 , 0 < t < 10 s 0.2 cos ( t ) , t 10 s , ζ 53 = 0 , 0 < t < 10 s 0.2 cos ( 0.5 t ) , t 10 s ρ 53 = 1 , 0 < t < 10 s 0.7 , t 10 s , ρ 54 = 1 , 0 < t < 10 s 0.4 , t 10 s , ζ 54 = 0 , 0 < t < 10 s 0.4 cos ( t ) , t 10 s
Under the initial conditions of P 1 = [ 0 , 0 , 0 ] T , P 2 = [ 0 , 1 , 0 ] T , P 3 = [ 1 , 0 , 0 ] T , P 4 = [ 1 , 1 , 0 ] T , and P 5 = [ 0 , 2 , 0 ] T , the simulation results are provided in Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9. Figure 4a–d show the moving trajectories (i.e., x ι , y ι , z ι , and ψ ι ) of the five UHs. Figure 5 depicts the formation-containment performance in the 3-orthogonal plane. Figure 6 depicts the formation-containment performance in the XY plane. From this, it can be concluded that the designed protocol can simultaneously make the multi-UHs reach and preserve the target geometric pattern from any position and track the global reference signal by a virtual leader with a specific position. In Figure 7, Figure 8 and Figure 9, it can be observed that the proposed fault estimation observer will approximate the lumped actuator faults Θ 21 n , Θ 22 n , Θ 31 n , Θ 32 n , Θ 51 n , and Θ 52 n well even though the faults occur. Through the above analysis of these helicopter formations and containment statuses, we can conclude that the system realizes the expected formation-containment flight performance under the designed inverse optimal fault-tolerant control protocol, and the fault effects can be compensated well based on the relevant estimations.
Further, to validate the advantages of the proposed inverse optimal fault-tolerant formation-containment control strategy, comparative results were conducted under the same initial conditions and parameters. Specifically, two schemes were considered: one using a non-optimal control law (Case I, i.e., u ι 1 n = ( k ¯ ι 1 n + 1 2 ) e ι P n π ι e ι P n Ψ i 1 n e ι P n Θ ^ ι 1 n and u ι 2 n = ( k ¯ ι 2 n + 1 2 ) e ι Φ n g ι n e ι Φ n ¯ Ψ i 2 n e ι Φ n Θ ^ ι 2 n ), and the other omitting the serial-parallel estimation model (Case II, i.e., r f i 1 n = r f i 2 n = r f i 3 n = 0 , r f k 1 n = r f k 2 n = r f k 3 n = 0 ). The corresponding formation error e ι P n and containment error e ι Φ n are presented in Figure 10, Figure 11 and Figure 12, where Figure 10 presents the performance of the proposed inverse optimal fault-tolerant formation-containment control strategy, Figure 11 displays the error trajectories under Case I while Figure 12 shows its performance. As illustrated, the proposed control protocol improves the tracking accuracy in both formation and containment tasks, compared with Case I and Case II. These improvements highlight the optimality and accurate estimation in enhancing system performance.
To quantitatively evaluate control performance, two reliability indices are introduced: the comprehensive performance index (CPI), defined as n = 1 3 e ι P n 2 + n = 1 3 e ι Λ n 2 1 / 2 , and the mean-squared error (MSE), given by 1 T t = 0 T n = 1 3 e ι P n 2 + n = 1 3 e ι Λ n 2 . The calculation results of these reliability indices also illustrate the effectiveness and the superiority of our proposed control algorithms on formation-containment flight.

6. Conclusions

This article investigated the composite learning-based inverse optimal fault-tolerant control algorithm for multiple UHs with unknown nonlinearities and actuator faults. To solve this problem, a two-layer control framework was constructed by the leaders’ formation control and the followers’ containment control. To improve the approximation accuracy of the compound nonlinearity, the network weight updating law was constructed based on the tracking error and the prediction error derived from the serial-parallel estimation model. Furthermore, the fault estimation observer was designed using the updating law and the prediction error. After that, the active fault-tolerant formation-containment controller was devised to realize the formation reference tracking of the leaders while the followers converged to the convex hull encompassed by the leaders. Finally, a numerical formation-containment experiment was carried out using five UHs to demonstrate the effectiveness and the advantage of the obtained theoretical results.
Future works will mainly concentrate on solving the observer-based control problem for more general helicopter models with disturbances and dynamic flapping, like [15,31,35].

Author Contributions

Conceptualization, Q.L. and K.Z.; methodology, Q.L.; software, Q.L.; validation, Q.L. and K.Z.; data curation, Q.L.; writing—original draft preparation, Q.L.; writing—review and editing, Q.L., K.Z. and B.J.; supervision, K.Z., B.J. and Y.T.; funding acquisition, K.Z. and Y.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China under Grants 62173180, 62020106003, and 62441310; the Science Center Program of National Natural Science Foundation of China under Grant 62188101; Natural Science Foundation of Jiangsu Province of China under Grants BZ2024037 and BK20222012; the National Key Laboratory Foundation of Helicopter Aeromechanics under Grant 2023-HA-LB-067-04; and the Postgraduate Research and Practice Innovation Program of Jiangsu Province (KYCX24_0589).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fang, X.; Wu, A.; Shang, Y.; Dong, N. Robust control of small-scale unmanned helicopter with matched and mismatched disturbances. J. Frankl. Inst. 2016, 353, 4803–4820. [Google Scholar] [CrossRef]
  2. Hernández-González, O.; Ramírez-Rasgado, F.; Farza, M.; Guerrero-Sánchez, M.-E.; Astorga-Zaragoza, C.-M.; M’Saad, M.; Valencia-Palomo, G. Observer for Nonlinear Systems with Time-Varying Delays: Application to a Two-Degrees-of-Freedom Helicopter. Aerospace 2024, 11, 206. [Google Scholar] [CrossRef]
  3. Marantos, P.; Bechlioulis, C.P.; Kyriakopoulos, K.J. Robust trajectory tracking control for small-scale unmanned helicopters with model uncertainties. IEEE Trans. Control Syst. Technol. 2017, 25, 2010–2021. [Google Scholar] [CrossRef]
  4. Jiang, T.; Lin, D.; Song, T. Finite-time control for small-scale unmanned helicopter with disturbances. Nonlinear Dyn. 2019, 96, 1747–1763. [Google Scholar] [CrossRef]
  5. Xian, B.; Zhang, X.; Zhang, H.; Gu, X. Robust adaptive control for a small unmanned helicopter using reinforcement learning. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 7589–7597. [Google Scholar] [CrossRef]
  6. Kuo, C.W.; Tsai, C.C.; Lee, C.T. Intelligent leader-following consensus formation control using recurrent neural networks for small-size unmanned helicopters. IEEE Trans. Syst. Man Cybern. 2019, 51, 1288–1301. [Google Scholar] [CrossRef]
  7. Wang, D.; Zong, Q.; Tian, B.; Shao, S.; Zhang, X.; Zhao, X. Neural network disturbance observer-based distributed finite-time formation tracking control for multiple unmanned helicopters. ISA Trans. 2018, 73, 208–226. [Google Scholar] [CrossRef]
  8. Wang, J.; Wang, D.; Wang, C.; Deng, F. Robust formation control for unmanned helicopters with collision avoidance. J. Frankl. Inst. 2020, 357, 11997–12018. [Google Scholar] [CrossRef]
  9. Zhao, Z.; Zhang, J.; Chen, S.; He, W.; Hong, K.S. Neural-network-based adaptive finite-time control for a two-degree-of-freedom helicopter system with an event-triggering mechanism. IEEE/CAA J. Autom. Sin. 2023, 10, 1754–1765. [Google Scholar] [CrossRef]
  10. Chen, M.; Shi, P.; Lim, C.-C. Adaptive neural fault-tolerant control of a 3-DOF model helicopter system. IEEE Trans. Syst. Man Cybern. Syst. 2015, 46, 260–270. [Google Scholar] [CrossRef]
  11. Liu, Q.; Zhang, K.; Jiang, B.; Xu, J. Prescribed-time fault-tolerant formation control for collision-free unmanned helicopters: A high-order fully actuated system approach. IEEE Trans. Aerosp. Electron. Syst. 2024, 60, 4715–4727. [Google Scholar] [CrossRef]
  12. Wang, Y.; Rotondo, D.; Puig, V.; Cembrano, G. Fault-tolerant control based on virtual actuator and sensor for discrete-time descriptor systems. IEEE Trans. Circuits Syst. I Regul. Pap. 2020, 67, 5316–5325. [Google Scholar] [CrossRef]
  13. Wang, B.; Zhang, Y. An adaptive fault-tolerant sliding mode control allocation scheme for multirotor helicopter subject to simultaneous actuator faults. IEEE Trans. Ind. Electron. 2017, 65, 4227–4236. [Google Scholar] [CrossRef]
  14. Yang, H.; Jiang, B.; Liu, H.H.; Yang, H.; Zhang, Q. Attitude synchronization for multiple 3-DOF helicopters with actuator faults. IEEE/Asme Trans. Mechatron. 2019, 24, 597–608. [Google Scholar] [CrossRef]
  15. Chen, M.; Yan, K.; Wu, Q. Multiapproximator-based fault-tolerant tracking control for unmanned autonomous helicopter with input saturation. IEEE Trans. Syst. Man Cybern. Syst. 2021, 52, 5710–5722. [Google Scholar] [CrossRef]
  16. Ai, S.; Song, J.; Cai, G.; Zhao, K. Active fault-tolerant control for quadrotor uav against sensor fault diagnosed by the auto sequential random forest. Aerospace 2022, 9, 518. [Google Scholar] [CrossRef]
  17. Wang, B.; Shen, Y.; Zhang, Y. Active fault-tolerant control for a quadrotor helicopter against actuator faults and model uncertainties. Aerosp. Sci. Technol. 2020, 99, 105745. [Google Scholar] [CrossRef]
  18. Wang, X.; Tan, C.P. Output feedback active fault tolerant control for a 3-DOF laboratory helicopter with sensor fault. IEEE Trans. Autom. Sci. Eng. 2023, 21, 2689–2700. [Google Scholar] [CrossRef]
  19. Wang, X.; Wang, Y.; Zhang, Z.; Wang, X.; Patton, R. Sensor fault tolerant control for a 3-dof helicopter considering detectability loss. IEEE Trans. Circuits Syst. I Regul. Pap. 2023, 70, 4112–4125. [Google Scholar] [CrossRef]
  20. Liu, C.; Jiang, B.; Zhang, K.; Ding, S.X. Hierarchical structure-based fault-tolerant tracking control of multiple 3-DOF laboratory helicopters. IEEE Trans. Syst. Man Cybern. Syst. 2021, 52, 4247–4258. [Google Scholar] [CrossRef]
  21. Li, K.; Fuzzy, Y.L. adaptive optimal consensus fault-tolerant control for stochastic nonlinear multiagent systems. IEEE Trans. Fuzzy Syst. 2021, 30, 2870–2885. [Google Scholar] [CrossRef]
  22. Liu, Q.; Zhang, K.; Jiang, B. Zero-sum differential game-based optimal fault-tolerant control for interconnected systems with actuator faults. IEEE Trans. Control Netw. Syst. 2023, 11, 1287–1299. [Google Scholar] [CrossRef]
  23. Li, R.; Yang, Z.; Yan, G.; Jian, L.; Li, G.; Li, Z. Robust approximate optimal trajectory tracking control for quadrotors. Aerospace 2024, 11, 149. [Google Scholar] [CrossRef]
  24. Lin, Z.; Liu, Z.; Zhang, Y.; Chen, C.P. Adaptive neural inverse optimal tracking control for uncertain multi-agent systems. Inf. Sci. 2022, 584, 31–49. [Google Scholar] [CrossRef]
  25. Liu, Y.; Li, Y. Application of inverse optimal formation control for euler-lagrange systems. IEEE Trans. Intell. Transp. Syst. 2023, 24, 5655–5662. [Google Scholar] [CrossRef]
  26. An, C.; Su, H.; Chen, S. Inverse-optimal consensus control of fractional-order multiagent systems. IEEE Trans. Syst. Man Cybern. Syst. 2021, 52, 5320–5331. [Google Scholar] [CrossRef]
  27. Yan, F.; Liu, X.; Feng, T. Distributed minimum-energy containment control of continuous-time multi-agent systems by inverse optimal control. IEEE/CAA J. Autom. Sin. 2024, 11, 1533–1535. [Google Scholar] [CrossRef]
  28. Chen, Z.; Yu, Z.; Li, S. Output feedback adaptive fuzzy inverse optimal security control against sensor and actuator attacks for nonlinear cyber-physical systems. IEEE Trans. Fuzzy Syst. 2024, 32, 2554–2566. [Google Scholar] [CrossRef]
  29. Lungu, M.; Dinu, D.-A.; Chen, M.; Flores, G. Inverse optimal control for autonomous carrier landing with disturbances. Aerosp. Sci. Technol. 2023, 139, 108382. [Google Scholar] [CrossRef]
  30. Fan, Z.; Adhikary, A.C.; Li, S.; Liu, R. Disturbance observer based inverse optimal control for a class of nonlinear systems. Neurocomputing 2022, 500, 821–831. [Google Scholar] [CrossRef]
  31. Ma, H.; Chen, M.; Feng, G.; Wu, Q. Disturbance-observer-based adaptive fuzzy tracking control for unmanned autonomous helicopter with flight boundary constraints. IEEE Trans. Fuzzy Syst. 2022, 31, 184–198. [Google Scholar] [CrossRef]
  32. Xu, J.; Cui, Y.; Xing, W.; Huang, F.; Du, X.; Yan, Z.; Wu, D. Distributed active disturbance rejection formation containment control for multiple autonomous underwater vehicles with prescribed performance. Ocean Eng. 2022, 259, 112057. [Google Scholar] [CrossRef]
  33. Yu, Z.; Qu, Y.; Zhang, Y. Distributed fault-tolerant cooperative control for multi-UAVs under actuator fault and input saturation. IEEE Trans. Control Syst. Technol. 2018, 27, 2417–2429. [Google Scholar] [CrossRef]
  34. Chen, L.; Liu, M.; Shi, Y.; Zhang, H.; Zhao, E. Adaptive fault estimation for unmanned surface vessels with a neural network observer approach. IEEE Trans. Circuits Syst. I Regul. Pap. 2020, 68, 416–425. [Google Scholar] [CrossRef]
  35. Campos-Mart, S.-N.; Hern, O.; Guerrero-Sánchez, M.-E.; Valencia-Palomo, G.; Targui, B.; López-Estrada, F.-R. Consensus tracking control of multiple unmanned aerial vehicles subject to distinct unknown delays. Machines 2024, 12, 337. [Google Scholar] [CrossRef]
Figure 1. The diagrammatic sketch of a single UH and the hierarchy structure of multiple UHs. (a) is the inertial reference frame and the body-fixed coordinate frame for a single UH. (b) is the formation-containment structure for multiple UHs.
Figure 1. The diagrammatic sketch of a single UH and the hierarchy structure of multiple UHs. (a) is the inertial reference frame and the body-fixed coordinate frame for a single UH. (b) is the formation-containment structure for multiple UHs.
Drones 09 00391 g001
Figure 2. The inverse optimal fault-tolerant formation-containment control framework for UHs.
Figure 2. The inverse optimal fault-tolerant formation-containment control framework for UHs.
Drones 09 00391 g002
Figure 3. The communication topology among the three leader UHs and two follower UHs.
Figure 3. The communication topology among the three leader UHs and two follower UHs.
Drones 09 00391 g003
Figure 4. Tracking trajectories of the position variables x ι , y ι , z ι , and the yaw angle ψ ι . (ac) are the lateral axis x ι , longitudinal axis y ι , and vertical axis z ι ; (d) is the yaw angle ψ ι .
Figure 4. Tracking trajectories of the position variables x ι , y ι , z ι , and the yaw angle ψ ι . (ac) are the lateral axis x ι , longitudinal axis y ι , and vertical axis z ι ; (d) is the yaw angle ψ ι .
Drones 09 00391 g004
Figure 5. Formation-containment trajectories in the 3-orthogonal plane.
Figure 5. Formation-containment trajectories in the 3-orthogonal plane.
Drones 09 00391 g005
Figure 6. Formation-containment trajectories in XY plane.
Figure 6. Formation-containment trajectories in XY plane.
Drones 09 00391 g006
Figure 7. Estimation trajectories of the lumped faults Θ 21 n and Θ 22 n in UH 2.
Figure 7. Estimation trajectories of the lumped faults Θ 21 n and Θ 22 n in UH 2.
Drones 09 00391 g007
Figure 8. Estimation trajectories of the lumped faults Θ 31 n and Θ 32 n in UH 3.
Figure 8. Estimation trajectories of the lumped faults Θ 31 n and Θ 32 n in UH 3.
Drones 09 00391 g008
Figure 9. Estimation trajectories of the lumped faults Θ 51 n and Θ 52 n in UH 5.
Figure 9. Estimation trajectories of the lumped faults Θ 51 n and Θ 52 n in UH 5.
Drones 09 00391 g009
Figure 10. Formation errors e i P n and e i Φ n , and containment errors e k P n and e k Φ n , under the proposed method.
Figure 10. Formation errors e i P n and e i Φ n , and containment errors e k P n and e k Φ n , under the proposed method.
Drones 09 00391 g010
Figure 11. Comparative results of formation errors e i P n and e i Φ n , and containment errors e k P n and e k Φ n , under Case I.
Figure 11. Comparative results of formation errors e i P n and e i Φ n , and containment errors e k P n and e k Φ n , under Case I.
Drones 09 00391 g011
Figure 12. Comparative results of formation errors e i P n and e i Φ n , and containment errors e k P n and e k Φ n , under Case II.
Figure 12. Comparative results of formation errors e i P n and e i Φ n , and containment errors e k P n and e k Φ n , under Case II.
Drones 09 00391 g012
Table 1. The physical parameters of each helicopter modeling.
Table 1. The physical parameters of each helicopter modeling.
SymbolValue and Unit
m ι 8.2 kg
g9.8 m / s 2
Z ω ι −0.7615 s 1
Z c o ι −131.4125  M / rad · s 2
N c o ι −0.3705 s 2
J ι diag { 0.18 , 0.34 , 0.28 } kg · m 2
A ι diag { 48.1757 , 25.5048 , 0.9808 } s 1

B ι
0 1689.5 0 894.5 0 0 0 0 135.8 s 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Q.; Zhang, K.; Jiang, B.; Tan, Y. Composite Learning-Based Inverse Optimal Fault-Tolerant Control for Hierarchy-Structured Unmanned Helicopters. Drones 2025, 9, 391. https://doi.org/10.3390/drones9060391

AMA Style

Liu Q, Zhang K, Jiang B, Tan Y. Composite Learning-Based Inverse Optimal Fault-Tolerant Control for Hierarchy-Structured Unmanned Helicopters. Drones. 2025; 9(6):391. https://doi.org/10.3390/drones9060391

Chicago/Turabian Style

Liu, Qingyi, Ke Zhang, Bin Jiang, and Yushun Tan. 2025. "Composite Learning-Based Inverse Optimal Fault-Tolerant Control for Hierarchy-Structured Unmanned Helicopters" Drones 9, no. 6: 391. https://doi.org/10.3390/drones9060391

APA Style

Liu, Q., Zhang, K., Jiang, B., & Tan, Y. (2025). Composite Learning-Based Inverse Optimal Fault-Tolerant Control for Hierarchy-Structured Unmanned Helicopters. Drones, 9(6), 391. https://doi.org/10.3390/drones9060391

Article Metrics

Back to TopTop