Next Article in Journal
Nondestructive Detection of Osmotic Damage in GFRP Boat Hulls Using Active Infrared Thermography Methods
Next Article in Special Issue
The Prediction and Dynamic Correction of Drifting Trajectory for Unmanned Maritime Equipment Based on Fully Connected Neural Network (FCNN) Embedding Model
Previous Article in Journal
Research on the Influence of Shallow Buried Tunnel Crossing on the Stability of Overlying Frame Structure Building
Previous Article in Special Issue
Trans-Media Motion Control of a Hybrid Aerial Underwater Vehicle Considering Actuator Dynamic Input Saturation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prescribed Performance-Based Formation Control for Multiple Autonomous Underwater Helicopters with Complex Dynamic Characteristics

1
Donghai Laboratory, Zhoushan 316021, China
2
Ocean College, Zhejiang University, Zhoushan 316021, China
3
Laboratory for Marine Geology, Qingdao Marine Science and Technology Center, Qingdao 266061, China
4
Hainan Institute of Zhejiang University, Sanya 572025, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(12), 2246; https://doi.org/10.3390/jmse12122246
Submission received: 15 November 2024 / Revised: 30 November 2024 / Accepted: 2 December 2024 / Published: 6 December 2024
(This article belongs to the Special Issue Unmanned Marine Vehicles: Perception, Planning, Control and Swarm)

Abstract

:
This research addresses the challenge of formation control among multiple homogeneous autonomous underwater helicopters (AUHs) in the presence of external disturbances and complex dynamic characteristics. The study introduces a novel approach by integrating both disturbance and state observers within the control law framework to manage external disturbances and the immeasurability of velocity, respectively. Concurrently, localized radial basis function neural networks (RBFNNs) of identical configurations are incorporated into the formation control law to assimilate model uncertainties. Building upon this integration, an experience-based formation control strategy is developed, leveraging accumulated knowledge to diminish computational demands while maintaining stipulated performance criteria. Furthermore, the incorporation of a finite-time prescribed performance control (FTPPC) technique enhances the learning process’s efficiency by expediting convergence. Numerical simulations are presented to validate the efficacy of the proposed methodology.

1. Introduction

Underwater exploration plays a vital role in advancing our understanding of marine ecosystems, resource management, and climate regulation [1]. Over the past few decades, humans have sought to explore the vast “universe under oceans” to uncover and utilize hidden resources for the benefit of mankind. However, despite these efforts, only 5% of the world’s oceans have been explored to date, leaving an immense frontier of unknowns [2]. However, conducting underwater exploration poses significant challenges, including harsh environmental conditions, limited visibility, and high-pressure zones, which can impede data collection and system functionality. The development of robust and efficient underwater vehicles, such as autonomous underwater vehicles (AUVs), gliders, and remotely operated systems, has been instrumental in overcoming these challenges [3,4].
Among these technologies, AUVs have attracted significant attention, particularly in formation control, which is critical for cooperative and efficient underwater missions. Researchers have applied various methods to achieve cooperative formation of AUVs, including the leader–follower, virtual structure, and behavior-based methods [5,6,7,8]. Compared to current AUVs, an autonomous underwater helicopter (AUH), as shown in Figure 1, provides more flexible steering, hovering, and landing due to its special shape and propulsion direction, which is more suitable for underwater formation operations [9,10,11].
The complexity of the underwater environment, including its unmeasurable speed, model dynamics, and disturbances, poses significant challenges in designing control law for AUVs. Much work has been carried out to address these issues. An AUV velocity estimation method derived from the Nussbaum state observer was investigated to tackle the problem of unmeasurable velocity [12]. Li et al. [13] presented a finite-time sliding-mode disturbance observer to estimate accurately unknown disturbances in finite time. For disturbances, Gao et al. [14] and Li et al. [15] combined external disturbances with model uncertainties as lumped disturbances and introduced a disturbances observer to estimate them in a finite time. In [16,17], the extended state observer was designed to estimate velocity and disturbances simultaneously. The neural network method has been developed to deal with model uncertainties due to its excellent generalization ability. RBFNNs were used to approximate nonlinear uncertainties in [18,19]. An estimation method derived from the neural-based robustness controller was used to approximate the lumped uncertainty composed of uncertain dynamics, disturbances, and approximate error [20]. Building upon coordinate transformation techniques and a hybrid linear–nonlinear differentiator to estimate the derivative of surge displacement, Zhang et al. [21] introduced an output-feedback adaptive backstepping control strategy tailored for AUVs operating without direct velocity measurements.
The NN methods mentioned above are based on online adaptive neural networks, which means that the learning ability of NN is ignored. Therefore, adaptive methods need to be reused even if AUVs are dealing with similar model uncertainties, which increases the computational burden. Wang et al. [22] proposed the deterministic learning theory to address this problem, which fully utilized the spatially localized learning capabilities of the localized RBFNN. Based on this, a distributed learning strategy composed of a distributed observer and learning controller was proposed in [23]. The cooperative learning controller was presented in [24], which can learn the model uncertainties cooperatively and use knowledge to tackle the uncertainties. In [25], reinforcement learning (RL) is employed to handle partially observable systems by compensating for the lack of state measurements. This algorithm uses a finite history of input–output data to approximate the system dynamics and optimize control actions. In [26], the RL algorithm is utilized to achieve optimal control at a low level for individual agents within the multi-agent formation. It transforms the time-varying system into an equivalent autonomous system, ensuring stability and optimality of the control strategy.
It is important to note that AUHs starts to learn only after getting into the steady state. Therefore, the error convergence rate will affect the learning speed, which poses demands on the convergence rate. The prescribed performance control limits the transient convergence process of tracking errors by performing the error transformation via performance function. In [27,28,29,30], prescribed performance control was proposed to achieve faster convergence, which is realized by using the exponential decay performance function. The traditional method was improved by introducing a finite-time performance function to ensure the trace error gets into the compact set in a specified time [31,32,33].
This study focuses on the formation control problem for homogeneous AUHs with external disturbance and model uncertainty. A cooperative formation event-triggered controller is established based on prescribed performance. Multiple observers estimate the unmeasurable high-order states and external disturbances separately. Finally, the dynamic uncertainties are learned cooperatively, and the learned knowledge is used to establish an empirically based controller.
To contextualize the advancements and challenges in formation control techniques, we conducted a comprehensive review of recent related works. Table 1 provides a detailed comparison of these studies, focusing on key aspects such as the dynamic models considered, robustness to disturbances, the integration of learning mechanisms, and performance metrics. This comparison highlights the strengths and limitations of existing approaches, serving as a foundation to underscore the unique contributions of the proposed method in addressing current gaps in the field.
Hybrid observer design: In contrast to the observer proposed in [14,15], which estimates the lumped uncertainties composed of model uncertainty and disturbance, and the extended state observer presented in [17], which estimates the higher-order states and disturbances simultaneously, the NN-based Luenberger observer and disturbance observer are used to estimate the high-order states and disturbance, respectively, which achieves higher estimation accuracy.
Experience-based control: Compared to the online adaptive neural network used in [18,19], this study maximizes the learning ability of local RBFNN, which enables AUH to learn dynamic uncertainties when moving along the periodic reference trajectory. Based on this, an experience-based control law is designed using the gained knowledge, which means the similar dynamic uncertainties is dealt with by experience rather than by reusing the adaptive method, thus reducing the calculation burden.
Finite-time prescribed performance control: Compared to the current results [23,24,26,34], the finite-time prescribed performance control (FTPPC) is proposed to guarantee that the tracking error converges to a compact set in a finite time, which accelerates the learning process.
The rest of this work is organized as follows. Section 2 contains the preliminaries and control objectives. Section 3 investigates the cooperative formation controller. Section 4 establishes an experience-based cooperative formation controller. Section 5 shows the effectiveness of the strategy through simulation experiments. Finally, Section 6 presents the conclusions.

2. Preliminaries and Problem Formulation

2.1. AUHs Dynamics

This study examines the control problem for a set of N AUHs. The dynamic model of the AUH is based on references [19,35], and its specific expressions are as follows:
η ˙ i = J ( η i ) ν i M ν ˙ i + C ( ν i ) ν i + D ( ν i ) ν i + Δ ( η i , ν i ) + g = τ d , i ( t ) + τ i
where the i 𝒩 = 1 , , N signifies the i-th AUH. M denotes the inertia matrix. η i = x i , y i , z i , φ i , θ i , ψ i T denotes the position and declination of the AUH in the earth-fixed frame. ν i = u i , υ i , w i , p i , q i , r i T represents the linear and angular velocities of the AUH in the body-fixed frame. C ( ν i ) represents the uncertain Coriolis force and the centripetal force matrix. D ( ν i ) denotes the uncertain hydrodynamic damping matrix. Δ ( η i , ν i ) represent unmodeled dynamics. g is the force and matrix of gravity and buoyancy. τ d , i ( t ) and τ i represent external disturbances and control inputs, respectively.
The rotation matrix J ( η i ) between the earth-fixed frame and body-fixed frame can be defined as:
J η i = J a 0 3 × 3 0 3 × 3 J b
with
J a = cos ψ cos θ cos ψ sin θ sin ϕ sin ψ cos ϕ sin ψ sin ϕ + cos ψ sin θ cos ϕ sin ψ cos θ sin θ sin ψ sin ϕ + cos ψ cos ϕ sin ψ sin θ cos ϕ cos ψ sin θ sin θ cos θ sin ϕ cos θ cos ϕ
and
J b = 1 sin ϕ tan θ cos ϕ tan θ 0 cos ϕ sin ϕ 0 sin ϕ / cos θ cos ϕ / cos θ
According to Fossen’s handbook [35], the representation singularity at θ ±   π / 2 in the expression for J b implies that the inverse matrix J 1 ( η i ) does not exist at this value.
The dynamic model can be reformatted as:
x ˙ 1 i = x 2 i x ˙ 2 i = J 1 ( η i ) τ i + J 1 ( η i ) τ d , i ( t ) + F i
where x 1 i , x 2 i T = η i , η ˙ i T , J 1 ( η i ) = J ( η i ) M 1 , and to simplify the notation, J ( η i ) , J 1 ( η i ) is denoted as J and J 1 , respectively, and F i = J J ˙ 1 η ˙ i J M 1 C ( ν i ) J 1 η ˙ i J M 1 D ( ν i ) J 1 η ˙ i J M 1 Δ ( η i , ν i ) J M 1 g represent the total set of dynamic uncertainties.

2.2. Formation Structure and Graph Theory

The ideal AUH formation consists of a virtual leader and N followers; the virtual leader follows the reference path η d and the i-th follower travels along the reference path η r , i .
η r , i = η d + η i *
where η i * is the relative placement of the i-th follower and the virtual leader that establish the formation structure and can be specified by the designer.
The communication topology is dictated by an undirected graph 𝒢 = 𝒱 , ε , where 𝒱 = ϑ 1 , , ϑ N represents the node set and ε ϑ i , ϑ k ϑ i , ϑ k 𝒱 , ϑ i ϑ k represents the edge set. The correlation adjacency matrix 𝒜 = [ a i k ] is determined as a i k = 1 if the i-th AUH is capable of receiving information from the k-th AUH; otherwise, a i k = 0 . Consider 𝒱 i = ϑ k ϑ i , ϑ k ε to be the neighbor set of ϑ i , with k 𝒩 i if ϑ k 𝒱 i . The Laplacian matrix is defined by = = l i k n × n where l i i = k 𝒩 i a i k , l i k = a i k .

2.3. RBFNN

Neural networks are frequently used to estimate unknown or complex functions. This study employed RBFNN to approximate the dynamic uncertainties F i ( x i ) .
F i ( x i ) = W i * T S x i + ε i
where x i = [ x 1 i T , x ^ 2 i T ] T 12 stands for the i-th RBFNN input variables and S x i = S 1 T x i , , S 6 T x i T represents the Gaussian basis function
S j x i = exp x i μ j T x i μ j σ 2
where μ j = μ j , 1 , , μ j , q T q indicates the center vector of the j-th node, σ denotes the basis width, W i * T = blockdiag W i , 1 * T , , W i , 6 * T 6 × 6 q is the ideal weight coefficient matrix, and W i j * q , ε i ε i , 1 , , ε i , 6 T represents the inherent error.
The control law design is based on the following assumptions.
Assumption 1. 
All followers can access the information of virtual leader, and the communication among the followers is described by a connected undirected graph, which means there exists an undirected path between every pair of nodes.
Assumption 2. 
The disturbances  τ d , i ( t )  and their derivative  τ ˙ d , i ( t )   are bounded.
Assumption 3. 
The tracking path  η d  and its derivative is bounded, periodic, or semi-periodic.
The control objectives are listed as follows: the AUHs are driven to maintain the preset formation configuration while tracking the reference trajectory, and the specified performance can be achieved within the limits of the performance function. The experience-based control law is designed to reduce the computing burden by using the experience obtained from learning to approximate the dynamic uncertainty.

3. NN-Based AUH Formation Control Mechanism

3.1. FTPPC

Definition 1 
[19]. A smooth function  β t  can be defined as a finite-time performance function (FTPF) if it satisfies  β t = β t f   for  t t f where  β t f   is a small constant and  t f   is the specified time.
According to Definition 1, an FTPF candidate is proposed:
β ˙ ( t ) = k 1 ( β ( t ) β ) 2 k 2 + ( β ( t ) β ) k 2 , t t f 0 , t > t f
where k 1 = 1 k 2 t f 1 tan 1 β 0 β 1 k 2 , 0 < k 2 < 1 , and β 0 and β are design parameters to constrain the overshoot and stable state error, respectively. Next, we will prove that (7) is a FTPF.
Proof of Definition 1. 
Consider the following Lyapunov function:
V β = 1 2 e β 2
where e β = β ( t ) β . Combined with (7), we obtain:
V ˙ β = k 1 e β ( β ( t ) β ) 2 k 2 + ( β ( t ) β ) k 2 = k 1 e β 2 × 3 k 2 2 + e β 2 × 1 + k 2 2 = k 1 μ 1 V β 1 k 2 + μ 2 V β 1 + k 2 2
where μ 1 = 2 3 k 2 2 and μ 2 = 2 1 + k 2 2 . 1 + k 2 2 < 1 since 0 < k 2 < 1 . Moreover, when e β 0 , V β 1 k 2 > 0 . Thus, e β may converge to 0 in finite time.
Let x β = V β 1 k 2 2 ; then, (9) can be rewritten as follows:
V β 1 + k 2 2 d V β d t = k 1 μ 1 V β 1 k 2 + μ 2 2 1 k 2 d V β 1 k 2 2 d t = k 1 μ 1 V β 1 k 2 + μ 2 2 1 k 2 d x β d t = k 1 μ 1 x β 2 + μ 2 1 μ 1 x β 2 + μ 2 d x β = 1 k 2 k 1 2 d t
By integrating both sides of (10), we obtain:
μ 1 μ 2 1 2 tan 1 μ 1 / μ 2 1 2 x β t = μ 1 μ 2 1 2 tan 1 μ 1 / μ 2 1 2 x β 0 1 k 2 k 1 2 t
Substituting μ 1 μ 2 = 4 into (11) yields:
tan 1 μ 1 / μ 2 1 2 x β t = tan 1 μ 1 / μ 2 1 2 x β 0 1 k 2 k 1 t
Therefore, there exists t f satisfying x β ( t f ) = 0 .
t f = 2 μ 1 μ 2 1 2 1 k 2 k 1 tan 1 μ 1 / μ 2 1 2 x β 0 = 1 1 k 2 k 1 tan 1 β 0 β 1 k 2
Then, we know that β ( t ) tends to β in a finite time t f .
The tracking error transformations are described in the following formula:
e 1 , i j = β i j ( t ) T i j ( z 1 , i j ) j = 1 , , 6
where β i j ( t ) is the FTPF, whose update law is provided in (7), z 1 , i j are the transformed errors, and T i j ( z 1 , i j ) refers to the error transformation function which can be provided by:
T i j ( z 1 , i j ) = exp z 1 , i j exp z 1 , i j exp z 1 , i j + exp z 1 , i j
Substituting (15) into (14) yields:
z 1 , i j = 1 2 ln 1 + e 1 , i j β i j ( t ) 1 2 ln 1 e 1 , i j β i j ( t )

3.2. State Observer Design

We introduce state observer to estimate the high-order states x 2 i . The RBFNN, disturbance observer, and Luenberger observer are combined to estimate the state variables x 2 i due to the uncertainties and disturbances in the dynamic model. The state observer based on RBFNN and disturbance observer are designed as follows:
x ^ ˙ 1 i = x ^ 2 i + L 1 i ( y i y ^ i ) x ^ ˙ 2 i = J 1 τ i + J 1 τ ^ d , i + F ^ i + L 2 i ( y i y ^ i ) y ^ i = x ^ 1 i
where x ^ 1 i , x ^ 2 i , and y ^ i represent the estimation of the state variables and output variables, respectively. τ ^ d , i denotes the estimation of disturbances, F ^ i stands for the approximation of complex uncertainties using RBFNN, and L 1 i and L 2 i represent the gain matrix to be designed. The state estimation error can be calculated as follows:
x ˜ ˙ 1 i = L 1 i x ˜ 1 i + x ˜ 2 i x ˜ ˙ 2 i = L 2 i x ˜ 1 i + J 1 τ ˜ d , i + F ˜ i
where x ˜ 1 i = x 1 i x ^ 1 i and x ˜ 2 i = x 2 i x ^ 2 i stand for state estimation errors, τ ˜ d , i stands for disturbances estimation error, and F ˜ i denotes the approximation error. If x ˜ i = x ˜ 1 i , x ˜ 2 i T , then (18) is expressed in matrix form.
x ˜ ˙ i = A x ˜ i + B F ¯ ˜ i
where A = L 1 i I L 2 i 0 , B = 0 I , F ¯ ˜ i = J 1 τ ˜ d , i + F ˜ i .
Remark 1. 
Although the Luenberger observer is traditionally applied to linear systems, this study extends its application by integrating it with an RBFNN to address the nonlinearities in the AUH model. Specifically, the RBFNN approximates the nonlinear components of the system dynamics, while the Luenberger observer estimates the linearized states. This hybrid design leverages the Luenberger observer’s efficiency in linear dynamics estimation and the RBFNN’s capability to handle nonlinearities.

3.3. NN-Based AUH Formation Controller

The tracking errors are defined as:
e 1 , i = η i η r , i = x 1 i η r , i
Combined with (16), z ˙ 1 , i is calculated as:
z ˙ 1 , i = Υ i e ˙ 1 , i Θ i e 1 , i = Υ i x 2 i η ˙ r , i Θ i e 1 , i
where Υ i = diag χ i 1 , , χ i 6 , Θ i = diag θ i 1 , , θ i 6 , χ i j = 1 2 1 β i j ( t ) + e 1 , i j , + 1 β i j ( t ) e 1 , i j , and θ i j = 1 2 β ˙ i j ( t ) β i j ( t ) β i j ( t ) + e 1 , i j + β ˙ i j ( t ) β i j ( t ) β i j ( t ) e 1 , i j .
Step 1: The Lyapunov function are selected as V 1 = 1 2 i = 1 N x ˜ i T x ˜ i + 1 2 i = 1 N z 1 , i T z 1 , i . According to (19) and (21), the derivative of V 1 can be calculated as:
V ˙ 1 = i = 1 N x ˜ i T A x ˜ i + x ˜ i T B F ¯ ˜ i + z 1 , i T Υ i x ^ 2 i + x ˜ 2 i η ˙ r , i Θ i e 1 , i
Consider x ^ 2 i as the virtual control variable, and z 2 , i is designed as:
z 2 , i = x ^ 2 i α i
where the virtual control law α i is designed as:
α i = Υ i 1 K 1 , i z 1 , i + η ˙ r , i + Υ i 1 Θ i e 1 , i
where K 1 , i = diag [ k 1 , i 1 , , k 1 , i 6 ]   >   0 is the gain matrix, and combining (22), (23), and (24), we obtain:
V ˙ 1 = i = 1 N x ˜ i T A x ˜ i + x ˜ i T B F ¯ ˜ i z 1 , i T K 1 , i z 1 , i + z 1 , i T Υ i z 2 , i + z 1 , i T Υ i x ˜ 2 , i
Next, a nonlinear observer is proposed to estimate the α ˙ i , which avoids the direct computation of α ˙ i .
˙ i = i ς 1 sig 1 1 / ς 3 ( i α i ) ˙ i = ς 2 sig 1 2 / ς 3 ( i α i )
where i and i are the estimates of α i and α ˙ i , respectively. The function sig is defined by sig a ( m ) = sign ( m ) m a . ς 1 > 0 , ς 2 > 0 , and ς 3 > 2 are the parameters.
Considering the estimate error ˜ i = i α i and ˜ i = i α ˙ i , the error dynamic can be calculated as:
˜ ˙ i = ˜ i ς 1 sig 1 1 / ς 3 ( ˜ i ) ˜ ˙ i = ς 2 sig 1 2 / ς 3 ( ˜ i ) α ¨ i
According to [36], the error dynamic (27) is stable, which means the ˜ i will converge to a compact set and satisfy ˜ i ¯ i with ¯ i being a small constant.
Remark 2. 
Note that direct computation of  α ˙ i  requires the use of higher order states  x 2 i , which is unmeasurable and unavailable. And the  α ˙ i  obtained by the differentiator directly is often not smooth. To address this problem, the nonlinear observer is used to estimate the signal  α ˙ i , which avoids both the direct computation of  α ˙ i  and the use of differentiator and is more practical for control law design.
Step 2: The derivative of z 2 , i can be computed as:
z ˙ 2 , i = J 1 τ i + J 1 τ d , i ( t ) + F i x ˜ ˙ 2 i α ˙ i = J 1 τ i + d i + W i * T S x i x ˜ ˙ 2 i α ˙ i
where d i = J 1 τ d , i + ε i represents the sum of disturbances and RBFNN inherent errors, which can be estimated by the disturbance observer.
ξ ˙ i = l i ξ i l i J 1 τ i i + W ^ i T S x i + p i ( z 2 , i ) d ^ i = ξ i + p i ( z 2 , i )
where l i > 0 is the design parameter, p i ( z 2 , i ) = l i z ˙ 2 , i d t , and ξ i is the state variable.
The estimation error dynamic is computed as follows:
d ˜ ˙ i = d ˙ i d ^ ˙ i = l i ξ i + l i J 1 τ i i + W ^ i T S x i + p i ( z 2 , i ) l i z ˙ 2 , i + d ˙ i = l i d ˜ i l i W ˜ i T S x i + l i x ˜ ˙ 2 i l i ˜ i + d ˙ i
where d ˜ i = d i d ^ i , and W ˜ i = W i W ^ i .
Based on the multi-agent consensus and graph theory, and taking full advantage of the information exchange between neighboring AUHs, the weight coefficient update law of cooperative neural network is devised as below:
W ^ ˙ i j = Γ 1 , i j S j x i z 2 , i j + σ i j W ^ i j γ 2 , i j k = 1 N a i k W ^ i j W ^ k j
where Γ 1 , i j , σ i j , and γ 2 , i j are positive constants, Γ 1 , i j S j x i z 2 , i j + σ i j W ^ i j represents the adaptive term for monomeric AUH, and γ 2 , i j k = 1 N a i k W ^ i j W ^ k j stands for the AUHs’ consensus cooperative adaptive term.
Remark 3. 
Although this work focuses on formation control, the consensus problem is introduced as a critical step to ensure consistent information sharing among the AUHs. Specifically, the connection and interaction facilitated through the cooperative neural network ensure rapid convergence of the weight coefficients, thereby enabling a robust and efficient formation control strategy. This perspective aligns with the broader understanding of consensus as a means to achieve coordinated behavior in multi-agent systems.
The NN-based cooperative formation controller is designed using a state observer (17), disturbance observer (29), and updated law (31).
τ i = J 1 1 K 2 , i z 2 , i Υ i z 1 , i + i W ^ i T S x i d ^ i
Theorem 1. 
If we consider the AUHs’ formation system (3), the state observer (17), the disturbance observer (29), the cooperative RBFNN adaptive update law (31), and the NN-based cooperative formation controller (32), then we have (i). All of the state variables of the system are uniformly ultimately bounded (UUB). (ii) The system achieves the specified performance, with trace errors converging in finite time within the limits of the performance function.
The relevant proof procedure for Theorem 1 can be seen in Proof A1 in Appendix A.

4. AUH Formation Control Using Experience

We propose the learning mechanism that enables the NN-based formation controller (32) to learn dynamic uncertainties F i ( x i ) and store the acquired knowledge during the adaptive process. Furthermore, the gained experience will be utilized to establish an experience-based cooperative formation controller which deals with similar dynamic uncertainties F i ( x i ) using experience instead of using the adaptive method repeatedly.

4.1. Learn from Formation Track Control

In this section, our goal is to establish a cooperative formation controller using the localized RBFNN which can learn dynamic uncertainties over a local range along the periodic reference trajectory.
Lemma 1 
([22]). In localized RBF networks, each individual basis function has a limited impact on the network output. In other words, localized RBFNN only has localized representation during deterministic learning. Therefore, for any bounded orbit  x i ( t ) , the model uncertainty  F i ( x i )  can be approximated using neurons located in a local region along the trajectory.
F i ( x i ) = W ζ * T S ζ ( x i ) + ε ζ , i
where ε ζ , i is the inherent error, S ζ ( x i ) q ζ is the regress subvector with the neural node center placed close to the trajectory, i.e., μ j x i < d N , where μ j is the center vector defined in Section 2.
Lemma 2 
([22]). Consider arbitrarily continuous periodic orbit  x i ( t ) : 0 , d  within a bounded set  Ω x i d . Then, for RBFNN  W T S ( x i )  with node centers located on a large enough regular lattice  Ω ζ  that satisfies  Ω x i Ω ζ , the regress subvector  S ζ ( x i )  defined in Lemma 1 satisfies the persistently exciting (PE) condition.
Remark 4. 
It is important to note that the satisfaction of the PE condition is one of the bases of deterministic learning. In other words, deterministic learning occurs when the system tracking the periodic orbit and regress subvector satisfies the PE condition simultaneously.
Theorem 2. 
Consider the AUH system (3) with Assumptions 1~3. For each AUH, if there exists a compact set large enough that  x i Ω x i  holds at any time, then the following conclusion stands if  [ x 1 , i T ( 0 ) , x 2 , i T ( 0 ) ] T Ω x i ( 0 )  and  W ^ i j ( 0 ) = 0 .
(1) The partial PE condition of the subvector is satisfied.
(2) Along the periodic reference trace  φ ζ , i ( x i ( t ) ) | t T i , the weight of the regress subvector will converge to the optimal value  W ζ , i j * .
(3) The local precise approximation of dynamic uncertainties  F i ( x i )   can be realized by  W ^ i T S x i   and  W ¯ i T S x i   simultaneously.
F i ( x i ) = W ^ i T S x i + W ˜ i T S x i + ε i = W ¯ i T S x i + ε ¯ i
where W ¯ i T = blockdiag W ¯ i , 1 T , , W ¯ i , 6 T 6 × 6 q
W ¯ i j = mean W ^ i j ( t ) , t t a , i , t b , i
where t a , i , t b , i is a time interval after the system (3) converges to the transient range, and ε ¯ i = ε ¯ i , 1 , , ε ¯ i , 6 T is close to ε i .
Remark 5. 
The time interval  t a , i , t b , i  should be at least one orbit period after the system converges, which means that enough dynamic information is expressed, and the RBFNN is sufficiently trained.
The relevant proof procedure for Theorem 2 can be seen in Proof A2 in Appendix A.

4.2. Experience-Based AUH Formation Controller

According to Theorem 2, the experience-based RBFNN can locally accurately approximate the dynamic uncertainties along the periodic reference trajectory. The RBFNN weight obtained by the adaptive method is learned by the controller and stored in the W ¯ i T .
By using experience, the experience-based disturbance observer can be specified as follows:
ξ ˙ i = l i ξ i l i J 1 τ i i + W ¯ i T S x i + p i ( z 2 , i ) d ^ i = ξ i + p i ( z 2 , i )
The experience-based cooperative formation controller can be designed as:
τ i = J 1 1 K 2 , i z 2 , i Υ i z 1 , i + i W ¯ i T S x i d ^ i
Theorem 3. 
If we consider the AUH formation system (3) with Assumptions 1~3, the state observer (17), the empirically based disturbance observer (36), the empirically based cooperative formation controller (37), then we have (i). All states in the system are UUB. (ii) The system achieves the specified performance, with trace errors converging in finite time within the limits of the performance function.
The relevant proof procedure for Theorem 3 can be seen in Proof A3 in Appendix A.

5. Simulation Results

To validate the proposed method, a group of four homogeneous AUHs was simulated in a lozenge formation tracking a predefined reference trajectory. The AUHs were modeled with identical dynamic parameters, operating in an environment with external disturbances and communication constraints. This setup represents practical applications such as collaborative underwater inspections, environmental monitoring, and resource exploration. The AUH’s model dynamics is shown in [14]. Table 2 presents the initial pattern of AUH formation. The desired formation shape is a lozenge, and the vertex of the lozenge represents the AUH’s barycenter. We choose the desired relative offsets as η 1 * = 0 , 5 , 0 T , η 2 * = 5 , 0 , 0 T , η 3 * = 0 , 5 , 0 T , and η 4 * = 5 , 0 , 0 T to achieve a lozenge formation shape. The neighboring sets 𝒱 1 = ϑ 2 , ϑ 3 , 𝒱 2 = ϑ 1 , ϑ 4 , 𝒱 3 = ϑ 1 , ϑ 4 , and 𝒱 4 = ϑ 2 , ϑ 3 describe the communication topology. The modeling uncertainties are provided by Δ η i , v i = Δ 1 , Δ 2 , Δ 3 , Δ 4 , Δ 5 , Δ 6 T , where Δ 1 = cos u i 2 + 1 , Δ 2 = sin v i 2 + 1 , Δ 3 = 0.1 w i 2 + 1 , Δ 4 = 1 cos p i 2 , Δ 5 = 1 sin q i 2 , and Δ 6 = 0.2 r i 2 .
We construct the RBFNNs W ^ i j T S j v i , i = 1 4 and j = 1 6 to approximate the modeling uncertainties using 200 nodes with the centers evenly spaced on [−4, 4] and the widths being 6.
The parameters in FTPPC are provided as ρ i 0 = 5 , ρ i = 0.03 , k 2 = 0.8 , and t f = 20 . The parameters in state observer (17), disturbance observer (29), adaptive law (31), and formation control law (32) are given as L 1 i = diag 15 , 15 , 15 , 15 , 15 , 15 , L 2 i = diag 50 , 50 , 50 , 50 , 50 , 50 , K 1 , i = diag 5 , 5 , 5 , 5 , 5 , 5 , K 2 , i = diag 4 , 4 , 4 , 4 , 4 , 4 , ς 1 = 4 , ς 2 = 4 , and ς 3 = 2.5 . The training interval [ t a , i , t b , i ] is set to be [ 20 , 20 + 20 π ] , where 20 is the converge time specified by FTPPC, and 20 π is the orbit period.
Figure 2 illustrates the shape and trajectory of AUH formation. Figure 3 shows the tracking error using the proposed experience-based cooperative formation control law (56), which converges in a finite time within the limit of the performance function. Furthermore, the tracking error remains in a compact set when t > 20 + 20 π , which means the tracking performance is guaranteed when experience is used to deal with the dynamic uncertainty. Figure 4 illustrates the RBFNN approximation error for dynamic uncertainty, which converges to a compact set during the adaptive process. Moreover, it can be seen that the approximation error remains in an admissible set after experience is used at t = ( 20 + 20 π ) s.
Several comparison experiments are provided to further demonstrate the superiority of the proposed algorithm. The experience-based control method in this paper is compared with the adaptive NN-based control method used in [18], which is conducted on a computer with Intel i5-11400 2.60 GHz and 16 GB RAM, and the solver step is set to be 0.01 s in MATLAB 2021a. The result in Table 3 shows that the solver running time is shorter using experience-based control method under the same tracking accuracy, which means the computing burden is reduced. The NN-based state observer (NN-SO) is compared with the ESO proposed in [17], and the result is shown in Figure 5. We can see that the NN-SO has a better estimate accuracy and smaller oscillation magnitude compared to the ESO. Figure 6 provides the results of comparison between the disturbance observer (DO) used in this paper and the ESO proposed in [17]. It can be seen that the disturbance observer has better estimate accuracy compared with the ESO. A comparison between the FTPPC used in our work and the PPC in [18] is shown in Figure 7. We can observe that the FTPPC leads to a faster convergence speed, which is more suitable for deterministic learning control methods.
To highlight the advantages of the proposed formation control algorithm, we conducted a comparative simulation study against a recently published method [37]. This method also employs a prescribed performance control strategy and was chosen for its relevance and novelty in addressing formation control challenges.
To ensure a fair and rigorous comparison, the following conditions were maintained for both algorithms: (1) both controllers were initialized with identical formation configurations and tracking errors. (2) The parameters defining the prescribed performance functions were set equivalently. (3) Both simulations included the same model uncertainty and external disturbances.
The comparative results, presented in Figure 8, demonstrate the following advantages of the proposed algorithm: the proposed method achieved faster error decay during the transient phase compared to the reference algorithm, indicating quicker adaptation to the desired trajectory. The steady-state tracking error achieved by the proposed method was significantly smaller, reflecting higher accuracy and stability under disturbances and uncertainties. These findings confirm that the proposed control algorithm outperforms the reference method in both transient and steady-state performance.

6. Conclusions

This study presents a novel formation control framework for AUHs that integrates prescribed performance control, hybrid observers, and experience-based learning. The proposed method offers several advantages: the hybrid observer, which combines RBFNN with a Luenberger observer, effectively addresses the nonlinear dynamics of AUHs, resulting in higher estimation accuracy. Additionally, the incorporation of experience-based learning reduces the need for continuous adaptation, thereby enhancing computational efficiency. Furthermore, the adoption of finite-time prescribed performance control ensures faster convergence of tracking errors compared to conventional approaches, contributing to improved overall performance. Nevertheless, the proposed method has certain limitations. It relies on reliable communication among AUHs, necessitates sufficient training data to support the experience-based framework, and imposes moderate computational demands on onboard systems. Moreover, a key limitation of this study is the absence of experimental validation using real robots.
To address these limitations, future research will focus on developing decentralized communication frameworks to enhance robustness and designing training methods suitable for highly dynamic environments. In parallel, we are actively pursuing the miniaturization and manufacturing of additional AUH prototypes. These efforts aim to facilitate future experiments with real robots, which will validate the proposed control strategy under realistic operating conditions and further demonstrate its practical applicability.

Author Contributions

Conceptualization, Z.W.; methodology, Z.W. and Z.S.; validation, Z.W. and H.H.; writing—original draft preparation, Z.W. and Z.S.; supervision, H.H.; project administration, H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Project of the Donghai Laboratory, grant number DH-2022ZY0004.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Proof A1. 
Consider the Lyapunov function with the following expressions:
V 2 = V 1 + i = 1 N 1 2 z 2 , i T z 2 , i + 1 2 d ˜ i T d ˜ i + 1 2 j = 1 6 W ˜ i j T Γ 1 , i j 1 W ˜ i j
whose derivative can be calculated as:
V ˙ 2 = i = 1 N { x ˜ i T A x ˜ i + x ˜ i T B F ¯ ˜ i z 1 , i T K 1 , i z 1 , i + z 1 , i T Υ i z 2 , i + z 1 , i T Υ i x ˜ 2 , i + z 2 , i T J 1 τ i + d i + W i * T S x i x ˜ ˙ 2 i α ˙ i + d ˜ i T l i d ˜ i l i W ˜ i T S x i + l i x ˜ ˙ 2 i l i ˜ i + d ˙ i j = 1 6 W ˜ i j T Γ 1 , i j 1 W ^ ˙ i j }
Substituting (32) and (31) into (A2) yields:
V ˙ 2 = i = 1 N { x ˜ i T A x ˜ i z 1 , i T K 1 , i z 1 , i d ˜ i T l i d ˜ i z 2 , i T K 2 , i z 2 , i + z 2 , i T d ˜ i + z 2 , i T ˜ i j = 1 6 σ i j W ˜ i j T W ^ i j z 2 , i T x ˜ ˙ 2 i + z 1 , i T Υ i x ˜ 2 , i d ˜ i T l i W ˜ i T S x i + d ˜ i T l i x ˜ ˙ 2 i d ˜ i T l i ˜ i + d ˜ i T d ˙ i + x ˜ 2 i T W ˜ i T S x i + x ˜ 2 i T d ˜ i j = 1 6 Γ 1 , j 1 γ 2 , j W ˜ j T L I W ˜ j }
where γ 2 , j = diag [ γ 2 , 1 j , , γ 2 , N j ] > 0 , Γ 1 , j 1 = diag [ Γ 1 , 1 j 1 , , Γ 1 , N j 1 ] > 0 , and W ˜ j = [ W ˜ 1 j T , , W ˜ N j T ] T . is positive semidefinite according to Assumption 1, which leads to j = 1 6 Γ 1 , j 1 γ 2 , j W ˜ j T I W ˜ j 0 . Equation (A3) can be further computed as:
V ˙ 2 i = 1 N x ˜ i T A x ˜ i z 1 , i T K 1 , i z 1 , i d ˜ i T l i d ˜ i z 2 , i T K 2 , i z 2 , i + z 2 , i T d ˜ i + z 2 , i T ˜ i j = 1 6 σ i j W ˜ i j T W ^ i j z 2 , i T x ˜ ˙ 2 i + z 1 , i T Υ i x ˜ 2 , i d ˜ i T l i W ˜ i T S x i + d ˜ i T l i x ˜ ˙ 2 i d ˜ i T l i ˜ i + d ˜ i T d ˙ i + x ˜ 2 i T W ˜ i T S x i + x ˜ 2 i T d ˜ i
By Young’s inequality, we obtain:
d ˜ i j T l i W ˜ i j T S j x i d ˜ i j l i W ˜ i j κ 1 l i 2 d ˜ i j 2 2 + W ˜ i j 2 2 κ 1 d ˜ i T d ˙ i κ 2 d ˜ i 2 2 + d ˙ ¯ i 2 2 κ 2 z 2 , i T d ˜ i κ 3 d ˜ i 2 2 + z 2 , i 2 2 κ 3 z 2 , i T ˜ i κ 4 ¯ i 2 2 + z 2 , i 2 2 κ 4 x ˜ 2 , i W ˜ i T S x i x ˜ i W ˜ i κ 5 x ˜ i 2 2 + W ˜ i 2 2 κ 5 x ˜ 2 , i T d ˜ i κ 6 d ˜ i 2 2 + x ˜ i 2 2 κ 6 d ˜ i T l i x ˜ ˙ 2 i κ 7 l i 2 d ˜ i j 2 2 + x ˜ ˙ i 2 2 κ 7 z 2 , i T x ˜ ˙ 2 i κ 8 z 2 , i 2 2 + x ˜ ˙ i 2 2 κ 8 d ˜ i T l i ˜ i κ 9 l i 2 d ˜ i j 2 2 + ¯ i 2 2 κ 9 z 1 , i T Υ i x ˜ 2 , i κ 10 z 1 , i 2 2 + Υ i x ˜ i 2 2 κ 10 σ i j W ˜ i j T W ^ i j σ i j W ˜ i j 2 2 + σ i j W i j * 2 2
where κ 1 , , κ 10 are positive constants, and d ˙ ¯ i is a constant satisfying d ˙ i d ˙ ¯ i . Substituting (A5) into (A4) yields:
V ˙ 2 i = 1 N j = 1 6 σ i j W i j * 2 2 + d ˙ ¯ i 2 2 κ 2 + x ˜ ˙ i 2 2 κ 6 + x ˜ ˙ i 2 2 κ 8 + κ 4 ¯ i 2 2 + ¯ i 2 2 κ 9 1 2 i = 1 N x ˜ i T K x ˜ , i x ˜ i + z 1 , i T K z 1 , i z 1 , i + d ˜ i T K d , i d ˜ i + z 2 , i T K z 2 , i z 2 , i + j = 1 6 W ˜ i j T K W , i W ˜ i j
where K x ˜ , i , K z 1 , i , K d ˜ , i , K z 2 , i , K W , i are provided as follows:
K x ˜ , i = diag 2 A , κ 5 , 1 κ 6 , Υ i κ 10 K z 1 , i = diag 2 K 1 , i , κ 10 K d ˜ , i = diag 2 l 1 , κ 1 l 1 2 , κ 2 , κ 3 , κ 6 , κ 7 l 1 2 , κ 9 l 1 2 K z 2 , i = diag 2 K 2 , i , 1 κ 3 , 1 κ 4 , κ 8 K W , i = diag 1 κ 1 , 1 κ 5 , σ i j
According to (A6) and (A1), we have:
V ˙ 2 J V 2 + δ
where J = min i N , j = 1 , , 6 λ min K x ˜ , i , 2 λ min K 1 , i , λ min K d , i , λ min K z , i , λ min K W , i λ min Γ 1 , i j δ = i = 1 N j = 1 6 σ i j W i j * 2 2 + d ˙ ¯ i 2 2 κ 2 + x ˜ ˙ i 2 2 κ 7 + x ˜ ˙ i 2 2 κ 8 + κ 4 ¯ i 2 2 + ¯ i 2 2 κ 9 .
Inequality (A7) implies that:
V 2 V 2 ( 0 ) e J t + C
where C = = δ J . According to (A1) and (A8), when t tends to infinity, we have:
x ˜ i 2 C , z 1 , i 2 C , z 2 , i 2 C , d ˜ i 2 C , W ˜ i j 2 C λ min Γ 1 , i j 1
According to (A8) and (A9), we have (i) The transformed error converges to the adjustable compact set 2 C , which can be adjusted by parameter A , K 1 , i , K 2 , i , l i , σ i j , and ν . (ii) The state estimation error x ˜ i , disturbance estimation error d ˜ i , weight coefficient estimation error W ˜ i j and virtual control error z 2 , i are UUB. (iii) According to (14), e 1 , i converges to the adjustable compact set in a finite time. □
Proof A2. 
From Theorem 1, we know that x 1 i can converges to η r , i accurately in predefined T i , which indicates x 1 i is periodic for t > T i since η r , i is periodic. In addition, x ^ 2 i will converge to η ˙ r , i since e 1 , i , z 1 , i , z 2 , i will converge to a compact set near 0, which means x ^ 2 i is periodic. Therefore, the input of RBFNN x i = [ x 1 i T , x ^ 2 i T ] T is periodic and restricted to a compact set Ω x i when t > T i . According to Lemma 1, the regress subvector S ζ ( x i ) is persistently exciting.
Next, we will prove that the partial weight coefficient W ^ ζ , i j corresponding to S ζ ( x i ) converges to the optimal value W ζ , i j * if the partial PE condition is satisfied.
Based on (21), (23), and (24), we have:
z ˙ 1 , i = K 1 , i z 1 , i + Υ i z 2 , i + Υ i x ˜ 2 i
According to (5), (28), and (32), we obtain:
z ˙ 2 , i = K 2 , i z 2 , i Υ i z 1 , i x ˜ ˙ 2 i + d ˜ i + F i ( x i ) ε i W ^ i T S x i
where W ^ i T S x i can be represented in the equation based on the localization property of Gaussian function:
W ^ i T S x i = W ^ ζ , i T S ζ x i + W ^ ζ ¯ , i T S ζ ¯ x i
where S ζ x i and S ζ ¯ x i are the regress subvectors with neural nodes center placed close to the trajectory and far from the trajectory, respectively; W ^ ζ , i T and W ^ ζ ¯ , i T are the corresponding weight.
Based on (33) and (A12), (A11) can be further expressed as:
z ˙ 2 , i = K 2 , i z 2 , i Υ i z 1 , i x ˜ ˙ 2 i + d ˜ i + W ˜ ζ , i T S ζ x i + ε ζ , i
where ε ζ , i is the approximation error using the localized RBFNN, which is computed as:
ε ζ , i = ε ζ , i W ^ ζ ¯ , i T S ζ ¯ x i ε i
where ε ζ , i is the inherent error defined in Lemma 1, and W ^ ζ ¯ , i T S ζ ¯ x i is the output component which is not persistently exciting and which is small according to the localization property of Gaussian function [38]. Therefore, we can conclude that ε ζ , i 0 with enough nodes.
According to W ˜ ζ , i j = W ζ , i j * W ^ ζ , i j and (31), we have:
W ˜ ˙ ζ , i j = Γ 1 , i j S ζ , j x i z 2 , i j + σ i j W ^ ζ , i j + γ 2 , i j k = 1 N a i k W ^ ζ , i j W ^ ζ , k j
Then, (47), (48), and (52) can be provided in the compact form:
z ˙ 1 , i z ˙ 2 , i W ˜ ˙ ζ , i , 1 W ˜ ˙ ζ , i , 6 = K 1 , i Υ i 0 Υ i K 2 , i i 0 Γ 1 , i , 1 S ζ , 1 x i γ 2 , i , 1 L I 0 Γ 1 , i , 6 S ζ , 6 x i γ 2 , i , 6 L I z 1 , i z 2 , i W ˜ ζ , i , 1 W ˜ ζ , i , 6 + Υ i x ˜ 2 i x ˜ ˙ 2 i + d ˜ i + ε ζ , i Γ 1 , i , 1 σ i , 1 W ^ ζ , i , 1 Γ 1 , i , 6 σ i , 6 W ^ ζ , i , 6
where i = diag S ζ , 1 x i , , S ζ , 6 x i . The exponential stability property of the nominal part of system (A16) is proved in [37] under the partial PE condition. In addition, ε ζ , i 0 from the previous discussion, Γ 1 , i j σ i j W ^ ζ , i j can be adjusted to a compact value via σ i j , and x ˜ 2 i and d ˜ i will converge to a compact set, which means the error part of (53) is small enough. According to the above discussion, z 1 , i , z 1 , i , and W ˜ ζ , i converge to a compact set, i.e., W ^ ζ , i j will converge to the optimal values W ζ , i j * .
When the AUH moves along the period reference trajectory, we have the following conclusion according to Lemma 1:
F i ( x i ) = W ζ , i * T S ζ x i + ε ζ , i = W ^ ζ , i T S ζ x i + ε ζ 1 , i
where ε ζ 1 , i = W ˜ ζ , i T S x i + ε ζ , i 0 , since W ˜ ζ , i j 0 and ε ζ , i 0 from the above analysis. Due to the convergence of W ^ ζ , i j , we can obtain the constant W ¯ ζ , i j by (35), based on the localized representation property of RBFNN [24]; then, (54) can be further expressed as:
F i ( x i ) = W ^ ζ , i T S ζ x i + ε ζ 1 , i = W ¯ ζ , i T S ζ x i + ε ζ 2 , i
where ε ζ 2 , i is the approximation error using the experience, and ε ζ 2 , i ε ζ 1 , i is small when the system tracks the periodic reference trajectory and reaches the steady state [22], which means ε ζ 2 , i is small.
For an RBFNN with centers far away from the periodic reference trace, the regress subvector S ζ ¯ x i is activated slightly due to the local properties of the Gaussian function, which means that W ^ ζ ¯ , i T S ζ ¯ x i will remain small during tracking reference trajectory; thus, W ¯ ζ ¯ , i T S ζ ¯ x i will also remain small according to (35).
From the above discussion and (A18), the global RBFNN W ^ i T S x i and W ¯ i T S x i also have the ability to provide a locally accurate approximation of the dynamic uncertainties during tracking trajectory φ ζ , i ( x i ( t ) ) | t T i .
F i ( x i ) = W ¯ ζ , i T S ζ x i + ε ζ 2 , i = W ¯ i T S x i + ε ¯ i
where ε ¯ i = ε ζ 2 , i W ¯ ζ ¯ , i T S ζ ¯ x i are small approximation errors. This ends the proof. □
Proof A3. 
According to (28), (34), and (37), we have:
z ˙ 2 , i = J 1 τ i + J 1 τ d , i ( t ) + F i α ˙ i x ˜ ˙ 2 i = J 1 τ i α ˙ i + d i + W ¯ i T S x i x ˜ ˙ 2 i
where d i = J 1 τ d , i ( t ) + ε ¯ i .
From (36) and (A20), we have:
d ˜ ˙ i = l i d ˜ i + l i x ˜ ˙ 2 i l i ˜ i + d ˙ i
The Lyapunov function is selected as:
V 3 = V 1 + i = 1 N 1 2 z 2 , i T z 2 , i + 1 2 d ˜ i T d ˜ i
whose derivative is calculated as:
V ˙ 3 = i = 1 N x ˜ i T A x ˜ i + x ˜ i T B F ¯ ˜ i z 1 , i T K 1 , i z 1 , i + z 1 , i T Υ i z 2 , i + z 1 , i T Υ i x ˜ 2 , i + z 2 , i T J 1 τ i α ˙ i + d i + W ¯ i T S x i x ˜ ˙ 2 i + d ˜ i T l i d ˜ i + l i x ˜ ˙ 2 i l i ˜ i + d ˙ i i = 1 N x ˜ i T A x ˜ i z 1 , i T K 1 , i z 1 , i d ˜ i T l 1 d ˜ i z 2 , i T K 2 , i z 2 , i + z 2 , i T d ˜ i + z 2 , i T ˜ i z 2 , i T x ˜ ˙ 2 i + z 1 , i T Υ i x ˜ 2 , i + d ˜ i T l i x ˜ ˙ 2 i d ˜ i T l i ˜ i + d ˜ i T d ˙ i + x ˜ 2 i T d ˜ i
According to (A23), we have:
V ˙ 2 J ¯ V 2 + δ ¯
where J ¯ = min i N , j = 1 , , 6 λ min K x ˜ , i , 2 λ min K 1 , i , λ min K d , i , λ min K z , i , δ ¯ = i = 1 N d ˙ ¯ i 2 2 κ 2 + x ˜ ˙ i 2 2 κ 7 + x ˜ ˙ i 2 2 κ 8 + κ 4 ¯ i 2 2 + ¯ i 2 2 κ 9 . Equation (61) implies that:
V 2 V 2 ( 0 ) e J ¯ t + C ¯
where C ¯ = = δ ¯ J ¯ . From (A22) and (A25), when t tends to infinity, the following inequalities hold:
x ˜ i 2 C ¯ , z 1 , i 2 C ¯ , z 2 , i 2 C ¯ , d ˜ i 2 C ¯
Following a proof that is similar to that of Theorem 1, (i) e 1 , i converges to a compact set in a finite time. (ii) All states in the system are UUB. (iii) Within the limits of the FTPF, the system meets the preset performance. The proof is thus accomplished. □

References

  1. Joshi, B.; Xanthidis, M.; Roznere, M.; Burgdorfer, N.J.; Mordohai, P.; Li, A.Q.; Rekleitis, I. Underwater exploration and mapping. In Proceedings of the 2022 IEEE/OES Autonomous Underwater Vehicles Symposium (AUV), Singapore, 19–21 September 2022; pp. 1–7. [Google Scholar]
  2. Mirza, J.; Kanwal, F.; Salaria, U.A.; Ghafoor, S.; Aziz, I.; Atieh, A.; Almogren, A.; Haq, A.U.; Kanwal, B. Underwater temperature and pressure monitoring for deep-sea SCUBA divers using optical techniques. Front. Phys. 2024, 12, 1417293. [Google Scholar] [CrossRef]
  3. Mirza, J.; Atieh, A.; Kanwal, B.; Ghafoor, S.; Almogren, A.; Kanwal, F.; Aziz, I. Relay aided UWOC-SMF-FSO based hybrid link for underwater wireless optical sensor network. Opt. Fiber Technol. 2025, 89, 104045. [Google Scholar] [CrossRef]
  4. Wibisono, A.; Piran, M.J.; Song, H.K.; Lee, B.M. A survey on unmanned underwater vehicles: Challenges, enabling technologies, and future research directions. Sensors 2023, 23, 7321. [Google Scholar] [CrossRef]
  5. Thuyen, N.A.; Thanh, P.N.N.; Anh, H.P.H. Adaptive finite-time leader-follower formation control for multiple AUVs regarding uncertain dynamics and disturbances. Ocean Eng. 2023, 269, 113503. [Google Scholar] [CrossRef]
  6. Wang, J.; Wang, C.; Wei, Y.; Zhang, C. Bounded neural adaptive formation control of multiple underactuated AUVs under uncertain dynamics. ISA Trans. 2020, 105, 111–119. [Google Scholar] [CrossRef] [PubMed]
  7. Zhang, Y.; Wang, Q.; Shen, Y.; Dai, N.; He, B. Multi-AUV cooperative control and autonomous obstacle avoidance study. Ocean Eng. 2024, 304, 117634. [Google Scholar] [CrossRef]
  8. Zhuang, Y.; Huang, H.; Sharma, S.; Xu, D.; Zhang, Q. Cooperative path planning of multiple autonomous underwater vehicles operating in dynamic ocean environment. ISA Trans. 2019, 94, 174–186. [Google Scholar] [CrossRef]
  9. Li, H.; An, X.; Feng, R.; Chen, Y. Motion control of autonomous underwater helicopter based on linear active disturbance rejection control with tracking differentiator. Appl. Sci. 2023, 13, 3836. [Google Scholar] [CrossRef]
  10. Wang, Q.; Wu, Z.; Xie, M.; Wu, F.; Huang, H. Finite-time prescribed performance trajectory tracking control for the autonomous underwater helicopter. Ocean Eng. 2023, 280, 114628. [Google Scholar] [CrossRef]
  11. Wu, Z.; Wang, Q.; Huang, H. Low-complexity tracking for autonomous underwater helicopters with event-triggered mechanism. Ocean Eng. 2023, 280, 114633. [Google Scholar] [CrossRef]
  12. Liu, X.; Zhang, M.; Yao, F.; Chu, Z. Observer-based region tracking control for underwater vehicles without velocity measurement. Nonlinear Dyn. 2022, 108, 3543–3560. [Google Scholar] [CrossRef]
  13. Li, J.; Tian, Z.; Zhang, H.; Li, W. Robust finite-time control of a multi-AUV formation based on prescribed performance. J. Mar. Sci. Eng. 2023, 11, 897. [Google Scholar] [CrossRef]
  14. Gao, Z.; Guo, G. Fixed-time sliding mode formation control of AUVs based on a disturbance observer. IEEE/CAA J. Automat. Sin. 2020, 7, 539–545. [Google Scholar] [CrossRef]
  15. Li, X.; Qin, H.; Li, L. Fixed-time formation control for AUVs with unknown actuator faults based on lumped disturbance observer. Ocean Eng. 2023, 269, 113495. [Google Scholar] [CrossRef]
  16. Li, Z.; Wang, M.; Ma, G.; Zou, T. Adaptive reinforcement learning fault-tolerant control for AUVs with thruster faults based on the integral extended state observer. Ocean Eng. 2023, 271, 113722. [Google Scholar] [CrossRef]
  17. Kong, S.; Sun, J.; Qiu, C.; Wu, Z.; Yu, J. Extended State Observer-Based Controller with Model Predictive Governor for 3-D Trajectory Tracking of Underactuated Underwater Vehicles. IEEE Trans. Ind. Inf. 2021, 17, 6114–6124. [Google Scholar] [CrossRef]
  18. Fang, K.; Fang, H.; Zhang, J.; Yao, J.; Li, J. Neural adaptive output feedback tracking control of underactuated AUVs. Ocean Eng. 2021, 234, 109211. [Google Scholar] [CrossRef]
  19. Wu, Z.; Wang, Q.; Huang, H. Adaptive neural networks trajectory tracking control for autonomous underwater helicopters with prescribed performance. Ocean Eng. 2022, 264, 112519. [Google Scholar] [CrossRef]
  20. Wang, J.; Wang, C.; Wei, Y.; Zhang, C. Observer-Based Neural Formation Control of Leader-Follower AUVs with Input Saturation. IEEE Syst. J. 2021, 15, 2553–2561. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Xu, O. Adaptive Backstepping Axial Position Tracking Control of Autonomous Undersea Vehicles with Deferred Output Constraint. Appl. Sci. 2023, 13, 2219. [Google Scholar] [CrossRef]
  22. Wang, C.; Hill, D.J. Learning from neural control. IEEE Trans. Neural Netw. 2006, 17, 130–146. [Google Scholar] [CrossRef] [PubMed]
  23. Yuan, C.; Licht, S.; He, H. Formation Learning Control of Multiple Autonomous Underwater Vehicles with Heterogeneous Nonlinear Uncertain Dynamics. IEEE Trans. Cybern. 2018, 48, 2920–2934. [Google Scholar] [CrossRef] [PubMed]
  24. Dai, S.-L.; He, S.; Ma, Y.; Yuan, C. Cooperative Learning-Based Formation Control of Autonomous Marine Surface Vessels with Prescribed Performance. IEEE Trans. Syst. Man Cybern. Syst. 2022, 52, 2565–2577. [Google Scholar] [CrossRef]
  25. Yaghmaie, F.A.; Modares, H.; Gustafsson, F. Reinforcement Learning for Partially Observable Linear Gaussian Systems Using Batch Dynamics of Noisy Observations. IEEE Trans. Autom. Control 2024, 69, 6379–6404. [Google Scholar] [CrossRef]
  26. Nguyen, K.; Dang, V.T.; Pham, D.D.; Dao, P.N. Formation control scheme with reinforcement learning strategy for a group of multiple surface vehicles. Int. J. Robust. Nonlinear Control 2024, 34, 2252–2279. [Google Scholar] [CrossRef]
  27. Huang, H.; He, W.; Li, J.; Xu, B.; Yang, C.; Zhang, W. Disturbance Observer-Based Fault-Tolerant Control for Robotic Systems with Guaranteed Prescribed Performance. IEEE Trans. Cybern. 2022, 52, 772–783. [Google Scholar] [CrossRef]
  28. Li, Z.; Ma, Y.; Yue, D.; Zhao, J. Adaptive Tracking for Uncertain Switched Nonlinear Systems with Prescribed Performance Under Slow Switching. IEEE Trans. Syst. Man Cybern. Syst. 2022, 52, 7279–7288. [Google Scholar] [CrossRef]
  29. Liu, L.; Liu, Y.-J.; Tong, S. Fuzzy-Based Multierror Constraint Control for Switched Nonlinear Systems and Its Applications. IEEE Trans. Fuzzy Syst. 2019, 27, 1519–1531. [Google Scholar] [CrossRef]
  30. Xu, Z.; Sun, C.; Liu, Q. Output-Feedback Prescribed Performance Control for the Full-State Constrained Nonlinear Systems and Its Application to DC Motor System. IEEE Trans. Syst. Man Cybern. Syst. 2023, 53, 3898–3907. [Google Scholar] [CrossRef]
  31. Qiu, J.; Wang, T.; Sun, K.; Rudas, I.J.; Gao, H. Disturbance Observer-Based Adaptive Fuzzy Control for Strict-Feedback Nonlinear Systems with Finite-Time Prescribed Performance. IEEE Trans. Fuzzy Syst. 2022, 30, 1175–1184. [Google Scholar] [CrossRef]
  32. Sui, S.; Chen, C.L.P.; Tong, S. A Novel Adaptive NN Prescribed Performance Control for Stochastic Nonlinear Systems. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 3196–3205. [Google Scholar] [CrossRef] [PubMed]
  33. Wang, H.; Bai, W.; Zhao, X.; Liu, P.X. Finite-Time-Prescribed Performance-Based Adaptive Fuzzy Control for Strict-Feedback Nonlinear Systems with Dynamic Uncertainty and Actuator Faults. IEEE Trans. Cybern. 2022, 52, 6959–6971. [Google Scholar] [CrossRef] [PubMed]
  34. Dai, S.-L.; Wang, M.; Wang, C. Neural Learning Control of Marine Surface Vessels with Guaranteed Transient Tracking Performance. IEEE Trans. Ind. Electron. 2016, 63, 1717–1727. [Google Scholar] [CrossRef]
  35. Fossen, T.I. Handbook of Marine Craft Hydrodynamics and Motion Control; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  36. Yi, S.; Wang, J.; Li, B. Composite backstepping control with finite-time convergence. Optik 2017, 142, 260–272. [Google Scholar] [CrossRef]
  37. Xie, M.; Wu, Z.; Huang, H. Low-complexity formation control of marine vehicle system based on prescribed performance. Nonlinear Dyn. 2024, 112, 18311–18332. [Google Scholar] [CrossRef]
  38. Wang, C.; Hill, D.J. Deterministic Learning Theory for Identification, Recognition, and Control; CRC Press: Boca Raton, FL, USA, 2009. [Google Scholar]
Figure 1. Autonomous underwater helicopter (AUH).
Figure 1. Autonomous underwater helicopter (AUH).
Jmse 12 02246 g001
Figure 2. The shape and trajectory of AUH formation.
Figure 2. The shape and trajectory of AUH formation.
Jmse 12 02246 g002
Figure 3. The tracking error. (a) Surge. (b) Lateral. (c) Heave. (d) Pitch. (e) Roll. (f) Yaw.
Figure 3. The tracking error. (a) Surge. (b) Lateral. (c) Heave. (d) Pitch. (e) Roll. (f) Yaw.
Jmse 12 02246 g003aJmse 12 02246 g003b
Figure 4. The NN approximation error. (a) AUH 1 in translational degrees, (b) AUH 1 in rotation degrees, (c) AUH 2 in translational degrees, and (d) AUH 2 in rotation degrees.
Figure 4. The NN approximation error. (a) AUH 1 in translational degrees, (b) AUH 1 in rotation degrees, (c) AUH 2 in translational degrees, and (d) AUH 2 in rotation degrees.
Jmse 12 02246 g004
Figure 5. The observer estimation for high-order state using NN-SO and ESO. (a) Surge. (b) Lateral. (c) Heave. (d) Pitch. (e) Roll. (f) Yaw.
Figure 5. The observer estimation for high-order state using NN-SO and ESO. (a) Surge. (b) Lateral. (c) Heave. (d) Pitch. (e) Roll. (f) Yaw.
Jmse 12 02246 g005
Figure 6. The observer estimation for disturbance using DO and ESO. (a) Estimation error in translational degrees. (b) Estimation error in rotation degrees.
Figure 6. The observer estimation for disturbance using DO and ESO. (a) Estimation error in translational degrees. (b) Estimation error in rotation degrees.
Jmse 12 02246 g006
Figure 7. The tracking error using PPC method and FTPPC method. (a) Tracking error in translational degrees. (b) Tracking error in rotation degrees.
Figure 7. The tracking error using PPC method and FTPPC method. (a) Tracking error in translational degrees. (b) Tracking error in rotation degrees.
Jmse 12 02246 g007
Figure 8. Comparison with state-of-the-art formation control [37]. (a) Tracking error in surge. (b) Tracking error in lateral. (c) Tracking error in heave. (d) Tracking error in yaw.
Figure 8. Comparison with state-of-the-art formation control [37]. (a) Tracking error in surge. (b) Tracking error in lateral. (c) Tracking error in heave. (d) Tracking error in yaw.
Jmse 12 02246 g008
Table 1. Comparison of related works on formation control techniques.
Table 1. Comparison of related works on formation control techniques.
StudyDynamic Model
Consideration
Robustness to
Disturbances
Learning
Mechanism
Performance
Metrics
[12,18,22]Unknown dynamicsModerateNoneUUB errors
[13]Feedback linearizationHighNoneFinite-time
convergence
[14,15]Unknown dynamics,
external disturbances
HighNoneFixed-time
convergence
[16]Thruster faults, unknown
disturbances
ModerateReinforcement learningUUB errors
[17,27,29]Unknown disturbancesModerateNoneUUB errors
[19]Thruster faults, unknown
disturbances
LowNoneSpecified convergence time
[20]Unknown dynamics,
input saturation
LowNoneUUB errors
[23]UncertaintiesHighDeterministic learningUUB errors
[24]Unknown disturbances,
uncertainties
HighCooperative learningUUB errors
[26]Dynamic modelsModerateReinforcement learningUUB errors
[30]Full-state constraints,
disturbances
ModerateNoneSpecified convergence time
ProposedUnknown dynamics, uncertainties, external disturbancesHighExperience-based learningSpecified convergence time
Table 2. The initial information on AUH formation.
Table 2. The initial information on AUH formation.
TermsValues
The initial states η i ( 0 ) = [ 20 , 30 , 0 , 0 , 0 , 0 ] T v i ( 0 ) = [ 0 , 0 , 0 , 0 , 0 , 0 ] T
The reference trajectory of the virtual leader η d η d = [ 30 sin 0.1 t + 20 cos 0.1 t , 20 sin 0.1 t 30 cos 0.1 t , 0 , 0 , 0 , 0.1 t ] T
The external disturbance τ d , i τ d , i = [ 20 + 5 cos 0.2 t + π 4 , 15 + 10 sin 0.2 t , 0 , 5 cos 0.2 t + π 4 , 0 , 10 + 5 sin 0.5 t ] T
Table 3. The comparison experiment results.
Table 3. The comparison experiment results.
MethodsThe Adaptive NN-Based Control MethodThe Experience-Based
Control Method
Simulation time setting (s)200200
Actual running time (s)97.6283.67
Tracking error
in translation degree (m)
<0.03<0.03
Tracking error
in rotation degree (rad)
<0.03<0.03
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, Z.; Song, Z.; Huang, H. Prescribed Performance-Based Formation Control for Multiple Autonomous Underwater Helicopters with Complex Dynamic Characteristics. J. Mar. Sci. Eng. 2024, 12, 2246. https://doi.org/10.3390/jmse12122246

AMA Style

Wu Z, Song Z, Huang H. Prescribed Performance-Based Formation Control for Multiple Autonomous Underwater Helicopters with Complex Dynamic Characteristics. Journal of Marine Science and Engineering. 2024; 12(12):2246. https://doi.org/10.3390/jmse12122246

Chicago/Turabian Style

Wu, Zheyuan, Zilong Song, and Haocai Huang. 2024. "Prescribed Performance-Based Formation Control for Multiple Autonomous Underwater Helicopters with Complex Dynamic Characteristics" Journal of Marine Science and Engineering 12, no. 12: 2246. https://doi.org/10.3390/jmse12122246

APA Style

Wu, Z., Song, Z., & Huang, H. (2024). Prescribed Performance-Based Formation Control for Multiple Autonomous Underwater Helicopters with Complex Dynamic Characteristics. Journal of Marine Science and Engineering, 12(12), 2246. https://doi.org/10.3390/jmse12122246

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop