Next Article in Journal
Experimental Study on the Mechanical Properties of Vertical Corrugated Pipe Grout Anchor Connection Joints
Previous Article in Journal
Cooperative Computing Offloading Scheme via Artificial Neural Networks for Underwater Sensor Networks
Previous Article in Special Issue
High-Performance Flux Tracking Controller for Reluctance Actuator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Optimal Control of UAV Brake Based on Parameter Estimation

College of Intelligent Equipment, Shandong University of Science and Technology, Tai’an 271001, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(21), 11888; https://doi.org/10.3390/app132111888
Submission received: 8 July 2023 / Revised: 28 July 2023 / Accepted: 14 October 2023 / Published: 30 October 2023
(This article belongs to the Special Issue New Technology for Autonomous UAV Monitoring)

Abstract

:
A new parameter estimation method for a UAV braking system with unknown friction parameters is suggested. The unknown part containing friction is separated from the coupling system. The law of parameter updating is driven by the estimation error extracted by the auxiliary filter. A controller is developed to guarantee the convergence of the parameter estimation and tracking errors simultaneously. Furthermore, a designed performance index function enables the system to track the desired slip rate optimally. The optimal value of the performance function is updated by a single critic neural network (NN) to obtain comprehensive optimization of the control energy consumption, dynamic tracking error, and filtering error during braking. In addition, the unknown bounded interference and neural network approximation errors are compensated by robust integral terms. Simulation results are presented to confirm the effectiveness of the proposed control scheme.

1. Introduction

With the rapid development of unmanned aerial vehicles (UAVs), the safety and efficiency of braking systems have drawn significant attention. Due to the coupling mechanism of vertical and longitudinal torques, the highly nonlinear aircraft dynamics model, and time-varying friction, it is challenging to establish an accurate mathematical model [1]. The most common control strategy for brake systems is slip rate control, which obtains the maximum tire–road friction by regulating the slip rate. Several methods have been proposed, such as dynamic surface control [2], adaptive control [3], and extreme value search control [4]. These methods focus on adjusting the slip rate in real time to track the expected slip rate but only partially consider an excellent dynamic performance and the system’s robustness. The ideal brake control should ensure that the brake operating point is always in the stable region. A rule-based control strategy has been studied in [5,6] to avoid the brake operating point being in an unstable region, which requires high model accuracy and has a complex controller structure. In [6], an adaptive radial basis function (RBF) neural network (NN) is used to predict the upper bound of the system’s complex external interference to improve the brake control’s robustness. The RBFNN and error backpropagation NN were used, respectively, to approximate the nonlinear interference of reduction gear deformation and switched reluctance motor in an Electro Mechanical Actuator (EMA) [7,8]. However, these models need to be trained offline in advance.
The friction coefficient affecting the binding force between the tire and runway in brake systems is influenced by external environments and various factors, making it challenging to measure directly or calculate accurately. Previous studies [9,10] proposed a fitting function between aircraft speed and slip rate. Still, this method requires a large amount of data for offline fitting and may not reflect the complex and changeable runway environment in real time and reduce control accuracy. Ref. [11] proposed the estimation of aircraft speed and ground–tire joint force based on the Kalman filter, but this requires additional sensors or algorithmic complexity, leading to practical difficulties. In addition, uncertain nonlinear systems with unmodeled dynamics can also be approximated by neural networks (NNs) [12,13,14] and fuzzy logic systems (FLSs) [15,16,17]. In these methods, the unknown friction is considered part of the nonlinearity of the unknown system and approximated by NN or FLS. Ref. [18] proposes an adaptive feedforward NN control scheme to deal with the pandybat friction effect, assuming that the friction model parameters are known. Moreover, traditional adaptive control design often adopts methods such as gradient [19], least squares [20], and the modified parameter estimation method [21]. However, none of the mentioned achieve convergence of the estimated parameters and tracking errors simultaneously. A new parameter estimation method [22] is proposed to ensure the convergence of parameters. In the controller design, the adaptive rate based on the parameter estimation method ensures that the unknown parameter converges to the true value and the tracking error converges to zero.
Moreover, as the balance criterion of resource control and control effect plays an increasingly important role in the design process, optimal control has attracted wide attention [23,24,25,26,27]. At present, the optimal control theory related to the design of the controller has been a significant research result. It can meet the tracking tasks or control objectives, while minimizing some performance indicators. In the classical optimal control method, the derived Riccati or Hamilton–Jacobi–Bellman (HJB) equation is usually solved offline iteratively [28]. In recent years, adaptive dynamic programming (ADP) has been proven to be a feasible method to solve the optimal control solution online [29]. In the classical critic–actor-based ADP framework [30], the complex ADP structure of multiple NN approximators increases the computational burden in the implementation. Thus, a simple and effective ADP structure based on a single critic NN is proposed in [31,32], in which the convergence of critic NN weight can be retained so that control actions can be calculated directly through the weight of a critic NN. Due to the fact that the approximate error of the NN cannot be compensated for, the steady-state error can only be guaranteed to converge to a small set near zero. A new error sign robust integral (RISE) control is developed in [33] to compensate for bounded uncertainties and perturbations. The key feature of the RISE method is that asymptotic stability can be proved by using an integral of the sign of the error feedback term. In addition, compared with the traditional Sliding Mode Control (SMC) method [34], this scheme generates continuous control signals, thus avoiding the requirement of infinitely fast control switching.
Motivated by the above discussions, the aim of this article is to develop a new adaptive law to ensure convergence of estimation parameters and tracking errors in UAV braking systems. A performance index function of error and input is established for the braking system, considering the balance criterion of control resources and control effect, and the optimal solution is updated online with ADP. Moreover, the improved adaptive optimal control strategy includes robust feedback items. Their purpose is to compensate bounded disturbance, modeling uncertainty, and NN estimation error, minimizing the cost function while tracking slip rate, reducing the use of control resources, and improving control performance. Comparative simulations are given to validate the effectiveness.
The contributions of this paper are as follows.
(1)
A parameter-updating law driven by parameter estimation error is designed for UAV braking systems with unknown parameters. The contradiction between system robustness and parameter convergence is solved, and the tracking error and estimation error converge simultaneously.
(2)
The performance index function about the error and input of the UAV braking system is established. The adaptive dynamic programming based on a single critic NN is adopted to solve the function so that the control energy consumption, dynamic tracking error, and filtering error in the braking process are optimized comprehensively.

2. Dynamic Model

Braking Dynamics of UAV

The dynamics model of the UAV braking system generally includes the body dynamics model, braking wheel dynamics model, tire–runway friction, and brake actuator. Brake control design should be based on a reasonable and simplified mathematical model. According to [35], consider the system formulated as
λ ˙ = 1 v x R 2 J w + n ( 1 λ ) m A F Z μ + B w R ω v x J w 1 λ v x F X m A + K b R v x J w P A , P ˙ A = c b L 0 ω m / 2 π + Δ 1 , ω ˙ m = 1 J L 0 2 π P A B v J ω m + K T J i + Δ 2 .
where Δ 1 and Δ 2 are unknown bounded disturbances. The above parameters are described in Table 1. The slip rate in the system is defined as
λ = v x R ω v x
The slip rate represents the ratio of the sliding motion of the wheel to the runway; thus, 0 λ 1 . The braking process depends on the binding force between the brake wheel and the ground, which is defined as
F z = μ F f
It is affected by many factors and cannot be directly measured or accurately calculated.
We define the state of the system as x 1 , x 2 , x 3 = λ , P A , ω m T , and then the system is described as
x ˙ 1 = f x 1 + K 1 x 2 x ˙ 2 = K 2 x 3 + Δ 1 x ˙ 3 = K 3 u K 4 x 2 K 5 x 3 + Δ 2 y = λ
where u = i , K 1 = R k b v x J w , K 2 = c b L 0 2 π , K 3 = k T J m , K 4 = L 0 2 π J m , and K 5 = B v J m are positive constants.
Rewrite the derivative of f x 1 in (4) as f ˙ x 1 = Θ T Ψ x 1 , where Θ is the unknown parameter vector, and Ψ x 1 is the known regressor vector.
Let x 1 d be the expected slip rate set, and define the tracking error e 1 as
e 1 = x 1 x 1 d
Assumption 1.
The perturbation Δ 1 , Δ 2 and its derivatives Δ ˙ 1 , Δ ˙ 2 , and Δ ¨ 1 , Δ ¨ 2 are bounded.
Assumption 2.
Linear velocity v x , angular velocity of the wheel ω, brake pressure P A , motor speed ω m , and armature current i can be measured.
The problems to be solved in this paper are as follows:
  • Online estimation of unknown parameters of UAV braking system to ensure convergence.
  • When t > 0 , tracking errors converge quickly and steady-state tracking errors remain within a small range.
  • The system control input and filtering error meet the performance function.

3. Adaptive Optimal Control Design with Guaranteed Parameter Estimation

This section proposes an adaptive optimal control based on the parameter estimation error for the system (4). We will introduce a new adaptive law driven by the estimation error to estimate unknown parameters and an optimal control item for minimizing performance function. The control structure is shown in Figure 1.

3.1. Filter Design

The traditional adaptive control derives the adaptive law through the gradient algorithm and least square method to minimize tracking error, but the robustness is poor. Based on the previous research, an auxiliary filter is used to extract the parameter estimation error and drive parameter updating to obtain a better-estimated response [36].
To facilitate the design of the adaptive law, the first equation in (4) is considered:
x ¨ 1 = Θ T Ψ x 1 + F x 2
where Θ is the unknown parameter vector, Ψ x 1 is a known nonlinear function matrix, and F x 2 is a known nonlinear. The filtered variables χ f and Ψ f , F f of the dynamics x 1 and Ψ , F are defined as [22]
χ ¨ f + χ ˙ f = x ˙ 1 , χ ˙ f 0 = 0 Ψ ˙ f + Ψ f = Ψ , Ψ f 0 = 0 F ˙ f + F f = F , F f 0 = 0
where is a filter constant. Then it is obtained from (6) and (7) that
χ ¨ f t = x ˙ 1 ( t ) χ ˙ f ( t )
and
x ˙ 1 ( t ) χ ˙ f ( t ) F f = Θ T Ψ x 1
The matrix G and vector Q are defined as
G ˙ = l G + Ψ f Ψ f T , G 0 = 0 Q ˙ = l Q + Ψ f x ˙ 1 ( t ) x ˙ 1 f ( t ) F f , Q 0 = 0
where l > 0 is a design parameter. The solution of (9) is derived as
G = 0 t e l t r Ψ f r Ψ f T r d r Q = 0 t e l t r Ψ f r x ˙ 1 ( t ) x ˙ 1 f ( t ) F f d r
Then, another auxiliary variable H is defined as
H = G Θ ^ T Q
where the estimation of the unknown parameter Θ is Θ ^ . Substituting (8) into (10), one has
H = G Θ ^ T G Θ T = G Θ ˜ T
where Θ ˜ = Θ Θ ^ is defined as the estimation error.

3.2. Adaptive Control with Finite-Time Parameter Estimation

For the position tracking control, we define filter tracking error variables as
e 2 = e ˙ 1 + k 1 e 1
r 1 = e ˙ 2 + k 2 e 2
where k 1 and k 2 are positive constants, and r 1 is the auxiliary filter error.
From (4), (5), (6), (13), (14), one has
r 1 = Θ T Ψ x 1 + K 1 K 2 x 3 + Δ 1 ( t ) x ¨ 1 d + k 1 e ˙ 1 + k 2 e 2
x 3 will be regarded as virtual input, the design virtual control rate is α 2 , and we define the input error e 3 as
e 3 = x 3 α 2
In order to reduce the impact of noise in measurement, the expected trajectory x 1 d is used to replace x 1 , so r 1 can be rewritten by using the expected trajectory as
r 1 = K 1 K 2 α 2 + Θ T Ψ x 1 d + S 1 + Δ 1 t x ¨ 1 d
where the auxiliary function S 1 is
S 1 = k 1 e ˙ 1 + k 2 e 2 + Θ T Ψ x 1 Θ T Ψ x 1 d + K 1 K 2 e 3
The design virtual control rate is
α 2 = 1 K 1 K 2 Θ ^ T Ψ x 1 d k 1 e 2 k 1 e 1 k 2 e 2 + x ¨ 1 d + μ 1
where μ 1 is the feedback item of RISE and
μ 1 = k s 1 e 2 ( t ) + k s 1 e 2 ( 0 ) 0 t k s 1 k 2 e 2 ( σ ) + β 1 sgn e 2 ( σ ) d σ
The time derivative of μ 1 is
μ ˙ 1 = k s 1 r 1 β 1 e 2
where k s 1 is a positive feedback gain, β 1 is a positive constant, and sgn is a function of signum. The extracted estimation error is added to the traditional adaptive law. This method makes the error of parameter estimation and system control converge eventually.
The parameter-updating law of Θ is
Θ ^ ˙ = Γ Ψ ˙ d r 1 M
with
M = σ H + σ G T H H
where Ψ d is the regressor and Γ > 0 , σ > 0 are the learning gains. Since r 1 is not to directly implement online learning, inspired by [37], we write (22) in the following form:
Θ ^ ( t ) = Θ ^ ( 0 ) + σ 0 t H ( τ ) + G T ( τ ) H ( τ ) H ( τ ) d τ + Γ Ψ ˙ d ( τ ) e 2 ( τ ) 0 t Γ 0 t Ψ ¨ d e 2 ( τ ) k 2 Ψ ˙ d e 2 ( τ ) d τ
As follows from (24), the variables r 1  can be replaced by e ˙ 2 + k 2 e 2 , where the Newton–Leibniz formula is used to calculate e ˙ 2 . Therefore, the parameters Θ can be calculated according to (24). The auxiliary filtering signal is designed as
r 2 = e ˙ 3 + k 3 e 3
The error variable r 2 combined with (4), (16), and (25) can be rewritten by using the desired trajectory x 2 d and x 3 d as
r 2 = K 4 x 2 d K 5 x 3 d + K 3 u + S 2 + Δ 2 t α ˙ 2
where the auxiliary function S 2 is
S 2 = K 4 x 2 x 2 d K 5 x 3 x 3 d + k 3 e 3
The brake control effect is also affected by brake performance. Thus, torque and energy consumption need to be considered in the brake control system, to realize efficient and stable control of the aircraft brake. According to (26), the controller is
u = 1 K 3 K 4 x 2 d + K 5 x 3 d + α ˙ 2 + μ 2 + u o
where u o is the auxiliary optimal controller for the given secondary performance index of the system and μ 2 is the feedback item of RISE with
μ 2 = k s 2 e 3 t + k s 2 e 3 0 0 t k s 2 k 3 e 3 σ + β 2 s g n e 3 σ d σ
The time derivative of μ 2 is
μ ˙ 2 = k s 2 r 2 β 2 s g n e 3
where k s 2 , β 2 are normal numbers.

3.3. Optimal Control Auxiliary Item Design

Considering the optimal synthesis of control energy consumption, dynamic tracking error, and filtering error, we design the performance index functions for the following dynamic error closed-loop system of the system, which can be obtained from (4) and (16) such that
e ˙ 3 = x ˙ 3 α ˙ 2 = K 3 u K 4 x 2 K 5 x 3 + Δ 2 t α ˙ 2 = K 3 u K 4 x 2 K 5 ( e 3 + α 2 ) + Δ 2 t α ˙ 2
By substituting controller (28) into (31), the dynamic equation of filtering error is
e ˙ 3 = K 5 e 3 + u o + Δ 3
where Δ 3 denotes the lumped disturbances and compensation error. The cost function for the system (32) is designed as
J e 3 = min u o t e 3 T Q e 3 + u o T R u o d τ
According to Bellman’s optimality principle [38], the solution of the optimal cost function is the solution of the HJB equation:
min u o e 3 T Q e 3 + u o T R u o + J ( e 3 ) ( K 5 e 3 + u o ) = 0
And the optimal control solution can be obtained as
u o = 1 2 R 1 J ( e 3 )
where  J ( x ) = J ( x ) / x , and it is a necessary condition for solving the optimal control, which is approximated by NN:
J ( x ) = W T ϖ ( e 3 ) + ε r
where W T  and ϖ ( x ) are the optimal weight and activation function of the evaluation NN, and ε r ( x ) represents the approximate error that converges to zero after proper training under continuous excitation. In practice, the ideal NN weights W are unknown, combined with (36) and (37); thus, we have the actual optimal control as
u ^ o = 1 2 R 1 W ^ T ϖ ( x ) = Λ W ^ T ϖ ( x )
In order to make the weight estimation W ^ of the NN converge quickly to the ideal value, the control mode of parameter updating is driven by the extraction estimation error in (13)–(21). This mode avoids using an executive NN, which significantly reduces the amount of training computation, speeds up the weight convergence, and improves the estimation accuracy. In order to determine the critical NN weight, the HJB equation can be expressed as
e 3 T Q e 3 + u o T R u o + W ^ T ϖ ( x ) [ K 5 e 3 + u o ] + ε H J B = 0
where ε H J B is a bounded error caused by the approximation error of critic NN, model linearization error, and system uncertainty. Suppose ρ = σ ( x ) [ K 5 e 3 + u o ] ; we obtain
η = W ^ T ρ ε H J B
where η = e 3 T Q e 3 + u o T R u o is a utility function. We define regression matrix A and regression vector B. One has
A ˙ = l A + p p T , A ( 0 ) = 0 B ˙ = l B + p η , B ( 0 ) = 0
where the regression parameter l is the normal number. The auxiliary vector Ξ based on A and B is introduced as
Ξ = A W ^ c + B
According to (19)–(21), we obtain
Ξ = A W ^ A W + v = A W ˜ + v
where W ˜ = W W ^ is the estimation error. We have v = 0 t e l ( t τ ) ε H J B ( τ ) p T ( τ ) d τ for υ ε υ 2 . W ˜ is the weight error, which is related to the auxiliary vector Ξ . The designed update rule is
W ^ ˙ = Υ ϖ ˙ ( x ) r 2 Ξ
where Υ is learning gain, and the calculation of W ^ ˙ is the same as (24).

4. Stability and Convergence Analysis

To analyze the stability of the closed-loop system, we obtain r ˙ 1 by (28) and (22) as
r ˙ 1 = Θ ˜ T Ψ ˙ d + N ˜ 1 + N d 1 + μ ˙ s 1 e 2
where
N ˜ 1 = Γ Ψ ˙ d Ψ d r 1 + Γ Ψ ˙ d σ H + G T H H + S ˙ 1 + e 2
The auxiliary function N d 1 is
N d 1 = Δ ˙ 1
Let Λ = 1 2 R 1 . Substituting control law (35) and (43) into (26), the closed-loop system can be written as
r 2 = S 2 + Δ 2 t + μ 2 + Λ σ ( x ) W ^ T
The time derivative of r 2 is
r ˙ 2 = Λ W ˜ σ ˙ ( x ) + N ˜ 2 + N d 2 + μ ˙ 2 e 3
where
N ˜ 2 = S ˙ 2 + e 3 Λ Υ σ ( x ) σ ˙ ( x ) r 2 + Λ Υ σ ˙ ( x ) A W ˜ + v
The auxiliary function N d 2 is
N d 2 = Δ ˙ 2 + ε r
Lemma 1
([2]). Define the following two auxiliary functions for the convenience of analysis:
L 1 t = r 1 Δ ˙ 1 β 1 s g n e 2
L 2 t = r 2 Δ ˙ 2 β 2 s g n e 3
where β 1 , β 2 meet the following conditions:
β 1 ζ 1 + 1 k 2 ζ 2
β 2 ζ 3 + 1 k 3 ζ 4
where ζ 1 , ζ 2 , ζ 3 , and ζ 4 are positive constants, and the following defined functions P ˙ 1 t , P ˙ 2 t are always positive:
P ˙ 1 t = L 1
P ˙ 2 t = L 2
Select a Lyapunov function candidate as
V L = 1 2 e 1 2 + 1 2 e 2 2 + 1 2 e 3 2 + 1 2 r 1 2 + 1 2 r 2 2 + P 1 + P 2 + 1 2 θ ˜ Γ 1 θ ˜ T + 1 2 W ˜ T K W W ˜
Let us define a vector
ξ t = z T t , P 1 t , P 2 t , θ ˜ T t , W ˜ T t
where
z T t = e 1 t , e 2 t , e 3 t , r 1 t , r 2 t
The Lyapunov function satisfies the inequality
U 1 ξ V L ξ U 2 ξ
The continuous, positive definite functions U 1 ξ and U 2 ξ are defined as
U 1 ξ = λ 1 ξ , U 2 ξ = λ 2 ξ
where λ 1 λ 2 are positive constants. We substitute (5), (13), (16), (22,) (43), (44), (48), (55), and (56) into the derivative of V L . It follows that
V ˙ L = e 1 e ˙ 1 + e 2 e ˙ 2 + e 3 e ˙ 3 + r 1 r ˙ 1 + r 2 r ˙ 2 + L ˙ 1 + L ˙ 2 + θ ˜ Γ 1 θ ˜ ˙ + Λ W ˜ Υ 1 W ˜ ˙ = k 1 e 1 2 k 2 e 2 2 k 3 e 3 2 k s 1 r 1 2 k s 2 r 2 2 + e 1 e 2 + r 1 N ˜ 1 + r 2 N ˜ 2 θ ˜ ψ ˙ d 1 r 1 σ H 1 σ P 1 T H 1 H 1 W ˜ Λ A W ˜ + W ˜ Λ v
Using the Young’s inequality, one has
e 1 e 2 1 2 e 1 2 + 1 2 e 2 2
r 1 N ˜ 1 ρ z z r 1 k s 1 r 1 2 + ρ 2 z z 2 4 k s 1
r 2 N ˜ 2 ρ z z r 2 k s 2 r 2 2 + ρ 2 z z 2 4 k s 2
where ρ is the function of z , and it may have a constant term, a first term, a second term, or a higher term.
W c T A W c + W c T v Υ W c 2
It is shown that the proposed NN weight estimation method can make the error converge to zero exponentially, and the convergence is determined by continuous excitation and learning rate. Substitute (64), (65), (66) into (63) to obtain
V ˙ L λ 3 z 2 + ρ 2 z z 2 4 k s 1 + ρ 2 z z 2 4 k s 2 U y
where λ 3 = min k 1 1 2 , k 2 1 2 , k 3 , k s 1 , k s 2 , Υ is a positive constant, and U ( y ) = c z 2 is a continuous positive function over a set defined as
Ξ y t | y ρ 1 2 λ 3 k s
Therefore, inequalities (67) and (68) can be used to prove V L L above Ξ, which further implies that e 1 , e 2 , e 3 P 1 , P 2 , r 1 , r 2 L in Ξ . e ˙ 1 , e ˙ 2 , e ˙ 3 L in Ξ can be proven through z T t L . Meanwhile, the fact that x ˙ d and x ¨ d are bounded implies that x 1 , x ˙ 1 L , given that r 1 , r 2 L in Ξ . A similar analysis to that given in [39] can prove that μ ˙ 1 , μ ˙ 2 L  in Ξ . Then, we can also verify that r ˙ 1 , r ˙ 2 L . The U ξ is uniformly continuous in Ξ due to e ˙ 1 , e ˙ 2 , e ˙ 3 , r ˙ 1 , r ˙ 2 L in Ξ. Hence, we can verify that U ξ 0 by an analysis similar to that shown in [36]. Then, from the definition of r, z 1 , z 2 , we can obtain that e 1 0 for t , and the proof of convergence of parameter estimation error can be seen [1].

5. Simulations

In order to verify the effectiveness of the proposed control law, simulation tests are carried out on the MATLAB platform with the model parameters in [35]. The reference slip rate is set as 0.1, the initial landing velocity v x = 87 m/s, and the initial main wheel angular velocity w ( 0 ) = 210  rad/s, and the initial slip rate is set as λ 0 = 0 .
The parameters in the controller are designed to k 1 = 3 , k 2 = 25 , k s 1 = 0.7 , k s 2 = 10 , Θ = 1, β 1 = 4.5 , β 2 = 8 , and the critic NN weights converge to [−1.982, −1.553, −4.006, −3.86, −1.210, −0.567, 0.72, −0.9234, −0.34, 0.765, 1.076, 0.365, 3.023].
The changes in the speed and distance of the UAV during braking are shown in Figure 2 and Figure 3. It takes 21.6 s from the start of braking to the end of the anti-skid control, and the braking distance is 702.5 m. No significant overshoot occurs during the braking process, and the velocity and wheel speed decrease smoothly without large fluctuations or deep skid of the wheel.
It can be seen from Figure 4 that the controller can track the optimal slip rate of real-time changes quickly and accurately and is always in the stable region. Figure 5 shows that the tracking error of the optimized controller based on RISE correction converges asymptotically to zero, meeting the high-precision requirements compared to the generic control with the gradient algorithm.
To show the improved estimation and control performance, the method proposed in this paper is shown in Figure 6. Compared with the conventional control algorithm shown in Figure 7, the parameter estimation error can converge to zero and the convergence speed is fast.
The quadratic performance index function of the algorithm is designed in this paper and the conventional algorithm is estimated by MATLAB. The performance index function of the optimized controller is larger than that of the conventional controller, which indicates that the RISE-modified adaptive optimal controller ensures the comprehensive quasi-optimal of the system dynamic error and control energy consumption during the whole braking process. The control input is shown in Figure 8.
Figure 9 shows the estimated weight of the NN. It can be seen that the weights converge to constant values. This result shows the convergence of the adaptive law and verifies the effectiveness of the proposed learning algorithm.
To show the improved performance of the proposed single critic NN-based ADP, compared with a critic–actor NN-based ADP method mentioned in [31], it is clear that the proposed single critic NN-based can achieve faster transient error convergence and smaller steady-state tracking errors.
The simulation results show that the designed controller has a good control effect, realizes the tracking of the unknown optimal slip rate, and has a robust ability. By comparing the size of the secondary performance index, the modified optimized controller can meet the performance index, and the system’s control input and control error are optimized comprehensively.

6. Conclusions

This paper investigated unknown parameter estimation for the UAV brake system and tracked the optimal slip rate at minimum cost. Parameter estimation error is extracted by an auxiliary filter and drives a parameter-updating law incorporated into adaptive control synthesis. Compared with traditional estimation methods, this method can simultaneously guarantee the convergence of parameter estimation and track errors. A performance index function of control input and filtering error was designed, considering the balance of control resources and control effect. This paper made up for the lack of consideration of resource control in the previous literature on UAV brake systems. Based on the ADP idea, an NN was introduced to approximate the performance index function online, and an adaptive law was designed to update the NN weights and optimal control online. Simulation results prove that the proposed single critic NN-based method can achieve faster transient error convergence and smaller steady-state tracking errors. Through theoretical and simulation analysis, the modified controller can converge to the optimal controller, compensating for external perturbations, modeling uncertainties, and NN estimation errors by combining the principles of adaptive control and optimal control with the RISE robust term.

Author Contributions

Conceptualization, X.G. and H.L.; methodology, Q.L. and H.L.; software, Y.W.; validation, Y.W., F.L. and Z.L.; formal analysis, Q.L.; invetigation, H.L.; data curation, F.L.; writing—original draft preparation, H.L.; writing—review and editing, X.G.; visualization, Y.W.; funding acquisition, X.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request. The data are not publicly available due to privacy or ethical restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hoseinnezhad, R.; Bab-Hadiashar, A. Efficient Antilock Braking by Direct Maximization of Tire–Road Frictions. IEEE Trans. Ind. Electron. 2011, 58, 3593–3600. [Google Scholar] [CrossRef]
  2. Peng, Z.; Jiang, Y.; Wang, J. Event-Triggered Dynamic Surface Control of an Underactuated Autonomous Surface Vehicle for Target Enclosing. IEEE Trans. Ind. Electron. 2021, 68, 3402–3412. [Google Scholar] [CrossRef]
  3. Hu, J.; Zheng, M. Disturbance observer-based control with adaptive neural network for unknown nonlinear system. Meas. Control 2022, 56, 287–294. [Google Scholar] [CrossRef]
  4. Dinçmen, E.; Güvenç, B.A.; Acarman, T. Extremum-Seeking Control of ABS Braking in Road Vehicles with Lateral Force Improvement. IEEE Trans. Control Syst. Technol. 2014, 22, 230–237. [Google Scholar] [CrossRef]
  5. Capra, D.; Galvagno, E.; Ondrak, V.; van Leeuwen, B.; Vigliani, A. An ABS control logic based on wheel force measurement. Veh. Syst. Dyn. 2012, 50, 1779–1796. [Google Scholar] [CrossRef]
  6. Pasillas-Lépine, W. Hybrid modeling and limit cycle analysis for a class of five-phase anti-lock brake algorithms. Veh. Syst. Dyn. 2006, 44, 173–188. [Google Scholar] [CrossRef]
  7. Zhang, Y.X.; Lin, C.P. Backstepping Adaptive Neural Network Control for Electric Braking Systems of Aircrafts. Algorithms 2019, 12, 215. [Google Scholar] [CrossRef]
  8. Chen, H.; Zhang, X.; Xu, Y. Modeling, Simulation, and Experiment of Switched Reluctance Ocean Current Generator System. Adv. Mech. Eng. 2013, 5, 261241. [Google Scholar] [CrossRef]
  9. Yi, J.; Alvarez, L.; Horowitz, R. Adaptive emergency braking control with underestimation of friction coefficient. IEEE Trans. Control. Syst. Technol. 2002, 10, 381–392. [Google Scholar] [CrossRef]
  10. Braghin, F.; Cheli, F.; Sabbioni, E. Environmental effects on Pacejka’s scaling factors. Veh. Syst. Dyn. 2006, 44, 547–568. [Google Scholar] [CrossRef]
  11. Jiao, Z.; Liu, X.; Shang, Y.; Huang, C. An integrated self-energized brake system for aircrafts based on a switching valve control. Aerosp. Sci. Technol. 2017, 60, 20–30. [Google Scholar] [CrossRef]
  12. Xu, B.; Yang, D.; Shi, Z.; Pan, Y.; Chen, B.; Sun, F. Online Recorded Data-Based Composite Neural Control of Strict-Feedback Systems with Application to Hypersonic Flight Dynamics. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 3839–3849. [Google Scholar] [CrossRef]
  13. Gao, H.; He, W.; Zhou, C.; Sun, C. Neural Network Control of a Two-Link Flexible Robotic Manipulator Using Assumed Mode Method. IEEE Trans. Ind. Inform. 2019, 15, 755–765. [Google Scholar] [CrossRef]
  14. Wang, S.; Yu, H.; Yu, J.; Na, J.; Ren, X. Neural-Network-Based Adaptive Funnel Control for Servo Mechanisms with Unknown Dead-Zone. IEEE Trans. Cybern. 2020, 50, 1383–1394. [Google Scholar] [CrossRef]
  15. Liu, Y.J.; Tong, S.; Chen, C.L.P. Adaptive Fuzzy Control via Observer Design for Uncertain Nonlinear Systems with Unmodeled Dynamics. IEEE Trans. Fuzzy Syst. 2013, 21, 275–288. [Google Scholar] [CrossRef]
  16. Tong, S.; Zhang, L.; Li, Y. Observed-Based Adaptive Fuzzy Decentralized Tracking Control for Switched Uncertain Nonlinear Large-Scale Systems With Dead Zones. IEEE Trans. Syst. Man Cybern. Syst. 2016, 46, 37–47. [Google Scholar] [CrossRef]
  17. Li, Z.; Xiao, S.; Ge, S.S.; Su, H. Constrained Multilegged Robot System Modeling and Fuzzy Control with Uncertain Kinematics and Dynamics Incorporating Foot Force Optimization. IEEE Trans. Syst. Man Cybern. Syst. 2016, 46, 1–15. [Google Scholar] [CrossRef]
  18. Xia, D.; Wang, L.; Chai, T. Neural-Network-Friction Compensation-Based Energy Swing-Up Control of Pendubot. IEEE Trans. Ind. Electron. 2014, 61, 1411–1423. [Google Scholar] [CrossRef]
  19. Sastry, S.; Bodson, M. Adaptive Control: Stability, Convergence and Robustness. J. Acoust. Soc. Am. 1990, 88, 588–589. [Google Scholar] [CrossRef]
  20. Narendra, K.S.; Annaswamy, A.M.J.P.H. Stable Adaptive Systems; Prentice Hall: Hoboken, NJ, USA, 1989. [Google Scholar]
  21. Ioannou, P.; Sun, J. Robust Adaptive Control; PTR Prentice-Hall: Upper Saddle River, NJ, USA, 1995. [Google Scholar]
  22. Na, J.; Herrmann, G.; Ren, X.; Mahyuddin, M.N.; Barber, P. Robust adaptive finite-time parameter estimation and control of nonlinear systems. In Proceedings of the 2011 IEEE International Symposium on Intelligent Control, Denver, CO, USA, 28–30 September 2011; pp. 1014–1019. [Google Scholar] [CrossRef]
  23. Yang, X.; Liu, D.; Wang, D. Reinforcement learning for adaptive optimal control of unknown continuous-time nonlinear systems with input constraints. Int. J. Control 2013, 87, 553–566. [Google Scholar] [CrossRef]
  24. Liu, D.; Wang, D.; Wang, F.Y.; Li, H.; Yang, X. Neural-Network-Based Online HJB Solution for Optimal Robust Guaranteed Cost Control of Continuous-Time Uncertain Nonlinear Systems. IEEE Trans. Cybern. 2014, 44, 2834–2847. [Google Scholar] [CrossRef]
  25. Peng, Z.; Luo, R.; Hu, J.; Shi, K.; Nguang, S.K.; Ghosh, B.K. Optimal Tracking Control of Nonlinear Multiagent Systems Using Internal Reinforce Q-Learning. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 4043–4055. [Google Scholar] [CrossRef] [PubMed]
  26. Peng, Z.; Zhao, Y.; Hu, J.; Luo, R.; Ghosh, B.; Nguang, S. Input-Output Data-Based Output Antisynchronization Control of Multi-Agent Systems Using Reinforcement Learning Approach. IEEE Trans. Ind. Inform. 2021, 17, 7359–7367. [Google Scholar] [CrossRef]
  27. Ding, L.; Li, S.; Gao, H.; Liu, Y.J.; Huang, L.; Deng, Z. Adaptive Neural Network-Based Finite-Time Online Optimal Tracking Control of the Nonlinear System With Dead Zone. IEEE Trans. Cybern. 2021, 51, 382–392. [Google Scholar] [CrossRef] [PubMed]
  28. Allwright, J. A lower bound for the solution of the algebraic Riccati equation of optimal control and a geometric convergence rate for the Kleinman algorithm. IEEE Trans. Autom. Control 1980, 25, 826–829. [Google Scholar] [CrossRef]
  29. Jiang, H.; Zhang, H.; Xie, X.; Han, J. Neural-network-based learning algorithms for cooperative games of discrete-time multi-player systems with control constraints via adaptive dynamic programming. Neurocomputing 2019, 344, 13–19. [Google Scholar] [CrossRef]
  30. Lewis, F.L.; Vrabie, D. Reinforcement learning and adaptive dynamic programming for feedback control. IEEE Circuits Syst. Mag. 2009, 9, 32–50. [Google Scholar] [CrossRef]
  31. Na, J.; Herrmann, G. Online Adaptive Approximate Optimal Tracking Control with Simplified Dual Approximation Structure for Continuous-time Unknown Nonlinear Systems. IEEE/CAA J. Autom. Sin. 2014, 1, 412–422. [Google Scholar]
  32. Dong, H.; Zhao, X.; Luo, B. Optimal Tracking Control for Uncertain Nonlinear Systems with Prescribed Performance via Critic-Only ADP. IEEE Trans. Syst. Man Cybern. Syst. 2022, 52, 561–573. [Google Scholar] [CrossRef]
  33. Xian, B.; Dawson, D.M.; De Queiroz, M.S.; Chen, J. A continuous asymptotic tracking control strategy for uncertain multi-input nonlinear systems. In Proceedings of the 2003 IEEE International Symposium on Intelligent Control, Houston, TX, USA, 5–8 October 2003; pp. 52–57. [Google Scholar] [CrossRef]
  34. Shuzhi, S.G.; Hang, C.C.; Woon, L.C. Adaptive neural network control of robot manipulators in task space. IEEE Trans. Ind. Electron. 1997, 44, 746–752. [Google Scholar] [CrossRef]
  35. Chen, X.; Dai, Z.; Lin, H.; Qiu, Y.; Liang, X. Asymmetric Barrier Lyapunov Function-Based Wheel Slip Control for Antilock Braking System. Int. J. Aerosp. Eng. 2015, 2015, 917807. [Google Scholar] [CrossRef]
  36. Wang, S.; Na, J. Parameter Estimation and Adaptive Control for Servo Mechanisms With Friction Compensation. IEEE Trans. Ind. Inform. 2020, 16, 6816–6825. [Google Scholar] [CrossRef]
  37. Sharma, N.; Bhasin, S.; Wang, Q.; Dixon, W.E. RISE-Based Adaptive Control of a Control Affine Uncertain Nonlinear System With Unknown State Delays. IEEE Trans. Autom. Control 2012, 57, 255–259. [Google Scholar] [CrossRef]
  38. Bellman, R. On the Theory of Dynamic Programming. Proc. Natl. Acad. Sci. USA 1952, 38 8, 716–719. [Google Scholar] [CrossRef]
  39. Makkar, C.; Hu, G.; Sawyer, W.G.; Dixon, W.E. Lyapunov-Based Tracking Control in the Presence of Uncertain Nonlinear Parameterizable Friction. IEEE Trans. Autom. Control 2007, 52, 1988–1994. [Google Scholar] [CrossRef]
Figure 1. Structure of the control system.
Figure 1. Structure of the control system.
Applsci 13 11888 g001
Figure 2. Aircraft speed and wheel speed.
Figure 2. Aircraft speed and wheel speed.
Applsci 13 11888 g002
Figure 3. Stopping distance.
Figure 3. Stopping distance.
Applsci 13 11888 g003
Figure 4. Wheel slip ratio.
Figure 4. Wheel slip ratio.
Applsci 13 11888 g004
Figure 5. The tracking error.
Figure 5. The tracking error.
Applsci 13 11888 g005
Figure 6. Textual method.
Figure 6. Textual method.
Applsci 13 11888 g006
Figure 7. Gradient descent.
Figure 7. Gradient descent.
Applsci 13 11888 g007
Figure 8. Gradient descent.
Figure 8. Gradient descent.
Applsci 13 11888 g008
Figure 9. The control input.
Figure 9. The control input.
Applsci 13 11888 g009
Table 1. Parameters of UAV and EMA.
Table 1. Parameters of UAV and EMA.
NameDescriptionValue
m A Weight of the UAV5000 kg
v x Longitudinal velocity-
ω m Wheel speed-
F X Aerodynamic drag-
nNumber of brake wheels2
F f The binding force-
c b Brake disc stiffness coefficient 5.0 × 10 6 N/m
L 0 Ball screw travel0.008 m
F z Main wheel support force-
μ Coefficient of static friction-
K T Motor torque constant0.4 Nm/A
B v Viscous damping coefficient0.02 Nm/rad
J w Moment of wheel0.95 kg · m 2
REffective radius of main wheel0.3 m
JMotor moment of inertia 6.38 × 10 4 kg · m 2
K b Braking torque and pressure conversion coefficient 3.7 × 10 4
B w Friction coefficient of axle0.01 Nm/rad
P A The pressure acting on the brake disc-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lv, H.; Liu, Q.; Wan, Y.; Gao, X.; Li, F.; Liu, Z. Adaptive Optimal Control of UAV Brake Based on Parameter Estimation. Appl. Sci. 2023, 13, 11888. https://doi.org/10.3390/app132111888

AMA Style

Lv H, Liu Q, Wan Y, Gao X, Li F, Liu Z. Adaptive Optimal Control of UAV Brake Based on Parameter Estimation. Applied Sciences. 2023; 13(21):11888. https://doi.org/10.3390/app132111888

Chicago/Turabian Style

Lv, Hao, Qinrui Liu, Yu Wan, Xuehui Gao, Feng Li, and Zhen Liu. 2023. "Adaptive Optimal Control of UAV Brake Based on Parameter Estimation" Applied Sciences 13, no. 21: 11888. https://doi.org/10.3390/app132111888

APA Style

Lv, H., Liu, Q., Wan, Y., Gao, X., Li, F., & Liu, Z. (2023). Adaptive Optimal Control of UAV Brake Based on Parameter Estimation. Applied Sciences, 13(21), 11888. https://doi.org/10.3390/app132111888

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop