Abstract
In this paper, we develop momentum-based adaptive update laws for parameter identification and control to improve parameter estimation error convergence and control system performance for uncertain dynamical systems. Specifically, we introduce three novel continuous-time, momentum-based adaptive estimation and control algorithms and evaluate their effectiveness via several numerical examples. Our proposed adaptive architectures show faster parameter convergence rates as compared to the classical gradient descent and model reference adaptive control methods.
1. Introduction
One of the fundamental problems in feedback control design is the ability to address discrepancies between system models and real-world systems. To this end, adaptive control has been developed to address the problem of system uncertainty in control-system design [1,2,3,4]. In particular, adaptive control is based on constant linearly parameterized system uncertainty models of a known structure but unknown variation and, in the case of indirect adaptive control, combines an online parameter estimation algorithm with a control law to improve performance in the face of system uncertainties. More specifically, indirect adaptive controllers utilize parameter update laws to identify unknown system parameters and adjust feedback gains to account for system uncertainty.
The parameter estimation algorithms that have been typically used in adaptive control are predicated on two gradient descent type methods that update the system parameter errors in the direction that minimizes a prediction error for a specified cost function. Namely, for a gradient descent flow algorithm, the cost function involves an instantaneous prediction error between an estimated system output and the actual system output of the most recent data, whereas for an integral gradient descent algorithm the cost function captures an aggregate of the past prediction errors in an integral form while placing more emphasis on recent measurements [2].
Additionally, the recursive least squares (RLS) algorithm [2] has also been used for parameter estimation and has been shown to give superior convergence rates as compared to the classical gradient descent algorithms. RLS algorithms can be viewed as a gradient descent method with a time-varying learning rate [5]. For static (i.e., time-invariant) cost functions, momentum-based methods, also known as higher-order tuners, such as the Nesterov method, have been shown to achieve faster parameter error convergence as compared to traditional gradient descent algorithms [6]. Since faster convergence can lead to improved transient system performance, there has been a significant interest in extending momentum-based methods to time-varying cost functions, which naturally appear in adaptive control formulations [7,8,9].
Specifically, in [7] momentum-based architectures were merged with the standard gradient descent method using a variational approach to guarantee boundedness for the parameter estimation and model reference adaptive control problems involving time-varying regressors. Higher-order tuners were also shown to provide an additional degree of freedom in the design of adaptive control laws [10,11]. Given that integral gradient and recursive least squares algorithms provide superior noise rejection properties and improved system performance to that of standard gradient descent methods, there has been a recent interest in integrating momentum-based architectures into adaptive laws for parameter identification and control [5,7].
In this paper, we develop new continuous-time, momentum-based adaptive laws for identification and control by augmenting higher-order tuners into the integral gradient, recursive least squares and composite gradient algorithms. Specifically, in Section 2, we first review the existing gradient-based and momentum-based update laws and introduce three new higher-order tuner architectures. Next, in Section 3, we show how these update laws can be applied for parameter identification, and then, in Section 4, we introduce a momentum-based recursive least squares (MRLS) algorithm for model reference adaptive control. Finally, Section 5 provides several numerical examples that highlight the improved parameter error convergence rate of the proposed momentum-based update laws.
The notation used in this paper is standard. Specifically, we write to denote the set of column vectors, to denote the set of real matrices, and to denote transpose. The equi-induced 2-norm of a matrix is denoted as where denotes the maximum eigenvalue and denotes the maximum singular value. We write (resp., ) for the minimum eigenvalue (resp., singular value), for the gradient of a scalar-valued function V with respect to a vector-valued variable x, and for the Laplace transform of , that is . Finally, we write to denote the proper rational functions with coefficients in (i.e., SISO proper rational transfer functions), to denote polynomials with real coefficients, to denote the space of square-integrable Lebesgue measurable functions on , and to denote the space of bounded Lebesgue measurable functions on .
2. From First-Order to Higher-Order Tuners for Parameter Estimation
Consider the system given by
where, for , is the system output, is a time-varying regressor vector, and is an unknown system parameter vector. Here we assume that is bounded, and hence, can be normalized via the scaling factor so that, with a minor abuse of notation, . Since is unknown, we design an estimator of the form , where , is an estimate of , so that the prediction error between the estimated output , and the actual system output , is given by
where is the parameter estimation error.
For the error system (2), consider the quadratic loss cost function
and note that the gradient of L with respect to the parameter estimate is given by . In this case, the gradient flow algorithm is expressed as
where is a weighting gain parameter. As shown in [7], for all .
It is well known that the rate of convergence for the gradient flow algorithm (4) can be slow [6]. However, if the regressor vector satisfies a persistency of the excitation condition, then converges to exponentially with a convergence rate proportional to . Although in this case the convergence rate can improve significantly, large values of can result in a stiff initial value problem for (4) necessitating progressively smaller sampling times for implementing (4) on a digital computer. Furthermore, small sampling times can increase the sensitivity of the adaptive law to measurement noise and modeling errors [2]. To address these limitations, momentum-based accelerated algorithms predicated on higher-order tuner laws are introduced in [7] to expedite parameter estimation error convergence.
In [12], an array of accelerated methods for continuous-time systems are derived using a Lagrangian variational formulation. Building on this work, [7] extends the variational framework of [12] to address a time-varying instantaneous quadratic loss function of the form given by (3). Specifically, a Lagrangian of the form
is introduced, where , is a friction coefficient, and , and the functional
where is a finite-time interval, is considered. A necessary condition for minimizing (6) is shown in [7] to satisfy
Note that (7) can be rewritten as
where is an auxiliary parameter and . The stability properties of (8) and (9) are given in [7]. Taking the limiting case as , the gradient flow algorithm (4) can be recovered as a special case of (8) and (9).
Rather than relying solely on the instantaneous system measurement , the loss cost function (3) can be modified to incorporate a weighted sum of past system measurements. Specifically, we can consider loss cost functions in the form of (see [2])
where is a forgetting factor that prevents the degeneration of system information updates in some direction by placing more emphasis on the more recent measurements. In this case, the gradient of the loss cost function (10) is given by
where
Now, setting
we obtain the integral gradient algorithm ([2])
The stability properties of (15)–(17) are addressed in [2] ([Thm. 3.6.7]).
Next, motivated by the variational formulation of [7] we introduce a momentum-based integral gradient algorithm. Specifically, defining the Lagrangian
where , and using the Euler-Lagrange equation, a necessary condition for minimizing (6) yields
where, for , and are given by (12) and (13). Note that (19) can be rewritten as
where is an auxiliary parameter and . Furthermore, (20) and (21) can be rewritten in terms of the parameter estimation error , and the auxiliary parameter estimation error as
The following definition and lemmas are needed for the main results of this section.
Definition 1.
A vector signal is persistently excited (PE) if there exist a positive constant ρ and a finite-time such that
Lemma 1.
Consider the system (22). Then, the following statements hold:
If is persistently excited, then
Proof.
Lemma 2
([3]). Let . If and , then .
Theorem 1.
Let , and consider the momentum-based integral gradient algorithm (20)–(23). Then, the following statements hold:
, , and .
If is persistently excited, then converges to exponentially as with a convergence rate of , where .
Proof.
To show , consider the positive-definite function and note that
which, since , implies . Thus, and .
Remark 1.
Note that unlike gradient flow and recursive least squares algorithms, the momentum-based integral gradient algorithm (20)–(23) does not require the assumption that in order to guarantee that [7].
Remark 2.
Note that a time-varying forgetting factor can also be employed in (22) and (23). In this case, the PE condition can be replaced with a less restrictive excitation condition where (26) holds for a fixed-time t and not for all [13].
Next, we consider the cost function
where for all is an nonnegative-definite matrix function called the general forgetting matrix and is an positive-definite matrix. It can be shown that a necessary condition for minimizing (33) gives the recursive least squares algorithm ([14])
Note that setting and , where is positive-definite for all (see [2]), we recover, respectively, the pure recursive least squares and the recursive least squares with exponential forgetting algorithms discussed in [2].
Note that the RLS algorithms (34) and (35) do not involve a gradient flow architecture, and the variational formulation cannot be used to generate a momentum-based recursive least squares (MRLS) algorithm. Despite this fact, higher-order tuner laws can still be incorporated into the RLS architecture to give the MRLS architecture
where and , . Note that (36) and (37) can be rewritten as
Noting that and using the fact that
it follows that
Lemma 3.
where .
Consider the system (42) with . Then, the following statements hold:
If is persistently excited, then
where .
Proof.
Next, to show (, note that if is PE, then
and
Thus,
This proves the lemma. □
Theorem 2.
Let , and consider the momentum-based recursive least squares algorithm (36)–(38). Then, given by (2) satisfies . Furthermore, the following statements hold:
If there exists such that (44) holds, then , . If, in addition, , then and .
If and is persistently excited, then the zero solution to (39) and (40) is exponentially stable.
Proof.
Consider the function and note that
Now, by the Cauchy–Schwarz inequality,
and hence,
Next, integrating over yields
and hence, .
To show (i), note that (49) and (44) imply that and . In addition, (37) implies that , and for , it follows from (2) and Lemma 2 that and .
Finally, to show , note that if is persistently excited, then (44) holds by Lemma 1. Now, using the fact that
where is a positive-definite matrix with eigenvalues , it follows from the Schur decomposition that
and hence,
Remark 3.
The constraints on parameter in Theorems 1 and 2 limit the amount of momentum that can be applied, thereby limiting the potential advantage gained by using a momentum-based algorithm. In the context of static optimization, Lyapunov-like functions are typically constructed as the sum of two terms; one term representing the norm of the error of the parameter estimate and the other term corresponding to the loss cost function [12]. This generalized structure allows for adjustments to the parameter estimate , that may not decrease or but can reduce the loss cost function as shown in Figure 1.
Figure 1.
Visualization of momentum-based learning for an ill-conditioned problem; is updated in a direction that increases while still decreasing the loss cost function.
In order to incorporate (10) in our Lyapunov function, we introduce the momentum-based composite gradient algorithm
where and are given by (12) and (13). Note that this composite update law includes an integral gradient term as well as a recursive least squares term in (55).
Theorem 3.
Let , and consider the momentum-based composite gradient algorithm (55)–(57). Then, the following statements hold:
If there exists such that (44) holds, then , , and .
If and is persistently excited, then the zero solution to (55)–(57) is exponentially stable.
3. System Parameter Identification
In this section, we address the problem of parameter identification using the first- and higher-order tuner algorithms developed in Section 2. First, the following definition is needed.
Definition 2
([2]). A signal is stationary and sufficiently rich of order n if the following statements hold:
The limit
exists uniformly in .
The support of the spectral measure , of u contains at least n points.
Consider the stable single-input, single-output (SISO) plant, with output and input being the only available signals for measurement, given by
where , , and . Since there are an infinite number of triples () yielding the same input–output measurements, we cannot uniquely define the coefficients in (). However, we can determine the parameters corresponding to the coefficients of the stable and strictly proper rational transfer function given by
where, for , is the system output, is the system input, and the coefficients , , and , , where , are unknown system parameters. The goal of the system parameter identification problem is to identify the system parameter vector
containing the unknown coefficients of the plant transfer function that satisfies
where and .
Note that (67) has the form of (1) in the frequency domain; however, in most applications the only signals that are available for measurement are the input , and the output , and not their derivatives. To address this, we filter (67) through a stable filter , where is an arbitrary monic Hurwitz polynomial of degree n, to obtain, for ,
where and . The time signals , and , can be generated by the state equations ([2])
where, for , , , , and
and where .
Theorem 4.
Proof.
The proof follows from Theorems 1, 2, and 3 by noting that if is stationary and sufficiently rich of order and G has no pole-zero cancellations, then is PE [2] ([Thm. 5.2.4]). □
Remark 4.
Note that if , then (68) would involve an additional bias error term. As shown in [2], this term converges to the origin exponentially fast, and hence, Theorem 4 remains valid.
4. Momentum-Based Recursive Least Squares and Model Reference Adaptive Control
Recursive least squares type algorithms were first introduced in adaptive control in [16] and extended in [2,17,18]. In this section, we show how the MRLS algorithm proposed in Section 2 can be extended to the problem of model reference adaptive control for systems with relative degree one.
Consider the linear SISO system given by
where , are unknown polynomials, is an unknown constant, is the control system input, and is the system output. As in [18], we make the following assumptions.
is minimum phase and has relative degree one.
The degree of is n.
The sign of is known.
The control objective is to design an appropriate control law , such that all the signals of the closed-loop system are bounded and , tracks the output , of a reference model given by
where , and are known constants, and is the Laplace transform of a bounded piecewise continuous signal .
To address this problem, consider the filter system ([2])
where, for , , ,
, and , where is an arbitrary monic Hurwitz polynomial of degree . Here, we use , and , to form the regressor vector
Note that the existence of a constant parameter vector such that the transfer function of the SISO system (76) with where , , matching the reference model transfer function is guaranteed by the choice of and Assumptions –.
Next, consider the control law
where , , and , and note that the tracking error satisfies
where . Note that (83) has a nonminimal state-space realization given by
where , , , and . Furthermore, note that
where , is a strictly positive real transfer function. In this case, it follows from the Meyer–Kalman–Popov lemma [2] ([Lem. 3.5.4]) that there exist positive-definite matrices and such that
Since for every nonsingular matrix and constant , the realizations and , , ) are equivalent, choosing and we can ensure that is a realization that satisfies (87) and (88) with and is normalized so that .
Next, consider the recursive least-squares algorithm given by
where and , , and . Note that for , (89) and (90) recover the recursive least-squares algorithm of [19].
Here, we modify the recursive least-squares algorithm (89) and (90) to construct our MRLS algorithm as
where , , , ,
and . Note that (91) and (92) can be rewritten in terms of the error parameters , and as
Theorem 5.
Proof.
Next, using the Cauchy–Schwarz and Young inequalities we obtain
Now, using (98) along with triangle inequality , , it follows from (97) that
which shows that . Hence, , , . Next, integrating over yields
and hence, . Similarly, and, since , .
Finally, if , then satisfies
or
Hence, , and thus, . Now, if , then (85) implies that and, by Lemma 2, . □
Remark 5.
Note that unlike many MRAC schemes (e.g., [20]), (91)–(93) does not necessitate a projection operator to guarantee boundedness of the tracking error. However, unlike [19], wherein only a lower bound is necessary for , we require knowledge of a lower and upper bound for .
Remark 6.
It is important to note that there exists an alternative MRAC framework in the litterature known as the normalized adaptive laws [2] whose design is not based on the tracking error but rather on a normalized estimation error of a particular parametrization of the plant. For this framework, the momentum-based integral gradient algorithm (20)–(23), the momentum-based recursive least squares algorithms (36)–(38), and momentum-based composite gradient algorithm (55)–(57) can be used directly for strictly proper plants without a relative degree restriction. In this case, the parametrization of the ideal controller is given by , where satisfies , and .
5. Illustrative Numerical Examples
5.1. System Parameter Identification
Consider the third-order transfer function representing a servo control system for the pitch control of an aircraft given by
where is the pitch angle, is the system input, and , , , , , and are the unknown system plant parameters. Note that, (68) gives with and .
Let , , , , , , , and choose . Note that is a stationary and sufficiently rich signal of order five, and hence, Theorem 4 guarantees that the estimated system parameters will converge exponentially to as using both the momentum-based integral gradient algorithm (20)–(23) and the MRLS algorithm (36)–(38).
First, we compare the performance of the RLS algorithm (34) and (35) with the MRLS algorithm (20)–(23). For this comparison, we set , , , , , , and . Figure 2 shows the system parameter estimate versus time. It can be seen that for both the RLS and the MRLS algorithms the parameter estimate converges to as expected by Theorem 2. Moreover, as seen in Figure 3 the MRLS algorithm provides faster convergence of the system parameters to the true values as compared to the standard RLS algorithm.
Figure 2.
Parameter estimate values versus time using the RLS and MRLS algorithms. In both cases the parameter estimate values converge to the true values with the MRLS algorithm converging faster.
Figure 3.
Norm of the parameter estimate error versus time using the RLS and MRLS algorithms.
Next, we compare the momentum-based composite gradient algorithm (55)–(57) with the composite gradient algorithm given by
Let , , , , , , and . Figure 4 shows the system parameter estimates versus time. It can be seen that for both the composite gradient and the momentum-based composite gradient algorithms the parameter estimate converges to as expected by Theorem 3. However, as seen in Figure 5, our proposed algorithm provides faster convergence of the system parameters to their true values. In particular, the momentum-based composite gradient settles around a value of at s as compared to 20 s for the composite gradient. Note that the convergence rate due to the addition of momentum is more pronounced in the case of the composite algorithm as compared to the RLS algorithm.
Figure 4.
Parameter estimate values versus time using the composite gradient and the momentum-based composite gradient algorithm. In both cases the parameter estimate values converge to the true values with the momentum-based integral algorithm converging slightly faster.
Figure 5.
Norm of the parameter estimate error versus time using the composite and the momentum-based composite algorithms.
5.2. Momentum-Based Recursive Least Squares and Model Reference Adaptive Control
Here, we consider the short-term dynamics of the aircraft as an example, as shown in Figure 6, where is the angle of attack, is the flight path angle, and is the equilibrium linear axial velocity. The transfer function describing the short-term dynamics of the reduced-order longitudinal state equation of the aircraft with angle of attack output and elevator deflection input is given by [21]
where , , , and , and , , and are the concise partial derivatives (see [21]) of the pitching moment with respect to the elevator angle , the normal velocity w, and the pitch rate q. Moreover, and denote the concise partial derivatives of the normal force on the aircraft with respect to the elevator angle and the normal velocity w. Here, we set , , , and [21].
Figure 6.
Schematic of an aircraft.
The desired system performance is to track the reference output
where is a reference command signal. Next, we use the filter system (78) and (79) with and , and the control law (82) whose state space realization is given by
where For our simulations, we select the reference signal to be a square wave with a frequency and amplitude 1. Furthermore, the system initial conditions are set to , , , and .
Next, we compare the RLS algorithm (89) and (90) with our proposed MRLS algorithm (91)–(93). For this comparison, we set , , , , , , , and . Note that all the conditions of Theorem 5 are satisfied. Figure 7 shows the system parameters versus time for the MRLS and the RLS algorithms. It can be seen that in both cases the system parameters converge to the true values with the MRLS providing faster convergence as compared to the standard RLS update law. Finally, Figure 8 shows the moving average (MA) of the absolute value of the tracking error with a sliding window of period . Note that the MRLS algorithm provides slightly better tracking accuracy as compared to the RLS algorithm after the first 10 s. Finally, we note that the difference in runtime complexity between the different algorithms addressed in this paper is negligible.
Figure 7.
Parameter estimate values versus time using the RLS and MRLS algorithms. In both cases the parameter estimate values converge to the true values with the MRLS algorithm converging faster.
Figure 8.
Moving average of the absolute value of the error for the RLS and MRLS algorithms with a sliding window of . The tracking error for the MRAC scheme predicated on the MRLS algorithm decreases faster than that predicated on the RLS algorithm.
6. Conclusions
In this paper, we developed three new momentum-based update laws for online parameter identification and model reference adaptive control. Specifically, we augmented higher-order tuner architectures into the integral gradient, recursive least squares, and composite gradient algorithms to achieve faster error convergence of the system parameters. Several numerical examples were provided to show the efficacy of the proposed approach. Future work will focus on developing new adaptive update laws for identification and control that guarantee finite time and fixed time convergence of the system parameters.
Author Contributions
L.S.: Conseptualization, Formal analysis, Software, Visualization, Writing-original draft. W.M.H.: Conceptualization, Formal analysis, Writing-review and editing, Supervision, Funding Acquisition. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported in part by the Air Force Office of Scientific Research under Grant FA9550-20-1-0038.
Data Availability Statement
Data are contained within the article.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Åström, K.J.; Wittenmark, B. Adaptive Control; Dover Publications: Mineola, NY, USA, 2008. [Google Scholar]
- Ioannou, P.; Sun, J. Robust Adaptive Control; Dover Publications: Garden City, NY, USA, 2012. [Google Scholar]
- Narendra, K.; Annaswamy, A.M. Stable Adaptive Systems; Prentice-Hall: Englewood Cliffs, NJ, USA, 1989. [Google Scholar]
- Krstic, M.; Kanellakopoulos, I.; Kokotovic, P. Nonlinear and Adaptive Control Design; Wiley: New York, NY, USA, 1995. [Google Scholar]
- Cui, Y.; Annaswamy, A.M. Discrete-Time High Order Tuner with A Time-Varying Learning Rate. In Proceedings of the 2023 American Control Conference, San Diego, CA, USA, 31 May–2 June 2023; pp. 2993–2998. [Google Scholar]
- Nesterov, Y. Introductory Lectures on Convex Optimization; Springer: New York, NY, USA, 2004. [Google Scholar]
- Gaudio, J.E.; Gibson, T.E.; Annaswamy, A.M.; Bolender, M.A. Provably Correct Learning Algorithms in the Presence of Time-Varying Features Using a Variational Perspective. arXiv 2019, arXiv:1903.04666. [Google Scholar]
- Boffi, N.M.; Slotine, J.J.E. Implicit Regularization and Momentum Algorithms in Nonlinearly Parameterized Adaptive Control and Prediction. Neural Comput. 2021, 33, 590–673. [Google Scholar] [CrossRef] [PubMed]
- Online accelerated data-driven learning for optimal feedback control of discrete-time partially uncertain systems. Int. J. Adapt. Control Signal Process. 2023, 38, 848–876.
- Costa, R.R. Model-reference adaptive control with high-order parameter tuners. In Proceedings of the 2022 American Control Conference, Atlanta, GA, USA, 8–10 June 2022; pp. 3370–3375. [Google Scholar]
- Costa, R.R. Least-squares model-reference adaptive control with high-order parameter tuners. Automatica 2024, 163, 111544. [Google Scholar] [CrossRef]
- Wibisono, A.; Wilson, A.C.; Jordan, M.I. A variational perspective on accelerated methods in optimization. Proc. Natl. Acad. Sci. USA 2016, 113, 7351–7358. [Google Scholar] [CrossRef] [PubMed]
- Cho, N.; Shin, H.S.; Kim, Y.; Tsourdos, A. Composite Model Reference Adaptive Control with Parameter Convergence Under Finite Excitation. IEEE Trans. Autom. Control 2018, 63, 811–818. [Google Scholar] [CrossRef]
- Shaferman, V.; Schwegel, M.; Glück, T.; Kugi, A. Continuous-time least-squares forgetting algorithms for indirect adaptive control. Eur. J. Control 2021, 62, 105–112. [Google Scholar] [CrossRef]
- Haddad, W.M.; Chellaboina, V. Nonlinear Dynamical Systems and Control: A Lyapunov-Based Approach; Princeton University Press: Princeton, NJ, USA, 2008. [Google Scholar]
- Goodwin, G.; Mayne, D. A parameter estimation perspective of continuous time model reference adaptive control. Automatica 1987, 23, 57–70. [Google Scholar] [CrossRef]
- Gaudio, J.E.; Annaswamy, A.M.; Lavretsky, E.; Bolender, M.A. Parameter Estimation in Adaptive Control of Time-Varying Systems Under a Range of Excitation Conditions. IEEE Trans. Autom. Control 2022, 67, 5440–5447. [Google Scholar] [CrossRef]
- Costa, R.R. Lyapunov design of least-squares model-reference adaptive control. IFAC-PapersOnLine 2020, 53, 3797–3802. [Google Scholar] [CrossRef]
- Costa-Gomes, M.A.; Crawford, V.P.; Iriberri, N. Comparing models of strategic thinking in Van Huyck, Battalio, and Beil’s coordination games. J. Eur. Econ. Assoc. 2009, 7, 365–376. [Google Scholar] [CrossRef]
- Naik, S.; Kumar, P.; Ydstie, B. Robust continuous-time adaptive control by parameter projection. IEEE Trans. Autom. Control 1992, 37, 182–197. [Google Scholar] [CrossRef]
- Cook, M. Flight Dynamics Principles; Elsevier: Oxford, UK, 2007. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).