Next Article in Journal
Adaptive Maximum Torque per Ampere Control of Sensorless Permanent Magnet Motor Drives
Previous Article in Journal
Unconventional Excess Heat Sources for District Heating in a National Energy System Context
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Stability of Perturbed Nonlinear Switched Systems with Adaptive Reinforcement Learning

1
Department of Automatic Control, Hanoi University of Science and Technology, 1 Dai Co Viet Road, Hanoi 100000, Vietnam
2
Department of Automation, Thai Nguyen University of Technology, 666, 3/2 Street, Tich Luong Ward, Thai Nguyen City 251750, Vietnam
3
Department of Electrical Engineering, Chonnam National University, Gwangju 61186, Korea
*
Authors to whom correspondence should be addressed.
Energies 2020, 13(19), 5069; https://doi.org/10.3390/en13195069
Submission received: 19 August 2020 / Revised: 21 September 2020 / Accepted: 24 September 2020 / Published: 27 September 2020
(This article belongs to the Section F: Electrical Engineering)

Abstract

:
In this paper, a tracking control approach is developed based on an adaptive reinforcement learning algorithm with a bounded cost function for perturbed nonlinear switched systems, which represent a useful framework for modelling these converters, such as DC–DC converter, multi-level converter, etc. An optimal control method is derived for nominal systems to solve the tracking control problem, which results in solving a Hamilton–Jacobi–Bellman (HJB) equation. It is shown that the optimal controller obtained by solving the HJB equation can stabilize the perturbed nonlinear switched systems. To develop a solution to the translated HJB equation, the proposed neural networks consider the training technique obtaining the minimization of square of Bellman residual error in critic term due to the description of Hamilton function. Theoretical analysis shows that all the closed-loop system signals are uniformly ultimately bounded (UUB) and the proposed controller converges to optimal control law. The simulation results of two situations demonstrate the effectiveness of the proposed controller.

1. Introduction

Many power electronic converters play a remarkable role in industrial applications, such as electrical drives, renewable energy systems, etc. [1,2,3,4,5]. The modelling technique in power electronic converters is usually implemented by considering the average small-signal analysis. This method enables us to design the voltage, current controller, as well as implement the space vector modulation technique. However, the average small-signal model method is disadavantageous for complicated applications, with the requirement of tracking control systems. Switched systems have been known as a special class of hybrid systems describing an active application area in the field of power electronics, such as DC–DC converter, power source, etc. [6,7,8,9,10,11,12,13,14,15]. The approach of employing switched systems representation in control design has been the efficient method to model many practical systems based on multi-subsystem. Basic issues in the control of switched systems are to find the control input under arbitrary switching signal and/or to find the appropriate switching signal obtaining the tracking, stability of closed-loop systems. Additionally, the identification of the active mode based on estimating the delay to analyze the switched systems was implemented in [6]. The authors in [7] proposed a novel Lyapunov function combining linear matrix inequalities (LMIs) for switched linear systems in the presence of exogenous disturbance. The control design problem with stochastic stability based on a complete probability space under the general random switching signal was investigated in [8]. The stabilizing switching signal was designed using average dwell time with proposed existence conditions [9]. Moreover, several conditions to characterize the mode-dependent dwell-times have been given in [10]. The switching event triggered and back stepping technique-based tracking controller achieved the bound of all signals and tracking error in the closed-loop system [11]. However, the optimal control with effectiveness in dealing with constraint has not been mentioned in switched systems yet [6,7,8,9,10,11]. For the optimal control design, it is necessary to find the result of a Ricatti equation for linear systems or a Hamilton–Jacobi–Bellman (HJB) equation for nonlinear systems [16]. However, in the general case, it is hard to solve the HJB (Hamilton–Jacobi–Bellman) equation for finding the optimal controller. In [16,17,18,19], thanks to the approximation description of Neural Networks (NNs), the weight of the actor and critic were updated simultaneously under the optimization problem for continuous/discrete time systems. The convergence effectiveness of weights in actor/critic policy is guaranteed by the persistence excitation (PE) condition. It is worth emphasizing that to deal with the uncertainties of nonlinear continuous systems, the identifier was inserted into the control structure [19] and the off-policy technique under the separation between actor/critic policy and control input in the first stage is presented [20]. Moreover, the constraint in input/output was mentioned in the adaptive/approximate dynamic programming (ADP) technique [17,18,21] by using the appropriate cost function and the dynamic programming principle. The framework of ADP was utilized in control scheme for surface vessel systems [22], a spring-Mass-Damper [23], two mass-spring systems [20], Wheeled mobile robotic systems [24]. Based on approximating the modified critic term in each control loop, the stability of whole cascade systems was guaranteed in surface vessel systems due to the consideration of the derivative of Lyapunov candidate function along the closed system [22,25]. However, all the ARL-based controllers presented in [20,22,23,24] are implemented under the consideration of nonlinear systems, which are not mentioned in the switching signals. The model-free problem, as well as the connection between continuous time systems and appropriate discrete time systems, ensure the development of off-policy algorithm [26,27]. To deal with the complete dynamic uncertainties, the modified cost function and critic enable us to obtain the approximately optimal control law [28,29,30,31]. Additionally, the general policy iteration (GPI) in linear discrete time systems [32] achieves the simultaneous computation and appropriate data collection.
To the best knowledge of the authors, adaptive reinforcement learning (ARL) for uncertain switched systems is still an open problem. Therefore, this paper attempts to implement an ARL algorithm for uncertain continuous-time nonlinear switched systems under arbitrary switching signals based on an optimization problem in NNs training. Inspired by the above works, this paper studies the ARL-based optimal control problem for a class of perturbed switched nonlinear continuous systems. First, the optimal control algorithm is established for switched systems. Then, the neural network is employed to approximate the Critic part in policy. Based on optimization principle, the training law is proposed to develop the optimal control strategy. The main contributions of this paper are as follows:
(1)
In comparison with the previous papers [12,13,14,15,18,22,29,30,33], an optimal control obtained from the nominal system is proposed for perturbed switched nonlinear continuous systems based on dynamic programming principle.
(2)
The Neural Network training law with optimization principle is developed to achieve the ARL-based optimal control strategy.
(3)
The strict proof concerning UUB stability of the closed system and the convergence of the controller to optimal control input are given based on Lyapunov stability theory and reinforcement learning scheme.
The rest of this article is organized as follows. The preliminaries and problem statements are presented in Section 2. The main results are given in Section 3. Two simulation cases are presented in Section 4 to illustrate the effectiveness of the proposed solution. Finally, the conclusions are exhibited in Section 5.

2. Problem Statement and Preliminaries

In [15], it can be seen that switched systems can represent a useful framework for modelling general power electronics converters. Therefore, in this section, we consider the perturbed continuous-time nonlinear switched system described by:
x ˙ = f σ x + g σ x u + Δ x , t
where x t Ω x R n and u t Ω u R m are the vector of state variables and the control inputs vector, respectively. The function σ : 0 , + Ω = 1 , 2 , . . . , n is an unknown switching signal and n is the number of subsystems. f i x i Ω are unknown smooth vector functions satisfying f i 0 = 0 and g i x i Ω are known smooth vector functions such that G min g i x G max .
Assumption 1.
There exists a known function ρ x satisfying that: Δ x , t ρ x .
Regarding the perturbed continuous-time nonlinear switched system (1), we introduce the cost function formulated as:
J x t , u t = t r x τ , u τ d τ
where r x τ , u τ = x T Q x + u T R u and Q = Q T 0 ; R = R T > 0
Control Objective: This article aims at designing the optimal guaranteed cost control scheme u = arg min u Ω u K x , u despite arbitrary switching law, in which the feedback control law u and a finite upper bound function K x , u are satisfied, not only is the closed-loop system (1) robustly stable but the cost function (2) also satisfies the condition: J x , u K x , u M .
Definition 1.
The function K ( x , u ) can be known as the guaranteed cost function. Therefore, the control law u with u = arg min u Ω u K x , u is known as the optimal guaranteed cost control law.
Remark 1.
It is worth emphasizing that the main objective of this work is to find the optimal control for the equivalent nominal system with the performance index being modified. Compared with the control objective in [33], this work investigates the tracking control for switched systems, for which it is hard to develop the optimal control algorithm. Additionally, it is obviously different from the existing work of robust ADP in [31], since the proposed optimal control design is implemented based on the upper bound function K ( x , u ) .

3. Adaptive Reinforcement Learning-Based Control Design

In this section, we investigate the ARL-based optimal control for perturbed nonlinear switched systems. Due to the difficulties in implementing directly the optimal control scheme for perturbed nonlinear systems, the strategy is proposed by three steps as follows. First, based on Dynamic Programming, we obtain the optimal control design for the corresponding nominal switched systems obtained by eliminating the uncertainties. Then, the ARL algorithm is developed by Neural Network technique for this nominal system. Finally, we carry out a stability analysis of the closed-loop system involving the perturbed switched system and the proposed ARL controller.
According to perturbed system (1), the nominal system is obtained by eliminating the term of uncertainties as:
x ˙ = f σ x + g σ x u
Because the purpose is to consider the tracking problem of optimal control law and the influence of uncertainties, the cost function is modified to satisfy that J x t , u t J 1 x t , u t . Therefore, the cost function associated with (3) can be represented as:
J 1 x t , u t = t r x τ , u τ + λ ρ x 2 d τ
It should be noted that J 1 x t , u t with λ R is one of the guaranteed cost function associated with system (1). According to (2) and (4), it can be seen that:
J x t , u t J 1 x t , u t
Based on the Dynamic programming principle, establishing the Bellman function as: V t = min u Ω u J 1 x t , u t , we have the cost function formulated as:
V t = min u Ω u t r x τ , u τ + λ ρ x 2 d τ
V t = min u Ω u t t + Δ t r x τ , u τ + λ ρ x 2 d τ + min u Ω u t + Δ t r x τ , u τ + λ ρ x 2 d τ
V x = min u Ω u t t + Δ t r x τ , u τ + λ ρ x 2 d τ + V t + Δ t
min u Ω u 1 Δ t t t + Δ t r x τ , u τ + λ ρ x 2 d τ + V t + Δ t V t Δ t = 0
As the convergence of Δ t 0 + , we can derive that:
min u Ω u r x t , u t + λ ρ 2 x + V T f σ x + g σ x u = 0
where V = V x .
Consider the Hamilton function obtaining from the nominal system and performance index (4):
H x , u , V = r x t , u t + λ ρ 2 x + V T f σ x + g σ x u
The control input is computed by minimizing this function with u Ω u as
H x , u , V = min u Ω u H x , u , V = 0
Therefore, we find that (13) yields
H x , u , V u u = u = 0 u = 1 2 R 1 g σ x T V
We implement to develop this control algorithm (13) for nonlinear switched system (1) and achieve the following result:
Theorem 1.
Take system (1) with the feedback control law u x = 1 2 R 1 g σ x T V into consideration. Then, the cost function V t = t r x τ , u τ + λ ρ x 2 d τ is the Lyapunov function candidate with λ R can guarantee that the system (1) is stable.
Proof. 
Consider the derivative of V ( t ) along the solution of Equation (1), we achieve the result formulated as:
V ˙ t = V T f i x + g i x u + Δ x , t
It follows from u x = 1 2 R 1 g i x T V and according to (14), (11) and (12), we have:
8 V ˙ t = r x t , u t λ ρ 2 x + g i x T V Δ x , t = x T Q x u T R u λ ρ 2 x 2 u T R Δ x , t
= x T Q x λ ρ 2 x u T + Δ x , t T R u + Δ x , t + Δ x , t T R Δ x , t
= x T Q x λ ρ 2 x Δ x , t T R Δ x , t u + Δ x , t T R u + Δ x , t
According to Assumption 1 and λ R , it can be seen that:
V ˙ t x T Q x λ ρ 2 x R Δ x , t 2 x T Q x λ R ρ 2 x
V ˙ t x T Q x
Therefore, the system (1) is stable under the optimal control for equivalent nominal system. □
It is noteworthy that Theorem 1 has a new extension in switched systems compared with nonlinear systems in the existing result [33].
However, it is impossible to find the Bellman function V x based on solving analytically the HJB nonlinear Equation (12). Hence, to solve it, we construct a critic network under the framework of adaptive critic learning. Using the approximation property of neural networks described in [28], the critic associated with system (3) can be described as:
V x = w T σ x + ε x
where σ x : R n R N ; σ 0 = 0 is the vector activation function including N linearly independent elements, N is the number of neurons in the hidden layer of Radial Basis Function (RBF) network [28] and it is also the number of dimensions of σ x : R n R N , and ε x is the function reconstruction error playing the role in finding the training law the in next steps, w R N is the idea weight vector and is generally unavailable. As N converges to infinity, we obtain that ε x 0 and ε x 0 . The following assumption is considered for each fixed N.
Assumption 2.
The estimation of terms in NNs can be described as: ε x ε max ; ε x ε max ; σ min σ x σ max ; w w max
According two Equations (12) and (13), it holds that:
H x , u , V = r x t , u t + λ ρ 2 x + V T f i x + g i x u = 0
= x T Q x + λ ρ 2 x + V T f i x 1 4 V T g i x R 1 g i x T V = 0
It follows from (20) to:
V = σ x T w + ε x
The NN-based HJB equation can be represented as follows:
e N N = x T Q x + λ ρ 2 x + σ x T w T f i x 1 4 σ x T w T g i x R 1 g i x T σ x T w
e N N = x T Q x + λ ρ 2 x + w T σ x f i x 1 4 w T σ x g i x R 1 g i x T σ x T w
Thus, the residual error is formulated by the function approximation error:
e N N = ε x T f i x + g i x u + 1 4 ε x T g i x R 1 g i x T ε x
It should be noted that, as N goes to infinity, e N N converges uniformly to zero. Hence, the residual error e N N is bounded for each fixed N. Under the framework of ADP-based approximate optimal control design, a critic neural network is established by an estimated weight vector w ˆ given as:
V ˆ = w ˆ T σ x = σ x T w ˆ ; u ˆ = 1 2 R 1 g i x T V ˆ
The approximate error of the critic network can be formulated as:
e H J B = r x t , u ˆ t + λ ρ 2 x + σ x T w ˆ T f i x + g i x u ˆ
e H J B = x T Q x + λ ρ 2 x + w ˆ T σ x f i x 1 4 w ˆ T σ x g i x R 1 g i x T σ x T w ˆ
The weight vector is determined by a steepest descent algorithm:
d d t w ˆ = α E w ˆ
with E = 1 2 e H J B T e H J B
Remark 2.
The weight vector w ˆ is designed as in (30) to minimize the quadratic function E = 1 2 e H J B T e H J B . The fact is that the following inequality:
E ˙ = E t = E w ˆ . w ˆ t = E w ˆ . d d t w ˆ = α E w ˆ 2 0
Theorem 2.
The proposed optimal control (27) with the weight vector of the critic network (30) can force the system (1) to be uniform ultimate bounded (UUB) stability.
Proof. 
Define: w ˜ = w w ˆ d d t w ˜ = d d t w ˆ . Consider the Lyapunov function:
V t = V 1 t + V 2 t , where : V 1 t = 1 2 α w ˜ t T w ˜ t , V 2 t = V
For deriving the term V 1 ( t ) , we obtain that:
d d t V 1 t = 1 α w ˜ t T d d t w ˜ t = 1 α w ˜ t T d d t w ˆ t = w ˜ t T E w ˆ
V ˙ 1 = w ˜ T e H J B σ x f i x + g i x u ˆ
According to (27) and (13), it follows that:
u ˆ u = 1 2 R 1 g i x T V ˆ V
= 1 2 R 1 g i x T σ x T w ˆ σ x T w ε x
= 1 2 R 1 g i x T σ x T w ˜ + ε x
Moreover, we have:
σ x f i x + g i x u ˆ = σ x f i x + g i x u + σ x g i x u ˆ u
σ x f i x + g i x u ˆ = σ x f i x + g i x u + 1 2 g i x R 1 g i x T σ x T w ˜ + ε x
From (26) and (29), it follows:
e H J B e N N = σ x T w ˆ T f i x σ x T w T f i x + 1 4 w T σ x g i x R 1 g i x T σ x T w w ˆ T σ x g i x R 1 g i x T σ x T w ˆ + 1 2 σ x T w ˆ T g i x R 1 g i x T σ x T w ˜ + ε x
e H J B e N N = σ x T w ˜ T f i x + 1 4 w T w ˆ T σ x g i x R 1 g i x T σ x T w + w ˆ T σ x g i x R 1 g i x T σ x T w w ˆ
e H J B e N N = w ˜ T σ x f i x + 1 4 w ˜ T σ x g i x R 1 g i x T σ x T w + w ˆ T σ x g i x R 1 g i x T σ x T w ˜
e H J B e N N = w ˜ T σ x f i x + 1 4 w ˜ T σ x g i x R 1 g i x T σ x T w + w w ˜ T σ x g i x R 1 g i x T σ x T w ˜
e H J B e N N = w ˜ T σ x f i x + 1 2 w ˜ T σ x g i x R 1 g i x T σ x T w 1 4 w ˜ T σ x g i x R 1 g i x T σ x T w ˜
Because u = 1 2 R 1 g i x T σ x T w + ε x , we have:
e H J B e N N = w ˜ T σ x f i x + w ˜ T σ x g i x u 1 2 R 1 g i x T ε x 1 4 w ˜ T σ x g i x R 1 g i x T σ x T w ˜
e H J B = e N N w ˜ T σ x f i x + g i u 1 2 w ˜ T σ x g i x R 1 g i x T ε x 1 4 w ˜ T σ x g i x R 1 g i x T σ x T w ˜
Assumption 3.
f i x + g i x u μ max
Define: μ i = f i x + g i x u ; G i = g i x R 1 g i x T ; σ = σ x ; ε = ε x
From (35) we obtain:
V ˙ 1 t = w ˜ T e N N + w ˜ T σ μ i + 1 2 w ˜ T σ G i ε + 1 4 w ˜ T σ G i σ T w ˜ σ x μ i + 1 2 G i σ T w ˜ + ε
V ˙ 1 t = 1 8 w ˜ T w ˜ T σ G i σ T w ˜ σ G i σ T w ˜ 1 2 w ˜ T w ˜ T σ μ i σ G i σ T w ˜ + 1 4 w ˜ T w ˜ T σ G i ε σ G i σ T w ˜ + 1 4 w ˜ T w ˜ T σ G i σ T w ˜ σ μ i w ˜ T w ˜ T σ μ i + 1 2 w ˜ T σ G i ε σ μ i + 1 2 w ˜ T e N N σ G i σ T w ˜ + w ˜ T e N N σ μ i
V ˙ 1 t = 1 8 w ˜ T σ G i σ T w ˜ 2 1 2 w ˜ T σ μ i + 1 4 w ˜ T σ G i ε + 1 4 w ˜ T σ μ i + 1 2 e N N w ˜ T σ G i σ T w ˜ w ˜ T σ μ i 2 + e N N 1 2 w ˜ T σ G i ε w ˜ T σ μ i
V ˙ 1 t = 1 8 w ˜ T σ G i σ T w ˜ 2 3 4 w ˜ T σ μ i + 1 4 w ˜ T σ G i ε + 1 2 e N N w ˜ T σ G i σ T w ˜ w ˜ T σ μ i 2 + e N N 1 2 w ˜ T σ G i ε w ˜ T σ μ i
Define: A = w ˜ T σ G i σ T w ˜ ; B = 3 4 w ˜ T σ μ i + 1 4 w ˜ T σ G i ε + 1 2 e N N ; C = w ˜ T σ μ i ; D = e N N 1 2 w ˜ T σ G i ε
It is obvious that:
V ˙ 1 t = 1 8 A 2 B A C 2 D C = 1 8 A + 4 B 2 + 2 B 2 C + D 2 2 + D 2 4 1 8 A + 4 B 2 16 B 2 + 2 D 2
We have:
A + 4 B w ˜ 2 G min 2 λ min R 1 σ min 2 w ˜ 3 σ max μ max + σ max G max 2 λ max R 1 ε max 2 e max
16 B 2 + 2 D 2 w ˜ 3 σ max μ max + σ max G max 2 λ max R 1 ε max + 2 e max 2 + 2 e max + 1 2 w ˜ σ max σ max G max 2 λ max R 1 ε max 2
According to (49) and (50), the inequality can be obtained
A + 4 B 2 16 B 2 + 2 D 2 π 1
It can be obtained that π 1 > 0 with a big enough variable w ˜ and the highest order coefficient G min 2 λ min R 1 σ min 2 2 > 0 . Therefore, we can determine the positive number ϑ 1 such that: w ˜ > ϑ 1 , we imply that A + 4 B 2 16 B 2 + 2 D 2 π 1 , and from (48) we obtain: V ˙ 1 t π 1 .
Regarding the term of V 2 ( t ) , from (20) we compute the derivative of them:
V ˙ 2 = V T f i + g i u ˆ + Δ = x T Q x + λ ρ 2 x + 1 4 V T g i x R 1 g i x T V + V T g i 1 2 R 1 g i x T σ x T w ˆ + Δ = x T Q x + λ ρ 2 x 1 4 V T g i x R 1 g i x T V + 1 2 V T g i R 1 g i x T V σ x T w ˆ + V T g i Δ = x T Q x + λ ρ 2 x 1 4 V T g i x R 1 g i x T V + 1 2 V T g i R 1 g i x T σ x T w ˜ + ε x + V T g i Δ = x T Q x + λ ρ 2 x 1 4 V T g i R 1 g i T V + 1 2 V T g i R 1 g i T σ x T w ˜ + ε x + V T g i Δ
It can be considered that ρ x = ϖ x . According to (51), we obtain:
V ˙ 2 λ min Q + λ ϖ x 2 + θ 2
with
θ 2 = 1 4 V T g i R 1 g i T V + 1 2 V T g i R 1 g i T σ x T w ˜ + ε x + V T g i Δ
According to the assumption 1 and assumption 2, it follows that:
θ 2 1 4 w max σ max + ε max 2 g max 2 λ max R 1 + 1 2 ϑ σ max + ε max 2 g max 2 λ max R 1 + w max σ max + ε max g max ϖ x
It can be seen that λ min Q + λ ϖ x 2 θ 2 π 2 with π 2 > 0 also leads to other inequalities, including the polynomial quadratic form with the variable x and the highest order coefficient λ min Q + λ ϖ > 0 . Hence, we can find the positive number ϑ 2 such that: x > ϑ 2 we get λ min Q + λ ϖ x 2 θ 2 π 2 , and from (52) we obtain
V ˙ 2 t π 2 .
 □
Remark 3.
The numbers ϑ 1 ; ϑ 2 can be changed by establishing the neural network of the optimal cost function. Additionally, for any switching index, the variable x and w ˜ converge to the specified domains. The approximate optimal control law u ˆ is considered in the equation as (27), which converges to the neighbourhood of the optimal control as u . Different from the controller in [33], the switched system with unknown switching signal is mentioned in this work by proposed adaptive optimal control.
Proof. 
It should be noted that according to (34), we imply that:
u ˆ u = 1 2 R 1 g i x T σ x T w ˜ + ε x 1 2 λ max R 1 . G max . σ max . υ 1 + ε max = ϑ 3
 □

4. Simulation Results

In this section, we verify the effectiveness and performance of the proposed controller. The ARL algorithm is developed for solving the optimal and tracking problem by RBF network-based ADP, where a single critic Neural Network is used to approximate the Bellman function. The ARL control law and the weight of the Critic part are established as in (27) and (30) with the appropriate learning rate α and the coefficient λ . Moreover, to carry out the ARL algorithm (27) and (30), we need to employ the term of model g i ( x ) , the term of cost function Q , R and the appropriate function σ ( x ) in Neural Network. To verify the proposed method of our algorithm, two different situations are implemented to do simulations with all the parameters and functions being listed in each case as follows:

4.1. The Second-Order Switched Nonlinear Systems

In this simulation experiment, we consider the case of switched systems including N = 2 subsystems to be described as Equations (56) and (57):
x ˙ 1 = x 1 3 2 x 2 + u + Δ 1 x , t x ˙ 2 = x 1 + 0.5 cos x 1 2 sin x 2 3 u + Δ 1 x , t
x ˙ 1 = x 1 5 sin x 2 + u + Δ 2 x , t x ˙ 2 = 1 2 x 1 cos x 1 cos x 2 3 u + Δ 2 x , t
The initial value of state vector is selected as
x 0 = 5 5 T
and choosing the parameters and matrices in cost function as follows: R = 2 0 0 2 ; Q = 1 0 0 3 ; α = 0.1 ; λ = 5 . Additionally, the function σ ( x ) in RBF Neural Network can be chosen as σ x = x 1 2 , x 1 x 2 , x 2 2 T .
Under the arbitrary switching law (Figure 1), the results are shown in Figure 2, Figure 3, Figure 4, Figure 5 and Figure 6 by using proposed ARL law, in which the responses of state variables are described in Figure 2 and Figure 3 under the control input (Figure 5) and the weights are trained as in Figure 6. It is obvious that the closed system is stable due to the convergence of state variables to zero (Figure 2 and Figure 3). On the other hand, the convergence is also shown in the weights of Critic part (Figure 6).
Remark 4.
The second-order switched nonlinear system was also mentioned in [14] by using nonlinear control law after the equivalent transform operation (Figure 1, Figure 2, Figure 3, Figure 4, Figure 5 and Figure 6). However, it is noteworthy that the response of control input of proposed method has a better transient and steady performance compared with the nonlinear controller in [14] because of the effectiveness of handling the constraint in optimal control (Figure 4 and Figure 5).

4.2. The Third-Order Switched Nonlinear Systems

Here, we continue to investigate the case of switched systems including N = 2 subsystems as follows:
x ˙ = x 1 + x 2 0.2 x 2 sin x 1 cos x 3 x 1 x 2 sin x 3 + 1 1 2 u + Δ 1 x , t x ˙ = x 1 + x 2 2 x 3 x 2 sin x 1 cos x 3 sin x 2 x 3 + 1 1 2 u + Δ 2 x , t
where x 0 = 5 ; 3 ; 2 T ; Δ i x , t = 1 10 d 1 x 1 d 2 x 2 d 3 x 3 ; Δ i 1 10 x with d 1 , d 2 , d 3 are the random functions. The parameters of proposed control design can be chosen as R = 1 0 0 1 ; α = 0.01 ; Q = 1 0 0 0 2 0 0 0 3 ; λ = 5 . Moreover, the function σ ( x ) in RBF Neural Network can be chosen as σ x = x 1 2 , x 2 2 , x 3 2 , x 1 x 2 , x 2 x 3 , x 1 x 3 T . In this case, under the arbitrary switching law in Figure 7, the results are obtained in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13 with the responses of state variables are shown in Figure 8, Figure 9 and Figure 10 based on the control input (Figure 11 and Figure 12), and the training weights are shown in Figure 13. It can be seen that the closed system is stable (Figure 8, Figure 9 and Figure 10) and the convergence is also shown in the weights of Critic part (Figure 13).
Remark 5.
It is worth noting that the third-order switched nonlinear system was also considered in [15] based on the adaptive backstepping output feedback nonlinear control scheme combined with the transform operation obtaining the response of state variables, control input in Figure 8, Figure 9, Figure 10 and Figure 11 and adaptation law (Figure 14). However, thanks to the property of optimal control, the control input response of the proposed controller has a better transient compared with the adaptive backstepping nonlinear controller in [15] (Figure 11 and Figure 12).

5. Conclusions

This paper investigated the optimal control design for perturbed switched nonlinear systems based on the adaptive dynamic programming technique. The optimal control is first designed under the consideration of nominal systems. Then, the ADP technique is developed by using the Neural Networks. Due to the description of nonlinear systems, as well as unknown switch index, the neural network is considered to approximate the critic part of the iterative algorithm. Moreover, the UUB stability problem of the closed-loop system and the convergence of weight training are guaranteed under this solution. Finally, two simulation examples are given to verify the effectiveness of the presented ARL algorithm.

Author Contributions

P.N.D. was originally responsible for conceptualization, methodology, simulation and prepared the original draft of the article. H.Q.N., M.-D.N. verified the simulation data, data curation, investigation, and draft writing. S.-J.A. contribute a formal analysis of review, editing, and supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Korea Electric Power Corporation (Grant number: R18XA04).

Acknowledgments

The authors gratefully acknowledge Thai Nguyen University of Technology for supporting this work.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ARLAdaptive Reinforcement Learning
ADPAdaptive Dynamic Programming
UUBUniformly Ultimately Bounded
NNsNeural Networks
HJBHamilton–Jacobi–Bellman

References

  1. Loschi, H.; Smolenski, R.; Lezynski, P.; Nascimento, D.; Demidova, G. Aggregated Conducted Electromagnetic Interference Generated by DC/DC Converters with Deterministic and Random Modulation. Energies 2020, 13, 3698. [Google Scholar] [CrossRef]
  2. Bakeer, A.; Chub, A.; Vinnikov, D. Step-Up Series Resonant DC–DC Converter with Bidirectional-Switch- Based Boost Rectifier for Wide Input Voltage Range Photovoltaic Applications. Energies 2020, 13, 3747. [Google Scholar] [CrossRef]
  3. Premkumar, M.; Subramaniam, U.; Haes Alhelou, H.; Siano, P. Design and Development of Non-Isolated Modified SEPIC DC-DC Converter Topology for High-Step-Up Applications: Investigation and Hardware Implementation. Energies 2020, 13, 3960. [Google Scholar] [CrossRef]
  4. Korkh, O.; Blinov, A.; Vinnikov, D.; Chub, A. Review of Isolated Matrix Inverters: Topologies, Modulation Methods and Applications. Energies 2020, 13, 2394. [Google Scholar] [CrossRef]
  5. Chen, B.-Y.; Shangguan, X.-C.; Jin, L.; Li, D.-Y. An Improved Stability Criterion for Load Frequency Control of Power Systems with Time-Varying Delays. Energies 2020, 13, 2101. [Google Scholar] [CrossRef] [Green Version]
  6. Zhang, L.; Xiang, W. Mode-identifying time estimation and switching-delay tolerant control for switched systems: An elementary time unit approach. Automatica 2016, 64, 174–181. [Google Scholar] [CrossRef]
  7. Yuan, S.; Zhang, L.; De Schutter, B.; Baldi, S. A novel Lyapunov function for a non-weighted L2 gain of asynchronously switched linear systems. Automatica 2018, 87, 310–317. [Google Scholar] [CrossRef] [Green Version]
  8. Xiang, W.; Lam, J.; Li, P. On stability and H control of switched systems with random switching signals. Automatica 2018, 95, 419–425. [Google Scholar] [CrossRef]
  9. Lin, J.; Zhao, X.; Xiao, M.; Shen, J. Stabilization of discrete-time switched singular systems with state, output and switching delays. J. Frankl. Inst. 2019, 356, 2060–2089. [Google Scholar] [CrossRef]
  10. Briat, C. Convex conditions for robust stabilization of uncertain switched systems with guaranteed minimum and mode-dependent dwell-time. Syst. Control. Lett. 2015, 78, 63–72. [Google Scholar] [CrossRef]
  11. Lian, J.; Li, C. Event-triggered control for a class of switched uncertain nonlinear systems. Syst. Control. Lett. 2020, 135, 104592. [Google Scholar] [CrossRef]
  12. Vu, T.A.; Nam, D.P.; Huong, P.T.V. Analysis and control design of transformerless high gain, high efficient buck-boost DC-DC converters. In Proceedings of the 2016 IEEE International Conference on Sustainable Energy Technologies (ICSET), Hanoi, Vietnam, 14–16 November 2016; pp. 72–77. [Google Scholar]
  13. Nam, D.P.; Thang, B.M.; Thanh, N.T. Adaptive Tracking Control for a Boost DC–DC Converter: A Switched Systems Approach. In Proceedings of the 2018 4th International Conference on Green Technology and Sustainable Development (GTSD), Ho Chi Minh City, Vietnam, 23–24 November 2018; pp. 702–705. [Google Scholar]
  14. Thanh, N.T.; Sam, P.N.; Nam, D.P. An Adaptive Backstepping Control for Switched Systems in presence of Control Input Constraint. In Proceedings of the 2019 International Conference on System Science and Engineering (ICSSE), Quang Binh, Vietnam, 19–21 July 2019; pp. 196–200. [Google Scholar]
  15. Chiang, M.-L.; Fu, L.-C. Adaptive stabilization of a class of uncertain switched nonlinear systems with backstepping control. Automatica 2014, 50, 2128–2135. [Google Scholar] [CrossRef]
  16. Vamvoudakis, K.G.; Vrabie, D.; Lewis, F.L. Online adaptive algorithm for optimal control with integral reinforcement learning. Int. J. Robust Nonlinear Control. 2014, 24, 2686–2710. [Google Scholar] [CrossRef]
  17. Bai, W.; Zhou, Q.; Li, T.; Li, H. Adaptive reinforcement learning neural network control for uncertain nonlinear system with input saturation. IEEE Trans. Cybern. 2019, 50, 3433–3443. [Google Scholar] [CrossRef] [PubMed]
  18. Zhang, T.; Xu, H. Adaptive optimal dynamic surface control of strict-feedback nonlinear systems with output constraints. J. Frankl. Inst. 2020, 30, 2059–2078. [Google Scholar] [CrossRef]
  19. Lv, Y.; Na, J.; Yang, Q.; Wu, X.; Guo, Y. Online adaptive optimal control for continuous-time nonlinear systems with completely unknown dynamics. Int. J. Control 2016, 89, 99–112. [Google Scholar] [CrossRef] [Green Version]
  20. Chen, C.; Modares, H.; Xie, K.; Lewis, F.L.; Wan, Y.; Xie, S. Reinforcement learning-based adaptive optimal exponential tracking control of linear systems with unknown dynamics. IEEE Trans. Autom. Control 2019, 64, 4423–4438. [Google Scholar] [CrossRef]
  21. Yang, X.; Wei, Q. Adaptive Critic Learning for Constrained Optimal Event-Triggered Control with Discounted Cost. IEEE Trans. Neural Netw. Learn. Syst. 2020. [Google Scholar] [CrossRef]
  22. Wen, G.; Ge, S.S.; Chen, C.L.P.; Tu, F.; Wang, S. Adaptive tracking control of surface vessel using optimized backstepping technique. IEEE Trans. Cybern. 2018, 49, 3420–3431. [Google Scholar] [CrossRef]
  23. Wang, D.; Mu, C. Adaptive-critic-based robust trajectory tracking of uncertain dynamics and its application to a spring–mass–damper system. IEEE Trans. Ind. Electron. 2017, 65, 654–663. [Google Scholar] [CrossRef]
  24. Li, S.; Ding, L.; Gao, H.; Liu, Y.-J.; Huang, L.; Deng, Z. ADP-based online tracking control of partially uncertain time-delayed nonlinear system and application to wheeled mobile robots. IEEE Trans. Cybern. 2019, 50, 3182–3194. [Google Scholar] [CrossRef]
  25. Wen, G.; Chen, C.L.P.; Ge, S.S.; Yang, H.; Liu, X. Optimized adaptive nonlinear tracking control using actor–critic reinforcement learning strategy. IEEE Trans. Ind. Inform. 2019, 15, 4969–4977. [Google Scholar] [CrossRef]
  26. Vamvoudakis, K.G.; Ferraz, H. Model-free event-triggered control algorithm for continuous-time linear systems with optimal performance. Automatica 2018, 87, 412–420. [Google Scholar] [CrossRef]
  27. Gao, W.; Jiang, Y.; Jiang, Z.-P.; Chai, T. Output-feedback adaptive optimal control of interconnected systems based on robust adaptive dynamic programming. Automatica 2016, 72, 37–45. [Google Scholar] [CrossRef] [Green Version]
  28. Mu, C.; Zhang, Y.; Gao, Z.; Sun, C. ADP-based robust tracking control for a class of nonlinear systems with unmatched uncertainties. IEEE Trans. Syst. Man Cybern. Syst. 2019. [Google Scholar] [CrossRef]
  29. Huang, Y. Optimal guaranteed cost control of uncertain non-linear systems using adaptive dynamic programming with concurrent learning. IET Control Theory Appl. 2018, 12, 1025–1035. [Google Scholar] [CrossRef]
  30. Tang, D.; Chen, L.; Tian, Z.F.; Hu, E. Modified value-function-approximation for synchronous policy iteration with single-critic configuration for nonlinear optimal control. Int. J. Control 2019, 1–13. [Google Scholar] [CrossRef]
  31. Fan, B.; Yang, Q.; Tang, X.; Sun, Y. Robust ADP design for continuous-time nonlinear systems with output constraints. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2127–2138. [Google Scholar] [CrossRef]
  32. Chun, T.Y.; Lee, J.Y.; Park, J.B.; Choi, Y.H. Adaptive dynamic programming for discrete-time linear quadratic regulation based on multirate generalised policy iteration. Int. J. Control 2018, 91, 1223–1240. [Google Scholar] [CrossRef]
  33. Mu, C.; Wang, D. Neural-network-based adaptive guaranteed cost control of nonlinear dynamical systems with matched uncertainties. Neurocomputing 2017, 245, 46–54. [Google Scholar] [CrossRef]
Figure 1. The Switching Law of the second-order Switched System.
Figure 1. The Switching Law of the second-order Switched System.
Energies 13 05069 g001
Figure 2. The response of state variable 1.
Figure 2. The response of state variable 1.
Energies 13 05069 g002
Figure 3. The response of state variable 2.
Figure 3. The response of state variable 2.
Energies 13 05069 g003
Figure 4. The response of Control Input.
Figure 4. The response of Control Input.
Energies 13 05069 g004
Figure 5. The response of Control Input using Adaptive Reinforcement Learning.
Figure 5. The response of Control Input using Adaptive Reinforcement Learning.
Energies 13 05069 g005
Figure 6. The Training Parameters of adaptive dynamic programming (ADP).
Figure 6. The Training Parameters of adaptive dynamic programming (ADP).
Energies 13 05069 g006
Figure 7. The Switching Law of the third-order Switched System.
Figure 7. The Switching Law of the third-order Switched System.
Energies 13 05069 g007
Figure 8. The response of state variable 1.
Figure 8. The response of state variable 1.
Energies 13 05069 g008
Figure 9. The response of state variable 2.
Figure 9. The response of state variable 2.
Energies 13 05069 g009
Figure 10. The response of state variable 2.
Figure 10. The response of state variable 2.
Energies 13 05069 g010
Figure 11. The response of Control Input.
Figure 11. The response of Control Input.
Energies 13 05069 g011
Figure 12. The response of Control Input using Adaptive Reinforcement Learning.
Figure 12. The response of Control Input using Adaptive Reinforcement Learning.
Energies 13 05069 g012
Figure 13. The Training Parameters of ADP.
Figure 13. The Training Parameters of ADP.
Energies 13 05069 g013
Figure 14. The Adaptation Law of Nonlinear Control in [15].
Figure 14. The Adaptation Law of Nonlinear Control in [15].
Energies 13 05069 g014

Share and Cite

MDPI and ACS Style

Dao, P.N.; Nguyen, H.Q.; Ngo, M.-D.; Ahn, S.-J. On Stability of Perturbed Nonlinear Switched Systems with Adaptive Reinforcement Learning. Energies 2020, 13, 5069. https://doi.org/10.3390/en13195069

AMA Style

Dao PN, Nguyen HQ, Ngo M-D, Ahn S-J. On Stability of Perturbed Nonlinear Switched Systems with Adaptive Reinforcement Learning. Energies. 2020; 13(19):5069. https://doi.org/10.3390/en13195069

Chicago/Turabian Style

Dao, Phuong Nam, Hong Quang Nguyen, Minh-Duc Ngo, and Seon-Ju Ahn. 2020. "On Stability of Perturbed Nonlinear Switched Systems with Adaptive Reinforcement Learning" Energies 13, no. 19: 5069. https://doi.org/10.3390/en13195069

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop