Next Article in Journal
A Blockchain-Based Privacy-Preserving Healthcare Data Sharing Scheme for Incremental Updates
Next Article in Special Issue
YOLOv7-SN: Underwater Target Detection Algorithm Based on Improved YOLOv7
Previous Article in Journal
Asymmetries Caused by Nonparaxiality and Spin–Orbit Interaction during Light Propagation in a Graded-Index Medium
Previous Article in Special Issue
Industrial Machinery Components Classification: A Case of D-S Pooling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Data-Driven Control Based on Information Concentration Estimator and Regularized Online Sequential Extreme Learning Machine

1
School of Vehicle and Mobility, Tsinghua University, Haidian District, Beijing 100084, China
2
School of Automation, Beijing Institute of Technology, Beijing 100084, China
3
National Key Lab of Autonomous Intelligent Unmanned Systems, Beijing Institute of Technology, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Symmetry 2024, 16(1), 88; https://doi.org/10.3390/sym16010088
Submission received: 10 February 2023 / Revised: 9 May 2023 / Accepted: 10 May 2023 / Published: 10 January 2024
(This article belongs to the Special Issue Artificial Intelligence, Adaptation and Symmetry/Asymmetry)

Abstract

:
Due to the complexity of digital equipment and systems, it is quite difficult to obtain a precise mechanism model in practice. For an unknown discrete-time nonlinear system, in this paper, a semi-parametric model is used to describe this discrete-time nonlinear system, and this semi-parametric model contains a parametric uncertainty part and a nonparametric uncertainty part. Based on this semi-parametric model, a novel data-driven control algorithm based on an information concentration estimator and regularized online sequential extreme learning machine (ReOS-ELM) is designed. The information concentration estimator estimates the parametric uncertainty part; The training data of ReOS-ELM network is obtained, based on symmetry and information concentration estimator, then the training of ReOS-ELM network and the estimate of nonparametric uncertainty part using ReOS-ELM network are carried out online, successively. A stability analysis and three simulation examples were performed, and the simulation results show that the proposed data-driven control algorithm is effective in improving the control accuracy.

1. Introduction

Data-driven control (DDC) algorithms rely on the input data and output data of control systems, and are very suitable for some nonlinear systems whose mechanism models cannot be established. After several years of development, some data-driven control techniques have been investigated, such as model free adaptive control (MFAC) [1,2,3,4,5], iterative learning control (ILC) [6,7,8,9], iterative feedback tuning (IFT) [10,11], unfalsified control (UC) [12,13], virtual reference feedback tuning (VRFT) [14], lazy learning control [15], some control algorithms based on neural networks [16,17,18,19], reinforcement learning [20,21,22,23,24], and so on. These DDC methods can be divided into two fundamental categories. The first one is that the controller structure is assumed to be known, then the controller design issue is transformed into a parametric synthesis, such as VRFT and IFT; the second one considers a system as an almost black-box model, it approaches the corresponding relationship between the input and output of systems as almost completely independent of the specific structure and parameters of systems, such as neural networks (NN) and MFAC.
For some nonlinear systems there is prior information about the control systems, these can be treated as linear systems to a certain extent, but this prior information is not enough for the first kind of DDC algorithm. NN do not consider prior information about control systems, but it is very difficult to explain their theory. In [25], a novel class of DDC algorithm was developed successfully for a simple first-order discrete-time nonlinear system. It is different from the previous two kinds of DDC algorithm, and divides a nonlinear system into a parametric uncertainty part and a nonparametric uncertainty part. The nonparametric uncertainty part was eliminated by the nearest-neighbor method. In [26], an extended version of an information concentration (IC) estimator was discussed for a second-order discrete-time nonlinear system. This IC estimator is effective when the parametric uncertainty part contains two unknown parameters. In [27], two independent IC estimators were adopted to estimate two unknown parameters of a speed control system of a robot arm. However, the nonparametric uncertainty part is assumed to be bounded, and this assumption is generally considered as a strict assumption. Furthermore, the stability of discrete-time nonlinear systems based on the extended version of the IC estimator was not guaranteed by theoretical analysis. In this paper, a data-driven control algorithm based on an IC estimator and regularized online sequential extreme learning machine (ReOS-ELM) is investigated for a discrete-time nonlinear system. Firstly, an improved IC estimator, which does not need the assumption that the nonparametric uncertainty part is bounded, is introduced. Secondly, a binary updating algorithm of ReOS-ELM is introduced, for stability analysis of the control system. Thirdly, the stability of the control system based on the proposed control algorithm is guaranteed by theoretical analysis. Finally, three simulation examples are performed. Compared with the existing results, there are two contributions. The first contribution is that this method does not need the assumption that the nonparametric uncertainty part is bounded. The second contribution is that the control system stability based on the proposed data-driven control algorithm based on the IC estimator is guaranteed by theoretical analysis.
This paper is structured as follows. The related knowledge is briefly introduced in Section 2. In Section 3, the proposed data-driven control algorithm based on an IC estimator and ReOS-ELM, including the parameter tuning for the data model as well as the stability issue, are stated. Simulation examples are stated in Section 4. Some conclusions are shown in Section 5.

2. Related Knowledge

2.1. Semi-Parametric Model

Modeling accurate mechanism models of actual systems is difficult due to their complex dynamic characteristics. Inspired by the research on the maximum capability and limitation of adaptive feedback mechanisms, some scholars designed a semi-parametric model. A semi-parametric model is different from a mechanism model; it is established depending on the historical data of a system and is used to describe discrete-time nonlinear systems. Semi-parametric models contain a parametric uncertainty part and a nonparametric uncertainty part. A semi-parametric model is described as
y k + 1 = ϕ k T θ + u k + g ( ψ k ) + τ k + 1
where θ and g ( · ) : R d 2 R denote an unknown parameter vector and an unknown vector-valued function, respectively, and τ k denotes random noise. Here, ϕ k T θ and u k + g ( ψ k ) are considered as parametric uncertainty part and nonparametric uncertainty part, respectively. From Equation (1), it is derived that y k + 1 ϕ k T θ τ k + 1 and nonparametric uncertainty part are symmetrical.
The regression vectors ϕ k and ψ k , are composed of historical data,
ϕ k = y k , y k 1 , , y k p 1 + 1 , u k 1 , , u k q 1 T
and
ψ k = y k , y k 1 , , y k p 2 + 1 , u k 1 , , u k q 2
Here, d 1 and d 2 are the dimensions of the regression vectors ϕ k and ψ k ,
d 1 = q 1 + p 1
and
d 2 = q 2 + p 2

2.2. Information Concentration Estimator

The classic self tuning regulator based on least squares (LS) can be applied to autoregressive exogenous systems, while the recursive least-squares algorithm can estimate unknown parameters of linear systems. The performance of traditional real-valued parameter identification on nonlinear systems which contain large nonparametric uncertainty is not very outstanding. Different from traditional real-valued parameter identification methods, the IC estimator regards prior knowledge at each step as some constraint on unknown parameters or functions, and reduces parameter or function uncertainties based on the prior knowledge step by step by using geometry knowledge. Its basic idea, which is depicted in Figure 1, is to test each vertex of v k to see whether to keep the original vertex or to generate a new vertex. At time k, the center point of the polygon domain I k is generally taken as θ ^ k . When the straight line intersects the polygon v, and splits it into two sub-polygons, the n k of the new polygon v may become large or small. Figure 1b shows that n k will become large.

2.3. Regularized Online Sequential Extreme Learning Machine

The online sequential extreme learning machine (OS-ELM) [28] is an online learning algorithm, and it often is used to train single-hidden-layer feed forward neural (SLFN) networks online. ReOS-ELM [29] is an improved version of OS-ELM, and it adds a regular term to the cost function.
For a single-hidden-layer neural network which is composed of n input nodes, L hidden nodes, and one output node, the jth output corresponding to the jth input pattern can be represented by
o j = i = 1 L β i G ( w i , b i , x j ) , j = 1 , , N 0
where x j and o j denote the jth input pattern and the jth output, respectively; b i is a bias of the active function G ( · ) ; and w i and β i are two weights which are used to connect the input node and the ith hidden node, and the ith hidden node and the output node, respectively.
For an initial training dataset with N 0 training patterns ( x j , t j ) R n × R , where x j = [ x j 1 , x j 1 , , x j n ] T , OS-ELM adopts Equation (7) as the cost function for the goal of determining the output weights in the training process.
H 0 β 0 T 0 2
The cost function (7) is sensitive to noise, and noise would lead to a poor generalization and over-fitting. In order to reduce the influence caused by noise, ReOS-ELM [30], which adds a regularization factor λ to the cost function for the purpose of controlling the trade-off between the output weight norm and the training error, is proposed based on Tikhonov regularization. The improved cost function is
H 0 β 0 T 0 2 + λ β 0 2
where
T 0 = t 1 , , t N 0 T
and
H 0 = G ( w 1 , b 1 , x 1 ) G ( w L , b L , x 1 ) G ( w 1 , b 1 , x N 0 ) G ( w L , b L , x N 0 ) N 0 × L
The solution of β 0 is
β 0 = P 0 H 0 T T 0
where
P 0 = ( H 0 T H 0 + λ I L ) 1
and I L is an identity matrix of size L (number of hidden nodes).
When the kth chunk of the dataset is received
N k = { ( x i , t i ) } i = j = 0 k 1 N j + 1 i = j = 0 k N j
ReOS-ELM takes a similar form to recursive least squares (RLS) to update the weight connecting the hidden node and output node, and the weight updating algorithm is as follows:
P k = P k 1 P k 1 H k T H k P k 1 I N k + H k P k 1 H k T
and
β k = β k 1 + P k H k T ( T k H k β k 1 )
where T k and H k indicate the target of the kth arriving training data and the hidden layer output for the kth arriving training data, respectively, and I N k is an identity matrix of size  N k .

2.4. Problem Formulation

In this paper, the following second-order nonlinear system is considered:
y k + 1 = ϑ 1 y k + ϑ 2 y k 1 + u k + f ( y k , y k 1 )
where y k and u k are the output signal and input signal of the system (16), respectively. ϑ 1 and ϑ 2 are two unknown parameters, f ( · ) is an unknown nonlinear function, and f ( · ) F ( L ) , which indicates the nonparametric uncertainty part of the system (16). Here, F ( L ) (the set F ( L ) will be defined later) represents a priori knowledge on possible f ( · ) .
In this paper, the following assumptions need to be guaranteed for making further study.
Assumption 1. 
The reference signal { y k * } is bounded, i.e.,
y k * B ω , k 0
where B ω is a positive constant.
Assumption 2. 
Each element of the unknown vector θ is bounded by two known constants B ¯ θ , j and B ̲ θ , j , i.e., for j = 1 , 2
B ̲ θ , j ϑ j B ¯ θ , j
Assumption 3. 
The unknown function f belongs to the following uncertainty set
F ( L ) = { f : | f ( x 1 , y 1 ) f ( x 0 , y 0 ) | L M ( x 1 , y 1 ) ( x 0 , y 0 ) + L c }
where L M < 3 2 + 2 .

3. The Design of Data-Driven Control Using the Proposed IC Estimator and ReOS-ELM

In [26], the nonparametric uncertainty part was supposed to be bounded. As time passes, n k may become very large, and so it takes a lot of time to calculate θ ^ k . In this paper, an improved IC estimator is introduced and used to estimate the unknown parameters of the parametric uncertainty part. The improved IC estimator adopts a constraint form which is from the constraint form in [26], but it consists of two steps of history data, which is different from the previous result, and the previous result needs step history data as well as the assumption that the nonparametric uncertainty part of the actual system is bounded. The improved IC estimator only needs to meet with Assumption 3. In addition, the problem that n k may be very large is considered, the number of vertexes is restricted to 10, and if n k > 10 , the case of Figure 1b is ignored. Equation (16) can be rewritten as
y k + 1 = ϑ ^ 1 , k y k + ϑ ^ 2 , k y k 1 + u k + f 1 ( y k , y k 1 )
where
f 1 ( y k , y k 1 ) = ϑ ˜ 1 , k y k + ϑ ˜ 2 , k y k 1 + f ( y k , y k 1 )
and ϑ ˜ 1 , k = ϑ 1 ϑ ^ 1 , k , ϑ ˜ 2 , k = ϑ 2 ϑ ^ 2 , k , and ϑ ^ 1 , k and ϑ ^ 2 , k are the estimated values of ϑ 1 and ϑ 2 by the IC estimator at time k. As time passes and more data is accumulated, more and more constraints on the true parameters ϑ 1 and ϑ 2 can be obtained, and the parameter or function uncertainties become smaller, and finally the values ϑ ^ 1 , k and ϑ ^ 2 , k become fixed values.
Assumption 4. 
For f 1 ( y k , y k 1 ) , suppose that there is β * , then f 1 ( y k , y k 1 ) can be rewritten as
f 1 ( y k , y k 1 ) = H k β * + k
where
H k β * = i = 1 L β i * G ( w i , b i , x k )
x k = y k , y k 1
and sup | k | , is a given upper bound, and w i , b i can be randomly generated and kept constant.
In this paper, the unknown parameters ϑ 1 and ϑ 2 are estimated by the IC estimator, and the unknown nonlinear term f 1 ( y k , y k 1 ) is approached by a single-hidden-layer NN. The controller is developed by
u k = u k l u k c
where u k l indicates the output of a controller which is used to stabilize linearized dynamics,
u k l = y k + 1 * ϑ ^ 1 , k y k ϑ ^ 2 , k y k 1 ,
and here u k c is the output of a single-hidden-layer NN.
u k c = H k β ^ k
where β ^ k is the estimated value of the ideal weight vector β * . The detailed design for the proposed improved IC estimator is shown in Section 3.1. The updating algorithm for the ideal weight vector connecting hidden nodes and output nodes adopts a binary form, and the detailed content of the binary form is introduced in Section 3.2.

3.1. The Design for the Proposed Improved IC Estimator

Define
J k y k u k 1
For the discrete-time nonlinear system (16) with a priori knowledge θ Θ R 2 , at the kth step ( k 0 ), with the data y k coming, the so-called information set I k is defined as
I k θ Θ : J k J k 1 ( ϕ k 1 T ϕ k 2 T ) θ F
where the initial set I 0 = Θ .
Combined with Assumption 2, Assumption 3, and (29), we have the following form of parametric uncertainty at time k
J k J k 1 L M ( y k 1 , y k 2 ) ( y k 2 , y k 3 ) L c ( ϕ k 1 T ϕ k 2 T ) θ ( ϕ k 1 T ϕ k 2 T ) θ J k J k 1 + L M ( y k 1 , y k 2 ) ( y k 2 , y k 3 ) + L c
Hence, we can obtain the specific constraint form I k about parameter vector θ , and it consists of two independent inequalities:
( ϕ k 1 T ϕ k 2 T ) θ c k , 1 , ( ϕ k 1 T ϕ k 2 T ) θ c k , 2
where
c k , 1 = J k J k 1 + L M ( y k 1 , y k 2 ) ( y k 2 , y k 3 ) + L c c k , 2 = ( J k J k 1 L M ( y k 1 , y k 2 ) ( y k 2 , y k 3 ) L c )
Then, the so-called concentrated information set C k at step k is
C k = j = 0 k I k = C k 1 I k
where C 0 = I 0 .
Generally, C k is a domain in R 2 , and the area of C k would shrink over time. The actual parameter vector θ may be located at anywhere in C k , and in the simulation experiment, the center point of C k is adopted as an estimated value of the parameter vector θ at the kth step.
The key problem of the IC estimator is to calculate the information set I k and the concentrated information set C k at each step. According to the basic idea of information concentration, these two constraints (32) are added one by one. i.e.,
v k 1 = AddLinearCons 2 D ( v k 1 , ( ϕ k 1 T ϕ k 2 T ) , c k , 1 ) , v k = AddLinearCons 2 D ( v k 1 , ( ϕ k 1 T ϕ k 2 T ) , c k , 2 )
where v k = { Q i , i = 1 , 2 , , n k } denotes the different vertexes of the domain C k , and n k denotes the number of vertexes. AddLinearCons 2 D is the core of the IC estimator. The pseudo code of AddLinearCons 2 D is shown in Algorithm 1.
Algorithm 1: AddLinearCons2D( v , ϕ , c ): Add linear constraint ϕ T θ c ( ϕ R 2 ) to a polygon v.
Input:  v : represented by clock-wise arranged vertexes Q 1 , Q 2 , , Q n
Output:  v : clock-wise arranged vertexes
1: Denote the number of vertexes in v by n
2: for  j = 1 to n do
3:    Denote the jth vertex by Q j
4:    Let F i = sign ( c ϕ Q j ) .
5: end for
6: if  F j 0 , 1 j n  then
7:     v v and return
8: else if  F j < 0 , 1 j n  then
9:     v and return
10: end if
11: for  j = 2 to n 1  do
12:    if ( Q j and Q j 1 are in different side) ∧ ( Q j and Q j + 1 are in different sides) then
14:        Q 1 i the intersection of l i and Q j 1 Q j ¯
15:        Q 2 i the intersection of l i and Q j + 1 Q j ¯
16:          if ( F j > 0 ) then
17:              v = { Q 1 i , Q j , Q 2 i }
18:          else
19:              v = { Q 1 , , Q j 1 , Q 1 i , Q 2 i , Q j + 1 , · , Q n }
20:          end if
21:       break
22:       else if ( Q j and Q j 1 are in different side) ∧ ( Q j , Q j + 1 , ,   Q j + a 1 are on
               the same side) ∧ ( Q j + a 1 , Q j + a are on different side) then
23:           Q 1 i the intersection of l i and Q j 1 Q j ¯
24:           Q 2 i the intersection of l i and Q j + a 1 Q j + a ¯
25:             if  F j > 0  then
26:                 v ={ Q 1 i , Q j , , Q j + a 1 , Q 2 i }
27:             else
28:                 v = { Q 1 , , Q j 1 , Q 1 i , Q 2 i Q j + a , , Q n }
29:             end if
30:       break
31:          else if ( Q j and Q j 1 are in different side) ∧ ( Q j 1 Q j 2 , , Q 1 are on the,
               same side) ∧ ( Q j , , Q n are the same side) then
32:              Q 1 i the intersection of l i and Q 1 Q n ¯
33:              Q 2 i the intersection of l i and Q j 1 Q j ¯
34:                   if  F j > 0  then
35:                       v ={ Q 1 i , Q j , , Q n , P 2 i }
36:                   else
37:                       v ={ Q 2 i , Q 1 , , Q j 1 , Q 1 i }
38:                   end if
39:          break
40:          end if
41:       end if
42:    end if
43: j j + 1
44: end for
46: if (size ( v ) > 10) then
47:     v = v
48: else
49:     v = v
50: end if
51: return  v

3.2. The Design of the Data-Driven Control Algorithm

The proposed data-driven control algorithm based on the IC estimator and ReOS-ELM network is performed in two main stages: the first stage is system initialization and the second stage is parameter learning. The process of the proposed data-driven control algorithm is shown in Figure 2.
Initialization of System
  • According to a priori knowledge on θ , we can obtain the initial polygon V 1 (usually a quadrangle) and the value of θ ^ 1 , and define V 3 V 2 V 1 , θ ^ 3 θ ^ 2 θ ^ 1 , set k m = 50 .
  • Set y 1 = 0 , and y 2 = u 1 ; u 1 , u 2 are two random numbers, | u 1 | 0.5 , and | u 2 | 0.5 ; set ε = 0.06 , δ 1 = 0.4 , δ 2 = 0.05 , and k = 3 .
Initialization of ReOS-ELM
  • For ReOS-ELM, define the hidden node number L, and the regularization factor λ .
  • Assign random parameters ( w i , b i ) for ReOS-ELM, where i = 1 , 2 , , L .
  • Measure the output y 3 of the plant (16).
  • The first sampling data can be obtained as x 1 = ( y 2 , y 1 ) and t 1 = e 3 = y 3 ϑ ^ 1 , 1 y 2 ϑ ^ 1 , 2 y 1 u 2 .
  • Obtain P 1 and β 1 ^ using the following equations
    P 1 = ( λ I L + H 1 T H 1 ) 1
    and
    β ^ 1 = P 1 H 1 T e 3
    Define β ^ 3 β ^ 2 β ^ 1 , and P 3 P 2 P 1 P 0 .
  • Set F l a g = 0 , which means that the relearning process has not been activated.
Whether or not the IC estimator updates
  • Calculate σ k and
    σ k = [ σ n k ( 1 ) , σ n k ( 2 ) ]
    where
    σ n k ( 1 ) = var ( X Q 1 , X Q 2 , , X Q n k ) σ n k ( 2 ) = var ( Y Q 1 , Y Q 2 , , Y Q n k )
    and X Q and Y Q denote the two coordinates of Q.
  • if σ n k ( 1 ) > δ 1 σ n k ( 2 ) > δ 1 , then execute ➅ or ➃.
Whether or not it is the first time
  • Calculate ξ k ( 1 ) and ξ k ( 2 ) . If k < 11 , execute
    ξ k ( 1 ) = var ( θ ^ 1 , k 10 , , θ ^ 1 , k ) ξ k ( 2 ) = var ( θ ^ 2 , k 10 , , θ ^ 2 , k )
    or ξ k ( 1 ) = 1 , ξ k ( 2 ) = 1 .
  • If ( σ n k ( 1 ) < δ 2 σ n k ( 2 ) < δ 2 ) ( ξ k ( 1 ) ε ξ k ( 2 ) ε k k m F l a g = 0 ) , then set F l a g = 1 .
  • if F l a g = 0 , execute ➄ or ➆.
Initialize ReOS-ELM and save the estimated value of the IC estimator
  • Relearning: using the data x k = y k , y k 1 and e k + 1 initialize ReOS-ELM, and obtain P k and β ^ k + 1 using the following equations
    P k = ( λ I L + H k T H k ) 1
    and
    β ^ k + 1 = P k H k T e k + 1
  • θ ^ k = θ ^ k 1 .
IC estimator update
  • Calculate c k , 1 and c k , 2 using (32), obtain the current polygon domain C k using (34), and take the center point of I k as the value of θ ^ k .
Load the estimated value of the IC estimator
  • θ ^ k = θ ^ k 1 .
Obtain control signal
  • Using the kth sample data x k = y k , y k 1 , calculate the output of the SLFN networks H k β ^ k , and
    u k c = H k β ^ k
  • Calculate the control signal u k
    u k = u k l u k c
  • Measure the output y k + 1 of the system (16).
The binary updating algorithm which is used for updating ReOS-ELM
  • Obtain the kth sample data x k = y k , y k 1 and e k + 1 ,
    e k + 1 = y k + 1 y k + 1 *
  • In this paper, an amended updating of the formulas in [30] is adopted, and if e k + 1 2 > 1 + H k P k 1 H k T 2 2 , execute (45) and (46) or (47)
    P k 1 = P k 1 1 + H k T H k
    β ^ k + 1 = β ^ k + P k H k T e k + 1
    and
    P k = P k 1 , β ^ k + 1 = β ^ k
End or not
  • If k < N , set k k + 1 , and execute ➉, or the proposed algorithm ends.

3.3. Stability Analysis

Define
γ k 1 1 + H k P k 1 H k T
Combining (46), Equation (49) can be obtained, and
β ˜ k + 1 = β ˜ k γ k P k 1 H k T e k + 1
where β ˜ k + 1 = β * β ^ k + 1 .
Introducing the Lyapunov candidate function as
V k + 1 = β ˜ k + 1 T P k 1 β ˜ k + 1
When e k + 1 2 1 + H k P k 1 H k T 2 2 , we can obtain that
V k + 1 V k 0
When e k + 1 2 > 1 + H k P k 1 H k T 2 2 , Equation (52) can be obtained from (20), (22), (50), and (49), and
H k β ˜ k + 1 + k = H k β ˜ k + 1 + f 1 ( y k , y k 1 ) H k β * = H k β ˜ k H k γ k P k 1 H k T e k + 1 + y k + 1 ϑ ^ 1 , k y k ϑ ^ 2 , k y k 1 u k l H k β * + u k c = y k + 1 ϑ ^ 1 , k y k ϑ ^ 2 , k y k 1 u k l H k γ k P k 1 H k T e k + 1 = e k + 1 H k P k 1 H k T e k + 1 1 + H k P k 1 H k T = γ k e k + 1
Hence, we can obtain
V k + 1 = β ˜ k + 1 T P k 1 β ˜ k + 1 = β ˜ k + 1 T P k 1 1 β ˜ k + 1 + β ˜ k + 1 T H k T H k β ˜ k + 1 = [ β ˜ k γ k P k 1 H k T e k + 1 ] T P k 1 1 [ β ˜ k γ k P k 1 H k T e k + 1 ] + β ˜ k + 1 T H k T H k β ˜ k + 1 = V k 2 γ k H k β ˜ k e k + 1 + γ k 2 e k + 1 2 H k P k 1 H k T + ( H k β ˜ k + 1 ) 2
Substituting (52) into the above equation, then
V k + 1 = V k 2 γ k H k β ˜ k e k + 1 + γ k 2 e k + 1 2 H k P k 1 H k T + ( H k β ˜ k + 1 ) 2 = V k 2 H k β ˜ k ( H k β ˜ k + 1 + k ) + H k P k 1 H k T ( H k β ˜ k + 1 + k ) 2 + ( H k β ˜ k + 1 ) 2 = V k 2 H k ( β ˜ k + 1 + γ k P k 1 H k T e k + 1 ) ( H k β ˜ k + 1 + k ) + H k P k 1 H k T ( H k β ˜ k + 1 + k ) 2 + ( H k β ˜ k + 1 ) 2 = V k ( H k β ˜ k + 1 ) 2 2 H k β ˜ k + 1 k H k P k 1 H k T ( H k β ˜ k + 1 + k ) 2 V k ( H k β ˜ k + 1 ) 2 2 H k β ˜ k + 1 k = V k + k 2 ( H k β ˜ k + 1 + k ) 2
Then,
V k + 1 V k + k 2 ( H k β ˜ k + 1 + k ) 2 = V k + k 2 γ k 2 e k + 1 2 = V k ( γ k 2 e k + 1 2 k 2 ) V k ( γ k 2 e k + 1 2 2 )
V k + 1 V k ( γ k 2 e k + 1 2 2 ) 0
Based on (51) and (56), if enough data is obtained, the system tracking error will infinitely approach zero.

4. Analysis of Experimental Results

In this paper, a second-order unknown discrete-time nonlinear system is considered, and three simulation examples are performed. In order to indicate the effectiveness of the IC estimator, in the first simulation example, f ( y k , y k 1 ) = 0 ; in the second simulation example, the proposed data-driven algorithm based on the IC estimator and ReOS-ELM is compared with the MFAC algorithm based on RLS ([31], Equation 5.49); in the third example, the performance of ReOS-ELM when the training data contains noise is compared with ReOS-ELM when the training data does not contain noise, by approaching a nonlinear function.

4.1. The First Simulation Example

In this first example, the parameters of the system (16) are
ϑ 1 = 1.75 , ϑ 2 = 1.25 , f ( y k , y k 1 ) = 0
In this example, the a priori knowledge is listed as follows:
L M = 1.5 , L c = 0.5
and
B ̲ θ , 1 = 10 , B ¯ θ , 1 = 10 B ̲ θ , 2 = 10 , B ¯ θ , 2 = 15
The desired reference signal y k + 1 * is
y k + 1 * = 2 sin ( 2 π k T s ) + 1.5 cos ( π k T s )
For ReOS-ELM, these parameters are: the number of hidden nodes of ReOS-ELM is L = 25 , λ = 0.001 , the activation function of ReOS-ELM is sigmoid, T s = 0.01 , and = 0.15 . The main results, which include the estimated values of θ by the improved IC estimator and the tracking error, are depicted in Figure 3 and Figure 4. The curves in Figure 3a,b show the change trends and and estimated values of ϑ 1 and ϑ 2 , respectively. The tracking curves and tracking error are shown in Figure 4a and Figure 4b, respectively. The estimated values shown in Figure 3 show that the values of θ estimated by the improved IC effectively approach the actual value, and the small tracking error in Figure 4b also indicates its effective performance in estimating the unknown parameters.

4.2. The Second Simulation Example

The proposed data-driven algorithm based on the IC estimator and ReOS-ELM is a novel kind of DDC method. In order to verify its performance, it is compared with the controller of MFAC based on RLS ([31], Equation 5.49) for a second-order unknown discrete-time nonlinear system which contains a parametric uncertainty part and a nonparametric uncertainty part. In this simulation example,
f ( y k , y k 1 , u k ) = 0.5 y k sin ( log e 10 y k 1 + 1 ) + 0.3 y k 1 sin ( log e 10 y k + 1 )
In order to check the learning performance of ReOS-ELM in the proposed data-driven control algorithm, the following cost function is introduced
H k β k T k 2 + λ β k 2
Figure 5, Figure 6, Figure 7 and Figure 8 are the second simulation’s example results. Figure 5 plots the change curves of the estimated values ϑ ^ 1 , k and ϑ ^ 2 , k . In Figure 6, the curves denote the simulation experimental results of the control algorithm based on the proposed IC estimator and ReOS-ELM, and Figure 6a and Figure 6b show the tracking curves and error curve, respectively. The curve in Figure 7 indicates the change trend of the cost function, Equation (62), which diminishes through time. In Figure 8, the curves denote the simulation experimental results of MFAC based on RLS, and the curves in Figure 8a and Figure 8b show the output signal and error signal, respectively. The change trends of Figure 6b and Figure 8b show that when this system reaches a stable state, the maximum error in the control algorithm based on the proposed IC estimator and ReOS-ELM is 0.0152, the maximum error in MFAC based on RLS is 0.0596. This result indicates that the performance of the proposed algorithm is better than the performance of the MFAC based on RLS. The change curves in Figure 5 show that the final estimated values ϑ ^ 1 , k and ϑ ^ 2 , k have some degree of deviation from the values ϑ 1 and ϑ 2 , and it indicates that the semi-parametric model is established depending on the historical data and is different from mechanism models.

4.3. The Third Simulation Example

In order to analyze the influence caused by noise in the training data, two estimated errors of Equation (61) were compared, one is the estimated error by ReOS-ELM in which the training data do not contain noise, and the other is the estimated error by ReOS-ELM in which the training data contain random noise whose range is [−0.2, 0.2]. The errors are shown in Figure 9, which indicates that the noise in the training data would make the estimated error become large, but would not change the performance of approaching Equation (61).

5. Conclusions

In this paper, first, an improved IC estimator is introduced. This improved IC estimator adopts a constraint form which consists of two steps of history data. This is different from the previous results, and the previous results need one step history data. This improvement makes this estimator fit in more systems, and those systems do not need to meet the assumption that the nonparametric uncertainty part of actual systems is bounded. Second, a data-driven control algorithm based on an improved IC estimator and ReOS-ELM is designed. For stability analysis of control systems, a binary updating algorithm of ReOS-ELM is introduced, and the stability of the proposed data-driven control algorithm is guaranteed by a theoretical analysis based on a Lyapunov candidate function. Finally, in order to indicate the effectiveness of the IC estimator, an ideal second-order unknown discrete-time linear system is considered, and the experimental results show that the proposed IC estimator is effective in estimating the unknown parameters of a second-order unknown linear system. Furthermore, the proposed data-driven algorithm is compared with the MFAC based on RLS ([31], Equation 5.49) for the system (16), and the second simulation’s results show that the maximum tracking error in the control algorithm based on the proposed IC estimator and ReOS-ELM is smaller than the maximum tracking error in MFAC based on RLS, which indicates that the proposed data-driven algorithm is effective at improving the control accuracy.
The contribution of this paper is that a novel data-driven control algorithm is introduced, and the stability of this algorithm is guaranteed by theoretical analysis. A semi-parametric model divides a system into a parametric uncertainty part and a nonparametric uncertainty part, and this is different from previous methods, which consider the whole system almost as a black-box model. A semi-parametric model only needs a neural network to estimate the nonparametric uncertainty part rather than the whole system, and this may contribute to reducing the network complexity and calculation workload. However, if the parametric uncertainty part and nonparametric uncertainty part of a semi-parametric model are simultaneously unknown, it is challenging to make use of this model. The proposed data-driven control algorithm can estimate parametric uncertainty part, firstly. Then training data of ReOS-ELM network is obtained, based on symmetry, and the training of ReOS-ELM network and the estimate of nonparametric uncertainty part using ReOS-ELM network are carried out online, successively. The proposed data-driven control algorithm provides a possibility of making use of this model. This algorithm may be used in second-order systems, such as motor control systems. The disadvantage of the proposed data-driven control algorithm is that it is only suitable for second-order systems. The improved IC estimator adopts a polygon to indicate the possible domain of the unknown parameters of the parametric uncertainty part, and this polygon can only indicate those data which belong to it. Because of the nature of the polygon, the improved IC estimator is not suitable for higher-order systems. In future research, the IC estimator may be improved by changing the way in which it is used in the possible domain of the unknown parameters of the parametric uncertainty part, for example, for third-order systems a cube may be adopted to indicate those data which belong to it. For higher-order systems, it is a challenging thing to indicate the possible domain of unknown parameters by using geometry knowledge. However, with the developments of computer technology and parallel computing, there may be some solutions by estimating the unknown parameters of the parametric uncertainty part in parallel.

Author Contributions

Methodology, X.Z.; investigation, H.Z.; writing—original draft preparation, X.Z. and H.M.; writing—review and editing, X.Z.; funding acquisition, H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially funded by the National Key Research and Development Plan of China (No. 2018AAA0101000) and the National Natural Science Foundation of China under grant 62076028.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

No potential conflict of interest was reported by the authors.

References

  1. Hou, Z.; Jin, S. A novel data-driven control approach for a class of discrete-time nonlinear systems. IEEE Trans. Control Sytems Technol. 2011, 19, 1549–1558. [Google Scholar] [CrossRef]
  2. Liao, Y.; Du, T.; Jiang, Q. Model-free adaptive control method with variable forgetting factor for unmanned surface vehicle control. Appl. Ocean. Res. 2019, 93, 101945. [Google Scholar] [CrossRef]
  3. Hou, Z.; Zhu, Y. Controller-Dynamic-Linearization-Based Model Free Adaptive Control for Discrete-Time Nonlinear Systems. IEEE Trans. Ind. Inform. 2013, 9, 2301–2309. [Google Scholar] [CrossRef]
  4. Aghaei Hashjin, S.; Pang, S.; Ait-Abderrahim, K.; Nahid-Mobarakeh, B. Data-Driven Model-Free Adaptive Current Control of a Wound Rotor Synchronous Machine Drive System. IEEE Trans. Transp. Electrif. 2020, 6, 1146–1156. [Google Scholar] [CrossRef]
  5. Zhang, X.; Ma, H. Data-Driven Model-Free Adaptive Control Based on Error Minimized Regularized Online Sequential Extreme Learning Machine. Energies 2019, 12, 3241. [Google Scholar] [CrossRef]
  6. Yu, M.; Zhou, W.; Liu, B. On iterative learning control for MIMO nonlinear systems in the presence of time-iteration-varying parameters. Nonlinear Dyn. 2017, 89, 2561–2571. [Google Scholar] [CrossRef]
  7. He, W.; Meng, T.; He, X.; Ge, S.S. Unified iterative learning control for flexible structures with input constraints. Automatica 2018, 96, 326–336. [Google Scholar] [CrossRef]
  8. Mandra, S.; Galkowski, K.; Rogers, E.; Rauh, A.; Aschemann, H. Performance-enhanced robust iterative learning control with experimental application to PMSM position tracking. IEEE Trans. Control Syst. Technol. 2019, 27, 1813–1819. [Google Scholar] [CrossRef]
  9. Choi, W.; Volpe, F.A. Simultaneous iterative learning control of mode entrainment and error field. Nucl. Fusion 2019, 59, 056011. [Google Scholar] [CrossRef]
  10. Hjalmarsson, H. Iterative feedback tuning—An overview. Int. J. Adapt. Control Signal Process. 2002, 16, 373–395. [Google Scholar] [CrossRef]
  11. Heertjes, M.F.; Vander Velden, B.; Oomen, T. Constrained iterative feedback tuning for robust control of a wafer stage system. IEEE Trans. Control Syst. Technol. 2016, 24, 56–66. [Google Scholar] [CrossRef]
  12. Sanchez Pena, R.S.; Colmegna, P.; Bianchi, F. Unfalsified control based on the controller parameterisation. Int. J. Syst. Sci. 2015, 46, 2820–2831. [Google Scholar] [CrossRef]
  13. Safonov, M.; Tsao, T. The unfalsified control concept and learning. IEEE Trans. Autom. Control 1997, 42, 843–847. [Google Scholar] [CrossRef]
  14. Jeng, J.C.; Lin, Y.W. Data-driven nonlinear control design using virtual-reference feedback tuning based on the block-oriented modeling of nonlinear systems. Ind. Eng. Chem. Res. 2018, 57, 7583–7599. [Google Scholar] [CrossRef]
  15. Kobayashi, M.; Konishi, Y.; Ishigaki, H. A lazy learning control method using support vector regression. Int. J. Innovtive Comput. Inf. Control 2007, 3, 1511–1523. [Google Scholar]
  16. Jia, C.; Li, X.; Wang, K.; Ding, D. Adaptive control of nonlinear system using online error minimum neural networks. Isa Trans. 2016, 65, 125–132. [Google Scholar] [CrossRef] [PubMed]
  17. Li, X.; Jia, C.; Liu, D.; Ding, D. Adaptive Control of Nonlinear Discrete-Time Systems by Using OS-ELM Neural Networks. Abstr. Appl. Anal. 2014, 2014, 1–11. [Google Scholar] [CrossRef]
  18. Rahmani, B.; Belkheiri, M. Adaptive neural network output feedback control for flexible multi-link robotic manipulators. Int. J. Control 2019, 92, 2324–2338. [Google Scholar] [CrossRef]
  19. Li, X.; Fang, J.; Li, H. Exponential stabilisation of memristive neural networks under intermittent output feedback control. Int. J. Control 2018, 91, 1848–1860. [Google Scholar] [CrossRef]
  20. Chen, C.; Modares, H.; Xie, K.; Lewis, F.L.; Wan, Y.; Xie, S. Reinforcement Learning-based Adaptive Optimal Exponential Tracking Control of Linear Systems with Unknown Dynamics. IEEE Trans. Autom. Control 2024. to be published. [Google Scholar] [CrossRef]
  21. Nguyen, N.D.; Nguyen, T.; Nahavandi, S. Multi-agent behavioral control system using deep reinforcement learning. Neurocomputing 2019, 359, 58–68. [Google Scholar] [CrossRef]
  22. Wang, Z.; Li, H.; Wang, J.; Shen, F. Deep reinforcement learning based conflict detection and resolution in air traffic control. IET Intell. Transp. Syst. 2019, 13, 1041–1047. [Google Scholar] [CrossRef]
  23. Khalatbarisoltani, A.; Soleymani, M.; Khodadadi, M. Online control of an active seismic system via reinforcement learning. Struct. Control Health Monit. 2019, 26, e2298. [Google Scholar] [CrossRef]
  24. Yang, X.; Liu, D.; Wang, D. Reinforcement learning for adaptive optimal control of unknown continuous-time nonlinear systems with input constraints. Int. J. Control 2014, 87, 553–566. [Google Scholar] [CrossRef]
  25. Ma, H.; Lum, K.Y.; Ge, S.S. Adaptive control for a discrete-time first-order nonlinear system with both parametric and non-parametric uncertainties. In Proceedings of the IEEE Conference on Decision & Control, New Orleans, LA, USA, 12–14 December 2007. [Google Scholar]
  26. Zhou, H.; Ma, H.; Li, N.; Yang, C. Semi-parametric adaptive control of discrete-time systems using extreme learning machine. In Proceedings of the 2017 9th International Conference on Modelling, Identification and Control (ICMIC 2017), Piscataway, NJ, USA, 26–28 July 2017. [Google Scholar]
  27. Zhang, X.; Ma, H.; Luo, M.; Liu, X. Adaptive sliding mode control with information concentration estimator for a robot arm. Int. J. Syst. Sci. 2020, 51, 217–228. [Google Scholar] [CrossRef]
  28. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  29. Liang, N.Y.; Huang, G.B.; Saratchandran, P.; Sundararajan, N. A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans. Neural Netw. 2006, 17, 1411–1423. [Google Scholar] [CrossRef]
  30. Huynh, H.T.; Won, Y. Regularized online sequential learning algorithm for single-hidden layer feedforward neural networks. Pattern Recognit. Lett. 2011, 32, 1930–1935. [Google Scholar] [CrossRef]
  31. Hou, Z. Nonparametric Model and Adaptive Control Theory, 1st ed.; Science Press of China: Beijing, China, 1999. [Google Scholar]
Figure 1. The process of the IC estimator at time k, the polygon v indicates the possible candidate regions of unknown parameters at the previous moment. The straight line comes from the information at step k; it splits v into two sub-polygons, the left-hand one of which will become the new polygon v . Four typical situations of IC estimator are shown in subfigures (a), (b), (c) and (d), respectively.
Figure 1. The process of the IC estimator at time k, the polygon v indicates the possible candidate regions of unknown parameters at the previous moment. The straight line comes from the information at step k; it splits v into two sub-polygons, the left-hand one of which will become the new polygon v . Four typical situations of IC estimator are shown in subfigures (a), (b), (c) and (d), respectively.
Symmetry 16 00088 g001
Figure 2. The process of the proposed data-driven control algorithm. The green area indicates the initialization stage and the yellow area indicates the parameter learning stage.
Figure 2. The process of the proposed data-driven control algorithm. The green area indicates the initialization stage and the yellow area indicates the parameter learning stage.
Symmetry 16 00088 g002
Figure 3. The estimated values of θ by the improved IC estimator in the first simulation example. ϑ 1 and ϑ 2 are two actual values of θ , and the estimated curves ϑ ^ 1 , k and ϑ ^ 2 , k are shown in (a) and (b), respectively.
Figure 3. The estimated values of θ by the improved IC estimator in the first simulation example. ϑ 1 and ϑ 2 are two actual values of θ , and the estimated curves ϑ ^ 1 , k and ϑ ^ 2 , k are shown in (a) and (b), respectively.
Symmetry 16 00088 g003
Figure 4. The performance of the control system in the first simulation example. (a) and (b) show the tracking curves and error curve, respectively.
Figure 4. The performance of the control system in the first simulation example. (a) and (b) show the tracking curves and error curve, respectively.
Symmetry 16 00088 g004
Figure 5. The estimated values of θ by the improved IC estimator in the second simulation. ϑ 1 and ϑ 2 are two actual values of θ . The estimated curves ϑ ^ 1 , k and ϑ ^ 2 , k are shown in (a) and (b), respectively.
Figure 5. The estimated values of θ by the improved IC estimator in the second simulation. ϑ 1 and ϑ 2 are two actual values of θ . The estimated curves ϑ ^ 1 , k and ϑ ^ 2 , k are shown in (a) and (b), respectively.
Symmetry 16 00088 g005
Figure 6. The results of the proposed novel data-driven control algorithm based on the IC estimator and ReOS-ELM in the second simulation example. (a) and (b) show the tracking curves and error curve, respectively.
Figure 6. The results of the proposed novel data-driven control algorithm based on the IC estimator and ReOS-ELM in the second simulation example. (a) and (b) show the tracking curves and error curve, respectively.
Symmetry 16 00088 g006
Figure 7. Cost function change curve.
Figure 7. Cost function change curve.
Symmetry 16 00088 g007
Figure 8. The results of the MFAC algorithm based on RLS in the second simulation example. (a) and (b) show the tracking curves and error curve, respectively.
Figure 8. The results of the MFAC algorithm based on RLS in the second simulation example. (a) and (b) show the tracking curves and error curve, respectively.
Symmetry 16 00088 g008
Figure 9. Two estimated errors of Equation (61). E y 1 , k indicates the estimated error by ReOS-ELM in which the training data does not contain noise, and E y 2 , k indicates the estimated error by ReOS-ELM in which the training data contains random noise whose range is [−0.2, 0.2].
Figure 9. Two estimated errors of Equation (61). E y 1 , k indicates the estimated error by ReOS-ELM in which the training data does not contain noise, and E y 2 , k indicates the estimated error by ReOS-ELM in which the training data contains random noise whose range is [−0.2, 0.2].
Symmetry 16 00088 g009
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Ma, H.; Zhang, H. Data-Driven Control Based on Information Concentration Estimator and Regularized Online Sequential Extreme Learning Machine. Symmetry 2024, 16, 88. https://doi.org/10.3390/sym16010088

AMA Style

Zhang X, Ma H, Zhang H. Data-Driven Control Based on Information Concentration Estimator and Regularized Online Sequential Extreme Learning Machine. Symmetry. 2024; 16(1):88. https://doi.org/10.3390/sym16010088

Chicago/Turabian Style

Zhang, Xiaofei, Hongbin Ma, and Huaqing Zhang. 2024. "Data-Driven Control Based on Information Concentration Estimator and Regularized Online Sequential Extreme Learning Machine" Symmetry 16, no. 1: 88. https://doi.org/10.3390/sym16010088

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop