Next Article in Journal
Extraction of Partial Discharge Pulses from the Complex Noisy Signals of Power Cables Based on CEEMDAN and Wavelet Packet
Previous Article in Journal
Parameter Analysis and Optimization of Class-E Power Amplifier Used in Wireless Power Transfer System
Previous Article in Special Issue
Combination of Data-Driven Active Disturbance Rejection and Takagi-Sugeno Fuzzy Control with Experimental Validation on Tower Crane Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Data-Driven Model-Free Adaptive Control Based on Error Minimized Regularized Online Sequential Extreme Learning Machine

1
School of Automation, Beijing Institute of Technology, Beijing 100081, China
2
State Key Laboratory of Intelligent Control and Decision of Complex Systems, Beijing Institute of Technology, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Energies 2019, 12(17), 3241; https://doi.org/10.3390/en12173241
Submission received: 18 July 2019 / Revised: 6 August 2019 / Accepted: 7 August 2019 / Published: 22 August 2019
(This article belongs to the Special Issue Energy Efficiency and Data-Driven Control)

Abstract

:
Model-free adaptive control (MFAC) builds a virtual equivalent dynamic linearized model by using a dynamic linearization technique. The virtual equivalent dynamic linearized model contains some time-varying parameters, time-varying parameters usually include high nonlinearity implicitly, and the performance will degrade if the nonlinearity of these time-varying parameters is high. In this paper, first, a novel learning algorithm named error minimized regularized online sequential extreme learning machine (EMREOS-ELM) is investigated. Second, EMREOS-ELM is used to estimate those time-varying parameters, a model-free adaptive control method based on EMREOS-ELM is introduced for single-input single-output unknown discrete-time nonlinear systems, and the stability of the proposed algorithm is guaranteed by theoretical analysis. Finally, the proposed algorithm is compared with five other control algorithms for an unknown discrete-time nonlinear system, and simulation results show that the proposed algorithm can improve the performance of control systems.

1. Introduction

It is difficult to obtain an accurate mechanism model of a physical system when the production technologies and processes are very complex. Data-driven control (DDC) [1] relies on input/output (I/O) data of control systems and does not need to consider mechanism models of systems. After several years of development, some data-driven control techniques have been investigated, such as, proportional-integral derivative (PID) [2], fuzzy logic control [3], unfalsified control (UC) [4,5], model free adaptive control (MFAC) [6,7,8,9,10,11,12], iterative learning control (ILC) [13,14,15], iterative feedback tuning (IFT) [16,17], some control algorithms based on neural network [18,19,20,21,22,23] and so on.
MFAC is a class of DDC, and it builds a virtual equivalent dynamic linearized model [24] by using a dynamic linearization technique. The virtual equivalent dynamic linearized model contains some time-varying parameters. In practice, it is often not easy to obtain the resulting time-varying parameters, but they can be estimated by using historical data of control systems. Those time-varying parameters usually include high nonlinearity implicitly, and the performance will degrade if the nonlinearity of the resulting time-varying parameters is too severe [25]. The traditional methods of getting those time-varying parameters are projection algorithm, recursive least squares, and so on. Besides every time-varying parameter in the virtual equivalent dynamic linearized model can be considered as a nonlinear function, and in [25], radial basis function neural network (RBFNN) is used to estimate those time-varying parameters.
Extreme learning machine (ELM) is developed for offline learning and training a single-hidden-layer feed forward neural networks (SLFNs) [26]. Online sequential extreme learning machine (OS-ELM) [27] is an online learning algorithm, it adjusts the output weights online, besides the input weights and hidden biases are randomly chose. In recent years OS-ELM has gained a large amount of interests and been uesed to estimate unknown parameters of systems [28,29]. Some improved OS-ELM algorithms were introduced by some scholars, such as regularized online sequential extreme learning machine (REOS-ELM) [30], initial-training-free online extreme learning machine (ITF-OELM) [31], etc. Although, the learning speed of OS-ELM is extremely high, it could yield singular and ill-posed problens. To overcome adverse effects, which cuased by the noisy data of control systems, REOS-ELM, which is an improvement of OS-ELM, was investigated. In [30], REOS-ELM was introduced, however the stability of REOS-ELM for unknown discrete-time nonlinear systems was not analysed, and how to select the hidden node number of neural networks is unknown.
In this paper, first, a updating formula which contains dead-zone characteristics for REOS-ELM is introduced. Second, in order to obtain a more compact network structure, error minimized regularized online sequential extreme learning machine (EMREOS-ELM) is investigated, EMREOS-ELM is different from EMOS-ELM [28], and it updates the output weights using the pseudoinverse of a partitioned matrix during the growth of the networks. Finally, a novel model-free adaptive control method based on EMREOS-ELM is proposed. This paper is structured as follows, in Section 2 REOS-ELM is briefly introduced. Dynamic linearization technique and the updating formula which contains dead-zone characteristics for REOS-ELM are introduced in Section 3. The model-free adaptive control method based on EMREOS-ELM and the stability analysis of the proposed method for unknown discrete-time nonlinear systems are stated in Section 4. Section 5 shows simulation experiments. In Section 6 some conclusions are brought.

2. REOS-ELM

In [32], the universal approximation capability of ELM was analyzed. ELM is developed for offline learning, however training data could arrive one-by-one or chunk-by-chunk; OS-ELM is an online version of ELM, and it is sensitive to noise data; REOS-ELM is an improvement of OS-ELM. In this part REOS-ELM will be introduced.
Suppose that there is an initial training dataset with N 0 training patterns ( x j , t j ) R n × R where x j = [ x j 1 , x j 1 , , x j n ] T . For a single hidden layer neural network with n input nodes and one output node, the jth output corresponding to the jth input pattern is
o j = i = 1 L β i G ( w i , b i , x j ) , j = 1 , , N 0
where L denotes the number of hidden nodes, and the subscript i indicates the ith hidden node; G ( · ) and o j are the active function of hidden nodes and the output of the network, respectively, b i and β i are the hidden bias of the ith hidden node and the output weight connecting the ith hidden layer and the output node, respectively.
OS-ELM is a novel online learning algorithm due to its highlighted features, and its highlighted features are that the input weights and hidden biases are randomly chose, and the main goal of training process is to determine the output weights. In order to obtain output weights, they minimize the error function defined by
H 0 β 0 T 0 2 .
The collected data for training process often contains noise, and the only objective of empirical risk minimization as shown in (2) may lead to poor generalization and over-fitting [30]. REOS-ELM is different from OS-ELM, and REOS-ELM tries to minimize the empirical error and obtain small norm of network weight vector [30]. In [30] the followin cost function was considered, and the cost function is
H 0 β 0 T 0 2 + λ β 0 2
where λ is a regularization factor.
The solution of β 0 is
β 0 = P 0 H 0 T T 0 ,
and
P 0 = ( H 0 T H 0 + λ I L ) 1
here I L is an identity matrix, the size of I L is L, T 0 = t 1 , , t N 0 T , and
H 0 = G ( w 1 , b 1 , x 1 ) G ( w L , b L , x 1 ) G ( w 1 , b 1 , x N 0 ) G ( w L , b L , x N 0 ) N 0 × L .
When the kth chunk of data is recerived
N k = { ( x i , t i ) } i = j = 0 k 1 N j + 1 i = j = 0 k N j
The weight updating algorithm, which is used in REOS-ELM, takes a similar form to recursive least squares (RLS) algorithm, and
P k = P k 1 P k 1 H k T H k P k 1 I N k + H k P k 1 H k T ,
and
β k = β k 1 + P k H k T ( T k H k β k 1 )
where I N k is an identity matrix, and the size of I N k is N k ; T k and H k are the target of kth arriving training data and the hidden layer outputs for the kth arriving training data, respectively.

3. Dynamic Linearization Technique and the New Updating Formula for REOS-ELM

3.1. Dynamic Linearization Technique

MFAC builds a virtual equivalent dynamic linearized model by using a dynamic linearization technique.
The following unknown discrete-time nonlinear system is considered
y k + 1 = f ( y k , , y k L y , u k , , u k L u )
where f ( · ) represents an unknown nonlinear function, and here L y and L u indicate the orders of the system input and the system output, respectively.
To make further study, the following assumptions are used.
Assumption A1.
The system (10) is observable and controllable.
Assumption A2.
The f ( · ) is a smooth nonlinear function, and the partial derivative of f ( · ) with respect to u k is continuous.
Assumption A3.
Suppose that for all k 0 , 1 , , T , and if u k 0 , then system (10) satisfies the generalized Lipschitz condition along the iteration axis, that is
| y k + 1 | L b | u k | ,
where y k + 1 = y k + 1 y k , u k = u k u k 1 , L b is a constant, and L b > 0 .
Remark 1.
This assumption means that bounded change of the input can not cause unbounded change of the system output, which can be guaranteed by many industrial processes [25].
Theorem 1.
For the unknown discrete-time nonlinear system (10), satisfying Assumptions (Remarks 1–3), there must be ψ k called pseudo-partial-derivative (PPD), when u k 0 , the system (10) can be rewritten as
y k + 1 = ψ k u k
where | ψ k | L b .
Proof. 
See ([33] theorem 4.1). □
Assumption A4.
For the unknown discrete-time nonlinear system (10), satisfying Assumptions (Remarks 1–3), there is β * , when u k 0 , the system (10) can be rewritten as
y k + 1 = y k + H k β * u k + k u k
where
H k β * = i = 1 L β i * G ( w i , b i , x k ) ,
sup | k | , and ▵ is a given upper bound; w i , b i are randomly generated, then they kept in constant.

3.2. The Updating Formula Based on Dead-Zone Characteristics

The discrete time nonlinear system (10) is a class of special system where the training data comes one by one, and N k = 1 .
When H k β ^ k 0 , and | y k + 1 * y k H k β ^ k | M , the control law is
u k = y k + 1 * y k H k β ^ k + u k 1
where M is a constant, M > 0 , y k + 1 * is the desired signal and β ^ k is defined in below.
When H k β ^ k = 0 , or | y k + 1 * y k H k β ^ k | > M , or u k = 0 , a new re-learning process begins.
The updating formula, which contains dead-zone characteristics, for REOS-ELM is
P k 1 = P k 1 1 + H k T H k σ k ,
and
β ^ k + 1 = β ^ k + σ k P k H k T e k + 1 * ,
where
σ k = 1 ( | e k + 1 * | 1 + σ k H k P k 1 H k T ) 2 > 2 0 otherwise
and
e k + 1 * = y k + 1 y k + 1 * u k .
Lemma 1.
If
P k 1 = P k 1 1 + H k T H k σ k
where the scalar σ k > 0 , then P k is related to P k 1 via
P k = P k 1 σ k P k 1 H k T H k P k 1 1 + σ k H k P k 1 H k T ,
and
P k H k T = P k 1 H k T 1 + σ k H k P k 1 H k T .
Proof. 
It is a direct conclusion of [33]. □

4. The MFAC Method Based on EMREOS-ELM

In this section, we will introduce the MFAC method based on EMREOS-ELM, and the MFAC method based on EMREOS-ELM is performed in three main phases, and they are parameter initialization, parameter learning and the adjustment of network structure. EMREOS-ELM is an online learning algorithm and it can train a single hidden layer neural network, besides it can adjust the network structure; When the system is initialized or the network structure is adjusted, EMREOS-ELM need much training data, which is used for the training network structure. In this paper, the size of those training data is 200, than is, K m = 200 I m = 200 ; The maximum number of hidden nodes and the minimum number of hidden nodes are usually selected based on the complexity of unknown discrete-time nonlinear systems. Figure 1 illustrates the process of the proposed algorithm.

4.1. Initialization Phase

  • u 1 and u 2 are two random values, | u 1 | 1 | u 2 | 1 , and set k = 3 .
  • Set M = 100 , L m = 20 , and L 1 = L 2 = 8 . L m denotes the maximum number of hidden nodes, and L 1 and L 2 denote the minimum number of hidden nodes.
  • Measure the output y 2 and y 3 of system (10).
  • Assign random parameters of hidden nodes ( w i , b i ) where i = 1 , , L 2 , λ = 0 . 1 .
  • Using the first sample data x 1 = y 2 , u 1 , initialize P 1 and β ^ 2 , and
    P 0 = ( λ I L 2 ) 1 ,
    P 1 = ( λ I L 2 + H 1 T H 1 ) 1 ,
    and
    β ^ 1 = P 1 H 1 T e 3 * .
    Define
    β ^ 2 β ^ 1 .

4.2. Parameter Learning

  • Using the kth sample data x k = y k , u k 1 , calculate H k , and
    when H k β ^ k 0 , and | y k + 1 * y k H k β ^ k | M , then
    u k = y k + 1 * y k H k β ^ k + u k 1 .
    when H k β ^ k = 0 , or | y k + 1 * y k H k β ^ k | > M or | u k | = 0 , a new re-learning process begins.
  • Measure the output y k + 1 of system (10).
  • Set E m = 0 . 05 , E v = 50 , and calculate E k
    E k = i = k E V k 1 ( y * ( i ) y ( i ) ) 2 .
    where E V and E m are two values, and they are important for adjusting network structure; The appropriate E V and E m can improve the tracking effect of systems, and too small E V or too big E m make EMREOS-ELM become invalid; We can choose those values E V and E m based on performance requirements of systems.
  • When the tracking error does not meet the requirements of systems, network structure will be adjusted. The core of the proposed EMREOS-ELM is the adjustment of network structure. if E k E m L k L m I I m k K m then execute II; or I.
    (I)
    Using the kth training data x k = y k , u k 1 and e k + 1 * , update the output weights
    P k = P k 1 σ k P k 1 H k T H k P k 1 1 + σ k H k P k 1 H k T
    and
    β ^ k + 1 = β ^ k + σ k P k H k T e k + 1 * ,
    where
    σ k = 1 ( | e k + 1 * | 1 + σ k H k P k 1 H k T ) 2 > 2 0 otherwise ,
    and
    e k + 1 * = y k + 1 y k + 1 * u k .
  • k k + 1 , L k = L k 1 , I I + 1 , and go to Section 4.2.
    (II)

4.3. The Adjustment of Network Structure

  • Set L k = L k 1 + 1 , and assign random parameters of the L k th hidden node ( w L k , b L k ) .
  • When network structure is adjusted, it is equivalent to add a new column to H k , and
    δ = G ( w L k , b L k , x k ) .
    Then
    H k * = H k | δ .
  • The pseudo inverse of the new ( H k * ) is
    ( H k * ) = ( H k ) + d b b
    where
    d = ( H k ) + δ
    and
    b = c 1 if c 0 ( 1 + d T d ) 1 d T ( H k ) if c = 0 ,
    and
    c = δ H k d .
  • Calculate β ^ k + 1 , and
    β ^ k + 1 = β ^ k d b δ b δ .
  • Initialize P k and
    P k = λ × I L k
    where I L k denotes L k × L k identify matrix.
  • Set u k as a random number, and | u k | 1 .
  • Set k k + 1 , I = 1 , and go to Section 4.2.

4.4. Stability Analysis

For EMREOS-ELM, the adjustment of network structure and online learning are independent of each other, when the structure of the network is adjusted, online learning is not executed in this period. If the structure, i.e., the hidden node number is settled to a fixed value, EMREOS-ELM will be changed into REOS-ELM. Thus, the stability analysis of the model-free adaptive control method based on EMREOS-ELM can be similarly established the stability analysis of the model-free adaptive control method based on REOS-ELM. In this section, the stability of the model-free adaptive control method based on REOS-ELM is analyzed.
Theorem 2.
For the system (10), if the updating formulas described by (16) and (17) are adopted, the following results can be obtained.
(1)
β * β ^ k 2 ι 1 β * β ^ 1 2 where ι 1 = λ max ( P 0 1 ) λ min ( P 0 1 ) , and λ max ( P 0 1 ) and λ min ( P 0 1 ) are the maximum eigenvalue and the minimum eigenvalue of the matrix P 0 1 , respectively.
(2)
lim N k = 1 N σ k [ ( | e k + 1 * | 1 + σ k H k P k 1 H k T ) 2 2 ] <
and this implies
(a)
lim k σ k [ ( | e k + 1 * | 1 + σ k H k P k 1 H k T ) 2 2 ] = 0
(b)
lim sup k ( | e k + 1 * | 2 ( 1 + σ k H k P k 1 H k T ) 2 ) 2
Proof. 
Define
γ k 1 1 + σ k H k P k 1 H k T
Equation (42) can be obtained from (22) and (17), and
β ˜ k + 1 = β ˜ k σ k γ k P k 1 H k T e k + 1 *
where β ˜ k + 1 = β * β ^ k + 1 . Equation (43) can be obtained from Assumption A4 and (42), and
H k β ˜ k + 1 + k = H k β ˜ k + 1 + y k + 1 y k u k H k β * = H k β ˜ k H k σ k γ k P k 1 H k T e k + 1 * + y k + 1 y k u k H k β * = y k + 1 y k u k H k β ^ k H k σ k γ k P k 1 H k T e k + 1 * = e k + 1 * H k σ k P k 1 H k T e k + 1 * 1 + σ k H k P k 1 H k T = γ k e k + 1 *
Introducing Lyapunov candidate as V k + 1 = β ˜ k + 1 T P k 1 β ˜ k + 1 , taking (20) into it, hence, we can obtain
V k + 1 = β ˜ k + 1 T P k 1 β ˜ k + 1 = β ˜ k + 1 T P k 1 1 β ˜ k + 1 + σ k β ˜ k + 1 T H k T H k β ˜ k + 1 = [ β ˜ k σ k γ k P k 1 H k T e k + 1 * ] T P k 1 1 [ β ˜ k σ k γ k P k 1 H k T e k + 1 * ] + σ k β ˜ k + 1 T H k T H k β ˜ k + 1 = V k 2 σ k γ k H k β ˜ k e k + 1 * + σ k 2 γ k 2 ( e k + 1 * ) 2 H k P k 1 H k T + σ k ( H k β ˜ k + 1 ) 2
Substitute (43) in the above equation then
V k + 1 = V k 2 σ k γ k H k β ˜ k e k + 1 * + σ k 2 γ k 2 ( e k + 1 * ) 2 H k P k 1 H k T + σ k ( H k β ˜ k + 1 ) 2 = V k 2 σ k H k β ˜ k ( H k β ˜ k + 1 + k ) + σ k 2 H k P k 1 H k T ( H k β ˜ k + 1 + k ) 2 + σ k ( H k β ˜ k + 1 ) 2 = V k 2 σ k H k ( β ˜ k + 1 + σ k γ k P k 1 H k T e k + 1 * ) ( H k β ˜ k + 1 + k ) + σ k 2 H k P k 1 H k T ( H k β ˜ k + 1 + k ) 2 + σ k ( H k β ˜ k + 1 ) 2 = V k σ k ( H k β ˜ k + 1 ) 2 2 σ k H k β ˜ k + 1 k σ k 2 H k P k 1 H k T ( H k β ˜ k + 1 + k ) 2 V k σ k ( H k β ˜ k + 1 ) 2 2 σ k H k β ˜ k + 1 k = V k + σ k k 2 σ k ( H k β ˜ k + 1 + k ) 2
then
V k + 1 V k + σ k k 2 σ k ( H k β ˜ k + 1 + k ) 2 = V k + σ k k 2 σ k γ k 2 ( e k + 1 * ) 2 = V k σ k ( γ k 2 ( e k + 1 * ) 2 k 2 ) V k σ k ( γ k 2 ( e k + 1 * ) 2 2 )
V k + 1 V k σ k ( γ k 2 ( e k + 1 * ) 2 2 ) 0 ,
and V k is nonnegative.
Then
β ˜ k + 1 T P k 1 β ˜ k + 1 β ˜ k T P k 1 1 β ˜ k β ˜ 1 T P 0 1 β ˜ 1 .
The following inequalities can be obtained from the matrix inversion lemma,
λ min ( P k 1 ) λ min ( P k 1 1 ) λ min ( P 0 1 ) ,
and this implies
λ min ( P 0 1 ) β ˜ k + 1 2 λ min ( P k 1 ) β ˜ k + 1 2 β ˜ k + 1 T P k 1 β ˜ k + 1 β ˜ 1 T P 0 1 β ˜ 1 λ max ( P 0 1 ) β ˜ 1 .
This establishes part (1).
Summing both sides of (47) from 1 to N with V k being nonnegative, we have
V k + 1 V 1 lim N k = 1 N σ k ( γ k 2 ( e k + 1 * ) 2 2 ) ,
and we immediately get
lim N k = 1 N σ k [ ( e k + 1 * ) 2 ( 1 + σ k H k P k 1 H k T ) 2 2 ]
(2) holds, then
lim k σ k [ ( e k + 1 * ) 2 ( 1 + σ k H k P k 1 H k T ) 2 2 ] = 0
(a) and (b) hold. □
We also immediately have
lim sup k [ e k + 1 * ( 1 + σ k H k P k 1 H k T ) ] 2 2 .
Because the activation function G ( w i , b i , x j ) is bounded, e k + 1 * is bounded. Because u k = | y k + 1 * y k H k β ^ k | M , e k + 1 = y k + 1 y k + 1 * is bounded.

5. Analysis of Experimental Results

In this paper, EMREOS-ELM is investigated, and EMREOS-ELM is adopted for estimating the value ψ k ; EMREOS-ELM is an online learning algorithm and can train a single hidden layer neural network; besides it can adjust the network structure. In order to further show the effectiveness of the MFAC method based on EMREOS-ELM, the MFAC method based on another online learning algorithm is compared, and the online learning algorithm adjusts network structure using an incremental form. In order to describe this online learning algorithm easily, in this paper, this online learning algorithm is named as IREOS-ELM. The MFAC method based on IREOS-ELM also has three phases, parameter initialization, parameter learning and the adjustment of network structure. During the adjustment of network structure, all input weights and all hidden biases are reinitialized, and the network is retrained by using the k 1 th training data. The MFAC method based on EMREOS-ELM is different from the MFAC method based on IREOS-ELM, it only initializes the input weight of the increased hidden node and the hidden bias of the increased hidden node, and it updates the output weights using the pseudoinverse of a partitioned matrix.
In this part, a simulation experiment is did, besides the MFAC algorithm based on EMREOS-ELM is compared with the MFAC algorithm based on RLS ([33] Equation 5.49), the MFAC method based on IREOS-ELM and RBF algorithm ([34] Equation 10.22).
In this numerical simulation experiment, the following single-input single-output unknown discrete-time nonlinear system is considered, and
y k + 1 = y k 1 + y k 2 + u k 3 .
The discrete-time nonlinear system (55) comes from [35].
The reference signal is
y k + 1 * = 4 + 2 sin ( π T s k ) + 1.5 cos ( π T s k )
where T s is the sampling time, and T s = 0.01 .
The parameters of MFAC algorithm based on RLS are ρ k = 1 , λ 1 = 0.01 , and a 0 = 0.95 . The activation function of EMREOS-ELM is sigmoid, λ = 0.1 , and = 0.002 . Figure 2, Figure 3, Figure 4, Figure 5 and Figure 6 show the main results of this numerical simulation experiment. Figure 2 plots the tracking curves, and in Figure 3, the curve denotes the tracking error change trend. Figure 4 shows the change trends of E k and L k . Figure 4 indicates that the final value of L k is 9, when L k = 9 , the tracking error satisfies E k E m , and the result of Figure 4 shows that EMEROS-ELM is effective in obtaining a compact network structure. Figure 5 indicates the change curve of ψ ^ k , ψ ^ k = H k β ^ k , and ψ ^ k is the estimated value of ψ k . In Section 4.4, the stability analysis of the MFAC algorithm based on EMREOS-ELM for unknown discrete-time nonlinear was stated, and we get the inequality (47). In order to verify the inequality (47), a new variable as
Δ v k σ k ( γ k 2 ( e k + 1 * ) 2 2 ) .
Figure 6 shows the change curve of of Δ v k , and it indicates that Δ v k 0 .
Besides the error change curves of e k ( 1 ) , e k ( 2 ) , e k ( 3 ) and e k ( 4 ) are plotted in Figure 7, Figure 8, Figure 9 and Figure 10, respectively. Here e k ( 1 ) indicates the error change curve of the REOS-ELM which has a fixed network structure, and the number of hidden nodes is 8; e k ( 2 ) indicates the error change curve of the REOS-ELM which has a fixed network structure, and the number of hidden nodes is 20; e k ( 3 ) is the error change curve of the MFAC algorithm based on RLS; and e k ( 4 ) indicates the error change curve of RBF which has a fixed network structure, and the number of hidden nodes is 9. Figure 7, Figure 8, Figure 9 and Figure 10 indicate that the tracking effect of EMREOS-ELM is better than four other algorithms.
In order to further show the effectiveness of algorithms, the integral square error (ISE) index of predicted output is introduced, and
e ISE = l = 5000 10000 ( y l * y l ) 2 .
Twenty groups of simulation experiments were done, and the ISE of those six control algorithms are shown in Table 1. The final hidden node numbers of simulation experiments are showed in Table 2, and Table 2 shows that the MFAC method based on EMREOS-ELM is effective in adjusting network structure. The average values of ISE indicate that compared with five other control algorithms, the proposed algorithm has an improvement in the performance of control systems.

6. Conclusions

In order to analyse the stability of the MFAC algorithm based on REOS-ELM for unknown discrete-time nonlinear systems, an updating formula, which contains dead-zone characteristics, is introduced, and EMREOS-EM is investigated for the purpose of getting a more compact network structure and improving the performances of control systems. The proposed MFAC method based on EMREOS-ELM is compared with five other algorithms, and the simulation results show that the proposed algorithm has an improvement in the performance of control systems.
In this paper, the MFAC method based on EMREOS-ELM is introduced for single-input single-output unknown discrete-time nonlinear systems. In the future, we plan to do some work for multiple-input, multiple-output unknown discrete-time nonlinear systems based on the proposed algorithm, and in order to improve the performance of robot control systems, we will also apply this algorithm to robot control systems which are complex nonlinear systems.

Author Contributions

Conceptualization and Methodology, X.Z. and H.M.; Writing-draft version, X.Z.; Writing—review editing, X.Z. and H.M.

Funding

This work is partially supported by Ministry of Science and Technology of the People’s Republic of China under Grant No. 2017YFF0205306. This work is also partially supported by National natural Science Foundation of China under Grant No. 91648117.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MFACmodel free adaptive control
I/Oinput/output
DDCdata-driven control
PIDproportional-integral derivative
UCunfalsified control
RBFNNradial basis function neural network
EMREOS-ELMerror minimized regularized online sequential extreme learning machine
ELMextreme learning machine
OS-ELMonline sequential extreme learning machine
REOS-ELMregularized online sequential extreme learning machine
EMOS-ELMerror minimized online sequential extreme learning machine

References

  1. Hou, Z.S.; Xu, J.X. On data-driven control theory: The state of the art and perspective. Acta Autom. Sin. 2009, 35, 650–667. [Google Scholar] [CrossRef]
  2. Caponetto, R.; Dongola, G.; Fortuna, L.; Gallo, A. New results on the synthesis of FO-PID controllers. Commun. Nonlinear Sci. 2010, 15, 997–1007. [Google Scholar] [CrossRef]
  3. Roman, R.C.; Precup, R.E.; David, R.C. Second order intelligent proportional-integral fuzzy control of twin rotor aerodynamic systems. Procedia Comput. Sci. 2018, 139, 372–380. [Google Scholar] [CrossRef]
  4. Sanchez-Pena, R.S.; Colmegna, P.; Bianchi, F. Unfalsified control based on the controller parameterisation. Int. J. Syst. Sci. 2015, 46, 2820–2831. [Google Scholar] [CrossRef]
  5. Safonov, M.G.; Tsao, T.C. The unfalsfied control concept and learning. IEEE Trans. Autom. Control 1997, 42, 2819–2824. [Google Scholar] [CrossRef]
  6. Hou, Z.S.; Huang, W.H. The model-free learning adaptive control of a class of SISO nonlinear systems. In Proceedings of the 1997 American Control Conference, Albuquerque, NM, USA, 6 June 1997; Volume 1, pp. 343–344. [Google Scholar]
  7. Hou, Z.S.; Jin, S.T. A novel data-driven control approach for a class of discrete-time nonlinear systems. IEEE Trans. Control Syst. Technol. 2011, 19, 1549–1558. [Google Scholar] [CrossRef]
  8. Tutsoy, O.; Barkana, D.E.; Tugal, H. Design of a completely model free adaptive control in the presence of parametric, non-parametric uncertainties and random control signal delay. ISA Trans. 2018, 76, 67–77. [Google Scholar] [CrossRef]
  9. Safaei, A.; Mahyuddin, M.N. Adaptive model-free control based on an ultra-local model with model-free parameter estimations for a generic SISO system. IEEE Access 2018, 6, 4266–4275. [Google Scholar] [CrossRef]
  10. Hung-Yi, C.; Jin-Wei, L. Model-free adaptive sensing and control for a piezoelectrically actuated system. Sensors 2010, 10, 10545–10559. [Google Scholar]
  11. Xia, Y.; Dai, Y.; Yan, W.; Xu, D.; Yang, C. Adaptive-observer-based data driven voltage control in islanded-mode of distributed energy resource systems. Energies 2018, 11, 3299. [Google Scholar] [CrossRef]
  12. Hou, Z.S.; Xiong, S.S. On model free adaptive control and its stability analysis. IEEE Trans. Autom. Control 2019, in press. [Google Scholar] [CrossRef]
  13. He, W.; Meng, T.T.; He, X.Y.; Samge, S.Z. Unified iterative learning control for flexible structures with input constraints. Automatica 2018, 96, 326–336. [Google Scholar] [CrossRef]
  14. Xuan, Y.; Ruan, X. Reinforced gradient-type iterative learning control for discrete linear time-invariant systems with parameters uncertainties and external noises. IMA J. Math. Control Inf. 2016, 34, 1117–1133. [Google Scholar]
  15. Miao, Y.; Wei, Z.; Liu, B.B. On iterative learning control for MIMO nonlinear systems in the presence of time-iteration-varying parameters. Nonlinear Dynam. 2017, 89, 2561–2571. [Google Scholar]
  16. Hjalmarsson, H. Iterative feedback tuning—An overview. Int. J. Adapt. Control Signal Process. 2010, 16, 373–395. [Google Scholar] [CrossRef]
  17. Heertjes, M.F.; Velden, B.V.D.; Oomen, T. Constrained iterative feedback tuning for robust control of a wafer stage system. IEEE Trans. Control Syst. Technol. 2016, 24, 56–66. [Google Scholar] [CrossRef]
  18. RiosPatron, E.; Braatz, R. On the identification and control of dynamical systems using neural networks. IEEE Trans. Neural Netw. 1997, 8, 452. [Google Scholar]
  19. Ren, X.M.; Lewis, F.L.; Zhang, J.L. Neural network compensation control for mechanical systems with disturbances. Automatic 2009, 45, 1221–1226. [Google Scholar] [CrossRef]
  20. Wang, F.; Chen, B.; Lin, C.; Zhang, J.; Meng, X. Adaptive neural network finite-time output feedback control of quantized nonlinear systems. IEEE Trans. Cybern. 2018, 48, 1839–1848. [Google Scholar] [CrossRef]
  21. Sun, C.Y.; He, W.; Ge, W.L.; Chang, C. Adaptive neural network control of biped robots. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 315–326. [Google Scholar] [CrossRef]
  22. Wang, M.S.; Tsai, T.M. Sliding mode and neural network control of sensorless PMSM controlled system for power consumption and performance improvement. Energies 2017, 10, 1780. [Google Scholar] [CrossRef]
  23. Faria, J.; Pombo, J.; Calado, M.D.R.; Mariano, S. Power management control strategy based on artificial neural networks for standalone PV applications with a hybrid energy storage system. Energies 2019, 12, 902. [Google Scholar] [CrossRef]
  24. Hou, Z.S.; Liu, S.D.; Tian, T.T. Lazy-learning-based data-driven model-free adaptive predictive control for a class of discrete-time nonlinear systems. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 1–15. [Google Scholar] [CrossRef] [PubMed]
  25. Zhu, Y.M.; Hou, Z.S. Data-driven MFAC for a class of discrete-time nonlinear systems with RBFNN. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 1013–1020. [Google Scholar] [PubMed]
  26. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  27. Liang, N.Y.; Huang, G.B.; Saratchandran, P.; Sundararajan, N. A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans. Neural Netw. 2006, 17, 1411–1423. [Google Scholar] [CrossRef] [PubMed]
  28. Jia, C.; Li, X.L.; Wang, K.; Ding, D.W. Adaptive control of nonlinear system using online error minimum neural networks. ISA Trans. 2016, 65, 125–132. [Google Scholar] [CrossRef] [PubMed]
  29. Li, X.L.; Jia, C.; Liu, D.X.; Ding, D.W. Adaptive control of nonlinear discrete-time systems by using OS-ELM neural networks. Abstr. Appl. Anal. 2014, 2014, 1–11. [Google Scholar] [CrossRef]
  30. Huynh, H.T.; Won, Y. Regularized online sequential learning algorithm for single-hidden layer feedforward neural networks. Pattern Recognit. Lett. 2011, 32, 1930–1935. [Google Scholar] [CrossRef]
  31. Gao, X.H.; Wong, K.I.; Wong, P.K.; Chi, M.V. Adaptive control of rapidly time-varying discrete-time system using initial-training-free online extreme learning machine. Neurocomputing 2016, 194, 117–125. [Google Scholar] [CrossRef]
  32. Huang, G.B.; Chen, L.; Siew, C.K. Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans. Neural Netw. 2006, 17, 879–892. [Google Scholar] [CrossRef] [PubMed]
  33. Hong, Z.S. NonParametric Model and Adaptive Control Theory; Science Press: Beijing, China, 1999. [Google Scholar]
  34. Kun, L.J. RBF Neural Network Adaptive Control Matlab Simulation; Tsinghua University Pre: Beijing, China, 2014. [Google Scholar]
  35. Narendra, K.S. Identification and control for dynamic systems using neural networks using neural network. IEEE Trans. Inform. Theory 1990, 1, 4–27. [Google Scholar]
Figure 1. The process of the proposed MFAC method based on EMREOS-ELM.
Figure 1. The process of the proposed MFAC method based on EMREOS-ELM.
Energies 12 03241 g001
Figure 2. The tracking curves of simulation experiment.
Figure 2. The tracking curves of simulation experiment.
Energies 12 03241 g002
Figure 3. The changing trend of e k .
Figure 3. The changing trend of e k .
Energies 12 03241 g003
Figure 4. The changing trends of E k and L k .
Figure 4. The changing trends of E k and L k .
Energies 12 03241 g004
Figure 5. The changing trend of ψ ^ k .
Figure 5. The changing trend of ψ ^ k .
Energies 12 03241 g005
Figure 6. The changing trend of Δ v k .
Figure 6. The changing trend of Δ v k .
Energies 12 03241 g006
Figure 7. The changing trend of e k ( 1 ) .
Figure 7. The changing trend of e k ( 1 ) .
Energies 12 03241 g007
Figure 8. The changing trend of e k ( 2 ) .
Figure 8. The changing trend of e k ( 2 ) .
Energies 12 03241 g008
Figure 9. The changing trend of e k ( 3 ) .
Figure 9. The changing trend of e k ( 3 ) .
Energies 12 03241 g009
Figure 10. The changing trend of e k ( 4 ) .
Figure 10. The changing trend of e k ( 4 ) .
Energies 12 03241 g010
Table 1. The simulation result e ISE .
Table 1. The simulation result e ISE .
The MFAC
Algorithm
Based on
REOS-ELM(L = 8)
The MFAC
Algorithm
Based on
REOS-ELM(L = 20)
The MFAC
Algorithm
Based on
EMREOS-ELM
RBFNN
(L = 9)
The MFAC
Algorithm
Based
on RLS
The MFAC
Algorithm
Based on
IREOS-ELM
10.1506720.0025150.0015620.5500534.4499640.012889
20.0194590.0409140.0017050.5024262.5341290.433985
30.7032820.0149890.0054510.5643612.2744060.016331
40.1726100.1507500.0272220.5049681.9091200.017123
50.0831490.2440610.0048060.4028551.8242170.007621
60.6968410.0091120.0064750.4118572.5483980.004997
71.5680390.0512200.0147300.5919082.2222980.182836
80.3229190.0240960.002950.5847211.8395720.009986
90.0516060.5763310.002950.5248701.7998290.048388
100.1582930.6786170.0035590.6961582.2314480.001695
110.0552510.0044480.0142550.5516502.4099470.015570
120.0087330.0041310.0190710.4930701.9133700.046355
130.3769440.0081840.0046490.5531472.9466020.012967
140.2974520.014590.0028630.6336311.9063570.054175
150.1320940.3296500.0021160.7305322.4569590.016178
160.0998350.1943840.0369170.7799083.4463120.017376
170.2229070.0020910.0114950.4313271.9524560.049845
180.0212880.0026290.0057040.5617052.0455760.049984
190.4814430.0045170.0250930.5822502.3803520.007150
200.0186250.2126890.0017240.6649682.0121020.016439
average value0.2820720.1284960.0097290.5658182.3548290.051094
Table 2. The final number of hidden nodes.
Table 2. The final number of hidden nodes.
1234567891011121314151617181920
The MFAC algorithm
based on EMREOS-ELM
9910111014991010101091110111111109
The MFAC algorithm
based on IREOS-ELM
101110912111111101012109111014119910

Share and Cite

MDPI and ACS Style

Zhang, X.; Ma, H. Data-Driven Model-Free Adaptive Control Based on Error Minimized Regularized Online Sequential Extreme Learning Machine. Energies 2019, 12, 3241. https://doi.org/10.3390/en12173241

AMA Style

Zhang X, Ma H. Data-Driven Model-Free Adaptive Control Based on Error Minimized Regularized Online Sequential Extreme Learning Machine. Energies. 2019; 12(17):3241. https://doi.org/10.3390/en12173241

Chicago/Turabian Style

Zhang, Xiaofei, and Hongbin Ma. 2019. "Data-Driven Model-Free Adaptive Control Based on Error Minimized Regularized Online Sequential Extreme Learning Machine" Energies 12, no. 17: 3241. https://doi.org/10.3390/en12173241

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop