Next Article in Journal
Optimization of Actuator Arrangement of Cable–Strut Tension Structures Based on Multi-Population Genetic Algorithm
Previous Article in Journal
Improving Hydrodynamics and Energy Efficiency of Bioreactor by Developed Dimpled Turbine Blade Geometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Iterative Learning Control with Forgetting Factor for MIMO Nonlinear Systems with Randomly Varying Iteration Lengths and Disturbances

1
College of Electrical Engineering, Henan University of Technology, Zhengzhou 450001, China
2
Information Engineering University, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(5), 694; https://doi.org/10.3390/sym17050694
Submission received: 1 April 2025 / Revised: 24 April 2025 / Accepted: 27 April 2025 / Published: 1 May 2025
(This article belongs to the Section Computer)

Abstract

In this paper, a PD-type iterative learning control algorithm with a forgetting factor is developed for MIMO nonlinear systems with randomly varying iteration lengths, initial state shifts and disturbances. Firstly, considering the randomly varying iteration lengths, a modified tracking error is designed. Secondly, for the initial state shift and disturbances, a PD-type iterative learning control algorithm with a forgetting factor (PDILCFF) method is proposed. A contraction mapping method is exploited to obtain the convergence property of the proposed control scheme, which can guarantee that the tracking error is bounded. Considering the iteration-varying trial lengths, the proposed PDILCFF algorithm is closely related to symmetry. Symmetry can provide prior information about the system’s structure and characteristics for the proposed PDILCFF method, which is helpful for designing more efficient control algorithms. On the other hand, the proposed PDILCFF method exploits system symmetries across different intervals through iterative processes to achieve accurate control and performance optimization of MIMO unknown nonlinear systems. Finally, two simulations are presented, one with a subway train tracking control system and the other with a two-degree-of-freedom robot manipulator system, to verify the effectiveness of the theoretical studies.

1. Introduction

Iterative learning control (ILC) [1,2,3,4], proposed by Arimoto, is an effective control for repetitive systems to achieve perfect tracking control. After 40 years of development, ILC has developed rapidly both in theory, including traditional ILC [5,6], adaptive ILC [7,8], data-driven ILC [9,10,11], point-to-point ILC [12,13,14], and iterative learning fault-tolerant control [15] and in practical applications, such as robot control [16,17], industrial systems [13,18], train systems [19,20], and transportation systems [21,22].
In most ILC methods, identical trial length is a basic assumption, meaning the operation time is the same for each iteration. However, due to the complexity and uncertainty of the actual system, this assumption does not hold. In other words, the duration of each experiment may vary in the iteration domain. For instance, based on foot strike times, the gait issue of humanoid robots [23] is divided into phases. In the learning process, the duration of each phase usually varies across iterations. Due to the influence of various uncertainties, disturbances and emergencies, trains [24] usually arrive earlier or later than their scheduled times in daily operation, which results in random changes in running length. ILC has been applied to functional electrical stimulation of upper limb movement and gait assistance [25]. When stroke patients walk on a treadmill relying on their own strength and ability, they usually shorten the number of steps by suddenly lowering their feet, resulting in different trial lengths. In [26], ILC was used for lab-scale gantry crane trajectory tracking, and the duration of each iteration ended when the constraint conditions were violated. These applications clearly demonstrate a violation of the standard experimental length assumption typically used in ILC.
In recent years, the problem of iteration-varying trial lengths for ILC has received a lot of attention. For linear time-invariant systems and nonlinear continuous-time systems, an iteration-average-operator-based ILC in [27,28] was proposed to address information loss caused by randomly varying iteration lengths. Based on the lifted method, a classic P-type ILC with almost sure and mean-square convergence was presented for a linear time-varying system with varying iteration lengths in [29]. In [30], for discrete-time linear systems with randomly varying trial lengths, two novel improved ILC methods with a searching mechanism were designed to improve learning speed. In [31], for nonlinear systems with iteration-varying trial lengths, a sampled-data ILC was investigated. All the references reviewed above are contraction-mapping-based ILC algorithms. In addition, composite-energy-function-based ILC schemes have been developed. In [32], a robust adaptive ILC method was presented for uncertain nonlinear systems with varying trial lengths. In [33], an adaptive ILC method with a one-link robotic manipulator simulation was proposed for nonlinear systems with randomly varying iteration lengths. In [34], for state-constrained nonlinear systems with randomly varying iteration lengths subjected to actuator faults, an adaptive iterative learning fault-tolerant control scheme was designed to achieve tracking error convergence. In [35], to address the problems of the modeling difficulties and varying iteration lengths of nonlinear systems, a model-free adaptive ILC was investigated. In [36], for discrete-time multi-agent systems with randomly varying trials, a model-free predictive iterative learning control method was also investigated.
All the works summarized above are based on the ILC method designed under the identical initial condition, but initial conditions are not the same in practice due to various factors. In [37], a P-type ILC with initial state shift error and randomly varying iteration lengths for nonlinear systems was presented, and strict convergence in the pointwise sense was guaranteed by using a modified supremum norm technique. In [38], a robust higher-order ILC was also investigated to address initial state shifts and varying iteration lengths for nonlinear discrete-time systems, which can insure that tracking errors are bounded in mathematical expectation. For the formation control of discrete-time multi-agent systems with iteration-varying initial conditions and nonuniform actual trial lengths, a P-type iterative learning protocol was proposed in [39]. In [40], a robust anti-windup (AW) controller was proposed for a nonlinear delayed system with exogenous perturbations. However, these ILC methods do not take into account the effect of disturbances on the system. In an ILC system, there are not only initial state shifts, but also various disturbances, such as state disturbances, measurement noise and input disturbances, which will affect the convergence of the system.
For randomly varying iteration lengths, existing ILC methods do not have the ability to address the issues of initial state shifts and disturbances. Inspired by these observations, finding a way to tackle these problems simultaneously in actual systems has important theoretical significance and practical value. The contributions of this paper are summarized as follows.
(1)
Randomly varying iteration lengths, initial state shifts and disturbances are dealt with simultaneously for MIMO nonlinear systems using the proposed method.
(2)
For randomly varying iteration lengths, a modified tracking error is designed. For the initial state shifts, state disturbances and measurement disturbances, a PD-type iterative learning control method with a forgetting factor, utilizing the modified tracking error, is proposed. The boundedness of errors is demonstrated using the contraction mapping method.
(3)
Two simulations, one with a subway train tracking control system and the other with a two-degree-of-freedom robot manipulator system are shown to verify the effectiveness of the theoretical studies.
The paper is organized as follows. Section 2 presents the problem formulation. The design and analysis of the control method for MIMO nonlinear systems with randomly varying iteration lengths, initial state shifts and disturbances are developed in Section 3. In Section 4, two simulations are used to verify the effectiveness of the proposed control algorithm. Section 5 concludes this paper.
In this paper, for notational convenience, the time argument t is omitted when no confusion arises.

2. Problem Formulation

Consider the following MIMO continuous-time nonlinear system.
x . i ( t ) = f ( x i ( t ) , t ) + B u i ( t ) + w i ( t ) y i ( t ) = C x i ( t ) + v i ( t )
where i is the iteration index,  t [ 0 , T i ]  is time, and  T i  is the actual operation length.  x i ( t ) R n u i ( t ) R r y i ( t ) R m B R n × r C R m × n w i ( t ) R n  represents state disturbances, and  v i ( t ) R m  represents measurement disturbances.  w i ( t ) b w v i ( t ) b v 1 v . i ( t ) b v 2 f ( , )  is globally Lipschitz with respect to x. That is,  f ( x 1 , t ) f ( x 2 , t ) k f x 1 ( t ) x 2 ( t ) , t [ 0 , T i ]  where  k f  is the bounded Lipschitz constant.
Assume that for the desired output trajectory  y d ( t ) t [ 0 , T r ]  and initial state  x d ( 0 ) , there exists a unique input  u d ( t )  such that
x . d ( t ) = f ( x d ( t ) , t ) + B u d ( t ) y d ( t ) = C x d ( t )
where  T r  is the desired operation length. The tracking error is defined as  e i ( t ) = y d ( t ) y i ( t ) .

3. Controller Design and Convergence Analysis

3.1. Randomly Varying Iteration Lengths

The primary difficulty in the design of the ILC system (1) is that the actual trial length is iteration varying, which is different from the desired operation length.
In practical applications, the actual operation length  T i  may be smaller or greater than the desired operation length  T r . Assume that the minimum iteration length is  T min  and the maximal iteration length is  T max T i [ T min , T max ]  is a stochastic variable, which denotes the actual length at the i-th iteration.
For the first case,  T i < t < T r , the system output data information is missing and cannot be utilized for control input updating. For the second case,  T r < t < T max , the system output data after the desired length are redundant and can be discarded because they have no use for learning control. In the second case, the full desired iteration length information is available. Hence, the second case can be considered a normal learning process.
Definition 1.
E { μ }  is the mathematical expectation of the stochastic variable  μ P [ T i t ]  is the occurrence probability of the event  μ t .
Let the probability of a stochastic variable  T i  occurring at time t be q(t). Its probability distribution function is given as follows:
Θ T i ( t ) = P [ T i t ] = { 0 t [ 0 , T min ] q ( t ) t [ T min , T max ] 1 t [ T max , ]
where  0 q ( t ) 1  is a known function.
To state the randomness of iteration length, a stochastic variable  υ i ( t ) R , t [ 0 , T max ]  is given, which satisfies Bernoulli distribution by taking binary values of 0 and 1. The event  υ i ( t ) = 1  indicates that the learning process can operate at or after the time instant t with a probability of  p ( t )  satisfying  0 < p ( t ) 1 . The event  υ i ( t ) = 0  indicates that the learning process cannot operate at the time instant k with a probability of  1 p ( t ) . It is clear that the system output data are always available for any iteration in the time interval  [ 0 , T min ] . Because the event  υ i ( t ) = 1 t [ 0 , T min ]  holds at any iteration, the probability is  p ( t ) = 1 . For  t [ T min , T max ] , the event  υ i ( t ) = 1  indicates that the learning process stops at or after the time instant t, i.e.,  T i t .
Therefore, it yields
P [ υ i ( t ) = 1 ] = P [ T i t ] = 1 P [ T i < t ] = 1 P [ T i < t ] P [ T i = t ] = 1 P [ T i t ] = 1 F T i ( t )
where  P [ T i = t ] = 0 .
Hence, we have
p ( t ) = 1 Θ T i ( t ) = { 1 t [ 0 , T min ] 1 q ( t ) t [ T min , T max ] 0 t [ T max , ] .
Because  υ i ( t )  satisfies Bernoulli distribution, the mathematical expectation of  υ i ( t )  is
E { υ i ( t ) } = 1 p ( t ) + 0 ( 1 p ( t ) ) = p ( t ) .
Define the modified tracking error as  e i ( t ) = υ i ( t ) e i ( t ) , t [ 0 , T r ] .
If  T i < T r , we have
e i ( t ) = { e i ( t ) t [ 0 , T i ] 0 t [ T i + 1 , T r ] .
If  T i T r , it yields
e i ( t ) = e i ( t ) t [ 0 , T r ] .

3.2. Controller Design

According to the modified tracking error, the PD-type iterative learning control law with the forgetting factor (PDILCFF) is designed as follows.
u i + 1 ( t ) = ( 1 β ) u i ( t ) + β u 0 ( t ) + H e . i ( t ) + K e i ( t )
where  0 β < 1 β  is the forgetting factor. The initial input correction item  β u 0 ( t )  can avoid the large oscillation of the iterative trajectory and accelerate the rate of convergence.  H  and  K  are the gain matrices of appropriate dimensions.
Figure 1 is a block diagram of the proposed PDILCFF method with randomly varying iteration lengths, the initial state shift and disturbances. In this paper, considering iteration-varying trial lengths, the proposed PDILCFF algorithm is closely related to symmetry. Symmetry can provide prior information about the system structure and characteristics of the proposed PDILCFF method, which is helpful for designing more efficient control algorithms. On the other hand, the proposed PDILCFF method exploits system symmetries in different intervals through iterative processes to achieve accurate control and performance optimization of MIMO unknown nonlinear systems.

3.3. Convergence Analysis

Definition 2.
Define the  λ  norm for a function  F : [ 0 , T ] R n  by  F ( ) λ = sup t [ 0 , T i ] e λ t F ( ) .
Assumption 1.
The initial state error is bounded, i.e.,  x d ( 0 ) x i ( 0 ) b x 0 , where  b x 0  is a positive constant.
Lemma 1.
If the sequence  c p R , satisfies  | c p + 1 | α | c p | + ε , 0 α < 1 , where  ε R ,  p Z + , the following inequation is held.
lim sup p | c p | 1 1 α ε .
Proof: 
According to  | c p + 1 | α | c p | + ε , we have
| c 1 | α | c 0 | + ε | c 2 | α 2 | c 0 | + ( 1 + α ) ε | c p | α p | c 0 | + j = 1 p α p j ε = α p | c 0 | + 1 α p 1 α ε .
Due to  0 α < 1 , we obtain  | c p | | c 0 | + 1 1 α ε .
Therefore, when  p , it leads to  lim sup p | c p | 1 1 α ε . □
Lemma 2 (Gronwall’s lemma).
 Let  x ( t )  and  y ( t )  be real-valued continuous functions defined on the interval [0, T],  a 0 . If  x ( t ) c + 0 t ( a x ( τ ) + b y ( τ ) ) d τ , then  x ( t ) c + 0 t ( a x ( τ ) + b y ( τ ) ) d τ .
Theorem 1.
For nonlinear system (1), if  ( 1 β ) I HCB α < 1  and assumption 1 are satisfied, the PDILCFF algorithm (9) can ensure that the tracking error is uniformly ultimately bounded as the iteration number i approaches infinity.
Proof: 
The proof consists of two parts. Part A demonstrates that the error between  u d ( t )  and  u i ( t )  is bounded. Part B shows that the state and output errors are bounded.
Part A:
According to (2) and (9), we have
u d u i + 1 = u d [ ( 1 β ) u i + β u 0 + H υ i ( y . d y . i ) + K υ i ( y d y i ) ] = ( 1 β ) ( u d u i ) + β ( u d u 0 ) H υ i [ C ( f d + B u d ) C ( f i + B u i + w i ) v . i ] K υ i ( C x d C x i v i ) = [ ( 1 β ) I H υ i CB ] ( u d u i ) + β ( u d u 0 ) H υ i C [ ( f d + B u d ) ( f i + B u d + w i ) ] + H υ i v . i K υ i C ( x d x i ) + K υ i v i .
Taking the norm of both sides of (10), we have
u d u i + 1 ( 1 β ) I H υ i CB u d u i + β u d u 0 + H υ i C [ k f x d x i ] + H υ i C w i + H υ i v . i + K υ i C x d x i + K υ i v i ( 1 β ) I H υ i CB u d u i + β u d u 0 + ( H υ i C k f + K υ i C ) x d x i + H υ i C b w + H υ i b v 2 + K υ i b v 1 .
According to system (1) and the globally Lipschitz condition for  f ( , ) , the following can be obtained:
x d x i = x d ( 0 ) x i ( 0 ) + 0 t ( ( f d + B u d ) ( f i + B u i + w i ) ) d τ = x d ( 0 ) x i ( 0 ) + 0 t ( f d f i + B ( u d u i ) w i ) d τ x d ( 0 ) x i ( 0 ) + 0 t ( f d f i + B u d u i + w i ) d τ x d ( 0 ) x i ( 0 ) + 0 t ( k f x d x i + b B u d u i + b w ) d τ
where  B b B .
Define  Δ x i = x d x i  and  Δ u i = u d u i . According to Gronwall’s lemma, we have
Δ x i Δ x i ( 0 ) + 0 t ( k f Δ x i + b B Δ u i + b w ) d τ Δ x i ( 0 ) e k f t + 0 t e k f ( t τ ) ( b B Δ u i ( τ ) + b w ) d τ = Δ x i ( 0 ) e k f t + b B 0 t e k f ( t τ ) Δ u i ( τ ) d τ + b w 0 t e k f ( t τ ) d τ .
Substituting (13) into (11) yields the following:
Δ u i + 1 ( 1 β ) I H υ i CB Δ u i + β Δ u 0 + ( H υ i C k f + K υ i C ) Δ x i + H υ i C b w + H υ i b v 2 + K υ i b v 1 ( 1 β ) I H υ i CB Δ u i + β Δ u 0 + ( H υ i C k f + K υ i C ) × ( Δ x i ( 0 ) e k f t + b B 0 t e k f ( t τ ) Δ u i ( τ ) d τ + b w 0 t e k f ( t τ ) d τ ) + H υ i C b w + H υ i b v 2 + K υ i b v 1 = ( 1 β ) I H υ i CB Δ u i + β Δ u 0 + ( H υ i C k f + H υ i C ) Δ x i ( 0 ) e k f t + ( H υ i C k f + H υ i C ) b B 0 t e k f ( t τ ) Δ u i ( τ ) d τ + ( H υ i C k f + H υ i C ) b w 0 t e k f ( t τ ) d τ + H υ i C b w + H υ i b v 2 + K υ i b v 1 .
Multiplying (14) by  e λ t , leads to
e λ t Δ u i + 1 ( 1 β ) I H υ i CB e λ t Δ u i + β e λ t Δ u 0 + ( H υ i C k f + K υ i C ) Δ x i ( 0 ) e ( k f λ ) t + ( H υ i C k f + K υ i C ) b B 0 t e λ τ e ( k f λ ) ( t τ ) Δ u i ( τ ) d τ + ( H υ i C k f + K υ i C ) b w 0 t e λ τ e ( k f λ ) ( t τ ) d τ + H υ i C b w e λ t + H υ i b v 2 e λ t + K υ i b v 1 e λ t .
Therefore, we obtain
Δ u i + 1 λ ( 1 β ) I H υ i CB Δ u i λ + ( H υ i C k f + K υ i C ) b B 0 t e ( k f λ ) ( t τ ) d τ Δ u i λ + β Δ u 0 λ + ( H υ i C k f + K υ i C ) Δ x i ( 0 ) e ( k f λ ) t + ( H υ i C k f + K υ i C ) b w 0 t e ( k f λ ) ( t τ ) d τ + H υ i C b w + H υ i b v 2 + K υ i b v 1 .
Taking mathematical expectation of both sides of (16) and noting that  υ i  is the only stochastic variable leads to
Δ u i + 1 λ E { ( 1 β ) I H υ i CB } Δ u i λ + E { ( H υ i C k f + K υ i C ) } b B 0 t e ( k f λ ) ( t τ ) d τ Δ u i λ + β Δ u 0 λ + E { ( H υ i C k f + K υ i C ) } Δ x i ( 0 ) e ( k f λ ) t + E { ( H υ i C k f + K υ i C ) } b w 0 t e ( k f λ ) ( t τ ) d τ + E { H υ i C } b w + E { H υ i } b v 2 + E { K υ i } b v 1 = ( 1 β ) I H p CB Δ u i λ + ( H p C k f + K p C ) b B 0 t e ( k f λ ) ( t τ ) d τ Δ u i λ + β Δ u 0 λ + ( H p C k f + K p C ) Δ x i ( 0 ) e ( k f λ ) t + ( H p C k f + K p C ) b w 0 t e ( k f λ ) ( t τ ) d τ + H p C b w + H p b v 2 + K p b v 1 = ħ 1 Δ u i λ + ħ 2 b B 0 t e ( k f λ ) ( t τ ) d τ Δ u i λ + β Δ u 0 λ + ħ 2 Δ x i ( 0 ) e ( k f λ ) t + ħ 3 b w 0 t e ( k f λ ) ( t τ ) d τ + H p C b w + H p b v 2 + K p b v 1
where  ħ 1 = ( 1 β ) I H p CB ħ 2 = H p C k f + K p C ħ 3 = H p C k f + K p C .
Define  ħ = max { ħ 2 b B , k f } . Assume that  λ > ħ . It yields
Δ u i + 1 λ ħ 1 Δ u i λ + ħ 0 t e ( ħ λ ) ( t τ ) d τ Δ u i λ + β Δ u 0 λ + ħ 2 Δ x i ( 0 ) e ( k f λ ) t + ħ 3 b w 0 t e ( k f λ ) ( t τ ) d τ + H p C b w + H p b v 2 + K p b v 1 ( ħ 1 + ħ λ ħ ( 1 e ( ħ λ ) T i ) ) Δ u i λ + β Δ u 0 λ + ħ 2 Δ x i ( 0 ) + ħ 3 b w 1 k f ( 1 e ( k f λ ) T i ) + H p C b w + H p b v 2 + K p b v 1 = ( ħ 1 + ħ λ ħ ( 1 e ( ħ λ ) T i ) ) Δ u i λ + β Δ u 0 λ + ħ 2 Δ x i ( 0 ) + ( H p C + ħ 3 1 k f ( 1 e ( k f λ ) T i ) ) b w + H p b v 2 + K p b v 1
Defining  ħ 3 = ħ 1 + ħ λ ħ ( 1 e ( ħ λ ) T i )  and  ħ 4 = H p C + ħ 3 1 k f ( 1 e ( k f λ ) T i ) , we obtain
Δ u i + 1 λ ħ 3 Δ u i λ + β Δ u 0 λ + ħ 2 Δ x i ( 0 ) + ħ 4 b w + H p b v 2 + K p b v 1 .
Further, it yields
Δ u i + 1 λ ħ 3 Δ u i λ + σ
where  σ = β Δ u 0 λ + ħ 2 Δ x i ( 0 ) + ħ 4 b w + H p b v 2 + K p b v 1 , which includes the norm bounds of the initial state errors, state disturbances, and bias contribution.
Due to  0 < p 1 , we have  ħ 1 < 1 . There must exist a sufficiently large  λ  such that  λ > ħ . Therefore, we have  ħ 3 < 1 . In terms of (20) and lemma 1, we obtain
lim i sup Δ u i + 1 λ 1 1 ħ 3 σ .
Part B:
According to (13), we obtain
Δ x i λ Δ x i ( 0 ) + 0 t e k f ( t τ ) ( b B Δ u i ( τ ) + b w ) d τ Δ x i ( 0 ) + 0 t e ( k f λ ) ( t τ ) ( b B Δ u i λ + b w ) d τ b x 0 + 1 λ k f ( 1 e ( k f λ ) T i ) ( b B Δ u i λ + b w ) .
From (21) and (22), we obtain
lim i sup Δ x i λ b x 0 + 1 λ k f ( 1 e ( k f λ ) T i ) ( b B 1 ħ 3 σ + b w ) .
Because  e i = y d y i = C Δ x i ( t ) v i ( t ) , we have
e i λ C Δ x i ( t ) + v i ( t ) c Δ x i λ + b v 1
where  C c , c is a constant.
According to (24),  e i λ  is bounded.
According to (21), (23) and (24), it is clear that the initial state error, state disturbance and bias term influence the errors. If  σ = 0   b x 0 = 0 b w = 0  and  b v 1 = 0 , the tracking errors can converge to zero.
The proof of Theorem 1 is complete. □
Remark 1.
In this paper, the forgetting factor can enable the system to maintain stability and good performance when factors such as external interference and initial state shifts occur. When the system is disturbed, the forgetting factor enables the algorithm to remove the influence of the interference on the information of past iterations more quickly and readjust the control input to deal with the new situation, thereby enhancing the robustness and anti-interference ability of the system.

4. Simulation Study

Two simulation examples are provided to illustrate the effectiveness of the proposed PDILCFF scheme in this section. The PILC algorithm (25) is considered for comparison.
u i + 1 ( t ) = u i ( t ) + λ p e i ( t )
where  λ p  is the gain matrix of the appropriate dimension.
Example 1.
Subway train tracking control
The parameters of the subway train dynamic model in [41], such as basic resistance coefficients and the mass and distribution constant for each vehicle, are given and omitted here for the sake of simplification. The disturbance is given by  w ( t ) = 0.1 M s i n ( t ) e 3 i t . The initial values are  u i ( 0 ) = 0.01 x i ( 0 ) = 0.01 , and  x ˙ i ( 0 ) = 0 T min = 185 s T min = 191 s .
The parameters of the proposed PDILCFF algorithm (9) and PILC algorithm (25) are  β = 0.001 H = 11 K = 1.9 × 10 5 , and  λ p = 4600 ( 1 β ) I HCB = 0 < 1  satisfies the condition in Theorem 1.
Figure 2 shows the unit additional resistance profiles of the subway train system at the 2nd, 5th, 8th and 10th iterations. The additional resistance includes tunnel resistance, gradient resistance, and curve resistance.
Figure 3 and Figure 4 plot the speed tracking profiles with PDILCFF and PILC at the 2nd, 5th, 8th and 10th iterations in the time domain. Figure 5 and Figure 6 show the speed and position tracking profiles at the 10th iteration. Figure 7 and Figure 8 exhibit the maximum speed and position tracking error profiles in the iteration domain. Figure 9 and Figure 10 show the force contribution profiles for PDILCFF and PILC. From Figure 3, Figure 4, Figure 5, Figure 6, Figure 7 and Figure 8, we can see that the presented PDILCFF method has better convergence and control performance than the PILC method under the randomly varying iteration lengths for different iterations and disturbances.
Table 1 and Table 2 show a comparison of the maximum speed and position errors of the different algorithms in Example 1. It is clear that the proposed control algorithm has better tracking performance and convergence.
Table 1. Comparison of the maximum speed errors of different algorithms in Example 1.
Table 1. Comparison of the maximum speed errors of different algorithms in Example 1.
Control AlgorithmThe Maximum Speed Tracking Error of Different Iterations
12345678910
PDILCFF22.237.582.781.430.600.730.510.380.500.42
PILC58.5832.0457.4635.0028.4929.7029.7123.7338.0251.19
Table 2. Comparison of the maximum position errors of different algorithms in Example 1.
Table 2. Comparison of the maximum position errors of different algorithms in Example 1.
Control AlgorithmThe Maximum Position Tracking Error of Different Iterations
12345678910
PDILCFF1687641242125566743334534
PILC20146412239253817271188883676514796
Example 2.
Two-degree-of-freedom robot manipulator [42].
D ( x 1 ) x ˙ 2 + C ( x 1 , x 2 ) x 2 + G ( x 1 ) + w ( t ) = τ
where  x 1 = [ x 11 , x 12 ]  is the joint position vector.  x 2 = [ x 21 , x 22 ]  is the velocity vector. τ = [ τ 1 , τ 2 ]  is the control input vector of two joint links.  D ( x 1 ) = [ d 11 d 12 d 21 d 22 ]  is the inertia matrix.
d 11 = m 1 l c 1 2 + m 2 ( l 1 2 + l c 2 2 + 2 l 1 l c 2 cos ( x 12 ) ) + I 1 + I 2 d 12 = d 21 = m 2 ( l c 2 2 + l 1 l c 2 cos ( x 12 ) ) + I 2 , d 22 = m 2 l c 2 2 + I 2 .
C ( x 1 , x 2 ) = [ c 11 c 12 c 21 c 22 ]  is the Coriolis–centripetal matrix.
c 11 = m 2 l 1 l c 2 x 22 sin ( x 12 ) c 21 = m 2 l 1 l c 2 x 21 sin ( x 12 )
c 12 = m 2 l 1 l c 2 x 22 sin ( x 12 ) ( x 21 + x 22 ) c 22 = 0 .
G ( x 1 ) = [ G 1 , G 2 ] T  is gravity.
G 1 = ( m 1 l c 2 + m 2 l 1 ) g cos ( x 11 ) + m 2 l c 2 g cos ( x 11 + x 12 )
G 2 = m 2 l c 2 g cos ( x 11 + x 12 ) .
The parameters of the robot manipulator are  l 1 = l 2 = 1   m l c 1 = l c 2 = 0.5   m I 1 = I 2 = 0.2   kg m 2 , and  g = 9.81   m / s 2 m 1 = 0.9   kg  and  m 2 = 1   kg  represent the mass. The disturbance is  w ( t ) = [ 0.3 sin ( t ) , 0.1 ( 1 e t ] . The desired trajectories are  x 1 d = [ sin ( 3 t ) , cos ( 3 t ) ] T , and  x 2 d = [ 3 cos ( 3 t ) , 3 sin ( 3 t ) ] T .
The parameters of PDILCFF and PILC are given by  β = 0.02 K = [ 0.3 0 0 0.5 ] H = [ 200 0 0 20 ] λ p = [ 0.8 0 0 0.8 ] T min = 4.9   s , and  T min = 5.05   s x 1 ( 0 ) = [ 0 , 2.8 ] x 2 ( 0 ) = [ 1.01 , 0 ] , and  τ ( 0 ) = [ 0.1 , 0.1 ] .
Figure 11 and Figure 12 plot the position and velocity tracking profiles of two links with PDILCFF at the 1st and 5th iterations in the time domain. Figure 13 and Figure 14 exhibit the position and velocity tracking profiles for PDILCFF and PILC at the 5th iteration. Figure 15 and Figure 16 give the maximum speed and position tracking error profiles in the iteration domain. Figure 11, Figure 12, Figure 13, Figure 14 and Figure 15 demonstrate that when the operation length of the robot manipulator system with disturbances varies randomly, the proposed PDILCFF method has better convergence and control performance than PILC.
Table 3 and Table 4 show a comparison of the maximum speed and position errors of the different algorithms in Example 2. It can be seen that the proposed control algorithm has better tracking performance and convergence.
The simulation results of Section 4 demonstrate that the proposed PDILCFF has good control effect and robustness and verifies the correctness of Theorem 1.
Remark 2.
Considering iteration-varying trial lengths, the proposed PDILCFF algorithm is closely related to symmetry.
(1) The PDILCFF method takes advantage of time symmetry. The essence of iterative learning control is to learn and optimize the process of repeating the same task, and each iteration can be regarded as a time period. If the system meets certain conditions, different iteration cycles have similarities or symmetries in the time dimension. Time symmetry enhances the effectiveness of iterative learning control, enabling it to make better use of historical information. For example, for the control task of a robotic arm with periodic motion, each complete motion cycle can be regarded as a repetition of the previous cycle, but the control input is constantly optimized during the iterative process. This time symmetry enables iterative learning control to effectively learn from historical iterative data and symmetrically apply the error information of the previous period to the control and adjust subsequent periods, gradually reducing the error and improving system performance.
(2) The PDILCFF method utilizes dynamic symmetry. There is a certain symmetrical relationship between the treatment methods of errors and their rates of change in the proportional link and the differential link. The proportional link is adjusted according to the current error magnitude, while the differential link is adjusted according to the error change rate. The two cooperate with each other to control system dynamics from a symmetrical perspective. Dynamic symmetry enhances the controller’s response to the dynamic changes of the system.
(3) The PDILCFF method utilizes spatial symmetry. For example, during the movement of the robotic arm, the movements of each joint axis cooperate with each other to jointly complete complex spatial tasks, and the movement relationship among them contains spatial symmetry. Meanwhile, iterative learning control can utilize spatial symmetry to share and transfer control experience and error information of different axes at the same task stage, optimizing the overall control performance. Spatial symmetry ensures the stability and accuracy of the overall motion of the system in systems involving spatial motion.
The comprehensive application of these symmetries enables the PDILCFF method to work more efficiently in complex control systems, enhancing the robustness and reliability of the system.

5. Conclusions

This paper has proposed a PDILCFF method for nonlinear systems with randomly varying iteration lengths and state and measurement disturbances. Firstly, considering the randomly varying iteration lengths, a modified tracking error was designed. Secondly, a PDILCFF scheme was proposed to deal with the problems of initial state shift, state disturbances and measurement disturbances. The compression mapping method was exploited to obtain a convergence property of the proposed PDILCFF scheme, which guaranteed that the tracking error was bounded convergence, and when the disturbances disappeared, the tracking error converged to zero. Finally, two simulation examples were provided to verify the effectiveness of the theoretical studies. In future work, by drawing on the control algorithm and analysis method proposed in this paper, we will study multi-agent consensus and formation problems under randomly varying iteration lengths, initial state shifts and disturbances.

Author Contributions

Conceptualization, G.L.; methodology, G.L. and Y.W.; software, G.L. and J.L.; validation, G.L., Y.W., J.L. and Q.W.; formal analysis, G.L. and Y.W.; investigation, G.L.; writing—original draft preparation, G.L.; writing—review and editing, G.L. and Y.W.; supervision, G.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant 62203151, Scientific and Technological Project of Henan Province of China under Grant 252102240132, and High-level Talents Fund Project of Henan University of Technology under Grant 2021BS084.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to thank the anonymous reviewers for their helpful comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Arimoto, S.; Kawamura, S.; Miyazaki, F. Bettering operation of robots by learning. J. Field Robot. 1984, 1, 123–140. [Google Scholar] [CrossRef]
  2. Xu, J.-X. A survey on iterative learning control for nonlinear systems. Int. J. Control. 2011, 84, 1275–1294. [Google Scholar] [CrossRef]
  3. Tan, W.; Hou, Z.; Li, Y.X. Robust data-driven iterative learning control for high-speed train with aperiodic DoS attacks and communication delays. IEEE Trans. Intell. Vehicl. 2024, 1–10. [Google Scholar] [CrossRef]
  4. Li, S.; Li, X. Finite-time extended state observer-based iterative learning control for nonrepeatable nonlinear systems. Nonlinear Dyn. 2025, 113, 16531–16543. [Google Scholar] [CrossRef]
  5. Arimoto, S. Learning control theory for robotic motion. Int. J. Adapt. Control. Signal Process. 1990, 4, 543–564. [Google Scholar] [CrossRef]
  6. Saab, S.S. On the P-type learning control. IEEE Trans. Automat. Contr. 1994, 39, 2298–2302. [Google Scholar] [CrossRef]
  7. Xu, J.X.; Jin, X. State-Constrained Iterative Learning Control for a Class of MIMO Systems. IEEE Trans. Automat. Contr. 2013, 58, 1322–1327. [Google Scholar] [CrossRef]
  8. Bai, L.; Feng, Y.-W.; Li, N.; Xue, X.-F.; Cao, Y. Data-Driven Adaptive Iterative Learning Method for Active Vibration Control Based on Imprecise Probability. Symmetry 2019, 11, 746. [Google Scholar] [CrossRef]
  9. Hou, Z.S.; Chi, R.; Gao, H. An overview of dynamic-linearization-based data-driven control and applications. IEEE Trans. Ind. Electron. 2017, 64, 4076–4090. [Google Scholar] [CrossRef]
  10. Yu, X.; Chen, T. Distributed Iterative Learning Control of Nonlinear Multiagent Systems Using Controller-Based Dynamic Linearization Method. IEEE Trans. Cybern. 2023, 54, 4489–4501. [Google Scholar] [CrossRef]
  11. Zhang, H.; Chi, R.; Huang, B.; Hou, Z. Compensatory Data-Driven Networked Iterative Learning Control with Communication Constraints and DoS Attacks. IEEE Trans. Autom. Sci. Eng. 2025, 22, 10728–10740. [Google Scholar] [CrossRef]
  12. Chi, R.; Hou, Z.S.; Huang, B.; Jin, S. A unified data-driven design framework of optimality-based generalized iterative learning control. Comput. Chem. Eng. 2015, 77, 10–23. [Google Scholar] [CrossRef]
  13. Chi, R.; Liu, X.; Zhang, R.; Hou, Z.S.; Huang, B. Constrained data-driven optimal iterative learning control. J. Process Contr. 2017, 55, 10–29. [Google Scholar] [CrossRef]
  14. Huang, Y.; Tao, H.; Chen, Y.; Rogers, E.; Paszke, W. Point-to-point iterative learning control with quantised input signal and actuator faults. Int. J. Control. 2024, 97, 17. [Google Scholar] [CrossRef]
  15. Wang, L.; Dong, L.; Chen, Y.; Wang, K.; Gao, F. Iterative Learning Control for Actuator Fault Uncertain Systems. Symmetry 2022, 14, 1969. [Google Scholar] [CrossRef]
  16. Tayebi, A. Adaptive iterative learning control for robot manipulators. Automatica 2004, 40, 1195–1203. [Google Scholar] [CrossRef]
  17. Yin, C.; Xu, J.-X.; Hou, Z. A high-order internal model based iterative learning control scheme for nonlinear systems with time-iteration-varying parameters. IEEE Trans. Automat. Contr. 2010, 55, 2665–2670. [Google Scholar] [CrossRef]
  18. Chi, R.H.; Hou, Z.S.; Huang, B. Optimal iterative learning control of batch processes: From model-based to data-driven. Zidonghua Xuebao/Acta Autom. Sin. 2017, 43, 917–932. [Google Scholar]
  19. Liu, G.; Hou, Z.S. Cooperative adaptive iterative learning fault-tolerant control scheme for multiple subway trains. IEEE Trans. Cybern. 2022, 52, 1098–1111. [Google Scholar] [CrossRef]
  20. Liu, G.; Hou, Z.S. RBFNN-based adaptive iterative learning fault-tolerant control for subway trains with actuator faults and speed constraint. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 5785–5799. [Google Scholar] [CrossRef]
  21. Hou, Z.S.; Yan, J.; Xu, J.-X.; Li, Z. Modified iterative-learning-control-based ramp metering strategies for freeway traffic control with iteration-dependent factors. IEEE Trans. Intell. Transp. Syst. 2012, 13, 606–618. [Google Scholar] [CrossRef]
  22. Hou, Z.; Xu, X.; Yan, J.; Xu, J.X.; Xiong, G. A complementary modularized ramp metering approach based on iterative learning control and ALINEA. IEEE Trans. Intell. Transp. Syst. 2011, 12, 1305–1318. [Google Scholar] [CrossRef]
  23. Longman, R.W.; Mombaur, K.D. Investigating the use of iterative learning control and repetitive control to implement periodic gaits. Lect. Notes Control Inform. Sci. 2006, 340, 189–218. [Google Scholar]
  24. Yu, Q.; Hou, Z.S. Adaptive fuzzy iterative learning control for high-speed trains with both randomly varying operation lengths and system constraints. IEEE Trans. Fuzzy Syst. 2021, 29, 2408–2418. [Google Scholar] [CrossRef]
  25. Seel, T.; Schauer, T.; Raisch, J. Iterative learning control for variable pass length systems. In Proceedings of the the 18th IFAC World Congress, Milano, Italy, 28 August–2 September 2011; pp. 4880–4885. [Google Scholar]
  26. Guth, M.; Seel, T.; Raisch, J. Iterative learning control with variable pass length applied to trajectory tracking on a crane with output constraints. In Proceedings of the 52nd IEEE Conference on Decision and Control, Florence, Italy, 10–13 December 2013; pp. 6676–6681. [Google Scholar]
  27. Li, X.; Xu, J.-X.; Huang, D. An iterative learning control approach for linear time-invariant systems with randomly varying trial lengths. IEEE Trans. Automat. Contr. 2014, 59, 1954–1960. [Google Scholar] [CrossRef]
  28. Li, X.; Xu, J.-X.; Huang, D. Iterative learning control for nonlinear dynamic systems with randomly varying trial lengths. Int. J. Adapt. Control Signal Process 2015, 29, 1341–1353. [Google Scholar] [CrossRef]
  29. Shen, D.; Zhang, W.; Wang, Y.; Chien, C.J. On almost sure and mean square convergence of P-type ILC under randomly varying iteration lengths. Automatica 2016, 63, 359–365. [Google Scholar] [CrossRef]
  30. Li, X.; Shen, D. Two novel iterative learning control schemes for systems with randomly varying trial lengths. Syst. Control. Lett. 2017, 107, 9–16. [Google Scholar] [CrossRef]
  31. Wang, L.; Li, X.; Shen, D. Sampled-data iterative learning control for continuous-time nonlinear systems with iteration-varying lengths. Int. J. Robust Nonlinear Control. 2018, 28, 3073–3091. [Google Scholar] [CrossRef]
  32. Shen, D.; Xu, J.-X. Robust learning control for nonlinear systems with nonparametric uncertainties and nonuniform trial lengths. Int. J. Robust Nonlinear Control. 2019, 29, 1302–1324. [Google Scholar] [CrossRef]
  33. Shen, D.; Xu, J.-X. Adaptive learning control for nonlinear systems with randomly varying iteration lengths. IEEE Trans. Neural. Networ. 2019, 30, 1119–1132. [Google Scholar] [CrossRef] [PubMed]
  34. Liu, G.; Hou, Z. Adaptive iterative learning fault-tolerant control for state constrained nonlinear systems with randomly varying iteration lengths. IEEE Trans. Neural. Networ. 2024, 35, 1735–1749. [Google Scholar] [CrossRef]
  35. Bu, X.; Wang, S.; Hou, Z.S.; Liu, W. Model free adaptive iterative learning control for a class of nonlinear systems with randomly varying iteration lengths. J. Frankl. Inst. 2019, 356, 2491–2504. [Google Scholar] [CrossRef]
  36. Xu, R.; Chen, H.; Tang, Y.; Long, X. Model-free predictive iterative learning safety-critical consensus control for multi-agent systems with randomly varying trial lengths. Syst. Control. Lett. 2025, 196, 105987. [Google Scholar] [CrossRef]
  37. Shen, D.; Zhang, W.; Xu, J.-X. Iterative learning control for discrete nonlinear systems with randomly iteration varying lengths. Syst. Control. Lett. 2016, 96, 81–87. [Google Scholar] [CrossRef]
  38. Wei, Y.; Li, X. Robust higher-order ILC for non-linear discrete-time systems with varying trail lengths and random initial state shifts. IET Control. Theory Appl. 2017, 11, 2440–2447. [Google Scholar] [CrossRef]
  39. Liu, Y.; Fan, Y.; Ao, Y.; Jia, Y. An iterative learning approach to formation control of discrete-time multi-agent systems with varying trial lengths. Int. J. Robust Nonlinear Control. 2022, 32, 9332–9346. [Google Scholar] [CrossRef]
  40. Hussain, M.; Muslim, F.B.; Khan, O.; Saqib, N.u. Robust anti-windup control atrategy for uncertain nonlinear system with time delays. Arab. J. Sci. Eng. 2022, 47, 3847–3860. [Google Scholar] [CrossRef]
  41. Liu, G.; Hou, Z. Adaptive iterative learning control for subway trains using multiple-point-mass dynamic model under speed constraint. IEEE Trans. Intell. Transp. Syst. 2021, 22, 1388–1400. [Google Scholar] [CrossRef]
  42. Jin, X. Iterative learning control for MIMO nonlinear systems with iteration-varying trial lengths using modified composite energy function analysis. IEEE Trans. Cybern. 2021, 51, 6080–6090. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed PDILCFF method.
Figure 1. Block diagram of the proposed PDILCFF method.
Symmetry 17 00694 g001
Figure 2. Unit additional resistance.
Figure 2. Unit additional resistance.
Symmetry 17 00694 g002
Figure 3. Speed tracking with PDILCFF.
Figure 3. Speed tracking with PDILCFF.
Symmetry 17 00694 g003
Figure 4. Speed tracking with PILC.
Figure 4. Speed tracking with PILC.
Symmetry 17 00694 g004
Figure 5. Speed tracking with PDILCFF and PILC at the 150th iteration.
Figure 5. Speed tracking with PDILCFF and PILC at the 150th iteration.
Symmetry 17 00694 g005
Figure 6. Position tracking with PDILCFF and PILC at the 150th iteration.
Figure 6. Position tracking with PDILCFF and PILC at the 150th iteration.
Symmetry 17 00694 g006
Figure 7. Maximum speed tracking error profile with trial length satisfying uniform distribution in time interval [185, 191] s with PDILCFF and PILC.
Figure 7. Maximum speed tracking error profile with trial length satisfying uniform distribution in time interval [185, 191] s with PDILCFF and PILC.
Symmetry 17 00694 g007
Figure 8. Maximum position tracking error profile with trial length satisfying uniform distribution in time interval [185, 191] s with PDILCFF and PILC.
Figure 8. Maximum position tracking error profile with trial length satisfying uniform distribution in time interval [185, 191] s with PDILCFF and PILC.
Symmetry 17 00694 g008
Figure 9. Traction/braking force contributed by motor vehicles with PDILCFF and PILC.
Figure 9. Traction/braking force contributed by motor vehicles with PDILCFF and PILC.
Symmetry 17 00694 g009
Figure 10. Braking force contributed by trailers.
Figure 10. Braking force contributed by trailers.
Symmetry 17 00694 g010
Figure 11. Position tracking with PDILCFF.
Figure 11. Position tracking with PDILCFF.
Symmetry 17 00694 g011
Figure 12. Velocity tracking with PDILCFF.
Figure 12. Velocity tracking with PDILCFF.
Symmetry 17 00694 g012
Figure 13. Position tracking with PDILCFF and PILC at the 5th iteration.
Figure 13. Position tracking with PDILCFF and PILC at the 5th iteration.
Symmetry 17 00694 g013
Figure 14. Velocity tracking with PDILCFF and PILC at the 5th iteration.
Figure 14. Velocity tracking with PDILCFF and PILC at the 5th iteration.
Symmetry 17 00694 g014
Figure 15. Maximum position tracking error profile with trial length satisfying uniform distribution in time interval [4.9 5.05] with PDILCFF and PILC.
Figure 15. Maximum position tracking error profile with trial length satisfying uniform distribution in time interval [4.9 5.05] with PDILCFF and PILC.
Symmetry 17 00694 g015
Figure 16. Maximum velocity tracking error profiles with trial length satisfying uniform distribution in time interval [4.9 5.05] with PDILCFF and PILC.
Figure 16. Maximum velocity tracking error profiles with trial length satisfying uniform distribution in time interval [4.9 5.05] with PDILCFF and PILC.
Symmetry 17 00694 g016
Table 3. Comparison of the maximum velocity errors of different algorithms in Example 2.
Table 3. Comparison of the maximum velocity errors of different algorithms in Example 2.
Control AlgorithmLink 1 of Different IterationsLink 2 of Different Iterations
1234512345
PDILCFF0.420.200.210.200.200.550.350.230.220.22
PILC0.680.910.270.580.252.282.150.742.171.25
Table 4. Comparison of the maximum position errors of different algorithms in Example 2.
Table 4. Comparison of the maximum position errors of different algorithms in Example 2.
Control AlgorithmLink 1 of Different IterationsLink 2 of Different Iterations
1234512345
PDILCFF0.530.130.250.040.030.580.300.100.040.03
PILC1.280.800.130.190.080.730.840.240.550.40
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, G.; Wang, Y.; Li, J.; Wang, Q. Iterative Learning Control with Forgetting Factor for MIMO Nonlinear Systems with Randomly Varying Iteration Lengths and Disturbances. Symmetry 2025, 17, 694. https://doi.org/10.3390/sym17050694

AMA Style

Liu G, Wang Y, Li J, Wang Q. Iterative Learning Control with Forgetting Factor for MIMO Nonlinear Systems with Randomly Varying Iteration Lengths and Disturbances. Symmetry. 2025; 17(5):694. https://doi.org/10.3390/sym17050694

Chicago/Turabian Style

Liu, Genfeng, Yangyang Wang, Jinhao Li, and Qinghe Wang. 2025. "Iterative Learning Control with Forgetting Factor for MIMO Nonlinear Systems with Randomly Varying Iteration Lengths and Disturbances" Symmetry 17, no. 5: 694. https://doi.org/10.3390/sym17050694

APA Style

Liu, G., Wang, Y., Li, J., & Wang, Q. (2025). Iterative Learning Control with Forgetting Factor for MIMO Nonlinear Systems with Randomly Varying Iteration Lengths and Disturbances. Symmetry, 17(5), 694. https://doi.org/10.3390/sym17050694

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop