Next Article in Journal
A New Hyperparameter Tuning Framework for Regression Tasks in Deep Neural Network: Combined-Sampling Algorithm to Search the Optimized Hyperparameters
Previous Article in Journal
A New Random Coefficient Autoregressive Model Driven by an Unobservable State Variable
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Varying-Factor Finite-Time Recurrent Neural Network to Solve the Time-Varying Sylvester Equation Online

1
School of Electric and Electronic Enginnering, Shanghai University of Engineering Science, Shanghai 201620, China
2
School of Mathematics and Computer Sciences, Nanchang University, Nanchang 330031, China
3
Jiangxi Provincial Key Laboratory of Intelligent Systems and Human-Machine Interaction, Nanchang 330031, China
4
School of Information Engineering, Nanchang University, Nanchang 330031, China
5
School of Automation Science and Engineering, South China University of Technology, Guangzhou 510640, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(24), 3891; https://doi.org/10.3390/math12243891
Submission received: 11 November 2024 / Revised: 4 December 2024 / Accepted: 5 December 2024 / Published: 10 December 2024

Abstract

:
This paper presents a varying-parameter finite-time recurrent neural network, called a varying-factor finite-time recurrent neural network (VFFTRNN), which is able to solve the solution of the time-varying Sylvester equation online. The proposed neural network makes the matrix coefficients vary with time and can achieve convergence in a finite time. Apart from this, the performance of the network is better than traditional networks in terms of robustness. It is theoretically proved that the proposed neural network has super-exponential convergence performance. Simulation results demonstrate that this neural network has faster convergence speed and better robustness than the return to zero neural networks and can track the theoretical solution of the time-varying Sylvester equation effectively.

1. Introduction

As we all know, the Sylvester equation U P P V + W = 0 , as a kind of linear matrix equation, plays a crucial role in many fields, especially linear algebra and matrix theory. For some mathematics, dynamic control systems, and computer vision problems, the Sylvester equation can be used as a solution, like designing a feedback control system by pole assignment optimization, linear least-squares regression, image fusion [1], clustering [2,3], and so on. In recent years, many methods have been proposed by other scholars to solve real-time linear algebra problems [4,5,6,7,8,9,10,11,12,13,14,15,16,17,18]. Among these methods, several are efficient and representative, like gradient-based neural network (GNN) methods and zeroing neural network (ZNN) methods. It is notable that GNN and ZNN methods both use the idea to force the time derivative of the error function decline and further to force the error function gain toward zero. The proposal of GNNs, as emblematic recurrent neural networks, is very representational for solving real-time linear algebra problems [19,20,21]. Generally speaking, different GNNs are designed to solve different linear algebra problems. By itself, for solving the Sylvester equation, the GNN method is roughly stated as follows:
(1)
In the first step, the error function is defined as E = | | U P P V + W | | 2 / 2 . In order to acquire the solution of the Sylvester equation, the error function should converge to zero.
(2)
In the next step, for the purpose of acquiring the minimum point of the error function, GNN methods use a negative gradient descent direction. In addition to this, GNN methods employ a constant scalar-type parameter, which accelerates the convergence rate of the error function to solve the Sylvester equation efficiently.
However, when the task is time-varying, the traditional GNN method, which is based on gradient-based optimization, has difficulty tracking the ideal solution of the time-varying Sylvester equation. This is because the error, as defined by E = | | U P P V + W | | 2 2 , cannot converge to zero in such cases, resulting in residual errors [22]. Therefore, a GNN cannot solve the time-varying Sylvester function effectively. But in practice, time-varying problems are everywhere, so we need a neural network to better track locus solutions to time-varying issues in real-time. Under this background, many kinds of neural networks have been proposed to solve this kind of time-varying algebraic problem. Among these neural networks, the ZNN proposed by Zhang et al. proved to be efficient [12]. The ZNN is a kind of implicit dynamic neural network that skillfully replaces the gradient dynamic with an implicit dynamic, so that the computation error can achieve better convergence; that is, with the increase in time, the solution of the network can approach the theoretical solution in real time [23,24,25,26,27,28,29,30,31]. However, the ZNN also has some shortcomings, such as its convergence speed not being fast enough, convergence parameters being difficult to set, and other issues. Aiming to address the above shortcomings, Li et al. proposed a finite-time ZNN [23]. The sign-bi-power activation function is used in this network, as it effectively activates this network, makes the convergence speed much faster, and can achieve finite-time convergence. In addition, Zhang et al. proposed a variable-parameter convergence differential neural network (VP-CDNN) [32]. Different from the initial ZNN, the convergence parameter of the VP-CDNN changes with time, which makes the network hardware implementation much more reasonable than the ZNN, and the convergence speed reaches the super-exponential convergence speed. But despite this, the VP-CDNN cannot achieve finite-time convergence. Based on the above discussion, this paper proposes a recurrent neural network (VFFTRNN) for solving the time-varying Sylvester equation. The network not only has the speed of super-exponential convergence but also can achieve finite-time convergence.
The rest of this paper is divided into four parts. The second section introduces some tools needed to solve the Sylvester equation, and briefly describes the solution steps of the ZNN. In the third section, the implicit dynamic equation of the proposed VFFRTNN and the general steps of solving time-varying problems with the network are given, and the convergence and robustness of VFFTRNN under different conditions are analyzed theoretically. In the fourth section, the simulation results of MATLAB are given and compared with the simulation results of the ZNN. In the last section, the thesis is summarized.
Before concluding this section, some highlights of the paper are listed below:
(1)
We propose a VFFTRNN algorithm for solving the Sylvester equation.
(2)
Compared with VP-CDNN, this network can achieve finite-time convergence.
(3)
We theoretically prove the convergence time of the network in the case of three activation functions.
(4)
In the situation considering the existence of disturbances, the robustness of VFFTRNN is discussed.
(5)
Data simulation results show that the proposed VFFTRNN has better convergence and robustness than the traditional ZNN model under the same initial conditions.

2. Problem Formulation and Knowledge Preparation

Let us observe the following Sylvester equation with smooth time-varying characteristics:
U ( t ) P ( t ) P ( t ) V ( t ) + W ( t ) = 0 , t [ 0 , + )
where time is expressed as the variable t. The time-varying smooth coefficient matrices are W ( t ) m × n , V ( t ) n × n , as well as U ( t ) m × m . In addition, W ˙ ( t ) = d W ( t ) /dt, V ˙ ( t ) = d V ( t ) /dt, and U ˙ ( t ) = d U ( t ) /dt are their time derivatives, respectively, assuming these time derivatives are known or can be found exactly. Furthermore, P ( t ) m × n is uncharted but real, and our purpose is to find it to approach or even obtain the unique ideal solution P * ( t ) in finite time for the above Sylvester equation.
In order to facilitate the subsequent discussion and to solve Equation (1) more simply, (1) is best to be transformed into a vector form. The following Theorem (1) is introduced.
Theorem 1.
The Sylvester equation in matrix form (i.e., Equation (1)) can be transformed into vector form as follows:
I n U ( t ) V T ( t ) I m v e c ( P ( t ) ) + v e c ( W ( t ) ) = 0
where I n and I m represent the n-dimensional and m-dimensional identity matrices, respectively. Additionally, the symbol T represents the transpose of the matrix. The symbol denotes the Kronecker product. In other words, U V m n × m n represents a large matrix, which is equivalent to transforming the i t h row and j−column position element u i j in the matrix U into a matrix u i j V . The operation v e c ( P ( t ) ) m n represents a column vector obtained by stacking the columns of the matrix P ( t ) m × n in sequence.
Proof. 
See Section 2 and [12], Appendix. □
Furthermore, if we want to obtain the unique solution of Equation (1), then it must meet the regularity condition [12,33], which is described in the following Theorem (2).
Theorem 2.
If the Sylvester Equation (1) satisfies the following regularity condition, then it has a unique solution.
I n U T ( t ) U ( t ) V T ( t ) U T ( t ) V ( t ) U ( t ) + V ( t ) V T ( t ) I m α I m n , α > 0 , t [ 0 , + )
where I m n represents the mn-dimensional identity matrix.
Proof. 
See Section 2 and [12], Appendix. □
With the continuous development of neural dynamics and the proposal of ZNNs, the concept of ZNNs has been applied in many engineering and scientific fields to solve time-varying problems. To clarify the details of this paper, we will next briefly introduce how to solve the Sylvester Equation (1) using a typical ZNN method.
In the first step, an error function for Equation (1) is defined as follows:
E ( t ) : = U ( t ) P ( t ) P ( t ) V ( t ) + W ( t )
Next, since our goal is to obtain a unique solution to Equation (1), when we set the error function to approach 0, the unique solution of Equation (1) will be obtained naturally. In light of the design principles of neurodynamics, the following differential equations for the error function E ( t ) can be designed:
d E ( t ) d t = E ˙ ( t ) = μ Φ ( E ( t ) )
where the scale-valued factor μ is a positive design constant (i.e., μ > 0 ), which can adjust the convergence velocity of this network (5) by varying its value. Φ ( · ) : m × n m × n is an array consisting of the same activation functions. Φ ( · ) can represent different types of activation functions, such as linear-type, bipolar-sigmoid-type, power-type, sign-bi-power-type, and so on. When we use the activation function of the linear type, (5) becomes a typical ZNN.
Based on (5), we can obtain the following implicit dynamic equation:
U ( t ) P ˙ ( t ) P ˙ ( t ) V ( t ) + U ˙ ( t ) P ( t ) P ( t ) V ˙ ( t ) + W ˙ ( t ) = μ Φ U ( t ) P ( t ) P ( t ) V ( t ) + W ( t )
where the starting value of P ( t ) is P ( 0 ) = P 0 R m × n . Additionally, the state matrix P ( t ) is an approximate solution that will gradually approach the theoretical solution, and it will converge to the unique solution of ( 1 ) . In light of the following lemma [12], the performance of the ZNNs in solving the Sylvester equation can be guaranteed to exhibit exponential convergence.
Lemma 1.
Considering the Sylvester Equation (1), with its time-varying coefficient matrices U ( t ) m × m , V ( t ) n × n , and W ( t ) m × n , and assuming that they satisfy inequality (3), if the activation function of the implicit dynamic Equation (6) is chosen to be linear, then its state matrix P ( t ) will exponentially converge from any initial state P 0 to the time-varying ideal solution P * ( t ) of (1).

3. Varying-Factor Finite-Time Recurrent Neural Networks

ZNN can achieve better convergence performance by using different activation functions or setting a larger constant factor [34], but its convergence still has room for improvement. Inspired by classic ZNN design principles, a novel recurrent neural network is proposed in this paper. It can achieve a neural network with super-exponential convergence within a limited time, and because the factor before its activation function changes with time, it is called VFFTRNN. It can be expressed as
E ˙ ( t ) = Γ ( t ) Φ ( E ( t ) ) + Φ ( E λ ( t ) ) = ( t α + α ) Φ ( E ( t ) ) + Φ ( E λ ( t ) )
where Γ ( t ) is a function of t, and it can be set to different functions. For simplicity, we use exponential time-varying parameters ( t α + α ) , where α is a positive scalar-valued factor (i.e., α > 0 ), and the positive scalar-valued factor λ = s q ( 0 , 1 ) , where s and q are odd numbers. In light of (7), we can reformulate its implicit dynamic equation as
U ˙ ( t ) P ( t ) + U ( t ) P ˙ ( t ) P ˙ ( t ) V ( t ) P ( t ) V ˙ ( t ) + W ˙ ( t ) = ( t α + α ) ( Φ U ( t ) P ( t ) P ( t ) V ( t ) + W ( t ) + Φ U ( t ) P ( t ) P ( t ) V ( t ) + W ( t ) λ )
In addition, in light of Theorem 1 and Equation (2) in Section 1, we are able to obtain the vector form of the above implicit dynamic Equation (8):
N ( t ) p ˙ ( t ) = N ˙ ( t ) p ( t ) v e c W ˙ ( t ) ( t α + α ) ( Φ N ( t ) p ( t ) + v e c W ( t ) + Φ N ( t ) p ( t ) + v e c W ( t ) λ )
where N ( t ) : = I n U ( t ) V T ( t ) I m m n × m n is a matrix, p ( t ) : = v e c ( P ( t ) ) m n is a vector, and Φ ( · ) : m n × 1 m n × 1 is an activation function array. It is noteworthy that the transformation dimensions of Φ ( · ) in (9) are different from those in (5). The block diagram of VFFTRNN can be drawn based on the implicit dynamic Equation (8). As shown in Figure 1, ∫ and ∑ represent the integrator and accumulator, respectively. The symbols L and R represent left and right matrix multiplication, respectively.

3.1. Convergence Analysis

The following theorem will lay the foundation for later reference when discussing the global convergence of VFFTRNN.
Theorem 3.
For the time-varying Sylvester Equation (1), its coefficient matrices U ( t ) m × m , V ( t ) n × n , and W ( t ) m × n are given. If Equation (1) satisfies inequality (3), and the activation function sequence is chosen to be strictly increasing, then the state matrix P ( t ) of the VFFTRNN model (8) will globally converge from any initial state P 0 to the time-varying ideal solution P * ( t ) of the Sylvester Equation (1).
Proof. 
First, we can convert the VFFTRNN system (7) into scalar form as follows:
e ˙ i j ( t ) = ( t α + α ) ϕ e i j ( t ) + ϕ e i j λ ( t ) i { 1 , 2 , m } , j { 1 , 2 , n }
where e i j ( t ) represents the i t h row and j− column position element of E ( t ) . In addition, ϕ ( · ) represents the activation subfunction for the matrix elements.
Next, we will prove its global convergence. Let us define a Lyapunov function as shown below:
v i j ( t ) = 1 2 e i j 2 ( t ) , i { 1 , 2 , m } , j { 1 , 2 , n }
According to the above, we can calculate its time derivative as follows:
v ˙ i j ( t ) = e i j ( t ) e ˙ i j ( t ) = ( t α + α ) e i j ϕ e i j ( t ) + ϕ e i j λ ( t )
It can be seen from the previous definition that α > 0 and ϕ ( · ) is monotonic and strictly increasing. Thus, for any t 0 and e i j ( t ) , we have
v ˙ i j ( t ) = 0 e i j ( t ) = 0 0 e i j ( t ) 0 .
In light of Lyapunov stability theory, the error function (10) is globally asymptotically convergent. Therefore, for any i and j in the domain, the error function e i j ( t ) in scalar form will asymptotically converge to zero. In other words, the online solution P ( t ) using VFFTRNN will approach its theoretical solution P * ( t ) when t + .
In conclusion, the proof of Theorem 3 has been completed. That is, VFFTRNN exhibits global asymptotic convergence. □
In the next step, we will demonstrate the finite-time convergence of VFFTRNN under various activation functions (i.e., linear, bipolar-sigmoid, and power functions). See the following theorem for the specific convergence time.
Theorem 4.
For the time-varying Sylvester Equation (1), the coefficient matrices U ( t ) m × m , V ( t ) n × n and W ( t ) m × n have been given. If Equation (1) satisfies the regularity condition (3) and the activation function sequence is monotonic and strictly increasing, then the VFFTRNN system (8) can converge to the theoretical solution P * ( t ) of Equation (1) in finite time from any arbitrary initial state. In addition, the specific convergence time of the neural system ( 7 ) in the cases of different activation functions is as follows:
(1) 
When we choose the linear activation function (i.e., ϕ ( x ) = x ), the convergence time of VFFTRNN t c < t L .
(2) 
When we choose the bipolar-sigmoid activation function (i.e., ϕ ( x ) = ( 1 e x p ( ξ x ) ) / ( 1 + e x p ( ξ x ) ) , ξ > 0 ), the convergence time of VFFTRNN t c < t B .
(3) 
When we choose the power activation function (i.e., ϕ ( x ) = x s , λ s < 1 and s is an odd number), the convergence time of VFFTRNN t c < t P .
The above t L , t B and t P are, respectively,
t L = α + 1 1 λ l n ( χ 1 1 λ + 1 ) α + 1 t B = α + 1 ω ( 1 λ ) l n ( χ 1 1 λ + 1 ) α + 1 i f   χ 1 1 α + 1 ω ( 1 λ ) l n ( 2 ) + ( α + 1 ) ( χ 1 1 ) ω α + 1 o t h e r w i s e . t P = α + 1 2 ( 1 λ s ) χ 1 1 λ s α + 1 i f   χ 1 1 α + 1 2 ( 1 λ s ) + ( α + 1 ) ( χ 1 1 s 1 ) 2 ( 1 s ) α + 1 o t h e r w i s e .
where χ 1 represents m a x ( | e i j ( 0 ) | ) .
Proof. 
In the following, we will prove the convergence time of the VFFTRNN system under three different activation functions, one by one.
(1)
Linear-type: For the linear-type case, the following equation can be obtained from (10).
e ˙ i j ( t ) = ( t α + α ) e i j ( t ) + e i j λ ( t )
In light of differential equation theory [34], we can obtain the solution of (12).
e i j ( t ) = ( e 0 1 λ + 1 ) e x p ( λ 1 ) ( α t + t α + 1 α + 1 ) 1 1 1 λ
where e 0 = e i j ( 0 ) . Let e i j ( t ) in (13) be equal to zero, and we can obtain the following:
α t + t α + 1 α + 1 = l n ( e 0 1 λ + 1 ) 1 λ t α + 1 α + 1
then we can easily find t α + 1 1 λ l n ( e 0 1 λ + 1 ) α + 1 . In other words, for any i and j in the definitional domain, we have that as t α + 1 1 λ l n ( e 0 1 λ + 1 ) α + 1 , the scalar value e i j ( t ) 0 . Note that by the definition of e i j ( t ) above, at this time, the error function E ( t ) in matrix form will also approach zero, which guarantees that the state matrix P ( t ) P * ( t ) .
To sum up, we can obtain the convergence time when the VFFTRNN system uses the linear-type function as follows:
t L = α + 1 1 λ l n ( χ 1 1 λ + 1 ) α + 1
where χ 1 = m a x ( | e i j ( 0 ) | ) . This shows that in the case where the activation function is linear, the VFFTRNN method can achieve finite-time convergence in solving the Sylvester equation.
(2)
Bipolar-Sigmoid-Type: In this case, we can easily see that the activation function is ϕ ( x ) = ( 1 e x p ( ξ x ) ) / ( 1 + e x p ( ξ x ) ) , where ϕ 1 and x R are scalar factors. The following equation can be obtained from (10).
e ˙ i j ( t ) = ( t α + α ) ( R ( t ) + Q ( t ) ) = ( t α + α ) ( 1 exp ( ξ e i j ( t ) ) 1 + exp ( ξ e i j ( t ) ) + 1 exp ( ξ e i j λ ( t ) ) 1 + exp ( ξ e i j λ ( t ) ) )
where R e i j ( t ) = 1 exp ( ξ e i j ( t ) ) 1 + exp ( ξ e i j ( t ) ) , Q e i j ( t ) = 1 exp ( ξ e i j λ ( t ) ) 1 + exp ( ξ e i j λ ( t ) ) . In light of the relevant theory of differential equations, we can rewrite (15) as follows:
d y d t = ( t α + α ) R ( y ) + Q ( y )
where y = e i j ( t ) . According to the Lyapunov function of this system, we know that e i j ( t ) either increases or decreases monotonically as t increases. For simplicity, we will next discuss the case where e i j ( 0 ) = y 0 0 . In this case, we can derive the following:
0 t ( t α + α ) d t = y 0 0 1 R ( y ) + Q ( y ) d y
At this point, we consider the following two situations.
a.
When y 0 1 , noticing that R ( y ) and Q ( y ) are convex, solving (16) gives
α t + t α + 1 α + 1 = 0 y 0 1 R ( y ) + Q ( y ) d y 1 ω 0 y 0 1 y + y λ d y = 1 ω ( 1 λ ) l n ( y 0 1 λ + 1 )
where ω = R ( 1 ) = Q ( 1 ) = ( 1 e x p ( ξ ) ) / ( 1 + e x p ( ξ ) ) is a constant, determined by the scalar factor ξ . Similar to linear-type, solving for (17), we obtain
t α + 1 ω ( 1 λ ) l n ( y 0 1 λ + 1 ) α + 1
b.
When y 0 > 1 , based on Equations (16) and (17), Equation (19) be derived:
α t + t α + 1 α + 1 = 0 1 1 R ( y ) + Q ( y ) d y + 1 y 0 1 R ( y ) + Q ( y ) d y 1 ω ( 1 λ ) l n 2 + 1 y 0 1 R ( y ) + Q ( y ) d y
Notice that R ( y ) and Q ( y ) are monotonically increasing over the interval y > 0 , so for y > 1 , R ( y ) > R ( 1 ) > ω and Q ( y ) > Q ( 1 ) > ω . Therefore, we have
1 y 0 1 R ( y ) + Q ( y ) d y 1 y 0 1 ω d y = y 0 1 ω
Substituting (20) into (19) and solving, we obtain
t α + 1 ω ( 1 λ ) l n 2 + ( α + 1 ) ( y 0 1 ) ω α + 1
In light of the results (18) and (21) from the two situations, it can be concluded that when t α + 1 ω ( 1 λ ) l n ( y 0 1 λ + 1 ) α + 1 or t α + 1 ω ( 1 λ ) l n 2 + ( α + 1 ) ( y 0 1 ) ω α + 1 , the scalar value e i j ( t ) 0 for any i and j in the defined domain. This means that the convergence of the state matrix P ( t ) P * ( t ) is guaranteed. Thus, the convergence time for the bipolar-sigmoid-type activation function can be expressed as
t < α + 1 ω ( 1 λ ) l n ( χ 1 1 λ + 1 ) α + 1 if   χ 1 1 α + 1 ω ( 1 λ ) l n 2 + ( α + 1 ) ( χ 1 1 ) ω α + 1 if   otherwise . = t B
(3)
Power-Type: For the power-type case, the activation function is defined as ϕ ( x ) = x s , where s is an odd number and λ s < 1 . In this case, the following equation can be derived from (10).
e ˙ i j ( t ) = ( t α + α ) e i j s ( t ) + e i j λ s ( t )
  • Similar to (16), we have
    t α + 1 α + 1 + α t = 0 y 0 1 y s + y λ s d y
  • The discussion is divided into the following two situations:
a.
When y 0 1 , it can be found that 1 > λ s > s , so y s > y λ s . In light of (24), we can obtain
t α + 1 α + 1 + α t 0 y 0 1 2 y λ s d y = 1 2 ( 1 λ s ) y 0 1 λ s
Then, the result (25) can be obtained:
t < α + 1 2 ( 1 λ s ) y 0 1 λ s α + 1
b.
When y 0 > 1 , we know that for all y > 1 , we have y p < y λ s . So similarly,
t α + 1 α + 1 + α t 0 1 1 2 y λ s d y + 1 y 0 1 2 y s d y = 1 2 ( 1 λ s ) + y 0 1 s 1 2 ( 1 s )
Then, the following result (26) can be obtained:
t < α + 1 2 ( 1 λ s ) + ( α + 1 ) ( y 0 1 s 1 ) 2 ( 1 s ) α + 1
In conclusion, when t α + 1 2 ( 1 λ s ) y 0 1 λ s α + 1 or t α + 1 2 ( 1 λ s ) + ( α + 1 ) ( y 0 1 s 1 ) 2 ( 1 s ) α + 1 , the scalar value e ( i j ( t ) 0 for any i and j on the definitional domain. This means that the convergence of state matrix P ( t ) P * ( t ) is guaranteed. The convergence time for the power-type activation function can be represented as
t < α + 1 2 ( 1 λ s ) χ 1 1 λ s α + 1 if   χ 1 1 α + 1 2 ( 1 λ s ) + ( α + 1 ) ( χ 1 1 s 1 ) 2 ( 1 s ) α + 1 if   otherwise . = t P
In light of the above analysis, it can be concluded that VFFTRNN achieves finite-time convergence in solving the Sylvester Equation (1), regardless of the choice of activation function. Therefore, the proof of Theorem 4 on finite-time convergence is completed. □
Remark 1.
Regardless of the chosen activation function, the convergence time of VFFTRNN is determined by the initial error E ( 0 ) (i.e., E ( 0 ) = U ( 0 ) P ( 0 ) + P ( 0 ) V ( 0 ) W ( 0 ) ) and the design factors α and λ .
Next, we will analyze the influence of the factors α and λ on the convergence time for different activation functions.
(1)
Linear-Type:
a.
For α , in light of (14), t L can be rewritten as
t L = ( χ 2 α 1 ) 1 α 1
where χ 2 = ln ( χ 1 1 λ + 1 ) 1 λ > 0 and α 1 = α + 1 > 1 . Then, let us take the partial derivative with respect to α 1 .
t L α 1 = ( χ 2 α 1 ) 1 α 1 1 ln ( χ 2 α 1 ) α 1 2
We know that if α 1 is large enough, 1 l n ( χ 2 α 1 ) will become less than zero. Therefore, for t L α 1 , there are two possible situations:
t L α 1 < 0 α 1 > 1 t L α 1 > 0 α 1 > 1
In the first situation, t L decreases as α 1 increases. In other words, as α increases, t L decreases simultaneously. Similarly, in the second situation, t L will first increase and then decrease as α increases. This means that when α reaches a specific value, determined by the initial error E ( 0 ) , t L will attain its maximum.
b.
For λ , in light of (14), it can easily be seen that we only need to analyze ln ( χ 1 1 λ + 1 ) 1 λ . Next, take the partial derivative of this expression:
η λ 1 = λ 1 χ 1 λ 1 ln χ 1 ( χ 1 λ 1 + 1 ) ln ( χ 1 λ 1 + 1 ) λ 1 2
where η = ln ( χ 1 1 λ + 1 ) 1 λ and λ 1 = 1 λ ( 0 , 1 ) . Since the denominator of (30) is always positive, we only need to analyze the numerator:
λ 1 χ 1 λ 1 ln χ 1 ( χ 1 λ 1 + 1 ) ln ( χ 1 λ 1 + 1 )
If χ 1 1 , then λ 1 χ 1 λ 1 ln χ 1 0 , and λ 1 χ 1 λ 1 ln χ 1 ( χ 1 λ 1 + 1 ) ln ( χ 1 λ 1 + 1 ) < 0 . Therefore, the value of (30) is always less than zero. If χ 1 > 1 , let χ 1 λ 1 = λ 2 > 1 , then λ 1 = ln λ 2 ln χ 1 . The numerator can be rewritten as follows:
λ 1 χ 1 λ 1 ln χ 1 ( χ 1 λ 1 + 1 ) ln ( χ 1 λ 1 + 1 ) = λ 2 ln λ 2 ( λ 2 + 1 ) ln ( λ 2 + 1 )
Note that the function f ( x ) = x ln x is monotonically increasing when x > 1 . Therefore, the value of Equation (32) is always less than zero. In other words, the value of Equation (30) is always less than zero.
Based on the above discussion, we can conclude that for any λ 1 , it is always true that η λ 1 < 0 . Considering the relationship between λ 1 and λ , we can conclude that when λ is larger, t L is smaller.
(2)
Bipolar-Sigmoid-Type:
a.
For α , in light of (22), t B can be rewritten as
t B = ( χ 3 α 1 ) 1 α 1
where χ 3 = ln ( χ 1 1 λ + 1 ) ω ( 1 λ ) > 0 , χ 1 1 ln 2 ω ( 1 λ ) + χ 1 1 ω > 0 , χ 1 > 1 . Similar to the analysis of the linear-type above, there are two situations:
t B α 1 < 0 α 1 > 1 t B α 1 > 0 α 1 > 1
For the first situation, when α increases, t B decreases. For the second situation, there exists a specific value where t B is maximized.
b.
For λ , by analyzing the structure of Equation (22), we focus on ln ( χ 1 1 λ + 1 ) 1 λ when χ 1 1 and ln 2 1 λ when χ 1 > 1 . As λ increases, the value of ln 2 1 λ increases, Based on the derivation in Section 1), it is evident that for any λ ( 0 , 1 ) , as λ increases, t B also increases.
In addition, consider the influence of the factor ξ on the convergence time t B . We know that ω = 1 e ξ 1 + e ξ , and as ξ increases, ω also increases. Furthermore, according to Equation (22), the larger ω is, the smaller t B becomes.
(3)
Power-Type:
a.
For α , let χ 4 = χ 1 1 λ s 2 ( 1 λ s ) > 0 , χ 1 1 1 2 ( 1 λ s ) + 1 χ 1 s 1 2 ( s 1 ) > 0 , χ 1 > 1 , then we have
t P = ( χ 4 α 1 ) 1 α 1
Similar to the analysis of the linear-type case above, there are two possible situations:
t P α 1 < 0 α 1 > 1 t P α 1 > 0 α 1 > 1
For the first situation, when α increases, t P decreases. For the second situation, there exists a certain value where t P is maximized.
b.
For λ , according to (27), we can conclude that the larger λ is, the larger t P becomes, especially when χ 1 > 0 . In addition, when χ 1 1 , we only need to consider χ 1 λ s 1 λ s . Let λ 3 = 1 λ s ( 0 , 1 ) , then
λ 3 χ 1 λ 3 λ 3 = χ 1 λ 3 ( λ 3 ln χ 1 1 ) λ 3 2 < 0
So in conclusion, it is clear that for λ ( 0 , 1 ) , as λ increases, t P increases.
In light of the above derivation, the following conclusions can be drawn:
(1)
For α , its relationship to the convergence time can always be divided into two cases. In the first case, as α increases, the convergence time decreases. In the second case, the convergence time first increases and then decreases. It is worth noting that even when α = 0 , the VFFTRNN system will still converge in finite time.
(2)
For λ , the convergence time always increases with the increase in λ , regardless of the chosen activation function. Clearly, the factor λ accelerates the convergence process of the proposed VFFTRNN. If λ = 1 , the proposed VFFTRNN becomes VP-CDNN. Furthermore, if the factor α = 0 , the proposed VFFTRNN reduces to ZNN.

3.2. Robustness Analysis

As we all know, in practical applications, circumstances are rarely perfect. Due to various reasons, such as errors in digital implementations or higher-order residues from circuit elements (e.g., diodes) in hardware implementations, many factors continuously interfere with the VFFTRNN model. Among these, three common negative factors include coefficient matrix disturbances, differentiation errors, and model implementation errors. Next, we will discuss the robustness of the VFFTRNN model in the presence of these negative factors.
(1)
Coefficient Matrix Perturbation:
Theorem 5.
If uncharted, smooth coefficient matrix perturbations Δ U ( t ) m × m , Δ V ( t ) n × n and Δ W ( t ) m × n exist in the VFFTRNN model, and if they satisfy the following conditions:
| | N 1 ( t ) | | F υ 1 , | | N ^ 1 ( t ) | | F υ 2 | | W ( t ) | | F ι 1 , | | Δ W ( t ) | | F ι 2 , t [ 0 , + )
where the mass matrices are defined as N 1 ( t ) : = I n U ( t ) V ( t ) T I m 1 , N ^ 1 ( t ) : = I n ( U ( t ) + Δ U ( t ) ) ( V ( t ) + Δ V ( t ) ) T I m 1 . Therefore, the calculation error | | P ˜ ^ ( t ) | | F = | | P ^ ( t ) P * ( t ) | | F is bounded.
Proof. 
The following proof uses the linear activation function as an example. First, we define a variable P ˜ ( t ) = P ( t ) P * ( t ) which represents the error in the state matrix P ( t ) P * ( t ) . Then, its Frobenius norm can be written as
| | P ˜ ( t ) | | F = i = 1 m j = 1 n p ˜ i j 2 ( t ) = | | p ˜ ( t ) | | 2
where p ˜ ( t ) : = v e c P ˜ ( t ) m n × 1 . Furthermore, in light of (1), we can obtain an expression of W:
W ( t ) = U ( t ) P * ( t ) + P * ( t ) V ( t )
Substitute (36) into (4):
E ( t ) = U ( t ) P ˜ ( t ) P ˜ ( t ) V ( t )
In light of Theorem 1, we can acquire its vector form:
e ( t ) = N ( t ) p ˜ ( t ) = N ( t ) p ( t ) p * ( t )
where e ( t ) : = v e c E ( t ) and p * ( t ) : = v e c ( P * ( t ) ) . From Theorem 2 and inequality (3), it is clear that the matrix N ( t ) is invertible. Therefore, we conclude that
| | P ˜ ( t ) | | F = | | p ˜ ( t ) | | 2 | | N 1 ( t ) | | F | | e ( t ) | | 2 υ 1 i = 1 m j = 1 n e i j 2 ( t )
Substituting (13) into (38), we can obtain
| | P ˜ ( t ) | | F υ 1 i = 1 m j = 1 n ( ( e i j 1 λ ( 0 ) + 1 ) e ( λ 1 ) ( α t + t α + 1 α + 1 ) 1 ) 2 1 λ υ 1 m n ( ( χ 1 1 λ + 1 ) e ( λ 1 ) ( α t + t α + 1 α + 1 ) 1 ) 1 1 λ υ 1 τ 1 e ( α t + t α + 1 α + 1 )
where τ 1 = m n ( χ 1 1 λ + 1 ) 1 1 λ > 0 . The results demonstrate that the VFFTRNN model exhibits super-exponential convergence in solving the Sylvester equation when employing a linear-type activation function.
Next, we consider the case where perturbations in the coefficient matrices exist. Based on Equation (8), it is straightforward to derive the implicit dynamic equation of the VFFTRNN system under the influence of these perturbations.
U ^ ˙ ( t ) P ^ ( t ) + U ^ ( t ) P ^ ˙ ( t ) P ^ ˙ ( t ) V ^ ( t ) P ^ ( t ) V ^ ˙ ( t ) + W ^ ˙ ( t ) = ( t α + α ) ( Φ ( U ^ ( t ) P ^ ( t ) P ^ ( t ) V ^ ( t ) + W ^ ( t ) ) + Φ ( U ^ ( t ) P ^ ( t ) P ^ ( t ) V ^ ( t ) + W ^ ( t ) ) λ )
where U ^ ( t ) = U ( t ) + Δ U ( t ) , V ^ ( t ) = V ( t ) + Δ V ( t ) and W ^ ( t ) = W ( t ) + Δ W ( t ) . P ^ ( t ) denotes the solution to the VFFTRNN system with perturbation.
Furthermore, in light of (36), its vector form can be written as
w ( t ) = N ( t ) p * ( t )
where w ( t ) : = v e c ( W ( t ) ) . Then, we can acquire
p * ( t ) = N 1 ( t ) w ( t )
Therefore, we can derive the inequality for the theoretical solution P * ( t ) .
| | P * ( t ) | | F = | | p * ( t ) | | 2 = | | N 1 ( t ) w ( t ) | | 2 | | N 1 ( t ) | | F | | w ( t ) | | F υ 1 ι 1
Let us assume that the theoretical solution of the VFFTRNN system with perturbation is P ^ * ( t ) . Similar to (40), we can easily deduce that
| | P ^ * ( t ) | | F | | N ^ 1 ( t ) | | F | | W ^ ( t ) | | F υ 2 ( | | W ( t ) | | F + | | Δ W ( t ) | | F ) υ 2 ( ι 1 + ι 2 )
In addition, similar to (39), we can obtain
| | P ^ ( t ) P ^ * ( t ) | | F υ 2 τ 2 e ( α t + t α + 1 α + 1 )
where τ 2 = m n ( χ 5 1 λ + 1 ) 1 1 λ and χ 5 = max | ( U ^ ( 0 ) P ^ ( 0 ) P ^ ( 0 ) V ^ ( 0 ) + W ^ ( 0 ) ) i j | .
Then, based on (41), (42) and (43), we can derive an upper bound for the computation error | | P ˜ ^ ( t ) | | F .
| | P ˜ ^ ( t ) | | F = | | P ^ ( t ) P * ( t ) | | F | | P ^ ( t ) P ^ * ( t ) | | F + | | P ^ * ( t ) P * ( t ) | | F υ 2 τ 2 e ( α t + t α + 1 α + 1 ) + | | P ^ * ( t ) | | F + | | P * ( t ) | | F υ 2 τ 2 e ( α t + t α + 1 α + 1 ) + υ 1 ι 1 + υ 2 ( ι 1 + ι 2 )
In light of the above analysis, we conclude that when using a linear-type activation function, the computation error | | P ˜ ^ ( t ) | | F of the VFFTRNN system with perturbation is bounded. Additionally, it is worth mentioning that the solution process exhibits a super-exponential convergence rate.
Similarly, for sigmoid or power activation functions, it is not difficult to prove that the computation error of the VFFTRNN system with perturbation is bounded. Due to the limited space of this paper, the analysis for these two activation functions will not be presented here.
Thus, the proof of the robustness of Theorem 5 regarding matrix perturbation is complete. □
(2)
Differentiation and Model Implementation Errors:
In the process of hardware implementation, dynamics implementation errors and differential errors related to U ( t ) , V ( t ) and W ( t ) are inevitable. These errors are collectively referred to as model implementation errors [35]. In this section, we analyze the robustness of the VFFTRNN system in the presence of these errors. Let us assume that the differential errors for the time derivatives of the matrices U ˙ ( t ) and V ˙ ( t ) are Λ ( t ) m × m and Ω ( t ) n × n , respectively, and the model implementation error is denoted as Δ ( t ) R m × n . Then, based on Equation (8), the implicit dynamic equation for the VFFTRNN system with these errors can be written as
U ( t ) P ˙ ( t ) + ( U ˙ ( t ) + Λ ( t ) ) P ( t ) P ˙ ( t ) V ( t ) P ( t ) ( V ˙ ( t ) + Ω ( t ) ) + W ˙ ( t ) Δ ( t ) = ( t α + α ) ( Φ ( U ( t ) P ( t ) P ( t ) V ( t ) + W ( t ) ) + Φ ( U ( t ) P ( t ) P ( t ) V ( t ) + W ( t ) ) λ )
Theorem 6.
If there exist unknown smooth differentiation errors Λ ( t ) , Ω ( t ) and a model implementation error Δ ( t ) in the VFFTRNN model, which satisfy the following conditions:
| | Λ ( t ) | | F ε 1 | | Ω ( t ) | | F ε 2 | | Δ ( t ) | | F ε 3 | | N 1 ( t ) | | F υ 1 | | W ( t ) | | F ι 1 , t [ 0 , + )
then its computation error | | P ( t ) P * ( t ) | | F is bounded.
Proof. 
Let us rewrite (45) from matrix form to vector form for easier analysis.
N ( t ) p ˙ ( t ) = ( N ˙ ( t ) + Ψ ( t ) ) p ( t ) w ˙ ( t ) + δ ( t ) ( t α + α ) ( Φ ( N ( t ) p ( t ) + w ( t ) ) + Φ ( N ( t ) p ( t ) + w ( t ) ) λ )
where Ψ ( t ) : = I n Λ ( t ) Ω T ( t ) I m and δ ( t ) : = v e c ( Δ ( t ) ) . In addition, we can easily observe that e ( t ) = N ( t ) p ( t ) + w ( t ) , and its time derivative is given by e ˙ ( t ) = N ˙ ( t ) p ( t ) + N ( t ) p ˙ ( t ) + w ˙ ( t ) . Therefore, Equation (46) can be rewritten as
e ˙ ( t ) = ( t α + α ) Φ ( e ( t ) ) + Φ ( e λ ( t ) ) + δ ( t ) Ψ ( t ) p ( t )
In light of (37), we can acquire
x ( t ) = N 1 ( t ) e ( t ) + p * ( t )
Substituting (47) into (48), we have
e ˙ ( t ) = ( t α + α ) Φ ( e ( t ) ) + Φ ( e λ ( t ) ) + H ( t ) e ( t ) + K ( t )
where H ( t ) = Ψ ( t ) N 1 ( t ) m n × m n , as well as K ( t ) = δ ( t ) Ψ ( t ) p * ( t ) m n × 1 .
Next, we construct a Lyapunov function candidate in the form v : = 1 2 e T ( t ) e ( t ) , and its time derivative can be computed as follows:
v ˙ = e T ( t ) e ˙ ( t ) = ( t α + α ) e T ( t ) Φ ( e ( t ) ) + Φ ( e λ ( t ) ) + e T ( t ) H ( t ) e ( t ) + e T ( t ) K ( t )
By observing the composition of (50) and (41), we can derive the following:
e T ( t ) H ( t ) e ( t ) | | e T ( t ) | | 2 | | H ( t ) | | F | | e ( t ) | | 2 e T ( t ) e ( t ) | | H ( t ) | | F | | Ψ ( t ) | | F | | M 1 ( t ) | | F e T ( t ) e ( t ) υ 1 ( | | I n Λ ( t ) | | F + | | Ω T ( t ) I m | | F ) e T ( t ) e ( t ) υ 1 ( ε 1 n + ε 2 m ) i = 1 m j = 1 n e i j 2 ( t )
and
e T ( t ) K ( t ) | | e T ( t ) | | 2 | | K ( t ) | | F | | e T ( t ) | | 2 ( | | Δ ( t ) | | F + | | Ψ ( t ) | | F | | P * ( t ) | | F ) ε 3 + ( ε 1 n + ε 2 m ) υ 1 ι 1 i = 1 m j = 1 n e i j 2 ( t ) ε 3 + ( ε 1 n + ε 2 m ) υ 1 ι 1 i = 1 m j = 1 n | e i j ( t ) |
By substituting the above two equations into (50), we can obtain the following:
v ˙ i = 1 m j = 1 n | e i j ( t ) | ( ( t α + α ) ϕ ( | e i j ( t ) | ) + ϕ ( | e i j λ ( t ) | ) ζ 1 υ 1 | e i j ( t ) | ζ 2 ) = Y ( t )
where ζ 1 = ε 1 n + ε 2 m and ζ 2 = ε 3 + ζ 1 υ 1 ι 1 .
We know that ϕ ( | e i j ( t ) | ) | e i j ( t ) | for any t, and that t α + α increases as t increases. Therefore, there always exists a value t 0 , such that for all t > t 0 , ( t α + α ) ϕ ( | e i j ( t ) | ) + ϕ ( | e i j λ ( t ) | ) ζ 1 υ 1 | e i j ( t ) | ζ 2 ) > 0 . Thus, for Equation (51), we need to consider the following two cases:
α ( ϕ ( | e i j ( 0 ) | ) + ϕ ( | e i j λ ( 0 ) | ) ) ζ 1 υ 1 | e i j ( 0 ) | ζ 2 0 i { 1 , , m } , j { 1 , , n } α ( ϕ ( | e i j ( 0 ) | ) + ϕ ( | e i j λ ( 0 ) | ) ) ζ 1 υ 1 | e i j ( 0 ) | ζ 2 < 0 i { 1 , , m } , j { 1 , , n }
(1)
For the first situation, it is easy to see that Y ( t ) < 0 , i.e., v ˙ 0 . Therefore, the error e ( t ) will decrease monotonically as time t increases, which indicates that P ( t ) will eventually converge to P * ( t ) as time increases, i.e., | | P ( t ) P * ( t ) | | F will approach 0.
(2)
For the second situation, due to the uncertainty of the sign of Y ( 0 ) , we need to further subdivide it into two cases. If Y ( 0 ) 0 , then the analysis follows the same reasoning as in case 1. Moreover, if Y ( 0 ) > 0 , there exists a time t 0 such that Y ( t ) 0 for all t > t 0 . This indicates that the error e ( t ) will initially increase and then decrease. Based on the results in [36], we can obtain the upper bound of | | P ˜ ( t ) | | F = | | P ( t ) P * ( t ) | | F when using a linear activation function:
| | P ˜ ( t ) | | F υ 1 ζ 2 ( m n + m n ) 2 ( ρ ( t α + α ) υ 1 ζ 1 )
Since ζ 2 > 0 and ( m n + m n ) > 0 , it is guaranteed that (52) holds true. The design factor ρ should be set greater than υ 1 ζ 2 α in order for the denominator of the fraction to be positive. Therefore, as t + , the computation error | | P ˜ ( t ) | | F will approach 0.
Summing up the above, as time increases, the computation error | | P ˜ ( t ) | | F tends to zero in both cases. The difference is that in the first situation, the computation error | | P ˜ ( t ) | | F continuously decreases, whereas in the second situation, the computation error | | P ˜ ( t ) | | F first increases and then decreases. Additionally, there exists an upper bound given by υ 1 ζ 2 ( m n + m n ) 2 ( ρ ( t α + α ) υ 1 ζ 1 ) . □

4. Illustrative Example

This section presents simulations using MATLAB 2021b for solving the Sylvester equation with the proposed VFFTRNN. All experiments are conducted on a personal computer equipped with a 64-bit operating system, 16.00 GB RAM, and an Intel(R) Core(TM) i5-12600KF Processor (20M Cache, up to 4.90 GHz).
We will verify the performance of the proposed VFFTRNN and compare it with ZNN by providing a specific example. The coefficients of the time-varying Sylvester Equation (1) are set as follows:
U ( t ) = sin t 10 cos t 10 cos t 10 sin t 10 , V ( t ) = sin t 100 0 0 cos t 50 W ( t ) = sin 2 t 10 1 cos 2 t 5 sin t cos t 10 sin t cos t 5 1 .
Then, its theoretical solution P * ( t ) can be easily obtained:
P * ( t ) = 10 sin t 10 cos t 10 cos t 10 sin t .
In addition, we write P ( t ) solved by the VFFTRNN in the following form:
P ( t ) = p 11 p 12 p 21 p 22
where p i j denotes the i t h row and j column element of P ( t ) . For the sake of discussion, let us set the initial state P 0 as follows:
P 0 = p a p b p c p d
where p a , p b , p c and p d are numbers generated by MATLAB which lie within the range [ 25 , 25 ] .
Generally, the system is considered to have converged in solving the Sylvester Equation (1) when the computation error | | P ( t ) P * ( t ) | | F is less than 0.1 % of the range (i.e., | | P ( t ) P * ( t ) | | 50 × 0.1 % = 0.05 ). In other words, we only need to measure the time when the calculation error | | P ( t ) P * ( t ) | | F falls below 0.05, which allows us to determine the convergence time.

4.1. Convergence Discussion

In this section, we will discuss the convergence performance of the VFFTRNN and ZNN models in solving the Sylvester equation under various conditions.
First, we observe the effects of different factors on the convergence performance of ZNN and VFFTRNN with the same activation function. Table 1 presents a comparison of the convergence times of ZNN and VFFTRNN under various design factors.
Figure 2 and Figure 3 show detailed simulation results of both models under different conditions for α = μ = 0.1 , α = μ = 1 , and α = μ = 5 . As seen, when μ is sufficiently small (e.g., μ = 0.01 ), ZNN fails to converge. However, with μ = 0.1 , the convergence time of ZNN is 67.55 s. As the design factor μ increases, the convergence performance of ZNN improves gradually. From Figure 2, we observe that when μ = 1 , the computation error | | P ( t ) P * ( t ) | | F converges to zero in 6.98 s, nearly ten times faster than with μ = 0.1 . When μ = 5 , the convergence time further decreases to just 1.37 s.
Next, we examine the convergence time of the computation error using VFFTRNN. With α = 0.01 or α = 0.1 , the computation error | | P ( t ) P * ( t ) | | F converges to zero in just 2.39 or 2.35 s, which is significantly faster than ZNN.
Furthermore, the comparison results in Table 1 and Figure 2 and Figure 3 clearly show that the convergence performance of VFFTRNN is significantly better than that of ZNN. It is worth noting that these simulation results use a linear-type activation function.
Second, we observed the effects of different activation functions on the convergence performance of ZNN and VFFTRNN when the same factors were set.
Figure 4 and Figure 5 show the computation error | | P ( t ) P * ( t ) | | F curves for ZNN and VFFTRNN under three activation functions (i.e., linear, bipolar-sigmoid, and power-type).
When using the linear activation function, it can be observed that ZNN takes 6.98 s to converge to zero, whereas VFFTRNN converges in just 1.05 s. Similarly, when the bipolar-sigmoid or power activation functions are used, the convergence time of VFFTRNN is significantly shorter than that of ZNN.
This demonstrates the superior performance of the VFFTRNN system compared to ZNN and also verifies its finite-time convergence property for solving the Sylvester equation under different activation functions. It is worth noting that the factors were set to μ = α = 1 and λ = 1 5 in the above simulation results.

4.2. Robustness Discussion

In this section, we will verify the robustness of the VFFTRNN and compare it with the ZNN. To ensure both generality and simplicity, we consider the effects of differentiation and model implementation errors. For the simulation, we set the differentiation and model implementation errors as follows:
Λ ( t ) = β 1 cos 2 t 0 0 sin 2 t , Ω ( t ) = β 2 0 cos 2 t sin 2 t 0 Δ ( t ) = β 3 sin 2 t 0 0 cos 2 t .
where β 1 , β 2 and β 3 are adjustable factors that control the magnitude of these disturbances.
In this section, we will compare the robustness of VFFTRNN and ZNN using the four cases shown in Figure 6. From Figure 6, it can be observed that when β 1 = β 2 = β 3 , the disturbances are small, but the computation error | | P ( t ) P * ( t ) | | F of ZNN does not converge to zero. In contrast, for VFFTRNN, we can see that the computation error converges within approximately 4 s.
Examining the other three cases, it is evident that the computation error of VFFTRNN still converges to zero, even with much larger disturbances. Notably, the convergence time of VFFTRNN decreases as the design factor α increases or λ decreases.
In conclusion, although the presence of disturbances can affect the convergence performance of VFFTRNN, it still converges to zero in finite time, regardless of the size of the disturbance, confirming the robustness of the model.

5. Conclusions

In this paper, we propose a novel neural network called VFFTRNN and investigate its finite-time convergence and robustness for solving the Sylvester equation under three different activation functions. Compared to the conventional ZNN, the proposed VFFTRNN introduces a time-varying coefficient and a fractional power term, which accelerate the convergence speed. More importantly, VFFTRNN not only guarantees finite-time convergence but also achieves a super-exponential convergence rate. Simulation results demonstrate the superior performance of the proposed neural network in solving the Sylvester equation. Therefore, VFFTRNN offers significant advantages for real hardware implementation.
In the future, for hardware implementation, we will consider using FPGA to implement VFFTRNN networks. The flexibility and parallel processing capabilities of FPGA make it highly suitable for real-time applications, enabling efficient computation of network operations. This method will help achieve high performance and low power consumption, making it a feasible solution for the practical deployment of VFFTRNN networks.

Author Contributions

Methodology, H.T.; conceptualization, J.W.; software, H.G.; validation, Z.Z.; formal analysis, L.T.; investigation, Q.Z.; project administration, C.L. All authors have read and agreed to published version of this manuscript.

Funding

This work was supported in part by the Jiangxi Provincial Key Laboratory of Intelligent Systems and Human–Machine Interaction under grant 2024SSY03121, and in part by the Science and Technology Department of Jiangxi Province of China under grants 20204ABC03A39, 20161ACB21007, 20171BBE50071, and 20171BAB202033.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wei, Q.; Dobigeon, N.; Tourneret, J.Y.; Bioucas-Dias, J.; Godsill, S. R-FUSE: Robust fast fusion of multiband images based on solving a Sylvester equation. IEEE Signal Process. Lett. 2016, 23, 1632–1636. [Google Scholar] [CrossRef]
  2. Wang, L.; Li, D.; He, T.; Xue, Z. Manifold regularized multi-view subspace clustering for image representation. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 283–288. [Google Scholar]
  3. Hu, H.; Lin, Z.; Feng, J.; Zhou, J. Smooth representation clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 3834–3841. [Google Scholar]
  4. Jang, J.S.; Lee, S.Y.; Shin, S.Y. An optimization network for matrix inversion. In Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 1987. [Google Scholar]
  5. Fa-Long, L.; Zheng, B. Neural network approach to computing matrix inversion. Appl. Math. Comput. 1992, 47, 109–120. [Google Scholar] [CrossRef]
  6. Cichocki, A.; Unbehauen, R. Neural networks for solving systems of linear equations and related problems. IEEE Trans. Circuits Syst. I Fundam. Theory Appl. 1992, 39, 124–138. [Google Scholar] [CrossRef]
  7. Xia, Y.; Wang, J. A recurrent neural network for solving linear projection equations. Neural Netw. 2000, 13, 337–350. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Wang, J. Recurrent neural networks for nonlinear output regulation. Automatica 2001, 37, 1161–1173. [Google Scholar] [CrossRef]
  9. Li, S.; Li, Y. Nonlinearly activated neural network for solving time-varying complex Sylvester equation. IEEE Trans. Cybern. 2013, 44, 1397–1407. [Google Scholar] [CrossRef] [PubMed]
  10. Xiao, L.; Liao, B.; Luo, J.; Ding, L. A convergence-enhanced gradient neural network for solving Sylvester equation. In Proceedings of the 2017 36th Chinese Control Conference (CCC), Dalian, China, 26–28 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 3910–3913. [Google Scholar]
  11. Zhang, Z.; Zheng, L.; Weng, J.; Mao, Y.; Lu, W.; Xiao, L. A new varying-parameter recurrent neural-network for online solution of time-varying Sylvester equation. IEEE Trans. Cybern. 2018, 48, 3135–3148. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Jiang, D.; Wang, J. A recurrent neural network for solving Sylvester equation with time-varying coefficients. IEEE Trans. Neural Netw. 2002, 13, 1053–1063. [Google Scholar] [CrossRef]
  13. Yan, X.; Liu, M.; Jin, L.; Li, S.; Hu, B.; Zhang, X.; Huang, Z. New zeroing neural network models for solving nonstationary Sylvester equation with verifications on mobile manipulators. IEEE Trans. Ind. Inform. 2019, 15, 5011–5022. [Google Scholar] [CrossRef]
  14. Zhang, Z.; Zheng, L. A complex varying-parameter convergent-differential neural-network for solving online time-varying complex Sylvester equation. IEEE Trans. Cybern. 2018, 49, 3627–3639. [Google Scholar] [CrossRef]
  15. Deng, J.; Li, C.; Chen, R.; Zheng, B.; Zhang, Z.; Yu, J.; Liu, P.X. A Novel Variable-Parameter Variable-Activation-Function Finite-Time Neural Network for Solving Joint-Angle Drift Issues of Redundant-Robot Manipulators. IEEE/ASME Trans. Mechatron. 2024, 1–12. [Google Scholar] [CrossRef]
  16. Xu, X.; Sun, J.; Endo, S.; Li, Y.; Benjamin, S.C.; Yuan, X. Variational algorithms for linear algebra. Sci. Bull. 2021, 66, 2181–2188. [Google Scholar] [CrossRef] [PubMed]
  17. Li, Y.; Chen, W.; Yang, L. Multistage linear gauss pseudospectral method for piecewise continuous nonlinear optimal control problems. IEEE Trans. Aerosp. Electron. Syst. 2021, 57, 2298–2310. [Google Scholar] [CrossRef]
  18. Wang, X.; Liu, J.; Qiu, T.; Mu, C.; Chen, C.; Zhou, P. A real-time collision prediction mechanism with deep learning for intelligent transportation system. IEEE Trans. Veh. Technol. 2020, 69, 9497–9508. [Google Scholar] [CrossRef]
  19. Zhang, Z.; Li, S.; Zhang, X. Simulink comparison of varying-parameter convergent-differential neural-network and gradient neural network for solving online linear time-varying equations. In Proceedings of the 2016 12th World Congress on Intelligent Control and Automation (WCICA), Guilin, China, 12–15 June 2016; pp. 887–894. [Google Scholar]
  20. Zhang, Y.; Chen, D.; Guo, D.; Liao, B.; Wang, Y. On exponential convergence of nonlinear gradient dynamics system with application to square root finding. Nonlinear Dyn. 2015, 79, 983–1003. [Google Scholar] [CrossRef]
  21. Wang, S.; Dai, S.; Wang, K. Gradient-based neural network for online solution of lyapunov matrix equation with li activation function. In Proceedings of the 4th International Conference on Information Technology and Management Innovation, Shenzhen, China, 12–13 September 2015; Atlantis Press: Amsterdam, The Netherlands, 2015; pp. 955–959. [Google Scholar]
  22. Ma, W.; Zhang, Y.; Wang, J. Matlab simulink modeling and simulation of zhang neural networks for online time-varying sylvester equation solving. Comput. Math. Math. Phys. 2008, 47, 285–289. [Google Scholar]
  23. Li, S.; Chen, S.; Liu, B. Accelerating a recurrent neural network to finite-time convergence for solving time-varying sylvester equation by using a sign-bi-power activation function. Neural Process. Lett. 2013, 37, 189–205. [Google Scholar] [CrossRef]
  24. Shen, Y.; Miao, P.; Huang, Y.; Shen, Y. Finite-time stability and its application for solving time-varying sylvester equation by recurrent neural network. Neural Process. Lett. 2015, 42, 763–784. [Google Scholar] [CrossRef]
  25. Jin, L.; Zhang, Y.; Li, S.; Zhang, Y. Modified znn for time-varying quadratic programming with inherent tolerance to noises and its application to kinematic redundancy resolution of robot manipulators. IEEE Trans. Ind. Electron. 2016, 63, 6978–6988. [Google Scholar] [CrossRef]
  26. Mao, M.; Li, J.; Jin, L.; Li, S.; Zhang, Y. Enhanced discrete-time zhang neural network for time-variant matrix inversion in the presence of bias noises. Neurocomputing 2016, 207, 220–230. [Google Scholar] [CrossRef]
  27. Xiao, L.; Liao, B. A convergence-accelerated zhang neural network and its solution application to lyapunov equation. Neurocomputing 2016, 193, 213–218. [Google Scholar] [CrossRef]
  28. Liao, B.; Zhang, Y.; Jin, L. Taylor discretization of znn models for dynamic equality-constrained quadratic programming with application to manipulators. Neural Netw. Learn. Syst. IEEE Trans. 2016, 27, 225–237. [Google Scholar] [CrossRef] [PubMed]
  29. Xiao, L. A finite-time convergent zhang neural network and its application to real-time matrix square root finding. Neural Comput. Appl. 2017, 31, 793–800. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Mu, B.; Zheng, H. Link between and comparison and combination of zhang neural network and quasi-newton bfgs method for time-varying quadratic minimization. IEEE Trans. Cybern. 2013, 43, 490–503. [Google Scholar] [CrossRef]
  31. Yan, D.; Li, C.; Wu, J.; Deng, J.; Zhang, Z.; Yu, J.; Liu, P.X. A novel error-based adaptive feedback zeroing neural network for solving time-varying quadratic programming problems. Mathematics 2024, 12, 2090. [Google Scholar] [CrossRef]
  32. Zhang, Z.; Lu, Y.; Zheng, L.; Li, S.; Yu, Z.; Li, Y. A new varying-parameter convergent-differential neural-network for solvingtime-varying convex QP problem constrained by linear-equality. IEEE Trans. Autom. Control 2018, 63, 4110–4125. [Google Scholar] [CrossRef]
  33. Horn, R.A.; Johnson, C.R. Topics in Matrix Analysis, 1991; Cambridge University Presss: Cambridge, UK, 1991; Volume 37, p. 39. [Google Scholar]
  34. Strang, G.; Freund, L. Introduction to applied mathematics. J. Appl. Mech. 1986, 53, 480. [Google Scholar] [CrossRef]
  35. Mead, C.; Ismail, M. Analog VLSI Implementation of Neural Systems; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 80. [Google Scholar]
  36. Zhang, Y.; Ruan, G.; Li, K.; Yang, Y. Robustness analysis of the zhang neural network for online time-varying quadratic optimization. J. Phys. A Math. Theor. 2010, 43, 245202. [Google Scholar] [CrossRef]
Figure 1. Block diagram of VFFTRNN model ( 8 ) for solving the Sylvester equation.
Figure 1. Block diagram of VFFTRNN model ( 8 ) for solving the Sylvester equation.
Mathematics 12 03891 g001
Figure 2. Solving Sylvester Equation (1) with given coefficients online with ZNN and its computational error | | P ( t ) P * ( t ) | | F . Several red solid curves represent different initial states while the unique black dotted curve represents the theoretical solution. (a) Solution to ZNN with μ = 0.1 and error | | P ( t ) P * ( t ) | | F of ZNN with μ = 0.1 . (b) Solution to ZNN with μ = 1 and error | | P ( t ) P * ( t ) | | F of ZNN with μ = 1 . (c) Solution to ZNN with μ = 5 and error | | P ( t ) P * ( t ) | | F of ZNN with μ = 5 .
Figure 2. Solving Sylvester Equation (1) with given coefficients online with ZNN and its computational error | | P ( t ) P * ( t ) | | F . Several red solid curves represent different initial states while the unique black dotted curve represents the theoretical solution. (a) Solution to ZNN with μ = 0.1 and error | | P ( t ) P * ( t ) | | F of ZNN with μ = 0.1 . (b) Solution to ZNN with μ = 1 and error | | P ( t ) P * ( t ) | | F of ZNN with μ = 1 . (c) Solution to ZNN with μ = 5 and error | | P ( t ) P * ( t ) | | F of ZNN with μ = 5 .
Mathematics 12 03891 g002
Figure 3. Solving Sylvester Equation (1) with given coefficients online with VFFTRNN and its computational error | | P ( t ) P * ( t ) | | F ( λ = 5 7 ). Several red solid curves represent different initial states while the unique black dotted curve represents the theoretical solution. (a) Solution to VFFTRNN with α = 0.1 and error | | P ( t ) P * ( t ) | | F of VFFTRNN with α = 0.1 . (b) Solution to VFFTRNN with α = 1 and error | | P ( t ) P * ( t ) | | F of VFFTRNN with α = 1 . (c) Solution to VFFTRNN with α = 5 and error | | P ( t ) P * ( t ) | | F of VFFTRNN with α = 5 .
Figure 3. Solving Sylvester Equation (1) with given coefficients online with VFFTRNN and its computational error | | P ( t ) P * ( t ) | | F ( λ = 5 7 ). Several red solid curves represent different initial states while the unique black dotted curve represents the theoretical solution. (a) Solution to VFFTRNN with α = 0.1 and error | | P ( t ) P * ( t ) | | F of VFFTRNN with α = 0.1 . (b) Solution to VFFTRNN with α = 1 and error | | P ( t ) P * ( t ) | | F of VFFTRNN with α = 1 . (c) Solution to VFFTRNN with α = 5 and error | | P ( t ) P * ( t ) | | F of VFFTRNN with α = 5 .
Mathematics 12 03891 g003
Figure 4. The curves in the figure represent the convergence errors under different initial conditions. When ZNN uses different activation functions without perturbation to solve Sylvester Equation (1), its convergence error | | P ( t ) P * ( t ) | | F is as shown above ( μ = 1 ). (a) Linear. (b) Sigmoid. (c) Power ( s = 3 ).
Figure 4. The curves in the figure represent the convergence errors under different initial conditions. When ZNN uses different activation functions without perturbation to solve Sylvester Equation (1), its convergence error | | P ( t ) P * ( t ) | | F is as shown above ( μ = 1 ). (a) Linear. (b) Sigmoid. (c) Power ( s = 3 ).
Mathematics 12 03891 g004
Figure 5. The curves in the figure represent the convergence errors under different initial conditions. WhenVFFTRNN uses different activation functions without perturbation to solve Sylvester Equation (1), its convergence error | | P ( t ) P * ( t ) | | F is as shown above ( α = 1 and λ = 1 5 ). (a) Linear. (b) Sigmoid. (c) Power ( s = 3 ).
Figure 5. The curves in the figure represent the convergence errors under different initial conditions. WhenVFFTRNN uses different activation functions without perturbation to solve Sylvester Equation (1), its convergence error | | P ( t ) P * ( t ) | | F is as shown above ( α = 1 and λ = 1 5 ). (a) Linear. (b) Sigmoid. (c) Power ( s = 3 ).
Mathematics 12 03891 g005
Figure 6. Robustness of ZNN and VFFTRNN with varying degrees of differential error and model implementation error. The arrows point to the original location of the magnified detail in each figure. (a) β 1 = β 2 = β 3 = 0.05 . (b) β 1 = β 2 = 0.05 and β 3 = 0.15 . (c) β 1 = β 3 = 0.05 . and β 2 = 0.15 . (d) β 1 = 0.15 and β 2 = β 3 = 0.05 .
Figure 6. Robustness of ZNN and VFFTRNN with varying degrees of differential error and model implementation error. The arrows point to the original location of the magnified detail in each figure. (a) β 1 = β 2 = β 3 = 0.05 . (b) β 1 = β 2 = 0.05 and β 3 = 0.15 . (c) β 1 = β 3 = 0.05 . and β 2 = 0.15 . (d) β 1 = 0.15 and β 2 = β 3 = 0.05 .
Mathematics 12 03891 g006
Table 1. Comparisons of convergence time of ZNN and VFFTRNN when solving Sylvester equation.
Table 1. Comparisons of convergence time of ZNN and VFFTRNN when solving Sylvester equation.
Parameter ( α = μ ) T ZNN ( s ) T VFFTRNN ( s )
0.01Cannot converge2.39
0.167.552.25
0.234.201.99
0.513.881.69
16.981.41
23.481.02
51.370.49
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tan, H.; Wu, J.; Guan, H.; Zhang, Z.; Tao, L.; Zhao, Q.; Li, C. A New Varying-Factor Finite-Time Recurrent Neural Network to Solve the Time-Varying Sylvester Equation Online. Mathematics 2024, 12, 3891. https://doi.org/10.3390/math12243891

AMA Style

Tan H, Wu J, Guan H, Zhang Z, Tao L, Zhao Q, Li C. A New Varying-Factor Finite-Time Recurrent Neural Network to Solve the Time-Varying Sylvester Equation Online. Mathematics. 2024; 12(24):3891. https://doi.org/10.3390/math12243891

Chicago/Turabian Style

Tan, Haoming, Junyun Wu, Hongjie Guan, Zhijun Zhang, Ling Tao, Qingmin Zhao, and Chunquan Li. 2024. "A New Varying-Factor Finite-Time Recurrent Neural Network to Solve the Time-Varying Sylvester Equation Online" Mathematics 12, no. 24: 3891. https://doi.org/10.3390/math12243891

APA Style

Tan, H., Wu, J., Guan, H., Zhang, Z., Tao, L., Zhao, Q., & Li, C. (2024). A New Varying-Factor Finite-Time Recurrent Neural Network to Solve the Time-Varying Sylvester Equation Online. Mathematics, 12(24), 3891. https://doi.org/10.3390/math12243891

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop