You are currently viewing a new version of our website. To view the old version click .
Symmetry
  • Article
  • Open Access

30 October 2025

Error-Driven Varying-Gain in Zeroing Neural Networks for Solving Time-Varying Quadratic Programming Problems

,
,
and
School of Artificial Intelligence, Guangzhou Maritime University, Guangzhou 510725, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Computer

Abstract

Convergence is a critical performance metric when zeroing neural networks (ZNNs) are employed to solve time-varying problems. As an effective strategy to enhance the convergence of ZNN models, gain adjustment is widely adopted to accelerate convergence speed, with the core objective of enabling the residual error to converge to zero rapidly and accurately. However, existing gain-tuning methods are decoupled from the residual error. This decoupling may lead to oscillation when the residual error approaches zero. To address this limitation, this paper proposes a novel error-driven varying-gain scheme. In this scheme, the gain value dynamically adapts to the residual error; a large residual error triggers a large gain, while a small residual error corresponds to a small gain, ensuring the gain changes synchronously with the residual error. Theoretical analyses and experimental results collectively demonstrate that integrating this error-driven varying-gain into ZNNs yields superior convergence performance, providing a more reliable solution for time-varying problem solving with ZNNs.

1. Introduction

In the practical engineering applications and scientific research, many problems can be synthesized into a quadratic programming (QP) problem to be solved [1,2,3,4,5]. For example, in [1,2], the trajectory tracking of the robot manipulator can be expressed as a solution to a time-varying QP (TVQP) mathematical problem. An improved sequential quadratic programming (ISQP) algorithm is proposed to solve the receding horizon optimization problem by the authors in [6]. Note that, during the solution the robot manipulator [1,7,8], by making full use of the symmetry characteristics, the performance index can be formulated into a symmetric mathematical model, which is convenient for the computing [9]. As for its solution of the QP problem, there are also many methods, mainly including the numerical solutions and the neural network models [1,2,10,11]. As an effective parallel computing method during the hardware realization, recently, neural networks have been drawing increasing attention due to their outstanding performance in image recognition and natural language processing [12,13].
Inspired by Zhang’s design method [14], a zeroing neural network (ZNN) is also used for solutions to TVQP problems [1,2,7]. In [15], the authors presented a discrete sliding-mode reaching-raw ZNN model to solve the TVQP problem. In [7], since the nonlinear complementary method would expend the dimensions of the matrix, to overcome this deficiency, the authors proposed a lower ZNN model to solve the TVQP problem. In fact, many ZNN models are developed and have already been widely used in many other fields, e.g., robot control [2,7], unmanned aerial vehicle (UAV) control [8,16,17], and noise control [18,19]. During the solutions of these practical problems, convergence is a performance index used for neural networks. To accelerate the convergence of the neural network models, many authors have presented activation functions to enhance the convergence performance [20,21,22,23,24]. Based on the fuzzy logic strategy, a fuzzy activation function-based ZNN model was proposed and used for dynamic Arnold map image cryptography [21]. In [20], an activation function based on the super-twisting algorithm (STA) was encapsulated with the ZNN models to solve the time-variant Sylvester equation with the superior finite time convergence and noise tolerance. This is one way to accelerate the convergence speed for ZNN models.
Another way to improve the convergence is to change the value of the gain [25,26,27,28,29,30,31]. For instance, the authors in [25] presented a changing-parameter complex-valued ZNN model to solve the linear time-varying complex matrix equation with the finite time convergence. Evidently, this varying-gain scheme can achieve better convergence performance than fixed parameters [25,26,27,28,29], since the value of the varying gain would increase over time. However, this value would continue to increase even if the residual error is already very small. This would lead to oscillation of the residual error. In addition, with a varying gain, it is difficult to control a suitable value. In other words, if the gain value is set too small, the convergence performance may be unsatisfactory, which would result in slow error reduction that fails to meet the real-time requirements of time-varying problems. Conversely, if the gain value is set too high, it may exceed the voltage-bearing capacity of the circuit components used to implement the neural network, potentially causing component malfunctions or even irreversible burnout. This trade-off between convergence speed and hardware safety is a critical limitation of fixed or improperly designed varying-gain schemes [29]. Moreover, the gain setting has nothing to do with the residual error.
Considering these issues, in this paper, we present a varying gain driven by the residual error, which includes a constant item and an exponential item. Generally speaking, In the initial stage of the iterative solution, the residual error is typically much larger than in subsequent phases of the process. At this stage, the exponential term dominates, resulting in a sufficiently large gain value to rapidly reduce the residual error. Therefore, the initial gain value derived from the error-driven varying-gain scheme is also sufficiently large, which facilitates the rapid reduction of the residual error. As the iterative solution progresses, the residual error gradually decreases. At this point, the constant term takes over to further minimize the small residual error. Through this adaptive gain-tuning mechanism, the ZNN model integrated with the error-driven varying gain can solve the TVQP problem effectively and accurately. Both theoretical analysis and experimental results further validate the effectiveness of the proposed error-driven varying gain.
The remainder of this paper is organized as follows. Section 2 presents the mathematical formulations of existing varying-gain schemes and the proposed error-driven gain for ZNN models. Section 3 provides rigorous theoretical analyses of the proposed model. Section 4 conducts simulations, comparisons, and analyses of an illustrative example to validate the scheme’s performance. Section 5 concludes the paper with final remarks. The main contributions of this work are summarized as follows.
(i)
An error-driven varying-gain scheme is proposed for the ZNN model. The initial gain is large due to the large initial residual error, which facilitates the rapid reduction of the large initial residual error.
(ii)
Rigorous theoretical analysis, along with detailed proofs, is presented for the varying gain zeroing neural network (VGZNN) model. These analyses demonstrate that the VGZNN model achieves superior convergence performance when integrated with the proposed error-driven varying-gain scheme.
(iii)
In the experimental section, a numerical TVQP example is simulated, analyzed, and compared with existing methods. These results further validate that the proposed error-driven varying-gain scheme outperforms the other three varying-gain schemes, making it the optimal choice.

2. Varying Gain for Solving TVQP Problems

Suppose that the TVQP problem subjected to non-stationary linear equality constraints can be formulated as follows:
min x T ( t ) Q ( t ) x ( t ) / 2 + p T ( t ) x ( t ) , s .   t . A ( t ) x ( t ) = b ( t ) ,
where x ( t ) R n is an unknown state to be solved to satisfy the TVQP problem (1), and all of the known time-varying coefficients Q ( t ) R n × n , p ( t ) R n , A ( t ) R m × n , and b ( t ) R m are assumed to be smooth, which ensures their time derivatives can be computed during the solution process. Moreover, TVQP (1) is supposed to have a unique solution.
Then, by the Karush–Kuhn–Tucker conditions [32], we can define a Lagrangian function for TVQP (1) as follows:
L ( x ( t ) , ν ( t ) , t ) = x T ( t ) Q ( t ) x ( t ) / 2 + p T ( t ) x ( t ) + ν T ( t ) ( A ( t ) x ( t ) b ( t ) ) ,
where ν ( t ) R m is expressed as a Lagrange multiplier. As a result, the time-varying solution to the TVQP problem can be obtained by solving the following system of coupled equations:
L ( x ( t ) , ν ( t ) , t ) x ( t ) = Q ( t ) x ( t ) + p ( t ) + A T ( t ) ν ( t ) = 0 , L ( x ( t ) , ν ( t ) , t ) ν ( t ) = A ( t ) x ( t ) b ( t ) = 0 ,
which can be further re-written as the following linear time-varying matrix-vector equation:
W ( t ) y ( t ) = n ( t ) ,
with the following defined time-varying coefficients:
W ( t ) : = Q ( t ) A T ( t ) A ( t ) 0 m × m R ( n + m ) × ( n + m ) , y ( t ) : = x ( t ) ν ( t ) R n + m , n ( t ) : = p ( t )   b ( t ) R n + m .
As for the solution of Problem (2), inspired by Zhang et al.’s design idea in [14], an indefinite (i.e., negative or positive) error function e ( t ) = W ( t ) y ( t ) n ( t ) is set as follows:
e ˙ ( t ) = γ ϕ ( e ( t ) ) .
By extending the design Formula (3), we can derive the following fixed-gain zeroing neural network (FGZNN) model:
W ( t ) x ˙ ( t ) = γ ϕ ( W ( t ) x ( t ) n ( t ) ) W ˙ ( t ) x ( t ) + n ˙ ( t ) .
where γ generally is a constant (i.e., time-invariant) parameter used to regulate the convergence speed, and ϕ ( · ) is referred to as an activation function array, which serves to enhance the convergence performance [28,29]. In this study, the following activation functions are investigated [28,29]:
(1)
the linear activation function: ϕ ( u ) = u ;
(2)
the power function: ϕ ( u ) = u ξ with ξ 3 ;
(3)
and the sign-bi-power activation function:
ϕ ( u ) = ( | u | ϵ + | u | 1 / ϵ ) sgn ( u ) / 2 , 0 < ϵ < 1
with sgn ( · ) denoting the following signum function:
sgn ( u ) = 1 , if   u > 0 , = 0 , if   u = 0 , = 1 , if   u < 0 .
Generally speaking, assuming all other parameters remain identical, the sigmoid function and power-sigmoid function, two widely adopted and extensively investigated activation functions in existing studies, enable the model to achieve favorable convergence performance. In contrast, under the same conditions of other parameters, the linear activation function and the power function yield the poorest convergence performance, which is consistent with the findings reported in [14]. Therefore, to achieve significantly much better convergence performance even when using the latter two activation functions, the authors of [28] proposed a varying-gain scheme (i.e., γ γ ( t ) ) to enhance convergence performance. The FGZNN model with a varying gain can thus be transformed into the following varying gain ZNN (VGZNN):
W ( t ) y ˙ ( t ) = γ ( t ) ϕ ( W ( t ) y ( t ) n ( t ) ) W ˙ ( t ) y ( t ) + n ˙ ( t ) .
In this paper, the primary object is to find a neural solution y ( t ) that converges to the theoretical solution y * ( t ) of (2). That is, we can find x ( t ) from y ( t ) , and ensure that x ( t ) converges to the theoretical solution x * ( t ) of TVQP (1), where x * ( t ) is obtained from y * ( t ) . When a varying-gain scheme is used, the convergence performance of the model can be enhanced. For example, in [28], the authors presented the following varying-gain expression:
γ ( t ) = t k + k , k > 0 ,
As shown in Figure 4 of [28], the convergence performance is significantly improved when the gain is changed from a constant γ to a varying gain γ ( t ) defined in Equation (7). Meanwhile, another varying-gain scheme incorporating exp ( t ) was proposed in [29] to enhance the convergence performance of the model; this scheme is repeated as follows:
γ ( t ) = γ 0 + γ 1 exp ( γ 2 t ) ,
where γ 0 > 0 , γ 1 > 0 , and γ 2 > 0 are the known constant parameters, which are used to adjust the value of the varying-gain defined in (8). As for the varying-gain scheme (8), a large initial varying gain (at t = 0 ) can be achieved by setting a large value for γ 1 ; γ 0 is configured to determine the steady-state gain, while γ 2 is used to regulate the decay rate of the varying gain γ ( t ) . In this way, the network model can achieve a superior convergence performance, as illustrated in Figure 3 in [29].
Remark 1.
In practical applications, the initial computational error is generally relatively large. Therefore, a relatively large design parameter should be set to reduce the computational error rapidly. At this moment, if the varying parameter is designed as a linear-power function, e.g., Equation (7), or an exponential function exp ( t ) , e.g, Equation (8), the varying-gain will be relatively small due to their inherent characteristics of these functions. of this paper. Since the time t is extremely small at the initial stage (e.g., t = 10 5 s), this will result in an extremely slow convergence speed.
Therefore, to address this issue, we propose an error-driven varying-gain scheme, which is expressed as follows:
γ ( t ) = γ 0 + γ 1 exp ( e ( t ) t ) ,
where · is defined as the Frobenius norm [28,29], γ 0 and γ 1 have the same definitions and values in this paper.
Remark 2.
It is evident that the varying-gain defined in (9) is driven by the computation error. At the initial stage, the computational error is extremely large. As a result, a large gain is required to reduce the error rapidly. By following the function’s characteristics, the value of exp ( e ( t ) t ) will be very large initially (i.e., as t 0 ), given that the norm error e ( t ) is large at this stage. Moreover, as time t , the error will become extremely small, and the varying gain should be small to ensure the error reduction is smooth. It is worth noting that, if the varying gain remains relatively large at this stage, the residual error will exhibit an oscillation phenomenon.
Remark 3.
As for the VGZNN model (6), under the same conditions, in general, the larger the gain value, the faster the convergence speed that can be achieved. As a result, compared to the scheme (8), when e ( t ) 2 γ 2 , a faster convergence performance can be achieved by using the scheme (9) because of the different item e x p ( · ) . Compared to the scheme (7), since t is small at the beginning, the scheme (6) can achieve a faster convergence. However, as t , the gain value of Scheme (7) will become larger than that of Scheme (9), and the resulting convergence will be faster than that of Scheme (9).

3. Theoretical Analysis

In this section, we would like to present the theoretical analysis on the VGZNN model (6) when integrated with the error-driven varying gain specified in Equation (9).
Theorem 1.
We take as given that γ 0 > 0 and γ 1 > 0 . If the VGZNN model (6) is activated by an arbitrary monotonically increasing odd function, the neural solution y ( t ) would globally and asymptotically converge to the theoretical solution y * ( t ) of (2).
Proof. 
Let y ˜ ( t ) = y ( t ) y * ( t ) denote the difference of the neural state y ( t ) and y * ( t ) . Then
W ( t ) y ˜ ˙ ( t ) = γ ( t ) ϕ ( W ( t ) y ˜ ( t ) ) W ˙ ( t ) y ˜ ( t ) .
Additionally, from W ( t ) y ( t ) n ( t ) = 0 and W ( t ) y * ( t ) n ( t ) = 0 , we would have W ( t ) y ˜ ( t ) = 0 . Since the error function e ( t ) = W ( t ) y ( t ) n ( t ) and W ( t ) y * ( t ) n ( t ) = 0 , e ( t ) = W ( t ) y ˜ ( t ) and e ˙ ( t ) = W ˙ ( t ) y ˜ ( t ) + W ( t ) y ˜ ˙ ( t ) . By following the design Formula (3), Model (10) can be also expressed as e ˙ ( t ) = γ ( t ) ϕ ( e ( t ) ) , which can be further transformed into the following n + m element-wise form:
e ˙ i ( t ) = γ ( t ) ϕ i ( e i ( t ) ) , i { 0 , 1 , 2 , 3 , , n + m }
where ϕ i ( · ) denotes the ith element of the activation function array ϕ ( · ) . Therefore, as for the model (11), a candidate Lyapunov function can be defined as v i ( t ) = e i 2 ( t ) / 2 , so v ˙ i ( t ) = e i ( t ) e ˙ i ( t ) = γ ( t ) e i ( t ) ϕ i ( e i ( t ) ) . Given that ϕ i ( · ) is a monotonically increasing odd activation function. Then
e i ( t ) ϕ i ( e i ( t ) ) > 0 if   e i ( t ) > 0 , = 0 if   e i ( t ) = 0 , < 0 if   e i ( t ) < 0 .
It is evident that the Lyapunov function v i ( t ) satisfies v i ( t ) 0 for all t, and its time derivative v ˙ i ( t ) satisfies v ˙ i ( t ) 0 for all t. By the Lyapunov stability [28,29], each error element e i ( t ) from e ( t ) would globally and asymptotically converge to zero; that is, e ( t ) = W ( t ) y ˜ ( t ) would converge to zero. In turn, this implies that y ˜ ( t ) = y ( t ) y * ( t ) (the difference between the neural solution and the theoretical solution) would also converge to zero. In other words, the neural solution y ( t ) will converge to y * ( t ) when time t . Theorem 1 is thus proved. □
Theorem 2.
We take as given that γ 0 > 0 and γ 1 > 0 . If linear activation ϕ ( e ( t ) ) = e ( t ) is adopted for the VGZNN model (6), then the neural state y ( t ) will exponentially converge to y * ( t ) with the exponential rate is 2 γ 0 or so when time t .
Proof. 
From the proof of Theorem 1, we know that v ˙ i ( t ) = e i ( t ) e ˙ i ( t ) = γ ( t ) e i ( t ) ϕ i ( e i ( t ) ) . If the linear function ϕ ( e ( t ) ) = e ( t ) is used, then v ˙ i ( t ) = γ ( t ) e i 2 ( t ) = 2 γ ( t ) v i ( t ) since v i ( t ) = e i 2 ( t ) / 2 is defined in Theorem 1. Therefore, we have
v ˙ i ( t ) = 2 γ ( t ) v i ( t )
Recall the error-driven varying gain (9). When time t , γ ( t ) would be approximately equal to γ 0 . Therefore, solving Equation (12), we have the following inequality:
v i ( t ) κ exp ( 2 γ 0 t )
with κ = v i ( 0 ) = e i 2 ( 0 ) / 2 .
By leveraging Lyapunov stability theory, the energy function v i ( t ) will converge to zero at an exponential rate of 2 γ 0 ; this exponential convergence property is also applied to the error element e i ( t ) . It implies that the neural solution y ( t ) converge to the theoretical solution y * ( t ) at the same exponential rate 2 γ 0 as t . The proof is thus completed. □
Theorem 3.
We take as given that γ 0 0 and γ 1 > 0 , and power function ϕ ( u ) = u ξ is used with ξ 3 used to govern the nonlinearity of the activation function. When the fixed gain γ = γ 0 is used for the FGZNN (4) and the varying gain (9) is used for VGZNN (6), the convergence speed of VGZNN (6) would be faster than that of FGZNN (4).
Proof. 
Let V F ( t ) and V V ( t ) denote the Lyapunov function for the FGZNN (4) and VGZNN (6), respectively. By Theorem 1, it is known that V ˙ ( t ) = γ e ( t ) ϕ ( e ( t ) ) , so we have
V ˙ V ( t ) V ˙ F ( t ) = ( γ ( t ) + γ 0 ) e p + 1 = γ 1 exp ( γ 2 e ( t ) ) 0 .
That is, V ˙ V ( t ) descends faster than V ˙ F ( t ) in the negative direction. This further substantiates that the sufficiently large gain can accelerate the decay rate of the residual error e ( t ) . As a result, numerous researchers have adopted sufficiently large gain values to achieve superior convergence performance for neural network models [28,29]. However, if the power function ϕ ( u ) = u ξ (with ξ ≥ 3) is adopted, the convergence performance of the neural network model is difficult to improve, even if the gain is set to a sufficiently large value. Moreover, it is not allowed in piratical circuit realization if the gain is set too large, which will result in the instability or even irreversible burnout of circuit components. Fortunately, if the error-driven varying gain specified in Equation (9) is adopted, the model can automatically obtain a sufficiently large initial gain to significantly enhance the convergence performance of the neural network. □
Theorem 4.
We take as given that γ 0 > 0 and γ 1 > 0 . If the sign-bi-power activation function (5) is adopted, the state solution y ( t ) will be globally convergent to the theoretical solution y * ( t ) for Model (6) with the following finite convergence time t expressed as
t 2 e i ( t ) 1 ϵ ( 1 ϵ ) ( γ 0 + γ 1 ) .
Proof. 
By following Theorem 1, it is known that v i ( t ) = e i 2 ( t ) / 2 and v ˙ i ( t ) = γ ( t ) e i ( t ) ϕ i ( e i ( t ) ) . Therefore, when the activation function (5) is exploited, we have
v ˙ i ( t ) = 1 2 γ ( t ) e i ( t ) ( e i ( t ) ϵ + e i ( t ) 1 ϵ ) sgn ( e i ( t ) ) = 1 2 γ ( t ) ) e i ( t ) ( e i ( t ) ϵ + e i ( t ) 1 ϵ ) , e i ( t ) 0 1 2 γ ( t ) ) e i ( t ) ( e i ( t ) ϵ + e i ( t ) 1 ϵ ) , e i ( t ) 0 = 1 2 γ ( t ) ( e i 2 ( t ) ) 1 + ϵ 2 + ( e i 2 ( t ) ) 1 + ϵ 2 ϵ = 1 2 γ ( t ) ( 2 v i ( t ) ) 1 + ϵ 2 + ( 2 v i ( t ) ) 1 + ϵ 2 ϵ
Therefore, by considering the value range of v i ( t ) and recalling the error-driven varying gain specified in Equation (9) for (14), we can derive the following inequality:
v ˙ i ( t ) 1 2 γ ( t ) ( 2 v i ( t ) ) 1 + ϵ 2 .
Based on Equation (15) with 0 < ϵ < 1 , we have the following analytical solution:
0 ( 2 v i ( t ) ) 1 ϵ 2 ( 2 v 0 ( t ) ) 1 ϵ 2 1 ϵ 2 ( γ 0 + γ 1 ) t .
By following Theorem 1 and Lyapunov stability, v i ( t ) would converge to zero along with time t . Thus, Equation (16) can be further transformed into the following:
1 ϵ 2 ( γ 0 + γ 1 ) t ( 2 v 0 ( t ) ) 1 ϵ 2 .
Therefore, the convergence time t is satisfied with the following expression:
t 2 e i ( t ) 1 ϵ ( 1 ϵ ) ( γ 0 + γ 1 ) .
The proof is thus completed. □

4. Illustrative Examples

In the previous sections, we have already presented the different varying-gain schemes and active functions tailored for the VGZNN model (6), along with a comprehensive theoretical analysis of the model’s convergence behavior. To demonstrate the superior convergence performance of the VGZNN model when adopting the error-driven varying gain specified in Equation (9), we conduct comparative simulations with benchmarks including other varying-gain schemes and different activation functions.

4.1. Example 1: Numerical Example

To verify the effectiveness of the proposed varying-gain scheme, we investigate and compare the following illustrative time-varying quadratic programming (TVQP) problem. The coefficients of this TVQP problem are consistent with those used in [1], ensuring the realism of the experimental setup and the comparability of the results.
min ( 1 + 1 4 sin t ) x 1 2 ( t ) + ( 1 + 1 4 cos t ) x 2 2 ( t ) + ( cos t ) x 1 ( t ) x 2 ( t ) + ( sin 3 t ) x 1 ( t ) + ( cos 3 t ) x 2 ( t ) , s .   t . ( sin 4 t ) x 1 ( t ) + ( cos 4 t ) x 2 ( t ) = cos 2 t ,
which can be expressed as the standard form like (1) with the following time-varying coefficients for (18)
Q ( t ) = 0.5 sin t + 2 cos t cos t 0.5 cos t + 2 , p ( t ) = sin 3 t cos 3 t , A ( t ) = sin t cos 4 t , b ( t ) = cos 2 t .
Then the above TVQP problem can be easily transformed into the linear time-varying Equation (2) with the time-varying coefficients expressed nearby (2). For the comparison and analysis purposes, by using the MATLAB R2016a tools, we can obtain the theoretical solution y * ( t ) = [ x * T ( t ) , ν * T ( t ) ] T : = W 1 ( t ) n ( t ) R n + m .
When the sign-bi-power activation function (5) with ϵ = 1 / 3 is exploited for the VGZNN model (6) to solve the TVQP problem (18), together with γ 0 = 10 and γ 1 = 100 in the error-driven gain (9), the convergence performance is as shown in Figure 1. The neural state y ( t ) , denoted by the blue curve, is fast overlapped with the theoretical solution y * ( t ) = W 1 ( t ) n ( t ) , which can be easily computed by the MATLAB tool and is denoted by the red curve. As shown in Figure 1a, x 1 ( t ) and x 2 ( t ) are the neural solution of (18), and ν ( t ) is a Lagrangian factor designed for the construction from TVQP problem (1) to the linear time-varying Equation (2). In addition, the residual norm error W ( t ) y ( t ) n ( t ) 2 also quickly converges to zero within 0.025 s. This indicates that, as the neural state y ( t ) approaches the theoretical solution y * ( t ) , the residual error is convergent to zero accordingly.
Figure 1. Trajectories of the solution for (18) solved by the VGZNN model (6) with Scheme (9).
In addition, different parameters and activation functions will also affect the convergence of the VGZNN model (6). As shown in Figure 2a and Table 1, when the error-driven varying gain (9) and linear activation function ϕ ( u ) = u are used for the VGZNN model (6) under the same experimental conditions, the residual error W ( t ) y ( t ) n ( t ) would converge to zero at a much faster rate if the parameters are set much larger; for example, when γ 0 = γ 1 = 1 , the residual error is convergent to 9.813 × 10 3 at t = 3.148 s; in contrast, when γ 0 = γ 1 = 10 , the residual error is 7.909 × 10 3 at t = 0.293 s. The activation function is another important factor influencing the convergence performance of the VGZNN model. It can be seen in Figure 2b that the best convergence can be achieved when the sign-bi-power activation function is exploited for the VGZNN model (6) among the three functions (i.e., linear, sign-bi-power (which is denoted by sgn), and power function).
Figure 2. Convergence performance for (18) solved by VGZNN model (6) with different parameters and activation function when scheme (9) is used.
Table 1. Comparisons of the convergence performance of the VGZNN (6) integrated with Scheme (9) using different parameters.
As for the VGZNN model (6) used to solve the TVQP problem (18), if we utilize the different varying-gain schemes, then we can obtain a different convergence effect, which can be seen in Figure 3 and Table 2. Under the same simulation conditions, as shown in Figure 3a, the same linear function and the same parameters ( γ = 10 ,   γ 0 = 1 , γ 1 = 10 and γ 2 = 20 ) are adopted, and it is evident that the best convergence effect is achieved for the error-driven varying gain (9) (denoted by VG (9) in Figure 3) among the four cases (i.e., the constant case γ , varying gain (7), denoted by VG (7), and varying gain (8), denoted by VG (8)). It is noted that the residual errors are finally convergent to zero in all cases. However, if the power function is encapsulated in the VGZNN model (6), the residual errors do not converge to zero in all cases within 10 s. Certainly, if given enough time, the residual error would converge to zero finally. Or, in another case, if the parameters are set large enough to accelerate the convergence speed, the residual error would finally converge to zero [28,29]. Furthermore, in most cases, the power activation function is generally not investigated because of its poor convergence performance. This issue is also discussed in [14] and a power-extension function is also presented for the superior convergence. It is worth noting that, as shown in Table 2, when the power function is used, the convergence performance with VG (7) is superior to that with V G (9). This is because the value of VG (7) increases progressively as time elapses.
Figure 3. Convergence performance for (18) solved by the VGZNN model (6) with different parameters and activation functions.
Table 2. Comparisons of the convergence performance of the VGZNN (6) for solving the TVPQ problem (18) with different gain schemes.

4.2. Example 2: TVQP Problem Solution for a Robot Manipulator

As described in the Introduction section, robot arm manipulation can be synthesized as a solution to a TVQP problem. In this section, the VGZNN model (6) integrated with the varying-gain scheme (9) is used for the robot arm PUMA560 for tracking [19]. If the robot end-effector position d ( t ) R m and the joint angle θ ( t ) R n , then the forward kinematics dynamical equation in Cartesian space can be modeled as the following mathematical expression:
d ( t ) = ψ ( θ ( t ) )
where ψ ( · ) denotes the continuous nonlinear mapping with the known structure parameters of the robot arm. Evidently, Equation (19) is difficult to solve directly. Therefore, a TVQP-based scheme at velocity level is presented in [1], as follows:
min 1 2 θ ˙ ( t ) θ ˙ T ( t ) , s .   t . J ( θ ( t ) ) θ ˙ ( t ) = d ˙ ( t ) .
where J ( θ ( t ) ) = f ( θ ( t ) ) / θ ( t ) R m × n is a Jacobian matrix. Then Equation (20) can be transformed into a standard TVQP problem (1), where x ( t ) : = θ ˙ ( t ) R n , Q ( t ) : = I R n × n , A ( t ) : = J ( θ ( t ) ) R m × n , b ( t ) : = d ˙ ( t ) R m , and p ( t ) : = 0 R n . Therefore, we can adopt the same transformation method to convert the time-varying quadratic programming (TVQP) problem defined in Equation (1) into a linear time-varying equation, as shown in Equation (2).
In this example, the VGZNN model (6) integrated with our proposed error-driven varying scheme (9) is adopted to control the PUMA560 to track a circle. As shown in Figure 4, the actual trajectory of the robot arm coincides with the desired circular trajectory, as illustrated in Figure 4a. The real time trajectory of θ ( t ) is drawn in Figure 4b. Moreover, as shown in Figure 4c, the position error can reach 10 7 . The real time control trajectory can be seen in Figure 4d.
Figure 4. VGZNN model (6) used to control PUMA560 with the proposed varying-gain scheme (9).

5. Conclusions

This study focuses on enhancing the convergence performance of zeroing neural networks (ZNNs) for the online solution of time-varying matrix equation problems, with a particular emphasis on time-varying quadratic programming (TVQP) problems. Two core strategies for improving network convergence are systematically investigated and validated through theoretical analyses and experimental simulations. The selection of the activation functions significantly impacts the convergence behavior of ZNN models. Among the three typical functions, i.e., the linear function, power function (with ξ = 3 ), and sign-bi-power function (with 0 < ϵ < 1 ), the sign-bi-power activation function demonstrates superior performance in accelerating convergence. As verified by comparative simulations (Table 2 and Figure 3), when integrated with the varying-gain ZNN (VGZNN) model, this function enables faster convergence to the theoretical solution compared to linear and power functions, confirming its effectiveness as a priority choice for optimizing ZNN convergence.
In another way, the design of gain adjustment schemes is identified as a critical factor in determining ZNN performance. In this paper, to address the limitations of existing varying-gain schemes, such as decoupling from residual errors, potential oscillation when errors approach zero, and trade-offs between convergence speed and hardware safety, we propose an error-driven varying-gain scheme for the VGZNN model. Unlike traditional schemes that adjust gain independently of residual errors, the proposed scheme dynamically adapts the gain value to the real-time residual error. A large initial residual error triggers a sufficiently large gain to rapidly reduce errors, while a decreasing residual error corresponds to a reduced gain to avoid oscillation and ensure hardware safety. This adaptive mechanism ensures synchronous variation between gain and residual error, fundamentally overcoming the drawbacks of conventional gain-tuning methods.
Experimental results further validate the superiority of the proposed scheme. In numerical simulations of TVQP problems (i.e., Example 1), the error-driven varying-gain scheme reduces the time to reach a residual error of 10 3 to 0.4601 s (with linear activation functions), outperforming fixed-gain ( γ = 10 , 5.286 s) and other varying-gain schemes (e.g., VG (7), 1.566 s; VG (8), 4.829 s). In practical applications (i.e., Example 2), applying the scheme to the trajectory tracking control of the PUMA560 redundant robot arm results in a position error as low as 10 7 , and the actual trajectory of the robot end-effector perfectly matches the desired circular trajectory, demonstrating the scheme’s practical engineering value.
Future research will focus on two directions: (1) further optimizing the error-driven varying-gain scheme to enhance its robustness against external noise and complex time-varying characteristics, which is also the limitations in this article, and (2) expanding the application scope of the scheme to more fields involving solutions to time-varying problems, such as unmanned aerial vehicle (UAV) control and real-time signal processing, to fully exploit its potential in practical engineering scenarios.

Author Contributions

Conceptualization, Y.C.; methodology, J.Z.; software, Y.C.; validation, J.L.; formal analysis, C.S.; investigation, Y.C. and C.S.; writing—original draft preparation, Y.C.; writing—review and editing, J.Z.; visualization, C.S.; supervision, C.S.; project administration, C.S.; funding acquisition, C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Guangdong Basic and Applied Basic Research Foundation under Grant (2025A1515010466), the Guangdong Key Construction Discipline Scientific Research Ability Enhancement Project (2024ZDJS056), and the Guangzhou Programs (2024312151, 2024312000 and 2024312374).

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cai, J.; Yi, C. An adaptive gradient-descent-based neural networks for the on-line solution of linear time variant equations and its applications. Inf. Sci. 2023, 622, 34–45. [Google Scholar] [CrossRef]
  2. Sun, Z.; Tang, S.; Jin, L.; Zhang, J.; Yu, J. Nonconvex activation noise-suppressing neural network for time-varying quadratic programming: Application to omnidirectional mobile manipulator. IEEE Trans. Ind. Inform. 2023, 19, 10786–10798. [Google Scholar] [CrossRef]
  3. Thang, T.V.; Anh, P.N.; Truong, N.D. Convergence of the projection and contraction methods for solving bilevel variational inequality problems. Math. Methods Appl. Sci. 2023, 46, 10867–10885. [Google Scholar] [CrossRef]
  4. Chen, R.; Yang, Z.; Ye, J. Kernel-Free Nonlinear Support Vector Machines for Multiview Binary Classification Problems. Int. J. Intell. Syst. 2023, 2023, 6259041. [Google Scholar] [CrossRef]
  5. Shi, D.; Yin, J.; Bai, C. An effective global optimization algorithm for quadratic programs with quadratic constraints. Symmetry 2019, 11, 424. [Google Scholar] [CrossRef]
  6. Yang, C.; Wang, M.; Wang, W.; Pu, Z.; Ma, M. An efficient vehicle-following predictive energy management strategy for PHEV based on improved sequential quadratic programming algorithm. Energy 2021, 219, 119595. [Google Scholar] [CrossRef]
  7. Li, W.; Wu, H.; Jin, L. A lower dimension zeroing neural network for time-variant quadratic programming applied to robot pose control. IEEE Trans. Ind. Inform. 2024, 20, 11835–11843. [Google Scholar] [CrossRef]
  8. He, Y.; Xiao, L.; Zuo, Q.; Cai, H.; Wang, Y. A Predefined-Time Robust Sliding Mode Control Based on Zeroing Neural Dynamics for Position and Attitude Tracking of Quadrotor. IEEE Trans. Syst. Man Cybern. Syst. 2025, 55, 2438–2450. [Google Scholar] [CrossRef]
  9. Yi, C.; Chen, J.; Li, L. A Direct Discrete Recurrent Neural Network with Integral Noise Tolerance and Fuzzy Integral Parameters for Discrete Time-Varying Matrix Problem Solving. Symmetry 2025, 17, 1359. [Google Scholar] [CrossRef]
  10. Li, W.; Tian, X. Numerical solution method for general interval quadratic programming. Appl. Math. Comput. 2008, 202, 589–595. [Google Scholar] [CrossRef]
  11. Zhu, G.; Li, Y.; Zhang, S.; Duan, X.; Huang, Z.; Yao, Z.; Wang, R.; Wang, Z. Neural Networks With Linear Adaptive Batch Normalization and Swarm Intelligence Calibration for Real-Time Gaze Estimation on Smartphones. Int. J. Intell. Syst. 2024, 2024, 2644725. [Google Scholar] [CrossRef]
  12. Gorokhovatskyi, V.; Tvoroshenko, I.; Yakovleva, O.; Hudáková, M.; Gorokhovatskyi, O. Application a committee of Kohonen neural networks to training of image classifier based on description of descriptors set. IEEE Access 2024, 12, 73376–73385. [Google Scholar] [CrossRef]
  13. Miah, A.S.M.; Hasan, M.A.M.; Okuyama, Y.; Tomioka, Y.; Shin, J. Spatial–temporal attention with graph and general neural network-based sign language recognition. Pattern Anal. Appl. 2024, 27, 37. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Yi, C. Zhang Neural Networks and Neural-Dynamic Method; Nova Science Publishers, Inc.: Hauppauge, NY, USA, 2011. [Google Scholar]
  15. Zhang, C.; Gong, X.; Hu, Y.; Chen, H. A Discrete Sliding-Mode Reaching-Law Zeroing Neural Solution for Dynamic Constrained Quadratic Programming. IEEE Trans. Ind. Inform. 2025, 21, 4724–4733. [Google Scholar] [CrossRef]
  16. Fahad, N.M.; Thuseethan, S.; Azid, S.I.; Azam, S. An Innovative Coverage Path Planning Approach for UAVs to Boost Precision Agriculture and Rescue Operations. Int. J. Intell. Syst. 2025, 2025, 4700518. [Google Scholar] [CrossRef]
  17. Chen, D.; Jiang, J.; Wang, Z.; Li, S. A new visual-inertial odometry scheme for unmanned systems in unified framework of zeroing neural networks. Neurocomputing 2025, 617, 129017. [Google Scholar] [CrossRef]
  18. Liao, B.; Han, L.; Cao, X.; Li, S.; Li, J. Double integral-enhanced Zeroing neural network with linear noise rejection for time-varying matrix inverse. CAAI Trans. Intell. Technol. 2024, 9, 197–210. [Google Scholar] [CrossRef]
  19. Guo, D.; Zhang, C.; Cang, N.; Jia, Z.; Xue, S.; Zhang, W.; Li, S.; Wang, Y. Harmonic noise rejection zeroing neural network for time-dependent equality-constrained quadratic program and its application to robot arms. IEEE Trans. Ind. Inform. 2024, 21, 1279–1288. [Google Scholar] [CrossRef]
  20. Cai, J.; Zhang, W.; Zhong, S.; Yi, C. A super-twisting algorithm combined zeroing neural network with noise tolerance and finite-time convergence for solving time-variant Sylvester equation. Expert Syst. Appl. 2024, 248, 123380. [Google Scholar] [CrossRef]
  21. Jin, J.; Lei, X.; Chen, C.; Li, Z. A fuzzy activation function based zeroing neural network for dynamic Arnold map image cryptography. Math. Comput. Simul. 2025, 230, 456–469. [Google Scholar] [CrossRef]
  22. Qi, Z.; Ning, Y.; Xiao, L.; Luo, J.; Li, X. Finite-time zeroing neural networks with novel activation function and variable parameter for solving time-varying Lyapunov tensor equation. Appl. Math. Comput. 2023, 452, 128072. [Google Scholar] [CrossRef]
  23. Gao, Y.; Tang, Z.; Ke, Y.; Stanimirović, P.S. New activation functions and Zhangians in zeroing neural network and applications to time-varying matrix pseudoinversion. Math. Comput. Simul. 2024, 225, 1–12. [Google Scholar] [CrossRef]
  24. Han, C.; Xu, J.; Zheng, B. A Novel Robust and Predefined-Time Zeroing Neural Network Solver for Time-Varying Linear Matrix Equation. Math. Methods Appl. Sci. 2025, 48, 6048–6062. [Google Scholar] [CrossRef]
  25. Xiao, L.; Tao, J.; Dai, J.; Wang, Y.; Jia, L.; He, Y. A parameter-changing and complex-valued zeroing neural-network for finding solution of time-varying complex linear matrix equations in finite time. IEEE Trans. Ind. Inform. 2021, 17, 6634–6643. [Google Scholar] [CrossRef]
  26. Tang, W.; Cai, H.; Xiao, L.; He, Y.; Li, L.; Zuo, Q.; Li, J. A predefined-time adaptive zeroing neural network for solving time-varying linear equations and its application to UR5 robot. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 4703–4712. [Google Scholar] [CrossRef]
  27. Ning, B.; Han, Q.L.; Ge, X.; Sanjayan, J. A zeroing neural network-based approach to parameter-varying platooning control of connected automated vehicles. IEEE Trans. Intell. Veh. 2023, 9, 1977–1988. [Google Scholar]
  28. Chen, Y.; Chen, J.; Yi, C. Time-varying learning rate for recurrent neural networks to solve linear equations. Math. Methods Appl. Sci. 2022. [Google Scholar] [CrossRef]
  29. Fu, J.; Zhang, Y.; Geng, G.; Liu, Z. Recurrent neural network with scheduled varying gain for solving time-varying QP. IEEE Trans. Circuits Syst. II Express Briefs 2023, 71, 882–886. [Google Scholar] [CrossRef]
  30. Gerontitis, D.; Behera, R.; Tzekis, P.; Stanimirović, P. A family of varying-parameter finite-time zeroing neural networks for solving time-varying Sylvester equation and its application. J. Comput. Appl. Math. 2022, 403, 113826. [Google Scholar] [CrossRef]
  31. Katsikis, V.N.; Stanimirović, P.S.; Mourtas, S.D.; Xiao, L.; Karabašević, D.; Stanujkić, D. Zeroing neural network with fuzzy parameter for computing pseudoinverse of arbitrary matrix. IEEE Trans. Fuzzy Syst. 2021, 30, 3426–3435. [Google Scholar] [CrossRef]
  32. Ciano, T.; Ferrara, M. Karush-kuhn-tucker conditions and lagrangian approach for improving machine learning techniques: A survey and new developments. Atti Accad. Peloritana -Pericolanti-Cl. Sci. Fis. Mat. Nat. 2024, 102, 1. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.