Next Article in Journal
Impact of COVID-19 on Supply Chains: A Hybrid Trade Credit Policy
Next Article in Special Issue
On the Efficiency of Staggered C-Grid Discretization for the Inviscid Shallow Water Equations from the Perspective of Nonstandard Calculus
Previous Article in Journal
On a High-Precision Method for Studying Attractors of Dynamical Systems and Systems of Explosive Type
Previous Article in Special Issue
A Tensor Splitting AOR Iterative Method for Solving a Tensor Absolute Value Equation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Zeroing Neural Network for Pseudoinversion of an Arbitrary Time-Varying Matrix Based on Singular Value Decomposition

by
Mariya Kornilova
1,
Vladislav Kovalnogov
1,
Ruslan Fedorov
1,
Mansur Zamaleev
1,
Vasilios N. Katsikis
2,
Spyridon D. Mourtas
2 and
Theodore E. Simos
1,3,4,5,6,*
1
Laboratory of Inter-Disciplinary Problems in Clean Energy Production, Ulyanovsk State Technical University, 32 Severny Venetz Street, 432027 Ulyanovsk, Russia
2
Department of Economics, Division of Mathematics and Informatics, National and Kapodistrian University of Athens, Sofokleous 1 Street, 10559 Athens, Greece
3
Department of Medical Research, China Medical University Hospital, China Medical University, Taichung 40402, Taiwan
4
Data Recovery Key Laboratory of Sichun Province, Neijing Normal University, Neijiang 641100, China
5
Section of Mathematics, Department of Civil Engineering, Democritus University of Thrace, 67100 Xanthi, Greece
6
Department of Mathematics, University of Western Macedonia, 50100 Kozani, Greece
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(8), 1208; https://doi.org/10.3390/math10081208
Submission received: 11 March 2022 / Revised: 5 April 2022 / Accepted: 6 April 2022 / Published: 7 April 2022
(This article belongs to the Special Issue Numerical Analysis and Scientific Computing II)

Abstract

:
Many researchers have investigated the time-varying (TV) matrix pseudoinverse problem in recent years, for its importance in addressing TV problems in science and engineering. In this paper, the problem of calculating the inverse or pseudoinverse of an arbitrary TV real matrix is considered and addressed using the singular value decomposition (SVD) and the zeroing neural network (ZNN) approaches. Since SVD is frequently used to compute the inverse or pseudoinverse of a matrix, this research proposes a new ZNN model based on the SVD method as well as the technique of Tikhonov regularization, for solving the problem in continuous time. Numerical experiments, involving the pseudoinversion of square, rectangular, singular, and nonsingular input matrices, indicate that the proposed models are effective for solving the problem of the inversion or pseudoinversion of time varying matrices.

1. Introduction and Preliminaries

In this paper, the zeroing neural network (ZNN) approach is used to address the problem of calculating the inverse or pseudoinverse of an arbitrary time-varying (TV) real matrix. On the one hand, the pseudoinverse, or Moore–Penrose inverse, of A R m × n is the unique matrix A , such that the system of Penrose equations holds for X : = A [1,2,3]:
A = A X A , X = X A X , ( X A ) T = X A , ( A X ) T = A X ,
where A T denotes the transpose of A. Note that, if A is a nonsingular square matrix, A becomes the usual inverse A 1 . On the other hand, the singular value decomposition (SVD) of A R m × n is a factorization of the form [4]:
A = U S V T ,
where U R m × m and V R n × n are orthogonal matrices, i.e., U T = U 1 and V T = V 1 , while S R m × n is a rectangular (or square, in the case m = n ) diagonal matrix with the singular values of A on its main diagonal. SVD is frequently used to compute the inverse or pseudoinverse of a matrix, while commonly existing in fields of scientific research, such as medical treatment and industrial applications, lattice computing [5], automatic classification of electromyograms [6], and face recognition [7]. In a recent work [8], the authors provided a zeroing neural network for computing the singular value decomposition of an arbitrary matrix. This work move things one step further, by designing a new ZNN model for calculating the inverse or pseudoinverse of an arbitrary TV matrix based on the singular value decomposition. For comparison purposes, we build another model based on direct pseudoinversion in accordance with the paper [9], and the experiments section demonstrates the efficacy of the proposed SVD model.
Zhang et al. in [10], developed a ZNN design for generating online solutions to TV problems. It is worth noting that most ZNN based dynamical systems fall under the category of recurrent neural networks (RNN) that are designed to find equation zeros. As a consequence, numerous valuable research findings have been presented in the literature. Addressing generalized inversion problems [11,12], tensor and matrix inversion problems [13], systems of linear equations [14,15], systems of matrix equations [14,16], quadratic optimization problems [17], and diverse matrix functions approximation [18,19] are the main applications of ZNNs. The first stage in developing ZNN dynamics is to design an error function E ( t ) that is tailored to the underlying problem, commonly known as the Zhang function [20]. The second stage takes advantage of the proper dynamical evolution that follows:
E ˙ ( t ) = d E ( t ) d t = λ F E ( t ) ,
where E ˙ ( t ) R m × n is the time derivative of E ( t ) R m × n , λ > 0 is the design parameter that is used for scaling the convergence, while F ( · ) : R m × n R m × n means elementwise utilization of an odd and increasing activation function on E ( t ) . In our research, we will consider the ZNN evolution (3) under the linear activation function. That is,
E ˙ ( t ) = d E ( t ) d t = λ E ( t ) .
This work’s key points may be summarized as below:
  • A novel ZNN approach, which is based on SVD, is employed for solving the problem of calculating the pseudoinverse of an arbitrary TV real matrix.
  • Two ZNN models for calculating the pseudoinverse of an arbitrary TV matrix are offered: one called ZNNSVDP, which is based on SVD, and the other called ZNNP, which is based on a more direct approach to the problem and is offered for comparison purposes.
  • Four numerical experiments, involving the pseudoinversion of square, rectangular, singular, and nonsingular input matrices, indicate that both models are effective for solving the problem and that the ZNNSVDP model converges to the problem’s solution faster than the ZNNP model.
Additionally, it is worth mentioning some of the paper’s general notations: the symbols 1 n , 0 n denote a vector in R n consisting of ones and zeros, respectively; O n × n R n × n denotes a zero matrix of n × n dimensions; I n R n × n denotes the identity n × n matrix; ⊗ denotes Kronecker product; vec ( · ) denotes the vectorization technique; ⊙ denotes the Hadamard (or element wise) product; and · F denotes the matrix Frobenius norm.
The paper is constituted as follows. Section 2 and Section 3, respectively, define and analyse the ZNNSVDP and ZNNP models. Section 4 presents and discusses the results of four numerical experiments employing the pseudoinversion of square, rectangular, singular, and nonsingular input matrices. Lastly, the final remarks and conclusions are offered in Section 5.

2. Time-Varying Pseudoinverse Computation Based on SVD

This section presents and analyses the ZNNSVDP model for calculating the pseudoinverse of an arbitrary TV real matrix. Considering a smooth TV matrix A ( t ) R m × n , the inverse or pseudoinverse of A ( t ) based on SVD (2) is the following:
A 1 ( t ) = V ( t ) S 1 ( t ) U T ( t ) , m = n = rank ( A ( t ) ) A ( t ) = V ( t ) S ( t ) U T ( t ) , otherwise ,
where U ( t ) R m × m and V ( t ) R n × n are TV orthogonal matrices, and S ( t ) R m × n is a rectangular (or square, in the case m = n ) diagonal matrix with the singular values of A ( t ) on its main diagonal. Here, we consider decomposition so that the singular values of A ( t ) are in descending order on the main diagonal of S ( t ) . Based on (2) and (5), the ZNNSVDP model considers the following group of error functions for calculating the inverse or pseudoinverse of A ( t ) :
E 1 ( t ) = A ( t ) V ( t ) U ( t ) S ( t ) E 2 ( t ) = U T ( t ) U ( t ) I m E 3 ( t ) = V T ( t ) V ( t ) I n E 4 ( t ) = X ( t ) V ( t ) Y ( t ) U T ( t ) ,
where X ( t ) is the desired solution of the problem, i.e., the inverse or pseudoinverse of A ( t ) , and
Y ( t ) = S 1 ( t ) , m = n = rank ( A ( t ) ) S ( t ) , otherwise .
The following proposition about the structure and construction of the pseudoinverse of a diagonal matrix is offered, whereas [21] provides a full examination of this proposition.
Proposition 1.
For a rectangular (or a square singular) diagonal matrix B R m × n , let b 1 , b 2 , b w with w = rank ( B ) signify the elements of the main diagonal of B. Then, the pseudoinverse matrix of B is the following:
B = B ¯ 1 O w × ( n w ) O ( m w ) × w O ( m w ) × ( n w ) , with B ¯ 1 = 1 b 1 0 0 0 1 b 2 0 0 0 1 b w .
In addition, the first time derivative of (6) is the following:
E ˙ 1 ( t ) = A ˙ ( t ) V ( t ) + A ( t ) V ˙ ( t ) U ˙ ( t ) S ( t ) U ( t ) S ˙ ( t ) E ˙ 2 ( t ) = U ˙ T ( t ) U ( t ) + U T ( t ) U ˙ ( t ) E ˙ 3 ( t ) = V ˙ T ( t ) V ( t ) + V T ( t ) V ˙ ( t ) E ˙ 4 ( t ) = X ˙ ( t ) V ˙ ( t ) Y ( t ) U T ( t ) V ( t ) Y ˙ ( t ) U T ( t ) V ( t ) Y ( t ) U ˙ T ( t ) ,
where the first time derivative of Y ( t ) is the following [22]:
Y ˙ ( t ) = S ˙ 1 ( t ) = S 1 ( t ) S ˙ ( t ) S 1 ( t ) , m = n = rank ( A ( t ) ) S ˙ ( t ) = S ( t ) S ˙ ( t ) S ( t ) + ( S ( t ) ( S ( t ) ) T ) S ˙ T ( t ) ( I m S ( t ) S ( t ) ) + ( I n S ( t ) S ( t ) ) S ˙ T ( t ) ( ( S ( t ) ) T S ( t ) ) , otherwise ,
or equivalent
Y ˙ ( t ) = Y ( t ) S ˙ ( t ) Y ( t ) + ( Y ( t ) Y T ( t ) ) S ˙ T ( t ) ( I m S ( t ) Y ( t ) ) + ( I n Y ( t ) S ( t ) ) S ˙ T ( t ) ( Y T ( t ) Y ( t ) ) .
Notice that (11) is simplified to Y ˙ ( t ) = Y ( t ) S ˙ ( t ) Y ( t ) for m = n = rank ( A ( t ) ) . Then, combining (6), (9) and (11) with the ZNN design under the linear activation function (4), the following may be acquired:
A ˙ ( t ) V ( t ) + A ( t ) V ˙ ( t ) U ˙ ( t ) S ( t ) U ( t ) S ˙ ( t ) = λ E 1 ( t )        U ˙ T ( t ) U ( t ) + U T ( t ) U ˙ ( t ) = λ E 2 ( t )        V ˙ T ( t ) V ( t ) + V T ( t ) V ˙ ( t ) = λ E 3 ( t )        X ˙ ( t ) V ˙ ( t ) Y ( t ) U T ( t ) V ( t ) ( Y ( t ) S ˙ ( t ) Y ( t ) + ( Y ( t ) Y T ( t ) ) S ˙ T ( t ) ( I m S ( t ) Y ( t ) )        + ( I n Y ( t ) S ( t ) ) S ˙ T ( t ) ( Y T ( t ) Y ( t ) ) ) U T ( t ) V ( t ) Y ( t ) U ˙ T ( t ) = λ E 4 ( t ) .     
Using vectorization and the Kronecker product, the dynamics of (12) are modified as follows:
( S T ( t ) I m ) vec ( U ˙ ( t ) ) + ( I n A ( t ) ) vec ( V ˙ ( t ) ) ( I n U ( t ) ) vec ( S ˙ ( t ) ) = vec ( λ E 1 ( t ) A ˙ ( t ) V ( t ) ) ( U T ( t ) I m ) vec ( U ˙ T ( t ) ) + ( I m U T ( t ) ) vec ( U ˙ ( t ) ) = vec ( λ E 2 ( t ) ) ( V T ( t ) I n ) vec ( V ˙ T ( t ) ) + ( I n V T ( t ) ) vec ( V ˙ ( t ) ) = vec ( λ E 3 ( t ) ) ( I m V ( t ) Y ( t ) ) vec ( U ˙ T ( t ) ) ( U ( t ) Y T ( t ) I n ) vec ( V ˙ ( t ) ) K 1 ( t ) + I m n vec ( X ˙ ( t ) ) = vec ( λ E 4 ( t ) ) ,
where
K 1 ( t ) = U ( t ) Y T ( t ) V ( t ) Y ( t ) vec ( S ˙ ( t ) ) + U ( t ) ( I m S ( t ) Y ( t ) ) T V ( t ) Y ( t ) Y T ( t ) vec ( S ˙ T ( t ) ) + U ( t ) Y T ( t ) Y ( t ) V ( t ) ( I n Y ( t ) S ( t ) ) vec ( S ˙ T ( t ) )
Note that (13) must be simplified in order to produce a simple and explicit dynamical model that may easily calculate U ( t ) , V ( t ) , S ( t ) , and X ( t ) . As a result, the following lemmas about vectorization and the Kronecker product are offered, whereas [23] provides a full examination of Lemmas’ 1 and 2 content.
Lemma 1.
( L T A ) vec ( X ) = vec ( Y ) .
Lemma 2.
For B R m × m , let vec ( B ) R m 2 signify the matrix B vectorization. The following occurs:
vec ( B T ) = Q m vec ( B ) ,
where Q m R m 2 × m 2 is a constant permutation matrix defined exclusively by m.
The following, Algorithm 1, presents an algorithmic process for obtaining the permutation matrix Q m in (15), which corresponds to a matrix of m × m dimensions. Note that the notations eye(.) and reshape(.) in Algorithm 1 have the typical notion of the related MATLAB functions [24].
Algorithm 1 Permutation matrix calculation
Require: The rows or columns number m of a square matrix B R m × m .
 1: procedurePermutation_Matrix(m)
 2:      Set a = eye ( m 2 ) and b = reshape ( 1 : m 2 , m , m )
 3:      return  Q = a ( : , reshape ( b , 1 , m 2 ) )
 4: end procedure
Ensure: Q m , i.e., the permutation matrix.
Furthermore, because S ( t ) and S T ( t ) are rectangular (or square, in the case m = n ) diagonal matrices, just the nonzero elements of S ˙ ( t ) and S ˙ T ( t ) that are placed in their main diagonal must be obtained. By doing so, we may confine S ( t ) to being a diagonal matrix, while also reducing the dimensions of (13). Hence, employing the nonzero elements on the main diagonal of S ( t ) and S T ( t ) , whose number is w = rank ( A ( t ) ) , we utilize the equations vec ( S ˙ ( t ) ) = G 1 s ˙ ( t ) and vec ( S ˙ T ( t ) ) = G 2 s ˙ ( t ) , respectively, to replace S ˙ ( t ) and S ˙ T ( t ) in (14), where the matrices G 1 , G 2 R m n × w are operational matrices that can be calculated using the algorithmic procedure presented Algorithm 2. Additionally, the notation sum(.), min(.), zeros(.), mod(.) and floor(.) in Algorithm 2 have the typical notion of the related MATLAB functions [24].
Algorithm 2 Operational matrix calculation
Require: The number of the rows and columns, respectively, m and n of a matrix B R m × n ,
   and w = rank ( B ) .
 1:  procedureOperational_Matrix( m , n , w )
 2:     if  w < min ( m , n )  then
 3:        Set h = w
 4:    else
 5:        Set h = m
 6:    end if
 7:    Set G = zeros ( m n , w )
 8:    for  k = 1 : m n  do
 9:        Set c = mod ( k 1 , h ) + 1 and d = floor ( k 1 h ) + 1
 10:        if  d = = c  then
 11:           Set G ( k , c ) = 1
 12:        end if
 13:    end for
 14:    return G
 15: end procedure
Ensure: The operational matrix G.
Based on the aforementioned discussion, (13) can be reformulated as follows:
( S T ( t ) I m ) vec ( U ˙ ( t ) ) + ( I n A ( t ) ) vec ( V ˙ ( t ) ) ( I n U ( t ) ) G 1 s ˙ ( t ) = vec ( λ E 1 ( t ) A ˙ ( t ) V ( t ) ) ( U T ( t ) I m ) Q m vec ( U ˙ ( t ) ) + ( I m U T ( t ) ) vec ( U ˙ ( t ) ) = vec ( λ E 2 ( t ) ) ( V T ( t ) I n ) Q n vec ( V ˙ ( t ) ) + ( I n V T ( t ) ) vec ( V ˙ ( t ) ) = vec ( λ E 3 ( t ) ) ( I m V ( t ) Y ( t ) ) vec ( U ˙ T ( t ) ) ( U ( t ) Y T ( t ) I n ) vec ( V ˙ ( t ) ) K 2 ( t ) s ˙ ( t ) + I m n vec ( X ˙ ( t ) ) = vec ( λ E 4 ( t ) ) ,
where
K 2 ( t ) = U ( t ) Y T ( t ) V ( t ) Y ( t ) G 1 + U ( t ) ( I m S ( t ) Y ( t ) ) T V ( t ) Y ( t ) Y T ( t ) G 2 + U ( t ) Y T ( t ) Y ( t ) V ( t ) ( I n Y ( t ) S ( t ) ) G 2
As a result, setting
Z 1 ( t ) = ( S T ( t ) I m ) ( U T ( t ) I m ) Q m + ( I m U T ( t ) ) 0 n 2 × m 2 ( I m V ( t ) Y ( t ) ) , Z 3 ( t ) = ( I n U ( t ) ) G 1 0 m 2 × w 0 n 2 × w K 2 ( t ) , Z 2 ( t ) = ( I n A ( t ) ) 0 m 2 × n 2 ( V T ( t ) I n ) Q n + ( I n V T ( t ) ) ( U ( t ) Y T ( t ) I n ) , Z 4 ( t ) = 0 m n × m n 0 m 2 × m n 0 n 2 × m n I m n ,     Z ( t ) = Z 1 ( t ) Z 2 ( t ) Z 3 ( t ) Z 4 ( t ) ,                  
q ( t ) = vec ( λ E 1 ( t ) A ˙ ( t ) V ( t ) ) vec ( λ E 2 ( t ) ) vec ( λ E 3 ( t ) ) vec ( λ E 4 ( t ) ) T , x ˙ = vec ( U ˙ ( t ) ) vec ( V ˙ ( t ) ) vec ( s ˙ ( t ) ) vec ( X ˙ ( t ) ) , x = vec ( U ( t ) ) vec ( V ( t ) ) vec ( s ( t ) ) vec ( X ( t ) ) ,
we propose the following ZNN model:
Z T ( t ) Z ( t ) x ˙ ( t ) = Z T ( t ) q ( t ) ,
where Z T ( t ) Z ( t ) is a singular mass matrix. To solve the singularity problem, the Tikhonov regularization is used and (20) is converted into:
( Z T ( t ) Z ( t ) + β I m 2 + n 2 + w + m n ) x ˙ ( t ) = Z T ( t ) q ( t ) ,
where β 0 signifies the regularization parameter. The ZNN model (21) is termed as the ZNNSVDP model and can be solved efficiently with an appropriate ode Matlab solver. The exponential convergence of the ZNNSVDP model (21) to the theoretical TV inverse or pseudoinverse of the input matrix A ( t ) is proven in Theorem 1.
Remark 1.
According to MATLAB’sodesolvers syntax [24], the mass matrix is a symmetric matrix M that expresses the connection between the time derivative x ˙ of the generalized coordinate vector x of a system, by the equation:
M x ˙ = x .
Theorem 1.
Let U ( t ) R m × m , V ( t ) , R n × n , S ( t ) R m × n be differentiable and S ( t ) be a rectangular diagonal matrix. The ZNNSVDP model (21), starting from any initial value x ( 0 ) , converges exponentially to the theoretical TV inverse or pseudoinverse of the input matrix A ( t ) .
Proof. 
In order to obtain the solution x ( t ) , which corresponds to the TV inverse or pseudoinverse of the input matrix A ( t ) , the error matrix equation group is defined as in (9), inline with the ZNN design. Following that, by adopting the linear design formula for zeroing (9), the model (12) is obtained. From [Theorem 1] [10], each error matrix equation in the error matrix equation group (12) converges exponentially to the theoretical solution when t . As a result, the solution of (12) converges exponentially to the theoretical TV inverse or pseudoinverse of the input matrix A ( t ) when t . Furthermore, from the derivation procedure of (21) from (12), the proof is completed.    □

3. Alternative Time-Varying Pseudoinverse Computation

This section presents and analyzes a ZNN model, namely, ZNNP, for calculating the pseudoinverse of any TV real matrix, based on a recent work [9] on ZNN pseudoinverse computation, and this new model will serve as a strong and fair competitor to the proposed ZNNSVDP model. Considering a smooth TV matrix A ( t ) R m × n , then, if rank ( A ( t ) ) = n < m , the MP inverse A ( t ) becomes the left inverse A ( t ) = A L 1 ( t ) ( A T ( t ) A ( t ) ) 1 A T ( t ) of A ( t ) , where A T ( t ) A ( t ) A ( t ) = A T ( t ) . Otherwise, if rank ( A ( t ) ) = n > m , the MP inverse A ( t ) becomes the right inverse A ( t ) = A R 1 ( t ) A T ( t ) ( A ( t ) A T ( t ) ) 1 , where A ( t ) A ( t ) A T ( t ) = A T ( t ) . Therefore, we can design a ZNN model according to the following equations:
A T ( t ) A ( t ) A 1 ( t ) = A T ( t ) , m = n = rank ( A ( t ) ) A T ( t ) A ( t ) A ( t ) = A T ( t ) , rank ( A ( t ) ) n < m A ( t ) A ( t ) A T ( t ) = A T ( t ) , rank ( A ( t ) ) m < n .
Based on (22), the ZNNP model considers the following error function for calculating the inverse or pseudoinverse of A ( t ) :
E D ( t ) = A T ( t ) A ( t ) X ( t ) A T ( t ) , rank ( A ( t ) ) n m X ( t ) A ( t ) A T ( t ) A T ( t ) , rank ( A ( t ) ) m < n .
where X ( t ) is the desired solution of the problem, i.e., the inverse or pseudoinverse of A ( t ) . Furthermore, the first time derivative of (23) is the following:
E ˙ D ( t ) = A ˙ T ( t ) A ( t ) X ( t ) + A T ( t ) A ˙ ( t ) X ( t ) + A T ( t ) A ( t ) X ˙ ( t ) A ˙ T ( t ) , rank ( A ( t ) ) n m X ˙ ( t ) A ( t ) A T ( t ) + X ( t ) A ˙ ( t ) A T ( t ) + X ( t ) A ( t ) A ˙ T ( t ) A ˙ T ( t ) , rank ( A ( t ) ) m < n .
Then, combining (23) and (24) with the ZNN design (4), under the linear activation function, the following can be obtained:
A ˙ T ( t ) A ( t ) X ( t ) + A T ( t ) A ˙ ( t ) X ( t ) + A T ( t ) A ( t ) X ˙ ( t ) A ˙ T ( t ) = λ ( A T ( t ) A ( t ) X ( t ) A T ( t ) ) , rank ( A ( t ) ) n m X ˙ ( t ) A ( t ) A T ( t ) + X ( t ) A ˙ ( t ) A T ( t ) + X ( t ) A ( t ) A ˙ T ( t ) A ˙ T ( t ) = λ ( X ( t ) A ( t ) A T ( t ) A T ( t ) ) , rank ( A ( t ) ) m < n .
Using vectorization and the Kronecker product, the dynamics of (25) are modified as follows:
( I m A T ( t ) A ( t ) ) vec ( X ˙ ( t ) ) = vec ( λ ( A T ( t ) A ( t ) X ( t ) A T ( t ) ) A ˙ T ( t ) A ( t ) X ( t ) A T ( t ) A ˙ ( t ) X ( t ) + A ˙ T ( t ) ) , rank ( A ( t ) ) n m ( A ( t ) A T ( t ) I n ) vec ( X ˙ ) ( t ) = vec ( λ ( X ( t ) A ( t ) A T ( t ) A T ( t ) ) X ( t ) A ˙ ( t ) A T ( t ) X ( t ) A ( t ) A ˙ T ( t ) + A ˙ T ( t ) ) , rank ( A ( t ) ) m < n .
As a result, setting
L ( t ) = I m A T ( t ) A ( t ) + β I m n , rank ( A ( t ) ) < n m I m A T ( t ) A ( t ) , rank ( A ( t ) ) = n m A ( t ) A T ( t ) I n + β I m n , rank ( A ( t ) ) < m < n A ( t ) A T ( t ) I n , rank ( A ( t ) ) = m < n ,             r ( t ) = vec ( λ ( A T ( t ) A ( t ) X ( t ) A T ( t ) ) A ˙ T ( t ) A ( t ) X ( t ) A T ( t ) A ˙ ( t ) X ( t ) + A ˙ T ( t ) ) , rank ( A ( t ) ) n m vec ( λ ( X ( t ) A ( t ) A T ( t ) A T ( t ) ) X ( t ) A ˙ ( t ) A T ( t ) X ( t ) A ( t ) A ˙ T ( t ) + A ˙ T ( t ) ) , rank ( A ( t ) ) m < n . x ˙ ( t ) = vec ( X ˙ ( t ) ) , x ( t ) = vec ( X ( t ) ) ,                   
where β 0 signifies the Tikhonov regularization parameter, we have the next ZNN model:
L ( t ) x ˙ ( t ) = r ( t ) ,
where L ( t ) is a mass matrix. Note that the Tikhonov regularization is used in L ( t ) to solve the singularity problem of the cases rank ( A ( t ) ) < n m and rank ( A ( t ) ) < m < n , respectively, because the products A ( t ) A T ( t ) and A T ( t ) A ( t ) result to singular matrices. The ZNN model (28) is termed as the ZNNP model and can be solved efficiently with an ode Matlab solver, while its exponential convergence to the theoretical TV inverse or pseudoinverse of the input matrix A ( t ) is proven in Theorem 2.
Theorem 2.
The ZNNP model (28) starting form any initial value x ( 0 ) , converges exponentially to the theoretical TV inverse or pseudoinverse of the input matrix A ( t ) .
Proof. 
In order to obtain the solution x ( t ) , which corresponds to the TV inverse or pseudoinverse of the input matrix A ( t ) , the error matrix equation group is defined as in (23), inline with the ZNN design. Following that, by adopting the linear design formula for zeroing (23), the model (25) is obtained. From [Theorem 1] [10], each error matrix equation in the error matrix equation group (25) converges to the theoretical solution when t . As a consequence, the solution of (25) converges to the theoretical TV inverse or pseudoinverse of the input matrix A ( t ) when t . Moreover, from the derivation procedure of (28), we know it is (25) in a different form. The proof is, thus, completed.    □

4. Numerical Experiments

This section compares and contrasts the performances of the ZNNSVDP model (21) with the ZNNP model (28) on four numerical experiments (NE), involving the pseudoinversion of square, rectangular, singular, and nonsingular input matrices. In all NE, the time interval is restricted to [ 0 , 10 ] during the computation, which indicates that the starting time is t 0 = 0 and the ending time is t f = 10 , while the ZNN design parameter has been set to λ = 10 and the Tikhonov regularization parameter has been set to β = 1 e 8 . It is worth mentioning that the notation ZNNSVDP and ZNNP in the legends of Figure 1, respectively, denote the solutions produced by the ZNNSVDP and ZNNP models. Lastly, the MATLAB solver ode45 has been used, while the initial value for both models has been set to x ( 0 ) = sign ( x * ( 0 ) ) , where x * ( 0 ) is the theoretical solution at t = 0 and sign is the signum function.

4.1. Experiment 1

This NE deals with the inversion of the following square matrix:
A ( t ) = 4 / ( t + 18 ) sin ( t ) + 2 ( t + 18 ) / ( 2 t + 2 ) cos ( t ) 20 .
Note that A ( t ) is a full rank matrix with dimensions 2 × 2 .

4.2. Experiment 2

This NE deals with the pseudoinversion of the following rectangular matrix:
A ( t ) = 3 sin ( t ) + 7 3 + cos ( t ) sin ( 2 t ) + 4 5 sin ( t ) 4 / ( t + 8 ) 7 sin ( t ) 7 / 2 + sin ( 3 t ) 4 + cos ( t ) 6 + cos ( 3 t ) 3 sin ( t ) cos ( t ) 20 7 / 2 + sin ( 5 t ) 5 sin ( t ) sin ( t ) + 1 3 + cos ( t ) .
Notice that A ( t ) is a full column rank matrix with dimensions 5 × 3 .

4.3. Experiment 3

The pseudoinversion of the following rectangular matrix is the subject of this NE:
A ( t ) = 5 cos ( π t ) 3 + sin ( π t ) 4 cos ( t ) 1 + 3 sin ( t ) T 1 4 × 2 .
The matrix A ( t ) is rank deficient, with rank ( A ( t ) ) = 1 , and its dimensions are 4 × 2 .

4.4. Experiment 4

This NE is related to the pseudoinversion of the rectangular matrix given below:
A ( t ) = 2 + sin ( t ) 2 + 1 / 2 sin ( t ) 2 + 1 / n sin ( t ) 1 m × n .
With rank ( A ( t ) ) = 1 , the matrix A ( t ) is rank deficient, and its dimensions are m × n , where m = 4 and n = 9 .

4.5. Analysis of Numerical Experiments—Results and Comparison

The performance of the ZNNSVDP and ZNNP models for calculating the inverse or pseudoinverse of an arbitrary matrix A ( t ) is investigated through the four NE defined in Section 4.1 Section 4.2, Section 4.3 and Section 4.4. For all the experiments, the results produced by the ZNNSVDP and ZNNP models are depicted in Figure 1. It is worth noting that Figure 1 has the following layout: the first column figures show the convergence of the error function, i.e., E i ( t ) , i = 1 , , 4 , of the ZNNSVDP model and E D ( t ) of the ZNNP model; the second column figures show the convergence of the models according to the appropriate error function, i.e., residual errors; the third column figures show the trajectories of the solutions generated by the models.
The following can be deduced from the NE of this section. Overall, the error functions of the ZNNSVDP model, i.e., E i ( t ) , i = 1 , , 4 , receive lower values than the error function of the ZNNP model, i.e., E D ( t ) , in all NE, as depicted in Figure 1a,d,g,j. When X ( t ) corresponds to the solution of the ZNNSVDP model rather than the solution of the ZNNP model, the convergence in Figure 1b,h,k is faster, while the convergence in Figure 1e is almost identical. It is worth noting that Figure 1b depicts the residual error I A ( t ) X ( t ) F in the case of NE Section 4.1, Figure 1e depicts the residual error I X ( t ) A ( t ) F in the case of NE Section 4.2, Figure 1h depicts the residual error A T ( t ) X ( t ) A ( t ) A T ( t ) F in the case of NE Section 4.3, and Figure 1b depicts the residual error A T ( t ) A T ( t ) A ( t ) X ( t ) F in the case of NE Section 4.4. Finally, Figure 1c,f,i,l show that both models’ solutions match the theoretical inverse in the case of NE Section 4.1, and the theoretical pseudoinverse in the cases of NE Section 4.2, Section 4.3 and Section 4.4.
According to the presented NE, the following are some general generalizations that can be drawn. The ZNNSVDP model presented in this paper, which is based on the SVD method, shows better performances than the ZNNP model, which is based on a more direct approach for calculating the inverse or pseudoinverse. In addition, the ZNNSVDP model generates the minimum amount for the Frobenius norm of both the error functions and the residual errors. It is also important to note that, for both models, the larger the value of the design parameter λ , the higher the degree of convergence.

5. Conclusions

The problem of calculating the inverse or pseudoinverse of an arbitrary TV real matrix is addressed using the ZNN approach in this paper. Two ZNN models for calculating the inverse or pseudoinverse of an arbitrary TV matrix, one called ZNNSVDP, which is based on SVD, and the other called ZNNP, which is based on a more direct approach to the problem, are defined, analysed and compared. Four numerical experiments, involving the pseudoinversion of square, rectangular, singular, and nonsingular input matrices, indicate that both models are effective for solving the problem and that the ZNNSVDP model converges to the problem’s solution faster than the ZNNP model.
Some potential study areas can be identified:
  • It is possible to explore the streams of the ZNNSVDP and ZNNP models that are accelerated by a nonlinear activation function, as well as nonlinear ZNNSVDP and ZNNP model flows, with a terminal convergence in this direction.
  • Another option is to use carefully chosen fuzzy parameters to define future ZNN dynamics upgrades.
  • The presented ZNNSVDP and ZNNP models have the drawback of not being noise tolerant, because all types of noise have a substantial impact on the accuracy of the proposed ZNN approaches. As a consequence, future research could focus on adapting the ZNNSVDP and ZNNP models to an integration enhanced and noise-handling ZNN class of dynamical systems.

Author Contributions

M.K.: conceptualization, methodology. V.K.: validation, investigation. R.F.: formal analysis, investigation. M.Z.: methodology, investigation. V.N.K.: conceptualization, methodology, validation, formal analysis, investigation, writing—original draft. S.D.M.: conceptualization, methodology, validation, formal analysis, investigation, writing—original draft. T.E.S.: methodology, formal analysis, investigation. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by a Mega Grant from the Government of the Russian Federation within the framework of federal project No. 075-15-2021-584.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Penrose, R. A generalized inverse for matrices. Proc. Cambridge Philos. Soc. 1955, 51, 406–413. [Google Scholar] [CrossRef] [Green Version]
  2. Sayevand, K.; Pourdarvish, A.; Machado, J.A.T.; Erfanifar, R. On the Calculation of the Moore-Penrose and Drazin Inverses: Application to Fractional Calculus. Mathematics 2021, 9, 2501. [Google Scholar] [CrossRef]
  3. Chien, M.T. Numerical Range of Moore-Penrose Inverse Matrices. Mathematics 2020, 8, 830. [Google Scholar] [CrossRef]
  4. Crane, D.K.; Gockenbach, M.S. The Singular Value Expansion for Arbitrary Bounded Linear Operators. Mathematics 2020, 8, 1346. [Google Scholar] [CrossRef]
  5. Valverde-Albacete, F.J.; Peláez-Moreno, C. The Singular Value Decomposition over Completed Idempotent Semifields. Mathematics 2020, 8, 1577. [Google Scholar] [CrossRef]
  6. Hazarika, A.; Barthakur, M.; Dutta, L.; Bhuyan, M. F-SVD based algorithm for variability and stability measurement of bio-signals, feature extraction and fusion for pattern recognition. Biomed. Signal Process. Control 2019, 47, 26–40. [Google Scholar] [CrossRef]
  7. Wang, J.; Le, N.T.; Lee, J.; Wang, C. Illumination compensation for face recognition using adaptive singular value decomposition in the wavelet domain. Inf. Sci. 2018, 435, 69–93. [Google Scholar] [CrossRef]
  8. Chen, J.; Zhang, Y. Online singular value decomposition of time-varying matrix via zeroing neural dynamics. Neurocomputing 2020, 383, 314–323. [Google Scholar] [CrossRef]
  9. Katsikis, V.N.; Stanimirović, P.S.; Mourtas, S.D.; Xiao, L.; Karabasević, D.; Stanujkić, D. Zeroing Neural Network with Fuzzy Parameter for Computing Pseudoinverse of Arbitrary Matrix. IEEE Trans. Fuzzy Syst. 2021; Early Access. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Ge, S.S. Design and analysis of a general recurrent neural network model for time-varying matrix inversion. IEEE Trans. Neural Netw. 2005, 16, 1477–1490. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Wang, X.; Che, M.; Wei, Y. Recurrent neural network for computation of generalized eigenvalue problem with real diagonalizable matrix pair and its applications. Neurocomputing 2016, 216, 230–241. [Google Scholar] [CrossRef]
  12. Stanimirović, P.S.; Katsikis, V.N.; Zhang, Z.; Li, S.; Chen, J.; Zhou, M. Varying-parameter Zhang neural network for approximating some expressions involving outer inverses. Optim. Methods Softw. 2020, 35, 1304–1330. [Google Scholar] [CrossRef]
  13. Ma, H.; Li, N.; Stanimirović, P.S.; Katsikis, V.N. Perturbation theory for Moore–Penrose inverse of tensor via Einstein product. Comput. Appl. Math. 2019, 38, 111. [Google Scholar] [CrossRef]
  14. Katsikis, V.N.; Mourtas, S.D.; Stanimirović, P.S.; Zhang, Y. Solving Complex-Valued Time-Varying Linear Matrix Equations via QR Decomposition With Applications to Robotic Motion Tracking and on Angle-of-Arrival Localization. IEEE Trans. Neural Netw. Learn. Syst. 2021; Early Access. [Google Scholar] [CrossRef] [PubMed]
  15. Stanimirović, P.S.; Katsikis, V.N.; Li, S. Hybrid GNN-ZNN models for solving linear matrix equations. Neurocomputing 2018, 316, 124–134. [Google Scholar] [CrossRef]
  16. Stanimirović, P.S.; Katsikis, V.N.; Li, S. Integration enhanced and noise tolerant ZNN for computing various expressions involving outer inverses. Neurocomputing 2019, 329, 129–143. [Google Scholar] [CrossRef]
  17. Zhang, Z.; Yang, S.; Zheng, L. A Penalty Strategy Combined Varying-Parameter Recurrent Neural Network for Solving Time-Varying Multi-Type Constrained Quadratic Programming Problems. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 2993–3004. [Google Scholar] [CrossRef] [PubMed]
  18. Katsikis, V.N.; Stanimirović, P.S.; Mourtas, S.D.; Li, S.; Cao, X. Chapter Towards Higher Order Dynamical Systems. In Generalized Inverses: Algorithms and Applications; Mathematics Research Developments, Nova Science Publishers, Inc.: Hauppauge, NY, USA, 2021; pp. 207–239. [Google Scholar]
  19. Katsikis, V.N.; Mourtas, S.D.; Stanimirović, P.S.; Zhang, Y. Continuous-Time Varying Complex QR Decomposition via Zeroing Neural Dynamics. Neural Process. Lett. 2021, 53, 3573–3590. [Google Scholar] [CrossRef]
  20. Zhang, Y.; Guo, D. Zhang Functions and Various Models; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar] [CrossRef]
  21. Ben-Israel, A.; Greville, T.N.E. Generalized Inverses: Theory and Applications, 2nd ed.; CMS Books in Mathematics; Springer: New York, NY, USA, 2003. [Google Scholar] [CrossRef]
  22. Golub, G.H.; Pereyra, V. The Differentiation of Pseudo-Inverses and Nonlinear Least Squares Problems Whose Variables Separate. SIAM J. Numer. Anal. 1973, 10, 413–432. [Google Scholar] [CrossRef]
  23. Graham, A. Kronecker Products and Matrix Calculus with Applications; Courier Dover Publications: Mineola, NY, USA, 2018. [Google Scholar]
  24. Gupta, A.K. Numerical Methods Using MATLAB; MATLAB Solutions Series; Springer: New York, NY, USA, 2014. [Google Scholar]
Figure 1. The convergence of ZFs and the solutions’ convergence and trajectories in NEs Section 4.1, Section 4.2, Section 4.3 and Section 4.4. (a) NE Section 4.1: Convergence of ZFs. (b) NE Section 4.1: Solutions convergence. (c) NE Section 4.1: Solutions trajectories. (d) NE Section 4.2: Convergence of ZFs. (e) NE Section 4.2: Solutions convergence. (f) NE Section 4.2: Solutions trajectories. (g) NE Section 4.3: Convergence of ZFs. (h) NE Section 4.3: Solutions convergence. (i) NE Section 4.3: Solutions trajectories. (j) NE Section 4.4: Convergence of ZFs. (k) NE Section 4.4: Solutions convergence. (l) NE Section 4.4: Solutions trajectories.
Figure 1. The convergence of ZFs and the solutions’ convergence and trajectories in NEs Section 4.1, Section 4.2, Section 4.3 and Section 4.4. (a) NE Section 4.1: Convergence of ZFs. (b) NE Section 4.1: Solutions convergence. (c) NE Section 4.1: Solutions trajectories. (d) NE Section 4.2: Convergence of ZFs. (e) NE Section 4.2: Solutions convergence. (f) NE Section 4.2: Solutions trajectories. (g) NE Section 4.3: Convergence of ZFs. (h) NE Section 4.3: Solutions convergence. (i) NE Section 4.3: Solutions trajectories. (j) NE Section 4.4: Convergence of ZFs. (k) NE Section 4.4: Solutions convergence. (l) NE Section 4.4: Solutions trajectories.
Mathematics 10 01208 g001
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kornilova, M.; Kovalnogov, V.; Fedorov, R.; Zamaleev, M.; Katsikis, V.N.; Mourtas, S.D.; Simos, T.E. Zeroing Neural Network for Pseudoinversion of an Arbitrary Time-Varying Matrix Based on Singular Value Decomposition. Mathematics 2022, 10, 1208. https://doi.org/10.3390/math10081208

AMA Style

Kornilova M, Kovalnogov V, Fedorov R, Zamaleev M, Katsikis VN, Mourtas SD, Simos TE. Zeroing Neural Network for Pseudoinversion of an Arbitrary Time-Varying Matrix Based on Singular Value Decomposition. Mathematics. 2022; 10(8):1208. https://doi.org/10.3390/math10081208

Chicago/Turabian Style

Kornilova, Mariya, Vladislav Kovalnogov, Ruslan Fedorov, Mansur Zamaleev, Vasilios N. Katsikis, Spyridon D. Mourtas, and Theodore E. Simos. 2022. "Zeroing Neural Network for Pseudoinversion of an Arbitrary Time-Varying Matrix Based on Singular Value Decomposition" Mathematics 10, no. 8: 1208. https://doi.org/10.3390/math10081208

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop