Next Article in Journal
Approximation Properties of Solutions of a Mean Value-Type Functional Inequality, II
Next Article in Special Issue
Strong, Strongly Universal and Weak Interval Eigenvectors in Max-Plus Algebra
Previous Article in Journal
Detecting Inverse Boundaries by Weighted High-Order Gradient Collocation Method
Previous Article in Special Issue
Characteristic Polynomials and Eigenvalues for Certain Classes of Pentadiagonal Matrices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tensor Global Extrapolation Methods Using the n-Mode and the Einstein Products

1
LABMIA-SI, Department of Mathematics, University Mohammed V Rabat, Rabat 10000, Morocco
2
Department of Mathematics LMPA, 50 rue F. Buisson, ULCO, 62100 Calais, France
3
Laboratory of Modeling Simulation, Mohammed VI Polytechnic University, Hay Moulay Rachid, Ben Guerir 43150, Morocco
*
Author to whom correspondence should be addressed.
Mathematics 2020, 8(8), 1298; https://doi.org/10.3390/math8081298
Submission received: 9 June 2020 / Revised: 21 July 2020 / Accepted: 2 August 2020 / Published: 5 August 2020
(This article belongs to the Special Issue Matrix Structures: Numerical Methods and Applications)

Abstract

:
In this paper, we present new Tensor extrapolation methods as generalizations of well known vector, matrix and block extrapolation methods such as polynomial extrapolation methods or ϵ -type algorithms. We will define new tensor products that will be used to introduce global tensor extrapolation methods. We discuss the application of these methods to the solution of linear and non linear tensor systems of equations and propose an efficient implementation of these methods via the global-QR decomposition.

1. Introduction

Scalar extrapolation methods have been developed to accelerate the convergence of some sequences of numbers in R or C . This process consists in transforming a sequence ( s n ) converging to s to new ones by applying some transformation T k , k = 1 , defined as follows
T k : s n T k ( n ) = s n + i = 1 k 1 a i ( n ) g i ( n ) ,
where g i ( n ) , i = 1 , , k 1 is a scalar sequence that defines the method. One of the earlier extrapolation method is the well known Aitken’s Δ 2 process [1] defined by
T 2 ( n ) = s n Δ s n Δ 2 s n ,
where the first and the second forward differences are defined by Δ s n = s n + 1 s n and Δ 2 s n = Δ s n + 1 Δ s n and in that case we have g 1 ( n ) = Δ s n . Such a process was first proposed in 1926 by Aitken in [2]. It is well known that under some assumptions, the new Aitken sequence T 2 ( n ) will converge faster than ( s n ) to the same limit s; see [1] for more details. Vector extrapolation methods have been proposed the last decades and among them are the minimal polynomial extrapolation (MPE) [3,4], the reduced rank extrapolation (RRE) method [5] and the modified minimal polynomial extrapolation (MMPE) [6,7]. For more details, see [1,8,9,10,11,12]. A second class of vector sequence transformations contains the topological ϵ -algorithm (TEA) [6]. Applications for solving large linear and nonlinear systems of equations have been considered in [8,13].
Vector extrapolation methods were used in many applications such as google page rank by Golub et al. [14,15] and in other fields such as in statistics [16] or for solving discretized Navier–Stokes problems [17]. In the present paper, we consider tensor sequence transformations and propose new tensor extrapolation methods that generalize the classical vector ones. Using the Einstein product, we define some new tensor products that will allow us to develop the new methods based on orthogonal or oblique projections onto subspaces of small dimensions. It will be shown that when the tensor sequence is generated linearly then the proposed methods are theoretically equivalent to some tensor Krylov subspace methods such as the tensor version of GMRES and Lanczos methods developed recently in [18,19].
The remainder of this paper is organized as follows. In Section 2, we give notations, some basic definitions and properties related to tensors. In Section 3, we introduce the tensor versions of the vector polynomial extrapolation methods namely the Tensor Global Reduced Rank Extrapolation (TG-RRE), the Tensor Global Minimal Polynomial Extrapolation (TG-MPE) and the Tensor Global Modified Minimal Polynomial Extrapolation (TG-MMPE). We also give a Global tensor version of the topological ϵ -transformation. Section 4 describes the application of the proposed methods for solving linear and nonlinear tensor system of equations. In Section 5, we introduce efficient implementations via the tensor global-QR decomposition and in the last section, we present some numerical experiments.

2. Preliminaries and Notations

In this section, we briefly review some concepts and notions that are used throughout the paper. A tensor is a multidimensional array of data and a natural extension of scalars, vectors and matrices to a higher order, a scalar is a 0th order tensor, a vector is a 1th-order tensor and a matrix is 2th-order tensor. The tensor order is the number of its indices, which is called modes or ways. For a given N-mode tensor X R I 1 × I 2 × × I N , the notation x i 1 , , i N (with 1 i j I j , j = 1 , N ) stands for element i 1 , , i N of the tensor X . Corresponding to a given tensor X R I 1 × I 2 × I N , the notation
X : , : , , : ( N 1 ) t i m e s , k k = 1 , 2 , , I N
denotes a tensor in R I 1 × I 2 × × I N 1 which is obtained by fixing the last index and is called frontal slice. Fibers are the higher-order analogue of matrix rows and columns. A fiber is defined by fixing every index but one. A matrix column is a mode-1 fiber and a matrix row is a mode-2 fiber. Figure 1 shows the frontal, horizontal and lateral slices of a third order tensor and also a mode-3 tube fiber.
The n-mode matrix of a tensor X R I 1 × I 2 × I 3 × I N is denoted by X ( n ) and arranges the mode-n fibers to be the columns of the resulting matrix X ( n ) R I n × ( I 1 I 3 I n 1 I n + 1 I N ) with n = 1 , , N . We have:
X ( n ) ( i n , j ) = x i 1 , , i N
where j = 1 + k = 1 k n N ( i k 1 ) J k with J K = m = 1 , m n k 1 I m .
An important operation for a tensor is the tensor-matrix multiplication [20], also known as n-mode product of a tensor X R I 1 × I 2 × × I N with a matrix U R J × I n , the n-mode product is a tensor of size I 1 × I 2 × I n 1 × J × I n + 1 × × I N whose entries are given by:
( X × n U ) i 1 , , i n 1 , j , i n + 1 , , i N = i n = 1 I n x i 1 i N u j i n
The n-mode vector product of a tensor X R I 1 × I 2 × × I N with a vector v R I n is denote by ( X × ˜ n v ) is a tensor of size I 1 × I 2 × I n 1 × I n + 1 × × I N whose entries are given by:
( X × ˜ n v ) i 1 , , i n 1 , i n + 1 , , i N = i n = 1 I n x i 1 i N v i n
We recall the definition and some properties of the Einstein product, see [19].
Definition 1.
Let A R I 1 × × I L × K 1 × × K N , B R K 1 × × K N × J 1 × × J M , the Einstein product of the tensors A and B is the tensor of size ( I 1 × × I L × J 1 × × J M ) defined as:
( A N B ) i 1 i L j 1 j M = k 1 = 1 K 1 k 2 = 1 K 2 k 3 = 1 K 3 k N = 1 K N a i 1 i L k 1 k N b k 1 k N j 1 j M
Definition 2
([19]). Let A = a i 1 , , i N , j 1 , , j M R I 1 × × I N × J 1 × × J M , B = b i 1 , , i M , j 1 , , j N R J 1 × × J M × I 1 × × I N , if b i 1 , , i M , j 1 , , j N = a j 1 , , j N , i 1 , , i M then B is called the transpose of A and denoted by A T .
A tensor D = d i 1 , , i N , j 1 , , j N C I 1 × × I N × I 1 × × I N is diagonal if d i 1 , , i N , j 1 , , j N = 0 in the case that the indices i 1 , , i N are different from j 1 , , j N . If all the entries of tensor are zero, we say this tensor is zero tensor, denoted by O .
The trace of the tensor A C I 1 × × I N × I 1 × × I N is given by t r ( A ) = i 1 i N a i 1 i N i 1 i N .
The inner product of two tensors of the same order A , B R I 1 × × I N × J 1 × × J m is given by
A , B = t r ( B T N A ) ,
and the associated norm is defined by:
| | A | | 2 = t r ( A T N A ) .
Definition 3
([21]). A tensor X R I 1 × × I N × I 1 × × I N is called the inverse of the square tensor A R I 1 × × I N × I 1 × × I N and denoted by A 1 if it satisfies:
A N X = X N A = I
where I the identity tensor is called a unit tensor or identity tensor which all this entries are zero except for the diagonal entries I i 1 , , i N , i 1 , , i N = 1 .
In the following, we introduce the new ( N + M + 1 ) product between two tensors which is a generalization of diamond product introduced in [22].
Definition 4.
The ( N + M + 1 ) tensor product between two ( M + N + 1 ) -mode tensors X = X 1 , , X R I 1 × I 2 × I 3 × I N × J 1 × J 2 × J 3 × J M × and Y = Y 1 , , Y p R I 1 × × I N × J 1 × × J M × p where X i R I 1 × × I N × J 1 × × J M and Y j R I 1 × × I N × J 1 × × J M , is the × p matrix whose ( i , j ) entries, i = 1 , , l and j = 1 , , p , is given by
( X ( N + M + 1 ) Y ) i , j = X i , Y j = t r ( Y j T N X i ) .
Next, we give a proposition that will be used later.
Proposition 1.
Assume that U = [ U 1 , , U m ] R I 1 × I 2 × × I N × J 1 × J 2 × × J M × m is an ( N + M + 1 ) -mode tensor such that U 1 , , U m R I 1 × I 2 × × I N × J 1 × J 2 × × J M and let y = ( y 1 , , y m ) T R m . Then for an arbitrary ( N + M + 1 ) -mode tensor Y = [ Y 1 , , Y m ] such that Y 1 , , Y m R I 1 × I 2 × × I N × J 1 × J 2 × × J M , we have:
Y ( N + M + 1 ) ( U × ˜ ( N + M + 1 ) y ) = ( Y ( N + M + 1 ) U ) y ,
Proof. 
The proof comes easily from the definition of the two involved tensor products. □
Definition 5.
The set of ( N + M )-mode tensors U 1 , , U m R I 1 × × I N × J 1 × × J M is called orthonormal if
U i , U j = δ i , j ( = 1 i f i = j a n d 0 e l s w h e r e ) .
Remark 1.
Suppose that U = [ U 1 , , U m ] R I 1 × I 2 × × I N × J 1 × J 2 × × J M × m is an ( N + M + 1 ) -mode tensor such that U 1 , , U m R I 1 × I 2 × × I N × J 1 × J 2 × × J M . If the ( M + N )-mode tensors U 1 , , U m are orthonormal, then we have:
U ( N + M + 1 ) U = I m .

3. Tensor Extrapolation Methods

In this section, we present two classes of new tensor global extrapolation methods. The first class contains the tensor polynomial based methods while the second class is devoted to the global tensor topological ϵ -algorithm.

3.1. Tensor Global-Polynomial Extrapolation Methods

Let ( S n ) be a sequence of tensors of R I 1 × I 2 × I N × K 1 × K 2 × K M and consider the transformation T k , k = 1 , 2 , from R I 1 × I 2 × I N × K 1 × K 2 × K M onto R I 1 × I 2 × I N × K 1 × K 2 × K M and defined by
T k ( n ) = T k ( S n ) = S n + j = 0 k 1 α j ( k ) G j ( n )
= S n + H k ( n ) × ˜ ( N + M + 1 ) α ( k )
where α ( k ) = ( α 0 ( k ) , , α k 1 ( k ) ) T R k and H k ( n ) = [ G 0 ( n ) , , G k 1 ( n ) ] is the (N + M + 1)-mode tensor defined from the given auxiliary tensor sequences ( G i ( n ) ) n R I 1 × × I N × K 1 × × K M ; i = 0 , , k 1 . We will see later how to choose the vector α ( k ) R k .
Let T ˜ k denote the new transformation obtained from ( T k ) as follows:
T ˜ k ( S n ) = S n + 1 + j = 0 k 1 α j ( k ) G j ( n + 1 )
= S n + 1 + H k ( n + 1 ) × ˜ ( N + M + 1 ) α ( k ) .
Notice that the scalars α j ( k ) are the same in the expressions of T k ( S n ) and T ˜ k ( S n ) . We define now the generalized residual of T k ( n ) by
R ˜ k n = R ˜ ( T k ( n ) ) = T ˜ k ( S n ) T k ( S n ) = ( S n + 1 S n ) + j = 0 k 1 α j ( k ) ( G j ( n + 1 ) G j ( n ) ) = Δ S n + j = 0 k 1 α j ( k ) ( Δ G j ( n ) ) .
Then we get
R ˜ ( T k ( n ) ) = Δ S n + Δ H k ( n ) × ˜ ( N + M + 1 ) α ( k ) ,
where the first forward differences Δ S n = S n + 1 S n and Δ H k ( n ) = [ Δ G 0 ( n ) , , Δ G k 1 ( n ) ] R I 1 × I 2 × I N × K 1 × K 2 × K M × k . The vector α ( k ) is obtained from the orthogonality relation:
R ˜ ( T k ( n ) ) ( s p a n { Y 0 ( n ) , , Y k 1 ( n ) } )
where Y 0 ( n ) , , Y k 1 ( n ) R I 1 × I 2 × × I N × K 1 × K 2 × × K M are given tensors.
Here, the s p a n { Y 0 ( n ) , , Y k 1 ( n ) } is the tensor subspace generated by the tensors Y 0 ( n ) , , Y k 1 ( n ) .
If H ˜ k , n and L ˜ k , n denote the tensor subspaces H ˜ k , n = s p a n { Δ G 0 ( n ) , , Δ G k 1 ( n ) } and L ˜ k , n = s p a n { Y 0 ( n ) , , Y k 1 ( n ) } , then from (6) and (7), the generalized residual satisfies the following relations:
R ˜ ( T k ( n ) ) Δ S n H ˜ k , n
and
R ˜ ( T k ( n ) ) ( L ˜ k , n ) .
Let L k , n = [ Y 0 ( n ) , , Y k 1 ( n ) ] R I 1 × I 2 × I N × K 1 × K 2 × K M × k , then the relations (8) and (9) could be expressed as follows:
R ˜ ( T k ( n ) ) Δ S n = Δ H k ( n ) × ˜ ( N + M + 1 ) α ( k )
and
L k , n ( N + M + 1 ) R ˜ ( T k ( n ) ) = 0 .
Assuming that ( L k , n N + M + 1 Δ H k ( n ) ) is nonsingular, the vector α ( k ) appearing in the expression (6) of the generalized residual R ˜ ( T k ( n ) ) is given by
α ( k ) = ( L k , n ( N + M + 1 ) Δ H k ( n ) ) 1 ( L k , n ( N + M + 1 ) Δ S n ) .
Therefore, the approximation T k ( n ) is computed as follows
T k ( n ) = S n + H k ( n ) × ˜ ( N + M + 1 ) α ( k )
For the tensor global polynomial extrapolation methods, namely Tensor Global MPE (TG-MPE), Tensor Global RRE (TG-RRE) and Tensor Global MMPE (TG-MMPE), the auxiliary sequences are given as
G ( n ) = Δ S ( n + ) , = 0 , , k 1 ; n 0
Let Δ i V k ( n ) be the tensor defined by
Δ i V k ( n ) = [ Δ i S n , , Δ i S n + k 1 ] R I 1 × I 2 × I N × K 1 × K 2 × K M × k , i = 1 , 2 ,
where Δ 2 is the second forward difference Δ 2 S n = Δ S n + 1 Δ S n .
In this case, using the relation (12) and the fact that H k ( n ) = Δ V k ( n ) , and Δ H k ( n ) = Δ 2 V k ( n ) , the approximations T k ( n ) given in (13) can be expressed as:
T k ( n ) = S n Δ V k ( n ) × ˜ ( N + M + 1 ) ( ( L k , n ( N + M + 1 ) Δ 2 V k ( n ) ) 1 ( L k , n ( N + M + 1 ) Δ S n ) ) .
It is clear that T k ( n ) exists and is unique if and only if the square matrix L k , n ( N + M + 1 ) Δ 2 V k ( n ) is nonsingular. The generalized residual given in the relation (6) can be expressed as follows:
R ˜ ( T k ( n ) ) = Δ S n Δ 2 V k ( n ) × ˜ ( N + M + 1 ) ( ( L k , n ( N + M + 1 ) Δ 2 V k ( n ) ) 1 ( L k , n ( N + M + 1 ) Δ S n ) ) .
The choice of the tensors Y 0 ( n ) , , Y k 1 ( n ) R I 1 × I 2 × × I N × K 1 × K 2 × × K M required in the orthogonality relation (7) determines the global-polynomial tensor extrapolation method. For the three well known extrapolation polynomial-based methods, we have the following choices
Y ( n ) = Δ S n + for TG-MPE , Y ( n ) = Δ 2 S n + for TG-RRE , Y + 1 ( n ) = Y n + l for TG-MMPE .
Next, we will propose an efficient implementation of these methods. For this purpose, we first express the approximation T k ( n ) given in relation (2) in a different way.
Using the relation ( 2 ) , the fact that G ( n ) = Δ S ( n + ) , = 0 , , k 1 ; n 0 , and R ˜ ( T k ( n ) ) ( s p a n { Y 0 ( n ) , , Y k 1 ( n ) } ) , the (TG-RRE), (TG-MPE) and (TG-MMPE) extrapolation methods produce approximations T k ( n ) of the form
T k ( n ) = j = 0 k γ j ( k ) S n + j ,
where
j = 0 k γ j ( k ) = 1 and j = 0 k η , j γ j ( k ) = 0 0 < k ,
with η , j = Y ( n ) , Δ S n + j .
The system of linear Equation (17) can be written as
γ 0 ( k ) + γ 1 ( k ) + + γ k ( k ) = 1 γ 0 ( k ) Y 0 ( n ) , Δ S n + γ 1 ( k ) Y 0 ( n ) , Δ S n + 1 + + γ k ( k ) Y 0 ( n ) , Δ S n + k = 0 γ 0 ( k ) Y 1 ( n ) , Δ S n + γ 1 ( k ) Y 1 ( n ) , Δ S n + 1 + + γ k ( k ) Y 1 ( n ) , Δ S n + k = 0 γ 0 ( k ) Y k 1 ( n ) , Δ S n + γ 1 ( k ) Y k 1 ( n ) , Δ S n + 1 + + γ k ( k ) Y k 1 ( n ) , Δ S n + k = 0
Let β l ( k ) = γ 0 ( k ) / γ k ( k ) for 0 l k , then
γ l ( k ) = β l ( k ) l = 0 k β l ( k ) for 0 l < k and β k ( k ) = 1 .
With these notations, the linear system of Equation (18) becomes
β 0 ( k ) Y 0 ( n ) , Δ S n + + β k 1 ( k 1 ) Y 0 ( n ) , Δ S n + k 1 = Y 0 ( n ) , Δ S n + k β 0 ( k ) Y k 1 ( n ) , Δ S n + + β k 1 ( k ) Y k 1 ( n ) , Δ S n + k 1 = Y k 1 ( n ) , Δ S n + k .
The above system of equations can also be expressed in the following compact form
( L k , n ( N + M + 1 ) Δ V k ( n ) ) β ( k ) = ( L k , n ( N + M + 1 ) Δ S n + k )
where β ( k ) = [ β 0 ( k ) , , β k 1 ( k ) ] T . Assume now that γ 0 ( k ) , γ 1 ( k ) , , γ k ( k ) have been calculated and introduce the new variables
δ 0 ( k ) = 1 γ 0 ( k ) , δ j ( k ) = δ j 1 ( k ) γ j ( k ) , 1 j < k and δ k 1 ( k ) = γ k ( k ) ,
then the tensor approximation T k ( n ) can be expressed as
T k ( n ) = S n + j = 0 k 1 δ j ( k ) Δ S n + j = S n + Δ V k ( n ) × ˜ ( N + M + 1 ) δ ( k )
where δ ( k ) = [ δ 0 ( k ) , , δ k 1 ( k ) ] T .

3.2. The tensor Global Topological ϵ -Transformation

In [6], Brezinski proposed a generalization of the scalar ϵ -algorithm for vector sequences called the topological ϵ -algorithm (TEA). The matrix case has been introduced by Jbilou and Sadok in ([9]). In this section we define the tensor global topological ϵ -transformation (TG-TET).
Let ( S n ) be a sequence of tensors of R I 1 × I 2 × I N × K 1 × K 2 × K M and consider approximations E k ( S n ) = E k ( n ) of the limit of the tensor sequence ( S n ) n N such that
E k ( n ) = S n + i = 1 k a i ( n ) Δ S n + i 1 , n 0 .
where a i ( n ) R , for i = 1 , k . We introduce the new tensor transformation E ˜ k , j ( n ) , j = 1 , k defined by
E ˜ k , j ( n ) = S n + j + i = 1 k a i ( n ) Δ S n + i + j 1 j = 1 , , k .
We set E ˜ k , 0 ( n ) = E k ( n ) and define the j-th tensor generalized residual as follows
R ˜ j ( E k ( n ) ) = E ˜ k , j ( n ) E ˜ k , j 1 ( n ) .
The coefficients a i ( n ) involved in the expression (24) of E k ( n ) are computed such that each j-th generalized residual is orthogonal to some chosen tensor Y R I 1 × I 2 × I N × K 1 × K 2 × K M , that is
Y , R ˜ j ( E k ( n ) ) = 0 ; j = 1 , , k .
Let D k , n denote the following matrix
D k , n = Y , Δ 2 S n . . . . Y , Δ 2 S n + k 1 Y , Δ 2 S n + 1 . . . . Y , Δ 2 S n + k . . . . . . . . . . . . . . . . . . Y , Δ 2 S n + k 1 . . . . Y , Δ 2 S n + 2 k 2 .
Then, from the orthogonality relation (27), a ( n ) = ( a 1 ( n ) , , a k ( n ) ) T R k is given by
a ( n ) = D k , n 1 z ( n )
where z ( n ) = ( Y , Δ S n , , Y , Δ S n + k 1 ) T . We assume that the matrix D k , n is nonsingular. Hence, the approximation E k ( n ) exists, is unique and is expressed as
E k ( n ) = S n + Δ W k ( n ) × ˜ ( N + M + 1 ) β ( n )
where W k ( n ) = Δ S n , , Δ S n + k 1 . If the matrix D k , n is singular, then the approximation E k ( n ) is not defined and in that case we will have a breakdown of the method. One possibility to overcome this drawback is, instead of computing a ( n ) by using (29), we can solve the least squares problem
min a ( n ) R k D k , n a ( n ) + z ( n ) .
In the next sections, we will see how these tensor global extrapolation methods could be applied for solving linear and nonlinear tensor equations.

4. Application to Tensor Linear/Non Linear Systems of Equations

4.1. Application to Tensor Linear Systems

Consider the following tensor linear system of equations
A N X = B ,
where A is a tensor in R I 1 × I 2 × × I N × I 1 × I 2 × × I N , B R I 1 × I 2 × × I N × J 1 × J 2 × × J M and X R I 1 × I 2 × × I N × J 1 × J 2 × × J M is the unknown tensor to be determined. We assume here that the square tensor A is nonsingular so that (31) has a unique solution X = A 1 B . The purpose here is to apply tensor extrapolation methods to the problem (31) to compute approximate solutions to the solution X of (31). Starting from an initial tensor guess S 0 , we construct the linear sequence ( S n ) n as follows
S n + 1 = ( I A ) N S n + B
Notice that if the sequence ( S n ) n is convergent, then its limit S = X is the solution of the system (31).
In [18,19], tensor Krylov subspaces via the Einstein product have been introduced including the tensor GMRES and some tensor based Lanczos methods. The next theorem shows that, when applied to the generated linear sequence (32), the proposed tensor extrapolation methods are Tensor Krylov subspace methods and are mathematically equivalent to some well known Krylov based methods such as the tensor GMRES.
Theorem 1.
When applied to the sequence generated by (32), TG-RRE, and TG-MPE are tensor Krylov subspace methods and are mathematically equivalent to the tensor global GMRES and the tensor global Arnoldi methods, respectively.
Proof. 
Notice first that for tensor linear sequences (32) the generalized residual becomes the true residual. In fact, from (32), we have Δ S n = B A N S n = R ( S n ) the residual of the tensor S n . Since  Δ 2 S n = A N Δ S n we have Δ 2 V k ( n ) = [ Δ 2 S n , , Δ 2 S n + k 1 ] = A N Δ V k ( n ) where Δ V k ( n ) = [ Δ S n , , Δ S n + k 1 ] . Consequently using (6) and (32), the generalized residual of the approximation ( T k ( n ) ) is the true residual
R ˜ ( T k ( n ) ) = R ( T k ( n ) ) = B A N T k ( n ) .
For simplicity and unless specified otherwise, we set n = 0 , T k ( 0 ) = T k , S 0 = X 0 the initial guess and drop the index n in all our notations. When applied to the sequence generated by the linear relation (32), the TG-RRE, TG-MMPE and the TG-MPE above produce approximations X k = T k such that the corresponding residual R k = B A N T k satisfies the relations
R k Δ S 0 H ˜ k = A N V ˜ k R k ( L ˜ k )
where V ˜ k = s p a n Δ S 0 , , Δ S k 1 and L ˜ k H ˜ k for TG-RRE, L ˜ k V ˜ k for TG-MPE and L ˜ k Y ˜ k = s p a n Y 0 , , Y k 1 for TG-MMPE where Y 0 , , Y k 1 are some chosen tensors.
Notice that, since V ˜ k = K m ( A , R 0 ) (the tensor Krylov subspace defined in [18,19]), the extrapolation methods above are tensor global Krylov subspace methods. TG-RRE is an orthogonal projection and is theoretically equivalent to the tensor global GMRES while TG-MPE is oblique projection method and is equivalent to the tensor global Arnoldi method. □
As the TG-RRE is an orthogonal projection method, we also have the classical minimization property for the residual
| | R T G R R E | | = min T k X 0 + K m ( A , R 0 ) | | B A N T k | | .
When the linear process (32) is convergent, it is more useful in practice to apply the tensor extrapolation methods after some fixed number of iterations. To save memory, we can also use the algorithm in a cycling mode which means that the iterations are restarted after a fixed number m of iterations. Algorithm 1 is summarized as follows:
Algorithm 1 TG-RRE, TG-MPE and TG-MMPE Algorithms
Step 1. k = 0 choose X 0 and the numbers p and m.
Step 2. Basic iteration
    Set T 0 = X 0
      Z 0 = T 0
      Z j + 1 = ( I A ) N Z j + B , j = 0 , , p 1
Step 3. Extrapolation schema
      S 0 = Z p
      S n + 1 = ( I A ) N S n + B , n = 0 , , m
     Compute the approximation T m ( 0 ) by TG-RRE, TG-MPE or TG-MMPE.
Set X 0 = T m ( 0 ) , k = k + 1 and go to step 2.

4.2. Application to Non Linear Tensor Systems

Consider the nonlinear system of tensor equations
G ( X ) = B
with G ( X ) an operator from R I 1 × I 2 × × I N × J 1 × J 2 × × J M onto R I 1 × I 2 × × I N × J 1 × J 2 × × J M , and X R I 1 × I 2 × × I N × J 1 × J 2 × × J M the solution of the system of Equation (33). For any arbitrary tensor X , the residual is given by
R ( X ) = G ( X ) X
Let ( S n ) n be the sequence of tensors generated from an initial guess S 0 as follows
S n + 1 = G ( S n ) , j = 0 , 1 .
To get approximate solutions to a solution of (33), we apply the tensor extrapolation methods above to the sequence ( S n ) . As for the linear case, the different steps are summarized in the following Algorithm 2.
Algorithm 2 TG-RRE, TG-MPE and TG-MMPE for nonlinear systems
Step 1. k = 0 choose X 0 and the numbers p and m.
Step 2. Basic iteration
    Set T 0 = X 0
      W 0 = T 0
      W j + 1 = G ( W j ) j = 0 , , p 1
Step 3. Extrapolation scheme
      S 0 = W p
     If | | S 1 S 0 | | F < ϵ stop
     otherwise generate S n + 1 = G ( S n ) , n = 0 , , m
     Compute the approximation T m ( 0 ) by TG-RRE, TG-MPE or TG-MMPE.
Set X 0 = T m ( 0 ) , k = k + 1 and go to step 2.
Remark 2.
As we stated earlier, when applied to linearly generated tensor sequences, the proposed tensor extrapolation methods produce the same iterates as some well known tensor Krylov subspace methods but differ in the way that the approximations are computed. In the case where the process of generating the sequence ( S n ) n is not known and what is known is only the terms of this sequences, Krylov-based methods or Newton-type methods could not be used and in that case, extrapolation methods are well come. Such a problem can be found for example in some statistical-problems; see for example [16] for the application of the vector ϵ-algorithm of Wynn [23] in the expectation–maximization (EM) algorithm to find maximum likelihood estimates from incomplete or missing data.

5. The Global-QR Implementation of TG-MPE/TG-RRE

The purpose of this section is to give an efficient implementation of TG-MPE and TG-RRE using a generalisation of the technique based on the QR decomposition and given in [11] for vector MPE and vector RRE methods. We first introduce the Tensor Global-QR decomposition. Let U R I 1 × I 2 × × I M × J 1 × J 2 × × J N × k , be an ( M + N + 1 ) -mode tensor with the column tensors U 0 , , U k 1 R I 1 × I 2 × × I M × J 1 × J 2 × × J N . Then, there is an ( M + N + 1 ) -mode orthogonal tensor Q = [ Q 0 , , Q k 1 ] R I 1 × I 2 × × I M × J 1 × J 2 × × J N × k satisfying Q ( N + M + 1 ) Q = I k × k and an upper triangular matrix R R k × k such that
U = Q × ( M + N + 1 ) R T
The tensor decomposition (35) will be called the Tensor Global-QR (TG-QR) decomposition of A and is summarized in the following Algorithm 3.
Algorithm 3 The global-QR decomposition
  • Given U 0 , and compute the scalar r 00 and the tensor Q 0 by r 0 , 0 = U 0 , U 0 1 2 and Q 0 = U 0 r 00 .
  • For j = 1 , , k
    (a)
    W = U j
    (b)
    For i = 0 , , j 1 do
    • r i , j = Q i , W ; W = W r i , j Q i ,
    • End for
    (c)
    r j , j = W , W 1 2
    (d)
    Q j = W r j , j
    (e)
    End for
  • End
To show that the tensor Q is orthogonal, we just proceed by induction on k. The coefficients of the matrix R are the r i , j ’s given by Algorithm 3. Next, we give a proposition to be used.
Proposition 2
([24]). Let X R I 1 × I 2 × × I N , A R J n × I n and y R J n , then we have
X × ( n ) A × ˜ ( n ) y = X × ˜ ( n ) ( A T y ) .
For simplicity, we set here n = 0 and then the approximation T k ( 0 ) (defined earlier in (23)) is given by
T k ( 0 ) = S 0 + j = 0 k 1 δ j ( k ) Δ S j = S 0 + Δ V k ( 0 ) × ˜ ( N + M + 1 ) δ ( k )
Substituting now V k = V k ( 0 ) = Q × ( N + M + 1 ) R T and using Proposition 2, we get
T k ( 0 ) = S 0 + Δ V k × ˜ ( N + M + 1 ) δ ( k ) = S 0 + Q × ˜ ( M + N + 1 ) ( R δ ( k ) ) ,
which gives
T k ( 0 ) = S 0 + j = 0 k 1 θ j Q j ,
where θ j is the ( j + 1 ) component of the column vector R δ ( k ) ; the matrix R and the tensors Q j R I 1 × I M × J 1 × × J N , j = 0 , , k 1 are given by Algorithm 3. To compute δ ( k ) , we have first to compute β ( k ) = ( β 0 ( k ) , , β k 1 ( k ) ) T and use the relations given in (22).
For TG-MPE, the coefficients γ l ( k ) ’s of the vector γ ( k ) are determined by computing the vector β ( k ) , the solution of the linear system of equations (20)
( Δ V k ( N + M + 1 ) Δ V k ) β ( k ) = ( Δ V k ( N + M + 1 ) Δ S k )
where
( Δ V k ( N + M + 1 ) Δ V k ) = Δ S 0 , Δ S 0 . . . . Δ S 0 , Δ S k 1 Δ S 1 , Δ S 0 . . . . Δ S 1 , Δ S k 1 . . . . . . . . . . . . . . . . . . Δ S k 1 , Δ S 0 . . . . Δ S k 1 , Δ S k 1 R k × k
and ( Δ V k ( N + M + 1 ) Δ S k ) = ( Δ S 0 , Δ S k , , Δ S k 1 , Δ S k ) T R k .
For TG-RRE, β ( k ) are determined by solving the linear system of equation
( Δ 2 V k ( N + M + 1 ) Δ V k ) β ( k ) = ( Δ 2 V k ( N + M + 1 ) Δ S k )
where
( Δ 2 V k ( N + M + 1 ) Δ V k ) = Δ 2 S 0 , Δ S 0 . . . . Δ 2 S 0 , Δ S k 1 Δ 2 S 1 , Δ S 0 . . . . Δ 2 S 1 , Δ S k 1 . . . . . . . . . . . . . . . . . . Δ 2 S k 1 , Δ S 0 . . . . Δ 2 S k 1 , Δ S k 1 R k × k ,
and ( Δ 2 V k ( N + M + 1 ) Δ S k ) = ( Δ 2 S 0 , Δ S k , , Δ 2 S k 1 , Δ S k ) T R k .
Finally, once β ( k ) is computed, the coefficients γ l ( k ) ’s are given by
γ l ( k ) = β l ( k ) l = 0 k β l ( k ) for 0 l < k and β k ( k ) = 1 .
The following Algorithm 4 summarizes the main steps.
Algorithm 4 Implementation of TG-MPE/TG-RRE via the global-QR decomposition
  • S 0 is a given initial guess and k is a fixed index.
  • Apply Algorithm 3 to compute the global-QR decomposition of the tensor Δ V k .
  • Compute β ( k ) by solving the linear system (38) or (39).
  • Compute the coefficients γ j ( k ) from (40).
  • Compute δ ( k ) from the relation (22) and θ j is ( j + 1 ) component of the column vector R δ ( k ) .
  • Compute T k ( 0 ) by T k ( 0 ) = S 0 + j = 0 k 1 θ j Q j
As a numerical test, we consider the linear system of tensor equations given by (31) where A is a random tensor N = 3 , I 1 = I 2 = 20 , I 3 = 10 , J 1 = 20 , J 2 = 10 and J 3 = 5 as in [19]. The exact solution is the tensor X whose elements are all one and the right hand side B is given by B = A N X . The computations are carried out using MATLAB 7.4 with machine epsilon about 2 · 10 16 .
We took m = 10 , p = 1 and stopped the iteration when the relative error norm of the residual was less than 10 6 . Then after k = 6 iterations, Algorithm 4 for the tensor RRE method, gives the relative residual norm R k R 0 = 8.5 × 10 7 where the initial guess is X 0 = O . We notice that the convergence of the original sequence ( S n ) is very slow and at n = 200 , we get a relative residual R ( S n ) / R 0 of 10 1 . Generally, extrapolation methods are more effective than Krylov-based methods for nonlinear problems. More experimental studies and applications in some areas such as mutilinear google pagerank should be considered in the future.

6. Conclusions

In this paper, we introduced new tensor global extrapolation methods to accelerate the convergence of some tensor sequences. The proposed methods are generalisations to the tensor case of some well known vector extrapolation methods such as the reduced rank extrapolation or the topological epsilon algorithm. The new methods were defined as orthogonal or oblique projection processes using the Einstein product and also some new interesting tensor products. We showed how to apply the derived algorithms to tensor linear and nonlinear systems of tensor equations. Application to some interesting problems such as the multilinear page rank is still under investigation.

Author Contributions

Formal analysis, A.E.I.; Methodology, R.S.; Writing—original draft, K.J. All authors have read and agreed to the published version of the manuscript.

Funding

For the first author, this project was financially supported by Ministry of Europe and Foreign Affairs, Ministry of Higher Education, Research and Innovation and the French Institute of Rabat (PHC TOUBKAL 20XX (French-Morocco bilateral program) Grant Number: 12345AB).

Acknowledgments

We would like to thank the two referees for valuable remarks and helpful comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Brezinski, C.; Redivo Zaglia, M. Extrapolation Methods. Theory and Practice; North Holland Publishing: Amsterdam, The Netherlands, 1991. [Google Scholar]
  2. Aitken, A.C. On Bernoulli’s numerical solution of algebraic equations. Proc. R. Soc. Edinb. 1926, 46, 289–305. [Google Scholar] [CrossRef]
  3. Cabay, S.; Jackson, L.W. A polynomial extrapolation method for finding limits and antilimits for vector sequences. SIAM J. Numer. Anal. 1976, 13, 734–752. [Google Scholar] [CrossRef]
  4. Mesina, M. Convergence acceleration for the iterative solution of = Ax + f. Comput. Meth. Appl. Mech. Eng. 1977, 10, 165–173. [Google Scholar] [CrossRef]
  5. Eddy, V.P. Extrapolation to the limit of a vecteur sequence. In Information Linkage between Applied Mathematics and Industry; Wang, P.C.C., Ed.; Academic Press: New York, NY, USA, 1979; pp. 387–396. [Google Scholar]
  6. Brezinski, C. Généralisation de la transformation de Shanks, de la table de Padéé et l’epsilon algorithm. Calcolo 1975, 12, 317–360. [Google Scholar] [CrossRef]
  7. Pugatchev, B.P. Acceleration of the convergence of iterative processes and a method for solving systems of nonlinear equations. USSR Comput. Math. Math. Phys. 1978, 17, 199–207. [Google Scholar] [CrossRef]
  8. Jbilou, K.; Sadok, H. Analysis of some vector extrapolation methods for linear systems. Numer. Math. 1995, 70, 73–89. [Google Scholar] [CrossRef]
  9. Jbilou, K.; Sadok, H. VMatrix polynomial and epsilon-type extrapolation methods with applications. Numer. Algorithms 2015, 68, 107–119. [Google Scholar] [CrossRef]
  10. Jbilou, K.; Messaoudi, A.; Tabaa, K. Some Schur complement identities to matrix extrapolation methods. Linear Algebra Appl. 2004, 392, 195–210. [Google Scholar] [CrossRef] [Green Version]
  11. Sidi, A. Efficient implementation of minimal polynomial and reduced rank extrapolation methods. J. Comput. Appl. Math. 1991, 17, 305–337. [Google Scholar] [CrossRef] [Green Version]
  12. Sidi, A.; Ford, W.F.; Smith, D.A. Acceleration of convergence of vector sequences. SIAM J. Numer. Anal. 1986, 23, 178–196. [Google Scholar] [CrossRef] [Green Version]
  13. Jbilou, K.; Sadok, H. Vector extrapolation methods. Application and numerical comparison. J. Comput. Appl. Math. 2000, 122, 149–165. [Google Scholar] [CrossRef] [Green Version]
  14. Brezinski, C.; Redivo Zaglia, M.; Serra-Capizzano, S. Extrapolation methods for PageRank computation. Comptes Rendus Math. 2005, 340, 393–397. [Google Scholar] [CrossRef]
  15. Kamvar, D.; Haveliwala, T.H.; Manning, C.D.; Golub, G.H. Extrapolations methods for accelerating PageRank computations. In Proceedings of the WWW 2003—Twelfth International World Wide Web Conference, Budapest, Hungary, 20–24 May 2003. [Google Scholar]
  16. Kuroda, M.; Sakakihara, M. Accelerating the convergence of the EM algorithm using the vector ϵ-algorithm. Comput. Stat. Data Anal. 2006, 51, 1549–1561. [Google Scholar] [CrossRef]
  17. Duminil, S.; Sadok, H.; Silvester, D. Fast solvers for discretized Navier–Stokes problems using vector extrapolation. Numer. Algorithms 2014, 66, 89–104. [Google Scholar] [CrossRef] [Green Version]
  18. Elguide, M.; El Ichi, A.; Jbilou, K.; Beik, F.P.A. Tensor GMRES and Golub-Kahan Bidiagonalization methods via the Einstein product with applications to image and video processing. arXiv 2020, arXiv:2005.07458. [Google Scholar]
  19. Huang, B.; Xie, Y.; Ma, C. Krylov subspace methods to solve a class of tensor equations via the Einstein product. Numer. Linear Algebra Appl. 2019, 26, e2254. [Google Scholar] [CrossRef]
  20. Kolda, T.G.; Bader, B.W. Tensor Decompositions and Applications. SIAM Rev. 2009, 3, 455–500. [Google Scholar] [CrossRef]
  21. Liang, M.; Zheng, B. Further results on Moore–Penrose inverses of tensors with application to tensor nearness problems. Comput. Math. Appl. 2019, 77, 1282–1293. [Google Scholar] [CrossRef]
  22. Bouyouli, R.; Jbilou, K.; Sadaka, R.; Sadok, H. Convergence properties of some block Krylov subspace methods for multiple linear systems. J. Comput. Appl. Math. 2006, 196, 498–511. [Google Scholar] [CrossRef] [Green Version]
  23. Wynn, P. Acceleration techniqes for iterated vector and matrix problems. Math. Comput. 1962, 16, 301–322. [Google Scholar] [CrossRef]
  24. Beik, F.P.A.; Movahed, F.S.; Ahmadi-Asl, S. On the Krylov subspace methods based on tensor format for positive definite Sylvester tensor equations. Numer. Linear Algebra Appl. 2016, 16, 444–466. [Google Scholar] [CrossRef]
Figure 1. (a) Frontal, (b) horizontal, and (c) lateral slices of a third order tensor. (d) A mode-3 tube fibers.
Figure 1. (a) Frontal, (b) horizontal, and (c) lateral slices of a third order tensor. (d) A mode-3 tube fibers.
Mathematics 08 01298 g001

Share and Cite

MDPI and ACS Style

El Ichi, A.; Jbilou, K.; Sadaka, R. Tensor Global Extrapolation Methods Using the n-Mode and the Einstein Products. Mathematics 2020, 8, 1298. https://doi.org/10.3390/math8081298

AMA Style

El Ichi A, Jbilou K, Sadaka R. Tensor Global Extrapolation Methods Using the n-Mode and the Einstein Products. Mathematics. 2020; 8(8):1298. https://doi.org/10.3390/math8081298

Chicago/Turabian Style

El Ichi, Alaa, Khalide Jbilou, and Rachid Sadaka. 2020. "Tensor Global Extrapolation Methods Using the n-Mode and the Einstein Products" Mathematics 8, no. 8: 1298. https://doi.org/10.3390/math8081298

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop