Next Article in Journal
Mean Square Finite-Approximate Controllability of Semilinear Stochastic Differential Equations with Non-Lipschitz Coefficients
Next Article in Special Issue
Adaptive Fuzzy Predictive Approach in Control
Previous Article in Journal
Algorithms for Approximating Solutions of Split Variational Inclusion and Fixed-Point Problems
Previous Article in Special Issue
Challenging the Curse of Dimensionality in Multidimensional Numerical Integration by Using a Low-Rank Tensor-Train Format
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Aggregation-Based Algebraic Multigrid Method with Deflation Techniques and Modified Generic Factored Approximate Sparse Inverses

by
Anastasia A. Natsiou
1,
George A. Gravvanis
2,
Christos K. Filelis-Papadopoulos
3 and
Konstantinos M. Giannoutakis
4,*
1
School of Computing, Technological University of Dublin, D07 H6K8 Dublin, Ireland
2
Department of Electrical and Computer Engineering, School of Engineering, Democritus University of Thrace, Kimmeria, 67100 Xanthi, Greece
3
Department of Computer Science, University College Cork, T12 XF62 Cork, Ireland
4
Department of Applied Informatics, School of Information Sciences, University of Macedonia, 54636 Thessaloniki, Greece
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(3), 640; https://doi.org/10.3390/math11030640
Submission received: 29 November 2022 / Revised: 16 January 2023 / Accepted: 24 January 2023 / Published: 27 January 2023
(This article belongs to the Special Issue Advanced Numerical Analysis and Scientific Computing)

Abstract

:
In this paper, we examine deflation-based algebraic multigrid methods for solving large systems of linear equations. Aggregation of the unknown terms is applied for coarsening, while deflation techniques are proposed for improving the rate of convergence. More specifically, the V-cycle strategy is adopted, in which, at each iteration, the solution is computed by initially decomposing it utilizing two complementary subspaces. The approximate solution is formed by combining the solution obtained using multigrids and deflation. In order to improve performance and convergence behavior, the proposed scheme was coupled with the Modified Generic Factored Approximate Sparse Inverse preconditioner. Furthermore, a parallel version of the multigrid scheme is proposed for multicore parallel systems, improving the performance of the techniques. Finally, characteristic model problems are solved to demonstrate the applicability of the proposed schemes, while numerical results are given.

1. Introduction

Lately, research efforts for efficiently solving large sparse linear systems have focused on multigrid methods [1,2]. Their applicability and efficiency are based on the use of a stationary method as a smoother for higher frequency components of the error, while the lower frequency components are transferred to a coarser level with higher frequency, which can reduce the error [3,4]. Multigrid algorithms consider a recursive application with two-grid coarse grid-correction schemes [5]. In the coarse level, a smoother is used for defecting correction, which is interpolated to the finer level and then added to the solution [3,4].
An algebraic multigrid (AMG) requires the information of the coefficient matrix of the linear system [4]. The only difference between the geometrical and algebraic approaches is the coarsening algorithm; however, they both consist of a smoother and a coarse grid correction. The smoother, is usually the application of some iterations of a fixed-point iterative method (such as Jacobi and Gauss–Seidel) and coarse grid correction is the transformation of the approximate residual to the finest grid. It should be noted that the notion of aggregation has multiple applications in several scientific fields, including gene regulatory networks [6,7].
Moreover, approximate inverses have been extensively used for approximating the inverse of the coefficient matrix of a linear system, resulting in efficient preconditioners for Krylov subspace iterative methods. Their efficiency relies on the fact that they have inherent parallelism and can be computed quickly in parallel [8,9,10]. Additionally, approximate inverse techniques have been efficiently used in conjunction with multigrid methods for solving partial differential equations discretized with finite difference methods [11], and for a variety of other applications [12,13].
In this paper, a class of Modified Generic Factored Approximate Sparse Inverses (MGenFAspI) based on approximate inverse sparsity patterns is used to solve two-dimensional and three-dimensional partial differential equations, discretized with high-order finite difference schemes. For the computation of these approximate inverses, a fill-in parameter is required which affects the density, in terms of nonzero elements, of the preconditioner [10]. A deflation technique is also utilized in this work, splitting the problem at each multigrid level into two complementary subspaces, aiming to further improve the convergence rate.
The paper is organized as follows: Section 2 introduces the deflation technique and the corresponding algorithm used for the solution of linear systems. Section 3 presents the aggregation-based algebraic multigrid method in conjunction with the deflation techniques and the Modified Generic Factored Approximate Sparse Inverses. The algorithmic procedures for the pairwise aggregation and the preconditioned multigrid are also given. The parallel aspects and techniques for accelerating the computations required by the Modified Generic Factored Approximate Sparse Inverses in hybrid parallel environments are discussed in Section 4. Finally, numerical results regarding the efficiency and convergence behavior of the proposed techniques are presented and discussed in Section 5, by solving characteristic model problems. Concluding remarks and directions for future research are provided in Section 6.

2. Deflation Techniques

Deflation techniques were introduced by Nicolaides in 1987 for improving the rate of convergence of the conjugate gradient method [14]. The main idea is based on the projection of the linear system into two complementary subspaces, in order to separately solve the two produced linear systems. The efficiency of this technique is based on the fact that the projected linear systems could be solved faster than the initial system, and then their solution can be combined for formulating the solution of the initial problem [15].
Let us consider an n × n linear system of the form:
A x = b ,
where A is a symmetric and positive definite matrix, b is the right-hand side vector and x is the solution vector. The deflation subspace V R can be represented by a full-rank matrix V R n × k with k n such that:
E = V T A V .
Then, the projection matrix Q can be defined as follows:
Q = I V E 1 V T A ,
where I is the identity matrix. The matrix Q is the orthogonal projection on the subspace V , while I Q is the orthogonal projection on the subspace V. For the selection of the subspace V, we can use the deflation technique for the generation of the basis of V, proposed in [15], where the domain Ω is decomposed into m nonoverlapping subdomains Ω j , j = 1 , , m , where:
v i , j = 1 , if x i Ω j 0 , if x i Ω j .
The solution of the linear system can be computed as follows
x = ( I Q ) x c + Q x s ,
where x c is the solution vector of the projection to subspace V, while x s is the solution vector from the projection to the subspace V . These vectors can be computed as follows:
x c = ( I Q ) x and x s = Q x i t ,
where x is an initial approximation for the solution. By applying matrix operations, the solution vector x c can be defined as:
x c = ( I Q ) x = V E 1 V T A x = V ( V T A V ) 1 V T A x = V ( V T A V ) 1 V T b .
Then, the vector x i t can be computed by solving iteratively the following linear system:
A ( I V ( V T A V ) 1 V T A ) x i t = b A x c .
The procedure for the solution of the linear system with the deflation technique can be summarized by Algorithm 1.
Algorithm 1 Deflation technique for solving linear systems
1:
Select the basis of the subspace V.
2:
Project vector V based on the subspace V.
3:
Compute E 1 = ( V T A V ) 1 .
4:
Compute the solution from the subspace V, using the transformation matrix I Q , such that x c = V ( V T A V ) 1 V T b .
5:
Solve iteratively the linear system A ( V ( V T A V ) 1 V T ) x i t = b A x c .
6:
Multiply the solution x i t with Q and define the solution x s .
7:
Compute the approximate solution by adding the two solution vectors from the two complementary subspaces.
The use of the deflation technique in conjunction with an efficient preconditioning scheme can accelerate the convergence behavior of the solution of the system [14]. Preconditioning techniques yield the solution of a linear system of the form:
M A x = M b ,
where M is a symmetric positive definite preconditioner. An efficient preconditioner should be efficiently computed in parallel, M A should have a “clustered” spectrum, and the M × vector should be fast to compute in parallel [9,10,16]. Then, the preconditioner can be used in conjunction with a Krylov-subspace iterative method, such as the conjugate gradient or the generalized minimum residual methods for the solution of the linear system [9,10].

3. Aggregation Multigrid Method

Multigrid methods have been widely used by the scientific community for solving linear systems, due to their convergence behavior and efficiency [8,17]. The design of these methods is based on the fact that high-frequency error components can be damped by a stationary iterative method, while low-frequency error components are handled by coarser grids with a higher discretization step, where the high-frequency modes of the error are more oscillatory and can be damped efficiently by a stationary iterative method [4,18]. Multigrid methods are composed of four components: (a) the stationary iterative method (iterative methods such as the Richardson, Jacobi and Gauss–Seidel methods), (b) the restriction operator (transfer operators from finer to coarser grids), (c) the prolongation operator (transfer operators from coarser to finer grids) and (d) the cycle strategy (the sequence in which the grids are visited until a solution with the predefined tolerance is achieved) [16].
The method utilized for grid coarsening in this paper is the aggregation multigrid, first introduced in [19] and developed by Notay [5]. The nodes of the grid are combined into pairs, creating aggregates, based on a weighting rule.
The coarsening algorithm constructs a prolongation matrix P from the coefficient matrix A, which is the transformation matrix from the coarse to the fine grid. The dimensions of the prolongation matrix are n × n c , where n c < n is the number of the nodes of the coarser grid. The transformation from the fine to the coarser grid can be managed by the restriction matrix R, which is computed as follows:
R = P T ,
and the coarse coefficient matrix A c is estimated by Galerkin’s formula:
A c = R A P .
For the formulation of the prolongation matrix, the aggregates G i , which are disjoint subsets of the variable set, have to be defined. The number of these variables, n c , is, then, the number of the subsets and is used to define the prolongation matrix P:
P i , j = 1 , if i G j 0 , otherwise ( 1 i n , 1 j n c ) .
Therefore, P is a Boolean matrix with one nonzero entry in each row. In order to define the aggregates, the algorithm searches for the set of nodes S i to which i is strongly negatively coupled, using a strong/weak coupling threshold β :
S i = j i | a i , j < β max a i , k < 0 | a i , k | .
Then, one unmarked node at a time is selected, giving priority to the node with the minimal m i , where m i is the number of unmarked nodes which are strongly negatively coupled to I [5,20]. Next, one can pick an unmarked node j, which consists of the column of the matrix A with the minimal element a i , j in line i. Then, an aggregate is created between the node i and the node j. The  algorithm of the pairwise aggregation is given in Algorithm 2 [5]. It should be noted that the C h e c k D D parameter is optional and confirms if matrix A is diagonal-dominant [5].
For the solution of a linear system with an algebraic aggregation multigrid, the V-cycle strategy can be used. The linear system is smoothed and the residual is transferred to the coarser level until the resulting linear system is small enough to be solved directly. Then, the solution vector is subject to the procedure of coarse grid correction, by means of the appropriate prolongation [5,20].
During each iteration of the V-cycle, the deflation technique is used to split the problem into two complementary subspaces, and the approximate solution is computed, which is transferred to the coarser level of the grid. The combination of the multigrid and deflation techniques is depicted in Figure 1.
Algorithm 2 Pairwise aggregation
Inputs: matrix A = ( a i , j ) with n rows,
          Threshold β = 0.25 ,
          Logical parameter C h e c k D D
Outputs: number of coarse variables n c ,
             Aggregates G i , i = 1 , n c (such as G i G j = )
1:
if C h e c k D D then
2:
     U = [ 1 , n ] { i | a i , i > 5 j i | a i , j | }
3:
else U = [ 1 , n ]
4:
end if
5:
for i = 1 , , n do
6:
     S i = j U { i } | a i , j < β max a i , k < 0 | a i , k |
7:
end for
8:
for i = 1 , , n do
9:
     m = | { j | i S j } |
10:
end for
11:
n c = 0
12:
while U do
13:
    Select i U with minimum m i , n c = n c + 1
14:
    Select j U such that a i , j = min k U a i , k
15:
    if  j U  then
16:
         G n c = { i , j }
17:
    else  G n c = { i }
18:
    end if
19:
     U = U G n c
20:
    For all k G n c update m l = m l 1 for l S k
21:
end while
Algorithm 3 depicts the procedure of using a multigrid as preconditioner for a V-cycle.
Algorithm 3 Preconditioned V cycle
Inputs: Residual vector r k
         Grid matrix at level k: A k
         Deflation matrix M k
         Prolongation matrix P k
         Matrix of next grid A k 1
Outputs: z k
1:
function M G p r e c ( r k , k)
2:
    Compute z k ( 1 ) = M k 1 r k
3:
    Calculate the new residual vector r k = r k A k z k ( 1 )
4:
    Compute r k 1 = P k T r k
5:
    if k=1 then
6:
         x k 1 = A k 1 1 r k 1
7:
    else
8:
         x k 1 = M G p r e c ( r k 1 , k 1 )
9:
    end if
10:
     z k ( 2 ) = P k x k 1
11:
     r k = r k A k z k ( 2 )
12:
     z k ( 3 ) = M k 1 r k
13:
     z k = z k ( 1 ) + z k ( 2 ) + z k ( 3 )
14:
end function
The function M G p r e c is executed once at the higher level for k = l using the matrix A l = A , and iteratively calls itself until the parameter k reduces to 1. For smoothing, the symmetric Gauss–Seidel method is used, where
M k = l o w ( A k ) d i a g ( A k ) 1 u p p ( A k ) ,
where l o w , d i a g and u p p are the lower, diagonal and upper submatrices of A k [5].
In this paper, the coarse coefficient matrix A c is inverted by the Modified Generic Factored Approximate Sparse Inverse (MGenFAspI) technique [10]. For a coefficient matrix A, we consider its incomplete L U factorization, i.e.,
A = L U + E ,
where L and U are, respectively, the upper and lower triangular factors and E is the error matrix. The GenFAspI matrix can be computed using the following procedure:
A = L U A 1 = U 1 L 1 A 1 = M = G H ,
considering G = U 1 and H = L 1 . In order to compute the G and H, an “a-priori” sparsity pattern is required, [21]. These patterns occur when utilizing a predetermined drop tolerance (droptol) and then raising to a predefined power (level of fill or lfill) the upper and lower triangular factors. After that, the GenFAspI matrix is constructed by solving the following system:
M = G droptol lfill H droptol lfill U g : , j = e : , j L h : , j = e : , j , 0 j < m ,
where g : , j denote the elements of G, h : , j , the elements of H; e : , j are the elements of the identity matrix and m the order of the coefficient matrix of the linear system [22]. A modified version of GenFAspI (MGenFAspI) is used in order to enhance the performance during the computation of the approximate inverse [10]. The MGenFAspI matrix is used to compute each column of the factor of the approximate inverse, separately, by a restricted solution process of only the factors. The elements that do not belong to G and H factors are set to zero; thus, Equation (17) is rewritten as:
U g : , j = e : , j L h : , j = e : , j 0 j < m g i , j = 0 , ( i , j ) G droptol lfill h i , j = 0 , ( i , j ) H droptol lfill .
The advantage of the modified version of GenFAspI is that the construction of the approximate inverse can be easily parallelized, since the elements of the matrices G and H or L and U are not required to be identified. The algorithmic procedure of MGenFAspI is given in [10].

4. Parallel Techniques

In this section, the parallel algorithm for the Modified Generic Factored Approximate Sparse Inverse and the parallel V-cycle are discussed. The matrices G and H can be computed by solving the following equations:
U G = I L H = I ,
where U and L are the upper and lower triangular matrices that are derived from the incomplete LU factorization of the coefficient matrix A. Thus, for calculating the columns of the matrices G and H, the linear system of Equation (18) should be solved.
The MGenFAspI algorithm allows the individual computation of the columns of matrices G and H, and, due to this inherent parallelism, it can be efficiently implemented on hybrid parallel systems that consist of distributed and multicore CPUs. Each node of the distributed system can be assigned into a block of columns of the matrices, while, within each node, the multiple cores can further distribute the workload.
The computation of the elements of each column of matrices G and H does not require all rows of matrices U or L, respectively, but only the rows that correspond to nonzero elements of each column. For this reason, there is no need to broadcast the whole matrices U and L, yielding a reduced communication cost, especially for large-scale matrices. A visual representation of the distribution of matrix G on the available nodes is given in Figure 2, while the workload of two cores of a distributed node is depicted in Figure 3.
The same process was implemented for the computation of the elements of matrix H, where the rows of matrix L were used. Since this technique does not require any communication between the distributed processes, the parallel speedup and parallel efficiency are expected to be high. The derived matrices (G and H) are distributed on the nodes; however, they can be used on any distributed system for executing a column-wise matrix×vector multiplication. The algorithm for the parallelization of the MGenFAspI has been presented in [10]. In the context of this work, the parallelization on multicore systems was considered.

5. Numerical Results

The applicability, performance and convergence behavior of the proposed schemes are given in this section. Various model problems are considered and numerical results are obtained, experimentally demonstrating the behavior of the proposed methods. It should be noted that the experimental results were obtained on an Intel Xeon 2420v2 with 6 cores and 12 threads, 48GB of RAM memory, running Linux CentOS 7. All algorithmic techniques were implemented in C++ 11.
For the evaluation of the aggregation-based multigrid method with V-cycle, the 2D Poisson problem was initially selected with the number of elements per direction n x = 50 , 100 , 200 , 500 . Four iterations of the Jacobi method were performed as a pre-smoother and four iterations of the Jacobi method were used for post-smoothing in each level of grid hierarchy. In the last level of the multigrid, the linear system is solved by a direct method, such as LU factorization. The termination criterion is | | r i | | 2 < t o l | | r 0 | | 2 , where | | r i | | 2 is the norm of the residual vector of the i-th iteration, | | r 0 | | 2 is the norm of the residual of the initial problem and t o l = 10 8 . In the parallel experiments, the resolution at the lowest level was held by the PARDISO method (parallel direct solver) contained in the parallel libraries of Intel [23].
It should be noted that the choice of pre-smoothing and post-smoothing iterations affects the convergence behavior, e.g., increased values lead to better overall performance. However, since the stationary iterative methods in the context of multigrids are used to damp the high-frequency components of the error, an arbitrary increase does not guarantee substantial improvements. In the context of an algebraic multigrid, the choices for pre-smoothing and post-smoothing iterations vary between 1 and 4 for smoothers, such as Jacobi or Gauss–Seidel, in the literature. Thus, determination of the optimal parameters in the context of a multigrid depends on the problem to be solved and can be determined by experimentation [5,18].
The convergence behavior and the performance of the AGAMG method for different numbers of the unknowns per dimension for the 2D Poisson problem are given in Table 1.
The number of iterations required for convergence slightly increases as the size of matrix A increases, for the different number of levels. Moreover, increasing the number of levels while increasing the number of points per dimension ( n x ) yields reduced execution time. This is due to the size of the matrix at the last level, which is solved by a direct method. Thus, for large-scale systems, an increased number of levels is recommended.
In order to examine the applicability and efficiency of the deflation techniques, the same model problem was executed by considering deflation with V-cycle. For the integration of the two methods, the four Jacobi pre-smoothing iterations were replaced with one deflation iteration. The deflation algorithm computes, using a direct method, the solution in subspace I Q :
x c = ( I Q ) x = P E 1 R A x = P ( R A P ) 1 R A x ,
by inverting the matrix E with the Jacobi method (i.e., inverting only the main diagonal). The solution in subspace Q is computed with six iterations of the BiCGStab method for the following linear system:
A ( P E 1 R ) x i t = b A x c .
The efficiency and convergence behavior of the aggregation-based algebraic multigrid method with deflation are given in Table 2, for various numbers of levels and n x . It should be noted that the Jacobi method with n 2 = 4 was used as post-smoother.
To further improve the convergence behavior, the Modified Generic Factored Approximate Sparse Inverse (MGenFAspI) was used for approximating matrix E 1 . The l f i l l parameter was set to 1 and the d r o p t o l to zero. The corresponding results are given in Table 3, for various values of n x and numbers of levels for the 2D Poisson problem. It can be seen that in most of the cases, the number of iterations is reduced.
In Table 4, the convergence behavior and the performance of the proposed scheme with the use of MGenFAspI preconditioner for various values of n x and number of levels for solving the 3D Poisson problem are given. Moreover, in Table 5 and Table 6, numerical results are given for the 2D and 3D Poisson problem with l f i l l = 2 and d r o p t o l = 0.0 for various numbers of levels.
The convergence behavior and the performance of the proposed schemes with the use of MGenFAspI for various numbers of the l f i l l parameter and number of levels for the 2D Poisson problem with n x = 300 is given in Table 7.
The performance and applicability were also tested against the Sherman problem (https://sparse.tamu.edu/HB/sherman1 (accessed on 28 November 2022)). In Table 8, the convergence behavior and performance of the proposed schemes with the use of the MGenFAspI preconditioner, for various values of the parameter l f i l l and number of levels for the Sherman1 problem, are given in Table 8. In Table 9, the parallel aggregation-based algebraic multigrid method in conjunction with deflation and Modified Generic Factored Approximate Sparse Inverses for the 3D Poisson problem, with n x = 70 and l f i l l = 2 , is given.
The addition of deflation is expected to increase the computational time, since it involves matrix inversion operations (approximate) and matrix-by-matrix products. However, for difficult problems, e.g., Sherman1, where iterations increase substantially when following the original AGMG approach, deflation leads to substantial improvements and even ensures convergence within the prescribed maximum iterations. Moreover, the additional computational work is composed of inherently parallel operations which can be accelerated on modern hardware, leading to an efficient solution approach.
Direct parallelization of algebraic multigrid algorithms has been studied extensively in the literature and has been shown to be limited [24]. The advantages of a truncated approach, which is similar to the one adopted in this manuscript, for the parallelization of the V cycle have been discussed extensively in [25].
Finally, the speedups and efficiency of the parallel aggregation-based algebraic multigrid method in conjunction with deflation and Modified Generic Factored Approximate Sparse Inverses for the 3D Poisson problem, with n x = 70 and l f i l l = 2 , are given in Table 10 and Table 11, respectively. The corresponding figures for the execution times, speedups and efficiency are given in Figure 4, Figure 5 and Figure 6, respectively.

6. Discussion and Conclusions

The Modified Generic Factored Approximate Sparse Inverse technique based on incomplete factorization has been recently proposed and used as a preconditioner for Krylov subspace methods. In this work, algebraic multigrid methods were examined that are based on aggregation, while a V cycle was used in conjunction with deflation techniques to improve the convergence behavior. The proposed schemes were discussed and numerical results from solving two model problems were presented.
The evaluation of the proposed methods demonstrates the improvement in the rate of convergence of the multigrid techniques using deflation, for various values of the parameters n x and l f i l l . The extensive experimentation that was carried out validated that the V cycle of AGAMG in combination with deflation and MGenFAspI was the most efficient approach in terms of convergence behavior; however, with greater execution time. For this, the paper took advantage of the inherent parallelism of the approximate inverse scheme and the aggregation-based multigrid method, and resulted in reduced execution times on a shared memory parallel system. The parallel numerical results obtained demonstrated good parallel results with good potential for large-scale parallel systems.
Future work includes the use of the aggregation-based multigrid method in conjunction with deflation techniques and approximate inverses as preconditioners for Krylov-subspace iterative methods. Moreover, the use of domain decomposition methods will be examined, in order to further improve the parallel efficiency of the methods.

Author Contributions

Conceptualization, G.A.G. and C.K.F.-P.; Investigation, A.A.N., G.A.G., C.K.F.-P. and K.M.G.; Methodology, A.A.N., G.A.G., C.K.F.-P. and K.M.G.; Software, A.A.N. and C.K.F.-P.; Supervision, G.A.G.; Validation, A.A.N., C.K.F.-P. and K.M.G.; Visualization, C.K.F.-P. and K.M.G.; Writing—original draft, A.A.N., G.A.G., C.K.F.-P. and K.M.G.; Writing—review and editing, A.A.N., G.A.G., C.K.F.-P. and K.M.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AMGAlgebraic multigrid
AGAMGAggregation-based algebraic multigrid
CPUCentral processing unit
GenFAspIGeneric Factored Approximate Sparse Inverse
MGenFAspIModified Generic Factored Approximate Sparse Inverse

References

  1. Paludetto Magri, V.A.; Franceschini, A.; Janna, C. A Novel Algebraic Multigrid Approach Based on Adaptive Smoothing and Prolongation for Ill-Conditioned Systems. SIAM J. Sci. Comput. 2019, 41, A190–A219. [Google Scholar] [CrossRef]
  2. Demidov, D. AMGCL: An efficient, flexible, and extensible algebraic multigrid implementation. Lobachevskii J. Math. 2019, 40, 535–546. [Google Scholar] [CrossRef] [Green Version]
  3. Briggs, W.L.; Henson, V.E.; McCormick, S.F. A Multigrid Tutorial; SIAM: Philadelphia, PA, USA, 2000. [Google Scholar]
  4. Wesseling, P. Theoretical and Practical Aspects of a Multigrid Method. SIAM J. Sci. Stat. Comput. 1982, 3, 387–407. [Google Scholar] [CrossRef] [Green Version]
  5. Notay, Y. An aggregation-based algebraic multigrid method. Electron. Trans. Numer. Anal. 2008, 37, 123–146. [Google Scholar]
  6. Li, H.; Liu, Y.; Wang, S.; Niu, B. State Feedback Stabilization of Large-Scale Logical Control Networks via Network Aggregation. IEEE Trans. Autom. Control. 2021, 66, 6033–6040. [Google Scholar] [CrossRef]
  7. Wang, S.; Li, H. Aggregation method to reachability and optimal control of large-size Boolean control networks. Sci. China Inf. Sci. 2022, 66, 1869–1919. [Google Scholar] [CrossRef]
  8. Bröker, O.; Grote, M.J.; Mayer, C.; Reusken, A. Robust Parallel Smoothing for Multigrid Via Sparse Approximate Inverses. SIAM J. Sci. Comput. 2001, 23, 1396–1417. [Google Scholar] [CrossRef] [Green Version]
  9. Giannoutakis, K.M.; Gravvanis, G.A. High performance finite element approximate inverse preconditioning. Appl. Math. Comput. 2008, 201, 293–304. [Google Scholar] [CrossRef]
  10. Filelis-Papadopoulos, C.; Gravvanis, G. A class of generic factored and multi-level recursive approximate inverse techniques for solving general sparse systems. Eng. Comput. 2016, 33, 74–99. [Google Scholar] [CrossRef]
  11. Filelis-Papadopoulos, C.K.; Gravvanis, G.A. On the Multigrid Method Based on Finite Difference Approximate Inverses. Comput. Model. Eng. Sci. 2013, 90, 233–253. [Google Scholar]
  12. Gravvanis, G.; Filelis-Papadopoulos, C.; Matskanidis, P. Algebraic Multigrid Methods Based on Generic Approximate Banded Inverse Matrix Techniques. Comput. Model. Eng. Sci. 2014, 100, 323–345. [Google Scholar]
  13. Kyziropoulos, P.; Filelis-Papadopoulos, C.; Gravvanis, G. Parallel N-body simulation based on the PM and P3M methods using multigrid schemes in conjunction with generic approximate sparse inverses. Math. Probl. Eng. 2015, 2015, 450980. [Google Scholar] [CrossRef] [Green Version]
  14. Nicolaides, R.A. Deflation of Conjugate Gradients with Applications to Boundary Value Problems. SIAM J. Numer. Anal. 1987, 24, 355–365. [Google Scholar] [CrossRef]
  15. Frank, J.; Vuik, C. On the Construction of Deflation-Based Preconditioners. SIAM J. Sci. Comput. 2001, 23, 442–462. [Google Scholar] [CrossRef] [Green Version]
  16. Gravvanis, G.; Filelis-Papadopoulos, C. On the multigrid cycle strategy with approximate inverse smoothing. Eng. Comput. Int. J. Comput.-Aided Eng. 2014, 31, 110–122. [Google Scholar] [CrossRef]
  17. Haelterman, R.; Vierendeels, J.; Van Heule, D. Non-stationary two-stage relaxation based on the principle of aggregation multi-grid. In Proceedings of the Computational Fluid Dynamics 2006; Deconinck, H., Dick, E., Eds.; Springer: Berlin/Heidelberg, Germany, 2009; pp. 243–248. [Google Scholar]
  18. Hackbusch, W. Multi-Grid Methods and Applications; Springer: Berlin/Heidelberg, Germany, 1985; Volume 4. [Google Scholar] [CrossRef]
  19. Bulgakov, V.E. Multi-level iterative technique and aggregation concept with semi-analytical preconditioning for solving boundary-value problems. Commun. Numer. Methods Eng. 1993, 9, 649–657. [Google Scholar] [CrossRef]
  20. Emans, M. Aggregation schemes for k-cycle AMG. Proc. Conf. Algoritm. 2015, 141–150. Available online: http://www.iam.fmph.uniba.sk/amuc/ojs/index.php/algoritmy/article/view/324 (accessed on 28 November 2022).
  21. Chow, E. A Priori Sparsity Patterns for Parallel Sparse Approximate Inverse Preconditioners. SIAM J. Sci. Comput. 2000, 21, 1804–1822. [Google Scholar] [CrossRef]
  22. Grylonakis, E.N.G.; Filelis-Papadopoulos, C.K.; Gravvanis, G.A. On the Numerical Solution of the Generalized Dirichlet-Neumann Map for the 2D Laplace Equation Using Modified Generic Factored Approximate Sparse Inverse Preconditioning. In Proceedings of the 19th Panhellenic Conference on Informatics, Athens, Greece, 1–3 October 2015; Association for Computing Machinery: New York, NY, USA, 2015; pp. 13–18. [Google Scholar] [CrossRef]
  23. Schenk, O.; Gärtner, K. PARDISO. In Encyclopedia of Parallel Computing; Padua, D., Ed.; Springer: Boston, MA, USA, 2011; pp. 1458–1464. [Google Scholar] [CrossRef]
  24. Trottenberg, U.; Oosterlee, C.V.; Schuller, A. Multigrid; Academic Press: Cambridge, MA, USA; Elsevier: Amsterdam, The Netherlands, 2000. [Google Scholar]
  25. Filelis-Papadopoulos, C.K.; Gravvanis, G.A. Parallel multigrid algorithms based on generic approximate sparse inverses: An SMP approach. J. Supercomput. 2014, 67, 384–407. [Google Scholar] [CrossRef]
Figure 1. A V-cycle of aggregation multigrid in conjunction with deflation.
Figure 1. A V-cycle of aggregation multigrid in conjunction with deflation.
Mathematics 11 00640 g001
Figure 2. Distribution of columns of matrix G on four tasks.
Figure 2. Distribution of columns of matrix G on four tasks.
Mathematics 11 00640 g002
Figure 3. Computations performed from the two cores of Task 2 for the formulation of fourth and fifth columns of matrix G of the MGenFAspI.
Figure 3. Computations performed from the two cores of Task 2 for the formulation of fourth and fifth columns of matrix G of the MGenFAspI.
Mathematics 11 00640 g003
Figure 4. Execution times (in seconds) of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Figure 4. Execution times (in seconds) of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Mathematics 11 00640 g004
Figure 5. Speedups of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Figure 5. Speedups of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Mathematics 11 00640 g005
Figure 6. Efficiency of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Figure 6. Efficiency of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Mathematics 11 00640 g006
Table 1. Convergence behavior and performance of V cycle of AGAMG for 2D Poisson model problem.
Table 1. Convergence behavior and performance of V cycle of AGAMG for 2D Poisson model problem.
Level nx
50100200300500
2Execution time0.09670.80538.752940.4147275.5410
Iterations2527272828
3Execution time0.11180.54464.299115.694690.7970
Iterations4145525557
4Execution time0.1250.61483.755211.752959.7511
Iterations4959636465
5Execution time0.14830.76743.958210.457239.3368
Iterations618598105115
Table 2. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique for 2D Poisson model problem.
Table 2. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique for 2D Poisson model problem.
Level nx
50100200300500
2Execution time1.69711.274911.161742.9594296.577
Iterations2026292729
3Execution time0.32211.51837.845422.2745117.9680
Iterations4245474656
4Execution time0.33512.07279.132724.3004100.505
Iterations4060596166
5Execution time0.39642.988516.97641.8845170.785
Iterations4688124130181
Table 3. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 1 for 2D Poisson model problem.
Table 3. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 1 for 2D Poisson model problem.
Level nx
50100200300500
2Execution time0.62324.241524.009172.6027374.277
Iterations1725272728
3Execution time1.334916.395439.329993.8537378.846
Iterations3296545467
4Execution time2.904433.258952.0096379.832561.296
Iterations6518570218105
Table 4. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 1 for 3D Poisson model problem.
Table 4. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 1 for 3D Poisson model problem.
Level nx
50100200300500
2Execution time9.153947.4512573.9462146.985649.41
Iterations810162018
3Execution time9.942948.6524571.8371261.923664.08
Iterations810262123
4Execution time10.467754.8827327.732535.4492954.17
Iterations811282357
Table 5. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 2D Poisson model problem.
Table 5. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 2D Poisson model problem.
Level nx
50100200300500
2Execution time1.32016.944336.079397.401441.648
Iterations2126282728
3Execution time1.922611.448450.453135.453454.7931
Iterations2740424852
4Execution time2.09313.048373.6803191.0220555.9120
Iterations2843596564
5Execution time1.935417.4081100.6990282.1470860.0930
Iterations25567996103
Table 6. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem.
Table 6. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem.
Level nx
50100200300500
2Execution time8.901248.5778709.29401974.34005529.2900
Iterations810151616
3Execution time9.98649.2396519.01401704.42004300.770
Iterations810152022
4Execution time10.231354.0935427.95301100.9602115.110
Iterations811172325
5Execution time13.509651.6422605.79501135.47001956.0400
Iterations1010242527
Table 7. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and different values of l f i l l for 2D Poisson model problem with n x = 300 .
Table 7. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and different values of l f i l l for 2D Poisson model problem with n x = 300 .
Level lfill
1234
2Execution time72.602798.6010132.5990174.9720
Iterations27272727
3Execution time93.8537138.0190208.5200310.3720
Iterations54484850
4Execution time379.8320191.1260318.5270398.9170
Iterations218657161
5Execution time-283.9880407.6990730.2840
Iterations-9689111
Table 8. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and different values of l f i l l for Sherman1 problem.
Table 8. Convergence behavior and performance of V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and different values of l f i l l for Sherman1 problem.
Level lfill
1234
2Execution time0.87570.98761.02881.0522
Iterations369367369357
3Execution time1.10421.20731.35061.1743
Iterations365368366341
4Execution time1.32351.50631.60071.6300
Iterations369371369354
5Execution time1.55041.68231.89671.8641
Iterations375360367339
Table 9. Convergence behavior and performance of a parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Table 9. Convergence behavior and performance of a parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Level lfill
1234
2Execution time2867.371804.981125.15820.11
Iterations54374340
3Execution time2129.461410.571087.21866.76
Iterations34413533
4Execution time3542.252145.361621.551299.05
Iterations53374136
5Execution time3083.262145.081506.311201.97
Iterations45463034
Table 10. Speedup of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Table 10. Speedup of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
LevelThreads
246
21.592.553.77
31.511.962.46
41.652.182.73
51.442.052.57
Table 11. Efficiency of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
Table 11. Efficiency of parallel V cycle of AGAMG in combination with deflation technique and MGenFAspI with d r o p t o l = 0 and l f i l l = 2 for 3D Poisson model problem with n x = 70 .
LevelThreads
246
20.790.640.58
30.750.490.41
40.830.5580.45
50.720.510.43
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Natsiou, A.A.; Gravvanis, G.A.; Filelis-Papadopoulos, C.K.; Giannoutakis, K.M. An Aggregation-Based Algebraic Multigrid Method with Deflation Techniques and Modified Generic Factored Approximate Sparse Inverses. Mathematics 2023, 11, 640. https://doi.org/10.3390/math11030640

AMA Style

Natsiou AA, Gravvanis GA, Filelis-Papadopoulos CK, Giannoutakis KM. An Aggregation-Based Algebraic Multigrid Method with Deflation Techniques and Modified Generic Factored Approximate Sparse Inverses. Mathematics. 2023; 11(3):640. https://doi.org/10.3390/math11030640

Chicago/Turabian Style

Natsiou, Anastasia A., George A. Gravvanis, Christos K. Filelis-Papadopoulos, and Konstantinos M. Giannoutakis. 2023. "An Aggregation-Based Algebraic Multigrid Method with Deflation Techniques and Modified Generic Factored Approximate Sparse Inverses" Mathematics 11, no. 3: 640. https://doi.org/10.3390/math11030640

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop