DocumentCode :
234507
Title :
Deflation Strategies to Improve the Convergence of Communication-Avoiding GMRES
Author :
Yamazaki, Ichitaro ; Tomov, Stanimire ; Dongarra, Jack
Author_Institution :
Dept. of Electr. Eng. & Comput. Sci., Univ. of Tennessee, Knoxville, TN, USA
fYear :
2014
fDate :
17-17 Nov. 2014
Firstpage :
39
Lastpage :
46
Abstract :
The generalized minimum residual (GMRES) method is a popular method for solving a large-scale sparse nonsymmetric linear system of equations. On modern computers, especially on a large-scale system, the communication is becoming increasingly expensive. To address this hardware trend, a communication-avoiding variant of GMRES (CA-GMRES) has become attractive, frequently showing its superior performance over GMRES on various hardware architectures. In practice, to mitigate the increasing costs of explicitly orthogonalizing the projection basis vectors, the iterations of both GMRES and CAGMRES are restarted, which often slows down the solution convergence. To avoid this slowdown and improve the performance of restarted CA-GMRES, in this paper, we study the effectiveness of deflation strategies. Our studies are based on a thick restarted variant of CA-GMRES, which can implicitly deflate a few Ritz vectors, that approximately span an eigenspace of the coefficient matrix, through the standard orthogonalization process. This strategy is mathematically equivalent to the standard thick-restarted GMRES, and it requires only a small computational overhead and does not increase the communication or storage costs of CA-GMRES. Hence, by avoiding the communication, this deflated version of CA-GMRES obtains the same performance benefits over the deflated version of GMRES as the standard CA-GMRES does over GMRES. Our experimental results on a hybrid CPU/GPU cluster demonstrate that thick-restart can significantly improve the convergence and reduce the solution time of CA-GMRES. We also show that this deflation strategy can be combined with a local domain decomposition based preconditioner to further enhance the robustness of CA-GMRES, making it more attractive in practice.
Keywords :
convergence of numerical methods; eigenvalues and eigenfunctions; iterative methods; linear systems; matrix algebra; vectors; CA-GMRES; Ritz vectors; coefficient matrix eigenspace; communication-avoiding GMRES convergence; deflation strategy; generalized minimum residual method; hardware architectures; hybrid CPU-GPU cluster; large-scale sparse nonsymmetric linear system of equations; local domain decomposition based preconditioner; projection basis vectors; small computational overhead; standard orthogonalization process; standard thick-restarted GMRES variant; storage costs; Convergence; Eigenvalues and eigenfunctions; Graphics processing units; Linear systems; Sparse matrices; Standards; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Latest Advances in Scalable Algorithms for Large-Scale Systems (ScalA), 2014 5th Workshop on
Conference_Location :
New Orleans, LA
Type :
conf
DOI :
10.1109/ScalA.2014.6
Filename :
7016732
Link To Document :
بازگشت