Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 90% Match Research Paper Researchers in scientific computing,HPC engineers,Machine learning practitioners in scientific domains 2 days ago

Learning Sparse Approximate Inverse Preconditioners for Conjugate Gradient Solvers on GPUs

graph-neural-networks › graph-learning
📄 Abstract

Abstract: The conjugate gradient solver (CG) is a prevalent method for solving symmetric and positive definite linear systems Ax=b, where effective preconditioners are crucial for fast convergence. Traditional preconditioners rely on prescribed algorithms to offer rigorous theoretical guarantees, while limiting their ability to exploit optimization from data. Existing learning-based methods often utilize Graph Neural Networks (GNNs) to improve the performance and speed up the construction. However, their reliance on incomplete factorization leads to significant challenges: the associated triangular solve hinders GPU parallelization in practice, and introduces long-range dependencies which are difficult for GNNs to model. To address these issues, we propose a learning-based method to generate GPU-friendly preconditioners, particularly using GNNs to construct Sparse Approximate Inverse (SPAI) preconditioners, which avoids triangular solves and requires only two matrix-vector products at each CG step. The locality of matrix-vector product is compatible with the local propagation mechanism of GNNs. The flexibility of GNNs also allows our approach to be applied in a wide range of scenarios. Furthermore, we introduce a statistics-based scale-invariant loss function. Its design matches CG's property that the convergence rate depends on the condition number, rather than the absolute scale of A, leading to improved performance of the learned preconditioner. Evaluations on three PDE-derived datasets and one synthetic dataset demonstrate that our method outperforms standard preconditioners (Diagonal, IC, and traditional SPAI) and previous learning-based preconditioners on GPUs. We reduce solution time on GPUs by 40%-53% (68%-113% faster), along with better condition numbers and superior generalization performance. Source code available at https://github.com/Adversarr/LearningSparsePreconditioner4GPU
Authors (6)
Zherui Yang
Zhehao Li
Kangbo Lyu
Yixuan Li
Tao Du
Ligang Liu
Submitted
October 31, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

This paper proposes a novel learning-based method using GNNs to construct GPU-friendly Sparse Approximate Inverse (SAI) preconditioners for Conjugate Gradient solvers. This approach avoids the computationally expensive triangular solves inherent in traditional methods and relies only on matrix-vector products, enabling better GPU parallelization and faster convergence.

Business Value

Significantly speeds up simulations in scientific and engineering domains by accelerating the solution of large linear systems, leading to faster design cycles and reduced computational costs.