GPU-accelerated preconditioned iterative linear solvers

Ruipeng Li, Yousef Saad

Research output: Contribution to journalArticlepeer-review

190 Scopus citations


This work is an overview of our preliminary experience in developing a high-performance iterative linear solver accelerated by GPU coprocessors. Our goal is to illustrate the advantages and difficulties encountered when deploying GPU technology to perform sparse linear algebra computations. Techniques for speeding up sparse matrix-vector product (SpMV) kernels and finding suitable preconditioning methods are discussed. Our experiments with an NVIDIA TESLA M2070 show that for unstructured matrices SpMV kernels can be up to 8 times faster on the GPU than the Intel MKL on the host Intel Xeon X5675 Processor. Overall performance of the GPU-accelerated Incomplete Cholesky (IC) factorization preconditioned CG method can outperform its CPU counterpart by a smaller factor, up to 3, and GPU-accelerated The incomplete LU (ILU) factorization preconditioned GMRES method can achieve a speed-up nearing 4. However, with better suited preconditioning techniques for GPUs, this performance can be further improved.

Original languageEnglish (US)
Pages (from-to)443-466
Number of pages24
JournalJournal of Supercomputing
Issue number2
StatePublished - Feb 2013

Bibliographical note

Funding Information:
Acknowledgements This work is supported by DOE under grant DE-FG 08ER 25841 and by the Minnesota Supercomputer Institute.


  • GPU computing
  • Preconditioned iterative methods
  • Sparse matrix computations


Dive into the research topics of 'GPU-accelerated preconditioned iterative linear solvers'. Together they form a unique fingerprint.

Cite this