Restarting techniques for the (Jacobi-)Davidson symmetric eigenvalue methods

Andreas Stathopoulos, Yousef Saad

Research output: Contribution to journalArticlepeer-review

36 Scopus citations

Abstract

The (Jacobi-)Davidson method, which is a popular preconditioned extension to the Arnoldi method for solving large eigenvalue problems, is often used with restarting. This has significant performance shortcomings, since important components of the invariant subspace may be discarded. One way of saving more information at restart is through "thick" restarting, a technique that involves keeping more Ritz vectors than needed. This technique and especially its dynamic version have proved very efficient for symmetric cases. A different restarting strategy for the Davidson method has been proposed in [14], motivated by the similarity between the spaces built by the Davidson and Conjugate Gradient methods. For the latter method, a three term recurrence implicitly maintains all required information. In this paper, we consider the effects of preconditioning on the dynamic thick restarting strategy, and we analyze both theoretically and experimentally the strategy based on Conjugate Gradient. Our analysis shows that, in some sense, the two schemes are complementary, and that their combination provides an even more powerful technique. We also describe a way to implement this scheme without additional orthogonalizations or matrix multiplications.

Original languageEnglish (US)
Pages (from-to)163-181
Number of pages19
JournalElectronic Transactions on Numerical Analysis
Volume7
StatePublished - 1998

Keywords

  • Conjugate Gradient methods
  • Davidson
  • Deflation
  • Eigenvalue
  • Implicit restarting
  • Jacobi-Davidson
  • Lanczos
  • Preconditioning

Fingerprint

Dive into the research topics of 'Restarting techniques for the (Jacobi-)Davidson symmetric eigenvalue methods'. Together they form a unique fingerprint.

Cite this