TY - JOUR
T1 - Restarting techniques for the (Jacobi-)Davidson symmetric eigenvalue methods
AU - Stathopoulos, Andreas
AU - Saad, Yousef
PY - 1998
Y1 - 1998
N2 - The (Jacobi-)Davidson method, which is a popular preconditioned extension to the Arnoldi method for solving large eigenvalue problems, is often used with restarting. This has significant performance shortcomings, since important components of the invariant subspace may be discarded. One way of saving more information at restart is through "thick" restarting, a technique that involves keeping more Ritz vectors than needed. This technique and especially its dynamic version have proved very efficient for symmetric cases. A different restarting strategy for the Davidson method has been proposed in [14], motivated by the similarity between the spaces built by the Davidson and Conjugate Gradient methods. For the latter method, a three term recurrence implicitly maintains all required information. In this paper, we consider the effects of preconditioning on the dynamic thick restarting strategy, and we analyze both theoretically and experimentally the strategy based on Conjugate Gradient. Our analysis shows that, in some sense, the two schemes are complementary, and that their combination provides an even more powerful technique. We also describe a way to implement this scheme without additional orthogonalizations or matrix multiplications.
AB - The (Jacobi-)Davidson method, which is a popular preconditioned extension to the Arnoldi method for solving large eigenvalue problems, is often used with restarting. This has significant performance shortcomings, since important components of the invariant subspace may be discarded. One way of saving more information at restart is through "thick" restarting, a technique that involves keeping more Ritz vectors than needed. This technique and especially its dynamic version have proved very efficient for symmetric cases. A different restarting strategy for the Davidson method has been proposed in [14], motivated by the similarity between the spaces built by the Davidson and Conjugate Gradient methods. For the latter method, a three term recurrence implicitly maintains all required information. In this paper, we consider the effects of preconditioning on the dynamic thick restarting strategy, and we analyze both theoretically and experimentally the strategy based on Conjugate Gradient. Our analysis shows that, in some sense, the two schemes are complementary, and that their combination provides an even more powerful technique. We also describe a way to implement this scheme without additional orthogonalizations or matrix multiplications.
KW - Conjugate Gradient methods
KW - Davidson
KW - Deflation
KW - Eigenvalue
KW - Implicit restarting
KW - Jacobi-Davidson
KW - Lanczos
KW - Preconditioning
UR - http://www.scopus.com/inward/record.url?scp=0000776543&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=0000776543&partnerID=8YFLogxK
M3 - Article
AN - SCOPUS:0000776543
SN - 1068-9613
VL - 7
SP - 163
EP - 181
JO - Electronic Transactions on Numerical Analysis
JF - Electronic Transactions on Numerical Analysis
ER -