next up previous contents index
Next: Orthogonal Deflating Transformation Up: Implicitly Restarted Lanczos Method Previous: Computational Costs and Tradeoffs   Contents   Index


Deflation and Stopping Rules

Deflation is an important concept in the practical implementation of the QR iteration and therefore equally important to the IRLM. In the context of the QR iteration, deflation amounts to setting a small subdiagonal (and corresponding superdiagonal) element of the tridiagonal matrix $T$ to zero. This is called deflation because it splits the tridiagonal matrix into two smaller subproblems which may be independently refined further.

However, in the context of IRLM, the usual QR deflation techniques are not always appropriate. Additional deflation capabilities specific to implicit restart are needed. It is often desirable to deflate at a coarser accuracy level $\epsilon_D$ with $ 1 > \epsilon_D > \epsilon_M$, where $\epsilon_M$ is machine precision. In theory (i.e., in exact arithmetic), when $A$ has multiple eigenvalues it would be impossible for IRLM to compute more than one eigenvector to such a multiple eigenvalue. This is because it is a ``single vector" rather than a block method. However, in practice, there is usually little difficulty in computing multiple eigenvalues, because the method deflates itself as convergence takes place, and roundoff usually introduces components in new eigenvector directions in the subsequent starting vectors. Nevertheless, this can be unreliable and miss a multiple eigenvalue. In any case, this approach typically will require a stringent convergence tolerance in order to find all instances of a multiple eigenvalue.

It is far more efficient to deflate (i.e., lock) an approximate eigenvalue once it has converged to a certain level of accuracy and to force subsequent Lanczos vectors to be orthogonal to the converged subspace. With this capability, additional instances of a multiple eigenvalue can be computed to the same specified accuracy without the expense of forcing them to converge to unnecessarily high accuracy.

In the Lanczos process, as with a QR iteration, it is possible for some of the leading $k$ subdiagonals to become small during the course of implicit restart. However, it is usually the case that there are converged Ritz values appearing in the spectrum of $T$ long before small subdiagonal elements appear. This convergence is usually detected through observation of a small last component in an eigenvector $y$ of $T$. When this happens, we are able to construct an orthogonal similarity transformation of $T$ that will give an equivalent Lanczos factorization with a slightly perturbed $T$ that does indeed have a zero subdiagonal, and this is the basis of our deflation scheme.

In the context of IRLM, there are two types of deflation required:

Locking:
If a Ritz value $\theta $ has converged (meaning $\Vert Ax - x\theta\Vert < \epsilon_D$) and is thought to be a member of the wanted set of eigenvalues, we wish to declare it converged, decouple the eigenpair $(x, \theta)$, and continue to compute remaining eigenvalues with no further alteration of $x$ or $\theta $. This process is called locking.

Purging:
If a Ritz value $\theta $ has converged but is not a member of the wanted set, we wish to decouple and remove the eigenpair $(x, \theta)$ from the current Krylov subspace spanned by the Lanczos vectors. This process is called purging.
Techniques for this deflation were first developed in [289,294]. The scheme we present here is based on an improvement developed in [420]. This scheme allows for stable and efficient deflation (or locking) of Ritz values that have converged with a specified relative accuracy of $\epsilon_D$ which may be considerably larger than machine precision $\epsilon_M$. This is particularly important when a SI is not available to accelerate convergence. Typically in this setting the number of matrix-vector products will be large, and it will be highly desirable to lock converged Ritz values at low tolerances. This will avoid the expense of the matrix-vector products that would be required to achieve an accuracy that would allow normal QR-type deflation. Also, it is important to be able to purge converged but unwanted Ritz values. As pointed out in  [289], the forward instability of the QR bulge chase process discovered by Parlett and Le  [359] will prevent implicit restart being used for purging converged unwanted Ritz values.


next up previous contents index
Next: Orthogonal Deflating Transformation Up: Implicitly Restarted Lanczos Method Previous: Computational Costs and Tradeoffs   Contents   Index
Susan Blackford 2000-11-20