 
 
 
 
 
 
 
 
 
 
 and
 and  .
.
An interesting particular case of oblique projection methods is the
situation in which  is chosen as
 is chosen as  .
Similar to previous notation, let
.
Similar to previous notation, let  be a basis
of
 be a basis
of  . Assuming that
. Assuming that  is nonsingular, 
we can take as a basis of
 is nonsingular, 
we can take as a basis of  the system of vectors
 the system of vectors 
 . 
The approximate eigenvector
. 
The approximate eigenvector  to be extracted  from the subspace
 to be extracted  from the subspace
 can be expressed in the form
 can be expressed in the form 
 
 is an
 is an  -dimensional vector. 
The approximate eigenvalue
-dimensional vector. 
The approximate eigenvalue  is obtained from the Petrov-Galerkin
condition, which yields
 is obtained from the Petrov-Galerkin
condition, which yields
 
 for which the
left-hand-side matrix  is   Hermitian positive definite.    A standard
eigenvalue  problem can  be  obtained by  requiring   that
 for which the
left-hand-side matrix  is   Hermitian positive definite.    A standard
eigenvalue  problem can  be  obtained by  requiring   that  be an
orthonormal system. In this case, eq:HarmGen becomes
 V^ A^ V y =  W^A^-1 W y =1  y .
  
Since the matrix
 be an
orthonormal system. In this case, eq:HarmGen becomes
 V^ A^ V y =  W^A^-1 W y =1  y .
  
Since the matrix   is  orthonormal,  this leads to the  interesting
observation that  the method is mathematically  equivalent to using an
orthogonal projection process  for computing eigenvalues of
 is  orthonormal,  this leads to the  interesting
observation that  the method is mathematically  equivalent to using an
orthogonal projection process  for computing eigenvalues of  .
The subspace of approximants in this case is
.
The subspace of approximants in this case is  .
For this  reason the approximate eigenvalues
.
For this  reason the approximate eigenvalues  are
referred to as harmonic Ritz values. 
Note that one does not have to invert
 are
referred to as harmonic Ritz values. 
Note that one does not have to invert  , but if one maintains the
formal relation
, but if one maintains the
formal relation  by carrying out the orthogonal transformations on
 by carrying out the orthogonal transformations on  also on
also on  , then one can use the left-hand side of (3.10) for
the computation of the reduced matrix.
Since the harmonic Ritz values are Ritz values for
, then one can use the left-hand side of (3.10) for
the computation of the reduced matrix.
Since the harmonic Ritz values are Ritz values for  (although
with respect to a subspace that is generated for
 (although
with respect to a subspace that is generated for  ), they tend to be
increasingly better approximations for interior eigenvalues (those closest
to the origin). One can show, for Hermitian
), they tend to be
increasingly better approximations for interior eigenvalues (those closest
to the origin). One can show, for Hermitian  , that the harmonic Ritz
vectors maximize Rayleigh quotients for
, that the harmonic Ritz
vectors maximize Rayleigh quotients for  so that they can be
interpreted as the best information that one has for the smallest (in absolute
value) eigenvalues. This makes the harmonic Ritz vectors suitable for restarts
and this was proposed, for symmetric matrices, by Morgan [331].
 so that they can be
interpreted as the best information that one has for the smallest (in absolute
value) eigenvalues. This makes the harmonic Ritz vectors suitable for restarts
and this was proposed, for symmetric matrices, by Morgan [331].
The formal introduction of harmonic Ritz values and vectors was given
in [349], along with interesting relations between the Ritz
pairs and the harmonic Ritz pairs. It was shown that the computation of
the projected matrix 
 can be obtained as a rank-one update
of the matrix
 can be obtained as a rank-one update
of the matrix  , in the case of Krylov subspaces, so that the
harmonic Ritz pairs can be generated as cheap side-products of  the regular
Krylov processes. The generalization of the harmonic Ritz values for more
general subspaces was published in [411].
, in the case of Krylov subspaces, so that the
harmonic Ritz pairs can be generated as cheap side-products of  the regular
Krylov processes. The generalization of the harmonic Ritz values for more
general subspaces was published in [411].
Since the projection of  is carried out on a subspace that is generated
for
 is carried out on a subspace that is generated
for  , one should not expect
this method to do as well as a projection on a Krylov subspace that has been
generated with
, one should not expect
this method to do as well as a projection on a Krylov subspace that has been
generated with  . In fact, the harmonic Ritz values are increasingly
better approximations for interior eigenvalues, but the improvement for
increasing subspace can be very marginal (although steady). Therefore,
they are in general no alternative for shift-and-invert techniques, unless
one succeeds in constructing suitable subspaces, for instance by using
cheap approximate shift-and-invert techniques, as in the (Jacobi-) Davidson
methods.
. In fact, the harmonic Ritz values are increasingly
better approximations for interior eigenvalues, but the improvement for
increasing subspace can be very marginal (although steady). Therefore,
they are in general no alternative for shift-and-invert techniques, unless
one succeeds in constructing suitable subspaces, for instance by using
cheap approximate shift-and-invert techniques, as in the (Jacobi-) Davidson
methods.
We will discuss the behavior of harmonic Ritz values and Ritz value in 
more detail for the Hermitian case  .
Assume that the eigenvalues of
.
Assume that the eigenvalues of  are ordered by magnitude:
 are ordered by magnitude:
 
 .
. 
As has been mentioned before, the
Ritz values approximate  eigenvalues of a Hermitian
matrix  ``inside out,'' in the sense that the rightmost eigenvalues
are approximated from below and the leftmost ones are approximated
from above, as is illustrated in the following diagram.
 ``inside out,'' in the sense that the rightmost eigenvalues
are approximated from below and the leftmost ones are approximated
from above, as is illustrated in the following diagram. 
 
 ,
it follows that the harmonic Ritz eigenvalues
,
it follows that the harmonic Ritz eigenvalues  obtained from 
the  process  will approximate corresponding eigenvalues
 obtained from 
the  process  will approximate corresponding eigenvalues 
 of
 of  in an inside-out fashion. 
For positive eigenvalues we have,
 in an inside-out fashion. 
For positive eigenvalues we have, 
 
 . In a sense, they provide the optimal information
for eigenvalues of
. In a sense, they provide the optimal information
for eigenvalues of  that one can derive from a given Krylov  subspace.
For  example, a lower   and  upper bound to   the
(algebraically)  largest positive 
eigenvalue  can    be obtained by    using an
orthogonal   projection  method and a harmonic   projection method,
respectively.
 that one can derive from a given Krylov  subspace.
For  example, a lower   and  upper bound to   the
(algebraically)  largest positive 
eigenvalue  can    be obtained by    using an
orthogonal   projection  method and a harmonic   projection method,
respectively.
We conclude our discussion on harmonic Ritz values with the observation
that they can be computed also for the shifted matrix  , so that
one can force the approximations to improve for eigenvalues close to
, so that
one can force the approximations to improve for eigenvalues close to  .
.
We  must  be cautious   when  applying  this  principle  for  negative
eigenvalues that the order is reversed.  Therefore, we label positive
and negative eigenvalues  separately.  The above inequalities  must be
reversed   for  the negative      eigenvalues,   labeled  from     the
(algebraically)  smallest   to  the    largest   negative  eigenvalues
(
 ).   The  result is summarized  in  the
following diagram.
).   The  result is summarized  in  the
following diagram.
 
 is SPD and
define:
 is SPD and
define: 
 
 . Then the Courant characterization 
becomes,
. Then the Courant characterization 
becomes, 
 
 
 
 
 
 
 
 
 
 
