28
Eigenvalue Problems

Eigenvalue Problems - University of Utah

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Eigenvalue Problems - University of Utah

Eigenvalue

Problems

Page 2: Eigenvalue Problems - University of Utah

Last Time …

Social Network Graphs

Betweenness

Girvan-Newman Algorithm

Graph Laplacian

Spectral Bisection

𝜆2, 𝑤2

Page 3: Eigenvalue Problems - University of Utah

Today …

Small deviation into eigenvalue problems …

Page 4: Eigenvalue Problems - University of Utah

Formulation

Standard eigenvalue problem: Given a 𝑛 × 𝑛 matrix 𝐴, find scalar 𝜆and a nonzero vector 𝑥 such that

𝐴𝑥 = 𝜆𝑥

𝜆 is a eigenvalue, and 𝑥 is the corresponding eigenvector

Spectrum = 𝜆(𝐴) = set of eigenvalues of 𝐴

Spectral radius = 𝜌 𝐴 = max 𝜆 ∶ 𝜆 ∈ 𝜆 𝐴

Page 5: Eigenvalue Problems - University of Utah

Characteristic Polynomial

Equation 𝐴𝑥 = 𝜆𝑥 is equivalent to

𝐴 − 𝜆𝐼 𝑥 = 0

Eigenvalues of 𝐴 are roots of the characteristic polynomial

det 𝐴 − 𝜆𝐼 = 0

The characteristic polynomial is a powerful theoretical tool but

usually not useful computationally

Page 6: Eigenvalue Problems - University of Utah

Considerations

Properties of eigenvalue problem affecting choice of algorithm

Are all eigenvalues needed, or only a few?

Are only eigenvalues needed, or are corresponding eigenvectors also needed?

Is matrix real or complex?

Is matrix relatively small and dense, or large and sparse?

Does matrix have any special properties, such as symmetry?

Page 7: Eigenvalue Problems - University of Utah

Problem Transformations

Shift: If 𝐴𝑥 = 𝜆𝑥 and 𝜎 is any scalar, then 𝐴 − 𝜎𝐼 𝑥 = 𝜆 − 𝜎 𝑥

Inversion: If 𝐴 is nonsingular and 𝐴𝑥 = 𝜆𝑥, then 𝜆 ≠ 0 and 𝐴−1𝑥 =1

𝜆𝑥

Powers: If 𝐴𝑥 = 𝜆𝑥, then 𝐴𝑘𝑥 = 𝜆𝑘𝑥

Polynomial: If 𝐴𝑥 = 𝜆𝑥 and 𝑝(𝑡) is a polynomial, then 𝑝 𝐴 𝑥 = 𝑝 𝜆 𝑥

Page 8: Eigenvalue Problems - University of Utah

Similarity Transforms

𝐵 is similar to 𝐴 if there is a nonsingular matrix 𝑇, such that

𝐵 = 𝑇−1𝐴𝑇

Then,

𝐵𝑦 = 𝜆𝑦 ⇒ 𝑇−1𝐴𝑇𝑦 = 𝜆𝑦 ⇒ 𝐴 𝑇𝑦 = 𝜆 𝑇𝑦

Similarity transformations preserve eigenvalues and eigenvectors are easily recovered

Page 9: Eigenvalue Problems - University of Utah

Diagonal form

Eigenvalues of a diagonal matrix are the diagonal entries and the

eigenvectors are columns of the Identity matrix

The diagonal form is highly desirable in simplifying eigenvalue

problems for general matrices by similarity transformations

But not all matrices are diagonalizable by similarity transformations

Page 10: Eigenvalue Problems - University of Utah

Triangular form

Any matrix can be transformed into a triangular form by similarity

The eigenvalues are simply the diagonal values

Eigenvectors are not as obvious, but still easy to compute

Page 11: Eigenvalue Problems - University of Utah

Power iteration

Simplest method for computing one eigenvalue-eigenvector pair

𝑥𝑘 = 𝐴𝑥𝑘−1

Converges to multiple of eigenvector corresponding to dominant

eigenvalue

We have seen this before while computing the Page Rank

Proof of convergence ?

Page 12: Eigenvalue Problems - University of Utah

Convergence of Power iteration

Express starting vector 𝑥0 in terms of the eigenvectors of 𝐴

𝑥0 =

𝑖=1

𝑛

𝛼𝑖 𝒗𝑖

Then,

𝑥𝑘 = 𝐴𝑥𝑘−1 = 𝐴2𝑥𝑘−2 = ⋯ = 𝐴

𝑘𝑥0

𝑖=1

𝑛

𝜆𝑖𝑘 𝛼𝑖𝒗𝑖 = 𝜆𝑛

𝑘 𝛼𝑛𝒗𝑛 +

𝑖=1

𝑛−1𝜆𝑖𝜆𝑛

𝑘

𝛼𝑖𝒗𝑖

Since 𝜆𝑖

𝜆𝑛< 1, successively higher powers go to zero

Page 13: Eigenvalue Problems - University of Utah

Power iteration with shift

Convergence rate of power iteration depends on the ratio 𝜆𝑛−1

𝜆𝑛

It is possible to choose a shift, 𝐴 − 𝜎𝐼, such that

𝜆𝑛−1 − 𝜎

𝜆𝑛 − 𝜎<𝜆𝑛−1𝜆𝑛

so convergence is accelerated

Shift must be added back to result to obtain eigenvalue of original

matrix

Page 14: Eigenvalue Problems - University of Utah

Inverse iteration

If the smallest eigenvalues are required rather than the largest, we

can make use of the fact that the eigenvalues of 𝐴−1 are reciprocals of those of 𝐴, so the smallest eigenvalue of 𝐴 is the

reciprocal of the largest eigenvalue of 𝐴−1

This leads to the inverse iteration scheme

𝐴𝑦𝑘 = 𝑥𝑘−1𝑥𝑘 = 𝑦𝑘/ 𝑦𝑘 ∞

Inverse of 𝐴 is not computed explicitly, but some factorization of 𝐴 is

used to solve the system at each iteration

Page 15: Eigenvalue Problems - University of Utah

Shifted inverse iteration

As before, the shifting strategy using a scalar 𝜎 can greatly improve

convergence

It is particularly useful for computing the eigenvector corresponding

to the approximate eigenvalue

Inverse iteration is also useful for computing the eigenvalue closest

to a given value 𝛽, since if 𝛽 is used as the shift, then the desired

eigenvalue corresponds to the smallest eigenvalue of the shifted

matrix

Page 16: Eigenvalue Problems - University of Utah

Deflation

Once the dominant eigenvalue and eigenvector 𝜆𝑛, 𝑤𝑛 have

been computed, the remaining eigenvalues can be computed using deflation, which effectively removes the known eigenvalue

Let 𝐻 be any nonsingular matrix such that 𝐻𝑥 = 𝛼𝑒1

Then the similarity transform determined by 𝐻 transforms 𝐴 into,

𝐻𝐴𝐻−1 = 𝜆𝑛 𝑏𝑇

0 𝐵

Can now work with 𝐵 to compute the next eigenvalue,

Process can be repeated to find additional eigenvalues and eigenvectors

Page 17: Eigenvalue Problems - University of Utah

Deflation

Alternate approach: let 𝑢𝑛 be any vector such that 𝑢𝑛𝑇𝑤𝑛 = 𝜆𝑛

Then 𝐴 − 𝑤𝑛𝑢𝑛𝑇 has eigenvalues 𝜆𝑛−1, … , 𝜆1, 0

Possible choices for 𝑢𝑛

𝑢𝑛 = 𝜆𝑛𝑤𝑛, if 𝐴 is symmetric and 𝑤𝑛 is normalized so that 𝑤𝑛 2 = 1

𝑢𝑛 = 𝜆𝑛𝑦𝑛, where 𝑦𝑛 is the corresponding left eigenvector (𝐴𝑇𝑦𝑛 = 𝜆𝑛𝑦𝑛)

𝑢𝑛 = 𝐴𝑇𝑒𝑘, if 𝑤𝑛 is normalized such that 𝑤𝑛 ∞ = 1 and the 𝑘𝑡ℎ

component of 𝑤𝑛 is 1

Page 18: Eigenvalue Problems - University of Utah

QR Iteration

Iteratively converges to a triangular or block-triangular form, yielding

all eigenvalues of 𝐴

Starting with 𝐴0 = 𝐴, at iteration 𝑘 compute QR factorization,

𝑄𝑘𝑅𝑘 = 𝐴𝑘−1

And form the reverse product,

𝐴𝑘 = 𝑅𝑘𝑄𝑘

Product of orthogonal matrices 𝑄𝑘 converges to matrix of

corresponding eigenvectors

If 𝐴 is symmetric, then symmetry is preserved by the QR iteration, so

𝐴𝑘 converges to a matrix that is both symmetric and triangular

diagonal

Page 19: Eigenvalue Problems - University of Utah

Preliminary reductions

Efficiency of QR iteration can be enhanced by first transforming the

matrix to be as close to triangular form as possible

Hessenberg matrix is triangular except for one additional nonzero

diagonal immediately adjacent to the main diagonal

Symmetric Hessenberg matrix is tridiagonal

Any matrix can be reduced to Hessenberg form in finite number of

steps using Householder transformations

Work per iteration is reduced from 𝒪 𝑛3 to 𝒪(𝑛2) for general

matrices and 𝒪(𝑛) for symmetric matrices

Page 20: Eigenvalue Problems - University of Utah

Krylov subspace methods

Reduces matrix to Hessenberg or tridiagonal form using only matrix-vector products

For arbitrary starting vector 𝑥0, if𝐾𝑘 = 𝑥0 𝐴𝑥0 ⋯ 𝐴

𝑘−1𝑥0

then𝐾𝑛−1𝐴𝐾𝑛 = 𝐶𝑛

where 𝐶𝑛 is upper Hessenberg

To obtain a better conditioned basis for span(𝐾𝑛), compute the QR factorization,

𝑄𝑛𝑅𝑛 = 𝐾𝑛

so that𝑄𝑛𝐻𝐴𝑄𝑛 = 𝑅𝑛𝐶𝑛𝑅𝑛

−1 ≡ 𝐻

with 𝐻 is upper Hessenberg

Page 21: Eigenvalue Problems - University of Utah

Krylov subspace methods

Equating 𝑘𝑡ℎ columns on each side of the equation 𝐴𝑄𝑛 = 𝑄𝑛𝐻 yields

𝐴𝑞𝑘 = ℎ1𝑘𝑞1 +⋯+ ℎ𝑘𝑘𝑞𝑘 + ℎ𝑘+1,𝑘𝑞𝑘+1

relating 𝑞𝑘+1 to preceding vectors 𝑞1, … , 𝑞𝑘

Premultiplying by 𝑞𝑗𝐻 and using orthonormality

ℎ𝑗𝑘 = 𝑞𝑗𝐻𝐴𝑞𝑘 , 𝑗 = 1,… , 𝑘

These relationships yield the Arnoldi iteration

Page 22: Eigenvalue Problems - University of Utah

Arnoldi iteration

𝑥0 : arbitrary nonzero starting vector

𝑞1 = 𝑥0/ 𝑥0 2

for 𝑘 = 1,2, …

𝑢𝑘 = 𝐴𝑞𝑘

for 𝑗 = 1 to 𝑘

ℎ𝑗𝑘 = 𝑞𝑗𝐻𝑢𝑘

𝑢𝑘 = 𝑢𝑘 − ℎ𝑗𝑘𝑞𝑗

ℎ𝑘+1,𝑘 = 𝑢𝑘 2

if ℎ𝑘+1,𝑘 = 0 then stop

𝑞𝑘+1 = 𝑢𝑘/ℎ𝑘+1,𝑘

Page 23: Eigenvalue Problems - University of Utah

Arnoldi iteration

If𝑄𝑘 = 𝑞1 ⋯ 𝑞𝑘

then𝐻𝑘 = 𝑄𝑘

𝐻𝐴𝑄𝑘

is a upper Hessenberg matrix

Eigenvalues of 𝐻𝑘, called Ritz values, are approximate eigenvalues of 𝐴, and Ritz vectors given by 𝑄𝑘𝑦, where 𝑦 is an eigenvector of 𝐻𝑘, are corresponding approximate eigenvectors of 𝐴

Eigenvalues of 𝐻𝑘 must be computed by another method, such as QR iteration, but this is an easier problem as 𝑘 ≪ 𝑛

Page 24: Eigenvalue Problems - University of Utah

Arnoldi iteration

Is fairly expensive in both work and storage because each new

vector 𝑞𝑘 must be orthogonalized against all previous columns of 𝑄𝑘, and all must be stored for that purpose

Is usually restarted periodically with a carefully chosen starting

vector

Ritz vectors and values produced are often good approximations to

eigenvalues and eigenvectors of 𝐴 after relatively few iterations

Page 25: Eigenvalue Problems - University of Utah

Lanczos iteration

Work and storage costs drop dramatically if the matrix is symmetric or Hermitian, since the recurrence has only three terms and 𝐻𝑘 is tridiagonal

𝑞0, 𝛽0 = 0 and 𝑥0 = arbitrary nonzero starting vector

𝑞1 = 𝑥0/ 𝑥0 2

for 𝑘 = 1,2, …

𝑢𝑘 = 𝐴𝑞𝑘

𝛼𝑘 = 𝑞𝑘𝐻𝑢𝑘

𝑢𝑘 = 𝑢𝑘 − 𝛽𝑘−1𝑞𝑘−1 − 𝛼𝑘𝑞𝑘

𝛽𝑘 = 𝑢𝑘 2

if 𝛽𝑘 = 0 then stop

𝑞𝑘+1 = 𝑢𝑘/𝛽𝑘

Page 26: Eigenvalue Problems - University of Utah

Lanczos iteration

𝛼𝑘 and 𝛽𝑘 are diagonal and subdiagonal entries of symmetric

tridiagonal matrix 𝑇𝑘

As with Arnoldi, Lanczos does not produce eigenvalues and

eigenvectors directly, but only the tridiagonal matrix 𝑇𝑘, whose

eigenvalues and eigenvectors must be computed by another

method to obtain Ritz values and vectors

If 𝛽 = 0, then the invariant subspace has already been identified,

i.e., the Ritz values and vectors are already exact at that point

Page 27: Eigenvalue Problems - University of Utah

Lanczos iteration

In principle, if we let Lanczos run until 𝑘 = 𝑛, the resulting tridiagonal

matrix would be orthogonally similar to 𝐴

In practice, rounding errors cause loss of orthogonality

Problem can be overcome by reorthogonalizing vectors

In practice, this is usually ignored. The resulting approximations are

still good

Page 28: Eigenvalue Problems - University of Utah

Krylov subspace methods

Great advantage of Arnoldi and Lanczos is their ability to produce

good approximations to extreme eigenvalues for 𝑘 ≪ 𝑛

They only require one matrix-vector product per step and little

auxiliary storage, so are ideally suited for large sparse matrices

If eigenvalues are needed in the middle of the spectrum, say near

𝜎, then the algorithms can be applied to 𝐴 − 𝜎𝐼 −1, assuming it is

practical to solve systems of the form 𝐴 − 𝜎𝐼 𝑥 = 𝑦