<< numerical_analysis

Eigenvalues

1. Introduction

Recall that for matix A, if Av=λv, we define λ as the eigenvalue and v as the corresponding eigenvector of A.

To compute the eigenvalues and eigenvectors by definition, we need to solve equation |λIA|=0. The order of this equation is at most the order of matrix. For instance, to compute the eigenvalue of A=[0111], we must solve the euqation |λ11λ1|=λ(λ1)1=0 However, Abel–Ruffini theorem states that there is no solution in radicals to general polynomial equations of degree five or higher with arbitrary coefficients. Hence computing the eigenvalue/eigenvector by definition is not an universal approach.

The dominant eigenvalue, or largest magnitude eigenvalue of A is an eigenvalue λ whose magnitude is greater than all other eigenvalues. If it exists, an eigenvector associated to λ is called a dominant eigenvector.

The dominant eigenvalues is much more significant than others in many practical problems. For instance, whether the iteration matrix converges when solving linear equations depends on the spectral radius (a.k.a. dominant eigenvalue) of matix.

Power method is one of the numerical technique to compute dominant eigenvalue, and based on which we developed inverse power method to compute smallest magnitude eigenvalue.

2. Power Method

# Power Method Introduction

Any vector X(0) (as the initial vector) can be expressed by n linear independent eigenvector of A X(0)=α1v1+α2v2++αnvn

And the multiplication AX(0) can be expressed by X(1)=AX(0)=A(α1v1+α2v2++αnvn)=Aα1v1+Aα2v2++Aαnvn=λ1α1v1+λ2α2v2++λnαnvn

Applying mathematical induction , we have X(k)=AX(k1)=λ1kα1v1+λ2kα2v2++λnkαnvn This expression indicates that how X(k) change when iterating depends on the distribution of eigenvalues. We’ll discuss two simple cases.

# Two Cases when Applying Power Method

Case 1. If the ratio x(k+1)x(k) converges, the largest magnitude eigenvalue is unique and is a simple root. The dominant eigenvalue and the dominant eigenvector can be computed by {λ1xi(k+1)xi(k),i=1,2,,nv1X(k)

Proof. For the unique dominant eigenvalue λ1, X(k)=λ1kα1v1+λ2kα2v2++λnkαnvn=λ1k[λ1kα1v1+(λ2λ1)kα2v2++(λnλ1)kαnvn] For sufficiently large k, we have |(λiλ1)k|0 for i1 since λ1 is the dominant eigenvalue. Hence X(k)λ1kα1v1X(k+1)λ1k+1α1v1=λ1λkα1v1=λ1X(k)

Case 2. If the ratio x(k+1)x(k) does not converges but x(2k+2)x(2k) does, then dominant eigenvalues are the real roots with different signs. The dominant eigenvalue and eigenvector can be computed by {λ1=x1(k+2)xi(k),i=1,2,,nv1=X(k+1)+λ1X(k)v2=X(k+1)λ1X(k)

Proof. For two dominant eigenvalues λ1,λ2 and λ1=λ2 X(k)=λ1kα1v1+λ2kα2v2++λnkαnvn=λ1k[λ1kα1v1+(1)kα2v2+(λ3λ1)kα3vn++(λnλ1)kαnvn] For sufficiently large k, we have X(k)λ1k(α1v1+(1)kα2v2)X(k+1)λ1k+1(α1v1+(1)k+1α2v2)X(k+2)λ1k+2(α1v1+(1)k+2α2v2)

Those trends that do not fit the two cases above require other special processings.

# Normalized Power Method

If the dominant eigenvalue is larger or smaller than 1, some components of X(k) will increase or decrease continuously when iterating, which leads to the precision problem due to IEEE754. Hence we can normalize the vector after each step: Y0=X(0)X(0){X(k+1)=AY(k)Y(k+1)=X(k+1)X(k+1) The normalization process gurantees Y(k)=1, that is, the component that represents dominant eigenvalue of A remains ±1.

There are three cases for normalized power method:

  1. If {Xk} converges, there is unique positive dominant eigenvalue {λ1max1in|xi(k)|=|xj(k)|v1Y(k)

  2. If {X(2k)},{X(2k+1)} converges to two opposite vectors, there is unique negative dominant eigenvalue {λ1max1in|xi(k)|=|xj(k)|v1Y(k)

  3. If {X(2k)},{X(2k+1)} converges to two different vectors, there are two different dominant eigenvectors with opposite sign. For sufficiently large k, we exectue one more unnormalized operation X(k+1)=AX(k)=A2Y(k1), then {λ1xi(k+1)yi(k1)λ2=λ1v1=X(k+1)+λ1X(k)v2=X(k+1)λ1X(k)

  4. It requires special pocessing if {X(k)} does not fit any cases mentioned above.

3. Inverse Power Method

Inverse power method is used to compute the smallest magnitude eigenvalue.

In equation Av=λv we multiple A1 on both sides: A1v=λ1v That is, the smallest magnitude eigenvalue of A is exactly the reciprocal of dominant eigenvalue of A1.

In practice, we do apply power method after computing A1 but keep solving equation AX(k+1)=X(k) decomposition method is applied to decompose A=LU first then we solve {LZ(k+1)=X(k)UX(k+1)=Z(k+1) We can also normalize the inverse power method as well to avoid the precision problem: {Y(k)=X(k)X(k)AX(k+1)=Y(k)