Algebraic Eigenvalue Ch5

download Algebraic Eigenvalue Ch5

of 13

description

Algebraic Eigenvalue_ch5.pdf

Transcript of Algebraic Eigenvalue Ch5

  • Chapter 5

    Algebraic EigenvalueProblems

    Eigenvalue analysis is an important practice in many elds of engineering orphysics. We also have seen in this course that eigenvalue analysis play animportant role in the performance of many numerical algorithms. A generaleigenvalue problem is stated as follows:

    Denition 5.0.1 Given nn matrices A and B, nd numbers such that theequation

    Ax = Bx (5.1)

    is satised for some nontrivial vector x = 0.

    If B is invertible, then (5.1) can be reduced to

    Cx = x. (5.2)

    Even if both A and B are real-valued, it is likely that and x are complex-valued. Finding the solution of eigensystems is a fairly complicated procedure.It is at least as dicult as nding the roots of polynomials. Therefore, anynumerical method for solving eigenvalue problems is expected to be iterative innature. Algorithms for solving eigenvalue problems include the power method,subspace iteration, the QR algorithm, the Jacobi method, the Arnoldi methodand the Lanczos algorithm.

    Some major references in this eld are given at the end of this note.

    5.1 Perturbation Theory

    Given A = (aij) Rnn, dene ri :=n

    j =i |aij | for all i. Let Ci := {x C| |x aii| ri}. Then

    1

  • 2 CHAPTER 5. ALGEBRAIC EIGENVALUE PROBLEMS

    Theorem 5.1.1 (Gerschgorin) Every eigenvalue of A belongs to one of thecircles Ci. Moreover, if m of the circles form a connected set S disjoint fromthe remaining nm circles, then S contains exactly m eigenvalues.

    (pf): Let be a eigenvalue of A and let x be the corresponding eigenvector,i.e. Ax = x. Let k be the index for which x = |xk|. Then akjxj = xkimplies ( akk)xk =

    j =k akjxj . It follows that | akk| rk, i.e., Ck.

    We know that eigenvalues are roots of the characteristic polynomial p() =det(I A) whereas coecients of p() are polynomials in aij . We also knowthat the zeros of a polynomial depend continuously on its coecients. So weconclude that the eigenvalues of a matrix depend continuously on its elements.Consider A() := D + (A D) where D is the diagonal of A and 0 1.When = 0, all circles degenerate into points since A(0) = D. As increases,all circles expand. Note A(1) = A. As is changed, each eigenvalue () isvaried continuously in the complex plane and mark out a path from (0) to(1). Circles that are disjoint for = 0 are necessarily disjoint for all < 0.By the continuity of i() in , eigenvalues cannot go from a connected set toanother disjoint set. The theorem is proved. Theorem 5.1.2 (Bauer-Fike) If is an eigenvalue of A + E Cnn andP1AP = D = diag {1, . . . , n}. Then

    min | | (P )E (5.3)where denotes any induced norm.

    (pf): Consider (A + E)x = x. Then (I A)x = Ex, (I D)P1x =(P1EP )P1x. Assuming = i, then it follows that P1x (I D)1P1EPP1x. If an induced norm is used, then ID)1 = max 1|| .The theorem is proved. Example. Let A =

    [101, 90110, 98

    ], E =

    [ , 0, 0

    ]. The eigenvalues of A

    are {1, 2}. The eigenvalues of A + E are { 31838+22 }. When = 0.001,

    the eigenvalues are approximately {1.298, 1.701}. Thus min | | 0.298.This example shows that a small perturbation E can lead to relative largeperturbation in the eigenvalues of A.Remark. When A is a normal matrix, i.e., when AA = AA (This class ofmatrices include symmetric matrices, orthogonal matrices, hermitian matrices,etc.), it is known that P may be chosen to be a unitary matrix (i.e., PP =I). If we use the L2-norm, then P2 = P12 = 1. In this case, we havemin | E2. This implies that eigenvalue problems for normal matricesare well-conditioned.Remark. When talking about the perturbation of eigenvectors, some cautionsshould be taken.

    (1) Eigenvectors are unique up to multiplicative constants. Continuity shouldbe discussed only under the assumption that all vectors are normalized in thesame way.

  • 5.2. POWER METHOD AND ITS VARIANTS 3

    (2) There are diculties associated with multiply eigenvalues. For example,

    the matrix

    [1 + , 00, 1

    ]has eigenvectors [1, 0]T and [0, 1]T whereas the matrix

    I has eigenvectors everywhere in R2.Examples. (1) Consider the matrix

    A =

    1 1 0 0 00 1 1 0 00 0 1 1 00 0 0 1 10 0 0 0 1

    which is already in the Jordan Canonical form. It is not diagonalizable andhas eigenvalues = 1 as an n-ford eigenvalues. Consider a slightly perturbedmatrix

    A+E =

    1 1 0 0 00 1 1 0 00 0 1 1 00 0 0 1 1 0 0 0 1

    .

    The eigenvalues of A + E are roots of p() = (1 )n (1)n = 0. So1 i = ()1/ni where ()1/ni is the i-th of the n-th roots of . Note that|i() i| = |1/n|. With n = 10 and = 1010, we nd that |i() i| = .1.Note also that the matrix A has only one eigenvector whereas A + E has acomplete set of eigenvectors.

    (2) Consider the matrix

    [2, 10100,2

    ]. A is not diagonalizable. The exact

    eigenvalue of A are {2, 2} Suppose that = 1 is an approximate eigenvalue witheigenvector x = [1, 1010[T . Then we nd the residue r := Axx = [0, 1010]T .Obviously, the residue is misleading.

    5.2 Power Method and Its Variants

    Given a matrixA Cnn and x(0) Cn. We dene an iteration x(k+1) := Ax(k)for k = 0, 1, . . . What will happen to the sequence {x(k)}. If A is convergent,then x(k) = Akx(0) 0. If A is not convergent, then x(k) may grow unboundly.Either case is undesirable. So we modify the process as follows.

    Algorithm 5.2.1 (Power Method)Given x(0) Cn arbitrary,For k = 1, 2, . . ., do

    w(k) := Ax(k1)

    x(k) :=w(k)

    w(k)

  • 4 CHAPTER 5. ALGEBRAIC EIGENVALUE PROBLEMS

    There is nothing special about doing a sup-norm normalization. In fact, anykind of normalization will be ne. We now examine the convergence propertiesof the power method.

    For illustration purpose, we shall assume A is diagonalizable with eigenvalues|1| > |2| . . . |n|. Let the corresponding eigenvectors be denoted byx1, . . . xn. Detail discussion for other cases can be found in [1]. Suppose x

    (0) =ni=1 ixi. Then

    Ax(0) =ni=1

    iixi

    Akx(0) =ni=1

    iki xi =

    k1

    (1x1 +

    ni=2

    i

    (i1

    )kxi

    ).

    Assume 1 = 0 (This is guaranteed if x(0) is selected randomly). We observe

    x(k) = Ax(k1)

    Ax(k1) =A(

    w(k1)w(k1)

    )A(

    w(k1)w(k1)

    )

    = A2x(k2)

    A2x(k2) =Akx(0)

    Akx(0) .

    It is clear that as k , the vector Akx(0)behaves like 1k1x1 in the sense thatcontribution from x2, . . . xn becomes less and less signicant. Normalization

    makes x(k) 1k1|1k1 |x1

    x1 . That is, x(k) converges to an eigenvector associated

    with the eigenvalue 1. Also, w(k+1) = Ax(k) 1x(k). So (Ax

    (k))j

    x(k)j

    1.Remark. It is clear that the rate of convergence of power method depends onthe ratio 21 .Remark. The eigenvalues of the matrix A bI for a scalar b are i b if andonly if i are eigenvalues of A. With this shift in mind, we can work on thematrix AbI instead of A with the hope that the ratio of the rst two dominanteigenvalues i b will become smaller. This is called a shifted power method.It is not hard to nd that the application is very limited. For example, assumeall eigenvalues are real and are distributed as follows: Then with all choices of

    -2 -1 0 2 4

    3 4 2 1converges to 1 converges to 3

    b the shifted power method will converge to either 1 or 3, but not any othereigenvalues. The idea is plausible, however, in the following setting.

  • 5.2. POWER METHOD AND ITS VARIANTS 5

    Suppose A is nonsingular, then A1 has eigenvalues 11 , . . . ,1n. We apply

    the power method to A1, i.e., we dene

    Algorithm 5.2.2 (Inverse Power Method)Given x(0) Cn arbitrary, For k = 1, 2, . . ., do

    Aw(k) := x(k1)

    x(k) :=w(k)

    w(k) .

    If we assume |1| . . . |n1| > |n|, then {x(k)} converges to an eigen-vector associated with n at the rate depending on |n1n |. Now we incorporatewith the shift idea to obtain the following algorithm.

    Algorithm 5.2.3 (Shifted Inverse Power Method)Given x(0) Cn arbitrary,For k = 1, 2, . . ., do

    (A bI)w(k) := x(k1)

    x(k) :=w(k)

    w(k)The eigenvalues of (A bI)1 are 11b , . . . , 1nb . So whenever b is chosen

    close enough to the eigenvalue i, the sequence {x(k)} by the shifted inversepower method converges to an eigenvector associate with the eigenvalue 1ib .Example. Suppose eigenvalues are distributed as follows: Then the longer

    -2 -1 0 2 4

    3 4 2 1

    vertical bars separates the regions of b by which the shifted inverse power methodwill be able to nd, respectively, dierent eigenvalues.Remark. There are several ways to choose the shift b:

    (1) If some estimate of i has been found, we may use it for b.

    (2) We may generate b(0) at random. Then dene

    b(k+1) :=(w(k))Aw(k)

    (w(k))w(k)(5.4)

    successively. This is a variable-shift inverse power method, and is known as theRayleigh Quotient Iteration. It can be shown that the rate of convergence iscubic [2].

  • 6 CHAPTER 5. ALGEBRAIC EIGENVALUE PROBLEMS

    Remark. In the inverse power method, we have to solve a linear system

    (A bI)w(k) = x(k1). (5.5)

    When b is close to some eigenvalue i, the matrix A bI is nearly singular. Weexpect that small errors in x(k1) would lead to large errors in the computationof w(k). Fortunately, this error mainly occurs in the direction of x(k). Thus afternormalization the method is still useful in practice. (See, [2], for a justication.)

    5.3 The OR algorithm

    The basic idea behind the QR algorithm is from a sequence of matrices {Ai}that are isospectral to the original matrix A1 := A and that converge to a specialform from which the eigenvalues can readily be known. For simplicity, we shallrestrict our discussion for real matrices in the sequel.

    Algorithm 5.3.1 (Basic QR algorithm)

    Given A Rnn, dene A1 := A.For k = 1, 2, . . ., do

    Calculate the QR decomposition Ak = QkRk,

    Dene Ak+1 := RkQk.

    Remark. We observe that Ak+1 = RkQk = QTkAkQk. So Ak+1 is orthogonally

    similar to Ak. By induction, the sequence {Ak} is isospectral to A.

    Denition 5.3.1 We say matrix A Rnn is upper Hessenberg if and only ifAij = 0 for all i 2 j.

    The QR algorithm is expensive because of the large number of computationsthat must be performed at each step. To save this overhead, the matrix Ausually is simplied rst by orthogonal similarity transformations to an upperHessenberg matrix. This can be accomplished as follows:

    Given A, let U2 be the unitary matrix U2 =

    1, 0, . . . , 00,

    U20,

    where U2 is

    formed by the Householder transformation for the column vector [a21, . . . an1]T .

    Thus U2A =

    unchangedx x . . . x0 x x...

    ......

    0 x . . . x

    . It is clear that U2AU

    T2 does not

    change the rst column of U2A. Continuing this procedure, A is transformedby orthogonal similarity transformations into an upper Hessenberg matrix.

  • 5.3. THE OR ALGORITHM 7

    Denition 5.3.2 An orthogonal matrix of the form

    1 0 0 0 0 00 cos 0 0 sin 00 0 1 0 0 00 0 0 1 0 00 sin 0 0 cos 00 0 0 0 0 1

    j th row

    k th row

    is called a rotation matrix in the (j, k)-plane and is denoted as jk.

    Example. Consider the 2 2 matrix A =[a11, a12a21, a22

    ]. Then

    [c ss c

    ] [a11, a12a21, a22

    ]=

    [ca11 + sa21, ca12 + sa22sa11 + ca21, sa12 + ca22

    ].

    Thus if we choose c := a11a211+a

    221

    and s = a21a211+a

    221

    , then the (2, 1)-position

    of the product becomes zero. The matrix

    [c ss c

    ]amounts to the rotation

    of the coordinate axes by an angle determined by tan = sc =a21a11

    . This ideahas been incorporated into the so called Givens method for solving eigenvalueproblems:

    Suppose A is an upper Hessenberg matrix. Then the QR decomposition ofA can be calculated by a sequence of plane rotations. This is demonstrated asfollows:

    A1 =

    x x x xx x x x0 x x x0 0 x x

    12A1 = A2 =

    0 0 x x x0 0 x x

    23A2 = A3 =

    x x x x0 0 0 0 0

    34A3 = A4 =

    x x x x0 x x x0 0 0 0 0

    = R.

    Working with upper Hessenberg matrices, the QR decomposition will only takeO(n2) ops, while with general matrices, the QR decomposition requires O(n3)ops.Remark. It is important to note that if A1 = A is upper Hessenberg, thenthe orthogonal matrix Q1 in the QR decomposition of A1 = Q1R1 is also upper

  • 8 CHAPTER 5. ALGEBRAIC EIGENVALUE PROBLEMS

    Hessenberg (Prove this!). Therefore, A2 = R1Q1 maintains to be upper Hessen-berg. It follows that the upper Hessenberg structure is maintained throughoutthe QR algorithm.

    We have seen that

    Ak+1 = QTkAkQk = Q

    Tk . . . Q

    T1 AQ1 . . .Qk.

    Let us call

    Pk : = Q1 . . . Qk (5.6)

    Uk : = Rk . . . R1. (5.7)

    Note that Pk is still orthogonal and Uk upper triangular. Moreover,

    PkUk = (Q1 . . .Qk)(Rk . . . R1) = (Q1 . . . Qk1)Ak(Rk1 . . .R1)= (Q1 . . .Qk1)(QTk1 . . .Q

    T1 AQ1 . . . Qk1)(Rk1 . . .R1)

    = APk1Uk1 = . . . = Ak. (5.8)

    In other words, we have found that PkUk is the QR decomposition of the matrixAk

    Theorem 5.3.1 Suppose A Rnn and suppose |1| > . . . > |n| > 0. Thenthe sequence {Ak} converges to an upper triangular matrix.

    (pf): From the assumption, we know all eigenvalues of A are real and dis-tanct. So A is diagonalizable. Let P be the nonsingular matrix such thatP1AP = D := diag {1, . . . , n}. Then Ak = PDkP1. Let P1 = LUbe the LU decomposition of P1 with value 1 along the diagonal of L (We as-sume this LU decomposition exists with some prior pivoting if necessary). ThenAk = PDkLU = P (DkLDk)DkU . Consider DkLDk. The (i, j)-componentis (ij )klij . Since L is lower triangular, only the lower triangular i.e., i j,portion needs to be considered.

    Let DkLDk := I + Ek. By the ordering of i, we see that Ek 0 ask . Let P := QR. Then

    Ak = QR(DkLDk)DkU= QR(I +Ek)D

    kU = Q(I +REkR1)RDkU

    = Q(QkRk)RDkU = (QQk)(RkRD

    kU). (5.9)

    where QkRk is the QR decomposition of I +REkR1.

    We recall the issue of uniqueness of the QR decomposition of a matrix.Suppose a nonsingular matrix B has two QR decompositions, say B = QR =QR. Then QT Q = RR1. But QT Q is orthogonal whereas RR1 is uppertriangular. An orthogonal and upper triangular matrix is diagonal with absolutevalue 1 along its diagonal. If we call RR1 := D, then Q = QD and R = D1R.So the QR decomposition of B is unique if the diagonal elements of R arerequired to be positive.

  • 5.3. THE OR ALGORITHM 9

    We have already known Ak = PkUk. Thus Ak has two QR decompositions.

    It follows that

    Pk = QQkDk

    Uk = D1k RkRD

    KU = DkRkRDkU

    where Dk is a diagonal matrix with elements either 1 or -1. Now

    Ak+1 = PTk APk = (QQkDk)

    TA(QQkDk) = DkQTk (Q

    TAQ)(QkDk)

    = DkQTk (RP

    1APR1)(QkDk)= DkQ

    Tk (RDR

    1)(QkDk). (5.10)

    Note thatRDR1 is an upper triangular matrix with diagonal elements 1, . . . , n.Observe also that I + REkR

    1 converges to I as k since Ek 0. SoRk = Qk(I + REkR

    1) QTk as k . Since Rk is upper triangularwith positive diagonal elements, Rk and Qk I as k . Thus we seefrom (5.10) that Ak+! an upper triangular matrix with 1, . . . , n along thediagonal. Remark. From the above proof, we see that the rate of convergence dependson the ration maxj |j+1j |. It is, therefore , desired to use the shift technique toaccelerate the convergence.

    Algorithm 5.3.2 (QR algorithm with origin shift)(Explicit shift)Given A Rnn , dene A1 := A.For k = 1, 2, . . ., do

    Select a shift factor ck;Calculate the QR decomposition,

    Ak ckI = QkRk; (5.11)

    DeneAk+1 := RkQk + ckI. (5.12)

    Remark. We note that Rk = QTk (Ak ckI). So Ak+1 = QTk (Ak ckI)Qk +

    ckI = QTkAkQk. That is, Ak+1 and Ak are orthogonally similar.

    Denition 5.3.3 Suppose Ak is upper Hessenberg. The shift factor ck may bedetermine from the eigenvalues, say k or k, of the bottom 2 2 submatrix ofAk. If both k and k are real, we take ck to be k or k according to whether

    |k a(k)nn | or |k a(k)nn | is smaller. If k = k, then we choose ck = k. Sucha choice of shift factor is known as the Wilkinson shift.

    Theorem 5.3.2 Given any matrix A and the relationship B = QTAQ. SupposeB is upper Hessenberg with positive and diagonal elements and suppose Q isorthogonal. Then matrices Q and B are uniquely determined from the rstcolumn of Q.

  • 10 CHAPTER 5. ALGEBRAIC EIGENVALUE PROBLEMS

    (pf): This is homework problem.Remark. Let A be any given matrix and let B = QTAQ be an upper Hessen-berg matrix. Let PT be an orthogonal matrix whose rst column agrees withthat of Q. Suppose PAPT is reduced to upper Hessenberg, that is, supposeU1, . . . , Un2 are orthogonal matrices such that

    B := Un2 . . . U1(PAPT )U1 . . . UTn2

    becomes an upper Hessenberg matrix (Recall how this is done!). Let Q =PTUT1 . . . U

    Tn2. Then we have Q

    TAQ = B. Note that the rst column of Q is

    the same as that of Q. From Theorem 5.3.1, we conclude that B = B.Suppose Ak is upper Hessenberg. There is an easier way to compute Ak+1 =

    QTkAkQk from Ak based on the above theorem and remark: Since Ak is upperHessenberg, the rst column ofQk is known, i.e., it is the column [c,s, 0, . . . , 0]Twith

    c :=a(k)11 ck

    (a(k)11 ck)2 + a(k)21

    s :=a(k)21

    (a(k)11 ck)2 + a(k)21

    .

    So the rst column of PT is known and PT can be chosen to be the matrix

    PT :=

    c s 0 0s c 0 00 0 1 00 0 0 1

    It is easy to see that PAkPT is at most a matrix of the form

    PAkPT =

    x x x x xx x x x xx x x x x0 0 x x x0 0 0 x x

    So we can make a rotation in the (2.3)-plane to eliminate the nonzero elementat the (3, 1)-position. In doing so, the (4, 2)-position becomes nonzero. Wecontinue to chase the nonzero elements by n 2 rotations to produce Ak+1.

    Suppose we have applied one QR step with shift c1 to obtain A2 = QT1 A1Q1

    and other QR step with shift c2 to obtain A3 = QT2 A2Q2 = Q

    T2Q

    T1 A1Q1Q2.

    Then we may as well using the rst column of Q1Q2 to compute A3 fromA1 by using Theorem 5.3.1 and the above remark. We recall that A1 c1I =Q1R1, A2 = R1Q1+c1I, A2c2I = Q2R2 and A3 = R2Q2+c2I. It follows thatT2 = (A3 c2I)QT2 = (QT2QT1 A1Q1Q2 c2I)QT2 = QT2QT1 (A1 c2I)Q1. ThusR2R1 = Q

    T2Q

    T1 (A1 c2I)Q1R1, or equivalently Q1Q2R2R1 = (A1 c2I)(A

    c1I) = A2 (c0 + c1)A + c0c1I. That is, the rst column of Q1Q2 is seem to

    be the normalized rst column of the matrix A2 (c0 + c1)A + c0c1I. This isimplicit double shift.

  • 5.4. EXERCISES 11

    5.4 Exercises

    1. Let p be a chosen integer satisfying 1 p n. Given an n p matrix Q0with orthonormal columns, consider the iteration

    Zk = AQk1QkRk = Zk (QR factorization)

    for k = 1, 2, . . .. Explain why this iteration can usually be used to computethe p largest eigenvalues of A in absolute value. How then should youmodify the iteration when the p smallest eigenvalues of A in absolutevalue are needed.

    2. Prove that the upper Hessenberg structure is maintained throughout theQR algorithm. That is, if Ak = QkRk is an upper Hessenberg matrix,show that Ak+1 = RkQk is also upper Hessenberg.

    3. Prove Theorem 5.3.2.

    4. Compute a QR step with the matrix

    A =

    [2 1

    ]

    (a) without shift;

    (b) with shift k = 1.

    5. Assume A = diag(1, 2) with 1 > 2. Assume

    xk =

    (cksk

    )

    with c2k + s2k = 1 is an approximate eigenvector.

    (a) Compute the Rayleigh quotient of xk.

    (b) Compute, by pen and paper, the next Raleigh quotient iterationxk+1.

    (c) Comparing xk and xk+1, give an explanation why the convergence ofthe Rayleigh quotient iteration is ultimately cubic.

  • 12 CHAPTER 5. ALGEBRAIC EIGENVALUE PROBLEMS

  • Bibliography

    [1] D. K. Faddeev and V. N. Faddeva, Computational Methods of Linear Al-gebra, W. H. Freeman and Co., San Francisco, 1963.

    [2] B. N. Parlett, The Symmetric Eigenvalue Problems, Prentice-Hall, Engle-wood Clis, N.J., 1980.

    [3] B. T. Smith et al, Matrix Eigensystem Routines, EISPACK Guide, Lecturenotes in computer science, Vol. 6, Berlin, Springer-Verlag, 1974.

    [4] G.W. Stewart and Ji-guang Sun, Matrix Perturbation Theory, AcademicPress, Boston, 1990.

    [5] D. S. Watkins, Understanding the QR algorithm, SIAM Review, 24(1982),427-440.

    [6] J. H. Wilkinson and C. Reinsch, Linear Algebra, Handbook for AutomaticComputation, Vol. II, ed. F. L. Bauer, Berlin, Springer-Verlag, 1971.

    [7] J.H. Wilkinson, The Algebraic Eigenvalue Problem, Oxford, ClarendonPress, 1988.

    13