Power method with shift Normal Power method dominant eigenvalue Inverse Power method least dominant e-value Now consider a small change. Suppose, we now change our matrix A to A qI where q is a scalar. That is we add q to all elements on its diagonal. Now, let us use the Inverse Power Method (IPM) on it. Suppose we now get the smallest eigenvalue . But is an eigenvalue of A qI . Therefore, q is one of the eigenvalues of A . This is the shifting method. In physics, for example, this amounts to the assertion that energy of a system is defined up to an additive constant – there is no absolute zero for an energy. And energy levels in Quantum Mechanics are eigenvalues of a Hamiltonian matrix. Any example? __________________ Stability of numerical eigenvalue problems. Problem. Given Ax x , find the eigenvectors x and the corresponding eigenvalues of the operator A. Note 1. If A and B are almost equal their eigenvalues need not be even close. That is, small changes in A do not produce small changes in its eigenvalue spectrum. The issue here: Truncated data in machine storage may cause serious problems. Ex. Consider 1000 1 1000 1 A and B 0.001 0 1 1 A 1,1 B 0, 2 1000 1 consider this time also C 0.001 1 C has no real eigenvalue since its characteristic polynomial 2 2 2 has no real root. Therefore, Bad News: ■ errors will creep in matrix entries (almost unavoidable) ■ eigenvalues of such a matrix may be suspect if the matrix is no symmetric. and the Good News (from Bad News) ■ If A is symmetric over real field, small perturbations in A will render small changes in eigenvalues. Changes in Eigenvalues Changes in a symmetric Matrix Numerical methods are generally successful in situations when we deal with essentially symmetric matrices to compute eigenvalues. Define Frobenius norm of a matrix A as AF n 2 aij i , j 1 Stability Theorem: Let Ann our actual matrix and Enn is the error matrix (both real and symmetric). Let  A E be the error-version of A . If A 1 ,2 ,3 ,...,n are the eigenvalues of A and  ˆ1 , ˆ 2 , ˆ 3 ,...,ˆ n are the eigenvalues of  then it appears that sum of the deviations of the eigenvalues is bounded. That is 2 i ˆ i E n i 1 2 F With the same hypothesis of the stability theorem, on individual eigenvalue k ˆ k E F 2 2 Note that k ˆ k i ˆ i E F 2 n i 1 Thus, the above constraint. Example. We need to calculate eigenvalues of A where 1 1 2 A 1 2 7 2 7 5 suppose the machine stores it as 1.01 1.05 2.1  1.05 1.97 7.1 7.1 4.9 2.1 Therefore, the error matrix is in our case 0.01 0.05 0.1 E 0.05 0.03 0.1 0.1 0.1 0.1 Therefore, E F ( 0.01 )2 2( 0.05 )2 5( 0.1 )2 0.032 0.23664 So we are guaranteed that | k ˆ k | 0.23664 Even if the absolute error is small, the relative error could be large. Suppose, we have in a specific case | k ˆ k | and k 3 Then the relative error is about 300%. Gramm-Schmidt orthogonalization process. Given a bunch of n arbitrary vectors vi ,i n produce an orthogonal set of vectors ui ,i n The dot product of two vectors a and b would be computed as (assume both are column vectors) b1 b 2 a.b a t b = a1 a2 ... an ... b n a1b1 a2b2 ... anbn The dot product of a with itself is a .a a12 a2 2 ... an 2 a Let u1 v1 and e1 2 u1 u1 Now, let u 2 v2 ( u1 .v2 )u1 and e2 u2 u2 Given this, u1 .u 2 u1 .( v2 ( u1 .v2 )u1 ) = 0 Therefore, u1 is perpendicular to u 2 . Similarly construct u3 using u1 ,u 2 and v3 Again, define u3 v3 ( u1 .v3 )u1 ( u2 .v3 )u2 and u e3 3 u3 Now, u1 .u3 u1 .( v3 ( u1 .v3 )u1 ( u2 .v3 )u2 ) = 0 And, u2 .u3 = 0 Therefore, all u1 ,u 2 and u3 are perpendicular to each other. We continue in this way to define the rest of the orthogonal basis vector set. uk vk ( u1 .v k )u1 ( u2 .vk )u2 ... ( uk 1 .vk 1 )uk 1 uk with ek uk This is Gramm-Schmidt orthogonalization process. ei .e j 1 if i j Note that we have 0, otherwise Therefore the matrix Q e1 | e2 | e3 | ...| en would be an orthonormal matrix i.e. QQ t I We can proceed further. We note that v1 u1 v2 u 2 ( u1 .v2 )u1 v3 u3 ( u1 .v3 )u1 ( u2 .v3 )u2 v4 u4 ( u1 .v4 )u1 ( u2 .v4 )u2 ( u3 .v4 )u3 … This shows that v v1 v2 .. vn QR where Q u1 u2 a11 0 and R 0 0 a12 a22 0 0 .. un .. a1n .. a2 n .. akn .. ann Here Q is an orthogonal matrix and R is an upper triangular one. The entries in R are projections of the original vector v in the orthogonal basis space u . We can convert Q to the orthonormal basis space E where the vectors ei are orthonormal to each other. Example. Three vectors. 2 1 1 v1 1 , v2 1 and v3 3 3 2 2 Assume they span a 3 space. Find an orthogonal basis set that spans the same space. Given that a matrix can be decomposed into QR product, we get for a matrix A at the m th iteration: A( m ) Q ( m ) R ( m ) A( m 1 ) R( m )Q( m ) for m 1,2 ,.. Since R (m) ( m1 ) A Q ( m )t A( m ) we get recursive definition ( m )t A( m )Q( m ) Q As we progress the sequence A( m ) will converge to a triangular matrix with its eigenvalues on the diagonal (or to a nearly triangular matrix whence we can compute the egienvalues).
© Copyright 2026 Paperzz