REU 2013: Apprentice Program Summer 2013 Lecture 7: July 8, 2013 Madhur Tulsiani 1 Scribe: Young Kun Ko, David Kim Gram-Schmidt Orthononalization Exercise 1.1 If U is unitary and λ is a (complex) eigenvalue, prove that |λ| = 1. Recall that b1 , . . . , bk form an orthonormal basis if ( 0 if i 6= j hbi , bj i = 1 if i = j Also it is easy to check that b1 , . . . , bk are such that hbi , bj i = 0 for i 6= j then b1 , . . . bk are linearly independent. Theorem 1.2 (Gram-Schmidt Orthogonalization) If v1 , . . . , vk are linearly independent, then there exist vectors b1 , . . . , bk such that 1. b1 , . . . , bk form an orthonormal basis. 2. For all i ≤ k, Span(b1 , . . . , bi ) = Span(v1 , . . . , vk ). Proof: The proof is algorithmic and we claim the following process constructs b1 , . . . , bk with the above property: - Start with b1 = v1 kv1 k . - For each i = 2, . . . , k, set vi0 = vi − hvi , b1 i · b1 − · · · − hvi , bi−1 i · bi−1 and b0i = vi0 kvi0 k Prove by induction on i that the vectors b1 , . . . , bk satisfy the above properties. 2 The Spectral Theorem We can use Gram-Schmidt orthogonalization to prove the following theorem for every square matrix A. The spectral theorem then follows as an easy corollary. Recall that the chracteristic polynomial of any matrix A ∈ Mn (C) has n complex roots. 1 Theorem 2.1 (Schur’s Theorem) Let ∀A ∈ Cn×n be a matrix with eigenvalues λ1 , . . . , λn . Then there exists a unitary matrix U and an upper-triangular matrix T such that A = U T U ∗ . Also, the diagonal entries of T are equal to λ1 , . . . , λn . Proof Sketch: A must have at least one eigenvector corresponding to the eigenvalue λ1 . Consider u1 such that Au1 = λ1 u1 and ku1 k = 1. Complete it to a basis u1 , v2 , . . . , vn for Cn and use GramSchmidt to obtain an orthonormal basis. Note that the first element of basis given by Gram-Schmidt will still be u1 . Let the orthonormal basis be u1 , . . . , un and consider a unitary matrix U1 with columns u1 , . . . , un . Then U1∗ AU1 must be of the form λ1 B1 0 ∗ U1 AU1 = . .. A 1 0 Since A ∼ U1∗ AU1 , the characteristic polynomial of U1∗ AU1 must be the same as that of A. Also, since det(tI − U1∗ AU1 ) = (t − λ1 ) · det(tI − A1 ), A1 must have eigenvalues λ2 , . . . , λn . As above, we can find a unitary matrix V2 ∈ Mn−1 (C) such that B2 λ2 0 V2∗ A1 V2 = . . . A 2 0 Take U2 to be the matrix U2 = 1 0 ··· 0 .. . V2 0 0 , and note that U2∗ U1∗ AU1 U2 is of the form U2∗ U1∗ AU1 U2 = λ1 ∗ 0 λ2 .. . B3 A3 , 0 Continuing this process, we get unitary matrices U1 , . . . , Un such that Un∗ · · · U1∗ AU1 · · · Un = T , where T is upper triangular with diagonal entries λ1 , . . . , λn . Taking U = U1 · · · Un proves the theorem. A matrix A is called normal if AA∗ = A∗ A. Schur’s theorem gives the spectral theorem for normal matrices as an easy corollary. 2 Corollary 2.2 (Spectral Theorem) Let A be a normal matrix with eigenvalues λ1 , . . . , λn . Then ∃U such that A = U DU ∗ where D is a diagonal with entries λ1 , . . . λn . Proof: Assuming Schur’s theorem, proceed in following steps. 1. If A = U T U ∗ as in Schur’s theorem, then T T ∗ = T ∗ T . 2. Show that if T is a triangular matrix which is normal, then T must be diagonal. Example 2.3 Suppose A is normal and thus A = U DU ∗ . Let U = [u1 , ..., un ], Dii = λi . Then Aui = λi ui , and the u0i s form an orthornormal basis of eigenvectors. Exercise 2.4 Show that A is diagonalizable iff for all eigenvalues, algebraic multiplicity = geometric multiplicity. Exercise 2.5 A = U DU ∗ ⇒ A = Pn ∗ i=1 λi ui ui . Definition 2.6 (Hermitian and Symmetric Matrices) An n × n matrix A is Hermitian if A∗ = A. A is called symmetric if AT = A. Note that a Hermitian matrix is normal. Also, a real and symmetric matrix is Hermitian. Using the Spectral Theorem gives the following important conclusion for Hermitian matrices. Proposition 2.7 Let A ∈ Mn (C) be a Hermitian matrix. Then all eigenvalues of A must be real. Proof: If A is Hermitian, then using A = U DU ∗ and A = A∗ gives D = D∗ . Since D is diagonal with entries λ1 , . . . , λn , we get that for each i ∈ [n], λi = λi , which means that all eigenvalues must be real. Also, note that for a real symmetric matrix A, all the eigenvalues of A are real and the proof of Schur’s theorem can be carried out over Rn . This gives that a real symmetric matrix can be written as n X A= λi ui uTi , i=1 where λ1 , . . . , λn are real eigenvalues and u1 , . . . , un ∈ Rn form an orthonormal basis. 3 Adjacency matrices of graphs Definition 3.1 (Adjcency Matrix) Let G = (V, E) be a graph. Then the adjacency matrix A is defined as ( 1 Aij = 0 if (i, j) ∈ E otherwise which is symmetric if G is undirected. 3 If A is the adjacency matrix of an undirected graph G then all eigenvalues of A are real. Let µ1 , . . . , µn be the eigenvalues sorted so that µ1 ≥ µ2 ≥ ... ≥ µn . As before, we can find vectors u1 , . . . , un ∈ Rn forming an orthonormal basis such that A= n X µi ui uTi . i=1 Note that this implies Aui = µi ui and thus u1 , . . . , un are orthonormal (real) eigenvectors corresponding to (real) eigenvalues µ1 , . . . , µn . 4
© Copyright 2026 Paperzz