0% found this document useful (0 votes)
97 views21 pages

Algebra

The document discusses orthogonal diagonalization of matrices and proves several equivalent conditions: 1) A matrix A has an orthonormal set of eigenvectors. 2) A is orthogonally diagonalizable. 3) A is symmetric. It proves that if a matrix meets one of these conditions, it meets all three. The key steps are using an orthogonal matrix P to diagonalize A into D, and showing transpose properties to prove symmetry.

Uploaded by

Samin Khadka
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
97 views21 pages

Algebra

The document discusses orthogonal diagonalization of matrices and proves several equivalent conditions: 1) A matrix A has an orthonormal set of eigenvectors. 2) A is orthogonally diagonalizable. 3) A is symmetric. It proves that if a matrix meets one of these conditions, it meets all three. The key steps are using an orthogonal matrix P to diagonalize A into D, and showing transpose properties to prove symmetry.

Uploaded by

Samin Khadka
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd

Algebra

Orthogonal diagonalization
• A matrix A is said to be orthogonally diagonalizable if there is
an orthogonal matrix P and a diagonal matrix D such that:
• D=PTAP
• where the elements in the diagonal of the matrix D are the
eigenvalues of the matrix A.
Theorem
• The following conditions are equivalent for an n×n matrix A.
•  1. A has an orthonormal set of n eigenvectors. 
• 2. A is orthogonally diagonalizable. 
• 3. A is symmetric.
If A has an orthonormal set of n eigenvectors, A is
orthogonally diagonalizable. 
• (1) ⇔ (2)    (If 1, then 2. If 2, then 1.)
• Let x1, x2, ..., xn be orthonormal set of n eigenvectors of A. Then, P =
[x1 x2 ... xn] is an orthogonal matrix. 
• Since the columns of P are orthogonal and hence independent, P is
invertible. 
• A is diagonalizable if and only if it has eigenvectors x1, x2, . . . , xn
such that the matrix P=[x1 x2 . . . xn] is invertible(Theorem i). 
• By theorem i, A is diagonalizable. The diagonal matrix D=P-1AP.
• Since P is an orthogonal matrix, P-1=PT. 
• Hence, D=P-1AP= PTAP i.e A is orthogonally diagonalizable. 
If A is orthogonally diagonalizable, A has an
orthonormal set of n eigenvectors, 
• Suppose that A is orthogonally diagonalizable. Then there exists an
orthogonal matrix P such that PTAP is a diagonal matrix. If P has
columns {x1,x2,…xn}, then B = {x1,x2, …xn} is a set of n orthonormal
vectors in Rn.
• Since B is a set of n orthonormal vectors, B is independent. Let PTAP =
diag(l1,l2,….ln) = D (Here li are the eigenvalues of A).
• Then AP = PD, so
• Thus Axi = lixi for each i, 1≤ I ≤ n, implying that B consists of eigenvectors
of A. Therefore, A has an orthonormal set of n eigenvectors.
If A is orthogonally diagonalizable, A is
symmetric. 
•  (2)⇒ (3)   (If 2, then 3)
•  Since A is orthogonally diagonalizable, let D=PTAP be the diagonal
matrix, where P-1 =PT.
• Then, A = PDPT.
• But DT = D because the transpose of a diagonal matrix is the diagonal
matrix itself.
• So, this gives AT =(PT)TDTPT = PDPT = A
• Hence, A is symmetric. 
If A is symmetric, A is orthogonally
diagonalizable.
• Third,
• (3)⇒ (2)  (If 3,then 2)
• We need an intermediate theorem to prove this.
• Lemma
• Suppose that λ1 is an eigenvalue of the nxn symmetric matrix A, with
corresponding unit eigenvector v1. Then there is an orthogonal matrix
P whose 1st column is v1, such that

• where A1 is an (n-1)x (n-1) symmetric matrix and O is a zero matrix.


• Proof: 
• First, we know that we can extend the set {v1} to a basis {v1,
…….,vn} for Rn. And then we can use the Gram-Schmidt
Procedure on this basis to get an orthonormal basis {w1,
…….,wn} for Rn. But, since 
• v1  is already a unit vector, the Gram-Schmidt Procedure will
set  w1= v1, so 
• B = { v1, w2, ……wn} is an orthonormal basis for Rn. Let P =[ v1
w2..... wn].  
• Since λ1  is a unit eigenvector of v1, we have that v1. Av1 =  λ1
v1. v1= λ1 . 
• So we've shown that (PTAP)11 =1. Since B is orthonormal, we
get that w1 .Av1= w1 .( λ1 v1)= λ1  w1. v1 = 0. 
• So (PTAP)i1 = 0 for all 2≤ i ≤ n. 
• And the fact that (PTAP) is symmetric tells us that (PTAP)1i= 0 =
0 for all 2 ≤ i ≤ n. And 
• This means that PTAP=

•     where A1 must be symmetric because PTAP is symmetric


Proof
• The proof is by induction on n, the size of our symmetric matrix A.
The base case is when n = 1, which means A = [a], and A is diagonalized
by the orthogonal matrix P = [1] to PTAP = [1][a][1] = [a].
• For our induction step, suppose we know that all (n-1)x(n-1) symmetric
matrices are orthogonally diagonalizable, and let A be an nxn
symmetric matrix.
• Let λ be the eigenvalue of A with corresponding unit eigenvector  v1.
• By Lemma, we know that there is an orthogonal matrix Q=[v1 w2....wn] such that
QTAQ=  

    where A' is an (n-1)x(n-1) symmetric matrix.


• By our inductive hypothesis, there is an (n-1)x (n-1)   orthogonal matrix Q' and an
(n-1) x (n - 1) diagonal matrix D' such that Q'TA'Q'= D'.
• Also, A′=Q′D′Q′T .
• Hence,

• It follows that
• We obtained a diagonal matrix in the middle but we can't yet say that
A has been orthogonally diagonalized. 
• However, if we let U=                        , D=                    , we obtain A=UDUT

•  If we can show that U is an orthogonal,  A becomes orthogonally


diagonalizable. 

                                                                              
• Indeed, it is so. Thus, A=UDUT (orthogonally diagonalizable).
• If 1,then 2.
• If 2,then 1.
• If 2,then 3.
• If 3,then 2.

• Therefore,
• If 1,then 3.
• If 3,then 1.
Theorem: The Principal Axes Theorem
• Let A be a n x n symmetric matrix. Then there is an orthogonal change
of variable, x = Py, that transforms the quadratic form xTAx, into a
quadratic form yTDy with no cross-product term.
• Suppose we have a quadratic form
• Q= xTAx  where xT = [x x ...x ]  
, 1 2 n       (*)

• Here A is symmetric. So, we know that we can find an orthogonal


matrix P (i.e., PT = P-1) that diagonalizes A,
(i.e. P-1AP = D = PTAP).
• If we replace A with PDPT in (*) then we obtain
• Q  = xT(PDPT)x  
  = (xTP)D(PTx)
• Now, if we define a new set of vector yT= [y1 y2 ...yn] via the
substitution x= Py then the previous line becomes
• Q  =  ((Py)TP)D(PTPy)
•   =  yT (PTP) D (PTP) y
  = yT D y = λ1y12 + λ2 y22 +…+ λn yn2       
• There are no cross product terms here.
• Sources
• https://people.math.carleton.ca/~kcheung/math/notes/MAT
H1107/wk10/10_orthogonal_diagonalization_proof.html
• LINEAR ALGEBRA with Applications (W. Keith Nicholson)

• Thank you!

You might also like