Chapter02 - 2024-2025 Num - Ana
Chapter02 - 2024-2025 Num - Ana
2024/2025
Contents
1
Chapter 1
Chapter 2 : Eigenvalue
Problems
Eigenvalue problems are a fundamental concept in linear algebra and have sig-
ni…cant applications in various …elds such as physics, engineering, computer
science, and …nance. At their core, eigenvalue problems involve studying linear
transformations represented by matrices, speci…cally how these transformations
a¤ect vectors in a vector space.
In this chapter, we provide a brief overview of two numerical methods for
solving the standard eigenvalue problem, which involves …nding the eigenvalues
and eigenvectors v of an n n matrix A. The …rst method is a simple iterative
technique for identifying the dominant eigenvalue of a matrix, known as the
power method. The second method, which is more computationally intensive,
aims to …nd all eigenvalues of a general matrix and is referred to as the QR
method.
Both methods can be enhanced through an appropriate (inverse) shift. Spe-
cial orthogonal transformations, speci…cally Householder re‡ections, are pivotal
in several widely used eigenvalue algorithms and Singular Value Decomposition
(SV D) methods.
The algebraic eigenvalue problem is de…ned as follows:
Given a matrix A 2 Rn n , …nd a nonzero vector x 2 Rn and the scalar
such that
Ax = x
Note that this says that the vector Ax is parallel to x, with being an
ampli…cation factor, or gain. Note also that the above implies that (A I) x =
0; showing that (A I) is singular matrix. Hence,
p ( ) = det jA Ij = 0;
2
this is not a good way to compute the eigenvalues, it does give us some insight
into their properties.
Basic Eigenvalue Properties
Eigenvalue Count: A matrix A has exactly n eigenvalues, counted with
their algebraic multiplicities. Any complex eigenvalues will appear in conjugate
pairs.
Independence of Eigenvectors: Eigenvectors associated with distinct
eigenvalues of A are linearly independent.
Diagonalizability: If A has n linearly independent eigenvectors, there ex-
ists a nonsingular matrix P such that:
1
P AP = D
Here, D is a diagonal matrix with eigenvalues on its diagonal, and the
columns of P are the corresponding eigenvectors of A.
Properties of Symmetric Matrices: If A is symmetric, then all eigen-
values are real. We can also choose the eigenvectors to be real and orthogonal.
Orthogonal Diagonalization: For symmetric matrices, there exists an
orthogonal matrix Q (where QT = Q 1 ) such that
QT AQ = D
Here, D is a diagonal matrix containing the eigenvalues of A.
Eigenvalues of Triangular Matrices: For an upper or lower triangular
matrix A, the eigenvalues are simply the diagonal entries aii = i .
Example 1 Let 0 1
4 1 0
A=@ 1 4 1 A
0 1 4
The characteristic polynomial for this matrix is
3 2
P( )= 12 + 46 56
and the eigenvalues are
1 = 5:4142; 2 = 4; 3 = 2:5858:
Since the matrix is symmetric, we can …nd a set of orthonormal eigenvectors:
0 1 0 1 0 1
0:5 0:7071 0:5
x1 = @ 0:7071 A ; x2 = @ 0 A ; x1 = @ 0:7071 A :
0:5 0:7071 0:5
Since these vectors are orthonormal, when we arrange them as the columns of
a matrix, we get an orthogonal matrix:
0 1
0:5 0:7071 0:5
Q = @ 0:7071 0 0:7071 A :
0:5 0:7071 0:5
3
1
Moreover, we have Q AQ = QT AQ = D; diagonal
0 1
5:4142 0 0
Q AQ = @
T
0 4 0 A
0 0 2:5858
(A) = f 2 C; Ax = x; x 6= 0g
EA ( ) = f0 6= x 2 Cn ; Ax = xg :
j 1j >j 2j ::: j nj
Ak z
lim k
= cx1 ; for some c 6= 0; and (1)
k!1
1
z; Ak z
lim = 1; (2)
k!1 hz; Ak 1 zi
where
P
n
z= i xi ; 1 6= 0:
i=1
Proof. We have
n
X
k P
n
k k
A z= iA xi = i i xi :
i=1 i=1
4
So, that
k
Ak z P
n
i
k
= 1 x1 + i xi ;
1 i=2 1
where i
1
< 1; for all k 2:
Thus
k
P
n
i
lim i xi = 0;
k!1 i=1 1
k P
n k
1 z; 1 x1 + i
i
1
xi
i=2
=
k 1 Pn k 1
1 z; 1 x1 + i
i
1
xi
i=2
k
Since each of the i
1
;i 2 terms goes to zero, then
z; Ak z hz; 1 x1 i
lim = 1 = 1:
k hz; Ak 1 zi hz; 1 x1 i
Convergence Result
Regarding this algorithm, we have the following result concerning its con-
vergence:
j 1j >j 2j ::: j nj :
5
3) The vector z (0) has a nonzero component in the direction of x1 . This
means that z (0) 2 Rn can be expressed as
P
n
z (0) = i xi ; with 1 6= 0:
i=1
lim k = 1;
k!1
y (k) Az (k 1)
A y (k 1)
A Az (k 2)
z (k) = = = =
k k k k 1 k k 1
A2 z (k 2)
Ak z (0)
= = = = Ck Ak z (0) ;
k k 1 k k 1 ::: 1
1
where Ck = :
k k 1 ::: 1
We then expand in terms of the eigenvector basis, as we did before, to get
P
n P
n Pn
k
z (k) = Ck Ak i xi = iA
k
xi = Ck i i xi
i=1 i=1 i=1
!
k
k P
n
i
= Ck 1 1 x1 + i xi :
i=2 1
= lim Ck k1 k 1 x1 k1 ;
k !1
which implies
k 1
lim Ck 1 = < 1:
k !1 j 1 j kx 1 k1
6
So that there is at least c = j
1
1 jkx1 k1
6= 0 such that
Example 6 Assume
0 1 !
1 3 2 T
3 1
A=@ 4 4 1 A has a dominant eigenpair 7; ; ;1
10 15
6 3 5
7
0 1 0 1
2:311 7 0:350 40
y (4) = Az (3) = @ 0:038 96 A = 6:5974 @ 0:00591 A = 4z
(4)
;
6:5974 1
Since j 4 3j
0:05; then
0 1 0 1
2:3327 0:32762
y (5) = Az (4) = @ 0:42524 A = 7:1201 @ 0:05972 A = 5 = z (5) ;
7:1201 1
Since j 5 4j 0:05; then
0 1 0 1
2:1485 0:3007
y (6) = Az (5) = @ 0:54936 A = 7:1449 @ 0:07689 A = 6z
(6)
:
7:1449 1
Since j 6 5j = 0:0248 < 0:05; then
T
1 ' 6 = 7:1449; and x1 ' z (6) = (0:3007; 0:07689; 1) :
T
z (0) = (1; 0) : Using …ve-digits ‡oating-point decimal arithmetic with rounding.
We have
6 5 1 6 1
y (1) = Az (0) = = =6 = 1 z (1) ;
4 5 0 4 0:66667
9:3334 1
y (2) = Az (1) = = 9:3334 = 2 z (2) ;
7:3334 0:78572
9:9286 1
y (3) = Az (2) = = 9:9286 = 3 z (3) ;
7:9286 0:79856
9:9928 1
y (4) = Az (3) = = 9:9928 = 4 z (4) ;
7:9928 0:799 86
9:9993 1
y (5) = Az (4) = = 9:9993 = 5 z (5) ;
7:9993 0:799 99
10 1
y (6) = Az (5) = = 10 = 6 z (6) ;
8 0:8
10 1
y (7) = Az (6) = = 10 = 6 z (6) :
8 0:8
Hence
T
1 = 10; and x1 = (1; 0:8) :
Remark 8 The power method relies on several hypotheses for the theory to
hold. It’s worth considering whether all of these assumptions are strictly neces-
sary. What occurs if some of these conditions are violated ?
8
1: It is not necessary for the matrix A to have a complete set of eigenvectors
for the Power method to converge. Recall that if a matrix A does not pos-
sess n linearly independent eigenvectors, it is considered defective. Eigenvalue
computations for defective matrices are inevitably more challenging. However,
just as all singular matrices are close to nonsingular matrices, it follows that
all defective matrices can be made arbitrarily close to non-defective matrices.
Consequently, applying the power method to a defective matrix typically leads
to, at worst, slow convergence. In some cases, the convergence may be very
slow, particularly if the dependent eigenvectors correspond to the two largest
eigenvalues, 1 and 2 :
Example of a Defective Matrix : Consider the following 2 2 defective
matrix:
1 1 1
A= ; eigenvectors : $1
0 1 0
T
Let’s apply the power method starting with an initial vector z (0) = (1; 1) :
Iteration Steps:
1 1 1 2 1
y (1) = = =2 = 1z
(1)
;
0 1 1 1 0:5
1
y (2) = Az (1) = 1:5 = 2z
(2)
;
0:33333
1
y (3) = Az (2) = 1:3333 = 3z
(3)
;
0:25
1
y (4) = Az (3) = 1:25 = 4 z (4) ;
0:2
:::::::::::::::::::::::::::::::::::::::::
Conclusion. As we can see, despite starting with a vector that isn’t aligned
with the dominant eigenvector, the power method is converging slowly. Ulti-
mately, you can observe that the results will oscillate without converging to
a well-de…ned eigenvector direction because the matrix A is defective, lacking
su¢ cient independent eigenvectors.
In this case, slow convergence may arise, especially if we need to deal with
eigenvalues that are close in value, which could cause signi…cant delays in reach-
ing an accurate approximation of the dominant eigenvector.
2: If the two largest eigenvalues (in absolute value) are not well-separated,
meaning that j 1 j = j 2 j ; the iteration will cycle through a family of vectors
within the subspace spanned by the eigenvectors x1 and x2 associated with these
eigenvalues. This phenomenon can lead to oscillations instead of converging to
a single vector.
Example: Consider a matrix
2 0
A=
0 2
9
For the matrix A, the eigenvalues and corresponding eigenvectors are as
follows:
T T
1 = 2; x1 = (1; 0) and 2 = 2; x2 = (0; 1) :
T
If we start with an initial vector z (0) = (1; 1) , the iterations will involve the
linear combination of x1 and x2 . Due to the equal magnitudes of the eigenvalues,
the iterations will move between vectors like
2 0 1 2 1
y (1) = = =2 = 1z
(1)
;
0 2 1 2 1
2 0 1 2 1
y (2) = = =2 = 2z
(2)
;
0 2 1 2 1
ultimately cycling between vectors in the subspace spanned by x1 and x2 without
converging to a …xed point.
3: The Power method works by repeatedly multiplying the initial vector
by the matrix A. If z (0) has no component in the direction of the dominant
eigenvector, the results will not converge to the eigenvector corresponding to the
dominant eigenvalue. Instead, they could remain in the span of the components
of the other eigenvectors.
In most situations, particularly those involving non-integer components,
rounding errors during computations will typically introduce some component
in the direction of x1 into the initial guess, allowing the iteration to begin con-
verging toward the correct eigenvalue and eigenvector. However, if the initial
guess contains absolutely no component in the direction of x1 , there is a risk of
converging to the wrong eigenpair. Therefore, using random non-integer initial
guesses is generally more e¤ective to ensure convergence.
Example: Suppose we have a matrix
4 1
A=
2 3
The eigenvalues and eigenvectors of this matrix are:
T T
1 = 5; x1 = (1; 1) and 2 = 2; x2 = (1; 2) :
Initial Guesses
Non-integer guess (e¤ective): Let’s use the initial guess z (0) = (0:1; 0:3)T :
10
Due to rounding errors and the in‡uence of the dominant eigenvector x1 ; z (0)
will still converge to the eigenpair associated with 1 :
Integer guess (ine¤ective): Now consider an initial guess that is entirely
aligned with x2 ; z (0) = ( 1; 2)T :
1
4 1 1 2
y (1) = = =4 2 = 1z
(1)
;
2 3 2 4 1
1 1
4 1 1
y (2) = 2 = =2 2 = 2z
(2)
;
2 3 1 2 1
This guess has no component in the direction of x1 . As a result, it may lead
to convergence towards the eigenpair associated with 2 , which is undesirable
if our goal is to …nd the dominant eigenvalue.
Conclusion. In this example, the initial guess z (0) introduces a component
in the direction of the dominant eigenvector x1 , promoting convergence toward
the desired eigenpair ( 1 ; x1 ) : In contrast, the guess z (0) fails to do so and risks
converging toward the smaller eigenvalue.
11
Convergence Result
Concerning the Invesre Power Mathod Algorithm, we have the following
result of convergence.
j 1j j 2j ::: j n 1j >j nj :
Then the inverse power method converges in the sense that there is c 6= 0
such that
lim z (k) = cxn ; and
k!1
1
lim k =
k!1 n
with the errors
!
k
n
z (k) cxn = O ; k ! 1;
1 n 1
!
k
1 n
k = O ; k ! 1:
n n 1
1
Proof. This is equivalent to applying the previous theorem to A , so every-
thing follows, with the largest (in absolute value) eigenvalue of A 1
being 1n .
Example 11 Do four iterations of the inverse Power method to …nd the ap-
proximate smallest eigenvalue and the corresponding eigenvector of the matrix
2 12 T
A= ; with z (0) = (1; 1) :
1 5
n o
T T
eigenpairs : 2; (3; 1) ; 1; (4; 1) :
U sin g …ve-digits ‡oating-poin t decim al arithm etic w ith roun din g.
We have
5
1 2 6
A = 1 :
2 1
Hence
12
5
6 1 3:5 1
y (1) = A 1 (0)
z = 2
1 = = 3:5 =
2 1 1 0:5 0:14286
(1)
1z ;
5
6 1 1:6428 1
y (2) = A 1 (1)
z = 2
1 = = 1:6428 =
2 1 0:14286 0:35714 0:21740
(2)
2z ;
5
6 1 1:1956 1
y (3) = A 1 (2)
z = 2
1 = = 1:1956 =
2 1 0:21740 0:282 6 0:23637
(3)
3z ;
5
(4) 1 (3) 2 6 1 1:081 8 1
y =A z = 1 = = 1:0818 =
2 1 0:23637 0:26364 0:2437
(4)
4z ;
Thus
1 1 T
2 ' = = 0:92439; and x1 ' z (4) = (1; 0:2437) :
4 1:0818
Ay (k) = z (k 1)
13
1.1.3 Comparison between the Power Method and the In-
verse Power Method
Both the Power Method and the Inverse Power Method rely on iterative processes
to re…ne approximations of eigenvalues and eigenvectors. Each method begins
with an initial vector and performs successive iterations to converge towards an
eigenvector.
While both approaches are e¤ective in addressing eigenvalue problems, they
have distinct characteristics that in‡uence their performance:
Iteration Process: Both methods iteratively update a vector to approxi-
mate an eigenvector, with convergence in‡uenced by the properties of the matrix
and the choice of the initial vector.
Convergence Rates:The convergence rate can vary signi…cantly based on
the matrix’s spectral properties. For the Power Method, the choice of initial
vector is particularly crucial, as it can impact how quickly the method converges
to the dominant eigenvalue.
The Inverse Power Method, on the other hand, allows for accelerated con-
vergence toward smaller eigenvalues and can be adapted for various eigenvalues
by incorporating shifts.
Application Focus:The Power Method excels at identifying the dominant
eigenvalue and its corresponding eigenvector, making it suitable for applications
where the largest eigenvalue is of primary interest.
In contrast, the Inverse Power Method o¤ers greater ‡exibility by enabling
convergence toward almost any desired eigenvalue, which is bene…cial in situa-
tions involving multiple eigenvalues.
Conclusion
In conclusion, while both methods are valuable tools for solving eigenvalue
problems, the choice between them should be guided by the speci…c eigenval-
ues of interest and the characteristics of the matrix in question. The Power
Method is ideal for extracting the dominant eigenvalue, whereas the Inverse
Power Method provides enhanced ‡exibility and speed for targeting other eigen-
values e¤ectively.
S=A I
14
where I is the identity matrix of the same size as A. Therefore, if i are the
eigenvalues of A; then the eigenvalues of S are given by i for each i:
3: Apply the Inverse Power Method: Use the Inverse Power Method on
the shifted matrix S to …nd the eigenvalue and corresponding eigenvector closest
to the shift :
Then the shifted inverse power method converges in the sense that there is
c 6= 0 such that
lim z (k) = cxJ ; and
k!1
1
lim k =( J ) :
k!1
y (k) S 1 (k 1)
z S 1
y (k 1)
S 1
S 1 (k 2)
z
z (k) = = = =
k k k k 1 k k 1
2 (k 2) k (0)
S z S z k (0)
= = = = Ck S z ;
k k 1 k k 1 ::: 1
1
where Ck = :
k k 1 ::: 1
15
We then expand in terms of the eigenvector basis, as we did before, to get
k
P
n P
n 1
z (k) = Ck S k
i xi = Ck i xi
i=1 i=1 i
!
k k
1 P
n
J
= Ck J xJ + i xi :
J i=1;i6=J i
which implies
k
1 1
lim Ck = < 1:
k !1 J j J j kxJ k1
16
By introducing a shift , we can target an eigenvalue more e¤ectively. If we
designate J as the eigenvalue such that J is the smallest, the method
will converge to the eigenvector corresponding to the eigenvalue closest to the
shift :
Moreover, the convergence rate is signi…cantly in‡uenced by the proximity
of the shift to an actual eigenvalue. When the shift is chosen to be close to
J , it generally results in much faster convergence than if the shift is far from
the eigenvalue. This capability makes the Inverse Power Method highly ‡exible
and e¢ cient in targeting various eigenvalues, enhancing its utility in practical
applications.
17
(c) Deduce an approximate eigenvalue 2 based on the previous results.
0 1 0 1
4 1 0 1
A = @ 1 4 1 A ; z (0) = @ 0:5 A
0 1 4 1
np p T
o n o n p p T
o
T
Knowing that the eigenvectors: 2 + 4; 1; 2; 1 ; 4; ( 1; 0; 1) ; 4 2; 1; 2; 1 :
D o al l calculation in …ve decim al places.
18
0 1
0:70709
y (4) = S 1 (3)
z = 11:658 @ 1 A= 4z
(4)
;
0:70709
The, the eigenvalues of de‡ate matrix B are 0; 2 ; :::; n and the eigenvectors
of B are the eigenvectors of A: Here, 0 corresponds to the removed eigenvalue
1:
Proof. The proof of this theorem is left as an exercise in the worksheet02.
To …nd 2 , we can apply the Power Method to B. However, it is important
to note that since 1 is not exact, some error will be introduced in the Power
Method when applied to B. The use of the Power Method to …nd is referred to
as the method of de‡ation.
Example 17 Let
5 2
A=
2 8
(a) Use the Power Method to …nd the dominant eigenvalue 1, and its eigen-
T
vector v, with z (0) = (0; 1) :
19
T
(b) Apply De‡ation Method to …nd 2; with z (0) = (0; 1) : Use ( 1 ; x1 )
otained in (a) :
Stop when k+1 k < 0:03:
x1 xT1
B=A 8:9784 2;
kx1 k2
so 0 10 1T
@
0:49520 A@
0:49520 A
5 2 1 1 3:2319 1:5705
B= (8:9784) 0 1 2 = :
2 8 0:49520 1:5705 0:78973
@ A
1 2
The Power Method applied to B yields
3:2319 1:5705 0 1:0
y (1) = Az (0) = = 1:5705 ; 1 =
1:5705 0:78973 1 0:502 85
1:570 5;
3:231 9 1:5705 1 1:0
y (2) = Az (1) = = 4:0216 ; 2 =
1:5705 0:78973 0:502 85 0:489 26
4:0216;
with j 2 1j 0:03:
20
3:2319 1:570 5 1 1:0
y (3) = Az (2) = = 4:0003 ; 3 =
1:570 5 0:789 73 0:48926 0:489 19
4:0003;
Since j 3 2j = 0:021 3 < 0:03; then
T
2 ' 4:0003; and v2 = (1; 0:489 19) :
Advantages
1. More stable than direct methods based on determinants.
2. E¢ cient for large matrices.
3. Can be parallelized for high-performance computing.
Disadvantages
1. Requires multiple matrix factorizations.
1. Computationally expensive for very large matrices.
21
Convergence
Theorem 20 If A be a symmetric matrix or has distinct eigenvalues, then QR
Algorithm converges to an upper triangular matrix whose diagonal elements are
the eigenvalues of A:
5 2
A= ; the true eigenvalues are f4; 9g :
2 8
Solution 22 We use the formulas for the 2 2 case each time we need a QR
factorization.
5 2
A1 = A = ;
2 8
0:928 0:371 5:385 4:828
Q = Q1 = ; R = R1 =
0:371 0:928 0 6:685
6:788 5 2:482 5
A2 = R1 Q1 = ;
2:480 1 6:203 7
8:325 1 1:708 6
A3 = R2 Q2 = ;
1:708 6 4:675 0
8:850 0:852
A4 = R3 Q3 =
0:852 4:149
::::::::::::::::::::::
4
9 5:955 9 10
A13 = 4 :
5:955 6 10 4
22