0% found this document useful (0 votes)
464 views5 pages

Understanding Singular Value Decomposition

The singular value decomposition (SVD) is the factorization of a matrix A into three matrices: A = UΣV^T, where U and V are orthogonal matrices and Σ is a diagonal matrix containing the singular values of A. The SVD provides orthogonal bases for the row and column spaces of A such that the transformation of one basis into the other is a scaling by the singular values. The document provides examples to demonstrate calculating the SVD of matrices.

Uploaded by

FayazKhanPathan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
464 views5 pages

Understanding Singular Value Decomposition

The singular value decomposition (SVD) is the factorization of a matrix A into three matrices: A = UΣV^T, where U and V are orthogonal matrices and Σ is a diagonal matrix containing the singular values of A. The SVD provides orthogonal bases for the row and column spaces of A such that the transformation of one basis into the other is a scaling by the singular values. The document provides examples to demonstrate calculating the SVD of matrices.

Uploaded by

FayazKhanPathan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd

Singular value decomposition

The singular value decomposition of a matrix is usually referred to as the SVD.


This is the nal and best factorization of a matrix:
A = UV T
where U is orthogonal, is diagonal, and V is orthogonal.
In the decomoposition A = UV T , A can be any matrix. We know that if A
is symmetric positive denite its eigenvectors are orthogonal and we can write
A = QQ T . This is a special case of a SVD, with U = V = Q. For more general
A, the SVD requires two different matrices U and V.
Weve also learned how to write A = SS1 , where S is the matrix of n
distinct eigenvectors of A. However, S may not be orthogonal; the matrices U
and V in the SVD will be.

How it works
We can think of A as a linear transformation taking a vector v1 in its row space
to a vector u1 = Av1 in its column space. The SVD arises from nding an
orthogonal basis for the row space that gets transformed into an orthogonal
basis for the column space: Avi = i ui .
Its not hard to nd an orthogonal basis for the row space the Gram-
Schmidt process gives us one right away. But in general, theres no reason
to expect A to transform that basis to another orthogonal basis.
You may be wondering about the vectors in the nullspaces of A and A T .
These are no problem zeros on the diagonal of will take care of them.

Matrix language
The heart of the problem is to nd an orthonormal basis v1 , v2 , ...vr for the row
space of A for which

A v1 v2 vr = 1 u1 2 u2 r ur

1
2
= u1 u2 ur ,

. ..
r
with u1 , u2 , ...ur an orthonormal basis for the column space of A. Once we
add in the nullspaces, this equation will become AV = U. (We can complete
the orthonormal bases v1 , ...vr and u1 , ...ur to orthonormal bases for the entire
space any way we want. Since vr+1 , ...vn will be in the nullspace of A, the
diagonal entries r+1 , ...n will be 0.)
The columns of U and V are bases for the row and column spaces, respec
tively. Usually U = V, but if A is positive denite we can use the same basis
for its row and column space!

1
Calculation


4 4
Suppose A is the invertible matrix . We want to nd vectors v1
3 3
and v2 in the row space R2 , u1 and u2 in the column space R2 , and positive
numbers 1 and 2 so that the vi are orthonormal, the ui are orthonormal, and
the i are the scaling factors for which Avi = i ui .
This is a big step toward nding orthonormal matrices V and U and a di
agonal matrix for which:
AV = U.
Since V is orthogonal, we can multiply both sides by V 1 = V T to get:

A = UV T .

Rather than solving for U, V and simultaneously, we multiply both sides by


A T = V T U T to get:

A T A = VU 1 UV T
= V2 V T
2
1
22
T
=V V .

..
.
n2

This is in the form QQ T ; we can now nd V by diagonalizing the symmetric


positive denite (or semidenite) matrix A T A. The columns of V are eigenvec
tors of A T A and the eigenvalues of A T A are the values i2 . (We choose i to be
the positive square root of i .)
To nd U, we do the same thing with AA T .

SVD example

4 4
We return to our matrix A = . We start by computing
3 3

T 4 3 4 4
A A =
4 3 3 3

25 7
= .
7 25

The eigenvectors of this matrix will give us the vectors vi , and the eigenvalues
will gives us the numbers i .

2

1 1
Two orthogonal eigenvectors of A T A are and . To get an or
1 1

1/2 1/2
thonormal basis, let v1 = and v2 = . These have eigen
1/ 2 1/ 2
values 12 = 32 and 22 = 18. We now have:


A U T
V
4 4 4 2 0 1/2 1/2
= .
3 3 0 3 2 1/ 2 1/ 2

We could solve this for U, but for practice well nd U by nding orthonor
mal eigenvectors u1 and u2 for AA T = U2 U T .

T 4 4 4 3
AA =
3 3 4 3

32 0
= .
0 18

1
Luckily, AA T happens to be diagonal. Its tempting to let u1 = and u2 =
0

0 0
, as Professor Strang did in the lecture, but because Av2 = we
1 3 2

0 1 0
instead have u2 = and U = . Note that this also gives us a
1 0 1
chance to double check our calculation of 1 and 2 .
Thus, the SVD of A is:


A U T
V
4 4 1 0 4 2 0 1/2 1/2
= .
3 3 0 1 0 3 2 1/ 2 1/ 2

Example with a nullspace



4 3
Now let A = . This has a one dimensional nullspace and one dimen
8 6
sional row and column spaces.
4
The row space of A consists of the multiples of . The column space
3
4
of A is made up of multiples of . The nullspace and left nullspace are
8
perpendicular to the row and column spaces, respectively.
.8
Unit basis vectors of the row and column spaces are v1 = and u1 =
.6

3

1/5
. To compute 1 we nd the nonzero eigenvalue of A T A.
2/ 5

4 8 4 3
A T A =
3 6 8 6

80 60
= .
60 45

Because this is a rank 1 matrix, one eigenvalue must be 0. The other must equal
the trace, so 12 = 125. After nding unit vectors perpendicular to u1 and v1
(basis vectors for the left nullspace and nullspace, respectively) we see that the
SVD of A is:

4 3 1 2 125 0 .8 .6
= 1 .
8 6 5 2 1 0 0 .6 .8
A U VT

The singular value decomposition combines topics in linear algebra rang


ing from positive denite matrices to the four fundamental subspaces.
v1 , v2 , ...vr is an orthonormal basis for the row space.

u1 , u2 , ...ur is an orthonormal basis for the column space.

vr+1 , ...vn is an orthonormal basis for the nullspace.

ur+1 , ...um is an orthonormal basis for the left nullspace.

These are the right bases to use, because Avi = i ui .

4
MIT OpenCourseWare
[Link]

18.06SC Linear Algebra


Fall 2011

For information about citing these materials or our Terms of Use, visit: [Link]

You might also like