0% found this document useful (0 votes)
41 views

Econ 2001

This document outlines a lecture on eigenvectors and eigenvalues, diagonalization of matrices, and quadratic forms. Some key points: - Eigenvalues are solutions to the characteristic equation det(A - λI) = 0, and eigenvectors are non-trivial solutions to (A - λI)x = 0. - Diagonalization involves finding matrices P and D, where D is diagonal, such that P-1AP = D. This relates a complicated matrix A to a simple diagonal matrix D. - Diagonalization allows solving linear difference equations, by changing variables such that the equations become independent. - The diagonalization theorem states that if a matrix A has distinct eigenvalues or is
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
41 views

Econ 2001

This document outlines a lecture on eigenvectors and eigenvalues, diagonalization of matrices, and quadratic forms. Some key points: - Eigenvalues are solutions to the characteristic equation det(A - λI) = 0, and eigenvectors are non-trivial solutions to (A - λI)x = 0. - Diagonalization involves finding matrices P and D, where D is diagonal, such that P-1AP = D. This relates a complicated matrix A to a simple diagonal matrix D. - Diagonalization allows solving linear difference equations, by changing variables such that the equations become independent. - The diagonalization theorem states that if a matrix A has distinct eigenvalues or is
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 35

Lecture 8

Econ 2001

2015 August 19
Lecture 8 Outline

1 Eigenvectors and eigenvalues


2 Diagonalization
3 Quadratic Forms
4 De…niteness of Quadratic Forms
5 Uniqure representation of vectors
Eigenvectors and Eigenvalues
De…nition
An eigenvalue of the square matrix A is a number such that A I is singular.
If is an eigenvalue of A, then any x 6= 0 such that (A I)x = 0 is called an
eigenvector of A associated with the eigenvalue .

Therefore:
1 Eigenvalues solve the equation
det(A I) = 0:
2 Eigenvectors are non trivial solutions to the equation
Ax = x

Why do we care about eigenvalues and their corresponding eigenvectors?


They enable one to relate complicated matrices to simple ones.
They play a role in the study of stability of di¤erence and di¤erential equations.
They make certain computations easy.
They let us to de…ne a way for matrices to be positive or negative and that
matters for calculus and optimization.
Characteristic Equation
De…nition
If A is an n n matrix, the characteristic equation is de…ned as
0 1
11 12 1n
B C
B 21 22 2n C
f ( ) = det(A I) = 0 or f ( ) = det B .. .. .. C=
@ . . . A
m1 m2 mn

This is a polinomial equation in .

Example
For a two by two matrix:

a11 a12
A I = ) det A = (a11 )(a22 ) a12 a21
2 2 2 2 a21 a22
Hence the characteristic equation is
2
(a11 + a22 ) + a11 a22 a12 a21 = 0
Which typically has two solutions.
Characteristic Polynomial
The characteristic equation f ( ) = det(A I) = 0 is a polynomial degree n.
By the Fundamental Theorem of Algebra, it has n roots (not necessarily
distinct and not necessarily real).
That is
f( )=( c1 )( c2 ) ( cn )
where c1 ; : : : ; cn 2 C (the set of complex numbers) and the ci ’s are not
necessarily distinct.
Notice that f ( ) = 0 if and only if 2 fc1 ; : : : ; cn g, so the roots are all the
solutions of the equation f ( ) = 0.
if = ci 2 R, there is a corresponding eigenvector in Rn .
if = ci 62 R, the corresponding eigenvectors are in Cn n Rn .
Another way to write the characteristic polynomial is
m1 mk
P( ) = ( r1 ) ( rk ) ;
where r1 ; r2 : : : ; rk are distinct roots (ri 6= rj when i 6= j) and mi are positive
integers summing to n.
mi is called the multiplicity of root ri .
Distinct Eigenvectors
FACT
The eigenvectors corresponding to distinct eigenvalues are distinct.

Proof By Contradiction.
Let 1 ; : : : ; k be distinct eigenvalues and x1 ; : : : ; xk the associated eigenvectors.
Suppose these vectors are linearly dependent (why is this a contradiction?).
WLOG, let the …rst k 1 vectors be linearly independent, while xk is a linear
combination of the others.
Pk 1
Thus, 9 i i = 1; : : : ; k 1 not all zero such that: i =1 i xi = xk
Multiply both sides by A and use the eigenvalue property (Ax = x):
k 1
X
i i xi = k xk
i =1
Multiply the …rst equation by k and subtract it from the second:
k 1
X
i( i k )xi = 0
i =1

Since the eigenvalues are distinct i 6= k for all i; hence, we have a


non-trivial linear combination of the …rst k 1 eigenvectors equal to 0,
contradicting their linear independence.
Diagonalization

De…nition
We say B is diagonalizable if we can …nd matrices P and D, with D diagonal, such
that
P 1 BP = D

If a matrix is diagonalizable
1 1 1
PP BPP = PDP
or
1
B = PDP
where D is a diagonal matrix.

Why do we care about this?


We can use simple (i.e. diagonal) matrices to ‘represent’more complicated
ones.
This property is handy in many applications.
An example follows: linear di¤erence equations.
Di¤erence Equations Detour

A di¤erence equation is an equation in which “discrete time” is one of the


independent variables.

For example, the value of x today depends linearly on its value yesterday:
xt = axt 1 8t = 0; 1; 2; 3; : : :
This is a fairly common relationship in time series data and macro.
Given some initial condition x0 , this equation is fairly easy to solve using
x1 = ax0
x2 = ax1 = a2 x0
‘recursion’: ...
xt 1 = axt 2 = at 1
x0
xt = axt 1 = at x0
Hence:
xt = at x0 8t = 0; 1; 2; 3; : : :
Di¤erence Equations Detour Continued
Consider now a two-dimensional linear di¤erence equation:
ct+1 b11 b12 ct
= 8t = 0; 1; 2; 3; : : :
kt+1 b21 b22 kt
given some initial condition c0 ; k0 .
ct b11 b12
Set yt = 8t and B = and rewrite this more
kt b21 b22
compactly as
yt+1 = Byt 8t = 0; 1; 2; 3; : : :
where bij 2 R each i; j.
We want to …nd a solution yt ; t = 1; 2; 3; : : : given the initial condition y0 .
Such a dynamical system can arise as a characterization of the solution to a
standard optimal growth problem (you will see this in macro).
This is hard to solve since the two variables interact with each other as time
goes on.
Things would be much easier if there were no interactions (b12 = 0 = b21 )
because in that case the two equations would evolve independently.
Di¤erence Equations Detour: The End
We want to solve
yt+1 = Byt 8t = 0; 1; 2; 3; : : :

If B is diagonalizable, there exist an invertible 2 2 matrix P and a diagonal


2 2 matrix D such that
d1 0
P 1 BP = D =
0 d2
Then
1 1
yt+1 = Byt 8t () P yt+1 = P Byt 8t
1 1 1
() P yt+1 = P BPP yt 8t
() ^
yt+1 = D^
yt 8t
1
where we de…ned ^
yt = P yt 8t (this is just a change of variable).
Since D is diagonal, after this change of variable to ^
yt we now have to solve
two independent linear univariate di¤erence equations
yit = dit ^
^ y0 8t for i = 1; 2
which is easy because we can use recursion.
Diagonalization Theorem
Theorem
If A is an n n matrix that either has n distinct eigenvalues or is symmetric, then
there exists an invertible n n matrix P and a diagonal matrix D such that
1
A = PDP
Moveover, the diagonal entries of D are the eigenvalues of A, and the columns of
P are the corresponding eigenvectors.

Note
1
Premultiply by P and postmultiply by P , the theorem says:
1
P AP = D

De…nition
1
Two square matrices A and B are similar if A = P BP for some invertible matrix
P.

The theorem says that some square matrices are similar to diagonal matrices
that have eigenvalues on the diagonal.
Diagonalization Theorem: Idea of Proof

We want to show that for a given A there exist a matrix P and a diagonal matrix
D such that A = PDP 1 , where the diagonal entries of D are the eigenvalues of A
and the columns of P are the corresponding eigenvectors.

Idea of Proof (a real proof is way too di¢ cult for me)
Suppose is an eigenvalue of A and x is an eigenvector. Thus Ax = x.
If P is a matrix with column j equal to the eigenvector associated with i, it
follows that AP = PD.
The result would then follow if one could guarantee that P is invertible.
The proof works by showing that when A is symmetric, A only has real
eigenvalues, one can …nd n linearly independent eigenvectors even if the
eigenvalues are not distinct (these results use properties of complex numbers).
See a book for details.
A Few Computational Facts For You To Prove
Facts

det AB = det BA = det A det B

If D is a diagonal matrix, then det D is equal to the product of its diagonal


elements.
det A is equal to the product of the eigenvalues of A.

De…nition
The trace ofPa square matrix A is given by the sum of its diagonal elements. That
n
is, tr (A) = i =1 aii :

Fact
n
X
tr (A) = i;
i =1
where i is the ith eigenvalue of A (eigenvalues counted with multiplicity).
Unitary Matrices

Remember
At is the transpose of A: the (i; j)th entry of At is the (j; i)th entry of A.

De…nition
An n n matrix A is unitary if At = A 1
.

REMARK
By de…nition every unitary matrix is invertible.
Unitary Matrices
Notation
A basis V = fv1 ; : : : ; vn g of Rn is orthonormal
1 if each basis element has unit length (vi vi = 1 8i), and
2 distinct basis elements are orthogonal (vi vj = 0 for i 6= j).
1 if i = j
Compactly, this can be written as vi vj = ij = .
0 if i =
6 j

Theorem
An n n matrix A is unitary if and only if the columns of A are orthonormal.

Proof.
Let vj denote the j th column of A.
At = A 1
() At A = I
() vi vj = ij
() fv1 ; : : : ; vn g is orthonormal
Symmetric Matrices Have Orthonormal Eigenvectors

Remember
A is symmetric if aij = aji for all i; j, where aij is the (i; j)th entry of A.

Theorem
If A is symmetric, then the eigenvalues of A are all real and there is orthonormal
basis V = fv1 ; : : : ; vn g of Rn consisting of the eigenvectors of A. In this case, P in
the diagonalization theorem is unitary and therefore:
A = PDPt

The proof is also beyond my ability (uses the linear algebra of complex vector
spaces).
Quadratic Forms

Think of a second-degree polynomial that has no constant term:


n
X X
2
f (x1 ; : : : ; xn ) = ii xi + ij xi xj
i =1 i <j

Let 0 1
( 11 1n
ij
2 if i < j B .. .. .. C
ij = ji
and A = @ . . . A
2 if i > j
n1 nn
Then,
f (x) = xt Ax
This inspires the idea of a quadratic form.
Quadratic Forms

De…nition
A quadratic form in n variables is a function Q : Rn ! R that can be written as
Q(x) = xt Ax
where A is a symmetric n n matrix.

Why do we care about quadratic forms? They show up in many places.


For example, think about a function from Rn to R.
We will see that the …rst derivative of this function is a vector (we take the
derivative one component at a time).
Thus, the second derivative is a matrix (we take the derivative of the …rst
derivative one component at a time).
This matrix can be thought of as a quadratic form.
Thus, a function’s shape could be related to the “sign” of a quadratic form.
What is the sign of a quadratic form anyhow?
Examples of Quadratic Forms

Example
When n = 1 a quadratic form is a function of the form ax 2 .

Example
When n = 2 it is a function of the form
a11 x12 + 2a12 x1 x2 + a22 x22
(remember a12 = a21 by symmetry).

Example
When n = 3, it is a function of the form
a11 x12 + a22 x22 + a33 x32 + 2a12 x1 x2 + 2a13 x1 x3 + 2a23 x2 x3
Sign of A Quadratic Form

De…nition
A quadratic form Q(x) = xt Ax is
1 positive de…nite if Q(x) > 0 for all x 6= 0.
2 positive semi de…nite if Q(x) 0 for all x.
3 negative de…nite if Q(x) < 0 for all x 6= 0.
4 negative semi de…nite if Q(x) 0 for all x.
5 inde…nite if there exists x and y such that Q(x) > 0 > Q(y).

In most cases, quadratic forms are inde…nite.


What does all this mean? Hard to tell, but we can try to look at special cases
to get some intuition.
Positive and Negative De…niteness

Idea
Think of positive and negative de…niteness as a way one applies to matrices the
idea of “positive” and “negative”.

In the one-variable case, Q(x) = ax 2 and de…niteness follows the sign of a.


Obviously, there are lots of inde…nite matrices when n > 1.
Diagonal matrices also help with intuition. When A is diagonal:
Xn
Q(x) = xt Ax = aii xi2 :
i =1
therefore the quadratic form is:
positive de…nite if and only if aii > 0 for all i , positive semi de…nite if and only
if aii 0 for all i
negative de…nite if and only if aii < 0 for all i , negative semi de…nite if and only
if aii 0 for all i , and
inde…nite if A has both negative and positive diagonal entries.
Quadratic Forms and Diagonalization

For symmetric matrices, de…niteness relates to the diagonalization theorem.


Assume A is symmetric.
By the diagonalization theorem:
A = Rt DR;

where D is a diagonal matrix with (real) eigenvalues on the diagonal and R is


an orthogonal matrix.
For any quadratic form Q(x) = xt Ax, by de…nition, A is symmetric.
Then we have
t
Q(x) = xt Ax = xt Rt DRx = (Rx) D (Rx) :

The de…niteness of A is thus equivalent to the de…niteness of its diagonal


matrix of eigenvalues, D.
Think about why.
Quadratic Forms and Diagonalization: Analysis
A quadratic form is a function Q(x) = xt Ax where A is symmetric.
Since A is symmetric, its eigenvalues 1 ; : : : ; n are all real.
Let V = fv1 ; : : : ; vn g be an orthonormal basis of eigenvectors of A with
corresponding eigenvalues 1 ; : : : ; n . By an earlier theorem, the P in the
diagonalization theorem is unitary.
0 1
1 0 0
B 0 0 C
B 2 C
Then: A = Ut DU where D = B . . . .. C and U is unitary
@ .. .. .. . A
0 0 n
n
Pn
We know that any x 2 R can be written as x = i =1 i vi .
Then, one can rewrite a quadratic form as follows:
X X t X X t X
Q(x) = Q i vi = i vi A i vi = i vi Ut DU i vi
X t X X t X
= U i vi D U i vi = i Uvi D i Uvi
0 1
1
B C X
= ( 1 ; : : : ; n )D @ ... A = i i
2

n
Quadratic Forms and Diagonalization

The algebra on the previous slide yields the following result.

Theorem
The quadratic form Q(x) = xt Ax is
1 positive de…nite if i > 0 for all i.
2 positive semi de…nite if i 0 for all i.
3 negative de…nite if i < 0 for all i.
4 negative semi de…nite if i 0 for all i.
5 inde…nite if there exists j and k such that j >0> k.

REMARK
We can check de…niteness of a quadratic form using the eigenvalues of A.
Principal Minors

De…nition
A principal submatrix of a square matrix A is the matrix obtained by deleting any
k rows and the corresponding k columns.

De…nition
The determinant of a principal submatrix is called the principal minor of A.

De…nition
The leading principal submatrix of order k of an n n matrix is obtained by
deleting the last n k rows and column of the matrix.

De…nition
The determinant of a leading principal submatrix is called the leading principal
minor of A.
Principal minors can be used in de…nitess tests.
Another De…niteness Test

Theorem
A matrix is
1 positive de…nite if and only if all its leading principal minors are positive.
2 negative de…nite if and only if its odd principal minors are negative and its
even principal minors are positive.
3 inde…nite if one of its kth order leading principal minors is negative for an even
k or if there are two odd leading principal minors that have di¤erent signs.

This classi…es de…niteness of quadratic forms without …nding the eigenvalues


of the corresponding matrices.
Think about these conditions when applied to diagonal matrices and see if
they make sense in that case.
Back to Linear Algebra De…nitions
All vectors below are elements of X (a vector space) and all scalars are real
numbers.
The linear combination of x1 ; : : : ; xn with coe¢ cients 1 ; : : : ; n :
Xn
y= i xi
i =1
The set of all linear combinations of elements of V = fv1 ; : : : ; vk g
k
X
spanV = fx 2 X : x = i vi with v 2V g
i =1
A set V X spans X if spanV = X .
A set V X is linearly dependent if
n
X
9v1 ; : : : ; vn 2 V and 1; : : : ; n not all zero such that i vi =0
i =1
A set V X is linearly independent if it is not linearly dependent.
Thus, V X is linearly independent if and only if:
n
X
i vi = 0 with each vi 2 V ) i = 0 8i
i =1
.
A basis of X is a linearly independent set of vectors in X that spans X .
Vectors and Basis

Any vector can be uniquely written as a …nite linear combination of the


elements of some basis of the vector space to which it belongs.

Theorem
Let V be a basis for a vector space X over R. Every vector x 2 X has a unique
representation as a linear combination of a …nite number of elements of V (with all
coe¢ cients nonzero).

Haven’t we proved this yet? Not at this level of generality.


P
The unique representation of 0 is 0 = i 2; i vi .
Any vector has a unique representation as linear combination of …nitely many
elements of a basis.
Proof.
Since V spans X , any x 2 X can be written as a linear combination of elements of
V . We need to show this linear combination is unique.
X X
Let x= s vs and x= s vs
s 2S 1 s 2S 2
where S1 is …nite, s 2 R, s 6= 0, and vs 2 V for each s 2 S1 and
where S2 is …nite, s 2 R, s 6= 0, and vs 2 V for each s 2 S2 .
De…ne S = S1 [ S2 ,
s = 0 for s 2 S2 n S1 and s = 0 for s 2 S1 n S2
Then
X X X X X
0=x x= s vs s vs = s vs s vs = ( s s )vs
s 2S 1 s 2S 2 s 2S s 2S s 2S

Since V is linearly independent, we must have s s = 0, so s = s, for all


s 2 S. s 2 S1 , s 6= 0 , s 6= 0 , s 2 S2
So S1 = S2 and s = s for s 2 S1 = S2 , and the representation is unique.
A Basis Always Exists
Theorem
Every vector space has a (Hamel) basis.

This follows from the axiom of choice (did we talk about this?).
An equivalent result says that if a linearly independent set is not a basis, one
can always “add” to it to get a basis.

Theorem
If X is a vector space and V X is a linearly independent set, then V can be
extended to a basis for X . That is, there exists a linearly independent set W X
such that
V W spanW = X

There can be many bases for the same vector space, but they all have the
same number of elements.

Theorem
Any two Hamel bases of a vector space X have the same cardinality (are
numerically equivalent).
Standard Basis
De…nition
The standard basis for Rn consists of the set of N vectors ei , i = 1; : : : ; N, where
ei is the vector with component 1 in the ith position and zero in all other positions.
0 1 0 1 0 1 0 1
1 0 0 0
B0C B1C B0C B0C
B C B C B C B C
B C B C B C B C
e1 = B0C e2 = B0C en 1 = B ... C en = B ... C
B .. C B .. C B C B C
@.A @.A @1A @0A
0 0 0 1

1 A standard basis is a linearly independent set that spans Rn .


2 Elements of the standard basis are mutually orthogonal. When this happens,
we say that the basis is orthogonal.
3 Each basis element has unit length. When this also happens, we say that the
basis is orthonormal.

Verify all these.


Orthonormal Bases

We know an orthonormal basis exists for Rn (the standard basis).

Fact
One can always …nd an orthonormal basis for a vector space.

Fact
If fv1 ; : : : ; vk g is an orthonormal basis for V then for all x 2 V ,
k
X k
X
x= i vi = (x vi ) vi
i =1 i =1
This follows from the properties on the previous slide (check it).
Dimension and Basis

De…nition
The dimension of a vector space X , denoted dim X , is the cardinality of any basis
of X .

Notation Reminder
For V X , jV j denotes the cardinality of the set V .

Fact
Mm n , the set of all m n real-valued matrices, is a vector space over R.
A basis is given by
1 if k = i and ` = j
fEij : 1 i m; 1 j ng where (Eij )k ` =
0 otherwise
The dimension of the vector space of m n matrices is mn.
Proving this is an exercise.
Dimension and Dependence
Theorem
Suppose dim X = n 2 N. If A X and jAj > n, then A is linearly dependent.

Proof.
If not, A is linearly independent and can be extended to a basis V of X :
A V ) jV j jAj > n
a contradiction

Intuitively, if A’s dimension is larger than the dimension of X there must be


some lineraly dependent elements in it.

Theorem
Suppose dim X = n 2 N, V X , and jV j = n.
If V is linearly independent, then V spans X , so V is a basis.
If V spans X , then V is linearly independent, so V is a basis.

Prove this as part of Problem Set 8.


Tomorrow

We illustrate the formal connection between linear functions and matrices. Then
we move to some useful geometry.
1 Linear Functions
2 Linear Functions and Matrices
3 Analytic Geometry in Rn

You might also like