Econ 2001
Econ 2001
Econ 2001
2015 August 19
Lecture 8 Outline
Therefore:
1 Eigenvalues solve the equation
det(A I) = 0:
2 Eigenvectors are non trivial solutions to the equation
Ax = x
Example
For a two by two matrix:
a11 a12
A I = ) det A = (a11 )(a22 ) a12 a21
2 2 2 2 a21 a22
Hence the characteristic equation is
2
(a11 + a22 ) + a11 a22 a12 a21 = 0
Which typically has two solutions.
Characteristic Polynomial
The characteristic equation f ( ) = det(A I) = 0 is a polynomial degree n.
By the Fundamental Theorem of Algebra, it has n roots (not necessarily
distinct and not necessarily real).
That is
f( )=( c1 )( c2 ) ( cn )
where c1 ; : : : ; cn 2 C (the set of complex numbers) and the ci ’s are not
necessarily distinct.
Notice that f ( ) = 0 if and only if 2 fc1 ; : : : ; cn g, so the roots are all the
solutions of the equation f ( ) = 0.
if = ci 2 R, there is a corresponding eigenvector in Rn .
if = ci 62 R, the corresponding eigenvectors are in Cn n Rn .
Another way to write the characteristic polynomial is
m1 mk
P( ) = ( r1 ) ( rk ) ;
where r1 ; r2 : : : ; rk are distinct roots (ri 6= rj when i 6= j) and mi are positive
integers summing to n.
mi is called the multiplicity of root ri .
Distinct Eigenvectors
FACT
The eigenvectors corresponding to distinct eigenvalues are distinct.
Proof By Contradiction.
Let 1 ; : : : ; k be distinct eigenvalues and x1 ; : : : ; xk the associated eigenvectors.
Suppose these vectors are linearly dependent (why is this a contradiction?).
WLOG, let the …rst k 1 vectors be linearly independent, while xk is a linear
combination of the others.
Pk 1
Thus, 9 i i = 1; : : : ; k 1 not all zero such that: i =1 i xi = xk
Multiply both sides by A and use the eigenvalue property (Ax = x):
k 1
X
i i xi = k xk
i =1
Multiply the …rst equation by k and subtract it from the second:
k 1
X
i( i k )xi = 0
i =1
De…nition
We say B is diagonalizable if we can …nd matrices P and D, with D diagonal, such
that
P 1 BP = D
If a matrix is diagonalizable
1 1 1
PP BPP = PDP
or
1
B = PDP
where D is a diagonal matrix.
For example, the value of x today depends linearly on its value yesterday:
xt = axt 1 8t = 0; 1; 2; 3; : : :
This is a fairly common relationship in time series data and macro.
Given some initial condition x0 , this equation is fairly easy to solve using
x1 = ax0
x2 = ax1 = a2 x0
‘recursion’: ...
xt 1 = axt 2 = at 1
x0
xt = axt 1 = at x0
Hence:
xt = at x0 8t = 0; 1; 2; 3; : : :
Di¤erence Equations Detour Continued
Consider now a two-dimensional linear di¤erence equation:
ct+1 b11 b12 ct
= 8t = 0; 1; 2; 3; : : :
kt+1 b21 b22 kt
given some initial condition c0 ; k0 .
ct b11 b12
Set yt = 8t and B = and rewrite this more
kt b21 b22
compactly as
yt+1 = Byt 8t = 0; 1; 2; 3; : : :
where bij 2 R each i; j.
We want to …nd a solution yt ; t = 1; 2; 3; : : : given the initial condition y0 .
Such a dynamical system can arise as a characterization of the solution to a
standard optimal growth problem (you will see this in macro).
This is hard to solve since the two variables interact with each other as time
goes on.
Things would be much easier if there were no interactions (b12 = 0 = b21 )
because in that case the two equations would evolve independently.
Di¤erence Equations Detour: The End
We want to solve
yt+1 = Byt 8t = 0; 1; 2; 3; : : :
Note
1
Premultiply by P and postmultiply by P , the theorem says:
1
P AP = D
De…nition
1
Two square matrices A and B are similar if A = P BP for some invertible matrix
P.
The theorem says that some square matrices are similar to diagonal matrices
that have eigenvalues on the diagonal.
Diagonalization Theorem: Idea of Proof
We want to show that for a given A there exist a matrix P and a diagonal matrix
D such that A = PDP 1 , where the diagonal entries of D are the eigenvalues of A
and the columns of P are the corresponding eigenvectors.
Idea of Proof (a real proof is way too di¢ cult for me)
Suppose is an eigenvalue of A and x is an eigenvector. Thus Ax = x.
If P is a matrix with column j equal to the eigenvector associated with i, it
follows that AP = PD.
The result would then follow if one could guarantee that P is invertible.
The proof works by showing that when A is symmetric, A only has real
eigenvalues, one can …nd n linearly independent eigenvectors even if the
eigenvalues are not distinct (these results use properties of complex numbers).
See a book for details.
A Few Computational Facts For You To Prove
Facts
De…nition
The trace ofPa square matrix A is given by the sum of its diagonal elements. That
n
is, tr (A) = i =1 aii :
Fact
n
X
tr (A) = i;
i =1
where i is the ith eigenvalue of A (eigenvalues counted with multiplicity).
Unitary Matrices
Remember
At is the transpose of A: the (i; j)th entry of At is the (j; i)th entry of A.
De…nition
An n n matrix A is unitary if At = A 1
.
REMARK
By de…nition every unitary matrix is invertible.
Unitary Matrices
Notation
A basis V = fv1 ; : : : ; vn g of Rn is orthonormal
1 if each basis element has unit length (vi vi = 1 8i), and
2 distinct basis elements are orthogonal (vi vj = 0 for i 6= j).
1 if i = j
Compactly, this can be written as vi vj = ij = .
0 if i =
6 j
Theorem
An n n matrix A is unitary if and only if the columns of A are orthonormal.
Proof.
Let vj denote the j th column of A.
At = A 1
() At A = I
() vi vj = ij
() fv1 ; : : : ; vn g is orthonormal
Symmetric Matrices Have Orthonormal Eigenvectors
Remember
A is symmetric if aij = aji for all i; j, where aij is the (i; j)th entry of A.
Theorem
If A is symmetric, then the eigenvalues of A are all real and there is orthonormal
basis V = fv1 ; : : : ; vn g of Rn consisting of the eigenvectors of A. In this case, P in
the diagonalization theorem is unitary and therefore:
A = PDPt
The proof is also beyond my ability (uses the linear algebra of complex vector
spaces).
Quadratic Forms
Let 0 1
( 11 1n
ij
2 if i < j B .. .. .. C
ij = ji
and A = @ . . . A
2 if i > j
n1 nn
Then,
f (x) = xt Ax
This inspires the idea of a quadratic form.
Quadratic Forms
De…nition
A quadratic form in n variables is a function Q : Rn ! R that can be written as
Q(x) = xt Ax
where A is a symmetric n n matrix.
Example
When n = 1 a quadratic form is a function of the form ax 2 .
Example
When n = 2 it is a function of the form
a11 x12 + 2a12 x1 x2 + a22 x22
(remember a12 = a21 by symmetry).
Example
When n = 3, it is a function of the form
a11 x12 + a22 x22 + a33 x32 + 2a12 x1 x2 + 2a13 x1 x3 + 2a23 x2 x3
Sign of A Quadratic Form
De…nition
A quadratic form Q(x) = xt Ax is
1 positive de…nite if Q(x) > 0 for all x 6= 0.
2 positive semi de…nite if Q(x) 0 for all x.
3 negative de…nite if Q(x) < 0 for all x 6= 0.
4 negative semi de…nite if Q(x) 0 for all x.
5 inde…nite if there exists x and y such that Q(x) > 0 > Q(y).
Idea
Think of positive and negative de…niteness as a way one applies to matrices the
idea of “positive” and “negative”.
n
Quadratic Forms and Diagonalization
Theorem
The quadratic form Q(x) = xt Ax is
1 positive de…nite if i > 0 for all i.
2 positive semi de…nite if i 0 for all i.
3 negative de…nite if i < 0 for all i.
4 negative semi de…nite if i 0 for all i.
5 inde…nite if there exists j and k such that j >0> k.
REMARK
We can check de…niteness of a quadratic form using the eigenvalues of A.
Principal Minors
De…nition
A principal submatrix of a square matrix A is the matrix obtained by deleting any
k rows and the corresponding k columns.
De…nition
The determinant of a principal submatrix is called the principal minor of A.
De…nition
The leading principal submatrix of order k of an n n matrix is obtained by
deleting the last n k rows and column of the matrix.
De…nition
The determinant of a leading principal submatrix is called the leading principal
minor of A.
Principal minors can be used in de…nitess tests.
Another De…niteness Test
Theorem
A matrix is
1 positive de…nite if and only if all its leading principal minors are positive.
2 negative de…nite if and only if its odd principal minors are negative and its
even principal minors are positive.
3 inde…nite if one of its kth order leading principal minors is negative for an even
k or if there are two odd leading principal minors that have di¤erent signs.
Theorem
Let V be a basis for a vector space X over R. Every vector x 2 X has a unique
representation as a linear combination of a …nite number of elements of V (with all
coe¢ cients nonzero).
This follows from the axiom of choice (did we talk about this?).
An equivalent result says that if a linearly independent set is not a basis, one
can always “add” to it to get a basis.
Theorem
If X is a vector space and V X is a linearly independent set, then V can be
extended to a basis for X . That is, there exists a linearly independent set W X
such that
V W spanW = X
There can be many bases for the same vector space, but they all have the
same number of elements.
Theorem
Any two Hamel bases of a vector space X have the same cardinality (are
numerically equivalent).
Standard Basis
De…nition
The standard basis for Rn consists of the set of N vectors ei , i = 1; : : : ; N, where
ei is the vector with component 1 in the ith position and zero in all other positions.
0 1 0 1 0 1 0 1
1 0 0 0
B0C B1C B0C B0C
B C B C B C B C
B C B C B C B C
e1 = B0C e2 = B0C en 1 = B ... C en = B ... C
B .. C B .. C B C B C
@.A @.A @1A @0A
0 0 0 1
Fact
One can always …nd an orthonormal basis for a vector space.
Fact
If fv1 ; : : : ; vk g is an orthonormal basis for V then for all x 2 V ,
k
X k
X
x= i vi = (x vi ) vi
i =1 i =1
This follows from the properties on the previous slide (check it).
Dimension and Basis
De…nition
The dimension of a vector space X , denoted dim X , is the cardinality of any basis
of X .
Notation Reminder
For V X , jV j denotes the cardinality of the set V .
Fact
Mm n , the set of all m n real-valued matrices, is a vector space over R.
A basis is given by
1 if k = i and ` = j
fEij : 1 i m; 1 j ng where (Eij )k ` =
0 otherwise
The dimension of the vector space of m n matrices is mn.
Proving this is an exercise.
Dimension and Dependence
Theorem
Suppose dim X = n 2 N. If A X and jAj > n, then A is linearly dependent.
Proof.
If not, A is linearly independent and can be extended to a basis V of X :
A V ) jV j jAj > n
a contradiction
Theorem
Suppose dim X = n 2 N, V X , and jV j = n.
If V is linearly independent, then V spans X , so V is a basis.
If V spans X , then V is linearly independent, so V is a basis.
We illustrate the formal connection between linear functions and matrices. Then
we move to some useful geometry.
1 Linear Functions
2 Linear Functions and Matrices
3 Analytic Geometry in Rn