Chapter 003
Chapter 003
1
Partial derivatives and the Jacobian
determinant
Matrix Recap: for systems of linear equation
2
But, how do we
where the symbol fm denotes the mth function (and not the function
raised to the mth power), we can derive a total mth partial
derivatives. Together they will give rise to the Jacobean matrix.
The Jacobian matrix is associated with a system of equations and is
the matrix composed of all first-order partial derivatives, arranged
in ordered sequence.
5
Notice that the elements of the ith row are partial
derivatives of the ith function with respect to each of the
independent variables.
The Jacobian determinant (often simply called the
Jacobian) is the determinant of the Jacobian matrix.
6
Example 1. Consider
are linearly dependent if and only if the determinant |A|=0. This result
can now be interpreted as a special application of the Jacobian criterion
of functional dependence. For systems of linear equations |A|=|J|.
8
Example 2. Test the functional dependency using
Jacobian Metrix.
Solution:
9
Solution of Non-linear Equation by Jacobian
Matrix Method
A system of linear equations can be solved using .
How about for system of non-linear equations?
To solve non-linear equations using matrix method, they must linearize
the equations. The coefficients of the linearized equations (specifically
the partial derivatives taken with respect to the unknowns variables)
will be gathered to form the Jacobian Matrix.
For example: For two non-linear equations F and G, its Jacobian Matrix
is
10
Generally, let be the matrix of unknown corrections be
the first order Jacobian matrix and be the matrix of
constants, then we can form the matrix equation:
Solution:
12
1st iteration:
= . Since the correction are positive we add them on the initial values of until the both are zero.
2nd iteration:
= Since values are zero, the solution converged. Iteration will now stop. And the values of are therefore: .
13
3.2 The Hessian Determinant (for unconstraint
function)
It is associated with a single equation of optimizing unconstrained
function.
Let be a real function of two real variables.
The FOC and SOC for extremum can be expressed in terms of
derivatives and/or total differential.
Derivative: FOC:
SOC: min) if:for a minimum
max) if: for a maximum
A convenient test for this second-order condition is the (plain)
Hessian. The Hessian |H| is a determinant composed of all the
second-order partial derivatives, with the second-order direct
partials on the principal diagonal and the second-order cross partials
off the principal diagonal. Thus,
14
Cross (mixed) partial derivatives measure the rate of change
of one first order partial derivative with respect to the other
variable.
The Young’s theorem states that cross partial derivatives are
identical with each other as long as the two cross partials are
continuous.
Let A be a symmetric n × n matrix. A minor of A of order k is
principal if it is obtained by deleting n − k rows and the n − k
columns with the same numbers. The leading principal minor
of A of order k is the minor of order k obtained by deleting the
last n − k rows and columns.
If the first element on the principal diagonal, the first principal
minor, is positive and the second principal minor, the second
order condition for minimum is met. If , the Hessian |H| is
called positive definite. A positive definite Hessian fulfils15 the
B. Minimum
All principal minors |Hi | > 0 (Positive definite)
18
Economic Application of Hessian Optimization
problem
1. A firm produces two goods sold in (choice variables) that maximize profit
two separate markets where the (objective function).
average revenues are given as: 1. Define objective function
19
2. Finding the turning linear case,
points
FOC=0:
You can see that the bordered Hessian HB has one more row and one
more column than the ordinary Hessian. Its borders in the last row
and last column, apart from the 0 element in the bottom right
position, are the first-order partial derivatives of g.
22
The second-order conditions for optimization of a Lagrangian with one
constraint require that for
Maximization
If there are two variables in the objective function (i.e. HB is 3×3) then
the determinant of the bordered principal minor = |HB| > 0. And Hessian
is negative definite
If there are three variables in the objective function (i.e. HB is 4×4) then
the determinant |HB| < 0 and the determinant of the naturally ordered
principal minor of |HB| > 0. (The naturally ordered principal minor is the
matrix remaining when the last row and column have been eliminated
from HB.)
Minimization
If there are two variables in the objective function (i.e. HB is 3×3) the
determinant of the bordered principal minor = |HB| < 0. And Hessian is
positive definite
If there are three variables in objective function then the determinant
23 |
HB| < 0 and the determinant of the naturally ordered principal minor of |
Constrained optimization with any number of variables and
constraints
Second-order conditions requirements for optimization for the general case
with n variables xi in the objective function and r constraints are that the
naturally ordered border preserving principal minors of dimension n of HB
must have the sign
‘Border preserving’ means not eliminating the borders added to the basic
Hessian, i.e. the last column and the bottom row, which typically show the
prices of the variables. 24
These requirements only apply to the principal minors of order ≥ (1 + 2r). For example, if
the problem was to maximize a utility function U = U(X1,X2,X3) subject to the budget
constraint M = P1X1 + P2X2 + P3X3 then, as there is only one constraint, r = 1. Therefore
we would just need to consider the principal minors of order greater than three since (1 +
2r) = (1 + 2) = 3.
As the full-bordered Hessian in this example with three variables is 4th order then only HB
itself plus the first principal minor need be considered, as this is the only principal minor
with order equal to or greater than 3.
For the full bordered Hessian n = 4 and so |HB| must have the sign = = = −1 < 0 and the
determinant of the 3rd order naturally ordered principal minor of |HB| must have the sign =
= = +1>0 These are the same as the basic rules for the three variable case stated earlier.
25
Numerical examples
1. Optimize
Subject to:
L(w; x; y; z; λ; µ) = f + λg + µh
26
3.3 Eigenvalues and Eigenvectors
Almost all vectors change direction, when they are multiplied by A. Certain
exceptional vectors x are in the same direction as Ax. Those are the
“eigenvectors”. Multiply an eigenvector by A, and the vector Ax is a number
27
The eigenvalue could be zero! Then means that this eigenvector is in the nullspace.
For a given matrix A of order n, what are the vectors x that satisfy the equation for some
scalar ?
The vector x = 0 (that is, the vector whose elements are all zero) satisfies this equation.
With such a trivial answer, we might ask the question again in another way: For a given
matrix A, what are the nonzero vectors x that satisfy the equation Ax = λx for some scalar
λ?
To answer this question, we first perform some algebraic manipulations upon the equation
Ax = λx. We note first that, if I = In, then we can write
Ax = λx ⇔ Ax − λx = 0
⇔ Ax − λIx = 0
⇔ (A − λI) x = 0.
28
Remember that we are looking for nonzero x that satisfy this last equation. But A −
λI is an n × n matrix and, should its determinant be nonzero, this last equation will
have exactly one solution (i.e., trivial), namely x = 0.
Thus our question above has the following answer: The equation Ax = λx has
nonzero (or nontrivial) solutions for the vector x if and only if the matrix A − λI has
zero determinant (is singular or not invertible). .
For a given matrix A there are only a few special values of the scalar λ for which A −
λI will have zero determinant, and these special values are called the eigenvalues of
the matrix A. We do allow for the possibility that λ=0.
29
A scalar λ is an eigenvalue of an n × n matrix A if and only if λ
satisfies the characteristic equation det (A − λI) = 0. It can be shown
that if A is an n × n matrix, then det(A − λI) is a polynomial in the
variable λ of degree n. We call this polynomial the characteristic
polynomial of A.
32
Method of Finding Eigenvalues and Eigenvectors
34
•So the augmented matrix of the system is ,
35
Thus, the eigenvector corresponding to λ1 = 1 is . This tells
us that the eigenvectors corresponding to the eigenvalue 1
are precisely the set of scalar multiples of the vector .
When λ2 = 6, the eigenvector is given by the system
36
E.g. 2 Find the eigenvalue and eigenvectors of
When λ = −2, x1 + 7x2 + x3 = 0 → x1=-7x2-x3→x1=-x3
-20x2 = 0→x2=0
Here we have three unknown with two equations. In
principle, we’re finished with the problem in the sense that
we have the solution in hand. But it’s customary to rewrite
the solution in vector form so that its properties are more
evident. Let x3=t, then x1=-t, x2=0. Hence, eigenvector
corresponding to λ1 = −2 is
37
When λ = 3, x1 + 2x2 + x3 = 0
5x2 + 5x3 = 0
Therefore, the eigenvector corresponding to λ2 = 3 is
When λ = 6, x1 + 2x2 + x3 = 0
5x2 + 5x3 = 0
Therefore, the eigenvector corresponding to λ3 = 6 is
If a reduced system with n equations in m unknowns,
where n < m, has p entirely zeros rows in the reduced
augmented matrix, then there are m − n + p number of
free variables in the solution set.
38
E.g. 3 Find the eigenvalues and associated eigenvectors of the matrix
39
Properties of Eigenvalues
A square matrix A and its transpose have the same
eigenvalues.
If |A|≠0, then all eigenvalues are nonzero.
The sum of the eigenvalues of a matrix is equal to
the sum of the principal diagonal elements of A.
The product of the eigenvalues of a matrix A is equal
to |A|.
If λ1, λ2, . . . , λn are the eigenvalues of a matrix A,
then
a. kλ1, kλ2. . . kλn are the eigenvalue of the matrix
kA. 40
The eigenvalues of a triangular matrix are the entries of the main
diagonal. A triangular matrix has the property that either all of its
entries below the main diagonal are 0 or all of its entries above the
main diagonal are 0.
Ex. , λ = 0, 2, or 3
41
3.4 Quadratic Forms
The FOC and SOC for extremum can be expressed in terms of
derivatives and/or total differential.
Derivative: FOC:
SOC: for a minimum
for a maximum
Total differential: FOC:
SOC:
42
, where
exemplifies what are known as quadratic forms, for which there exist
established criteria for determining whether their signs are always positive,
negative, non-positive or non-negative for any values of dx and dy, not both
zero.
Since second order condition for extremum pivots directly on the sign of
those criteria are of direct interest.
43
The special case of polynomials where each term has a
uniform degree (i.e., where the sum of the exponents in
each term is the same) is called a form.
E.g. 4x-9y+z is a linear form while is a quadratic form.
Given , find dz and d2z.
The device that will do the trick of making u and v appear only in
some squares is that of completing the square.
46
We can now predicate the sign of q entirely on the
values of the coefficients a, b and h as follows:
The determinant |a|=a is the first leading principal minor of |D| and the
determinant is the second leading principal minor of |D| and their signs
will serve to determine the positive or negative definiteness of q.
We can test for the definiteness of the matrix in the following fashion:
1. A is positive definite iff all of its n leading principal minors are strictly
positive.
2. A is negative definite iff all of its n leading principal minors alternate
in sign.
3. If some kth order leading principal minor of A is nonzero but does not
fit either of the above sign patterns, then A is indefinite.
48
If the matrix A would meet the criterion for positive or negative
definiteness if we relaxed the strict inequalities to weak inequalities
(i.e. we allow zero to fit into the pattern), then although the matrix
is not positive or negative definite, it may be positive or negative
semi-definite.
In this case, we employ the following tests:
1. A is positive semi-definite iff every principal minor of A is ≥ 0.
2. A is negative semi-definite iff every principal minor of A of odd
order is ≤ 0 and every principal minor of even order is ≥ 0.
Notice that for determining semi-definiteness, we can no longer
check just the leading principal minors, but we must check all
principal minors. What a pain!
The cases of positive and negative definiteness of are related to
the second order sufficient condition for a minimum and a
maximum respectively. The cases of semi-definiteness, on the 49other
hand, relate to second order necessary conditions. When is
E.g. 1 Is either positive or negative definite? The
discriminant of q is , with principal minors 5>0, and
=7.75>0. Therefore, q is positive definite.
E.g. 2 Given fxx= -2, fxy=1, and fyy= -1at a certain
point on function z=f(x,y) does d2z have a definite
sign at that point regardless of the values of dx and
dy? The discriminant of the quadratic form d2z is ,
with principal minors -2<0, and =1>0. Thus, d2z is
negative definite.
50
E.g. 3 Determine whether is either positive or negative definite.
The discriminant of q is , with principal minors 1>0, =5>0 and
=11>0. Thus, the quadratic form is positive definite.
51
Determine the definiteness of the following
matrices:
1.
2.
3.
4.
5.
6.
52
• Express the quadratic form as a matrix product involving a symmetric
coefficient matrix.
54