0% found this document useful (0 votes)
164 views17 pages

Unconstrained Optimization Methods

This document introduces direct search methods for unconstrained optimization. It discusses that optimization methods typically handle constraints by converting problems to unconstrained problems. For unconstrained problems, a search direction is determined at each iteration and the best solution in that direction is found using line search. Common direct search methods described include coordinate descent, Hooke and Jeeves, and Powell's method. Powell's method is noted as the most efficient pattern search method.

Uploaded by

Saheera Hazarika
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
164 views17 pages

Unconstrained Optimization Methods

This document introduces direct search methods for unconstrained optimization. It discusses that optimization methods typically handle constraints by converting problems to unconstrained problems. For unconstrained problems, a search direction is determined at each iteration and the best solution in that direction is found using line search. Common direct search methods described include coordinate descent, Hooke and Jeeves, and Powell's method. Powell's method is noted as the most efficient pattern search method.

Uploaded by

Saheera Hazarika
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd

Introduction to

unconstrained optimization
- direct search methods

Jussi Hakanen
Post-doctoral researcher
[Link]@[Link]

spring 2014 TIES483 Nonlinear optimization


Structure of optimization methods
Typically
– Constraint handling Constraint handling
converts the problem to (a method
series of) unconstrained Unconstrained
problems optimization
– In unconstrained
optimization a search Line
search
direction is determined at
each iteration
– The best solution in the
search direction is found
with line search

spring 2014 TIES483 Nonlinear optimization


Group discussion
1. What kind of optimality conditions there exist for
unconstrained optimization (𝑥 ∈ 𝑅𝑛 )?
2. List methods for unconstrained optimization?
– what are their general ideas?
Discuss in small groups (3-4) for 15-20 minutes
Each group has a secretary who writes down the
answers of the group
At the end, we summarize what each group found

spring 2014 TIES483 Nonlinear optimization


Reminder: gradient and hessian
Definition: If function 𝑓: 𝑅𝑛 → 𝑅 is differentiable, then the
𝜕𝑓(𝑥)
gradient 𝛻𝑓(𝑥) consists of the partial derivatives i.e.
𝜕𝑥𝑖
𝜕𝑓(𝑥) 𝜕𝑓(𝑥) 𝑇
𝛻𝑓 𝑥 = ,…,
𝜕𝑥1 𝜕𝑥𝑛
Definition: If 𝑓2 is twice differentiable, then the matrix
𝜕 𝑓(𝑥) 𝜕2 𝑓(𝑥)

𝜕𝑥1 𝜕𝑥1 𝜕𝑥1 𝜕𝑥𝑛
𝐻 𝑥 = ⋮ ⋱ ⋮
𝜕2 𝑓(𝑥) 𝜕2 𝑓(𝑥)

𝜕𝑥𝑛 𝜕𝑥1 𝜕𝑥𝑛 𝜕𝑥𝑛
is called the Hessian (matrix) of 𝑓 at 𝑥
Result: If 𝑓 is twice continuously differentiable, then
𝜕2 𝑓(𝑥) 𝜕2 𝑓(𝑥)
=
𝜕𝑥𝑖 𝜕𝑥𝑗 𝜕𝑥𝑗 𝜕𝑥𝑖

spring 2014 TIES483 Nonlinear optimization


Reminder: Definite Matrices
Definition: A symmetric 𝑛 × 𝑛 matrix 𝐻 is positive
semidefinite if ∀ 𝑥 ∈ ℝ𝑛
𝑥 𝑇 𝐻𝑥 ≥ 0.
Definition: A symmetric 𝑛 × 𝑛 matrix 𝐻 is positive
definite if
𝑥 𝑇 𝐻𝑥 > 0 ∀ 0 ≠ 𝑥 ∈ ℝ𝑛
Note: If ≥ → ≤ (> → <), then 𝐻 is negative semidefinite
(definite). If 𝐻 is neither positive nor negative
semidefinite, then it is indefinite.
Result: Let ∅ ≠ 𝑆 ⊂ 𝑅𝑛 be open convex set and
𝑓: 𝑆 → 𝑅 twice differentiable in 𝑆. Function 𝑓 is convex if
and only if 𝐻(𝑥 ∗ ) is positive semidefinite for all 𝑥 ∗ ∈ 𝑆.
Unconstraint problem
min 𝑓 𝑥 , 𝑠. 𝑡. 𝑥 ∈ 𝑅𝑛

Necessary conditions: Let 𝑓 be twice differentiable in 𝑥 ∗ . If


𝑥 ∗ is a local minimizer, then
– 𝛻𝑓 𝑥 ∗ = 0 (that is, 𝑥 ∗ is a critical point of 𝑓) and
– 𝐻 𝑥 ∗ is positive semidefinite.
Sufficient conditions: Let 𝑓 be twice differentiable in 𝑥 ∗ . If
– 𝛻𝑓 𝑥 ∗ = 0 and
– 𝐻(𝑥 ∗ ) is positive definite,
then 𝑥 ∗ is a strict local minimizer.
Result: Let 𝑓: 𝑅𝑛 → 𝑅 is twice differentiable in 𝑥 ∗ . If
𝛻𝑓 𝑥 ∗ = 0 and 𝐻(𝑥 ∗ ) is indefinite, then 𝑥 ∗ is a saddle point.
Unconstraint problem

Adopted from Prof. L.T. Biegler (Carnegie Mellon


University)
Descent direction
Definition: Let 𝑓: 𝑅𝑛 → 𝑅. A vector 𝑑 ∈ 𝑅𝑛 is a descent
direction for 𝑓 in 𝑥 ∗ ∈ 𝑅𝑛 if ∃ 𝛿 > 0 s.t.
𝑓 𝑥 ∗ + 𝜆𝑑 < 𝑓(𝑥 ∗ ) ∀ 𝜆 ∈ (0, 𝛿].
Result: Let 𝑓: 𝑅𝑛 → 𝑅 be differentiable in 𝑥 ∗ . If ∃𝑑 ∈ 𝑅𝑛
s.t. 𝛻𝑓 𝑥 ∗ 𝑇 𝑑 < 0 then 𝑑 is a descent direction for 𝑓 in
𝑥∗.

spring 2014 TIES483 Nonlinear optimization


Model algorithm for unconstrained
minimization
Let 𝑥 ℎ be the current estimate for 𝑥 ∗
1) [Test for convergence.] If conditions are satisfied,
stop. The solution is 𝑥 ℎ .
2) [Compute a search direction.] Compute a non-
zero vector 𝑑 ℎ ∈ 𝑅𝑛 which is the search direction.
3) [Compute a step length.] Compute 𝛼ℎ > 0, the
step length, for which it holds that
𝑓 𝑥 ℎ + 𝛼ℎ 𝑑 ℎ < 𝑓(𝑥 ℎ ).
4) [Update the estimate for minimum.] Set
𝑥 ℎ+1 = 𝑥 ℎ + 𝛼ℎ 𝑑 ℎ , ℎ = ℎ + 1 and go to step 1.
From Gill et al., Practical Optimization, 1981, Academic Press
spring 2014 TIES483 Nonlinear optimization
On convergence
Iterative method: a sequence {𝑥 ℎ } s.t. 𝑥 ℎ → 𝑥 ∗
when ℎ → ∞
Definition: A method converges
– linearly if ∃𝛼 ∈ [0,1) and 𝑀 ≥ 0 s.t. ∀ ℎ ≥ 𝑀
𝑥 ℎ+1 − 𝑥 ∗ ≤ 𝛼 𝑥 ℎ − 𝑥 ∗ ,
– superlinearly if ∃𝑀 ≥ 0 and for some sequence 𝛼ℎ → 0 it
holds that ∀ℎ ≥ 𝑀
𝑥 ℎ+1 − 𝑥 ∗ ≤ 𝛼ℎ 𝑥 ℎ − 𝑥 ∗ ,
– with degree 𝑝 if ∃𝛼 ≥ 0, 𝑝 >𝑝
0 and 𝑀 ≥ 0 s.t. ∀ℎ ≥ 𝑀
𝑥 ℎ+1 − 𝑥 ∗ ≤ 𝛼 𝑥 ℎ − 𝑥 ∗ .
If 𝑝 = 2 (𝑝 = 3), the convergence is quadratic
(cubic).

spring 2014 TIES483 Nonlinear optimization


Summary of group discussion for
methods
1. Newton’s method
1. Utilizes tangent
2. Golden section method
1. For line search
3. Downhill Simplex
4. Cyclic coordinate method
1. One coordinate at a time
5. Polytopy search (Nelder-Mead)
1. Idea based on geometry
6. Gradient descent (steepest descent)
1. Based on gradient information

spring 2014 TIES483 Nonlinear optimization


Direct search methods
Univariate search, coordinate descent, cyclic
coordinate search
Hooke and Jeeves
Powell’s method

spring 2014 TIES483 Nonlinear optimization


Coordinate descent

From Miettinen: Nonlinear optimization, 2007 (in Finnish)


𝑓 𝑥 = 2𝑥12 + 2𝑥1 𝑥2 + 𝑥22 + 𝑥1 − 𝑥2

spring 2014 TIES483 Nonlinear optimization


Idea of pattern search

From Miettinen: Nonlinear optimization, 2007 (in Finnish)


spring 2014 TIES483 Nonlinear optimization
Hooke and Jeeves

From Miettinen: Nonlinear optimization, 2007 (in Finnish)


𝑓 𝑥 = 𝑥1 − 2 4 + x1 − 2x2 2

spring 2014 TIES483 Nonlinear optimization


Hooke and Jeeves with fixed step length

From Miettinen: Nonlinear optimization, 2007 (in Finnish)


𝑓 𝑥 = 𝑥1 − 2 4 + x1 − 2x2 2

spring 2014 TIES483 Nonlinear optimization


Powell’s method
Most efficient pattern search method
Differs from Hooke and Jeeves so that for
each pattern search step one of the coordinate
directions is replaced with previous pattern
search direction.

spring 2014 TIES483 Nonlinear optimization

You might also like