Dynamics and Control of Trajectory Tubes, Alexander B. Kurzhanski Pravin Varaiya
Dynamics and Control of Trajectory Tubes, Alexander B. Kurzhanski Pravin Varaiya
Alexander B. Kurzhanski
Pravin Varaiya
Dynamics and
Control of
Trajectory Tubes
Theory and Computation
Systems & Control: Foundations & Applications
Series Editor
Tamer Başar, University of Illinois at Urbana-Champaign, Urbana, IL,
USA
Editorial Board
Karl Johan Åström, Lund University of Technology, Lund, Sweden
Han-Fu Chen, Academia Sinica, Beijing, China
Bill Helton, University of California, San Diego, CA, USA
Alberto Isidori, Sapienza University of Rome, Rome, Italy
Miroslav Krstic, University of California, San Diego, CA, USA
H. Vincent Poor, Princeton University, Princeton, NJ, USA
Mete Soner, ETH Zürich, Zürich, Switzerland;
Swiss Finance Institute, Zürich, Switzerland
Roberto Tempo, CNR-IEIIT, Politecnico di Torino, Italy
v
vi Preface
type. However, impulse inputs are ideal elements and their physical realization by
“ordinary” bounded functions may be achieved through approximation by double-
bounded controls which are functions whose bound tends to infinity.
Chapter 7 is concerned with dynamics and control of systems under state
constraints. Previously discussed problems are now subject to additional “viability”
restrictions on the state variables. The description covers solution approaches to
problems of reachability (forward and backward) and emphasizes specifics of
related mathematical techniques including ellipsoidal approximations. The case
of linear systems with convex hard bounds on controls and state coordinates are
worked out in detail.
The contents of Chaps. 1–7 indicate that one of the main items in treating
considered problems are trajectory tubes and the means of their calculation. The
same is true for the rest of the chapters. So Chap. 8 begins with fundamentals of
a general vision—the theory of trajectory tubes with models of their evolutionary
dynamics. Indicated results, together with considerations of the previous chapter,
are further applied to closed-loop control under state constraints, with techniques
borrowed from both Hamiltonian approach and duality theory of nonlinear analysis.
This brings forward the discussion to complex state constraints in the form of
obstacle problems, wherein constrained trajectories must simultaneously lie within
one set and outside another.
The next two chapters consider uncertainty, which is inherent in realistic
problems of control. These chapters may serve as an introduction to a more thorough
description of uncertainty.
Chapter 9 is a concise explanation of the theory of guaranteed state estimation,
also known as the set-membership bounding approach to external disturbances in
estimation models. In contrast with conventional descriptions, the present exposition
involves Hamiltonian methods and is applicable to nonlinear systems. Dynamic
estimation of system trajectories under unknown but bounded errors is also formu-
lated as a problem with state constraints, which now are not known in advance, but
arrive online, in real time. In the linear case the proposed deterministic “filtering”
equations demonstrate connections and differences when compared with stochastic
(Kalman) filtering. Both continuous and discrete measurements are considered.
The results of Chap. 9 have a natural application to problems of output feedback
control under unknown but bounded disturbances in system and measurement
inputs. These problems are addressed in Chap. 10. The solutions introduced there
are based on the notions of generalized state and information tubes which describe
the overall system dynamics. For the linear case with convex constraints on controls
and uncertainties items, computation schemes based on ellipsoidal approximations
are presented. Several examples are worked out.
Finally, Chap. 11 is confined to verification problems and hybrid systems, with
a description of exact solutions and ellipsoidal schemes for their computation. The
discussion is accompanied by some examples. Special attention is given to possible
involvement of impulse inputs in formal mathematical models of hybrid systems.
The aim of this chapter is to emphasize the applicability of techniques presented in
this book to the investigation of hybrid systems.
Preface ix
References to prior literature and research results related to this book are given
in the introduction to each chapter.
The mathematical level of the book presumes reasonable knowledge of advanced
calculus, linear algebra and differential equations, as well as basics of variational
analysis with optimization methods and computational mathematics.
Throughout the previous years we had the pleasant opportunity for useful
discussions on the topics arranged in this book with K. Åström, J.-P. Aubin,
J. Baras, T. Başar, F. Chernousko, P. Kokotovic, A. Krener, A.A. Kurzhanskiy,
Yu. Ledyayev, G. Leitmann, A. Lindquist, J. Lygeros, M. Milanese, I. Mitchell,
S. Mitter, A. Rantzer, J. Sousa, C. Tomlin, I. Vályi, and V. Veliov. Their valuable
comments helped to crystallize the contents.
We thank A. Daryin, M. Gusev, T. Filippova, I. Rublev, and P. Tochilin for
reading parts of the manuscript, their useful comments, and contributed illustrations.
Our thanks surely goes to the authors of illustrations for the examples of this
book. Their names are indicated throughout the text.
The authors are grateful to the US national Science Foundation, the first author is
also grateful to the Russian Foundation for Basic Research for the support of their
work.
xi
xii Contents
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 431
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 443
Notations
h.P; Q/ D maxfhC .P; Q/; hC .Q; P /:g—the Hausdorff distance between sets
P; Q
xvii
xviii Notations
N
I.ljB/—indicator function of set B: I.ljB/ D 0 if l 2 B, I.ljB/ D 1 if x 2B
Lm
p ŒT —the space of m-dimensional vector functions f W T ! R integrable with
m
power p
Abstract This chapter gives an exposition of control theory for linear systems
with emphasis on items and techniques given in a form appropriate for topics in
forthcoming chapters. It introduces problems of reachability and optimal target
control under constraints, as well as time-optimal control. Indicated are solution
approaches to open-loop control that involve the moment problem, the maximum
principle, and the duality methods of convex analysis.
P / D f .t; x; u/;
x.t (1.1)
in which t is time, the vector x 2 Rn is the state, and the vector u 2 Rp is the
control. The function f .t; x; u/ is defined in a domain
D D T D P ; t 2 T; x 2 D ; u 2 P ;
@
G.t; s/ D A.t /G.t; s/; G.s; s/ D I (1.4)
@t
1
We use the terms measurable and integrable for Lebesgue-measurable and Lebesgue-integrable
functions.
1.1 The Controlled System 3
@
G.t; s/ D G.t; s/A.s/; G.t; t / D I; (1.5)
@s
in s.
G.t; s/ is nonsingular for all t; s and has the following properties:
An integral formula for the adjoint equation, similar to (1.3), is given in (1.80).
When A.t / A is constant, G.t; s/ D exp.A.t s//; in which exp A is the
matrix exponential.
Formula (1.3) may be checked by direct substitution. The existence of a solution
to system (1.2) is thus a standard property of linear differential equations (see [38,
105, 120, 274]).
Exercise 1.1.1. Check formula (1.3).
System (1.2) can be put into a simpler form. Transforming the state by
with
Thus there is a one-to-one correspondence (1.6) between the solutions x.t / and
z.t / to Eqs. (1.2) and (1.7), respectively. Their initial values are related through (1.8).
The state z.t / satisfies a particularly simple version of (1.2),
with
Thus we may consider system (1.9) rather than (1.2). In other words, in the notation
of (1.2) we may take A.t / 0 with no loss of generality. Note however that the
matrix function B.t / is now time-variant (even when (1.2) is time-invariant).
4 1 Linear Control Systems
One should realize, however, that the transformation (1.6) allows us to take
A.t / 0 only within the time range ft t0 g, a different t0 leads to a different
transformation (1.6). A similar result is obtained by the substitution
With this choice of state the original system again gives A.t / 0, but this is correct
only for ft t1 g.
We shall sometimes make use of these transformations to demonstrate some
basic techniques with a simpler notation. The reader will always be able to return to
A.t / 6 0 as an exercise.
A shrewd reader may have now realized that there should exist a transformation
of the state that takes a given linear homogeneous equation
zP D A0 .t /z; (1.11)
for any preassigned matrix A0 .t / with the same initial condition, x.t0 / D z.t0 /.
Exercise 1.1.2. Indicate the transformation that converts Eq. (1.10) into (1.11) and
vice versa.
Among the equations of interest are linear matrix differential equations of the
form
@ @
G.t; s/ D A.t /G.t; s/; G.s; s/ D I I G1 .t; s/ D A1 .t /G1 .t; s/;
@t @t
G1 .s; s/ D I: (1.13)
with A.t; u/ D A.t /U C D.t /, with matrix of controls U 2 Rnn and continuous
matrices of coefficients A.t /; D.t / 2 Rnn . In a simple case we may have A.t; u/ D
A.t /u C D.t / with a scalar u.
One objective of this book is to develop solutions for systems subject to constraints
on the control and state. The values u of the control may be restricted by various
types of bounds. Here are some typical types.
u 2 P .t /; (1.18)
Hard bounds may also be specified as the level set (at level ) of a function
®.t; u/:
U .t / D fu W ®.t; u/ g:
Suppose ®.t; u/ is continuous in t; u and convex in u. Then its level set P .t / will
be closed and convex. As indicated in [237], for U .t / 6D ; to be bounded, hence
compact, for any , it is necessary and sufficient that
0 2 int Dom® .t; / : (1.19)
Here
is the effective domain of the function ® .t; l/ for fixed t , see [237] (interiorQ
denotes the collection of all interior points of a set Q). Under condition (1.19) the
level sets P .t / will be bounded, ®.t; u/ ! 1 as hu; ui ! 1; and in the definition
of ® .t; l/ the operation of “sup” may be replaced by “max,” see [72]. Under the
given conditions the convex compact set-valued function P .t / will be Hausdorff-
continuous.
In practice, hard bounds represent limits imposed by equipment or considerations
of safety on a control variable such as voltage, force, or torque. Linear control syn-
thesis techniques, like those based on frequency domain methods, which formulate
the control as a linear function of the state, cannot satisfy hard bounds.
Integral bounds, also known as soft bounds, are constraints of the form
Z t1
jju.t /jjq dt ; > 0; q 1; (1.20)
t0
in which jjujj is a norm and q is usually an integer. A more general integral bound
is expressed as
Z t1
®.t; u.t //dt ; (1.21)
t0
in which ®.t; u/ is continuous in t; u and convex in u, with int Dom® .t; /
6D ;.
1.2 Control and State Constraints: Open-Loop and Closed-Loop Control 7
Z t1 X
k
kd i u.t /=dt i k dt ; > 0: (1.23)
t0 iD0
z.t / 2 Z .t /; t 2 Tz ;
with ®.t; x/ being convex in x; or mixed integral bounds on both control and state
of the form
Z t1
®.t; x; u/dt ; (1.27)
t0
In addition to constraints, one needs to specify the class of functions from which the
controls are to be selected. The appropriate class of functions should be considered
from both the mathematical and control design points of view.
Mathematically, we will consider classes that are functions of time only, u D
u.t /, and those that are functions of both time and state, u D u.t; x/. The two
classes are called open-loop and closed-loop controls, respectively.
An open-loop control, u D u.t /, is required to be measurable. For each open-loop
control, the solution to (1.1) reduces to solving the equation
which, in the linear case, is immediate through the integral formula (1.3). Thus,
substituting an open-loop control u.t / in (1.1) or (1.2) gives a unique solution xŒt D
x.t; t0 ; x 0 / for any starting position ft0 ; x 0 g; x.t0 / D x 0 . From the point of view of
control design, taking u D u.t / as a function of time only means that the selection
of the control action for each time t is fixed, implanted in the system design, and
cannot be changed throughout the process, which lasts from t0 to t1 : The class of
open-loop controls is denoted by UO .
A closed-loop control, u D u.t; x/, depends on both t and x. The function u.t; x/
may be nonlinear, even discontinuous, in x. The class of admissible closed-loop
controls u.t; x/ must however be so restricted that nonlinear differential equations
of type
or
would have a solution in some reasonable sense. That is, there should be an existence
theorem ensuring that Eq. (1.28) can be solved.
From the point of view of design and implementation, a closed-loop control
selects the control action u.t; x/ at each time t depending on the value of the state
x.t /. For on-line control (that is, when the value of the control action u.t; x.t //
is calculated in real time), the design presupposes that the state x.t / is exactly
measured for all t and continuously communicated to the control device. A control
of this form leads to a feedback loop (also “closed loop”), hence the name closed-
loop or feedback control. They are also referred to as synthesized controls in
contrast with controls of class UO which are known as control programs or open-
loop controls. The class of closed-loop controls to be used, complemented with an
existence theorem, is denoted by UC .
If the disturbance term v.t / is fixed, there is no essential difference between UO
and UC , as the following exercise indicates.
Exercise 1.2.1. Suppose the disturbance v.t / in (1.28) is fixed. Let xŒt D
x.t; t0 ; x 0 / be a given solution of (1.28) under a closed-loop control g.t; x/, i.e.,
Show that there is a corresponding open-loop control u.t / such that xŒt is also the
solution of
Show also that the corresponding open-loop control depends on the disturbance, i.e.,
if the disturbance input v.t / changes, so does the open-loop control in order to keep
the same solution xŒt .
Thus the difference in system performance achieved by using open-loop and closed-
loop controls becomes apparent for systems subject to unknown disturbances. The
control of systems with unknown disturbances is mentioned in Chap. 10.
In practice, exact measurement of the state vector x may not be possible. The
measurement may be incomplete because only a part of the vector x or a function of
x can be measured, or because the measurements can be taken only at discrete time
instants, or because the measurements may be corrupted by unknown disturbances
(“noise”). Consideration of such situations will bring us to new classes of feedback
control.
In many cases treated later we allow the closed-loop control u D U .t; x/
to be a set-valued map, with values U .t; x/ 2 compRp , measurable in t and
10 1 Linear Control Systems
upper-semicontinuous in x [7, 8, 48, 238]. Upon substituting u D U .t; x/, Eq. (1.2)
then becomes a nonlinear differential inclusion,
whose solution exists and in general is set-valued [75, 76]. The class of such set-
valued controls is denoted by UCS .
In the next section we consider several problems of optimal control within the
class of open-loop controls UO . These are the “norm-minimal” controls that can be
obtained through simple considerations.
The moment problem is a simple way of presenting the two-point boundary problem
of control. It naturally allows to present the optimality criterion as one of minimizing
the norm of the control u.t / in an appropriate functional space.
Consider system (1.2) on a finite time interval T D Œt; ª. We start with one of
the simplest problems of optimal control.
Problem 1.3.1. Given system (1.2) and two points, xŒt D x, xŒª D x .1/ , find the
optimal control that moves the system trajectory xŒs D x.s; t; x/ from x to x .1/
with minimum cost
Z ª
.p/
J .t; x/ D min hu.s/; N.s/u.s/ids j u./ 2 L2 .T / ; (1.30)
t
in which the continuous p p matrix N.s/ is positive definite, N.s/ D N 0 .s/ > 0.
We shall solve this problem applying basic Hilbert space techniques. Using
formula (1.3) the boundary constraints xŒt D x; xŒª D x .1/ may be rewritten
in the form
Z ª Z ª
G.ª; s/B.s/u.s/ds D x .1/
G.ª; t /x G.ª; s/v.s/ds D c: (1.31)
t t
1.3 Optimal Control with Norm-Minimal Cost: The Controllability Property 11
.p/
to be an inner product in L2 .T /—the space of square-integrable Rp -valued
.p/
functions. Consider both d .i/ ./ and u./ as elements of the Hilbert space L2 .T /
space, with norm
Z ª 1=2
ku./kN D hu.s/; N.s/u.s/ids :
t
The cost of a control u is the square of its norm and Problem 1.3.1 is to find
the element of a Hilbert space with minimum norm that satisfies certain linear
constraints.2 We can thus reformulate Problem 1.3.1.
Problem 1.3.1-A: Minimize hu./; u./iN under the constraints
We assume that the n functions d .i/ ./ are linearly independent, which means
that for any n-dimensional row-vector œ, œD.t / D 0, a.e. for t 2 T , only if œ D 0.
Now consider controls of the form u.s/ D N 1 .s/D 0 .s/l; for some l 2 Rn , and
substitute in (1.32), to get
Z ª
W .ª; t /l D c; W .ª; t / D D.s/N 1 .s/D 0 .s/ds:
t
The Gramian or Gram matrix W .ª; t / is nonsingular if and only if the functions
d .i/ are linearly independent [82]. As we shall observe later in this section the
determinant jW .ª; t /j 6D 0 (for any matrix N.s/ D N 0 .s/ > 0) ensures that
system (1.2) is controllable (see below, Lemma 1.3.1).
The particular control
satisfies the constraint (1.32). We now show that u0 .s/ minimizes the cost (1.30).
.p/
2
We could also regard functions d .i / .s/ as elements of Lq ŒT , q 1. Later, while dealing
with impulse controls, it will be more convenient to treat functions d .i / .s/ as elements of the
space C .p/ ŒT .
12 1 Linear Control Systems
Indeed, suppose u.s/ satisfies (1.32), but u.s/ 6D u0 .s/. Then ue .s/ D u.s/
u .s/ 6D 0 satisfies
0
hN 1 ./D 0 ./; ue ./iN D hN 1 ./D 0 ./; N./ue ./i D hD 0 ./; u.s/ u0 .s/i D 0;
and
since hu0 ./; N./ue ./i 0 due to (1.33). This gives the next result.
Theorem 1.3.1. The optimal control for Problem 1.3.1 is given by the continuous
function
Remark 1.3.1. If x .1/ D 0; v.s/ 0, then c D G.ª; t /x, so that the optimal
control
Problem 1.3.2. Given system (1.2), starting position xŒt D x, and convex compact
terminal set M and terminal time ª, find the optimal control that moves the system
trajectory xŒs D x.s; t; x/ from x to some point xŒª 2 M with minimum cost
Again take D.s/ D G.ª; s/B.s/. We treat its rows—the p-dimensional functions
.p/
d .i/ ./—as elements of L1 D L1 ŒT and treat the control u./ as an element of its
.p/
conjugate space L1 D L1 ŒT . Denote the bilinear functional
Z ª
hu./; d./i D hd.s/; u.s/ids
t
and norms
Z ª
kd./kL1 D kd.s/kds; ku./kL1 D ess supfku.s/k j s 2 T g:
t
and find the smallest D 0 for which there exists u./ 2 B1 that satisfies (1.36).
Since M is convex and compact, x.ª/ 2 M is equivalent to the system of
inequalities
in which
Z ª
c D G.ª; t /x C G.ª; s/v.s/ds:
t
Hence, for some u./ 2 B1 to satisfy (1.36), it is necessary and sufficient that
Z ª
min l 0 D.s/u.s/ds j u./ 2 B1 C hl; c i ¡.l j M /; 8l 2 Rn ;
u t
14 1 Linear Control Systems
or
Z ª
minfl 0 D.s/u.s/ j ku.s/k 1gds C hl; c i ¡.l j M /g; 8l 2 Rn ;
t
or
Z ˇ
ª ˇ
.0 /1 D min kl 0 D.s/kds ˇˇ hl; c i ¡.l j M / D 1 : (1.40)
l t
We are thus led to the next result whose proof follows from (1.40) and the definition
of the norms involved.
Theorem 1.3.2. The minimum magnitude control u0 that solves Problem 1.3.2
satisfies the maximum condition
0 0 0
maxfl 0 D.s/u j kuk 0 g D l 0 D.s/u0 .s/ D 0 kl 0 D.s/k; s 2 T; (1.41)
or
Z ª
maxfl 0 D.s/u j kuk 1gds C hl; c i ¡.l j M /; 8l 2 Rn ;
t
from which
¡.l j M / hl; c i
D min : (1.42)
l kl 0 D./kL1
(As before, kuk is the Rp -dimensional norm conjugate to the norm used in
kl 0 D.s/k.)
Theorem 1.3.3. The solution to Problem 1.3.2-A is given by formula (1.42).
Exercise 1.3.1. Solve Problem 1.3.1 with minimum cost
Z ª ˇ
ˇ
J .t; x/ D min ˇ
.u.s/; N.s/u.s// ds ˇ u./ 2 Lq .T / ; 1 < q < 1; q 6D 2:
q=2 .p/
t
(1.43)
instead of (1.30).
Observe that the existence of solutions to Problems 1.3.1, 1.3.2 (and as one may
check, also to (1.43)) requires the solvability of Eq. (1.32) for any vector c D
.c1 ; : : : ; cn /0 . This is guaranteed if and only if the functions d .i/ ./; i D 1; : : : ; n;
which are the rows of the n p matrix
1.3.3 Controllability
B; AB; : : : ; An1 B
xP D ax C u;
consider the problem of finding a control that moves the state from xŒt D x 0 6D 0
to x.ª/ D x .1/ D 0 with minimum cost
Z ª
.1/
minf ju.s/jds j u./ 2 L1 ŒT g:
t
Show that if a ¤ 0 the minimum is not attained, but is attained in the broader class
that includes delta functions.
An interesting issue is therefore to solve an optimization problem like Exercise
1.3.1, but with u./ D d U./=dt being the generalized derivative of U (in the sense
of the theory of distributions, [242]) and with cost as the total variation VarU./
of the function U.t / over the interval T . The solution, which may include delta
functions and its derivatives, is considered later in a separate section.
.p/ .p/
linearly independent. Then (1.32) is solvable by some u./ 2 Lq ŒT ; ku./kq
, if and only if
Z !1=q
ª
0 0 q =2
hD .s/l; D .s/li ds hl; ci; 8l 2 Rn :
t
1.4.1 Reachability
The reach set X Œª D X .ª; t0 ; X 0 / from set-valued position ft0 ; X 0 g at given time
ª is the union
[
X .ª; t0 ; X 0 / D fX .ª; t0 ; x 0 / j x 0 2 X 0 g:
Lemma 1.4.1. With X 0 and P .s/; s 2 T D Œt0 ; ª, convex and compact, the reach
set X λ is also convex and compact.
Proof. Consider two points x 0 2 X λ reached through control u0 .s/ 2 P .s/, and
x 00 2 X Œª, reached through u00 .s/ 2 P .s/; s 2 Œt0 ; ª. The convexity of X 0 and
P .s/ implies that x .’/ D ’x 0 C .1 ’x 00 / ; ’ 2 Œ0; 1 is reached through u.’/ .s/ D
’u0 .s/C.1’/u00 .s/ 2 P .s/ with initial condition x .’/ Œt0 D ’x 0 Œt0 C.1’/x 00 Œt0 :
This shows convexity. Boundedness of X λ is obvious. That X λ is closed follows
.p/
from the fact that the set of functions u./ 2 P ./ is weakly compact in L2 , so if
x is reached through u ./ and x ! x, then x is reached by a weak limit u./ of
n .n/ n
fu.n/ ./g. t
u
Problem 1.4.1. Find the reach set X λ.
Since X Œª D X .ª; t0 ; x 0 / is convex and compact, we shall describe it through its
support function
Introducing the column vector §Œt D §.t; ª; l/ D G 0 .ª; t /l as the solution to the
adjoint equation
xP 2 A.t /x C B.t /P .t /;
20 1 Linear Control Systems
which is a differential equation with set-valued right-hand side [7, 75]. From a
starting position ft0 ; X 0 g it has a set-valued solution X Œt D X .t; t0 ; X 0 / which
is nothing else than the reach set X Œt . This set X Œt may be represented as the set-
valued integral
Z t
X Œt D G.t; t0 /X 0 C G.t; s/B.s/P .s/ds:
t0
The set-valued integral above may be interpreted in the sense of either Riemann or
Lebesgue.
The support functions in (1.46), (1.48) yield the boundary @X λ of the set X λ. We
may also need to find a control that leads to each point of this boundary. We will
do that by calculating, for each direction l, the points of the boundary that are the
points of support xl for X λ in the direction l. For any l, xl is the solution to the
optimization problem
support of X Œª. Since, with l fixed, all the vectors ’l; ’ > 0, give the same point
of support xl , it suffices to deal only with vectors l of unit norm, hl; li D 1.
The calculation of xl decomposes into separate calculations of the two terms
in (1.49). The second term, which gives the appropriate initial condition, is obtained
through a convex optimization problem. The first term, which gives the control ul .t /
corresponding to xl from formula (1.46) is considered next.
Lemma 1.4.3. The point of support xl for X λ in direction l satisfies rela-
tion (1.49). The control ul .t / that steers the system to point xl satisfies the maximum
condition
l 0 G.ª; s/B.s/ul .s/ D maxfl 0 G.ª; s/B.s/u j u 2 P .s/g D ¡.l j G.ª; s/B.s/P .s//:
(1.50)
1.4 The Reachability Problem: Time-Optimal Control 21
Fig. 1.1 The hyperplane H.l/ supports X at xl and X H .l/ (left). Inner and outer
approximation to X (right)
Exercise 1.4.1. Prove formula (1.50) and also the assertions stated in Remark 1.4.1.
Definition (1.49) of xl has the following geometric interpretation. Since
H .l/ D fx j hl; xi ¡l g:
Thus we have the inner and outer approximation to X λ of the following exercise.
Exercise 1.4.2. Let l1 ; ; lk be nonzero vectors. Let ¡i D ¡.li j X Œª/, and
xi D xli . Show that
Since
we have
Z ª 1
q
¡.l j X Œª/ D .§0 Œt B.t /B 0 .t /§Œt /q =2 dt C §0 Œt0 x 0 ; (1.52)
t0
where 1=q C1=q D 1; so that q D q.q1/1 : Recall that §Œt D .G 1 .t; ª//0 l D
G 0 .ª; t /l:
To obtain (1.52) we used Hoelder’s inequality with d./ 2 Lq ; u./ 2 Lq 3
Z ª
hu.s/; d.s/ids ku./kL .p/ kd./kL .p/ ;
t0
q q
1=q
with equality if ku.s/kq ’kd.s/kq , u.s/ D ’kd.s/kq q d.s/, for some
constant ’ > 0.
Lemma 1.4.4. The reach set X λ under constraint (1.51) is given by
formula (1.52).
In particular, if q D 2, then also q D 2, we come to the next result.
Corollary 1.4.1. With q D 2 one has (§Œt is a column vector)
Z ª 1=2
¡.l j X Œª/ D §0 Œt B.t /B 0 .t /§Œt dt C §0 Œt0 x 0 D hl; W .ª; t0 /li1=2 C l 0 c;
t0
(1.53)
with c D G.ª; t0 /x 0 .
Exercise 1.4.3. For the support function
prove that
3
Instead of the traditional letter p paired with q, we use q , since p is used to denote for the
dimension of u.
1.4 The Reachability Problem: Time-Optimal Control 23
Calculation of the reach set is more difficult for the bilinear system (1.17), which
with v.s/ D 0 becomes
The difficulty is due to the fact that even if u is constrained to a convex compact set
with x 0 6D 0 the reach set X λ may be nonconvex. This is seen in the next example.
Exercise 1.4.4. Show that the differential inclusion for x 2 R2 ,
with X 0 D fx W x10 D x1 .0/ D 0; x20 D x2 .0/ 2 Œ’; ’; ’ > 0g; has a nonconvex
reach set X Œª D X .ª; 0; X 0 /; ª > 0:
Remark 1.4.2. A more general problem than finding the reach sets X Œt at time t is
S find reach sets within a given time interval t 2 T . The calculation of such sets,
to
fX Œt j t 2 T g, requires more complex operations and is treated in Sect. 2.6 of
Chap. 2.
Exercise 1.4.5. Find a two-dimensional S autonomous system x.t P / D Ax.t / and a
convex, compact initial set X 0 so that fX Œt j t 2 Œ0; 1g is not convex.
The next problem is basic. It initiated research in optimal control.
Problem 1.4.3. Given starting position ft0 ; x 0 g and final point x F , find control
u.s/ 2 P .s/; s 2 Œt0 ; ª that steers the system from x 0 to x F in minimum time,
i.e., ª t0 is minimized.
We shall solve this problem as follows. Consider the reach set X Œª D X .ª; t0 ; x 0 /
with ª > t0 . Assuming x F 6D x 0 ; let us look for the first instant of time ª0 when
x F 2 X Œª. Namely, denoting the (Euclidean) distance as –.ª/ D d.x F ; X Œª/, and
assuming –.t0 / > 0, we will find the smallest root ª0 of the equation –.ª/ D 0: Then
ª0 t0 will be the minimum time.
Taking B .0/ D fx W hx; xi 1g and expressing x F 2 X Œª C –B .0/ in terms of
its support function,
hl; x F i ¡.l j X Œª C ©B .0// D ¡.l j X Œª/ C ©hl; li1=2 ; 8fl W hl; li 1g;
we find that the distance d.x F ; X Œª/ D –.ª/ is the smallest © satisfying this
inequality,
Hence
Recall that the support function ¡.l j X Œª/ was calculated earlier (see
Lemma 1.4.2).
Theorem 1.4.1. The minimal time for transferring system (1.44) from starting point
xŒt0 to final point xŒª0 D x F is ª0 t0 , where ª0 is the minimal root of the equation
–.ª/ D 0:
Let l 0 be the optimizer of problem (1.56) with ª D ª0 .
Theorem 1.4.2. The time-optimal control u0 .t /; which steers the system from x 0 to
point x F satisfies the maximum condition of type (1.50) with l; ul ; t; ª substituted
by l 0 ; u0 .t /; t0 ; ª0 of Problem 1.4.3.
With l 0 6D 0 this maximum condition is known as the maximum principle,
discussed in more detail in the next section.
Remark 1.4.3. The optimal time ª0 t0 may be discontinuous in the boundary
points x 0 ; x F .
Exercise 1.4.6. Construct an example of Problem 1.4.3 for which ª0 t0 is
discontinuous in both x 0 ; x F .
We shall further deal with various versions of the time-optimal control problem
in detail, when solving specific cases.
In the next section we first introduce the maximum principle for linear systems as
the necessary condition of optimality, then indicate when it is a sufficient condition.
We shall mostly emphasize the nondegenerate case which is the main situation in
applications, the one for which the maximum principle was indeed introduced and
for which the optimal control may be found precisely from this principle. But we
shall also indicate degenerate abnormal or singular situations, which may occur and
for which the maximum principle is noninformative.
Here we deal with control problems for linear-convex systems, namely those of
type (1.2), with convex constraints on the control. In this section the problems are
solved in the class of open-loop controls UO , with no disturbance term, v.t / 0.
Problem 1.5.1. Consider system (1.2), with T D Œt0 ; ª. Given starting position
ft0 ; x 0 g, find
under constraints (1.2), (1.18). Here ®.t; x/ is continuous in ft; xg and convex
in x, bounded from below, with bounded level sets.
The level sets of ®.t; x/ are bounded iff 0 2 int Dom® .t; / ; 8t 2 T: Recall
that
and ® .t; l/, the (Fenchel) conjugate of ®.t; x/ in the second variable, is defined by
Substituting ª for t and xŒª for x in (1.58) and using the integral representa-
tion (1.3) for xλ, we come to
for system (1.2). The second term in the right-hand side of (1.61) comes from the
relations
Z ª
min hl; G.ª; s/B.s/u.s/ids ju./ 2 P ./
u t0
Z ª
D minfhl; G.ª; s/B.s/ui j u 2 P .s/gds
t0 u
Z ª
D maxfhl; G.ª; t /B.t /ui j u 2 P .s//ds
t0 u
Z ª
D ¡.§Œt j B.t /P .t //dt: (1.63)
t0
In the above lines the order of operations of min and integration may be
interchanged (see [238, p. 675]).
Let l 0 be a maximizer of ˆ0 .l j t0 ; x 0 / and let u0 ./ be an optimal control, i.e., a
minimizer of ®.ª; xŒª/ D maxl ˆ.u./; l j t0 ; x 0 /:
Recall the definition of saddle point for ˆ.u./; l j t0 ; x 0 / D ˆ.u; l/:
Definition 1.5.1. A pair fus ; l s g is said to be a saddle point for ˆ.u; l/ if
Lemma 1.5.1. The pair .u0 ./; l 0 / is a saddle point of ˆ.u./; l j t0 ; x 0 / D ˆ.u; l/
and ˆ.u0 ./; l 0 j t0 ; x 0 / D ˆ.u0 ; l 0 / D J .t0 ; x 0 /:
1.5 Optimal Open-Loop Control for Linear-Convex Systems: The Maximum. . . 27
Here equality is attained at u0 ./ if and only if for almost all t 2 T the following
pointwise maximum condition holds:
where @®.ª; xu0 /; @® .ª; l 0 / are the subdifferentials in the second variable of
®.ª; xu0 / and ® .ª; l 0 / at xu0 and l 0 :
1.5 Optimal Open-Loop Control for Linear-Convex Systems: The Maximum. . . 29
@®.ª; xu0 /
l 0 D ®x .ª; xu0 / D :
@x
Exercise 1.5.1. Prove Lemma 1.5.4. (See Lemma 1.5.7 below.)
Relations (1.69) are simpler under some additional conditions.
Lemma 1.5.5 ([238], Sect. 11.C). Suppose function ®.x/ is finite, coercive (this
means lim inff®.x/=kxkg ! 1 as kxk ! 1/, and also convex, of class C 2
(twice continuously differentiable) on Rn and its Hessian matrix r 2 ®.x/ is positive
definite for all x. Then its conjugate ® .l/ satisfies the same properties, namely, it
is also a finite, coercive, convex function of class C 2 on Rn with its Hessian matrix
r 2 ® .l/ positive definite for every l.
Corollary 1.5.2. Suppose function ®.ª; x/ satisfies conditions of Lemma 1.5.5 in
the second variable. Then relations (1.69) of Lemma 1.5.4 have the form
is degenerate. It is trivially fulfilled, but it is not the maximum principle. The optimal
control then has to be found from considerations other than the maximum principle.
As we shall now observe, this case occurs when Assumption 1.5.1 is not fulfilled.
Example 5.1. For the one-dimensional system
Here
1
max min ˆ.u; l/ D maxfl 0 x.0/ ªjlj hl; li2 g D max ˆ0 .l/:
l u l 2 l
The pair l 0 D 0 with any u D u0 satisfies the maximum condition but not the
maximum principle. Among these there exists an optimal control u0 which must
satisfy condition
Z ª
u.t /dt C x.0/ D 0
0
This is a singular control which has to be found from conditions other than the
maximum principle (Fig. 1.2).
We shall now see that there is a second degenerate case.
The second degenerate case
This is when §0 Œt 6D 0, but B 0 .t /§0 Œt 0:
Example 5.2. Consider system
Take ®.ª; xŒª/ D 12 hxŒª; xŒªi and x1 .0/ > 0; x2 .0/ < 0: We have
Z ª 1
ˆ 0 D max min ˆ.u; l/ D max minf.l1 x1 .0/ C l2 x2 .0// C .l1 C l2 /u.t /dt hl; lig D
l u l u 0 2
1
D maxf.l1 x1 .0/ C l2 x2 .0// ªjl1 C l2 j hl; lig:
l 2
1
ˆ0 D hx.ª/; x.ª/i; x.ª/ D x.0/ C bc.u.//; c.u.//
2
Z ª
D u.t /dt; x1 .ª/ D x2 .ª/ D x D x2 .0/=2 > 0:
0
The optimal control u0 must satisfy c.u0 .// D x ; and with ª jx2 .0/j=2 it is
singular, while u0 .t / 6D 0: There are many such controls (see Fig. 1.3).
We have thus indicated two types of degenerate problems. A protection from the
first type was ensured by Assumption 1.5.1. A protection from the second type is
ensured by the following assumption.
Let b .i/ .t / be the i -th column of B.t / and ui the i -th coordinate of u, i D
1; : : : ; p.
32 1 Linear Control Systems
Z ª
D l 0 G.ª; s/B.s/u0l .s/ds > 0; (1.70)
t0
We shall now indicate, under Assumption 1.5.1, some conditions when the
maximum principle turns out to be sufficient for optimality of the solution to
Problem 1.5.1.
Theorem 1.5.2. Suppose Assumption 1.5.1 is satisfied and the following conditions
for a function u .t / are true:
(a) u .t / satisfies the maximum principle of Definition 1.5.1 for l 0 6D 0—the
maximizer of ˆ0 .l/I
(b) u .t / satisfies the condition
Proof. Denote x 0 Œª; x Œª to be the vectors generated from xŒt0 D x 0 under an
optimal control u0 .t / and under control u .t /, respectively, with xλ being the one
generated from the same point under any admissible control u.t /:
Then, on the one hand, since u .t / satisfies condition (a) (the maximum principle
for l 0 ), we have
Z ª
ˆ.u ./; l / D h§ Œt0 ; x i C
0 0 0
§0 Œt 0 B.t /u .t /dt ® .ª; l 0 / D
t0
Z ª
D h§ Œt0 ; x i C
0 0
minf§0 Œt 0 B.t /u j u 2 P .t /gdt ® .ª; l 0 / D ˆ.u0 ./; l 0 / D
t0
1
maxfhl; x.0/i C k.l1 C l2 /ª hl; lig D ®.x .ª//;
l 2
we find
®.x .ª// D .x1 .0/ C k/2 C .x2 .0/ C k/2 > ®.x 0 .ª// D x12 .0/ C x22 .0/;
which indicates that the controls u .t / k with k 6D 0 are not optimal, though they
satisfy the maximum condition.
1.5 Optimal Open-Loop Control for Linear-Convex Systems: The Maximum. . . 35
find the optimal control that minimizes ®.xŒª/ D maxi jxi .ª/j; i D 1; 2:
Recall that here ® .l/ D I .l j B .0//; where B .0/ D fl W jl1 j C jl2 j 1g and
I .l j B / is the indicator function for set B (as defined in convex analysis), so that
I .l j B / D 1 if l 2 B and I .l j B / D C1 otherwise.
We have
ˆ 0 D ˆ.u0 ; l 0 / D
Z ª
D max hl; x.0/i C minfhl; u.s/i j ju.s/j gds I .l j B .0// D max ˆ0 .l/ D
0 l
minfhl 0 ; u.t /i j ju.t /j 1g D .jl10 j C jl20 j/; u1 .t / 2 Œ1; 1; u2 .t / 1;
where l 0 is the maximizer of ˆ0 .l/: The maximum principle is thus satisfied with
u2 .t / 1 and any control u1 .t /:
Let us now check condition (b) of Theorem 1.5.2, calculating
Z ª
maxfhl; x.0/i C hl; u .s/ids j jl1 j C jl2 j 1g D max ˆ.u ; l/:
l 0 l
This gives, for u2 .t / 1; and any u1 2 Œ1; 0; maxl ˆ.u ; l/ D
maxfjx1 .1/j; jx2 .1/jg D 1; which means all such controls are optimal.
At the same time, for u2 .t / 1; and any u1 2 Œk; 1; k 2 .0; 1 we have
maxl ˆ.u ; l/ D maxfjx1 .1/j; jx2 .1/jg 1 C k > ˆ.u ; l 0 / so such controls are
not optimal (Fig. 1.4).
36 1 Linear Control Systems
The next example illustrates the case when the optimal control is unique.
Exercise 1.5.3. For the two-dimensional system
consider the problem of minimizing ®.ª; xŒª/ D 12 kxŒªk2 : For ª < 1, show that
the unique maximizer of ˆ0 .l j t0 ; x 0 / is l0 D .1; 0/0 . Hence u.t /; t 2 Œ0; ª,
satisfies the maximum principle iff u1 .t / 1. The optimal control in addition
satisfies u02 .t / 0. Prove that the optimal control u01 .t / 1; u02 .t / 0 is unique.
It is interesting to observe that the notion of saddle point allows us to give the
following interpretation of sufficiency.
Theorem 1.5.4. Suppose fus ./; l s g is a saddle point of ˆ.u./; l j t0 ; x 0 /. Then
us ./ is an optimal control of Problem 1.5.1 which satisfies the maximum condition
of Definition 1.5.2 with §s Œª D l s :
Note that Theorem 1.5.4 covers both degenerate cases. If Assumption 1.5.1 is
satisfied, we moreover have l s 6D 0:
Exercise 1.5.4. Prove Theorem 1.5.4.
if and only if
If l 0 ¤ 0, then xu0 is the point of support for the set fx W ®.ª; x/ ®.ª; xu0 /g in
direction l 0 , i.e. l 0 2 @®.ª; xu0 /:
Proof. Evidently
hl 0 ; xu0 i ® .ª; l 0 / D ˆ.u0 ; l 0 / D max ˆ.u0 ; l/ D maxfhl; xu0 i ® .ª; l/g D ®.ª; xu0 /;
l l
is equivalent to
if and only if xu0 is the point of support for the reach set X λ in direction l 0 , i.e.,
and by Lemma 1.5.8 (the maximum principle in the form of a minimum rule),
whence
so u0 is optimal. t
u
Remark 1.5.4. Be careful, since, as indicated in Examples 5.1 and 5.2, the control
u0 of Lemma 1.5.9 may turn out to be singular and the maximum principle for u0
may be degenerate.
Recall that a compact set P is said to be strictly convex, if its boundary @P does not
contain any line segment. That is, if p 0 ; p 00 2 @P , then œp 0 C.1œ/p 00 62 @P ; 8œ 2
.0; 1/:
Strictly convex sets may be presented as level sets of strictly convex functions.
A function ®.x/ is strictly convex in the convex domain D if for points p 0 ; p 00 2 D
one has
4
Recall that here, as mentioned in Sect. 1.2, set P .t / is assumed to have an interior point for all t:
1.5 Optimal Open-Loop Control for Linear-Convex Systems: The Maximum. . . 39
E .p ; P / D fp W hp p ; P 1 .p p /i 1g; P D P 0 > 0:
maxfhd; ui j u 2 P g D hd; u0 i
is unique.
Lemma 1.5.12. Suppose the continuous function d.t / 6D 0 almost everywhere in
T and function P .t / is Hausdorff-continuous with values in strictly convex compact
sets. Then the optimizer u0 .t / to
hd.t /; u0 .t /i D maxfhd.t /; ui j u 2 P .t /g
Here and in the sequel all conjugates ® ; ®i and all sets Dom./ are taken
with respect to the second variable unless otherwise indicated. Next, we reduce
Problem 1.6.1 to a dual problem of optimization. We first rewrite functions
®1 .t; z/; ®.t; x/; ®2 .t; u/ in terms of their conjugates:
Using formula (1.3), substitute for xŒs; xŒª first in (1.74) and (1.75), and
then substitute these results into (1.73). This leads to the following problem of
minimaximization:
in which the variables range over u.s/ 2 Rp ; œ.s/ 2 Rm ; s 2 Œt; ª, and l 2 Rn ,
and
Z ª
ˆ.u./; l; œ./ j t; x/ D hl; G.ª; t /xi C hl; G.ª; Ÿ/B.Ÿ/u.Ÿ/id Ÿ ® .ª; l/C
t
Z ª Z s
C œ.s/; H.s/G.s; t /xCH.s/ G.s; Ÿ/B.Ÿ/u.Ÿ/d Ÿ ®1 .s; œ.s//C®2 .s; u.s// ds:
t t
(1.78)
We now introduce a new adjoint system for Problem 1.6.1,
P
§.s/ D A0 .s/§.s/ H 0 .s/œ.s/; §Œª D l: (1.79)
ˆ.u./; l; œ./ j t; x/
Z ª
D h§Œt ; xi C §ŒŸ0 B.Ÿ/u.Ÿ/ C ®2 .Ÿ; u.Ÿ// ®1 .Ÿ; œ.Ÿ// d Ÿ ® .ª; l/:
t
which yields
is bounded for each s 2 T D Œt; ª. Together with the continuity of ®2 .s; u/
in s this yields the Hausdorff-continuity of U 0 .s/, which means that in the
minimaximization (1.77) the functions u.s/ may be selected in a weakly compact
1.6 Duality Methods of Convex Analysis in Problems of Optimal Control 43
subset set of L2 ŒT . The last fact allows us to apply Ky Fan’s minmax theorem
to (1.77), which yields
in which
Z ª
ˆ0 .l; œ./ j t; x/ D h§Œt ; xi ®2 .s; B 0 .s/§Œs/C®1 .s; œ.s// ds® .ª; l/:
t
(1.82)
Here we have used the relations
Z ª Z ª
0
min .§Œs B.s/u.s/C®2 .s; u.s///ds D min.§Œs0 B.s/u.s/C®2 .s; u.s///ds D
u./ t t u
Z ª
D .®2 .s; B 0 .s/§Œs/ds: (1.83)
t
In summary we have two optimization problems: the Primal and the Dual.
Primal Problem: Find the optimal control u0 .s/; s 2 Œt; ª that minimizes the
primal functional
Z ª
….t; x; u/ D .®1 .s; H.s/xŒs/ C ®2 .s; u.s///ds C ®.ª; xŒª/;
t
Dual Problem: Find the optimal pair fl; œ.s/; s 2 Œt; ªg that maximizes the
dual functional
Z ª
ˆ0 .l; œ./ j t; x/ D h§Œt ; xi ®2 .s; B 0 .s/§Œs/C®1 .s; œ.s// ds® .ª; l/;
t
P
§.s/ D A0 .s/§.s/ H 0 .s/œ.s/;
The attainability of the maximum in fl; œ./g in (1.81), (1.82) and of the
minimum in u./ in (1.83) is ensured by Assumption 1.6.1. This pointwise min-
imaximization in u.s/; œ.s/; s 2 Œt; ª with a further maximization in l yields
functions u0 .s/; œ0 .s/ and also a vector l 0 . The indicated optimizers need not
be unique. However, due to the properties of systems (1.2), (1.79) and functions
®1 .t; x/; ®.t; x/; ®2 .t; u/, there always exists a pair of realizations u0 .s/; œ0 .s/ that
are integrable and, in fact, may even turn out to be continuous.
Exercise 1.6.1. Prove the existence of integrable realizations u0 .s/; œ0 .s/; s 2
Œt; ª, for the optimizers of (1.81)–(1.83).
Exercise 1.6.2. Prove that whatever be l 0 6D 0; one has §.t / 6 0; t 2 Œt0 ; ª;
whatever be œ.t /:
The relations above indicate the following necessary conditions of optimality of
the control realization u0 .s/; s 2 Œt; ª.
Theorem 1.6.1. Suppose u0 .t / is the optimal control for Problem 1.6.1 and
fl 0 ; œ0 ./g are the optimizers of the dual functional ˆ0 .l; œ./ j t; x/ that generate
the solution §0 Œs D §.s; ª/ under l D l 0 ; œ./ D œ0 ./. Then u0 .t / satisfies the
following pointwise minimum condition
§0 Œs0 B.s/u0 .s/C®2 .s; u0 .s// D minf§0 Œs0 B.s/uC®2 .s; u/ j u 2 Rp g; s 2 Œt; ª;
u
(1.85)
which is equivalent to the pointwise maximum condition
Proof. Indeed, if inclusion (1.87) does not hold, it would contradict the assumption
that fl 0 ; œ0 ./g 6D 0: This could be checked by direct substitution. t
u
Note that inclusion (1.87) is also the necessary and sufficient condition for the
level sets of the function
to be bounded and hence for the attainability of the minimum in u in (1.85) (which
is also the maximum in (1.86)) for s 2 T .
Lemma 1.6.5. Under the assumptions of Lemma 1.6.4 the minimum in (1.85) is
attained and the minimizer u0 .s/; s 2 T; may be selected as a piecewise continuous
realization.
This brings us to the sufficient conditions for optimality of the control realization
u0 .s/; s 2 Œt; ª.
Theorem 1.6.2. Suppose Assumptions 1.6.2, 1.6.3(i) are true, the requirements of
Lemma 1.6.4 hold and
(i) a control u .s/; s 2 T , satisfies the maximum principle (1.86), with fl 0 6D 0;
œ0 .t / 6D 0g as the maximizer for the Dual Problem,
46 1 Linear Control Systems
is true.
Then u .s/; s 2 T , is the optimal control for the Primal Problem.
Proof. The proof is similar to that of the previous section (see Theorem 1.5.2). t
u
An additional requirement for functions ®1 .t; z/; ®.t; x/; ®2 .t; u/ to be strictly
convex in z; x and u results in the uniqueness of the optimal control.
Theorem 1.6.3. Suppose under the conditions of Theorem 1.6.2 Assumption 1.6.4
is also true. Then solutions u0 .s/; and fl 0 ; œ0 .s/g to both Primal and Dual Problems
are unique.
Exercise 1.6.5. Suppose that functions ®1 .t; z/; ®.t; x/; ®2 .t; u/ are quadratic
forms in z and u:
1 1 1
®1 .t; z/ D hz; M.t /zi; ®.t; x/ D hx m; L.x m/i; ®2 .t; u/ D hu; N.t /ui;
2 2 2
(i) Solve the linear-quadratic problem (1.88) by duality theory of this subsection.
(ii) Solve the same problem through methods of classical variational calculus.
Problem 1.5.1 of open-loop optimal control was solved in Sect. 1.5 for a fixed
starting position ft0 D t; x 0 D xg with solution given as an optimal cost J .t0 ; x 0 /
under control u0 .Ÿ/ D u0 .Ÿ j t0 ; x 0 /; Ÿ 2 Œt0 ; ª; in the class u0 ./ 2 UOO :
Now, solving this problem of feedforward control for any interval Œt; ª and any
starting position ft; xg; such solution may be used to create a model-predictive
control uf .t; x/ D u0 .t j t; x/ which will be a function of ft; xg: However, such
moves require a more general approach to variational problems of control that would
indicate solutions of the same problem in the class UC C of closed-loop control
strategies. Such are the Dynamic Programming techniques discussed in Chap. 2.
Chapter 2
The Dynamic Programming Approach
1
The general nondifferentiable case for the value function is discussed later in Sect. 5.1.
with starting position f£; xg; x.£/ D x: Here we have either U.t / D Rp or
U.t / D P .t /, where P .t / is compact-valued and Hausdorff-continuous.
The control will be sought for as either open-loop u.t /; u./ 2 UO , or closed-
loop u.t; x/; u.; / 2 UC :
We wish to find the optimal control u that minimizes the Mayer–Bolza functional
of the general form
Zª
J .£; x j u.// D L.t; xŒt ; u.t //dt C ®.ª; xŒª/: (2.2)
£
In (2.2) ª is a fixed terminal time, and xŒt D x.t; £; x/ is the trajectory of (2.1)
starting from initial position f£; xg. It is customary to call L.t; x; u/ the running
cost and ®.ª; x/ the terminal cost. To begin, we suppose that the functions L; ® are
differentiable and additional conditions are satisfied that justify the mathematical
operations considered in this section. We will explicitly specify these conditions
later.
Definition 2.1.1. The value function for the problem of minimizing (2.2) is
defined as
0 ª 1
Z ˇ
ˇ
V .£; x£ / D inf @ L.t; xŒt ; u.t //dt C ®.ª; xŒª/ ˇˇ u./ 2 U./; xŒ£ D x£ A ;
u./
£
(2.3)
where U./ D fu./ W u.t / 2 U.t /; t 2 T£ g:
Thus V .£; x£ / is the infimum of the cost (2.2) incurred by any feasible control u./ 2
U./ and starting in state xŒ£ D x£ . For any u.t / 2 U.t / and resulting trajectory
xŒt D x.t; £; x£ / we must have
0 1
Z¢ Zª
V .£; x£ / L.t; xŒt ; u.t //dt C @ L.t; xŒt ; u.t //dt C ®.ª; xŒª/A ;
£ ¢
2.1 The Dynamic Programming Equation 49
for £ < ¢ < ª. Minimizing the second term with respect to u gives
V .£; x£ /
Z¢
L.t; xŒt ; u.t //dt
£
8 ª 9
<Z ˇ =
ˇ
C inf L.t; xŒt ; u.t //dt C ®.ª; xŒª/ ˇˇ u./ 2 U./; xŒ¢ D x.¢; £; x£ /
: ;
¢
Z¢
D L.t; xŒt ; u.t //dt C V .¢; x.¢//: (2.4)
£
On the other hand, since V .£; x£ / is the infimum, for each © > 0 there exists uQ ./
such that
Z¢
D Q ; uQ .t //dt C V .¢; xŒ¢/;
L.t; xŒt Q (2.5)
£
Q D x.t;
with xŒt Q £; x£ / being the trajectory resulting from control uQ .t /. From (2.4)
and (2.5)
Z¢
V .£; x£ / Q ; uQ .t //dt C V .¢; xŒ¢/
L.t; xŒt Q V .£; x£ / C ©:
£
Denote by V .t; x j ª; ®.ª; // the value function for the problem to minimize (2.2)
with terminal cost ®.ª; / (and the same running cost). With this notation, the value
function (2.3) is V .t; x j ª; ®.ª; // and the terminal cost serves as the boundary
condition
Lastly, recognizing the right-hand side of (2.6) as the value function for the
problem with terminal cost V .¢; / allows us to express (2.6) as the Principle of
Optimality:
Observe that (2.8) is the semigroup property for the mapping V .£; / 7! V .ª; /:
Substituting t for £, t C ¢ for ¢ and s for t , rewrite (2.6) as
0 tC¢ 1
Z ˇ
ˇ
V .t; x/ D inf @ L.s; xŒs; u.s//ds C V .t C ¢; xŒt C ¢/ ˇˇ u./ 2 U./A :
t
(2.9)
Here ¢ 0 and xŒs D x.s; t; x/ is the trajectory of (2.1) starting from ft; xg. For
¢ 0, (2.9) can be written as
8 tC¢ 9
<Z ˇ =
ˇ
inf L.s; xŒs; u.s//ds C V .t C ¢; xŒ£ C ¢/ V .t; x/ ˇˇ u./ 2 U./ D 0;
: ;
t
and so
0 1
Z
tC¢ ˇ
1 V .t C ¢; xŒt C ¢/ V .t; x/ ˇˇ
lim inf @ L.s; xŒs; u.s//ds C A
ˇ u./ 2 U./ D 0:
¢!C0 ¢ ¢
t
(2.10)
Suppose now that V .t; x/ is continuously differentiable at .t; x/. Then reversing the
order of the operations lim and inf and passing to the limit with ¢ ! C0, we get
ˇ
d V .t; xŒt / ˇˇ
inf L.t; xŒt ; u/ C ˇ u 2 U.t / D 0: (2.11)
u dt
Expanding the total derivative d V .t; xŒt /=dt in terms of partial derivatives and
noting that @V .t; xŒt /=@t does not depend on u, (2.11) may be rewritten as
ˇ
@V .t; x/ @V .t; x/ ˇ
C inf ; f .t; x; u/ C L.t; x; u/ ˇˇ u 2 U.t / D 0; (2.12)
@t @x
2.1 The Dynamic Programming Equation 51
with boundary condition (2.7). This partial differential equation is known as the
(backward) Hamilton–Jacobi–Bellman equation or simply the HJB equation. It is
“backward” because the boundary condition (2.7) is at the terminal time; the
“forward” HJB equation has boundary condition at the initial time.
Lemma 2.1.1. If the value function V .t; x/ given by (2.3) is differentiable, it
satisfies the HJB equation (2.12) with boundary condition (2.7).
Exercise 2.1.1. Justify the reversal of lim and inf in (2.10).
Theorem 2.1.1. (i) Suppose the value function V .t; x/ for problem (2.3) is differ-
entiable at point ft; xg. Then it satisfies the inequality
@V .t; x/ @V .t; x/
C infw ; f .t; x; w/ C L.t; x; w/ j w 2 UŒt 0:
@t @x
(2.13)
(ii) If in addition to (i) there exists an optimal control u0 .s/ 2 U.s/, such that
u0 .s/ ! w0 2 U.t / as s ! t C 0, then
@V @V .t; x/
C ; f .t; x; w0 / C L.t; x; w0 / D 0: (2.14)
@t @x
Since ¨.ª; x.ª// D ®.ª; xŒª/, moving the integral to the left side of the last
relation gives
Z ª
¨.t; x/ L.s; x.s/; u.s//ds C ®.ª; xŒª/ D J .t; x j u.//:
t
Remark 2.1.1. Formally, given a closed-loop control u0 .t; x/, one may find the
corresponding open-loop control u0 D u0 Œs by first obtaining the solution xŒs D
x.s; t; x/ of the differential equation
and then substituting to get u0 Œs D u0 .s; xŒs/: However, one must ensure
that (2.17) has a solution in some reasonable sense, particularly, when the
feedback control u0 .s; x/ is nonlinear (even discontinuous) in x, but the original
system is linear in x.
Consider now the problem of finding u that minimizes the inverse Mayer–Bolza
functional
Zt
J0 .t; x j u.// D ®.t0 ; x.t0 // C L.s; x.s/; u.s//ds; (2.18)
t0
with given final position ft; xg; xŒt D x; and initial time t0 t: The value function
for this problem is
Mimicking the previous argument now leads to the Dynamic Programming Equation
Z t
V0 .t; x/ D inffV0 .t ¢; xŒt ¢/C L.s; xŒs; u.s//ds j u./ 2 U./; x.t / D xg;
t¢
from which, supposing V0 .t; x/ is differentiable, one obtains the forward HJB
equation
ˇ
@V0 @V0 .t; x/ ˇ
.t; x/ C sup ˇ
; f .t; x; u/ L.t; x; u/ ˇ u 2 U.t / D 0; (2.20)
@t @x
Exercise 2.1.2. Derive the Dynamic Programming Equation in forward time and
then the forward HJB equation (2.20).
Theorem 2.1.3 (The Second Verification Theorem). Suppose a differentiable
function ¨0 .t; x/ satisfies the HJB equation (2.20) in domain D D Tt0 Rn ; together
with boundary condition (2.21). Then in this domain ¨0 .t; x/ V0 .t; x/. Moreover,
if control u0 ./, with trajectory x 0 Œs D x 0 .s; t; x/; x 0 Œt D x; s t , satisfies
@¨0
L.s; x 0 Œs; u0 .s// C .s; x 0 Œs/; f .s; x 0 Œs; u0 .s// D
@x
ˇ
@¨0 ˇ
D max L.s; x 0 Œs; u/ C .s; x 0 Œs; u/; f .s; x 0 Œs; u/ ˇ u 2 U.s/ ;
@x ˇ
(2.23)
then u0 ./ is optimal and ¨0 .t; x/ D V0 .t; x/.
Exercise 2.1.3. Prove Theorem 2.1.3.
Corollary 2.1.2. The optimal feedback (closed-loop) control for problem (2.19) is
ˇ
@V0 ˇ
u0 .t; x/ 2 Arg max ; f .t; x; u/ L.t; x; u/ ˇˇ u 2 U.t / : (2.24)
@x
54 2 The Dynamic Programming Approach
Remark 2.1.2. The optimal closed-loop control u0 .t; x/ found here is feasible only
when accompanied by an existence theorem for the equation
Minimizing a quadratic integral cost over the trajectories of a linear control system
is the basic problem in all courses on linear control. It has a solution in explicit
form, fairly easily obtained and smooth enough to satisfy all requirements for a
classical solution to the HJB equation of the previous section. This problem has
a vast literature, but it is not the subject of this book, and is presented here as a
necessary transitionary passage to main topics.
Consider the linear system (1.2) or (1.44), namely
Zt1
J.t0 ; x0 ; u.// D hx; M.t /xi C hu; N.t /ui dt C hx.t1 /; T x.t1 /i; (2.26)
t0
1 @V
u.t; x/ D N 1 .t /B 0 .t / : (2.30)
2 @x
Substituting (2.30) into (2.28), we come to
@V @V 1 @V 1 0 @V
C ; Ax ; BN B C hx; M xi D 0: (2.31)
@t @x 4 @x @x
We follow Theorem 2.1.2 and look for V .t; x/ as a quadratic form V .t; x/ D
hx; P .t /xi, with P .t / D P 0 .t /. Then
@V @V
D hx; PP .t /xi; D 2P .t /x:
@t @x
Substituting these into (2.31) gives
and so we arrive at the (backward) matrix differential equation of the Riccati type
PP C PA C A0 P P 0 BN 1 B 0 P C M D 0; (2.32)
P .t1 / D T: (2.33)
Consider the special case of Eq. (2.32) with M.t / 0. Differentiating the
identity I D P .t /P 1 .t / gives PP P 1 C P PP 1 D 0 or
PP 1 D P 1 PP P 1 :
Multiplying both sides of Eq. (2.32) by P 1 , and using the last relation, we have
PP 1 D AP 1 C P 1 A0 BN 1 B 0 ; P 1 .t1 / D T 1 : (2.34)
56 2 The Dynamic Programming Approach
The solution of this linear matrix equation is given by the integral formula
Z t1
P 1 .t / D G.t; t1 /T 1 G0 .t; t1 / C G.t; s/B.s/N 1 .s/B 0 .s/G0 .t; s/ds;
t
(2.35)
in which G.t; s/ is defined in (1.11), (1.12).
Theorem 2.2.1. The value function for Problem 2.2.1 has the quadratic form
V .t; x/ D hx; P .t /xi, in which P .t / satisfies the matrix Riccati equation (2.32),
with boundary condition (2.33). The optimal control, given by the linear feedback
function (2.30), is
u.t; x/ D N 1 .t /B 0 .t /P .t /x:
Zt1
J0 .t1 ; x1 ; u.// D hx.t0 /; Lx.t0 /i C hx; M.t /xi C hu; N.t /ui dt; (2.36)
t0
1 1 @V0
u.t; x/ D N .t /B 0 .t / ; (2.40)
2 @x
2.2 The Linear-Quadratic Problem 57
The solution to Problem 2.2.2 may be used to determine the reach set X Œt of
system (2.25) under the integral constraint
In fact,
J .t0 ; x0 ; u.// D
Zt1
D hxx .t /; M.t /.xx .t //iChu; N.t /ui dt Chx.t1 /a; T .x.t1 /a/i; x.t0 / D x 0 ;
t0
(2.44)
Zt1
J0 .t1 ; x1 ; u.// D maxfhx.t0 /; Lx.t0 /i; hu; N.t /uidt g; x.t1 / D x .1/ : (2.45)
t0
We shall now use the Dynamic Programming approach to the calculation of reach
sets under hard bounds on the controls.
58 2 The Dynamic Programming Approach
For problems of reachability and closed-loop control we derive the main Dynamic
Programming equation which is a first-order partial differential equation with
appropriate boundary conditions.
We extend Definition 1.4.1 of reach sets to the nonlinear system (2.1), under the
hard bound u 2 P .t /.
Definition 2.3.1. The reach set X Œª D X .ª; t0 ; x 0 / of system (2.1) at given time
ª, from position ft0 ; x 0 g, is the set of all points x, for each of which there exists a
trajectory xŒs D x.s; t0 ; x 0 /, generated by a control subject to the given constraint,
that transfers the system from position ft0 ; x 0 g to position fª; xg; x D xŒª:
The reach set X Œª D X .ª; t0 ; X 0 / from set-valued position ft0 ; X 0 g at given time
ª is the union
[
X .ª; t0 ; X 0 / D fX .ª; t0 ; x 0 / j x 0 2 X 0 g:
d 2 .x; X / D minfhx z; x zi j z 2 X g
is the square of the distance d.x; X / from point x to set X . Observe that (2.46) is a
special case of (2.19) with zero running cost in (2.18).
2.3 Reachability Through the HJB Equation: Hard Bounds 59
As seen from Fig. 2.1, (2.47) follows from the definition of the reach set
X .£; t0 ; X 0 /, which thus turns out to be the level set (at level zero) of the function
V0 .£; /.
For the value function V0 .t; x/ we again use the notation
This theorem together with Lemma 2.3.1 also implies the following assertion.
Lemma 2.3.2. The reachability set X Œt D X .t; t0 ; X 0 / satisfies the next relation
The solution of the reachability problem, namely the one of calculating the reach set,
now depends on the properties of either the “classical” or the generalized “viscosity”
solutions of the forward HJB equation
@V0 .t; x/ @V0 .t; x/
C max ; f .t; x; u/ j u 2 P .t / D 0; (2.50)
@t u @x
Equation (2.50) follows from (2.48) or it may be seen as a special case of (2.20).
This equation is solvable in the classical sense (that is, (2.50) holds everywhere),
if the partials of V .t; x/ in t; x exist and are continuous. Otherwise (2.50) is a
symbolic relation for the generalized HJB equation which has to be described in
terms of subdifferentials, Dini derivatives or their equivalents. However, the typical
situation is that V is not differentiable. The treatment of Eq. (2.50) then involves
the notion of generalized “viscosity” or “minmax” solution for this equation (see
[16, 50, 247]). The last notion will be later treated in a separate Sect. 5.1 which
deals with nondifferentiable value functions and generalized solutions. In the case
of linear systems with convex constraints the value functions are convex in the
state space variables and hence directionally differentiable these solutions belong
the generalized, which property is checked directly.
The calculation of reach sets and tubes can thus be reduced to the calculation
of the value function V0 .t; x/ or its level sets X Œt D fx W V0 .t; x/ 0g and their
evolution in time t . On the other hand, we note that X Œt may also be treated as
the cross-section (“cut”) X Œt D X .t; t0 ; X 0 / of the solution tube X .; t0 ; X 0 / to the
differential inclusion
Theorem 2.3.2. Under Assumption 2.3.1 the forward reach set X Œt is compact and
the set-valued tube X Œ is the unique solution to the following “funnel” equation for
the differential inclusion (2.52):
Whereas the reach set is the set of all states that can be reached from a given initial
set, the backward reach set comprises all those states, taken at instant £, from which
it is possible to reach a given “target” set M taken to be a convex compact in Rn .
Definition 2.3.2. Given a closed target set M Rn , the backward reach
(solvability) set W Œ£ D W .£; t1 ; M / at given time £, from set-valued position
ft1 ; M g; is the set of states x 2 Rn for each of which there exists a control
u.t / 2 P .t / that steers system (2.1) from state x.£/ D x to x.t1 / 2 M . The set-
valued function W Œt D W .t; t1 ; M /; £ t t1 ; is the solvability (or backward
reach) tube from ft1 ; M g.
Backward reach sets W Σ are also known as weakly invariant sets relative to
position ft1 ; M g:
62 2 The Dynamic Programming Approach
Definition 2.3.3. A set W Σ is said to be weakly invariant relative to system (2.1)
[(1.2), (1.7)] and position ft1 ; M g, if it consists of all those points x D x.£/
(positions f£; xg), from which set M is reachable at time t1 by some of the possible
controls.
Hence, from each position f£; xg 2 W Œ£ there exists at least one trajectory that
stays in the backward reach tube and reaches M at t D t1 : The backward reach set
W .£; t1 ; M / is the largest or, in other words, the inclusion maximal weakly invariant
set relative to M .
The calculation of backward reach sets may be achieved through the value
function
in which the minimization is over all u./ 2 P ./. Similarly to Lemma 2.3.1, the
backward reach set is a level set.
Lemma 2.3.3. The following relation is true:
(in backward time) and the value function V .t; xjt1 ; V .t1 ; // satisfies the Principle
of Optimality in the semigroup form
Associated with the Principle of Optimality (2.58) is the backward HJB equation
@V .t; x/ @V
C min ; f .t; x; u/ D 0; (2.59)
@t u @x
Associated with the semigroup (2.57) is the “backward” funnel equation for the
differential inclusion (2.52):
2
Problems of reachability under unknown but bounded disturbances are beyond the scope of this
book, along the approaches of which they are treated in papers [133, 176, 183].
64 2 The Dynamic Programming Approach
Theorem 2.3.5. If Eq. (2.64) has a unique (viscosity or classical) solution Vs .t; x/,
the strongly invariant set Ws Œ£ D Ws .£; t1 ; M / is the level set
In this section we calculate the value functions through methods of convex analysis
and also find the support functions for the forward and backward reachability sets.
with continuous matrix coefficients A.t /; B.t / and hard bound u.t / 2 P .t /, with
P .t / convex, compact, and Hausdorff-continuous.
Problem 2.4.1. Given starting set-valued position ft0 ; X 0 g, terminal time ª and
condition x.ª/ D x, calculate the value function V0 .ª; x/ of (2.46) at any position
fª; xg:
2.4 Reachability for Linear-Convex Systems 65
After solving this problem we may use (2.47) to find the forward reach set X Œt D
X .t; t0 ; X 0 / as the level set X Œt D fx W V0 .ª; x/ 0g.
The HJB equation (2.50) now has the form
ˇ
@V0 .t; x/ @V0 .t; x/ ˇ
C max ; A.t /x C B.t /u ˇˇ u 2 P .t / D 0;
@t @x
or, since
maxfhl; ui j u 2 P .t /g D ¡.l j P /
Instead of solving the HJB equation, V0 .£; x/ may be calculated through duality
techniques of convex analysis along the scheme given, for example, in [174,
Sect. 1.5].
Indeed, observing that the Fenchel conjugate for function ®.£; x/ D d 2 .x; X 0 /
in the second variable is ® .£; l/ D ¡.l j X 0 / C 14 hl; li, we may write, making use
of the minmax theorem of [72] to interchange min and max below,
1
D min maxfhl; x.t0 /i hl; li ¡.l j X 0 / j x.£/ D xg
u l 4
D max min ‰.£; x; l; u.//; (2.68)
l u
where
Z £
1
‰.£; x; l; u.// D hs.£; t0 ; l/; xi hs.t; t0 ; l/; B.t /u.t /idt hl; li ¡.l j X 0 /:
t0 4
with
We now show by a direct substitution of V0 .t; x/ into (2.66) that it satisfies this
equation with boundary condition (2.67). Indeed, let l 0 D l 0 .t; x/ be the maximizer
of function ˆ0 Œt; x; l in l (see (2.70)). The structure of ˆ0 Œt; x; l indicates that l 0
is unique. (Prove this assertion.)
Next we calculate the partial derivatives @V0 .t; x/=@t ; @V0 .t; x/=@x. According
to the rules of differentiating the maximum of a function like in (2.70), (see [61]),
we have
1
V .t0 ; x/ D maxfhl; xi hl; li ¡.l j X 0 / j l 2 Rn g D
4
Proof. The second assertion follows from the strict convexity of ˆ0 .£; x; l/ in l due
to the quadratic term and from the continuity of ˆ.£; x; l/ in £; x. Property (iii) may
be checked directly from the explicit formulas for the partials of V0 .t; x/, while (iv)
is verified through direct substitution (Check whether V0 .t; x/ is unique). t
u
Remark 2.4.1. Note that for a time-invariant system one has
V0 .£; x/ D V0 .£; x j t0 ; X 0 / D V0 .£ t0 ; x j 0; X 0 /:
Having calculated V0 .t; x/ according to (2.70), let us now find its conjugate in the
second variable, denoted by V0 .t; l/. We have
Z t
D min maxfhl; xi hs.t; t0 ; œ/; xi C ¡.B 0 .Ÿ/s.Ÿ; t0 ; œ/ j P .Ÿ//d Ÿ
œ x t0
1
C hœ; œi C ¡.œ j X 0 / j x; œ 2 Rn g:
4
Theorem 2.4.2. (i) The conjugate V0 .t; l/ to the value function V0 .t; x/ is given
by (2.71).
(ii) The forward reach set X Œt D X .t; t0 ; X 0 / is given by
A scheme similar to the previous one works for the solution V .t; x/ of the backward
equation (2.59) written for linear system (2.65) as
ˇ
@V .t; x/ @V .t; x/ @V .t; x/ ˇ
C ; A.t /x ¡ ˇ B.t /P .t / D 0: (2.73)
@t @x @x ˇ
with
Z t1
1
ˆŒ£; x; l D hs.£; t1 ; l/; xi ¡.s.t; t1 ; l/ j B.t /P .t //dt hl; li ¡.l j M /:
£ 4
Theorem 2.4.3. The value function V .£; x/ for the backward HJB equa-
tion (2.73), (2.74) has the following properties:
(i) V .£; x/ is given by (2.75).
(ii) The maximizer l 0 .£; x/ D arg maxl ˆŒ£; x; l is unique and continuous in £; x.
(iii) V .£; x/ is a proper convex function in x and is directionally differentiable for
any £; x.
(iv) V .t; x/ satisfies the HJB equation (2.73) for all t; x and also the boundary
condition (2.74).
Exercise 2.4.1. Find the conjugate function V .t; l/:
Formulas of type (2.70), (2.75) may be also calculated with criteria d.x.t0 /; X 0 /;
d.x.t1 /; M / instead of d 2 .x.t0 /; X 0 /; d 2 .x.t1 /; M / in the definitions of
V0 .t; x/; V .£; x/. In this case the term 14 hl; li has to be omitted and the maxima
taken over a unit sphere hl; li 1. These calculations lead to the next result.
2.4 Reachability for Linear-Convex Systems 69
Theorem 2.4.4. Let Vf .t; x/; Vb .t; x/ be the solutions of HJB equations (2.66),
(2.73) with boundary conditions
with I .ljB .0// being the indicator function for the unit ball B .0/ and
Z £
ˆf Œ£; x; l D hs.£; t0 ; l/; xi ¡.s.t; t0 ; l/ j B.t /P .t //dt ¡.l j X 0 /:
t0
Similarly,
with
Z t1
ˆb Œ£; x; l D hs.£; t1 ; l/; xi ¡.s.t; t1 ; l/ j B.t /P .t //dt ¡.l j M /:
£
xP 1 D x3 ; xP 3 D .k1 C k2 /x1 C k2 x3 ;
xP 2 D x4 ; xP 4 D k1 x1 .k1 C k2 /x3 C u;
Colliding Tubes
Consider two set-valued positions of system (2.25): the starting position ft0 ; X 0 g
and the terminal position ft1 ; M g with t0 t1 . For each of these one may specify a
reach tube—in forward and backward time, respectively:
X Œt \ W Œt 6D ;; 8t 2 Œt0 ; t1 :
The proof of this statement follows from the definitions of forward and backward
reach sets. This yields the next conclusion.
2.5 Colliding Tubes: Calculating All Reachable Points 71
Fig. 2.2 Two-dimensional projections of the four-dimensional value function with their level sets
72 2 The Dynamic Programming Approach
Z Œt D X Œt \ W Œt ; t 2 Œt0 ; t1 ;
consists of all trajectories of system (2.1) that connect the set X 0 at t D t0 to the set
M at t D t1 : Here
Z Œt D fx W V0 .t; x/ 0; V .t; x/ 0g D X Œt \ W Œt ;
for all t 2 Œt0 ; t1 ; V0 .t; x/ is the value function for the forward reach set (2.46) and
V .t; x/ is the value function for the backward reach set (2.55).
In particular Z Œt0 D X 0 \ W Œt0 ; Z Œt1 D X Œt1 \ M : Note that the knowledge of
tube Z ./ does not require any calculation of the corresponding controls.
The problem of calculating all reachable points may be separated into two
classes, namely the one of calculating the boundary points of the reach set X Œt1 D
X .t1 ; t0 ; x 0 / and that of calculating its interior points.
The first class of problems is easily solved in the following way. Suppose x 1 is
a point on the boundary of X Œt1 . The corresponding support vector (or vectors) l
are given by
Taking any l 2 L and applying the maximum principle in its simplest form (see
Sect. 1.4, (1.46)), we observe that the desired control u .t / satisfies the relation
3
With additional information on P .t / (see Remark 1.5.3), in degenerate cases the control u .t /
may be written down in more detail.
2.5 Colliding Tubes: Calculating All Reachable Points 73
and
Z t1
¡.l j W Œ£/ D hl; G.£; t1 /x i C 1
¡.l j G.£; s/B.s/P .s//ds; (2.86)
£
where tube X Œ£ develops from X Œt0 D x 0 with £ increasing and tube W Œ£
develops from W Œt1 D x 1 , with £ decreasing.
We thus have two intersecting tubes (X Œ£ \ W Œ£ 6D ;; 8£ 2 Œt0 ; ª), heading
towards each other in opposite directions. If x 0 lies on the boundary of W Œt0 , then
we arrive at the situation described in the previous case, Lemma 2.5.2, with x 0 and
x 1 interchanged and problem solved in backward time. Hence we further assume
x 0 2 intW Œt0 .
Then we have x 1 2 intX Œt1 and x 0 2 intW Œt0 and there exists an instant £ 2
Œt0 ; t1 when the boundaries of the two tubes collide. Namely, developing W Œt from
t1 backwards towards t0 , we first have W Œt X Œt till we reach instant t D £ ,
when W Œt bumps into the boundary of X Œt , touching it from inside at point x , so
that
with
k.£; l; x 0 ; x 1 /
Z £
D hl; G.£; t0 /x 0 G.£; t1 /x 1 i C ¡.l j G.£; s/B.s/P .s/ds
t0
Z t1
¡.l j G.£; s/B.s/P .s//ds;
£
74 2 The Dynamic Programming Approach
and
With x 0 ; x 1 fixed, instant £ will then be the largest root of equation ”.£; x 0 ; x 1 / D
0—the first time when W Œ£ touches X Œ£ from inside.
Function ”.£; x 0 ; x 1 / is defined for £ 2 Œt0 ; t1 , with ”.t0 ; x 0 ; x 1 / <
0; ”.t1 ; x 0 ; x 1 / > 0, and it is continuous in £. Therefore, point £ , where
”.£ ; x 0 ; x 1 / D 0 exists. The maximizer l of (2.88) is the support vector to X Œ£
at touching point, which is precisely the vector x D x.£ /, that ensures (2.87).
Lemma 2.5.3. The first instant £ of collision for W Œ£ X Œ£ with boundary of
tube X Œ£ from inside is the largest root of equation ”.£ ; x 0 ; x 1 / D 0. The point x
of such collision satisfies the maximum rule
The calculation of the control u .t / which steers x.t / from x.t0 / D x 0 to x.t1 / D
1
x is now reduced to the class of points considered in Lemma 2.5.2, namely the
points that lie on the boundaries of the respective reach sets.
Here we first specify the control u .t / for interval t 2 Œt0 ; £ , working with
point x , which lies on the boundary of set X Σ . Then u .t / satisfies the maximum
principle
Applying similar reasoning to the interval Œ£ ; t1 ; with “starting point” x 1 D x.t1 /
and x .£ / being on the boundary of reach set W Œ£ with same support vector
l , we may calculate u .t / for t 2 Σ ; t1 . This leads again to relation (maximum
principle) (2.89), but now taken for t > £ .
Theorem 2.5.1. The control u .t / that steers the trajectory x.t / from x.t0 / D x 0
to x.t1 / D x 1 satisfies the maximum principle relation (2.89), where l is the
minimizer of problem (2.88), with £ D £ being the largest root of equation
”.£; x 0 ; x 1 / D 0:
Here Fig. 2.3 illustrates the idea of colliding tubes.4 Namely, given in green is
the backward reach tube from the terminal point x m at target set M (marked in red).
Given in blue is the forward reach tube emanating from the starting point x 0 located
in set X 0 : The intersection of these two tubes is the union of all system trajectories
that connect the starting point with the terminal point.
4
This example is animated in the toolbox [132].
2.6 The Closed-Loop Target Control 75
with 2 u 2; t 2 Œ0; 5: The time of switching from one tube to the other is
£ D 0:886:
We shall now indicate how to construct feedback strategies for target control by
using backward reach sets.
Here we introduce a scheme for finding the closed-loop control for minimizing at
given time the distance from a given target set. This would require to know the
backward reach set. Once this set is known, there would be no need of integrating
the HJB equation which produces it as a level set. The problem will be therefore
shifted to computation of these level sets without solving the HJB equation. Such
computation could be done through ellipsoidal methods, as explained in the next
chapter.
Consider system (1.2) with target set M . Let W Œ£ D W .£; ª; M / be its
backward reach set (the solvability set) from position fª; M g.
76 2 The Dynamic Programming Approach
Problem 2.6.1. In the class of set-valued feedback strategies UC D fU.t; x/g find
feedback strategy U .t; x/ that ensures the inclusion xλ 2 M , whatever be the
starting position f£; xg; x 2 W Œ£.
From the definition of W Σ it follows that the problem of reaching target set
W Œª D M from position f£; xg is solvable if and only if x 2 W Œ£.
We now construct a solution strategy U .t; x/ for Problem 2.6.1. Let
and let d VC .t; x/=dt ju be the total derivative of VC .t; x/ at position ft; xg, along
the trajectory of Eq. (1.2) under control u.
The solution strategy is defined as
8 ˇ
< arg min d V .t; x/=dt j ˇˇ u 2 P .t / ; if V .t; x/ > 0;
u ˇ
U .t; x/ D
C C
(2.90)
:
P .t /; if VC .t; x/ D 0:
d VC .t; x/ ˇˇ
ˇ 0;
dt u
in which
Note that we need to ensure d VC .t; x/=dt ju 0 only where VC .t; x/ > 0: Now
suppose VC .t; x/ > 0:
2.6 The Closed-Loop Target Control 77
Then let
l 0 .t; x/ D arg maxfhl; G.ª; t /xi ¡.l j G.ª; t /W Œt / j hl; li 1g: (2.93)
Under the constraint hl; li 1 the vector l 0 D l 0 .t; x/ is unique. Applying the rules
for differentiating a “maximum function,” [61], we further have
@VC .t; x/
D G 0 .ª; t /l 0 :
@x
Since
Z ª
xP D A.t /x C B.t /u; W Œt D G.t; ª/M G.t; s/B.s/P .s/ds;
t
so that
@¡.l 0 j G.ª; t /W Œt /
D ¡.l 0 j G.ª; t /B.t /P .t //:
@t
Substituting the results into Eq. (2.92) we finally have
Taking
we get
To prove that U .t; x/ is a solution to Problem 2.6.1, first integrate d VC .t; x/=dt
from £ to ª along the trajectories x Œt of (2.91) emanating from x .£/ D x. Then,
with U .t; x/ selected according to (2.90), one gets
Z ª
d VC .t; x Œt / D VC .ª; x Œª/ VC .£; x/ 0;
£
78 2 The Dynamic Programming Approach
so VC .ª; x Œª/ VC .£; x/, which implies that once VC .£; x/ 0 (which means
x Œ£ D x 2 W Œ£), one will have VC .ª; x Œª/ 0 (which means x Œª 2
W λ D M ).
Note that under VC .t; x.t // 0, with control selected due to (2.90), the
trajectory x.t / never leaves the tube W Œt . So if, say VC .t; x.t // D 0, but at some
point t 0 D t C ¢ we have
then, due to the directional differentiability of function VC .t; x.t //, we would have
at some point t 00 2 .t; t 0 that the derivative d VC .t 00 /=dt > 0; which is not possible
due to (2.90).
Theorem 2.6.1. The closed-loop strategy U .t; x/ which solves Problem 2.6.1 is
defined by (2.90), (2.94), in which l 0 D l 0 .t; x/ is the optimizer of (2.93) and W Œt
is the backward reach tube from position fª; M g.
Remark 2.6.1. A strategy U .t; x/ given by (2.94), (2.93), is said to be produced
by the extremal aiming rule (see [121]).
Exercise 2.6.1. Find the closed-loop strategy that steers point x 2 W Σ to a given
point x 2 M (provided x; x are such that the problem is solvable).
Problem 2.7.1 may be solved by calculating the set WŒ£; ‚. Clearly,
[
WŒ£; ‚ D fW .£; ª; M / j ª 2 ‚g: (2.95)
Let
V.£; x/ D
Lemma 2.7.1.
Hence
where ª0 D ª0 .£; x/: The solution to Problem 2.7.2 is now reduced to Problem 2.6.1
with ª D ª0 :
Exercise 2.7.1. Solve Problem 2.7.2 in detail, by applying the scheme of Sect. 2.6
to WŒ£; ‚:
The notion of strong invariance is applicable also to reachability within an
interval.
Definition 2.7.1. A set Ws Œ£; ‚ is strongly invariant relative to set M within
interval ‚, if for each of its points x the entire reach set X .ª; £; x/ D X Œª M
for some ª 2 ‚:
If Ws Œ£; ‚ contains all such points, then it is said to be the maximal strongly
invariant set relative to M , within interval ‚.
80 2 The Dynamic Programming Approach
Let us now find the support function ¡.l j Ws .£; ª; M //: Following the scheme of
calculating (2.78), (2.80), but with minu substituted for maxu , we get
with
Z ª
ˆs Œ£; ª; x; l D hs.£; ª; l/; xi C ¡.s.t; ª; l/ j B.t /P .t //dt ¡.l j M /:
£
Hence
Ws .£; ª; M / D fx W ˆs Œ£; ª; x; l 0g
for all hl; li 1 and therefore, since the function ˆs Œ£; ª; x; l is homogeneous in
l, also for all l 2 Rn :
After a substitution similar to (2.71) we observe that 8x 2 Ws .£; ª; M /,
Z ª
hl; xi ¡.l j G.t; ª/M / ¡.l j G.t; Ÿ/B.Ÿ/P .Ÿ//d Ÿ D k.t; ª; l/: (2.104)
t
Theorem 2.7.1. The value function Vs .£; x/, (2.99), that determines Ws Œ£; ‚
through (2.100) is given by (2.101)–(2.103).
The next item is reachability within an interval (see Remark 1.4.2).
Definition 2.7.2. The reach set XŒ‚ D X.‚; t0 ; X 0 / of system (1.44) under
constraint u.s/ 2 P .s/; s t0 ; within given time-interval ‚ D Œ’; “; “
’ t0 , from set-valued position ft0 ; X 0 g, is the set of all points x, for each of
which there exists a trajectory xŒs D x.s; t0 ; x 0 /, for some x 0 2 X 0 ; generated
by some control subjected to the given constraint, that transfers the system from
position ft0 ; x 0 g to position fª; xg; x D xŒª, for some ª 2 ‚:
Lemma 2.7.4. The reach set XŒ‚ within a given time interval ‚ is the level set
The exact calculation of sets XŒ‚; WŒ£; ‚ may be achieved through formu-
las (2.106), (2.96) applied to functions V0 .£; x/; V .£; x/ given in (2.68), (2.77).
Definition 2.7.4. The total backward reachability set from position fª; M g is the
union
[
W Œª D W .ª; M / D fW .£; ª; M / j £ 2 .1; ªg:
2.8 Dynamic Programming: Time-Optimal Control 83
This is the set of all points from which set M may be reached at time ª from
some finite £ 2 .1; ª. For an autonomous system the set W Œª does not depend
on ª.
Denote V .ª; x/ D inf£ fV .£; xjª; V .ª; // j £ 2 .1; ªg; V .ª; x/ D
d 2 .x; M /.
Lemma 2.7.6. The total backward reachability set W λ is the level set
Definition 2.7.5. The total forward reachability set XC Œt0 from position ft0 ; X 0 g
is the union
[
XC Œt0 D XC .t0 ; X 0 / D fX .ª; t0 ; X 0 / j ª 2 Œt0 ; 1/g:
This is the set of all points that can be reached at some time ª t0 . For an
autonomous system the set XC Œt0 does not depend on t0 .
Denote VC .t0 ; x/ D inffV0 .ª; x jt0 ; V0 .t0 ; // j ª 2 Œt0 ; 1/g; where V0 .t0 ; x/ D
d 2 .x; X 0 /:
Lemma 2.7.7. The total reachability set XC Œt0 is the level set
We shall now use the notion of reachability to solve the problem of closed-loop
time-optimal control.
Problem 2.8.1. Given system (2.25), hard bound u 2 P , starting position ft; xg and
target set M ; x … M , find
Here target set M is taken similar to the above, being convex and compact.
In particular, one may have M D fmg as an isolated point.
84 2 The Dynamic Programming Approach
The solution scheme is as follows: for the given starting position ft; xg; we
construct the reachability tube X Œ£ D X .£; t; x/; £ t; then find the first instant
of time £ .t; x/ when X Œ£ touches M . The value £ .t; x/ t will be the minimal
time for reaching set M from position ft; xg, namely, £ .t; x/ D £0 .t; x/.
To calculate X Σ, recall problem (2.46), (2.47). Then
where function
Let us now look how to find the first instant of time, when X Œt touches M .
From Sect. 2.3 it follows that for inclusion m 2 X Σ to be true it is necessary and
sufficient that V0 .£; m/ D 0. Now, having noted that V0 .t; m/ > 0, solve equation
V0 .£; m/ D 0 and find the smallest root £0 > t of this equation. The number £0
exists if point m is reachable in finite time.
Suppose set M D fmg; then £0 D £0 .t; x/ is already the solution to
Problem 2.8.1. If not, consider equation
Here f .£/ D f .£ j t; x/ > 0 and f .£/ D 0 for some £ > 0, provided set M is
reachable in finite time.5
Taking the smallest root £0 .t; x/ of equation
f .£ j t; x/ D 0; (2.113)
5
A closed set Q is said to be reachable in finite time if the intersection Q \ XC .t; x/ 6D ; for
some £ > t. Here XC .t; x/ is the total (forward) reachability set from position ft; xg:
2.8 Dynamic Programming: Time-Optimal Control 85
Example 8.1
xP D u C f .t /; juj ;
Z t Z t
D maxfl.z x/ jljds lf .s/ds j jlj 1g:
l 0 0
This gives:
Rt Rt
z x t 0 f .s/ds if z > x C 0 f .s/ds;
V0 .t; z/ D Rt Rt
z C x t C 0 f .s/ds if z < x C 0 f .s/ds:
Then one may notice that the minimal time for reaching m D 0 is £0 .t; x/ D
x=. However, with starting position ft; x g instead of ft; xg, where x D x C • >
x; • > 0; the minimal time will be £0 .t; x / D £0 .t; x/ C 1 C .– C •/=.r C /.
Hence, function £0 .t; x/ is discontinuous in x. This example yields the next remark.
Remark 2.8.1. In general the minimal time £0 .t; x/ is not robust relative to the
starting point x.
Exercise 2.8.1. Indicate additional conditions for Problem 2.8.1 that ensure conti-
nuity of function £0 .t; x/ in ft; xg (Fig. 2.4).
Chapter 3
Ellipsoidal Techniques: Reachability
and Control Synthesis
Abstract This chapter describes the ellipsoidal techniques for control problems
introduced in earlier chapters. We derive formulas for reachability sets using the
properties of ellipsoids and relations from convex analysis. The formulas are derived
through inductive procedures. They allow calculation of both external and internal
ellipsoidal approximations of forward and backward reachability sets with any
desired level of accuracy. The approximations are illustrated on examples explained
in detail, then followed by ellipsoid-based formulas for problems of reachability and
control synthesis.
We thus start to introduce an array of methods for treating trajectory tubes rather
than isolated trajectories. Of special importance are methods that permit systems of
high dimensions in limited time. Such types of examples are presented in Chap. 4,
being solved with the aid of the ellipsoidal toolbox, [132].
The values u of the controls are restricted for almost all t by hard bounds
u 2 Q .t /; (3.2)
under condition
x .0/ D x.t0 / 2 X 0 ;
where X 0 2 convRn .
In this chapter we further restrict the constraints on u; x 0 to be ellipsoidal-valued,
namely:
and
where continuous functions q.t / and vector x 0 are given together with continuous,
positive, symmetric matrix-valued function Q0 .t / D Q.t / > 0 and matrix .X 0 /0 D
X 0 > 0. Here the respective ellipsoids are nondegenerate.
In terms of inclusions we have
8l 2 Rn :
Remark 3.1.1. Note that constraints in the form (3.8), (3.9) also allow degenerate
matrices Q.t / 0; X 0 0. In this case ellipsoids E .q.t /; Q.t // or E .x 0 ; X 0 / turn
out to be elliptical cylinders. This case will be discussed separately in Chap. 4.
Let us now start with the reachability issue described in Chap. 1, Sect. 1.4. Note
that we further take q.t / 6D 0 and presume f .t / 0, since keeping it here does not
add much to the procedure. The case f .t / 6D 0 will be important in Chap. 10, while
considering systems under uncertainty.
The reach set X Œ£ may be treated as the cut X Œ£ D X .£; t0 ; E .x 0 ; X 0 // of the
solution tube X ./ D fX Œt W t t0 g to the differential inclusion .t t0 /;
D< l ; G.£; t /B.t /q.t / > C < l ; G.£; t /B.t /Q.t /B 0 .t /G 0 .£; t /l >1=2
3.2 Ellipsoidal Approximation of Reach Sets 91
for all t 2 Œt0 ; £, and the “maximum” (transversality) condition for the initial state:
Here l is the support vector for set X Œt at point x that satisfies relation (3.14).
The product l 0 G.£; t / D sŒt D s.t; £; l/ may be presented as the backward solution
to the “adjoint system”
sP D sA.t /; sΣ D l 0 ;
where s is a row-vector.
Remark 3.1.2. The ellipsoidal nature of the constraints on u.t /; x .0/ yield unique-
ness of the optimal control and trajectory. (Prove this fact.)
Observe that although the initial set E .x 0 ; X 0 / and the control set E .q.t /; Q.t //
are ellipsoids, the reach set X Œt D X .t; t0 ; E .x 0 ; X 0 // will not generally be an
ellipsoid. (Since already the sum of two ellipsoids, E1 C E2 , is generally not an
ellipsoid.)
As indicated in [174], the reach set X Œt may be approximated both externally
and internally by ellipsoids E and EC , with E X Œt EC .
Definition 3.2.1. An external approximation EC of a reach set X Œt is tight if there
exists a vector l 2 Rn such that
The last definition is relevant for reach sets which are compact convex bodies,
symmetrical around the center x ? .t / (see formula (3.14) of the above). However it
does not produce a unique ellipsoid. A more precise notion is given through the next
definitions, where tightness is defined within a certain subclass EC of ellipsoids that
does not coincide with the variety of all possible ellipsoids.
Definition 3.2.2. An external approximation EC is tight in the class EC , if for any
ellipsoid E 2 EC , the inclusions X Œt E EC imply E D EC .
This paper is concerned with external approximations, where class EC D fEC g
is described through the following definition.
92 3 Ellipsoidal Techniques: Reachability and Control Synthesis
Definition 3.2.3. The class EC D fEC g consists of ellipsoids that are of the form
EC Œt D E .x ? ; XC Œt /, where x ? .t / satisfies the equation1
and
XC Œt D XC .t jpŒ/ D (3.18)
Z t Z t
D p.s/ds C p0 .t / p 1 .s/G.t; s/B.s/Q.s/B 0 .s/G 0 .t; s/ds
t0 t0
Cp01 .t /G.t; t0 /X 0 G 0 .t; t0 / :
Here X 0 ; Q.s/; s 2 Œt0 ; t are any positive definite matrices with function Q.s/
continuous, q.t / is any continuous function, p.s/ > 0; s 2 Œt0 ; t ; is any integrable
function and p0 .t / > 0.
In particular, this means that if an ellipsoid E .x ? ; XC /
X Œt is tight in EC , then
there exists no other ellipsoid of type E .x ; kXC /; k 1 that satisfies the inclusions
?
Moreover,
1
In the following formulas symbol pŒ stands for the pair fp0 ; p.s/; s 2 Œt0 ; t g.
3.2 Ellipsoidal Approximation of Reach Sets 93
Relations (3.19), (3.20) are true for all p0 0 and p.s/ ranging over all
continuous positive functions. These results follow from ellipsoidal calculus (see
[174], Sects. 2.1, 2.7). What also holds is the following characterization of tight
external ellipsoids.
Theorem 3.2.2. For a given time £ the tight external ellipsoids EC Œ£ D
E .x ? .£/; XC Œ£/, with EC Œ£ 2 EC , are those for which the functions p and
p0 are selected as
so that
¡.ljX Œ£/ D< l; x ? .£/ > C < l; XC Œ£l >1=2 D ¡.ljE .x ? .£/; XC Œ£//:
Remark 3.2.1. Note that the result above requires the evaluation of the integrals
in (3.18) for each time £ and vector l. If the computation burden for each evaluation
of (3.18) is C , and we estimate the reach tube via (3.19) for N values of time £ and
L values of l, the total computation burden would be C N L. A solution to the
next problem would reduce this burden to C L.
Problem 3.2.1. Given a unit-vector function l .t /; < l ; l >D 1, continuously
differentiable in t , find an external ellipsoid EC Œt
X Œt that would ensure for all
t t0 ; the equality
Cp01 .t /G.t; t0 /X 0 G 0 .t; t0 / ;
and
pt .s/ D< l .t /; G.t; s/B.s/Q.s/B 0 .s/G 0 .t; s/l .t / >1=2 ; (3.25)
D< l; G.t0 ; s/B.s/Q.s/B 0 .s/G 0 .t0 ; s/l >1=2 D p .s/I p0 .t / D< l; X 0 l >1=2 D p0 ;
and
XC Œt D G.t; t0 /XC .t /G 0 .t; t0 /; (3.27)
Z t
XC Œt D p .s/ds C p0 ‰.t /; (3.28)
t0
where
‰.t / D
Z t
D < l; G.t0 ; s/B.s/Q.s/B 0 .s/G 0 .t0 ; s/l >1=2 G.t0 ; s/B.s/Q.s/B 0 .s/G 0 .t0 ; s/dsC
t0
C < l; X 0 l >1=2 X 0 :
In this particular case pt .s/ does not depend on t (pt0 .s/ D pt00 .s/ for t 0 6D t 00 ) and
the lower index t may be dropped.
Direct differentiation of XC Œt yields
XP C Œt D (3.29)
where
Z t 1
.t / D p .t / p .s/ds C p0 :
t0
Calculating
Z t Z t 2
< l; XC Œt l >D p .s/ds C p0 < l; ‰.t /l >D p .s/ds C p0 ;
t0 t0
In order to pass to the matrix function XC Œt we note that
XP C
Œt D A.t /G.t; t0 /XC Œt G 0 .t; t0 /CG.t; t0 /XC Œt G 0 .t; t0 /A0 .t /CG.t; t0 /XP C Œt G 0 .t; t0 /:
XP C
D A.t /XC 0
C XC A .t /C (3.31)
C .t /XC
C 1 .t /B.t /Q.t /B 0 .t /; XC
.t0 / D X 0 :
We shall denote the ellipsoid constructed with matrix XC .t / described by
?
Eqs. (3.31), (3.30), as E.x .t /; XC .t //. We now summarize the last results.
Theorem 3.3.1. Under Assumption 3.3.1 the solution to Problem 3.2.1 is given by
the ellipsoid EC Œt D E.x ? .t /; XC Œt /, where x ? .t / satisfies Eq. (3.17) and XC Œt
is a solution to Eqs. (3.31), (3.30).
Since set XC Œt depends on vector l 2 Rn , we further denote XC Œt D XC Œt l ,
using it in the following text whenever it will be necessary.
Theorem 3.3.2. For any t t0 the reach set X Œt may be described as
This is a direct consequence of Theorem 3.2.2 and of the selection of good curves
for representing the solution. Differentiating .t /, according to (3.30), we arrive
at the next result.
Corollary 3.3.1. If B.t / B const and function .t / is differentiable, then it
satisfies the differential equation
where
P /B 0 .t //G 0 .t0 ; t /l >
< l; G.t0 ; t /.A.t /B.t /Q.t /B 0 .t / B.t /Q.t /B 0 .t /A0 .t / C B.t /Q.t
f .t / D 0 0
:
2 < l; G.t0 ; t /B.t /Q.t /B .t /G .t0 ; t /l >
f .t / D œ:
In the general case, differentiating relation (3.24) for XC Œt , we have
XP C
Œt D
Z t Z t
D pt .t /C .@pt .s/=@t /dsCpP0 pt1 .s/G.t; s/B.s/Q.s/B 0 .s/G 0 .t; s/ds
t0 t0
Cp01 .t /G.t; t0 /X 0 G 0 .t; t0 /
Z t
C pt .s/ds C p0 .t / .pt .t //1 B.t /Q.t /B 0 .t /C
t0
Z t
[email protected] .s//1 G.t; s/B.s/Q.s/B 0 .s/G 0 .t; s//=@t /dsC
t0
Cd.p01 .t /G.t; t0 /X 0 G 0 .t; t0 //=dt ;
0 0
G.t; s/B.s/Q.s/B .s/G .t; s/ds C p01 .t /G.t; t0 /X 0 G 0 .t; t0 / ;
Z t 1
.t / D pt .t / pt .s/ds C p0 .t / ; (3.34)
t0
gives
XP C
D (3.35)
0
D A.t /XC C XC A .t / C .t /XC
C 1 .t /B.t /Q.t /B 0 .t /
Here
Z t Z t
1 1
C pt .s/ds Cp0 .t / [email protected] .s// =@t /P .t0 ; s/ds C.d.p0 .t // =dt /P0
t0 t0
and
It is important to note that under Assumption 3.3.1 (l .t / D G 0 .t0 ; t /l) the terms
@pt .s/=@t D 0; pP0 .t / D 0, so that ¥.t; l.t /; B./Q./B 0 .// D 0.
Recall that an external ellipsoid constructed in the general case, with XC .t /
?
taken from Eq. (3.35), (3.34), is denoted as E .x .t /; XC .t //, in contrast with an
ellipsoid constructed along good curves, with XC .t / taken from Eqs. (3.31), (3.30)
and denoted earlier as EC Œt :
Theorem 3.3.3. The solution EC Œt D E .x ? .t /; XC .t // to Problem 3.2.1 is given
?
by vector function x .t / of (3.17) and matrix function XC .t / that satisfies Eq. (3.35),
with .t / defined in (3.34) and pt .s/; p0 .t / in (3.25). Under Assumption 3.3.1
Eq. (3.35) appears with term ¥ 0.
Throughout the previous discussion we have observed that under Assump-
tion 3.3.1 the tight external ellipsoidal approximation E .x ? .t /; XC .t // is governed
by the simple ordinary differential equations (3.31). Moreover, in this case the
points x l .t / of support for the hyperplanes generated by vector l.t / run along a
system trajectory of (3.1) which is generated by a control that satisfies the Maximum
Principle (3.15) and the relation (3.16) of Theorem 3.1.1.
In connection with this fact the following question arises:
Problem 3.3.1. Given is a curve l.t / such that the supporting hyperplane generated
by vectors l.t / touches the reach set X Œt at the point of support x l .t /. What should
be the curve l.t /, so that x l .t / would be a system trajectory for (3.1)?
Let us investigate this question when x 0 D 0; q.t / 0: Suppose a curve l.t / is
such that (xtl0 D x l .t0 /)
Z t
x .t / D
l
G.t; t0 /xtl0 C G.t; £/B.£/ut .£/d £ (3.37)
t0
is a vector generated by the pair fxtl0 ; ut ./g. In order that x l .t / be a support vector
to the supporting hyperplane generated by l.t / it is necessary and sufficient that this
3.3 Recurrent Relations: External Approximations 99
pair satisfy at each time t the maximum principle (3.15) and the condition (3.16)
under constraints
which means
with t0 £ t; and
X 0 G 0 .t; t0 /l.t /
xtl0 D : (3.39)
< l.t /; G.t; t0 /X 0 G 0 .t; t0 /l.t / >1=2
Note that since x l .t / varies in t , the variables xtl0 D x l .t0 /; ut .£/ should in general
also be taken as being dependent on t .
Now the question is: what should be the function l.t / so that x l .t / of (3.37)–
(3.39) is a trajectory of system (3.1)?
Differentiating (3.37), we come to
xP l .t / D A.t / G.t; t0 /xtl0 C (3.40)
Z t
C G.t; £/B.£/ut .£/d £ C B.t /ut .t / C Y .t /;
t0
where
Z t
Y .t / D G.t; t0 /.@xtl0 =@t / C G.t; £/B.£/.@ut .£/=@t /d £: (3.41)
t0
Yl .t / D G.t; t0 /[email protected] 0 G 0 .t; t0 /l.t /.p0 .t; l.t ///1 /=@t /C (3.42)
Z t
C G.t; £/B.£/[email protected].£/B 0 .£/G 0 .t; £/l.t /.pt .£; l.t ///1 /=@t /d £;
t0
100 3 Ellipsoidal Techniques: Reachability and Control Synthesis
where, as in (3.25),
pt .s; l.t // D< l.t /; G.t; s/B.s/Q.s/B 0 .s/G 0 .t; s/l.t / >1=2 ; (3.43)
where
D E
ˆ0 .t; t0 ; l/ D P0 .t; t0 /.A0 .t /l C d l=dt / l; G.t; t0 /X 0 G 0 .t; t0 /l
P0 .t; t0 /l .l 0 A.t / C d l 0 =dt /P0 .t; t0 /l ;
and
D E
ˆ.t; £; l/ D P .t; £/.l 0 A.t / C d l 0 =dt /0 l; P .t; £/l
P .t; £/l .l 0 A.t / C d l 0 =dt /P .t; £/l :
D .l00 .t /A.t / C d l00 .t /=dt /P0 .t; t0 /.A0 .t /l0 .t / C d l0 .t /=dt / < l0 .t /; P0 .t; t0 /l0 .t / >
.l00 .t /P0 .t; t0 /.A0 .t /l0 .t / C d l0 .t /=dt //2 p0 .t; l0 .t //3 C
Z t
C .l00 .t /A.t / C d l00 .t /=dt /P .t; £/.A0 .t /l0 .t / C d l0 .t /=dt / < l0 .t /; P .t; £/l0 .t / >
t0
..l00 .t /A.t / C d l00 .t /=dt /P .t; £/l/2 .pt .£; l0 .t ///3 d £;
should be equal to zero. But due to the Hölder inequality (applied in finite-
dimensional version for the first term and in infinite-dimensional version for the
integral term) the above expression is equal to zero if and only if l0 .t / satisfies
relation (3.45) for some k.t /: .A0 .t /l C d l.t /=dt / is collinear with l.t /.2 This
contradicts the assumption Yl .t / 0 and proves the following issues.
Theorem 3.3.4. Given a curve l.t /, the function xl .t / of (3.37)–(3.39), formed of
support vectors to the hyperplanes generated by l.t /, is a system trajectory for (3.1)
if and only if l.t / is a solution to the differential equation (3.45).
Through similar calculations the following assertion may be proved
Theorem 3.3.5. In order that G.t; t0 /¥.t; l.t /; B./Q./B 0 .//G 0 .t; t0 / 0; it is
necessary and sufficient that l.t / satisfies Eq. (3.45).
Exercise 4.1. Work out the proofs of Theorems 3.3.4, 3.3.5 in detail.
We have thus come to the condition that l.t / should satisfy (3.45). This equation
may be interpreted as follows. Consider the transformation
Z t
lk .t / D exp. k.s/ds/l.t /: (3.46)
t0
2
Here we also take into account that .l 0 .t /A.t / C d l.t /=dt /Yl .t / is bounded with p0 .t / > 0 and
pt .£/ > 0 almost everywhere due to the controllability assumption.
102 3 Ellipsoidal Techniques: Reachability and Control Synthesis
Then
Z t Z t
d lk .t /=dt D exp. k.s/ds/d l.t /=dt k.t / exp. k.s/ds/l.t / D
t0 t0
Z t Z t
D .A0 .t / C k.t /I / exp. k.s/ds/l.t / k.t / exp. k.s/ds/l.t / D A0 .t /lk .t /
t0 t0
which means that lk .t / is again a solution to the adjoint equation d l=dt D A0 .t /l,
but in a new scale. However, the nature of transformation (3.46) is such that lk .t / D
”.t /l.t /; ”.t / > 0.
Lemma 3.3.1. The solutions lk .t / to Eq. (3.45) generate the same support hyper-
planes to the reach set X Œt and the same points of support x lk .t /, whatever be
the functions k.t / (provided all these solutions have the same initial condition
lk .t0 / D l0 ).
In particular, with k.t / 0, we come to functions l.t / of Assumption 3.3.1 and
with Al D kl (k is an eigenvalue of constant matrix A) to equation d l=dt D 0 and
to condition l.t / D l D const:
In general the function l.t / of Assumption 3.3.1 is not normalized so that <
l.t /; l.t / >6D 1. In order to generate a function lu .t / with vectors lu .t / of unit length,
take the substitution lu .t / D l.t / < l.t /; l.t / >1=2 , where d l=dt C A0 .t /l D 0.
Direct calculations indicate the following statement.
Lemma 3.3.2. Suppose function lu .t / satisfies Eq. (3.45) with
where Q .t / D E .q.t /; Q.t //, so that the set-valued function X Œt satisfies for all t
the relation
lim ©1 h.X Œt C ©; .I C ©A.t //X Œt C ©E .B.t /q.t /; B.t /Q.t /B 0 .t /// D 0;
©!0
(3.48)
with initial condition X Œt0 D E 0 [158].
Here, as before, h.Q ; M / stands for the Hausdorff distance between sets Q ; M .
Recall that
lim ©1 h .E Œt C ©; .I C ©A.t //E Œt C ©E .B.t /q.t /; B.t /Q.t /B 0 .t /// D 0
©!0
(3.49)
with initial condition E Œt0 D E .x 0 ; X 0 / D E 0 .
Definition 3.4.1. A set-valued function EC Œt is said to be the solution to
Eq. (3.49) if
(i) EC Œt satisfies (3.49) for all t ,
(ii) EC Œt is ellipsoidal-valued and EC Œt 2 EC .
A solution to Eq. (3.49) is said to be minimal in EC if together with
conditions (i), (ii) it satisfies condition
(iii) EC Œt is a minimal in EC solution to (3.49) with respect to inclusion.
Condition .i i i / means that there is no other ellipsoid E Œt 6D EC Œt ; E Œt 2 EC
that satisfies both Eq. (3.49) and the inclusions EC Œt
E Œt
X Œt . This also
means that among the minimal solutions EC Œt to (3.49) there are non-dominated
(tight) ellipsoidal tubes EC Œt that contain X Œt . For a given initial set E 0 D E Œt0
the solutions to (3.49) as well as its minimal solutions may not be unique.
Remark 3.4.1. Note that Eq. (3.49) is written in terms of Hausdorff semidistance
h rather than Hausdorff distance h as in (3.48).
Denote an ellipsoidal-valued tube EŒt D E .x ? .t /; X Œt /, that starts at
ft0 ; E 0 g; E 0 D E .x 0 ; X 0 / and is generated by given functions x ? .t /; X Œt (matrix
X Œt D X 0 Œt > 0) as EŒt D E.t jt0 ; E 0 / D E .x ? .t /; X Œt /.
The evolution Eq. (3.49) defines a generalized dynamic system in the following
sense.
104 3 Ellipsoidal Techniques: Reachability and Control Synthesis
Lemma 3.4.1. Each solution E.t jt0 ; E 0 / D EŒt to Eq. (3.49), in the sense of
Definition 3.4.1, defines a map with the semigroup property
This follows from Definition 3.4.1. Equation (3.49) implies that a solution EŒt
to this equation should satisfy the inclusion
where –1 o.–/ ! 0 with – ! 0 and B .0/ D fx W hx; xi 1g is the unit ball in Rn .
Further denote qB .t / D B.t /q.t /; QB .t / D B.t /Q.t /B 0 .t /.
Given a continuously differentiable function l.t /; t t0 ; and assuming that
EŒt D E .x ? .t /; X Œt / is defined up to time t , let us select EŒt C – as an external
ellipsoidal approximation of the sum
requiring that it touches this sum Z.t C –/ (which, by the way, is not an ellipsoid),
according to relation
that is at those points z.t C–/ where Z.t C–/ is touched by its supporting hyperplane
generated by vector l.t C –/. Namely,
fz.t; –/ W< l.t C –/; z.t; –/ >D< l.t C –/; x ? .t C –/ > C < l.t C –/; X Œt C –l.t C –/ >1=2 D
D< l.t C –/; x ? .t C –/ > C < l.t C –/; .I C –A.t //X Œt .I C –A.t //0 l.t C –/ >1=2 C
In the latter case, according to the definitions of tightness, the ellipsoid EŒt C –
is among the tight external approximations for Z.t C –/ (relative to terms of order
higher than –).3
The requirement above is ensured if EŒt C – externally approximates the sum
Z tC–
Z .t C –/ D G.t C –; t /E .x .t /; X Œt / C
?
G.t C –; s/E .qB .s/; QB .s//ds
t
3
Here and in the sequel the terms of type oi .–/ are assumed to be such that oi .–/–1 ! 0 with
– ! 0.
3.4 The Evolution of Approximating Ellipsoids 105
X Œt C – D .1 C p 1 .t C –//G.t C –; t /X Œt G 0 .t C –; t /C (3.52)
where
Indeed, with p.t C–/ chosen as in (3.53), a direct substitution of p.t C–/ into (3.52)
gives
< l.t C–/; X Œt C–l.t C–/ >1=2 D< l.t C–/; G.t C–; t /X Œt G 0 .t C–; t /l.t C–/ >1=2 C
(3.54)
C– < l.t C –/; G.t C –; t /QB .t /G 0 .t C –; t /l.t C –/ >1=2 :
In view of (3.53), (3.55), and the fact that G.t C –; t / D I C –A.t / C o2 .–/, we arrive
at the desired result below.
Lemma 3.4.3. With p.t C –/ selected as in (3.53), relations (3.52)–(3.55) reflect
equalities
where
Remark 3.4.2. The ellipsoid EŒt of 3.4.1 solves Problem 3.2.1 for any (non-
normalized) curve l.t /. In order to have a solution for a normalized curve with
< l.t /; l.t / >D 1; one has to follow Lemmas 3.3.1, 3.3.2.
In the general case an ellipsoidal tube EC Œt D E .x ? ; XC Œt / may be such that
it touches X Œt along some curve according to the requirement of Problem 3.2.1.
But this curve may not be a good one. Then equations for XC Œt will differ
from (3.31), (3.59), but will coincide with (3.35), where ¥ 6D 0. And EC Œt need
not satisfy the semigroup property. An example of such case is when the touching
curve is “drawn” on the surface of exact reach tube by an external minimum-volume
ellipsoid. This touching curve is not a good one. The failure to distinguish good
curves from any curves led to confusion and controversial statements in some
publications in the 1980s on minimum-volume external ellipsoidal approximations.
But what if Eqs. (3.31), (3.59) are still used to approximate the exact reach set
from above, disregarding the effect of being tight? Then the ellipsoid EC Œt under
consideration will be a conservative upper estimate for the exact reach set so that
X Œt EC Œt , with no promise of being tight.
We may now summarize the earlier results emphasizing that they were directed
at the calculation of reachability tubes rather than fixed-time reach sets. Theo-
rems 3.4.1, 3.4.2, lead to the next proposition.
Theorem 3.5.1. The solution of Problem 3.2.1 allows the following conclusions:
(i) the ellipsoid EC Œt D E .x ? .t /; XC .t // constructed from Eqs. (3.24), (3.25)
for any given l .t / D l.t /; XC .t0 / D X 0 ; is a solution to Problem 3.2.1 and
XC .t / satisfies the “general” equation (3.35) with parameters .t / defined
in (3.34).
(ii) The ellipsoid EŒt D E.x ? .t /; XC .t // constructed from Eqs. (3.31), (3.30)
(or (3.59), (3.60), which are the same) is an upper bound for the solution
EC Œt D E .x ? .t /; XC .t // of (3.24), (3.25), so that for any given function
l .t / we have
¡.l.t /jE.x ? .t /; XC .t /// ¡.l .t /jE .x ? .t /; XC
.t ///; (3.61)
Remark 3.5.1. If in the previous theorem l .t / is chosen with k.t / 0, and XC .t /
is its related matrix with .t / being its related parameterizing function, then we
further drop the asterisks, assuming l .t / D l.t /; XC .t / D XC .t /; .t / D .t /
and are using Eqs. (3.31), (3.30) (or (3.59), (3.60)). For vector x ? .t / of the center of
ellipsoid E .x ? .t /; XC .t // the asterisk remains since it does not depend on l:
The solutions to the problem of the last theorem are tight in the sense of
Definitions 3.2.1, 3.2.4. It is also useful to note the following.
Lemma 3.5.1. Each ellipsoidal tube of type E.t jt0 ; E 0 / D EŒt , generated as a
solution to Eqs. (3.31), (3.30), defines a map with the semigroup property
We now indicate how to calculate the reach tube with the results above.
An application of Theorem 3.1.1 under Assumption 3.3.1 yields the following
conclusion.
Theorem 3.5.2. Suppose Assumption 3.3.1 is fulfilled. Then the points x l .t / of
support for vector l D l .t /, namely, those for which the equalities
< l .t /; x l .t / >D ¡.l .t /jX Œt / D ¡.l .t /jE.x ? .t /; XC Œt // (3.63)
are true for all t t0 , are reached from initial state x l .t0 / D x l0 using a system
trajectory with control u D u .t / which satisfies the maximum principle (3.15) and
condition (3.16) which now have the form l .t / D G 0 .t0 ; t /l
< l; G.t0 ; s/B.s/u .s/ >D maxf< l; G.t0 ; s/B.s/u > ju 2 E .q.t /; Q.t //g;
(3.64)
and
for any t t0 . For all s t0 the control u .s/; s 2 Œt0 ; t may be taken to be the
same, whatever be the value t .
From the maximum principle (Theorem 3.1.1) one may directly calculate the
optimal control u .s/ which is
where
X 0l
x l .t0 / D x l0 D C x 0 D x ? .t0 /: (3.68)
< l; X 0 l >1=2
These relations are not very convenient for calculation, being expressed in a non-
recurrent form.
However, one should note that due to (3.38), (3.39), the trajectory x l .t / of
Theorem 3.5.2 also satisfies the following “ellipsoidal” maximum principle.
Theorem 3.5.3. The next condition holds
and is attained at
and XC Œt may be calculated through Eq. (3.31) with explicitly known parame-
ter (3.30).
Thus the “trick” is to substitute at each point of the boundary of the reach set
X Œt the “original” maximum principle (3.15) of Theorem 3.1.1 by the “ellipsoidal”
maximum principle (3.69).
Theorem 3.5.4. The trajectory x l .t / that runs along the boundary @X Œt of the
reach set X Œt and touches X Œt at the points of support for the vector l .t / D
G 0 .t; t0 /l is given by equalities (3.71), (3.70), where XC Œt is the solution to
Eqs. (3.31), (3.30) (or, what is the same, to Eqs. (3.59), (3.60)).
Due to Lemma 3.3.1 the last theorem may be complemented by
Corollary 3.5.1. Given vector l 0 D l .t /, the related ellipsoid E.x ? .t /; XC Œt /
and vector x l .t / 2 @X Œt ; generated by function l .t / of Eq. (3.45), do not depend
on the choice of function k.t / in this equation.
Denoting x l .t / D xŒt; l, we thus come to a two-parametric surface xŒt; l that
defines the boundary @X Œ of the reachability tube X Œ D [f@X Œt ; t t0 g. With
t D t 0 fixed and l 2 S D fl W< l; l >D 1g varying, the vector xŒt 0 ; l runs along the
boundary @X Œt 0 . On the other hand, with l D l 0 fixed and with t varying, the vector
xŒt; l 0 moves along one of the system trajectories x l .t / that touch the reachability
set X Œt of system (3.1) with control u .t / of (3.66) and x l0 of (3.68). Then
Remark 3.5.2. Relation (3.70) is given in a recurrent form and throughout the
calculation of curves x l Œt and surface xŒt; l one need not compute the respective
controls u .t /.
Remark 3.5.3. This remark concerns the question of optimality of external ellip-
soids. Suppose for each time t a globally volume-optimal ellipsoid E .x ? .t /; Xv .t //
is constructed. It will obviously touch the reach set X Œt at each instant t and thus
produce a certain curve xv .t / on the surface @X Œt of the reach tube X Œ. However,
an example given in [174, Sect. 2.7, Example 2.7.1] indicates that the volume-
optimal ellipsoid may not exist in the class EC . The conclusion which follows is
that in general the volume-optimal curve xv .t / is not a “good one” in the sense
of Assumption 3.3.1. Therefore, one should not expect that matrix Xv .t / of the
volume-optimal ellipsoid to be described by Eq. (3.31) ((3.59)), as claimed by some
authors.
On the other hand, the tight ellipsoids of Sect. 3.2 do satisfy Definition 3.2.1.
This implies
where d Œl is the length of the exact projection of set X Œt on the direction l. Here
matrix XCl .t / is calculated due to (3.31), (3.30), for the given vector l 2 Rn .
3.6 Example 3.6 111
xP 1 D x2 ; xP 2 D u; (3.72)
Here we have:
Z t
x1 .t / D x10 C x20 t C .t £/u.£/d £;
0
Z t
x2 .t / D x20 C u.£/d £:
0
The boundary of the reach set X Œt may be calculated from the formula ([174])
Then the direct calculation of the minimum gives a parametric representation for
the boundary @X Œt by introducing parameter ¢ D l20 = l10 , where l10 ; l20 are the
minimizers in (3.73). This gives two curves
and
x1 .t / D x10 C x20 t .t 2 =2 ¢ 2 /; (3.75)
4
Note that for l20 = l10 > 0 or l20 = l10 < t we have ¢ 62 Œt; 0. For such vectors the point of support
x l Œt will be at either of the vertices of set X Œt .
112 3 Ellipsoidal Techniques: Reachability and Control Synthesis
and
With vector l 2 R2 fixed, this gives a parametric family of curves x l .t / that cover
the surface of the reach tube X Œ and are the points of support for the hyperplanes
generated by vectors l 2 R2 through formula l.t / D .l1 ; t l1 C l2 /. These curves
are shown in thick lines in Fig. 3.2.
3.6 Example 3.6 113
Each of these pairs generates a tight ellipsoid E Œt and E Œt due to Eq. (3.71),
where the elements xi;j of XC satisfy the equations
xP 11 D x12 C x21 C .t /x11 ; xP 12 D x22 C x12 ;
xP 21 D x22 C .t /x21 ; xP 22 D .t /x22 C ..t //1 2 ;
with XC .0/ D 0 and
.t / D f1 .t /=f2 .t /;
with
2
f1 .t / D jl2 t l1 jI f2 .t / D .x11 l1 C 2x12 l1 .l2 t l1 / C x22 .l2 t l1 /2 /1=2 :
E Œt \ E Œt X Œt
an internal approximation of the sum of any finite number of ellipsoids and going
further of an integral of ellipsoidal-valued functions. It is also important, as in the
external case, to single out families of tight internal approximations of reach tubes
or their neighborhoods through such ellipsoidal-valued functions that would touch
their boundary from inside at any point on their surface. Both of the questions are
answered in the forthcoming sections.
Thus, given as before is system (3.1) or differential inclusion (3.3), with
A.t / 0. The problem is to approximate the respective reach sets and tubes
internally, through appropriate tight ellipsoidal sets and ellipsoidal-valued tubes.
The text of this section relies on the next constructions introduced earlier in
[177, 182]. Though described there for nondegenerate ellipsoids, they are true for
degenerate ellipsoids as well.
Consider the internal approximation of sum
Z t
E .x 0 ; X0 / C E .B.s/q.s/; B.s/Q.s/B 0 .s//ds;
t0
X .t / D (3.78)
Z t 0 Z t
1=2 1=2 1=2 1=2
D S0 X 0 C S.£/QB .£/d £ S0 X 0 C S.£/QB .£/d £ :
t0 t0
turns into an equality iff matrices S0 ; S.£/ may be chosen such that for some
scalar function œ.£/ > 0 the equality
1=2 1=2
S.£/QB .£/l D œ.£/S0 X0 l (3.80)
This definition may be appropriate for the reach sets of the present paper, but the
respective ellipsoids may not be unique. A more general definition follows.
Definition 3.7.2. An internal approximation E is tight in the class E , if for any
ellipsoid E 2 E , X Œt
E
E implies E D E .
This section is concerned with internal approximations, where class E D fE g
is described within the following definition.
3.7 Reachability Sets: Internal Approximations 117
Definition 3.7.3. The class E D fE g consists of ellipsoids that are of the form
E Œt D E .x ? .t /; X .t //, where x ? .t / satisfies the equation
of Sect. 3.1.
Then the problem consists in finding the internal ellipsoid E .x ? .t /; X .t // for
the reach set
Z t
X Œt D G.t; t0 /E .x 0 ; X0 / C G.t; £/B.£/E .q.£/; Q.£//d £:
t0
t0
Z t
S.£/QB .£/G .t0 ; £/ G 0 .t; t0 /;
0
1=2 1=2
S0 .t0 /X0 C (3.82)
t0
and
Z t
x ? .t / D G.t; t0 /x 0 C G.t; £/qB .£/d £: (3.83)
t0
Theorem 3.7.2. The internal ellipsoids for the reach set X Œt satisfy the inclusion
where X .t /; x ? .t / are given by (3.82), (3.83), with S0 ; S.£/ being any orthogonal
matrices and S.£/ continuous in time.
The tightness conditions now transfer into the next proposition.
Theorem 3.7.3. For a given instant t the internal ellipsoid E .x ? .t /; X .t // will
be tight and will touch X Œt at the point of support x l of the tangent hyperplane
generated by given vector l , namely,
¡.l jX Œt / D (3.85)
Z t
D< l ; x ? .t / > C < l ; X0 l >1=2 C < l ; G.t; £/QB .£/G 0 .t; £/l >1=2 d £ D
t0
œ.£/ D< l ; G.t; £/QB .£/G 0 .t; £/l >1=2 < l ; G.t; t0 /X0 G 0 .t; t0 /l >.1=2/ ; t0 £ t:
(3.87)
The previous Theorems 3.7.2, 3.7.3 were formulated for a fixed instant of time
t and a fixed support vector l . It is important to realize what would happen if l
varies in time.
Consider system
xP 1 D x2 ; xP 2 D u; (3.88)
Here
Z t
x1 .t / D x10 C x20 t C .t £/u.£/d £;
0
Z t
x2 .t / D x20 C u.£/d £:
0
The boundary @X Œt of the reach set X Œt is the set of vectors such that
This leads to the next parametric presentation of @X Œt through two bounding curves
(one with sign plus in ˙ and the other with minus)
taken for values of parameter ¢ 2 .t; 0/.The values ¢ 62 .t; 0/correspond to two
points—the two vertices of X Œt .
Here ¢ D l20 = l10 where l10 ; l20 are the minimizers in (3.89).
Solving the problem for all t > 0, set l 0 D l.t /. Then
x1 .t / D (3.91)
D –.l1 .t /.t 2 C 1/ C l2 .t /t /=.l12 .t / C .l1 .t /t C l2 .t //2 /1=2 ˙ .t 2 =2 l22 .t /= l12 .t //;
x2 .t / D
where l 2 R2 ; t 0.
Here, for each t , vector l 0 D l.t / is the support vector to X Œt at point x.t / 2
@X Œt . Moreover, with t fixed, and x D x .l/ .t / running along @X Œt (which is a
120 3 Ellipsoidal Techniques: Reachability and Control Synthesis
closed curve in Rn for a given t ), its support vector l 0 D l.t / will sweep out all
directions in Rn . Therefore, considering any function l.t / of t with t t0 , we may
be sure that there exists a corresponding trajectory x .l/ .t / 2 @X Œt for t t0 .
Proceeding further, we select l.t / satisfying Assumption 3.3.1, namely, as l.t / D
0
e A t l . This transforms into l1 .t / D l1 ; l2 .t / D l2 t l1 and (3.91) simplifies to
x1 .t / D (3.92)
D –.l1 C t l2 /=.l12 C l22 /1=2 ˙ .t 2 =2 .t l1 l22 .t //2 = l12 .t //;
0
1
0.2
0.5
0.4
0
0.6
−0.5
0.8
−1
1
−1
−2
−3
X .£/ D
Z t Z t
QB .£/S 0 .£/d £
1=2 1=2
D –I C –I C S.£/QB .£/d £ ; (3.94)
0 0
–2 œ2 .£/ D hl ; QB .£/l ihl ; l i1 D .l1 £ C l2 /2 .l12 C l22 /1 : (3.96)
Denote
£2 l1 C £l2 cos ¥p .£/
p.£/ D QB .£/l D .1 C £2 /1=2 D rp .£/
1=2
;
£l1 C l2 sin ¥p .£/
where
and also
1=2 cos ¥l .£/
l D hl ; l i ; ¥l D ˙ arccos.l1 =.l12 C l22 /1=2 /:
sin ¥l .£/
where
The internal ellipsoids for the reach set X Œt D X .t; 0; X 0 / are shown in Figs. 3.6,
3.7, and 3.8 for X 0 D E .0; –I /; with epsilon increasing from – D 0 (Fig. 3.6) to
– D 0:175 (Fig. 3.7), and – D 1 (Fig. 3.8). The tube in Fig. 3.5 corresponds to the
epsilon of Fig. 3.7. One may also observe that the exact reach sets X .t; 0; f0g/ taken
for – D 0 are located within the sets X .t; 0; X 0 / D X Œt , calculated for – 6D 0 (see
Figs. 3.7 and 3.8).
We start with a question similar to Problem 3.2.1, but now formulated for internal
approximations.
Problem 3.9.1. Given a vector function l .t /, continuously differentiable in t , find
an internal ellipsoid E .x .t /; X .t // X Œt that would ensure for all t t0 ; the
equality
0.5
x2
−0.5
−1
−1.5
−0.8 −0.6 −0.4 −0.2 0 0.2 0.4 0.6
x1
X .t / D (3.102)
Z t
D G.t; t0 / X0 S0t0 .t0 / C G.t0 ; £/QB .£/St0 .£/d £
1=2 1=2
t0
Z t
St .£/QB .£/G .t0 ; £/ G 0 .t; t0 /;
0
1=2 1=2
S0t .t0 /X0 C
t0
124 3 Ellipsoidal Techniques: Reachability and Control Synthesis
0.5
x2
−0.5
−1
−1.5
−0.8 −0.6 −0.4 −0.2 0 0.2 0.4 0.6 0.8
x1
and S0t0 S0t D I; St0 .£/St .£/ I for all t t0 ; £ 2 Œt0 ; t , where
œt .£/ D (3.104)
D< l .t /; G.t; £/QB .£/G 0 .t; £/l .t / >1=2 < l .t /; G.t; t0 /X0 G 0 .t; t0 /l .t / >.1=2/ :
The proof follows by direct substitution. The last relations are given in a “static”
form and Theorem 3.9.1 indicates that the calculation of parameters S0t ; St .£/; œt .£/
has to be done “afresh” for every new instant of time t . We shall now investigate
whether the calculations can be made in a recurrent form, without having to perform
the additional recalculation.
In all the ellipsoidal approximations considered in this book the center of the
approximating ellipsoid is always the same, being given by x ? .t / of (3.83). The
discussions shall therefore concern only the relations for X .t /.
3.9 Reachability Tubes: Recurrent Relations—Internal Approximations 125
1.5
x2 0.5
−0.5
−1
−1.5
−2
−2.5
−2 −1.5 −1 −0.5 0 0.5 1 1.5 2
x1
Remark 3.9.1. Without conditions (3.103), (3.104) equality (3.101) turns into an
inequality
and
œt .£/ D< l; G.t0 ; £/QB .£/G 0 .t0 ; £/l >1=2 < l; X0 l >1=2 : (3.107)
Here the known vectors and functions used for calculating S0t ; St .£/; œt .£/ do
not depend on t . But then the matrix S0t and functions St .£/; œt .£/ do not depend
on t either, no matter what is the interval Œt0 ; t . Therefore, the lower indices t in
S0t ; St ; œt may be dropped and the matrix X .t / D X.l/ will depend only on l.
Then we have
where
K .t / D K 0 .t /K.t /; KP D KP 0 K C K 0 K;
P
and
Z t
S.£/QB .£/G 0 .t0 ; £/d £:
1=2 1=2
K.t / D S0 X0 C (3.108)
t0
XP .l/ D A.t /X.l/ C X.l/ A0 .t / C G.t; t0 /.KP 0 .t /K.t / C K.t /KP 0 .t //G 0 .t; t0 /;
(3.109)
where
E .x ? .t /; X.l/ .t // X Œt ; t t0 ;
X .t0 / D X0 ;
where
0
‰.t; / D ‰ .t; / C ‰ .t; /
and
Z t
1=2 0 1=2 0
‰ .t; / D G.t; t0 / X0 .@S0t .t0 /=@t / C G.t0 ; £/QB .£/.@St .£/=@t /d £
t0
Z t
St .£/QB .£/G 0 .t0 ; £/d £/ G 0 .t; t0 /:
1=2 1=2
S0t .t0 /X0 C
t0
E .w? .£/; W.l/ .£// W Œ£ E .w? .£/; WC.l/ .£//: (3.114)
.l/
while matrix WC .t / satisfies equation
where
.t / D< l.t /; QB .t /l.t / >1=2 < l.t /; WC .t /l.t / >1=2 ; WC .t1 / D M;
.l/ .l/
(3.117)
with l.t / D G 0 .t1 ; t /l; l 2 Rn :
(iii) For each vector l 2 Rn the next equality is true
.l/
¡.ljE .w? .t /; WC .t /// D ¡.ljW Œt /; (3.118)
.l/
(ellipsoid E .w.t /; WC .t // is tight).
(iv) For any vector l an internal ellipsoid E .w? .£/; W.l/ .£// W Œ£ is defined
through matrix
3.10 Backward Reachability: Ellipsoidal Approximations 129
W.l/ .t / D
Z t1
D G.t; t1 / M 1=2
Sm0 G.t1 ; t /QB .£/ 1=2 0
S .£/d £
t
Z t1
Sm M 1=2 S.£/QB .£/1=2 G 0 .t1 ; £/d £ G 0 .t; t1 /;
t
WP .l/ D (3.119)
with
and
S .l/ .t /QB .t /1=2 G 0 .t1 ; t /l < l; G.t1 ; t /QB .t /1=2 G 0 .t1 ; t /l >1=2 D
.l/
[ fE .w? .t /; W.l/ .£//j < l; l > 1g D W Œ£ D \fE .w? .t /; WC .£//j < l; l > 1g:
(3.122)
Fig. 3.9 Reachability tube with one external and one internal approximation
5
Thus example was worked out by P. Gagarinov.
3.11 The Problem of Control Synthesis: Solution Through Internal Ellipsoids 131
0 1 0 1
0 1 0 0 0 0 0 0 1
B 1 0 0 0 0 0 0 0C B 0 C
B C B C
B 0 0 0 0C B 0 C
B 0 1 0 0 C B C
B C B C
B 1 0 9 0 0 0 0 0C B 0 C
A.t / D B C; B.t / D B C;
B 0 0 0 0 0 1 0 0C B 0:3 C
B C B C
B 0 0 1 0 3 0 0 0C B 0 C
B C B C
@ 0 0 0 0 0 0 0 1A @ 0 A
0 0 0 0 1 0 2 0 0
and target set M D E .m; M /. One is to find a synthesizing strategy u D U .t; x/,
that transfers system (3.124) from any position f£; x.£/g; x.£/ 2 W Œ£, to the set
M at time t1 , so that x.t1 / 2 M . The specified strategy u D U .t; x/ must ensure
the existence of solutions to the differential inclusion
The solution to the given problem is well known and indicated in Sect. 2.6. It may
be found, once we know the solvability tube W Œt ; t 2 Œ£; t1 .
Namely, denoting
VC .t; x/ D d.x; W Œt /
and taking the total derivative d VC .t; x/=dt ju at position ft; xg, due to Eq. (3.124),
under control u, one gets the desired strategy as
8
< u W d V .t; x/=dt j 0j u 2 Q .t / ; if V .t; x/ > 0 .x 62 W Œt /;
Ue .t; x/ D
C u C
:
Q .t /; if VC .t; x/ D 0 .x 2 W Œt /:
(3.125)
It will be further indicated that also
arg min fd VC .t; x/=dt ju j u 2 Q .t /g ; if VC .t; x/ > 0 .x 62 W Œt /;
Ue .t; x/ D
Q .t /; if VC .t; x/ D 0 .x 2 W Œt /:
(3.126)
This is the so-called extremal aiming rule (due to Krasovski, see [121, 123]).
We shall now demonstrate that the solution to Problem 3.11.1 may be obtained
by operating only with ellipsoidal functions.
Consider position f£; x.£/g, and a parameterized family of ellipsoidal tubes
E .w? .t /; W.l/ .t //; t 2 Σ; t1 , constructed according to Theorem 3.10.1. For a given
position f£; xg; x D x.£/ the solution to Problem 3.11.1 exists if x 2 W Œ£.
With x 2 int W Œ£ one may choose any control u 2 Q .£/, so the selection of
control solutions will evolve around the boundary of W Σ and beyond. We shall
start by supposing that given is position f£; x g with x D x 2 @W Œ£ on the
boundary of W Œ£. We may then introduce an internal ellipsoid E .w? .£/; W .£//
that touches its boundary at this x : To do this we first note that with A.t / 0
the “good” curves that cover the tube W Œt ; t 2 Œ£; t1 are such that l.t / l ; is
constant. Hence we may further take l D l to be the support vector of x 2 @W Σ
so that the ellipsoid that touches would be with W .£/ D W.l / .£//, namely
hl ; x.t /i D ¡.l j W.l / Œ£/: (3.127)
And such property will hold within the whole interval t 2 Σ; t1 :
¡.l j E .w? .t /; W .l / .t // D hl ; w? .t /i C hl ; W.l / .t /l i1=2 ;
The selection of vector l D l and the related ellipsoid E .w? .t /; W.l / .£// is a
separate element of the solution process which is discussed further. But first recall
the following. Suppose for W .t / D W?0 .t /W? .t / we have
Z t1
1=2
W? .t / D Sm M 1=2 S.Ÿ/QB .Ÿ/d Ÿ; (3.128)
t
Then for any orthogonal-valued matrix Sm and matrix function S.t /; t 2 Σ; t1 the
ellipsoid E .w? .t /; W.l / .t // W Œt will be an internal approximation of W Œt :
However, if for a given vector l l we select Sm , S.t / coordinated according to
relations
(3.129)
1=2
Then, due to these relations, vectors Sm M 1=2 l and S.t /QB .t /l; 8t 2 Σ; t1 will
be collinear. Hence, taking such Sm D Sm ; S.t / D S .t / together with l D l ; and
taking W .t / D W.l / .t /; we will have
Z t1
hl ; W.l / .t /l i1=2
D hl ; M l i 1=2
C hl ; QB .s/l i1=2 ds; (3.130)
t
and since
Also note that if items l.t / and Sm ; S.t / are not coordinated, as in (3.129),
then (3.130), (3.131) will turn into inequalities with D substituted by .
Also observe that for any fl W hl ; l i D 1g, coordinated with related Sm ; S .t /
according to (3.129), we also have
hl ; W.l / .t //l i hl; W.l / .t //li (3.132)
Assumption 3.11.1. Given set E.l / Œ£ D E .w? .£/; W.l / .£// that touches W Œ£
internally, there exists © > 0 such that W.l / .£/ ©I; 8t 2 Œ£; t1 :
Checking whether x D x.£/ 2 W Œ£.
We first find if the distance d Œ£; x D d.x.£/; W Œ£/ > 0. Then
and
or (ii) we may select among internal ellipsoids E.l/ Σ that internally touch the
boundary @W Σ the one that does it for l D l D l 0 . This may not need to calculate
the whole reach set.6 Then vector l with related W.l / Σ are those that satisfy
6
If W Œt is already calculated, the procedure is simple.
3.11 The Problem of Control Synthesis: Solution Through Internal Ellipsoids 135
Here
Z t1
W .t / D W?0 .t /W? .t /;
1=2
W? .t / D W? .t1 / S.Ÿ/QB .Ÿ/d Ÿ;
t
D maxfhl; x w? .t /i ¡.l j E .0; W.l / .t // j hl; li D 1g D hl 0 ; xi hl 0 ; W.l / .t /l 0 i1=2 ;
l
(3.135)
we observe that the maximizer l 0 of this problem is unique and due to (3.132) we
have
hl 0 ; xi D hl 0 ; w? .t /i C hl 0 ; W.l / .t /l 0 i1=2 D ¡.l 0 j E .w? .t /; W.l / .t ///
¡.l 0 j W .t //
Then, following the reasoning of Sect. 2.6, we get for d Œx.t /; E .w? .t /;
W.l / .t //
> 0 the strategy
8 ˇ
< arg min d V .t; x/=dt j ˇˇ u 2 E .q.t /; Q.t // ; if x 62 E .w? .t /; W .l / .t //;
u ˇ
Ue .t; x/
e
D
: /
E .q.t /; Q.t //; if x 2 E .w .t /; W .t //:
? .l
(3.136)
Since l D l 0 this gives
(
q.t / Q.t /B 0 .t /l 0 < l 0 ; QB .t /l 0 >1=2 ; if x 2
0
6 E .w? .t /; W.l / .t //;
Ue .t; x/ D 0
E .q.t /; Q.t //; if x 2 E .w? .t /; W.l / .t //:
(3.137)
We further show that the ellipsoidal strategy Ue .t; x/ of (3.137) does solve the
problem of control synthesis if the starting position f£; xg is such that x D x.£/ 2
0
@E .w? .£/; W.l / .£// for related l 0 and also x.£/ 2 @W Œ£: Indeed, suppose x.£/ 2
E.l / Œ£, and xŒt D x.t; £; x.£// that satisfies Eq. (3.124) with u 2 Ue .t; x/; £
0
t t1 ; is the respective trajectory that emanates from f£; x.£/g. We will demonstrate
that any such solution xŒt of (3.124) will guarantee the inclusion xŒt1 2 E .m; M /.
To do that we will need, as in Sect. 2.6, to know the total derivative d Ve .t; x/=dt
for Ve .t; x/ > 0.
136 3 Ellipsoidal Techniques: Reachability and Control Synthesis
Recalling
0
Ve .t; xŒt / D d Œt D d0 .t; xŒt / D< l 0 ; xŒt > ¡.l 0 j E.l / Œt /;
d d 0
d Œt D .< l 0 ; xŒt > ¡.l 0 jE.l / Œt //;
dt dt
so that
d @ 0
d Œt D< l 0 ; xŒt
P > ¡.l 0 jE .w? .t /; W .l / .t // D (3.138)
dt @t
d 0
D< l 0 ; B.t /.u q.t // > < l 0 ; W.l / .t /l 0 >1=2 :
dt
Ue .t; x/ D arg minf< l 0 ; B.t /u > ju 2 E .q.t /; Q.t //g: (3.140)
Ve .t1 ; x.t1 // D d0 .t1 ; x.t1 // D d.x.t1 /; E .m; M // d0 .£; x.£// D Ve .£; x.£// D 0
0
which means x.t1 / 2 E .m; M /, provided x.£/ 2 E.l / Œ£ W Œ£, whatever be the
solution to
We thus came to
Theorem 3.11.1. The solution to Problem 3.11.1 (with A.t / 0) is given by the
“ellipsoidal strategy” Ue .t; x/ of (3.140).
Exercise 3.11.1. Prove an analogy of Theorem 3.11.1 for system (3.1) with
A.t / 6D 0:
We now indicate some techniques for finding l 0 .
Calculating l 0
The scalar product hx; zi taken above, in solving Problem 3.11.1, may be substituted
for one of more general kind, namely by hx; T zi; T D T 0 > 0; hx; T xi D kxkT ;
so that in the above we took T D I: However we will also need to consider T 6D I;
since the calculation of vector l 0 is especially simple if one takes T 1 D W.l / .t /.
0
Therefore, before moving towards the proof that control (3.136) does indeed
solve Problem 3.11.1, we first look at how to calculate l 0 . Namely, we have to find
the distance d Œt > 0 in the more general form as
d Œt D VT .t; x/ D dT Œt; x D dT .x; E .w? .t /; W.l / .t /// D
D maxf< T l; x w? .t / > hT l; W.l / .t /T li1=2 / j < l; T l > 1g; (3.141)
In view of
we find
Lemma 3.11.1. For dT Œt > 0 the distance between given point x 2 Rn and
ellipsoid E D E.y ; Y / Rn ; Y D Y 0 > 0 is the solution to optimization
problem (3.142), with T D T 0 > 0:
Finding the solution to (3.142) may be pursued with various matrices T .
We emphasize two cases–when T D I and T D Y 1 :
Case (a) : T D I
Referring to Lemma 3.11.1, we use the standard Euclid metric, taking, y D
w ; Y D T . Then relations (3.137)–(3.140) remain without formal change.
?
D d.x w? ; E .0; W.l / // D minfkz pk j hp; .W.l / /1 pi 1; z D x w? g:
p
For kzkT > 1; T D .W.l / /1 ; this may be reduced to finding minp fkz
pk2 j hp; .W.l / /1 pi D 1g > 0 through conventional Lagrangian techniques.
The related Lagrangian is
L .p; œ/ D hz p; z pi C œ.hp; .W.l / /1 pi 1/
with
Lp0 .p; œ/ D 2..z p/ C œ.W.l / /1 p/ D 0; so that p D .I œ.W.l / /1 /1 z:
Here h.0/ > 0 and with œ ! 1 we have lim h.œ/ D 1; so there exists a root
œ0 > 0 of equation h.œ/ D 0. This root œ0 is unique, since h0 .œ/ < 0: We have thus
found
d.x w? ; E .0; W.l / // D kz p 0 k; p 0 D .I œ0 .W.l / /1 /1 z
and the element of E .w? ; W.l / / closest to x is x D w? C p 0 :
3.11 The Problem of Control Synthesis: Solution Through Internal Ellipsoids 139
D kx .y C q 0 /kY 1 :
140 3 Ellipsoidal Techniques: Reachability and Control Synthesis
(ii) The optimizer lT0 in problem (3.147) under metric kx ykT ; T D Y 1 is given
by (3.146).
(iii) The element of E closest to x under metric kx ykT is
y 0 y D q 0 ; kq 0 kT D 1:
Exercise 3.11.3. In the previous lemma show that vectors lT0 of (ii) and p 0 of (iii)
are collinear.
We now return to the original Problem and apply the previous Lemma 3.11.3 to
find the synthesizing control Ue .t; x/: In terms of Problem 3.11.1 we have: x D
x.t /; y D w? .t /; Y D W.l / .t /:
Now, in order to rewrite these results in terms of Lemma 3.11.3, we need to
substitute its l 0 by T l 0 D lT0 ; T D .W.l / /1 , with metric kx ykT D hx y;
.l / 1
.W / .x y/i : We thus get
1=2
0; if kx.t / w? .t /kT 1;
lT0 D (3.148)
.x.t / w? .t //kx.t / w? .t /k1
T ; otherwise:
The related control uT e .t; x/, for x.t / … E ; is then found due to (3.140), as
After following (3.140) and recalling QB .t / D B.t /Q.t /B 0 .t /, this yields the
solution strategy
8
< q.t / if x 2 E .w? .t /; W.l / .t //;
UTe .t; x/ D q.t / Q.t /B 0 .t /.x.t / w? .t //kQ1=2 .t /B 0 .t /.x.t / w? .t //k1
T ;
: .l /
if x 62 E .w .t /; W .t //:
?
(3.149)
Exercise 3.11.4. By direct substitution of u D UTe .t; x/ into (3.124) prove the
following:
(i) the existence of solution to (3.124) under such substitution,
(ii) that u D UTe .t; x/ does solve Problem 3.11.1.
3.12 Internal Approximations: The Second Scheme 141
Remark 3.11.1. A solution in the absence of Assumption 3.11.1 has to deal with
degenerate internal ellipsoids. The calculation of related controls then has to
undergo a procedure of regularization. This issue will be treated at the end of
Chap. 4.
The given scheme thus describes the applicability of ellipsoidal techniques
to the calculation of trajectory tubes for controlled systems with further design
of synthesizing controls within a fairly simple scheme as compared to existing
methods. This scheme is realized in the ellipsoidal toolbox used further to solve
the presented examples (see [45, 132]).
Apart from the techniques for internal approximation of sets and tubes given in
Sects. 3.7–3.10, there exists another set of formulas which may be used for the same
purpose. The restrictive element in this “second approach” is that here the ellipsoidal
approximations are derived through operations with only a pair of ellipsoids rather
than any number of these as in the above. The treatment of internals for sums of
ellipsoids may therefore be achieved only through an inductive sequence of pairwise
operations rather than through a single operation. Nevertheless this scheme turns out
to be useful and is also being recommended. The scheme was discussed in [174] (see
also [45]).
We shall calculate the reachability set X Œt D X .t; t0 ; X0 / for system (3.1) with
v.t / 0: Using the funnel equation (2.54), we have
¢ 1 o1 .¢/ ! 0 if ¢ ! 0:
X Œt D E .x ? .t /; X .t //; X .t / > 0;
we may apply the following formula. For two given ellipsoids E .a.1/ ; X1 /;
E .a.2/ ; X2 /; X1 D X10 > 0; X2 D X20 > 0; it indicates the matrix
where H 2 † and
† D fH 2 Rnn W H 0 D H; jH j 6D 0g:
and
[
E1;2 D fE .a.1/ C a.2/ ; X .H // j H 2 †g D E .a.1/ ; X1 / C E .a.2/ ; X2 /:
(3.153)
If one of the two ellipsoids is degenerate, say X2 0; then the union E1;2 at the
left-hand side of (3.153) should be substituted by its closure E1;2 .
Applying the last relations to (3.150), with X Œt D E .x ? .t /; X .t // being an
ellipsoid, we take
X Œt C ¢ C o1 .¢/B .0/
E .x ? .t C ¢/; X .t C ¢//;
where
X .t C ¢/ D H 1 .t /ŒH.t /X .t /H 0 .t / C
C ¢.A.t /X .t / C X .t /A0 .t // C ¢ 2 A.t /X .t /A0 .t / C
1 1
C ¢.H.t /.I C ¢A.t //X .t /.I C ¢A0 .t //H 0 .t // 2 .H.t /QB .t /H 0 .t // 2 C
1 1
C ¢.H.t /QB .t /H 0 .t // 2 .H.t /.I C ¢A.t //X .t /.I C ¢A0 .t //H 0 .t // 2 C
C ¢ 2 H.t /QB.t /H 0 .t /H 01 .t /;
and
From here, after subtracting X .t / from both sides, then dividing them by ¢ and
applying a limit transition with ¢ ! 0, we come to ordinary differential equations
1 1
C .H.t /QB .t /H 0 .t // 2 .H.t /X .t /H.t // 2 /.H 1 /0 ; (3.154)
x ? .t0 / D x 0 ; X .t0 / D X 0 :
X Œt
E .x ? .t /; X .t //; (3.156)
where the union is taken over all measurable matrix-valued functions with values
in †.
Remark 3.12.1. The closure of the union of sets at the right-hand side of equal-
ity (3.157) appears due to the fact that matrix QB D B.t /Q.t /B 0 .t / may turn out
to be degenerate. Then the exact reachability set X Œt may have kinks while being
approximated by internal ellipsoids all of which are nondegenerate. This effect is
absent in the “first approach” of Sects. 3.7–3.10 which allows degenerate elements
among the internal approximating ellipsoids.
One should note that internal approximations of reachability sets by nondegen-
erate ellipsoids require that the “reach set” itself would be nondegenerate.
Definition 3.12.1. A reachability set X Œt is said to be nondegenerate if there exists
an n-dimensional ball B© .c/ D fx W< x c; x c > ©g of radius © > 0, such that
B© .c/ X Œt :
In other words, set X Œt a nonempty interior. The last property may arise either from
the properties of the system itself and its constraints or from the degeneracy of the
starting set X Œt0 , or from both factors. To check degeneracy due to the system itself
it suffices to deal with reach sets of type X Œt D X .t; t0 ; f0g/; emanating from point
x.t0 / D 0:
Consider system
x.t0 / D 0;
where B.t / is continuous and q.t / 2 Rp ; Q.t / 2 Rpp ; p < n are also continuous.
144 3 Ellipsoidal Techniques: Reachability and Control Synthesis
Zt
X Œt D G.t; s/B.s/E .0; Q.s//ds; (3.159)
t0
for all l 2 Rn ; t t0 .
With Q.t / > 0; t > t0 ; this assumption implies that the reachability domain
X Œt of system (3.158) has a nonempty interior (intX Œt ¤ ;) and is therefore
nondegenerate. Assumption 3.12.1 is actually equivalent to the requirement that
system (3.158) with unbounded control u.t / would be completely controllable
[109, 195], on every finite time interval Œt0 ; t .
Exercise 3.12.1. Prove that under Assumption 3.12.1 and with Q.t / > 0; t > 0;
x.t0 / D 0 the reachability set X Œt for system (3.158) is nondegenerate. (Use
Exercise 1.4.2.)
In a similar way the “second approach” may be applied to the approximation of
backward reachability sets W Σ. A reasoning similar to the above in this section
gives the result.
Theorem 3.12.2. The internal approximation of the backward reachability set
W Œt is given by the inclusion
E .w? .t /; W Œt / W Œt ;
where the union is taken over all measurable matrix-valued functions with values
in †.
3.12 Internal Approximations: The Second Scheme 145
7
Such a proof is indicated in [174, Sect. 3.6, p. 212].
Chapter 4
Solution Examples on Ellipsoidal Methods:
Computation in High Dimensions
Abstract In this chapter we describe solution examples for controlled systems that
illustrate the contents of Chaps. 1–3. These include the multiple integrator, planar
Newtonian motions and calming down a chain of springs. Special sections are
devoted to relevant computational formulas and high-dimensional systems. Also
discussed are possible degeneracy effects in computation and the means of their
avoidance.
In this chapter we first present examples of control problems discussed above. All
these problems are treated in finite time. The solutions are reached through the
theory given in Chaps. 1, 2 and the computation is done using methods of Chap. 3.1
The second part deals with specifics of calculation for systems of high
dimensions.
We begin with the examples.
The calculation of reach sets for a two-dimensional system with one integrator was
treated in detail in Chap. 3, Sect. 3.9. We continue now by dealing with the multiple
integrator with single input given by system (i D 1; : : : ; n 1I n 2)
xP i D xiC1 ;
(4.1)
xP n D b.t /u
1
The illustrations for the presented solutions were provided by N.Bairamov, A.Mesiats,
D.Odinokov, and V.Stepanovich.
with b.t / taken continuous. The fundamental transition matrix for this system is
0 1
1 t s : : : .t s/n1 =.n 1/Š
B0 1 : : : .t s/n2 =.n 2/Š C
B C
G.t; s/ D B : : :: :: C
@ :: :: : : A
0 0 0 1
X 0 D fx W hx x 0 ; .X 0 /1 .x x 0 /i 1g:
In order to calculate the reach sets for this system we shall follow Chap. 1 and
transform the coordinates so that the new system has A 0 and a new matrix
B. Applying transformation z.t / D G.t0 ; t /x.t /; t t0 to (4.2), (see Sect. 1.1,
formula (1.6)), we get
We consider this system without loss of generality, under the same starting set X 0
and same constraint on control u as before. Now, with t t0 , the forward reach
set is
Z t
Xz Œt D X 0 C B.s/Q .s/ds; (4.4)
t0
and
0 1
.t0 t /n1 =.n 1/Š
B .t0 t /n2 =.n 2/Š C
B C
B.t / D B :: C b.t /:
@ : A
1
with its internal and external approximating ellipsoids being E .xz .t /; Xz .t //;
E .xz .t /; XzC .t //: These are given by
Z t
xz .t / Dx C 0
qb .s/ds; (4.7)
t0
Z t
0
1=2
Xz .t / D Xz .t /Xz .t /; Xz .t / D .X 0 /1=2 C S.s/QB .s/ds; .S0 D I /;
t0
(4.8)
Z Z !
0 1=2
t
1=2 X0 t QB .s/
XzC .t / D hl; X li C hl; QB .s/li ds C ds ;
t0 hl; X 0 li1=2 t0 hl; QB .s/li1=2
(4.9)
along good curves l.t / 2 Rn : (See Sect. 3.3.) Here the good curves lz .t / lz .t0 / D
l turn out to be constant. Therefore the previous relations turn into equalities for
any l under proper selection of parameterizing coefficients .t /; S.t /; in ODEs that
describe functions xz .t /, Xz .t /, XzC .t / for the approximating ellipsoids.
These ODEs are
( ( 8
.t / D .X 0 /1=2 ; < XzC .t0 / D X 0 ;
xz .t0 / D x 0 ; Xz 0
P D /XzC C 1 .t /QB .t /
xP z D qb .t /I XP z
D S.t /Q1=2 .t /I
B
X
: zC .t
(4.11)
with
Z t 1
.t / D hl; QB .t /li 1=2
hl; X li C
0 1=2
hl; QB .s/li ds
1=2
t0
and with matrices S 0 .t /S.t / D I; selected such that vectors .X 0 /1=2 l; S.t /QB .t /l
1=2
are collinear.
4.1 The Multiple Integrator 151
E .x .t /; XC Œt / X Œt E .x .t /; X Œt /; (4.12)
Remark 4.1.1. In further relations below we omit the lower indices z for all the
related variables in (4.12).
For the backward reach sets the passage to new coordinates fzg is given by
z.t / D G.ª; t /x.t /; t ª: Then, with given target set M D E .m; M /, we
again have Eq. (4.6) and in these new coordinates the backward reach set W Œt D
W .t; ª; M /; M D E .m; M / also allows tight ellipsoidal approximations
Here
Z !1
ª
.t / D hl; QB .t /li 1=2
hl; M li 1=2
C hl; QB .s/li 1=2
ds
t
with
Z ª
0
W D W W ; W D .W 0 /1=2
1=2
S.t /QB .t /dt;
£
Demonstrated here are the projections of reachability sets for system (4.2) of
dimension Dim = 4 and 8 on the plane fx1 ; x2 g at time ª: N stands for the number
152 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
The directions l that ensure tight approximations are selected for time ª to be a
uniform partition on the unit circle in the plane fx1 ; x2 g (Figs. 4.1, 4.2, 4.3, 4.4, 4.5,
4.6, 4.7, and 4.8).
Our first task will be to find forward and backward reach sets
under constraints
u1
u.t / D u D 2 Q .t / D E .q.t /; Q.t //; x.t0 / D x .0/ 2 X 0 :
u2
Denote
0 1 0 1 0 1
x1 0100 0 0
B x2 C B0 0 0 0C u1 B b1 0 C
xDB C B C B C
@ x3 A ; A D @ 0 0 0 1 A ; u D u2 ; B D @ 0 0 A :
x4 0000 0 b2
4.2 A Planar Motion Under Newton’s Law 157
Then system (4.14) has the form (4.2), here with B D const: The fundamental
matrix of its homogeneous part has the form G.t; £/ D exp A.t £/; where Ak D
0; k 2: Hence
0 1
1t £ 0 0
B0 1 0 0 C
G.t; £/ D e A.t£/ D I C A.t £/ D B
@0
C:
0 1t £A
0 0 0 1
with
0 1
.t0 t /b1 0
B b1 0 C
B.t / D G.t0 ; t /B D B
@
C ; qB .t / D B.t /q;
0 .t0 t /b2 A
0 b2
B.t /Q.t /B0 .t / D QB .t /:
¡.l j Xz Œt / D (4.17)
Z t
hl; x 0 i C hl; X 0 li1=2 C .hl; qB .t /i C hl; QB .t /li1=2 /dt:
t0
Similarly, for the backward reach set we use transformation z D G.ª; t /x; t ª;
and take B.t / D G.ª; t /B: Then system (4.14), without changing notations, may
be again transformed into (4.17), but now with t ª: Here we also return from
boldface B to B:
For target set M D E .m; M / the backward reach set will be
Z ª
Wz Œt D Wz .t; ª; M / D M B.s/Q .s/ds
t
158 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
¡.l j Wz Œt / D (4.18)
Z ª
hl; mi C hl; M li 1=2
.hl; qB .t /i hl; QB .t /li1=2 /dt:
t
Presented are calculations of the forward reach set X Œ1 D X .1; 0; E .x 0 ; X 0 // for
the four-dimensional system
and
0 1 0 1
0 3000
B0C B C
x0 D B C ; X0 D B 0 1 0 0 C ; q D 1 ; Q D 1 0 :
@0A @0 0 1 0A 1 04
0 0003
We are now passing to the calculation of control strategies. Our task in the simplified
coordinates (from z to x D G.t; ª/z) will be to transfer system (4.19) from given
position ft0 ; xg to the final one m 2 M within prescribed time ª t0 . As we have
seen in Sects. 2.6 and 3.11, the task will be solvable iff x 2 W Œt0 , where W Œt0 D
W .t0 ; ª; M / is the backward reachability set from target set M .
4.2 A Planar Motion Under Newton’s Law 159
We therefore have a problem in the next form: find feedback control strategy
u D U.t; x/ which ensures the following conditions
P / D B.t /u.t /; t 2 Œt0 ; ª;
x.t
(4.20)
x.t / 2 Rn ; x.ª/ 2 E .m; M /; u.t / 2 E .q.t /; Q.t //;
s 0 D argminfkx sk W s 2 E Œt ; x D x.t /g
so that l 0 is the gradient of the distance function d.x; E Œt / with t fixed and s 0 is
the nearest point of E Œt from x 62 E Œt .
162 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
Then s 0 will lie on the line that connects point x and w? , hence
l 0 D .x w? /=kx w? k:
1
xP 1 D x2 ; xP 2 D u C 1; M D E .m; M /; m D .1; 1/0 ; M D I; juj 1;
2
indicated here are the projections of the backward reach set W Œ£ for £ D 0; ª D 3
(Figs. 4.14, 4.15, 4.16, 4.17, 4.18, 4.19, 4.20, 4.21, and 4.22).
4.2 A Planar Motion Under Newton’s Law 163
Fig. 4.15 N D 5; DIM D 2. Backward reach set—internal estimates and their convex hull
164 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
Fig. 4.16 N D 30; DIM D 2. Backward reach set—internal and external estimates
Fig. 4.17 N D 30; DIM D 2. Backward reach set—internal estimates and their convex hull
4.2 A Planar Motion Under Newton’s Law 165
Fig. 4.19 N D 15; DIM D 2. Backward reach tube—coinciding internal and external estimates
166 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
Fig. 4.20 N D 5; Di m D 2.
Closed-loop chattering
control—aiming method (b)
in the ellipsoidal norm
In the previous figures the controls are of chattering type, ensuring sliding-type
regimes when the trajectory fluctuates along the boundary of the backward reach set
W Œt . However, one may introduce a solution with continuous control, when having
started from inside W Œ£ the trajectory does not reach the boundary of W Œt ; t > £.
Such control is constructed as follows (T D W1 .t /)
8
< q.t /; if hx w? .t /; W1 .t /.x w? .t //i 1;
Ue .t; x/ D otherwise
:
q.t / QB 0 .t /.x w? .t //kQ1=2 B 0 .t /.x w? .t //k1 :
168 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
with t > t0 D 0. Here n is the number of springs as numbered from top to bottom.
The loads are numbered similarly, so that the i -th load is attached to the lower end
of the i -th spring; wi is the displacement of the i -th load from the equilibrium, mi is
the mass of the i -th load, ki is the stiffness coefficient of the i -th spring. The gravity
force enters (4.24) implicitly, through the lengths of the springs at the equilibrium.
The initial state of the chain at time t0 D 0 is given by the displacements w0i and
the velocities of the loads wP 0i , with
wi .t0 / D w0i ; w
P i .t0 / D wP 0i ; i D 1; : : : ; n:
The control u D uj .t / in this equation is indicated at the j-th load (which may
be j D n) with its values confined to the interval Q D Œumin ; umax . In particular,
this constraint may be symmetric: Q D Œ; or one-sided: directed only down
(Q D Œ0; ) or only up (Q D Œ; 0).
We now rewrite system (4.24) in normal matrix form introducing the extended
vector x 2 R2n such that .x1 ; : : : ; xn /0 D w; .xnC1 ; : : : ; x2n /0 D w.
P Then
w0
P / D Ax.t / C b .j / uj .t /;
x.t x.t0 / D x 0 D (4.25)
P0
w
0 I 0
b .j / D .0; : : : ; 0; bnCj ; 0; : : : ; 0/; bnCj D m1
.j / .j /
AD ; j ;
M 1 K 0
0 1
m1
B :: C
M D@ : A;
mn
0 1
k1 C k2 k2
B k2 k2 C k3 k3 C
B C
B : : : C
K D B ::: :: :: :: ::: C :
B C
@ kn1 kn1 C kn kn A
kn kn
The emphasis of this book is not only to describe exact solutions but also to indicate
effective numerical methods. Based on ellipsoidal approximations such methods
rely on specially derived ordinary differential equations, indicated above in Chap. 3.
We now present some examples of feedback control for a chain of springs. These
are solved by schemes explained in Sect. 3.11.
Example 4.2. Chain of four springs with four loads with control at each load
Indicated are projections of the backward reachability set for system (4.24) with
four springs and four loads of different masses fmi g D f3; 1; 1:5; 5g and different
coefficients of stiffness fki g D f2; 1; 4; 1g.
We assume target set M D E .0; M /; M D 0:1I and the control in (4.25) to be
four-dimensional: Bu D fb .5/ u1 ; : : : ; b .8/ u4 g (applied to each load),
0 1
1 1 0 0
B1 2 0 0C
u 2 E .0; Q/; QDB
@0
C:
0 2 0A
0 0 0 1
The phase space for our system is of dimension 8, with unit orths of from
e10 D .1; 0; : : : ; 0/ 2 R 8 to e80 D .0; : : : ; 0; 1/ 2 R 8 : Our aim is to calculate
the solvability set (backward reach set) W Œt D W .t; ª; M / with ª D 3, seeking
for the projections of eight-dimensional W Œt on two-dimensional planes fw5 ; w6 g
and fw7 ; w8 g of system velocities. To find these projections we need the related
external and internal ellipsoidal approximations that are calculated for good curves
l.t / along which they would be tight. The boundary conditions l D l.ª/ for such
curves are taken along a partition of unit circles on the planes e5 ; e6 and e7 ; e8 for
the velocities. The evolution of these projections in time will reflect the dynamics of
tube W Œt . The result of intersecting approximating ellipsoids is given in Figs. 4.24,
4.25, 4.26, and 4.27.
Example 4.3. Target control trajectories for chain of four springs with four loads:
Control at each load
Indicated are the target controlled trajectories with four springs and four loads
of different masses fmi g D f5; 2; 4; 1g and different coefficients of rigidity fki g D
f2; 1; 1:5; 3g; taking M D E .0; 0:1I /: The aiming is done within scheme (b) of
Sect. 3.11, in the metric of internal ellipsoids. The results are illustrated in Figs. 4.28,
4.29, 4.30, and 4.31 with boundary of target set in blue.
Example 4.4. Oscillating springs with two loads. Considered is a system with two
loads and control applied to one of them.
4.3 Damping Oscillations 171
Fig. 4.30 Di m D 8:
Projection of phase trajectory:
velocities x5 ; x6
Fig. 4.31 Di m D 8:
Projection of phase trajectory:
velocities x7 ; x8
Calculated are:
– projections of approximations of backward reach set at final time by arrays of
ellipsoids (their boundaries are marked blue for externals and green for internals;
an accurate approximation through many ellipsoids indicates that the result is
seen as almost the same).
– solutions to the problem of ellipsoidal control synthesis with pictures showing
starting set is in green and boundary of target set in red.
(i) Two loads. Control at upper load.
Here, in system (4.24), the number of loads is N D 2, m1 D 2; m2 D 2, k1 D 1:3,
k2 D 2, w 2 R2 , t0 D 0; t1 D 7. The control force u 2 Œ1; 1 is applied only to
the upper load. In system (4.25) x 2 R4 : The approximated backward reach tube,
as calculated in backward variables tN D 7 t; t 2 Œ0; 7; emanates at time tN D 0;
4.3 Damping Oscillations 175
Fig. 4.32 Dim 4: Projections of approximation of backward reach set with control at upper load
on two-dimensional planes in axes (position, velocity) for each load
.t D 7/; from target set M —a ball of radius 0:01 centered at 0, and terminates at
t D 0; T D tN D 7: Reach sets at shown in backward time at T D 7.t D 0/, see
Fig. 4.32.
The controlled trajectory evolves from starting set X 0 , centered at point x.0/ D
x D Œ0:5688; 1:0900; 0:8921; 0:24230 to target set M —a ball of radius 0:1
0
Fig. 4.33 Dim 4: The target controlled trajectories for system of two loads with control at upper
load
get set M —a ball of radius 0:1 centered at m D 0. It is shown in Figs. 4.39 and 4.40.
Here the problem parameters are the same as with control at first (upper) load except
that now it is applied to the fourth (lowest ) load.
Projections of the reach set are shown in Fig. 4.41. The controlled trajectory now
evolves from starting set X 0 , centered at point
x.0/ D x 0
D Œ0:6296I 1:9940I 2:6702I 3:6450I 0:1013I 0:1054I 0:2913I 0:63830
to the same target set M —a ball of radius 0:1 centered at f0g. It is depicted in
Figs. 4.42 and 4.43.
4.3 Damping Oscillations 177
Fig. 4.34 Dim 4: Projections of phase trajectories of target controlled system with control at upper
load
In order to calculate exact solutions to our control problems one needs to heavily
rely on calculating fundamental matrix G.t; s/: Their calculation may also occur in
other elements of the solution process. The calculation of such matrices for large
dimensions is a special direction in numerical methods (see [89, 90, 92]). In the next
lines we indicate an algorithmic method applicable to differential equations for the
chain of springs.
178 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
Fig. 4.35 Projections of approximation of backward reach set with control at lower load on two-
dimensional planes in axes (position, velocity) for each load
Fig. 4.36 The target controlled trajectories for system of two loads with control at lower load
For stationary systems matrix function G.t; s/ D exp A.t s/ which is a matrix
exponent. Consider calculation of such an exponent X.t / D e At , assuming
0 I
AD ;
M 1 A0 0
4.3 Damping Oscillations 179
Fig. 4.37 Dim 4: Projections of phase trajectories of target controlled system with control at lower
load
Fig. 4.38 Projections of approximation of backward reach set with control at first (upper) load on
two-dimensional planes in axes (position, velocity) for each load
we further obtain
.1/k .M 1 A0 /k 0
A2k D ;
0 .1/k .M 1 A0 /k
0 .1/k .M 1 A0 /k
2kC1
A D kC1 1 kC1 :
.1/ .M A0 / 0
1 1
Introducing new matrix L D .M 2 A0 M 2 /, we have
1 1
!
.1/k M 2 L2k M 2 0
A 2k
D 1 1 ;
0 .1/k M 2 L2k M 2
1 1
!
0 .1/k M 2 L2kC1 L1 M 2
A 2kC1
D 1 1
.1/kC1 M 2 L2kC1 LM 2 0
4.3 Damping Oscillations 181
Fig. 4.39 The target controlled trajectories for system of four loads with control at first (upper)
load
and further on
1
X 1
e At
D .At /k D
kŠ
kD0
1
! 1
!
M2 0 cos.Lt / sin.Lt /L1 M2 0
D 1 1 : (4.26)
0 M2 sin.Lt /L cos.Lt / 0 M2
Fig. 4.40 Projections of phase trajectories of target controlled system with control at first (upper)
load
The presented algorithm relies on the special structure of matrix A and proves to
be far more effective than the general algorithms that do not rely on such structure.
A simulation experiment indicates an eight-times more effective performance of the
structure related algorithm as compared to the “ordinary” one.
As we have observed, the basic problem treated here is how to calculate ellipsoidal
estimates for reachability tubes of a control system. Its solution is a crucial element
for calculating feedback target control U .t; x/ that steers the system trajectory from
2
This section follows paper [58].
4.4 Computation in High-Dimensional Systems. Degeneracy and Regularization 183
Fig. 4.41 Projections of approximations of backward reach set with control at fourth (lowest) load
on two-dimensional plane in axes (position, velocity) for each load
given position ft0 ; X 0 g to given target set M . Namely, as indicated in Sect. 2.6, to
apply the aiming rule, one needs to calculate the backward reachability (solvability)
tube W Œt for the investigated system. For linear-convex systems of type (3.1) this
is done through ellipsoidal methods of Sect. 3.11 which proved successful for many
problems. This route is especially useful for systems of higher dimensions where
finding set W Œt directly is computationally too cumbersome. Thus we use its
internal ellipsoidal approximations. Here we require that constraints in the problem
are ellipsoidal: X 0 D E .x 0 ; X 0 /, Q .t / D E .q.t /; Q.t //. Otherwise sets Q .t / have
to be approximated by their internal or external ellipsoids. Functions q.t /, Q.t / for
the constraints are assumed to be continuous.
With W Œt given, the desired control strategy U D Ue .t; x/ is given
by (3.136), (3.137). Following Sects. 3.9, 3.11, and (3.123), we observe that the
internal ellipsoids are generated by the following ODEs:
Fig. 4.42 The target controlled trajectories for system of four loads with control at fourth (lowest)
load
1 1
H.t / D W 2 .t /S.t /QB2 .t /;
d l.t /
D A0 .t /l.t /; l.t0 / D l:
dt
The computation of external estimates is less difficult than of the internals which
are therefore less investigated. However, the internals are crucial for designing
feedback controls, so these difficulties, which increase with dimension, have to be
coped with. We therefore concentrate on internals.
4.4 Computation in High-Dimensional Systems. Degeneracy and Regularization 185
Fig. 4.43 Projections of phase trajectories of target controlled system with control at fourth
(lowest) load
Recall from Sect. 3.7 that the formula for an internal ellipsoidal estimate of the
geometrical sum of m ellipsoids E .qi ; Qi /, i D 1; : : : ; m, tight in direction ` 2 Rn :
E .q1 ; Q1 / C C E .qm ; Qm /
E .q; Q/; (4.29)
X
m
1 1 1
qD qi ; Q D QŒm0 QŒm; QŒm D Q12 C S2 Q22 C C Sm Qm2 ;
iD1
Corollary 4.4.1. The set of rank 1 matrices is closed with respect to opera-
tion (4.29).
This means a collection Q of matrices of rank 1, made according to (4.29), is also
of rank 1.
Remark 4.4.1. The degeneracy of estimates is the result of their tightness (the
requirement that they touch the exact set). This is not a property of the particular
formula (4.29). Indeed, if rankQi D 1, then the summed ellipsoids may be
presented as Ei D E .qi ; Qi / D convfqi ˙ ai g, where vectors ai are the only
nonzero semi-axes of these ellipsoids. Then the exact sum E1 C C Em will be a
polyhedron convfq1 C C qm ˙ a1 ˙ ˙ am g. And for almost all directions `
the only tight approximation will be one of the diagonals of this polyhedron. In this
case formula (4.29) describes exactly those diagonals.
Remark 4.4.2. For some choices of orthogonal matrix S the dimension of the inter-
nal ellipsoidal estimate may be strictly less than the dimension of ellipsoids being
added. For example, if Q1 D Q2 D I , ` D l .1/ , S D diagf1; 1; 1; : : : ; 1g,
then (4.29) gives Q D diagf1; 0; 0; : : : ; 0g.
Applying results of the above to systems of type (4.27), (4.28), we indicate how
several degenerate estimates may be combined (mixed) to get an estimate of higher
dimension. This approach is based on the formula for internal ellipsoidal estimates
of the convex hull for the union of ellipsoids [174].
Lemma 4.4.1. If ellipsoids Eiw D E .w; W.i/ /, i D 1; : : : ; m, are internal estimates
of a convex set X , then ellipsoid
X
m X
m
E’w D E .w; W ’ /; W’ D ’i W.i/ ; ’i 0; ’i D 1; (4.30)
iD1 iD1
X
m
1
’i .hqi ; `i C h`; W.i/ `i 2 / max ¡.` j Ew i /;
iD1;:::;m
iD1
188 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
hence
[
m
E’w conv Eiw X :
iD1
t
u
Theorem 4.4.2. Suppose w D 0 and the dimension of L (the linear hull of Eiw ) is r.
Then if ’i > 0, i D 1; : : : ; m, the following equality holds: the image of mapping
W’ is imW’ D L . In particular, the matrix W’ is of rank r.
Proof. This theorem means that
Due to the symmetricity of matrices W.i/ , the latter is equivalent in terms of matrix
kernels to
The inclusion of the right-hand side into the left is obvious. On the other hand, if
Since matrices W.1/ are non-negative definite and ’i is positive, we observe that
hx; W.i/ xi D 0. Hence x 2 ker W.1/ . t
u
Note some properties of estimates E’w :
1. Suppose ellipsoidal approximation E1w is tight in direction `, i.e. ¡.` j E1w / D
¡.` j X /. Let us now estimate the difference between its support function and
that of ellipsoid E’w in the same direction. After some calculation we have:
1
¡.` j E’w / hw; `i C ’1 h`; W.1/ `i 2
1 p
D hw; `i C h`; W.1/ `i 2 1 .1 ’1 / D
1 1
D ¡.` j X / .1 ’1 /h`; W.1/ `i 2 C O..1 ’1 /2 /:
2
4.4 Computation in High-Dimensional Systems. Degeneracy and Regularization 189
Therefore, the closer ’1 is to 1, the closer are ellipsoids E’w to the tight internal
ellipsoid along direction `.
2. If all the ellipsoids Eiw , i D 1; : : : ; m, are tight approximations of X along the
same direction `, then in (4.31) an equality is true and ellipsoid E’w is also a tight
approximation of X along `.
Example 4.6. Figure 4.44 shows ellipsoidal estimates of the sum of two degenerate
ellipsoids. Original ellipsoids Ei are two line segments (thick dotted line), and their
sum is a parallelogram (thin dotted line). Due to Theorem 4.4.1 tight approximations
of the sum are also degenerate ellipsoids (line segments) shown by thick solid line.
Regularized approximations with ’1 D 10 1 1 9
; 2 ; 10 are presented with thin solid lines.
These are non degenerate due to Theorem 4.4.2 and touch the parallelogram (note
that they are tight in direction of normals to the sides of the parallelogram, see
property 2). Besides that, for ’ D 10 1
and 10 9
thanks to property 1 the support
functions for the estimates are close to the support function of the parallelogram
along corresponding directions.
Figure 4.45 shows internal ellipsoidal approximations of solvability set, as
calculated by (4.31), for an oscillating system xP 1 D x2 , xP 2 D x1 C u on
time interval Œ0; with parameter ” D 20 1
(left) and 21 (right). Exact (tight)
approximations are degenerate ellipsoids (shown with thick lines).
’1 D 1 ¢ ” C ¢ ”“i ; ’i D ¢ ”“i ; i D 2; : : : ; m;
P
where “i 0 and m iD1 “i D 1; ” 0.
Here ¢ is sufficiently small, ensuring ’1 > 0. Then
!
X
m
¢ 1 .W’ W.1/ / D ” “i W.i/ W.1/ :
iD1
190 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
2 2
1.5 1.5
1 1
0.5 0.5
0 0
−0.5 −0.5
−1 −1
−1.5 −1.5
−2 −2
−2 −1 0 1 2 −2 −1 0 1 2
Fig. 4.45 Internal ellipsoidal estimates of the reachability set of a 2D system (left: ” D 1
20
, right:
” D 12 )
We use this result to mix m ellipsoidal estimates of the reachability tube as follows:
1 1
with H.t / D .W.i/ .t // 2 S.t /QB2 .t /.
Recall that Si .t / are arbitrary orthogonal matrices such that vectors
1 1
Si .t /.B.t /Q.t /B 0 .t // 2 li .t / and .W.i/ .t // 2 li .t / are directionally collinear. The
column vector functions li .t / are solutions to related adjoint systems
d li .t /
D A0 .t /li .t /; li .t0 / D `i :
dt
Remark 4.4.3. Parameter ” 0 controls the level of “mixed” approximations: the
higher is ”, the greater is the impact of mixing. Parameters “ij 0 control the
configuration of the mixture (namely, which ellipsoids are being mixed).
Remark 4.4.4. The choice of identical coefficients “ij D “O j (in particular, “ij D m1 )
P
reduces the number of operations, since in this case the sum m j D1 “ij W .t / D
.j /
Pm O .j /
j D1 “j W .t / does not depend on j and is calculated only once for each time
step.
Theorem 4.4.3. Suppose matrix-valued solutions W.i/ .t / to Eq. (4.28) are extend-
able to interval Œt0 ; t1 , and are positive definite on this interval. Then the set-value
function
[
m [
m
W Œt D conv Eiw Œt D conv E .w .t /; W.i/ .t //
iD1 iD1
4.4 Computation in High-Dimensional Systems. Degeneracy and Regularization 191
Let ¢ > 0 be sufficiently small, such that for • 2 Œ0; ¢ the maximum for given
1
direction ` is achieved at same i D i0 , i.e. ¡.` j W Œt / D h`; W.i0 / .t /`i 2 :
Assuming k`k D 1, the estimate for the support function of W Œt ¢ is
1
¡.` j W Œt ¢/ D h`; W.i0 / .t ¢/`i 2 D
¢
h`; W.i0 / .t /`i 2 h`; WP .i0 / .t /`i 2 C o.¢/:
1 1 1
Ch`; W.i0 / .t /`i 2
2
0 1
X
m
C” @ “i0 j h`; W.j / `i h`; W.i0 / `iA
j D1
1 1 1 1
2k.W.i0 / / 2 `k kS kkQB2 `k 2h`; W.i0 / `i 2 h`; QB `i 2 :
which is equivalent to the funnel equation (4.32) of the above (A.t / 0).
Corollary 4.4.2. Set-valued function W Œt is an internal estimate of the reacha-
bility tube W Œt , and functions W.i/ Œt are internal ellipsoidal estimates of W Œt .
This follows from the fact that the solvability set W Œt is the maximum solution to
funnel equation (4.32) with respect to inclusion.
192 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
102
101
Eigenvalues of W1/2
100
10−1
10−2
100 101
Number of Approximations (m)
101
(Volume of E)1/20
100
10−1
10−2
100 101
Number of Approximations (m)
Fig. 4.46 The size of ellipsoidal estimates depending on the number of mixed estimates m (top:
eigenvalues, bottom: volume in the power n1 )
1
all eigenvalues of the matrix .W.1/ / 2 (these eigenvalues are the sizes of semi-axes
of the estimates). On the bottom is the volume of the estimates plotted in power
1
n
D 201
(the geometrical mean of the axes).
Analyzing similar plots for a number of values of N , one may conclude that for
robust computation of the reachability tube the number of approximations m should
be chosen close to the system dimension n.
Equation (4.28) for the ellipsoidal approximation includes the operation of finding
an orthogonal matrix S D S.v1 ; v2 / 2 Rnn , such that S v2 is collinear with v1 for
some nonzero vectors v1 ; v2 2 Rn .
Note that with n 2 the matrix S.v1 ; v2 / is not unique (for n D 2 there are at
least two such matrices, while for n 3 there are infinitely many).
Function S.v1 ; v2 / should be defined as sufficiently smooth in variables v1 ; v2 , so
that integration schemes for higher-order ODEs could be applied to (4.28).
Matrix S.v1 ; v2 / may be calculated, for example, by computing the singular value
decomposition of vectors v1 ; v2 and by multiplying the corresponding orthogonal
matrices [132]. The computational burden for this procedure is of order O.n3 /, and
continuous dependence of S on v1 ; v2 is not guaranteed.
The following theorem gives explicit formulas for calculating S.v1 ; v2 /, which
relax the burden to O.n2 / operations and also ensure S to be sufficiently smooth.
Theorem 4.4.4. Take some nonzero vectors v1 ; v2 2 Rn . Then matrix S 2 Rnn
calculated as
S D I C Q1 .S I /Q1T ; (4.34)
p
c s vi
SD ; ; c D hOv1 ; vO 2 i; s D 1 c 2 ; vO i D ; (4.35)
s c kvi k
Note that the columns of Q1 and Q2 form orthonormal bases in plane Lv and its
orthogonal complement Lv? , respectively. With an additional constraint R11 > 0,
R22 > 0 matrix Q1 is unique, while matrix Q2 may be arbitrary, with orthonormal
columns orthogonal to those of Q1 . The indicated relations may be regarded as the
Gram—Schmidt orthogonalization procedure for finding matrix Q1 .
Set
S O
S DQ Q0 :
O I
Remark 4.4.6. It may be checked directly that for nonzero vectors v1 ; v2 function
S D S.v1 ; v2 / has continuous derivatives of any order everywhere except the cone
generated by collinear v1 ; v2 .
For the feedback problem the exact “extremal aiming” control should be
calculated as
( 1
Q.t /B 0 .t /lhl; QB .t /li 2 if B 0 .t /l ¤ 0I
Ue .t; x/ D
E .q.t /; Q.t //; if B 0 .t /l D 0;
(especially so, since parameters of the sets are stored across multiple processes).
196 4 Solution Examples on Ellipsoidal Methods: Computation in High Dimensions
[
m [
m
O Œt /; W
VO .t; x/ D d.G.t1 ; t /x; G.t1 ; t /W O Œt D Eiw Œt D E .w.t /; W.i/ .t //;
iD1 iD1
VO .t; x/ D min max fhq; G.t1 ; t /xi ¡.G 0 .t1 ; t /q j Eiw Œt /g: (4.38)
iD1;:::;m kqk1
Simulation Results
As indicated above, the solutions to many problems of control synthesis for systems
described by ODEs reduce to an investigation of first-order PDEs—the HJB type—
and their modifications [21, 24, 178, 198, 226, 248]. Similar equations may also be
used for calculating forward and backward reachability sets for control systems with
or without disturbances (the HJB equations).1
It is also well known that solutions to equations of the HJB type are rather
difficult to calculate, and their respective algorithms are still being developed
[221, 244]. However, for many problems, as those on reachability, on design of
safety zones for motion planning or on verification of control algorithms, one may
often be satisfied with approximate solutions that require a smaller computational
burden and may be achieved through substituting original HJB equations by
variational inequalities [23] due to certain Comparison Principles [48,95,149,186].
This chapter indicates such comparison theorems which are also applicable to
nonlinear systems for both smooth and nonsmooth solutions and to description of
nonconvex sets. They also allow to derive ellipsoidal methods through deductive
procedures in contrast with inductive approaches of Chap. 3. At the same time,
1
Such considerations are also true for systems with uncertain (unknown but bounded) disturbances
(the HJBI equation) [123, 176, 222], and for dynamic game-type problems. They involve more
complicated equations of Hamilton–Jacobi–Bellman–Isaacs (HJBI) type, [18, 19, 102, 123]. [121,
150,183]. These issues are mostly, except Chaps. 9 and 10, beyond the scope of the present volume.
In this section we deal with ordinary systems without disturbances. Consider first
the nonlinear equation
which coincides with Eq. (1.1) whose properties are described in the first lines of
Sect. 1.1.
As indicated in Sect. 2.3, given initial set X 0 (for t D t0 ) and a target set M
(for t D ª/, it makes sense to construct forward X Œt D X .t I t0 ; X 0 /; 8t t0 /;
and backward W Œt D W .t I ª; M /; t ª; reachability tubes for system (5.1),
emanating from set-valued positions ft0 ; X 0 g and fª; M g, respectively [121, 178].
Introduce notations
It is well known (see Sect. 2.3 and also [145, 178, 198]) that the solution V .t; x/
of the respective “forward” HJB equation of type
wC C
t C H.t; x; wx / .t /: (5.4)
Theorem 5.1.1. Let functions H.t; x; p/; wC .t; x/; .t / satisfy Assumption 5.1.1.
Then the following estimate is true
X Œt XC Œt ; (5.5)
where
Z t
XC Œt D fx W wC .t; x/ .s/ds C maxfwC .t0 ; x/jx 2 X 0 gg: (5.6)
t0
wC C C C
t .t; x / C H .t; x ; wx / wt .t; x / C H.t; x ; wx / .t /:
wC .t; x .t //
Z t Z t
.s/ds C wC .t0 ; x .t0 // .s/ds C maxfwC .t0 ; x/jx 2 X 0 g;
t0 t0
where V .b/ .t; x/ is the solution (classical or generalized) of the “backward” HJB
equation
.b/
Vt H .t; x; Vx.b/ / D 0; V .b/ .ª; x/ D d 2 .x; M /: (5.7)
Here UC is the class of closed-loop controls in which the last problem does have an
optimal closed-loop solution u0 .; / which generates an optimal trajectory x 0 λ D
x 0 .ªI £; x/:
Assumption 5.1.2. There exist functions H.t; x; p/, w.bC/ .t; x/ 2 C1 ; .t / 2 L1 ,
which satisfy the inequalities
Denote
where w.bC/ .ª; x 0 Œª/ V .b/ .ª; x 0 Œª/ according to the last condition of
Assumption 5.1.2.
However, along the optimal trajectory x 0 Œt ; t 2 Œ£; ª; we have
(see, for example, [48, Sects. 7.5–7.7]). This brings us to the next theorem.
5.1 The Comparison Principle for the HJB Equation 201
Theorem 5.1.2. Let functions H.t; x; p/; w.bC/ .t; x/; .t / satisfy Assumption
5.1.2. Then there exists a lower estimate for V .b/ .£; x/ W
Z ª
w.bC/ .£; x/ C .t /dt V .b/ .£; x/;
£
W Σ WC Σ;
where
Z ª
WC Σ D fx W w .bC/
.£; x/ C .t /dt 0g:
£
Exercise 5.1.1. (a) Prove a proposition similar to Theorem 5.1.1 for the backward
reach set W Σ:
(b) Prove a proposition similar to Theorem 5.1.2 for the forward reach set X Σ.
Let us now pass to the discussion of internal estimates for the reachability sets
and the related HJB equations. We shall consider the internal estimates for backward
reachability sets W Œt D W .t; x; M /. As in the above, here we do not necessarily
require differentiability of the value function V .b/ .t; x/.
Let function V .b/ .t; x/ of Eq. (5.7) be continuous in all the variables, being a
generalized (viscosity) solution of this equation (in particular, it may also turn out
to be classical).
Consider the next assumption.
Assumption 5.1.3. There exist function h.t; x; p/ and function w .t; x/ 2 C1 ;
which satisfy the inequalities
w
t .t; x/ h.t; x; wx .t; x// 0;
w
t H .t; x; wx / wt h.t; x; wx / 0:
Integrating the last inequality from £ to ª, along some optimal trajectory x 0 Œs D
x 0 .s; £; x/ which ends in x 0 Œª; we come to
However V .ª; x 0 Œª/ D V .£; x/, according to (5.9). Comparing this with (5.10),
we come to the next proposition.
Theorem 5.1.3. Under Assumption 5.1.3 the following estimate is true
We shall further observe, in the linear-convex case, how comparison theorems may
be used for obtaining ellipsoidal estimates of reachability sets.
E .pu ; P / D fu W hu pu ; P 1 .u pu /i 1g;
Then
.b/
Vt C H .b/ .t; x; Vx.b/ / D 0; V .b/ .ª; x/ D d 2 .x; M /: (5.14)
Under Assumption 5.1.2, as applied to system (5.1), the function w.t; x/ is taken
to be quadratic:
In order that satisfy this scheme we shall demand that the expression in curly
brackets of (5.16) is equal to zero for all x x :
Equalizing with zero the terms with multipliers of second order in x x ; then those
of first order in the same variable, we observe that the last equality will be fulfilled
if and only if the following equations are true
d w=dt ” 2 .t /:
W Œt WC Œt ; (5.22)
is true, where
Z ª
WC Œt D x W hx x .t /; K.t /.x x .t //i 1 C 2
” .s/ds :
t
W Œt WC Œt ;
where
we shall apply Theorem 5.1.3 and Corollary 5.1.1. In doing this we shall use the
inequality
hp; B.t /P .t /B 0 .t /pi1=2 hT .t /p; T .t /pi1=2 hp; S.t /.B.t /P .t /B 0 .t //1=2 pi;
where p 2 Rn is arbitrary and S.t / is any continuous matrix function, whose values
S.t / are orthogonal matrices, namely, S.t /S 0 .t / D I and T .t / D T 0 .t / is a matrix
function with T .t /p 6D 0:
Here an equality is attained under coordinated collinearity of vectors
S.t /.B.t /P .t /B 0 .t //1=2 p and T .t /p.
As in the “external” case, under Assumption 5.1.3, when applied to sys-
tem (5.12), function w .t; x/ is taken quadratic: w .t; x/ D hx x ; K .t /.x
x /i 1, with K .t / D .K /0 .t / > 0 and x .t / differentiable. Taken T .t / D
.K /1 ; we then have
w
t CH
.b/
.t; x; w 0 1=2
x /Dwt Chwx ; A.t /xCB.t /pu .t /ihwx ; B.t /P .t /B .t /wx i
w
t C hwx ; A.t /x C B.t /pu .t /i
2hxP .t /; K .t /.x x .t //i C 2hK .t /.x x .t //; A.t /x C B.t /pu .t /i
2r 1 hx x .t /; S.t /PB .t /K .t /.x x .t //i :
1=2
(5.24)
5.1 The Comparison Principle for the HJB Equation 207
The last relation will reflect the scheme of Theorem 5.1.4 with r sufficiently
large. Then, in order to satisfy this scheme, it will be sufficient that the expression
in curly brackets will be equal to zero in D.r/: Demanding this equality, we have
for all ft; x 2 D.r/g: Equalizing with zero the terms with multipliers of second
order in x x ; then those of first order in the same variable, we observe that the
last equality will be fulfilled if and only if the next equation
is true together with Eq. (5.18) for x .t / and with boundary conditions
d w =dt 0;
w .£; x/ w.ª; x 0 Œª/ D V .b/ .ª; x 0 Œª/ D V .b/ .t; x/; (5.27)
The inclusion
W Œt
W Œt ; (5.28)
is true, where
K D K 1 ; KP D K KP K ;
we come to equations
would be true.
Recall that due to Theorem 3.3.2 and Comparison in Principle as in Sect. 5.1.2,
we may write2 :
\
X Œt D fE .x .t /; XC
l
Œt / j hl; li D 1g
l
Then we have
[
XŒ‚ D fX Œt j t 2 ‚g D fx W V0 .‚; x/ 1g; (5.32)
where
\[
fE .x .t /; XC
l
Œt / j hl; li D 1; t 2 ‚g; (5.35)
l t
where x Œt ; XC
l
Œt are defined by Theorems 3.3.1, 3.3.2 (Eqs. (3.31), (3.32)).
2
In Sect. 5.1.2 we calculated the backward reach set W Œt due to value function V .b/ .t; x/. Here,
in a similar way, we calculate the forward reach set X Œt due to value function V 0 .t; x/.
210 5 The Comparison Principle: Nonlinearity and Nonconvexity
\[
fE .x .t /; XC
l
Œt / j l./ 2 SSg ./; t 2 ‚g; (5.36)
l./ t
where SSg ./ is the set of all good curves kl.t /k D 1; t 2 ‚; normalized according
to Sect. 3.3, to lie on a unit sphere.
Exercise 5.2.1. Indicate conditions when the operation of inclusion ( ) in (5.36)
may be substituted by an equality.
Exercise 5.2.2. Indicate internal ellipsoidal approximations for set XŒ‚:
Example 5.8.
Consider system
3
The illustrations for this example were done by Zakroischikov.
5.2 Calculation of Nonconvex Reachability Sets 211
l
Fig. 5.1 The set XC Œ‚ (in red) and its covering by intersection of nonconvex sets XC Œ‚
Shown in Fig. 5.1 is the set XC Œ‚ (in red) and its covering by intersection of
nonconvex sets XCl Œ‚; calculated along good curves l.t /: An example of such
covering
\
XC Œ‚ j l ; l D XC Œ‚; l XC Œ‚; l
by only two such sets calculated from good curves generated by l .£/ D
.1; 0/; l .£/ D .0:54; 0:83/ is indicated in Fig. 5.2.
Figures 5.3 and 5.4 show set XŒ‚ and its external approximation by an ellipsoidal
tube.
212 5 The Comparison Principle: Nonlinearity and Nonconvexity
Here we briefly indicate a result similar to Theorem 5.1.1 of Sect. 5.1 for a forward
reachability set. Take linear system (5.12), with constraints on x.t0 /; u.t / given by
nondegenerate ellipsoids:
Then
1
.t / D e .t; p/ D hp; B.t /P .t /B 0 .t /pi1=2 :
2
214 5 The Comparison Principle: Nonlinearity and Nonconvexity
Let us now look for w.t; x/ as a quadratic form, w.t; x/ D .hx x .t /; K.t /.x
x .t //i 1; where K.t / D K 0 .t / is differentiable and requires that w satisfies the
PDE
wt C hwx ; A.t /x C B.t /pu i C .4.t //1 hwx ; B.t /P .t /B 0 .t /wx i D 0 (5.40)
Here K.t / may be obtained through a standard procedure of solving the resulting
Riccati equation
1
KP D KA.t / A0 .t /K .K; B.t /P .t /B 0 .t /K/; (5.43)
.t /
with equation
convert into
1
KP C D A.t /KC KC A0 .t / C B.t /Q.t /B 0 .t / C .t /KC ; KC .t0 / D .X 0 /1 ;
.t /
(5.44)
Here
Z t
.t / D .t /=.1 C .s/ds/:
t0
An appropriate selection of i ./ ensures that ellipsoids ECi Œt are tight in the sense
that there is no other external ellipsoid of type (5.44) that could be squeezed in
between ECi Œt and X Œt .
with Hamiltonian
Due to a theorem of Lyapunov on the range of a vector measure (see [3, 202])4 a
vector x.t0 / of type
Z£
x.t0 / D x C Q.£/; Q.£/ D f .t; P .t //dt;
t0
From here it follows that V .£; x/ is closed, convex in x, being the conjugate of a
closed convex function V .£; x/ D ¡.l j Q.£// C ® .l/.
Lemma 5.3.1. Under Assumption 5.3.1 on ®.x/ the following equality is true:
We further have
Z £
D supfhl; xi min maxfhp; x f .t; u.t //dt i ® .p/gg
x u./ p t0
Z £
D max supfhl; xi maxfhp; x f .t; u.t //dt i ® .p/gg
u./ x p t0
Z £
D max sup minfhl p; xi C hp; f .t; u.t //idt C ® .p/g
u./ x p t0
Z £
D max min supfhl p; xi C hp; f .t; u.t //idt C ® .p/g
u./ p x t0
4
Do not confuse famous Lyapunov (1911–1973) with celebrated A.M. Lyapunov (1856–1918),
founder of modern stability theory.
5
Here and below the conjugates of V are taken only in the second variables with £ fixed.
5.3 Applications of Comparison Principle 217
Z £
D maxf hl; f .t; u.t //idt C ® .l/g
u./ t0
Z £
D H .t; l/dt C ® .l/:
t0
Assumption 5.3.1 allowed to interchange the sup and min operations above by
applying the minmax theorem of Ky Fan [72]. Also used was the property
Z £ Z £
max hl; f .t; u.t //idt D maxff .t; u/ j u 2 P .t /gdt;
u./ t0 t0
indicated in [238, Chap. 14, Sect. F]. The above results in the next proposition.
Theorem 5.3.1. Under Assumption 5.3.1 the following relation is true:
Z £
V .£; x/ D maxfhp; xi H .t; p/dt ® .p/g: (5.51)
p t0
With H independent of both t; x the last relation is a formula of the Lax–Hopf type,
[17, 247]. The requirements of Assumption 5.3.1 are clearly satisfied by function
®.x/ D d 2 .x; X 0 / with X 0 convex and compact.
Exercise 5.3.1. For system
H .t; p/ D maxfp1 sin uCp2 cos u j u 2 Œ; g D maxf.p12 Cp22 /1=2 sin.uC’/g;
u ’
where ’ D arc cos .p1 .p12 C p22 /1=2 /: This gives H .t; p/ D .p12 C p22 /1=2 : We also
have
p10 D.x1 x10 /..x1 x10 /2 C.x2 x20 /2 /1=2 ; p20 D.x2 x20 /..x1 x10 /2 C.x2 x20 /2 /1=2 :
Here Vt .£; x/ D 1; Vx .£; x/ D .p10 ; p20 /0 D .cos ®; sin ®/0 , so that H.t; p 0 / D 1:
Hence this equation is satisfied together with its boundary condition. We finally
calculate the reachability set for system (5.52) which is
equation is
Rather than solving the last equation, let us transform Eq. (5.53) to another
system of coordinates (system (5.53) may be linearized). Namely, taking zi D
log xi ; z0i D log xi0 , we have
under the same constraint (5.54). The reach set Z .£; 0; z0 / from point z0 for this
system may be written down using relations similar to (3.32), (2.68), with changes
in the definition of value function V .t; x/ from d 2 .z.t0 /; z0 / to d.z.t0 /; z0 /.
This leads to the value function
where
maxf.l1 log x1 Cl2 log x2 /˚.£; log x10 ; log x20 ; l1 ; l2 /jl W hl; li 1g D 0 (5.58)
l
in the variables x1 ; x2 , where xi0 D exp z0i . The solution to (5.55), (5.56) then
happens to be V .£; x/ D V ? .£; log x1 ; log x2 /.
Rather than calculating the exact V .£; x/ and X .£; 0; x 0 /, it may be simpler to
approximate these. To explore this option, let us first approximate function V ? .£; z/
by quadratic functions and set Z .£; 0; z0 / by ellipsoids.
220 5 The Comparison Principle: Nonlinearity and Nonconvexity
and
where
0 a12 10
AD ; PD :
a21 0 01
In order to ensure the property of tightness for ellipsoids E .z .t /; ZC .t //, and
therefore an equality in (5.46), we have to select the parameterizing functions .t /
as follows (see Chap. 3):
1 1
.t / D hl; F .t /QF 0 .t /li 2 hl; ZC .t /li 2 : (5.60)
Here
0 1
cosh ’t “ sinh ’t
F .t / D exp.At / D @ A:
1
“ sinh ’t cosh ’t
The reach tube ZŒt D Z.t; 0; z0 / will now be exactly described by the formula (see
Chap. 3)
1
z .t; l/ D z? C ZC Œt F .t /lhl; F 0 .t /ZC Œt F .t /li 2 ; (5.61)
E .z .t /; ZC .t // D fz W ˙i;j
2
D1 hzi zi Œt ; ZCij Œt .zj zj Œt //i 1g:
? ?
Here ZCij are the coefficients of ZC . Returning to the original coordinates x and
taking xi? .t / D exp z?i .t /, we observe that ellipsoids E .z .t /; ZC .t // transform into
star-shaped nonconvex sets
S Œt D fx W ˙i;j
2
D1 hlog xi log xi Œt ; ZCij Œt .log xj log xj Œt /i 1g;
? ?
5.3 Applications of Comparison Principle 221
so that now
The surface of the tube X Œt D X.t; 0; x 0 / will be totally covered by the family of
nonintersecting curves .i D 1; 2/
1
xi .t; l/ D exp.log.xi? Œt C .ZC Œt F .t /l/i hl; F 0 .t /ZC Œt F .t /li 2 /;
jxP1 jDu1 jx1 ja12 jx1 j log jx2 j; jxP2 jDu2 jx2 ja21 jx2 j log jx1 j; jx1 j.0/>0; jx2 j.0/>0;
which includes Eq. (5.53), if considered within the domain fx1 > 0; x2 > 0g. The
pictures for the quadrants fx1 < 0; x2 > 0g; fx1 < 0; x2 < 0g; fx1 > 0; x2 < 0g are
symmetrical with the one for fx1 > 0; x2 > 0g, relative to either the coordinate axes
or the origin.
Here Figs. 5.5, 5.6, and 5.7 illustrate the exact value function V .£; x/ for
£ D 0; £ D 0:5; £ D 1, respectively. Figure 5.8 shows four reach tubes X Œt each of
which originates in one of the four coordinate quadrants. Figures 5.9 and 5.10 show
the cuts of the value function (for each of the four reach tubes) surrounded by the
boundaries of sets S Œt —the preimages of ellipsoidal surfaces, for £ D 0:5; £ D 1.6
6
This example was worked out by O.L. Chucha.
222 5 The Comparison Principle: Nonlinearity and Nonconvexity
It is well known that reachability sets for nonlinear systems are typically nonconvex
and may have a peculiar form. But in some cases it may suffice to have ellipsoidal
estimates of their convex hulls instead. This may be given by an intersection
of an array of ellipsoids. Explained here is an example of such a convexifying
approximation. Consider the nonlinear system
5.3 Applications of Comparison Principle 223
xP 1 D x2 ;
xP 2 D r cos x5 ;
xP 3 D x4 ; (5.63)
xP 4 D r sin x5 ;
xP 5 D ’u;
As before, let X Œt D X .t; t0 ; X 0 / stand for the reachability set of system (5.63):
˚ ˇ
X t; t0 ; X 0 D x ˇ 9u./ 2 U ./ W x.t0 / D x .0/ 2 X 0 W x.t; t0 ; x .0/ / D x :
Problem 5.3.1. Find a family of external ellipsoidal estimates for the reachability
set X Œt of system (5.63) such that for each t 2 Œt0 ; t1 it touches the convex hull
convX Œt of X Œt at a certain point x.t /.
To solve this problem we apply Sect. 5.1 of this chapter.
so that
X Œt D fx j V .t; x/ 0 g
To find the necessary approximation for convX Œt we apply the comparison principle
(Theorem 5.1.1). Let wC .t; x/ 2 C1 , .t / 2 L1 , H .t; x; p/ satisfy conditions
X Œt XC Œt ; (5.68)
where
8 9
ˆ
ˆ Z t >
>
< =
C C
XC Œt D x w .t; x/ .£/d £ C max w .t0 ; x/ : (5.69)
ˆ x2X 0 >
>
:̂ t0
;
We now parameterize the family of functions wC .t; x/ by the pair fx./; p./g
which is the solution to the characteristic ODE system for our HJB PDE
equation (5.66) [51]. This ODE is written as
xP 1 D x2 ; pP1 D 0;
xP 2 D r cos x5 ; pP2 D p1 ;
xP 3 D x4 ; pP3 D 0;
xP 4 D r sin x5 ; pP4 D p2 ;
xP 5 D ’Nu; pP5 D r.p4 cos x5 p2 sin x5 /;
where uN D sgn.p5 / with p5 ¤ 0 and uN 2 Œ1; 1 with p5 D 0. We will look for such
ellipsoidal approximations XC Œt of reachability set X Œt , that for each t would touch
convX Œt at point x.t /. We further omit the argument “t ” in x.t N /; p.t
N /, denoting
them as x;N p:N
For arbitrary h > 0, œ1 , œ2 the next inequality is true
1 1
p2 cos x5 C p4 sin x5 h hp; e2 e20 C e4 e40 piC
2
N 2 C 2b.¥ ¥/
cos ¥ a.¥ ¥/ N C c;
226 5 The Comparison Principle: Nonlinearity and Nonconvexity
where
(
N 2.¥N / ; ¥N ¤
sin ¥=
aD ; b D a.¥N /; N
c D cos ¥; ¥N 2 Œ0; 2
1
2
; ¥N D
with equality at ¥ D ¥.N Upper estimates for functions cos ¥, sin ¥, sin ¥ may
be obtained similarly. Applying them, we come to
1 1
p2 cos x5 C p4 sin x5 h hp; e2 e20 C e4 e40 pi C h1 hœ1 e2 C œ2 e4 ; piC
2
with equality at x D x,
N p D p. N
Using this relation, we find an upper estimate H.t; x; p/ for H .t; x; p/:
1 1 1
A D e1 e20 Ce3 e40 ; B D rh .e2 e20 Ce4 e40 /C ’e5 e50 ; C D r.œ1 a1 Cœ2 a2 /e5 e50
2 2jpN5 j
1
D ’jpN5 j C hx x;
N C.x x/i
N C 2hf; x xi
N C r.œ1 c1 C œ2 c2 C ”/:
2
and write down for functions wC .t; x/ and H .t; x; p/ the inequality (5.67):7
7
This example was worked out by V.V.Sinyakov.
5.3 Applications of Comparison Principle 227
In order that this inequality would be true for any .x x .t // it is sufficient to have
the next equations satisfied:
max wC .t0 ; x/ D 1
x2X 0
K .t / D K 1 .t /; KP .t / D K .t /K.t
P /K .t /;
N
If under specific selection of fx./; N
p./g the solution to the last equations exists,
then we may introduce notation
0 1
Zt
KC .t / D @1 C .s/ds A K .t /;
t0
Fig. 5.11 The external ellipsoidal-valued estimate for the reachability tube (1 ellipsoid)
Fig. 5.12 The external ellipsoidal estimate for the reachability set at time t1 D 1 (3 ellipsoids)
5.4 Ellipsoidal Methods for Non-ellipsoidal Constraints 229
In the previous texts the main types of hard bounds on system parameters were taken
as ellipsoidal-valued. However one of the conventional types of such bounds are also
polyhedrons which led to specific algorithms based, for example, on linear program-
ming and related constructions. A detailed theory of approximating reachability
sets and control solutions for linear-convex systems by parallelotopes, including
the rectangular case (boxes), was developed by Kostousova [115]. However, in this
section, we indicate that symmetrical polyhedrons may be well treated by ellipsoidal
approximations, with reasonable computational burden, especially if one involves
parallel calculation.
Let us now assume that system (5.12) is subjected to hard bounds of the “box” type,
namely,
u.t / 2 P .t /; x.t0 / 2 X 0 ;
where
X
m
B .p; P / D fx W x D p C p .i/ ’i ; ’i 2 Œ1; 1g:
iD1
T B .p; P / D B .Tp; TP /:
X
m
B .0; P / D E .0; Qi i / E .0; Q.p//;
iD1
X
m
pi D jli ji if li 6D 0; pi D –m1 jjljj if li D 0; jjljj2 D li2 (5.74)
iD1
q
Here jli j D li2 , so
X
m X
m
¡.ljE .0; Q.p/// D hl; Q.p/li1=2 jli ji C–jjljj; jli ji D ¡.ljB .0; P //
iD1 iD1
(5.75)
and
and a partition ¢ŒN similar to the one of Sect. 2.2. As before, the m m diagonal
matrix P .t / and n m matrix B.t / are continuous.
Then
Z £ X
N X
m
B .0; B.t /P .t //dt D lim E .0; B.tj /Qjj .tj /B 0 .tj //¢i
t0 iD1 j D1
with N ! 1 we have ¢ŒN D maxf¢i g ! 0. Applying again the formula for the
external ellipsoidal approximation of the sum of ellipsoids, we have
8
In the next subsection related to zonotopes we indicate a regularization procedure that ensures
a numerical procedure that copes with such degeneracy. This is done by substituting cylinders for
far-stretched ellipsoids along degenerate coordinates.
232 5 The Comparison Principle: Nonlinearity and Nonconvexity
X
N X
m
E .0; B.tj /Qjj .tj /B 0 .tj //¢i E .0; XCN .pN Œ//;
iD1 j D1
X
N X
m X
N X
m
1
XCN .pN .// D pjj .ti / pjj .ti /Qjj .ti / ; pjj .ti / > 0:
iD1 j D1 iD1 j D1
n Z
X
£ X
n Z £
1
XC .pŒ/ D pjj .t /dt pjj .t /B.t /Qjj .t /B 0 .t /dt
j D1 t0 j D1 t0
where
X
n m Z
X
£
.0/
XC .£; pŒ/ D pkk C pjj .t /dt (5.81)
kD1 j D1 t
0
X
n m Z
X £
T00 C 1
.t /T .t /B.t /Qjj .t /B 0 .t /T 0 .t /dt :
.0/1
pkk 0
T0 Xkk pjj
j D1 j D1 t0
.0/
Here pŒ D fpkk ; pjj ./ j k D 1; : : : ; nI j D 1; : : : ; m; t 2 Œt0 ; £g:
5.4 Ellipsoidal Methods for Non-ellipsoidal Constraints 233
would be possible for a given l 2 Rn , we would formally have to choose XC .£; pŒ/
taking
(5.83)
But a nondegenerate matrix XC .£; pŒ/ would be possible only if pjj .t / ¤ 0 almost
.0/
everywhere and pkk ¤ 0. The equality is then checked by direct calculation.
Lemma 5.4.3. In order that for a given l 2 Rn there would be an equality (5.82), it
.0/
is necessary and sufficient that pjj .t /; pkk would be selected according to (5.83)
.0/
and both of the conditions pjj .t / ¤ 0 almost everywhere and pkk ¤ 0 would be
true.
Otherwise, either an equality (5.82) will still be ensured, but with a degenerate
E .0; XC .£; pŒ//, or, for any – given in advance, an inequality
Z £
¡.ljE .0; XC .£; pŒ// ¡.ljB .0; T0 X 0 // ¡.ljB .0; T .t /B.t /P .t ///dt –jjljj
t0
(5.84)
may be ensured with a nondegenerate E .0; XC .£; pŒ//. This may be done by
selecting
.0/
pŒ D p – Œ D fpkk ; pjj ./ j k D 1; : : : ; nI j D 1; : : : ; m; t 2 Œt0 ; £; g
as
–jjljj
pjj .t / D hl; T .t /B.t /Qjj .t /B 0 .t /T 0 .t /li1=2 C
.–/
;
2m.£ t /
–jjljj
T00 li1=2 C
.0–/
pkk D hl; T0 Xkk
0
:
2n
This follows from the definition of linearly independent functions. Note that with
– D 0 we have
X
n m Z
X
£
.0/
pkk C pjj .t /dt D hl; XC .£; pŒ/li1=2 ; (5.85)
kD1 j D1 t
0
X
n X
n
pkk D hl; T0 X 0 T00 li1=2 T00 li1=2 :
.0/
hl; T0 Xkk
0
kD1 kD1
The parameters of the ellipsoid E .0; XC .£; pŒ// may be expressed through a
differential equation. Taking XC Œ£ D XC .£; pŒ/ and differentiating it in £, we get
X
m X
n
.0/ 1
XP C Œ£ D pjj .£/ pkk 0
T0 Xkk T00
j D1 kD1
m Z
X
£
1
C pjj .t /T .t /B.t /Qjj .t /B 0 .t /T 0 .t /dt
j D1 t0
X
n m Z
X
£ X
m
1
.£/T .£/B.£/Qjj .£/B 0 .£/T 0 .£/
.0/
C pkk C pjj .t /dt pjj
kD1 j D1 t0 j D1
Denoting
X
n m Z
X
£ 1
D j jlj jhl; XC Σli1=2 ;
.0/
j .£/ D pjj .£/ pkk C pjj .t /dt
kD1 j D1 t
0
X
n X
n
pkk /1 ; XC D hl; T0 X 0 T00 li1=2 T00 li1=2 T0 Xkk T00
.0/ .0/ .0/
k D pkk . 0
hl; T0 Xkk
0 0
kD1 kD1
Consider system (5.12) under box-valued constraints (5.72). Its reach set will be
X Œt D G.t; t0 /X Œt ; where
Zt
X Œt D B .x ; X / C
0 0
G.t0 ; s/B.s/B .u0 .s/; P .s//ds: (5.87)
t0
Let us first apply the results of the previous section to the approximation of X Œt .
Taking T .s/ D G.t0 ; s/; T0 D I , we have
X Œt E .G.t0 ; t /x 0 .t /; XC Œt //;
where
X
m X
m
XP C Œt D j .t / XC Œt C j1 .£/G.t0 ; t /B.t /Qjj .t /B 0 .t /G 0 .t0 ; t /;
j D1 j D1
(5.88)
with initial condition
X
n
.k /1 Xkk
.0/
XC Œt0 D 0
; (5.89)
kD1
236 5 The Comparison Principle: Nonlinearity and Nonconvexity
Further on, denoting XC Œt D G.t; t0 /XC Œt G 0 .t; t0 /, we obtain
0
where now XC D XC Œt0
X
m X
m
XP C
Œt D A.t /XC 0
C XC A .t / C j .t / XC Œt C j1 .t /B.t /Qjj .t /B 0 .t /:
j D1 j D1
(5.92)
Theorem 5.4.4. The inclusion (5.91) is true, whatever be the parameters
j .t / > 0; k > 0 of Eq. (5.92).
Let us now presume that Assumption 3.3.1 of Chap. 3 is fulfilled, so that vector
function l.t / along which we would like to ensure the tightness property is taken as
l.t / D G.t0 ; t /l; l 2 Rn . Then, following the schemes of Sect. 3.8 of Chap. 3, we
come to the next results.
Theorem 5.4.5. Under Assumption 3.3.1 of Chap. 3, in order that equality
would be true for a given “direction” l, the external ellipsoids E .x 0 .t /; XC .t; pŒ//
should be taken with
0
instead of j .t /; XC . The inequality
X Œt D X0 Œt C Xu Œt ;
where X0 Œt D G.t; t0 /B .x 0 ; X 0 / is a (nonrectangular) box and
Zt
X u Œt D G.t; s/B.s/B .u0 ; P .s//ds;
t0
would be possible for an appropriately selected ellipsoid E .0; XC Œt /, it is sufficient
that the pair fA.t /; e g would be completely controllable for any k D 1; : : : ; m.
k
Then XC Œt will be correctly defined when described by Eq. (5.92), with parameters
0
j .t / and initial condition XC Œt0 D XC D 0; selected due to (5.93), (5.94).
This follows from the definition of complete controllability (see Sect. 1.3). Under
such condition the boundary of set XC Œt will have no faces (“platforms”) and the
set can be totally described by “tight” ellipsoids, as in Chap. 3, Sects. 3.7 and 3.9
(see Fig. 3.1).
Finally, a parametric presentation of set X Œt , similar to (5.92), (5.93) can be
produced. Then
x .t / D x 0 .t / C XC
Œt l hl ; XC
Œt l i1=2 ; (5.97)
with
x .t0 / D x 0 C XC
0
li ; l .t / D G.t0 ; t //0 l
0 1=2
lhl; XC (5.98)
.0–/
with either j .t /; k0 or j– .t /; k selected as indicated in Theorem 5.4.5. This
results in an array of external ellipsoids of either type E .x 0 .t /; XC Σ/ which yields
an equality similar to (5.82), namely
–
or type E .x 0 .t /; XC Œt /, which yields
¡.l jE .0; XC
.£; pŒ// ¡.l j X Œ£/ –jjl jj (5.100)
In Figs. 5.13 and 5.14, we return to Example 3.6 of Chap. 3. Here external
ellipsoidal approximations are used to construct reachability sets emanated from
a box X 0 D X Œt0 D fx W jxi j 1; i D 1; 2g rather than from an ellipsoid,
as in Chap. 3, Sect. 3.3. Calculations are made due to relations (5.90), (5.92), and
Theorems 5.4.4, 5.4.5, with m D 1; n D 2:9
t = 0.5
5
x2
−5
−5 −4 −3 −2 −1 0 1 2 3 4 5
x1
t=1
5
x2
−5
−5 −4 −3 −2 −1 0 1 2 3 4 5
x1
9
Illustrations to this example were worked out by M. Kirillin.
5.4 Ellipsoidal Methods for Non-ellipsoidal Constraints 239
5
x2
−5
0 0
0.2
0.4
0.6 x1
0.8 1 −5
t
Following Remark 5.4.1, we recall that the results of Chap. 3, Sects. 3.2–3.4, 3.7,
and 3.8 are all also true for degenerate ellipsoids. We may therefore directly apply
them to box-valued constraints
P .t / D B .u0 .t /; P .t //; X 0 D B .x 0 ; X 0 /;
X
m X
n
B .u0 .t /; P .t // D E .u0 .t /; Qi i .t //; B .x 0 ; X 0 / D E .x 0 ; Xkk /;
j D1 kD1
where
0 0
Qi i .t / D e.i/ e.i/ qi i ; qi i D 2i ; Xkk
0
D e.k/ e.k/ xkk
0 0
; xkk D k2 ;
and e.i/ ; e.k/ are unit orths in the respective spaces Rm ; Rn . This leads to the
following statements.
240 5 The Comparison Principle: Nonlinearity and Nonconvexity
where
Z X t !
X
n n
X .t / D 0 1=2 0
.Xkk / S0k C G.t0 ; s/.B.s/Qjj .s/B 0 .s//1=2 Sj0 .s/ds
kD1 t0 j D1
0
and S0k ; Sj are any orthogonal matrices of dimensions n n; .S0k S0k DI;
0
S Sj Sj D I / is an internal ellipsoidal approximation for the reach set X Œt
of (5.87).
(ii) In order that for a given “direction” l the equality
k D 1; : : : ; n; j D 1; : : : ; m; s 2 Œt0 ; t ;
E .x 0 .t /; X .t // X Œt ;
where
X
m
YP 0 D A.t /Y 0 C .B.t /Qjj B 0 .t //1=2 Sj0 .t /;
j D1
X
n
Y 0 .t0 / D 0 1=2 0
.Xkk / S0k :
kD1
0
and S0k ; Sj are orthogonal matrices of dimensions nn, (S0k S0k D I; Sj Sj0 D I ).
In order that equality ¡.l jX Œt / D ¡.l jE .x 0 .t /; X .t /// would be true, it is
necessary and sufficient that relations of type (5.102) would be satisfied.
5.5.1 Zonotopes
In this section we consider hard bounds on control u and initial set X.t0 / D X 0 in
the form of symmetric polyhedrons also known as zonotopes, denoted here as
X
m
Z .p; P / D fx W x D p C p i ’i ; ’i 2 Œ1; 1g:
iD1
X
m
¡.ljZ .p; P // D hp; li C jhp i ; lij:
iD1
0 2
X0 P(t)
0.2 1.5
0.4
1
0.6
0.5
0.8
1 0
1.2 0.5
1.4
1
1.6
1.5
1.8
2 2
0 0.5 1 1.5 2 2 1 0 1 2
Fig. 5.15 Hard bounds on control u and initial set X 0 for Example 5.9
Examples of zonotopes used below in Example 5.9 are given in Fig. 5.15. Note
that a particular case for the zonotope is a box which is initially defined by a set of
orthogonal vectors p i .
Definition 5.5.2. An ellipsoid E .q; Q/ with support function
X
mp
X
m0
u.t / 2 p.t / C E .0; Pjj .t //; x.t0 / 2 x 0 C E .0; Xi0i /; (5.104)
j D1 iD1
0 0
Here Pjj .t / D p j .t /p j .t /, Xi0i D x i x i
In order to calculate reachability sets for zonotopes one may follow the schemes
of the previous section as taken for boxes. These may be repeated for zonotopes as a
useful exercise. We formulate the main task beginning with internal approximation.
5.5 Ellipsoidal Methods for Zonotopes 243
Problem 5.5.1. (i) For the reachability set X Œt D X .t; t0 ; X 0 / of system (5.12),
under constraints 5.103 specify the family E D fE g of internal ellipsoidal
approximations such that
[
conv E j E 2 E D X Œt :
E
(ii) Indicate a subfamily of tight recurrent approximations El Œt X Œt , that touch
the reachability set along “good” curves” ensuring equalities
[
¡.l .t /jEl Œt / D ¡.l .t /jX Œt /; conv El Œt D X Œt :
l
Recall that such a tight approximation is unique if it is selected among the class
E , in the sense of Definitions 3.2.3, 3.7.3 of Chap. 3. Namely E Œt is unique if for
any ellipsoid E that satisfies E Œt E X Œt it follows that E D E Œt :
Following the reasoning of Sects. 5.4.3 and 5.4.4, but applying it our zonotopes
Z .p.t /; P .t //; Z .x 0 ; X 0 /; we come to a conclusion similar to Theorem 5.4.8.
Theorem 5.5.1. (i) The reachability set X Œt allows to have the following internal
approximation
Z t
E .x .t /; X .t // G.t; t0 /Z .x ; X /C0 0
G.t; s/B.s/Z .p.s/; P .s//ds D X Œt ;
t0
X
m0 Z tX
mp
1 1
X .t /D S0;i .Xi0i / 2 G 0 .t; t0 /C Sj .s/.B.s/Pjj .s/B 0 .s// 2 G 0 .t; s/ds;
iD1 t0 j D1
Z (5.105)
t
x .t / D G.t; t0 /x 0 C G.t; s/B.s/p.s/ds; (5.106)
t0
t
if only the time-varying parameters S0;i , Sjt ./ are selected such that
1 1
t
S0;i .Xi0i / 2 G 0 .t; t0 /l 0 D œt0;j p t ; Sjt .s/.B.s/Pjj .s/B 0 .s// 2 G 0 .t; s/l D œtj .s/p t ;
(5.108)
s 2 Œt0 ; t ; i D 1; : : : ; m0 ; j D 1; : : : ; mp ;
for certain scalar functions œtj ./ > 0; œt0;j > 0; j D 1; : : : ; mp ; and
functions p t 2 Rn , p t ¤ 0.
Differentiating relations (5.105), (5.106), and applying equality
Among these ellipsoidal-valued tubes we now have to specify those that ensure
tight approximations.
We now as usually (see Sects. 3.3 and 3.9) select a “good” curve as a solution to
equation
In order to satisfy this equality it is necessary to ensure, due to Theorem 5.5.1 (ii),
that orthogonal matrices S0;i t
, Sjt ./, i D 1; : : : ; m0 , j D 1; : : : ; mp would satisfy
conditions (5.108) for certain œt0;i > 0; œtj ./ > 0; p t 2 Rn , p t ¤ 0. Substituting
in these conditions l .t /, we arrive at
1 1
t
S0;i .Xi0i / 2 l 0 D œt0;i p t ; Sjt .s/B.s/.Pjj .s// 2 l .s/ D œtj .s/p t ; (5.111)
s 2 Œt0 ; t ; i D 1; : : : ; m0 ; j D 1; : : : ; mp :
5.5 Ellipsoidal Methods for Zonotopes 245
where
These relations represent the maximum principle and the ellipsoidal maximal
principle which hold simultaneously at x D x .l/ .t /; l.t / D l .t /: They also
indicate that at each time t point x .l/ .t // lies on the support plane to X Œt generated
by vector l .t / with equalities (5.112) reached at x .l/ .t /:
Remark 5.5.1. (a) Note that in formula (5.113) vector x .l/ .t / does not depend on
the length of l .t /. Then, denoting x .l/ .t / D xŒt; l .t /; we may consider the
surface „ Œt; l 0 D xŒt; l .t /; where l .t / depends on l 0 , which together with
t serve as a parameter of dimension n C 1. Hence, for t given, we have the
n-dimensional surface
[
„ Œt ; D fxŒt ; l0 j kl 0 k D 1g
and
[
„Œt ; D @X Œt D fxŒt ; l 0 j kl 0 k D 1g;
246 5 The Comparison Principle: Nonlinearity and Nonconvexity
u 2 Z .p; P /
x.0/ 2 Z .x0 ; X0 /
Problem 5.5.2. (i) For the reachability set X Œt D X .t; t0 ; X 0 / of system (5.12),
under constraints 5.103 specify the family E D fE Cg of external ellipsoidal
approximations such that for a given © > 0 we have
\ \
h.X Œt ; fEC j EC 2 Eg/ ©; X Œt fEC j EC 2 Eg:
l
(ii) Indicate a subfamily of tight recurrent approximations EC Œt
X Œt , that touch
the reachability set along “good” curves ensuring equality
\
l
¡.l.t /jEC Œt / ¡.l.t /jX Œt / C ©kl.t /k; h.X Œt ; ECl Œt / ©:
klk1
10
Examples illustrated in Figs. 5.16, 5.17, 5.18, and 5.19 of this section were worked out by M.
Kirillin.
5.5 Ellipsoidal Methods for Zonotopes 247
Fig. 5.18 Reachability tube X©z Œt with one tight external approximation
© X m
Pi©i D Pi i C I; P © D Pi©i D P C ©I:
m iD1
X
m
Z© .0; P / D Z .0; P © / D E .0; Pi©i / Z .0; P /
iD1
Pm
Here with X 0 D iD1 Xi0i ; X 0 C ©I D X 0© ; we have the ©-neighborhood of
Z .0; X 0 / denoted as
X
m
Z© .0; X 0 / D Z .0; X 0© / D E .0; Xi0©i / Z .0; X 0 /
iD1
Then the reachability set of system (5.12) under such constraints will be
denoted as
Z t
X© Œt D x .t / C G.t; t0 /Z .0; X / C
z 0©
G.t; s/B.s/Z .0; P © .s//ds; (5.115)
t0
Z t
x .t / D G.t; t0 /x 0 C G.t; s/B.s/p.s/ds: (5.116)
t0
Exact reach set X Œt will be defined through (5.115) where X 0© ; P © .s/ are
substituted by X 0 ; P .s/.
Lemma 5.5.1. For any © > 0 we have
(i) the inclusion X Œt X©z Œt ,
(ii) the inequality h.X Œt ; X©z Œt / k.t /©; where
Z t ˇ
ˇ
k.t / D sup kG 0 .t; t0 /lk1=2 C kB.s/G 0 .t; s/lk1=2 ds ˇ l 2 B .0/ :
ˇ
t0
250 5 The Comparison Principle: Nonlinearity and Nonconvexity
for any positive p0;i > 0, i D 1; : : : ; m0 ; and any continuous functions j .t / >
0, j D 1; : : : ; mp .
(ii) At each time t ellipsoid E .x .t /; XC
©
.t // touches the ©-neighborhood X©z Œt along
0
direction l .t / D G .t0 ; t /l , so that relation
0
Relation (5.119) yields the next property: set X©z Œt may be presented as the
following intersection of approximating ellipsoids, namely
11
Note that designing external approximations of reachability sets X Œt we have to apply them
to nondegenerate ©-neighborhoods of ellipsoids Ei i instead of exact Ei i , as in Sect. 5.4.2. This is
because all ellipsoids Ei i are degenerate, and their approximations may also turn out to be such. But
since we need all externals to be nondegenerate, this will be guaranteed by applying our scheme to
nondegenerate neighborhoods of Ei i .
5.5 Ellipsoidal Methods for Zonotopes 251
For a fixed vector l 0 2 Rn let xl .t / be the related good curve along which
ellipsoid E .x .t /; XC
©
.t // touches the tube X©z Œt ; ensuring equalities
hl .t /; xl .t /i D maxfhl .t /; xi j x 2 E .x .t /; XC
©
.t //g; (5.122)
along trajectory
XC ©
.t /l .t /
xl .t / D x .t / C :
hl .t /; XC©
.t /l .t /i1=2
cylinder.”
Example 5.10. Is the same as Example 5.9. Here Fig. 5.18 indicates the
©-neighborhood X©z Œt for the exact zonotopic reachability tube X Œt calculated as
an intersection of its tight external approximations one of which is shown touching
X©z Œt along a good curvel .t /. Figure 5.19 demonstrates the zonotopic cross-section
of the ©-neighborhood X©z Œt at fixed time t D 1; together with an array of its tight
ellipsoidal approximations.
Chapter 6
Impulse Controls and Double Constraints
Abstract In the first section of this chapter we deal with the problem of feedback
impulse control in the class of generalized inputs that may involve delta functions
and discontinuous trajectories in the state space. Such feedback controls are not
physically realizable. The second section thus treats the problem of feedback control
under double constraints: both hard bounds and integral bounds. Such solutions are
then used for approximating impulse controls by bounded “ordinary” functions.
Control inputs in the form of impulses of the •-function type have been studied since
the conception of control theory, being motivated by designing space trajectories,
automation, biomedical issues, and also problems in economics (see [23, 42, 43]).
However, impulse control problems were initially studied mostly as open-loop
solutions, [120, 219] with detours to closed-loop strategies being rather rare [23].
In the first section we describe the Dynamic Programming approach to problems
of feedback impulse control for any finite-dimensional linear system. This is
reached through variational inequalities that propagate HJB techniques to impulse
control. So, we begin with problems of open-loop control, then concentrate on
closed-loop solutions emphasizing their specifics and the type of HJB equation
that describes them. The solutions arrive as •-functions. However such functions
are ideal mathematical elements which require physically realizable approximation
by “ordinary” functions. This is done in the second section, where we solve the
problem with double constraints—joint hard instantaneous and soft L1 —integral
bounds on the controls. The achieved solution then yields realizable approximations
of impulsive feedback inputs. The contents of this chapter rely on investigations
[55, 152, 154].
Considering impulse control problems in finite time, it could at first seem reason-
able, say, in system (1.2), to minimize the norm ku./g of the control, taking it in the
space L1 . But due to the specifics of this space the minimum may not be attained
within its elements. It is therefore natural to apply the next scheme.
Consider the following generalization of the Mayer-Bolza problem in classical
Calculus of Variations (see [33]).
Problem 6.1.1. Minimize the functional
x.t0 0/ D x 0 : (6.3)
Here ®.x/ is a convex function, VarU./ stands for the total variation of function
U./ over the interval Œt0 ; ªC0, where U./ 2 Vp Œt0 ; ª is the space of vector-valued
functions of bounded variation. The generalized control U.t / attains its values in Rp
which means that each component Ui .t /; .i D 1; : : : ; p/; assumed left-continuous,
is a function of bounded variation on Œt0 ; ª C 0 and B.t / 2 Rnp are taken to be
continuous. The minimum over U./ will now be attained.1
Equation (6.2) with condition (6.3) is a symbolic relation for
Z t
x.t / D G.t; t0 /x 0 C G.t; Ÿ/B.Ÿ/d U.Ÿ/; (6.4)
t0
where the last term in the right-hand side is a Stieltjes or Lebesgue–Stieltjes integral
(see [234]). The terminal time is fixed and the terminal cost function ®.x/ W Rn !
R [ f1g is closed and convex.
A special selection of ®.x/ D I .x jx .1/ / yields
Problem 6.1.2. Steer x.t / from point x 0 D x.t0 / to point x .1/ D x.ª/ with
minimal variation of control U.t /:
1
In this book we give a concise description of impulse controls that are confined only to
•-functions, but not their derivatives. A general theory of impulse control that also involves
derivatives of •-functions is beyond the scope of this book and is presented in [152, 154]. Such
theory leads to the description of “fast” or “ultra fast” control solutions achieved on a quantum
level and in “nano”-time.
6.1 The Problem of Impulse Controls 255
Our main interest in this book lies in closed-loop control, but prior to that we start
with open-loop solutions presenting results in terms of the above. Such type of
problems in terms of the moment problem was indicated in Sect. 1.3 and addressed
in detail in [120, 219].
Define
as the value function for Problem 6.1.1. We shall also use notation V .t0 ; x 0 / D
V .t0 ; x j ª C 0; ®.//; emphasizing the dependence on fª; ®./g:
Let us start by minimizing V1 .t0 ; x 0 / D V .t0 ; x 0 j ª; I . j x .1/ //: Then we first
find conditions for the solvability of boundary-value problem (6.6) under constraint
VarU./ ; with given.
Since
and since the conjugate space for vector-valued continuous functions C p Œt0 ; ª is
Vp Œt0 ; ª; then treating functions B 0 .t /G.ª; t /l as elements of C p Œt0 ; ª; we have
hl; x .1/ i hl; G.ª; t0 /x 0 i C kB 0 ./G 0 .ª; /lkC Œt0 ;ª ; (6.7)
whatever be l 2 Rn :
Here
kB 0 ./G 0 .ª; /lkC Œt0 ;ª D maxfkB 0 .t /G 0 .ª; t /lk j t 2 Œt0 ; ªg D klkV ;
t
and klk in the braces is the Euclidean norm. We further drop the upper index p in
C p ; Vp while keeping the dimension p.
Denote c.t0 ; x .1/ / D x .1/ G.ª; t0 /x .0/ . Moving the first term in the right-hand
side of (6.7) to the left, then dividing both sides by kB 0 ./G 0 .ª; /lkC Œt0 ;ª , we come
to the next proposition.
256 6 Impulse Controls and Double Constraints
Theorem 6.1.1. Problem 6.2 of steering x.t / from x 0 D x.t0 / to x .1/ D x.ª C 0/
under constraint VarU.jŒt0 ; ª/ is solvable iff (6.7) is true for all l 2 Rn :
The optimal control U 0 .t / for this problem is of minimal variation
( ˇ )
hl; c.t0 ; x .1/ /i ˇ
ˇ
VarU ./ D D sup
0 0
ˇl 2R ; :
n
(6.8)
kB ./G 0 .ª; /lkC Œt0 ;ª
0 ˇ
Note that the criteria of controllability in the class of functions U./ 2 Vp Œt0 ; ª
for system (6.2) are the same as in Sect. 1.3.
Lemma 6.1.1. In (6.8) the maximum (supremum over l) is attainable if system (6.2)
is strongly completely controllable. (Assumption 1.5.2 is true.)
Indeed, if 0 6D 0 and Assumption 1.5.2 requiring strong complete controllability
is true, then klkV defines a finite-dimensional norm in Rn (check this property (!))
and (6.8) is equivalent to
Z ªC0
D hl 0 ; G.ª; Ÿ/B.Ÿ/d U 0 .Ÿ/i: (6.10)
t0
This is the maximum principle for impulse controls which is given in integral
form.
Denote §0 ŒŸ D §.Ÿ; ª; l/ D l 0 G.ª; Ÿ/: Then we can rewrite (6.10) as
(Z ˇ ) Z ªC0
ªC0 ˇ
ˇ
0 D max § ŒŸB.Ÿ/d U.Ÿ/ ˇ VarU./ D
0 0
§0 ŒŸB.Ÿ/d U 0 .Ÿ/:
t0 ˇ t0
(6.11)
Theorem 6.1.2. The optimal impulse control U 0 ./ that minimizes variation
VarU./ satisfies the maximum principle (6.11) under l D l 0 —the maximizer
of (6.9). With 0 > 0 and under Assumption 1.5.2 it is also sufficient for the
optimality of U 0 ./:
6.1 The Problem of Impulse Controls 257
Lemma 6.1.2. (i) Among optimal controls U 0 ./ there always exists one of type
U.t / D const; £i1 < t £i ; i D 1; m C 1;
U.t / W
U.£i C 0/ U.£i / D ’i ; i D 1; m:
d U 0 Œt Xm
D p .i/ •.t £i /; (6.12)
dt iD1
where
X
m
VarU ./ D
0
jp .i/ j D 0 I
iD1
2
The proof of this Lemma may be found in [120, 219].
258 6 Impulse Controls and Double Constraints
with positive or negative impulse (a strike), so as to calm down the motion to a total
stop at given finite time.The problem is to find the total number of strikes with their
intensities and directions.
The equation for the motion of the controlled ball is
sR C ¨2 s D u; (6.13)
where s is the length from the origin along the curve with t 2 Œ0; ™. The stopping
rule is
s.ª/ D 0;
sP .ª/ D 0:
Exercise 6.1.4. 1. What should be the curve that yields equation of motion without
friction to be (6.13)?
2. Write down the equation of motion for the ball when the curve is a circular arc.
Write down (6.13) as
dx1 D x2 dt;
;
dx2 D ¨2 x1 dt C Udt;
Hence
1
sin.¨t /
G.=¨; t /B D ¨ ;
cos.¨t /
with
0 1
x .1/
D x.=¨/ D ; x .0/
D x.0/ D ;
0 0
Then
1
c D x .1/ X.=¨; 0/x .0/ D :
0
l1 ¨
0 D max :
l max jl1 sin.¨t / C l2 cos.¨t /j
0t ¨
q
Introducing ® D arccos l1 = l1 C l2 we get
2 2
q
l1 sin.¨t / C l2 cos.¨t / D l12 C l22 sin.¨t C ®/;
so that
l1 ¨
0 D max q : (6.14)
l
l12 C l22 max j sin.¨t C ®/j
0t=¨
Note that the right hand side of (6.14) depends only on the ratio ® D arctg.l2 = l1 /.
Then, since l1 > 0 is obviously infeasible, we may take l1 D 1, arriving at the
minmax problem of finding
q
min 1 C l22 max j sin.¨t C ®/j;
l2 0t=¨
Its solution gives l20 D 0, 0 D ¨ with only one extremum t D =2¨. Then the
desired control is
u D u0 D ’• t :
2¨
260 6 Impulse Controls and Double Constraints
Z
ªC0
0
hl ; ci D
0
l 0 G.ª; t /B.t /d U 0 .t /;
t0
Z=¨ Z=¨
sin.¨t /u dt D ¨;
0
cos.¨t /u0 dt H) ’ sin =2 D ¨;
0 0
Returning to Problem 6.1.1 with arbitrary starting position, we shall look for the
value function V .t; x/; assuming klkV is a norm. (As indicated above this is
ensured if system (6.2) is strongly completely controllable.) Recall that 0 D
0 .t; xI ª; x .1/ /:
We have
V .t; x/ D
inff0 .t; x I ª; x .1/ /C®.x .1/ /g D inffkx .1/ G.ª; t /xkV C®.x .1/ / j x .1/ 2 Rn g:
(6.15)
Denote
ˆ.x .1/ / D kc.t; x .1/ /kV C ®.x .1/ /; c.t; x .1/ / D x .1/ G.ª; t /x;
6.1 The Problem of Impulse Controls 261
and suppose inffˆ.x .1/ /jx .1/ 2 Rn g > 1: Then, since kc.t; x .1/ /kV is a norm,
we have
Hence 0 2 int Domˆ which indicates that the level sets of ˆ.x/ are compact
(closed and bounded). Then clearly, V .t; x/ < 1; the “infimum” in x .1/ in the
above is attainable (let it be point x ) and can be substituted for “minimum.” This,
together with Lemma 6.1.2 brings us to the next conclusion.
Theorem 6.1.3. With V .t; x/ < 1 there exists a point x where
For every starting position ft; xg there exists an open-loop control U 0 .jt; x/
of type (6.12) which steers system (6.1) from ft; xg to fª; x g thus solving Prob-
lem 6.1.1.
In order to calculate the conjugate of V .t; x/ in the second variable we apply
formula
where the min and max are interchangeable (check this property). Changing the
order of these operations and using the definition of convex conjugate functions we
have
with
BV Œt; ª D fl W klkV 1g and klkV D maxfkB 0 .Ÿ/G 0 .ª; Ÿ/lk j Ÿ 2 Œt; ª/g;
Ÿ
Theorem 6.1.4. The value function V .t; x/ is convex in x and its conjugate in the
second variable is
Proof. If at time £ there is no jump, then V .£; x.£// D V .£ C 0; x.£ C 0//. So, by
applying formulas (6.16), (6.17) we have:
V .£; xjª C 0; ®.// D maxfhl; G.ª; £/xi ® .l/ I .ljBV Œ£; ª/ j l 2 Rn g;
l
V .t; xj£; V .£; // D maxfhl; G.£; t /xi V .£; l/ I .l j BV Œt; £/ j l 2 Rn g D
l
D maxfhG 0 .£; ª/l; G.ª; t /xi ® .G 0 .£; ª/l/ I .G 0 .£; ª/ljBV Œ£; ª/
l
V .t; x/ D V .£; x.£C0// D V .£; x.£//C’.£/I ’.£/ D V .£; x.£C0//V .£; x.£//:
is true, since
This inequality may even be a strict one. Thus, with ®.x/ D I .xjfx .1/ g/;
B.ª/ D I; we have V .ª; xjª; ®.// D maxl fhl; x x .1/ ijklk 1g D kx x .1/ k
< ®.x/:
The principle of optimality may now be used for deriving an analogy of the HJB
equation. However, here we have to count on two types of points £ 2 Œt; ª, namely,
those, where there is no control (U.t / D const; with t 2 Œ£ ©; £ C © for some
© > 0/ and those where there is an impulse u0 •.t £/:
In the first case take t D £; ¢ 2 Œ0; ©/: Then we come to inequality
d V .£; x/
H1 .£; x; Vt ; Vx / D D Vt .£; x.£//ChVx .£; x.£//; A.£/xi 0: (6.19)
d£
Further on, in the absence of differentiability the scalar product should be
interpreted as the directional derivative.
In the second case there is a jump at t D £: Then £ does not change and we take
Vt .£; x.£// D 0. Here denote the jump as u D ’h; where the unit vector h defines
the direction of the jump and ’ > 0—the size of the jump.
Now we have
where the minimizer h0 of unit length gives us the direction of the jump. Clearly,
here the multiplier ’ > 0 may be omitted.
The size ’ of the jump is obtained from condition
Here x.£ C 0/ D x.£/ C B.£/u0 ; u0 D ’h0 ; ’ > 0: With ®.x/ 0 it can also
arrive from the calculation ’ D V .£ C 0; x.£ C 0// V .£; x.£//:
264 6 Impulse Controls and Double Constraints
We have thus found that V .t; x/ must satisfy the relation (a variational inequality)
H1 .t; x; q; p/ D q C hp; A.t /xi; H2 .t; x; q; p/ D minfhp; B.t /hi C khk j khk D 1g:
h
V .t; x/ D
( ˇ )
hl; xi ˇ jxj
ˇ
D max ˇŸ 2 Œt; 2; l 2 R; l 6D 0 D
l maxŸ jb.Ÿ/lj ˇ maxŸ fb.Ÿ/jŸ 2 Œt; 2g
jxjb 1 .1/; t 1;
D
jxjb 1 .t /; t > 1:
6.1 The Problem of Impulse Controls 265
b.t /
H2 .t; x; Vt ; Vx / D 1 jb.t /Vx j D 1 D .1 t /2 > 0:
b.1/
while
minfhsign.x/; hi C 1g D h0 C 1 D 0;
h
and
Zt
x.t / D x b.t /x.t /d U 0 .t; £/;
0
Impulse controls introduced above are “ideal” elements. In order to allow a physical
realization of the proposed schemes we introduce the next “realistic” scheme. Here
the original control will be subjected to double constraints where in the previous
problem there is an additional hard bound on the norm of the control vector, while
the bounding parameter may tend to infinity.
Problem 6.2.1. Consider Problem 6.1.1 under additional constraint u.t / 2 B .0/,
so that there arrives the next problem: find
Z ª
J.t0 ; x 0 ju.// D ku.t /kdt C ®.x.ª// ! inf; (6.25)
t0
The solution to this problem exists since the set of all admissible controls u./ is now
weakly compact in L2 .Œt0 ; t1 I Rp / and convex, while the cost functional J.u.// is
weakly lower semicontinuous and convex.
The value function for the last problem is
Using the schemes of Chap. 2 (see Sect. 2.3.1), we come to the next proposition.
6.2 Realizable Approximation of Impulse Controls 267
Theorem 6.2.1. The value function V .t; x/ is a solution to the HJB equation
(* + ˇ )
@V .t; x; / @V .t; x; / ˇ
C min ; A.t /x.t / C B.t /u C kuk ˇˇ kuk D 0;
@t @x
(6.28)
with boundary condition
V .t; x; / D
Z ª
D min max .ku.Ÿ/k C hl; G.ª; Ÿ/B.Ÿ/u.Ÿ/i/d Ÿ C hl; G.ª; t /xi
u l t
® .l/ j l 2 Rn ; kuk D
where
Z ª
¥.t; x; l/ D minfhl; G.ª; s/B.s/ui C kuk j kuk gds:
t u
This yields
Lemma 6.2.1. The value function
Z ª
V .t; x; / D maxfhl; G.ª; t /xi fk§ŒsB.s/k 1gC ds ® .l/g D
l t
where
k§ŒsB.s/ 1k; if k§ŒsB.s/k 1 > 0
fk§ŒsB.s/k 1gC D
0; if k§ŒsB.s/k 1 0:
Exercise 6.2.1. (a) Check that function V .t; x; / satisfies Eqs. (6.28), (6.29).
(b) Prove that with ! 1 function V .t; x; / converges pointwise to V .t; x/
of (6.16).
(c) Prove that with ! 1 the HJB Eqs. (6.28), (6.29) converges to
Eqs. (6.22), (6.23).
Lemma 6.2.2. The conjugate function to V .t; x; / in the second variable is
Z ª
V .t; l; / D fkl 0 G.t; s/B.s/k 1gC ds C ® .l 0 G.t; ª//: (6.31)
t
dx
2 A.s/x C B.s/U 0 .s; x; /: (6.32)
ds
Example 6.1.3. We now deal with the same 2-dimensional system as in Exam-
ple 6.1.1, but looking for the solution in the form of feedback control strategies.
This will be done by specifying certain impulse domains which serve as regions
where the instantaneous impulses are activated. They are in a certain sense similar
to switching surfaces for bang-bang control.
We consider system
dx1 D x2 dt;
;
dx2 D x1 dt C d U;
where
sin t cos t
G.=2; t / D ;
cos t sin t
and
where b D Œ0; 10 and ® D arcsin l2 .l12 C l22 /1=2 ; ® 2 Œ=2; =2.
6.2 Realizable Approximation of Impulse Controls 271
and
(q
l12 C l22 l1 l2 > 0;
klkV D
maxfjl1 jj; jl2 jg otherwise:
Then
(q
z21 C z22 z1 z2 > 0;
.0; x/ D kzk D maxfhl; zi
0
jV klkV 1g D
maxfjz1 j C jz2 jg z1 z2 < 0;
where z D G.=2; 0/x D fx2 ; x1 g: The maximizer in the last problem is
8
< .1; 1/ z1 > 0; z2 < 0;
l D .1; 1/ z1 ; 0; z2 > 0
0
:
z=kzk z1 z2 > 0;
and the impulses arrive at £ D argmaxfjb 0 G 0 .; Ÿ/l 0 j j Ÿ 2 Œ0; =2g which here, at
z1 z2 < 0; are £1 D 0; £2 D =2: Then the optimal synthesizing strategy u0 .t; x/ at
point ft 2 .0; ª; xg; for ’1 > 0; ’2 > 0; is
6
wait jump down
4
x2
0
−2
−4
jump up wait
−6
−8
−10
−10 −8 −6 −4 −2 0 2 4 6 8 10
x1
2
x2
−2
−4
−6
−8
−10
−10 −8 −6 −4 −2 0 2 4 6 8 10
x1
6.2 Realizable Approximation of Impulse Controls 273
u = −µ
1
x2 0
−1
u=µ
−2
u=0
−3
Not Solvable
−4
−5
−5 −4 −3 −2 −1 0 1 2 3 4 5
x1
15 u=0 u = −µ
10
5
x2
−5
−10
−20
−25
−25 −20 −15 −10 −5 0 5 10 15 20 25
x1
Chapter 7
Dynamics and Control Under State Constraints
Abstract The topics of this chapter are problems of reachability and system
dynamics under state constraints in the form of reach tubes. Indicated are general
approaches based on the Hamiltonian formalism and a related Comparison Prin-
ciple. Further emphasis is on the dynamics of linear systems under hard bounds
on the controls and system trajectories. A detailed solution is presented based on
ellipsoidal approximations of bounded trajectory tubes.
The issue of control under state constraints is at the heart of many applied problems.
An abundant literature on such issues is published under the notion of “viability
theory” [5, 6]. The specifics of this book lies in general treatment of such problems
under time-dependent state constraints through Hamiltonian techniques with related
Comparison Principle, while the linear-convex case is covered in detail through
ellipsoidal approximations. Such a move provides a natural merger of theory with
appropriate computation schemes. The material of this chapter uses results of
papers [158, 185]. Its topics are also a useful preparation for further consideration
of problems on feedback control under state constraints and external obstacles
addressed in the next chapter.
This section deals with general theory of reach sets and tubes under hard bounds
on controls, and time-dependent geometric (hard) constraints on the states. The
solution relations are given here in Hamiltonian terms, while the linear-convex
case is described through tight external approximations by parametrized ellipsoid-
valued tubes. The result is an exact parametric representation of state-constrained
reachability tubes through families of external ellipsoidal tubes.
We begin with the general nonlinear case. Similar to (1.1) the controlled system is
thus described by the same differential equation, with same properties, namely
X Œ£ D X .£; t0 ; X 0 /
under the restriction x.£/ D x. Here the minimum is over all measurable functions
(controls) u.t / 2 P .t /, function x.t /; .t t0 /; is the corresponding trajectory
of (7.1), and
This follows from the definition of the reach set X .£; t0 ; X 0 /, which is thus a level
set of V .£; x/.
To state the important semigroup property of the value function, we extend
definition (7.3) to more general boundary conditions, namely,
under the restriction x.£/ D x. Function V .t0 ; / defines a given boundary condition.
In (7.2) this condition is V .t0 ; x/ D d 2 .x; X 0 /.
Theorem 7.1.1. The value function V .£; x/ satisfies the principle of optimality,
which has the semigroup form
with boundary condition V .t0 ; x/ D d 2 .x; X 0 /. Here Vt ; Vx stand for the partial
derivatives of V .t; x/. Note that the term d 2 .x; Y .t // 6D 0 only outside the state
constraint Y .t /.
An alternative scheme relies on the value function
The minimum in (7.7) is over all u.t / 2 P .t /; t 2 Œt0 ; £, under restrictions x.£/ D
x and ®.t; x/ 1; t 2 Œt0 ; £. We will use this formulation for the case when
0
®0 .t0 ; x/ D hx x 0 ; .X 0 /1 .x x 0 /i1=2 ; X 0 D X 0 > 0;
so that the initial condition and the state constraints are ellipsoidal:
Once x.£; t0 ; x.t0 // D x.£/ D x is fixed, one may figure out whether x 2
X .£; t0 ; X 0 / by looking at the value V .£; x/. Then the respective vector x.t0 / 2
E .x 0 ; X 0 / iff there exists a control u.t /; t 2 Œt0 ; £; that ensures V .£; x/ 1, so
that x.£/ 2 X .£; t0 ; X 0 / iff V .£; x/ 1.
Therefore the reach set at time £ is
which is thus a level set of V .£; x/, and the previously stated semigroup property of
the value function (the principle of optimality) holds again.
The solution of the reachability problem can again be cast in the form of a
solution of the forward HJB equation—now somewhat different from (7.6). Again
assume the functions V .t; x/ and ®.t; x/ are continuously differentiable.
Denote
and
with
or
With ®.t; x/ D 1 we apply the same principle but only through those controls that
do not allow the trajectory to move outside the state constraint. These are
Let u0 .t; x/ be the optimal control for the trajectory x 0 .t / that emanates from
x.£/ D x and minimizes ®0 .t0 ; x.t0 // under constraints ®.t; x.t // 1; t 2 Œt0 ; £.
Note that with ®.t; x 0 .t // D 1 we have two cases: either H .t; x 0 .t /; ®; u0 / D
0, which means the related optimal trajectory runs along the state constraint
boundary, or
so the optimal trajectory departs from the boundary, and for ¢ > 0 we have ®.t C
¢; x 0 .t C ¢// < 1. Relations H .t; x 0 .t /; ®; u0 / 0 and (7.15) allow one to find the
control u0 along the state constraint boundary.
If all the operations in (7.6), (7.10), (7.11) result in smooth functions, then these
equations may have a classical solution [51]. Otherwise (7.6), (7.7) form a symbolic
generalized HJB equation, which has to be described in terms of subdifferentials,
Dini derivatives, or their equivalents. However, the typical situation is that V is
not differentiable and as said before, in Chap. 5, the treatment of (7.6), (7.7) then
involves the notion of a generalized “viscosity” solution for these equations or their
equivalents [50, 80, 247]. One approach is to use the method of characteristics as
developed for this type of equation (see [51]) and modified for the nonsmooth
case [248]. But it is a fairly complicated procedure, especially in the nonsmooth
case for which the method requires additional refinement. Another approach is
to look for the solution through level set methods [221, 244]. However, for the
specific “linear-convex” problems of this book the function V .t; x/ is convex in
x, hence directionally differentiable in any direction f.1; x/g, which proves it to be
a generalized solution through classical technique (see [16, 123]). Moreover, in this
case an effective ellipsoidal technique may be applied, which allows one to bypass
the calculation of solutions to the HJB equation.
The next question is: on which evolution equation would the multivalued mapping
X Œt D X .t; / D X .t; jX .t0 ; // under state constraint be satisfied? An answer is
280 7 Dynamics and Control Under State Constraints
in papers [158, 164], where this was suggested to be taken as the integral funnel
equation for differential inclusion
under constraint
x 2 Y .t /; t 2 Œt0 ; ª:
hC .X 0 ; X 00 / D minf© W X 0 X 00 C ©B .0/g
Assumption 7.1.1. Functions V .£; x/ are continuous in all the variables, with
nonempty compact zero level-sets that have nonempty interiors of full dimension.
Such level sets
X Œ£ D fx W V .£; x/ 0g 6D ;
may be gotten due to the last assumption for Lemma 7.1.2. Such are those that solve
our problem for linear systems with continuous coefficients and convex constraints
continuous in time. The indicated class of function V .£; / will be denoted as KV .
Remark 7.1.1. The variety KV may be considered as a metric space with metric:
We shall now indicate the specifics of applying the comparison principle of Sect. 5.1
when the HJB equation is derived under state constraints. We further work with
system (7.1), (7.2), but here the state constraint Y .t / will be presented as
we observe, as in Sects. 2.3 and 2.4 (see also [178,198]), that the solution V .t; x/ of
the corresponding “forward” HJB equation
wC C
t C H .t; x; wx / .t /; (7.22)
Theorem 7.1.3. Suppose H C .t; x; p/; wC .t; x/; .t / satisfy Assumption 7.1.2.
Then the following estimate for the reachability set X Œt is true:
X Œt XC Œt ; (7.24)
where
Z t
XC Œt D x W wC .t; x/ .s/ds C V .t0 ; x.t0 // : (7.25)
t0
wC C C C C
t .t; x / C H.t; x ; wx / wt .t; x / C H .t; x ; wx / .t /;
7.1 State-Constrained Reachability and Feedback 283
hence
.i i / w
t .t; x/ C H .t; x; wx .t; x// 0;
Theorem 7.1.4. Suppose functions H .t; x; p/; w .t; x/ satisfy Assumption 7.1.3.
Then the following upper estimate for V .t; x/ is true:
D w
s .s; x.s// C H.s; x.s/; wx .s; x.s/// C d .y.s/ g.s; x.s//; R .s//
2
w
s .s; x.s// C H .s; x.s/; wx .s; x.s/// C d .y.s/ g.s; x.s//; R .s//
2
Z t
V .t0 ; x 0 / C d 2 .y.s/ g.s; x.s//; R .s//ds D V .t; x/:
t0
X Œt X Œt : (7.27)
.1/
Vt C œ™t C H1 .t; x; Vx ; ™x ; œ/ D 0; V .t0 ; x/ D d.x; X 0 / (7.28)
a generalized solution to Eq. (7.29). The comparison theorems and the respective
proofs for Version-B are similar to Version-A.
Remark 7.1.2. In this subsection we dealt with reach sets that are forward. For the
design of state-constrained feedback controls we may need to use backward reach
sets. These are designed similarly and discussed with detail within related schemes
of Sect. 5.1.
We now pass to the class of linear systems with convex constraints where the
previous results may be developed with greater detail.
7.1 State-Constrained Reachability and Feedback 285
in which A.t /; B.t / are continuous, and the system is completely controllable (see
[109]). The control set P .t / is a nondegenerate ellipsoid, P .t / D E .p.t /; P .t //,
and
with p.t / 2 Rp (the center of the ellipsoid) and the symmetric positive definite
matrix function P .t / 2 Rpp (the shape matrix of the ellipsoid) continuous in t .
The support function of the ellipsoid is
¡.l j E .p.t /; P .t /// D maxfhl; ui/ j u 2 E .p.t /; P .t //g D hl; p.t /iChl; P .t /li1=2 :
For each vector l the solution to the optimal control problem is attained at the
terminal point x .0/ .£/ of an optimal trajectory x .0/ .t /; t 2 Œt0 ; £; which starts from
a point x .0/ .t0 / 2 X 0 determined throughout the solution process.
We will solve Problem 7.1.2 by calculating the value function (7.4), namely,
for system (7.30), using the techniques of convex analysis as in paper [166].
To handle the state constraints X Σ, one usually imposes the following constraint
qualification.
Assumption 7.1.4. There exist a control u.t / 2 P .t /; t 2 Œt0 ; £; x 0 2 X 0 , and
© > 0, such that the trajectory xŒt D x.t; t0 ; x 0 / D x.t; t0 ; x 0 j u.// satisfies
where
Z £
C dƒ.t /y.t / C .dƒ.t /Y .t /dƒ0 .t //1=2 :
t0
Theorem 7.1.5. Under Assumption 7.1.4, the support function ¡.l j X Œ£/ is
given by
Let ƒ.0/ ./ 2 arg minf‰.£; t0 ; l; ƒ.// jƒ./ 2 Vn Œt0 ; £g be the minimizer
of (7.14) and let s .0/ Œt D s.t; £; l j ƒ.0/ .// be the solution to (7.30) with ƒ./ D
ƒ.0/ ./.
The application of formula (7.34) indicates a new approach to the next proposi-
tion. Alternate derivations of this approach can be found in [137,166, pp. 116–121].
Theorem 7.1.6 (The “Standard” Maximum Principle Under State Con-
straints). For Problem 7.1.2 under Assumption 7.1.4, the optimal control u.0/ .t /,
initial condition x .0/ , and trajectory x .0/ Œt D x.t; t0 ; x .0/ ju.0/ .// must satisfy the
“maximum principle”
Here we note that the minimum over ƒ./ 2 Vn Œt0 ; £ in (7.33) may be replaced by
the minimum over the pair fM.t /; œ.t /g, with dƒ.t / D l 0 M.t /d œ.t /, in which the
n n matrix M.t / is continuous, and œ./ 2 V1 Œt0 ; £ is a scalar function of bounded
variation. Moreover, M.t / may be chosen within a compact set C0 of continuous
functions. This new form of the multiplier dƒ.t / is a result of combining the earlier
schemes of [96, 158]. We summarize this result as follows.
Lemma 7.1.4. The multiplier ƒ.0/ .t / allows the representation
This is a symbolic expression for the linear differential equation. Under permutabil-
ity of A.t / with its integral, namely, under condition
Z t Z t
A.t / A.s/ds D A.s/dsA.t /
£ £
in which the second integral is a Stieltjes integral. In the absence of the permutation
property the solution is expressed through expanding a matrix series of the Peano
type [32]. Note that with A.t / const the permutability property is always true.
As in the scheme of [158], the single-valued functional ‰.£; t0 ; l; ƒ.// may be
substituted with the set-valued integral
which yields the next result whose analytical form differs from [158].
Lemma 7.1.5. The following equalities hold:
To achieve the desired results, in the following sections we shall use some properties
related to the structure of the optimal controls and the state constraints. These
properties and additional assumptions are summarized below in Sect. 7.2.1 and
are typical of the problems under discussion. They are borrowed from previous
investigations.
290 7 Dynamics and Control Under State Constraints
Starting with Assumptions 7.1.4 and 7.1.5 and presuming the conditions of
Lemma 7.1.6 are true, let us first restrict ƒ in (7.39) to satisfy the relation
dƒ.t / D l 0 M.t /d œ.t /, in which œ.t / is absolutely continuous. This means
dƒ.t / D l 0 M.t /˜.t /dt and ˜.t / D d œ.t /=dt 0 for t 2 TY D ft j xl .t / 2
.0/
.0/
intY .t /g, when xl .t / is the optimal trajectory for Problem 7.1.2, for the given l.
Denoting L.t / D S 1 .t; £ j M./; ˜.//M.t /˜.t /, we may replace
whose solution is SL .t; £/. Then S.t; £ j M./; l.// SL .t; £/; t 2 Œt0 ; £.
In this case, R.£; t0 ; M./; ˜./; X 0 / transforms into
Z £
R .£; t0 ; L.// D SL .t0 ; £/E .x 0 ; X 0 / C SL .t; £/.E .B.t /p.t /; B.t /P .t /B 0 .t //
t0
P / 2 .A.t / L.t //x C L.t /E .y.t /; Y.t // C E .B.t /p.t /; B.t /P .t /B 0 .t //;
x.t
(7.43)
t t0 ; x 0 2 E .x 0 ; X 0 /: (7.44)
Here also the compact set C0 of functions M.t / transforms into a compact set C00
of functions L.t /.
We thus arrive at the following important property, which is similar to those
proved in [158].
Lemma 7.2.1. The reach set X Σ is the intersection
with
¡.l j XL .£; t0 ; E .x 0 ; X 0 /// D hl; xL? Œ£i C hl; GL .£; t0 /X 0 GL0 .£; t0 /li1=2 (7.47)
Z £
C hl; GL .£; s/B.s/P .s/B 0 .s/GL0 .£; s/li1=2 ds
t0
Z £
C hl; GL .£; s/L.s/Y.s/L0 .s/GL0 .£; s/li1=2 ds:
t0
Here
Z £
xL? Œ£ D GL .£; t0 /x C 0
GL .£; s/.B.s/p.s/ C L.s/y.s//ds; (7.48)
t0
and GL .£; s/ D SL .s; £/ is the transition matrix for the homogeneous system
The significance of the last result is that in this specific problem the support
function of the intersection (7.45) is equal to the pointwise infimum (7.46) of the
support functions rather than to their infimal convolution as given by general theory
[237].
It is not unimportant to specify when the infimum in (7.46) is attained; that is,
it is actually a minimum. Indeed, it may happen that for a given l, the minimum
over L./ is in the class C00 (this, for example, is the case when œ.t / turns out to
be absolutely continuous, as in the above). But to ensure the minimum is always
reached, we have to broaden the class of functions L./.
To illustrate how to continue the procedure we will assume the following.
Assumption 7.2.1. For each l 2 Rn , the optimal trajectory xl0 .t / of Problem 7.1.2
visits the boundary @Y .t / only during one time interval Œt1 ; t2 ; t0 t1 t2 £.
(The case of finite or countable collection of such intervals is treated in a similar
way.)
Then, instead of the product M.t /˜.t /, in (7.42) we must deal with multipliers
of the form M .t / D M.t /˜.t / C M1 •.t t1 / C M2 •.t t2 /, where M1 ; M2 are
n n matrices.
By introducing a new multiplier L .t / D L.t / C L1 •.t t1 / C L2 •.t t2 / under
transformation L .t / D S 1 .t; £ j M .//M .t /, we shall match the formulas for
R.£; t0 ; M .// and its transformed version R .£; t0 ; L .//.
Applying the schemes of [96, 97, 158], to the specific case of this paper, it is
possible to rewrite the preceding assertions.
292 7 Dynamics and Control Under State Constraints
(7.49)
¡.l j X Œ£/ D ¡.l j X .£; t0 ; E .x ; X /// D minf¡.l j XL .£; t0 ; E .x ; X // j L ./g;
0 0 0 0
with
and
so that
x.t / D G .£; t0 /x 0 ;
Remark 7.2.1. Note that in the transition function G .t; s/ we have the difference
of a Riemann integral and a Riemann–Stieltjes integral. On the other hand, the last
7.2 State-Constrained Control: Computation 293
The difference between Lemmas 7.2.1 and 7.2.2 is that in the former it is
not guaranteed that the boundary @X Σ is touched at each point by one of the
intersecting sets XL .£; t0 ; E .x 0 ; X 0 //, whereas in the latter the boundary @X Œ£ is
indeed touched at each point by one of the sets XL .£; t0 ; E .x 0 ; X 0 //. This is because
the minimum in (7.47) is attained.
Under Assumptions 7.1.4, 7.1.5, 7.2.1, and Assumption 7.2.3 given below, in
Sect. 7.2.5, the reasoning of the above leads to the next result.
Lemma 7.2.3. The reach set X Œt D X .t; t0 ; E .x 0 ; X 0 // is a convex compact set in
Rn which evolves continuously in t.
The boundary of the reach set X Œt has an important characterization. Consider a
point x on the boundary @X Œ£ of the reach set X Œ£ D X .£; t0 ; E .x 0 ; X 0 //.1
Then there exists a support vector z such that
Then the control u .t /, which steers the system (7.30) from x .t0 / D x 0 to x .£/ D
x under constraints u.t / 2 E .p.t /; P .t //; x.t / 2 E .y.t /; Y.t // while ensuring
1
The boundary @X Σ of X Σ may be defined as the set @X Σ D X Σ n intX Σ. Under the
controllability assumption, X Œ£ has a nonempty interior, intX Œ£ 6D ;; £ > t0 .
294 7 Dynamics and Control Under State Constraints
satisfies the following pointwise “maximum principle” for the control .s 2 Œt0 ; £/:
0 0
z G 0 .£; s/B.s/; u .s/ D maxfz G 0 .£; s/B.s/u j u 2 E .p.s/; P .s//g (7.54)
u
0
D hz ; G 0 .£; s/B.s/p.s/iChz ; G 0 .£; s/B.s/P .s/B 0 .s/G 0 .£; s/z i1=2 ; s 2 Œt0 ; £;
Here G 0 .£; s/ stands for the matrix function G .£; s/ taken for L 0 .s/—the minimizer
of problem (7.53).
The function h.£; s/ D z 0 G 0 .£; s/B.s/ is taken right-continuous.2
Remark 7.2.2. Suppose we want to solve Problem 7.1.2, seeking ¡.l .t / j X Œt /
along a curve l .t /; t > t0 : Then, taking l .t / D z 0 .G 0 /1 .t; s/, one may observe
that the integrands in functional ˆ.l .t /; L 0 ./; t; t0 / of (7.47)–(7.51) will be
independent of t . This property ensures the existence of a recurrent computational
procedure, as indicated in the next section (see also Sect. 2.3). The modified
maximum principle of this section thus allows a solution in recurrent form. This
is not the case for the standard maximum principle.
We now pass to the construction of ellipsoidal approximations for the reach sets.
Despite the linearity of the system, the direct calculation of reach sets is rather
difficult. Among effective methods for such calculations are those that rely on
ellipsoidal techniques. Indeed, although the initial set E .x 0 ; X 0 / and the control
set E .p.t /; P .t // are ellipsoids, the reach set X Œt D X .t; t0 ; E .x 0 ; X 0 // will of
course not generally be an ellipsoid.
2
In the general case, under Assumption 7.2.1, the optimal trajectory may visit the smooth boundary
of the state constraint for a countable set of closed intervals, and function L0 ./ allows not more
than a countable set of discontinuities of the first order.
7.2 State-Constrained Control: Computation 295
As shown in Chap. 3, in the absence of state constraints the reach set X Œt may be
approximated both externally and internally by ellipsoids E and EC , with E
X Œt EC . Here we deal only with external approximations, but taken under state
constraints.
An approximation E .xC ; XC / is said to be tight if there exists a vector z 2 Rn
such that ¡.z j E .xC ; XC // D ¡.z j X Œt / (the ellipsoid E .xC ; XC / touches X Œt
along direction z). We shall look for external approximations that are tight, on the
one hand, and are also recursively computable, on the other.
In order to apply ellipsoidal techniques to state-constrained problems, recall
Lemma 7.2.1 and Corollary 7.2.1, which indicate how the reach set X Œt D
X .t; t0 ; X 0 / may be presented as an intersection of reach sets XL or XL for
system (7.30) without state constraints. We first study how to approximate sets XL
by ellipsoids.
To demonstrate the nature of the procedures suggested in this section we shall
introduce the ellipsoidal technique in two stages. First, transform the system
coordinates in (7.30) from x to z according to the formula z.t / D S.t; t0 /x.t /,
with
of points, namely (I) those that are reached by x .0/ .£/ without having visited the
boundary of the state constraint, then (II) those that are reached by x .0/ .£/ after
having visited the boundary of the state constraint, and (III) those that lie on the
boundary of the state constraint. Since case (I) has been investigated in detail in
Chap. 3, Sects. 3.2–3.4, and case (III) is trivial, our present investigation actually
deals only with case (II).
Apart from Assumptions 7.1.4 and 7.1.5, let us assume in this section that the
functions L./ in what follows do not have any delta function components. (The
case when such components are present is treated in the next section.)
The solution to Problem 7.2.1 is given within the following statement.
Theorem 7.2.3. With l .t / given, the solution to Problem 7.2.1 is an ellipsoid
ELC Œt D EL .xL .t /; XL Œt /, in which
Z t
XL Œt D .pu .t; s/ C pY .t; s//ds C p0 .t / (7.59)
t0
Z t
.pu .t; s//1 GL .t; s/B.s/P .s/B 0 .s/GL0 .t; s/ds
t0
Z t
0
C .pY .t; s//1 GL .t; s/Lt .s/Y.s/Lt .s/GL0 .t; s/dsCp01 .t /GL .t; t0 /X 0 GL0 .t; t0 / ;
t0
and
E .xL .t /; XL .t //
XL .t /
X Œt ;
may be described applying formulas of Sect. 3.3 to system (7.43). Each of these
ellipsoids will be an external estimate for X Œt .
But to describe the collection of tight ellipsoids for X Œt we have to ensure the
following property:
Under our assumptions the minimum over L./ in (7.47) is attained for any l D
l .t / D GL0 .t0 ; t /l 2 Rn , the minimizers being denoted as Lt ./ 2 C00 :3
But prior to moving ahead we have to investigate the following. Suppose element
Lt .s/ is the minimizer of functional
ˆ.GL0 .t0 ; t /l ; L./; t; t0 / D ˆ.l0 ; L./; t; t0 /; l0 D .GL0 .t0 ; t /l ;
The proof is achieved by contradiction. Note that under Assumption 7.2.2 and
due to Property 7.2.1 of Sect. 7.2.5 we may always take L.t / 0 when t < £1 ; t >
£2 , where £1 ; £2 are the points of arrival and departure at the state constraint.
Following Assumption 7.2.2, we proceed further by selecting L./ to be the
minimizer in L of functional ˆ.l0 ; L./; t; t0 /, namely, we now take L.s/ D
Lt .s/ D L .s/; s 2 Œt0 ; t ; which depends on l ; s, but, as indicated in
Lemma 7.2.4, does not depend on t .
Let G .t; s/ D GL .t; s/ under conditions A.t / 0; L.t / L .t /.
Then pu .t; s/; pY .t; s/; p0 .t / of (7.60) transform into
pu .t; s/ D hl ; G .t0 ; s/B.s/P .s/B 0 .s/G0 .t0 ; s/l i1=2 D pu .s/; (7.63)
pY .t; s/ D 0
hl ; G .t0 ; s/L .s/Y.s/L .s/G0 .t0 ; s/l i1=2 D pY .s/;
p0 .t / D hl ; X l i
0 1=2
D p0 I
3
The same minimum value is also attained here in classes of functions broader than C00 .
298 7 Dynamics and Control Under State Constraints
We may now differentiate XC Œt ; xC Œt . According to Sect. 7.2.5 “Specifics”,
below, in Sect. 7.2.5, the necessary condition (7.84) for a jump in L .t / is not
fulfilled, and therefore L .t / L 0 .t /.
Denoting
Z t 1
u .t / D pu .t / .pu .s/ C pY .s//ds C p0 ;
t0
Z t 1
Y .t / D pY .t / .pu .s/ C pY .s//ds C p0 ; (7.66)
t0
and differentiating XC Œt ; xC Œt , we arrive at
XP C
Œt D .u .t / C Y .t //XC Œt C .u .t //1 G .t0 ; t /B.t /P .t /B 0 .t /G 0 .t0 ; t /
C. .t //1 G .t ; t /L .t /Y.t /L0 .t /G 0 .t ; t /; (7.67)
Y 0 0
xP C Œt D G .t0 ; t /.B.t /p.t / C L .t /y.//; (7.68)
XC Œt0 D X 0 ; x Œt0 D x 0 : (7.69)
7.2 State-Constrained Control: Computation 299
Returning to function l .t / D GL0 .t0 ; t /l with GL .t; s/ D G .t; s/, we have
XC Œt D G .t; t0 /XC Œt G0 .t; t0 /
and
so that
and
and
hl .t /; x .t /i D ¡.l .t / j X Œt /; t £1 ; t £2 ;
so that
Remark 7.2.4. If it is not necessary to find the tight external approximations for the
exact reach set X Σ, but a conservative estimate of X Σ using only one ellipsoid
would suffice, then there is no need to solve optimization problem (7.46) in L, since
ellipsoids E .xŒt ; XC Œt /, defined by (7.72), (7.73), happen to be external estimates
of X Σ for all L./. One may single out one of these through some appropriate
criterion.
An obviously conservative estimate may also be obtained by intersecting the
reach set without state constraints with the ellipsoid that defines the state constraint.
The objective of this paper, however, is to indicate the calculation of the exact reach
set under state constraints, especially when the state constraint produces a reach set
different from the one without state constraints.
We now pass to the more general case, in which the multipliers may have delta
function components.
In this section, Assumptions 7.1.4 (of Sect. 7.1.3) and 7.2.1 (of Sect. 7.2.1) and the
conditions of Lemma 7.1.6 are taken to be true together with Assumption 7.1.5. This
allows us to consider ellipsoidal approximations without the additional requirement
of Property 7.2.2 of Sect. 7.2.5.
Now functional ˆ.l; L ./; £; t0 / has the form (7.65), and its minimizer for a given
l D l .£/ is of the form
for all l .£/ with L0 .t / absolutely continuous and L0 .t / 0 whenever x .t / 2
intY .t /. Then the respective transition matrix is G .t; s/ D G0 .t; s/ and l .t / D
0
G0 .t; s/l .
Here L 0 ./ may be interpreted as the generalized derivative of a generalized
Lagrange multiplier ƒ00 ./ similar to ƒ.0/ ./ of Sect. 7.2.1. ƒ00 ./ is piecewise
absolutely right-continuous, with possible jumps at points £1 ; £2 and possible jumps
at t0 and/or £ (when it happens that t0 D £1 and/or £ D £2 ). ƒ00 .t / const,
whenever x .t / 2 intY .t /.
Following the reasoning of the previous section, we may derive equations
for the approximating external ellipsoids similar to (7.72), (7.73). The necessary
prerequisites for such a move are similar to Lemma 7.2.4 and Theorems 7.2.3
and 7.2.4.
Thus, we come to equations for tight external ellipsoids:
X2
dxC D ..A.t /L0 .t //xC C.B.t /p.t /CL0 .t /y.t //dt L0i .xC y.t //d¦.t; £i /;
iD1
(7.76)
0 0
dXC Œt D ..A.t / L0 .t //XC Œt C XC Œt .A .t / L0 .t ///dt (7.77)
302 7 Dynamics and Control Under State Constraints
!
X
k
C u .t / C Y .t / C i XC Œt dt C .u .t //1 B.t /P .t /B 0 .t /dt
iD1
0
X
2
0
C.Y .t //1 L0 .t /Y.t /L0 .t /dt C i1 L0i Y .t /L0i d¦.t; £i /;
iD1
XC Œt0 D X 0 ; xC .t0 / D x 0 :
7.2.4 An Example
xP 1 D x2 ; xP 2 D u; x.0/ D 0; (7.78)
under constraints
This gives parametric equations for a part of the boundary @X Σ, which is:
x1 D .£ ¢/2 =2 C c; x2 D ¢ C d; £1 ¢ £:
Here £1 D = < £ is the first instance when the trajectory reaches the boundary
of the state constraint, while ¢ is a parameter that indicates the instance of £2 when
the trajectory leaves this boundary. Also c D £ 2 =2I d D £: Note that
we have used Assumption 7.1.4 of Sect. 7.2.1, which implies that optimizer ˜0 .t /
satisfies relations
˜0 .t / l1 ; t 2 Œ£1 ; £2 ; ˜0 .t / 0; t 62 Œ£1 ; £2 ;
and s1 .t / l1 .
A typical trajectory for case 1(b) is when it reaches the boundary of the state
constraint at time £1 and then runs along the boundary and leaves it at time £2 . Later
it runs toward the boundary @X Σ while staying in the interior of the constraint.
For the interval Œ0; £, the adjoint system given above, when written in matrix
form (7.42), is
Taking l D l .0/, we will have l0 M.t /˜0 .t / D .0; 1/˜0 .t / when t 2 Œ£1 ; £2 .
This yields M.t / D M0 .t / depending on ˜0 . Direct calculation then allows one to
find for each l the relation
for the optimizer L .t / of (7.53). This is
with L .t / 0 when t < £1 ; t > £2 . Here SM0 .t / is the solution to (7.81) with
initial condition SM0 .£/ D I and M.t / 0; t; £1 ; t > £2 .
Recalling the ellipsoidal equations, we have
Figure 7.1 shows the structure of the reach set boundary, given here by a thick
line. The system trajectory OAG reaches its end point G without visiting the
boundary of the state constraint. Point G is attained at time £. It lies on the reach
set boundary of type 1(a) (see beginning of this Sect. 7.2.4). Trajectory OBF visits
7.2 State-Constrained Control: Computation 305
the boundary of the state constraint at only one point £1 D £2 < £. Its end point
F is of both types 1(a) and 1(b). Trajectory OBCE lies on the boundary of the state
constraint during the interval Œ£1 ; £2 / with £1 < £2 < £, and its end point E lies on
the reach set boundary of type 1(b). Finally, trajectory OBD reaches the boundary
of the state constraint at time £2 < £ and lies on it until final time £, so that point D
is of both types 1(b) and 1(c). Thus, the part of the reach set boundary along points
BDEFGH consists of segments BD [of type 1(c)], DF [of type 1(b)], and FGH [of
type 1(a)].
Figure 7.2 shows the structure of the generalized Lagrange multiplier ˜.t / for the
case of trajectory x2 .t / of type OBCE shown in the top of the figure. Here ˜ 0 on
intervals Œ0; £1 / and Œ£2 ; £ and ˜ const on Œ£1 ; £2 /. Note that the related trajectory
x2 .t / runs along the boundary of the state constraint during the interval Œ£1 ; £2 /,
where ˜.t / 6D 0. Points £1 ; £2 potentially could have •-functions as components
of ˜. But the necessary conditions for the existence of such components are not
fulfilled in this example.
Figure 7.3 shows an external ellipsoid (thin line) for the boundary of the reach
set under state constraints (thick line), indicating its difference from the reach set
without state constraints (dash-dotted line). Figure 7.4 illustrates the final reach tube
with and without state constraints for this example.4
4
The computer illustrations for this subsection were made by M.N. Kirilin.
306 7 Dynamics and Control Under State Constraints
In this subsection we present some additional facts useful for proving the main
assertions of this paper.
One difficulty in solving the control problem under state constraints (Prob-
lem 7.1.2) is to determine the set of times ft W x.t / 2 @Y .t /g. This is a union of
closed intervals during which the optimal trajectory is on the boundary of the state
constraint.
Here are some helpful facts. For Problem 7.1.2 the Properties 7.2.1–7.2.3 are
given below under Assumptions 7.1.4 and 7.2.1,
Property 7.2.1. For any l 2 Rn , the minimizer of (7.35) is ƒ.0/ .t / D const and
in (7.39) the corresponding minimizer is œ.0/ .t / D const during time intervals for
which x.t / 2 intY .t /, the interior of Y .t /.
Since l 0 M .0/ .t /d œ.0/ .t / D dƒ.0/ .t /; we may track whether the trajectory is on the
boundary of Y .t / by the multiplier œ.0/ .t /. Thus, we need not be interested in values
of M .0/ .t / when the trajectory is not on the boundary @Y .t /.
Property 7.2.2. Suppose l 2 Rn is given and x .0/ .t / is the solution of
Problem 7.1.2. For the function ƒ.0/ .t / to have a jump at t , under the smoothness
conditions on the state constraint of Lemma 7.1.6, t must be a time of arrival
308 7 Dynamics and Control Under State Constraints
or departure from the boundary of the tube E .y.t /; Y.t //; t0 t £, and the
trajectory x .0/ .t / must be tangent to the tube E .y.t /; Y.t // at t . Thus, x .0/ .t / is
differentiable at t and
For example, if x .0/ .t ¢/; ¢ > 0, lies inside the interior of the constraint set and
x.t C ¢/ lies on the boundary, then
and (7.84) will be fulfilled only if this inner product equals zero.
We also need the following assumption.
Assumption 7.2.3. For Problem 7.1.2, with given l, there exists no control u .s/
that satisfies the maximum principle
for fs j l 0 S.s; £ j M .0/ ./; œ.0/ .//B.s/ 6D 0g and at the same time ensures for these
values of s that the corresponding trajectory x .s/ 2 @Y .s/.
Property 7.2.3. Assumption 7.2.3 holds.
This means that if the control u.0/ .s/ is determined by the maximum principle,
with h.0/ .£; s/ D l 0 S.s; £ j M .0/ ./; œ.0/ .//B.s/ 6D 0, and therefore attains its
extremal values under given hard bounds, then this control cannot also keep the
corresponding trajectory x .0/ .s/ along the boundary @Y .s/. In other words, in this
case the maximum principle is degenerate along the state constraint, i.e. h.0/ .£; s/
0 and does not help to find the control when the trajectory lies on the boundary
@Y .s/.
This assumption excludes the case when solution of Problem 7.1.2 without state
constraints already satisfies the given state constraints, which therefore turn out to
be passive.
Lemma 7.2.5. Under Assumption 7.83 applied to Problem 7.1, the function
Remark 7.2.6. Recall that equations for tight external ellipsoids without state
constraints are explained in detail in Chap. 3, Sects. 3.2–3.3, being described by
equations (see (3.59), (3.60), (3.17))
XP C
D A.t /XC C XC A0 .t / C .t /XC C 1 .t /P .t /; XC .t0 / D X 0 ;
l .t / D A.t /0 l .t /; l .t0 / D l ;
and
Abstract This chapter begins with the theory of trajectory tubes which are
necessary elements of realistic mathematical models for controlled processes and
their evolutionary dynamics. We then deal with the evolution in time of state-
constrained forward and backward reachability tubes also known as “viability
tubes.” The backward tubes are then used to design feedback controls under state
constraints that may also appear in the form of obstacles to be avoided by system
trajectories.
In this chapter we study the evolution of trajectory tubes. These tubes are needed
for problems of dynamics and control, especially those that deal with incompletely
described systems. Such a theory of trajectory tubes was introduced in [158]
and is described here in appropriate form for the topics of this book. The theory
is then applied to state-constrained reachability (“reach”) tubes, also known as
“viability tubes.” We introduce the technical tools in set-valued dynamics that are
used for calculating such tubes. These backward tubes are then used to define state-
constrained closed-loop strategies for target control. Finally we study similar issues
for obstacle problems that generate a class of more complex state constraints: the
target-oriented trajectories should develop within a bounded control set while also
avoiding obstacles that lie on route to the target set. Throughout the investigations
we assume that the notions of continuity and measurability of single-valued and
multivalued maps are taken in the sense of [41, 238].
We start with a definition of trajectory tubes that are at the object of further
discussion.
Consider the nonlinear differential inclusion derived from Sect. 7.1 , (7.1), (7.2),
x.t / 2 Y .t /; t 2 T: (8.4)
so that
1
Recall that a trajectory x.t / of system (8.1) generated by given initial condition ft0 ; x 0 g is marked
by square brackets, as xŒt , in contrast with an unspecified trajectory x.t /:
8.1 The Theory of Trajectory Tubes: Set-Valued Evolution Equations 313
to be the integral funnel for (8.1), [224]. One may observe that under our
assumptions „ 2 comp Rn [8, 26].
The subset of X .; t0 ; X0 / that consists of all solutions to (8.1) viable on Œt0 ; £
is denoted by Xy .; £; t0 ; X 0 / and its s-cross-sections as Xy .s; £; t0 ; X 0 /, s 2 Œt0 ; £.
We also use the notation
As indicated earlier, it is not difficult to observe that sets Xy Σ are actually
the reachable sets at instant £ for the differential inclusion (8.1) with state
constraint (8.7).
314 8 Trajectory Tubes State-Constrained Feedback Control
We will usually work under one of the two following groups of hypotheses, unless
otherwise noted.
Denote the graph of map F .t; / as grapht F (t is fixed):
x Œt 2 int Y .t /; 8t 2 T:
2
If Y .t / has no interior points, being located in space Rm ; m < n, then Y .t / may be substituted
by its neighborhood Yr .t / D Y .t / C rB.0/ > 0; with some additional regularization.
8.1 The Theory of Trajectory Tubes: Set-Valued Evolution Equations 315
Observe that Assumptions 8.1.1 and 8.1.2 are overlapping, but in general neither
of them implies the other.
One important property of trajectory tubes and viable trajectory tubes is to
preserve some geometrical characteristics of sets X Œt , Xy Œt along the system
trajectories, as in the following assertions.
Lemma 8.1.1. Under Assumption 8.1.1 the cross-sections X Œt , Xy Œt are convex
and compact for all t 2 T (X Œt , that is, Xy Œt 2 conv Rn .
Lemma 8.1.2. Under Assumption 8.1.2 the cross-sections X Œt , Xy Œt are star-
shaped and compact for all t 2 T , that is, X Œt , Xy Œt 2 St.Rn /.
Loosely speaking, a differential inclusion with a convex grapht F generates
convex-valued tubes X Œt , Xy Œt , while a star-shaped grapht F generates tubes with
star-shaped cross-sections X Œt , Xy Œt . Convexity and the star-shape property are
therefore the two simplest basic geometrical invariants for cross-sections of the
trajectory tubes and viable trajectory tubes.
Having defined the notion of viable trajectory tube and observed that mapping
Xy Œt D Xy .t; t0 ; X 0 / defines a generalized dynamic system, we come to the
following natural question: does there exist some sort of evolution equation that
describes the tube Xy Œt as solutions to a related generalized dynamic system?
It should be emphasized here that the space K D fcomp Rn g of all compact
subsets of Rn to which the “states” Xy Œt belong is only a metric space with
a rather complicated nonlinear structure. In particular, there does not exist even
an appropriate universal definition for the difference of sets A; B 2 comp Rn .
Hence the evolution equation for Xy Œt is devised to avoid using such “geometric”
differences. It is mainly due to this reason that the construction of an infinitesimal
generator for set-valued transition maps generated by such generalized system
dynamics is cumbersome.
In this subsection we rely on an approach to the evolution of trajectory tubes
through funnel equations for set-valued functions in contrast with Hamiltonian
techniques where this is done, as we have seen in Sect. 7.1, through single-valued
functions described by PDEs of the HJB type.
We shall further require that one of the following assumptions concerning the
mapping Y ./ is fulfilled.
Assumption 8.1.3. The graph graph Y 2 conv RnC1 .
Assumption 8.1.4. For every l 2 Rn the support function f .l; t / D ¡.l j Y .t // is
differentiable in t and its derivative @f .l; t /=@t is continuous in fl; t g.
316 8 Trajectory Tubes State-Constrained Feedback Control
Theorem 8.1.1. Suppose Assumption 8.1.1 or 8.1.2 holds and map Y satisfies
either Assumption 8.1.3 or 8.1.4. Then the multifunction Xy Œt D Xy .t; t0 ; X 0 / is
a set-valued solution to the following evolution equation
[
lim ¢ 1 h Xy Œt C ¢; fx C ¢ F .t; x/ j x 2 Xy Œt g \ Y .t C ¢/ D 0;
¢!C0
(8.8)
t0 t t1 ; Xy Œt0 D X 0 :
Rigorous proofs of this theorem in terms of the Hausdorff distance (see Sect. 2.3)
and of the following uniqueness theorem are rather lengthy and are given in detail
in [158, Sects. 6, 7]. The uniqueness of solution to Eq. (8.8) is indicated in the next
theorem.
Denote Z Œt0 ; t1 to be the set of all multivalued functions Z Œt (Z W Œt0 ; t1 !
comp Rn ) such that
[
lim ¢ 1 h Z Œt C ¢; fx C ¢ F .t; x/ j x 2 Z Œt g \ Y .t C ¢/ D 0; (8.9)
¢!C0
Z Œt0 D X 0 ; t0 t t1 ;
Theorem 8.1.2. Suppose the assumptions of Theorem 8.1.1 hold. Then the mul-
tivalued function Xy Œ£ D Xy .£; t0 ; X 0 / is the unique solution to the funnel
equation (8.8) in the class Z Œt0 ; t1 of all the set-valued mappings Z ./ that satisfy
this equation uniformly in t 2 Œt0 ; t1 .
Linear Systems
lim ¢ 1 h.Xy Œt C¢; ..I C¢A.t //Xy Œt C¢B.t /Q .t //\ Y .t C¢// D 0 (8.11)
¢!C0
t0 t t1 ; X Œt0 D X 0 :
and the right-hand side F .t; x/ D A .t /x C P .t / depends bilinearly upon the state
vector x and the set-valued map A .
We assume that A ./ is a continuous mapping from Œt0 ; t1 into the set conv Rnn
of convex and compact subsets of the space Rnn of n n-matrices.
Equation (8.12) is an important model of an uncertain bilinear dynamic system
with set-membership description of the unknown matrices A.t / 2 A .t /, and inputs
h.t / 2 P .t / and with x 0 2 X 0 ; [158, Sect. 3].
It is not difficult to demonstrate that here Assumption 8.1.1 does not hold, so that
sets X Œt D X .t; t0 ; X 0 / (and also Xy Œt D Xy .t; t0 ; X 0 /) need not be convex.
Example 8.1.1. Indeed, consider a differential inclusion in R2 :
xP 1 2 Œ1; 1x2 ; xP 2 D 0; 0 t 1;
318 8 Trajectory Tubes State-Constrained Feedback Control
Then
X Œ1 D X .1; 0; X 0 / D X 1 [ X 2 ;
where
t0 t t1 ; X Œt0 D X 0 :
A Nonlinear Example
We indicate one more example, which is given by a set-valued function F .t; x/ with
a star-shaped grapht F .
Namely, let F .t; x/ be of the form
We now mention an approach similar to the described, but with the evolution
equation (8.8) written using the Hausdorff semidistance
rather than the Hausdorff distance (a metric) h.A; B/ D maxfhC .A; B/; hC .B; A/g.
Here is such an equation (see [174] and also [163]) for system (7.1), (7.2):
( ˇ )!
[ ˇ
1 ˇ
lim hC ZŒt C ¢; x C ¢.f .t; x; P .t //ˇ x 2 ZŒt \ Y .t / D 0; Z.t0 / D X 0 ;
¢!C0 ¢ ˇ
(8.15)
Remark 8.1.1. The solution ZŒt to this evolution equation is a multivalued function
with ZŒt0 D X 0 ; which satisfies (8.15) almost everywhere. As a rule, this solution
is not unique. However, we may single out a solution X Œt which is the inclusion-
maximal among all solutions ZŒt to (8.15). Namely, X Œt ZŒt , where ZŒt
is any solution that starts from Z.t0 / D X 0 : Note that Eq. (8.15) makes sense
for any piecewise continuous function y.t /. As mentioned above, when dealing
with information tubes, we presume these functions to be right-continuous. The
hC -techniques may be more adequate for dealing with discontinuous set-valued
functions Xy Œt D Xy .t; t0 ; X0 /.
The evolution equation (8.15) delivers a formal model for the generalized
dynamic system generated by mapping Xy Œt D Xy .t; t0 ; Xy0 /. However, if the sets
Xy Œt are convex, there exists an alternative presentation of this evolutionary system.
Namely, each of the sets Xy Œt could be described by its support function
In this section we restrict our attention to a system for which Assumption 8.1.1 is
satisfied. In this case the tube Xy .; £; t0 ; X 0 / will be a convex compact subset of
C n Œt0 ; t1 . With set Xy Œ£ D Xy .£; t0 ; Xy0 / 2 conv Rn for every £ 2 Œt0 ; t1 .
320 8 Trajectory Tubes State-Constrained Feedback Control
where
®.l; t C ¢/ D (8.18)
where
On the other hand, since the function ®.l; t / is convex and positively homogeneous
in l 2 Rn with t fixed, we also observe
3
Recall that a set A Rn is defined to be symmetric with respect to 0 if A D A .
322 8 Trajectory Tubes State-Constrained Feedback Control
Lemma 8.2.1. Suppose assumption (8.24) holds and sets X 0 , grapht F , Y .t / are
symmetric for all t 2 Œt0 ; t1 . Then set X Œt is also symmetric whatever be the instant
t 2 Œt0 ; t1 .
Under the assumptions of Lemma 8.2.1 the inequality (8.22) turns into an
equality if and only if the vectors l, q are collinear (l D ’q and ’ 0). Then (8.17)
turns into
so that with ¡.l j X Œt / < ¡.l j Y .t // we have ’ D 1 and in (8.16) the set Q.l; t / D
flg. Otherwise, q D ’l, ’ 2 Œ0; 1 and
The last result will later be used in Sect. 8.4 in describing feedback solution
strategies for the problem of closed-loop control under state constraints.
In this subsection the description viable trajectory tubes is reduced to the treatment
of trajectory tubes for a variety of specially designed new differential inclusions
8.2 Viability Tubes and Their Calculation 323
without state constraints. These new inclusions are designed depending upon certain
parameters and have a relatively simple structure. The overall solution is then
presented as an intersection of parallel solution tubes for the parametrized inclusions
over all the parameters.
We start with the general nonlinear case, namely, with differential inclusion (8.1)
and set-valued state constraint (8.4), repeated here as
with constraints
x.t / 2 Y .t /; t0 t £; (8.29)
supposing also that the basic conditions of Sect. 8.1 on maps F , Y and set X 0 are
satisfied.
In this section we do not require however that either of the Assump-
tions 8.1.1, 8.1.2 or 8.1.3, 8.1.4 would hold.
We further need to introduce the restriction FY .t; x/ of map F .t; x/ to a set
Y .t / (at time t ). This is given by
F .t; x/; if x 2 Y .t /I
FY .t; x/ D
;; if x 62 Y .t /:
The next property follows directly from the definition of viable trajectories.
Lemma 8.2.2. An absolutely continuous function xŒt defined on the interval Œt0 ; £
with x 0 2 X 0 is a viable trajectory to (8.28) for t 2 Œt0 ; £ if and only if the inclusion
P 2 FY .t; xŒt / is true for almost all t 2 Œt0 ; £.
xŒt
We will now represent FY .t; x/ as an intersection of certain multifunctions. The
first step to achieve that objective will be to prove the following auxiliary assertion.
Lemma 8.2.3. Suppose A is a bounded set, B a convex closed set, both in Rn . Then
\
A if 0 2 B;
fA C LB j L 2 Rnn g D
; if 0 62 B;
T
Hence, A D fACLB j L 2 Rnn g. If 0 62 B, then there exists a hyperplane in Rn
that strictly separates the origin from B. By the boundedness of A, for a sufficiently
large number œ > 0, we then have
.A C œB/ \ .A œB/ D ;:
whose solution tubes parametrized by L./ are Z.£; t0 ; X 0 ; L.// However, the main
point for a linear system is that the second inclusion of Theorem 8.2.4 actually turns
into an equality, producing the next result.
Theorem 8.2.5. The following equality is true for any £ 2 Œt0 ; t1 :
\
Xy Œ£ D fZ.£; t0 ; X 0 ; L.// j L./ 2 Rnn Œt0 ; £g: (8.32)
The techniques of Chaps. 2 and 3 with those of Sects. 7.1 and 7.2 allow to approach
the problem of control synthesis under state constraints (see [136]). Such issues
were treated later under the term “viability problems,” with some renaming of earlier
terminology [5]. Therefore we may also interpret our problem as that of terminal
control “under viability constraints,” meaning that solutions subject to the state
constraint are “viable solutions.” As in the case of systems without state constraints,
the synthesized solution strategy described here will be nonlinear.
326 8 Trajectory Tubes State-Constrained Feedback Control
x 0 D x.t0 / 2 X 0 ; (8.35)
where p.t /, P.t / are continuous and y.t /, Y .t / are absolutely continuous. The target
set is also ellipsoidal: M D E .m; M /
We may present constraints (8.34) and x 2 E .m; M / as
that start from a given position f£; xg, x D xŒ£, xŒ£ 2 W .£; t1 ; M /, £ 2
Œt0 ; t1 , would satisfy the inclusion (8.34), £ t t1 , with x.t1 / 2 M .
Here UY is an appropriate class of functions defined below, in (8.42). The given
problem is non-redundant, provided W Œ£ D W .£; t1 ; M / ¤ ;.
The value function for Problem 8.3.1 may be defined as follows
ˇ
ˇ
V s .£; x/ D min max maxf®.t; xŒt / j t 2 Œ£; t1 g; ®M .xŒt1 / ˇˇ xŒ£ D x :
u t
(8.37)
Then
W Œ£ D fx W V s .£; x/ 1g
will be the backward reach set relative to M under state constraint Y .t /, which
is the set of points x for which there exists some control u.t / that steers trajectory
xŒt D x.t; £; x/ to M with x.t / 2 Y .t /; t 2 Œ£; t1 .
8.3 Control Synthesis Under State Constraints: Viable Solutions 327
This relation is true for linear systems of type (8.33), with a differentiable V s .t; x/.
Otherwise, the property of differentiability for V s may be relaxed to directional
differentiability which in our case always holds.
The class of solution strategies. As in Chap. 2, we shall look for the solution in the
class UY of all set-valued functions with convex compact values that are measurable
in t and upper semicontinuous in x, ensuring thus the solvability and extendability
of solutions to (8.36) for any x 0 2 X 0 2 conv Rn . But the difference will be in that
now UY of Problem 8.3.1 will be
8
ˆ
ˆ Pw .t / D E .p.t /; P .t // \ fu W d®.t; x/=dt ju 0g;
ˆ
ˆ
< if ®.t; x/ D 1;
UY D fU .t; x/g; U .t; x/ D Pw .t / D E .p.t /; P .t //;
ˆ
ˆ
ˆ if ®.t; x/ < 1:
:̂
(8.39)
Assume
Note that under control u the total derivative along the trajectory of (8.33) is
t0 t t1 ; W Œt1 D M :
8.3 Control Synthesis Under State Constraints: Viable Solutions 329
The tube W Œt will now be used to introduce a scheme for finding the
synthesizing control strategy U .t; x/.
Given W Œt , t 2 Œ£; t1 , let
ˇ
dhC .x.t /; W Œt /=dt ˇ.8:41/ (8.44)
We now have to prove that the set-valued strategy U .t; x/ ¤ ; for all ft; xg and
that it solves the problem of control synthesis.
Calculating hC .x; W Œt /, we have
hC .x; W Œt / D hl 0 ; xi ¡.l 0 j W Œt /
@ @
¡.l 0 j W Œt / D ¡.l 0 j W Œs/; s D t:
@.t C 0/ @.s 0/
Here the right-hand part can be calculated similarly to [158, Sect. 9], as done for
@¡.l j X Œt /[email protected] C 0/, but now in backward time.
Since W Œt Y .t /, this gives
@¡.l 0 j W Œt /
@.t 0/
8
ˆ
< ¡.l 0 j B.t /P .t //; if ¡.l 0 j W Œt / < ¡.l 0 j Y .t //;
D (8.45)
:̂ minf¡.l 0 j B.t /P .t //; @=@t .¡.l 0 j Y .t ///g; if ¡.l 0 j W Œt / D ¡.l 0 j Y .t //:
Corollary 8.3.1. The tube of all synthesized trajectories from a given point
x£ 2 W Œ£ may be defined as the tube of all solutions to the differential inclusion
lim ¢ 1 h W Œt ¢; ..I ¢A.t //W Œt .B.t /P .t //¢/ \ Y .t ¢/ D 0;
¢!0
(8.49)
t0 t t1 ; W Œt1 D M :
Exercise 8.3.5. Prove the results of this section for A.t / 6D 0, using (8.34) and the
results of Sect. 2.6.
Exercise 8.3.6. Calculate W Σ using the techniques of convex analysis, similar to
Sect. 2.4.
8.3.3 Example
Consider the double integrator system of Sect. 7.2.4, taken on the time interval t 2
Œ0; 2; namely,
xP 1 D x2 ; xP 2 D u; (8.50)
with control juj k. Let m D .m1 ; 0/, ®M .x/ D hxm; M.xm/i; M D M 0 > 0
and ®.x2 / D jx2 j2 .
332 8 Trajectory Tubes State-Constrained Feedback Control
−0.5
−1
−1.5
−3 −2.5 −2 −1.5 −1 −0.5 0 0.5 1
x1
The objective is to calculate at time £ 2 Œ0; 2/ the backward reach set W Œ£ D
W .£; 2; M / from target set M D fx W hx m; M.x m/i1=2 1g under state
constraint jx2 Œt j 1; t 2 Œ£; 2.
Such calculations may be done using schemes of Sect. 7.2.4 or following funnel
equation (8.43).
Based on such a scheme Fig. 8.1 shows a particular version of this example
worked out by Ian Mitchell, with solution approximated numerically, using level
set methods of [221, 244]. The chosen parameters are
12 4
k D 1; m D .0; 0:8/0 ; M D :
4 12
This figure shows final reach set at t1 D 2; without constraint (dashed line) and
with convex constraints jx2 .t /j 1 (thick solid line). The constrained reach set is
much smaller than the intersection of the state constraints and the unconstrained
reach set.
To present system (8.50) in the form of (8.41), we apply transformation z.t / D
G.0; t /x.t /
1 t
G.0; t / D ;
0 1
8.4 Obstacle Problems 333
or
which is of the form (8.41). Hence the related funnel equation (8.43) will be
lim ¢ 1 h W Œt ¢; .W Œt ¢B.t /P .t // \ Y .t ¢/ D 0; (8.51)
¢!0
t0 t t1 ; W Œt1 D M :
This is the set of points x from which some controlled trajectory xŒt D x.t; £; x£ /,
starting at time £, reaches M at time t D t1 under state constraint §.t; xŒt / 1 or,
equivalently, xŒt 2 Z .t /; 8t 2 Œ£; ª.
Note that Problem 8.4.1 requires x.t / to stay outside the interior of convex
compact set Z .t /. So function V c .t; x/ and sets W Œt in general lack the property
of convexity.
The value function V c .t; x/ also satisfies a semigroup property similar to (8.43).
This gives, for any u.s/ 2 E .p.s/; P .s//; s 2 Œt; t C ¢; ¢ 0; the inequality
max maxf§.s; xŒs/ C 2 j s 2 Œt; t C ¢g V c .t; x/;
s
ˇ
ˇ
c ˇ
V .t C ¢; xŒt C ¢/ V .t; x/ ˇ xŒt D x 0;
c
we have
maxfH .t; xŒt ; Vtxc .t; xŒt /; u/; H .t; xŒt ; §tx .t; xŒt /; u/ 0;
The next obstacle problem combines the previous two state constraints, namely
those of Problems 8.3.1 and 8.4.1. We thus have
Problem 8.4.2. Given time interval Œ£; t1 and functions ®.t; x/; §.t; x/; ®M .x/,
find the set
Wo Œ£ D x 2 Rn W 9u./; 8t 2 Œ£; t1 ; ®.t; xŒt / 1;
§.t; xŒt / 1; ®M .xŒt1 / 1I xŒ£ D x :
This is the set of points x from which some controlled trajectory xŒt D x.t; £; x/,
starting at time £, reaches M at time t D t1 and also satisfies the state constraints
xŒt 2 Y .t /; xŒt 2 Z .t /; 8t 2 Œ£; t1 .
Set W Σ is known as the reach-evasion set [203, 204].
Finally, in Problem 8.4.2, the value function V o .t; x/ satisfies an analog of
Lemma 8.3.1. We have, for any u.s/ 2 P .s/; s 2 Œt; t C ¢, ¢ 0, the relations
max maxfmaxf®.s; xŒs/; §.s; xŒs/ C 2g j s 2 Œt; t C ¢g V o .t; x/;
s
V .t C ¢; xŒt C ¢/ V .t; x/ j xŒt D x 0;
o o
(8.55)
Case (a-o). Assuming V o .t; x/ > ®.t; x/; V o .t; x/ > §.t; x/ C 2; we have
the HJB equation
Case (b-o). Assuming V o .t; x/ D ®.t; x/, but V 0 .t; x/ > §.t; x/ C 2; we have
0 D H .t; x .0/ Œt ; Vtxo .t; x .0/ Œt /; u0 / H .t; x .0/ Œt ; ®tx .t; x .0/ Œt /; u0 / (8.57)
and lastly,
Case (c-o). Assuming V o .t; x/ D §.t; x/ C 2; but V 0 .t; x/ > ®.t; x/; we have
Here we consider an illustrative example for the obstacle problem worked out by I.
Mitchell. We introduce a complementary convex constraint of Problem 8.4.1, then
add a convex constraint coming to Problem 8.4.2. On a finite interval t 2 Œ0; t1 we
consider system
xP 1 D u1 ; xP 2 D u2 ; (8.60)
£ 2 Œ0; ª/ the backward reach set W Œ£ D W .£; t1 ; M / from the terminal set
M D fx W hx m; x mi1=2 1g under state constraints ®.x/ 2; §.xŒt /
1; t 2 Σ; t1 .
For our system (8.60) we can calculate the related value function V o .t; x/ as
n
V .£; x/ D min max maxfmaxf®.xŒt / 1;
u t
ˇ
®2 .xŒt / C 2g j t 2 Œ£; t1 g; ®M .t1 / ˇ xŒ£ D x : (8.61)
k D 1; m D .0; 1:3/0 :
k = 1, m = [ 0 1.3 ]
2.5
1.5
1
x2
0.5
−0.5
−1
Fig. 8.2 Growth of reach W Œt set subject to convex and nonconvex constraints (solid lines; final
time is shown thicker). The target set (dotted circle) and constraints (shaded regions) are shown as
well. The unconstrained reach set at the same final time is shown for comparison (dashed line)
338 8 Trajectory Tubes State-Constrained Feedback Control
The figure shows the final reach set at t1 D 1 with no constraint (dashed line)
and subjectqto the combination of convex constraint jx2 Œt j 2 and nonconvex
constraint x12 C x22 1 (thick solid line).
Once the backward reachability sets Wo Œt are described, one may proceed to the
solution of the control synthesis problem, to obtain the closed-loop control.
Problem 8.4.3. Given set Wo Σ; find control strategy ue .t; x/.Ue .t; x// that steers
system (8.33) from any position f£; xg; x 2 Wo .t / to a position ft1 ; x.t1 /g under
following additional constraints:
(i=1) ®.t; x.t // 1; t 2 Œ£; t1 I ®M .x.t1 / 1;
(i=2) §.t; x.t // 1; t 2 Œ£; t1 I ®M .x.t1 // 1;
(i=3) ®.t; x.t // 1; §.t; x.t // 1; t 2 Œ£; ªI ®M .x.ª// 1.
Each of the strategies ue .t; x/.Ue .t; x// may be sought for directly, from the
respective HJB-type equations or inequalities given above for calculating Vo .t; x/.
However we will apply the generalized “aiming” scheme, used in Sect. 8.1.2 and
introduced earlier, for systems without state constraints, in Sect. 2.6 (see [123,247]).
Namely, considering function Vo .t; x/ D d 2 .x; Wo .t //, introduce either single-
valued strategies
ue .t; x/ 2 Ue .t; x/ D
depending on the type of system and the definition of solutions used. Here œ is the
Lipschitz constant in ft; xg for function f .
Theorem 8.4.3. The closed-loop strategy Ue .t; x/ for Obstacle Problem 8.4.2 is
given by (8.62).
The problem is that the proposed strategy ue .t; x/ or Ue .t; x/ must satisfy in
some appropriate sense the equation
For a general nonlinear system of type (8.63) the solution may be defined as
a “constructive motion” introduced in [123], while in the case of linear systems
with convex compact constraints on the controls the solutions Ue .t; x/ may be
taken in the class of upper semi-continuous set-valued strategies with synthesized
system (8.64) treated as a differential inclusion [121, 123, 174].
Chapter 9
Guaranteed State Estimation
The problem of model and system state estimation through incomplete observa-
tions under non-probabilistic noise (the “theory of guaranteed estimation”) was
introduced and developed in [25, 45, 119, 135–137, 210, 241, 271]. Various versions
of the problem had been worked out using different tools and serving numerous
applications. The specifics of this problem are that the system operates under
unknown but bounded input disturbances while the available observations, corrupted
by similar set-membership noise, satisfy the measurement equation which yields an
on-line state constraint (in contrast with such constraints given in advance). Hence
the problem to be considered is actually one of finding the reachability set for a
system subjected to state constraints which arrive on-line.
The solution to the state estimation problem (an “observer”) is set-valued, with
output given by “information sets” of states consistent with the system dynamics
and the available measurements. The calculation of such sets—the observer outputs,
which evolve in time as an “information tube,” pose a challenging computational
problem, especially in the nonlinear case.
In this chapter we proceed with a further application of Hamiltonian techniques
indicated previously in Chap. 2 for systems without state constraints and in Chap. 7
for those with state constraints given in advance. This gives the solution to the
guaranteed (set-membership) state estimation problem by describing information
tubes using HJB equations. The idea applied here is that these information sets may
be expressed as level sets of related value functions, called “information states”,
which are the solutions to special types of the HJB equation. Such a turn yields
a deeper insight into the investigated problem and opens new routes for designing
computational algorithms. It applies to both linear and nonlinear systems.
Since calculating solutions to HJB equations is not simple, we proceed, as in
Sect. 7.1, with substituting the original HJB equations by variational inequalities
due to a related comparison principle. As in Sects. 5.1 and 7.1, the comparison
theorems generated now by the guaranteed state estimation (filtering) problem are
applicable to both smooth and non-smooth solutions to related HJB equations. It is
also shown that in case of linear systems with convex constraints on the disturbances
this approach may lead to effective external and internal approximations of the
information sets and tubes using ellipsoidal techniques.
Further on, a later detour is to discrete-time systems introducing the reader to a
collection of formulas useful for computation. This is followed by interrelations
between guaranteed state estimation theory of this book, as taken for linear
systems, and the well-known stochastic approaches to state estimation based on
the Kalman filtering theory. Finally mentioned is the important issue of dealing
with discontinuous measurements which extend previous results of this chapter to a
broader array of applied problems.
xP D f .t; x; v/ (9.1)
Here Q .t / is a multivalued function with values in the set compRq of all compacts
of space Rq , continuous in the Hausdorff metric, and set X 0 is compact. The pair
ft0 ; X0 g is the “initial position” of system (9.1).
We also assume that set F .t; x/ D f .t; x; Q .t // is convex and compact. Then,
due to the indicated properties, the set-valued function F .t; x/ will be Hausdorff-
continuous in all the variables.
The on-line information on vector x is given through observations which arrive
through one of the two following schemes.
9.1 Set-Membership State Estimation: The Problem. The Information Tube 343
where y.t / 2 Rr is the measurement and Ÿ.t /—the unknown but bounded
disturbance (“noise”), which is restricted to the inclusion
where y.£i / 2 Rr ; x.£i / 2 Rn , £i < £iC1 , £i 2 Œt0 ; ª, and noise Ÿ.£i / is
unknown but bounded, restricted by (9.4). The measurement times £i are taken
to be given.
Remark 9.1.1. An interesting situation is the case of communication-type con-
straints where the measurement signals y.£i / are assumed to arrive at random time
instants £i distributed, (say) as a Poisson process, with frequency œ > 0. Such case
lies beyond the scope of this book and is treated in [56].
Taking Scheme I, we assume that given are: the initial position ft0 ; X 0 g, functions
f .t; x; v/; g.t; x/; set-valued functions Q .t /; R .t /, and available on-line measure-
ments y£ .¢/ D y.£ C ¢/ (¢ 2 Œt0 £; 0).
Definition 9.1.1. The information set X .£; y£ ; / D X Œ£ of system (9.1)–(9.4) is
the collection of all its states xŒ£ D x.£; t0 ; x.t0 //; x.t0 / 2 X 0 , consistent with its
parameters and with observed measurements y£ .¢/.
Thus, the actual on-line position of the system may be taken as the pair f£; X Œ£g.
Problem 9.1.1. Calculate sets X Œt ; t 2 Œt0 ; ª; and derive an equation for their
evolution in time.
The set-valued function X Œt is called the information tube. Such tubes are estimates
of the system dynamics due to on-line measurements. They will be calculated
through two approaches—Hamiltonian techniques and set-valued calculus.
A similar definition and a problem are introduced within Scheme II under
available on-line measurements yŒt0 ; £ D fy.£i / W £i 2 Œt0 ; £g. Then one may
define a similar information set Xd .£; yŒt0 ; £/ D Xd Œ£. The information tube Xd Œt
may turn out to be discontinuous. This depends on the properties of the function
y.t /. Under Scheme I the discontinuities may be caused by discontinuous noise
while under Scheme II they arrive naturally, since measurements y.£i / are made at
isolated times.
Concentrating on Scheme I, note that the information set X Σ is a guaranteed
estimate of the unknown actual state x.£/ of system (9.1), so we always have
344 9 Guaranteed State Estimation
x.£/ 2 X Œ£; whatever be the unknown noise. Given starting position ft0 ; X 0 g and
measurement y.t /; t 2 Œt0 ; £, it makes sense to construct for system (9.1), (9.2) a
forward reachability tube X Œt D X .t I t0 ; X 0 /; t t0 ; which consists of all those
solutions to the differential inclusion xP 2 F .t; x/ that emanate from the set-valued
initial position ft0 ; X 0 g and develop in time under the on-line state constraint
Problem 9.1.1 indicated above is not the one of optimization. But as before we
shall solve it through alternative formulations of dynamic optimization. Following
Sect. 7.1.1 these are given in two versions.
Here V .t; x/ is the value function related to Problem 9.2.1. This function V .t; x/
is henceforward referred to as the information state of system (9.1)–(9.3). Its level
set or “cross-section” X Œ£ at level 0, namely (9.8), is the reachability set for (9.1)
under on-line state constraints (9.3), given y.t /. Relation
defines the boundary condition for V .t; x/. And as in Theorem 7.1.1 we have the
“semigroup” property
with boundary condition (9.9). Note that the term d 2 .y.t /g.t; x/; R .t // 6D 0 only
if y.t / g.t; x/ is outside of the constraint R .t /.
Here again V .t; x/ is assumed to be differentiable and if not, then (9.10) is
a formal symbolic notation for an equation whose solution should be considered
in a generalized “viscosity” sense (see [16, 17, 50, 80]) or equivalent “minimax”
sense (see [247]). For linear systems with convex value functions V .t; x/ the total
derivative d V =dt D Vt C hVx ; f .t; x; v/i may be substituted for a directional
derivative of V .t; x/ along the direction f1; f .t; x; v/g. This directional derivative
here exists for any direction f1; f g.
The value function V may be also defined under additional assumptions on
smoothness and convexity of the constraints. Given proper continuously differen-
tiable functions ®0 .t; x/; ®.t; Ÿ/, convex in x; Ÿ respectively, the initial set and the
constraint on the disturbance Ÿ in (9.3), may be presented in the form
®0 .t0 ; x/ D d 2 .x; X 0 /C1; ®.t; Ÿ.t // D d 2 .Ÿ.t /; R .t //C1; Ÿ.t / D y.t /g.t; x/;
Version B
Assume given is the realization y./ D yŒ of the observation y on the interval
Œt0 ; t .
Problem 9.2.2. Given are system (9.1) and starting position ft0 ; X 0 g; according
to (9.11). Find the value function
346 9 Guaranteed State Estimation
V .1/ .t; x/ D
D minf®0 .t0 ; xŒt0 / j xŒt D x; ®.s; y.s/ g.s; xŒs// 1; s 2 Œt0 ; t g: (9.12)
v
.1/ .1/
Denoting hVx ; f .t; x; v/i D H .t; x; Vx ; v/; consider the equation
.1/
Vt CmaxfH .t; x; Vx.1/ ; v/ j v 2 Q ™ .t; x/g D 0; V .1/ .t0 ; x/ D ®0 .t0 ; x/ (9.13)
v
where
Q .t /; ™.t; x/ < 1;
Q ™ .t; x/ D
Q .t / \ fv W d ™.t; x/=dt jv 0g; ™.t; x/ 1
Remark 9.2.1. The suggested Versions A and B lead to different HJB equations
and serve to solve the original Problem 9.1.1 under different assumptions on the
problem data. Namely, Version-B requires additional assumptions on smoothness
and convexity of the state constraints produced by the measurement equation. In
the general nonlinear case this difference may affect the smoothness properties of
respective solutions V .t; x/; V .1/ .t; x/. On the other hand, the Comparison Principle
of the next subsection does not depend on the smoothness properties of original HJB
equations for Versions A and B.
We will now obtain an external approximation of X Œt which in the linear case will
be exact. We proceed as follows.
Assumption 9.2.1. Given are functions H.t; x; p/, wC .t; x/ 2 C1 and .t / 2 L1 ,
which satisfy the inequalities
wC C
t C H.t; x; wx / .t /: (9.17)
Theorem 9.2.2. Suppose H.t; x; p/; wC .t; x/; .t / satisfy Assumption 9.2.1. Then
the following estimate for the information set X Œt holds:
X Œt XC Œt ; (9.18)
where
Z t
XC Œt D x W wC .t; x/ .s/ds C maxfwC .t0 ; x/ j x 2 X 0 g : (9.19)
t0
Version-B
.1/
Vt C œ™t C H1 .t; x; Vx.1/ ; ™x ; œ/ D 0; V .1/ .t0 ; x/ D ®0 .t0 ; x/ (9.20)
.a/ v 2 Q .t / D E .q.t /; Q.t //I .b/ Ÿ.t / 2 R .t / D E .0; R.t //; (9.24)
x.t0 / 2 X 0 D E .x 0 ; X 0 /;
To solve Problem 9.1.1 for the linear case we now approach the versions for
Problems 9.2.1 and 9.2.2 of dynamic optimization.
Denote
and
Version LA
Vt C hVx ; A.t /x C C.t /q.t /i C hVx ; C.t /Q.t /C 0 .t /Vx i1=2 ¦.t /.k 2 .t; x/ 1/ D 0;
(9.28)
with boundary condition
Version LB
d ™.t; x/=dt D ™t .t; x/ 2hR1 .t /.yŒt G.t /x/; G.t /.A.t /x C C.t /v/i
.1/
Vt C Ϫt C maxfhVx.1/ C Ϫx ; A.t /x C C.t /vi j v 2 E .q.t /; Q.t //g D 0;
v
or, with additional notations G 0 .t /R1 .t /y.t / D z.t /; G 0 .t /R1 .t /G.t / D G.t /,
into
.1/
Vt C œ™t C œhVx.1/ 2œ.z.t / G.t //x; A.t /x C C.t /q.t /i C H 0 .t; x; œ/ D 0;
(9.30)
where
H 0 .t; x; œ/ D maxfhVx.1/ 2œ.z.t / G.t /x/; C.t /vi j v 2 E .0; Q.t //g D
v
D hVx.1/ 2œ.z.t / G.t /x/; C.t /Q.t /C 0 .t /.Vx.1/ 2œ.z.t / G.t /x//i1=2 :
350 9 Guaranteed State Estimation
where
The next move is to approximate the exact solutions given in this subsection by
using the comparison principle. This is done through an application of ellipsoidal
approximations. The next scheme will rely on a deductive approach to the derivation
of ellipsoidal estimates (see also Sect. 5.1.2) in contrast with Chap. 7, where such
estimates were achieved due to inductive procedures introduced in Chap. 3.
9.3.1 Version AE
we use relation
hp; C.t /Q.t /C 0 .t /pi1=2 ” 2 .t / C .4” 2 .t //1 hp; C.t /Q.t /C 0 .t /pi; 8p 2 Rn ;
(9.32)
with equality reached for ” 2 .t / D .1=2/hp; C.t /Q.t /C 0 .t /pi1=2 . Denoting
H.t; x; p/ D hp; A.t /x C C.t /q.t /i C ” 2 .t / C .4” 2 /1 hp; C.t /Q.t /C 0 .t /pi
we then have
H.t; x; p/ H.t; x; p/
wt C H.t; x; wx / D
2hxP .t /; P .t /.x x .t //i C 2hP .t /.x x .t //; A.t /x C C.t /q.t /iC
¦.t /.hy.t / G.t /x; R1 .t /.y.t / G.t /x/i 1/ C ¦.t /d.h2 .t //=dt; (9.34)
with ¦.t / D 1.
Continuing further with ¦.t / D 1, we demand that the right-hand side in (9.34)
is treated as follows. We separately equate with zero the terms with multipliers of
second order in x x , then those of first order in the same variable, then free
terms, excluding ” 2 .t / C ¦.t /. We finally observe that the equated elements of the
right-hand side of (9.34) will be zeros if and only if the following equations are true:
Functions H.t; x; p/; w.t; x/ satisfy Assumption 9.2.1. Hence we are in the position
to apply Theorem 9.2.2. We proceed as follows.
Definition 9.3.1. A measurement y.t / is said to be informative on the interval
Œt; t C ¢; ¢ > 0; if constraint (9.25) is active throughout this interval.
Let X 0 Œs D X 0 .s; t; X Œt / stand for the reachability set at time s t; for
system (9.22) without state constraints, emanating at time t from X Œt and recall
that X Œs D X .s; t; X Œt / is the information set gotten under active measurement
output state constraints. Then Definition 9.3.1 means that for all s 2 Œt; t C ¢; it
should be X Œs X 0 Œs: Namely, taking into account the measurement y.s/ must
diminish set X 0 Œs towards X Œs.
Assumption 9.3.1. The available measurement y.s/ D yŒs is informative on any
interval t < s t C ¢ taken within Œt0 ; ª.
Remark 9.3.1. Since we are actually approximating the reach set from above,
we always deal with informative measurements which yield solutions that satisfy
Assumption 9.3.1.
This means we further assume ¦.t / 1 throughout Œt0 ; ª.
Integrating inequality (9.38) from t0 to t; along a trajectory xŒs D x.s; t0 ; x 0 /;
xŒt D x; s 2 Œt0 ; t that runs as xŒs 2 X Œs; under disturbance v.s/ 2 E .q.s/;
Q.s//; with measurement y.s/ D yŒs under disturbance Ÿ.s/ 2 E .0; R.s//; we get
Z t
w.t; x/ .” 2 .s/ C 1/ds w.t0 ; xŒt0 / D V .t0 ; xŒt0 /: (9.39)
t0
E .x .t /; P 1 .t /“.t //;
9.3.2 Version-BE
.1/
Vt C œ™t C œhVx 2œ.z.t / G.t /x/; A.t /x C C.t /qi C H 0 .t; x; œ/
Cœ™t CœhVx 2œ.z.t /G.t /x/; A.t /xCC.t /qiC” 2 .t /C.4” 2 .t //1 .H 0 .t; x; œ//2 :
.1/
Vt
(9.41)
Then we again take w.t; x/ in the form (9.33). Substituting w.t; x/ into the right-
hand side of the last relation we further follow the previous scheme of Version-A
now applying it to (9.41). This time we come to equations
These equations should be taken with any multiplier œ < 0; that satisfies (9.31).
The set of all such multipliers is further denoted as ƒ. Therefore now P .t / D
P”œ .t /. The further procedures are similar to Version AE, transforming P into
P”œ .t / and thus allowing to formulate in the next form.
1
Theorem 9.3.2. The set X Œt E .x .t /; “.t /P”œ .t //; where
E .x .t /; “.t /P”œ
1
.t // D fx W hx x .t /; “1 .t /P”œ .t /.x x .t //i 1g;
and functions P”œ .t /; xP .t / are described by Eqs. (9.42)–(9.44), with ” 2 .t / > 0 and
œ 2 ƒ.
xP 1 D x2 ;
(9.45)
xP 2 D x1 C v; jvj 1
Following (4.5) the reachability set without measurement state constraint will be
described as X Œt D X .t; 0; X 0 / D X0 Œt C Xv Œt where
cos.t s/ sin.t s/
X0 Œt D G.t; 0/X ; G.t; s/ D
0
sin.t s/ cos.t s/
9.4 Discrete Observations 355
and
Z t
Xv Œt D fx W x D G.t s/Q .s/dsg; Q .s/ D Q D fq W q D bv; jvj 1g; b 0 D f0; 1g:
0
The illustrations below are done for y.t / 0. Then Y .t / D R and the recurrent
relation for calculating Xy Œt is
4 4
3 3
2 2
1 1
0
x2
x2
0
−1 −1
−2 −2
−3 −3
−4 −4
−4 −3 −2 −1 0 1 2 3 4 −4 −3 −2 −1 0 1 2 3 4
x1 x1
4
3 8
2 6
1 4
0 2
x2
x2
−1 0
−2 −2
−4
−3
−6
−4
−4 −3 −2 −1 0 1 2 3 4 −6 −4 −2 0 2 4 6 8
x1 x1
Fig. 9.1 Reachability set X Œ without constraint Y in green. Reachability set Xy Œ with
constraint Y in orange. State constraint Y in red
X
N.t/
V .t; x/ D minfd 2 .x.t0 /; X 0 / C d 2 .y.£i / g.£i ; x.£i //; R .£i // j x.t / D xg:
v
iD1
(9.48)
Following (9.26) , (9.25) denote
The solution to the last problem is to be taken as a recurrent procedure with V .s; x/
being the solution to equations
and
This results in
Theorem 9.4.1. The information set for Problem 9.4.1 is
X
N.t/ Z t X
N.t/
d 2 .y.£i / g.£i ; x.£i //; R .£i // D d 2 .y.s/ g.s; x.s//; R .s// •.s £i /ds;
iD1 t0 iD1
we may derive for V .t; x/ a related HJB-type equation, similar to (9.10), presenting
it in symbolic form as an equation in distributions, [242],
X
N.t/
Vs CmaxfhVx ; f .s; x; v/ijv 2 Q .s/g •.s£i /d 2 .y.s/g.s; x.s//; R .s// D 0;
iD1
(9.52)
seeking V .t; x/, according to the theory of distributions, as a weak solution to this
equation.
In the last case the formal procedures for deriving tight external ellipsoids may
follow those of Version-A. However, here we indicate a slightly different route by
directly applying Theorem 9.4.1. Namely, for s 2 Œt0 ; £1 ; take
PP CPA.s/CA0 .s/P 2 .s/P . 2 .s//1 P C.s/Q.s/C 0 .s/P D 0; P .£i C0/ D Pd .£i /;
where
and the array of external ellipsoids EdC may be found through one of the schemes
given in Chap. 3, Sect. 3.2.
The treatment of ellipsoidal techniques for internal approximations of informa-
tion sets in linear systems may be achieved along the schemes of Sects. 3.7 and 3.12.
Discrete-Time Processes
Here f .k; x/ is a given map from N Rn into Rn (N is the set of integers), C.k/
are given matrices. The vector-valued inputs v.k/ are taken in Rp ; p n. These
vectors, together with x 0 , are unknown but bounded, with hard bounds
where observations y.k/ 2 Rm and matrices G.k/ are known while measurement
disturbances Ÿ.k/ are unknown but bounded, with
Symbol xŒk D x.k; vŒk0 ; k 1; x 0 / will be the end of trajectory x.j / of (9.53),
that develops through Œk0 ; k with x 0 and vŒk0 ; k 1 given.
As before the information set Xy Œs D Xy .s j t; k; F/ is the collection of all
points xŒs 2 Rn that arrive at time s (k s) due to such trajectories of system
(9.53) that emanate from state x.k/ D x ; evolve throughout interval Œk; t and also
produce the measured realization yŒk; s for some triplet fx ; v; Ÿg constrained by
inclusions
Xy .s j t; k; F/ D Xy .s j t; l; Xy .l j £; k; F//:
The given conditions ensure that the mapping Xy .t; k; F/ W comp Rn ! comp Rn
defines a generalized dynamic system similar to Xy .t; t0 ; X 0 / of Sect. 8.3.1, but
under discrete time.
In the general case sets Xy Œt defined here may turn out to be nonconvex and
even disconnected. However the next property is true.
Lemma 9.4.2. Let F be a closed, convex set in Rn , and f .k; x/ D A.k/x be a
linear map. Then sets Xy .s j t; k; F/ are closed and convex.
As before, the problem is to describe information sets Xy Œs and the related
information tube Xy Œk; s; s 2 Œk0 ; t —now a sequence of sets Xy Œs.
Remark 9.4.1. Note that apart from X Œs, X Œs one may introduce other sequences
which at each respective stage may be chosen as either (9.59), or (9.60).
We first describe sets X Œs D X.s; k0 C 1; X0 /, using notation
The last equation indicates that the innovation introduced by each new
measurement y.k/ arrives through an intersection of set X? .k; k 1; X Œk 1/—
the estimate at stage k before arrival of measurement y.k/ with Y.k/ derived
from y.k/.
z 2 f .X/ C C P; Gz y 2 Q; (9.63)
9.4 Discrete Observations 361
and
Z coZ Z
where
These relations allow to approach exact solutions due to the next assertion, given in
a form similar to Chaps. 3, 7, and 8.
Theorem 9.4.2. The exact relations for Z; Z are:
\
ZD fR.M; f .X// j M 2 Rmn g; (9.65)
\
Z D fR.M; co f .X// j M 2 Rmn g: (9.66)
Denoting
\
Z.s; F./; X0 / D fZ.s; k0 ; F./; M./; X0 / j M Œk; s 2 Rmn g; (9.74)
Though the problem considered further is formulated under hard bounds on the
system inputs and system trajectories, the calculation of related trajectory tubes may
be also achieved through quadratic approximations.
Let us return to the problem of finding viability tube Xy Œt D Xy .t; t0 ; X 0 / for
system
G.t /x 2 Y .t /; (9.77)
where set-valued maps Q .t /, Y .t / and set X 0 are the same as in Sect. 9.1.
The support function for convex compact set Xy Œt may be calculated directly, as
a linear-quadratic problem, similar to Sects. 1.6 and 2.2.This gives
where
Z ˇ
t ˇ
0
‰t .l; œ.// D ¡ S .t0 ; t /l S .t0 ; £/G .£/œ.£/d £ ˇˇ X0 C
0 0
t0
Z Z t ˇ
t ˇ
C ¡ S 0 .£; t /l S 0 .£; s/G 0 .s/œ.s/ds ˇ Q .£/ d £C
ˇ
t0 £
Z t
C ¡.œ.£/ j Y .£//d £:
t0
where measurable functions v.t /, w.t / and vector x 0 satisfy the inclusions
as in (9.2), (9.3).
Instead of handling (9.80) we shall now consider a “perturbed” system, which is
G.t /z D w .t / C Ÿ.t /;
Elements d./ D f—0 ; ˜./; Ÿ./g represent the unknown disturbances bounded
jointly by quadratic inequality
Z £ Z £
0
—0 M —0 C ˜0 .t /R.t /˜.t /dt C Ÿ0 .t /H.t /Ÿ.t /dt 2 ; (9.83)
t0 t0
where fM; R./; H./g 2 = and symbol = stands for the product space
= D Rnn
C RC Œt0 ; £ RC
nn mm
Œt0 ; £;
with RC rr
Œt0 ; £ denoting the class of all r r-matrix functions N./ 2 Rrr Œt0 ; £
whose values N.t / are symmetric and positive definite.
For every fixed k ./, ƒ D fM; R./; H./g and denote ZŒ£ D Z.£; k ; ƒ; /
to be the set of all states z.£/ of system (9.82) that are consistent with con-
straint (9.83). The support function of this set is
z.t0 / D x0 ; t0 s £;
where ƒ D fM; R./; H./g 2 = are fixed and †./ is the matrix solution to Riccati
equation
†.t0 / D M 1 ; t0 s £:
which is the union of centers z0 Σ over all triplets k./ D k ./ from (9.81).
Set Z0 .£; ƒ/ is convex and compact, being the reachability set for system (9.85)
under constraints (9.80), or in other words, of the differential inclusion
Y .t / D y.t / R .t /; z.t0 / 2 X 0 ; t0 t £;
The support function for Z0 .£; ƒ/, which gives a complete description of this set, is
now to be calculated as in Sect. 2.4. It is a closed convex, positively homogeneous
functional
where
by relations
.Dƒ l/.t / D G.t / S.t0 ; t /M 1 S 0 .t0 ; £/C
Z t
C S.s; t /R1 .s/S 0 .s; £/ds l; l 2 Rn ; t0 t £;
t0
Z £
.K1 œ.//.t / D K.t; s/œ.s/ds;
t0
Since here we deal with a linear system, the further objective is to emphasize, as
in Theorem 8.2.5 of Sect. 8.2.2, that inclusion (9.91) is actually an equality.
Lemma 9.5.2. Suppose the m n-matrix G.t / is of full rank: r.G.t // D m for any
t 2 Œt0 ; £. Then for every l 2 Rn the following equalities are true
Combining formula (9.78), Lemma 9.5.2 and taking into account (9.91) as an
equality, we get
Theorem 9.5.1. Let r.G.t // D m for every t 2 Œt0 ; £. Then
\
Xy Œ£ D fZ0 .£; ƒ/ j ƒ 2 =g: (9.93)
9.6 Information Tubes vs Stochastic Filtering Equations. Discontinuous. . . 367
The two approaches to the state estimation problem, namely the stochastic and the
set-membership filtering, may seem rather different. However it turns out that, apart
from differences, there are useful connections between the techniques of calculating
the results.
differential equations
The estimate z .t / thus depends on the triplets k ./ and ƒ DfM ; R ./;
H ./g2=.
9.6 Information Tubes vs Stochastic Filtering Equations. Discontinuous. . . 369
which, with a given realization q.t /, is the reachability set for Eq. (9.99).
Theorem 9.6.1. Assume equalities
M D M 1 ; R .t / R1 .t /; H .t / H 1 .t / (9.101)
Then sets Z0 .t; ƒ/ of Sect. 9.5 and Z .t; ƒ / of this section coincide, namely,
Corollary 9.6.1. Under the assumptions of Theorem 9.6.1, the following equality
is true
\
Xy Œ£ D fZ .£; ƒ / j ƒ 2 =g: (9.103)
The proof of Theorem 9.6.1 follows from the fact that under the assumptions
of the last theorem equation (9.99), with k ./ 2 fQ ./ Y ./ X 0 g, and
the system (9.79)–(9.81), have the same reachability sets under constraint (9.83).
Corollary 9.6.1 then follows from Theorem 9.5.1.
Remark 9.6.1. The last results describe a clear connection between solutions to
the linear-quadratic Gaussian filtering problem (the Kalman filter) and those for
deterministic guaranteed state estimation under unknown but bounded “noise”
which may satisfy not only soft integral quadratic bounds on the uncertain items
but also the non-quadratic instantaneous (hard) bounds on the unknown items.
A more detailed discussion of this approach may be found in [138].
One should recall that while the theorems discussed in Sect. 9.1 were proved
under Assumptions 8.1.1, 8.1.3, or 8.1.4, or 8.1.6, those of Sect. 9.2.1 contained
additional conditions on smoothness and convexity of the on-line continuous state
constraints, whereas those of Sects. 9.2.2 and 9.2.3, allowed Y .t / to be only
continuous in time t . A certain relaxation of this condition would be to demand that
y.t / would be piecewise continuous. We shall now recall a technique of singular
perturbations for differential inclusions which will eventually allow to relax the
requirements on Y ./, accepting state constraints, when Y .t / is only measurable
on T .
A nonlinear filtering version of these results could be pursued through a
combination of the reasoning of Sect. 8.1 and of [14].
370 9 Guaranteed State Estimation
with x.t0 / 2 X 0 ; t0 t £.
Here the nn- and mn-matrices A.t /, G.t / are taken measurable and bounded
almost everywhere on Œt0 ; t1 , with set-valued maps Q W Œt0 ; t1 ! conv Rn and
R W Œt0 ; t1 ! conv Rm measurable and bounded.
Further on we consider a system of type
t0 t £; x 0 D x.t0 / 2 X 0 ; r 0 D r.t0 / 2 R 0 ;
and
Here Rmm Œt0 ; £ is the class of all continuous invertible matrix functions L./ 2
Rmm Œt0 ; £.
Denote z D fx; rg 2 Rn Rm and zŒt D z.t I £; t0 ; z0 ; L/ to be the solution
to (9.105) that starts at point zŒt0 D z0 D fx 0 ; r 0 g. Symbol Z.I £; t0 ; Z 0 ; L/ will
now stand for the solution tube to system (9.105) with ZŒ£ D Z.£; t0 ; Z 0 ; L/ D
Z.£I £; t0 ; Z 0 ; L/ and Z 0 2 Rn Rm .
Let x W denote the projection of set W Rn Rm into space Rn of x-variables.
Then the next assertion is true
Theorem 9.6.2. Under assumption x Z 0 D X 0 the following formula is true for
any £ 2 Œt0 ; t1
\
Xy Œ£ D x fZ.£I t0 ; Z 0 ; L/ j L 2 Rmm Œt0 ; £ :
Set Xy .£/ Rn therefore arrives as the intersection over all functions L./ 2
R mm
Œt0 ; £ of the projections of sets ZŒ£ R.nCm/.nCm/ on the space Rn of
x-variables. The proof of the last theorem is given in detail in [158, Sect. 16], with
examples given in [174, Sect. 4.6] (see also [159, 168]).
It is clear that conditions of smoothness or continuity of Y .t / in time are not
required in such setting.
Chapter 10
Uncertain Systems: Output Feedback Control
Abstract This chapter finalizes some results of earlier chapters for systems that
operate under set-membership uncertainty. Its aim is to emphasize a successful
application of previously described techniques to such systems. The chapter thus
gives a concise presentation of solution techniques for the problem of output feed-
back control based on available measurements under set-membership uncertainty.
This chapter gives a concise description of controlled systems that operate under
uncertainty—incomplete information on the system dynamics and uncertainty in
the system and measurement inputs due to unknown but bounded disturbances.
Considered are systems with mostly hard bounds on the uncertain items. The issue
of uncertainty is illustrated on the problem of output feedback control (OFC) for
such systems. The aim of the chapter is to demonstrate that techniques of the present
book may be successfully used in solving problems of feedback control under
imperfect knowledge of its model. A detailed discussion of feedback control under
realistic (inaccurate or incomplete) information on the system model, its on-line
dynamics and accompanying on-line measurement process is a broad topic left for
separate publication.
The basic problem here is of closed-loop target control under realistic
information. As everywhere in this book, the forthcoming results range from
theoretical issues to computational schemes. But the specifics of the derived solution
schemes lie in the combination of both approaches discussed in the previous
chapters, namely, of Hamiltonian methods in the form of dynamic programming
and of techniques taken from set-valued analysis and minmax approaches. The
overall general solution for the considered problem then appears as a combination
of two parts, which deal firstly with a finite-dimensional problem of guaranteed
state estimation and secondly—with an infinite-dimensional problem of feedback
control under set-membership uncertainty. For the first problem new types of set-
valued observers were introduced. For the second problem, which is especially
difficult to formalize and solve, the achieved solution is reduced to one of
finite-dimensions which facilitates calculation. For systems with linear structure and
convex constraints the solution is more precise. Here the computational procedure
is based on new developments in ellipsoidal calculus, which proved to be effective
and allowed to design complemental software, producing complete solutions with
useful examples (see [133]). Calculations for the nonlinear case are achieved by
using specifics of the problem in combination with comparison principles that allow
to relax the original equations or variational inequalities of the Hamilton–Jacobi
type to simpler, finite-dimensional relations.
toolbox [132]). The procedure also allows one to apply polyhedral techniques,
[114, 115, 117]. For the nonlinear case calculations may be facilitated by using
the specifics of the problem and applying modifications of the earlier suggested
comparison principles that allow to relax the original equations or variational
inequalities of the Hamilton–Jacobi type to simpler relations (see Sects. 5.1 and 7.1
of this book and also [149]).1
The results of this chapter follow the lines of [189].
The problem of OFC under set-membership noise is formulated for the next control
system model and available information.
The System Model. Given is an n-dimensional differential equation
where the functions f1 .t; x; u/, f2 .t; x; v/ are continuous with respect to the triplets
of independent variables and such that their sum satisfies standard conditions of
uniqueness and extendability of solutions throughout the interval Œt0 ; ª, for any
initial condition x 0 2 Rn , any admissible, bounded, closed-loop control strategies
u D U .t; / 2 P .t /; (10.2)
and uncertain disturbances v.t / from appropriate classes indicated below. Here,
as before, P .t / 2 compRp is a set-valued function, with compact values,
continuous in the Hausdorff metric. And as before, the system trajectory is denoted
xŒt D x.t; t0 ; x 0 /.
The Measurement Equation and the Noise. The observations of vector x are due
to measurement equation
1
Effective results may be reached for some classes of problems, especially those on an infinite
horizon, by Lyapunov-type methods [198].
374 10 Uncertain Systems: Output Feedback Control
x 0 2 X 0 ; v 2 Q .t /; Ÿ 2 R .t /; (10.4)
Xy Œt D fx W V .t; x/ ’g; ’ 0
is the level set of V .t; x/.
The overall basic problem of OFC now reads as follows.
10.3 The Overall Solution Scheme: General Case 375
dependent on the state ft; Xy Œt g or ft; V .t; /g, which for any starting position
ft0 ; X 0 g; t0 < ª; would bring xŒª D x.ª; t0 ; x 0 / to a preassigned neighborhood
M” of the given target set M ; at given time t D ª, whatever be the uncertain items
—t .; t0 /; constrained by (10.4). Here M” D M C ” BM .0/; where BM .0/ D fx W
hx; M xi 1g; M D M 0 > 0.
Note that control strategy U is a functional of the information set (or information
state) that depends on measurement y.t; / and given constraints on the uncertain
items. The notion of information sets and information states, as well as the
description of their properties, are the subject of a separate theory of guaranteed
estimation treated in Chap. 9.
Depending on what we use—the information set or the information state, there
are two basic interconnected approaches, both considered in Chap. 9. The first one
is to use the information state V .t; x/; calculated as the value function for a
Dynamic Programming problem of minimaximizing a functional borrowed from
H1 theory and producing solutions in the form of control strategies, as functionals
u D U .t; V .t; // (see [18, 27, 103, 123]). The second is to use the informa-
tion set described through set-valued calculus in terms of differential inclusions,
involving further the notion of invariant sets and calculating set-valued strategies
u D U .t; Xy Œt / through the “aiming rule” and its analogies (see related work at
[6,122,136,137,153,204]). We will use both approaches within a unified framework.
Suppose that on the interval Œt0 ; £ the control u D uŒt and the observation y D yŒt
have been realized and are therefore known.2
Problem OFC will be to find
2
Here and in the sequel square brackets in u D uŒt ; y D yŒt mean that for the interval under
consideration these functions are known, otherwise we use round brackets.
376 10 Uncertain Systems: Output Feedback Control
Z ª
d .y.t / g.t; x.t //; R .t //dt C d .x.ª/; M /
2 2
£
where
and
Z £
d 2 .yŒt g.t; x.t //; R .t //dt j x.£/ D x; u D u£ Œ; t0 g
t0
Z ª
d .y.t / g.t; x.t //; R .t //dt C d .x.ª/; M /
2 2
£
Note that on the interval Œt0 ; £; where the realization of the measurement noise
is ŸŒt D yŒt g.t; x.t //; the triplet — D —£ .; t0 / is constrained by inclusions
x 0 2 X 0 ; v.t / 2 Q.t /; yŒt g.t; x.t / 2 R .t /; with observation yŒt and control
uŒt being known (such triplets are denoted as —Œ), while on the interval .£; ª; the
measurements y.t / are unknown, so that — D —ª .; £/ has to be bounded only by
Xy Œ£; Q .t /; R .t /; t £; to satisfy all the possible future realizations of y.t /.
Then, on the interval .£; ª; functional (10.6) has to be maximized over all such
—ª .; £/ and minimized over u 2 P .t / in the class of strategies described further.
One may now observe that the overall Problem OFC may be separated into
two, namely, Problem E on guaranteed estimation (calculation of the information
state V .t; x/ D V .t; x/), solved within the interval Œt0 ; £ and Problem C of
feedback control (finding strategy U .t; V .t; // as a functional of the information
state), solved within the interval .£; ª. Note that V .£; x/ depends on the available
measurement y£ Œ; t0 D y.t C ¢/; ¢ 2 Œt0 £; 0; and that its role is similar
10.4 Guaranteed State Estimation Revisited: Information Sets and Information. . . 377
to sufficient statistics in stochastic control. Also note that the control strategy
U .t; V .t; // has to be selected within a class of functionals that ensure existence
and extendability of solutions to Eq. (10.1) with u D U .t; V .t; //. Therefore of
importance will also be a third goal which is to achieve a proper mathematical
formalization for the obtained results, ensuring the existence of solutions to the basic
Eq. (10.1) under strategies U .t; V .t; // obtained in the space of information states.
In the previous paragraph the information state V .t; x/ may be substituted by
information set Xy Œt D Xy .t; / due to the following: with set-valued hard bounds
on fx 0 ; v; Ÿg being given, it turns out that
therefore the information set X Œt would be the level set for the information state
V .t; /. This property thus justifies the fact that the generalized state is selected,
as indicated above, either as a function (the information state, which is the value
function V .t; / for an appropriate problem of Dynamic Programming) or as a
set (the information set). The dynamics of the overall OFC problem may be thus
described by either PDEs of the HJBI type (which give the dynamics of V .t; /
and its level sets ) or, for example, through a related integral funnel equation—an
evolution equation with set-valued solution which gives the dynamics of Xy Œt (see
Chaps. 8, 9, and also [158]). We further rely on either of these two definitions. The
fact is that to achieve effective solutions with natural geometrical interpretations it
appears necessary to combine both approaches. This allows us to work out a unified
vision of the general problem and open new routes to a detailed solution of specific
Problem OFC, especially with hard bounds and working within computation in
finite-dimensional space. Conditions for the latter to suffice, as given later, also
allow us to avoid the common restriction that u; v should satisfy certain matching
conditions. It is thus possible to solve the linear case completely and to apply
new computation methods to systems with set-valued solutions, enabling effective
computer animation of solution examples. Recall that under feedback control the
original linear system becomes nonlinear. We now describe the solution schemes
for Problems E and C the subproblems of the overall Problem OFC.
This definition of information set Xy Σ is similar to Definition 9.1.1, but the
difference lies in the system Eq. (10.1), where also present is the realization uŒt
known within Œt0 ; £/.
Therefore, the generalized state of the system may be taken as ft; Xy Œt g.
An equivalent definition of such a state may be also served by the triple
f£; y£ .; t0 /; u£ ./g. Then the passage from y£ .; t0 / to Xy Œ£ (now with given
u£ ./, is achieved through set-valued observers described in Chap. 9, Sects. 9.2
and 9.6 (see also [137, 151, 158]).
To solve basic problems E and C one should start by describing the evolution
of information sets Xy Œ£ D Xy .£; / in time, which now is also influenced by the
selected control u. Here we first explain this process through differential inclusions,
along the lines of Sects. 8.2, 9.2 and 9.6. Denote
h.YD .t /; YD .t 0 // œy jt 00 t 0 j: (10.11)
F .t; x; uŒt / D f1 .t; x; uŒt / C f2 .t; x; Q .t //; f2 .t; x; Q / D [ff2 .t; x; v/ j v 2 Q .t /g:
Then the information set Xy Œ£ will be the cross-section (cut) at time £ of the solution
tube (the “information tube”) Xy Œt D Xy .t; t0 ; X 0 / for the differential inclusion
(DI)
3
For Y .t / generated by given yŒt we use notation Y D Y Œt with square brackets.
10.4 Guaranteed State Estimation Revisited: Information Sets and Information. . . 379
zP 2 A.t /x CB.t /uŒt CC.t /Q .t /CL.t /.yŒt CG.t /zC R .t //; z0 D z.t0 / 2 X 0 ;
(10.15)
and inclusion (10.14) into an equality
\
Xy Œ£ D Xy .£; t0 ; X 0 / D fZu .t; t0 ; X 0 ; L/ j L D L./ 2 Cnr
1 Œt0 ; £g:
(10.16)
The last formula will be used below to obtain more detailed solutions under system
linearity. DIs (10.13) and (10.15), taken under given u£ ./; are actually the related
set-valued observers.
Assumption 10.4.1 implies that measurement y.t / is continuous. This require-
ment may be relaxed to piecewise continuous measurements y.t / D yŒt by
applying another set-valued observer generated due to the relaxed set-valued funnel
equation of Sect. 8.1.5. For solution tube Xy Œ£; of system (10.12), under x 2 Y Œt
the set-valued function Y .t / may then be taken as Hausdorff upper semi-continuous.
T Xy Œt then satisfies the evolution equation of the
With uŒt ; yŒt known, the tube
funnel type, with X Œt0 D X 0 Y Œt0 :
[ \
1
lim ¢ hC X Œt C ¢; fx C ¢ F .t; x; uŒt / j x 2 X Œt Y Œt C 0g D 0:
¢!0C0
(10.17)
Passing to the Hamiltonian approach, the dynamics of information sets Xy Σ
may be now calculated by solving the alternative problem E0 , formulated in terms
of value functions.
380 10 Uncertain Systems: Output Feedback Control
Problem E0 . Given are the starting position ft0 ; X 0 g, the available measurements
y£ Œ; t0 , and the control realization uŒt ; t 2 Œt0 ; £/. One is to specify
Z £
V .£; x/ D maxfd .xŒt0 ; X /
2 0
d 2 .yŒt g.t; x/; R .t //dt j xŒ£ D x 2 Rn ;
v t0
(10.18)
v.t / 2 Q .t /; t 2 Œt0 ; £g;
V .t; x/ may be non-smooth, hence the solution to Eq. (10.20) under condi-
tion (10.21) may not be classical, but must be redefined in a generalized sense,
for example, as a “viscosity” or “minmax” solution [48, 50, 80, 247].
The existence of a generalized solution to Eq. (10.20), (10.21) is necessary and
sufficient for the existence of solution to Problem E0 . In this case the mapping
V .t; / D V .t; jV .t0 ; // satisfies the evolution equation of type
As indicated in the above, the information space may be defined in either the space
of information sets or the space of value functions—the information states. We first
begin with the space of information sets.
The feedback control problem in the space of information sets reads as follows.
Problem C. Given starting position f£; Xy Œ£g; target set M and number ” > 0;
specify a control strategy u D u.t; X /; t 2 Œt0 ; £ that steers system (10.20), (10.21)
under u D u.t; X / into a ”-neighborhood M” of given terminal target set M ;
whatever be the function
Problem Copt . Among values ” > 0; for which Problem C is solvable, find
smallest ”0 .4
With ” > 0 given, Problem C is solvable if and only if Xy Œ£ belongs to
W Œ£ 6D ;, that is the set of subsets X£ Rn for which this problem with starting
position f£; Xy Œ£ D X£ g, does have a solution. Further we indicate that W Œ£ is
understood to be weakly invariant relative to target set M” .
Similarly, but for the information space of functions, we have the next problem.5
Problem Cv . Find value function
Z ª
V .£; V .£; // D min max max V .£; x/ d 2 .y.t / g.t; x.t //; R .t //dt C
u y x;v £
ˇ
ˇ
C d .xŒª; M” /ˇˇu 2 U ; y./ 2 Yª .; £/; v./ 2 Qª .; £/; x 2 X Œ£ ;
2
(10.23)
4
Note that if for ” D 0 Problem C is not solvable, then ”0 > 0:
5
Here and in (10.6) we use equivalent notations V .£; V .£; // and V .£; X Œ£/; since X Œ£ is the
zero-level set for V .£; /.
6
See [18] and [123] where this is done in finite-dimensional space.
10.5 Feedback Control in the Information Space 383
where d V .£; V .£; //=d £ is the total derivative of functional V .t; V .t; // due to
evolution equation (10.20), (10.21), or (10.22), which is the same. Equation (10.22)
may have a smooth solution, when the total derivative exists in the strong or weak
sense and the equation holds everywhere. Otherwise it has to be dealt with in terms
of generalized solutions, as mentioned above.
The solution strategy should then be formally determined through minimization
over u in (10.25), as u0 D u0 .t; V .t; // P .t /. Substituting this into (10.20),
instead of uŒt with y.t / instead of yŒt ; we have
˚
ˇ
Vt C max Vx ; f1 .t; x; u0 .t; V .t; /// C f2 .t; x; v/ ˇ v 2 Q .t /
v
The last relation may be treated as an evolution equation in the metric space
KV of functions fV .t; /g. The “trajectories” V Œt D V .t; /, issued at time t0
from V .t0 ; / D d 2 .x; X 0 /, may then be interpreted as “constructive motions”
V .t; I u0 .t; //; that arrive as a result of limit transition from infinite-dimensional
analogies of Euler broken lines, constructed under all possible partitions of the
interval Œt0 ; ª, by selecting piecewise-constant realizations uŒt 2 P .t / (see [123],
p. 11). Another option could be to treat fV .t; /g as a generalized (viscosity) solution
in infinite-dimensional space KV (see [98]).
Being based on dynamic programming, the overall Problem OFC allows to
consider linked solutions of problems E and C, where E is finite-dimensional, while
C is infinite-dimensional. In the linear case, as we shall see, these solutions are even
independent and may be separated. Such a scheme produces a good insight into the
problem and its complete solution. But it is difficult to calculate. However, once we
need to solve concrete problems, our interest is in computation schemes feasible for
applications and such that would allow procedures that do not go beyond finite-
dimensional dynamics with computational burden not greatly above the one for
problems with complete information.
As indicated above, the solution to Problem OFC is given through two types of
HJB, HJBI equations (see (10.20), (10.25))—one for state estimation and one for
feedback control. However, what we actually need here are not the value functions,
but only their level sets. Hence, instead of directly solving the infinite-dimensional
equation (10.25), we will use the notion of invariant sets (or backward reachability
sets) relative to target set M” . Continuing the previous subsection we deal with the
space of information sets. But to calculate them we shall use relations derived in the
space of information states. We follow the scheme below.
Formally, the set WV Œ£ D fV .£; / W V .£; V .£; // 0g that we need is the
(weakly) invariant set defined as follows.
384 10 Uncertain Systems: Output Feedback Control
Definition 10.5.1. (a) Set WV Œ£; weakly invariant relative to target set M” , is the
union of value functions V .£; / (in space KV ) for each of which Problem C is
solvable from initial position f£; Xy Œ£g, with Xy Œ£ D fx W V .£; x/ 0g.
Multi-valued function WV Œt ; t 2 Œ£; ª is the solvability tube in KV (an
infinite-dimensional analogy of the “Krasovski bridge”) for Problem C.
(b) Set W Œ£ weakly invariant relative to target set M” , is the union ( in space KX )
of compact sets X D X Σ for each of which Problem C is solvable from initial
position f£; X Œ£g.
Set-valued function W Œt ; t 2 Œ£; ª is the solvability tube in KX (another
infinite-dimensional analogy of “Krasovski bridge,” equivalent to case (a)).
Combining alternative descriptions in terms of value functions with those given
through set-valued calculus, we now pass from space KV to KX ; applying to
Problem C an infinite-dimensional modification of the rule of extremal aiming
towards weakly invariant sets W Σ (see finite-dimensional version in [123]). With
on-line position f£; X Œ£g given, we now have to calculate the Hausdorff semi-
distance HC .Xy Σ; W Σ/ between the actual state Xy Σ and the solvability set
W Σ, taken in the metric of space KX ; and arriving at
where the first (external) infimum is to be taken over all X£ 2 W Œ£. (This Hausdorff
semi-distance is similar to hC .X 0 ; X 00 / in Rn which was defined earlier.)
Then the solution to Problem C would be
Vh .£; Xy Œ£/ D hC .Xy Œ£; WŒ£/ D minf© > 0 W Xy Œ£ WŒ£ C ©B .0/g;
where
ˇ
[ [ ˇ
WŒ£ D ˇ
fx W x 2 X£ g ˇ X£ 2 W Œ£
The previous results described solution schemes that ensure inclusion X Œª M” ;
whatever be the uncertain items —./. Here M” is the guaranteed neighborhood of
the target set. The final problem is now to minimize this neighborhood M” under
uncertainty and output feedback. Then we need to solve the following subproblem.
We shall now discuss how far can the OFC problem be solved for linear systems
of type
As indicated in the previous section, we must find the feedback control u0 .£; Xy Œ£/,
having calculated Xy Σ and WΣ; where in the linear case both sets are convex
and compact. These sets may be expressed explicitly, through duality methods of
convex analysis, as the level sets of related value functions similarly to formulas of
Chap. 2, Sect. 2.4 and Chap. 3. And it would also allow us to calculate derivatives
of such value functions. Here, however, we indicate another scheme for such
calculations by using related evolution equations of the funnel type introduced in
10.6 More Detailed Solution: Linear Systems 387
Chap. 2, Sect. 2.3.1 and Chap. 8, Sect. 8.1.3. This move brings us, using notation
hC .X£ ; WŒ£/ D minf© > 0 W X£ 2 WŒ£ C ©B .0/g; to problem
Vh .£; Xy Œ£/ D maxf¡.l j G.ª; £/Xy Œ£/ ¡.l j G.ª; £/WŒ£/ j < l; l > 1g:
(10.31)
Then, following (10.28) for VH D Vh , we find the overall control solution u0 .
Hence, in order to calculate u0 we first need to calculate the total derivative of
Vh .£; Xy Œ£/ and ensure that it is non-positive under control u0 . Related evolution
equations for Xy Œt ; WŒt may be written as
the unique inclusion maximal solution, for the second we use the ordinary solution,
which is always unique.
Denote the unique maximizer in (10.31) as l 0 . We have
d Vh .t; Xy Œt /=dt D d¡.l 0 j G.ª; t /Xy Œt /=dt d¡.l 0 j G.ª; t /WŒt /=dt:
(10.34)
Calculating along the lines of Sect. 8.3, we get
d¡.l 0 j G.ª; t /Xy Œt /=dt hl 0 ; G.ª; t /B.t /ui C ¡.l 0 jG.ª; t /C.t /Q .t //;
d¡.l 0 j G.ª; t /WŒt /=dt D ¡.l 0 j G.ª; t /B.t /P .t // C ¡.l 0 j G.ª; t /C.t /Q .t //:
d Vh .t; Xy Œt /=dt hl 0 ; G.ª; t /B.t /ui C ¡.l 0 j G.ª; t /B.t /P .t //: (10.35)
Finally, selecting
and integrating d Vh .t; Xy Œt /=dt juDu0 from t0 to ª, we ensure the next condition.
Lemma 10.6.1. The control strategy (10.36) ensures the inequality
The next question is how to calculate the solutions numerically. Here we first
recall that Xy Σ; WΣ may be indicated in explicit form. Then, using this
knowledge, we suggest a computational approach based on approximating convex
sets by parametrized families of ellipsoids that evolve in time, following the
approximated sets.
xP L D A.t /xL C B.t /uŒt C C.t /q.t / C L.t /.yŒt G.t /xL /; xL .t0 / D x 0 ;
(10.37)
XP L D .A.t / L.t /G.t //XL C XL .A.t / L.t /G.t //0 C ..t / C ¦.t //XL C
The parametrizing functions ¨.t / D f.t / > 0; ¦.t / > 0; L.t /g may be taken
piecewise continuous.
Theorem 10.6.2. The next relation is true
\
Xy Œ£ D fEL Œ£ j ¨./g:
Among these parametrizers, one may select some optimal or some tight ones when
one could have, for some direction l 2 Rn ; a related triplet ¨0 ./, that ensures the
equality ¡.l j Xy Œ£/ D ¡.l j fEL Œ£ j ¨0 g/.
For WΣ we may also apply ellipsoidal approximations of the types indicated
in Chap. 5 (see also paper [149]), or use discretized versions for reachability under
uncertainty (see [133]). Here available are internal and external ellipsoidal bounds
10.6 More Detailed Solution: Linear Systems 389
that depend on internal and external parametrizing functions .t /; S.t / and tuning
parameter r.t / > 0. The equations for w .t /; WC .t /; W .t / are of the type (B.t / D
.B.t /P .t /B 0 .t //1=2 ; C.t / D .C.t /Q.t /C 0 .t //1=2 /:
With .t / > 0; r.t / > 0 and orthogonal matrices S.t /S 0 .t / D I; we have
E .w .t /; WC .t // D WC0 Œt
WŒt
W0 Œt D E .w .t /; W .t //;
where
whatever be the functions .t / and the orthogonal matrices S.t /. Following [180],
we have the next proposition.
Theorem 10.6.3. The next property is true
[
WŒ£ D fE .w .£/; W .£// j ./; S./g:
The parameter r.t / in (10.35) may be taken as r.t / D hle ; W Œt le i1=2 .
Suppose le is a unique support vector for Xy Œt ; WŒt . Then, along the lines of
Chap. 3, Sects. 3.2 and 3.7, one may describe the values of e .t /; ¦e .t /; Le .t /
for (10.33) and .t /; S.t /; r.t / for (10.35) that ensure the equalities
e .t/ D hle ; C.t/le i1=2 hle ; XL Œtle i1=2 ; ¦.t/ D hle ; L0e R.t/Le le i1=2 hle ; XL Œtle i1=2 ;
(10.42)
hle ; W .t /S.t /B.t /P 1=2 .t /le i1=2 D hle ; B.t /le i1=2 hle ; W Œt le i1=2 ; (10.43)
d¡.l 0 j EL .t //=dt D .hl 0 ; L0 G.t /XC .t /G 0 .t /L0 l 0 i1=2 Chl 0 ; L00 G.t /.x x /i/
Chl 0 ; L0 Ÿ.t /i hl 0 ; L00 R.t /L0 l 0 i1=2 C hl 0 ; C.t /vŒt i C hl 0 ; B.t /uŒt i: (10.45)
Similarly, we get
d¡.l 0 jES .t //=dt D hl 0 ; B.t /p.t / C C.t /q.t /i C hl 0 ; C.t /l 0 i1=2 hl 0 ; B.t /l 0 i:
(10.46)
Here x; Ÿ are those that generated the realized measurement y.t / D G.t /xCŸ.t /.
Substituting this in (10.44), we find the total derivative
d VhE .£; EL Œ£/=dt D hl 0 ; L0 G.t/XC .t/G 0 .t/L00 G.t/l 0 i1=2 Chl 0 ; L00 G.t/.xx /iC
Chl 0 ; B.t /.u p.t //i C hl 0 ; B.t /l 0 i1=2 C hl 0 ; L0 Ÿ.t /i hl 0 ; L00 R.t /L0l i1=2 C
Selecting
In order to simplify calculations for linear systems we shall treat the given Problem
OFC in another coordinate system. Following Sect. 1.1 and taking G.t; £/ to be the
fundamental transition matrix
@G.t; £/
D A.t /G.t; £/; G.£; £/ D I;
@t
and
with
where
Z t
x C w D x; z.t / D y.t / H.t / B.s/u Œsds:
t0
These examples illustrate the solution of Problem OFC under various types of
unknown disturbances. The system equation is
xP 1 D x2 C ©1 v
xP 2 D u C ©2 v
y.t / D x1 .t / C Ÿ.t /;
in Fig. 10.2 by
8
< ™1 ; with j™1 j ; ™1
N .0; 1/;
v.t / D ; with ™1 < ;
:
; with ™1 > ;
where N .0; 1/ stands for the normal distribution with zero mean and unit
variance,
in Fig. 10.3 by v.t / 0; Ÿ.t / 0 which means the disturbances are zero-valued,
but we do not know that. This is the worst-case disturbance.
Concluding Remarks
This chapter describes theoretical tools and computation approaches for the problem
of optimizing OFC under set-membership bounds on controls and uncertain items. It
indicates solution routes, which allow one to cope with uncertainty and incomplete
measurements within finite-dimensional techniques, with effective computational
schemes given for linear systems without imposing matching conditions for the hard
bounds on controls and input disturbances. It also demonstrates the applicability of
tools presented in this book to significant problems of control. The chapter is also
an invitation to a broad research on challenging problems in feedback control under
realistic information with increasing number of new motivations.
394 10 Uncertain Systems: Output Feedback Control
Abstract This chapter deals with a specific class of hybrid systems which combine
controlled continuous dynamics through switching from one available motion
to another due to discrete-time logical commands. Solutions to the reachability
problem and their verification are indicated, followed by computational schemes,
The application of impulse controls to the switching process is described. Examples
of various difficulty are worked out. The chapter is to demonstrate applicability of
methods of this book to hybrid systems.
The main part of this chapter deals with a specific class of hybrid systems
(see [39, 91, 204, 228, 258]). Their performance is due to an array of standard
systems with acting motion generated by only one of them and with instantaneous
switchings from one to another. The switchings are governed by a discrete time
process that generates logically based commands when crossing some prespecified
spatial domains (“the guards”). The reachability problem for such processes leads
to branching trajectory tubes. Their description is complemented by verification
problems whose solution, loosely speaking, should show which branch intersects
with a given target zone or misses. Hence the chapter starts with description of the
verification problem and methods of its solution, then passes to described target
controlled hybrid dynamics and its verification. Ellipsoidal computation schemes
for problems of reachability and verification in hybrid processes for isolated linear
systems are indicated. Finally discussed is the application of impulse controls to
such dynamics. Several examples of various difficulty are demonstrated. The aim
of this chapter is to demonstrate the applicability of methods given in this book to
hybrid processes.
The problems of reachability and solvability are closely related to that of verification
for the achieved solutions. The aim of verification is to check whether suggested
algorithms do solve the intended problem.
Problem 11.1.1. Given time ª, target set M 2 compRn and set X 0 D X Œt0 , verify
which of the following relations is true:
(i) X .ª; t0 ; X 0 / M (all the reachable points are in M ),
(ii) X .ª; t0 ; X 0 / \ M 6D ; (some of the reachable points are in M ),
(iii) X .ª; t0 ; X 0 / \ M D ; (set M is not reachable from ft0 ; X 0 g at time ª).
Since X .ª; t0 ; X 0 / is the level set of the value function at time t D ª (see Chap. 2,
Sect. 2.3):
V .t; x/ D V .t; x j t0 ; X 0 / D min fd 2 .x.t0 /; X 0 / j x.t / D xg;
u./;x.t0 /
V .t0 ; x/ D d 2 .x; X 0 /;
we may check conditions (i)–(iii) using the following relations:
Theorem 11.1.1. Following conditions (i)–(iii) of Problem 11.1.1 are, respectively,
equivalent to the next relations, with X .ª; t0 ; X 0 / D fx W V .ª; x/ 0g W
Corollary 11.1.1. If M is the level set for a certain convex function ¥, such that
M D fx W ¥.x/ 0g;
then conditions (11.1)–(11.3) are, respectively, equivalent to
maxf¥.x/jV .ª; x/ 0g 0;
x
minf¥.x/jV .ª; x/ 0g 0;
x
Remark 11.1.1. Note that if in the above and in the sequel we write minª over
domain ª t0 , then this presumes that the minimum should be attained at some
finite time ª .
Consider system (3.1) under ellipsoidal constraints (3.4), (3.6), (3.7) and ellipsoidal
target set M D E .m; M /. Dealing with formulation of the verification Prob-
lem 11.1.1 (i)–(iii) we shall solve it through ellipsoidal methods, further naming
it then as a new Problem 11.1-E (i)–(iii).
Recall that due to (3.23) we have
\
X λ D fE .x ? λ; XCl λ/j l W hl; li D 1g: (11.7)
where E Œl; ª D E .x ? Œª; Xl Œª/ are tight internal approximations of X Œª along
direction l.
Beginning with Problem 11.1.1-E(i) we have to check the inclusion
or
T
For Problem 11.1.1-E(ii) we T have to check condition X λ E .m; M / 6D ;.
This will be ensured if E Œl; ª E .m; M / 6D ; for some tight internal approx-
imation E Œl; ª D E .x Œª; Xl .ª// of X Œª. The last requirement will be ensured
iff for some such E Œl; ª the inequality
Finally, for Problem 11.1.1-E(iii) we have to check that X λ \ E .m; M / D ;. This
will be ensured if EC Œl; ª \ E .m; M / D ; for some tight external approximation
EC Œl; ª of X Œª; or in other terms, if the distance d.EC Œl; ª; E .m; M // > 0 for
some such EC Œl; ª. The latter holds iff
VC .ª; x; l/ D hx x Œª; XC
l
λ.x x λ/i:
Then
Here both functions VC .ª; x; l/; V .ª; x; l/ may serve as quadratic approximations
of value function V .ª; x/ that defines X Œª D X .ª; t0 ; X 0 / (see Chap. 2, Sect. 2.3).
Namely,
O x D xO W Vm .x/
.i i / f9 l D l; O V .ª; x; O
O l/g;
O 1g > 1g:
.i i i / f9 l D lO W minfVm .x/ j VC .ª; x; l/
The previous chapters of this book deal with various types of problems on control for
systems whose mathematical models are described by standard ordinary differential
equations. However, recent applications may require treating such problems in a
more complicated setting, under complex dynamics. In the present chapter such
systems are of the hybrid type. The notion of hybrid system has various definitions
(see [36,39,204,229,258,259]). However, the main idea is that the system is defined
by an array of standard systems such that the active motion is due to only one of
them, with instantaneous switching from one to another. The process of switching
is usually logically controlled in discrete times, in such way that a possible switch
may occur only when passing through some spatial domains (the “guards”), or may
not occur. The performance range of such systems is then obviously broader than
that of standard systems. Hence, the overall controlled motions will develop in
time as those generated by alternating isolated continuous motions due to systems
whose sequence is controlled through logically based discrete commands, while
their individual contribution is designed by its own controllers. The switching from
one system to another may be also accompanied by an instantaneous change of some
phase coordinates.
Despite the fairly complicated overall dynamics, efficient computation of control
solutions for hybrid processes appears to be also available through ellipsoidal-
valued approximations for systems composed of individual participants of the types
described above. The discussion given below is devoted to linear controlled systems,
though the overall schemes may be also applied under individual nonlinearities.
Hj D fx W hc .j / ; xi ”j D 0g; c .j / 2 Rn ; ”j 2 R ; j D 1; : : : ; k:
1
The part with memory may be important for making the decision—“to switch” or “not to switch,”
for example, if the number of switchings is restricted.
11.2 Hybrid Dynamics and Control 403
At each new crossing a new term is added to this sequence. Thus the following
general rules should be observed:
(i) that crossing each hyperplane Hj results either in a reset to subsystem with
number j or in no reset at all,
(ii) that the crossing takes place in direction of support vectors c .j / , and at points
of crossing we have
(iii) that the state space variable after › crossings is ft; x; i Œ1s1 ; : : : ; i Œ›s› g where
each “boolean” index “s” is either C or .
(iv) that at a crossing with hyperplane Hj the sequence of type fi Œ1s ; : : : ; i Œ›C g,
describing the “discrete event” part of the state is complemented by a new term,
which is either i Œ› C 1C if there is a switching to system, or i Œ› C 1 , if there
is no switching.
Such a notation allows to trace back the array of subsystems used earlier from any
current position ft; xg. Thus, if the state is ft; xI 1C ; i Œ1 ; : : : ; i Œ› g with s D
for all i Œ1; : : : ; i Œ›, then the trajectory did not switch at all throughout any of the ›
crossings, having followed one and the same subsystem with i D 1 throughout the
whole process. Note that at each state ft; x; Œ1C ; i Œ1s ; : : : ; i Œ›s g the system is to
follow the subsystem whose number coincides with that of the last term with index
s D C.
We will be interested in the reachability problem for such systems.
Remark 11.2.1. The system under consideration is one of the possible types of
hybrid systems. It differs from the so-called switching systems in that the time
instants for crossing are not given, but are located during the spatial course of the
trajectory as intervals of crossing some specified domains (the “guards”) where it is
also possible to reset the phase coordinates. The suggested scheme also allows to be
propagated to a broad variety of options with different information requirements.
Remark 11.2.2. In this chapter the guards are taken as hyperplanes and the time
of each crossing is unique for every trajectory. An example of propagating the
suggested scheme is when the guards are taken as a domain bounded by two parallel
hyperplanes.The switching is then assumed possible at any time within this domain.
This situation with an example is indicated at the end of the present section.
The problems considered in this chapter deal with reachability under piecewise—
open-loop controls with possible resets of controlled systems at given guards,
taken as hyperplanes, so that between these zones the control is open-loop. For
further considerations the restriction on starting set X 0 is of ellipsoidal type:
X 0 D E .x 0 ; X 0 /.
404 11 Verification: Hybrid Systems
where G .i/ .t; s/ is the transition function for the i -th subsystem.
(b) To be precise, suppose that before reaching Hj we have
maxfhc .j / ; xi j x 2 X .1/ Œt g D ¡C
j .t / < ”j :
T
The first instant of time when X .1/ Œt Hj 6D ; is £0j . It is found as the
smallest root of the equation
”j ¡C
j .t / D 0; t t0 :
minfhc .j / ; xi j x 2 X .1/ Œt g D ¡
j .t /;
T
we may also observe that condition X .1/ Œt Hj 6D ; will hold as long as
C
¡
j .t / ”j ¡j .t /;
11.2 Hybrid Dynamics and Control 405
and the point of departure from Hj is the smallest positive root £00j of the
equation
”j ¡ 0
j .t / D 0; t £j :
Condition (11.17) ensures that points £0j ; £00j are unique. Note that £00j is the
time instant when the entire reach set X .1/ Œt leaves Hj .
T .1/
Denote X .1/ Œt Hj D Zj .t /.
(c) After the crossing we have to envisage two branches:
() with no reset—then nothing changes and
Here in case () the reach tube develops further along the “old” subsystem (1),
while in case (+) it develops along “new” subsystem (j ).
For each new crossing we may now repeat the described procedure. Thus,
we may obtain the reach set X .t; t0 ; X 0 I 1C ; i Œ1s1 ; : : : ; i Œlsl / for a branch
f1C ; i Œ1s1 ; : : : ; i Œlsl g.
We further assume the next condition
0 00
Assumption 11.2.1. The neighboring intervals of crossing £› ; £› ; › D 1;
: : : ; l; do not intersect. It is presumed that £00› < ª; where Œt0 ; ª is the interval
under consideration.
Taking an interval Œt0 ; t , with £00iŒk1 t £0iŒk and £00iŒm1 t £0iŒm ; i Œl <
i Œm, one may observe the following semigroup type property.
Lemma 11.2.1. Each branch f1C ; i Œ1s1 ; : : : ; i Œksm g yields the following superpo-
sition property (1 k m)
and
Here
[
Z .j / Œ£00j D fX .j / .£00j ; t; Hj \ X .1/ .t; £0j ; X //j t 2 Œ£0j ; £00j g
t
Ti1C f£0i1 ; X .1/ .£0i1 ; £; X /; Œ1C g D f£00i1 ; Z .i1/ Œ£00i1 ; Œ1C ; i Œ1C g
Ti2 f£0i2 ; X .i1/ Œ£0i2 ; Œ1C ; i Œ1C g D f£00i2 ; X .i1/ .£00i2 ; £0i2 ; X .i1/ Œ£0i2 /; Œ1C ; i Œ1C ; i Œ2 g
ft I X .i1/ .t; £0i2 ; X .i1/ .£0i2 ; £00i1 ; Z .i1/ Œ£00i1 //; Œ1C ; i Œ1C ; i Œ2 g:
This is a branch of the overall reach set. The continuous variables for this branch
at t 2 Œ£00i2 ; £0i3 / produce the equality
An alternative scheme for calculating reach sets is to describe them through value
functions of optimization problems. Its advantage is that it is not restricted to linear
systems.
where
and x .i/ .t / D x.t / is the trajectory of the i-th system. We further also use the
notation V .1/ .t; x/ D V .1/ .t; x1 ; x2 ; : : : ; xn /.
.1/
(b) At the crossing we have X .1/ Œt \ Hj D Zj .t / which can be calculated as
follows.
.j /
Without loss of generality we may presume c1 D 1. Then
X
n
Zj.1/ .t / D fx W x1 D —.x/; V .1/ .t; —.x/; x2 ; : : : ; xn / 0g \ Hj ; —.x/ D ”1 .j /
c i xi :
i D2
¡C
j .t / D maxf.c
.j /
; x/ j V .1/ .t; x/ 0g;
¡
j .t / D minf.c
.j /
; x/ j V .1/ .t; x/ 0g:
C 0
This happens within the time interval Œ£0j ; £00j ; ¡ 00
j .£j / D ”j ; ¡j .£j / D ”j .
408 11 Verification: Hybrid Systems
With t > £00j this union may be calculated as the level set for function
where
V .j / .t; s; x/ D minfV .1/ .s; —.x.s//; x2 .s/; : : : ; xn .s//j s 2 Œ£0j ; £00j ; x.t / D xg;
u.j /
so that
Let us first calculate the reachability set after a one-stage crossing transformation.
(a) Starting with X 0 D E .x 0 ; X 0 /; i D 1; the reach set X .t I t0 ; X 0 ; Œ1C / is due
to equation
where
0 0
XP C D A.1/ .t /XC CXC A.1/ .t /C.t /XC C..t //1 B .1/ .t /P .1/ .t /B .1/ .t /;
.1/ .1/ .1/ .1/
(11.19)
0
XP .1/ D A.1/ .t /X.1/ C X.1/ A.1/ .t /C
0 .1/ 0
CX S.t /B .1/ .t /.P .1/ /1=2 .t / C .P .1/ /1=2 .t /B .1/ .t /S 0 .t /X ;
.1/
Here
(11.20)
Moreover,
[ \
E .x .1/ .t /; X.1/ .t jl// D X .t I t0 ; X 0 ; Œ1C / D E .x .1/ .t /; XC.1/ .t jl//:
l l
(11.21)
over all fl W hl; li 1g.
.j / .1/
(b) Let us now discuss the crossings E1 .t jl/ D E .x .1/ .t /; XC .t jl// \ Hj . Let
.i/
e be the unit orths for the original coordinate system.
Introducing the linear map
.j / .j /
E1 .t jl/ D E .z.j / .t /; Z1C .t j l// D
.1/ .1/
D E .x .1/ .t /; XC .t j l// \ H1 .j / D fx W VC .t; x j l/ 1; x1 D 0g;
where
.j /
The intersection E1 .t jl/ is a degenerate ellipsoid whose support function
is
.j / .j /
¡.ljE1 .t j l// D hl; z.j / .t /i C hl; Z1C .t j l/li1=2 ;
where
2
From here on it is important to emphasize the dependence of ellipsoids, reach sets, and value
functions on l. Therefore we further include l in the arguments of respective items.
11.2 Hybrid Dynamics and Control 411
Note that though generated by an ellipsoid, the set X .j / λ; t in general is not an
ellipsoid. However, it may be externally approximated by an array of parametrized
ellipsoids according to Sects. 3.2 and 3.3. Namely the following property is true.
.j /
Lemma 11.2.3. The exact reachability set from position ft; Z1C Œt g is
\ .j /
XC.j / λ; t D fXC λ; t j l j hl; li 1g:
0 0
XP C D A.j / .t /XC CXC A.j / .t /Cq .t /XC C.q .t //1 B .j / .t /P .j / .t /B .j / .t /;
.j / .j / .j / .j /
(11.23)
xP .j / D A.j / .t /x .j / C B .j / .t /p .j / .t / C C .j / .t /v.j / .t /;
.j / .j /
x .j / .£/ D z.j / .£/; XC .£/ D Z1 Œ£:
.j /
Note that the approximating sets E£ .x .j / .ª/; XC .ª j q; l// depend on two vector
parameters: l 2 Rn (calculated with D l ; from Eq. (11.20) and responsible for
.j /
finding Z1 Σ) and q 2 Rn1 (calculated with q ; from Eq. (11.19) and responsible
for finding X .j / Œª; £). Function q .t / is calculated by formulas similar to l .
We finally come to the following conclusion
Theorem 11.2.1. The following equality is true:
\
XC.j / λ; t j Z1.j / Σ D .j /
fE£ .x .j / .ª/; XC .ªjq; l// j klk 1; kqk 1g:
The last formula indicates the possibility of using parallel calculations through an
array of parametrized identical procedures and synchronization of their results.
412 11 Verification: Hybrid Systems
X .ª; t0 ; X 0 / D
[\ .j /
fE£ .x .j / .ª/; XC .ª j q; l// j klk 1; kqk 1; £ 2 T g: (11.25)
£ l;q
The above relation is written in the form of set-valued functions. But it may be
also written in terms of single-valued functions. Indeed, since
define
Then
Theorem 11.2.3. The reachability set X .ª; t0 ; X 0 / after one crossing with an active
switching is the level set
The above value functions thus allow both external approximations and an exact
description of the reach set. A similar scheme is true for internal approximations.
Recall that calculation of unions of reachability sets was discussed in Sect. 5.2
where an example was also given. For computational purposes we refer to following
considerations from this subsection.
According to the theory of minimax problems we have
\[ .j /
fE£ .x .j / .ª/; XC .ª j q; l// j klk 1; kqk 1; £ 2 T g D XC .ª; t0 ; X 0 /
l;q £
where
\
X .t; 0; X 0 / D fE .x .c/ .t /; XC .t j l// j hl; li 1g:
.1/
is an intersection of tight ellipsoids produced using (11.19), with x .1/ D x .c/ ; XC D
XC .t j l/.
Denote Vl .t; x/ D fx W hx x .c/ ; XC .t j l/.x x .c/ /i 1g. Then, due to
relations
we may find the exact nonconvex reachability set or its external approximation (also
nonconvex), namely,
Figure 11.1 demonstrates the intersection of nonconvex level sets over all fl W
hl; li 1g; which coincides with the external approximation XC .ª; 0; X .0/ /.
414 11 Verification: Hybrid Systems
Fig. 11.1 Nonconvex level sets of function minfVl .t; x/ j t 2 ‚g for fixed values of l and their
intersection
minfhc .j / ; xi j x 2 X Œt g D ” .t /;
T
we observe that condition X Œt Dj 6D ; will hold as long as ” C .t / ” .1/ and
” .t / ” .2/ , and the point of departure from Dj will be the smallest root £00j £0j
of equation ” .2/ ” .t / D 0.
11.2 Hybrid Dynamics and Control 415
minfhc .j / ; xi j x 2 X Œt g D ”r .t /;
and the time of departure of X Œt from Dj will be the smallest root £00jr of equation
” .2/ ”r .t / D 0.
Condition (11.17) ensures that the points £0j ; £00jr are unique. What follows is now
similar to what was written above for guards taken as hyperplanes.
We discuss an example.
Exercise 11.2.2. A branching reachability tube.
Consider the system
with parameters
0 1 0 2
A .1/
D ; A D
.2/
; B .1/ D B .2/ D I;
1 0 2 0
7 0
f .t /
1/
; f .t /
.2/
;
2 20
Fig. 11.3 Reachability tubes for both branches (left) and the overall branching reachability set
(right)
Indicated in Figs. 11.2 and 11.3 are external approximations for the two branches of
the reachability set after crossing the gap D at time t D 3. Here for i D 1 there is
no reset and the reachability set is a convex set X Œt , while for i D 2 there is a reset
from i D 1 to i D 2 and X Œt is a nonconvex union of convex sets of type (11.28).
11.3 Verification of Hybrid Systems 417
Let us investigate the solution of this problem for a given branch I .j / D Œ1C ;
.s /
i1 1 ; : : : ; ij j , assuming k D j; t > £00j .
.s /
(B) Suppose the sets X .j / Œt; s are convex, but their union set X .j / Œt D
[fX .j / Œt; s j s 2 Tj D Œ£0 ; £00 g is not convex. Then case (i) of Problem 11.3.1
holds if and only if
reachable within interval Œª0 ; ª00 ; ª00 £00 ; then one has to repeat operations similar
to (11.34)–(11.36), but for X Œt taken within this interval.
If there is more than one crossing, then each new crossing adds a new parameter
s D sk ; with range within related interval of type Œ£0k ; £00k . So, after j such active
intervals, it would be necessary to compute related unions of nonconvex sets involv-
ing optimization of parametrized value functions over parameters sk ; k D 1; : : : ; j .
The realization of such computations brings us to ellipsoidal approximations.
(ii) For a given s 2 Tk condition X .k/ Œt; s \ E .m; M / 6D ; holds if and only if
d2 Œt D minfd1 Œt; s j s 2 T g 0:
s
(iii) For a given s 2 Tk condition X .k/ Œt; s E .m; M / holds if and only if
d3 Œt D minfd1 Œt; s j s 2 T g 0:
s
Relations of this lemma allow us to compute the union X .k/ Œt in all the considered
cases.
If it is further required to compute
[
fX .ª; t0 ; X 0 / j ª 2 Œª0 ; ª00 g; ª £00 ;
then the previous schemes have to be applied once more, involving one more
parametrized array of ellipsoids.
Such overall procedures should be repeated after each active crossing, designing
a branching process, whose calculation would involve effective parallelization.
Computing sequential arrays of ellipsoids that correspond to related directions q; l;
through procedures of such parallelization and increasing the number of directions,
one may approach the exact solutions with any preassigned degree of accuracy.
In Sect. 11.2 we considered resets occurring only in the system model. We now
indicate how to treat resets in both system model and the state space variables. This
leads to the use of impulse controls for describing the resets.
420 11 Verification: Hybrid Systems
We now discuss situations when the hybrid system involves resets both in the system
number (marked as Rs .j / ) and in the phase coordinates (marked as Rcrd .i /). Upper
index s D with no crossing, s D C with reset in system number and s D with
reset in phase coordinates.
Returning to (11.19), consider hybrid system
Hj D fx j hc .j / ; xi ”j D 0g; c .j / 2 Rn ; ”j 2 R; j D 1; : : : ; m:
3
In more complicated problems, for example, with complex constraints on the number of
switchings or other outputs of resets, this memorized part may be logically controlled and its
knowledge may be important. Such components are also important in the design of feedback
controls and computation of backward reach sets for hybrid systems. Situations mentioned in this
footnote mostly lie beyond the scope of this book.
11.4 Impulse Controls in Hybrid System Models 421
We give some examples of hybrid systems with reset of their phase coordinates.
In order to formalize such situations we will introduce additional system inputs in
the form of exogenous impulse controls. We begin with two simple examples.
Example 11.2. The two-dimensional bouncing ball.
Consider a small freely falling heavy ball which bounces when striking the floor
(a plane, inclined to horizontal level), due to an elastic impact. It then moves further
along a trajectory which depends on the inclination angle of the floor at each new
bounce. Such angles, which act as controls, should be selected so that the ball ends
up in a fixed hole on the horizontal axis—the target. The process of bouncing may
cause some small losses of energy.
Following is an analytical, two-dimensional description of such bouncing
motion. A more complicated three-dimensional model of a bouncing motion is
indicated later.
The system equations are
where ® D ®.x; y/ are the values of the discrete control inputs which are presented
as inclination angles of the floor to horizontal line x2 D 0 at impact times ti . Such
inputs are modeled as delta functions of type v•.x2 .ti //.
The last equations are symbolic. Another representation of the same system,
according to Sect. 9.3.1 is
dx2 D y2 dt; dy2 D kv2 .x; y; ®/d U 0 .x2 ; 0/ gdt; 0 < k < 1;
where function U 0 .x; h/ D 1h .x/ (with scalar variable x) is a unit jump, namely,
1h .x/ D 0; if x < h and 1h .x/ D 1; if x h.
The solution of the last equations may be interpreted as a solution of the
next vector-valued integral equation with integral taken in the Lebesgue–Stiltjes
sense [234].
Z t Z t
x1 .t / D x10 C y1 .s/ds; y1 .t / D y10 C v1 .x.t /; y.t /; ®.x.t /; y.t ///d U 0 .x2 .t /; 0/;
t0 t0
Z t Z t
x2 .t / D x20 C y2 .s/ds; y2 .t / D y20 k v2 .x.t /; y.t /; ®.x.t /; y.t ///d U 0 .x2 .t /; 0/:
t0 t0
(11.38)
422 11 Verification: Hybrid Systems
Here, Fig. 11.4 shows the bouncing trajectory where at times of impact, with
x2 .ti / D 0; the vertical velocity y2 of the ball instantaneously changes its sign while
velocities y1 ; y2 change their values. This leads to an instantaneous jump (a reset)
of the values of vector fxP 1 .t /; xP 2 .t /g as soon as x2 .t / D 0. The inclination angle ®
of the floor at x2 .ti / D 0 therefore depends on the state space position of the system
and generates a feedback impulse control, thus reflecting the discrete component of
the hybrid process. The reset of variables fxP 1 .t / D y1 .t /; xP 2 .t / D y2 .t /g is due to
virtual •-valued controls of intensities v1 .x; y; ®/ and kv2 .x; y; ®/.
Example 11.3. One inflow for two water tanks.
Another is the well-known simple example of controlling two water tanks
through one inflow [104]. The feedback dynamics of this system is now formalized
using impulsive inputs.
This system is a model with two “motions”: (q1 ), (q2 ), with alternating resets
from one to the other. Here
.q1 / W xP 1 .t / D w v1 ; xP 2 D v2 ; x2 V2 ;
.q2 / W xP 1 .t / D v1 ; xP 2 D w v2 ; x1 V1 :
These equations describe a water supply at speed w through inflow from only
one pipe directed alternatingly to two water tanks from which water leaves with
11.4 Impulse Controls in Hybrid System Models 423
respective speeds v1 and v2 . The water levels to be maintained at the tanks are V1
and V2 . At each time t the pipe fills one of the tanks, then switching to the other
according to the following feedback rule.
Suppose at first the water is filling tank q2 with x1 > V1 ; x2 < V2 . Then, when
x1 decreases to level V1 the pipe switches from q2 to q1 (at that time it should be at
x2 > V2 ). Then later, after x2 had decreased to V2 , the pipe is reversed again to q2 .
The speeds and the water levels have to be coordinated to ensure sustainability of
the required water levels V1 ; V2 .
Exercise 11.4.1. Indicate the range of speeds w; v1 ; v2 and water levels V1 ; V2 when
these levels are sustainable.
The two systems may be combined into one, switching from one to the other through
two impulsive control inputs generated due to feedback control signals. Denoting
the right-hand sides of systems (q1 ) and (q2 ) by F1 .x/; F2 .x/ and introducing a
third variable — that has only two possible values ˙1; consider the following system
in symbolic form
8
< xP D F .x/ 1Csign— C F .x/ 1sign—
1 2 2 2 ;
: P
— D 2•.x1 V1 C 0/ 2•.x2 V2 C 0/
or as
8
< dx D F .x/ 1Csign— C F .x/ 1sign— dt
1 2 2 2 :
:
d — D 2d U .x1 V1 ; 0/ 2d U .x2 V2 ; 0/
0 0
The solution of this equation is defined through integral equations taken in the sense
of the Lebesgue–Stiltjes integral and similar to the type (11.37) (Fig. 11.5).
—P D u— ; (11.40)
x 2 Rn ; — 2 f1; : : : ; kg:
impulse control responsible for reset of the phase coordinates and has the form
.imp/ .imp/
u.imp/ D u.imp/ .t; x; —/I uj D uj .t; x; —/ D ’.t; x; —.t 0//•.f .x; —//:
The scalar u— controls the number — of system Eq. (11.39), and is the input in
Eq. (11.40) for finding the integer — 2 f1; : : : ; kg and has the form
t 2 R; x 2 x D f.x1 ; : : : ; x6 / W xi 2 R; x3 0g:
4
This example was worked out by P.A. Tochilin.
426 11 Verification: Hybrid Systems
Here ” 2 Œ0; 1 is the “recovery” coefficient: with ” D 1 we have the model for an
absolutely elastic impact, and with ” D 0 for a nonelastic.
Such a model may now be also formalized involving •-functions.
With ®2 D 0 (when plane H may rotate only around axis l1 ), then (11.42), taken in
coordinate form, is
8
ˆ
ˆ xP 1 D x4
ˆ
ˆ
ˆ
ˆ xP 2 D x5
ˆ
<
xP 3 D x6
;
ˆ
ˆ xP 4 D 0
ˆ
ˆ
ˆ
ˆ x5 D .1 C ”/.sin .®1 /x5 C sin.®1 / cos.®1 /x6 /•.x3 /
2
:̂
x6 D .1 C ”/.sin.®1 / cos.®1 /x5 C cos2 .®1 /x6 /•.x3 / g
and with ®1 D 0 (when H may rotate only around l2 ), Eq. (11.42) in detail is:
8
ˆ
ˆ xP 1 D x4
ˆ
ˆ
ˆ
ˆ xP 2 D x5
ˆ
<
xP 3 D x6
:
ˆ xP 4
ˆ D .1 C ”/.sin2 .®1 /x4 sin.®1 / cos.®1 /x6 /•.x3 /
ˆ
ˆ
ˆ
ˆ x5 D0
:̂
x6 D .1 C ”/. sin.®1 / cos.®1 /x4 C cos2 .®1 /x6 /•.x3 / g
At each time t the on-line controls ®1 ; ®2 are determined by the running position
x.t / of the system.
On horizontal plane H D H0 D fx W x3 D 0g consider a hole (target set)
Also specify a starting position x 0 D x.t0 / D .x10 ; x20 ; x30 ; 0; 0; 0/0 , where x30 > 0.
The problem will be to direct the trajectory of the bouncing ball, by choice
of angles ®1 ; ®2 (the exogenous controls) to the target set M . Here the system
parameters and the matrix C are independent of time.
Considered are the following optimization problems.
Problem 11.4.1. Find minimal time t1 t0 , for which there exists a control
®1 ./; ®2 ./, that ensures inclusion x.t1 I t0 ; x0 /j®1 ;®2 2 M R3 with any final
velocity.
Problem 11.4.2. Given ft0 ; x 0 g; find the minimal number k of bounces (resets)
k 2 ZC (the set of positive integers), such that there exist a time interval Œt0 ; t1 ; and
a control ®1 ./; ®2 ./ for which the inclusion xŒt1 D x.t1 I t0 ; x 0 /j®1 ;®2 2 M R3 is
achieved with number of bounces k k .
11.4 Impulse Controls in Hybrid System Models 427
CI .x3 .tI t0 ; x/j®1 ;®2 j 0/ ;
V .2/
.x; k/ D min min Œ.x1 .tI t0 ; x/j®1 ;®2 m1 /2 C.x2 .tI t0 ; x/j®1 ;®2 m2 /2 r 2 C D
t t0 ®1 ./;®2 ./
CI .x3 .tI t0 ; x/j®1 ;®2 j 0/ C ™.x.I t0 ; x/j®1 ;®2 ; t; k/ :
Here Œh.t; x/C D 0 if h.t; x/ 0; and Œh.t; x/C D h.t; x/ if h.t; x/ > 0;
function ™.x.I t0 ; x/j®1 ;®2 ; t; k/ D 0 within interval £ 2 Œt0 ; t , if the number of
bounces does not exceed k, otherwise it is C1.5
To compute solutions of Problems 11.4.1 and 11.4.2, we may use the following
facts.
Lemma 11.4.1. (1) With x 0 given, the solution to Problem 11.4.1—the minimal
time t1 D t1 —is:
For computing V .1/ .t1 ; x/; V .2/ .x; k/, there exist related HJB equations with appro-
priate boundary conditions (see [171]).
Exercise 11.4.2. Indicate the HJB equations with boundary conditions for comput-
ing value functions V .1/ .t1 ; x/; V .2/ .x; k/.
Remark 11.4.3. Value functions V .1/ .t1 ; x/; V .2/ .x; k/ may be also computed
through methods of nonlinear analysis and ellipsoidal techniques, along the lines of
Chaps. 2 and 3.
With V .1/ .t1 ; x/; V .2/ .x; k/ computed, one may specify the solutions to Prob-
lems 11.4.1 and 11.4.2.
5
As before I .x j D/ D 0, if x 2 D and C1 otherwise.
428 11 Verification: Hybrid Systems
will be of type
.®1 .x/; ®2 .x// W V .1/ .t1 .x/; C.®1 ; ®2 /x/ V .1/ .t1 .x/; x/;
.®1 .x/; ®2 .x// W V .2/ .C.®1 ; ®2 /x; k/ V .2/ .x; k .x//;
until k D k 1.
Shown in Fig. 11.6 is the trajectory of the 3-d bouncing ball (a hybrid system)
which reaches the target set with minimum resets (bounces). The problem parame-
ters are: ’1 D 0:5, ’2 D 0:5, m1 D 1, m1 D 2, r 2 D 0:2, ” D 0:8, and the starting
position is x 0 D .4; 1; 4; 0; 0; 0/0 .
Shown in Fig. 11.7 is the time-optimal target-oriented trajectory of the bouncing
ball with parameters ’1 D 1, ’2 D 0:4, m1 D 1, m2 D 2, r 2 D 0:2, ” D 0:9 and
starting position x 0 D .2; 3; 4; 0; 0; 0/0 .
1. Akhiezer, N.I.: The Classical Moment Problem and Some Related Questions in Analysis.
Oliver & Boyd, Edinburgh (1965)
2. Albert, A.: Regression and the Moor-Penrose Pseudoinverse. Mathematics in Science and
Engineering, vol. 94. Academic, New York (1972)
3. Artstein, Z.: Yet another proof of the Lyapunov convexity theorem. Proc. Am. Math. Soc.
108(1), 89–91 (1990)
4. Åström, K.J.: Introduction to Stochastic Control Theory. Academic, New York (1970)
5. Aubin, J.-P.: Viability Theory. SCFA. Birkhäuser, Boston (1991)
6. Aubin, J.-P., Bayen, A.M., St.Pierre, P.: Viability Theory: New Directions. Springer,
Heidelberg (2011)
7. Aubin, J.-P., Cellina, A.: Differential Inclusions. Springer, New York (1984)
8. Aubin, J.-P., Frankowska, H.: Set-Valued Analysis. SCFA 2. Birkhäuser, Boston (1990)
9. Aumann, R.J.: Integrals of set-valued functions. J. Math. Anal. Appl. 12, 1–12 (1965)
10. Bai, E.W., Cho, H., Tempo, R.: Optimization with few violated constraints for linear bounded
error parameter estimation. IEEE Trans. Automat. Control 47(7), 1067–1077 (2002)
11. Balakrishnan, A.V.: Applied Functional Analysis. Springer, New York (1976)
12. Balluchi, A., Benvenuti, L., Di Benedetto, M.D., Sangiovanni-Vincentelli, A.: The design
of dynamical observers for hybrid systems: theory and application to an automotive control
problem. Automatica 49(4), 915–925 (2013)
13. Baras, J.S., Bensoussan, A., James, M.R.: Dynamic observers and asymptotic limits of
recursive filters: special case. SIAM J. Appl. Math. 48(5), 1147–1158 (1988) 1
14. Baras, J.S., Kurzhanski, A.B.: Nonlinear filtering: the set-membership (bounding) and the
H1 techniques. In: Proceedings of the 3rd IFAC Symposium NOLCOS, pp. 409–418.
Pergamon Press, Oxford (1995)
15. Barbashin, E.A.: On the theory of generalized dynamic systems. Sci. Notes Moscow State
Univ. 135, 110–133 (1949)
16. Bardi, M., Capuzzo Dolcetta, I.: Optimal Control and Viscosity Solutions of Hamilton–
Jacobi–Bellman Equations. SCFA. Birkhäuser, Boston (1997)
17. Barron, E.N., Jensen, R.: Semicontinuous viscosity solutions for Hamilton–Jacobi equations
with convex Hamiltonians. Commun. Partial Differ. Equ. 15(12), 1713–1742 (1990)
18. Başar, T., Bernhard, P.: H1 Optimal Control and Related Minimax Design Problems. SCFA,
2nd edn. Birkhäuser, Basel (1995)
19. Ba sar, T., Olsder, J.: Dynamic Noncooperative Game Theory. Academic, New York (1982)
20. Basile, G., Marro, G.: Controlled and Conditioned Invariants in Linear System Theory.
Prentice-Hall, Englewood Cliffs (1992)
21. Bellman, R., Dreyfus, S.: Applied Dynamic Programming. Princeton University Press,
Princeton (1962)
22. Bellman, R.: Introduction to the Mathematical Theory of Controlled Processes, vol. 1/2.
Academic, New York (1967/1971)
23. Bensoussan, A., Lions, J.-L.: Contrôle impulsionnel et inéquations quasi-variationnelles.
Dunod, Paris (1982)
24. Bertsekas, D.P.: Dynamic Programming and Optimal Control, vol. 1/2. Athena Scientific,
Belmont (1996/2012)
25. Bertsekas, D.P., Rhodes, I.: Recursive state estimation for a set-membership description of
uncertainty. IEEE Trans. Automat. Control 16(2), 117–128 (1971)
26. Blagodatskih, V.I, Filippov, A.F.: Differential inclusions and optimal control. Proc. Steklov
Math. Inst. 169, 194–252 (1985)
27. Blanchini, F., Miani, S.: Set-Theoretic Methods in Control. Springer, Boston (2008)
28. Blaquière, A., Leitmann, G. (eds.): Dynamical Systems and Microphysics: Control Theory
and Mechanics. Academic, Orlando (1982)
29. Bliss, G.A.: Calculus of Variations. Mathematical Association of America, Chicago (1944)
30. Boltyanski, V.G.: Mathematical Methods of Optimal Control. Holt, New York (1971)
31. Boltyanski, V.G.: Tent methods in optimal control. ISNM International Series of Numerical
Mathematics, vol. 111, pp. 3–20
32. Bohner, M., Peterson, A.: Dynamic Equations on Time Scales: An Introduction with
Applications. Birkhauser, Boston (2001)
33. Bolza, O.: Lectures on Calculus of Variations. Hafer Pub. Co, New York (1946). Dover reprint
34. Bonneuil, N.: Computing the viability kernel in large state dimensions. J. Math. Anal. Appl.
323, 1444–154 (2006)
35. Boyd, S., Ghaoui, L.E., Feron, E., Balakrishnan, V.: Linear Matrix Inequalities in System and
Control Theory. Studies in Applied Mathematics, vol. 15. SIAM, Philadelphia (1994)
36. Branicky, M.S., Borkar, V.S., Mitter, S.K.: A unified framework for hybrid control: model and
optimal control theory. IEEE Trans. Automat. Control 43(1), 31–45 (1998)
37. Bressan, A., Piccoli, B.: Introduction to the Mathematical Theory of Control, vol. 2. American
Institute for Mathematical Sciences, Springfield (2007)
38. Brockett, R.W.: Finite-Dimensional Linear Systems. Wiley, New York (1970)
39. Brockett, R.W.: Hybrid models for motion control systems. In: Trentelman, H., Willems,
J. (eds.) Essays in Control: Perspectives in the Theory and Its Applications, pp. 29–53.
Birkhäuser, Boston (1993)
40. Callier, F.M., Desoer, C.A.: Linear System Theory. Springer, New York (1991)
41. Castaing, C., Valadier, M.: Convex Analysis and Measurable Multifunctions. Lecture Notes
in Mathematics, vol. 580. Springer, Berlin (1977)
42. Carter, T.E.: Optimal impulsive space trajectories based on linear equations. J. Optim. Theory
Appl. 70(2), 277–297 (1991)
43. Carter, T.E., Brient, J.: Linearized impulsive rendezvous problem. J. Optim. Theory Appl.
86(3), 553–584 (1995)
44. Chernousko, F.L.: Ellipsoidal bounds for sets of attainability and uncertainty in control.
Optim. Control Appl. Methods 3(2), 187–202 (1982)
45. Chernousko, F.L.: State Estimation for Dynamic Systems. CRC Press, Boca Raton (1994)
46. Chutinan, A., Krogh, B.H.: Verification of polyhedral-invariant hybrid automata using
polygonal flow pipe approximations. In: Hybrid Systems: Computation and Control. Lecture
Notes in Control Science, vol. 1569, pp. 76–90. Springer, Berlin (1999)
47. Clarke, F.H.: Optimization and Nonsmooth Analysis. Wiley, New York (1983)
48. Clarke, F.H., Ledyaev Yu, S., Stern, R.J., Wolenski, P.R.: Nonsmooth Analysis and Control
Theory. Graduate Texts in Mathematics, vol. 178. Springer, New York (1998)
49. Crandall, M.G., Evans, L.C., Lions, P.-L.: Some properties of solutions of Hamilton–Jacobi
equations. Trans. Am. Math. Soc. 282(2), 487–502 (1984)
References 433
50. Crandall, M.G., Lions, P.-L.: Viscosity solutions of Hamilton–Jacobi equations. Trans. Am.
Math. Soc. 277(1), 1–41 (1983)
51. Courant , R., Hilbert, D.: Methods of Mathematical Physics, vol. 1 & 2. Wiley, New York
(1989) (Wiley GmbH & Co., KGaA, 2008)
52. Dar’in, A.N., Kurzhanskii, A.B.: Nonlinear control synthesis under two types of constraints.
Differ. Equ. 37(11), 1549–1558 (2001)
53. Daryin, A.N., Kurzhanskii, A.B.: Control under indeterminacy and double constraints. Differ.
Equ. 39(11), 1554–1567 (2003)
54. Daryin, A.N., Kurzhanskii, A.B., Seleznev, A.V.: The dynamic programming method in
impulsive control synthesis. Differ. Equ. 41(11), 1566–1576 (2005)
55. Daryin, A.N., Kurzhanski, A.B., Seleznev, A.V.: A dynamic programming approach to
the impulse control synthesis problem. In: Proceedings of Joint 44th IEEE Conference
on Decision and Control and European Control Conference ECC 2005, Seville, Spain,
pp. 8215–8220 (2005)
56. Daryin, A.N., Digailova, I.A., Kurzhanski, A.B.: On the problem of impulse measurement
feedback control. Proc. Steklov Math. Inst. 268(1), 71–84 (2010)
57. Daryin, A.N., Kurzhanski, A.B.: Estimation of reachability sets for large-scale uncertain
systems: from theory to computation. In: Proceedings of CDC-2012 (2012)
58. Daryin, A.N., Kurzhanski, A.B.: Parallel algorithm for calculating the invariant sets of high-
dimensional linear systems under uncertainty. Comput. Math. Math. Phys. 53(1), 34–43
(2013)
59. Davis, M.H.A., Varaiya, P.: Information states for linear stochastic systems. J. Math. Anal.
Appl. 37, 384–402 (1972)
60. Demyanov, V.F.: Minimax: Directional Differentiation. Leningrad University Pub. Leningrad
(1974)
61. Demyanov, V.F., Malozemov, V.N.: Introduction to Minimax. Wiley, New York (1974)
62. Demyanov, V.F., Rubinov, A.M.: Quasidifferential Calculus. Optimization Software,
New York (1986)
63. Digailova, I.A., Kurzhanski, A.B.: The joint model and state estimation problem under
set-membership uncertainty. In: Proceedings of the 15-th IFAC Congress, Barcelona (2002)
64. Digailova, I.A., Kurzhanski, A.B.: Reachability analysis under control-dependent stochastic
noise. In: Proceedings of 16th IFAC World Congress, Prague (2005)
65. Digailova, I.A., Kurzhanski, A.B., Varaiya, P.: Stochastic reachability and measurement feed-
back. In: Proceedings of 17th World Congress IFAC, Seoul, 6–11 July 2008, pp. 14336–14341
66. Dontchev, A.L.: Perturbations, Approximations, and Sensitivity Analysis of Optimal Control
Systems. Lecture Notes in Control and Information Sciences, vol. 52. Springer, New York
(1983)
67. Dunford, N., Schwartz, J.T.: Linear Operators. Interscience, New York/London (1958)
68. Dykhta, V.A., Samsonyuk, O.N.: Optimal Impulse Control and Its Applications. Fizmatlit,
Moscow (2003) [in Russian]
69. Ekeland, I., Temam, R.: Analyse Convexe et Problemes Variationelles. Dunod, Paris (1973)
70. Fan, K.: Minimax theorems. Proc. Natl. Acad. Sci. USA 39(1), 42–47 (1953)
71. Fan, K.: On systems of linear inequalities. In: Kuhn, H.W., Tucker, A.W. (eds.) Linear
Inequalities and Related Systems. Annals of Mathematical Studies, vol. 38, pp. 99–156.
Princeton University Press, Princeton (1956)
72. Fan, K.: Existence theorems and extreme solutions for inequalities concerning convex
functions or linear transformations. Mathematische Zeitschrift 68(1), 205–216 (1957)
73. Fenchel, W.: On conjugate convex functions. Can. J. Math. 1, 73–77 (1949)
74. Fenchel, W.: Convex Cones, Sets and Functions. Princeton University Press, Princeton (1951)
75. Filippov, A.F.: On certain questions in the theory of optimal control. SIAM J. Control. 1,
76–84 (1962)
76. Filippov, A.F.: Differential Equations with Discontinuous Righthand Sides. Kluwer, Dor-
drecht (1988)
434 References
77. Filippova, T.F.: A note on the evolution property of the assembly of viable solutions to a
differential inclusion. Comput. Math. Appl. 25(2), 115–121 (1993)
78. Filippova, T.F., Kurzhanski, A.B., Sugimoto, K., Valýi, I.: Ellipsoidal calculus, singular
perturbations and the state estimation problems for uncertain systems. Working Paper
WP-92-051, IIASA, Laxenburg (1992)
79. Filippova, T.: Approximation techniques in impulsive control problems for the tubes of solu-
tions of uncertain differential systems. In: Anastassiou, G.A., Duman, O. (eds.) Advances in
Applied Mathematics and Approximation Theory: Contributions from AMAT, pp. 385–396.
Springer, New York (2012)
80. Fleming, W.H., Soner, H.M.: Controlled Markov Processes and Viscosity Solutions. Springer,
New York (1993)
81. Gamkrelidze, R.V.: Principles of Optimal Control Theory. Mathematical Concepts and
Methods in Science and Engineering, vol. 7. Plenum Press, New York (1978)
82. Gantmacher, F.R.: Matrix Theory, vol. 2. Chelsea Publishing, New York (1964)
83. Gao, Y., Lygeros, L., Quincampoix, M.: On the reachability problem for uncertain hybrid
systems. IEEE Trans. Automat. Control 52(9), 1572–1586 (2007)
84. Garcia, C.E., Prett, D.M., Morari, M.: Model predictive control: theory and practice-a survey.
Automatica 25(3), 335–348 (1989)
85. Gayek, J.E.: A survey of techniques for approximating reachable and controllable sets, Proc.
of the CDC Conf. Brighton, England, 1724–1729 (1991)
86. Gelfand, I.M., Fomin, S.V.: Calculus of Variations. Academic, New York (1964)
87. Gelfand, I.M., Shilov, G.E.: Generalized Functions. Dover, New York (1991)
88. Giannessi, F., Maugeri, A. (eds.): Variational Methods and Applications. Kluwer, New York
(2004)
89. Godunov, S.K.: Guaranteed Accuracy in Numerical Linear Algebra. Springer, Dordrecht
(1993)
90. Godunov, S.K.: Ordinary Differential Equations with Constant Coefficients, vol. 169. AMS
Bookstore, Providence (1997)
91. Goebel, R., Sanfelice, R.G., Teel, A.R.: Hybrid Dynamical Systems. Princeton University
Press, Princeton (2012)
92. Golub, G.H., Van Loan, C.F.: Matrix Computations. John Hopkins Un. Press, Baltimore, MI
(1996)
93. Grant, M., Boyd, S.: Graph interpretation for nonsmooth convex programs. In: Recent
Advances in Learning and Control, pp. 95–110. Springer, Berlin (2008)
94. Greenstreet, M.R., Mitchell, I.: Reachability analysis using polygonal projections. In:
Hybrid Systems: Computation and Control. Lecture Notes in Control Science, vol. 1569,
pp. 103–116. Springer, Berlin (1999)
95. Gurman, V.I.: The Extension Principle in Problems of Control. Nauka, Moscow (1997)
96. Gusev, M.I., Kurzhanski, A.B.: Optimization of controlled systems with bounds on the
controls and state coordinates, I. Differ. Equ. 7(9), 1591–1602 (1971) [translated from
Russian]
97. Gusev, M.I., Kurzhanski, A.B.: Optimization of controlled systems with bounds on the
controls and state coordinates, II. Differ. Equ. 7(10), 1789–1800 (1971) [translated from
Russian]
98. Helton, J.W., James, M.R.: Extending H1 Control to Nonlinear Systems. SIAM, Philadel-
phia (1999)
99. Ioffe, A.D., Tikhomirov, V.M.: Theory of Extremal Problems. North-Holland Amsterdam
(1979)
100. Isidori, A., Astolfi, A.: Disturbance attenuation and H1 control via measurement feedback
in nonlinear systems. IEEE Trans. Automat. Control 37, 1283–1293 (1992)
101. Isidori, A., Marconi, L.: Asymptotic analysis and observer design in the theory of nonlinear
output. In: Nonlinear Observers and Applications. Lecture Notes in Control and Information
Sciences, vol. 363, pp. 181–210. Springer, Berlin (2007)
102. Isaacs, R.: Differential Games. Wiley, New York (1965)
References 435
103. James, M.R., Baras, J.S.: Partially observed differential games, infinite-dimensional
Hamilton–Jacobi–Isaacs equations and nonlinear H1 control . SIAM J. Control Optim 34(4),
1342–1364 (1996)
104. Johansson, M.: Piecewise Linear Control Systems. Lecture Notes in Control and Information
Sciences, vol. 284. Springer, Berlin (2003)
105. Kailath, T.: Linear Systems. Prentice Hall, Englewood Cliffs (1980)
106. Kalman, R.E.: Contributions to the theory of optimal control. Bol. Soc. Mat. Mexicana. 5,
102–119 (1959)
107. Kalman, R.E.: On the general theory of control systems. In: Proceedings of the 1-st IFAC
Congress, Butterworth, London, vol. 1 (1960)
108. Kalman, R.E.: A new approach to linear filtering and prediction problems. Trans. ASME
82(Series D), 35–45 (1960)
109. Kalman, R.E., Ho, Y.-C., Narendra, K.S.: Controllability of linear dynamical systems.
Contrib. Theory Differ. Equ. 1(2), 189–213 (1963)
110. Kolmogorov, A.N., Fomin, S.V.: Elements of the Theory of Functions and Functional
Analysis. Courier Dover, New York (1999)
111. Komarov, V.A.: Estimates of reachability sets for differential inclusions. Math. Notes
(Matematicheskiye Zametki) 37(6), 916–925 (1985)
112. Komarov, V.A.: Equation of reachability sets for differential inclusions under state constraints.
Proc. Steklov Math. Inst. 185, 116–125 (1989)
113. Koscheev, A.S., Kurzhanski, A.B.: On adaptive estimation of multistage systems under
uncertainty. Izvestia Academy Sci. 2, 72–93 (1983)
114. Kostousova, E.K.: State estimation for dynamic systems via parallelotopes: optimization and
parallel computations. Optim. Methods Softw. 9(4), 269–306 (1998)
115. Kostousova, E.K.: Control synthesis via parallelotopes: optimization and parallel computa-
tions. Optim. Methods Softw. 14(4), 267–310 (2001)
116. Kostousova, E.K.: External polyhedral estimates for reachable sets of linear discrete-time
systems with integral bounds on controls. Int. J. Pure Appl. Math. 50(2), 187–194 (2009)
117. Kostousova, E.K.: On tight polyhedral estimates for reachable sets of linear differential
systems. In: ICNPAA 2012 World Congress: 9th International Conference on Mathematical
Problems in Engineering, Aerospace and Sciences (2012)
118. Kostousova, E.K., Kurzhanski, A.B.: Theoretical framework and approximation techniques
for parallel computation in set-membership state estimation. In: Proceedings of Symposium
on Modelling, Analysis and Simulation. CESA’96 IMACS Multiconference, CESA’96, Lille-
France, vol. 2, pp. 849–854 (1996)
119. Krasovski, N.N.: On the theory of controllability and observability of linear dynamic systems.
Appl. Math. Mech. (PMM) 28(1), 3–14 (1964)
120. Krasovski, N.N.: Motion Control Theory. Nauka, Moscow (1968) [in Russian]
121. Krasovski, N.N.: Rendezvous Game Problems. National Technical Information Service,
Springfield (1971)
122. Krasovskii, A.N., Krasovskii, N.N.: Control Under Lack of Information. Birkhäuser, Boston
(1995)
123. Krasovskii, N.N., Subbotin, A.I.: Game-Theoretical Control Problems. Springer Series in
Soviet Mathematics. Springer, New York (1988)
124. Krein, M., Smulian, V.: On regular convex sets in the space conjugate to a Banach space. Ann.
Math. 41, 556–583 (1940)
125. Krener, A.J.: Eulerian and Lagrangian observability of point vortex flows. Tellus A 60,
1089–1102 (2008)
126. Krener, A.J., Ide, K.: Measures of unobservability. In: Proceedings of the IEEE Conference
on Decision and Control, Shanghai (2009)
127. Krener, A.J., Wang, K.: Rational minmax filtering. In: Proceedings of IFAC-2011 World
Congress, Milan (2011)
128. Krener, A.J., Kang, W.: Linear time-invariant minmax filtering. Ann. Rev. Control 35,
166–171 (2011)
436 References
129. Kruzhkov, S.N.: Generalized solutions of nonlinear first-order PDE’s with many independent
variables. Math. Sb. 70((112)3), 394–415 (1966)
130. Kuhn, H.W., Tucker, A.W. (eds.): Linear Inequalities and Related Systems. Annals of
Mathematical Studies, vol. 38. Princeton University Press, Princeton (1956)
131. Kuratowski, K.: Topology, vol. 1/2. Academic, New York (1966/1968)
132. Kurzhanskiy, A.A., Varaiya, P.: Ellipsoidal Toolbox. University of California at Berkeley.
http://code.google.com/p/ellipsoids/ (2005)
133. Kurzhanskiy, A.A., Varaiya, P.: Ellipsoidal techniques for reachability analysis of discrete-
time linear systems. IEEE Trans. Automat. Control 52(1), 26–38 (2007)
134. Kurzhanskiy, A.A., Varaiya, P.: Active traffic management on road networks. Philos. Trans.
R. Soc. A 368, 4607–4626 (2010)
135. Kurzhanski, A.B.: On duality of problems of optimal control and observation. Appl. Math.
Mech. 34(3), 429–439 (1970) [translated from the Russian]
136. Kurzhanski, A.B.: Differential games of observation. Russ. Math. Dokl. (Dok. Akad. Nauk
SSSR) 207(3), 527–530 (1972)
137. Kurzhanski, A.B.: Control and Observation Under Uncertainty Conditions, 392 pp. Nauka,
Moscow (1977)
138. Kurzhanskii, A.B.: On stochastic filtering approximations of estimation problems for systems
with uncertainty. Stochastics 23(6), 109–130 (1982)
139. Kurzhanski, A.B.: Evolution equations for problems of control and estimation of uncertain
systems. In: Proceedings of the International Congress of Mathematicians, pp. 1381–1402
(1983)
140. Kurzhanski, A.B.: On the analytical description of the set of viable trajectories of a differential
system. Dokl. Akad. Nauk SSSR. 287(5), 1047–1050 (1986)
141. Kurzhanski, A.B.: Identification: a theory of guaranteed estimates. In: Willems, J.C. (ed.)
From Data to Model, pp. 135–214. Springer, Berlin (1989)
142. Kurzhanski, A.B.: The identification problem: a theory of guaranteed estimates. Autom.
Remote Control (Avtomatica i Telemekhanika) 4, 3–26 (1991)
143. Kurzhanski, A.B.: Set-Valued Analysis and Differential Inclusion. Progress in Systems and
Control Theory, vol. 16. Birkhäuser, Boston (1993)
144. Kurzhanski, A.B.: Modeling Techniques for Uncertain Systems. Progress in Systems and
Control Theory, vol. 18. Birkhäuser, Boston (1994)
145. Kurzhanski, A.B.: Set-valued calculus and dynamic programming in problems of feedback
control. Int. Ser. Numer. Math. 124, 163–174 (1998)
146. Kurzhanski, A.B.: The principle of optimality in measurement feedback control for linear
systems. In: Ranter, A., Byrnes, C.V. (eds.) Directions in Mathematical Systems Theory
and Optimization. Lecture Notes in Control Science, vol. 286, pp. 193–202. Springer, Berlin
(2003)
147. Kurzhanski, A.B.: The problem of measurement feedback control. J. Appl. Math. Mech.
(PMM) 68(4), 487–501 (2004) [translated from the Russian]
148. Kurzhanski, A.B.: The diagnostics of safety zones in motion planning. Optim. Methods Softw.
20(2/3), 231–239 (2005)
149. Kurzhanski, A.B.: Comparison principle for equations of the Hamilton–Jacobi type in control
theory. Proc. Steklov Math. Inst. 253(1), 185–195 (2006)
150. Kurzhanski, A.B.: “Selected Works of A.B. Kurzhanski”, 755 pp. Moscow State University
Pub., Moscow (2009)
151. Kurzhanski, A.B.: Hamiltonian techniques for the problem of set-membership state estima-
tion. Int. J. Adapt. Control Signal Process. 25(3), 249–263 (2010)
152. Kurzhanski, A.B.: On synthesizing impulse controls and the theory of fast controls. Proc.
Steklov Math. Inst. 268(1), 207–221 (2010)
153. Kurzhanski, A.B., Baras, J.: Nonlinear filtering: the set-membership (bounding) and the
H-infinity techniques. In: Proceedings of the IFAC Symposium NOLCOS-95 at Lake Tahoe,
USA, pp. 409–418. Pergamon Press, Oxford (1995)
References 437
154. Kurzhanski, A.B., Daryin, A.N.: Dynamic programming for impulse controls. Ann. Rev.
Control 32(2), 213–227 (2008)
155. Kurzhanski, A.B., Digailova, I.A.: Attainability problems under stochastic perturbations.
Differ. Equ. 40(4), 1573–1578 (2004)
156. Kurzhanski, A.B., Filippova, T.F.: On the set-valued calculus in problems of viability and
control for dynamic processes: the evolution equation. Annales de l’institut Henri Poincaré.
Analyse non linéaire S6, 339–363 (1989)
157. Kurzhanski, A.B., Filippova, T.F.: On the method of singular perturbations for differential
inclusions. Dokl. Akad. Nauk SSSR 321(3), 454–459 (1991)
158. Kurzhanski, A.B., Filippova, T.F.: On the theory of trajectory tubes: a mathematical formalism
for uncertain dynamics, viability and control. In: Advances in Nonlinear Dynamics and
Control. Progress in Systems and Control Theory, vol. 17, pp. 122–188. Birkhäuser, Boston
(1993)
159. Kurzhanski, A.B., Filippova, T.F., Sugimoto, K., Valyi, I.: Ellipsoidal state estimation for
uncertain dynamical systems. In: Milanese, M., Norton, J., Piet-Lahanier, H., Walter, E. (eds.)
Bounding Approaches to System Identification, pp. 213–238. Plenum Press, New York (1996)
160. Kurzhanski, A.B., Kirilin, M.N.: Ellipsoidal techniques for reachability problems under
nonellipsoidal constraints. In: Proceedings of the NOLCOS-01, St. Petersburg, pp. 768–768
(2001)
161. Kurzhanski, A.B., Mitchell, I.M., Varaiya, P.: Optimization techniques for state-constrained
control and obstacle problems. J. Optim. Theory Appl. 128(3), 499–521 (2006)
162. Kurzhanski, A.B., Mitchell, I.M., Varaiya, P.: Control synthesis for state constrained systems
and obstacle problems. In: Proceedings of the IFAC Eds. Allgower F., Zeitz M., Oxford.
NOLCOS 6-th Symposium at Suttgart, Elsevier, Amsterdam (2005–2014)
163. Kurzhanski, A.B., Nikonov, O.I.: On the control strategy synthesis problem. Evolution
equations and set-valued integration. Dokl. Akad. Nauk SSSR. 311(4), 788–793 (1990)
164. Kurzhanski, A.B., Nikonov, O.I.: Evolution equations for trajectory bundles of synthesized
control systems. Russ. Acad. Sci. Dokl. Math. 333(5), 578–581 (1993)
165. Kurzhanski, A.B., Osipov, Y.S.: The problem of control with bounded phase coordinates.
Appl. Math. Mech. (PMM) 32(2), 188–195 (1968)
166. Kurzhanski, A.B., Osipov, Y.S.: On optimal control under state constraints. Appl. Math.
Mech. (PMM) 33(4), 705–719 (1969)
167. Kurzhanski, A.B., Sivergina, I.F.: Method of guaranteed estimates and regularization prob-
lems for evolutionary systems. J. Comput. Math. Math. Phys. 32(11), 1720–1733 (1992)
168. Kurzhanski, A.B., Sugimoto, K., Vályi, I.: Guaranteed state estimation for dynamical systems:
ellipsoidal techniques. Int. J. Adapt. Control Signal Process. 8(1), 85–101 (1994)
169. Kurzhanski, A.B., Tanaka, M.: Identification: deterministic versus stochastic models. Öster-
reichische Zeitschrift für Statistik und Informatik 19(1), 30–56 (1989)
170. Kurzhanski, A.B., Tochilin, P.A.: Weakly invariant sets of hybrid systems. Differ. Equ. 44(11),
1585–1594 (2008)
171. Kurzhanski, A.B., Tochilin, P.A.: Impulse controls in models of hybrid systems. Differ. Equ.
45(5), 731–742 (2009)
172. Kurzhanski, A.B., Tochilin, P.A.: On the problem of control synthesis under uncertainty from
the outputs of finite observers. Differ. Equ. 47(11), 1–9 (2011)
173. Kurzhanski, A.B., Vályi, I.: Ellipsoidal calculus for estimation and feedback control. In:
Systems and Control in the Twenty-First Century, pp. 229–243. Birkhäuser, Boston (1997)
174. Kurzhanski, A.B., Vályi, I.: Ellipsoidal Calculus for Estimation and Control. SCFA.
Birkhäuser, Boston (1997)
175. Kurzhanski, A.B., Varaiya, P.: Ellipsoidal techniques for reachability analysis. In: Proceed-
ings of the Pittsburg Conference “Hybrid Systems-2000”. Lecture Notes in Control Science,
vol. 1790, pp. 202–214. Springer, Berlin (2000)
176. Kurzhanski, A.B., Varaiya, P.: On the reachability problem under persistent disturbances.
Dokl. Math. 61(3), 3809–3814 (2000)
438 References
177. Kurzhanski, A.B., Varaiya, P.: Ellipsoidal techniques for reachability analysis. Internal
approximation. Syst. Control Lett. 41, 201–211 (2000)
178. Kurzhanski, A.B., Varaiya, P.: Dynamic optimization for reachability problems. J. Optim.
Theory Appl. 108(2), 227–251 (2001)
179. Kurzhanski, A.B., Varaiya, P.: Reachability under state constraints - the ellipsoidal technique.
In: Proceedings of the 15-th IFAC Congress, Barcelona (2002)
180. Kurzhanski, A.B., Varaiya, P.: Reachability analysis for uncertain systems—the ellipsoidal
technique. Dyn. Continuous Dis. Impulsive Syst. Ser. B 9(3), 347–367 (2002)
181. Kurzhanski, A.B., Varaiya, P.: Ellipsoidal techniques for reachability analysis. Part I: external
approximations. Optim. Methods Softw. 17(2), 177–206 (2002)
182. Kurzhanski, A.B., Varaiya, P.: Ellipsoidal techniques for reachability analysis. Part II: internal
approximations. Box-valued constraints. Optim. Methods Softw. 17(2), 207–237 (2002)
183. Kurzhanski, A.B., Varaiya, P.: On reachability under uncertainty. SIAM J. Control Optim.
41(1), 181–216 (2002)
184. Kurzhanski, A.B., Varaiya, P.: On some nonstandard dynamic programming problems of
control theory. In: Giannessi, F., Maugeri, A. (eds.) Variational Methods and Applications,
pp. 613–627. Kluwer, New York (2004)
185. Kurzhanski, A.B., Varaiya, P.: Ellipsoidal techniques for reachability under state constraints.
SIAM J. Control Optim. 45(4), 1369–1394 (2006)
186. Kurzhanski, A.B., Varaiya, P.: The Hamilton–Jacobi equations for nonlinear target control
and their approximation. In: Analysis and Design of Nonlinear Control Systems (in Honor of
Alberto Isidori), pp. 77–90. Springer, Berlin (2007)
187. Kurzhanski, A.B., Varaiya, P.: On synthesizing team target controls under obstacles and
collision avoidance. J. Franklin Inst. 347(1), 130–145 (2010)
188. Kurzhanski, A.B., Varaiya, P.: Impulsive inputs for feedback control and hybrid system
modeling. In: Sivasundaram, S., Vasundhara Devi, J., Udwadia, F.E., Lasiecka, I. (eds.)
Advances in Dynamics and Control: Theory, Methods and Applications. Cambridge Scientific
Publications, Cottenham (2009)
189. Kurzhanski, A.B., Varaiya, P.: Output feedback control under set-membership uncertainty.
J. Optim. Theory Appl. 151(1), 11–32 (2011)
190. Kvasnica, M., Grieder, P., Baotic, M., Morari, M.: Multiparametric toolbox (MPT). In: Hybrid
Systems: Computation and Control, pp. 448–462. Springer, Berlin (2004)
191. Kwakernaak, H., Sivan, R.: Linear Optimal Control Systems. Wiley-Interscience, New York
(1972)
192. Lancaster, P.: Theory of Matrices. Academic, New York (1969)
193. Lattès, R., Lions, J.-L.: Méthode de quasi-réversibilité et applications. Travaux et recherches
mathématiques, vol. 15. Dunod, Paris (1967)
194. Laurent, P.J.: Approximation and Optimization. Hermann, Paris (1972)
195. Lee, E.B., Markus, L.: Foundations of Optimal Control Theory. Wiley, New York (1967)
196. Lefschetz, S.: Differential Equations: Geometric Theory. Dover, Mineola (1977)
197. Le Guernic, C., Girard, A.: Nonlinear Anal. Hybrid Syst. 4(2), 250–262 (2010)
198. Leitmann, G.: Optimality and reachability with feedback controls. In: Blaquière, A., Leit-
mann, G. (eds.) Dynamical Systems and Microphysics: Control Theory and Mechanics.
Academic, Orlando (1982)
199. Lempio, F., Veliov, V.: Discrete approximations of differential inclusions. Bayreuther Mathe-
matische Schriften 54, 149–232 (1998)
200. Lions, P.-L.: Viscosity solutions and optimal control. In: Proceedings of ICIAM 91,
pp. 182–195. SIAM, Washington (1991)
201. Lions, P.-L., Souganidis, P.E.: Differential games, optimal control and directional derivatives
of viscosity solutions of Bellman’s and Isaac’s equations. SIAM J. Control Optim. 23,
566–583 (1995)
202. Lyapunov, A.A: On fully additive vector functions. Izvestia USSR Acad. Sci., Ser. Math.,
4(6), 465–478 (1940)
References 439
203. Lygeros, J.: On the relation of reachability to minimum cost optimal control. In: Proceedings
of the 41st IEEE CDC, Las Vegas, pp. 1910–1915 (2002)
204. Lygeros, J., Tomlin, C., Sastry, S.: Controllers for reachability specifications for hybrid
systems. Automatica 35(3), 349–370 (1999)
205. Lygeros, J., Quincampoix, M., Rzezuchowski, T.: Impulse differential inclusions driven by
discrete measures. In: Hybrid Systems: Computation and Control. Lecture Notes in Computer
Science, vol. 4416, pp. 385–398. Springer, Berlin (2007)
206. Margellos, K., Lygeros, J.: Hamilton-Jacobi formulation for reach-avoid differential games.
IEEE Trans. Automat. Control 56(8), 1849–1861 (2011)
207. Mayne, D.Q., Rawlings, J.B., Rao, C.V., Scokaert, C.O.M.: Constrained model predictive
control: stability and optimality. Automatica 36, 789–814 (2000)
208. Mayne, D.Q., Seron, M.M., Rakovic, S.V.: Robust model predictive control of constrained
linear systems with bounded disturbances. Automatica 41, 219–224 (2005)
209. Mazurenko, S.S.: A differential equation for the gauge function of the star-shaped attainability
set of a differential inclusion. Doklady Math., Moscow 86(1), 476–479 (2012) [original
Russian Text published in Doklady Akademii Nauk 445(2), 139–142 (2012)]
210. Milanese, M., Norton, J., Piet-Lahanier, H., Walter, E. (eds.): Bounding Approach to System
Identification. Plenum Press, London (1996)
211. Milanese, M., Tempo, R.: Optimal algorithms theory for robust estimation and prediction.
IEEE Trans. Automat. Control 30(8), 730–738 (1985)
212. Miller, B.M., Rubinovich, E.Ya.: Impulsive Control in Continuous and Discrete-Continuous
Systems. Kluwer, New York (2003)
213. Mitchell, I.M.: Application of level set methods to control and reachability problems in
continuous and hybrid systems. Ph.D. thesis, Stanford University (2002)
214. Mitchell, I.M., Tomlin, C.J.: Overapproximating reachable sets by Hamilton-Jacobi projec-
tions. J. Sci. Comput. 19(1), 323–346 (2003)
215. Mitchell, I.M., Templeton, J.A.: A toolbox of Hamilton-Jacobi solvers for analysis of
nondterministic in continuous and hybrid systems. In: Hybrid Systems: Computation and
Control, pp. 480–494. Springer, Berlin (2005)
216. Morari, M., Thiele, L. (eds.): Hybrid Systems: Computation and Control. Lecture Notes in
Computer Sciences, vol. 3414. Springer, Berlin/Heidelberg (2005)
217. Moreau, J.J.: Fonctionelles Convexes. College de France, Paris (1966)
218. Motta, M., Rampazzo, F.: Dynamic programming for nonlinear system driven by ordinary
and impulsive control. SIAM J. Control Optim. 34(1), 199–225 (1996)
219. Neustadt, L.W.: Optimization, a moment problem and nonlinear programming. SIAM J.
Control Optim. 2(1), 33–53 (1964)
220. Nikolski, M.S.: On the approximation of a reachability set of a differential inclusion. Herald
of the Moscow Univ. (Vestnik MGU), Ser. Comput. Math. Cybern. 4, 31–34 (1987)
221. Osher, S., Fedkiw, R.: Level Set Methods and Dynamic Implicit Surfaces. Applied Mathe-
matical Sciences, vol. 153. Springer, New York (2003)
222. Osipov, Yu.S.: On the theory of differential games in distributed parameter systems. Sov.
Math. Dokl. 223(6), 1314–1317 (1975)
223. Osipov, Yu.S., Kryazhimski, A.V.: Inverse Problems for Ordinary Differential Equations:
Dynamical Solutions. Gordon and Breach, London (1995)
224. Panasyuk, V.I., Panasyuk, V.I.: On one equation generated by a differential inclusion . Math.
Notes 27(3), 429–437 (1980)
225. Patsko, V.S., Pyatko, S.G., Fedotov, A.A.: Three-dimensional reachability set for a nonlinear
control system. J. Comput. Syst. Sci. Int. 42(3), 320–328 (2003)
226. Pontryagin, L.S., Boltyansky, V.G., Gamkrelidze, P.V., Mischenko, E.F.: The Mathematical
Theory of Optimal Processes. Wiley Interscience, New York (1962)
227. Pschenichniy, B.N.: Convex Analysis and Extremal Problems. Nauka, Moscow (1980)
228. Puri, A., Varaiya, P.: Decidability of hybrid systems with rectangular differential inclusions.
In: Proceedings of the 6th Workshop on Computer-Aided Verification. Lecture Notes in
Control Science, vol. 818, pp. 95–104. Springer, New York (1994)
440 References
229. Puri, A., Borkar, V.S., Varaiya, P.: ©-Approximation of differential inclusions. In:
Hybrid Systems. Lecture Notes in Control Science, vol. 1066, pp. 362–376. Springer,
New York (1996)
230. Rakovic, S.R., Kerrigan, E., Mayne, D., Lygeros, J.: Reachability analysis of discrete-time
systems with disturbances. IEEE Trans. Automat. Control 51, 546–561 (2006)
231. Rakovic, S.R., Kouvritakis, B., Cannon, M., Panos, Ch., Findeisen, R.: Parametrized tube
model predictive control. IEEE Trans. Automat. Control 57, 2746–2761 (2012)
232. Rakovic, S.R., Kouvritakis, B., Findeisen, R., Cannon, M.: Homothetic model predictive
control. Automatica 48, 1631–1637 (2012)
233. Rakovic, S.R., Kouvritakis, B., Cannon, M.: Equi-normalization and exact scaling dynamics
in homothetic tube control. Syst. Control Lett. 62, 209–217 (2013)
234. Riesz, F., Sz-.Nagy, B.: Leçons d’analyse fonctionnelle. Akadémiai Kiadó, Budapest (1972)
235. Rockafellar, R.T.: Level sets and continuity of conjugate convex functions. Trans. Am. Math.
Soc. 123, 46–63 (1966)
236. Rockafellar, R.T.: State constraints in convex problems of Bolza. SIAM J. Control Optim
10(4), 691–715 (1972)
237. Rockafellar, R.T.: Convex Analysis, 2nd edn. Princeton University Press, Princeton (1999)
238. Rockafellar, R.T., Wets, R.J.: Variational Analysis. Springer, Berlin (2005)
239. Roxin, E.: On generalized dynamical systems defined by contingent equations. J. Differ. Equ.
1(2), 188–205 (1965)
240. Saint-Pierre, P.: Approximation of the viability kernel. Appl. Math. Optim. 29(2), 187–209
(1994)
241. Schlaepfer, F.M., Schweppe, F.C.: Continuous-time state estimation under disturbances
bounded by convex sets. IEEE Trans. Automat. Control 17(2), 197–205 (1972)
242. Schwartz, L.: Théorie des Distributions. Hermann, Paris (1950)
243. Schweppe, F.C.: Uncertain Dynamic Systems. Prentice Hall, Englewood Cliffs (1973)
244. Sethian, J.A.: Level Set Methods and Fast Marching Methods. Cambridge University Press,
Cambridge (1999)
245. Skelton, R.E.: Dynamic Systems and Control. Wiley, New York (1988)
246. Sorensen, A.J.: A survey of dynamic positioning control systems. Ann. Rev. Control 35,
123–136 (2011)
247. Subbotin, A.I.: Generalized Solutions of First-Order PDE’s. The Dynamic Optimization
Perspective. SCFA. Birkhäuser, Boston (1995)
248. Subbotina, N.N., Kolpakova, E.A., Tokmantsev, T.B., Shagalova, L.G.: Method of Character-
istics for the Hamilton-Jacobi-Bellman Equation. Ural Scientific Center/Russian Academy of
Sciences, Yekaterinburg (2013)
249. Summers, S., Kamgarpour, M., Tomlin, C.J., Lygeros, J.: A stochastic reach-avoid problem
with random obstacles. In: Belta, C., Ivancic, F. (eds.) Hybrid Systems: Computation and
Control. ACM, New York (2013)
250. Summers, S., Kamgarpour, M., Tomlin, C.J., Lygeros, J.: Stochastic hybrid system con-
troller synthesis for reachability specifications encoded by random sets. Automatica 49(9),
2906–2910 (2013)
251. Tikhonov, A.N., Arsenin, V.Y.: Solutions of Ill-Posed Prolems. Winston, New York (1975)
252. Tikhonov, A.N., Goncharsky, A.V.: Ill-Posed Problems in the Natural Sciences. Oxford
University Press, Oxford (1987)
253. Tolstonogov, A.A.: Differential Inclusions ina Banach Space. Kluwer Academic, Boston
(2000)
254. Tomlin, C.J., Lygeros, J., Sastry, S.: A game theoretic approach to controller design for hybrid
systems. Proceedings of IEEE 88, 949–969 (2000)
255. Ushakov, V.N.: Construction of solutions in differential games of pursuit-evasion. Lecture
Notes in Nonlinear Analysis, vol. 2(1), pp. 269–281. Polish Academic Publishers, Warsaw
(1997)
256. Usoro, P.B., Schweppe, F., Wormley, D., Gould, L.: Ellipsoidal set-theoretic control synthesis.
J. Dyn. Syst. Meas. Control 104, 331–336 (1982)
References 441
257. Ustyuzhanin, A.M.: On the problem of matrix parameter identification. Probl. Control Inf.
Theory 15(4), 265–273 (1986)
258. Van der Schaft, A., Schumacher, H.: An Introduction to Hybrid Dynamical Systems. Lecture
Notes in Control and Information Sciences, vol. 25. Springer, Berlin (2000)
259. Varaiya, P.: Reach set computation using optimal control. In: Proceedings of the KIT
Workshop on Verification of Hybrid Systems, Grenoble, October 1998
260. Varaiya, P., Lin, J.: Existence of saddle points in differential games. SIAM J. Control Optim.
7(1), 142–157 (1969)
261. Vazhentsev, A.Yu.: On internal ellipsoidal approximations for problems of feedback control
under state constraints. Proc. Russ. Acad. Sci. Theory Syst. Control 3, 70–77 (2000)
262. Veliov, V.: Second order discrete approximations to strongly convex differential inclusions.
Syst. Control Lett. 13(3), 263–269 (1989)
263. Veliov,V.: Stability-like properties of differential inclusions. Set Valued Anal. 5(1), 73–88
(1997)
264. Vidal, R., Schaffert, S., Lygeros, J., Sastry, S.: Controlled invariance of discrete time systems.
In: Hybrid Systems: Computation and Control. Lecture Notes in Computer Science, vol. 1790,
pp. 437–450. Springer, Berlin (2000)
265. Vinter, R.B.: A characterization of the reachable set for nonlinear control systems. SIAM J.
Control Optim. 18(6), 599–610 (1980)
266. Vinter, R.B., Wolenski, P.: Hamilton–Jacobi theory for optimal control problems with data
measurable in time. SIAM J. Control Optim. 28(6), 1404–1419 (1990)
267. Vinter, R.B.: Optimal Control. Birkhäuser, Boston (2010)
268. Vostrikov, I.V., Dar’in, A.N., Kurzhanskii, A.B.: On the damping of a ladder-type vibration
system subjected to uncertain perturbations. Differ. Equ. 42(11), 1524–1535 (2006)
269. Whittaker, E.T.: Analytical Dynamics. Dover, New York (1944)
270. Willemsm, J.C. (ed.): From Data to Model. Springer, New York (1989)
271. Witsenhausen, H.: Set of possible states of linear systems given perturbed observations. IEEE
Trans. Automat. Control 13(1), 5–21 (1968)
272. Wolenski, P.: The exponential formula for the reachable set of a lipschitz differential
inclusion. SIAM J. Control Optim. 28(5), 1148–1161 (1990)
273. Wonham, W.M.: Linear Multivariable Control: A Geometrical Approach. Springer, Berlin
(1985)
274. Zadeh, l.A., Desoer, C.A.: Linear System Theory: The State Space Approach. Dover,
New York (2008)
275. Zakroischikov, V.N.: Reachability sets of hybrid systems under sequential switchings. Herald
MSU 2, 37–44 (2009)
Index
C
Closed loop target control, 75–78, 167, 371 D
Colliding tubes, 70–75, 330 Differential inclusion, 10, 19, 23, 60, 61, 63,
Comparison principle 76, 88, 89, 115, 132, 268, 270, 280,
convex reach sets, 198 290, 312, 313, 315–317, 322, 324,
nonconvex reach sets, 208–213 326, 331, 339, 344, 365, 367, 369,
under state constraints, 281–284 370, 375, 378, 385
Constraints (bounds) Disturbance, 2, 9, 25, 63, 196–198, 341–343,
ellipsoidal, quadratic, 128, 369, 388 345, 352, 358, 364, 371–373,
joint (double), 7 392–394
magnitude, integral, 5–8 Dynamic programming, 46–86, 253, 260, 371,
Control 375, 377, 383, 424
closed-loop, 1, 5–10, 46, 52, 53, 54, 58, 63, Dynamic programming equation, 48–54
75, 78, 167, 200, 255, 322, 326–331,
338–339, 373
feedback, 9, 47, 52, 128, 161, 170, 184,
195–196, 260, 269, 270, 272, 275, E
284, 311–339, 371–394, 420, Ellipsoidal approximations
423 box-valued constraints, 125, 229–231,
feedforward, model predictive, 46 235–241, 285
minimal magnitude, 12–16 ellipsoidal constraints, 88–91, 202, 388,
norm-minimal, minimal energy, 10 398
open-loop, 1, 8–10, 24–40, 52, 63, 253, external, tight, 96, 98
261, 402, 403 internal, tight, 120, 131, 161, 246
optimal, linear, 1–46 zonotope constraints, 241–251, 388
F
Function L
conjugate (Fenchel), 28, 68, 261, 268 Legendre transformation, 29
convex, 29, 38, 39, 66, 68, 81, 216, 254, Lipschitz condition, 312, 378
396
second conjugate (Fenchel), 81
strictly convex, 38, 39 M
support, 19, 20, 22–24, 39, 47, 64, 65, Matching conditions, 196, 393
69, 80, 89, 90, 111, 119, 120, Maximum condition, 14, 24, 28–31, 34, 36, 44,
129, 144, 150, 157, 158, 185, 188, 287, 294
189, 191, 202, 241, 242, 280, 285, Maximum principle, 1, 14, 24–40, 44, 45, 72,
287, 289–292, 315, 319, 320, 330, 74, 90, 99, 108, 109, 245, 256, 287,
363–365, 410 289–294, 308, 325
ellipsoidal, 107–110, 251
Measurement equation, 312, 341, 345, 346,
G 348, 373
Generalized state, 372–375, 377, 378 Metric space, 281, 315, 381, 383
Gram matrix, 11, 15, 16 Minimum condition, 28, 35, 44
Guaranteed state estimation, 341–370, 372, Moment problem, 10, 12, 17–18, 255
377–381
O
H Obstacle problem
Hausdorff distance, 61, 103, 105, 281, 319 closed-loop control, 338–339
Hausdorff semidistance, 61, 103, 280, 281, 319 complementary convex constraint, 336
Hybrid system reach-evasion set, 335–336
branching trajectory tubes, 395 Optimization problem
ellipsoidal methods, 398–401, 418–419 linear-quadratic, 46, 54–57, 363, 367, 368
enabling zones, resets, 402, 414–415 primal, dual, 40–46
impulsive feedback control, 420, 423 Output feedback control
reachability, 395, 396, 401–416 ellipsoidal approximations, 388–392
value functions, 396, 397, 399, 407–416, linear-convex systems, 372, 379, 381,
418, 427 386–388
overall solution, 323, 372, 375–377
I
Impulse control P
approximating motion, 269–273 Parallel computation, 127, 195–196
closed loop, 260–266 Polyhedral techniques, output feedback
delta function, number of impulses, 7, 269, control, 373
421 Principle of optimality, 53, 59, 62, 263, 277,
HJB variational inequality, 260–266 278, 327
Index 445