Linear Systems
Linear Systems
by Sanand D
“Sometimes the truth isn’t good enough, sometimes people need more...” -Batman
————————————————————————————————————————-
1 Introduction
The various types of systems are
1. Linear/Non-linear
2. Continuous/Discrete
3. Time invariant/Time varying
4. Deterministic/Stochastic
5. Lumped/Distributed
6. Finite/Infinite dimensional
7. Autonomous/Non-autonomous
8. Dissipative/Lossless.
In here, we study linear systems which one must understand before studying more general non-
linear systems and complex systems occuring in many real life applications. Some examples of
linear systems are electrical circuits with linear elements (e.g., RLC circuits), spring-mass-damper
systems, linear filters and so on. One can study non-linear systems locally around an equilibrium
point by local linearization. A system is said to be linear if the superposition and the homogeneity
principles hold.
Study of systems mainly involve: modelling, analysis and synthesis and design. Modelling is
done using first principles i.e., using Euler-Lagrange equations, Kirchoff/Newton’s laws, conserva-
tion laws (e.g., Maxwell’s equation, heat equation, wave equation) etc. We consider state-space
models which can be obtained from n−th order odes by introducing auxiliary variables. The number
of states is given by the number of independent energy storing elements in the system. Sometimes,
linear systems can also be represented in the input-output form using a convolution operator e.g.,
signal processing systems in communication engineering. The other form of modeling is known as
system identification where one tries to estimate the system equations by probing the system with
known input signals and observing the corresponding output. Modeling allows us to represent a
physical system using mathematical expression and allows us to determine the system properties
using mathematical concepts. It allows us to predict the system response to various input signals.
Analysis and synthesis involves studying intrinsic system properties such as controllability,
observability, stabilizability, controllable/unobservable subspaces, controllability indices, poles and
zeros etc. Designing control systems involve: control, estimation and optimization. We consider
continuous time systems of the form
ẋ = Ax + Bu
y = Cx + Du (1)
and discrete time systems of the form
x(t + 1) = Ax(t) + Bu(t)
y(t) = Cx(t) + Du(t) (2)
where x ∈ Rn , u ∈ Rm , y ∈ Rp , A ∈ Rn×n , B ∈ Rn×m , C ∈ Rp×n , D ∈ Rp×m . In other words, these
are the models for systems under consideration, if A, B, C, D are time dependent, then the system
is called linear time varying system (LTV) whereas, if they are time independent, then it is called
linear time invariant (LTI) system. These are called non-autonomous systems i.e. systems with
inputs. The simplest example of LTI systems ´ t is given by an integrator system where the output
is the integration of the input i.e., y = 0 udt. Let x = y, then ẋ = u. This is a scalar LTI
system
´ t ´where A = 0, B = 1, C = 1 and D = 0. One can also consider a double integrator where
s
y = 0 0 udsdt. Let x1 = y and x2 = ẋ1 , then ẋ1 = x2 and ẋ2 = u. (Find A, B, C, D.) Autonomous
systems are of the for ẋ = Ax where there is no input (e.g., LC circuit with no current or voltage
sources). The advantages of state space models is that both LTI and LTV can be studied. For
LTI systems and signal processing system, transfer function approach is also applicable (using the
Laplace and Fourier transforms). State space models also allows us to study the internal dynamics
of a system.
In general, one can consider the following questions for linear systems
1. How to drive the state from an initial condition to the desired state, how to choose a control
input?
3. What is stabilizability? How to check stabilizability and how to make a system stable?
4. What is the optimal control input for a state transfer, e.g., least energy input?
We need linear algebra (refer short notes on linear algebra) and linear odes to address the above
questions.
2 Linear ODEs
Consider a linear ode p(D)y = u where p(D) is a polynomial involving differential operator D.
Consider a homogeneous ode p(D)y = 0 and let yh be a solution of this homogeneous ode. Let
yp be a particular solution of the ode p(D)yp = u (where yp depends on u). A general solution
can be written as y = yh + yp . This is true for both time varying and time invariant odes as long
as they are linear. This not only holds for scalar odes above but also holds for vector odes. By
using additional variables, scalar odes can be converted to vector odes which are in a first order
form (i.e. only first order derivatives are involved). Roots of the scalar ode and eigenvalues of the
corresponding matrix in the first order form are the same. (This can be shown using ideas from
Quotient spaces from linear algebra and the proof is given in the linear algebra notes. Moreover, the
matrix obtained in the first order form is a companion matrix. Thus, its characteristic polynomial
is same as the characteristic equation for ode again from linear algebra.)
For linear time invariant odes, one can also apply Laplace transform methods to find a solution.
One can write down a solution of linear odes explicitly. This is called as a closed form solution
which is the sum of the two terms as explained above. By the existence and uniqueness theorem
for odes, a unique solution always exists. The terms such as a solution of an ode, a trajectory, an
integral curve are all synonymous. We assume that the reader is familiar with definitions of linear
ode, impulse response, causality, LTI systems and transfer functions. Let u(t) = δ(t) be an input
to a system = Ax + Bu. The homogeneous solution is xh = eAt x(0) and a particular solution
´ tẋ A(t−τ
is xp (t) = 0 e ) Bδ(τ )dτ = eAt B. Thus x(t) = eAt x(0) + eAt B. Suppose u(t) = u ejωt . Then,
0
xp (t) = xm e . Substituting xp (t) in the system equation, xm = (jωI − A)−1 Bu0 and x(t) =
jωt
eAt x(0) + (jωI − A)−1 Bu0 ejωt . The first term represents the transient response and the second
term represents the steady state response. Let y(t) = Cx(t) be the output of the system. Hence,
the impulse response is CeAt B and the frequency response is C(jωI − A)−1 Bu0 ejωt (Kwakernaak
and Sivan).
As far as the existence and uniqueness of solutions of an ode is concerned we refer the reader
to any standard textbook on ode. It turns out that for the linear odes under consideration, there
exists a unique solution which can be written in a closed form using the state transition matrix as
shown in the next section.
Definition 3.1. The state transition matrix for a linear homogeneous system (or autonomous
system i.e. a system without any inputs) ẋ = A(t)x is defined as
ˆ t ˆ t ˆ s1
Φ(t, t0 ) := I + A(s1 )ds1 + A(s1 ) A(s2 )ds2 ds1 + . . . . (3)
t0 t0 t0
For LTI systems, A(t) = A and therefore from the definition of the state transition matrix (STM),
Φ(t, t0 ) = eA(t−t0 ) . Thus, by Cayley-Hamilton theorem, we can express eA(t−t0 ) as a linear combi-
nation of I, A, . . . , An−1 with time dependent coefficients. Note that AeAt = eAt A.
For non-homogeneous linear ODEs,
ˆ t
x(t) := Φ(t, t0 )x(t0 ) + Φ(t, τ )B(τ )u(τ )dτ. (4)
t0
ˆ t
⇒ ẋ(t) = A(t)Φ(t, t0 )x(t0 ) + A(t) Φ(t, τ )B(τ )u(τ )dτ + Φ(t, t)B(t)u(t)
t0
= A(t)x(t) + B(t)u(t).
Thus, the closed form expression for x(t) satisfies the ODE hence, it is a solution.
´ t Observe that a so-
lution is a sum of a homogeneous response Φ(t, t0 )x(t0 ) and a forced response t0 Φ(t, τ )B(τ )u(τ )dτ .
sn−1 Ik×k
Z1
Z2 sn−2 Ik×k
1
. =
d(s) . .
. .
Zn Ik×k
Now, CZ(s) = C(sI − A)−1 B = Gsp (s). It can be verified that CZ(s) gives (5) giving the desired
transfer matrix. This realization is called controller canonical realization of G(s).
Now suppose
−α1 Ik×k Ik×k 0 ··· 0
N1
−α2 Ik×k 0 0 ··· 0
N2
. . . ··· .
A= , B = . , C = Ik×k 0 ··· 0 0 . (9)
. . . ··· .
.
−αn−1 Ik×k 0 0 ··· Ik×k
Nn
−αn Ik×k 0 0 ··· 0
One can verify that this also realizes Gsp (s). This is called observer canonical realization.
where q is a generalized position vector, M (q) > 0 is the mass matrix, B(q, q̇) represents cen-
trifugal/Coriolis/friction matrix, and G(q) is a vector due to conservative
forces. Using F =
q
B(q, q̇)q̇ + G(q) + M (q)v, one obtains M (q)q̈ = M (q)v. Now using x = , one gets
q̇
0 I 0
ẋ = x+ v, y = 0 I x.
0 0 I
Using v = −KP q − KD q̇ = − KP KD x, one obtains
0 I
ẋ = x, y = 0 I x.
−KP −KD
Systems in strict feedback form: Consider a system in the following form called strict-feedback
form
ẋ1 = f1 (x1 ) + x2
ẋ2 = f2 (x1 , x2 ) + u. (16)
In this case, one cannot cancel the nonlinearity f1 (x1 ) using u. Consider a change of variable
z2 := f1 (x1 ) + x2 . Thus,
ẋ1 = z2
∂f ∂f (x1 )
ż2 = ẋ1 + ẋ2 = (f1 (x1 ) + x2 ) + f2 (x1 , x2 ) + u. (17)
∂x1 ∂x1
Now, we can cancel the nonlinearity using u and linearize the system.
6 Stability
We now study stability properties of linear systems in absence and presence of inputs. First we
study the internal stability of a system in the absence of inputs. Then, we consider bounded input
bounded output stability in the presence of inputs. This discussion is borrowed from Hespanha
([1]).
• Exponentially stable if in addition to stability, there exists c, λ > 0 such that kx(t)k ≤
ce−λ(t−t0 ) kx(t0 )k.
Theorem 6.2. Let ẋ = Ax. Then, the following statements are equivalent.
1. All eigenvalues of A lie strictly in the LHP.
4. For every positive definite matrix Q, there exists a unique positive definite matrix P such that
AT P + P A = −Q (Lyapunov equation).
Note that by definition, keA(t−t0 ) k = max{y(t),ky(t)k=1} keA(t−t0 ) y(t)k. Let ŷ(t) be such that
keA(t−t0 ) k = keA(t−t0 ) ŷ(t)k. For simplicity, assume that A has only one eigenvalue λ lying in the
left half plane. Therefore,
(t−t0 )2 λ(t−t0 ) (t−t0 )n−2 λ(t−t0 ) (t−t0 )n−1 λ(t−t0 )
ŷ1 (t)
eλ(t−t0 ) (t − t0 )eλ(t−t0 ) 2! e ... (n−2)! e (n−1)! e
(t−t0 )n−2 λ(t−t0 ) ŷ2 (t)
eλ(t−t0 ) (t − t0 )eλ(t−t0 )
0 ... 0 (n−2)! e
.
A(t−t0 ) . . . ... . .
e ŷ(t) = .
. . . ... . .
.
eλ(t−t0 ) (t − t0 )eλ(t−t0 )
0 0 0 ... ŷn−1 (t)
0 0 0 ... 0 eλ(t−t0 ) ŷn (t)
keA(t−t0 ) ŷ(t)k = kŷ1 (t)v1 + . . . + ŷn (t)vn k ≤ |ŷ1 (t)|kv1 (t)k + . . . + |ŷn (t)|kvn (t)k (19)
Note that kŷ(t)k = 1, therefore |ŷi (t)| ≤ 1 for 1 ≤ i ≤ n. Hence the previous equation reduces to
the following
Observe that |g(t − t0 )eλ(t−t0 ) | is bounded and goes to zero as t → ∞. Since g(t − t0 )eλ(t−t0 )
is decaying to zero, we can choose γ > 0 such that |g(t − t0 )eλ(t−t0 ) | ≤ e−γ(t−t0 ) for t ≥ t1
for some t1 ∈ R. Let c1 = supt≥0 |g(t − t0 )eλ(t−t0 ) |. Thus by choosing an appropriate c ∈ R,
|g(t − t0 )eλ(t−t0 ) | ≤ ce−γ(t−t0 ) for all t ≥ 0.
Hence, kx(t)k ≤ ce−γ(t−t0 ) kx(t0 )k. The general case for multiple Jordan blocks follows similarly.
´∞ T
Suppose the system has all eigenvalues strictly in the LHP. Let P = 0 eA t QeAt dt. Observe
that ˆ ∞ ˆ ∞
T T AT t At T
A P + PA = A e Qe dt + eA t QeAt Adt.
0 0
But d AT t QeAt ) T T
= AT eA t QeAt + eA t QeAt A. Therefore,
dt (e
ˆ ∞
T d AT t At T
A P + PA = (e Qe )dt = limt→∞ (eA t QeAt ) − Q = −Q.
0 dt
Thus, P satisfies the Lyapunov equation. Observe that P = P T and P > 0 since Q > 0. To
show that P is unique, let P̄ be another solution. Therefore, AT P + P A = AT P̄ + P̄ A ⇒ AT (P −
T T d T
P̄ ) + (P − P̄ )A = 0 ⇒ eA t AT (P − P̄ )eAt + eA t (P − P̄ )AeAt = 0 ⇒ dt (eA t (P − P̄ )eAt ) = 0.
T
Thus, eA t (P − P̄ )eAt remains constant with time, but it goes to zero as t → ∞. Therefore,
T
eA t (P − P̄ )eAt = 0 ⇒ P = P̄ . Hence, 1 ⇒ 4. Now 4 ⇒ 5 is trivial.
Finally, we show that 5 ⇒ 3. Let Q := −(AT P +P A) and v(t) = xT P x > 0. Now v̇ = xT (AT P +
P A)x = −xT Qx < 0. Thus, v is a decreasing signal and v(t) = xT (t)P x(t) < v(0) = xT (0)P x(0).
v(t)
v(t) = xT (t)P x(t) ≥ λmin (P )kx(t)k2 ⇒ kx(t)k2 ≤ (21)
λmin (P )
λmin (Q)
v̇ = −xT (t)Qx(t) ≤ −λmin (Q)kx(t)k2 ≤ − v(t). (22)
λmin (P )
Lemma 6.3 (comparison lemma). Let v(t) be a scalar signal such that v̇ ≤ µv(t), ∀t ≥ t0 , for
some µ ∈ R. Then, v(t) ≤ eµ(t−t0 ) v(t0 ).
Therefore, s is nonincreasing and s(t) = e−µ(t−t0 ) v(t) ≤ s(t0 ) = v(t0 ) which proves the lemma.
λmin (Q)
− (t−t0 )
Now applying the comparison lemma to (22), v(t) ≤ e λmin (P ) v(t0 ). Since v(t) converges
to 0 exponentially, from (21), x(t) also converges to 0 exponentially.
2. Every eigenvalue of A has a real part strictly less than zero or equal to zero and eigenvalues
with zero real part have trivial Jordan structure (i.e., they are semi-simple).
Proof. Note that eAt is bounded iff the condition on the eigenvalues of A mentioned above is
satisfied. Therefore, 1 ⇔ 2.
Thus, to check stability, it is enough to find eigenvalues of the system matrix. From the
characteristic equation of the underlying system, one can find a matrix in companion form having
the same characteristic polynomial as the characteristic equation.
Remark 6.5. For stability/asymptotic stability of time varying systems, we must have Φ(t, t0 )
bounded for all t ≥ t0 or Φ(t, t0 ) → 0 as t → ∞. For exponential stability, Φ(t, t0 ) must go to zero
exponentially and so on.
Geometry of Lyapunov functions Any positive definite matrix P defines an energy func-
tions on the state space where the energy v(t) = xT P x. This defines ellipsoids which are energy
surfaces on the state space. Consider a set {x ∈ Rn |xT P x = c} where c ∈ R. This is an ellipsoid
which contains all states whose energy v(t) = xT P x = c. Thus, the entire state space is partitioned
as a union of these ellipsoids (energy surfaces). As x(t) evolves as a function of time (say ẋ = f (x)
is the underlying dynamical system), its energy varies accordingly since x(t) is varying as a function
of time. If v̇ = 0, then the rate of change of energy is zero. Thus, the energy of x(t) is constant.
Hence, it does not leave the energy surface (ellipsoid) on which it is already lying. If v̇ > 0, then
the rate of change is positive and energy of the state trajectory is increasing hence, it goes to a
higher energy surface. If v̇ < 0, then the trajectory moves to a lower energy surface. If v̇ ≤ 0 for all
points in the set {x ∈ Rn |xT P x ≤ c}, then once a trajectory enters this set, it can not leave this
set. This is the idea to prove invariance of a set or stability/asymptotic stability related arguments.
Consider a nonlinear autonomous system ẋ = f (x) and its linearization around the equilibrium
point x∗ .
Theorem 6.6 (local stability via linearization). Consider ẋ = f (x) where f is at least twice
differentiable on a closed ball B̄ ⊂ Rn around the equilibrium point x∗ . If the linearized system
∂f (x∗ )
δ ẋ = Aδx, A := ∂x is exponentially stable, then there is exists a ball B ⊂ Rn around x∗ and
constants c, λ > 0 such that for every solution x(t) of the nonlinear system that starts at x(t0 ) ∈ B,
kx(t) − x∗ k ≤ ce−λ(t−t0 ) kx(t0 ) − x∗ k, ∀t ≥ t0 . (23)
Proof. Let r(x) := f (x) − f (x∗ ) − Aδx. By Taylor’s formula, there exists d > 0 such that
kr(x)k ≤ dkδxk2 , ∀x ∈ B̄. (24)
Since, the linearized system is exponentially stable, there exists P > 0 such that AT P + P A = −I.
Let v(t) := δxT P δx.
v̇ = f (x)T P δx + δxT P f (x)
= (Aδx + r(x))T P δx + δxT P (Aδx + r(x))
= δxT (AT P + P A) + 2δxT P r(x)
= −kδxk2 + 2δxT P r(x)
≤ −kδxk2 + 2kP kkδxkkr(x)k. (25)
We want to make the rhs negative. For example,
1
−kδxk2 + 2kP kkδxkkr(x)k ≤ − kδxk2 .
2
To achieve this, let > 0 be small enough such that the ellipsoid
E := {x ∈ Rn | (x − x∗ )T P (x − x∗ ) ≤ } (26)
satisfies the following properties
1. The ellipsoid E is contained in B̄. When x is inside this ellipsoid, (24) holds, hence,
x(t) ∈ E ⇒ v̇ ≤ −kδxk2 + 2dkP kkδxk3 = −(1 − 2dkP kkδxk)kδxk2 .
1 1
2. We further shrink such that inside the ellipsoid E, (1 − 2dkP kkδxk) ≥ 2 ⇔ kδxk ≤ 4dkP k .
For this choice of , x(t) ∈ E ⇒ v̇ ≤ − 12 kδxk2 .
Once x(t) ∈ E, v(t) ≤ , v̇ ≤ 0. Thus, v cannot increase above , hence, it cannot exit E.
Note that
δxT P δx 1 1 δxT P δx 1 v(t)
δxT P δx ≤ kP kkδxk2 ⇒ kδxk2 ≥ ⇒ − kδxk2 ≤ − =− . (27)
kP k 2 2 kP k 2 kP k
v(t)
Now if x(0) ∈ B ⊂ E, by (27) v̇ ≤ − 12 kδxk2 ≤ − 12 kP k . Hence, by the comparison lemma, v and
∗
consequently δx = x − x decrease to zero exponentially.
Theorem 6.7 (unstability via linearization). Consider ẋ = f (x) where f is twice differentiable
(x∗ )
and f (x∗ ) = 0. If the linearized system δ ẋ = Aδx, A := ∂f∂x is unstable, then there are solutions
that start arbitrarily close to x∗ , but do not converge to this point as t → ∞.
6.2 Input-output stability
Note that the forced output of a linear system due to the input u(t) is given by
ˆ t
yf (t) = C(t)Φ(t, τ )B(τ )u(τ )dτ + D(t)u(t).
0
Definition 6.8. A system is said to be uniformly BIBO stable if there exists a finite constant g
such that
supt∈[0,∞] kyf (t)k ≤ gsupt∈[0,∞] ku(t)k (28)
Proof. Sketch: Clearly, 2 ⇒ 1 since, taking sup over the rhs, we can bound the forced response
yf (t). On the other hand, if 2 is not satisfied, then we cannot bound yf (t) over all bounded inputs
as there will be bounded inputs corresponding to unbounded entries in either D(.) or g(t, τ ) which
make the forced response yf (t) unbounded.
Theorem 6.10 (Frequency domain characterization). Following statements are equivalent for LTI
systems.
2. Every pole of every entry of the transfer function of the system has a strictly negative real
part.
Proof. Sketch: The second statement in the time-domain characterization is equivalent to the
statement that every entry Gij (s) of the transfer matrix G(s) corresponds to either an exponentially
decaying term in the time domain.
Note that exponential/asymptotic or internal stability implies BIBO stability but the converse
need not be true.
In this case eAt becomes unbounded, hence the system is not internally stable. But CeAt B = e−2t
which is bounded, hence the system is BIBO stable.
7 Controllability
We now formalize one of the central properties of control system which is controllability. The
other one being observability. We also study an equally important notion of stabilizability and pole
placement. We will see that control problems are basically least norm problems.
Definition 7.1. A system ẋ = A(t)x(t) + B(t)u(t) is said to be controllable if for any initial
condition x(t0 ) ∈ Rn (or Cn ) and for any final condition xf ∈ Rn (or Cn ), there exists an input
u(t) which drives the state from x(t0 ) to xf in finite time t ∈ R. If t0 = 0 and x(0) = 0 in the
above case and xf is arbitrary, then we say that the system is reachable. Moreover, if xf = 0 and
x(t0 ) is arbitrary in the above case, then we say that the system is 0−controllable.
Consider the following equation for linear systems
ˆ t
x(t) = Φ(t, t0 )x(t0 ) + Φ(t, τ )B(τ )u(τ )dτ (30)
t0
ˆ t
⇒ x(t) − Φ(t, t0 )x(t0 ) = Φ(t, τ )B(τ )u(τ )dτ (31)
t0
´t
Thus, the system if controllable ⇔ the range space of t0 Φ(t, τ )B(τ )u(τ )dτ spans the entire state
space. Note that if the system is controllable, then we have a least norm problem (please refer my
linear algebra notes). There could be multiple u(t)s driving the state from the initial to the final
condition and we would like to find the input with least norm (or least energy). It is clear that for
reachability and 0−controllability, we have the same equivalent condition. Hence all these notions
are equivalent for continuous time systems. For discrete time systems, state transition matrix is
not invertible if the system matrix A is non-invertible. For discrete time systems, these notions are
equivalent when A is invertible. If A is not invertible for discrete time systems, then the system
cannot be controllable or reachable. But it can be 0−controllable.
Theorem 7.2. Following statements are equivalent for LTI systems:
1. (A, B) is controllable.
´t T
2. The Controllability Gramian Wt (A, B) := t0 eA(t−τ ) BB T eA (t−τ ) dτ > 0 for some t > t0 .
Proof. Follows
from the rank test of the
Controllability matrix. Suppose the Controllability matrix
C(A, B) = B AB . . . An−1 B has rank n. Let Ā = T AT and B̄ = T −1 B. Consider
−1
Proof. Follows by applying PBH test to (A, B) and (A − BF, B). Assuming that the matrix for
PBH test is rank deficient for one system, one can show that the corresponding matrix for other
system is also rank deficient.
Now we show that for controllable systems, one can do an arbitrary pole placement to have
the desired system poles. Controllability is unaffected by the previous theorem. The following
lemma and the pole placement theorem is taken from Wonham (Linear Multi-variable control: A
geometric approach).
Lemma 7.6. Let 0 6= b ∈ column span of B. If (A, B) is controllable, then there exists a feedback
matrix F such that (A + BF, b) is controllable.
Proof. Let b1 = b. Let X1 be a cyclic subspace generated by b1 under the action of A i.e.
X1 =< b1 , Ab1 , . . . , An1 −1 b1 > where {b1 , Ab1 , . . . , An1 −1 b1 } forms a linearly independent set. Let
x1 = b1 , x2 = Ax1 + b1 , . . . , xn1 = Axn1 −1 + b1 . Clearly, xi s are linearly independent since the
set {b1 , Ab1 , . . . , An1 −1 b1 } is linearly independent. Choose b2 ∈ / X1 and generate a cyclic subspace
X2 =< b2 , Ab2 , . . . , A n2 −1 b2 > where {b2 , Ab2 , . . . , A n2 −1 b2 } forms a linearly independent set. De-
fine xn1 +1 = Axn1 + b2 , xn1 +2 = Axn1 +1 + b2 and so on. Thus, {x1 , . . . , xn1 +n2 } is a basis for
X1 + X2 . Continuing this way we obtain a basis {x1 , . . . , xn } for the entire state space. (Since
(A, B) is controllable, rank of C(A, B) is n. Therefore, the above procedure generates the entire
state space.)
Let xi+1 = Axi + b̂i (1 ≤ i ≤ n − 1). Let b̂i = Bηi . Define a linear map F on the state space
such that F (xi ) = ηi (1 ≤ i ≤ n). Thus, xi+1 = Axi + b̂i = Axi + Bηi = Axi + BF xi = (A + BF )xi .
Therefore, {x1 , (A + BF )x1 , . . . , (A + BF )n−1 x1 } generate the entire state space where x1 = b1 .
Theorem 7.7 (Pole placement). (A, B) is controllable ⇔ for every symmetric set Λ of complex
numbers, there exists a feedback matrix F such that eigenvalues of (A + BF ) are given by Λ.
Proof. (⇒) Suppose we have a single input system and (A, B) is controllable. Therefore, B = b
is a cyclic vector. Suppose pA (s) = sn − an sn−1 − . . . − a2 s − a1 . Let Λ = {λ1 , . . . , λn } and
(s − λ1 ).(s − λ2 ) . . . (s − λn ) = sn − ân sn−1 − . . . − â2 s − â1 the desired characteristic polynomial.
By change of basis (since (A, b) is controllable, span{b, Ab, . . . , An−1 b} = Rn , choose v1 = An−1 b −
an An−2 b − . . . − a2 b, v2 = An−2 b − an An−3 b − . . . − a3 b, . . . , vn−1 = Ab − an b, vn = b as a new
basis), we may assume that (A, b) is in the following companion form
0 1 ... 0 0
0 0 ... 0 .
A= . . ... . ,b = . .
. . ... 1 0
a1 a2 . . . an 1
Choosing f = â1 − a1 â2 − a2 . . . ân − an , A + bf has the desired eigenvalues.
For multi-input systems, from the previous lemma, there exists F1 and b ∈ column span of B
i.e. b = Bη such that (A + BF1 , b) is cyclic. Thus by the single input case above, there exists f¯
such that A + BF1 + bf¯ = A + BF1 + Bη f¯ = A + B(F1 + η f¯) = A + BF has the desired eigenvalues.
(⇐) Choose Λ to be a set of n real and distinct numbers {λ1 , . . . , λn } such that λi is not an
eigenvalue of A (1 ≤ i ≤ n). Choose F such that {λ1 , . . . , λn } are eigenvalues of (A + BF ) and
{x1 , . . . , xn } are the corresponding eigenvectors.
((λi I − A)) is invertible for (1 ≤ i ≤ n) since λi is not an eigenvalue of A. Note that (λi I − A)−1
is an analytic function since λi is not an eigenvalue of A and hence it is a power series expansion.
By Cayley-Hamilton theorem, this power series can be expressed as follows:
n
X
(λi I − A)−1 = pj (λi )Aj−1 (32)
j=1
Pn
Thus, xi = (λi I − A)−1 BF xi = j=1 pj (λi )Aj−1 BF xi . Therefore, the basis vectors xi ∈ column
span of C(A, B) = B AB . . . An−1 B which implies that rank of C(A, B) is n. Hence,
(A, B) is controllable.
λ 0 1 0
Example 7.8 (MIMO pole placement). Let A = ,B = . Following the proof
0 λ 0 1
of Lemma 7.6, X1 =< b1 > and X2 =< b2 >. X = X1 ⊕ X2 . Moreover, a basis for X is
x1 = b1 , x2 = Ax1 + b2 . Note that in the notation of Lemma 7.6, x2 = Ax1 + b̂1 = Ax1 + Bη1 where
b̂1 = b2 and b2 = Be2 . Hence, η1 = e2 and F x1 = e2 . Let F x2 = 0. Thus,
λ 0 1 0 0 0 λ 0
A + BF = + = . (33)
0 λ 0 1 1 0 1 λ
Clearly, (A + BF, b1 ) is cyclic. Now we can do an pole placement using f1 such that A + BF + b1 f1
has the desired characteristic polynomial. (A + BF + b1 f1 = A + BF + Be1 f1 = A + B(F + e1 f1 ) =
A + BK).
λ 1 0 0 0 0
0 λ 0 0 1 0
Example 7.9 (MIMO pole placement). Let A = 0 0 λ 1 and B = 0 0 . Suppose
0 0 0 λ 0 1
we want to place poles at arbitrary locations. Following the proof of Lemma 7.6, X1 =< b1 , Ab1 >
and X2 =< b2 , Ab2 >. X = X1 ⊕ X2 . Moreover, a basis for X is x1 = b1 , x2 = Ax1 + b1 , x3 =
Ax2 + b2 , x4 = Ax3 + b2 . Note that xi+1 = Axi + b̂i 1 ≤ i ≤ 3 where b̂1 = b1 , b̂2 = b2 , b̂3 = b2 .
Furthermore, b1 = Be1 and b2 = Be2 . Hence, η1 = e1 , η2 = e2 , η3 = e2 . Recall that F was defined
by its action of xi s i.e.,
F x1 = η1 = e1 , F x2 = η2 = e2 , F x3 = η3 = e2 , F x4 = η (34)
where η is some vector whose value will be specified later. Now we find F by its action on old basis
vectors e1 , . . . , e4 of X since we have (A, B) in the standard basis representation.
−(λ + 1) 1 0 (λ + 1)2
Choose η such that F e3 = 0. Thus F = and
1 0 0 −2λ
λ 1 0 0 0 0
−(λ + 1) 1 0 (λ + 1)2
0 λ 0 0
+ 1 0
A + BF =
0 0 λ 1 0 0 1 0 0 −2λ
0 0 0 λ 0 1
λ 1 0 0 0 0 0 0
0 λ 0 0 −(λ + 1)
1 0 (λ + 1)2
=
0 +
0 λ 1 0 0 0 0
0 0 0 λ 1 0 0 −2λ
λ 1 0 0
−(λ + 1) λ + 1 0 (λ + 1)2
= . (41)
0 0 λ 1
1 0 0 −λ
One can check that (A + BF, b1 ) = (A + BF, e2 ) is cyclic. Now we can use SISO trick for pole
placement.
We give the following algorithm for MIMO pole placement based on Lemma 7.6.
1. Select b = b1 i.e. the first column of B and generate a maximal cyclic subspace X1 . Thus
x1 = b1 , . . . , xn1 = Axn1 −1 + b1 .
2. Select b2 i.e. the second column of B and generate X2 . xn1 +1 = Axn1 + b2 , . . . , xn1 +n2 =
Axn1 +n2 −1 + b2 and so on. Thus, we obtain a basis {x1 , . . . , xn } for X.
3. b1 = Be1 , b2 = Be2 , . . . , bk = Bek , xi+1 = Axi + b̂i . For i = 1, . . . , n1 − 1 b̂i = b1 = Be1 . For
i = n1 , . . . , n1 + n2 − 1, b̂i = b2 = Be2 and so on.
Remark 7.10. Suppose one wants to drive the state from x(t0 ) to x(tf ) = xf and hold it there
for all t ≥ tf . For a controllable system, one needs to choose u|[t0 ,tf ] such that x(tf ) = xf . Then,
for all t ≥ tf , u(t) must be chosen such that ẋ = Ax + Bu = 0. Since x(t) = xf for all t ≥ tf ,
0 = Axf + Bu(t) i.e., choose u(t) such that Bu(t) = −Axf . This implies that Axf must lie in the
column span of B if one wants to hold the state at xf for all t ≥ tf .
Lemma 7.11 (Wonham). Let B be a subspace with minimal polynomial β w.r.t. A. Then, there
exists b ∈ B with minimal polynomial β.
Theorem 7.12 (Wonham). Suppose (A, B) is controllable. Let α1 , . . . , αk be the invariant factors
of A. Then, there exists A−invariant subspaces Xi ⊆ X and vectors bi ∈ column span of B such
that
• X = X1 ⊕ . . . ⊕ Xk .
• If Bi =< b1 > + . . . + < bi > (1 ≤ i ≤ k), then < Bi , ABi , . . . , An−1 Bi >= X1 ⊕ . . . ⊕ Xi .
Proof. Since (A, B) is controllable, the minimal polynomial of B is same as the minimal polynomial
of A. By the previous lemma, there exists b1 ∈ B such that the minimal polynomial of b1 is same
as the minimal polynomial of B which is mA . Let X1 =< b1 , Ab1 , . . . , An−1 b1 >. Since X1 is
maximaly cyclic, X = X1 ⊕ Y1 where AY1 ⊂ Y1 . Let Q be a projection on Y1 along X1 . Since
both X1 and Y1 are A−invariant, QA = AQ and (I − Q)A = A(I − Q). Therefore,
By the previous lemma and controllability, there exists b2 ∈ B such that Qb2 under the action of
A forms a maximal cyclic subspace X2 ⊂ Y1 = X2 ⊕ Y2 . Note that b2 ∈ X1 ⊕ X2 . Furthermore,
B2 =< b1 > + < b2 > and < B2 , AB2 , . . . , An−1 B2 >= X1 ⊕ X2 . Continuing this way, one obtains
that if Bi =< b1 > + . . . + < bi > (1 ≤ i ≤ k), then < Bi , ABi , . . . , An−1 Bi >= X1 ⊕ . . . ⊕ Xi .
It is clear that the basis used in the proof above to convert A into Rational
canonical form
is {b1 , Ab1 , . . . , b2 , Ab2 , . . . , bk , Abk , . . . , Ank −1 bk }. Let B = b̂1 b̂2 . . . b̂m . Recall that un-
Now we study some properties of the Controllability Gramian when (A, B) is controllable.
ˆ t ˆ t
T T T
Wt (A, B) = eA(t−τ ) BB T eA (t−τ ) dτ = eAt ( e−Aτ BB T e−A τ dτ )eA t
0 0
d
⇒ Wt (A, B) = AWt (A, B) + BB T + Wt (A, B)AT ( differential Lyapunov equation).
dt
Suppose A is stable, i.e. all eigenvalues of A lie strictly in the LHP. Then,
ˆ ∞ ˆ ∞
T T
W∞ (A, B) = eA(t−τ ) BB T eA (t−τ ) dτ = eAρ BB T eA ρ dρ
0
ˆ ∞ 0
T d Aρ T AT ρ
⇒ AW∞ (A, B) + W∞ (A, B)A = (e BB e )dρ = −BB T
0 dρ
Thus, W∞ (A, B) satisfies a Lyapunov equation. Moreover, the solution to this Lyapunov equation
is unique. Suppose W̄ is some other solution of the Lyapunov equation.
Tt
A(W∞ (A, B) − W̄ ) + (W∞ (A, B) − W̄ )AT = 0 ⇒ eAt (A(W∞ (A, B) − W̄ ) + (W∞ (A, B) − W̄ )AT )eA =0
ˆ ∞
d At AT t d At T
⇒ (e (W∞ (A, B) − W̄ )e ) = 0 ⇒ (e (W∞ (A, B) − W̄ )eA t )dt = 0
dt 0 dt
Thus, evaluating the integral at boundary points, eAt vanishes at ∞, hence W∞ (A, B) = W̄ .
Observe that eigenvectors of A and −µI − A are the same. One can choose µ > 0 such that
eigenvalues of −µI − A are strictly in the LHP. By the PBH test, (−µI − A, B) is controllable ⇔
(A, B) is controllable. Thus, from the above theorem, there exists W > 0 such that
Note that by pole placement theorem, controllability implies stabilizability. The following result
is from Hespanha (Theorem 14.4).
Theorem 7.17. (A, B) is stabilizable ⇔ there exists a positive definite solution P to the Lyapunov
matrix inequality AP + P AT − BB T < 0.
Proof. (⇒) Suppose (A, B) is in the following form after a change of basis
A1 A12 B1
A= ,B = .
0 A2 0
Since, (A1 , B1 ) is controllable, by Equation (44), there exists positive definite P1 such that A1 P1 +
P 1 AT T
1 −B1 B1 = −Q1 < 0. Moreover, since A2 is stable, there exists P2 > 0 such that A2 P2 +P2 AT
2 =
P1 0
−Q2 < 0. Let P = , ρ > 0. Thus, AP + P AT − BB T =
0 ρP2
T
B1 B1T 0
A1 A12 P1 0 P1 0 A1 0 Q1 −ρA12 P2
+ − =− .
0 A2 0 ρP2 0 ρP2 AT12 A2
T 0 0 −ρP2 AT12 ρQ2
By making ρ sufficiently small, the RHS can be made negative definite.
(⇐) Let v ∗ be a left eigenvector of A associated with an unstable eigenvalue λ. Thus, Re(λ) > 0
v ∗ (AP + P AT − BB T )v < 0 ⇒ 2Re(λ)v ∗ P v − kv ∗ Bk2 < 0 ⇒ kv ∗ Bk2 > 0. (46)
Therefore, v ∗ ∈
/ kerB. Hence, (A, B) is stabilizable by PBH test for stabilizability.
Let K = 21 B T P −1 where P > satisfies the Lyapunov matrix inequality above.
1 1
AP + P AT − BB T = (A − BB T P −1 )P + P (A − BB T P −1 )T = (A − BK)P + P (A − BK)T < 0.
2 2
Multiplying on the left and right by Q = P −1 , we get a Lyapunov inequality hence, (A − BK) must
be stable. Thus, for stabilizable systems, there exists a feedback law u = −Kx which asymptotically
stabilizes the system.
8 Observability
We show that the observability problem is least squares problem. We also give a canonical Kalman
decomposition of a linear system into controllable/uncontrollable and observable/unobservable
parts.
Definition 8.1. Consider a linear system ẋ = Ax + Bu and y = Cx + Du. We say that a system
is observable if it is possible to uniquely determine the state x(t) from the knowledge of the output
y(t) and the input u(t).
For simplicity, assume that D = 0 and y = Cx. Thus, taking derivatives on both sides of the
output equation,
y = Cx
d d
y = C x = CAx + CBu
dt dt
.
.
d n−1 d d
( ) y = C( )n−1 x = CAn−1 x + CAn−2 Bu + . . . + CB( )n−2 u
dt dt dt
y C 0 ... 0 0 u
d d
dt y CA CB ... 0 0 dt u
. =
. x +
. . . . . .
.
(47)
. . . ... . . .
d n−1 CAn−1 CAn−2 B . . . CB 0 d n−1
( dt ) y ( dt ) u
C
CA
Let O(C, A) :=
. be the Observability matrix and let
.
CAn−1
y 0 ... 0 0 u
d d
dt y CB ... 0 0 dt u
z := . − . ... . . .
. . ... . . .
d n−1 CAn−2 B . . . CB 0 d n−1
( dt ) y ( dt ) u
Thus, we have a linear equation O(C, A)x = z. Observe that this is a Least squares problem.
Observe that by construction, z ∈ Im(O(C, A)). Thus, this system of linear equations has a
solution. The solution is unique ⇔ O(C, A) has full column rank. Thus, a system is observable ⇔
O(C, A) has full column rank.
Assuming that O(C, A) has full column rank, O(C, A)T O(C, A)x = O(C, A)T z and
x = (O(C, A)T O(C, A))−1 O(C, A)T z.
Lemma 8.2. O(C, A) is observable ⇔ (AT , C T ) is controllable
Proof. Follows from the rank test above for observability and the corresponding rank test for
controllability.
Observe that if O(C, A) is not full column rank, then it is not possible to uniquely determine
x which implies unobservability. Ker(O(C, A)) = ker(C)∩ker(CA) ∩ . . . ∩ker(CAn−1 ) forms unob-
servable states. If a state x is unobservable, then Cx = CAx = . . . = CAn−1 x = 0. Hence, by
Cayley-Hamilton theorem, CAi x = 0 for all i ∈ N∪0. Thus, if Ō = ker(O(C, A)) is the unobservable
subspace, then AŌ ⊆ Ō. Therefore, the unobservable subspace is A−invariant.
Consider a following equation for y with D = 0 (for simplicity)
ˆ t
y(t) = Cx(t) = CeAt x(0) + CeA(t−τ ) Bu(τ )dτ
0
ˆ T ˆ T ˆ T ˆ t
T T T
⇒ eA t C T y(t)dt = eA t C T CeAt dtx(0) + eA t C T CeA(t−τ ) Bu(τ )dτ dt. (48)
0 0 0 0
´T T
Define WT (C, A) = 0 eA t C T CeAt dt as the Observability Gramian. Thus, from the previous
equation, one can uniquely determine x(0) from the knowledge of outputs and inputs ⇔ WT (C, A)
is positive definite (hence invertible). From x(0), one can determine x(t) using the closed form
expression for x(t).
1. (C, A) is observable.
Note that in presence of noise, taking derivatives of the output would increase the effect of
noise. Determining the state using the Observability Gramian would work better in the presence
of noise. Moreover, it works for LTV too.
´t
Corollary 8.4. For LTV systems, (C(t), A(t)) is observable ⇔ t0 ΦT (t, τ )C T (t)C(t)Φ(t, τ )dτ > 0
for some t > t0 .
Theorem 8.7. (C, A) is observable ⇔ for every symmetric set Λ of complex numbers, there exists
a feedback matrix K such that eigenvalues of (A + KC) are given by Λ.
Proof. Follows from (C, A), (AT , C T ) duality (Lemma 8.2) and the corresponding controllability
theorem for pole placement.
Thus, one use a feedback matrix such that all eigenvalues of the system matrix lie in the LHP
i.e. (A − KC) has eigenvalues in the LHP.
´T T
Recall that WT (C, A) = 0 eA t C T CeAt dt.
ˆ t ˆ t
y(t) = CeAt x(0) + CeA(t−τ ) Bu(τ )dτ = CeAt x(0) + CeAt e−Aτ Bu(τ )dτ
0 0
ˆ t
T T T
therefore, y T (t) = xT (0)eA t C T + ( uT (τ )B T e−A τ dτ )eA t C T
0
´t T
Let z(t) = 0 e−Aτ Bu(τ )dτ . Thus, y(t) = CeAt x(0) + CeAt z(t) and y T (t) = xT (0)eA t C T +
T
z T (t)eA t C T .
T T
⇒ y T (t)y(t) = xT (0)eA t C T CeAt x(0) + z T (t)eA t C T CeAt x(0)
T T
+xT (0)eA t C T CeAt z(t) + z T (t)eA t C T CeAt z(t)
ˆ T ˆ T
T T T T
⇒ y (t)y(t)dt = (xT (0)eA t C T CeAt x(0) + 2z T (t)eA t C T CeAt x(0) + z T (t)eA t C T CeAt z(t))dt
0 0
ˆ T
T T T
= x (0)WT (C, A)x(0) + (2z T (t)eA t C T CeAt x(0) + z T (t)eA t C T CeAt z(t))dt(49)
0
d T
Note that dT WT (C, A) = eA t C T CeAt . Therefore,
ˆ T ˆ T
d d
y T (t)y(t)dt = xT (0)WT (C, A)x(0) + (2z T (t) WT (C, A)x(0) + z T (t) WT (C, A)z(t))dt(50)
0 0 dT dT
This gives as expression for the output energy in terms of the Observability Gramian.
Suppose all eigenvalues of A lie strictly in the LHP. Thus, one can show using similar arguments
used for the Controllability Gramian that W∞ (C, A) satisfies the Lyapunov equation AT W∞ (C, A)+
W∞ (C, A)A = −C T C.
Theorem 8.8. Suppose all eigenvalues of A lie strictly in the LHP. Then, (C, A) is observable ⇔
there exists a unique positive definite solution W to the Lyapunov equation AT W + W A = −C T C.
Suppose (C, A) is not observable. We saw that the unobservable subspace is A−invariant and
belongs to the kernel of C. Choose a basis for the unobservable subspace and extend it to a basis
for the entire state space. By arranging the basis vectors such that vectors in the kernel of C are
placed after the vectors which are not in the kernel of C, we get a following decomposition for A
and C ,
A1 0
A= ,C = C1 0 .
A21 A2
Now further subdivide both Vc and Vc̄ into observableand unobservable states i.e. Vc = [Vco Vcō ]
xco
xcō
and Vc̄ = [Vc̄o Vc̄ō ]. Thus, T = [Vco Vcō Vc̄o Vc̄ō ] and x =
xc̄o .
xc̄ō
With this similarity transform, recalling the observability decomposition,
ẋco Aco 0 xco A∗o 0 xc̄o Bco
ẋc = = + + u
ẋcō Ac∗ Acō xcō A∗∗ A∗ō xc̄ō Bcō
ẋc̄o Ac̄o 0 xc̄o
ẋc̄ = = (51)
ẋc̄ō Ac̄∗ Ac̄ō xc̄ō
It follows from the Kalman decomposition that the transfer function of a system only captures the
controllable and observable part of the system. This can be verified by substituting (A, B, C) triple
above in G(s) = C(sI − A)−1 B. Thus, G(s) = C(sI − A)−1 B = G(s) = Cco (sI − Aco )−1 Bco .
9 State estimation
The feedback law u = −Kx can not be implemented when the state x is not directly accessible.
For observable systems, one can uniquely determine the state vector. However, to implement a
feedback law, we need x(t) for all time steps. Building an observer which estimates the state vector
asymptotically at each time step lets one implement a feedback law. If (C, A) is unobservable but
detectable, it is possible to estimate x from the system output up to an error e → 0 as t → ∞.
Consider a following model for an observer
Theorem 9.1. If there exists L such that A − LC has eigenvalues in the LHP, then the state
estimation error converges to 0 exponentially for all input signals u.
For observable and or detectable systems, it is always possible to choose L such that eigenvalues
of (A − LC) lie strictly in the LHP.
Suppose we want to implement a state feedback law using state estimates x̂ i.e. u = −K x̂.
The above equation is known as separation principal where one can design a stable feedback gain
K and an output injection gain L independently such that the closed loop system is stable and the
state estimation error converges to zero. (Eigenvalues of the matrix above are given by the union
of eigenvalues of (A − BK) and (A − LC). If (A, B) is controllable and (C, A) is observable, one
can choose K and L such that both x and e converge to zero at infinity. Moreover, if (A, B) is
stabilizable and (C, A) is detectable, one can still construct an observer such that x and e converge
to zero at infinity.) This is also called stabilization through output feedback. This determines the
observer design to implement state feedback laws.
The dynamics of the state estimate x̂ under the feedback u = −K x̂ are as follows:
x̂˙ = Ax̂ + Bu + LC(x − x̂) = Ax̂ − BK x̂ + LC(x − x̂) = (A − LC − BK)x̂ + LCx (57)
Let (Ā, B̄, C̄) be a matrix representation in the old basis. Thus, A = T −1 ĀT , B = T −1 B̄ and
C = C̄T . Let T = [T1 T2 ]. Thus [C̄T1 C̄T2 ] = [I 0]. (Therefore,
T1 is the right inverse of C̄
x1
and T2 is orthogonal complement of rows of C̄.) Let x = be partition of x according to the
x2
partition [T1 T2 ]. Observe that y = [I 0]x = x1 . Thus, we need to estimate x2 .
x̂˙ 2 = A21 x1 + A22 x̂2 + L(ẋ1 − A11 x1 − B1 u − A12 x̂2 ) + B2 u = A21 x1 + A22 x̂2 + LA12 (x2 − x̂2 ) + B2 u(60)
⇒ ẋ2 − x̂˙ 2 = A22 (x2 − x̂2 ) − LA12 (x2 − x̂2 ) = (A22 − LA12 )(x2 − x̂2 )(61)
10 Feedback
Feedback can be an output feedback or a state feedback for state space models. Moreover, it can
be static or dynamic. Static feedback does not involve any dynamics. Note that with a static
output feedback u = Ky = KCx, due to the presence of the matrix C, one cannot do an arbitrary
pole placement. This is a limitation of static output feedback. However, using an output feedback
and (A, B, C, D) matrices, one can build an observer/state estimator (observer based controller)
as shown above which asymptotically estimates states and allows us to implement state feedback
laws/arbitrary pole placement with an error converging to zero. Note that static feedback can be
identified with P (proportional) controller.
Dynamic feedback controller involves dynamics in its system model. Assuming that (A, B)
is stabilizable and (C, A) is detectable (where ẋ = Ax + Bu, y = Cx), one can design a linear
dynamic output feedback controller
ż = F z + Gy, u = Lz + M y (62)
11 Well-posed interconnection
Consider an interconnection of a system and a controller with following dynamics
ẋs = A1 xs + B1 u1 , ys = C1 xs + D1 u1 (63)
ẋc = A2 xc + B2 u2 , yc = C2 xc + D2 u2 . (64)
1
Thanks to Pragada Shivaramkrishna for this issue.
The interconnection implies that u2 = ys and u1 = yc . Therefore, one obtains
u2 = ys = C1 xs + D1 u1 = C1 xs + D1 (C2 xc + D2 u2 ) = C1 xs + D1 C2 xc + D1 D2 u2
⇒ u2 = (I − D1 D2 )−1 (C1 xs + D1 C2 xc ) (65)
u1 = yc = C2 xc + D2 u2 = C2 xc + D2 (C1 xs + D1 u1 ) = C2 xc + D2 C1 xs + D2 D1 u1
⇒ u1 = (I − D2 D1 )−1 (C2 xc + D2 C1 xs ) (66)
A1 + B1 (I − D2 D1 )−1 D2 C1 B1 (I − D2 D1 )−1 C2
ẋs xs
= .
ẋc B2 (I − D1 D2 )−1 C1 A2 + B2 (I − D1 D2 )−1 D1 C2 xc
An interconnection which is not well posed is called ill posed. It turns out that if either the system
or the controller is strictly proper (either D1 = 0 or D2 = 0), then the interconnection is always
well posed. It follows that interconnection of the system with an observer based controller is well
posed since D2 = 0.
12 Applications/Design examples
Example 12.1 (Feedback linearization and a stabilizing control law). Consider an oscillating
pendulum θ̈ + gl sin(θ) = 0. Find a stabilizing control law u for the system θ̈ + gl sin(θ) = u which
drives the pendulum from an arbitrary initial position with an arbitrary initial velocity to θ = 0 and
θ̇ = 0 asymptotically.
Observe that introducing state variables x1 = θ, x2 = θ̇, this can be converted into a first order
system.
g
ẋ1 = x2 , ẋ2 = − sin(θ) + u
l
(If we consider θ̇ as a variable of interest, this system can be modeled as a first order system where
the order of the ode is equal to the number of energy storing elements which is one (mass) in this
case.)
Using u to for feedback linearization, choose u = gl sin(θ) + v. Thus we have the following linear
system:
˙
x1 0 1 x1 0
= +
x2 0 0 x2 v
Clearly, the above system is controllable (by the rank test). Thus, an arbitrary pole placement is
possible. Choose v = k1 x1 + k2 x2 . Thus we have,
˙
x1 0 1 x1
=
x2 k1 k2 x2
Now choose k1 , k2 such that the eigenvalues of the feedback matrix lie in the LHP.
One can consider other practical examples of non-linear systems too which are feedback lin-
earizable and controllable after feedback linearization. Find out such systems in robotics, power
electronics or systems in other engineering domains and find appropriate control laws for their
designs. Build observers for these systems to implement feedback laws. Do simulations for these
models.
Example 12.2 (Tracking). Consider a SISO system (A, b, c) which is both controllable and observ-
able. Find u such that y tracks a constant reference signal which is unit step. Let u = −Kx + v
such that v(t) = 1 for t ≥ 0. Thus,
ˆ t ˆ t
(A−BK)t (A−BK)(t−τ ) (A−BK)t
y(t) = Ce x(0) + Ce Bv(τ )dτ = Ce x(0) + Ce(A−BK)(t−τ ) Bdτ
0 0
ˆ t
limt→∞ y(t) = limt→∞ {Ce(A−BK)t x(0) + Ce(A−BK)t ( e−(A−BK)(τ ) dτ )B}
0
´t
Since (A − BK) has eigenvalues strictly in the LHP, it is invertible. Hence, 0 e−(A−BK)(τ ) dτ =
−(A − BK)−1 (e−(A−BK)t − I). Thus,
limt→∞ y(t) = limt→∞ {−C(A − BK)−1 e(A−BK)t (e−(A−BK)t − I)B}
= limt→∞ {−C(A − BK)−1 (I − e(A−BK)t )B}
= −c(A − bK)−1 b (67)
Thus, we normalize v by the factor above so that y(t) tracks the unit step signal. (Observe that
the transfer function for the closed loop system is G(s) = c(sI − A + bK)−1 b. Thus G(0) =
−c(A − bK)−1 b.)
Example 12.3 (Feedback linearization and tracking). Consider Example 1. Suppose y(t) = θ(t).
Suppose we want to design a feedback law such that θ(t) tracks the position θ = π and θ̇ = 0.
Then one can use feedback linearization to convert the non-linear system to a linear system which
is controllable. Then one can use Example 2 to design an appropriate input signal.
One can combine the above methods to obtain methods which will give a strategy to track
reference signals for more general feedback linearizable non-linear systems. Construct/find such
examples in applications. One can consider an example of robot arm manipulator which is feedback
linearizable. One can have a network of single input systems where each system could be linear or
feedback linearizable. Then one can choose an appropriate input for each of these systems.
Example 12.4. Suppose we want to build an observer to implemented state feedback laws. If the
output is a non-linear function of states, linear observers won’t work. If the output is a linear
function of states, still there are difficulties due to non-linearities in the state equation and one
needs to use techniques from non linear control theory. However, there are feedback linearizable
systems for which it is possible to build linear observers to implement state feedback laws. They are
of the following form
ẋ1 = x2
.
.
ẋn−1 = xn
ẋn = f (x1 , . . . , xn ) + u
y = x1 .
Thus, all states can be derived from the output equation. Thus, one can implement state feedback
laws for such systems. Find more example of systems where it is possible to build an observer using
tools from linear systems theory.
13 Disturbance decoupling problem
Consider a following model for LTI system where q represents either disturbance or noise (Wonham).
ẋ = Ax + Bu + Sq (68)
y = Cx (69)
We say that a system in the form above is disturbance decoupled w.r.t. the pair (q, y) if for each
initial state x(0) ∈ X, the output y(t) is the same for every q. Note that with a state feedback
u = Kx,
ˆ t
(A+BK)t
ẋ = (A + BK)x + Sq ⇒ x(t) = e x(0) + e(A+BK)(t−τ ) Sq(τ )dτ
0
ˆ t
⇒ y(t) = Ce(A+BK)t x(0) + C e(A+BK)(t−τ ) Sq(τ )dτ
0
Hence, disturbance decoupling means that find u = Kx such that the forced response yf (t) =
´t
C 0 e(A+BK)(t−τ ) Sq(τ )dτ = 0. In other words, we want to find K such that the transfer function
from q to y is identically zero. Thus, we want T (s) = YQ(s)
(s)
= C(sI − (A + BK))−1 S = 0. Let K =
Ker(C) and S = Im(S). Let V =< A + BK, S > be a subspace generated by columns of S under
´ t (A+BK)(t−τ
the action of A + BK. It is clear from Cayley-Hamilton theorem that 0 e ) Sq(τ )dτ lies
in the subspace < A + BK, S >. This is called disturbance decoupling problem (DDP). Thus DDP
is solved ⇔ there exists K such that < A + BK, S >= V ⊂ K.
Observe that V is (A + BK)−invariant from construction.
Definition 13.1. (Wonham) A subspace W ⊂ X is said to be (A, B)−invariant, if there exists a
matrix K such that (A + BK)W ⊂ W. We denote the class of (A, B)−invariant subspaces of X
by γ(A, B, X).
It is clear that all A−invariant subspaces are (A, B)−invariant by choosing K = 0.
Lemma 13.2. Let W ⊂ X and B = Im(B). Then W ∈ γ(A, B, X) ⇔ AW ⊂ W + B.
Proof. (⇒) Suppose W ∈ γ(A, B, X). Therefore, there exists K such that (A + BK)W ∈ W. Let
w1 ∈ W. Therefore, (A + BK)w1 = w2 ∈ W ⇒ Aw1 = w2 − BKw1 ∈ W + B.
(⇐) Suppose AW ⊂ W + B. Let {w1 , . . . , wl } be a basis for W. Therefore, for each wi 1 ≤ i ≤ l,
there exists vi ∈ W and ηi such that Awi = vi − Bηi . Define a linear map K̂ on W such that
K̂wi = ηi (1 ≤ i ≤ l). Extending the basis for W to a basis for X, we can extend K̂ to a linear
map K on X. Then (A + BK)wi = Awi + BKwi = vi − Bηi + Bηi ∈ W. Thus, W is (A + BK)
invariant.
Lemma 13.3. The class of subspaces γ(A, B, X) is closed under subspace addition.
Proof. Follows from the above lemma.
ẋ = Ax + Bu, y = Cx
Find if possible a feedback matrix K such that y(t) = Ce(A+BK)t → 0 as t → ∞. From the
previous DDP problem, find V ∗ = sup(γ(A, B, Ker(C) = K)). Let (A + BF )V ∗ ⊆ V ∗ . If all
unstable modes of (A + BF, B) lie in the kernel of C, then this solves OSP by choosing u = F x.
Consider the case when there are unstable modes of A + BF lying outside ker(C). If these unstable
modes are controllable, then they can be made to lie in the LHP by an appropriate state feedback
i.e. for ẋ = (A + BF )x + Bv, use v = Kx such unstable (but controllable) eigenvalues of (A + BF )
are shifted to the LHP. Thus, OSP is solved. Thus, OSP is solvable ⇔ (unstable modes of A)
⊆ < A, B > +V ∗ . (Union of subspaces is not a subspace in general. Hence we take the sum of
subspaces to get a bigger subspace containing both the smaller subspaces.)
Uncontrollable modes of A are unaffected under state feedback. If they lie in the kernel of C or
are stable, OSP can be solved by shifting all controllable modes to the LHP. If there are unstable
uncontrollable modes of A lying outside the kernel of C, then OSP is not solvable.
Observe that by construction, controllability subspaces are A−invariant. Thus when A is re-
stricted to C(A, B1 ), arbitrary pole placement is possible for restriction of A on this subspace. A
collection of controllability subspaces forming a class of controllability subspaces of the state space
X is denoted by C(A, B, X). By choosing elements in B, we can generate different controllability
subspaces. It is clear from the construction of controllability subspaces that the class C(A, B, X) is
closed under subspace addition (just like the class γ(A, B, X) of (A, B)−invariant subspaces) and
it contains a unique maximal element.
Remark 13.7. Consider a controllability subspace C(A, B1 ) such that the initial condition x0 ∈
C(A, B1 ). Let {b1 , . . . , bk } be a basis for B1 . Let B1 = [b1 , . . . , bk ]. Taking action of A on this
set and taking only linearly independent vectors, generate a basis for C(A, B1 ) and extend it to a
basis of X. Let A1 be the restriction of A to C(A, B1 ). Clearly, (A1 , B1 ) is controllable and using
a feedback F1 , we can do arbitrary pole placement for (A1 + B1 F1 , B1 ). Note that (A, B1 ) need not
be controllable and w.r.t. the chosen basis for X, we have an upper triangular decomposition of
(A, B1 ) into the controllable and uncontrollable part. Since the initial condition x0 ∈ C(A, B1 ), the
component of the intial condition in uncontrollable part of the state space is zero. This implies that
the evolution of x(t) remains inside the controllable subspace C(A, B1 ) if x0 ∈ C(A, B1 ). This is
an important property of controllability subspaces.
DDP with stability (DDPS): We want to solve DDP using a state feedback such that the
feedback matrix is stable. For an (A, B)−invariant subspace W, we know from the definition that
there exists K such that (A + BK)W ⊂ W. Let F(W) denote the set of matrices K such that
(A + BK) has eigenvalues strictly in the LHP. Consider a class Γ(A, B, K) ⊂ γ(A, B, K) such that
if W ∈ Γ(A, B, K), then F(W) 6= ∅. Thus, we have the following natural analogue of the DDP
result.
Definition 14.1. A realization (A, B, C, D) of a transfer function G(s) (i.e. C(sI − A)−1 B + D =
G(s)) is said to be minimal if there does not exist a realization of a smaller order.
Definition 14.2. Two state space systems are said to be zero state equivalent if they realize the
same transfer function.
ti i ti
Observe that (sI − A)−1 = L(eAt ) = L{ ∞ −(i+1) . Therefore,
P
i=0 i! A }. Note that L(P
i! ) = s
−1 ∞ −(i+1) i −1 ∞ −(i+1) CAi B.
P
(sI − A) = i=0 s A . Hence, G(s) = C(sI − A) B + D = D + i=0 s
Definition 14.3. CAi B for i ≥ 0 are called Markov parameters of a realization (A, B, C, D).
d d di di
g(t) = L −1 (G(s)) = CeAt B + Dδ(t), g(t) = CAeAt B + Dδ(t), . . . i g(t) = CAi eAt B + i Dδ(t)
dt dt dt dt
d
Thus, taking the derivatives from the right as t → 0+ , g(0+ ) = CB, dt g(t)|t=0+ = CAB and so on.
Theorem 14.4. Two realizations (A, B, C, D) and (Ā, B̄, C̄, D̄) are zero state equivalent ⇔ they
have the same Markov parameters and D = D̄.
Proof. (⇒) Zero state equivalence implies that D = D̄. Moreover, since G(s) = Ḡ(s), Markov
parameters must be the same. Conversely, if two realizations have the same Markov parameters
and D = D̄, then they realize the same transfer function hence, are zero state equivalent.
Since (A, B, C, D) and (Ā, B̄, C̄, D̄) realize the same transfer function, they have the same Markov
parameters. Therefore, OC = ŌC̄. But rank(ŌC̄) = n̄ < n = rank(OC), which is a contradiction.
Therefore, the realization (A, B, C, D) is minimal.
Corollary 14.6. (A−KC, B) is controllable ⇔ (C, A−KC, B) is minimal. Similarly, (C, A−BF )
is observable ⇔ (C, A − BF, B) is minimal.
Observe that the controllability/observability matrix and the corresponding Gramian matrices
do not remain invariant under a change of basis. However, their rank remains invariant.
Suppose a matrix A is full column rank. Then AT A is positive definite, hence invertible. Let
Al = (AT A)−1 AT . Then Al A = I and Al is called the left inverse of A. Similarly, if A is full row
rank, then AAT > 0 and Ar = AT (AAT )−1 is called the right inverse of A i.e. AAr = I.
Theorem 14.8. All minimal realizations of a transfer function are algebraically equivalent.
Proof. Let (A, B, C, D) and (Ā, B̄, C̄, D̄) be two minimal realizations. By the previous theorem,
they are both controllable and observable and OC = ŌC̄. Note that C and C̄ have full row rank
and O, Ō have full column rank. Let T = C̄C r .
Thus, from previous two equations, T −1 B̄ = B and C̄T = C. Since OC = ŌC̄ and both realizations
have the same Markov parameters, OAC = ŌĀC̄ ⇒ Ol OACC r = (Ol Ō)Ā(C̄C r ). This implies
that A = T −1 ĀT . Thus, T provides a similarity transform between the two realizations and D = D̄
since they realize the same transfer function.
System norms: Substituting s = jω in the transfer function G(s), we obtain the frequency
response for an LTI system. The 2−norm is given by
ˆ ∞ ˆ ∞
1 1
kGk2 := ( 2
kG(jω)kF dω) = (
2 trace(G(jω)G∗ (jω))dω) 2 (73)
−∞ −∞
where σ1 is the maximum singular value of G(jω). Note that infinity norm is finite ⇔ G(s) has
no poles on the purely imaginary axis. For 2−norm to be finite, apart from the poles condition,
G(s) also has to be strictly proper. (We refer the reader to Doyle, Francis and Tannerbaum for
details.) System norms are useful in robust control to design controllers which minimize the effect
of the disturbance on the output by minimizing the norm of the corresponding transfer function
(from the disturbance to the output).
Matrix-fraction Descriptions (MFDs) Given a matrix transfer function G(s), one can write
−1 −1
G(s) = NR (s)DR (s) or G(s) = DL (s)NL (s). This is called right and left matrix fraction respec-
tively and it is not unique. Kailath shows that it is possible to obtain a controllable realization for
right MFDs and observable realization for left MFDs. For more on this topic, we refer the reader
to Kailath (Chapter 6). MFD is useful for design problems where one wants desired transfer func-
tions/loop transfer functions. Using MFDs, one can shape the transfer function of an underlying
system by choosing compensators with appropriate MFDs so that one obtains a desired transfer
function.
15 Poles-zeros and Smith McMillan form, Transmission zeros, In-
variant zeros, System inverse
Refer Hespanha. Applications of Invariant zeros: Modeling, detection and correction of cyber
attacks (Refer works of Pasqualetti, Fawzi)
G(s) = C(sI − A)−1 B + D = pA1(s) (CAdj(sI − A)B + DpA (s)). One can write G(s) = pA1(s) N (s)
and from the Smith form Σ(s) of N (s) where Σ = diag(σi (s)) such that σi |σi+1 , we obtain the
Smith-McMillan form SMG of G(s), where the nonzero diagonal entries of SMG are ψnii such that
ni σi ni+1 σi+1
ψi = pA and ψi+1 = pA . Therefore, ni |ni+1 and ψi+1 |ψi . Roots of Πi ni (s) are called
P
zeros of G(s)
and roots of Πi ψi (s) are called poles of G(s). The McMillan degree is given by i deg(ψi ).
Lemma 15.1. Poles of G(s) ⊆ eigenvalues of A.
1
Proof. Follows from the fact that G(s) = pA (s) (CAdj(sI − A)B + DpA (s)).
where G(s) is the transfer function. It is clear from this factorization that the rank of P (s) drops at
transmission zeros of G(s). Moreover, if there are pole-zero cancellations, then there are eigenvalues
of A which are not poles of G(s). At these complex numbers, P (s) again loses rank. Thus, if there
are no pole-zero cancellations i.e., for minimal realizations, transmission and invariant zeros are
the same.
Example 15.3. Consider the following system
1 0 1
A= ,B = ,C = 1 0 ,D = 0
0 2 1
s−1 0 1
P (s) = 0 s−2 1 .
−1 0 0
Rank(P (s)) = 3. At s = 2, P (s) loses its rank. Thus, there is an invariant zero at s = 2. Note
1
that G(s) = s−1 . Therefore, there are no transmission zeros hence the set of transmission zeros
= ∅ ⊂ the set of invariant
zeros.
If we choose C = 1 1 in the above example, then there are no invariant zeros.
Suppose there is a transmission zero at z0 then, G(z0 ) is rank deficient i.e., there exists u0 such
that G(z0 )u0 = 0. Suppose x0 = (z0 I − A)−1 Bu0 . Let u = ez0 t u0 . We will see that the output is
zero for this input and the intial condition x0 .
ˆ t ˆ t
At −Aτ z0 τ At
x(t) = e (x0 + e e Bu0 dτ = e (x0 + e(z0 I−A)τ dτ Bu0
0 0
At −1 (z0 I−A)t
= e (x0 + (z0 I − A) (e − I)Bu0 )
At −At z0 t −1
At
= e x0 + e e e (z0 I − A) Bu0 − eAt (z0 I − A)−1 Bu0
= eAt x0 + ez0 t x0 − eAt x0 = ez0 t x0
y(t) = Cez0 t x0 = ez0 t Cx0 = ez0 t G(z0 )u0 = 0.
Thus, for the initial condition x0 = (z0 I − A)−1 Bu0 , a change in the input u(t) of the form u(t) +
ez0 t u0 goes undetected at the output. Therefore, malicious attacks made at system transmission
zeros go undetected.
Note that if there is an invariant zero at z0 , then P (z0 ) is rank deficient i.e., there exists x0 , u0
such that
x0 zI − A B x0
P (z0 ) = =0
−u0 −C D −u0
⇒ (z0 I − A)x0 − Bu0 = 0 ⇒ x0 = (z0 I − A)−1 Bu0 , and − Cx0 − Du0 = 0.
Let u(t) = ez0 t u0 and x(0) = x0 . Thus, from the arguments seen previously, x(t) = ez0 t x0 and
y(t) = Cez0 t x0 + Dez0 t u0 = 0. For a connection between the set of invariant zeros and detection
and correction of cyber attacks, we refer the reader to [11], [12]. Absence of invariant zeros implies
that attacks are not possible. Attacks can be made on states via B matrix and on the output via
D matrix. We show below that invariant zeros are invariant under feedback.
Consider the Rosenbrock system matrix P (s). By a unimodular transformation,
sI − A B I 0 sI − A − BK B
=
−C D −K I −C − DK D
Thus, invariant zeros do not change under a state feedback/output feedback as pre and post mul-
tiplication by unimodular matrices do not affect invariant zeros of P (s). Note that the Smith form
of P (s) gives all invariant zeros of P (s). These are precisely those complex numbers where P (s)
loses rank.
sI − A B
Remark 15.4. Observe that the Rosenbrock system matrix P (s) = captures all
−C D
information about the underlying linear dynamical system.
• The controllability and stabilizability property can be checked using the PBH test from the
block matrices in the first row of P (s).
• The observability and detectability property can be checked using the PBH test from the block
matrices in the first column of P (s).
• Invariant zeros are given by the zeros of the Smith form associated with P (s).
• Taking Schur complement w.r.t. the block (sI − A), we obtain the transfer function G(s) =
C(sI − A)−1 B + D from which one can deduce the impulse response, frequency response,
Markov parameters, poles and transmission zeros. (From the knowledge of the transfer func-
tion, on can draw Nyquist/Bode plots too.)
• The state transition matrix can be found from the inverse Laplace transform of (sI − A)−1 .
Definition 15.5. A system (A, B, C, D) with transfer function G(s) is said to be invertible if there
exists a system (Â, B̂, Ĉ, D̂) with a proper transfer function Ĝ(s) such that GĜ = ĜG = I.
Proof. Refer Hespanha. If D is invertible, (Â, B̂, Ĉ, D̂) can be constructed explicitly using D−1 and
system equations. If D is not invertible, we get a contradiction by choosing v such that Dv = 0.
Theorem 15.7. A system inverse exists ⇔ the Rosenbrock system matrix is invertible for almost
all complex numbers.
Proof. Suppose that if z0 is not an eigenvalue of A and not an invariant zero of the Rosenbrock
system matrix P (s). Therefore,
I (z0 I − A)−1 B
z0 I − A B I 0 z0 I − A 0
= .
−C D −C(z0 I − A)−1 I 0 C(z0 I − A)−1 B + D 0 I
16 LQR/LQG
In this section, we elaborate a little on some topics from Hespanha.
Definition 16.1 (Feedback invariants). A functional H(x, u) that involves system’s input and state
is called a feedback invariant if computed along solution of the system, it depends only on the initial
condition x(0) and not on the input signal u.
is a feedback invariant.
where
ˆ tf
JLQR (x, u) := xT (tf )F (tf )x(tf ) + xT (t)Qx(t) + uT (t)Ru(t) + 2xT (t)N u(t)dt (77)
0
such that Q ≥ 0 and R > 0. Note that the term xT (tf )F (tf )x(tf ) represents the cost associated
with the deviation of x(tf ) from the origin. The terms inside the integral give cost attached to
inputs and deviation of x(t) from the origin along the state trajectory from t = 0 to t = tf .
Using the feedback invariant obtained above, we want to represent JLQR (x, u) as
ˆ tf
JLQR (x, u) = H(x, u) + Λ(x, u)dt (78)
0
where minu Λ(x, u) = 0. In this case, u(t) = argminu Λ(x, u) minimizes JLQR and the minimum
cost is equal to H(x, u). Note that there are infinitely many choices for a feedback invariant
depending on P (0) and P (t). We will see below that there is a specific feedback invariant for which
the corresponding P (t) which satisfies a matrix differential equation (differential Riccati equation
(DRE)) which will be our choice of feedback invariant to find minu {JLQR (x, u)}.
ˆ tf
T
JLQR = x (tf )F (tf )x(tf ) + xT (t)Qx(t) + uT (t)Ru(t) + 2xT (t)N u(t)dt
0
ˆ tf
T
= H(x, u) + x (tf )F (tf )x(tf ) + xT (t)Qx(t) + uT (t)Ru(t) + 2xT (t)N u(t)dt − H(x, u)
0
ˆ tf
= H(x, u) + xT (tf )F (tf )x(tf ) − xT (tf )P (tf )x(tf ) + (xT (t)Qx(t) + uT (t)Ru(t) + 2xT (t)N u(t) +
0
(Ax + Bu)T P x + xT P (Ax + Bu) + xT Ṗ x)dt
= H(x, u) + xT (tf )F (tf )x(tf ) − xT (tf )P (tf )x(tf ) +
ˆ tf
xT (AT P + P A + Q + Ṗ )x + uT Ru + 2uT (t)(B T P + N T )xdt (79)
0
Let K := R−1 (B T P + N T ).
Adding and subtracting the term xT (P B + N )R−1 (B T P + N T )x in Equation (79) and using Equa-
tion (80),
If we find P (t) such that AT P (t) + P (t)A + Q + Ṗ (t) − (P (t)B + N )R−1 (B T P (t) + N T ) = 0 and
P (tf ) = F (tf ), then Λ(x, u) = (uT + xT K T )R(u + Kx) which has a minimum equal to zero when
u = −Kx. This leads to a closed loop system ẋ = (A − BR−1 (B T P (t) + N T ))x. Therefore, choose
the feedback invariant among all possible choices such that P (t) satisfies the differential Riccati
equation with boundary conditions.
Theorem 16.3. Assume that there exists a symmetric solution to differential Riccati equation
AT P (t) + P (t)A + Q + Ṗ (t) − (P (t)B + N )R−1 (B T P (t) + N T ) = 0 satisfying the boundary condition
P (tf ) = F (tf ).Then the feedback law u(t) = −Kx(t) where K := R−1 (B T P + N T ) minimizes
JLQR (x, u).
If states are not directly accessible, we use a state estimator using an output feedback to im-
plement the feedback law of LQR.
Recall that we had considered an optimization problem on linear systems earlier as follows:
ˆ t
minimizeu uT (τ )u(τ )dτ
0
ˆ t
At
subject to ẋ = Ax + Bu, xf = e x(0) + eA(t−τ ) Bu(τ )dτ. (81)
0
We solved this optimization problem using tools from linear algebra. Cost functions of the above
problem and the LQR problem are different. Substituting Q = 0, F (tf ) = 0, N = 0 and R = I in the
LQR problem, we get the same cost function as above but the constraints are still different as there
is an additional constraint on (81). Therefore, clearly the optimal solutions to the two problems
are also different as there are different cost functions and different constraints. Both problems can
be solved independently using two different approaches as we have seen. However, they are closely
related to one another. They can be solved using Hamilton-Jacobi-Bellman equations (HJB) from
optimal control. Note further that for the LQR problem, we only gave sufficient conditions for an
optimal solution. HJB theory says that these conditions are also necessary for an optimal solution.
Infinite horizon LQR We refer the reader to Hespanha. The above approach is inspired from
Hespanha’s infinite horizon LQR approach.
ẋ = Ax + Bu + B̄d, y = Cx + n (82)
where d represents disturbance and n represents measurement noise with both of them unknown.
Thus any estimate of x would work for sufficiently large d and n. Thus, previous state estimator or
observer does not work in the presence of disturbance and noise. MEE consists of finding a state
trajectory
that starts at rest at t → −∞ and is consistent with the past measured output y and control input
u for the least amount of noise n and disturbance d, measured by
ˆ t
JM EE := (n(τ )T Qn(τ ) + d(τ )T Rd(τ ))dτ (84)
−∞
where Q, R > 0. Once x̄ is found, the minimum energy state estimate x̂ is the most recent value of
x̄ i.e. x̂(t) = x̄(t). The MEE problem is solved by minimizing the quadratic cost
ˆ t
JM EE := ((C x̄ − y)T Q((C x̄ − y)) + dT Rd)dτ (85)
−∞
by appropriately choosing d(.). Note that the dynamic constraint on x̄ is given by Equation (83).
Proposition 16.4. Suppose u(.) and y(.) are given up to some time t > 0. For every symmetric
matrix P , differentiable signal β : (−∞, t] → Rn and a scalar H0 that does not depend on (d, x̄),
the functional
ˆ t
H(x̄, d) := H0 + ((x̄˙ − β̇)T P (x̄ − β) + (x̄ − β)T P (x̄˙ − β̇))dτ − (x̄(t) − β(t))T P (x̄(t) − β(t))(86)
−∞
Proof.
ˆ t
d(x̄(τ ) − β(τ ))T P (x̄(τ ) − β(τ ))
H(x̄, d) := H0 + dτ − (x̄(t) − β(t))T P (x̄(t) − β(t))
−∞ dτ
= H0
Choose
1. P such that −AT P − P A + C T QC − P B̄R−1 B̄ T P = 0.
2. β such that −AT P β −P B̄R−1 B̄ T P β +P Bu+C T Qy −P β̇ = 0. Using −AT P −P B̄R−1 B̄ T P =
P A − C T QC,
P A − C T QC + P Bu + C T Qy − P β̇ = 0
⇒ β̇ = (A − P −1 C T QC)β + Bu + P −1 C T Qy (91)
´t
3. H0 = −∞ y
T Qy + 2β T P (Bu − β̇) − β T P B̄R−1 B̄ T P β.
Therefore,
ˆ t
T
JM EE = H(x̄, d) + (x̄(t) − β(t)) P (x̄(t) − β(t)) + (d − R−1 B̄ T P (x̄ − β))T R(d − R−1 B̄ T P (x̄ − β))]dτ.(92)
−∞
Solving this in backward time and using the fact that limτ →−∞ y(τ ) = 0 and limτ →−∞ β(τ ) = 0,
limt→∞ (x̄(τ ) − β(τ )) = 0 if −(A + B̄R−1 B̄ T P ) is asymptotically stable. The minimum value of
JM EE is given by H(x̄, d) = H0 . This proves the following theorem (Hespanha)
Theorem 16.5. Assume that there exists a symmetric solution P to the ARE: −AT P − P A +
C T QC − P B̄R−1 B̄ T P = 0 such that matrices −(A + B̄R−1 B̄ T P ) and −(A − P −1 C T QC) are
asymptotically stable. Then MEE estimator for (83) is given by
This is a deterministic estimator. Let e = x − x̂. Then ė = ẋ − x̂˙ = (A − LC)e + B̄d − Ln where
A − LC is asymptotically stable. Thus, we have BIBO stability of MEE estimator w.r.t. d and n
which ensures that the estimate x̂ doesn’t diverge from x. In the absence of both d and n, we have
asymptotically converging estimates.
E[d(t)dT (t)] = δ(t − τ )R−1 , E[n(t)nT (t)] = δ(t − τ )Q−1 , Q, R > 0. (95)
Then the MEE estimate x̂ minimizes JLQG := limt→∞ E[kx(t) − x̂(t)k2 ] which is same as the trace
of E(x(t) − x̂(t))(x(t) − x̂(t))T . The MEE estimator for the above models of d and n is also called
Kalman-Bucy filter.
The proof of the following result is taken from Murray (Lecture notes).
Theorem 16.6 (Kalman-Bucy filter). Let ẋ = Ax + Bu + B̄d, y = Cx + n where d and n are
uncorrelated zero mean Gaussian white-noise stochastic processes with covariance matrices
E[d(t)dT (t)] = δ(t − τ )R−1 , E[n(t)nT (t)] = δ(t − τ )Q−1 , Q, R > 0. (96)
ė = (A − LC)e + ξ
⇒ Ṗ (t) = (A − LC)P + P (A − LC)T + B̄R−1 B̄ T + LQ−1 LT . (99)
For the explanation of the last equation, we refer the reader to Appendix, Equation (108) (This
derivation is taken from Friedland: Section 10.7, derivation of equation (10.58)).
We need to find L such that P (t) is as small as possible. Since (LQ−1 −P C T )Q(LQ−1 −P C T )T ≥ 0,
to make P (t) small, the positive term in the expression for Ṗ can be made zero. Thus, Ṗ (t) =
AP + P AT − P C T QCP + B̄R−1 B̄ T and P (0) = E{x(0)x(0)T }.
Suppose we design a state feedback controller u = −Kx that solves an LQR problem and
constructed an LQG/MEE state estimator x̂˙ = (A − LC)x̂ + Bu + Ly. If states are not available,
we may use u = −K x̂. Therefore,
x̂˙ = (A − LC − BK)x̂ + Ly
ẋ = (A − BK)x + BKe + B̄d
ė = (A − LC)e + B̄d − Ln
˙
x A − BK BK x B̄ 0
= + d+ n. (100)
e 0 A − LC e B̄ −L
Thus, the separation principle holds for LQG problem as well. By choosing K and L, the closed loop
system can be made asymptotically stable. This ensures that both x and e do not diverge. Note that
the separation principle says that we can design optimal estimator and optimal regulator separately,
and combining them together solves an optimal regulation problem using output feedback.
• If the number of outputs is greater than the number of inputs, we have an overdetermined
system of equations and it may or may not have a solution.
• If the number of outputs is equal to the number of inputs, then the matrix in the previous
equation is the Rosenbrock system matrix evaluated at s = 0. Thus, if the origin is not an
invariant zero, then there is a unique equilibrium point.
• If the number of outputs is less than the number of inputs, we have an under-determined
−x∗
system of equations and it has multiple solutions. In general, we may write =
u∗
Fx
r.
Fu
Let x̄ = x − x∗ , ū = u − u∗ and ȳ = y − r. Consider the following cost functional
ˆ ∞
JLQR = (ȳ T Qȳ + ρūT Rū)dt, ρ > 0. (102)
0
Note that Q penalizes the deviation from the reference signal and R penalizes deviation from the
equilibrium input. Observe that
Thus, we now have a regulator problem and the optimal solution is given by ū = −K x̄. Thus
u = ū + u∗ = −K(x − x∗ ) + u∗ .
We showed that the tracking problem can be transformed into an equivalent LQR (regulator)
problem. If states are not available, then to implement feedback laws, we need an estimator
(Hespanha). In the presence of noise and disturbance, we need a linear quadratic estimator to
estimate states x̂ and we use u = −K x̂.
Tracking non constant reference signals: Suppose we want x(t) to track a trajectory r(t).
Suppose (A, B) is controllable and there exists u∗ (t) such that ṙ = Ar(t) + Bu∗ (t). Let x̄(t) =
x(t) − r(t) and ū = u − u∗ . Therefore,
and one needs to choose ū such that the above system is asymptotically stable. For output tracking,
suppose there exists x∗ , u∗ such that ẋ∗ = Ax∗ + Bu∗ and r(t) = Cx∗ + Du∗ . Then, using
x̄ = x − x∗ , ȳ = y − y ∗ and ū = u − u∗ , one can solve the problem. We refer the reader to ([7]) for
building controllers to track non constant reference signals.
17 Appendix
The following definitions are from Papoulis and Pillai [15].
Definition 17.1. A random variable is a function from a sample space to C or R. (A sample space
is the set of all possible events/outcomes of an experiment.)
All random variables have a probability density function (distribution) which allows us to
compute probabilities of different events. A normal or Gaussian distribution is the one where
−(x−µ)2
the probability density function f is of the form of a Gaussian f (x) = √ 1 e 2σ 2 where µ is
2πσ 2
the mean and σ 2 is the variance.
Definition 17.2. A random variable X is said to be zero mean if its expectation E(X) = 0, for a
random vector X̄, zero mean ⇒ E(X̄) = 0.
Definition 17.3. Two random variables X and Y are said to be uncorrelated if their covariance
C(X, Y ) = E(XY ) − E(X)E(Y ) = 0. For random vectors X̄, Ȳ , they are said to be uncorrelated
if their covariance matrix C(X̄, Ȳ ) = E(X̄ Ȳ T ) − E(X̄)E(Ȳ )T is diagonal.
Definition 17.4. A function x(t, ξ) is said to be a stochastic process if for a fixed t0 , x(t0 , ξ) is a
random variable.
A stochastic process x(t) is said to be white noise if its values x(ti ) and x(tj ) are uncorrelated
for every ti 6= tj . A white noise Gaussian stochastic process is the one where each random variable
is zero mean and Gaussian with any two random variables being uncorrelated.
A correlation matrix is given by Rx (t, τ ) = E{x(t)xT (τ )}.
Consider a following system
ẋ = Ax + F v, y = Cx
where v is a white Gaussian noise. Thus, E(v) = 0. Let Φ(t, t0 ) be the state transition matrix.
Thus,
ˆ t
x(t) = Φ(t, t0 )x(t0 ) + Φ(t, λ)F (λ)v(λ)dλ.
t0
Then,
ˆ τ
T T T
x(t)x (τ ) = Φ(t, t0 )x(t0 )x (t0 )Φ (τ, t0 ) + Φ(t, t0 )x(t0 ){ Φ(τ, λ)F (λ)v(λ)dλ}T +
t0
ˆ t ˆ tˆ τ
T T
Φ(t, λ)F (λ)v(λ)dλ.x (t0 )Φ (τ, t0 ) + Φ(t, λ)F (λ)v(λ)v T (ξ)F T (ξ)ΦT (τ, ξ)dξdλ. (103)
t0 t0 t0
Taking expectation on both sides in the previous equation and observing that E(v) = 0,
ˆ tˆ τ
Rx (t, τ ) = Φ(t, t0 )E(x(t0 )xT (t0 ))ΦT (τ, t0 ) + Φ(t, λ)F (λ)E(v(λ)v T (ξ))F T (ξ)ΦT (τ, ξ)dξdλ.(104)
t0 t0
Let E(v(λ)v T (ξ)) = Qδ(λ − ξ) and E(x(t0 )xT (t0 )) = P (t0 ). Thus,
ˆ tˆ τ
Rx (t, τ ) = Φ(t, t0 )P (t0 )ΦT (τ, t0 ) + Φ(t, λ)F (λ)Qδ(λ − ξ)F T (ξ)ΦT (τ, ξ)dξdλ
t0 t0
ˆ t
T
= Φ(t, t0 )P (t0 )Φ (τ, t0 ) + Φ(t, λ)F (λ)QF T (λ)ΦT (τ, λ)dλ (105)
t0
Observe that
ΦT (τ, λ) = [Φ(τ, t)Φ(t, λ)]T = ΦT (t, λ)ΦT (τ, t). (106)
Substituting (106) in (105),
ˆ t
T
Rx (t, τ ) = Φ(t, t0 )P (t0 )Φ (τ, t0 ) + Φ(t, λ)F (λ)QF T (λ)ΦT (t, λ)ΦT (τ, t)dλ
t0
= Rx (t, t)ΦT (τ, t) (107)
Let P (t) := E(x(t)xT (t)) = Rx (t, t). Therefore,
ˆ t
d d
Rx (t, t) = (Φ(t, t0 )P (t0 )ΦT (t, t0 ) + Φ(t, λ)F (λ)QF T (λ)ΦT (t, λ)dλ)
dt dt t0
⇒ Ṗ (t) = AP + P AT + F QF T (108)
where the last equality follows by using differentiation under the integral sign.
Loop shaping: The idea behind loop shaping is to manipulate block transfer functions such that
one obtains a desired transfer function (which gives desired characteristics such as time response,
stability etc.). Using the Nyquist criterion for stability of a closed loop transfer function using
an open loop transfer, one manipulates open loop transfer function to obtain a desired frequency
response. This method is used frequently in design problems. MFDs are used for loop shaping.
We refer the reader to Doyle, Francis and Tannerbaum, Skogestad and Postlaithwaite for further
details. While implementing feedback laws using state estimators (e.g. linear quadratic estimator
(LQE)), open loop gain gets affected. It is shown in Hespanha (Chapter 24) that one can recover
open loop gain if certain conditions are satisfied. This is called loop transfer recovery (LTR).
Kalman equality:
GT H = 0, K = R−1 B T P, L̂(s) = K(sI − A)−1 B, L̂T (−s) = −B T (sI + AT )−1 K T ,
T̂ (s) = G(sI − A)−1 B + H, T̂ T (−s) = −B T (sI + AT )−1 GT + H T
T̂ T (−s)T̂ (s) = −B T (sI + AT )−1 GT G(sI − A)−1 B + H T H (109)
T T −1 T T T −1 T
A P + P A + G G − P BR B P = 0 ⇒ A P + sP − sP + P A + G G − P BR B P =0
pre multiplying by -B T (sI + AT )−1 and post multiplying by (sI − A)−1 B,
T −1
T
⇒ −B (sI + A ) ((sI + A )P − P (sI − A) + GT G − P BR−1 B T P )(sI − A)−1 B = 0
T
References
[1] J. Hespanha, Linear systems theory, Princeton university press, 2009.
[2] W.M. Wonham, Linear Multi-variable Control A Geometric Approach, third edition, Springer-
Verlag, 1985.
[3] R. W. Brockett, Finite Dimensional Linear Systems, John Wiley and Sons, Inc., 1970.
[5] B. Friedland, Control system design : An introduction to state-space methods, Dover publica-
tions inc., 1986.
[7] A. Astolfi, Tracking and Regulation in Linear Systems, Encyclopedia of Systems and Control
Springer-Verlag London, 2014.
[8] H. Kwakernaak, R. Sivan, Linear Optimal Control Systems, Willey Interscience, 1972.
[9] S. Skogestad, I. Postlaithwaite, Multi-variable Feedback Control Analysis and design, John Wil-
ley and sons, 2007.
[10] J. Doyle, B. Francis, A. Tannerbaum, Feedback Control Theory, Macmillan Publishing Co.,
1990.
[11] F. Pasqualetti, A Systems and Control Perspective on Privacy, Safety, and Security in Large-
Scale Cyber-Physical Systems, Presentation slides from Disc summerschool talk, the Hague, the
Netherlands, 2017.
[12] H. Fawzi, P. Tabuada, S. Diggavi, Secure state-estimation for dynamical systems under active
adversaries, IEEE Transaction on Automatic Control, vol. 59, no.6, pp. 1454-1467, 2014.
[13] P. M. Van Dooren Numerical Linear Algebra for Signals Systems and Control, Lecture notes
for Graduate School in Systems and Control, 2003.
[14] K. Zhou, J. Doyle, K. Glover, Robust and Optimal Control, Prentice hall, 1996.
[15] A. Papoulis, U. Pillai, Probability, Random Variables and Stochastic Processes, fourth edition,
Tata-McGraw hill, 2002.
[16] A. Antoulas, Approximation of Large-Scale Dynamical Systems, Advances in Design and Con-
trol, SIAM, 2005.