Adaptive Lecture04 2005
Adaptive Lecture04 2005
In many applications, plant (model) structure may be known, but its parameters
may be unknown and time-varying due to change in operation conditions, aging of
equipment, etc. Thus, the off-line parameter estimation is inefficient.
On-line estimation schemes refer to those estimation schemes that provide frequent estimates of plant parameters by properly processing the plant I/O data on-line.
The essential idea behind is the comparison of the observed system response y(t) with
the output of a parameterized model y (, t ) , whose structure is the same as that of
plant model. Then, (t) is adjusted continuous so that y (, t ) approaches y(t) as t increases. (Under certain input conditions, y being close to y implies that (t) is close
to the unknown .)
*
Remark 4.0.1: In adaptive control, where the convergence of (t) to is usually not
*
one of the objectives, the first two steps are the most important ones.
(4.1.1)
where is unknown, and y(t) and u(t) are measurable. If u and y are measured in a
*
y (t )
, u (t ) 0 .
u (t )
(ii)
noise effect of measurements of y(t) and u(t) will cause wrong estimations.
~
1 = y y = u u = u,
~
where = , and define
1
1
J ( ) = 12 = ( y u ) 2 .
2
2
Then, adjustment of is trying to minimize J(), which naturally leads to the gradient method as:
& = J () = [ y u ]u = 1u , (0) = 0 ,
where > 0 is a scaling constant. For stability analysis of the estimator, construct the
following Lyapunov function candidate:
~
1 ~2
V ( ) =
,
2
~
~
~&
subject to = & = 1u , then V& ( ) = 12 0, which implies that L (or L ).
Since
~&
Barbalats Lemma, 1(t) 0 as t and hence (t ) 0 as t .
On the other hand, note that
t
u 2 ( ) d ~
~
(t ) = e 0
(0).
~
Therefore, (t ) converges to zero if and only if
t + T0
t
u 2 ()d 0T0 , t 0 ,
Consider the plant (4.1.1) again and assume u and y are piecewise continuous but not
u, y L .
An alternative is y u , y =
chosen so that
u
m
y
m
, u=
u
m
1
m2
~&
~
~
where , which implies that = & = u 2 . Construct
~2
~
V ( ) =
.
2
Then,
~
V& = 2u 2 = 2m2 0
~
~
~
such that , L , m L2. Because u , L , if follows that = mu and m
d
dt
~&
that m(t) 0 as t by Barbalats Lemma. Then, (t ) 0 as t .
Remark 4.2.1: Despite y, u may be unbounded, adaptive law can still be designed
x& = Ax + Bu,
y = C T x.
Two kinds of parameterizations: (i) y = T + 0 , or (ii) y = W ( s )T + 0 , where
= [bn1 , bn2 , K, b1 , b0 , an1 , an2 , K, a1 , a0 ]T , = b , b = [0, T ]T ,
u
= H1 ( s ) ,
y
u
= H (s) ,
y
and W(s) is one strictly proper transfer function with stable poles, stable zero (i.e.,
minimum phase) and relative-degree one. A general model is the form of
z = W (s )T , which is linear parameter model or linear regression model.
1
Choose L(s) so that L (s) is a proper stable transfer function and W(s)L(s) is a
proper SPR transfer function. Then
z = W ( s ) L( s )T ,
= L1 ( s ) ,
L , m 2 = 1 + ns2 .
m
(Note that the example of ns is ns2 = T , or ns2 = T P , where P > 0.) Thus,
~
= W ( s) L( s)( T ns2 ) ,
~
where . Now, let (Ac, Bc, Cc) be the state space representation of W(s)L(s),
then
~
e& = Ac e + Bc ( T ns2 ),
= CcT e,
and
W ( s ) L( s ) = CcT ( sI Ac ) 1 Bc .
Since W(s)L(s) is SPR, there is a matrix Pc > 0 such that
Pc Ac + AcT Pc = qq T Lc ,
Pc Bc = Cc ,
for some vector q, matrix Lc > 0, and a small constant > 0, by either KYL Lemma (if
(Ac, Bc, Cc) is minimal) or MKY Lemma (if (Ac, Bc, Cc) is nonminimal). To design the
adaptive law, construct a Lyapunov like function
~
~
~
e T Pc e T 1
V ( , e) =
+
,
2
2
where > 0 such that
~
~
~&
1
V& = eT qq T e e T Lc e + eT Pc Bc ( T ns2 ) + T 1
2
2
~
~&
1
= eT qq T e e T Lc e 2 ns2 + T 1 ( ),
2
2
(4.2.1)
~
~&
i.e., = . Obviously, e, , , L , and , ns L2 , so that m L2 and
& = (m ) L2 ,
m
, L ,
(ii) , ns , & L2 ,
known vector . Such a property is achieved for a special class vector signals described by the following definition.
Definition 4.2.1 (Persistence of Excitation (PE)): A piecewise continuous signal
n
vector : R + R n is PE in R with a level of excitation 0 > 0, if there are constants 1, T0 > 0 such that
1I
1
T0
t + T0
t
()T ()d 0 I , t 0 .
1
T0
t +T0
t
[q T ()]2 d 0 , t 0 , where q
Assume that there exits constants > 0, kv 0, such that for all t0 0, K(t) R
sat-
isfies
t0
nl
t0 + v
K () d kv
2
nn
,AR
KC ) is a UCO pair.
n
Lemma 4.2.2: If w: [0, ) R is PE, w& L , and H(s) is a stable, minimum phase,
satisfies
2I
1
T0
t + T0
f ()Tf ()d 1I , t 0
~&
=
= Cc e
(4.2.2)
that describe the stability properties of the adaptive law. In proving the Theorem 4.1.1,
we have also shown that the time derivative of
& = J () = ,
where = T > 0 and Pc = PcT > 0 satisfies
V& 2
for some constant > 0. Defining
A BcCcT ns2
A(t ) = c
T
Cc
Bc T
T
, C = [Cc
0
0]T , P =
1 Pc
2 0
0
1
we rewire (4.2.2) as
x& = A(t ) x, = C T x
and express the above Lyapunov-like function V and its derivative as
V = x T Px
V& = x T ( PA + AT P + P& ) x x TCC T x = 2
B n2
K c s ,
7
is bounded (see Lemma 4.1.1). We can therefore establish that (4.2.1) is UCO by
showing that is a UCO pair. The system corresponding to (C, A + KC ) is as
T
f () CcT e Ac ( ) Bc ()d
t
is also a SPR (refer to Lemma 4.1.2); therefore, there exists constants 1, 2, T0 > 0
such that
2I
1
T0
t + T0
t
f ()Tf ()d 1I , t 0
We can conclude that (C, A + KC ) is UCO (see Lemma 4.1.3) which implies (C, A) is
~
UCO. Hence, we conclude that the equilibrium xe = 0 (i.e., ee = 0 and e = 0) is e.s. in
T
the large.
1
If W(s) is minimum phase, one may choose L(s) = W (s) leading to W(s)L(s) = 1.
Then,
=
~T
1
=
.
2
m
m2
Consider
~ T 1~
~
V ( ) =
2
so that V& = 2 m 2 provided the adaptive law is chosen
& = .
2 m 2 ( z T ) 2
=
(Quadratic cost function),
2
2m 2
where
8
z T
m2
, ns , , & L ,
(ii) , ns , & L2 ,
~&
= .
(4.3.1)
~ T 1~
.
2
(4.3.2)
(4.3.3)
~
Hence, V , L , which implies that , m L . In addition, we establish from the
properties of V, V& that m L2 , which implies that , ns L2 . Now, from the
adaptive law, we have
~& &
= m
m
which together with
where
(4.3.4)
A(t ) =
T
T
T
,
C
(
t
)
=
,
m2
m
y0 = m
This system is analyzed using the Lyapunov-like function (4.3.2) that led to (4.3.3)
along the trajectory of this adaptive law. We need to establish that the equilibrium
~
e = 0 of (4.3.4) is e.s. We achieve that by using Theorem 3.3.4 (See Lecture 3) as
1
~T ~
P
2
and
~
~
~
1~
V& = T ( PA + AT P + P& ) = TCC T
2
where P& = 0 . This implies that
PA(t ) + AT (t ) P + 2C (t )C T (t ) 0
~
According to Theorem 3.3.4, e = 0 is e.s. provided (C, A) is UCO. Using Lemma
4.2.1, we have that (C, A) is UCO if (C, A + KC ) is UCO for some that satisfies the
T
(C, A + KC ), i.e.,
T
Y& = 0
y0 = C T (t )Y =
T
m
(4.3.5)
t +T
t
()()
d
m 2 ()
plete.
4.2.1 The Gradient Algorithm Based on Integral Cost
10
J () =
1 t ( t ) 2
e
(t , )m 2 ()d (Integral cost function),
0
2
where
(t , ) =
z () T (t )()
, (t , t ) =
m 2 ( )
is the normalized estimation error at time based on the estimate (t) of at time
t . We have
t
z () T (t )()
()d , (Integral adaptive law)
& = J () = e ( t )
0
m 2 ()
i.e. (t) is chosen at each time t to minimize the integral square of the errors on all
past data that are discounted exponentially. That is,
& = [ R(t ) + Q(t )]
R& = R + 2 ,
m
z
Q& = Q 2 ,
m
T
R(0) = 0,
R R n n
Q(0) = 0,
Q Rn
, ns , , & L ,
(ii) , ns , & L2 ,
(iii) lim & (t ) = 0 , and
t
Proof: Because
system with bounded input. Substituting for z = T in the differential equation for
Q, we verify that
t
Q = e
(t )
()T ()
d = R(t ) ,
2
m
and hence,
~&
~
& = = R(t ) .
(4.3.6)
(4.3.7)
such that
~
~
V& = T R (t ) 0 .
(4.3.8)
~
T
Since R(t) = R (t) 0, t 0, it follows that V, L ,
1~
~ ~ 1
( T R ) 2 = R 2 L2 .
~T
~
From = m 2 and ,
1~
which together with R L and R 2 L2 L imply that & L2 L . Since
~& &
~&
, R L , it follows from (4.3.6) that && L , which together with L2 , implies
that
~
lim & (t ) = lim R (t ) (t ) = 0 .
t
t
0
t~
t~
~ ~
~
~
2 m 2 d = T R + 2 T RR d + T R d .
0
1~
~
~
Because T (t ) R(t ) (t ) 0 as t 0 , and R 2 L2 , it follows that
lim 2 m 2 d = 2 m 2 d < .
t
i.e. m L2 .
The proof for (iv) is given now. In proving (i) to (iii), we have shown (4.3.8)
from (4.3.7). From the differential equation on R, we have
t
R (t ) = e ( t )
0
12
()T ()
d
m 2 ()
R (t ) =
t
t T0
(t )
0e T0
T
t T0
()T ()
( t ) ( ) ( )
d +
e
d
0
m 2 ()
m 2 ()
t
t T0
()T ()d
1e T0 I
for any t T0, where 1 = 00T0 , 0 = sup t
1
m2 ( t )
we have that
max ( ) ~
( t T )
(T0 ) e 2 0
min ( )
~
( t T )
(t ) 2 max ( )V (t0 )e 2 0
and m.
Remark 4.3.2: (t) converges to a trajectory that minimizes the integral asymptoti-
(t t 0 )
max ( ) ~
, t T0 ,
(t0 ) e 2
min ( )
1
m2 ( t )
defining PE of . Hence, larger 0 and larger min () will guarantee faster convergence of (t ) to zero.
Bad approach:
(t ) =
y ( )
d ( )
= + n
u ()
u ( )
J () =
1 t
( y () (t )u ())2 d,
2 0
t
(t ) = u 2 ()d
t
0
1 t
1 t
y
(
)
u
(
)
d
+
lim
d n ()d = .
t t 0
t t 0
lim (t ) = lim
t
For general linear model, z = T , the estimate z of z and the normalized estimation error are generated as: z = T , and
=
J () =
z z z T
2
=
, m2 = 1 + ns ,
m2
m2
1 t ( t ) ( z () T (t )()) 2
1
e
d + e t ( 0 )T Q0 ( 0 ) ,
2
2 0
m ()
2
z
m
over R at each time t. Hence, any local minimum is also global and satisfies
J ((t )) = 0 , t 0,
i.e.
J () = e t Q0 ((t ) 0 )
1 t ( t ) z () T (t )()
e
()d = 0
2 0
m 2 ( )
z ()()
(t ) = P(t ) e t Q00 + e ( t )
d ,
0
m 2 ()
where
1
()T ()
P(t ) = e t Q0 + e ( t )
d .
0
m 2 ( )
14
Because Q0 = Q0T > 0 , and () () is positive semi-definite, P(t) exists at each time
T
t. Then
d
d
PP 1 = P& P 1 + P P 1 = 0
dt
dt
1
P& = P P 2 P , P (0) = P0 = Q0 ,
m
T
& = P
Since
t
z ()()
d ,
(t ) = P(t ) e t Q00 + e ( t )
0
m 2 ()
it follows that
t
P 1 (t )(t ) = e t Q00 + e ( t )
0
()T ()
d ,
2
m ( )
Hence,
t
()T ()
(t )T (t )
+
.
0
m 2 ( )
m 2 (t )
Recall that
(t ) (t )
P& 1 (t ) = P 1 (t ) +
.
m 2 (t )
T
Therefore,
(t ) (t )
P (t )& (t ) = P& 1 (t )(t ) ( P 1 (t )(t )) +
m 2 (t )
T
(t )T (t )
()T ()
1
(
t
)
(
P
(
t
)
(
t
))
+
= P 1 (t ) +
m 2 (t )
m 2 ()
( z z )(t ) 1 (t )
=
= 2 (t ) = (t )(t ),
m 2 (t )
m (t )
and hence, the least square is given by
& (t ) = P(t )(t )(t ).
4.4.1 Pure Least-Squares Algorithm
Set = 0, we have
15
P P &
, = P ,
P& =
m2
T
which implies that P may grow without bound. In the matrix case, this means P may
become arbitrarily small and slow down adaptation in some directions. This is the
so-called covariance wind-up problem that constitutes one of the main drawbacks of
the pure least-squares algorithm.
Theorem 4.4.1: The pure least-squares algorithm guarantees that
(i)
, ns , , & , P L ,
(ii) , ns , & L2 ,
(iii) limt (t ) = , where is a constant vector, and
(iv) if ns, L, and is PE, then (t) converges to * as t .
Proof: We have that P& 0 , i.e. P(t) P0. Because P(t) is nonincreasing and bounded
limt P(t ) = P .
where P = P T 0 is a constant matrix. Let us now consider
T~
d 1~
&
1 & 1~
1~
+ = 0,
( P ) = P PP + P =
dt
m2
~
~
1~
1~
Hence, P 1 (t ) (t ) = P0 (0), and therefore, (t ) = P (t ) P0 (0), and
~
1~
lim (t ) = P P0 (0),
t
~
~
~
Because P(t) P0 and (t ) = P(t ) P01 (0) , we have , L , which, together with
~T
function. Consider
~T 1~
P
,
V=
2
The time derivative of along the trajectory of this adaptive law is given by
16
~T T~
2m2
2m2
~T
2 2
&
V = +
=
m
+
=
0
2m 2
2
2
which implies that V L, m L2; therefore, , ns L2, we have. From the adaptive law, we have
& P
m
m
Because P,
t +T
t
T
t
()T ()
T
d n0 0 0 I 1 0 0 I
2
0
m ()
m
m
T0
t
Therefore
T
t
T
t
P 1 (t ) P 1 (0) + 1 0 0 I 1 0 0 I , t T0
T0 m
T0 m
t
t
P (t ) 1 0 0 I = 1 0T0 m I , t T0
m
T0
T0
Since P(t) 0 for all t 0 and the right-hand side of the above inequality goes to zero
asymptotically, we can conclude that P(t) 0 as t as. Hence the proof of (iv) is
complete.
P (t )
m
~
P01m ~
I , (t )
(0) , t T0 ,
(t T0 ) 0
(t T0 ) 0
~
where m = sup t m 2 (t ) . i.e. (t ) is guaranteed to converge to zero with a speed of
1/t.
4.4.2 Pure Least-Squares with Covariance Resetting
To avoid covariance wind-up problem, we modify least-squares by incorporating covariance resetting mechanism:
& = P,
P P
P& =
, P(t r+ ) = P0 = 0 I ,
m2
T
where tr is the time for which min ( P (t )) 1 , and 0 > 1 > 0 are some design scalars
such that P (t ) 1I , t 0 .
Theorem 4.4.2: The pure least-squares algorithm with covariance resetting has the
following properties:
(i)
, ns , , & L ,
(ii) , ns , & L2 ,
(iii) if ns, L, and is PE, then (t) converges to exponentially fast.
*
Proof: The covariance matrix P(t) has elements that are discontinuous functions of
time due to the resetting. At the discontinuity or resetting point tr, P(tr+ ) = P0 = 0 I ,
therefore, P 1 (tr+ ) = 01I . Between discontinuities,
d
dt
P 1 (t ) 0 , i.e., P (t2)
18
1 ~ d ( P 1 ) ~ ~ T 1~&
1 ~ d ( P 1 ) ~
+ P = 2 m 2 + T
V& = T
dt
dt
2
2
Between the resetting points, it follows
1 (T ) 2
2m2
V& = 2 m 2 +
=
0
2 m2
2
t [t1, t2], where [t1, t2] is any interval in [0, ) for which tr [t1, t2]. On the other
hand, at the discontinuity of P, we have
V (tr+ ) V (tr ) =
~
1 ~ T 1 +
[ P (tr ) P 1 (tr )].
2
When > 0, the problem of P(t) becoming arbitrarily small in some directions no
longer exist. But P(t) may grow without bound since P& > 0 for P > 0 . Thus,
modification is the following:
& = P
PT P
P
m2
P& =
0,
if P(t ) P0
otherwise
for estimates of in the set where is located. The advantages lie in:
*
(i)
to speed up convergence,
(ii) to reduce large transients when (t) is far away from , and
*
(iii) to constrain (t) such that it always satisfies certain properties t (Con-
& = Pr(J )
gg T
J ,
otherwise
J
g T g
or
,
if Int( S ) or if ( S ) and ()T g 0
& = Pr(+) =
gg T
,
otherwise
g T g
Brief Proof: Whenever (S ) , & g 0 , which implies that & points either in-
side or along the tangent plane of (S ) at point . Because (0) S , it follows that
(t) will never leave S, t 0 .
Next, the difference of the adaptive law, when projection is applied, lies in the
additional term:
Q=
g g T
J .
g T g
20
To show that this additional term will not try to make V& more positive, we see that
~ T g g T
~ T 1
J ,
Q = g T g
0 ,
otherwise
if ( S ) and (J )T g > 0
~
and that Tg = ( )T g 0 when (S ) because S is convex. Therefore,
~
( T g )(J ) T g
~T 1
0,
Q=
g T g
0 ,
otherwise
if ( S ) and (J ) T g > 0
~
In other words, the additional term T 1Q in V& introduced by projection can only
make V& more negative.
& = Pr ( P) =
gg T
otherwise
P P g T g P,
PT P
P
,
&
P=
m2
0,
z = W ( s ) L( s )[ (T + z1 )]
~
~T which implies that
Now T T = T
~
~ n 2 ], = T + z
= W ( s ) L( s )[ T
s
1
= CcT e
(4.6.1)
where Pc = PcT > 0 satisfies the algebraic equations implied by the KYL Lemma, and
= T > 0 , > 0. Along the trajectory of the adaptive law, we have
~
~
~& ~
eT qq T e T
~
&
V& =
e Lc e T ~
2 ns2 + T 1 +
2
2
& = & =
we have
22
(4.6.2)
e T qq T e T
V& =
e Lc e 2 ns2 0
2
2
, , L ,
lowing the same procedure as the linear parametric model case and is left as an exercised for the students. The proof of (iii) is establish by using the results of Corollary
treated as an external input
4.2.1 to show the homogeneous part of (4.6.1) with ~
L2 and Ac is
together with the equation of (4.6.2) form an e.s. system. Since ~
~
stable, it follows that e(t ), (t ) 0 as t . The proof of (iv) follows from ,
L2 and the inequality
1
2
2
2
2
0 & () d 0 ()() d 0 ()d 0 ()d <
Remark 4.6.1: The lack of convergence of to is due to L2. If, however, are
such that is PE, then we can establish by following the same approach as in the proof
in Corollary 4.2.1 that converge to zero exponentially fast. For L2, the vector [ ,
T
z z z (T + z1 )
=
m2
m2
23
& =
& =
for
1 =
= sgn( )
n+1
because
of the dependence of on . Let us, however, ignore this dependence and treat as an
independent function of time.
The performance of this adaptive law is summarized in the following.
Theorem 4.6.2: This instantaneous adaptive law guarantees that:
(i)
, ns , , & , & L ,
The proof from that of the linear parametric model and Theorem 4.6.1 and is left as an
exercise for the students.
The extension of the integral adaptive law and least-squares algorithms to the
bilinear parametric model is more complicated and difficult to the implement due to
the appearance of the unknown in the adaptive laws. This problem is avoided the
t k +1
tk
T
1, m 1,
m2
and
~
where k = k . We have
~
~
~
V (k ) = (2 k + k )T 1 k
tk
tk
~T
Because m 2 = k (t ) and m (= max () ) , we have
2
V (k ) 2
t k +1
tk
t k +1
()
d .
()m ()d + m ()m()
t
m()
k
2
t k +1
t k +1 ( )
t k +1
()
d
()m()
d 2 ()m 2 ()d
t
tk
tk
m ( )
m()
k
Ts
t k +1
tk
such that
25
2 ()m 2 ()d
V (k ) = (2 Ts m )
t k +1
tk
2 ()m 2 ()d
t k +1
0
2 ()m 2 ()d
V (0) V (k + 1)
(2 Ts m )
t k +1
tk
2 ()m 2 ()d
such that
T
k
k =1
Ts m
which implies k l2 , and thus completes the proof of (i) and (ii).
The proof of (iii) is given now. From the proof of (1) and (ii), we have
V (k ) = V (k + 1) V (k ) = (2 Ts m )
t k +1
tk
2 ()m 2 ()d
which implies
n 1
V (k + n) V (k ) = (2 Ts m )
i =1
t k +i +1
t k +i
2 ()m 2 ()d
(4.7.1)
t k +i +1
t k +i
~
[ kT+1()]2
d
()m ()d =
t k +i
m 2 ( )
~T
~
~ T
2
t k +i +1 [ ( ) + (
k
k +1 k ) ( )]
d
=
t k +i
m 2 ()
2
t k +i +1
Using ( x + y ) 2 12 x 2 y 2 , we write
Since
t k +i +1
t k +i
( )
m ( )
2 ()m 2 ()d
~
~
~ T
2
t k +i +1 [(
1 t k +i+1 [ kT+1()]2
k +1 k ) ( )]
d
d
t k +i
2 t k +i
m 2 ()
m 2 ()
is bounded, we have
t k +i +1
t k +i
~
~
~
~
[( k +1 k )T ()]2
d cTs k + i k
2
m ( )
where
26
(4.7.2)
()
m 2 ()
2
c = sup
( )
m ( )
t k +n
~
~ 2
k + i k ciTs 2 ()m 2 ()d,
tk
(4.7.3)
t k +i +1
t k +i
~
t k +n
1 t k +i +1 [ kT+1()]2
()m ()d
d c 2iTs2 2 ()m 2 ()d
2
t k +n
2 t k +i
m ( )
2
which leads to
t k +n
tk
n 1
2 ()m 2 ()d =
i =0
t k +i +1
2 ( ) m 2 ( ) d
t k +i
~
t k +n
1 t k +i+1 [ kT+1()]2
d c 2iTs2 2 ()m 2 ()d
2
t
t
k +n
m ()
i = 0 2 k +i
n 1
t k +n
tk
~
~ T t k +i+1 [ kT+1()]2 ~
1
dk
()m ()d
k
2[1 + n(n 1)c 2Ts2 / 2] t k +i
m 2 ()
2
(4.7.4)
t + T0
t
~
[ kT+1()]2
d 1 I
m 2 ()
for any t. Hence, for any integer k, n where n satisfies nTs T0 we have
~
~ T t k +i +1 [ kT+1()]2 ~
~ ~ V (k )
k
1
dk 2 kT k
2
t k +i
m
m ( )
Using (4.7.4) and (4.7.5) in (4.7.1), we obtain the following inequality:
V ( k + n) V ( k ) =
(2 Ts m )1
V (k )
2[1 + n(n 1)c 2Ts2 / 2]
V (k + n) V (k )
with
=1
(2 Ts m )1
<1
2[1 + n(n 1)c 2Ts2 / 2]
27
(4.7.5)
Therefore,
( )k
m
m
~
yields that kn 0 exponentially, which together with the property of the hybrid
~
~
adaptive algorithm (i.e., k +1 k ), implies k converges to exponentially. The
ASSIGNING READING
28