0% found this document useful (0 votes)
37 views

Fault Tolerant Process Control c2

This document provides an overview of nonlinear systems and control theory, including definitions of stability, Lyapunov stability, and asymptotic stability. It also introduces common notation used in analyzing nonlinear systems and control. The review is intended to provide necessary background knowledge before presenting results in subsequent chapters.

Uploaded by

augur886
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
37 views

Fault Tolerant Process Control c2

This document provides an overview of nonlinear systems and control theory, including definitions of stability, Lyapunov stability, and asymptotic stability. It also introduces common notation used in analyzing nonlinear systems and control. The review is intended to provide necessary background knowledge before presenting results in subsequent chapters.

Uploaded by

augur886
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 21

Chapter 2

Background on Nonlinear Systems and Control


In this chapter, we review some basic results on the analysis and control of nonlinear
systems. This review is not intended to be exhaustive but to provide the reader with
the necessary background for the results presented in the subsequent chapters. The
results presented in this chapter are standard in the nonlinear systems and control
literature. For detailed discussion and proofs of the results, the reader may refer to
the classic books [72, 76].
2.1 Notation
Throughout this book, the operator | | is used to denote the absolute value of a scalar
and the operator is used to denote Euclidean norm of a vector, while we use

Q
to denote the square of a weighted Euclidean norm, i.e., x
Q
=x
T
Qx for all
x R
n
. The symbol
r
is used to denote the set
r
:={x R
n
: V(x) r} where
V is a scalar positive denite, continuous differentiable function and V(0) =0, and
the operator / denotes set subtraction, that is, A/B := {x R
n
: x A, x / B}.
The notation R = [r
1
r
2
] is used to denote the augmented vector R R
m+n
com-
prising the vectors r
1
R
m
and r
2
R
n
. The notation x(T
+
) denotes the limit of
the trajectory x(t ) as T is approached from the right, i.e., x(T
+
) = lim
t T
+ x(t ).
The notation L
f
h denotes the standard Lie derivative of a scalar function h() with
respect to the vector function f (), i.e., L
f
h(x) =
h
x
f (x).
2.2 Nonlinear Systems
In this book, we deal with a class of time invariant nonlinear systems that can be
described by the following state-space model:
x =f (x, u), (2.1)
P. Mhaskar et al., Fault-Tolerant Process Control, DOI 10.1007/978-1-4471-4808-1_2,
Springer-Verlag London 2013
9
10 2 Background on Nonlinear Systems and Control
where x R
n
denotes the vector of state variables, u R
m
denotes the vector of
control (manipulated) input variables, and f is a locally Lipschitz vector function
on R
n
R
m
such that f (0, 0) = 0. This implies that the origin is an equilibrium
point for the unforced system. The input vector is restricted to be in a nonempty
convex set U R
m
which is dened as follows:
U :=
_
u R
m
: u u
max
_
, (2.2)
where u
max
is the magnitude of the input constraint. Another version of the set that
we will use is
U
con
:=
_
u R
m
: u
min
i
u
i
u
max
i
, i =1, . . . , m
_
, (2.3)
where u
min
i
and u
max
i
denote the constraints on the minimum and maximum value
of the ith input.
In many chapters, we will restrict our analysis to a special case of the system of
Eq. (2.1) where the input vector u enters the dynamics of the state x in an afne
fashion as follows:
x =f (x) +G(x)u, (2.4)
where f is a locally Lipschitz vector function on R
n
such that f (0) =0 and G is an
n m matrix of locally Lipschitz vector functions on R
n
.
2.3 Stability of Nonlinear Systems
For all control systems, stability is the primary requirement. One of the most widely
used stability concepts in control theory is that of Lyapunov stability, which we
employ throughout the book. In this section, we briey review basic facts from
Lyapunovs stability theory. To begin with, we note that Lyapunov stability and
asymptotic stability are properties not of a dynamical system as a whole, but rather
of its individual solutions. We restrict our attention to the class of time-invariant
nonlinear systems:
x =f (x), (2.5)
where the control input u does not appear explicitly. This does not necessarily mean
that the input to the system is zero. It could be that the input u has been specied as
a given function of the state x, u =u(x), and could be considered as a special case
of the system of Eq. (2.1).
The solution of Eq. (2.5), starting from x
0
at time t
0
R, is denoted as
x(t ; x
0
, t
0
), so that x(t
0
; x
0
, t
0
) =x
0
. Because the solutions of Eq. (2.5) are invariant
under a translation of t
0
, that is, x(t +T ; x
0
, t
0
+T ) =x(t ; x
0
, t
0
), the stability prop-
erties of x(t ; x
0
, t
0
) are uniform, i.e., they do not depend on t
0
. Therefore, without
loss of generality, we assume t
0
=0 and write x(t ; x
0
) instead of x(t ; x
0
, 0).
2.3 Stability of Nonlinear Systems 11
Lyapunov stability concepts describe continuity properties of x(t ; x
0
, t
0
) with
respect to x
0
. If the initial state x
0
is perturbed to x
0
, then, for stability, the perturbed
solution x(t ; x
0
) is required to stay close to x(t ; x
0
) for all t 0. In addition, for
asymptotic stability, the error x(t ; x
0
) x(t ; x
0
) is required to vanish as t .
Some solutions of Eq. (2.5) may be stable and some unstable. We are particularly
interested in studying and characterizing the stability properties of equilibria, that
is, constant solutions x(t ; x
e
) x
e
satisfying f (x
e
) =0.
For convenience, we state all denitions and theorems for the case when the
equilibrium point is at the origin of R
n
; that is, x
e
=0. There is no loss of generality
in doing so since any equilibrium point under investigation can be translated to
the origin via a change of variables. Suppose x
e
= 0, and consider the change of
variables, z =x x
e
. The derivative of z is given by:
z = x =f (x) =f (z +x
e
) :=g(z),
where g(0) =0. In the new variable z, the system has an equilibrium point at the ori-
gin. Therefore, for simplicity and without loss of generality, we will always assume
that f (x) satises f (0) = 0 and conne our attention to the stability properties of
the origin x
e
=0.
2.3.1 Stability Denitions
The origin is said to be a stable equilibrium point of the system of Eq. (2.5), in the
sense of Lyapunov, if for every >0 there exists a >0 such that we have:
_
_
x(0)
_
_
=
_
_
x(t )
_
_
, t 0. (2.6)
In this case, we will also simply say that the system of Eq. (2.5) is stable. A similar
convention will apply to other stability concepts introduced below. The origin is said
to be unstable if it is not stable. The requirement for stability takes a challenge
answer form. To demonstrate that the origin is stable, for every value of that a
challenger may care to design, we must produce a value of , possibly dependent on
, such that a trajectory starting in a neighborhood of the origin will never leave
the neighborhood.
The origin of the system of Eq. (2.5) is said to be asymptotically stable if it is
stable and in Eq. (2.6) can be chosen so that (attractivity property of the origin):
_
_
x(0)
_
_
= x(t ) 0 as t . (2.7)
When the origin is asymptotically stable, we are often interested in determining
how far from the origin the trajectory can be and still converge to the origin as
t approaches . This gives rise to the denition of the region of attraction (also
called region of asymptotic stability, domain of attraction, and basin). Let (t ; x)
be the solution of Eq. (2.5) that starts at initial state x at time t =0. Then the region
12 2 Background on Nonlinear Systems and Control
of attraction is dened as the set of all points x such that lim
t
(t ; x) =0. If the
origin is a stable equilibrium and its domain of attraction is the entire state-space,
then the origin is called globally asymptotically stable.
If the system is not necessarily stable but has the property that all solutions with
initial conditions in some neighborhood of the origin converge to the origin, then
it is called (locally) attractive. We say that the system is globally attractive if its
solutions converge to the origin from all initial conditions.
The system of Eq. (2.5) is called exponentially stable if there exist positive real
constants , c, and such that all solutions of Eq. (2.5) with x(0) satisfy the
inequality:
_
_
x(t )
_
_
c
_
_
x(0)
_
_
e
t
, t 0. (2.8)
If this exponential decay estimate holds for any x(0) R
n
, the system is said to be
globally exponentially stable.
2.3.2 Stability Characterizations Using Function Classes K, K

,
and KL
Scalar comparison functions, known as class K, K

, and KL, are important stabil-


ity analysis tools that are frequently used to characterize the stability properties of a
nonlinear system.
Denition 2.1 A function : [0, a) [0, ) is said to be of class K if it is con-
tinuous, strictly increasing, and (0) =0. It is said to belong to class K

if a =
and (r) as r .
Denition 2.2 A function : [0, a) [0, ) [0, ) is said to be of class KL
if, for each xed t 0, the mapping (r, t ) is of class K with respect to r and, for
each xed r, the mapping (r, t ) is decreasing with respect to t and (r, t ) 0 as
t .
We will write K and KL to indicate that is a class K function and
is a class KL function, respectively. As an immediate application of these function
classes, we can rewrite the stability denitions of the previous section in a more
compact way. For example, stability of the system of Eq. (2.5) is equivalent to the
property that there exist a > 0 and a class K function, , such that all solutions
with x(0) satisfy:
_
_
x(t )
_
_

__
_
x(0)
_
_
_
, t 0. (2.9)
Asymptotic stability is equivalent to the existence of a > 0 and a class KL func-
tion, , such that all solutions with x(0) satisfy:
_
_
x(t )
_
_

__
_
x(0)
_
_
, t
_
, t 0. (2.10)
2.3 Stability of Nonlinear Systems 13
Global asymptotic stability amounts to the existence of a class KL function, , such
that the inequality of Eq. (2.10) holds for all initial conditions. Exponential stability
means that the function takes the form (r, s) =cre
s
for some c, >0.
2.3.3 Lyapunovs Direct (Second) Method
Having dened stability and asymptotic stability of equilibrium points, the next task
is to nd ways to determine stability. To be of practical interest, stability conditions
must not require that we explicitly solve Eq. (2.5). The direct method of Lyapunov
aims at determining the stability properties of an equilibrium point from the proper-
ties of f (x) and its relationship with a positive-denite function V(x).
Denition 2.3 Consider a C
1
(i.e., continuously differentiable) function V :
R
n
R. It is called positive-denite if V(0) = 0 and V(x) > 0 for all x = 0. If
V(x) as x , then V is said to be radially unbounded.
If V is both positive-denite and radially unbounded, then there exist two class
K

functions
1
,
2
such that V satises:

1
_
x
_
V(x)
2
_
x
_
(2.11)
for all x. We write

V for the derivative of V along the solutions of the system of
Eq. (2.5), i.e.:

V(x) =
V
x
f (x). (2.12)
The main result of Lyapunovs stability theory is expressed by the following state-
ment.
Theorem 2.1 (Lyapunov) Let x = 0 be an equilibrium point for the system of
Eq. (2.5) and D R
n
be a domain containing x = 0 in its interior. Suppose that
there exists a positive-denite C
1
function V : R
n
R whose derivative along the
solutions of the system of Eq. (2.5) satises:

V(x) 0, x D (2.13)
then x =0 of the system of Eq. (2.5) is stable. If the derivative of V satises:

V(x) <0, x D/{0} (2.14)


then x = 0 of the system of Eq. (2.5) is asymptotically stable. If in the latter case,
V is also radially unbounded, then x = 0 of the system of Eq. (2.5) is globally
asymptotically stable.
14 2 Background on Nonlinear Systems and Control
A continuously differentiable positive-denite function V(x) satisfying
Eq. (2.13) is called a Lyapunov function. The surface V(x) = c, for some c > 0,
is called a Lyapunov surface or a level surface. The condition

V 0 implies that
when a trajectory crosses a Lyapunov surface V(x) = c, it moves inside the set

c
= {x R
n
: V(x) c} and can never come out again. When

V < 0, the trajec-
tory moves from one Lyapunov surface to an inner Lyapunov surface with smaller
c. As c decreases, the Lyapunov surface V(x) = c shrinks to the origin, showing
that the trajectory approaches the origin as time progresses. If we only know that

V(x) 0, we cannot be sure that the trajectory will approach the origin, but we can
conclude that the origin is stable since the trajectory can be contained inside any
ball, B

, by requiring that the initial state x


0
lie inside a Lyapunov surface contained
in that ball.
The utility of a Lyapunov function arises from the need (or difculty) of specify-
ing a unique (necessary and sufcient) direction of movement of states for stability.
To understand this, consider any scalar system (whether linear or nonlinear). The
necessary and sufcient condition for stability is that, for any value of the state x,
the value of x should be opposite in sign to x, and greater than zero in magnitude
(unless x =0). A Lyapunov function that allows readily capturing this requirement
is V(x) =
x
2
2
, resulting in

V(x) = x x. If and only if the origin of the systems is
stable (i.e., x is opposite in sign to x), it will result in

V(x) <0.
For non-scalar systems, this unique direction of movement of states, while pos-
sible for linear systems (see Remark 2.1), is in general difcult to identify for non-
linear systems. For instance, if one considers a simple two state system, and restricts
the choice of the Lyapunov function to quadratic forms, it is clear that the square
of the distance to the origin (resulting in circles as level curves) is not necessarily
the only choice of the Lyapunov-function, and there is no unique way to nd a nec-
essary and sufcient direction of the movement of states to achieve stability. This
is the problem that lies at the core of the Lyapunov-stability theorythe inability
to dene (and/or construct) a unique Lyapunov function for a given system that is
necessary and sufcient to establish stability. Having recognized this limitation, it
is important to note that the Lyapunov-based analysis at least provides sufcient
conditions to ascertain stability.
In this direction, various converse Lyapunov theorems show that the conditions
of Theorem 2.1 are also necessary. For example, if the system is asymptotically
stable, then there exists a positive-denite C
1
function V that satises the inequality
of Eq. (2.14). The theorems, however, do not provide a way of constructing this
Lyapunov function.
Remark 2.1 It is well-known that for the linear time-invariant system
x =Ax (2.15)
asymptotic stability, exponential stability, and their global versions are all equivalent
and amount to the property that A is a Hurwitz matrix, i.e., all eigenvalues of A have
2.3 Stability of Nonlinear Systems 15
negative real parts. Fixing an arbitrary positive-denite symmetric matrix Q and
nding the unique positive-denite symmetric matrix P that satises the Lyapunov
equation
A
T
P +PA=Q,
one obtains a quadratic Lyapunov function V(x) = x
T
Px whose time derivative
along the solutions of the system of Eq. (2.15) is

V =x
T
Qx. The explicit formula
for P is
P =
_

0
e
A
T
t
Qe
At
dt.
Indeed, we have
A
T
P +PA=
_

0
d
dt
_
e
A
T
t
Qe
At
_
dt =Q,
because A is Hurwitz.
2.3.4 LaSalles Invariance Principle
With some additional knowledge about the behavior of solutions, it is possi-
ble to prove asymptotic stability using a Lyapunov function which satises the
nonstrict inequality of Eq. (2.13). This is facilitated by LaSalles invariance
principle. To state this principle, we rst recall the denition of an invariant
set.
Denition 2.4 A set M is called (positively) invariant with respect to the given
system if all solutions starting in M remain in M for all future times.
We now state a version of LaSalles theorem.
Theorem 2.2 (LaSalle) Suppose that there exists a positive-denite C
1
function V :
R
n
R whose derivative along the solutions of the system of Eq. (2.5) satises
the inequality of Eq. (2.13). Let M be the largest invariant set contained in the
set {x :

V(x) = 0}. Then the system of Eq. (2.5) is stable and every solution that
remains bounded for t 0 approaches M as t . In particular, if all solutions
remain bounded and M ={0}, then the system of Eq. (2.5) is globally asymptotically
stable.
To deduce global asymptotic stability with the help of this result, one needs
to check two conditions. First, all solutions of the system must be bounded.
This property follows automatically from the inequality of Eq. (2.13) if V is
chosen to be radially unbounded; however, radial boundedness of V is not nec-
essary when boundedness of solutions can be established by other means. The
16 2 Background on Nonlinear Systems and Control
second condition is that V be not identically zero along any nonzero solution.
We also remark that if one only wants to prove asymptotic convergence of
bounded solutions to zero and is not concerned with Lyapunov stability of the
origin, then positive-deniteness of V is not needed (this is in contrast to Theo-
rem 2.1).
While Lyapunovs stability theorem readily generalizes to time-varying systems,
for LaSalles invariance principle this is not the case. Instead, one usually works
with the weaker property that all solutions approach the set {x :

V(x) =0}.
2.3.5 Lyapunovs Indirect (First) Method
Lyapunovs indirect method allows one to deduce stability properties of the nonlin-
ear system of Eq. (2.5), where f is C
1
, from stability properties of its linearization,
which is the linear system of Eq. (2.15) with
A:=
f
x
(0). (2.16)
By the mean value theorem, we can write
f (x) =Ax +g(x)x,
where g is given componentwise by g
i
(x) :=
f
i
x
(z
i
)
f
i
x
(0) for some point, z
i
,
on the line segment connecting x to the origin, i = 1, . . . , n. Since
f
x
is contin-
uous, we have g(x) 0 as x 0. From this it follows that if the matrix A
is Hurwitz (i.e., all its eigenvalues lie in the open left half of the complex plane),
then a quadratic Lyapunov function for the linearization serveslocallyas a Lya-
punov function for the original nonlinear system. Moreover, its rate of decay in a
neighborhood of the origin can be bounded below by a quadratic function, which
implies that stability is, in fact, exponential. This is summarized by the following
result.
Theorem 2.3 If f is C
1
and the matrix of Eq. (2.16) is Hurwitz, then the system of
Eq. (2.5) is locally exponentially stable.
It is also known that if the matrix A has at least one eigenvalue with a posi-
tive real part, the origin of the nonlinear system of Eq. (2.5) is not stable. If A has
eigenvalues on the imaginary axis but no eigenvalues in the open right half-plane,
the linearization test is inconclusive. However, in this critical case, the system of
Eq. (2.5) cannot be exponentially stable since exponential stability of the lineariza-
tion is not only a sufcient but also a necessary condition for (local) exponential
stability of the nonlinear system.
2.3 Stability of Nonlinear Systems 17
2.3.6 Input-to-State Stability
It is of interest to extend stability concepts to systems with disturbance inputs. In
the linear case represented by the system
x =Ax +B,
it is well known that if the matrix A is Hurwitz, i.e., if the unforced system, x =Ax,
is asymptotically stable, then bounded inputs lead to bounded states while inputs
converging to zero produce states converging to zero. Now, consider a nonlinear
system of the form
x =f (x, ), (2.17)
where is a measurable bounded disturbance input. In general, global asymptotic
stability of the unforced system x = f (x, 0) does not guarantee input-to-state sta-
bility with respect to of the kind mentioned above. For example, the scalar system
x =x +x (2.18)
has unbounded trajectories under the bounded input 2. This motivates the fol-
lowing important concept, introduced by Sontag [151].
Denition 2.5 The system of Eq. (2.17) is called input-to-state stable (ISS) with
respect to if for some functions K

and KL, for every initial state x(0),


and every input , the corresponding solution of the system of Eq. (2.17) satises
the inequality
_
_
x(t )
_
_

__
_
x(0)
_
_
, t
_
+
_

s
[0,t ]
_
, (2.19)
where
s
[0,t ]
:=ess.sup{(s) : s [0, t ]} (supremum norm on [0, t ] except for a
set of measure zero).
Since the system of Eq. (2.17) is time-invariant, the same property results if we
write
_
_
x(t )
_
_

__
_
x(t
0
)
_
_
, t t
0
_
+
_

s
[t
0
,t ]
_
, t t
0
0. (2.20)
The ISS property admits the following Lyapunov-like equivalent characterization:
The system of Eq. (2.17) is ISS if and only if there exists a positive-denite radially
unbounded C
1
function V : R
n
R such that for some class K

functions and
we have
V
x
f (x, )
_
x
_
+
_

_
, x, . (2.21)
This is, in turn, equivalent to the following gain margin condition:
x
_

_
=
V
x
f (x, )
_
x
_
, (2.22)
18 2 Background on Nonlinear Systems and Control
where , K

. Such functions V are called ISS-Lyapunov functions. If the sys-


tem of Eq. (2.17) is ISS, then (t ) 0 implies x(t ) 0.
The system of Eq. (2.17) is said to be locally input-to-state stable (locally ISS)
if the bound of Eq. (2.19) is valid for solutions with sufciently small initial condi-
tions and inputs, i.e., if there exists a >0 such that Eq. (2.19) is satised whenever
x(0) and
s
[0,t ]
. It turns out that (local) asymptotic stability of the un-
forced system x =f (x, 0) implies local ISS.
2.4 Stabilization of Nonlinear Systems
This book is primarily about control design. Our objective is to create closed-loop
systems with desirable stability and performance properties, rather than analyze the
properties of a given system. For this reason, we are interested in an extension of
the Lyapunov function concept, called a control Lyapunov function (CLF).
Suppose that our problem for the time-invariant system
x =f (x, u), (2.23)
where x R
n
, u R (i.e., we consider the unconstrained problem), f (0, 0) = 0,
is to design a feedback control law (x) for the control variable u such that the
equilibrium x =0 of the closed-loop system
x =f
_
x, (x)
_
(2.24)
is globally asymptotically stable. We can pick a function V(x) as a Lyapunov func-
tion candidate, and require that its derivative along the solutions of the system of
Eq. (2.24) satises

V W(x), where W(x) is a positive-denite function. We
therefore need to nd (x) to guarantee that for all x R
n
V
x
(x)f
_
x, (x)
_
W(x). (2.25)
This is a difcult task. A stabilizing control law for the system of Eq. (2.23) may
exist, but it may fail to satisfy Eq. (2.25) because of a poor choice of V(x) and
W(x). A system for which a good choice of V(x) and W(x) exists is said to possess
a CLF. This notion is made more precise below.
Denition 2.6 A smooth positive-denite radially unbounded function V : R
n
R
is called a control Lyapunov function (CLF) for the system of Eq. (2.23) if
inf
uR
_
V
x
(x)f (x, u)
_
<0, x =0. (2.26)
The CLF concept of Artstein [9] is a generalization of Lyapunov design results
by Jacobson and Judjevic and Quinn. Artstein showed that Eq. (2.26) is not only
2.4 Stabilization of Nonlinear Systems 19
necessary, but also sufcient for the existence of a control law satisfying Eq. (2.25),
that is, the existence of a CLF is equivalent to global asymptotic stabilizability.
For systems afne in the control, namely,
x =f (x) +g(x)u, f (0) =0, (2.27)
the CLF inequality of Eq. (2.25) becomes
L
f
V(x) +L
g
V(x)u W(x). (2.28)
If V is a CLF for the system of Eq. (2.27), then a particular stabilizing control law
(x), smooth for all x =0, is given by Sontags formula [150]:
u =
s
(x) =
_
_
_

L
f
V(x)+

(L
f
V)
2
(x)+(L
g
V)
4
(x)
(L
g
V)
2
(x)
L
g
V(x), L
g
V(x) =0,
0, L
g
V(x) =0.
(2.29)
It should be noted that Eq. (2.28) can be satised only if
L
g
V(x) =0 = L
f
V(x) <0, x =0. (2.30)
The intuitive interpretation of the existence of a CLF is as follows: For any x such
that L
g
V(x) =0, since there are no constraints on the input,

V can be made negative
by picking a large enough control action, with an appropriate sign, to counter the
effect of possibly positive L
f
V(x) term. For all x such that L
g
V(x) =0, the control
action has no effect on the Lyapunov-function derivative. For it to be possible to
show stability using the CLF V , it should therefore be true that whenever L
g
V(x) =
0, we also have that L
f
V(x) < 0. This is the requirement that is formalized in
Eq. (2.30). With such a CLF, Eq. (2.29) results in
W(x) =
_
(L
f
V)
2
(x) +(L
g
V)
4
(x) >0, x =0. (2.31)
A further characterization of a stabilizing control law (x) for the system of
Eq. (2.27) with a given CLF V is that (x) is continuous at x = 0 if and only if
the CLF satises the small control property: For each >0 there is a () >0 such
that, if x =0 satises |x| , then there is some u with |u| < such that
L
f
V(x) +L
g
V(x)u <0. (2.32)
The main deciency of the CLF concept as a design tool is that for most nonlinear
systems a CLF is not known. The task of nding an appropriate CLF maybe as com-
plex as that of designing a stabilizing feedback law. In the next section, we review
one commonly used tool for designing a Lyapunov-based control law that utilizes
coordinate transformations. We also note that in the presence of input constraints,
the concept of a CLF needs to be revisited, and this issue is discussed in Sect. 2.6.
20 2 Background on Nonlinear Systems and Control
2.5 Feedback Linearization and Zero Dynamics
One of the popular methods for nonlinear control design (or alternatively, one way
to construct a Lyapunov-function for the purpose of control design) is feedback
linearization, which employs a change of coordinates and feedback control to trans-
form a nonlinear system into a system whose dynamics are linear (at least partially).
This transformation allows the construction and use of a Lyapunov function for the
control design utilizing results from linear systems analysis. A great deal of re-
search has been devoted to this subject over the last four decades, as evidenced by
the comprehensive books [72, 126] and the references therein. In this section, we
briey review some of the basic geometric concepts that will be used in subsequent
chapters. While this book does not require the formalism of differential geometry,
we will employ Lie derivatives only for notational convenience. If f : R
n
R
n
is a vector eld and h : R
n
R is a scalar function, the notation L
f
h is used for
h
x
f (x). It is recursively extended to
L
k
f
h(x) =L
f
_
L
k1
f
h(x)
_
=

x
_
L
k1
f
h(x)
_
f (x).
Let us consider the following nonlinear system:
x =f (x) +g(x)u,
y =h(x),
(2.33)
where x R
n
, u R, y R, f , g, h are analytic (i.e., innitely differentiable)
vector functions. The derivative of the output y =h(x) is given by
y =
h
x
(x)f (x) +
h
x
(x)g(x)u
= L
f
h(x) +L
g
h(x)u. (2.34)
If L
g
h(x
0
) = 0, then the system of Eq. (2.33) is said to have relative degree one at
x
0
(note that since the functions are smooth L
g
h(x
0
) =0 implies that there exists a
neighborhood of x
0
on which L
g
h(x) =0). In our terminology, this implies that the
output y is separated form the input u by one integration only. If L
g
h(x
0
) =0, there
are two cases:
(i) If there exist points arbitrarily close to x
0
such that L
g
h(x) =0, then the system
of Eq. (2.33) does not have a well-dened relative degree at x
0
.
(ii) If there exists a neighborhood B
0
of x
0
such that L
g
h(x) = 0 for all x B
0
,
then the relative degree of the system of Eq. (2.33) may be well-dened.
In case (ii), we dene

1
(x) =h(x),
2
(x) =L
f
h(x) (2.35)
2.5 Feedback Linearization and Zero Dynamics 21
and compute the second derivative of y
y =

2
x
(x)f (x) +

2
x
(x)g(x)u
= L
2
f
h(x) +L
g
L
f
h(x)u. (2.36)
If L
g
L
f
h(x
0
) =0, then the system of Eq. (2.33) is said to have relative degree two
at x
0
. If L
g
L
f
h(x) =0 in a neighborhood of x
0
, then we continue the differentiation
procedure.
Denition 2.7 The system of Eq. (2.33) is said to have relative degree r at the point
x
0
if there exists a neighborhood B
0
of x
0
on which
L
g
h(x) =L
g
L
f
h(x) = =L
g
L
r2
f
h(x) = 0, (2.37)
L
g
L
r1
f
h(x) = 0. (2.38)
If Eq. (2.37)(2.38) are valid for all x R
n
, then the relative degree of the system
of Eq. (2.33) is said to be globally dened.
Suppose now that the system of Eq. (2.33) has relative degree r at x
0
. Then
we can use a change of coordinates and feedback control to locally transform this
system into the cascade interconnection of an r-dimensional linear system and an
(n r)-dimensional nonlinear system. In particular, after differentiating r times the
output y =h(x), the control appears:
y
(r)
=L
r
f
h(x) +L
g
L
r1
f
h(x)u. (2.39)
Since L
g
L
r1
f
h(x) = 0 in a neighborhood of x
0
, we can linearize the inputoutput
dynamics of the system of Eq. (2.33) using feedback to cancel the nonlinearities in
Eq. (2.39):
u =
1
L
g
L
r1
f
h(x)
_
L
r
f
h(x) +v
_
. (2.40)
Then the dynamics of y and its derivatives are governed by a chain of r integra-
tors: y
(r)
= v. Since our original system of Eq. (2.33) has dimension n, we need
to account for the remaining n r states. Using differential geometry tools, it can
be shown that it is always possible to nd n r functions
r+1
, . . . ,
n
(x) with

i
x
(x)g(x) =0, for i =r +1, . . . , n such that the change of coordinates

1
=y =h(x),
2
= y =L
f
h(x), . . . ,
r
=y
(r1)
=L
r1
f
h(x),

1
=
r+1
, . . . ,
nr
=
n
(x)
(2.41)
22 2 Background on Nonlinear Systems and Control
is locally invertible and transforms, along with the feedback law of Eq. (2.40), the
system of Eq. (2.33) into

1
=
2
,
.
.
.

r
=v,

1
=
1
(, ),
.
.
.

nr
=
nr
(, ),
y =
1
,
(2.42)
where
1
(, ) =L
r+1
f
h(x),
nr
(, ) =L
n
f
h(x).
The states
1
, . . . ,
nr
have been rendered unobservable from the output y by
the control of Eq. (2.40). Hence, feedback linearization in this case is the nonlin-
ear equivalent of placing n r poles of a linear system at the origin and canceling
the r zeros with the remaining poles. Of course, to guarantee stability, the canceled
zeros must be stable. In the nonlinear case, using the new control input v to sta-
bilize the linear subsystem of Eq. (2.42) does not guarantee stability of the whole
system, unless the stability of the nonlinear part of the system of Eq. (2.42) has been
established separately.
When v is used to keep the output y equal to zero for all t > 0, that is, when

1

r
0, the dynamics of
1
, . . . ,
nr
are described by

1
=
1
(0, ),
.
.
.

nr
=
nr
(0, ).
(2.43)
They are called the zero dynamics of the system of Eq. (2.33) because they evolve
on the subset of the state-space on which the output of the system is identically
zero. If the equilibrium at
1
= =
nr
=0 of the zero dynamics of Eq. (2.43) is
asymptotically stable, the system of Eq. (2.33) is said to be minimum phase.
Remark 2.2 Most nonlinear analytical controllers emanating from the area of ge-
ometric control are inputoutput linearizing and induce a linear inputoutput re-
sponse in the absence of constraints [72, 81]. For the class of processes modeled by
equations of the form of Eq. (2.33) with relative order r and under the minimum
phase assumption, the appropriate linearizing state feedback controller is given by
u =
1
L
g
L
r1
f
h(x)
_
v L
r
f
h(x)
1
L
r1
f
h(x)
r1
L
f
h(x)
r
h(x)
_
(2.44)
2.6 Input Constraints 23
and induces the linear rth order response
d
r
y
dt
r
+
1
d
r1
y
dt
r1
+ +
r1
dy
dt
+
r
y =v, (2.45)
where the tunable parameters,
1
, . . . ,
r
, are essentially closed-loop time constants
that inuence and shape the output response. The nominal stability of the process is
guaranteed by placing the roots of the polynomial s
r
+
1
s
r1
+ +
r1
s +
r
in the open left-half of the complex plane.
2.6 Input Constraints
The presence of input constraints requires revisiting the concept of the CLF for both
linear and nonlinear systems. To understand this, consider a scalar linear system of
the form x =x +u, with u
min
u u
max
. For the sake of simplicity and without
loss of generality, let us assume u
min
< 0 < u
max
and > 0. For the case of scalar
systems, it is possible to determine the entire set of initial conditions from where
the system can be driven to the origin subject to input constraints (regardless of the
choice of the control law). This set is generally referred to as the null controllable
region (NCR). An explicit computation of the NCR is possible in this case because
for scalar systems (as discussed earlier) there exists a unique direction in which the
system states needs to move to achieve stability.
To determine this set, one can simply analyze the system trajectory to the left
and right of zero. Consider rst x > 0, and the requirement that for x > 0, x < 0.
If < 0, x < 0 x > 0 (and also x > 0 x < 0). On the other hand, if > 0,
x < 0 can only be achieved for x <
u
min

. Similarly, x > 0 can only be achieved


for x >
u
max

. The analysis reveals what was perhaps intuitive to begin with: For
linear systems, if the steady state is open-loop stable, the NCR is the entire state
space, while if the steady state is open-loop unstable, it has a nite NCR, which
in this case is {x :
u
max

< x <
u
min

}. The same result for the NCR can also


be obtained using a CLF V(x) =
x
2
2
and determining the states for which

V < 0 is
achievable using the available control action. Furthermore, it points to the require-
ment of additional considerations when dening CLFs for systems with constrained
inputs. In particular, requiring that

V(x) <0 x is simply not achievable for certain
cases, at best what is achievable is that

V(x) <0 x NCR{0}. The denition of
a CLF (or more appropriately, a constrained CLF) then becomes intricately linked
with the characterization of the NCR. The characterization of the NCR, however,
is an increasingly difcult (although possible, see [71]) problem when considering
non-scalar linear systems, and currently an open problem for nonlinear systems.
To understand the impact of the lack of availability of constrained CLFs
(CCLFs), let us rst consider again the linear scalar system under a feedback law
of the form u
c
(x) = kx, with k > 0 such that ( k) < 0 under two possible
scenarios: (i) < 0 (i.e., for the unforced system, there is an isolated equilibrium
24 2 Background on Nonlinear Systems and Control
point at the origin and the system is stable at that operating point) and (ii) >0 (i.e.,
for the unforced system, there is an isolated equilibrium point at the origin and the
system is unstable at that operating point). Due to the presence of input constraints,
the closed-loop system is no longer a linear system, but operates in three modes,
depending on the state, described by the following set of equations:
dx
dt
=x +u
c
, u
min
u
c
u
max
,
dx
dt
=x +u
max
, u
c
>u
max
,
dx
dt
=x +u
min
, u
min
>u
c
.
(2.46)
Let us analyze the three possible modes of operation of the closed-loop system
for scenario (i). For
|u
max
|
k
x
|u
min
|
k
, we have that
dx
dt
=x +u
c
=( k)x,
which establishes that for all initial conditions x
0
such that
|u
max
|
k
x
0

|u
min
|
k
,
the prescribed control action u
c
is within the constraints and the system state will
be driven to the origin. For
|u
min
|
k
<x
u
min

, u
c
>u
max
resulting in u =u
max
, in
turn resulting in x < 0. A similar result is obtained for
u
max

< x <
|u
max
|
k
. The
analysis shows that for scalar systems, while the region of unconstrained operation
for a particular control law might depend on the specic control law chosen, the
stability region under the control law might still possibly be the entire NCR.
The issue of directionality again crops up when considering non-scalar systems.
While it is relatively easy to determine the region of unconstrained operation for a
particular control law, and, in certain cases, the region of attraction for the closed-
loop system, it is not necessary that the region of attraction for the closed-loop
system match the NCR. This happens due to the fact that it is in general difcult
to determine, for a particular value of the state, the unique direction in which the
inputs should saturate to achieve closed-loop stability. To achieve this objective, re-
cent control designs have utilized the explicit characterization of the NCR [71] in
designing CCLF based control laws that ensure stabilization from all initial con-
ditions in the NCR [93, 94]. For nonlinear systems, where the characterization of
the NCR is still an open problem, a meaningful control objective is to be able to
explicitly account for the constraints in the control design and provide an explicit
characterization of the closed-loop stability region.
2.7 Model Predictive Control
One of the control methods useful for accounting for constraints and optimality si-
multaneously is that of model predictive control (MPC). MPC is an approach which
accounts for optimality considerations explicitly and is widely adopted in industry
as an effective approach to deal with large multivariable constrained optimal con-
trol problems. The main idea of MPC is to choose control actions by repeatedly
2.7 Model Predictive Control 25
solving an online a constrained optimization problem, which aims at minimizing a
performance index over a nite prediction horizon based on predictions obtained by
a system model. In general, an MPC design is composed of three components:
1. A model of the system. This model is used to predict the future evolution of the
system in open-loop and the efciency of the calculated control actions of an
MPC depends highly on the accuracy of the model.
2. A performance index over a nite horizon. This index is minimized subject to
constraints imposed by the system model, restrictions on control inputs and sys-
tem state, and other considerations at each sampling time to obtain a trajectory
of future control inputs.
3. A receding horizon scheme. This scheme introduces the notion of feedback into
the control law to compensate for disturbances and modeling errors, whereby
only the rst piece of the future input trajectory is implemented and the con-
strained optimization problem is resolved at the next sampling instance.
Consider the control of the system of Eq. (2.1) and assume that the state mea-
surements of the system of Eq. (2.1) are available at synchronous sampling time
instants {t
k0
}, a standard MPC is formulated as follows [60]:
min
uS()
_
t
k+N
t
k
__
_
x()
_
_
Q
c
+
_
_
u()
_
_
R
c
_
d +F
_
x(t
k+N
)
_
(2.47)
s.t.

x(t ) =f
_
x(t ), u(t )
_
, (2.48)
u(t ) U, (2.49)
x(t
k
) =x(t
k
), (2.50)
where S() is the family of piece-wise constant functions with sampling period
, N is the prediction horizon, Q
c
and R
c
are strictly positive denite symmetric
weighting matrices, x is the predicted trajectory of the system due to control input
u with initial state x(t
k
) at time t
k
, and F() denotes the terminal penalty.
The optimal solution to the MPC optimization problem dened by Eq. (2.47)
(2.50) is denoted as u

(t |t
k
) which is dened for t [t
k
, t
k+N
). The rst step value
of u

(t |t
k
) is applied to the closed-loop system for t [t
k
, t
k+1
). At the next sam-
pling time t
k+1
, when a new measurement of the system state x(t
k+1
) is available,
and the control evaluation and implementation procedure is repeated. The manipu-
lated input of the system of Eq. (2.1) under the control of the MPC of Eq. (2.47)
(2.50) is dened as follows:
u(t ) =u

(t |t
k
), t [t
k
, t
k+1
), (2.51)
which is the standard receding horizon scheme.
In the MPC formulation of Eq. (2.47)(2.50), Eq. (2.47) denes a performance
index or cost index that should be minimized. In addition to penalties on the state
and control actions, the index may also include penalties on other considerations;
for example, the rate of change of the inputs. Equation (2.48) is the model of the
26 2 Background on Nonlinear Systems and Control
system of Eq. (2.1) which is used in the MPC to predict the future evolution of the
system. Equation (2.49) takes into account the constraint on the control input, and
Eq. (2.50) provides the initial state for the MPC which is a measurement of the
actual system state. Note that in the above MPC formulation, state constraints are
not considered but can be readily taken into account.
It is well known that the MPC of Eq. (2.47)(2.50) is not necessarily stabilizing.
To understand this, let us consider a discrete time version of the MPC implementa-
tion, for a scalar system described by x(k + 1) = x(k) + u(k), in the absence of
input constraints. Also, let N = 1, q and r denote the horizon, penalty on the state
deviation and input deviation, respectively. The objective function then simplies to
q(
2
x(k)
2
+ u(k)
2
+ 2x(k)u(k)) + ru(k)
2
, and the minimizing control action is
u(k) =
qx(k)
q+r
, resulting in the closed-loop systemx(k+1) =
rx(k)
q+r
. The minimiz-
ing solution will result in stabilizing control action only if q > r( 1). Note that
for < 1, this trivially holds (i.e., the result trivially holds for stabilization around
an open-loop stable steady state). For > 1, the result establishes how large the
penalty on the set point deviation should be compared to the penalty on the control
action for the controller to be stabilizing. The analysis is meant to bring out the fact
that generally speaking, the stability of the closed-loop system in the MPC depends
on the MPC parameters (penalties and the control horizon) as well as the system
dynamics. Note also that even though we have analyzed an unconstrained system,
the prediction horizon we used was nite (in comparison to linear quadratic regula-
tor designs, where the innite horizon cost is essentially captured in computing the
control action, and therefore results in stabilizing controller in the absence of con-
straints). Finally, also note that for the case of innite horizon, the optimum solution
is also the stabilizing one, and it can be shown that such an MPC will stabilize the
system with the NCR as the stability region (albeit at an impractical computational
burden).
To achieve closed-loop stability without relying on the objective function pa-
rameters, different approaches have been proposed in the literature. One class of
approaches is to use well-designed terminal penalty terms that capture innite hori-
zon costs; please, see [16, 100] for surveys of these approaches. Another class
of approaches is to impose stability constraints in the MPC optimization problem
[3, 14, 100]. There are also efforts focusing on getting explicit stabilizing MPC laws
using ofine computations [92]. However, the implicit nature of MPC control law
makes it very difcult to explicitly characterize, a priori, the admissible initial con-
ditions starting from where the MPC is guaranteed to be feasible and stabilizing.
In practice, the initial conditions are usually chosen in an ad hoc fashion and tested
through extensive closed-loop simulations.
2.8 Lyapunov-Based MPC
In this section, we introduce Lyapunov-based MPC (LMPC) designs proposed in
[93, 108, 110] which allow for an explicit characterization of the stability region
and guarantee controller feasibility and closed-loop stability.
2.8 Lyapunov-Based MPC 27
For the predictive control of the system of Eq. (2.1), the key idea in LMPC-based
designs is to utilize a Lyapunov-function based constraint and achieve immediate
decay of the Lyapunov function. The set of initial conditions for which it is possible
to achieve an instantaneous decay in the Lyapunov function value can be computed
explicitly, and picking the (preferably largest) level curve contained in this set can
provide the explicitly characterized feasibility and stability region for the LMPC.
The following example of the LMPC design is based on an existing explicit con-
trol law h(x) which is able to stabilize the closed-loop system [108, 110]. The for-
mulation of the LMPC is as follows:
min
uS()
_
t
k+N
t
k
__
_
x()
_
_
Q
c
+
_
_
u()
_
_
R
c
_
d (2.52)
s.t.

x(t ) =f
_
x(t ), u(t )
_
, (2.53)
u(t ) U, (2.54)
x(t
k
) =x(t
k
), (2.55)
V(x(t
k
))
x
f
_
x(t
k
), u(t
k
)
_

V(x(t
k
))
x
f
_
x(t
k
), h
_
x(t
k
)
__
, (2.56)
where V(x) is a Lyapunov function associated with the nonlinear control law h(x).
The optimal solution to this LMPC optimization problem is denoted as u

l
(t |t
k
)
which is dened for t [t
k
, t
k+N
). The manipulated input of the system of Eq. (2.1)
under the control of the LMPC of Eq. (2.52)(2.56) is dened as follows:
u(t ) =u

l
(t |t
k
), t [t
k
, t
k+1
), (2.57)
which implies that this LMPC also adopts a standard receding horizon strategy.
In the LMPCdened by Eq. (2.52)(2.56), the constraint of Eq. (2.56) guarantees
that the value of the time derivative of the Lyapunov function, V(x), at time t
k
is
smaller than or equal to the value obtained if the nonlinear control law u = h(x)
is implemented in the closed-loop system in a sample-and-hold fashion. This is a
constraint that allows one to prove (when state measurements are available every
synchronous sampling time) that the LMPC inherits the stability and robustness
properties of the nonlinear control law h(x) when it is applied in a sample-and-hold
fashion; please, see [30, 125] for results on sampled-data systems.
Let us denote the stability region of h(x) as

. The stability properties of the


LMPC implies that the origin of the closed-loop system is guaranteed to be stable
and the LMPC is guaranteed to be feasible for any initial state inside

when the
sampling time is sufciently small. Note that the region

can be explicitly
characterized; please, refer to [110] for more discussion on this issue. The main
advantage of the LMPC approach with respect to the nonlinear control law h(x)
is that optimality considerations can be taken explicitly into account (as well as
constraints on the inputs and the states [110]) in the computation of the control
actions within an online optimization framework while improving the closed-loop
performance of the system. Since the closed-loop stability and feasibility of the
28 2 Background on Nonlinear Systems and Control
LMPC of Eq. (2.52)(2.56) are guaranteed by the nonlinear control law h(x), it is
unnecessary to use a terminal penalty term in the cost index (see Eq. (2.52) and
compare it with Eq. (2.47)) and the length of the horizon N does not affect the
stability of the closed-loop system but it affects the closed-loop performance.
2.9 Hybrid Systems
Hybrid systems are characterized by the co-existence of continuous modes of op-
eration along with discrete switches between the distinct modes of operation and
arise frequently in the design and analysis of fault-tolerant control systems. The
class of hybrid systems of interest to the focus of this bookswitched systemscan
be described by
x =f
i(x,t )
(x) +g
i(x,t )
(x)u
i(x,t )
, (2.58)
where x R
n
, u R
n
are the continuous variables and i N are the discrete vari-
ables indexing the mode of operation. The nature of the function i(x, t ) and, in par-
ticular, its two specic forms i(x) and i(t ) result in the so-called state-dependent and
time-dependent switching. What is of more interest from a stability analysis and de-
sign point of view (both when considering the design of control laws and, in the case
of time-dependent switching, the switching signal) is the possibility of innitely
many switches where it becomes crucial to explicitly consider the switched nature
of the system in the stability analysis. In particular, when the possibility of innitely
many switches exists, establishing stability in the individual modes of operation
is not sufcient [19], and additional conditions on the behavior of the Lyapunov-
functions (used to establish stability in the individual modes of operation) during
the switching (as well as of sufcient dwell-time [68]) need to be satised for the
stability of the switched system. For the case of nite switches, the considerations
include ensuring stability requirements at the onset of a particular mode are satised
and, in particular, satised for the terminal (last) mode of operation.
2.10 Conclusions
In this chapter, some fundamental results on nonlinear systems analysis and control
were briey reviewed. First, the class of nonlinear systems that will be considered
in this book was presented; then the denitions of stability of nonlinear systems
were introduced; and following that, techniques for stabilizing nonlinear systems,
for example, Lyapunov-based control, feedback linearization, handling constraints,
model predictive control and Lyapunov-based model predictive control and stability
of hybrid (switched) systems were discussed.
http://www.springer.com/978-1-4471-4807-4

You might also like