Linear Systems With Control Theory
Linear Systems With Control Theory
Differential Equations
not be used.
1
2 CHAPTER 1. DIFFERENTIAL EQUATIONS
A = {(x, t) : |t − τ | ≤ , |x − ξ| ≤ } (1.5)
and suppose that F (x; t) and ∂F/∂x are continuous functions in both x and
t on A. Then the differential equation (1.1) has a unique solution x = f (t),
satisfying the initial condition ξ = f (τ ) on the interval [τ − 1 , τ + 1 ], for
some 1 with 0 < 1 ≤ .
We shall not give a proof of this theorem, but we shall indicate an approach
which could lead to a proof. The differential equation (1.1) (together with the
initial condition) is equivalent to the integral equation
t
x(t) = ξ + F (x(u); u)du. (1.6)
τ
x(0) (t) = ξ,
t (1.7)
x(j+1) (t) = ξ + F (x(j) (u); u)du, j = 1, 2, . . . .
τ
The members of this sequence are known as Picard iterates. To prove Picard’s
theorem we would need to show that, under the stated conditions, the sequence
of iterates converges uniformly to a limit function x(t) and then prove that it
is a unique solution to the differential equation. Rather than considering this
general task we consider a particular case:
x(0) (t) = 1,
t
x(1) (t) = 1 + 1 du = 1 + t,
0
t
x(2) (t) = 1 + 1 2
(1 + u)du = 1 + t + 2! t ,
0
t (1.9)
(3) 1 2 1 2 1 3
x (t) = 1 + 1+u+ 2!
u du = 1 + t + 2!
t + 3!
t ,
0
.. ..
. .
x(j) (t) = 1 + t + 2!
1 2 1 j
t + · · · + j! t .
We see that, as j → ∞, x(j) (t) → exp(x), which is indeed the unique solution
of (1.8).
The MAPLE command for solving a differential equation is dsolve and the
MAPLE code which obtains the general solution for (1.11) is
> dsolve(diff(x(t),t$3)=x(t));
1√ 1√
x(t) = C1 et + C2 e(−1/2 t) sin( 3 t) + C3 e(−1/2 t) cos( 3 t)
2 2
> dsolve({diff(x(t),t$3)=x(t),x(0)=2,D(x)(0)=3,
> (D@@2)(x)(0)=7},x(t));
4 √ (−1/2 t) 1√ 1√
x(t) = 4 et − 3e sin( 3 t) − 2 e(−1/2 t) cos( 3 t)
3 2 2
Note that in the context of the initial condition the code D(x) is used for the
first derivative of x with respect to t. The corresponding n-th order derivative
is denoted by (D@@n)(x).
have the connotation of being specified at a fixed or initial time and boundary conditions at
fixed points in space at the ends or boundaries of the system.
1.3. GENERAL SOLUTION OF SPECIFIC EQUATIONS 5
sides of the equation. The solution is now more or less easy to find according
to whether it is easy or difficult to perform the integrations.
This gives
x
= C exp(µt), (1.16)
κ−x
which can be solved for x to give
C κ exp(µt)
x(t) = . (1.17)
1 + C exp(µt)
The MAPLE code for solving this equation is
> dsolve(diff(x(t),t)=mu*x(t)*(kappa-x(t))/kappa);
κ
x(t) =
1 + e(−µ t) C1 κ
for some m. The method of solution is to make the change of variable y(t) =
x(t)/t. Since
is an integrating factor.
Example 1.3.3 Find the general solution of
dx
+ x cot(t) = 2 cos(t). (1.35)
dt
The integrating factor is
exp cot(t)dt = exp [ln sin(t)] = sin(t), (1.36)
giving
dx
sin(t) + x cos(t) = 2 sin(t) cos(t) = sin(2t),
dt
d[x sin(t)]
= sin(2t). (1.37)
dt
So
x sin(t) = sin(2t)dt + C
= − 12 cos(2t) + C, (1.38)
giving
C − cos2 (t)
x(t) = . (1.39)
sin(t)
The MAPLE code for solving this equation is
> dsolve(diff(x(t),t)+x(t)*cot(t)=2*cos(t));
1
− cos(2 t) + C1
x(t) = 2
sin(t)
8 CHAPTER 1. DIFFERENTIAL EQUATIONS
We divide the problem of finding the solution to (1.42) into two parts. We first
describe a method for finding xc (t), usually called the complementary function,
and then we develop a method for finding a particular solution xp (t).
1.4. EQUATIONS WITH CONSTANT COEFFICIENTS 9
Then
Cj exp(λj t) + Cj+1 exp(λj+1 t) = exp(αt) Cj [cos(βt) + i sin(βt)]
+ Cj+1 [cos(βt) − i sin(βt)]
= exp(αt) C cos(βt) + C sin(βt) ,
(1.55)
where
In order to consider the case of equal roots we need the following result:
Theorem 1.4.2 For any positive integer n and function u(t)
Dn u(t) exp(λt) = exp(λt)(D + λ)n u(t) (1.57)
Proof: We prove the result by induction. For n = 1
Du(t) exp(λt) = exp(λt)Du(t) + u(t)D exp(λt)
= exp(λt)Du(t) + exp(λt)λu(t)
= exp(λt)(D + λ)u(t). (1.58)
Now suppose the result is true for some n. Then
Dn+1 u(t) exp(λt) = D exp(λt)(D + λ)n u(t)
= (D + λ)n u(t)D exp(λt)
+ exp(λt)D(D + λ)n u(t)
= exp(λt)λ(D + λ)n u(t)
+ exp(λt)D(D + λ)n u(t)
= exp(λt)(D + λ)n+1 u(t). (1.59)
and the result is established for all n.
as long as φ(b) = 0, that is, when b is not a root of the auxiliary equation (1.48).
To consider that case suppose that
φ(λ) = ψ(λ)(λ − b)m , ψ(b) = 0. (1.74)
That is b is an m-fold root of the auxiliary equation. Now try the trial function
T (t) = Atm exp(bt). From (1.60)
φ(D)T (t) = ψ(D)(D − b)m Atm exp(bt)
= A exp(bt)ψ(D + b)[(D + b) − b]m tm
= A exp(bt)ψ(D + b)Dm tm
= A exp(bt)ψ(b)m! (1.75)
Equating this with f (t), given by (1.71), we see that the trial function is a
solution of (1.42) if
a
A= . (1.76)
m!ψ(b)
Table 1.1 contains a list of trial functions to be used for different forms of f (t).
Trial functions when f (t) is a linear combination of the forms given are simply
the corresponding linear combination of the trial functions. Although there
seems to be a lot of different cases it can be seen that they are all special cases
of either the eighth or tenth lines. We conclude this section with two examples.
atn sin(bt) or atn cos(bt) [B1 sin(bt) + B2 cos(bt)][tn + An−1 tn−1 + · · · + A0 ] λ2 + b2 not a factor of φ(λ).
13
atn sin(bt) or atn cos(bt) tk [B1 sin(bt) + B2 cos(bt)][tn + An−1 tn−1 + · · · + A0 ] λ2 + b2 a factor of φ(λ) of multiplicity k.
14 CHAPTER 1. DIFFERENTIAL EQUATIONS
and
> dsolve(diff(x(t),t$2)-4*diff(x(t),t)+3*x(t)=6*t-11+8*exp(-t));
d3 x d2 x
+ 2 = 4 − 12 exp(2t). (1.84)
dt3 dt
The auxiliary equation is
λ2 (λ + 1) = 0, (1.85)
We have
and
dxn−1
= xn (t),
dt
dxn
= F (x1 , x2 , . . . , xn ; t)
dt
of n first-order equations with independent variable t and n dependent variables
x1 , . . . , xn . In fact this is just a special case of
dx1
= F1 (x1 , x2 , . . . , xn ; t),
dt
dx2
= F2 (x1 , x2 , . . . , xn ; t),
dt
.. ..
. . (1.91)
dxn−1
= Fn−1 (x1 , x2 , . . . , xn ; t),
dt
dxn
= Fn (x1 , x2 , . . . , xn ; t),
dt
where the right-hand sides of all the equations are now functions of the variables
x1 , x2 , . . . , xn .3 The system defined by (1.91) is called an n-th order dynamical
system. Such a system is said to be autonomous if none of the functions F is
an explicit function of t.
Picard’s theorem generalizes in the natural way to this n-variable case as
does also the procedure for obtained approximations to a solution with Picard
iterates. That is, with the initial condition x (τ ) = ξ ,
= 1, 2, . . . , n, we define
(j)
the set of sequences {x (t)},
= 1, 2, . . . , n with
(0)
x (t) = ξ ,
t (1.92)
(j+1) (j)
x (t) = ξ + F (x1 (u), . . . , x(j)
n (u); u)du, j = 1, 2, . . .
τ
for all
= 1, 2, . . . , n.
3 Of course, such a system is not, in general, equivalent to one n-th order equation.
16 CHAPTER 1. DIFFERENTIAL EQUATIONS
j = 0, 1, . . . . (1.97)
In the limit j → ∞ (1.96) becomes the MacLaurin expansion for sin(ωt) and
(1.97) the MacLaurin expansion for ω cos(ωt).
x(t)
x(t0 )
where
x1 G1 (x; t)
x2 G2 (x; t)
x= . , G(x; t) = .. , (1.103)
.. .
xm Gm (x; t)
where
p1 ∂x1 /∂t
p2 ∂x2 /∂t
p= . = .. . (1.105)
.. .
pm ∂xm /∂t
A rather more general case is when, for the system defined by equations (1.98)
and (1.99), there exists a scalar field U (x; t) with
over the range t = 0 to t = 10, with initial conditions x(0) = 1, y(0) = −1.
The MAPLE routine dsolve can be used for systems with the equations and
the initial conditions enclosed in curly brackets. Unfortunately the solution is
returned as a set {x(t) = · · · , y(t) = · · ·}, which cannot be fed directly into
the plot routine. To get round this difficulty we set the solution to some
variable (Fset in this case) and extract x(t) and y(t) (renamed as f x(t) and
f y(t)) by using the MAPLE function subs. These functions can now be plotted
parametrically. The complete MAPLE code and results are:
> Fset:=dsolve(
> {diff(x(t),t)=x(t)-y(t),diff(y(t),t)=x(t),x(0)=1,y(0)=-1},
> {x(t),y(t)}):
> fx:=t->subs(Fset,x(t)):
> fx(t);
1 (1/2 t) 1 √ √ 1 √
e (3 cos( t 3) + 3 3 sin( t 3))
3 2 2
> fy:=t->subs(Fset,y(t)):
> fy(t);
1 (1/2 t) √ 1 √ 1 √
e (3 3 sin( t 3) − 3 cos( t 3))
3 2 2
> plot([fx(t),fy(t),t=0..10]);
20 CHAPTER 1. DIFFERENTIAL EQUATIONS
250
200
150
100
50
For the system given by (1.108) it is clear that a equilibrium point is a sta-
tionary point of U (x) and for the conservative system given by (1.103)–(1.106)
equilibrium points have p = 0 and are stationary points of V (x). An equilib-
rium point is useful for obtaining information about phase behaviour only if
we can determine the behaviour of trajectories in its neighbourhood. This is a
matter of the stability of the equilibrium point, which in formal terms can be
defined in the following way:
The equilibrium point x∗ of (1.111) is said to be stable (in the sense of
Lyapunov) if there exists, for every ε > 0, a δ(ε) > 0, such that any solution
x(t), for which x(t0 ) = x(0) and
|x∗ − x(0) | < δ(ε), (1.113)
satisfies
|x∗ − x(t)| < ε, (1.114)
for all t ≥ t0 . If no such δ(ε) exists then x∗ is said to be unstable (in the
sense of Lyapunov). If x∗ is stable and
lim |x∗ − x(t)| = 0. (1.115)
t→∞
There is a warning you should note in relation to these definitions. In some texts
the term stable is used to mean what we have called ‘asymptotically stable’ and
equilibrium points which are stable (in our sense) but not asymptotically stable
are called conditionally or marginally stable.
a
x2 = a
x
0
Figure 1.2: The bifurcation diagram for Example 1.6.1. The stable and unstable
equilibrium solutions are shown by continuous and broken lines and the direction
of the flow is shown by arrows. This is an example of a simple turning point
bifurcation.
According to the Lyapunov criterion it is stable if, when the phase point is
perturbed a small amount from x∗ it remains in a neighbourhood of x∗ , asymp-
totically stable if it converges on x∗ and unstable if it moves away from x∗ .
We shall, therefore, determine the stability of equilibrium points by linearizing
about the point.5
Example 1.6.1 Consider one-variable non-linear system given by
ẋ(t) = a − x2 . (1.118)
The parameter a can vary over all real values and the nature of equilibrium
points will vary accordingly.
√
The equilibrium points are given by x = x∗ = ± a. They exist only when
a ≥ 0 and form the parabolic curve shown in Fig. 1.2. Let x = x∗ + x and
substitute into (1.118) neglecting all but the linear terms in x.
dx
= a − (x∗ )2 − 2x∗ x. (1.119)
dt
5 A theorem establishing the formal relationship between this linear stability and the Lya-
dx
= −2x∗ x, (1.120)
dt
which has the solution
gives
2
x(0) x(0)
x(t) = ẋ(t) = − . (1.123)
1 + tx(0) 1 + tx(0)
Then
0, as t → ∞ if x(0) > 0,
x(t) → (1.124)
−∞, as t → 1/|x(0)| if x(0) < 0.
In each case x(t) decreases with increasing t. When x(0) > 0 it takes ‘forever’
to reach the origin. For x(0) < 0 it attains minus infinity in a finite amount of
time and then ‘reappears’ at infinity and decreases to the origin as t → ∞. The
linear equation (1.120) cannot be applied to determine the stability of x∗ = 0
as it gives (dx/dt)∗ = 0. If we retain the quadratic term we have
dx
= −(x)2 . (1.125)
dt
So including the second degree term we see that dx/dt < 0. If x > 0 x(t)
moves towards the equilibrium point and if x < 0 it moves away. In the
∗
strict Lyapunov sense the√ equilibrium point x = 0 is unstable. But it is ‘less
∗
unstable’ that x = − a, for a > 0, since there is a path of attraction. It
is at the boundary between the region where there are no equilibrium points
and the region where there are two equilibrium points. It is said to be on the
margin of stability. The value a = 0 separates the stable range from the unstable
range. Such equilibrium points are bifurcation points. This particular type of
bifurcation is variously called a simple turning point, a fold or a saddle-node
bifurcation. Fig.1.2 is the bifurcation diagram.
24 CHAPTER 1. DIFFERENTIAL EQUATIONS
= ∓2x a + c2 c ± a + c2 . (1.128)
a
x2 = a
x
0
Figure 1.3: The bifurcation diagram for Example 1.6.2, c = 0. The stable and
unstable equilibrium solutions are shown by continuous and broken lines and the
direction of the flow is shown by arrows. This is an example of a supercritical
pitchfork bifurcation.
1.6. AUTONOMOUS SYSTEMS 25
a
x2 = a
c
0 x
−c2
Figure 1.4: The bifurcation diagram for Example 1.6.2, c > 0. The stable and
unstable equilibrium solutions are shown by continuous and broken lines and
the direction of the flow is shown by arrows. This gives examples of both simple
turning point and transcritical bifurcations.
c > 0. √
The √equilibrium point x = c + a + c2 is stable. The equilibrium point x =
c − a + c2 is unstable for a < 0 and stable for a > 0. The point x = c,
a = −c2 is a simple turning point bifurcation and x = a = 0 is a transcritical
bifurcation. That is the situation when the stability of two crossing lines of
equilibrium points interchange. The bifurcation diagram for this example is
shown in Fig.1.4.
c < 0.
This is the mirror image (with respect to the vertical axis) of the case c > 0.
Example 1.6.3
Ax = c, (1.140)
where
x1 c1
x2 c2
x=
..
,
c=
..
,
(1.141)
. .
xn cn
x = A−1 c. (1.142)
in the variable λ. Suppose that there are n distinct roots λ(1) , λ(2) , . . . , λ(n) .
Then Rank{A− λ(k) I} = n− 1 for all k = 1, 2, . . . , n. So there is, corresponding
to each eigenvalue λ(k) , a left eigenvector v (k) and a right eigenvector u(k) which
are solutions of the linear equations
The eigenvectors are unique to within the choice of one arbitrary component.
Or equivalently they can be thought of a unique in direction and arbitrary in
length. If A is symmetric it is easy to see that the left and right eigenvectors
are the same.7 Now
and since λ(k) = λ(j) for k = j the vectors v (k) and u(j) are orthogonal. In fact
since, as we have seen, eigenvectors can always be multiplied by an arbitrary
(k) (k)
constant we can ensure that the sets √ {u } and {v } are orthonormal by
dividing each for u(k) and v (k) by u(k) .v (k) for k = 1, 2, . . . , n. Thus
where
Kr 1, k = j,
δ (k − j) = (1.149)
j,
0, k =
F = Ax − c, (1.150)
Ax = c. (1.152)
7 The vectors referred to in many texts simply as ‘eigenvectors’ are usually the right eigen-
vectors. But it should be remembered that non-symmetric matrices have two distinct sets of
eigenvectors. The left eigenvectors of A are of course the right eigenvectors of AT and vice
versa.
1.6. AUTONOMOUS SYSTEMS 29
As we saw in Sect. 1.6.2 these can be either no solutions points, one solution
or an infinite number of solutions. We shall concentrate on the case where A is
non-singular and there is a unique solution given by
x∗ = A−1 c. (1.153)
As in the case of the first-order system we consider a neighbourhood of the
equilibrium point by writing
x = x∗ + x. (1.154)
Substituting into (1.151) and using (1.153) gives
dx
= Ax. (1.155)
dt
Of course, in this case, the ‘linearization’ used to achieve (1.155) was exact
because the original equation (1.151) was itself linear.
As in Sect. 1.6.2 we assume that all the eigenvectors of A are distinct and
adopt all the notation for eigenvalues and eigenvectors defined there. The vector
x can be expanded as the linear combination
x(t) = w1 (t)u(1) + w2 (t)u(2) + · · · + wn (t)u(n) , (1.156)
of the right eigenvectors of A, where, from (1.148),
wk (t) = v (k) .x(t), k = 1, 2, . . . , n. (1.157)
Now
Ax(t) = w1 (t)Au(1) + w2 (t)Au(2) + · · · + wn (t)Au(n)
= λ(1) w1 (t)u(1) + λ(2) w2 (t)u(2) + · · · + λ(n) wn (t)u(n) (1.158)
and
dx
= ẇ1 (t)u(1) + ẇ2 (t)u(2) + · · · + ẇn (t)u(n) . (1.159)
dt
Substituting from (1.158) and (1.159) into (1.155) and dotting with v (k) gives
ẇk (t) = λ(k) wk (t), (1.160)
with solution
wk (t) = C exp λ(k) t . (1.161)
So x will grow or shrink in the direction of u(k) according as λ(k) >, < 0.
The equilibrium point will be unstable if at least one eigenvalue has a positive
real part and stable otherwise. It will be asymptotically stable if the real part
of every eigenvalue is (strictly) negative. Although these conclusions are based
on arguments which use both eigenvalues and eigenvectors, it can be seen that
knowledge simply of the eigenvalues is sufficient to determine stability. The
eigenvectors give the directions of attraction and repulsion.
30 CHAPTER 1. DIFFERENTIAL EQUATIONS
Example 1.6.4 Analyze the stability of the equilibrium points of the linear
system
The matrix is
0 1
A= , (1.163)
4 3
(i) λ(1) and λ(2) both real and (strictly) positive. x grows in all directions.
This is called an unstable node.
(ii) λ(1) and λ(2) both real with λ(1) > 0 and λ(2) < 0. x grows in all direc-
tions, apart from that given by the eigenvector associated with λ(2) . This,
as indicated above, is called a saddle-point.
(iii) λ(1) and λ(2) both real and (strictly) negative. x shrinks in all directions.
This is called a stable node.
(iv) λ(1) and λ(2) conjugate complex with {λ(1) } = {λ(2) } > 0. x grows in
all directions, but by spiraling outward. This is called an unstable focus.
(v) λ(1) = −λ(2) are purely imaginary. Close to the equilibrium point, the length
of x remains approximately constant with the phase point performing a
closed loop around the equilibrium point. This is called an centre.
(vi) λ(1) and λ(2) conjugate complex with {λ(1) } = {λ(2) } < 0. x shrinks
in all directions, but by spiraling inwards. This is called an stable focus.
Example 1.6.5 Analyze the stability of the equilibrium points of the linear
system
with
x 2 −3 −4
x= , A= , c= . (1.166)
y −1 2 1
The matrix is
2 −3
A= , (1.167)
−1 2
with Det{A} = 1, has inverse
2 3
A−1 = . (1.168)
1 2
So the unique equilibrium point is
∗
2 3 −4 −5
x = = . (1.169)
1 2 1 −2
Linearizing√about x∗ gives an equation of the form (1.155). The eigenvalues of
A are 2 ± 3. Both these numbers are positive so the equilibrium point is an
unstable node.
giving
ar02
r02 + exp(−2at){a − r02 } , a = 0,
r2 (t) = (1.180)
r02
, a = 0.
1 + 2tr02
This gives
0, a ≤ 0,
r(t) −→ (1.181)
√
a, a > 0.
When a < 0 trajectories spiral with a constant angular velocity into the origin.
When a = 0 linear analysis indicates that the origin is a centre. However, the√full
solution shows that orbits converge to the origin as t → ∞, with r(t)
1/ 2t,
which is a slower rate of convergence than any exponential.
y y
x x
x2 + y 2 = a
(a) (b)
Figure 1.5: A Hopf bifurcation with (a) a ≤ 0, (b) a > 0.
34 CHAPTER 1. DIFFERENTIAL EQUATIONS
y
x
Problems 1
1) Find the general solutions of the following differential equations:
[There are each of one of the types described in Sects. 1.3.1-3. The first thing
to do is identify the type.]
[These are all equations with constant coefficients as described in Sect. 1.4.]
3) Find the general solution of the differential equation
2.1 Introduction
Consider a function x(t), where the variable t can be regarded as time. A linear
transform G is an operation on x(t) to produce a new function x̄(s). It can be
pictured as
are all examples of linear transformations. Now let us examine the case of the
transform given by the differential equation
dx̄(t) x̄(t) c
+ = x(t). (2.5)
dt T T
37
38 CHAPTER 2. LINEAR TRANSFORM THEORY
dx̄(t) x̄(t) d
exp(t/T ) + exp(t/T ) = [exp(t/T )x̄(t)]
dt T dt
c
= exp(t/T ) x(t). (2.6)
T
This gives
t
c
Z
x̄(t) = exp[−(t − u)/T ]x(u)du + x̄(0) exp(−t/T ). (2.7)
T 0
In the special case of a constant input x(t) = 1, with the initial condition
x̄(0) = 0,
where x(t) is periodic in t with period 2π and s now takes the discrete values
s = 0, ±1, ±2, . . .. The inverse of this transformation is the ‘usual’ Fourier series
s=∞
X
x(t) = x̄(s) exp(ist). (2.10)
s=−∞
Γ (z + 1) = zΓ (z), (2.12)
Γ (1) = 1. (2.13)
1 In the case of a light wave the Fourier series transformation determines the spectrum.
Since different elements give off light of different frequencies the spectrum of light from a star
can be used to determine the elements present on that star.
2.2. SOME SPECIAL FUNCTIONS 39
H(t − t0 )
0 t0 t
Clearly
(
0, if t ≤ t0 ,
H(t − t0 ) = (2.17)
1, if t > t0 ,
defined using the limit of a normal distribution curve as the width contacts and the height
increases, while maintaining the area under the curve.
3 Henceforth, unless otherwise stated, we shall use x̄(s) to mean the Laplace transform of
x(t).
2.3. LAPLACE TRANSFORMS 41
It is clear that the Laplace transform satisfies the linear property (2.1). That
is
A constant C.
Z ∞
C
L{C} = C exp(−st)dt = , <{s} > 0. (2.26)
0 s
0, if t0 < 0,
(
D
L{δ (t − t0 )} = <{s} > 0. (2.37)
exp(−st0 ), if t0 ≥ 0,
2.3. LAPLACE TRANSFORMS 43
dp x̄(s) dp exp(−st)
∞
Z
= x(t) dt (2.40)
dsp 0 dsp
Z ∞
= (−1)p tp x(t) exp(−st)dt, (2.41)
0
as long as the function is such as to allow differentiation under the integral sign.
In these circumstances we have, therefore,
dp L{x(t)}
L{tp x(t)} = (−1)p , for integer p ≥ 0. (2.42)
dsp
It is clear the (2.28) is the special case of this result with x(t) = 1.
dp x(t) dp x(t)
Z ∞
L = exp(−st)dt
dtp 0 dtp
p−1 ∞ Z ∞ p−1
d x(t) d x(t)
= p−1
exp(−st) + s exp(−st)dt
dt 0 0 dtp−1
p−1 p−1
d x(t) d x(t)
= − + sL . (2.43)
dtp−1 t=0 dtp−1
p−1
dp x(t) dj x(t)
X
L = sp L{x(t)} − sp−j−1 . (2.44)
dtp j=0
dtj t=0
44 CHAPTER 2. LINEAR TRANSFORM THEORY
As expected, when t0 ≤ 0 the presence of the Heaviside function does not affect
the transform. When t0 > 0 make the change of variable u = t − t0 . Then
L{x(t)}, if t0 ≤ 0,
L{x(t)H(t − t0 )} = (2.46)
exp(−st0 )L{x(t + t0 )}, if t0 > 0.
is called the convolution of x(t) and y(t). It is not difficult to see that
Z t Z t
x(u)y(t − u)du = y(u)x(t − u)du. (2.49)
0 0
Now define
Z λ Z t
Iλ (s) = dt du x(u)y(t − u) exp(−st). (2.51)
0 0
The region of integration is shown in Fig. 2.2. Suppose now that the functions
are such that we can reverse the order of integration4 Then (2.51) becomes
Z λ Z λ
Iλ (s) = du dt x(u)y(t − u) exp(−st). (2.52)
0 u
4 To do this it is sufficient that x(t) and y(t) are piecewise continuous.
2.3. LAPLACE TRANSFORMS 45
t=u
0 λ t
Now take the limit λ → ∞ and, given that x(t), y(t) and s are such that the
integrals converge it follows from (2.50), (2.51) and (2.53) that
Z t
L x(u)y(t − u)du = x̄(s)ȳ(s). (2.54)
0
A special case of this result is when y(t) = 1 giving ȳ(s) = 1/s and
Z t
x̄(s)
L x(u)du = . (2.55)
0 s
The results of Sects. 2.3.1 and 2.3.2 are summarized in Table 2.1
Cp!
Ctp p ≥ 0 an integer, <{s} > 0.
sp+1
Γ (ν + 1)
tν ν 6= −1, −2, −3, . . ., <{s} > 0.
sν+1
1
exp(−αt) <{s} > <{α}.
s+α
s
cosh(αt) <{s} > |<{α}|.
s2 − α 2
α
sinh(αt) <{s} > |<{α}|.
s2 − α 2
s
cos(ωt) <{s} > |={ω}|.
s2 + ω 2
ω
sin(ωt) <{s} > |={ω}|.
s2 + ω 2
x(c t) (1/c)x̄(s/c)
dp x̄(s)
tp x(t) (−1)p p ≥ 0 an integer.
dsp
p−1
dp x(t) p
X p−j−1 dj x(t)
s x̄(s) − s p ≥ 0 an integer.
dtp dtj t=0
j=0
x(t)H(t − t0 ) exp [−st0 H(t0 )] x̄1 (s) Where x1 (t) = x(t + t0 H(t0 )).
Z t
x(u)y(t − u)du x̄(s)ȳ(s) The convolution integral.
0
2.3. LAPLACE TRANSFORMS 47
This is the case of a particle of unit mass moving on a line with simple harmonic
oscillations of angular frequency ω, in a medium of viscosity ξω. Suppose that
x(0) = x0 and ẋ(0) = 0. Then from Table 2.1 lines 6 and 12
L {ẍ(t)} = s2 x̄(s) − sx0 ,
(2.57)
L {ẋ(t)} = sx̄(s) − x0 .
So the Laplace transform of the whole of (2.56) is
x̄(s)[s2 + 2ξωs + ω 2 ] = x0 (s + 2ξω). (2.58)
Giving
x0 (s + 2ξω)
x̄(s) = . (2.59)
(s + ξω)2 + ω 2 (1 − ξ 2 )
To find the required solution we must invert the transform. Suppose that ξ 2 < 1
and let θ2 = ω 2 (1 − ξ 2 ).5 Then (2.59) can be re-expressed in the form
s + ξω ξω
x̄(s) = x0 + . (2.60)
(s + ξω)2 + θ2 (s + ξω)2 + θ2
Using Table 2.1 lines 6, 7 and 10 to invert these transforms gives
ξω
x(t) = x0 exp(−ξωt) cos(θt) + sin(θt) . (2.61)
θ
Let ζ = ξω/θ and defined φ such that tan(φ) = ζ. Then (2.61) can be expressed
in the form
p
x(t) = x0 1 + ζ 2 exp(−ξωt) cos(θt − φ). (2.62)
This is a periodic solution with angular frequency θ subject to exponential
damping. We can use MAPLE to plot x(t) for particular values of ω, ξ and x0 :
> theta:=(omega,xi)->omega*sqrt(1-xi^2);
p
θ := (ω, ξ) → ω 1 − ξ2
> zeta:=(omega,xi)->xi*omega/theta(omega,xi);
ξω
ζ := (ω, ξ) →
θ(ω, ξ)
> phi:=(omega,xi)->arcsin(zeta(omega,xi));
> y:=(t,omega,xi,x0)->x0*exp(-xi*omega*t)/sqrt(1-(zeta(omega,xi))^2);#
x0 e(−ξ ω t)
y := (t, ω, ξ, x0 ) → p
1 − ζ(ω, ξ)2
> x:=(t,omega,xi,x0)->y(t,omega,xi,x0)*cos(theta(omega,xi)*t-phi(omega,xi));
> plot(
> {y(t,2,0.2,1),-y(t,2,0.2,1),x(t,2,0.2,1)},t=0..5,style=[point,point,line]);
0.5
0 1 2 3 4 5
t
–0.5
–1
In physical terms the function f (t) is a forcing term imposed on the behaviour
of the oscillator. As we saw in Sect. 1.4, the general solution of (2.63) consists of
the general solution of (2.56) (now called the complementary function) together
with a particular solution of (2.63). The Laplace transform of (2.63) with the
initial conditions x(0) = x0 and ẋ(0) = 0 is
¯
x̄(s)[s2 + 2ξωs + ω 2 ] = x0 (s + 2ξω) + f(s). (2.64)
Giving
x0 (s + 2ξω) ¯
f(s)
x̄(s) = + . (2.65)
(s + ξω) + ω (1 − ξ ) (s + ξω) + ω 2 (1 − ξ 2 )
2 2 2 2
2.3. LAPLACE TRANSFORMS 49
Comparing with (2.59), we see that the solution of (2.63) consists of the sum of
the solution (2.62) of (2.56) and the inverse Laplace transform of
¯
f(s)
x̄p (s) = . (2.66)
(s + ξω)2 + θ2
From Table 2.1 lines 7, 10 and 15
1 t
Z
xp (t) = f (t − u) exp(−ξωu) sin(θu)du. (2.67)
θ 0
So for a particular f (t) we can complete the problem by solving this integral.
However, this will not necessarily be the simplest approach. There are two other
possibilities:
(i) Decompose x̄p (s) into a set of terms which can be individually inverse-
transformed using the lines of Table 2.1 read from right to left.
(ii) Use the integral formula (2.23) for the inverse transform.
If you are familiar with the techniques of contour integration (ii) is often the most
straightforward method. We have already used method (i) to derive (2.60) from
(2.59). In more complicated cases it often involves the use of partial fractions.
As an illustration of the method suppose that
f (t) = F exp(−αt), (2.68)
for some constant F. Then, from Table 2.1,
F
x̄p (s) = . (2.69)
(s + α)[(s + ξω)2 + θ2 ]
Suppose now that (2.69) is decomposed into
A B(s + ξω) + C θ
x̄p (s) = + . (2.70)
(s + α) (s + ξω)2 + θ2
Then
xp (t) = A exp(−αt) + exp(−ξωt)[B cos(θt) + C sin(θt)]. (2.71)
It then remains only to determine A, B and C. This is done by recombining
the terms of (2.70) into one quotient and equating the numerator with that of
(2.69). This gives
s2 (A + B) + s[2A ξω + B(ξω + α) + C θ]
+ A θ2 + Bξωα + C θα = F. (2.72)
Equating powers of s gives
F
A = −B = − ,
2ξωα − α2 − θ2
(2.73)
F(ξω − α)
C= .
θ(2ξωα − α2 − θ2 )
50 CHAPTER 2. LINEAR TRANSFORM THEORY
In general the Laplace transform of the n-th order equation with constant co-
efficients (1.40) will be of the form
¯
x̄(s)φ(s) − w(s) = f(s). (2.74)
where φ(s) is the polynomial (1.43) and w(s) is some polynomial arising from
the application of line 12 of Table 2.1 and the choice of initial conditions. So
¯
w(s) f(s)
x̄(s) = + . (2.75)
φ(s) φ(s)
since the coefficients a0 , a1 , . . . , an−1 are real it has a decomposition of the form
( )
Y m Ỳ
2 2
φ(s) = (s + αj ) [(s + βr ) + γr ] , (2.76)
j=1 r=1
Recombining the quotients to form the denominator φ(s) and comparing co-
efficients of s in the numerator will give all the constants Aj , Br and Cr . If
αj = αj+1 = · · · = αj+p−1 , that is, φ(s) has a real root of degeneracy p then
in place of the p terms in the first summation in (2.77) corresponding to these
factors we include the terms
p (i)
X Aj
. (2.78)
i=1
(s + αj )i
In a similar way for a p-th fold degenerate complex pair the corresponding term
is
p (i) (i)
X Bj (s + βj ) + Cj
. (2.79)
i=1
[(s + βj )2 + γj2 ]i
(i)
Another, often simpler, way to extract the constants Aj in (2.78) (and Aj in
(2.77) as the special case p = 1) is to observe that
p
w(s) X (i)
(s + αj )p = (s + αj )p−i Aj
φ(s) i=1
+ (s + αj )p × [terms not involving (s + αj )]. (2.80)
2.4. THE Z TRANSFORM 51
Thus
p−i
(i) 1 d p w(s)
Aj = (s + α j ) , i = 1, .., p (2.81)
(p − i)! dsp−i φ(s) s=−αj
Once the constants have been obtained it is straightforward to invert the Laplace
transform. Using the shift theorem and the first line of Table 2.1
exp(−αt)ti−1
1
L−1 i
= . (2.83)
(s + α) (i − 1)!
This result is also obtainable from line 11 of Table 2.1 and the observation that
(−1)i−1 di−1
1 1
= . (2.84)
(s + α)i (i − 1)! dsi−1 s + α
The situation is somewhat more complicated for the complex quadratic factors.
However, the approach exemplified by (2.84) can still be used.6 As we saw in
the example given above. The second term of the right hand side of (2.75) can
be treated in the same way except that now f(s)¯ may contribute additional
factors in the denominator. Further discussion of Laplace transforms will be in
the context of control theory.
where conditions are applied to z to ensure convergence of the series. Again, for
later reference, we record the fact that, as a consequence of Cauchy’s theorem,
the inverse of the transform is
1
I
x(k) = z k−1 x̃(z)dz, (2.86)
2iπ C
6 We just have to be more careful because differentiation throws up linear terms in s in the
numerator.
52 CHAPTER 2. LINEAR TRANSFORM THEORY
Also
1
Z{cosh(αk)} = 2 [Z{exp(αk)} + Z{exp(−αk)}]
z[z − cosh(α)]
= 2
, |z| > exp (|<{α}|) (2.91)
z − 2z cosh(α) + 1
z sinh(α)
Z{sinh(αk)} = , |z| > exp (|<{α}|) , (2.92)
z 2 − 2z cosh(α) + 1
z[z − cos(ω)]
Z{cos(ωk)} = 2 , |z| > exp (|={ω}|) , (2.93)
z − 2z cos(ω) + 1
z sin(ω)
Z{sin(ωk)} = − 2 , |z| > exp (|={ω}|) . (2.94)
z − 2z cos(ω) + 1
Other important results can be derived from these using the general properties
derived below. The Kronecker delta function δ Kr (k) is defined in (1.149). The
terms of the sequence δ Kr (k − m), k = 0, 1, . . . are all zero except that for which
k = m, which is unity. Thus
1
Z{δ Kr (k − m)} = . (2.95)
zm
2.4. THE Z TRANSFORM 53
As a generalization of (2.88)
So
Xk
Z x(j)y(k − j) = x̃(z)ỹ(z). (2.103)
j=0
This is the Z transform analogue of the convolution formula (2.55). The results
of Sects. 2.4.1 and 2.4.2 are summarized in Table 2.2.
Cz
C ak |z| > |a|.
z−a
z
exp(−αk) |z| > exp (−<{α}).
z − exp(−α)
z[z − cosh(α)]
cosh(αk) |z| > exp (|<{α}|).
z 2 − 2z cosh(α) + 1
z sinh(α)
sinh(αk) |z| > exp (|<{α}|).
z 2 − 2z cosh(α) + 1
z[z − cos(ω)]
cos(ωk) |z| > exp (|={ω}|).
z 2 − 2z cos(ω) + 1
z sin(ω)
sin(ωk) |z| > exp (|={ω}|).
z 2 − 2z cos(ω) + 1
1
δ Kr (k − m)
zm
ak x(k) x̃(z/a)
p−1
X
x(k + p) z p x̃(z) − x(j)z (p−j) p > 0 an integer.
j=0
dx̃(z)
kx(k) −z
dz
k
X
x(j)y(k − j) x̃(z)ỹ(z) The convolution formula.
j=0
2.4. THE Z TRANSFORM 55
The method can also be used to study systems of difference equations. This is
an example based on a simple model for the buffalo population in the American
West starting in the year 1830.7
Example 2.4.2 Let x(k) and y(k) be the number of female and male buffalo
at the start of any one year (k = 0 is 1830). Five percent of adults die each
year. Buffalo reach maturity at two years and the number of new females alive
at the beginning of year k + 2, taking into account infant mortality, is 12% of
x(k). More male calves than female are born and the corresponding figure is
14% of x(k). Show that in the limit k → ∞ the population grows by 6.3% per
year.
The difference equations are
x(k + 2) = 0.95x(k + 1) + 0.12x(k),
(2.110)
y(k + 2) = 0.95y(k + 1) + 0.14x(k).
Applying the Z transform to these equations and using x(0) = x0 , x(1) = x1 ,
y(0) = y0 and y(1) = y1
z 2 x̃(z) − x0 z 2 − x1 z = 0.95[z x̃(z) − zx0 ] + 0.12x̃(z),
(2.111)
z 2 ỹ(z) − y0 z 2 − y1 z = 0.95[z ỹ(z) − zy0 ] + 0.14x̃(z)
7 Taken from Barnett and Cameron(1985) p. 21.
56 CHAPTER 2. LINEAR TRANSFORM THEORY
Since we are interested only in the long-time behaviour of the total population
p(k) = x(k) + y(k) we need extract simply the formula for p̃(z) from these
equations. With p0 = x0 + y0 , p1 = x1 + y1
(p0 z + p1 − 0.95p0 ) 0.26(x0 z + x1 − 0.95x0 )
p̃(z) = z + . (2.112)
z(z − 0.95) z(z − 0.95)(z 2 − 0.95z − 0.12)
The reason for retaining the factor z in the numerators (with a consequent z in
the denominators) can be seen by looking at the first line of Table 2.2. Factors
of the form z/(z − a) are easier to handle than 1/(z − a). We now resolve the
contents of the brackets into partial fractions. You can if you like use MAPLE
to do this. The code is
> convert(0.26*(x0*z+x1-0.95*x0)/(z*(z-0.95)*(z^2-0.95*z-0.12)),parfrac,z);
> convert((p0*z+p1-0.95*p0)/(z*(z-0.95)),parfrac,z);
p1 −1.052631579 p1 + p0
1.052631579 +
z − .9500000000 z
Thus we have
z(1.053p1 − 2.281x1 )
p̃(z) = (p0 − 1.053p1 − 2.167x0 + 2.281x1 ) +
z − 0.95
z(1.959x0 − 1.843x1 ) z(0.208x0 + 1.843x1 )
+ + (2.113)
z + 0.113 z − 1.063
and inverting the transform using lines 1 and 7 of Table 2.2 gives
p(k) = (p0 − 1.053p1 − 2.167x0 + 2.281x1 )δ Kr (k)
+ (1.053p1 − 2.281x1)(0.95)k
+(1.959x0 − 1.843x1 )(−0.113)k
+ (0.208x0 + 1.843x1 )(1.063)k . (2.114)
In the limit of large k this expression is dominated by the last term
p(k) ' (0.208x0 + 1.843x1 )(1.063)k . (2.115)
The percentage yearly increase is
p(k + 1) − p(k)
× 100 = 6.3. (2.116)
p(k)
2.4. THE Z TRANSFORM 57
Problems 2
1) Show, using the standard results in the table of Laplace transforms, that:
2ωs
(i) L{t sin(ωt)} = ,
(ω 2 + s2 )2
2ω 3
(ii) L{sin(ωt) − ωt cos(ωt)} = .
(ω 2 + s2 )2
Hence solve the differential equation
d3 x(t)
+ x(t) = H(t),
dt3
where x(0) = ẋ(0) = ẍ(0) = 0.
3) Given that x(t) = −t H(t) and
x̄(s)
ȳ(s) = ,
(s − 1)2
find y(t).
4) Show using your notes that
√
− 21 π
L{t }= 1 .
s 2
Show that
n 1
o 1
L Erf(t 2 ) = 1 .
s(s + 1) 2
5) Find the sequences x(0), x(1), x(2), . . . for which Z{x(k)} = x̃(z) are:
z
(i) ,
(z − 1)(z − 2)
z
(ii) 2 ,
z + a2
58 CHAPTER 2. LINEAR TRANSFORM THEORY
z 3 + 2z 2 + 1
(iii) .
z3
6) Use the Z transform to solve the following difference equations for k ≥ 0:
(i) 8x(k + 2) − 6x(k + 1) + x(k) = 9, where x(0) = 1 and x(1) = 32 ,
√
(ii) x(k + 2) + 2x(k) = 0, where x(0) = 1 and x(1) = 2.
Show that in the long time limit the person’s capital changes at a rate of
20% per annum.
Chapter 3
3.1 Introduction
Linear control theory deals with a linear time-invariant system having a set of
inputs {u1 (t), u2 (t), . . .} and outputs {x1 (t), x2 (t), . . .}. The input functions are
controlled by the experimenter, that is, they are known functions. The aim of
control theory is to
(ii) Devise a strategy for choosing the input functions and possibly changing
the design of the system (and hence the equations) so that the output have
some specific required form. If the aim is to produce outputs as close as
possible to some reference functions {ρ1 (t), ρ2 (t), . . .} then the system is
called a servomechanism. If each of the reference functions is constant the
system is a regulator.
Consider, for example, the simple case of one input function u(t) and one output
function x(t) related by the differential equation
dn x dn−1 x dx
+ a n−1 + · · · + a1 + a0 x =
dtn dtn−1 dt
m
d u dm−1 u du
bm m + bm−1 m−1 + · · · + b1 + b0 u, (3.1)
dt dt dt
59
60 CHAPTER 3. TRANSFER FUNCTIONS AND FEEDBACK
and with
i
dx
= 0, i = 0, 1, . . . , n − 1,
dti t=0
j (3.2)
d u
= 0, j = 0, 1, . . . , m − 1.
dtj t=0
Now taking the Laplace transform of (3.1) gives
φ(s)x̄(s) = ψ(s)ū(s), (3.3)
where
φ(s) = sn + an−1 sn−1 + · · · + a1 s + a0 ,
(3.4)
ψ(s) = bm sm + bm−1 sm−1 + · · · + b1 s + b0 .
(Cf (1.43).) Equation (3.3) can be written
x̄(s) = G(s)ū(s), (3.5)
where
ψ(s)
G(s) = (3.6)
φ(s)
is called the transfer function. This system can be represented in block dia-
grammatic form as
G(s) = K. (3.8)
du(t)
x(t) = K , (3.11)
dt
G(s) = K s. (3.12)
with equations
x̄(s) = G2 (s)ū(s),
(3.13)
ȳ(s) = G1 (s)x̄(s),
where G1 (s) is called the process transfer function and G2 (s) is called the con-
troller transfer function. Combining these equations gives
where
ū(s) ȳ(s)
G(s)
Let the control variable u(t) be the position on the heating dial of the oven and
suppose that this is directly proportional to the heat x(t) supplied to the oven by
the heat source. This is the situation of proportional control with x(t) = K1 u(t).
Let the output variable y(t) be the temperature difference at time t between the
oven and its surroundings. Some of the heat supplied to the oven will be lost by
radiation. This will be proportional to the oven temperature. So the heat used
to raise the temperature of the oven is x(t) − K2 y(t). According to the laws of
thermodynamics this is Qẏ(t), where Q is the heat capacity of the oven. Then
we have
From (3.18)–(3.19)
K1
ȳ(s) = ū(s). (3.20)
Qs + K2
Suppose the dial is turned from zero to a value u0 at t = 0. Then u(t) = u0 H(t)
and ū(s) = u0 /s. So
u0 K 1
ȳ(s) = . (3.21)
s(Qs + K2 )
K1
y(t) = u0 [1 − exp(−t/T)]H(t). (3.23)
K2
3.2. SYSTEMS IN CASCADE 63
Suppose now that the proportional control condition (3.16) is replaced by the
integral control condition
Z t
x(t) = K1 u(τ )dτ. (3.24)
0
K1
x̄(s) = ū(s). (3.25)
s
K1
ȳ(s) = ū(s). (3.26)
s(Qs + K2 )
If we now use the form u(t) = u0 H(t) this is in fact equivalent to x(t) = K1 u0 t
which implies a linear buildup of heat input over the time interval [0, t]. Formula
(3.21) is replaced by
u0 K 1
ȳ(s) = . (3.27)
s2 (Qs + K2 )
K1 1 T T
ȳ(s) = u0 − + −1 , (3.28)
K2 s2 s T +s
T K1 t
y(t) = u0 − 1 + exp(−t/T) H(t). (3.29)
K2 T
We now use MAPLE to compare the results of (3.23) and (3.29) (with u0 K1 /K2 =
1, T = 2).
> plot({t-2+2*exp(-t/2),1-exp(-t/2)
> },t=0..5,style=[point,line]);
64 CHAPTER 3. TRANSFER FUNCTIONS AND FEEDBACK
0 1 2 3 4 5
t
It will be see that with proportional control the temperature of the oven reaches
a steady state, whereas (if it were allowed to do so) it would rise steadily with
time for the case of integral control.
ū2 (s)
ū2 (s)
We shall also need to represent a device which receives an input and transmits
it unchanged in two (or more) directions. This will be represented by
ȳ(s) ȳ(s)
ȳ(s)
A simple example of the use of this formalism is the case where equations (3.13)
are modified to
ȳ(s) = G1 (s)x̄(s),
ū2 (s)
G3 (s)
x̄2 (s)
ū1 (s) x̄1 (s) + x̄(s) ȳ(s)
G2 (s) + G1 (s)
66 CHAPTER 3. TRANSFER FUNCTIONS AND FEEDBACK
3.4 Feedback
Feedback is present in a system when the output is fed, usually via some feedback
transfer function, back into the input to the system. The classical linear control
system with feedback can be represented by the block diagram
¯
f(s) ȳ(s)
H(s)
f¯(s) = H(s)ȳ(s),
ȳ(s) = G(s)v̄(s).
G(s)
ȳ(s) = ū(s). (3.32)
1 + G(s)H(s)
The block diagram for this problem is obtained by introducing feedback into
the block diagram of Example 3.2.1.
¯
f(s) ȳ(s)
H(s)
3.4. FEEDBACK 67
From (3.20)
K1
ȳ(s) = v̄(s) (3.33)
Qs + K2
and
v̄(s) = ū(s) − H(s)ȳ(s). (3.34)
Giving
K1
ȳ(s) = ū(s). (3.35)
Q s + K2 + H(s)K1
To complete this problem we need to make some assumptions about the nature
of the feedback and we must also know the form of u(t). Suppose as in Example
3.2.1 u(t) = u0 H(t), giving ū(s) = u0 /s and assume a proportional feedback.
That is H(s) = H, a constant. Then
u0 K 1
ȳ(s) = . (3.36)
s[Q s + (K2 + HK1 )]
Comparing with equations (3.21)–(3.23) we see that the effect of the feedback
is to replace K2 by K2 + HK1 . The solution is therefore
K1
y(t) = u0 [1 − exp(−t/T 0 )]H(t), (3.37)
K2 + HK1
where T 0 = Q/(K2 + HK1 ). With HK1 > 0, T 0 < T , so the feedback promotes
a faster response of the output to the input. A typical case is that of unitary
feedback where H(s) = 1.
Example 3.4.2 We have a heavy flywheel, centre O, of moment of inertia
I. Suppose that P designates a point on the circumference with the flywheel
initially at rest and P vertically below O. We need to devise a system such that,
by applying a torque Ku(t) to the wheel, in the long-time limit OP subtends an
angle y ∗ with the downward vertical.
O
y
Ku(t)
P
68 CHAPTER 3. TRANSFER FUNCTIONS AND FEEDBACK
d2 y
I = Ku(t). (3.38)
dt2
Let J = I/K and take the Laplace transform of (3.38). Remembering that
y(0) = ẏ(0) = 0
1
ȳ(s) = ū(s). (3.39)
Js2
and the block diagram is
ū(s) 1 ȳ(s)
Js2
Suppose that the torque is u(t) = u0 H(t). Then ū(s) = u0 /s and inverting the
transform
u0 2
y(t) = t . (3.40)
2J
The angle grows, with increasing angular velocity, which does not achieve the
required result. Suppose now that we introduce feedback H(s). The block
diagram is modified to
¯
f(s) ȳ(s)
H(s)
√
where ω0 = 1/ J. Again the objective is not achieved since y(t) oscillates about
u0 . Suppose we now modify the feedback to
Then
u0 u0 (s + 12 aω02 ) + 21 aω02
ȳ(s) = = − u 0 , (3.46)
s(Js2 + as + 1) s (s + 12 aω02 )2 + ω 2
aω02
1 2
y(t) = u0 1 − exp − 2 atω0 cos(ωt) + sin(ωt) . (3.47)
2ω
dy(t)
f (t) = a + y(t). (3.48)
dt
Problems 3
1) For the system with block diagram
ū(s) K ȳ(s)
+ s(s+Q)
−
show that
Kū(s)
ȳ(s) = .
s2 + Qs + K
H2
ū(s) − 1 1 ȳ(s)
+ +
− s+Q s
H1
show that
ū(s)
ȳ(s) = .
s2 + s(H2 + Q) + H1
(Hint: Put in all the intermediate variables, write down the equations associ-
ated with each box and switch and eliminate to find the relationship between
ū(s) and ȳ(s).)
3) Discrete time systems, where input u(k) is related to output (response) y(k)
by a difference equation can be solved by using the Z transform to obtain a
formula of the type ỹ(z) = G(z)ũ(z), where G(z) is the discrete time version
of the transfer function. For the following two cases find the transfer function
(i) y(k) − 2y(k − 1) = u(k − 1).
(ii) y(k) + 5y(k − 1) + 6y(k − 2) = u(k − 1) + u(k − 2).
Obtain y(k), in each case when u(k) = 1 for all k.
Chapter 4
Controllability and
Observability
4.1 Introduction
In Sect. 3.1 we introduced a system with inputs {u1 (t), u2 (t), . . .} and outputs
{x1 (t), x2 (t), . . .}. In Sect. 3.2 this was modified to take into account the fact
that the number of x-variables may be large or be difficult to handle. They may
also include unnecessary details about the system. If the x-variables are now
renamed state-space variables a new, possibly smaller, set {y1 (t), y2 (t), . . .} of
output variables can be introduced. In terms of transfer functions this situation
is represented by systems in cascade with one set of equations relating input
variables to state-space variables and a second set relating state-space variables
to output variables. The Laplace transformed equations for one variable of each
type are (3.13). In general if there are m input variables n state-space variables
and r output variables we need a matrix formulation to specify the system. The
most general form for a linear system with continuous time is
where u(t) is the m-dimensional column vector of input variables, x(t) is the
n-dimensional column vector of input variables and y(t) is the r-dimensional
column vector of output variables. The matrices A(t), B(t) and C(t) are re-
spectively n × n, n × m and r × n. Equation (4.1) is a system of n first-order
differential equations. They will be equations with constant coefficients if the
elements of the matrix A are not time-dependent. The control system repre-
sented by the pair of equations (4.1) and (4.2) is said to be a constant system if
none of the matrices A, B and C is time-dependent.
The concentration on first-order differential equations is not a serious re-
striction. As we saw in Sect. 1.5, a single n-th order differential equation can be
71
72 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
f (t) = aT x(t),
Equations (4.3) can be combined into one equation and rewritten in the form
ẋ(t) = Ax(t) + u(t)b, (4.6)
where
0 1
!
T
A = X − ba = . (4.7)
−J−1 −aJ−1
Equations (4.6) and (4.4) are a particular case of (4.1) and (4.2) with m = r = 1
and n = 2.
Of course, this formula is of practical use for determining x(t) only if we have
a closed-form expression for exp(At). One case of this kind would be when
A has a set of n distinct eigenvalues λ(j) , j = 1, 2, . . . , n and we are able to
calculate all the left eigenvectors p(j) and right eigenvectors q (j) , which satisfy
the orthonormality condition (1.148).1 Then the matrix P formed by having
T
the vectors p(j) (in order) as rows and the matrix Q formed by having the
P = Q−1 , (4.13)
P AQ = Λ, (4.14)
where Λ is the diagonal matrix with the eigenvalues (in order) as diagonal
elements. Then
1 2 1
exp(At) = I + tQΛP + t (QΛP )2 + · · · + tk (QΛP )k + · · ·
2! k!
1 1
= Q[I + tΛ + t2 Λ2 + · · · + tk Λk + · · ·]P
2! k!
= Q exp(Λt)P , (4.15)
where exp(Λt) is the diagonal matrix with diagonal elements exp(λ(j) t), j =
1, 2, . . . n (in order).
The problem with this method is that it involves calculating (or using
MAPLE to calculate) all the eigenvalues and the left and right eigenvectors.
It is also valid only when all the eigenvalues are distinct, so that the left and
right eigenvectors are orthonormalizible. We now develop a method of obtaining
exp(At) as a polynomial in A which depends only on deriving the eigenvalues
and which is valid even if some are degenerate. The characteristic equation of
A is
∆(λ) = 0, (4.16)
where
usage.
74 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
is called the characteristic polynomial. The zeros of ∆(λ) are the eigenvalues of
A. Suppose that the eigenvalues are λ(1) , λ(2) , . . . , λ(m) , where λ(j) is µ(j) -fold
degenerate. Then
m
X
µ(j) = n (4.18)
j=1
and
(j)
!
dµ −1 ∆(λ)
(j) d∆(λ)
∆(λ )= = ··· = = 0,
dλ λ=λ(j) dλµ(j) −1
λ=λ(j)
j = 1, 2, . . . , m (4.19)
This is easily proved when (4.13) and (4.14) are valid (no degenerate eigenvalues,
µj = 1, for j = 1, 2, . . . , m). Then
A = QΛP , (4.21)
As = QΛs P , s = 1, 2, . . . (4.22)
and
Since the eigenvalues satisfy the characteristic equation, the matrix obtained
by summing all the terms in the square brackets has every element zero, which
establishes the theorem. The result still holds for repeated eigenvalues but the
proof is a little longer. An important result for our discussion is the following:
Theorem 4.2.1 The power series for exp(zt) can be decomposed in the form
where
about each of its poles z = λ(1) , λ(2) , . . . , λ(m) in order. Thus, for λ(1) ,
(1)
µ ∞
X γi X h ik
g(z) = i + ρk z − λ(1)
i=1 z − λ(1) k=0
P1 (z)
= µ(1) + g1 (z), (4.27)
z − λ(1)
Then
∆(λ) = λ2 − λ − 6, (4.33)
with eigenvalues λ(1) = 3, λ(2) = −2. Now let
exp(zt) = β0 (t) + β1 (t)z + ∆(z)f (z), (4.34)
76 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
giving
Thus
1
exp(At) = 5
I[2 exp(3t) + 3 exp(−2t)] + 51 A[exp(3t) − exp(−2t)]
! !
1 2 1 1 3 −1
= 5 exp(3t) + 5 exp(−2t) . (4.36)
6 3 −6 2
and inverting the Laplace transform it follows, on comparison with (4.12) that
L−1 (sI − A)−1 = exp(At). (4.39)
s −1
!
sI − A = (4.40)
−6 s − 1
and
s−1 1
!
1
(sI − A)−1 = , (4.41)
∆(s) 6 s
where ∆(s) is given by (4.33). Inverting this Laplace transformed matrix ele-
ment by element gives (4.36).
where the n-dimensional column vector b replaces the matrix B and the n-
dimensional row vector cT replaces the matrix C. Let T be an n×n non-singular
matrix. Equations (4.42) and (4.43) can be expressed in the form
dT x(t)
= T AT −1 T x(t) + T b u(t), (4.44)
dt
With
x0 (t) = T x(t), T AT −1 = A0 ,
(4.46)
T b = b0 , cT T −1 = (c0 )T ,
this gives
dx0 (t)
= A0 x0 (t) + b0 u(t), (4.47)
dt
These equations still have the same input and output variables, but the state
variables and matrices have been transformed. These equivalent expressions of
the problem are called realizations and are denoted by [A, b, cT ] and [A0 , b0 , (c0 )T ].
The Laplace transforms of (4.42) and (4.43) are
where
dn y dn−1 y dy
+ a n−1 + · · · + a1 + a0 y =
dtn dtn−1 dt
dm u dm−1 u du
bm m
+ bm−1 m−1 + · · · + b1 + b0 u, (4.57)
dt dt dt
with n > m and initial conditions
i
dy
= 0, i = 0, 1, . . . , n − 1,
dti
j t=0 (4.58)
d u
= 0, j = 0, 1, . . . , m − 1.
dtj t=0
Then
dm x 1 dm−1 x1 dx1
y(t) = bm m
+ b m−1 m−1
+ · · · + b1 + b 0 x1 , (4.62)
dt dt dt
4.3. REALIZATIONS OF SYSTEMS 79
and define
x2 (t) = ẋ1 (t),
d2 x 1
x3 (t) = ẋ2 (t) = ,
dt2
.. .. (4.63)
. .
dn−1 x1
xn (t) = ẋn−1 (t) = .
dtn−1
Then (4.62) becomes
Let
n
X
ẋn (t) = κu(t) + γk−1 xk (t) (4.66)
k=1
which is equivalent to
1 dn x 1 γn−1 dn−1 x1 γ1 dx1 γ0
u(t) = n
− −··· − − x1 (t).
κ dt κ dtn−1 κ dt κ
(4.67)
We now substitute into (4.57) from (4.62) and (4.67). This gives
m
X dj w
bj = 0, (4.68)
j=0
dtj
where
n−1
dn x 1 X γ i di x 1
1
w(t) = 1− + a i + . (4.69)
κ dtn i=0
κ dti
Thus, by choosing
κ = 1, γi = −ai , i = 1, 2, . . . , n − 1, (4.70)
80 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
(4.68), and hence (4.57), are identically satisfied and (4.66) and (4.63) can be
combined in the matrix form (4.42) with
0 1 0 ··· ··· ··· 0 0
0 0 1 0 ··· ··· 0 0
.. .. .. .. .. .. .. ..
A = . . . . . . . . (4.71)
0 0 0 0 ··· ··· 0 1
−a0 −a1 · · · · · · · · · · · · · · · −an−1
0
0
..
b = . . (4.72)
0
1
When A, b and cT are of this type they are said to be in companion form. It is
not difficult to show that the characteristic function for A is
∆(λ) = φ(λ), (4.73)
again establishing that the poles of the transfer function are the eigenvalues of
A.
For any realization, the n × n matrix
U = b Ab A2 b · · · An−1 b ,
(4.74)
that is the matrix with j-th column Aj−1 b, is3 called the controllability matrix.
The n × n matrix
cT
cT A
c T A2
V = , (4.75)
..
.
cT An−1
that is the matrix with j-th row cT Aj−1 , is4 called the observability matrix.
3 For reasons which we shall see below.
4 Again for reasons which we shall see below.
4.3. REALIZATIONS OF SYSTEMS 81
Proof:
ξT A
T = .. (4.78)
.
ξ T An−1
ξ T Ab 0
Tb = .. = .. = b0 . (4.80)
. .
ξ T An−1 b 1
ξ T Ak ej = δ Kr (k + 1 − j), k = 0, 1, . . . , n − 1. (4.81)
82 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
So in each row of A0 apart from the last there is one non-zero element equal to
unity. In the k-th row the element is the k + 1-th which is exactly the form of
(4.71) if we define
aj = −ξT An Aej+1 , j = 0, 1, . . . , n − 1. (4.83)
Finally we note that
(c0 )T = c T e1 c T e2 · · · cT en (4.84)
b T AT
bT (AT )2
T
U = . (4.88)
..
.
bT (AT )n−1
It follows that the controllability and observability matrices of a realization
(4.42)–(4.43) are the transposes of the observability and controllability matrices
of it dual (4.55)–(4.56). This is, of course, a symmetric relationship because the
transpose of the transpose of a matrix is the matrix itself and the dual of the
dual of a realization is the realization itself. This idea can be developed further
by defining an alternative companion realization. This is one where the matrix
A has a form like the transpose of (4.71) and c replaces b in (4.72). We then
have the theorem:
Proof: We first take the dual [AT , c, bT ] of [A, b, cT ]. Since V T is the control-
lability matrix of [AT , c, bT ], there exists a matrix T giving a companion real-
ization [T AT T −1 , T c, bT T −1 ] if and only if V is nonsingular. Taking the dual of
this realization gives an alternative companion realization
−1
[T 0 AT 0 , T 0 b, cT (T 0 )−1 ], where T 0 = (T T )−1 = (T −1 )T .
−2 −2 −3 1 5
− 32 − 35 2
ξT = 3 , T =
1 2 −1
, (4.92)
−1 −2 2
with
6 6 1
1
Det{T } = 3 , −3 −2 0 .
T −1 = (4.93)
0 1 1
Then the companion realization is given by
0 1 0
0
A0 = T AT −1 = 0 1
,
−6 −11 −6 (4.94)
0
b0 = T b =
0 ,
(c0 )T = cT T −1 = 3 3 3 .
1
84 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
20
− 25 26
63 63 63
13 11 2
= − 63 (4.97)
63 63
5 1 25
− 63 63 63
and
−13 −31 16
−15 −30 18 .
Ta = (4.98)
−2 −5 5
Then the alternative companion realization is given by
0 0 −6
1 0 −11 ,
A00 = Ta A(Ta )−1 =
0 1 −6
(4.99)
3
b00 = Ta b =
3 ,
(c00 )T = cT (Ta )−1 = 0 0 1 .
3
Example 4.3.2 Determine a diagonal realization for the system given by ȳ(s) =
G(s)ū(s), where
3(s2 + s + 1)
G(s) = . (4.100)
(s + 1)(s + 2)(s + 3)
Then define state variables x1 (t), x2 (t) and x3 (t) with Laplace transforms re-
lated to the input variable by
Inverting the Laplace transforms and expressing the results in matrix form gives
(4.42) and (4.43) with
3
−1 0 0 1 2
A= 0 −2 0
, 1 ,
b= −9 .
c= (4.104)
0 0 −3 1 21
2
Let
ū(s) x̄3 (s) ū(s)
x̄3 (s) = , x̄2 (s) = = ,
s+1 s+1 (s + 1)2
(4.107)
x̄2 (s) ū(s) ū(s) ū(s)
x̄1 (s) = = , x̄4 (s) = , x̄5 (s) =
s+1 (s + 1)3 s+2 s+3
86 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
Then
sx̄1 (s) = −x̄1 (s) + x̄2 (s),
ȳ(s) = 12 x̄1 (s) − 34 x̄2 (s) + 78 x̄3 (s) − x̄4 (s) − 81 x̄5 (s). (4.109)
Inverting the Laplace transforms and expressing the results in matrix form gives
(4.42) and (4.43) with
−1 1 0 0 0 0
0 −1 1 0 0 0
A= 0 0 −1 0 0 , b = 1 ,
0 0 0 −2 0 1
(4.110)
0 0 0 0 −3 1
1
cT = 2
− 34 7
8
−1 1
8
.
In this case A is diagonal in the rows and columns containing the non-degenerate
eigenvalues −2 and −3 but has a 3 × 3 block corresponding to the 3-fold de-
generate eigenvalue −1. The matrix is said to be in Jordan canonical form and
this is an example of a Jordan realization.
We have seen that a system defined by its transfer function can lead to different
realization of the same dimensions related by a transformation (4.46). In fact
the realizations need not be of the same dimension.
Example 4.3.4 Consider the system with a realization for which
−1 0 1 1 1
A= −1 −2 −1 , b= 0 , c= 1 . (4.111)
−2 −2 −3 1 −1
We can now find the transfer function using (4.52). The code to do the calcu-
lation in MAPLE is
> with(linalg):
Warning, new definition for norm
4.3. REALIZATIONS OF SYSTEMS 87
−1 0 1
" #
A := −1 −2 −1
−2 −2 −3
> b:=array([[1],[0],[1]]);
1
" #
b := 0
1
> ct:=array([[1,1,-1]]);
ct := 1 1 −1
> II:=array([[1,0,0],[0,1,0],[0,0,1]]);
1 0 0
" #
II := 0 1 0
0 0 1
> W:=s->inverse(s*II-A);
W := s → inverse(s II − A)
> W(s);
s+4 1 1
−2
s2
+ 5s +6 s3
+ 6 s2
+ 11 s + 6 s2 + 4s +3
s2 + 4 s + 5
1 1
− 2 − 2
s3 + 6 s2 + 11 s + 6
s + 5s +6 s +4s+3
1 1 1
−2 2 −2 2
s + 5s + 6 s +5s+6 s+3
> G:=s->simplify(multiply(ct,W(s),b));
> G(s);
88 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
h 1 i
3
(s + 3) (s + 2)
It is not difficult to show that this system, with the transfer function
3
G(s) = , (4.112)
(s + 2)(s + 3)
also has the two-dimensional diagonal realization
! ! !
−2 0 1 3
A= , b= , c= . (4.113)
0 −3 1 −3
4.4 Controllability
As indicated above an essential step in dealing with many control problems is
to determine whether a desired outcome can be achieved by manipulating the
input (control) variable. The outcome is determined in terms of a particular set
of state variables and the controllability is that of the realization rather than
the system. In fact the outcome y(t) does not play a role and the equation of
interest is (4.42).
The realization given by (4.42) is controllable if, given a finite time interval
[t0 , tf ] and state vectors x0 and xf , we can find an input u(t) over the interval
[t0 , tf ] such that x(t0 ) = x0 and x(tf ) = xf .
Since we are concerned with a constant system we can, without loss of generality,
set t0 = 0. Then from (4.12),
Z tf
xf = exp (Atf ) x0 + exp(−Aτ )bu(τ )dτ . (4.114)
0
Since xf , x0 and tf are all arbitrary it is both sufficient and necessary for
controllability that for any state vector x∗ and time interval tf we can find
an input u(t) to satisfy
Z tf
x∗ = exp(−Aτ )bu(τ )dτ. (4.116)
0
Proof:
We use the polynomial formula (4.31) for the exponential matrix and thus, from
(4.116)
n−1
X Z tf
k
x ∗
= A b βk (−τ )u(τ )dτ
k=0 0
= UΩ (4.117)
where
ΩT = Ω0 Ω1 · · · Ωn−1 (4.118)
and
Z tf
Ωk = βk (−τ )u(τ )dτ. (4.119)
0
Thus
and
!
1 −t
exp(−At) = exp(2t) . (4.126)
0 1
We know because the system is controllable that for any given tf , x∗1 and x∗2 we
can find a form for u(t) which will satisfy (4.127) and (4.128). Suppose we try
the form
4.5 Observability
Allied to the concept of controllability is that of observability. The essence of
this concept is that by observing the the input and output of a system over a
period of time the state at the beginning of that time can be inferred. More
formally:
4.5. OBSERVABILITY 91
The realization given by (4.42) and (4.43) is observable if, given a time
t0 , there exists a time interval [t0 , tf ] such that the initial state x(t0 ) = x0 is
determined by the input function u(t) and the output function y(t) over [t0 , tf ].
Let
Z t
T
y (t) = y(t) − c exp (At)
∗
exp(−Aτ )bu(τ )dτ. (4.133)
0
Proof:
We use the polynomial form (4.31) for the exponential matrix and, from (4.134)
n−1
X
y ∗ (t) = βk (t)cT Ak x0
k=0
= β T (t)V x0 , (4.135)
where
β T (t) = β0 (t) β1 (t) · · · βn−1 (t) (4.136)
Sufficiency. Suppose V is non-singular and suppose that two states x0 and x00
both satisfy (4.135). Then
β T (t)V 4x0 = 0, (4.137)
where 4x0 = x0 − x00 . Since (4.137) holds for all t in the interval [0, tf ], it must
be the case that
V 4x0 = 0, (4.138)
Since V is non-singular (4.138) has the unique solution 4x0 = 0. That is
x0 = x00 , which means that the realization is observable.
92 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
Necessity. Suppose that V is singular. Then there exists a vector p such the
V p = 0. Then if x0 satisfies (4.135) so does x0 +µp for any µ and the realization
is not observable.
It is clear from these results that a realization is controllable if and only if its
dual realization is observable.
(i) controllable and observable then all realizations of dimension n are con-
trollable and observable.
They differ only in the vector c. We showed in Example 4.3.1 that (4.89) was
both controllable and observable and it is therefore minimal, with nmin = 3. In
Example 4.3.4 we derived the transfer function (4.112) corresponding to (4.111)
and showed that it also had a realization of dimension 2. This means, of course,
that (4.111) cannot be minimal. Since it is controllable (in this respect it is
identical to (4.89)) it cannot be observable. We confirm this by working out the
observability matrix. From (4.75) and (4.111)
1 1 −1
V =
0 0
3
(4.139)
−6 −6 −9
4.6. MINIMAL REALIZATIONS 93
cT = c−1 2−c 1−c .
Since A and b are not functions of c it is clear that the controllability or un-
controllability of the system is not affected by the value of c. From (4.74) and
(4.145)
1 −1 1
U = 0 1 −4 . (4.146)
1 −2 4
94 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
Since Det{U } = −1 the realization is controllable. Now from (4.75) and (4.145)
c−1 2−c 1−c
V = 1 − c 2c − 4 c . (4.147)
c − 1 8 − 4c −4
(4.148)
cT = 1 0 .
with
! !
0 1 1 0
U= , V = . (4.149)
1 −2 −2 1
For c = 2
! !
−1 0 1
A= , b= ,
0 −2 1
(4.150)
cT = 1 −1 .
with
! !
1 −1 1 −1
U= , V = . (4.151)
1 −2 −1 2
In each case the realization is both controllable and observable and this minimal
with nmin = 2.
It is clear, from this example, that the dimension of a realization, that is the
number of state variables, corresponds, when it is derived from a transfer func-
tion G(s), to the number of partial fractions of G(s). This in term is simply
4.6. MINIMAL REALIZATIONS 95
the degree of the polynomial denominator of G(s). In the cases we have ob-
served, the realization was minimal unless there was some cancellation in fac-
tors between the numerator and denominator of G(s). The following theorem,
therefore, comes as no surprise.
Theorem 4.6.3 Suppose that a system is given by (4.59)–(4.61) where m < n.
Then φ(s) and ψ(s) have no common factors, the degree n of the denominator
φ(s) is the dimension of minimal realizations of the system.
Problems 4
1) Let A be a 2 × 2 matrix with eigenvalues λ and µ. Show that, when λ 6= µ,
[λ exp(µt) − µ exp(λt)]I + [exp(λt) − exp(µt)]A
exp(At) = .
λ−µ
What is the corresponding result when λ = µ?
2) Consider the system with realization
0 1 0 0
ẋ(t) = 0 0 1
x(t) + 0
u(t),
2 1 −2 1
y(t) = 1 2 0 x(t).
Use the exponential matrix to find the output y(t), for u(t) = Kt, where K
is a constant and x(0) = 0.
3) For the system with realization
− 43 − 14
! !
1
ẋ(t) = x(t) + u(t),
− 21 − 12 1
y(t) = 4 2 x(t)
calculate the transfer function G(s), the controllability matrix U and the
observability matrix V .
4) For the system with realization
! !
−1 −1 1
ẋ(t) = x(t) + u(t),
2 −4 3
y(t) = −1 1 x(t)
verify that the controllability matrix is non-singular and find the matrix T
which transforms it to the companion realization.
96 CHAPTER 4. CONTROLLABILITY AND OBSERVABILITY
5) Obtain Jordan representations for the systems which have transfer functions:
2
(i) G(s) = s + s +31 ,
(s + 1)
(ii) G(s) = 4 .
(s + 1)2 (s + 3)
Chapter 5
Stability
The Euclidean norm of a vector of real elements is, of course, the ‘usual’ modulus
of the vector. For any two matrices P and Q (with appropriate dimensions)
and any scalar µ the following properties hold:
(i) ||Q|| > 0 unless Q = 0.
(ii) ||µQ|| = |µ| ||Q||.
(iii) ||P + Q|| ≤ ||P || + ||Q||.
(iv) ||P Q|| ≤ ||P || ||Q||.
97
98 CHAPTER 5. STABILITY
and defined what it meant to say that the equilibrium point x∗ , which satisfies
F (x∗ ) = 0, (5.5)
Proof:
According to the definition, x∗ = 0 is an asymptotically stable equilibrium
point of (5.6) if x(t) → 0 as t → ∞. The solution of (5.6) is
Proof: From (5.9) and (4.132), with the intial time t = t0 replacing t = 0,
Z t
y(t) = cT exp[A(t − t0 )]x(t0 ) + cT exp[A(t − τ )]bu(τ )dτ. (5.12)
t0
Since A is a stability matrix, it follows from Thm. 5.1.1 that there exist positive
constants K and k such that
Z t
||y(t)|| ≤ ||c|| || exp[A(t − t0 )]|| ||x(t0 )|| + ||c|| || exp[A(t − τ )]|| ||b|| |u(τ )|dτ
t0
Z t
≤ K||c|| ||x(t0 )|| exp[k(t0 − t)] + ||b||B1 exp[k(τ − t)]dτ
t0
= K||c|| ||x(t0 )|| exp[k(t0 − t)] + k −1 ||b||B1 [1 − exp[k(t0 − t)]
≤ K||c|| ||x(t0 )|| + k −1 ||b||B1 .
(5.13)
The proof of this theorem is somewhat more complicated and will be omitted.
We can see on the basis of these two theorems that the asymptotic stability of
x∗ = 0 is a stronger condition than bounded input–bounded output stability
and can be deduced from the bounded input–bounded output stability only for
minimal1 realizations.
where
ψ(s) θ(s)
G(s) = , G0 (s) = (5.15)
φ(s) φ(s)
are respectively the transfer function, as before, and the contribution from the
initial conditions, with
φ(s) = sn + an−1 sn−1 + · · · + a1 s + a0 ,
(5.18)
The way to build this determinant is as follows:
(i) Extend the range of the definition of the coefficients over all integer values
by defining a` = 0 if ` > n or ` < 0.
(ii) The i–j-th element is an−2i+j .
Thus
(a) For n = 2
φ(s) = a 2 s2 + a 1 s + a 0 , (5.19)
a1 a2 a1 a2
Φ2 = = . (5.20)
a−1 a0 0 a0
(b) For n = 3
φ(s) = a 3 s3 + a 2 s2 + a 1 s + a 0 , (5.21)
a2 a3 a4 a2 a3 0
Φ3 = a0 a1 a2 = a0 a1 a2 . (5.22)
a−2 a−1 a0 0 0 a0
(c) For n = 4
φ(s) = a 4 s4 + a 3 s3 + a 2 s2 + a 1 s + a 0 , (5.23)
a3 a4 a5 a6 a3 a4 0 0
a1 a2 a3 a4 a1 a2 a3 a4
Φ4 = = . (5.24)
a−1 a0 a1 a2 0 a0 a1 a2
a−3 a−2 a−1 a0 0 0 0 a0
102 CHAPTER 5. STABILITY
(k)
For any n we now define a hierarchy of subdeterminants Φn , k = 1, 2, . . . , n−1,3
(k)
where Φn is the (n − k) × (n − k) determinant obtained by deleting the last k
rows and columns from Φn . Thus, for example,
a3 a4
(2)
Φ4 = . (5.25)
a1 a2
From (5.18)
Φn(n−1) = an−1 , (5.27)
Φn = Φ(0)
n = a0 Φ(1)
n . (5.28)
These results mean that (5.26) can be confined to the range k = 1, 2, . . . , n − 2
with the additional conditions an−1 > 0 and a0 > 0.
The conditions a3 = 1 > 0 and a0 = 2 > 0 are satisfied and another necessary
condition for all the roots to have negative real part is a2 = κ > 0. The only
other condition for sufficiency is, from (5.22), given using
κ 1
(1)
Φ3 = = 3κ − 2. (5.30)
2 3
So according to the Routh-Hurwitz criterion the roots all have a negative real
part if κ > 2/3. We can check this out using MAPLE .
> phi:=(s,k)->s^3+k*s^2+3*s+2;
φ := (s, k) → s3 + k s2 + 3 s + 2
> fsolve(phi(s,1)=0,s,complex);
3 We (0)
shall also, for convenience, use Φn = Φn .
5.2. STABILITY AND THE TRANSFER FUNCTION 103
> fsolve(phi(s,3/4)=0,s,complex);
> fsolve(phi(s,2/3)=0,s,complex);
> fsolve(phi(s,1/2)=0,s,complex);
> fsolve(phi(s,0)=0,s,complex);
1
1+Q3 s
In this example we are given that an = a4 , an−1 = a3 and a0 are all positive.
The only remaining conditions for the zeros to have negative real parts are
a3 a4 0
(1)
= a1 a2 a3 = a1 (a2 a3 − a1 a4 ) − a0 a23 > 0,
Φ4 (5.34)
0 a 0 a1
a3 a4
(2)
Φ4 = = a2 a3 − a1 a4 > 0, (5.35)
a1 a2
K(1 + Q2 s)(1 + Q3 s)
G(s) = . (5.37)
(1 + Q1 s)(1 + Q3 s)Js2 + K(1 + Q2 s)
This simplifies to
which shows that (5.33) is a necessary condition for stability. The maximum
value of K is given, from (5.40) by
J(Q1 + Q3 )(Q2 − Q1 − Q3 )
> K. (5.41)
Q1 Q22 Q3
5.3. STABILITY AND FEEDBACK 105
f¯(s) ȳ(s)
H(s)
The transfer function in the absence of feedback would be GOL (s) and in the
presence of feedback it becomes
GOL (s)
GCL (s) = . (5.42)
1 + GOL (s)H(s)
GOL (s) and GCL (s) are often referred to as open-loop and closed-loop transfer
functions (hence the notation) and the signal
v̄(s) = ū(s) − f¯(s) = ū(s) − H(s)ȳ(s) (5.43)
is called the error. In the present context this type of feedback will be called
output feedback. For a system which is unstable it is sometimes possible, as we
saw in Example 3.4.2, to achieve stability by altering parameters in the feedback.
The way this happens can be seen if we let
ψ(s)
GCL (s) = , (5.44)
φ(s)
with
φ(s) = sn + an−1 sn−1 + · · · + a1 s + a0 ,
(5.45)
ψ(s) = bm sm + bm−1 sm−1 + · · · + b1 s + b0 .
Now suppose that the change in feedback leads to the modification of the closed-
loop transfer function to
ψ(s)
G0CL (s) = , (5.46)
φ(s) + χ(s)
where
χ(s) = hn−1 sn−1 + · · · + h1 s + h0 . (5.47)
106 CHAPTER 5. STABILITY
We know that the system is stable if the zeros of the denominator of the closed-
loop transfer function all have negative real parts and that this can be ascer-
tained using the Routh-Hurwitz criterion. If the system is unstable with φ(s)
alone in the denominator, then the introduction of the extra factor χ(s) can be
used to stabilize it. Now suppose that the introduction of χ(s) corresponds to
replacing the feedback H(s) in the block diagram by H(s) + 4H(s). Then
GOL (s) ψ(s)
= . (5.48)
1 + GOL (s)[H(s) + 4H(s)] φ(s) + χ(s)
From (5.42), (5.44) and (5.48)
χ(s)
4H(s) = . (5.49)
ψ(s)
Although we have considered 4H(s) simply to be an additive factor in the
feedback, it could also be applied as an additional feedback on the system.
Consider the block diagram
ū(s) ȳ(s)
+ + G (s)
− − OL
H(s)
4H(s)
From (5.42)
GCL (s) GOL (s)
G0CL (s) = = . (5.50)
1 + GCL (s)H(s) 1 + GOL (s)[H(s) + 4H(s)]
This result does, of course, illustrate the general point that feedback can be
added either as elements of one feedback loop or as a succession of separate
loops as shown in this block diagram.
thus giving
˙ = [A − b hT ]x(t) + bu(t).
x(t) (5.52)
The net effect of this change is to replace the coefficients aj in the characteristic
polynomial ∆(λ), or equivalently in φ(s) by aj + hj . Thus φ(s) is replaced by
φ(s) + χ(s) as was the case in going from the transfer function of (5.44) to that
of (5.46). It is clear that the state feedback can be used to ensure that A is a
stability matrix, or equivalently, that all the zeros of φ(s) + χ(s) have negative
real parts.
6 −11 6 1
(5.54)
cT = 1 0 0 .
is minimal but that the system is unstable. By the use of a state feedback obtain
a stable realization with eigenvalues −1, −2 and −3. Determine the transfer
function of the original system and the form of 4H(s) required to produce the
stabilization.
108 CHAPTER 5. STABILITY
From (4.74) and (4.75) the controllability and observability matrices are respec-
tively
0 0 1 1 0 0
0 1
U = 6
, 0 1 0 .
V = (5.55)
1 6 25 0 0 1
It is easy to see that this equation has one root λ = 1 and thus to extract the
remaining roots λ = 2 and λ = 3. It follows that the system is unstable. The
polynomial with the required roots is
Subtracting (5.56) and (5.57) we see that the coefficients of the state feedback
are h0 = 12, h1 = 0 and h2 = 12 with
We now calculate the transfer function from this realization using (4.53). It is
not difficult to show that
and, of course,
Giving
1
G(s) = . (5.61)
s3 − 6s2 + 11s − 6
1 1
G0 (s) = = 3 . (5.62)
φ(s) + χ(s) s + 6s2 + 11s + 6
This change from a condition on the real part of the eigenvalues to their mag-
nitudes is easily understood when when realize that we have changed from the
exponential matrix to the matrix itself.4 The definition of bounded input–
bounded output stability of Sect. 5.1 carries over to the discrete-time case with
the index k replacing t. Thm. 5.1.3, relating the asymptotic stability of x = 0
and bounded input–bounded output stability is also valid and can be proved in
a similar way from (5.67). A slightly different approach, which we shall outline
for the two-dimensional case, is to note that the characteristic function of A is
∆(λ) = λ2 + a1 λ + a0 . (5.71)
With a change of variable this is the denominator in (5.65), which on applying
partial fractions can be written in the form
C1 z ũ(z) C2 z ũ(z)
ỹ(z) = + , (5.72)
z − λ1 z − λ2
for constants C1 , C2 and the eigenvalues λ1 and λ2 of A. Using the last line of
Table 2.2 to invert the Z transform gives
k
X k
X
y(k) = C1 λj1 u(k − j) + C2 λj2 u(k − j). (5.73)
j=0 j=0
Example 5.4.1 Consider the stability of the female and male buffalo popula-
tions discussed in Example 2.4.2.
4 Note that | exp(ζ)| < 1 if and only if <{ζ} < 0.
5.4. DISCRETE-TIME SYSTEMS 111
In fact, of course these are precisely the numbers which appear in our solu-
tion (2.114). A is not a convergent matrix and the zero-population state is not
asymptotically stable. According to this simplified model, as we saw in Exam-
ple 2.4.2 the buffalo population would grow by 6.3% a year. In fact, due to
indiscriminate slaughter,5 the population of buffalo fell from 60 million in 1830
to 200 in 1887. Attempts are currently being made to reintroduce buffalo in
the plains of South Dakota. Even then of course it may ultimately be necessary
and economically desirable to implement a policy of culling.6 Suppose that γ%
of females and ξ% of males are culled each year. Then
x2 (k + 1) = 0.12x1 (k) + (0.95 − 0.01γ)x2 (k),
(5.79)
x4 (k + 1) = 0.14x1 (k) + (0.95 − 0.01ξ)x4 (k),
and A is replaced by
0 1 0 0
0.12 0.95 − 0.01γ 0 0
A0 = , (5.80)
0 0 0 1
0.14 0 0 0.95 − 0.01ξ
where
1 h p i
ω (±) (γ) = 95 − γ ± 13825 − 190γ + γ 2 . (5.82)
200
5 Encouraged by the United States Government.
6A single buffalo carcass will provide about 250 kg. of meat, enough for 10 people for year.
112 CHAPTER 5. STABILITY
It will be observed that the culling of male has no affect on the long-term pop-
ulation. This is because we have assumed, perhaps unrealistically, that the
number of calves born is proportional to the number of female adults, irrespec-
tive of the number of males. The root of ∆0 (λ) which leads to the explosion in
population is ω (+) (γ). This is a decreasing function of γ with ω (+) (7) = 1. So
a 7% cull of females will have the effect of stabilizing the population at a fixed
value. The size of this population and the proportions of males and females,
given particular initial populations can be calculated as in Example 2.4.2. This
is an example of stabilization by linear feedback.
Problems 5
1) Show that the system with state-space equations
is unstable. Derive the transfer function. Now investigate the situation when
the input is changed to u(t) − γx1 (t) for some constant γ and interpret this
change as an output feedback 4H(s). Show that the system is asymptot-
ically stable if γ > 2 and find y(t) given that γ = 5, u(t) = u0 H(t) and
x1 (0) = x2 (0) = 0.
2) Consider the system with block diagram given on page 105 of the notes and
K(α + βs)
GOL (s) = , H(s) = 1,
s(1 + 2s)2
where α and β are positive. Determine the closed loop transfer function and
using the Routh-Hurwitz stability criterion show that
(i) If β < α then the system is asymptotically stable for 0 < K < (α−β)−1 .
(ii) If α < β the system is asymptotically stable for all K > 0.
Find a minimal realization when α = 1, β = 2, K = −6.
3) Consider the system with block diagram given on page 105 of the notes and
1
GOL (s) = , H(s) = γ.
s3 + s 2 + s + 1
Determine the closed loop transfer function and show that the system is
unstable for all γ > 0. Show by including the output feedback 4H(s) =
αs2 + βs with suitable values of α β and γ the system can be stabilized with
poles of the closed-loop transfer function at s = −1, −2, −3. With these
values of α, β and γ determine the output when u(t) = u0 H(t). (The initial
values of y(t) and any derivatives may be assumed zero.)
Chapter 6
Optimal Control
For fixed τA and τB the curve γ, including in general the points A and B will
vary with the functional form of x(τ ), as will I, which known as a functional.
The technique for finding a form for x(τ ) which, for a specific f (x(τ ), ẋ(τ ); τ )
and designated constraints on A and B, gives an extreme value for I[x] is the
calculus of variations.
113
114 CHAPTER 6. OPTIMAL CONTROL
wire descends from one end of the wire at x = y = 0 to the other end for which
x = xB , y = yB . If the particle has mass m and the acceleration due to gravity
is g then the total energy is conserved with
( 2 )
2
1 dx dy
2
m + − mgy = 0. (6.3)
dt dt
Now suppose the equation of the path is x = w(y) with w(0) = 0. Then
2
dy
1 + [w0 (y)]2 = 2gy.
(6.4)
dt
This equation can be integrated to find the total time
Z yB s
1 1 + [w0 (y)]2
T [w] = √ dy (6.5)
2g 0 y
that the bead takes for the path x = w(y). The problem now is to find the
function w(y), describing the shape of the wire, which minimizes T [w] subject
to whatever constraint we impose on xB .
Since
dx(τ ) dẋ(τ )
= ξ(τ ), = ξ̇(τ ), (6.8)
dε dε
df
= ∇x f [x∗ (τ ) + ε ξ(τ ), ẋ∗ (τ ) + ε ξ̇(τ ); τ ] · ξ(τ )
dε
Now we need to substitute from (6.9) into (6.7). In doing so we apply integration
by parts to the second term
Z τB h iτB Z τB d(∇ f )
ẋ
∇ẋ f · ξ̇(τ )dτ = ∇ẋ f · ξ(τ ) − · ξ(τ )dτ. (6.10)
τA τA τA dτ
6.1. DIGRESSION: THE CALCULUS OF VARIATIONS 115
x∗ (τ ) + ε ξ(τ )
B
x∗ (τ )
Figure 6.1: Two paths (one the extremum) from A to B in the phase space Γn .
The first term on the right-hand side of (6.12) is zero and for (6.11) and
(6.12) to be satisfied for any ξ(τ ) satisfying (6.13) we must have
d(∇ẋ f )
− ∇x f = 0. (6.14)
dτ
In scalar form this equation is
d ∂f ∂f
− = 0, j = 1, 2, . . . , n. (6.15)
dτ ∂ ẋj ∂xj
116 CHAPTER 6. OPTIMAL CONTROL
(ii) A is a fixed point but the location of the final point on the path
is allowed to vary.
In this case the Euler-Lagrange equations must still be satisfied but we
must also have
or in scalar form
∂f
= 0, j = 1, 2, . . . , n. (6.17)
∂ ẋj τ =τB
For reasons which will be explained below (6.16) and (6.17) are known as
transversality conditions.
From (6.14)
d(∇ẋ f )
ẋ(τ )· − ẋ(τ )·∇x f = 0, (6.18)
dτ
giving
d[ẋ(τ )·∇ẋ f ]
− ẍ(τ )·∇ẋ f − ẋ(τ )·∇x f = 0. (6.19)
dτ
Now
df ∂f
= ẍ(τ )·∇ẋ f + ẋ(τ )·∇x f + , (6.20)
dτ ∂τ
so we have the alternative form
d ∂f
[ẋ(τ )·∇x f − f ] + = 0, (6.21)
dτ ∂τ
for the Euler-Lagrange equations.
In whatever form they are represented and given a particular function f (x, ẋ, τ ),
the Euler-Lagrange equations are a set of n second-order differential equations
for the variables x1 (τ ), x2 (τ ), . . . , xn (τ ). In two special cases first-integrals can
be derived immediately:
(a) When f is not a function of xj for some j it follows from (6.15) that
∂f
= constant. (6.22)
∂ ẋj
We first note that if xB is allowed to vary along the line y = yB then the
transversality condition (6.17) applies giving C = ∞, and thus ẋ(y) = 0. The
wire is vertically downwards with xB = 0, as would be expected. We now
suppose that both ends of the wire are fixed. Let ẋ(y) = cot(θ). Then, from
(6.25), and choosing the sign so that y ≤ 0,
y = − 21 C2 [1 + cos(2θ)]. (6.26)
Thus
dx
= −C2 [1 + cos(2θ)]. (6.27)
dθ
When y = 0, θ = 21 π, so integrating (6.27) and choosing the constant of inte-
gration so that x = 0 when θ = 12 π gives
> x:=(theta,c)->c^2*(2*theta-Pi+sin(2*theta))/2;
1 2
x := (θ, c) → c (2 θ − π + sin(2 θ))
2
> y:=(theta,c)->-c^2*(1+cos(2*theta))/2;
1
y := (θ, c) → − c2 (1 + cos(2 θ))
2
> plot([x(theta,1),y(theta,1),theta=Pi/2..2*Pi]);
118 CHAPTER 6. OPTIMAL CONTROL
1 2 3 4
0
–0.2
–0.4
–0.6
–0.8
–1
Given particular values for xB and xB the equations can be solved to obtain C
and θB .
Then Hamilton’s principle of least action states that the path in Γn which rep-
resents the configuration of the particles in a time interval [tA , tB ] from a fixed
6.1. DIGRESSION: THE CALCULUS OF VARIATIONS 119
Integral Constraints
In this case the constraint is of the form
Z τB
J [x] = g(x(τ ), ẋ(τ ); τ ) dτ = J, (6.35)
τA
where J is some constant. To solve this problem we use the method of Lagrange’s
undetermined multipliers. We find an extremum for
Z τB
I[x] + pJ [x] = [f (x(τ ), ẋ(τ ); τ ) + p g(x(τ ), ẋ(τ ); τ )] dτ, (6.36)
τA
Example 6.1.3 Consider all the curves in the x–y plane between (0, 0) and
(2, 0) which are of length π. Find the equation of the one which encloses the
maximum area between it and the x–axis.
which is the upper semicircle of the circle radius one centre (1, 0). This is an
example where we have needed a maximum rather than a minimum stationary
value. As is usually the case with variational problems the evidence that this
result does indeed correspond to the enclosing of a maximum area is not difficult
to find. You can reduce the area to almost zero with a smooth curve of length
π in the first quadrant from (0, 0) to (2, 0).
Non-Integral Constraints
In this case the constraint is of the form
Example 6.2.1 Suppose we have a system with the block diagram shown at
the beginning of Sect. 5.3.1 with
1
GOL (s) = , H(s) = 1. (6.54)
Qs
122 CHAPTER 6. OPTIMAL CONTROL
The aim it to run the system so that y(tI ) = yI and y(tF ) = yF while at the
same time minimizing the time average of the square error [v(t)]2 , where, from
(5.43)
Thus we have
Z tF
I[u, y] = [u(t) − y(t)]2 dt. (6.56)
tI
ÿ(t) = 0. (6.60)
The simplest way to do the problem is to use (6.68) to replace u(t) in (6.69).
Alternatively (6.68) can be included by using an undetermined multiplier. As
an exercise we shall try the second method. Thus (6.68) can be treated as a
constraint. From (6.52)
Z tF
[u(t)]2 + β[x(t)]2 + p(t)[ẋ(t) − αx(t) − u(t)] dt
Ip [u, x] = (6.70)
tI
subject to the constraint (6.78) with x(0) = 0, ẋ(0) = 1 and x(t) → 0 and
ẋ(t) → 0 as t → ∞.
The task is now to find an extremum for Ip [u, x]. Before deriving the equations
for this we reformulate the problem slightly by defining the Hamiltonian
H(u(t), x(t), p(t); t) = p(t) · X(u(t), x(t); t) − f (u(t), x(t); t). (6.92)
Then (6.91) becomes
Z tF
Ip [u, x] = [p(t) · ẋ(t) − H(u(t), x(t), p(t); t)] dt. (6.93)
tI
4 If in particular problems time derivatives appear in the integrand which do not correspond
to constraints then they can be removed by ‘inventing’ new variables, rather in the way that
we treated the second-order time derivative in Example 6.2.3.
126 CHAPTER 6. OPTIMAL CONTROL
∂H
= 0. (6.95)
∂u
In this context these are often referred to as the Hamilton-Pontriagin equations.
It is also interesting (but not particularly useful) to note that (6.89) can be
rewritten as
ẋ(t) = ∇p H. (6.96)
In scalar form (6.94) and (6.96) are
∂H
ṗj (t) = − , (6.97)
∂xj
∂H
ẋj (t) = , (6.98)
∂pj
for j = 1, 2, . . . , n. Examples 6.2.2 and 6.2.3 can both be formulated in the way
described here. In fact in these cases, since the integrand of the cost functional
does not involve time derivatives, no substitutions from the constraint conditions
are needed. For Example 6.2.2 the Hamiltonian is
H(u(t), x(t), p(t)) = p(t)[u(t) + αx(t)] − [u(t)]2 − β[x(t)]2 (6.99)
and (6.95) and (6.98) yield (6.71). For Example 6.2.3 the Hamiltonian is
H(u(t), x1 (t), x2 (t), p1 (t), p2 (t)) = p1 (t)x2 (t) + p2 (t)[u(t) − x2 (t)] − 4[u(t)]2 − [x1 (t)]2
(6.100)
and (6.95) and (6.98) yield (6.82).
In cases like Example 6.2.2 where the location of the final point on the path
is unrestricted we need transversality conditions. From (6.90) and (6.91) these
are given by
pj (tF ) = 0, if xj (tF ) is unrestricted for j = 1, . . . , n. (6.101)
in agreement with the transversality condition (6.72) of Example 6.2.2.
or in vector form
ẋ(t) = v(t). (6.103)
Then
L(x(t), v(t); t) = 12 mv 2 − V (x; t) (6.104)
and from (6.92) the Hamiltonian is
H(x(t), v(t), p(t); t) = p(t)·v(t) − L(x(t), v(t); t), (6.105)
minimum of f (u(t), x(t); t) and thus a maximum of H(u(t), x(t), p(t); t), which
we assume to be a continuous differentiable function of u over the range (6.110).
Then the condition for an optimum u∗ (t) for u(t) is that
H(u∗ (t), x(t), p(t); t) > H(u∗ (t) + δu, x(t), p(t); t), (6.111)
1 ∂2H
∗ ∗
∂H
(δu)2 + O [δu]3 < 0.
δu + 2 2
(6.112)
∂u ∂u
If there is an extremum of H in the allowed range (6.112) gives the usual con-
ditions for a maximum. Otherwise (6.112) can be approximated by its leading
term to give
∗
∂H
δu < 0. (6.113)
∂u
In these circumstances H is a monotonic function of u. If it is increasing then
δu < 0, which means that u∗ (t) = uu , if it is decreasing then δu > 0 which means
that u∗ (t) = ul . The optimum value of u will be at one of the boundaries of the
range. Since the sign of ∂H/∂u will depend on the other variables x(t) and p(t)
there may, as t varies, be a sudden change of u∗ (t) between ul and uu . This
sudden change is called bang-bang control and
∂H
S(t) = (6.114)
∂u
is called the switching function.
ṗ1 (t) = 0,
(6.118)
ṗ2 (t) = −p1 (t)
p1 (t) = A,
(6.119)
p2 (t) = B − At
and
∂H
= p2 (t) = B − At. (6.120)
∂u
Thus H is a monotonic strictly increasing or strictly decreasing function of u(t)
for all t except at t = B/A if this lies in the interval of time of the journey.
Since the vehicle starts from rest at t = 0 and comes to rest at t = tF it must
be the case that ẍ(0) = u(0) > 0 and ẍ(tF ) = u(tF ) < 0. So in the early part
of the journey u(t) = ub and in the later part of the journey u(t) = −ub . The
switch over occurs when p2 (t) changes sign. So p2 (t) is the switching function
S(t). For the first part of the journey
ẍ(t) = ub ,
ẋ(t) = ub t, (6.121)
x(t) = 21 ub t2 .
ẍ(t) = −ub ,
ẋ(t) = ub (tF − t), (6.122)
1 2
x(t) = xF − u (t
2 b F
− t) .
Since both ẋ(t) and x(t) are continuous over the whole journey the switch occurs
at
t = ts = tF /2, (6.123)
with
p
tF = 2 xF /ub ,
(6.124)
x(ts ) = xF /2.
These results give us the strategy for completing the journey in the minimum
time. Suppose alternatively we chose
u(t) = ub − µt (6.125)
130 CHAPTER 6. OPTIMAL CONTROL
ẋ(t) = ub t − 12 µt2 ,
(6.126)
x(t) = 21 ub t2 − 61 µt3 .
The conditions that the journey ends with zero velocity and x = xF gives
p
µ = 2u3b /3xF ,
p (6.127)
tF = 6xF /ub .
Comparing
p this result with (6.124) we see that this procedure yield a journey
time 3/2 longer. Plots of velocity against distance can be obtained using
MAPLE :
> tF1:=(uB,xF)->2*sqrt(xF/uB);
r
xF
tF1 := (uB , xF ) → 2
uB
> v1:=(uB,xF,t)->uB*t;
v1 := (uB, xF , t) → uB t
> x1:=(uB,xF,t)->uB*t^2/2;
1
x1 := (uB, xF , t) → uB t2
2
> v2:=(uB,xF,t)->uB*(tF1(uB,xF)-t);
v2 := (uB, xF , t) → uB (tF1(uB , xF ) − t)
> x2:=(uB,xF,t)->xF-uB*(tF1(uB,xF)-t)^2/2;
1
x2 := (uB, xF , t) → xF − uB (tF1(uB, xF ) − t)2
2
> tF2:=(uB,xF)->sqrt(6*xF/uB);
r
xF
tF2 := (uB , xF ) → 6
uB
> v3:=(uB,xF,t)->t*uB*(tF2(uB,xF)-t)/tF2(uB,xF);
6.4. PONTRYAGIN’S PRINCIPLE 131
t uB (tF2(uB, xF ) − t)
v3 := (uB , xF , t) →
tF2(uB , xF )
> x3:=(uB,xF,t)->t^2*uB*(3*tF2(uB,xF)-2*t)/(6*tF2(uB,xF));
1 t2 uB (3 tF2(uB, xF ) − 2 t)
x3 := (uB, xF , t) →
6 tF2(uB , xF )
> plot(
> {[x1(1,1,t),v1(1,1,t),t=0..tF1(1,1)/2],[x2(1,1,t),v2(1,1,t),
> t=tF1(1,1)/2..tF1(1,1)],[x3(1,1,t),v3(1,1,t),t=0..tF2(1,1)]
> },linestyle=1);
0.8
0.6
0.4
0.2
The upper curve corresponds to the optimum control and the lower to the control
with linearly decreasing thrust. Since the area under the plot of the reciprocal
of the velocity against distance would give the time of the journey the larger
area in this plot corresponds to a shorter journey time. The latter part of the
upper curve and its extension, which are given parametrically by the second and
third of equations (6.122) is called the switching curve. It represents the points
in the velocity-position space from which the final destination can be reached,
arriving with zero velocity, by applying maximum deceleration. So the point
where this curve is crossed by the first branch of the journey is the point when
switching from acceleration to deceleration must occur.
132 CHAPTER 6. OPTIMAL CONTROL
(6.130)
ṗ1 (t) = 0,
(6.131)
ṗ2 (t) = −p1 (t)
giving (6.118)
p1 (t) = A,
(6.132)
p2 (t) = B − At
and
∂H
= −p2 (t) = At − B. (6.133)
∂u
Now −p2 (t) is the switching function. H is a monotonic strictly increasing or
strictly decreasing function of u(t) for all t except at t = B/A if this lies in the
interval of time of the journey. Since the vehicle begins the landing process with
an upward velocity its engine thrust must be initially downwards. (Otherwise
the initial total downward thrust would be m(g − ub ) which is negative.) There
is one switch to an upward engine thrust to give a soft landing. For the first
part of the journey
ẍ(t) = −(ub + g),
ẋ(t) = vI − t(ub + g), (6.134)
x(t) = xI + vI t − 21 t2 (ub + g).
6.4. PONTRYAGIN’S PRINCIPLE 133
Since both ẋ(t) and x(t) are continuous over the whole journey the switch occurs
at
tF (ub − g) + vI
t = ts = , (6.136)
2ub
with
( s )
1 2ub [vI2 + 2xI (ub + g)]
tF = vI + . (6.137)
ub + g ub − g
These results give us the strategy for completing the journey in the minimum
time. The plot of velocity against distance is given by:
> v1:=(uB,xI,vI,g,t)->vI-t*(uB+g);
v1 := (uB , xI , vI , g, t) → vI − t (uB + g)
> x1:=(uB,xI,vI,g,t)->xI+vI*t-t^2*(uB+g)/2;
1 2
x1 := (uB, xI , vI , g, t) → xI + vI t − t (uB + g)
2
> tF:=(uB,xI,vI,g)->(vI+sqrt(2*uB*(vI^2+2*xI*(uB+g))/(uB-g)))/(uB+g);
r
uB (vI 2 + 2 xI (uB + g))
vI + 2
uB − g
tF := (uB , xI , vI , g) →
uB + g
> v2:=(uB,xI,vI,g,t)->(uB-g)*(t-tF(uB,xI,vI,g));
> x2:=(uB,xI,vI,g,t)->(uB-g)*(t-tF(uB,xI,vI,g))^2/2;
1
x2 := (uB, xI , vI , g, t) → (uB − g) (t − tF(uB, xI , vI , g))2
2
> tS:=(uB,xI,vI,g)->(tF(uB,xI,vI,g)*(uB-g)+vI)/(2*uB);
134 CHAPTER 6. OPTIMAL CONTROL
1 tF(uB , xI , vI , g) (uB − g) + vI
tS := (uB, xI , vI , g) →
2 uB
> plot(
> {[x1(2,10,5,1,t),v1(2,10,5,1,t),t=0..tS(2,10,5,1)],
> [x2(2,10,5,1,t),v2(2,10,5,1,t),t=tS(2,10,5,1)..tF(2,10,5,1)]
> });
0 2 4 6 8 10 12 14
–2
–4
Problems 6
1) Find the extremum of
Z 1 n o
1
I[x] = 2
[ẋ(τ )]2 + x(τ )ẋ(τ ) dτ
0
ẋ(t) = u(t).
ẍ(t) = u(t).
With x(0) = x(tF ) = x0 , where tF > 0, find the input u(t) which minimizes
Z tF
1
I[u] = [u(t)]2 dt.
2 0
Find the minimized value of I[u] and show that it is less than the value
of I obtained by putting x(t) = x0 over the whole range of t. Given that
the condition x(tF ) = x0 is dropped and x(tF ) is unrestricted show that the
minimum value of I is zero.
6) A system is governed by the equation
ẋ(t) = u(t) − 1.
when
136 CHAPTER 6. OPTIMAL CONTROL
(a) x(tF ) = 1.
(b) x(tF ) is unrestricted.
Show that for case (a)
cosh(t) sinh(t)
u(t) = 1 + , x(t) = ,
sinh(tF ) sinh(tF )
cosh(t) sinh(t)
u(t) = 1 − , x(t) = − .
cosh(tF ) cosh(tF )
Without evaluating the minimized I show from these results that it is smaller
in case (b) than in case (a). Think about why this is what you should expect.
7) The equation of motion of a flywheel with friction is
where C and B are constants. Deduce that there is at most one switch
between ub and −ub .
Show from the transversality condition that there is no switch if θ̇ is unre-
stricted at t = tF and that in this case tF is give by the implicit equation
8) A rocket is ascending vertically above the earth. Its equations of motion are
C u(t)
ẍ(t) = − g, ṁ(t) = −u(t).
m(t)
where the switching function S(t) satisfies the equation Ṡ(t) = D/m(t), for
some positive constant D. Given that the switch occurs at ts , show that
m(t)
D
−
uu ln , 0 ≤ t ≤ ts ,
m(ts )
S(t) =
D(ts − t)
− , ts ≤ t ≤ t F .
m(ts )
9) A vehicle moves along a straight road, its distance from the starting point
at time t being denoted by x(t). The motion of the vehicle is governed by
where u(t), the thrust per unit mass, is the control variable. At time t = 0,
x = ẋ = 0 and the vehicle is required to reach x = L > 0, with ẋ = 0,
in minimum time, subject to the condition that |u(t)| < ub , where ub > k.
Using the state-space variables x1 = x and x2 = ẋ, construct the Hamiltonian
for the Hamiltonian-Pontryagin method and show that during the motion
either u(t) = ub or u(t) = −ub .
Show that u(t) cannot switch values more than once and that a switch occurs
when
L(ub + k)
x= .
2ub
7.1 Introduction
At some stage in many problems in linear control theory we need to analyze a
relationship of the form
where ū(s) and ȳ(s) are respectively the Laplace transforms of the input u(t)
and output y(t). The transfer function G(s) is a rational function of s. That is
ψ(s)
G(s) = , (7.2)
φ(s)
where ψ(s) and φ(s) are polynomials in s. Our interest has been, not only in
obtaining y(t) for a given form for u(t) for which we have usually used partial
fraction methods, but in determining the stability of the system, for which we
have the Routh-Hurwitz criterion described in Sect. 5.2.1. In this chapter we
shall describe other methods which rely on using the properties of (7.1) in the
complex s-plane.
139
140 CHAPTER 7. COMPLEX VARIABLE METHODS
exp(s t)
F (s) = . (7.4)
(s − 5)3
1 3
+ 3! t (s − 5)3 exp(5 t) + O[(s − 5)4 ]. (7.5)
1 dj−1
(s − s0 )j F (s) ,
Res(F ; s0 ) = lim (7.7)
(j − 1)! s→s0 dsj−1
for any j ≥ m.
With j = 3 you will see that this gives you a quick way to obtain (7.6).2 A
closed contour is a closed curve in the complex plane with a direction. Given
any closed contour γ and some point s0 not lying on γ the winding number or
index of γ with respect to s0 , denoted by Ind(γ; s0 ) is the number of times the
contour passes around s0 in the anticlockwise direction minus the number it
passes around s0 in the clockwise direction.
Theorem 7.2.2 Let γ be a closed contour and s0 a point not lying on γ then
1 ds
Z
Ind(γ; s0 ) = . (7.8)
2iπ γ s − s0
1 Proofs for this theorem and the Cauchy residue theorem are given in any book on Complex
1 F 0 (s)
Z
ds = {Number of zeros in γ} − {Number of poles in γ}, (7.10)
2iπ γ F (s)
F 0 (s) dF (s)
ds = = d {ln[F (s)]} . (7.11)
F (s) F (s)
Now let
1 F 0 (s) 1
Z Z
ds = d {ln[F (s)]}
2iπ γ F (s) 2iπ γ
1 1
Z Z
= d {ln |F (s)|} + dΘ
2iπ γ 2π γ
1
Z
= dΘ. (7.13)
2π γ
This final term measures the change in argument (hence the name ‘argument
principle’ ) of F (s) along γ in units of 2π. Suppose now we consider the mapping
s =⇒ F (s). (7.14)
As s describes the curve γ, F (s) will describe some other closed curve ΓF and
the last term in (7.13) is just the number of times that ΓF passes around the
origin, or simply the winding number Ind(ΓF ; 0). Thus
α+i∞
1
Z
y(t) = G(s)ū(s) exp(st)ds, (7.16)
2πi α−i∞
where α > η and the integral is along the vertical line <{s} = α in the complex
s–plane. According to Sect. 2.3 the parameter η is such that the integral defining
the Laplace transform converges when <{s} > η. This might seem to be a
problem for solving the integral (7.16). However, we do have some information.
We know that G(s) is a meromorphic function, that is its only singularities are
poles. Suppose that these are located at the points s1 , s2 , . . . , sr in the complex
s–plane. The pole sj will contribute a factor exp(sj t) to the Laplace transform.
Thus for convergence we must have <{s} > <{sj } and so α > η > <{sj }. This
applies to all the poles of G(s). If we also assume that ū(s) is also a meromorphic
function these must also be included and we final have the conclusion that the
vertical line of integration in (7.16) must be to the right of all the poles of
G(s)ū(s).
The problem now is to evaluate the integral (7.16). We have assumed the
ū(s) is meromorphic, so the integrand is meromophic with poles denoted by
s1 , s2 , . . . , sn .5 Now define
1
Z
yR (t) = G(s)ū(s) exp(st)ds, (7.17)
2πi γR
where γR is
s2 R
•
s1
•
•s
3
−R
We take R to be sufficiently large so that all the poles of the integrand are
within the contour. Then since the winding number of γR with respect to each
5 These are all the poles of G(s) and all the poles of ū(s) unless one of these functions has
n
X
= Res(G(s)ū(s) exp(st); sj ). (7.18)
j=1
The final part of this argument, on which we shall not spend any time, is to show
that, subject to certain boundedness conditions on the integrand, in the limit
R → ∞ the contributions to the contour integral from the horizontal sections
and from the semicircle become negligible. Thus
α+i∞ n
1
Z X
y(t) = G(s)ū(s) exp(st)ds = Res(G(s)ū(s) exp(st); sj ). (7.19)
2πi α−i∞ j=1
The first term has a simple pole at the origin and the second has two simple
poles at
iaω02
1
1
= − 2 u0 1∓ exp ±iωt − 2
aω02 t . (7.22)
2ω
where
G(s)
GCL (s) = . (7.25)
1 + G(s)
We consider the mapping
s =⇒ G(s). (7.26)
Assuming that ψ(s) and φ(s) in (7.2) are both polynomials with real coefficients
with ψ(s) a lower degree that φ(s) then ΓG is a closed curve6 in the Z–plane
with G(i∞) = G(−i∞) = 0 and symmetry about the X–axis. This curve is
called the Nyquist locus7 of G(s). We now prove the following theorem:
Theorem 7.4.1 If the number of poles of G(s) with <{s} > 0 is equal to
Ind(ΓG ; −1), the index of the Nyquist locus of G(s) with respect to Z = −1,
then the closed-loop transfer function GCL (s) is asymptotically stable.
6 Although, in Example 7.4.2, we see a case where this ‘closed’ curve has a discontinuous
Proof: Let
1
F (s) = 1 + G(s) = . (7.28)
1 − GCL (s)
The poles of GCL (s) will be the zeros of F (s).
Let γR be the the closed contour (traversed in the clockwise direction) con-
sisting of the imaginary axis in the s-plane from s = −iR to s = iR, together
with a semicircle, centre the origin, of radius R to the right of the imaginary
axis. We assume for simplicity that G(s) has no poles on the imaginary axis.8
Then for sufficiently large R all the poles and zeros of G(s) and F (s) with
<{s} > 0 will be enclosed within γR . This closed curve in the s–plane is called
the Nyquist contour. Now plot F (iω) = U (ω) + iV (ω) in the W = U + iV plane.
The contour ΓF , produced by this is, apart from an infinitesimal arc near the
origin the image of γR in the s–plane. (With very large R the arc of radius R is
contracted into a very small arc around the origin in the Z–plane.) Thus, from
(7.15),9
Ind(ΓF ; 0) = {Number of poles of F (s) with <{s} > 0}
In this case G(s) has no poles so GCL (s) will be asymptotically stable if the
Nyquist plot does not pass around Z = −1. We first use MAPLE to find the
real and imaginary parts of G(iω).
> G:=(s,K,a,b)->K*(a+b*s)/(s*(1+2*s)^2):
> X:=(w,K,a,b)->simplify(evalc(Re(G(I*w,K,a,b)))):
> X(w,K,a,b);
K (−b + 4 b w 2 + 4 a)
−
1 + 8 w2 + 16 w4
> Y:=(w,K,a,b)->simplify(evalc(Im(G(I*w,K,a,b)))):
> Y(w,K,a,b);
7.4. THE STABILITY OF A SYSTEM WITH UNIT FEEDBACK 147
K (−a + 4 a w 2 − 4 b w2 )
w (1 + 8 w2 + 16 w4 )
We see that Y (−ω) = −Y (ω). The Nyquist plot is symmetric about the X–axis
with both ends of the curve at the origin (X(±∞) = Y (±∞) = 0). Unless
α = 0, Y (ω) → ∓0 × (Kα), as ω → ±0, giving an infinite discontinuity in
the plot as ω passes through zero. If α = 0 then Y (ω) → 0 as, ω → 0, with
X(ω) → Kβ.
In the case α 6= 0 the plot cuts the X–axis when Y (ω) = 0 giving
α
r
ω=± . (7.39)
4(α − β)
If α > 0 and α > β or α < 0 and β > α the two branches of the plot cross
at this value of ω. We calculate the point on √
the X–axis where this occurs for
β = α/2 > 0. In this case (7.39) gives ω = 1/ 2.
> X1:=(w,K,a)->simplify(X(w,K,a,a/2)):
> X1(w,K,a);
1 K a (7 + 4 w 2 )
−
2 1 + 8 w2 + 16 w4
> Y1:=(w,K,a)->simplify(Y(w,K,a,a/2)):
> Y1(w,K,a);
K a (−1 + 2 w 2 )
w (1 + 8 w2 + 16 w4 )
> simplify(X1(1/sqrt(2),K,a));
1
− Ka
2
So, for β = α/2 > 0, the closed-loop transfer function is stable if − 21 Kα > −1.
That is if K < 2/α, which is the result obtained by the Routh-Hurwitz method.
We compute the Nyquist plot for an unstable case when K = 1, α = 6, β = 3.
> with(plots):
> plot([X(w,1,6,3),Y(w,1,6,3),
> w=-infinity..infinity],X=-6..2,Y=-2..2,numpoints=1000);
148 CHAPTER 7. COMPLEX VARIABLE METHODS
1Y
–6 –5 –4 –3 –2 –1 0 1 2
X
–1
–2
When α = 0, the system will be stable if the point where the plot cuts the
X–axis is to the right of the origin. That is Kβ > 0. We plot the case K = 21 ,
β = 1.
> plot([X(w,0.5,0,1),Y(w,0.5,0,1),
> w=-infinity..infinity],X=-0.2..0.6,Y=-0.6..0.6,numpoints=1000);
0.6
0.4
Y
0.2
–0.4
–0.6
7.4. THE STABILITY OF A SYSTEM WITH UNIT FEEDBACK 149
Problems 7
1) For the system with block diagram:
ū(s) ȳ(s)
+ G(s)
−
K
G(s) = .
(1 + s)3
Determine the closed loop transfer function GCL(s) and use the Routh-
Hurwitz stability criteria to show that the system is stable for −1 < K < 8.
Find the functions X(ω) and Y (ω) so that G(iω) = X(ω) + iY (ω). Define
the Nyquist plot and state the Nyquist criterion which relates the form of
this curve to the stability of GCL (s). Show that, for the given example, the
result obtained from the Nyquist criterion confirms the result obtained by
the Routh-Hurwitz procedure.
Chapter 8
Non-Linear Systems
8.1 Introduction
In Sects. 1.5 and 1.6 we discussed systems of differential equations most of which
were non-linear. As we have seen, it is no restriction to concentrate on a first-
order system since higher-order equations governing a system can be expressed
as a system of first-order equations by introducing additional variables. For
simplicity we shall again consider single input/single output systems and we
shall also suppose the system is autonomous. A realization will then be of the
form
where x(t) is the n-dimensional state-space vector, just as in the linear version
(4.42)–(4.43) of these equations.
w̄(s)
ȳ(s) = ,
Qs2
w̄(s) = F{v̄(s)},
(8.3)
v̄(s) = ū(s) − q̄(s),
q̄(s) = (1 + H s)ȳ(s).
151
152 CHAPTER 8. NON-LINEAR SYSTEMS
q̄(s) ȳ(s)
1+Hs
where F is some non-linear operator. Assuming all the variables and necessary
derivatives are zero for t < 0 we eliminate all the intermediate variables to give
Q ÿ(t) = L−1 F L{u(t) − H ẏ(t) − y(t)} .
(8.4)
For a linear system the operator F would simply apply a multiplicative rational
function of s to the Laplace transform − H ẏ(t) − y(t) and the final effect
of u(t)
of the sequence of operators L−1 F L{·} would be to produce some linear
combination of u(t) and y(t) and their derivatives. For a non-linear system we
define the non-linear function
1
f (·) = L−1 F L{·} . (8.5)
Q
Now introduce the two state-space variables
x1 (t) = y(t) + H ẏ(t),
(8.6)
x2 (t) = −ẏ(t)
and we have the realization
ẋ1 (t) = H f (u(t) − x1 (t)) − x2 (t),
usually used the variable a. We examined the equilibrium points of the system to
determine their stability and showed that this can alter as u0 changes. This leads
to bifurcations at particular values of u0 , where the stability of an equilibrium
point changes and/or new equilibrium points appear. The simplest case we
considered was Example 1.6.1, which in our present notation is
It thus has all the properties of an Abelian (commutative) group apart from the
possible non-existence of an inverse; it is therefore an Abelian semigroup.
The important question concerning a solution x(t) of (??) is whether it is
stable. There are many different definitions of stability in the literature. As we
did for equilibrium points in Sect. 1.6 we shall use the one due to Lyapunov:
Lyapunov stability could be characterized by saying that, for stability, the two
solutions are forced to lie in a ‘tube’ of thickness ε, for t > tI , by the initial
condition (??). The following definitions are also useful:
A cluster (or limit) point x∞ of the solution x(t) to (??), with x(tI ) = xI ,
is such that, for all τ > 0 and ε > 0, there exists a t1 (ε) > τ with
|x∞ − x(t1 )| < ε. (8.18)
The set of cluster points is called the ω-limit set of the trajectory.
Given that the solution x(t) to (??) is defined for all (positive and negative)
t and x(0) = xI the reverse trajectory xR (t) is defined by xR (t) = x(−t).
The set of cluster points of the reverse trajectory is called the α-limit set of
the trajectory x(t).
x(tm ) → x∞ , as m → ∞. (8.19)
8.3. CONSTANT CONTROL-VARIABLE SYSTEMS 155
Let A be the ω-limit set of a particular solution x(t) to (??). If there exists
a region D(A), in Γn , which contains A and for which the trajectories with
x(0) = xI , for all xI in D(A), have A as their ω-limit set, then A is called an
attractor with basin (or domain) D(A). An α-limit with the same property
for reverse trajectories is called a repellor.
Theorem 8.3.1 Let x∗ (u0 ) be an equilibrium point of (??). Suppose that there
exists a continuous differentiable function L(x) such that
L(x∗ ) = 0 (8.20)
and for some µ > 0
L(x) > 0, when 0 < |x∗ − x| < µ. (8.21)
Then x∗ is
(i) stable if
(iii) unstable if
Try
L(x, y) = αx2 + βy 2 . (8.31)
For α and β positive (??) and (??) are satisfied and
X(x, y).∇L(x, y) = −{2αx(2x + y 2 ) + 2βy(y + x2 )}
With
∇V
∇L(x) = (8.36)
v
Since, from (??) L(x∗ , 0) = 0 it follows from (??) that the equilibrium point is
stable (but not asymptotically stable) if (??) holds. From (??) this will certainly
be the case if x∗ is a local minimum of V (u∗0 , x). It can be shown that such a
minimum of the potential is a centre, which is stable in the sense of Lyapunov.
So the origin is a saddle point in this plane when |u0 | > 1. However, the function
L(x1 , x2 , p1 , p2 ) = H(0, x1 , x2 , p1 , p2 ) (8.44)
has a minimum at the origin with
X(u0 , x1 , x2 , p1 , p2 ).∇L(x1 , x2 , p1 , p2 ) = 0. (8.45)
So we have found a Lyapunov function which establishes the stability of the
equilibrium point.
Theorem 8.4.2 Let C be a closed, bounded (i.e. compact) subset of the x 1 –x2
plane. If there exists a solution γ = (x1 (t), x2 (t)) of (??), which is contained
in C for all t ≥ 0, then it tends either to an equilibrium point or to a periodic
solution as t → ∞.
Proof: Consider the infinite sequence (x1 (t0 + nε), x2 (t0 + nε)) of points of γ,
with t0 > 0, ε > 0, n = 0, 1, 2, . . .. All these points lie in the compact set C so
it follows from the Bolzano-Weierstrass theorem that the sequence has at least
one limit point. This point must belong to the ω-limit set of γ, which is thus
non-empty. From Thm. ?? this ω-limit set is an equilibrium point or a periodic
solution to which γ tends.
Problems 8
1) Systems are given by
(i) ẋ(t) = −x − 2y 2 , ẏ(t) = xy − y 3 ,
(ii) ẋ(t) = y − x3 , ẏ(t) = −x3 .
Using a trial form of L(x, y) = xn + αy m for the Lyapunov function show
that, in each case the equilibrium point x = y = 0 is asymptotically stable.
2) A system is given by
ẋ(t) = x2 y − xy 2 + x3 , ẏ(t) = y 3 − x3
Show that x = y = 0 is the only equilibrium point and, using a trial form of
L(x, y) = x2 +αxy +βy 2 for the Lyapunov function, show that it is unstable.
160 CHAPTER 8. NON-LINEAR SYSTEMS
Let C be a closed bounded subset of the {x, y} plane. Show that if there
exists a solution γ = (x(t), y(t)) to these equations which is contained in C
for all t ≤ 0 then C contains either an equilibrium point or a periodic solution
of the system. For the particular case
show that the origin is the only equilibrium point and determine its type.
Express the equations in polar form and, by considering the directions in
which trajectories cross suitable closed curves, show that the system has at
least one periodic solution. As an optional extra solve the equations and
determine the equation of the periodic solution. Try plotting it in MAPLE .