Controltheory0000gree 1
Controltheory0000gree 1
modern
control
theory
*• r-'A
Digitized by the Internet Archive
in 2019 with funding from
Kahle/Austin Foundation
https://archive.org/details/controltheoryOOOOgree
Elements of Modern
Control Theory
dedicated
to the memory of
Max Greensite
and
Sholom Smilowitz
CONTROL THEORY: VOLUME I
Elements of Modern
/ Control Theory
Arthur L. Greensite
SPARTAN BOOKS
^ i-f
. u /
PREFACE
*Greensite, A., Analysis and Design of Space Vehicle Flight Control Systems, New York, Spar¬
tan Books, 1970.
v
PREFACE
VI
Arthur L. Greensite
San Diego, California
CONTENTS
IX
X CONTENTS
Index . 883
Chapter One
THEORY OF LINEAR SYSTEMS
The basic theory of linear systems goes back to Newton and includes
many refinements and extensions (Laplace transforms, linear operators)
developed over the years. In the period following World War II, the
rapid development of the theory of automatic control (feedback principle)
led to the need for special methods specifically suited to linear feedback
control systems. There emerged the frequency-response techniques
(Nyquist, Bode) and root locus (Evans). These permitted a rapid and
efficient analysis of what might be called “conventional” control systems.
Multiple loop, high order systems presented extreme difficulties for
“paper-and-pencil” analysis. The advent of the modern high-speed digital
computer has, for the most part, eliminated this problem, and the design
of linear automatic control systems has become virtually routine. Con¬
sequently, research studies in recent years have been in the areas of
nonlinear, adaptive, and optimal control. However, the well has not
run dry in linear theory. New ideas do appear.
A case in point is the theory of observability and controllability. One
important result of this theory is a new insight into the relationship be¬
tween transfer functions and the state variable representation. This has
clarified the phenomenon of “hidden oscillations” as well as the condi¬
tions under which a conventional transfer function is an accurate repre¬
sentation of a given dynamic system. The theory is perhaps most powerful
when applied to multivariable, i.e., multiple input/multiple output, sys¬
tems. There are many instances in the literature where the neglect of this
idea has led to erroneous results.
Another recent development of some importance is the Principle of
Invariance, whose basic ideas have been extensively advanced by the
Russians. Here, the fundamental idea, while simple and elegant, en¬
counters severe difficulties in practical application. However, the basic
approach is novel and will no doubt prove useful in the design of control
systems as current obstacles are progressively resolved.
1
2 ELEMENTS OF MODERN CONTROL THEORY
(1)
dt*
X ak\n~k= 0. (3)
k=0
where x2(t) is a particular solution of Eq. (1). These results are classical,
and details may be found in any standard text on differential equations.
Although, in principle at least, linear systems may be solved in a
straightforward manner, a multitude of highly specialized techniques
have been developed to yield results more efficiently and for specific
needs. Basically, these methods bypass the problem of solving Eq. (1)
repetitively for each new set of system parameters, yielding instead such
4 ELEMENTS OF MODERN CONTROL THEORY
have negative real parts if, and only if, all the determinants
ai a3 a6 ....
ao a2 3<4 .... &2k—2
0 ax a3 • • • • . &Vc~Z
ao 32
0 ai
ao
0
0 0 0 . a,
s3 + ais2 + a2s + a3 = 0
where
ai = K„
a2 = KaK0Krh„ ~ Ma
a3 = Kc(K4p0 — na).
= 2.5 Ac = 2
K c = 5 Aa ~ 1
Ivr = 1
we find
SLi = 5
a2 = 24
a3 = 20.
Therefore
Hi ax = 5
ai a3
H2
a0 a2
ai a3 0
ao a2 0 2000.
0 ai a3
Since all the H* are positive, the characteristic equation has no roots
with positive real parts, and the system is stable.
Theorem A may be reformulated in a manner that does not require
the evaluation of determinants. The essential result is contained in the
following.
*It is immaterial whether we use s (the Laplace operator) or X, since for present purposes this is
merely a dummy variable.
6 ELEMENTS OF MODERN CONTROL THEORY
ao a2 a4 a6
ai a3 a5 a7
bi b2 b3
Cl c2 C3
di d2
ei e2
fi
where
aia2 —
bia3 —
cib2 —
dic2 —
cid2 dic3 — Cid3
ex = e2
di di
eid2 —
die2
fi =
ei
etc.
Notice that two terms in the first column are used in each calculation.
As the term to be calculated shifts to the right, the additional two terms
in the formula also shift to the right. The formula for calculation of
terms in any given row uses only those terms in the two rows immediately
above. The process is contained for (n + 1) rows.
The number of changes in sign of the terms in the first column of the
above array is equal to the number of roots with positive real parts.
THEORY OF LINEAR SYSTEMS 7
Furthermore, if the a* are not all of the same sign, or if any a* is zero,
then some roots are either pure imaginaries or else have positive real
parts.
It may happen that the first column term in any row is zero but the
remaining terms in this row are not all zero. In this case, replace the
zero term by an arbitrarily small constant, e, and proceed as usual.
If all the coefficients of any row are zero, this indicates a pair of
complex roots with zero a real part.
The above results are classical and constitute the earliest attempts to
study the stability of linear systems in some rational manner. In recent
years, these criteria have been generalized in various ways. Perhaps the
most significant is the extension of the method to determine “relative
stability.” In this case, one derives, on the coefficients, a*, conditions
that ensure that all the roots of the characteristic equation lie to the
left of the shaded lines shown in Fig. 1.1.
Now any complex root pair may be expressed as
— =b j \/l — fB2]
from the origin into the left-half plane and making equal angles with
both halves of the imaginary axis. According to Fig. 1.1, the relative
damping factor defined by the shaded lines is given in terms of 6 by
X = ze ie (9)
whose roots are identical to those of Eq. (5), except that they are rotated
counterclockwise by d degrees. Consequently, any root of Eq. (5) that
is located in the sector between the shaded line and the negative imagi¬
nary axis will appear as a root in the right-half plane of Eq. (10). The
difficulty now is that the coefficients ofEq. (10) are (in general) complex.
What is needed, therefore, is a “Routh-Hurwitz” type criterion that is
valid for polynomials with complex coefficients. Various criteria of this
type have indeed been developed, and the simplest, perhaps, is the
Bilharz-Frank theorem given by Marden(9) in the following form.
Ao — 1
Bo = 0.
*There is obviously no loss of generality in assuming that the leading coefficient is unity.
THEORY OF LINEAR SYSTEMS 9
0 B* 0 Afc_i
where k = 1,2,3, • • •, n.
The number of roots of Eq. (11) having positive real parts is equal to
the number of changes of sign in the sequence, Ai, A2, • • •, A„.
Any term A^or B, whose subscript i > n is set equal to zero.
When the B* are all zero, i.e., Fx(z) is a real polynomial, then Ax = Hx
and A*, = H*, Ht_i where H* is the determinant defined by Eq. (6),
with a0 = 1. Then, since sgn (AXA2)= sgn A2, and sgn (A* Ak+1) = sgn
(Hh Ht+]) for k = 2,3, •••, n — 1, the theorem reduces to the con¬
ventional Routh-Hurwitz form.
The necessity of evaluating high order determinants is often awkward,
even though there are a large number of zero elements. An alternate
approach that permits the use of the simplified criteria of Theorem B is
the following/10) Substitute
X = ze* (12)
The roots of this polynomial are identical to those of Eq. (5), except
that they are rotated clockwise by B degrees. Forming the product
10 ELEMENTS OF MODERN CONTROL THEORY
results in a new polynomial having the property that any complex root
pair of Eq. (5) lying in the sector between the shaded lines and the
imaginary axis now appears as a complex root pair in the right-half
plane of Eq. (14). The coefficients of Eq. (14) are all real. In fact, for
any specific pair of indices, a = p and /3 = q, we have a term of the
form
apaqeiip~r)9z^2n~p~r)
.
2a/>a9z(2n—?—r)cos (p — r)6.
2n
b0 = a02
bi = 2a0ai cos 8
b2 = ai2 -f- 2a0a2 cos 28
b3 = 2aia2 cos 6 + 2a0a3 cos 3 8
b 4 = a22 + 2aia3 cos ‘28 + 2aoa4 cos 4 8
b5 = 2a2a3 cos 8 -f- 2aia4 cos 38 -(- 2aoa3 cos 58
b6 = a32 + 2a2a4 cos 28 + 2axa6 cos 4 8 + 2a0ae cos 68
b7 = 2a3a4 cos 8 2a2a6 cos 38 + 2axa6 cos 58 + 2a0a7 cos 78
b8 = a42 + 2a3a6 cos 28 + 2a2a6 cos -id + 2axa7 cos 68 -f 2a0a8 cos 88
b2 2a4a6 cos 8 4- 2a3a6 cos 36 -)- 2a2a7 cos 58 T 2axas cos 78
+ 2a0a9 cos 9^
bio = a6 2 + 2a4a6 cos 28 + 2a3a7 cos 4 8 + 2a2a8 cos 68 + 2aiag cos 88
+ 2a0ai0 cos 100
THEORY OF LINEAR SYSTEMS 11
(b)
0
where
A0 = 1 Bo = 0
Ai = 4.33 Bj = 2.5
A2 = 12 B2 = 20.78
A3 = 0 B3 = 20.
4.33 0 0 20.78 0
1 12 0 -2.5 -20
A3 — 0 4.33 0 0 -20.78
0 20.78 0 4.33 0
0 2.5 20 1 12
=
-24,994
4.33 0 -20.78
A2 = 1 12 -2.5
0 20.78 4.33
= 18.02
Ai — Ai — 4.33.
Since there is one change of sign in the sequence Ai, A2, A3,Theorem C
indicates that Eq. (b) has one root in the right-half plane, which, in turn,
means that Eq. (a) has a complex root pair whose relative damping
factor is less than 0.5.
To obtain the same result via another route, we calculate the bt of
Eq. ( 15), using the ak coefficients of Eq. (a). The result is
b6 = 400 b3 = 208
b5 = 831 b2 = 49
b4 = 676 bi = 8.66
bo = 1.
1 49 676
8.66 208 831
25 580 400
7 692
-1891 400
692
400.
There are two changes of sign in the first column, which indicates
that the equation
6
fc(z) = £ b*z6-* = 0
k=0
THEORY OF LINEAR SYSTEMS 13
has two roots in the right-half plane; therefore Eq. (a) has a complex
root pair with a relative damping factor less than 0.5.
C(s) _ KG(s)
(16)
R(s) 1 + KG(s)H(s)
= KG(s)H(s) (17)
E(s)
B(s)
H(s)
B(s)
K.h(^+«n(i>^+i)
i=l j=l\COj OJj /
(18)
E(s) 8 T / S2 2tl \
B-ner**
4=1
+ dii( —
2+—
fs+ ]) " '
m P
B(s) = KnAi(s)
E(s) s"A2(s)' 1
(23)
N = Z - P (24)
*This result, which follows from a simple application of Cauchy’s Residue theorem, is proved in
any standard text on complex variables.
16 ELEMENTS OF MODERN CONTROL THEORY
jOJ
Im
the finite poles and zeros of Y(s). The small semicircle about the origin
is drawn so that a pole of Y(s) at the origin is not on the contour.
Similarly, arbitrarily small semicircles are drawn on the imaginary axis
to avoid purely imaginary poles and zeros of Y(s).
We now investigate the form of r2 when a point on I\ approaches
zero on the positive jo> axis, i.e., with p —> 0. If Y(s) has an nth order
pole at the origin, then it may be written as
Y(s) (26)
THEORY OF LINEAR SYSTEMS 17
si = pe,e.
— e-»'»9 = Y(si)
Pnbo
jw
It follows that contour Tiof Fig. 1.4, described in a positive sense, will
map into a contour, r, in the L plane, that encircles the point ( —1/K„, jO)
in a positive sense N = Z — P times.
This is the Nyquist Stability Criterion. Note that if L(s) is expressed as
Ai(s)
L(s)
s"A2(s)
skA2(s) + K^A^s)
Y(s) (28)
K„snA2(s)
It is obvious that the roots of the numerator (zeros) of Eq. (28) are
also the roots of the characteristic Eq. (22).
Since the poles of L(s) are also the poles of Y(s), it follows that Eq.
(22) has
Z = N + P (29)
roots in the right-half plane, where P is the number of poles of L(s) in the
right-half plane and N is the number of positive encirclements of the
( —1/Kn, jO) point in the L plane.
L(s) = K”Al(s)
s"A2(s)
and stability is described in terms of the (—1, jO) point. However, it
is simpler to move the point 1/K„ than to redraw the KnAi(s)/s"A2(s)
locus for every new value of K„. Note also that if K„ is a negative
quantity, all the previous results hold except that the critical point
is (1/K„, jO) instead of ( —1/K„, jO).
The value of the exponent, n, in the above expression indicates the
so-called “system type.” (See Sect. 1.3.1.) It provides a measure of
the steady-state error in response to particular input signals (step,
ramp, etc.).
Table 1.1 shows frequency-response and root locus plots for some
typical open-loop transfer functions. For the cases shown:
THEORY OF LINEAR SYSTEMS 19
Case 1
Nyquist Plot
Case 2
K (r s+l)(rs+l)
KG(s)H(s) = - --
' sC1s+l)C2s+l)(r3s+l)Cr4s+l)
Case 3
K (T s+1)
KG(s) H(s) = -
s Cr1s+1)(r2s+1)
Note that when there is an nth order pole at the origin, the point o> = 0+
is connected to w = 0_ by n large counterclockwise semicircles. Further¬
more, an encirclement of the —1 point is positive if, in tracing the
Nyquist diagram as co varies from + °o to — <*>,the net encirclement is in a
counterclockwise direction. Otherwise, the encirclement is negative.
This follows from the discussions related to Figs. 1.3 and 1.4.
Two important figures of merit that may be obtained from the Nyquist
plot are the phase margin and gain margin. These are defined as follows.
Gain margin is the factor by which the gain, K„, must be multiplied to
make the locus pass through the ( —1, jO) point.
Phase margin is the amount of phase shift needed at unity gain to make
the locus pass through the (—1, jO) point.
These concepts may be clarified by considering the Nyquist plot of
R-tT'TaS “f- 1)
KG(s)H(s) =
*Gain margin is generally expressed in decibels. The decibel equivalent, N<rt, of a number, N, is
N<26 = 20 logio N.
THEORY OF LINEAR SYSTEMS 23
2
Wl
KG(s) (30)
s(s *F 2fiwi)
and
H(s) = 1.
Then obviously
KG(ju) = - — - = M/V
—co2+ 2jfiwiw
where M is the magnitude and <p is the phase angle of the complex
number KG(jco). It follows immediately that
if = TT + 6
where
2f ion
tan 6 =
co
tan d0
Wo
2fi
7 = tan-
-(V 1 + 4f? - 2f12)1/2J
This relation is very nearly linear for 0 < 7 < 50 0 and may be written as
7^110fi. (31)
For second order systems, the phase margin is therefore directly re¬
lated to the relative damping factor. The response of most systems of
engineering interest is, in fact, governed by a dominant pair of complex
poles. Consequently, in such cases, the phase margin is a measure of how
oscillatory the system is. (See also Sect. 1.3.1.)
It is often convenient to make use of the Simplified Nyquist Criterion,
which may be stated as follows.
If the open-loop transfer function, KG(s) H(s), contains no poles in
the right-hand s plane, and if the Nyquist locus does not encircle the
( —1, jO) point, the system is stable.
It is also possible to derive a kind of generalized Nyquist criterion as
follows. Instead of the s plane contour, Ti, of Fig. 1.4, consider the
contour, Ti, shown in Fig. 1.6. As before, we apply this to the transfer
function, Y(s), defined by Eq. (27).
In the conventional case, the Nyquist locus is obtained by replacing
s in L(s) by jw, and letting « vary from zero to infinity. To describe the
contour of Fig. 1.6, however, we must replace s by ( —f 0« + jco\Ai — f02),
where f0 is a prescribed constant. Now if the contour, TJ, is described in
a positive sense, then the corresponding contour, r', in the L plane will
encircle the point, ( —1/Ka, jO), N = Z - P times in a positive sense,
where Z and P are the Y(s) zeros and poles, respectively, within contour
T'i. Since the poles of L(s)and Y(s)are identical, this serves to determine
26 ELEMENTS OF MODERN CONTROL THEORY
jW
the number of closed loop transfer function poles that have either
positive real parts or a relative damping factor less than f 0-
This constitutes the Generalized Nyquist Criterion.
The computation of L(s0), where s0 = —f0co + jco\/l — for
0 < co< °° , may be simplified materially by using the relation
F(s) = E
k— 0
becomes
and(—
\w„2
+ ^
a>„
s + 1
Each of these terms, with s replaced by jco, results in a complex number,
M/i£, having a distinctive form when plotted for M in decibels and « on a
logarithmic scale. This is shown in Figs. 1.7-1. 9. A frequency-response
curve, when plotted in the coordinate scale shown, is called a Bode plot.
The fundamental advantage of this representation is that the general
form of the frequency response can be quickly visualized and displayed
with a minimum of effort, since addition of basic forms rather than
multiplication is required. This is, of course, due to the logarithmic,
rather than numeric, representation of magnitude.
Note: The Bode plot for (jw)n is obtained by taking the mirror image
of the above lines about w = 1. In this case, 0 =nir/2.
(deg)
0
Note: The Bode plot for (1 + j wt) is obtained by taking the mirror
image of the above curves about M = 0.
Consider again the feedback system shown in Fig. 1.2, for which the
closed-loop transfer function is given by Eq. (16) (repeated here for
convenience).
1 + KG(s)H(s) = 0 (35)
KG(s)H(s)| = 1 (36)
*A transfer function is minimum phase if it has no poles or zeros in the right-half s plane.
THEORY OF LINEAR SYSTEMS 31
(38)
(39)
where Gi(s), G2(s), Hi(s) and H2(s) are polynomials ins. For a physically
realizable system, G2(s) H2(s) is of equal or higher order in s than
Gi(s) Hi(s) .
Substituting Eqs. (38) and (39) in Eq. (34) results in
The number of separate root loci is equal to the order of the characteristic Eq. (35).
32 ELEMENTS OF MODERN CONTROL THEORY
C(s) = KQx(s)
R(s) Q2(s)
Kjl (s - z<)
n (s - P;)
7=1
Kli (s - Zi)
C(s) = — - . (42)
*11 (S - P;)
1=1
(43)
s ;=i * - Pi
Here*
Kli (-z i)
K„ = - 1=1 (44)
n (-Pi)
j=i
*In the terminology of complex variable theory, the Ko and Kt are the residues, at the
respective poles, of C(s).
THEORY OF LINEAR SYSTEMS 33
Kjj (s - z,•)
=
V
(45)
sir (s - py)
c(t) = K0 + H K i&vi
l. (46)
Suppose there exists a pair of dominant poles,* px and p2, such that
| p, | ^ | Pi | for i = 2, 3, • • •. In order to ascertain if the poles p;
(i = 2, 3, •••) have a negligible influence on the response, c(t), it is
necessary to examine first the magnitude of the term K, ep*‘ compared to
2 | Ki | e<r*1
cos (coit + ifi), and second, the influence of pion Ki and ipi.
For definiteness, let us consider the configuration shown in Fig. 1.1.
Equation (42), for this case, is written as
K
C(s) = (48)
s(s - Pi) (s - p2)(s - p3)
where
Pi = -ai + j/h
P2 - ai - j/3i
*A dominant pole in the s plane is defined as that which has the smallest absolute value and
therefore has the predominant influence on the time response.
34 ELEMENTS OF MODERN CONTROL THEORY
and
K = — pip2p3. (49)
Now
* A
C(s) = Z (51)
i= 1 s - p<
where
Ai = —/ -7, - 90 ° - 7,
abc "
A2 = ~/yx + 90 ° + 7,
abc
A3= —/180 °
c2d-
where
b = |pi — p2|
d = |p»|.
But
K = -pip2p3
= ~(ai2 + /?i2)p3
= a2d.
Therefore, A0 = 1. Furthermore
K _ ad
abc be
a
Ai
b
Also
A3 _K — —* 0 for c 2> a
c2d
and
K(s - zi)
C(s) (52)
s(s - px) (s - p2)
P1P2
K =
Zl
C(s) = E -Al_.
i—0 S-Pi
Im
Kd
Aq
a2
Ax —/y, - 90 ° - 7i
ab
A2 = — /- t2 + 90 ° + 7x.
ab
Also
K = PtP2 _
Zi d
and
Ao — 1
while
cK a c
Ai A2I
ab b d
38 ELEMENTS OF MODERN CONTROL THEORY
c (t)
Figure 1.13. Unit Step Response Showing Influence of Additional Pole or Zero.
THEORY OF LINEAR SYSTEMS 39
Poles
Complex
of
Pair
2:
Case
Real
1:
Two
Configuration
Pole-Zero
Selected
for
Response
Step
Unit
1.2.
Table
Q)
C and
Poles
Complex
of
Zero
One
Pair
4:
3:
Case
(Pole-Zero
Configuration
Selected
for
Response
Step
Unit
1.2.ont’d)
Table
Pole
Real
One
and
Poles
Complex
of
5:
Pair
6:
Case
Re
C with
Poles
Complex
of
8:
Case
Parr
Pair
7:
(Pole-Zero
Configuration
Selected
for
Response
Step
Unit
1.2.ont’d)
Table
KaKcKs^c
(s+K,)
(.+£)
I\G(s)H(s) = (53)
T
or
k + V<K*S
/
+ *>
KG(s)H(s) = (54)
Ka = 3 (nondimensional)
44 ELEMENTS OF MODERN CONTROL THEORY
CJ
0>
>
Q.
O
3
<
O
c
£
"O
s
a
§
£3
1
£
THEORY OF LINEAR SYSTEMS 45
Replacing s by jco in Eq. (35) and letting u>vary from zero to infinity,
we obtain the Nyquist plot of Fig. 1.15.
Since there is one open-loop pole in the right-half plane (P = 1), and
one clockwise encirclement of the (-1, jO) point as w varies from- « to
+ °°(N = —1), we find from Eq. (29) that Z = 0; hence the system is
stable.
An inspection of the diagram indicates^that the upper and lower gain
margins are Xi = 1/0.23 = 4.35 = 12.8 db and \2 = 1/6.32 = 0.158 =
- 16.2 db, respectively. Consequently, if KA is assumed to be the only
adjustable parameter, instability occurs for Kj > 3 X 4.35 = 13.1 or
KA < 3 X 0.158 = 0.474. The phase margin is found to be 26 degrees.
The Bode and Nichols plots for Eq. (35) are shown in Figs. 1.16 and
1.17, respectively. Also indicated are the gain and phase margins noted
above.
The Nichols plot is used extensively by many aerospace agencies to
represent complicated open-loop transfer functions. In order to discuss
the general features of such a diagram, we will include the effect of three
bending and two sloshing modes added to the present example. A
Nichols plot of the general form shown in Fig. 1.18 is then obtained.
The terminology in this figure is that in general use in the aerospace
industry for autopilot control of launch vehicles. The gain and phase
margins related to specific modes are immediately apparent. If additional
phase lag exceeding yi degrees is introduced at unity gain, the rigid body
mode will become unstable. Similarly, and 72 represent the factors
that, if exceeded either by an increase or decrease in gain, will make
the rigid body mode become unstable. Also, the addition of a phase
lead of 73 degrees (at unity gain) will cause the first bending mode to
become unstable.
Autopilot performance specifications are often expressed in terms of
gain and phase margins for specific modes. The ease with which these
quantities are determined on a Nichols plot makes this type of specifi¬
cation particularly attractive.
We may also analyze the system of Fig. 1.14 using root locus. The
poles and zeros of the open-loop transfer function are available from
inspection of Eq. (53), and the root locus is shown in Fig. 1.19.
Certain correlation with the results obtained via frequency-response
are immediately evident. Thus, it was found from the frequency-
response analysis that instability occurs for IC < 0.474 or IC > 13.1.
According to Fig. 1.19, there will be closed-loop poles in the right-half
plane if KA < 0.47 or KA > 13.2. That the agreement is quite good is not
surprising, since a lack of agreement would simply indicate an error
somewhere in the analysis.
46 ELEMENTS OF MODERN CONTROL THEORY
(ggp) 0)
Exampl
for
Plot
Bode
1.16.
Figure
(rad/
Frequen
sec)
48 ELEMENTS OF MODERN CONTROL THEORY
For = 3, the closed-loop poles and zeros, obtained from Fig. 1.19,
are shown in Fig. 1.20, where the pole at the origin represents a unit step
input. In calculating the response for this case, the dipole at (-0.2,
-0.36) and the pole at -29.9 are very nearly negligible. Therefore, from
Eqs. (46) and (47)
Pi — or T- j&>i— —2.42 + j 6. 1
P2 — — 4.93.
THEORY OF LINEAR SYSTEMS 49
Slosh
and
Include
Bendin
Modes
3with
Examp
for
Plot
Nichol
1.18.
Figure
<P
(deg)
ELEMENTS OF MODERN CONTROL THEORY
50
Im
THEORY OF LINEAR SYSTEMS 51
3.
Examp
for
Config
Pole-Z
1.20.
Figure
Loop
Closed
52 ELEMENTS OF MODERN CONTROL THEORY
(j 2 42
relative damping factor: ft = - = - 1- = 0.37
vV + C0X2 V 6.12 + 2.422
Kzx
K0 — - = 1
Pi P2
kJ — KA3
lm|
A0AxA2
K(p2 — zx)
k2 =
p2A22
Ao = |px = 6.56
Ax = 2wx = 12.2
The first of these expressions shows how the amplitude factor of the
oscillatory response is influenced by the presence of the additional pole
and zero.
The significant fact about K2 is that the exponentially decaying term
K2ep2‘ will be positive or negative, depending upon whether the zero or
the pole is nearer the imaginary axis. All the essential features of the
response are thereby delineated. y
We will conclude the discussions of this section by considering various
refinements in root locus technique developed in recent years. The usual
method of obtaining the root locus is by graphical trial and error, aided
by various systematized procedures. However, by analyzing the root
locus as a continuous function, certain features that are obscured by the
graphical approach become evident.
According to Eq. (35), the root locus is the function that satisfies
G(s)H(s)
= - I. (56)
Since K is always real, any value of s for which G(s) H(s) is real will
be a solution of Eq. (56). Thus, the root locus is defined by the con¬
dition*
Im [G(s)H(s)] = 0
or equivalently
Im [KG(s)H(s)] = 0. (57)
Equation (57) is valid for positive or negative Iv, which means that this
yields the root locus for - oo < K < <». We may add the relations
to distinguish the loci for positive or negative K, but this is not necessary
for present purposes.
We will, therefore, refer to Eq. (57) as the equation of the root locus,
regardless of the sign of K.
Usually, Y(s) may be written as a ratio of polynomials in s, viz.
P(s)
Y(s) = KG(s)H(s) = (58)
Q(s)'
which reduces to
This is the equation of the root locus. If P(s)= a real constant, then Eq.
(59) simplifies to
Im Q(s) = 0. (60)
Two features of the root locus may be derived immediately from Eq.
(59). First, replacing s by a + jo, setting « equal to zero, and then
solving the resulting equation for a yields the “breakaway points” for
the real axis. Also, if s is replaced by a + jo and <j is then set equal to
zero, the solution of the resulting equation for o will yield the points
of intersection with the imaginary axis.
THEORY OF LINEAR SYSTEMS
55
This is the equation of a circle of radius R = [(ai - «2)2 -ft ft2]1'2 and
center at ( —a2, jO). The root locus is shown in Fig. 1.21.
In similar manner, one finds that the root locus for
shown in Fig. 1.22 is also a circle, centered at the point a0 on the figure.
It is apparent that only in simplified cases is it possible to obtain
useful information with relatively modest effort. When s is of high
order, the resulting algebraic complexity makes this approach quite
unattractive.
In certain cases, however, it is possible to construct complex root
loci from simpler ones by making use of the following theorem due to
Steiglitz.C1 7>
“Let Ti be the root locus associated with Gfts), and let T2 be the
locus associated with-G2(s). Then the intersections of Ti and T2 are on
the root locus associated with Gfts) G2(s).”
This theorem is useful when the total open-loop transfer function can
be broken up into a product of two other transfer functions whose
root loci can be drawn immediately. The concept of an arbitrarily
located coincident pole-zero pair in combination with the above theorem
greatly facilitates the construction of various types of root loci. We will
illustrate the application of these ideas in the construction of the root
locus for
56 ELEMENTS OF MODERN CONTROL THEORY
Im
K(s + a-i)
Figure 1.21. Root Locus for Y(s) =
(s + ai)2 + Pi2
THEORY OF LINEAR SYSTEMS
57
(s T ai)[(s + C*2)2
+ ^2 ]
The pole configuration for this case is shown in Fig. 1.23a, which also
shows a coincident pole-zero pair at ( —7, jO). This arrangement may be
broken up into the two separate pole-zero configurations shown in Fig.
1.23b and c. For each of the two latter cases, the root locus may be
drawn immediately as shown. Superimposing these two loci yields points
of intersection that define a point on the root locus for the given transfer
function. Putting the coincident pole-zero pair at another location along
the real axis again yields two further points on the root locus of the
original function. By taking a sufficient number of locations for the
coincident pole-zero pair, a number of points on the root locus for the
given transfer function will be obtained so as to completely define the
curve. Figure 1.24 shows the final form of the root locus constructed in
this manner.
Im
Im
Im
D
O
Root
Final
1.24.
Figure
60 ELEMENTS OF MODERN CONTROL THEORY
where h(t)is a time function of the form shown in Fig. 1.26. By taking
the Laplace transform of both sides of this equation, we find
1 - e-7s
H(s) - £[h(t)] = (63)
s(l — e~TS)
Noting that
(1 - e~TS).
Sn JIlC05
n — 0, =t 1, =t 2, • • •
where
CO
r(t)
(b) Sampler
r*(t)
If we denote the residue of H(s) at the pole, s„, by k„(s„), then we have
1 — e-/"“sY
27rnj
n = 0, ±1, ±2, ••%
62 ELEMENTS OF MODERN CONTROL THEORY
h(t)
1 — e~inu°y
R*(s) = £ R(s — s„) (64)
2irnj
n = 0, rbl, ±2,
Since
lim
1 — e~J'"“vy 1
n— >0 2rrnj T
1 _ e-jno>sy
R*(s) = 1 R(s) + E R(s - s„) (65)
-L n 2xnj
II — rbl, rb2, • • • .
n = 1, 2, 3, •••
(67)
R*(s)
« ^ X R(s- jmo,)
n
(68)
But 1/T is precisely the residue of Ar(s) at the pole sn, where*
n = 0, 1, 2, • • •
(69)
Thus, in fact, it does appear that the sampler output is a series of unit
impulses, with each unit impulse, 5(t — nT) , being multiplied by the
factor 7r(t - nT). The purpose of the preceding development has been to
emphasize the fact that the duration of switch closure, 7, need not be
infinitesimal. For the mathematical representation of Eq. (70) to be
valid, it is merely necessary that higher order terms in the expansion for
e-,n^y be negligible. This situation prevails when (7/T)2 is small
compared to 7/T.
It is also true that many authors assume, ab initio, that the equation
n = 0, ±1, ±2, %• %.
A(s)
R(s) (74)
B(s)
where A(s) and B(s) are polynomials in s, and the roots of B(s) are all
simple, then* y
A(sjO
R*(s) E A r(s S*) (75)
k= 1 B'(st)
where
B'(st)
=^B(S)]s=ii
q = order of polynomial B(s)
Si = a pole of R(s) ; i.e., a root of B(s) = 0.
y-' A(s k) z
Z[r(t)] = R(z) = (77)
k=i B'(st) Lz - eT3k_
1 1 d™ / 1 \
- t- = (_l)m . (- ).
(s + a)m + 1 m! dam \s + a/
For example, if
l
Ri(s)
(s + a)2
then*
= Z
1
U
.(s + a)2_ da _ s +f aJ
R(s) - 18(3+2) .
(s + 3)(s + 6)2
24 2
R(s) +
(s + 6) 2 (s + 6) (s + 3)
Now, by letting
2
Ri(s)
s + 6
2
R2(s)
s + 3
2z
Ri(z) =
z — e~bT
2z
R2(z) = _ p— 3T ’
z — e
Also, if
24
R*(s)
(s + 6) 2
The Z transform operation is expressed by any one of the three equivalent symbols shown.
THEORY OF LINEAR SYSTEMS 67
then
R3(z)
or
R3(z) 24zTe_6r
(z - e_6r)2'
x 2z 2z 24zTe-6 T
R(z) = - — - 1- - - .
z — e-6T z — e~3T (z — e-6T)2
-nTs -n
e 6(t - nT) z
1 z
u(t)
s z - 1
i Tz
t
2 2
s (z - 1)
i -at z
e
s + a -aT
z - e
b z sin b T
sin bt
s
2 +
u2
b z2 - 2 z cos bT + 1
-aT
b “at * ht z e sin b T
e sm bt
2 -aT , ^ ^ -2aT
(s +a)2+b2 z - 2 ze cos bT + e
ELEMENTS OF MODERN CONTROL THEORY
68
n = 0, ± 1, ±2, ••• .
Therefore
R*(si - jws)
1_ R(si — jnw, — j«.)
T n
=^ L R[si
- + n)].
By letting m = 1 + n, and noting that
m = — °° when n = — »
* o- -
c*(t)
-%º
Noting that R*(s) satisfies the periodic property Eq. (78), this further
reduces to
It follows readily from Fig. 1.28 that the output may be represented
as*
t'i(s) = E r(nT)
n= 0
Is S
ce
= E r(nT)e_nrs
n= 0
It follows that
m = 1 — A.
r(t) c(t)
0(8) *> At
T
The output, c(t), in Figs. 1.28 and 1.30 is continuous, but the Z
transform of the output yields values only at the sampling instants.
However, by using the modified Z transform defined by Eq. (85), the
actual output can be obtained by varying A between 0 and T. One can,
in fact, develop a complete theoryU9) based on the modified rather than
conventional Z transform, but this is Jbeyond the scope of the present
discussion. The modified Z transform will be used only to facilitate the
analysis when transport lag is included in the control loop.
We note only that if G(s) may be expressed as
A(s)
G(s)
B(s)
where A(s) and B(s) are polynomials in s (the latter of order q) and all
the roots of B(s) are simple, then— analogous to Eq. (77)
1 s lT
r, , \ v' Ms*) (86)
G(z,m) = z 1 2^ 777- t .1 - e-7’(,-s/t)J
k=1 B (s^
Ri(s) = Gi(s)R*(s)
-nTs m-l-n
e 6 (t - nT) z
1 i
u(t)
s z - 1
i m T T
t
s
2 z - 1 ' ,(z - 1)
2
-amT
i -at e
s + a -aT
z - e
. . _ -aT . , , ^1 -amT
b -at . . . |zsinmbT +e sin(l-m)bT|e
e sin bt
2 -aT . _ -2aT
(s +a)2+b2 z - 2 z e cos b T + e
R(s)
OF LINEAR
22L G(s)
- -a
V'
I | o -
C(s)
—
75
c<z>
%º
o-
C(z)
r—** » o— %º
1
R(s) . R(z) .
1 C(s)
G (s) G (si
2V
C(z)
=R(z)Z[Gx(s)G2(s)j= R(z)
GJG,,(z)
(b) Sampled Data System With Cascaded Elements
Gi(z)G2(z) ^ GiG2(z)
Gi(z)G2(z) = Z[Gi(s)]Z[G2(s)]
G!G2(z) = Z[G1(s)G2(s)].
Using the relations shown in Fig. 1.31, the equations for the feedback
systems shown in Fig. 1.32 may be derived without difficulty.
When a pure lag appears in the control loop, the open-loop transfer
function may be written as
T0 = XT
1. 0 < X < 1.
2. x > l.
3. X = k = integer.
For Case 2
X — l -)- X'
l = integer
.—Pf'o-E®*-
z = esT (92)
w = — cr tan 7
where
S = — <7 JCO )/
7 = cos -1fi
— q—(TT(1+ j tan y)
so that
= e •’ cos 2ir
(I -wn2 a2\12 .
) + i sin
fa>n2
2ir[ --
—cr2Y/2
I
V w„2 / \ ws2 / _
where
2 7T
= "TTr-
T
Im
The use of such a diagram in connection with a Z plane root locus will
be discussed subsequently. Note that a different diagram is required for
each different value of sampling period, T.
In order to investigate the stability properties of a sampled data
feedback control system of the types shown in Fig. 1.32, it is necessary
to determine the location of the closed-loop poles of the system. In Fig.
1.32a, for example, one must determine the roots of
1 + GH(z) = 0.
Im
where
“3 — KxKcKj/jc
2 nio HcOiT
COb — -
OM Tc
f = 1, / Ke
" 2\ KaKr»c •
where
Do =
C
Id
0
i)
c ./
e~2aT — e~aT cos (bT — <p) sec <p
2G
Go —
Ua
Ci=u~2 +4^2- 0
( 0 \03b '
ELEMENTS OF MODERN CONTROL THEORY
86
a —
b = aWl -
*- ,an“(iLr^‘)
Rj = 2e~aT cos bT
R0 = e~2aT.
Do = -49.8 X 10~6
With these values, we see that the numerator polynomial of G(z) has
roots at -0.202 and essentially zero, while the roots of the denominator
quadratic are both essentially zero. Consequently, for purposes of
drawing the root locus, G(z) may be written as
T.04Ka (z + 0-202)
G-(z) =
. 0.485 z2(z — 1)
The root locus for this case is shown in Fig. 1.39. Choosing Kw= 0.2
yields closed-loop poles as indicated by the small squares. The response
is governed by the complex pole pair shown, which has an undamped
natural frequency equal to cos/8 or 0.63 rad/sec and a relative damping
factor of 0.35 as determined by an overlay on Fig. 1.35. Instability
occurs for Kw > 0.37.
THEORY OF LINEAR SYSTEMS 87
1.39.
Figure
Plane
Z
6.
Example
for
Locus
Root
88 ELEMENTS OF MODERN CONTROL THEORY
1.41.
Figure
Plane
Z
for
Locus
Root
6.
Examp
=0.25
K
Point
Operating
NOTE:
THEORY OF LINEAR SYSTEMS 91
a. Stability.
b. Sensitivity.
c. Noise.
d. Transient Response.
e. Static Accuracy.
It goes without saying that, first and foremost, the system must be
stable. Measures of stability are important from a practical point of
view and will be considered shortly. Sensitivity is usually measured in
terms of variation of system performance as a function of variations in
prescribed parameters. Since this is the subject of a separate chapter, it
will not be considered here. Similarly, the behavior of the system in the
presence of noise is analyzed in Chapter Four and will be dismissed from
further consideration here. We shall, therefore, be concerned exclusively
with stability, transient response, and static accuracy.
The classical interpretation of degree of stability stems from the use
of the Nyquist plot of the open-loop transfer function as the medium of
analysis. Since stability is related to encirclement of the 1” point, it is
natural to relate degree of stability to proximity of the Nyquist curve to
this point. The terms phase margin and gain margin were defined in Sect.
1.1.2. These represent the phase lag (or, in some cases, phase lead) that
can be added to the system (at unity gain) before instability occurs; and
the factor by which the gain can be increased (or decreased) before
instability occurs. Thus, one may interpret each of these quantities as a
“factor of safety” in the system.
In addition, phase margin is a measure of the response characteristics
of the closed-loop system to sinusoidal inputs. In Sect. 1.1.2, it was
shown that, for systems whose response is governed by a dominant pair
of complex poles, the phase margin is intimately related to the relative
damping factor. To examine the general case, in frequency-response
terms, consider the two Nyquist loci depicted in Fig. 1.42. In Fig. 1.42a,
the phase margin is relatively large, while in Fig. 1.42b it is small.
Consider now the closed-loop frequency response, viz.
ELEMENTS OF MODERN CONTROL THEORY
R(8) C(s)
KG(s)
+ n
Im^ (jw)
<p = tan = tan_V
(i«)
THEORY OF LINEAR SYSTEMS 93
M =
KG(jcd)
1 + KG(jco)
I =
and simdar types.(66) However, these are rarely used and will not be
discussed here.
THEORY OF LINEAR SYSTEMS 95
The discussion of this section will focus on the control system depicted
in Fig. 1.45. Here Gi(s) represents the fixed plant that has a transfer
function of the general form
M
K II (s - z»)
Gi(s) = — — - (93)
N
s" II (s - p3)
i—l
(96)
K0 = lim s2G(s)
s->0
which are known as the position, velocity, and acceleration error coeffi¬
cients, respectively, and where
G(s) = Gc(s)Gi(s)
1
lim[e(t)] = lim (97)
t—>CG 3— >0 _1 + G(s)_
1
lim[e(t)] = (98)
t-tx 1 + Kp
K IJ (s — z.)
i= 1
Gi(s) =- - (99)
n (s - pj)
;% =l
k n (-z.)
KP = — —- . (100)
N
n (-py)
y-i
Gi(s) = - (101)
(s - Pi)(s - p2)(s - p3)
In other words, for a unit step input, the final steady-state value of
the output is 1 —U.253 = U.747. This error may be reduced by increasing
K, which, in turn, decreases the relative damping factor and which, for
K sufficiently large, leads to instability. We seek to increase the open-
loop gain (and therefore decrease the steady-state error) without de¬
creasing the relative damping factor. For this purpose we introduce the
compensating lag network
A(TiS -(- 1) A (s z c)
(102)
(aTiS +1) a (s pc)
where
= — > 1
Pc
KA (s - zc)
G(s) = — (103)
a (s - pc)(s - pi)(s - p2)(s - p3)
K = n ISo- Vi (104)
ELEMENTS OF MODERN CONTROL THEORY
98
Im
3
KA So ~ Pd
II ISo- P. (105)
a
So zc i=l
and
A = a.
KA _ ( —z c)_
K V AKP. (108)
« (-Pc)(-pi)(-p2)(-p3)
1
lim [e(t)] =
t— *co 1 AK„
1
lim [e(t)J = 0.0328
1—>CG 1 + 29.5
be in the vicinity of the origin in order that the influence on the dominant
pole transient response be negligible.
The undamped natural frequency for the dominant mode, i.e., corre¬
sponding to the closed-loop pole at s0, is 4.65 rad/sec. Suppose now that
we wish to increase the response time of the system, say by increasing
the undamped natural frequency to 6.5 rad/sec without changing the
relative damping factor. In other words, it is desired to have the dominant
closed-loop pole at si in Fig. 1.46. It is readily ascertained that this may
be accomplished by adding a network which contributes 33° of phase
lead at si. There are an infinity of networks that will do this, but we add
the requirement that the position error coefficient have a prescribed
value.
The given plant now has the form
where
A = ^<1
pe
where
z0 s zc = 0.10
Po = pc = 0.01
Pi = -2
P2 = -4
P3 = —7
A = a = 10
K = 165.
II ISl - Pi
BKA _ |si —p£| 3'=0
a |Si —Z£| |Si — Z0|
(~P£) II (~Pi)
1=0
3
II |si - Pil
Si — p£| Iz£[ kl 1=0
(113)
Si — Z£| |pfI |si - z0|
II IP1
1=0
z£, pf, and B must be chosen such that si is a point on the root locus
for the system and the corresponding position error coefficient, K„" =
45.
This may be done as follows. Referring to Fig. 1.47, we write K/'as
(114)
a
ELEMENTS OF MODERN CONTROL THEORY
102
Im
where
X =
Ini c
IpeI d
Si — ze| = 8j
|Si —pe| = b
|z0| II
J=0
I —
si - z0| %n iPi1
THEORY OF LINEAR SYSTEMS 103
The quantity, L, may be calculated from the fixed poles and zeros and
is, therefore, a known quantity.
From the geometry of the figure, we have
yi
tan
C — Xi
yi
tan y
d - X!
<P = d — 7
tan /8 — tan 7
tan <p = tan (/? — 7)
1 + tan j8 tan 7
yi - ^Xi- —Jtanv
Applying the law of sines to triangle pe zi si, we have
a _ b
sin 7 sin d
or
b _ sin d _ sin /J
a sin 7 sin (/3 — 95)
_ yi _ (116)
yi cos <p— (c — Xi) sin ^
Substituting Eqs. (115) and ( 1 16) in Eq. (114) and solving for c yields
All the quantities on the right-hand side of this equation are known.
Having c, we then determine d from Eq. (115).
ELEMENTS OF MODERN CONTROL THEORY
104
L = 57.9 Kp = 45
toi = 6.5 xx = 2.0
v = 33 0 yi = 6.2.
We find, therefore
c = 5.96
d = 15.60
X = 0.382
pe = -15.60
z£ = —5.96.
BKA
660.
a
Since A = a
B 660 _ 660
K ~ 165
and
0 > <p.
c = 4.18
X = 0.414
d = 10.1
PC = —10.1
z£ = —4.18.
BKA
502
a
or
K" = 37.1.
tan /3 = % y —
C — Xj
„ sin (e ~ v)
tan /8 = - -- r-
1 — cos (6 — <p)
1 — cos (0 — <p)
cos /3 = / I=.
V2[l — cos (0 — <p)]
As shown in Fig. 1.47, the required pole and zero are then found after
a trivial geometric construction.
The discussions thus far have been concerned with Type 0 systems
and the position error coefficient. The procedure is, however, identical
for Type 1 and Type 2 systems where the parameters of interest are the
velocity and acceleration error coefficients.
The simple networks thus far discussed are not effective in all situations.
In the case just considered, the open-loop poles of the fixed plant were
on the real axis. This permitted considerable latitude in choosing a lead
network that would shift the closed-loop poles. In Fig. 1.48, the original
root locus, before compensation, is shown by dotted lines, and the root
locus of the compensated system is shown by the solid lines. (This figure
is not to scale.) The general qualitative effects are, however, apparent.
The compensator zero effectively cancels the pole, p2, which enables the
main locus to be shifted to the left by an appreciable amount.
THEORY OF LINEAR SYSTEMS 107
Im
Re
Figure 1.48. Comparison of Root Loci for Compensated and Uncompensated Systems.
Consider now the situation depicted in Fig. 1.49, where the fixed
plant has a pair of complex open-loop poles and one pole, p3, on the
real axis. The uncompensated root locus is shown by the dotted lines. A
simple lead compensator of the form (109) cannot alter the system
significantly, in the way of increasing either the relative damping factor
or the undamped natural frequency. This becomes evident by observing
a typical “compensated” root locus as shown by the solid lines in Fig.
1.49. A problem of this type is precisely the one encountered in launch
vehicle autopilots where the lightly damped pole pair is due to the vehicle
bending mode. Here a more general type of compensation is required.
A suitable compensating network in this case is given by
Gc(s) = (120)
where u, < The complex zero pair, zj, is chosen to effectively cancel
the complex pole pair, pi, and the complex pole pair, pe, is moved far to
108 ELEMENTS OF MODERN CONTROL THEORY
the left. The resulting situation is depicted in Fig. 1.50. Figure 1.51
shows a passive linear network for realizing the transfer function, Eq.
(120).
Im
R R
3 4
„ T T 1 + s + (T + T)s+1
E 12 R (R + R + R ) ' 1 27
out 1 2 3 47 J
In
R 3 (R 2 + R 4 )7 R
2
T T 1 + s + T 1 (1 + R
~) 7 + T 2
1 2 R (R + R + R )
12 3 4
T =
(z - l)R(z)\
lim [e*(t)] = lim
^—%ºoo 2— %º1 z[l + G(z)]j '
R(z) = -L-
z — 1
1
lim [e*(t)] = lim (125)
t—> oo z— %º1 _1 + G(z)_
k n (z - zo
Gi(z) = - —- . (126)
(z - l)n II (z ~ P>)
1=1
Gc(z) = (127)
(z - pc)
Figure 1.53. Pole- Zero Configuration in Z plane for Compensating Network, Eq. (127).
Gc(Z)
'G.(s)'
G„(z) = (1 - z_1)Z
L s J
or
G^s) G „(z)
s J (1 — z x)
Ga(s) (1 - ZCZ X)
s J (1 - Z !) (1 - PcZ 0
=( 1 ~ ZA 1 , (Zc- pA 1
\1 - pJ (1 - Z-1) \1 - pc/ (1 - PcZ-1)
Taking the inverse Z transform and solving for G^(s)
8/(1 Zc)
S +
1 - Pc (128)
G,(b) =
(s + a)
114 ELEMENTS OF MODERN CONTROL THEORY
where
a (129)
Eo(z) 1
Gc(z)
E;(z) 1 + Gi/Hx(z)
Ha(s)~
1 + (1 - z-!)Z
s
or
Ha(s) 1 - G c(z)
(132)
s (1 - z_1)Gc(z)
1_ 1
(zc - pc)b
H.4(s)
(1 - zc)(s + b)
where
b = (133)
Gi(s) = G*(s)Gj(s)
then one hold circuit may be eliminated by feeding the output of GH(s)
in Fig. 1.55 directly in G((s).
Given now the open-loop transfer function whose Z transform is de¬
scribed by Eq. (126), one may shape the resulting Z plane root locus,
using the compensator (127), to satisfy a variety of criteria. Nothing
essentially new in the way of technique is involved (other than inter¬
pretation), since the usual root locus rules carry over directly into the Z
plane. One may, for example, require a quicker rise time, increased
damping, and/or a decrease in steady-state error. The root locus shaping
methods are completely analogous to those for the continuous case,
except that results are interpreted via a diagram of the type shown in
Fig. 1.35. There is, consequently, more work involved, but no difference
in principle. For purposes of illustrating the basic approach, we will
take the transfer function of the fixed plant as
_ Ki _
s(0.1s + l)(0.05s + 1)
Tz
R(z)
(z - l)2
K* = 0.1 X Kx = 0.26
Im
z - 0.368
Ge(z)
z + 0.950
Im
compensation wherein the sampler following G„(s) in Fig. 1.52 does not
appear. Here, frequency-response methods are generally employed, and
these are more elaborate and complicated than the techniques discussed
above. Detailed expositions are available in the literature.!19)
It was noted in connection with continuous compensation methods
(Sect. 1.3.2) that a perennial problem in launch vehicle autopilots is the
presence of a lightly damped, complex pole pair adjacent to the imaginary
axis. When there is a digital computer in the loop, as depicted in Fig.
1.59, the Z plane root locus takes the form shown in Fig. 1.60. Here the
lightly damped pole pair is located adjacent to the unit circle, and the
corresponding locus is such that relatively small values of open-loop
gain will result in closed-loop poles outside the unit circle. The methods
of resolving this problem are similar to that used in the continuous case,
namely, a sampled data equivalent of the notch filter, Eq. (120). This
takes the form
Digital
Computer
120 ELEMENTS OF MODERN CONTROL THEORY
_ Uncompensated
It was noted in the previous section that the difficulties in this problem
are compounded because the bending pole locations vary with flight
time, thereby seriously limiting the usefulness of the above approach.
More sophisticated techniques are discussed in Chapter Six.
In the current generation of large launch vehicles (Atlas, Titan, Saturn),
the approach taken for this problem involves programmed gains and
filters. At specific times of flight, one switches to a different value of
open-loop gain (in the summing amplifier and/or rate gyro). This is
usually accompanied by the switching in or out of a prescribed filter.
This, of course, presupposes reasonably good data on the vehicle inertial
and bending mode properties as a function of flight time. For certain
complex vehicles, switching may be inadequate, or else the bending mode
data (especially for higher modes) may not be known with sufficient
precision. In this case, the use of one of the various adaptive control
techniques may be necessary. Needless to say, such a step must be
weighed against the added cost and complexity and the decreased
reliability that adaptive control entails.
THEORY OF LINEAR SYSTEMS 121
The traditional methods for the analysis and synthesis of linear feed¬
back control systems make extensive use of the concept of a “transfer
function,” whether in the frequency domain (Nyquist) or in the time
domain (root locus). In the rapidly developing fields of optimal and
adaptive control, however, it has been found more appropriate to deal
with linear systems via the “state variable” approach, in which matrix
techniques play a fundamental role. This concept of the “state of a
system” has also served to clarify various aspects of the transfer function
method, especially when applied to multivariable input-output systems.
The literature contains many incorrect results for multivariable systems
because various subleties involved in dealing with a so-called matrix
transfer function have been neglected. This subject will be discussed at
length in Sect. 1.5.
Actually, the terms “state variable” and “state space techniques” are
really new names for a body of ideas that have long been used in such
areas as analytical dynamics, quantum mechanics, ordinary differential
equations, and stability theory. These ideas were first introduced in
control theory by the Russians, mainly in the development of the
Lyapunov stability theory. In the United States, the work of Kalman
and Bellman (among others) contributed to the widespread adoption of
state variable methods in control system design.
The basis for all the subsequent discussions is contained in the follow¬
ing pair of matrix equations, which describe the motion of a given
dynamic system.
x = Ax + Bu (136)
v = Cx + Du (137)
Here
D = q X p constant matrix
d
(') - ( )
dt
n = order of the system.
X
P(b) _ ?=°
(138)
5(s)
X aiSn_l
i=0
e = 5 + Trx (139)
ao
o _ /3°
@n <2n
<2o
a _ 00
Pn—1 Oin—
1 (HO)
a 0
a Po
Pi — - <*1
«0
x2 = x3
(141)
Xn— i xn
1 r,
Xn [V 0!iXn —1 &n—1X2 ^nXi]
Ot 0
*We make the usual assumption that the initial conditions are zero.
THEORY OF LINEAR SYSTEMS 123
0 1 0 . o
/
ooi :
A = (142)
0
0
C = I, the unit matrix
B = (143)
0 D = 0, the null matrix
1
and
u == u„ = — = a scalar.
£20
where
«f(t)
At y AH* (145)
h k!
A° = I, the unit matrix.
If the initial condition vector, x(t0), is zero, then combining the above
two equations yields
Whde succinct answers have been provided to the three basic questions
posed in relation to the system, (136) and (137), a variety of details
remain to be filled in for purposes of design and analysis. In the first
place, the solution for x(t), as given by Eq. (144), requires the evaluation
of the transition matrix, <t>(t). Evaluating this quantity via Eq. (145) is
not the most efficient method, because of error uncertainty in the
truncation process. A practical and theoretically exact expression for
4>(t) is( 38)
THEORY OF LINEAR SYSTEMS 125
_ \l£
3X2<
<t>(t) = M M-1
(148)
„X„(
where the X, are the distinct eigenvalues* of the matrix, A, and M is the
corresponding modal matrix. In other words, X,- is a root of the character¬
istic equation
|xi - a| = o
[XJ — A] Ci = 0
a,.,-+i i = 1, 2, • • •, (n - 1)
*When there are multiple eigenvalues, the expression for <f>(t) is much more complicated. For a
complete treatment in this case, see Ref. 25 .
126 ELEMENTS OF MODERN CONTROL THEORY
are zero, then we say that the system matrix is in companion standard
form, and we denote it by F. It was noted earlier that the system matrix
for the set of equations, (141), is of this form.
If the input matrix, B, is of dimension n X 1 (i.e., an n vector), we
represent it by the symbol, f. In the special case when the nth component
of f is unity and all the other components of f are zero, we denote it by
the symbol, h. The system represented by F and h in the state-variable
form has a special name, the phase variable ( canonical) form.
We will also use the lambda and Vandermonde matrices, which are
defined, respectively, as follows.
A = (149)
T = (150)
x = Ax + fui (151)
where
Ui = a scalar.
The transformation
x = Jz (152)
z = Fz + hui (153)
THEORY OF LINEAR SYSTEMS 127
where*
J = MR_1T_1 (154)
x = 't'z ( 156)
—fs fa f2. fn 1
-fa f4. -fn 1 0
f4 :
! -fn 1 0
-fn 1 0
1 0 0
* = Eft. (159)
x = My (160)
y = Ay + Pu (161)
v = Ty + Du. (162)
Here
Noting that
THEORY OF LINEAR SYSTEMS 129
s —
s — \2
(Is - A)-1 =
(166)
• 1
S — X„
L(s)- i-1± S - X*
+ D (167)
where
K, = r,P,, (168)
9(s) = 3s2 + 4s + 1
5(s) s2 + 5s + 2
It is required to express this in the state variable matrix form, Eqs.
(136) and (137).
From a comparison with Eq. (138), we have
ao = 1 Po — 3
ai = 5 Pi = 4
£*2 = 2 Pi = 1.
We are still using the assumption that the eigenvalues, X,, are distinct.
130 ELEMENTS OF MODERN CONTROL THEORY
Xl = x2
(a)
x2 = — 2xi —5x2 + 5
"-5 '
7 = L—H_
Therefore, the transformation, Eq. (139), becomes
To see how the transfer function format is recovered from Eqs. (a)
and (b), note that from the former
xi = — 2xi — 5xi + 5
or
(s2 —
|—5s —
f- 2)xi — 5. (c)
d = 35 — 5xx — llx2
= 35 — 5xi — llxi
or
Eliminating xi between Eqs. (c) and (d) yields the original transfer
function.
x = Ax + fui
THEORY OF LINEAR SYSTEMS 131
where
_2 -1 r r
1 0 1 f = 1
-1 0 i_ _i_
and ui is a scalar. /
It is required to express the system in canonical (phase variable) form.
The characteristic equation is
Ai = 1
V = —1 + j
X3 = -1 - j.
5
M = •(3 + 4j) -(3 - 4j)
(2 + j) (2 - j) .
whose inverse is
-2j 2j 8j
M-1 - — : % (1- j) -1 1
10j 1 -1
(1 + j)
1 1 1
T = 1 -(1-j) -d + j)
1 —2j 2j
The inverse is
4j 4j 2j
T"1 (1 + 3j) (1 - 2j) -(2 + j)
10j -(1 - 3j) -(1 + 2j) (2 - j).
132 ELEMENTS OF MODERN CONTROL THEORY
8
M-T = -d - j)
10j
(1 + J')_
and
2j
T-'h = -(2 + j)
10j
(2 - j)_
_1J-H
fl o o 8j 2j
0 r22 0 -(1 - j) =
“(2 + j)
_0 0 r33_ _ (1 + j)_ (2 - j)_
1 0 0
0 2(1 + 3j) 0
0 0 2(1 - 3j)
whose inverse is
20 0 0
R-1 i 0 (1 — 3j) 0
Lo 0 (1 + 3j)
J = MR_1T_I
0 -1 1
0 3 1
2 1 1
THEORY OF LINEAR SYSTEMS 133
-1 2
' 1 0
1 0
0 1
J-*AJ 0 0 5= F
2 0
J-T
1 -2 2
E = [f Af A2f] 1 2 -2
1 0 2
ti = 2, f, = U, f, = - 1
0 1 1
ft = 1 1 0
1 0 0
6 -l 1
* = E ft = 0 3 1
2 1 1
ELEMENTS OF MODERN CONTROL THEORY
134
As a check, we calculate
N^-l -1
-1
0 1 0
0 0 1 = F
2 0 -1
and
0
'F~1f = 0 h.
1
x = Ax + Bu
v = Cx + Du
where
"-2 -1 1
A = 1 0 1 B =
-1 0 1
2 0 1
C
0 3 0
We seek the solution for v(t) when the initial condition vector, x(0), is
zero, and u(t)is a prescribed function of time.
This can be done in either of two ways:
In the first case, the crucial step is the calculation of the transition
matrix, <t>(t). The system matrix, A, is identical to that of Example 8.
THEORY OF LINEAR SYSTEMS 135
0 5 5
1 (-3 - 4j) (-3+ 4j)
1 (2 + j) (2 - j) _
~eXl‘ 0 0 -2j 2j 8j
0 eXs< 0 ( —1 + j) —1 1
0 0 eXs' _(1 +j) 1 —1
Assume now that the vector, u(t), is composed of a unit step and unit
ramp as follows.
1
u(t)
t
The second quantity on the right-hand side of Eq. (144), which repre¬
sents the particular solution, is a three-dimensional vector whose com¬
ponents thus become
f [4$3i(t — o) T t4>32
(t — cr) -)- 3t4>33(t — a)]da.
*’o
Using the f>iy(t) obtained above, these components are readily evaluated
by elementary methods. For example, a lengthy but straightforward inte¬
gration yields, for the first of the above components
v = Cx + Du.
To obtain v(t) via the matrix transfer function approach, we use Eq.
(167). This requires the values of r and P. These are readily calculated
as follows.
0 5 5
"2 o r
r = CM = 1 ( —3 — 4j) ( —3 + 4j)
0 3 0
J (2 + j) (2 - j) _
1 (12 + j) (12 - j)
.3 -3(3 + 4j) -3(3 - 4j)
-2j c7 <—»
•
GO 4 0
P = M-‘B = — -(1 - j) i i 0 1
10j
_ (1 + j) 1 -i _ _io CO
-4 13
1
2(1 + j) -j
5
_2(1 ~ j) j
THEORY OF LINEAR SYSTEMS
137
Therefore
K,
L(s) = D + z
(s - X.)
1 (1 + 12j)
+
2(11 - 13j)
5(s + 1 + j) L6(l + 7j) -3(4 + 3j)_
Since the last two matrices are complex conjugates, the above reduces
to
~2(lls - 2) (s + 13)
+
5(s2 + 2s + 2) [6(s + 8) —3 (4s + 7)
Now
138 ELEMENTS OF MODERN CONTROL THEORY
1
s
U(s) = £[u(t)] = £
1
S2
We obtain, therefore
V(s) = L(s)U(s)
1 ' "1
s , 1 1
hP*- 1—1_
1CO s
V (s)
1 5(s - 1) -12 39_ 1
_ S 2_ _ S2_
1
s
2(lls - 2) (s + 13)
+
5(s2 + 2s + 2) _6(s + 8) —3 (4s + 7) 1
S2
Let us now consider the case where the input vector, u(t), is sampled
periodically and constrained to be constant over each interval, i.e.
k = 0, 1, 2, ••• (169)
i (ar) = kr + £r — a (171)
In particular, for £ = 1
where
G= <t>(r)
= eAT (175)
H= f $(*)Bd^ (176)
and
xi+1 may be expressed as a function of the initial state and the input
sequence, m, by successive iteration of Eq. (173), viz.
x*4.i = G*+1xo +
(177)
k = 0, 1, 2, ••• .
140 ELEMENTS OF MODERN CONTROL THEORY
x* = G*x0. (178)
G — ml
G* =i^n (179)
i= 1 ;=1 -M* Mi-
f eXlT 0
E = eX2T
(181)
from Eqs. (175) and (148). The m. are obtained as the roots of
|g - Mi| = o. (182)
However
Therefore
THEORY OF LINEAR SYSTEMS 141
|G - MI| = |E - MI| = 0.
a. = eV. (183)
n k
G - ml
x* = z Mi*n X0.
i=l 3=1 lm. Aj
3^ i
kl <1. (184)
This is the stability condition for the discrete time system, Eqs. (173)
and (174).
We now seek to express the discrete time system, Eqs. (173) and
(174), in terms of Z transforms analogous to the Laplace transform
representation, Eqs. (146) and (147), of the continuous system, Eqs.
(136) and (137).
The sampled state vector, x*(t), may be written as
or
*The influence of the multiplying factor, y, in the discussion following Eq. (71), is not relevant
in the present case, since we are dealing with a sample-and-hold operation. See the discussion
following Eq. (84).
142 ELEMENTS OF MODERN CONTROL THEORY
00 CO
or
oo m— 1
where m is an integer.
Substituting Eq. (188) in (187), we find
We say that W(z) is the sampled data matrix transfer function for the
discrete time system, Eqs. (173) and (174).
The transformation
x* = My, (194)
where (as before) M is the modal matrix for A, converts the system, Eqs.
(173) and (174), to the normal form
Q = M-1H (197)
Example 10: Consider the discrete time system shown in Fig. 1.62.
We seek to obtain the solution for xi(t) at the sampling instants, first by
the state variable methods and then by the usual Z transform technique.
In terms of state variables, we may express the system dynamics by
Xi = x2
x = Ax + Bu
= Cx + D
where
0
A = B =
-2
C = [1 0] D = 0.
|XI - A| = X2 + 3X + 2 = 0
Xj = -1
X2 = —2.
1 -1
M
-1 2
M->
[? ;]•
Also
E
, =
fex>T 0
0 exA
while
<I>(t) = MEM 1= G
0.6005
-0.4652
0.2326
-0.0973 :]
and from Eq. (176)
0.1998
0.2325
r = CM = [1 -1]
(1 - e~”) 1
W(z) = Z
s (s2 + 3s + 2)_
= (1 - z-')Z
Ls(s2 + 3s + 2)J
146 ELEMENTS OF MODERN CONTROL THEORY
z 0.5z
(z - e !) (z - e 2)_
0.2(z + 0.368)
(z - 0.368) (z - 0.135)
V(s) _ s — 1
U(s) s + 5
which indicates that the system is stable. However, notice that a pole
and zero have been cancelled in the right-half plane. This operation,
which appears theoretically valid, must be carefully examined, as we will
show.
If we now choose state variables as shown in Fig. 1.63, we may
describe the system by the dynamic equations
xi = — 4xi 4- 5x2 — 5u
*2 = Xl + U
v = xi — x2 -F u
THEORY OF LINEAR SYSTEMS
147
or, equivalently, by
x = Ax + Bu
v = Cx -f Du
where
C = [ 1 -1] D = l.
The eigenvalues for the system matrix, A, are readily obtained; viz.
Xi = —5
X* = 1
x = Ax + Bu (199)
v = Cx + Du (200)
y = Ay + Pu (201)
v = Ty + Du (202)
transfer function matrix will not represent all the dynamic modes of the
system. If any of these modes is unstable, “hidden oscillations” will
occur.
It is important, therefore, to establish criteria for determining the
controllability and observability of given dynamic systems. The fol¬
lowing theorems serve this purpose.
We found that
-4 5
A = B =
1 0_
C = 1 -1] D = 1
\i = -5 n = 2
X2 = 1 q= 1
CB = -6 CAB = 30
= rank 1 = q
-5 25
[B i AB] 1 —5
= rank 1 < n
Furthermore
[Cr | ArCr] = _ 1 5
= rank 1 < n
1
M = modal matrix for A
1
1 1
M_1= k o 1 5
1
p = M-1B =
0
r = CM = [-6 0].
Since P has a zero row and r has a zero column, the system is neither
completely state controllable nor observable.
We would expect, therefore, that the transfer function calculated by
Eq. (203) would not contain all the dynamic modes of the system. We
obtain, in fact
_ 1 0
= [-6 0] s + 5 1 + 1
0 s — 1
6 s - 1
+ 1
s + 5 s + 5'
u 1 0 ‘ 0
A = 5 0 2 B = 0
-2 0 _2 _0.5_
C = [-2 1 0] D = 0.
THEORY OF LINEAR SYSTEMS
153
We readily compute
X» = - 1
X2 = -3
X3 = 2
0.5 0.5
M = -1.5 1
1 -0.25.
0.625
P = M-1B = 0.75
30
0.50
T = CM = [3 -2.5 0],
' 1 0 0
r 0.625~1
s + 1 1 0
[3 -2.5 0] 1 0.75
0 s -f- 3
0 0 s - 2
0.50 _
(s + l)(s + 3)
The transfer function does not contain the unstable mode corre¬
sponding to the eigenvalue equal to two.
Checking for the controllability and observability of large complex
systems may be quite tedious. The use of the following three theo-
rems(31> reduces the workload somewhat.
154 ELEMENTS OF MODERN CONTROL THEORY
For this case, the relevant sampled data matrix transfer function is
given by Eq. (193) or Eq. (198), viz.
Similarly, Theorem III is valid for the discrete time case if the test
matrix is replaced by
In other words, Theorems II and III are true for discrete time systems
of the type described by Eqs. (204) and (205), if A is replaced by G, and
B is replaced by H.
A natural question presents itself at this point. Suppose that the
system represented by Eqs. (199) and (200) is completely state control¬
lable and observable. How are these properties affected by the intro¬
duction of sampling? An answer to this question is contained in the
following^41)
Theorem VIII: Suppose that the system, Eqs. (199) and (200), is
completely state controllable and observable in the absence of sampling.
Then, if u(t) is constant over each sampling interval, r, a sufficient con¬
dition for the complete state controllability and observability of the
sampled system is
27rm
Im(Ai - Ay)
whenever
Re(Ai - Ay) = 0
Example 13: Consider the system, Eqs. (199) and (200), where
0 0
A = 0 1
- (a2 + tt2)
1
B = 0
7T
C = [1 1 0J
D = 0.
It is assumed that u(t) satisfies Eq. ( 169). In this case, we may represent
the system in the discrete-time form, Eqs. (204) and (205), proceeding as
follows.
THEORY OF LINEAR SYSTEMS
157
|XI - A| = 0
Xi = —7
X2 = —a + ]7r
X3 — — a — J"-
0 0
M = 1 1
— (a — j 7r) •(a + jjr).
2j 7T 0 O'
M-1 = —1^
0 (a + j7r) 1
2j ir 0 — (a — j 7r) -1
1
P = -0.5j
0.5j
r = [i 1 i]
<t>n(l) = e-7
$22(i) = —e~a
4*33(1)= —a
0 0
— e~a 0
0 -e~“
while
" (1 - e-7)
y Hi
tt (e-“ - 1)
H = = h2
a2 + 7r2
0 _0 _
158 ELEMENTS OF MODERN CONTROL THEORY
hi (1 - e~7)
7
7r (a + j7r)(e~a — 1)
a2 tt2
a( —a + j7r)(e~a — 1)
a2 + 7r2
Using Theorem IV for the continuous system we find that the system
is completely state controllable and completely observable, since P has
no zero rows and r has no zero columns.
However, by applying Theorem II to the discrete time system, we find
[H ; GH j G2H]
which is of rank 2 (< n = 3), so that the discrete time system is not
completely state controllable. Furthermore, using the discrete time
version of Theorem III
~1 e~7 e-27'
= 1 -e“« e-2"
.0 0 0
which is also of rank two. Therefore, the discrete time system is not
completely observable.
The same result could have been obtained directly from Theorem
VIII, since
0 0
z — e 27T,jHl
1
= [1 1 1] 0 0 (a + j7r)H2
z - eXz, 2ir]
( —a + j7r)H2_
0
z — e
(1 - e~y) 7r(e-a + 1)
y(z — e y) (a2 + 7r2)(z + e~a)
V(s) 1
+
U(s) J Ls + 7 (s + a)2 -(- 7r2_
W(z) = (1~e~Tr) +
t7(z — l)(z — e yT) (a 2 -f- 7r2) .(z - 1)
where
- .an-(f)
1 2
(s + 1) (s + 1)
L(s) (209)
-1 _ 1_
(s + l)(s + 2) (s + 2)
K,-
L(s) = z
(s - s.)
and the st- are the poles of the matrix elements in L(s). For the special
case of Eq. (209), we have
Si = — 1 s2 = -2
and
1 2
Ki
-1 0
(210)
0 0
K2 =
1 1
m K
L(s) = E + D (211)
i= 1 S - s,
where
Let the rank of matrix Ki be denoted by r,-. Then L(s) can be repre¬
sented by a system of differential equations, Eqs. (199) and (200) or
162 ELEMENTS OF MODERN CONTROL THEORY
n= Ir, (214)
i=l
Ki = PSi (215)
Pi = {piTPi)~lPiTK„ (216)
X2I2 (217)
**XmIn
*The matrix (p;r Pi), is nonsingular, since Gram determinant [pi7 p<| is nonzero if the columns of
pi are linearly independent.
THEORY OF LINEAR SYSTEMS 163
0i
02
P = (218)
0 m.
P [plP2* Pm]
(219)
I, = unit matrix of order r,
and\i =
Note that is an eigenvalue of multiplicity r,: and that the order of
m
system is n = X! in.
i= 1
51 — — 1
52 = —2
53 = —3
s4 = —4
s4 = —5.
"8 0 0 0
0 0 4 1
3 1 0 3
+
CO
<N
+
<N
CQ
lO
+
CO
+
CQ
CM
CO +
.1
ti¬
Tt<
iO +
+ CQ
CQ
+
CM
00
+
CO
+ +
CQ CM
ll
CQ
THEORY OF LINEAR SYSTEMS 165
8 0 0
Pi = 0 0 4
3 1 0
and
1 ' —3 0
(pirpi) 1 -3 73 0
64
0 0 4
1 0 0 0
Pi = 0 1 0 3
0 0 1 0.25
-4.5 -3 0 1
K2 0 0 -6 0 = rank two
0 0 0 0
-3 0
0 -6
0 0
1.5 1 0 -0.333
& =
0 0 1 0
0 0 1 f
k3 1 1 2 2 = rank two
-3 -3 1 1
0 1
1 2
-3 1
110 0
0 0 11
-0.5 9 1 0
K 0 0 0 0 = rank one
0 0 0 0
166 ELEMENTS OF MODERN CONTROL THEORY
P4 = 0
_0_
/?4 = 0.5 9 1 01
-
0 0 0 o'
k5 = -
1 0 0 -3 s= rank one
2 0 0 6_
-
o'
P5 =
—
1
2_
$5 = [1 0 0 3].
-Is
-21,
A = %3U
-41,
-51,
s o o -3 0 0 1 1
r = 0 0 4 0 -6 1 2 0
3 1 0 0 0 -3 1 0
1 0 0 0
0 1 0 3
0 0 1 0.25
1.5 1 0 0.333
P = 0 0 1 0
1 1 0 0
0 0 1 1
0.5 9 1 0
0 0 3
~~ = T*(s) (220)
R(s)
— = TjXs) (221)
D(s)
where
H(s) are compensation networks that are to be designed such that the
transfer functions of Eqs. (220) and (221) are realized. It is immediately
evident from Fig. 1.70 that
C(s) _ G2(s)
D(s) 1 + Gi(s)G2(s)H(s)
1 C(s) {ZZJJ
= T£7(s).
~ Gi(s) Lr(s)J
C(s) _ B(s)
(224)
D(s) A(s)
(228)
_ G2i(s)_
L(s) (231)
Gi(s)[Gii(s)G22(s) — Gi2(s)G2i(s)]
In contrast to the previous case, wherein the disturbance signal was fed
back directly, the use of a disturbance-dependent variable as an added
loop definitely affects system stability.
From the sensitivity point of view, transfer function G2i(s) and Gi^s)
behave open-loop with regard to parameter variations.!48) However,
sensitivity to parameter variation in Gn(s), G22(s), Cn(s), and L(s) is re¬
duced, because these systems behave “closed-loop” in this respect.
Nevertheless, parameter variations anywhere in the system affect the
invariance condition adversely.
While the use of the invariance principle is theoretically attractive, its
practical application is limited by four main problems:
D(s)
Figure 1.73. Feedback System with Absolute Invariance via Disturbance Feedback.
172 ELEMENTS OF MODERN CONTROL THEORY
a path from the disturbance directly. Figure 1.74 shows the use of an
additional loop from a disturbance-dependent variable. The main ad¬
vantage in using invariance with a feedback system is the reduced sensi¬
tivity to parameter variations. While in practice it is not feasible to achieve
a theoretically exact absolute invariance, it is possible to design high
quality control systems that have a high degree of “approximate invari¬
ance.”
In the case of direct disturbance feedback, as illustrated in Fig. 1.73,
stability is not affected by the additional path.* Here we find that
L2(s) = (233)
Gi(s)
_ G2l(s) _
L(s) (235)
Gl(s)[Gi2(s)G2l(s) — Gii(s)G22(s)]
a = a1 + aw (238)
, w
Y = - (239)
U
R(s) = 0C(s)
C(s) = d( s)
X(s) * «(s)
Y(s) S « '(B)
D(s) = aw(s)
174 ELEMENTS OF MODERN CONTROL THEORY
Also
K,KC (242)
Gi(s)
s + Kc
Eliminating a from Eqs. (236) and (237) via Eq. (238), we obtain,
after some reduction
/
a Gn(s) Gl2(s) aw
(244)
j _G2l(s) G22(s)_ J
where
Gn(s) (245)
(246)
Gl2(s)
G2l(s)
(247)
G22(s) (248)
L(s)
mU(s
+Kc)s (249)
T,K,K,
(l+|)
The result thus obtained, while theoretically attractive, suffers from
obvious practical limitations. The zeros in the numerator of L(s) intro¬
duce serious problems of noise and physical realizability. Furthermore, a
simplied mathematical model of the actual vehicle was used to obtain the
result, and it may be expected that a more complete representation of
the vehicle would introduce substantial complexities. Finally, the actual
vehicle parameters are time varying, and this would tend to vitiate the
invariance properties.
However, one succinct feature of the compensating network, L(s), is
quite apparent: the need to cancel the actuator dynamics and use a
precisely defined gain in the network. The use of acceleration feedback
for purposes of load reduction is, of course, not new. Nevertheless, the
present approach to the problem is novel, and it would appear that
attempting to achieve a kind of approximate invariance in the manner
outlined would be worthwhile. To our knowledge, no investigation along
these lines has been made.
The discussions thus far have been in terms of system transfer func¬
tions, a procedure that has the virtue of familiarity, since United States
control engineers have traditionally used this approach. Much of the
Russian literature on the invariance principle is, however, expressed in
the state variable format, leading to results that are sometimes more
convenient for special purposes.
We may formulate one type of invariance problem as follows. Suppose
we are given the system
x = Ax -f- fu (250)
v = Cx (25 1)
which is expressed in the state variable format of Sect. 1 .4, i.e.
u =e input; a scalar
What are the conditions that ensure that v is absolutely invariant with
respect to u? This problem has been solved by Rozenoer,(45> who
showed that a necessary and sufficient condition for this invariance is
that the relations*
CrAT = 0
(252)
k = 1, 2, - (n - 1)
be satisfied.
This result is typical of those derived in the Russian literature, which
nevertheless is of a mostly theoretical nature concerned with questions of
existence and realizability.
A simple application of the above result is contained in the following.
-3 o
A =
-4 a22
C = [2 1]
ficj -T f2c2 — 0
2f, + 10 = 0
fi — o
a22 = — 15.
Xj = —5
X2 = -13
NOMENCLATURE
Vi — mpiaT/\yy
(0 =
u negative slope of ith bending mode = —
(0 =
<f> normalized mode shape function for ith bending mode
f(i), w(i); a),; fc, fpi) wpi = relative damping factor and
undamped natural frequen¬
cy for: ith bending mode;
rate gyro; engine actuator;
i111slosh pendulum
(Al)
s20 = S2 + Ac 8 + fjact — y. viT i
XT /
(A2)
(s2 + 2r(i)co<i>s + [w(i,]2)q(t> = —— (m*£*s2 + Tc)5
M(l)
eF = e + 2>pJ0q<o +
Kffico«2s
[0 + (A4)
(s2 2£rCi)RS-f- A)ft2)
Note that this formulation provides for the possibility that the rate
and position gyros are not placed at the same location along the vehicle.
(A5)
5c= Kx(s+ K-^G,(s)04
*A. Greensite, Analysis and Design of Space Vehicle Flight Control Systems, 1970 .
180 ELEMENTS OF MODERN CONTROL THEORY
9e — dc dp (A6)
u E q(i)(t (A9)
while
— = - E q(1)(t)«r<*>(*) (A 10)
dl i
where
(All)
The vehicle geometry is depicted in Fig. 1.A1, and the sign convention
for bending parameters is shown in Fig. 1.A2. Figure 1 .A3 is a block
diagram of the pitch plane autopilot; it is assumed here that only the
lowest bending mode is significant.
The complete set of equations, (A1)-(A8), is too formidable for
pencil-and-paper analysis. By introducing appropriate simplifications,
the various essential features of the control system are placed in evidence.
Ultimately, a simulation of the complete system via computer serves
merely to refine the results obtained with the approximate analysis.
We assume first that perturbations in flight path angle are negligible.
This enables us to write
Furthermore, it is assumed that the rate and position gyros are located
sufficiently close together that
(*) (*)
(A12)
THEORY OF LINEAR SYSTEMS
181
where
^(1>TC (A 14)
Ai —
182 ELEMENTS OF MODERN CONTROL THEORY
2f<1)co(1) (A 16)
Cl
1 - A!
Te (A 17)
r =
nii
f/t — (A 18)
THEORY OF LINEAR SYSTEMS 183
+
184 ELEMENTS OF MODERN CONTROL THEORY
where
k; = MU - V (UY - i)]-1
- V (U)2- 1 = tR
Consequently
2U 2U
K* + K* K,
OiR COft
Therefore
(A 19)
S2 + 2fflWSS + Ur2 — KsO)fl2
where
- = [KxKcK*Mc(1 - AO]
s(s2 - mJ(s + Kc) (s2 + 2r(1)«(^s + [co^>]2)
(A22)
- = KaKcKrhc
S(S2 — Aa)(s + Kc)
With the inclusion of a simple lag filter in the loop, this becomes
(A23)
(s + KO
1
s
186 ELEMENTS OF MODERN CONTROL THEORY
0F
Y^aKcKruA s + i)
(A24)
dE
dF (A25)
6b
dF (A26)
1 + -L = 0.
6e
APPENDIX B
This has the usual meaning where « is the undamped natural frequency
and r is the relative damping factor. We form successive powers of s as
follows.
s2 = w2[(2f2 - i) - 2jrvr=T2]
(2D - 1)
(-4D + 3D
(8f4 - 8f2 + 1)
-2f
(4D - 1)
(-8D + 4D
187
188 ELEMENTS OF MODERN CONTROL THEORY
Tt(-f) = (-1)*T*(T)
u* (—r) = (-i)*u*(r)
we write this as
s*+i = - (i - r2)u,(r)]
= (-i)w«w[Ta) - ivr^u4+1(r)]
Q.E.D.
Consequently, if Eq. (B3) is substituted into the polynomial
f(s) = J2 a*s4
k= 0
THEORY OF LINEAR SYSTEMS 189
(Cl)
0
following which the elegant means of solving linear time invariant differ¬
ential equations via Eq. (Cl) are demonstrated. However, since no
motivation is given for choosing the peculiar form of (Cl) to begin
with, the procedure has the air of some esoteric ritual being conducted
by a sorcerer.
We will attempt, therefore, in what follows, to show that the function
defined by Eq. (Cl) arises quite naturally when one seeks to solve a
linear differential equation in a somewhat unconventional fashion. Our
line of departure is, of course, the original route taken by Laplace
almost 200 years ago—a route that has been almost completely sub¬
merged by the mathematical refinements developed subsequently.
The equation to be solved is
(C2)
y, a,D*x = f(t)
where the a < are known constants and D ! is the linear operator defined by
(C3)
i = 1,2, ••• n
X0 = x(0)
dx
]
(C4)
where
(C7)
Similarly
r d2x , dx r dx ,
/ e~st — dt = + s / e~s<— dt
dt2 dt. L dt
Assuming that
dx (CIO)
lim = 0
/% —&fCo dt.
foo drx
/ e_s( ——dt = — (sr_1x0 + s^xi + •
Jn dtr
(Cl 2)
+ SXr_2 + Xr_; /oo e- ‘xdt
d’x
lim = 0
dt’
i = 0, 1, 2, %• • , n- 1. (Cl 3)
For simplicity, assume that all the initial conditions are zero; i.e.
x,- = 0 for i = 0, 1, 2, • %• , n — 1.
THEORY OF LINEAR SYSTEMS 193
(Cl 4)
/ooo
/•oo
e_s‘xdt = / e~s(f(t)dt.
^0
F(s) (Cl 5)
X(s) =
*p(s)
'F(s) ^ (C16)
x(t) = £ HX(s)] £_1
-*P(
s)-
Consequently, the solution of the given equation is expressed in terms
of the inverse form of Eq. (Cl), requiring only that one tabulate a suit¬
able set of paired functions, g(t) <->G(s), in order to write the solution
virtually by inspection.
For example, if Eq. (Cl 5) is written as
F(s) _ A,
X(s)
(p(s) i=1 s — s,
/“ 0
e_"e5»'dt A
= -— -—
S — Si
N,(s) (D2)
H(s) =
D2(b)
The roots of Ni N2 = 0 in Eq. (D3) are called the zeros of the open-
loop transfer function G(s) H(s) and the roots of D,D2 = 0 are called
the poles of the open-loop transfer function G(s) H(s). Similarly, the
roots of 1 + NiN2/DiD2 = 0 are called the poles of the closed-loop
transfer function C(s)/R(s) and determine the operating root locations
corresponding to the system time constants. The numerator Ni/Di of
the control ratio C(s)/R(s) merely modifies the coefficients of the tran¬
sient components.
The characteristic equation of the system is from Eq. (D4)
1 + G(s)H(s) = 0 (D5)
194
THEORY OF LINEAR SYSTEMS 195
or
G(s)H(s) = - 1. (D6)
The linear factors in rectangular form can then be related to the polar
form as follows.
i<f>*i
s — Zi — rZ|e
s —Z2= rZ!e7^z2
s» = = 1>0Ne’N<t>Po (DIO)
s - Pj = rPley0pi
s - P2 = rP,e
Upon substituting Eq. (DIO) into Eq. (D7), the entire G(s) H(s)
transfer function can be written in polar form.
K(rZle^zi)(r22e^22)...(rZme^)
J*Z„
G(s)H (s) =
(Dll)
(rp^e,w0po)
'po (rPle^pi). . .(rP ey^p»)
= Kre^
where
%-rZm
r ( D 1 2)
r n r r
rPo rPl • • %rPn
THEORY OF LINEAR SYSTEMS 197
and
since Kr = 1 and j = 0 for G(s) H(s) to be a real number. From Eq. (D16)
the phase angle of G(s)H(s) becomes odd multiples of r
ForK < 0
\p = 0, ± 2t, ± 47r, • • •
(D19)
= 2k7r, k = 0, ± 1, ± 2, • • •.
198 ELEMENTS OF MODERN CONTROL THEORY
In summary, the two criteria that are required in plotting the root
loci of a system are the phase criterion and magnitude criterion.
Phase Criterion:
where
k — 0, zb 1, zb 2, • • • .
Magnitude Criterion:
K II rzj
i—1
- = 1 (D21)
n
rPon rPi
i—l
and plot the zeros Z, and poles P, in the s plane, using the
same scale for both the real axis and imaginary axis. Continu¬
ous curves start at each pole of G(s)H(s) for K = 0 and termi¬
nate on the zeros of G(s)H(s) for I\ = oo.
2. Draw the loci along those sections of the real axis which are to
the left of an odd number of real poles and zeros when K is
positive. When K is negative, the locus lies to the left of an
THEORY OF LINEAR SYSTEMS 199
N-\-n m
Ip.-Ez.
i=l i=J
centeroid = ac — — -- (D23)
(N + n) - (m)
^asymptotes = ± —, ± — k = 0, ±1, — 2.
4 4
K(s - 1)
G(s)H(s) =
THEORY OF LINEAR SYSTEMS 201
+jw +Jw
—X a o 1 o
K=0 K=0
K =0O| K - »
axis, the locus enters the real axis at the breakaway point and
terminates on the zero at I( = «. In both cases the loci are
perpendicular to the real axis at the breakaway point. The
breakaway point can be found by assuming values of a,, be¬
tween the poles or zeros until the following equality is satisfied.
T - - _ + T r_ 1_ 7 = T t_ -_ T + £ _ i _ T
cri, + Pf ah + Zr,-1 \0h + ZfJ |cr6 —
(- Pr,|
(D26)
where
or
To show the relationship between the root locus and the correspond¬
ing time solution, consider the following second order differential equa¬
tion relating the attitude, <p, to the commanded attitude,
ip T- 2£ci)ntp
+ u„2tp= w„2<pc (D27)
THEORY OF LINEAR SYSTEMS
203
Figure 1.D6. Relationship between Roots in the s Plane and the Corresponding Transient
Solution in the Time Domain.
204 ELEMENTS OF MODERN CONTROL THEORY
<p{
s) = _ u„2 (D28)
<pc{
s) s2 + 2fw„s + w„2
Si ,2 = — C0n, f = 1
si, 2 — ±jw?i, r — o.
When the roots Si, S2 are located to the left of the j« axis, the system
is stable, and the transient part of the response decays to zero. When
the roots are located on the jw axis, the system contains no damping, and
sustained oscillations result. As the roots move to the right of the jw
axis, the amplitude of the oscillations increases with time, and the
system is unstable. The above statements are true regardless of the
order of the system. In the complex s plane, horizontal lines represent
constant damped frequency, vertical lines represent a constant rate
of decay, <r; radial lines through the origin represent constant damping
ratio, and circles concentric about the origin represent constant
natural frequency, w„.
REFERENCES
a. s4 + 2s3 + 10s2 + 5s + 1 = 0.
208 ELEMENTS OF MODERN CONTROL THEORY
a. s4 + 2s3 + 8s2 + 4s + 3 = 0.
G(s)= - \- .
S(riS + 1)(t2S + 1)
1
a. G(s)
s(s + 4)'
b.
(s + 3)
G(s)
s(s + 5)
c. G(s)
s(s + l)(s + 2)
d. (s + 1)
G(s) =
s2(s2 + 6s + 25)
THEORY OF LINEAR SYSTEMS 209
e. G(s) = - --
(s + l)(s + 2)
For the cases a, b, and e, determine the gain and phase margins
when K = 10. In cases c, d, and f, determine the range of values of
K for which the system is stable.
G(s) =
s(s2 + 8s + 32)
Find that value of K for which the closed-loop pole has a relative
damping factor of 0.30. By making a sketch of the Nyquist curve,
determine the phase margin for this value of K.
1 + Tl = 0.
7. Show that the portion of the root locus which is off the real axis is
a hyperbola in the case of the unity feedback system for which
KG (a) = - -
s(s + ai)(s + a2)
K
G(s)
s(s + 1)
_ 1_
H(s)
(s + 2)
G(s) =
K
G(s)
s(s + 2)
and
H(s) = e~rDs.
KG (s) =
s(s + l)(s + 5)
Determine the value of K which gives a value of 0.5 for the relative
damping factor of the closed-loop pole. Using this value of K,
calculate the steady-state error in response to a unit step inpui.
Suppose this steady-state error must be reduced by a factor of at
least five, without materially altering the features of the closed-loop
response. Show how this can be done by using a lag circuit in
cascade.
13. In the previous problem, suppose that the value of K, which gives
the closed-loop pole a relative damping factor of 0.5, also results in
an unacceptably low value for the undamped natural frequency of
this closed-loop pole. It is required to double this undamped natural
frequency without materially changing the relative damping factor.
Design a suitable compensator for doing this.
K
K(Ks) =
s2(s + 5)
K
G(s) =
(s + 2)
a. 0_ -(- 3 0 T 4 0 -f~ 60 = 5 4“ 25
b. 1+ 50 + 20 + 100 = 35 + 45 + 6-
a + 2a = 5 -f- 3r/
0 + 40 + 50 = 25 + 4t? + r).
Xi = x2 + 2u
x2 = — 2xj — 3x2 + u
y = xi + 2u.
19. Calculate the transition matrix for the system of Problem 17, and
find the response to a step input assuming zero initial conditions.
21. For the discrete time system represented by Eqs. (173) and (174),
suppose that
~-l 0 0" "l 0
-1 -2 0 H = 1 1
-1 0 —3_ 1 0_
0 1 0" "0 (f
D =
-1 0 —3_ _1 4_
s + 1
Gi(s) = , G2(s)
s + 3 s —
f- 3
5(s)
b. parallel: = Gj(s) + G2(s).
5(s)
c. feedback: — = - —^ - .
S(s) 1 -f- Gi(s)G2(s)
23. For the system described by Eqs. (199) and (200) take
0 1 (f --T
A = 5 0 2 B = 1
— 2 0 —2_ _-l_
C = -2 1 0] D = 0.
1
Gx(s) = G2(s)
s + 1 s + 2
Li(s) = K o = constant.
a. absolute invariance.
b. steady state invariance.
c. conditional invariance.
214 ELEMENTS OF MODERN CONTROL THEORY
H(s) = K2(s)
Gu(s) = - Gi2(s) — 1
s
1
G2i(s) —1 G22(s)
S'
X = y
x = P(x, y)
(5)
y = Q(x, y).
dy = Q(x, y)
dx P(x, y)
*The knowledgeable reader will recognize this as the van der Pol equation, the workhorse of
nonlinear theory.
THEORY OF NONLINEAR SYSTEMS 219
m = 0
Figure 2.1. Graphical Construction of Phase Plane Trajectory Using the Isocline
Method.
let A be a point on the solution curve. (This may represent the initial
state x [o], y [o].) The slope at point A is equal to 1 since it lies on the
curve for m = 1. Point B on the system trajectory may be determined as
follows. We note first of all that the trajectory proceeds in a clockwise
direction, as may be readily verified by considering the relation between
x and y. Now since B is on the m = 0.5 curve, its slope equals 0.5, and
the average slope between A and B is, therefore, 0.75. Drawing a line of
slope = 0.75 through point A consequently determines point B. Pro¬
ceeding in this fashion, the entire trajectory is formed as shown in Fig.
2.1.
The accuracy of the method can be made as high as desired by using a
large number of isoclines (lines of constant slope).
Another method, which is somewhat more direct, proceeds by con¬
structing the system trajectory by means of small circular arcs. This
approach, known as the “Delta Method,” is applicable to equations of
220 ELEMENTS OF MODERN CONTROL THEORY
the form
x + V (x, x) = 0. (9)
x + 8 + x = 0. (10)
dy _ x + 3
(ID
dx y
where
5 = <p(x,y) - x. (12)
y2 + (x + 5)2 = R2 (13)
which thereby locates the center of the circular arc and determines the
radius R automatically. A short circular arc is then drawn to locate the
new point, xj and yx. The procedure is then repeated until the entire
trajectory is formed. Figure 2.3 illustrates the trajectory obtained for the
system
x + 25(1 + 0. fx2)x = 0
x = —1.8
x = y = —1.6.
In certain cases, the trajectory of the system in the phase plane may
be obtained by integrating Eq. (6) directly. For example, consider the
equation
ec = A, t > 0 (14)
= 0, t < 0
x — e — A
T = W0t
dy = _ 2fr + x (16)
dx y
x2 + y2 = c2 (17)
If f = 0 and the sign in front of the x term in Eq. (15) were minus
instead of plus, then Eq. (15) would integrate to
y2 — x2 = c2 (18)
or
x2 y2 — c2 (19)
/
x = y
y = vx (22)
_2f_ / y + fqjx \
<®(x,y) tan"
V\ - f2 \u)X\/l — fv
\i = — fw — wV f2 — l
\2 — fa; + w y/ f 2 — 1•
224 ELEMENTS OF MODERN CONTROL THEORY
From Eq. (21), observe that the slope changes sign aty = - x/2f. This
line is also shown in the figure.
226 ELEMENTS OF MODERN CONTROL THEORY
u = (coVi — b2)x
v = y + fwx
u2 + v2 = C'iev<
u = p cos 8
v = p sin 6
THEORY OF NONLINEAR SYSTEMS 227
l\ I < Ix21
Figure 2.8. Phase Portrait of the Damped Second Order System
x 4- 2£x + x = 0; £ = 1.4 .
and we obtain
P — C3ee
u = y — X2x
v = y - Xix
v = C4uXaA‘
Remark: The importance of the results derived in this section lies not
in their neatness and elegance but rather in the fact that they
are basic to the study of the piecewise linear systems considered
in Sect. 2.1.4. What is significant is that the phase portraits
are distinctive, depending upon whether the system is under¬
damped, overdamped, or undamped. These qualitative charac¬
teristics, taken in conjunction with properties related to singular
points and limit cycles, provide one of the most valuable tools
available for predicting the qualitative features of nonlinear
phenomena. These topics are taken up next.
P(x, y) = 0
(26)
Q(x, y) = o.
In a sufficiently small region about the singular point, the higher order
terms are negligible and the system behaves linearly; viz.
1. Xi and X2 are both real and negative. This case is termed a stable
node and all trajectories converge to the singular point, as shown
in Fig. 2.10a.
2. Xi and X2 are complex conjugate with negative real part. This
condition gives rise to a stable focus, with all trajectories con¬
verging to the singular point in the manner shown in Fig. 2.10b.
3. Xi and X2 are pure imaginary. For this case the motion is simple
harmonic, with the oscillation amplitude dependent on the initial
conditions. The trajectories are closed paths around the singular
point which in this case is termed a center. The phase plane
portrait is shown in Fig. 2.10c.
4. Xi and X2are both real and positive. This gives rise to an unstable
node with all trajectories diverging from the singular point, as
shown in Fig. 2.10d.
5. Xi and X2 are complex conjugate with positive real part. This case
results in oscillatory motion but with divergent amplitude; all tra¬
jectories leave the singular point, as shown in Fig. 2.10e. The
singular point here is termed an unstable focus.
230 ELEMENTS OF MODERN CONTROL THEORY
b. Stable Focus - The roots are complex e. Unstable Focus - Roots are complex
conjugates with negative conjugates with positive
real parts real parts
c. Center - The roots are complex conju- f. Saddle Point - The roots are both real
gates with zero real parts and of opposite sign
6. Aj and X2 are both real with one positive and one negative. The
singular point here is termed a saddle point. The trajectories are
depicted in Fig. 2.1 Of.
/u — aib2 — a2bi
v2 = 4/u.
SADDLE
POINT
2. 1.1. 2. Note that while the results obtained there were valid for the
whole phase plane, the case discussed here is limited to the immediate
vicinity of the singular point because of the linearizing process involved
in deriving Eqs. (27).
Theorem I: If outside a circle, Ci, in the phase plane all paths are
converging (the radial distance to the point moving along the path is
decreasing, the radial distance being measured to the center of the circle,
Ci), and inside a smaller circle, C2, with the same center as Ci, the
paths are diverging, then a limit cycle must exist within the region
bounded by Ci and C2.
/
Remark: Most systems of interest, which can be put in the form (5), may
be analyzed by invoking the concepts of singular points and
limit cycles, together with appropriate graphical methods. The
form of the phase portrait thus determined yields valuable
information relating to stability, points of equilibrium, bound¬
edness, or periodicity of the solution, etc. These phenomena
often cannot be predicted by linearized techniques alone. The
application of these ideas will be illustrated by two simple
examples.
x — e(l — x2)x + x = 0.
x = y = P(x, y)
dP dQ
= e(l - x2).
dx dy
By Theorem II, a limit cycle exists, since this quantity changes sign at
x = ±1. The phase portrait for this system is shown in Fig. 2.13. It is
worthy of note that this limit cycle is reached whatever the form of the
initial conditions. Linear systems do not display this phenomenon.
x + o)02 x — h2x3 = 0
* = y = P(x, y)
a. A center at x = y = 0.
b. A saddle at x = w0h,y = 0.
c. A saddle at x = — &>0h,y = 0.
time. This problem was discussed in Sect. 2.1 .1.2. Figure 2.10 depicted
the types of phase plane trajectories which characterize each type of
singular point. In Sect. 3.1.2, these trajectories are said to indicate the
form of the phase portrait in the vicinity of the singular point, since the
relevant equation was linearized about this point. However, if the system
is linear in the entire region, then these paths are valid in the entire
region. These considerations suggest that the phase plane be decomposed
into regions, in each of which the system is linear, then combining the
various paths in order to completely describe the motion. At all times,
the qualitative aspects of the system response will be kept in the fore¬
ground. The aim is to obtain the greatest possible degree of insight into
the problem with a minimum of labor. While exact numerical results may
be obtained with some additional effort, this is not our primary ob¬
jective at the moment. As will be seen, a wealth of information on
properties of the system motion will be derived with little computational
effort.
The approach will be described by applying the method to several
problems in the field of flight control. A few preliminary remarks are in
order. For any region in the plane where the system is linear, the motion
is governed by
x = e* - e
x = y = P
9C = A, t > 0
(35)
8 = B , 8 c > Scm
= 13 , 8C <C. 8cm %
Here
* K^(/iaA + a cB)
o l = -
Mar
** _ h«KaA (40)
O 2 — — -
(MoK^j — Ha)
*
8 3
Ma
The 8* are the singular points for their respective regions. Note
that a singular point need not be located inside its own region. When
240 ELEMENTS OF MODERN CONTROL THEORY
0 V'tyO— /Jic,8
8 = Ka(0c - K Re - f)
f = o - |0| > e0
O
O
2.18.
Figure
Stable
3(Aerod
Examp
for
Portrait
Phase
Case).
242 ELEMENTS OF MODERN CONTROL THEORY
b.
$* — AcK^(A T do)
(pcKa — na)
AcK^A
V2 = - -
Ma
The phase plane is again divided into,, three regions, as shown in Fig.
2.20. (9* is the critical point for Regions © and © while (9*is the critical
point for Region ©. When the quantity (ncKA — na) is positive and the
system gains are appropriately chosen, 6\ represents a stable focus andd%
is a saddle point. The trajectory starts at the origin (0 = 0= 0 initially)
and, on reaching the region boundary, takes a path corresponding to
the stable focus in Region ®. The critical point represents a point of
stable equilibrium. Note that in the present case, with positive step
input of magnitude A, 6 is never negative.
5 = KAdE - id
8 — na6 = nc5
Oe = 0e - d
f = Kakr |0®|< da
= o |0*| > do
Because of the fact that linear methods of analysis are so powerful and
universally applicable, it is natural to seek to “linearize” the nonlinear
systems encountered in control analysis. One well-established technique
involves the derivation of the (linear) equations characterizing small
perturbations about steady-state conditions. When the motions are not
small, this approach is invalid. The method of “describing functions,”
246 ELEMENTS OF MODERN CONTROL THEORY
where, in general, R„ is different from R„ and <pis the angle of phase lag.
The amplitude ratio, R0'/R0, and phase lag, <p,are functions of the input
frequency but not of the input amplitude. If the dynamic element is not
linear, then the response to a sinusoidal input is periodic but not sinusoi¬
dal. In seeking to define an equivalent transfer function for a nonlinear
THEORY OF NONLINEAR SYSTEMS 247
element, one is led to consider the frequency spectrum of the output and
to assume that only the first harmonic is significant.
Let the input to the nonlinear element be given by Eq. (41). Then the
output has the property
where
2
ra + T
/ 27rnt^ (45)
_ T /Ja y(t) cos\ T~ )) dt
^27rnt\ (46)
2
_ T % [ Ty(t)
sin( ) dt
~ 2xnt . „ 27rnt .
= Ln sin - cos in„ + Ln cos - - sin
T T
= Cn cos ipn
An = C* sin <pn.
If we put
a = -L
T = 2L
then we have
A„= i J y(t)cos dt
B„ = 0
and
An = 0
and
THEORY OF NONLINEAR SYSTEMS 249
we have
00
(52)
y(«t) = A + La: cos ncot/+ X sin ncot
^ n—1 n=l
where
1 ra^~27r
B„ = - / y(cot) [sin ncot]d (cot). (54)
7T *'«
If
y(wt) = y(-«t)
then
B, = 0
2 rir (55)
A„ = — / y(cot) [cos ncot] d (cot).
7r Jo
If
y( —cot) = - y(cot)
An = 0
(56)
Bn = — / y(cot) [sin ncot] d (cot).
7T Jn
where
Therefore
4 r*
4— I R0 (sin wt2 — sin wti) sin wt d(cot)
ir Jtt
2R„
ic
cot2 — o»ti -)- sin2cot2
2
sin2wtiJ 2
252 ELEMENTS OF MODERN CONTROL THEORY
and
Ax = 0.
Hence
wti = 0
so that
so that
or
y(«t) = — K R > 0
= K R < 0.
THEORY OF NONLINEAR SYSTEMS 253
and
Saturat
for
Functio
Zone.
Dead
Describ
2.24.
Figure
254 ELEMENTS OF MODERN CONTROL THEORY
R(t>
Hence
4 f 4K
B, = - / (—K)sin«td (wt)= - —
7T " o IT
Ai = 0
so that
Bt 4K (62)
R/0 7TR()
THEORY OF NONLINEAR SYSTEMS 255
A. = 0
2 fn
Bi = — / (R0 sin cot)" sin cot d (cot)
7T Jo
^ J d —
2R0”~
-
1r( 22) 7=- - 7- '
(63)
VTr(£+3)
Relay with Dead Zone and Hysteresis. The essential elements of this
nonlinearity are depicted in Fig. 2.27. Here we find that
277— 01/3
K K f
= —- / cos cot d (cot) — — I cos cot d (cot)
77 •'oia 77 77+coa
2K . . .
= —— (sin coa — sin cop)
r 77-COp 1-2T7-COS
B, = — / sin cot d (cot) — — / sin cot d (cot)
77 77 TT+GIOC
2K
(cos coa + cos co/3)
where
A T h
sin coa
2R
THEORY OF NONLINEAR SYSTEMS 257
a. NONUNEAR ELEMENT
A —h
sin oj/3
2R0 '
Now
Ci = VAi2 + B)2
258 ELEMENTS OF MODERN CONTROL THEORY
and
Ai
f>\ tan-1
Bi
Putting
9 = coj3
we find
(64)
a. No hysteresis (h = 0)
(65)
Go =
b. No dead zone (A = 0)
4K (66)
Go / ~ *
ttR„
8 =
J3|<j
00 (M «X>
Fun
De
Re
2.2
An
Fig
Pha
O 00 (M CO O t^QOCMCOO t*<
• I t-* •-* (Si 03 CO ri* t)*
I I I I I I I 1 I
(Sap) ch
THEORY OF NONLINEAR SYSTEMS 261
-4b -44 -40 -36 -32 -28 -24 -20 -16 -12 -8 -4 0
<P1
(deg)
Figure 2.30. Relay Describing Function, Amplitude versus Phase.
3ir
= R0 sin cut + 7T — cutl < Cut <C —
-
< cut < 2ir — cotl
where
262 ELEMENTS OF MODERN CONTROL THEORY
RP — a
sin wti -
Ro
Therefore
Ai
Ro j"n~utl a i
H- - cos cot d (wt) - / cos cot d (cot)
* J 27rKn
THEORY OF NONLINEAR SYSTEMS 263
Rc r3*h . a
+ — / sin wt cos wt d (cot) -
7r TT
—OJf 2?r
2TT
ITT- — COt
Ro 1 a
7T
(
J3t/2
cos cot d (cot) H-
27T
R0 r2* . a
d- / sin cot cos cot d (cot) cos cot d (cot)
7T •'27T— 2tt
1 wtl
Rq rvli a r7^2.
Bi = — 1 / sin2 cot d (cot) — — / sin cot d (cot)
i ' 2tt j0
TT —
^ 1,— ^ ^
—- / sin cot d (cot) — — / sin cot d (cot)
,r t{2 27r "r/2
R »3ir/2 a r3ir/2
d- - sin2 cot d (cot) ~h~ sin cot d (cot)
7T " TT-COf 27T JjT —COt
1 1
£ 27T-CO*! a 2TT
—OH1
- - / sin cot d (cot) + — / sin cot d (cot)
TT •'2t/9 2tt •'o-./o
R r27r a f27r
-f- —:- / sin2 cot d (cot) - / sin cot d (cot)
7T 007T—OH 2ir *2Tr — Ci)t
=^ [5+0,11
+1~(r0)cos
a,ti]
TT — COt
27 r
a
R0 sin cot — d (cot) = 0
2.
2 TT— Olti
and we have
VaTTb? -1Ai
Gd tan
R„ Bj
1/2
2 2
— —2 ) —|—( — —
|—coti -(- sin coti cos coti ) tan_1/3
rT2 R0 / \2 /
(68)
where
ii~2)
p = - .
R°^ +coti
+sincoti
coscoti
^
+ General Describing Function for Piecewise Linear Elements. Most
of the describing functions derived thus far are special cases of a more
general form (10) shown in Fig. 2.32. Applying Eqs. (53), (54), and
(58) to this case, we find
R0 .
+ — (sin 20i — 2 sin 202 — sin 203 + sin 204 + sin 206)
2
+ ^-[Ro(ir— 202) +
7T IV 0
R0 sin 202 — 4k3 cos 02]
and
THEORY OF NONLINEAR SYSTEMS 265
2.32.
Figure
Nonlin
Linear
Piecew
Genera
A
266 ELEMENTS OF MODERN CONTROL THEORY
n2 R
° (— cos 203 -f cos 206) + 2k2 (sin 05 — sin 03)
ttR0 L 2
where
Dc — Fe
k2 =
D - F
k3
n3
„ . , a . , b _ . _. c
0i = sin-1 — 02 = sin 1— 03 = sin 1 —
R0 R0 Ro
d g
04 = sin-1 — 06 = sin-1 — .
R0 R0
Note that the 0f ’s have been defined such that they all lie in the first
quadrant. A variety of special cases may be deduced from Eqs. (69) and
(70). An extensive tabulation of these is given in Table 2.1., which is
adapted from Ref. 10.
(71)
THEORY OF NONLINEAR SYSTEMS 267
Remark: All of the describing functions encountered thus far are inde¬
pendent of frequency. Occasionally, a situation arises in which
the describing function is dependent on input frequency. This
may occur, for example, when there is a velocity saturation.
The analysis in this case is extremely complex and the use of a
computer is mandatory. Various problems of this type are
discussed by Kochenburger.(U)
x = R0 sin cot.
be o
| i
£%
a a
S
§ B
s 3
Characteristic Describing function
g(Ro>= o
[*% Ro(262+e3' V +~T (- 2 8ln262
-C08V]
+iilo[Ro(e5-e3)
+^(8ln2e3
- sin 2 0 ) + 2k (cos 0 - cos 0 )
5 2 5 3
+~ 1r ( it - 2 0 ) + R sin 2 0„
ttR [ o 2 o 2
o
*4k300862]
"l f Ro
b(Ro)
=^ [~f (C0S
63' C°82V +28(8l“®3
o
'"‘“VI+ik [~T(-cos
20g o
d
b£> O
$ 2
£ 8
a a
§ «
a 3
Characteristic Describing function
g(Ro>
=^r[ V2 Ve3-e5) +!^(-28ln2e2
o *
- COS
0g) o(e5-e3)+T-(8ln2e3
4M .
—— cos 0„
irR 2
o
b(R0>
= 1~r (coe
2e3•cos
o
2e5)
-2a(sin03+sin©J +^ [-f- (-cos20g
3 o
+ cos 2 0 ) + 2k (sin 0 - sin 0 )
oc 5 o
THEORY OF NONLINEAR SYSTEMS 275
g(Ro>
- 7r (2V- VV+V81”26! o
15
- sin 2 02)+4a (cos8^ - cos S^J
n3 r
+ —— R ( ir - 2 0J + R sin 2 0„
irR l o 2' o 2
- 4k cos 0„
3 2
b(R ) = 0
o
g(Ro>
= I2V-V«2> + Ro(8in2ei
o
“i/i M - sin 2 02>+4a (cos8^ - cos 0^)j
16
ab
4M
^ —r— cos 0
»rR 2
o
b(R ) = 0
o
THEORY OF NONLINEAR SYSTEMS 279
GH
K
KGH =
Im
Re
K GH K GH
2
*The simplified Nyquist criterion is valid when all the poles of the open-loop transfer function
he in the left-hand plane.
282 ELEMENTS OF MODERN CONTROL THEORY
Figure 2.35. Nyquist Curve and Describing Function with Memory— Limit
Cycle Shown.
The analyses of the previous sections have been based on the assump¬
tion that there is only one nonlinearity in the system. For multiple
nonlinearities, various extensions of the describing function technique
284 ELEMENTS OF MODERN CONTROL THEORY
Im
have been described in the literature. (13-15) Perhaps the most straight¬
forward approach is to develop a composite describing function when
the nonlinear elements are in series. This problem has been studied by
Gronnerd15) If the nonlinearities are separated by linear dynamic
elements, the analysis is much more difficult. Even for relatively simple
systems some sort of computer assistance is required.
While a detailed computational approach is tedious, the broad outlines
of the method of analysis are relatively simple and afford a degree of
THEORY OF NONLINEAR SYSTEMS 285
x = R0 sin cot.
R0|G,(jco)[
|G2(jco)|GD(1)(R0,
«) sin (cot+ Z9l + ZZA
and this is used to determine the describing function for the second
nonlinearity, GD(2)(R0,w). This procedure is continued until all the
describing functions are formulated. Although these are, in general,
quite complex, they depend only on R0 and co. For either form, a set of
values of R0 and co will determine a linear system whose operating gain
and pole-zero locations are determined for these values of R0 and co.
The characteristic equation for the entire system is now written as
II [s + Zi(R0, co)]
1 + K0(R0> co) — - =0. (72)
M
XT Is T Pi(Ro>w)]
t=l
a. Select a value of R0 and co (say R0(1) and coi). Determine the root
locus and operating gain, K0(1).
b. Select a new value of R„ (say R^2)),and with the same value of coi.
A new pole-zero configuration and a new root locus result, along
with a new value for operating gain, K0(2).
c. On a separate diagram, connect all the operating points, Ko(0, re¬
sulting in an “input-dependent locus.” The limit cycle amplitude and
THEORY OF NONLINEAR SYSTEMS 287
The above procedure simplifies materially for the case of Fig. 2.38, since
the pole-zero locations and operating gains do not depend on frequency.
In any event, the input-dependent locus generated by this method may
be interpreted as follows.
The proof of the above assertions follows the conventional arguments for
root locus stability. As an example of the above approach, we consider
the following.
K0(R0)
(s + l)2(s + l)
where
}= 100Gp(1)(Ro)
10 G d(2)(R0) + 1
1 = 100 G g(2)(R0) + 1
" 10 G d(2)(R0) + 1
4Mj
GD'2>(R0)
ttRq
b = 1
Mi = 1000
M2 = 10.
In other words, the output from the nonlinearity is a different d-c bias
plus a sinusoid of different amplitude and with some phase angle, <pi,
with respect to the input. The quantity, w0>represents the limit cycle
frequency. We now define the following.
Ri <2
Rn 3
(77)
1
COo 3
where ws is the frequency of the input signal, x(t), and a>0is the limit
cycle frequency. A thorough discussion of these conditions is contained
in Gelb and Vander Velde. (16)
The requirement imposed on the frequency ratio stems from the as¬
sumption that the error (between input and output) be approximated by
a d-c bias over one limit cycle period. The requirement on amplitude
ratio is necessary to ensure that higher harmonics in the DIDF approxi¬
mation be negligible compared to the fundamental. Gelb and Vander
Velde show that when conditions, Eq. (77), are satisfied, the approxi¬
mate DIDF is within 5 percent error referred to the untruncated
expression for the DIDF.
The above ideas can be clarified by considering a particular case.
R sin cot
o
x = Ri + R0 sin &>0t.
-
l
irRi
4D
N0
7rR0
2D
Ns
ttRo
294 ELEMENTS OF MODERN CONTROL THEORY
This curve crosses the imaginary axis, {a = 0), when u>= on, which means
that the limit cycle frequency, o>0= an.
Now by substituting s = janin
1+ N0G(S) = 0
Im
we find
N0 — 2£i coi.
7r|l 0)^
N. « ^ &on.
2
In order to determine the value of the closed-loop pole on the real axis,
we put
S = a
N s = |l wi
1 + Ns G(s) = 0.
Time
Time
The DIDF for some of the more common types of nonlinearities are
listed below.
-6
1_
6
1 1 -D
c. Hysteresis
C-^)
n0=^E k
c*n!R0"-*-
(n — k) ! k !
*Ri* V
C-^)
In (^1
\ 2
N.= Ctn!-TRon_*Rifc_1
k (n — k) !k!
odd
Also
AT = transpose of matrix A
A-1 = inverse of matrix A
a •b = scalar product of vectors
Va = gradient of a
C) = d/dt( ).
x = Ax + f(t) (78)
Definition 2: The variables, xi, x2,. ,xn, which are the components
of the state vector, x, are referred to as the state variables, since at any
given instant of time they represent the condition or state of the system.
X = g(x, t) + (79)
and the linear system, given by Eq. (78), are said to be forced if f and f
are not zero.
Definition 5: The systems described by Eqs. (78) and (79) are said to be
free if f and f are zero.
Definition 6: The systems described by Eqs. (78) and (79) are said to be
stationary if the elements of A and g are not functions of the independent
variable, t; otherwise, they are called nonstationary.
2. ||x|| = J2 |x,-|.
i
n = xrBx
is called a quadratic form in the state variables xi, x2, . , x». The
n X n matrix, B, is symmetric.
Definition 10: The quadratic form, n, will be called positive definite if,
for all x X 0 and such that m = 0 for x = 0
n > 0
n < 0 .
a ^ 0
m <0 .
xrBx >0
is satisfied. Similar definitions hold for negative definite and positive and
negative semidefinite matrices.
302 ELEMENTS OF MODERN CONTROL THEORY
Example (a):
Example (b):
Example (c):
Example (d):
x = f(x, t) (80)
f(a, t) = 0 , t ;> 0
x = f(x) (81)
THEORY OF NONLINEAR SYSTEMS
303
x(o) = 0. (82)
Let be a region in the state space for which ||x|| < a, and letR(2)
be a similar region for which ||x|| < p. We may envision these as hyper¬
spheres centered at the origin with radii a and p, respectively. It is
assumed that a > p. Further, let the initial state of the system at time
t = 0 be x = b ^ 0. We say that the system is:
x = f(x) (83)
x(o) = 0. (84)
1. All the first partial derivatives of V(x), with respect to the components,
Xi, of x, exist and are continuous for all x.
V(o) = 0
V(x) <0x^0.
V(o) = 0
V(x) <0x^0.
THEORY OF NONLINEAR SYSTEMS 305
6. V(x) > 0.
The following theorems refer to the system described in Eqs. (83) and
(84) and the properties of an appropriate Lyapunov function.
Theorem VI: Let R(r) denote the region where V(x) < r. Within
this region, assume that V(x) satisfies 1 and 2 and that the region R(r) is
bounded. Then any solution of Eq. (83) which starts in R(r) is:
Example 8:
Xi = x2 — atXi(xx2 + x22)
X2 = - X! - ax2 (xx2 + x22)
V = Xl2+ x22
then
V = — 2a (Xi2 + X22)2
V —» * as | |x| | —» oo.
Example 9 :
Xi = x2
x2 = — a(l + x2)2x2 — Xi
a = positive constant.
THEORY OF NONLINEAR SYSTEMS 307
Again take
V = Xj2 + x22.
It follows that
V = - 2a(l + x2)2x22
/
which is negative semidefinite.* Also
2
= — «(1 + X2) _ 5l
dX;
But for x2 = 0 this slope is infinite, which means that the xi axis cannot be
a trajectory of the system. Furthermore, forx2 =— lwe have
which is never zero except at the origin. Therefore, the line x2 =— 1 can¬
not be a system trajectory.
Since we have now fulfilled the conditions of Theorem IV, we conclude
that the origin of the given system is asymptotically stable in the large.
Example 10:
Xi = x2
Again taking
V = xi2 + x22
we find
V = - 2x22 (1 - |x x| ) .
Consider now the region in the state space which is bounded by the
curve
xi2 + x22 = 1.
Example 1 1:
xi + Xi T x i3 == 0 .
We write this as
Xj = x2
X2 = — X2 — Xj3.
1 , 1 ,
V = iX3 +2X2
then there follows
V = — x22
« + g(e)« + h(e) = 0.
(
<p(o) = iZ'(o) = 0.
Xi = x2 — ^(Xi)
x2 = - h(xi).
V = i x22 + vKxx).
It follows that
V = - h(xi)^(xj).
We now suppose that there exist two positive constants, a and f), such that
h(x1)<p(xi)
> 0 for |xi| < a, xx ^ 0
iZ'(xi)< /3 implies|xi| - a.
If these two relations are satisfied, then the region defined by V < /3 is
bounded, and within this region V < 0 . Furthermore, V = 0 in this region
only when x, =0. But the x2 axis is not a trajectory of the system since
the slope
dx2 _ _ h(xj)
dxi x2 — <p(xi)
310 ELEMENTS OF MODERN CONTROL THEORY
e + M (1 - e2) * + « = 0
Xi = X2 — <p(Xi)
x2 = — h(xj)
with
<p(xi) = —
^(Xl)
=^-Xi2.
Taking
V= ~x22
+ iA(xi)
=\ (X22
+ Xj2)
we find
V = )
which is negative semidefinite for |xi| < \/3. Consequently, we take
a = V3. Now if iA(xi) < 0 is to imply that |xi| < V%, we must have
0 = 3/2. It follows, therefore, that the region bounded by x,2 + x22 = 3
is a region of asymptotic stability.
THEORY OF NONLINEAR SYSTEMS 311
<7
(85)
*>(o) = 0.
a 0 «J) 6
&f - 1 N.L. G(s)
i= 0
G(s) = - (86)
<P n
the denominator is unity. We also add the restriction that the degree of
the numerator is at least one less than that of the denominator. In
summary, it is stipulated that
«0 = 10„=O. (87)
x = Ax + b(p(<r) (88)
cr = — Xi
d = xi
Xn y] z-t-
1 Xi -j- bn^3 (90)
i=l
bQ = po = 0
bi = Pi — X ai-jbj (91)
i=o
i = 1, 2. . , n.
G(s) = - = - + g2
<P S2 + aiS + a2
bi = (h — axb0 = ft
b2 = /32 — a2b0 — aibi = /32 — ai/3i.
Xi — x2 +
Xi 0 1 Xi bi
=
+
_*2_ _ — a2 — «i_ _x2__ Xi 1—
0 — xi.
Notice that in the format of Eq. (88), the numerator dynamics (deriva¬
tives of the driving function) are effectively eliminated. In the analysis
to follow, we will consider a slightly generalized version of Eq. (88),
namely
x = Ax + h(p{<j) (92)
T
(J — V X
x = Ax + b<p(<r)
\ = <p{o) (93)
a = vrx + y£.
a%= vTx + 7£
x = Ax -f- b<^>
T i(p
a = vTx
7 = vrf
314 ELEMENTS OF MODERN CONTROL THEORY
or — VrX
x = Ty (94)
where
1 1 . 1
Xi X2 . X„
^l2 ^22 . X„2
T = : : : (95)
n — 1 1 n — 1
X1 x2a~ \
n
This matrix is known as the Vandermonde matrix; the X, are the eigen¬
values of A. Asa result of this operation, we find
y = Fy + ni*>(<r)
(96)
T
<T = C y
THEORY OF NONLINEAR SYSTEMS 315
where
£ O
F = ^2 = T_1AT (97)
.0
m = T-!b (98)
/
c = Tr V. (99)
cr
j
V = yrPy + /
= <p(a)da. (100)
Taking the derivative with respect to time using Leibnitz’s formula, and
making use of Eq. (96), with a = cry, one obtains
where
V > 0
V < 0
If P is positive definite, and if Eq. (85) holds, the first and third conditions
are automatically satisfied. To satisfy the second condition, Q must be
*P is a constant symmetric matrix whose elements are, for the moment, undetermined.
316 ELEMENTS OF MODERN CONTROL THEORY
Q = aar. (103)
r = — crm (104)
and substituting Eq. (103) into Eq. (101), and then completing the
square, results in
P = f eFT‘QeFldt.
Jo
(107)
Now
3Xlt
e* = 3X2t
0
eFTtqeFt
THEORY OF NONLINEAR SYSTEMS 317
IS
(X i+\ j)t
Furthermore
(X i+XyH dt 3j i&j
a,aJ
' (V + Xy)
P = _ a<a7 (108)
(Xj + Xy)
(s + 4)
G(s)
(s + l)(s + 2) <p
s + 4
s2 + 3s + 2
x = Ax + b<p
where
A =
b =
318 ELEMENTS OF MODERN CONTROL THEORY
and
a — vTx = — xi = —&
where
|A - AI| = 0.
However, in the present case, these eigenvalues are the poles of the
open-loop transfer function, G(s), viz.
Ai = -1
A2 = — 2.
-1
we then have
3
m
-2
THEORY OF NONLINEAR SYSTEMS 319
1
c =
from Eqs. (98) and (99), respectively. Furthermore, via Eq. (104), we find
r = — c7'm = 1
and
6a !2 4aia2
P
12
4aia2 3a 22_
ai2 -f - (2 — 3a2)a! + I =0
3 3
This is indeed the case, since we find that ai = -0.98 and a2 = -1 .41 .
For higher-order systems, it becomes increasingly more awkward to
solve Eq. (106) for the vector a. For an nth order system, the n com¬
ponents of the vector, a, are to be obtained by solving n simultaneous
algebraic equations. This presents serious difficulties at the computa¬
tional level. An approach offering some advantages in this respect is the
following.
We assume that
7i
Q = (109)
0
Here the t.-’s are positive constants. Substituting this expression for Q in
Eq. (101), and completing the square, yields
where
(111)
u = — (Pm + - Fc)
r ^ ur Q-1 u . (112)
X = Ax -f bip(a)
a = VrX
where
0 1 0
A 0 0 1
-30 -31 -10
b =
-1
v -1
0
THEORY OF NONLINEAR SYSTEMS 321
Xi = -2
X2 = -3
X3 = -5
so that
1 1 1
T = -2 -3 -5
4 9 25
30 16 2
1
T"1 = -30 -21 -3
6
6 5 1
and
8
1
m = T lb = -9
6
1
r = — crm = 8.
Taking Q as
0
72
7s.
where the y-s are (as yet undetermined) positive constants, we solve for
322 ELEMENTS OF MODERN CONTROL THEORY
0
l 7i
~2Xl
P =
1
~ 2 X2
The components of the vector, u, are now obtained from Eq. (11 1);
viz.
1
di = ~ (mt- Ti + X,)2
i = 1, 2, 3
71 = 1
72 = 6
73 = 50 .
<p = ktr
a. The root locus of the linear part of the system is not confined to the
left half of the s plane.
b. There are open-loop poles at the origin of the s plane.
c. The open-loop function has multiple poles.
d. The constant, r, in Eq. ( 106) or Eq. ( 1 12) is negative.
324 ELEMENTS OF MODERN CONTROL THEORY
Im
Im
a. Analyze a system whose open-loop gain does not fall below a certain
minimum value.
b. Separate multiple open-loop poles.
c. Analyze systems with open-loop poles in the right-half plane.
/
which obviously satisfies Eq. (85). The revised block diagram is shown
in Fig. 2.54. If we now deal with the new nonlinear parameter, <p', de¬
fined by Eq. (114) and the new linear transfer function
G(s) (116)
G(s) ~ 1 + aG(s)
N(s)
G(s) =
D(s)
-G(s) _~ D(s)
N(s)
+ <*N(s)'
a (118)
trV > o .
V = yrPy
yO <s)
I - 1
V = - yrQy + 2ymTPy
V = — (ary)2 + <p(2mrP) y .
mTP = 0 . (119)
dV dxj
where forx,-, we insert the system equations. Now this may be expressed
as
V = (VV)rx . (120)
THEORY OF NONLINEAR SYSTEMS
329
rX fX
V = J/0 (VV)Tdx
= J0/ VV-dx. (121)
This is a line integral to an arbitrary jaoint in the state space and is inde¬
pendent of the path of integration. The simplest such path is the fol¬
lowing.
(122)
n
In this expression, V,-( ) stands for the ith component of the vector, VV.
To obtain a unique scalar function, V, by a line integration of a vector
function, VV, it is necessary that(3°)
V X VV = 0 . (123)
Vi = Oil Xx 4* 0,2 X2 + ~b a in X*
330 ELEMENTS OF MODERN CONTROL THEORY
contain terms that have more than one state variable as factors. Evidently
such terms may be determined from terms such as ai3 (x) x;; therefore,
the a Unix) may be taken without loss of generality as aiill{xl). Further¬
more, for V to be positive definite in the neighborhood of the origin, aiik
must always be positive. Also, if the relation
<p = <r\p(a)
It will be shown here that this requirement is, in fact, unnecessary. The
differing results stem from the fact that the equations of Schultz and
Gibson contain derivatives in the forcing function, which may be elimi¬
nated by making the transformation of variables described in Sect. 2.3.3.
Proceeding in this way, we write the state equations as
xi = x2 — h(xO xi
x2 = — x2 — (0 — 1) h (xj.) X!
where
6 — Xi = — a
h(xi) - - \K-xi) %
auXi + ai2x2
VV
c*2iXi T 2x2
where we have arbitrarily set a22 = 2. Experience has shown that this
particular selection for a »* is appropriate for most systems of practical
interest. From Eq. ( 120), we find
or
a 12 = <*21
If we put
<212 = 0
and
an = 203 - 1) * to
'2*to(l8 - l)xi'
VV =
2x2
0 > 1
h(xi)xi > 0 .
P > 1
THEORY OF NONLINEAR SYSTEMS 333
and
h(xi)xi > 0 .
x = f (x)
(126)
f(o) = 0 .
V = xrQx (127)
where J(x) is the Jacobian matrix of f(x). Substituting this in Eq. (128),
we find
V= f xr[Jr(ax)Q
+ QJ(ax)]xda. (130)
Therefore, by Theorem III of Sect. 2.3.2, we may conclude that the
system (126) is asymptotically stable in the large if for some positive
definite matrix Q, the matrix [JT (x)Q + QJ(x)] is negative definite for
all x ^ 0.
As usual, the selection of a suitable matrix, Q, for the problem at hand
is the core of the matter, requiring a judicious blend of experience and
ingenuity.
We conclude our discussion of the Lyapunov theory by taking note of
some deceptive facets of the stability problem for linear nonstationary
systems. It is common practice, in the design of autopilots for launch
vehicles, to take so-called “time slices”; in effect, a linear, time varying
system is analyzed by assuming that it is valid to consider the system as
non-time varying during preselected small time intervals. Essentially, the
assumption may be stated as follows. If, in the system
x = A(t)x (131)
the eigenvalues of A(t) have negative real parts for all t, then the system
may be presumed stable. That this premise is not true in general may be
demonstrated by the following counter-example due to Zubov/35) For
the system described by
The eigenvalues are — 1 and —10 for all t. Yet the solution is
THEORY OF NONLINEAR SYCTEMS 335
V = xr C(t)x . (132)
To prove stability, one must select a positive definite matrix C(t) such
that V, as given by Eq. (133), is negative definite. In certain simple
cases, some useful results may be obtained as follows. Let the system of
Eq. (131) be of second order with the matrix, A(t), given by
0 1
A(t) =
ai a*.
where ai and a2 are functions to t. Assume that the matrix C(t)has the
form
1 0
C(t) =
0 c22(t)
1 aic22 — 0 (134)
336 ELEMENTS OF MODERN CONTROL THEORY
and
1 (136)
C22 = - %
ai
2a2 ai (137a)
— + — 2 < 0.
ai ai
Furthermore, the requirement that C22(t) be positive for all t means that
(137b)
— > 0 .
ai
We will now show that V f 0 along any solution of the given system
other than the origin. In fact
(2a2 ai\
- — + — ) x2‘
ax a /I
which is zero only along the xi axis. But the slope of the system tra¬
jectory is
dx2
— ai - i aj
dx! x2
1 (139)
- — > 0
THEORY OF NONLINEAR SYSTEMS 337
and
All of the quantities Tc, La, I, 4, and 4 are known functions of time.
From Fig. 2.59, we find (writing xi = 0)
Xi — X2
x2 = ai (t) xi + a2 (t) x2
where
2K AK R)Xc>
(KaHc — Ha)
The first of these is the usual condition derived from linear stationary
analysis. The second condition displays quantitatively the destabilizing
influence of rapid increase in/i«.
THEORY OF NONLINEAR SYSTEMS 339
TIME
DELAY LOAD
f #3)
(such as lead filter, time delay, dead zone, and hysteresis) affect the total
performance will be made evident.
The simplest form of the system of Fig. 2.60 is with no lead filter or
time delay and with no dead zone or hysteresis in the relay. In other
words, Ti = r = h = ju = 0. We will also assume that there is no input,
i.e., dc = 0. The system equation then takes the form*
Ie + M sgn e = 0 (HI)
and the initial conditions are
«(o) = fQ
e(o) = <r0.
M (142)
e
I 1
M (144)
e - a2 = - y (t - ai)
where
* sgn t = 1 if e > 0
= 1 if e < 0 .
THEORY OF NONLINEAR SYSTEMS 341
2M
<2= ~ ~Y (* “ a^- (145)
M
e = 7-
2M
(e-a'2) (146)
where
If the relay has a dead zone of magnitude 2n, then the trajectory, Eq.
(145), is valid as long as /3 > ^. When —n<p<n,M = 0, and the tra¬
jectory in the phase plane is described by « = h, where a is the value of
6 at Point 0 in Fig. 2.63. Since at this time e = ht + eh it is inferred that
e increases for positive h and vice versa. There is now no difficulty in
completing the phase plane representation for this case, as shown in Fig.
2.63. Again the magnitude of the limit cycle depends on the initial
conditions.
ELEMENTS OF MODERN CONTROL THEORY
342
Ti e + « = 0 (147)
after which the motion proceeds per Eq. (146). In this way, the phase
plane trajectory of Fig. 2.64 is constructed. When the trajectory reaches
Point CD in this diagram, an ambiguity arises. The motion tends to con¬
tinue in that portion of the phase plane to the right of the switching line.
However, the tangent to the trajectory parabola has a slope whose mag¬
nitude is less than that of the switching line. The parabola representing
motion to the right of the switching line appears to the left of the
switching line. This is shown by the dotted portion in Fig. 2.64. There
is apparently no way to describe the motion beyond this point (which is
obviously not a point of equilibrium). This paradox is, in fact, a result
of the assumption of a perfect relay, where it is implicitly assumed that
changes of state may take place at infinite speed. By introducing a small
time lag, it may be shown tnat the trajectory converges to the origin
«i — e(ti) (148)
(150)
M
(151)
(152)
The positive sign is used in front of the radical since tx is positive. Now
the time at which actual switching occurs is given by
t2 — ti T T. (153)
We write
«2 — *(^2)
«2 — *(^2) •
M
~ ~ 21 ^ r^2 c° '
/
( Mr2\
1 Y*+Tr) / 2i Y/2 (154)
M(Ta- t) + 11“ V1 + M to) %
From Eq. (151)
«2 — — — (ti + r)
M / 21 \1/2 “I
I (Tl+ m'")+TJ-
—Ti + ( T x2+
1 Mr (155)
«2 = «2+ OT (2Tl — t)
(Tj - r) [_' 1 21
_ i_ Mr (156)
«2 62 - (2Tl
(Ti - t)
Figure 2.65 depicts the character of the phase plane trajectory for this
case. We have indicated by U and the delayed switching lines corres¬
ponding to Eqs. (155) and (156), respectively. The case shown is typical
for the condition, Tx^> r. When r is on the same order of magnitude as
Ti, a limit cycle of the form shown in Fig. 2.66 is produced.
ELEMENTS OF MODERN CONTROL THEORY
346
DELAYED
SWITCHING
The characteristics of the relay were shown in Fig. 2.61 , and we assume
that the lead time constant is finite. For the present we assume that t = 0.
THEORY OF NONLINEAR SYSTEMS 347
Let (Ti € + e) be positive and decreasing. Then the motion follows Eq.
(144). When
Tlt + 6 = M (157)
Tie + e = - h (158)
For many years, most of the research studies in the area of nonlinear
controls have been concerned with extending the basic ideas of the
Figure 2.67. Phase Portrait for Relay with Dead Zone and Hysteresis.
ELEMENTS OF MODERN CONTROL THEORY
348
Delayed
switching
lines
Figure 2.68. Construction of Phase Portrait for Relay with Dead Zone,
Hysteresis, and Pure Time Lag.
Slope =
Figure 2.69. Final Limit Cycle Resulting From Construction of Fig. 2.68.
X = Ax -f b<p{(r) (159)
cr = VrX (160)
(161)
a
or
G<*>
- -^ - vr<ls
- A)_‘b- (162)
350 ELEMENTS OF MODERN CONTROL THEORY
Theorem: If, in the system described by Eqs. (159) and (160), the
following conditions are satisfied:
P(s)
G(s)
Q(s)
we have
_ Pl(<«))
+ jP2(to)
(JWj ~ Qi(co) + jQ2(co)
where the subscripts 1 and 2 are used to denote real and imaginary parts,
respectively. This reduces to
PiQi + P2Q2
(165)
1_ Ql2+Q22
P2Q1 — P1Q2
(166)
2= Qi2+ Q22 '
0 1"
A =
-10 -7
b =
v =
—<x(s) s + 3
G(s) = *>(s) = s2 + 7s + 10’
limit on k in
(T
which ensures complete stability of the system. From Eqs. (165) and
(166) we find
4a>2 + 30
Gl = a,4 + 29o2 + 100
—co(w2+11)
°2 = «4 + 29+ + 100
we see that this inequality is satisfied for all w > 0, for any nonnegative
q, and 0 < k < °o. Consequently, this system is completely stable as long
as the nonlinear function is restricted to the first and third quadrants.
1 0
A = 0 1
-29 -10
0
b = 0
1
v =
For third and higher order systems, the matrix inversion to be performed
in Eq. (162) becomes extremely awkward, and for all practical purposes
impossible, for only moderately high order systems. The essential dif¬
ficulty stems from the need to carry polynomials in s through the maze
THEORY OF NONLINEAR SYSTEMS 353
x = Ty
X2 = —4
X3 = 5
1 1 1
T = -1 -4 -5
1 16 25_
~-20 -9 -1
T-1 = - — 20 24 4
12
-12 -15 -3
and we find
y = Fy + rrup
a = cry
where
m = T-1 b
1
c = Trv = 1
1
_ _ 1_
(s + l)(s + 4) (s + 5)
U = Gi(u) (168)
W = - coG2(co). (169)
U + qW + i > 0
(170)
q ;> 0, k > 0 .
U + qW + - = 0 (171)
k
then it is easy to show that Eq. (170) represents those points in the U-W
plane which are to the right of the line, Eq. (171).
The procedure now consists of making a plot of G(jco) in the U-W plane
for 0 ^ w < oo. Popov<36) calls this the “modified phase amplitude char¬
acteristic” (MPAC). The line, Eq. (171), is now fitted to this curve such
that the curve remains wholly to the right of the line while the abscissa
356 ELEMENTS OF MODERN CONTROL THEORY
— 1/k is made as small as possible. Figure 2.73 illustrates the method for
a typical G(j«) locus.
It is apparent that the usefulness of the method is not seriously com¬
promised by taking systems of arbitrarily high order. The limiting factor
is computer time and storage, much the same as for linear system
analysis.
In the remaining sections devoted to the Popov method, we will relax
various restrictions and discuss some generalizations.
(172)
a
In all the systems considered thus far, there were no external forces
applied. As a matter of fact, there are very few results available on the
* Essentially this means that the slope of the nonlinearity is not permitted to be zero at the
origin in the a — ip plane.
358 ELEMENTS OF MODERN CONTROL THEORY
a. The applied force f(t) is bounded; that is, lim f(t) < <*>.
t - %º00
G(j«)
(173)
1 + rG(jo;)
^ d<p(a)
r + e * “ckT - e (174)
Note that the poles of G(s) are not restricted to lie in the left-half plane.
This constitutes a significant generalization over the results given in the
previous two sections.
In the case where all the poles ofG(s)do lie in the left-half plane, then
we may take r = 0. A further notable difference between the forced and
unforced systems is that the stability criteria for the former involve the
derivative of the nonlinear element rather than the function itself.
For purposes of practical application, the condition in Eq. (173) may
be manipulated to a form in which frequency response techniques may
be employed in a familiar manner. Defining a constant, B = k/r, we may
write Eq. (173) as
k G(jq>)
+ ^ 0
B + k G(jw) (175)
0 <C to < 00 .
Now let
REFERENCES
1. Using either the isocline or the delta method, sketch the phase plane
trajectories for
2x
a. x 4- j—j + 9x = 0
|*|
x(0) = 0, x(0) = 3.
b. x + (1 + 4x2)x = 0
x(0) = 0, x(0) = 2.
c. x + ^x|x| + 2x|x| = 0
x(0) = 0, x(0) = 2.
364 ELEMENTS OF MODERN CONTROL THEORY
d. X- (1- | X2)X
+ X= 0
x(0) = 1, x(0) = 0.
e. x + 5 sin x = 0
c + c + 4c = 4u
e = r — c
where r and c are the input and output signals, respectively. The
relation between u and e is given by
u = e —3 < e < + 3
= 3 e > 3
= - 3 e < -3
x = aix -f a2xy
y = — a3y + a4-xy.
x + 2f x + 16 x = 0
where
f = 10 -2 < x < 2
= 0 |x| > 2.
If x(0) = 0, discuss the nature of the transient response for large and
small initial deflections.
G«=srhy H<s>
-1
where the nonlinear element is an on-off device of the type shown
in Item 23, Table 2.1. Determine the amplitude and frequency of
the steady-state limit cycle using the describing function method.
Take M = 1.
u = e3+ e e2 .
_ 1
H(s) = 1.
= s(s2 + 1.2s + 9) ’
The nonlinearity is a relay with dead zone (Item 18, Table 2.1).
With the aid of appropriate loci in the Nyquist plane discuss
10. The Nyquist curves in Fig. 2.34 are representative of a unity feed¬
back system where the linear transfer function has the form
_ K_
^^ (s + a)(s + b)(s + c)
and the nonlinearity is a relay with dead zone. Show how the
conclusions relative to limit cycle properties derived from Fig.
2.34, may also be ascertained from a root locus plot.
K
G(s)
s(s + 1)
and
(f = <72for a > 0
y f (1 + 0.5y)y + y = 0.
x = —2x + xy2
y = -y + yx2.
where I*, Iy, lz, and M*, M„, Mz, are the moments of inertia and ex¬
ternal torques, respectively, about the indicated axes. The quantities
xi, x2, and x3 denote the angular velocities about the x, y, and z axes,
respectively. In order to eliminate the tumbling motion of the
satellite, it is proposed to use the control laws
Mj = — kiXj
M„ = — k2X2
Mz = — k3x3 .
15. A single axis attitude control system for a satellite subject to gravity
gradient effects may be described by
K
G(s) = 7— ,— . , .
(s + a) (s + b)
s + 2
G(s) =
s2(s + 5)
and
ip = 100(1 + lOtrV.
18. Let
1
G(s)
s(s + 1)
and
(p = cr3
in Fig. 2.5 1.
19. For the system described in Example 14, Sect. 2.3.3, suppose that
the nonlinear element satisfies the condition
0 < < k.
a
Using the Popov method, derive an upper limit for k which ensures
absolute stability.
THEORY OF NONLINEAR SYSTEMS
369
5(s + 1)
G(s)
s2(s + 3)
and
p = a3.
At the risk of belaboring the obvious, we may point out that a system
design and its hardware implementation are never completely identical.
Design values, when translated into equipment specifications, usually
include permissible deviations (tolerances) that reflect the degree of per¬
formance considered acceptable. Tolerances that are too “tight” are
generally wasteful of money and effort, while those that are too liberal
could compromise performance.
In control systems, once nominal values have been established for
component gains and time constants, there remains the problem of
determining the effect of deviations from them. A brute force approach
consists of what we may euphemistically call a “parameter study,” in
which all combinations of selected parameter values are examined with
respect to resultant system performance. When a large number of para¬
meters are involved, such a task becomes unmanageably large.
A rational approach to the problem is contained in the methods of
sensitivity theory. The concept of sensitivity is an old one in control
theory, and, indeed, one of the primary virtues of the principle of feed¬
back was that it provided a degree of insensitivity to parameter variations
in the forward loop. This idea is expanded upon in Sect. 3.1.1.
The modern theory of sensitivity is concerned with determining the
effects of parameter variations on system performance in a broad sense.
It embraces the usual frequency-response methods for single-input/single-
output systems, as well as state variable representations and sampled-
data, multivariable, and optimal control systems. The fact that significant
theoretical contributions still appear is indicative of the formative state
of the theory. Nevertheless, the foundation is sufficiently firm to be
extremely useful in system design.
An intelligent use of these methods provides a framework for the
rational specification of system component tolerances, and a quantitative
evaluation of how the system is affected by parameter deviations.
370
SENSITIVITY ANALYSIS 371
3. 1 MEASURES OF SENSITIVITY
(1)
o y =
(2)
d(Aix)
s T _ G dT 1 _ 1
G T aG ~ 1 + GH " 1 + L' (4)
/ tn[ SGr(joj)]dw
=0
Jo
*The argument, s, will be dropped whenever it is convenient and where no ambiguity results,
f Assuming that the open-loop transfer function has at least two more poles than zeros, or,
equivalently, if | GH | -> 0 at greater than 6 db/octave.
374 ELEMENTS OF MODERN CONTROL THEORY
K
G(s) = H(s) = 1 (5)
s(s + a)
Sr _ 1 _ s(s + a)
(6)
1 + G(s) s(s + a) + K
o2(co2+ a2)
isGT(jco)i
= (7)
K2 - 2Kco2 + arV + a2)
a aco
4 S/(j«) = tan 1 — tan-1 — - — (8)
—Cl) (K — to2)
Thus
ISg7"
(jco)|< 1 for co2< —
Obviously, the greater the value of K, the greater the frequency range
in which feedback is effective for reducing the sensitivity parameter
variations.
Evaluation of SGT(jw)is most conveniently accomplished from a Nyquist
plot of the open-loop transfer function. From Eq. (4)
1
SGr(jco) (9)
1 + L(jco)
This, in turn, means that |SGr(jw)| < 1 for < wi. Therefore, a well-
designed system, in terms of low sensitivity, will have a bandwidth less
than ui.
If a plot of the closed-loop frequency response is available, SGr(jcu)may
be read directly off the diagram, as shown in Fig. 3.3.
376 ELEMENTS OF MODERN CONTROL THEORY
Im
we have
G aT
(11)
tag
SENSITIVITY ANALYSIS 377
so that
T* = T + AT = T 1 + — SJ (12)
L G .
k n (s + zj)
Ka(s)
(13)
0(s)
n (s + p<)
C(s)
&f G(s)
+_ a
K II (s + zy)
7—1
C(s) G(s) Ka(s)
T(s) (14)
R(s) 1 + G(s) <3(s) + Ka(s) m-f-n
II (s + q»)
1=1
assuming m > 1.
Conventional root locus techniques may be used to determine how
the closed-loop poles are affected by changes in loop gain only. For
present purposes, we seek to determine the sensitivity of the closed-loop
poles to variations in K, zy, and p<. This particular sensitivity may be
defined in various ways. UrO takes
where x may be the open-loop gain, a pole, or a zero. On the other hand,
HuangO) uses
s . = x_ dq,
* q » dx
S*f = (15)
(16)
dq» .
S (17)
8py
to emphasize that it is a function of not only s and K, but also the pole
and zero locations. Forming the total differential, we have
ip 6G , I dG <3G,
dG = — ds + —-dK + E — dzy + E- — dp, . (19)
6s dK ,=1 dzj ,=i dpy
Since — q, is a root of
1 + G(s) = 0 (20)
the total differential of Eq. (19) must be zero for s = —qi. Setting
dG = 0 and s = —qi in Eq. (19), we obtain, after rearranging terms
/dG
dq, =
(§' dK +
.E (f)
' dZ; + E dp, \dPi
(21)
« ,°G\
(f). s = -qi
=Gt)J s=-q.
ds ) s = -q,-
1 12" 71 771+71
(26)
(27)
(?)
Note that the gain sensitivity is based on a fractional change in K,
while the pole and zero sensitivities are based on absolute shifts in p, and
z y. This apparent lack of harmony in the definitions is of little concern,
since these definitions lead to simple and instructive relations for the
respective sensitivities.
Recalling that G( —q,) = —1, while
3G = G
dK I\
dG G
dZj (s + Zj)
dG G
dP / (S + P;)
1
S
(28)
(29)
(30)
If we now let Q, denote the residue of T(s) at the pole, -q;, then
SENSITIVITY ANALYSIS 381
T(s)
G(s) _ Q< (32)
1 + G(s)~ £ (s + q<)
where(t6)
(s + q,-)G(s)
*<(s) (34)
1 + G(s)
Then, by definition
= Q< % (35)
where
Q.E.D.
Another important relation satisfied by the pole and zero sensitivities
is
n m-\-n
Es,/ + E s l. (36)
;=i 7=1
This is, in fact, a direct consequence of Eqs. (28)-(30) and the form of
G(s). We note that
(37)
But
i
/
a
s + Zj
m+n 1
P' = pE
l
s + Pi %
Therefore
1
SK' =
m-fn i (38)
E + E
7=1 -q* j=i q«— Pi
or, equivalently
Q * m+n C i
E — +L — = l .
zi - q« % q«-- Pi
The gain sensitivity, S«' (which is, in general, a complex number), has
a simple physical interpretation in the s plane. It is a vector, tangent to
the root locus at s = —q,-, and oriented in a sense opposite to increasing
K. This follows trom the fact that a root locus is a plot of the roots of
1 + G(s) = 0.
<3G dG
dG = — ds + — dK .
os dk
s=-<u
Since dK/K is a real number, the direction of d s along the locus for
positive dK/K is given by —SK4.
The classical and gain sensitivities are related in a very simple way.
From Eq. (4), with H(s) = 1
„ 1 G
Sg = 1 + G = 1 “ 1 + G ‘
In a given situation, the nominal values of K, q*, z and p> are known.
When the variation of G(s) is expressed in terms of dK, dzy, and dpy, one
may determine dqy if the gain sensitivity, Sr , is known. Various methods
of calculating this quantity are considered next.
m
SV — (42)
(s) + KV(s)J
This relation follows directly from Eqs. (13) and (28), making use of
the fact that G( —q.) = —1. For high order systems, the evaluation of
SK' via Eq. (42) becomes quite laborious. One may then use a graphical
approach (slightly less accurate), the simplest of which is the so-called
Gain Perturbation Method. Here we make use of the incremental ap¬
proximation to Eq. (15), i.e.
S (43)
(s + q<) G(s)~
1 + G(s)
SENSITIVITY ANALYSIS 385
Gain = K
s =
-qi
K II (zy - q.)
7=1
SU = (44)
m-\-n
II (qy - q<)
7=1
We may call the approach using Eq. (44) the Vector Method. To use
this, it is necessary to have a complete set of closed-loop poles. The gain
sensitivity is then expressed in terms of a product of vectors from the
open-loop zeros divided by the product of vectors from the closed-loop
poles. The accuracy is limited only by the accuracy of the graphical plot.
At this point, it is instructive to consider the application of the above
ideas in a specific case.
K(s 5) Ka(s)
G(s) = (45)
(s + l)(s + 2.75) 0(s)
386 ELEMENTS OF MODERN CONTROL THEORY
The corresponding root locus is depicted in Fig. 3.6, which also shows
the closed-loop poles corresponding to K = 7.25. It is required to deter¬
mine the shift in these closed-loop poles due to prescribed variations in
gain and open-loop poles and zero.
Applying Eq. (41) to this case, we have
Zl - qi = 3,00 /100°
qi — pi = 5.38 /~~33°
qi — p2 = 4.03 /~47°
— qi = —5.5 + j2.96 .
o o in o
LO O LO o 10
<M <N CO
c4 1-H © I I I
T n- i- 1- r T n- 1- 1
1.
Exam
for
Locus
Root
3.6.
Figure
7.25
20
388 ELEMENTS OF MODERN CONTROL THEORY
<u
3.7
Fig
Ve
for
Qu
De
of
Se
.25
21
SENSITIVITY ANALYSIS
389
AK = 1 Api = 0.25
0.75 0.25
"r 3.0 /T00° ^ 5.38 X -33°
0.5 1
~ 4.03 /~47°
= 1.16 /~71° .
8.‘25(s + 5.75)
G(s) =
(s + 1.25) (s + 2.25) '
We find
— q* = -5.875 + j 3.968 .
In view of the fact that the parameter deviations were not at all
infinitesimal, the agreement is excellent.
Remark: One of the main virtues of the method herein described is that
useful qualitative sensitivity features may be obtained with
little effort. An examination of Eq. (41) indicates that the
sensitivity of a closed-loop pole to variations in a particular
open-loop pole or zero diminishes with increasing distance
between the two. Thus, variations in open-loop poles or zeros
far removed from the closed-loop pole in question have a
minor influence on the latter. This is perhaps intuitively evi¬
dent, but Eq. (41) expresses this condition in precise fashion.
x = Ax + w (47)
x(0) = c (48)
e
At
c + / eA(l~
a,w (<r) d<7 (49)
(51)
where
vz- = row vector (1 X n matrix) whose elements are the ith row of
M"1.
(52)
da*/
AXy F*/(Xy)
Aa*/ n
i^i
where
AXy = Xy - Xy . (57)
SENSITIVITY ANALYSIS 393
Sk ’ = d\; _ E^(Xy)
" d&kl F«>(Xy) (58)
j.L (7 Ay
S‘ - the matrix whose (kf)‘ element is ~—
da k(
(60)
F(1)(Xj)'
We have also*
where uy and vy are the eigenvectors and the row vector previously
defined. Using this and Eq. (A5) of Appendix A, we find that the
eigenvalue sensitivity matrix of Eq. (60) may be expressed as
This is the basic result of the analysis. Equation (63) shows that a
knowledge of the eigenvalues and modal matrix for A yields all the infor¬
mation concerning the system sensitivity, in addition to that necessary
for determining the system response.
duy ^ , i
/ j b.k( Uy
d&hf i=l (65)
H* = VA u,-r (66)
(Xy- X,-)'
whose solution is
= 12 eX‘7UyViC
= X eX,T(S']
x = Ty
x = Ax
we have
Y = By
where
B = T-1 AT . (67)
and, therefore
Q.E.D.
But A may be written as
A=±\imT (68)
i=l
B = Z MS/f (69)
i=l
396 ELEMENTS OF MODERN CONTROL THEORY
where
x = Ax
where
-2 -1 1
A 1 0 1
-1 0 1
Xi = 1.0
X2 = —1 + j
X3 = — 1 — j .
5 5
—(3 + 4j) - (3 - 4j )
(2 + j) (2 - j)
and its inverse
M-1 = (1 +
(1 - j)
Therefore
0 5
Ui = 1 u2 = -(3 + 4j)
_ 1 _ (2 + j) _
5
u3 =
1_ Tv
CO,
04- 4j)
- j) J
SENSITIVITY ANALYSIS 397
and
Vi = -1
v2 - (1 + j)
10
v3 = (1 - J) -j
10
Now
aa23 - —0.8
1
(1 + j)' [5 - (3 + 4j) (2 + j)]
S2 = v2r u2r
10
j
L_j
5(1 + j) (1 - 7j) (1 + 3j)
1
5j (4 - 3j) -(1 - 2j)
To (1 - 2j)_
_-5j -(4 - 3j)
(1 ~ 2j)
s23a= 10
Consequently
(1 - 2j)
AX2 (-0.8) 0.08 - 0.1 6j .
10
398 ELEMENTS OF MODERN CONTROL THEORY
Thus, using Eq. (57), we find that the new value of X2is
_2 -1 1
A* = 1 0 0.2
-1 0 1
Xf = 0.815
X* = -0.9075 + 0.81 j
X* = -0.9075 - 0.81 j
The sensitivity measures considered thus far carry over with simple
and obvious modifications to sampled data systems.
Employing Z transforms, Eq. (4) becomes
(71)
SENSITIVITY ANALYSIS 399
which may be called the classical sensitivity for sampled data systems.
It may be evaluated in the manner shown in Figs. 3.2 and 3.3, except
that in the present case, we plot L(z) with* z = e’“T, and co runs from
zero to 7r instead of from zero to infinity. In other words, the usual
frequency-response methods for sampled data systems apply.
Furthermore, the discussion of Sect. 3.1.2 applies directly, except
that we deal with the z plane instead of s plane poles and zeros.
It is also easy to show that the eigenvalue sensitivities discussed in
Sect. 3.1.3 have a direct equivalent for discrete systems.
Consider, for example, the discrete system represented in state
variable form by
xt = A4c . (74)
A = M-1 A M (75)
Xl
A =
*T = sampling period.
ELEMENTS OF MODERN CONTROL THEORY
400
M = [ui u2
Vi
V2
M-1 =
A* = MA* M“ (76)
X* = [UiU2 Vi
V2
'% \n Vn
which reduces to
(77)
X*= X ^ UiVi
c .
O&kt
S' = v jr u jT (79)
Gc(s) = m X p matrix
Gi(s) = m X p matrix
H(s) = p X n matrix.
Subscripts o and c on vectors U(s) and C(s) refer to the vectors in the
open- and closed-loop systems, respectively. We define also
where G*(s) represents the plant matrix when the parameters differ from
nominal.
Referring to Figs. 3.9 and 3.10, we find
It will now be shown that the matrix relating E0(s) to Ec(s) has all the
properties of a sensitivity matrix. In this context, it is assumed that
Gc(s), Gi(s), and H(s)are such that Cc(s) = G0(s) when there are no plant
variations.
From Eqs. (83) and (84), we obtain (dropping the argument s for
simplicity)
/
Substituting Eqs. (80), (90), and (91) into Eq. (89), we obtain, after
some reduction
where
Co = TR . (94)
Also
Ec = (I + G*GiH)_1 E0 . (98)
404 ELEMENTS OF MODERN CONTROL THEORY
We define
S = (I + G*G1H)~1 (99)
Ec = Cc — C* = (Tc - T*)R
E0 = C0 - C* = (T0 - T*)R .
Now
T0 = GGc (101)
and
where
atc = t* - Tc . (104)
I.P. = (105)
e(t) = £ HE(s)]
r‘f {%‘f
/ e cr(t)ec(t) dt < / e0r(t)e0(t) dt (106)
\ **0
For the scalar case, it was shown in Sect. 3.1.1 that |S(j«)| < 1 was
sufficient to ensure that the feedback system has better sensitivity
properties than the open-loop system. The matrix equivalent of this is
exhibited in Eq. (107). For a well-designed system, it is, therefore,
necessary to satisfy Eq. (107) over the frequency band of interest.
406 ELEMENTS OF MODERN CONTROL THEORY
c ’— sk‘,
(108)
Z’ (zy - Qi
Sk'
V = (109)
(q< - Py)
m-f-n
1
Z + z (110)
7=1 7=1 (q* - py)
Im
1m
sin2 <p .
e2
^ d2
We have also
sin2 <p
e2 = 2r2 — 2r2 cos 2 =
~ d2
Solving for r
r =
2d
sin
e 2r sin <p =
~d
Q.E.L).
The fact that the compensating pole and zero are restricted to lie on
the negative real axis in the s plane means that the U vector locus will
traverse only a portion of the U circle. To determine these limits, con¬
sider Fig. 3.14a, in which the compensating zero is located at the origin.
To obtain the resulting U vector, draw IV parallel to QO. Now locate
point Jr by constructing VIJr = <p/2. From Jr, draw a line that passes
through the center, S2, of the M circle. Finally, draw IUr perpendicular
to JrS2. Point Ur on the U circle represents the right-hand limit of the U
vector locus.
The left-hand limit is obtained from Fig. 3.14b. Here the compen¬
sating pole is located at minus infinity. Locate point on the M circle
by constructing A I&Q = <p. Draw the line, Ifi, that intersects the U
circle at point U^. Point U/ represents the left-hand limit for the U vector
locus.
Thus, as the compensating pole and zero move from the extreme right
position (where the compensating zero is at the origin) to the extreme
left position (where the compensating pole is at minus infinity), the U
vector follows that portion of the U circle that is shown solid in Fig.
3.14b.
These results are easily established by elementary geometry.
It will now be shown how the U circle may be used to satisfy stipulated
specifications on the sensitivity of the desired root location. The required
sensitivity may be expressed in one of several ways:
412 ELEMENTS OF MODERN CONTROL THEORY
Im
Im
Im
s(s + 1) (s T 4)
1 1
r “ 2d = 2
e = 2r sin <p = 0.89 .
The results are shown in Fig. 3.16, in which the limit points on the
U circle have also been determined.
We now seek to obtain a unique location for the compensating pole
and zero by considering, in turn, three types of sensitivity specification.
We first take the case where the sensitivity of the root at Q is to be a
minimum with respect to variations in the open-loop poles. As noted
earlier, this simply means that the vector from Q to the U circle, i.e., the
U vector, must be a maximum. An inspection of Fig. 3.16 indicates
immediately that the maximum U vector is given by QUr, i.e., at the
SENSITIVITY ANALYSIS 415
3.
Example
for
Circles
U
and
M
3.16.
Figure
416 ELEMENTS OF MODERN CONTROL THEORY
q i SfC* _ _ 1_
* (q<- pO /_U
\Sjc7 (q» - Pa)
QU X QA
and
In other words, the U vector must be aligned with 0Q. This is the
same result obtained for constant relative damping factor with respect to
SENSITIVITY ANALYSIS 417
for
Constr
3.17.
Figure
Damp
Relati
Consta
418 ELEMENTS OF MODERN CONTROL THEORY
AP = P - P0 (111)
AT = T - To (112)
where the zero subscripts indicate values for nominal plant parameters
and T is the closed-loop transfer function
T = (113)
1 + L
For ease of writing, the argument, s, in the transfer functions will hereafter be dropped.
SENSITIVITY ANALYSIS 419
L = GHP . (114)
T AT P
(115)
P AP T '
By substituting Eqs. (Ill )—( 113) in Eq. (115), we find, after some
reduction
S/ = — ]— . (116)
1 + L0
ar = G
ap (1 + GHP)2
which was used to derive Eq. (4). Therefore, Eq. (116) is valid what¬
ever the magnitude of AP.
It was first pointed out by HorowitzG5/9) that a failure to appreciate
the far-ranging significance of the sensitivity relation for large parameter
variations, Eq. (116), has led to the formulation of ambiguous and
superficial “adaptive” systems that are, in fact, less effective than con¬
ventional linear types. He quotes several motivations in the literature to
justify the needs for adaptive methods:
1 + HP0
To = P0S
o
1 + HP0
L0 = y - 1. (117)
bp
Here there are two “degrees of freedom. ” For this case, “it appears
that it is theoretically possible to design an ordinary type of feedback
system to have a chosen sensitivity to any amount of plant variation.”!19)
Many different types of two-degree-of-freedom configurations are pos¬
sible/15) The essential point is that they are all essentially equivalent in
having the capability to realize T and SPTindependently. With this point
in mind, it becomes fruitless to search for new and exotic configurations
(i.e., model feedback, reference model, etc.) when what is really needed
is merely an additional degree of freedom.
As is generally the case, added benefits are obtained at a price. In the
situation considered above, the desired insensitivity of T to large varia¬
tions in P is obtained at the cost of large open-loop gain, which effectively
means compensating networks with large bandwidth. The design philoso¬
phy of the method will be explained in terms of familiar root locus
concepts.
P = _ __ . (119)
s(s2 + 2fpw„s + cop2)
Im
One way of achieving this is to place a complex zero pair (zh zi) in the
vicinity of (R, R) and to use a high open-loop gain. There are two funda¬
mental problems:
The use of high open-loop gain means that ultimately the root loci
associated with the compensator poles are such that a closed loop may
be in the vicinity of the dominant poles or else in the right-half plane
(unstable) if the gain is sufficiently high. Consequently, these compen¬
sator poles must be “sufficiently far” to the left. The further they are to
the left (and the higher the gain), the greater the bandwidth of the
system. Usually, one stipulates that the compensator closed-loop poles
should be to the left of some arbitrary line, UVY'U', in the s plane.
There is some freedom in choosing this line, depending on the particular
SENSITIVITY ANALYSIS 423
II (p;E)
i
K = = -
n we)
;
which is merely the usual expression for open-loop gain as the product
of pole vectors to the closed-loop pole divided by the product of zero
vectors to the same closed-loop pole. When K is increased to K', the
closed-loop pole moves to point E', and we have
= II (PiEQ
XI WE') '
424 ELEMENTS OF MODERN CONTROL THEORY
The conditions of the problem are such that the drift in E is very
small; therefore, p<E piE' and zyE'« zyE' except for the zero, zh near
E. Consequently
K' (siE)
(120)
K ~ (zxEO‘
(plE'KplEO II (p.E)
t^l
(ziEO n (zjE)
1*1
SENSITIVITY ANALYSIS 425
Therefore
(ziEQ^ (plEQ(pjEQ
(ziE) ~ (PlE)(PlE)
This approximation is valid whenever |PiE| ^ |p<E'| except possibly
for i = 1, and when |z*E| « |ZjE',1 except for j = 1. As noted earlier,
the design constrains the drift in E td be small, which means that these
conditions are generally satisfied. In this case, it is also permissible to
assume that piE'^piE andpiE'^piE. We then have
Equations (120) and (121) are used to find the approximate shape
and orientation of the region of variation of the dominant roots of
1 + L = 0 due to variations in gain and the poles of L. This is done as
follows.
In Fig. 3.21, Point O corresponds to the location of the compensating
zero, and X corresponds to the nominal location of the dominant root.
These two points may be selected at will, since only their relative location
determines the scale and orientation of Fig. 3.21. It is assumed in the
following that Kmin represents the nominal value of K. As drawn in Fig.
3.21, OX = 1.0 /0°.
Now suppose that K changes from its nominal value to K' = 4Kmin.
The new position of the dominant root (denoted by X') is given by Eq.
(120); viz.
_ ziX__ OX =
K ~ zjX' OX'
This serves to locate X' in Fig. 3.21; i.e., OX' = 0.25 OX. Thus,
X'X is the approximate locus of the dominant roots as K changes from
Emin to 4 Kmin.
With the open-loop gain equal to its nominal value, Kmin, we now
seek to determine the locus of the dominant root as the open-loop pole
traverses rectangle ADBC of Fig. 3.19. Consider first the effect of
the plant poles moving from A to D. We use Eq. (121)* with (see
Fig. 3.19)
Keeping in mind that E(E') in Fig. 3.20 or Eq. (121) corresponds to X(X') of Fig. 3.21.
426 ELEMENTS OF MODERN CONTROL THEORY
piE 4
piE « —4 + j20
p,'E sa - 10
piE « -10 + j20 .
and
OM = 2.70 ,/15°
z(E = (M
ZjE' OM'
o
O /
-30
Figur
3.21.
of
Regio
Varia
Domi
Roots
428 ELEMENTS OF MODERN CONTROL THEORY
point R is a root of 1 -f- L = 0 when K = Kmin and when the plant poles
are at approximately —5 ± lOj. The latter is obtained by noting that in
Fig. 3.21, R lies on lineXM, which corresponds to line AD in Fig. 3.19.
By measuring on Fig. 3.21, it is found that XR ;^(0.18)XM. Since
(0. 18) (AD) « 1, point —5 + jlO is thus determined.
With the above information, the required location of the zeros, zh zi,
of Fig. 3.19 can be obtained fairly closely. It is known that the net
angle of the vectors from the poles and zeros of L to —10 + jlO must be
180°. Since the far-off poles of L are not known as yet, a few degrees
may be assigned for their contribution, or, as a first approximation, they
may be neglected altogether. Therefore (see Fig. 3.19), 180° = ZOR +
ZpR + ZpR — ZziR — ZziR. Since ZziR ^ 90°, this leads to ZziR ^
146°. It has previously been determined that |ziR| = 1.2. Since R is at
—10 + j 10, this locates zi at —9 + j 9.3. With this information, it is pos¬
sible to turn to the problem of locating the far-off poles of L.
4 Kmin= |(QW)(QW)(OW)|
. (122)
However, for a closed-loop pole to appear at R (Fig. 3.19), we must
have in addition
SENSITIVITY ANALYSIS 429
(PlR)(pill)(01l)(QR)(QR) (123)
(ziRXzrR)
430 ELEMENTS OF MODERN CONTROL THEORY
x = f(t, x, u, a) (124)
x(t0) = c (125)
*The optimal control shown is closed-loop; that is, it depends on the current state of the system.
For open-loop control, c would replace x(t) in Eq. (127).
432 ELEMENTS OF MODERN CONTROL THEORY
If the actual parameter vector is used instead of the nominal a0, the
increment in performance index is given by
(130)
— &>iQ •
c j _ dJ*(tp, X, a)
(131)
f da i
y = Gpx (133)
x(t0) = c
r‘f
j (to, X, u, a0) = / (yrQy + uTRu)dt + xr(t/)Mx(t/) . (134)
SENSITIVITY ANALYSIS 433
Po(t/) = M (137)
where
where <t>(t,t0) is the transition matrix!10) for the system, Eq. (139).
Substituting Eq. (135) and (141) in Eq. (126), we obtain
J*(t0, x, a) = cT J W(t,t0)Qi<t>(t,
t0)dt+ &T(tf,
to)M4>(t/,
to) (142)
*We also make the usual assumption that the weighting matrices, Q, R, and M, are symmetric,
positive definite. Actually it is sufficient for Q and M to be positive semidefinite.
434 ELEMENTS OF MODERN CONTROL THEORY
where
where
r‘f
P(t) = Jt/ 4>r(r,
t)Q1(r)#(r,
t)dr + 4>r(t,,
t)M$(t,,t) . (145)
Noting that 4>(t, r) = I (the unit matrix) for all r, we find the boundary
condition
P(t/) = M . (146)
rpi/d<J>r(r,
_ t) t)
P =J - dt
—-Qi(t)4>(t,
t)dr+ vi/'’
dt
M«f>(t/,
t)
(147)
+ /Jt t)Qi(r)
—T-
dt
t}-dr
+ 4>nt/,
t)M dt
- Q,(t).
Noting that the transition matrix satisfies the relations*:10)
d4>r(r, t)
= - FT(t)<t>r(r,t) (148)
dt
d<t>(r,t)
= - *(t, t)F(t) (149)
dt
P + FrP + PF + Q, = 0 (150)
P(t,) = M . (151)
SENSITIVITY ANALYSIS 435
SJ = cr —
* 3a i
c
(153)
= ai0
*-(**)
3at- \dt /
+ ft**+
d&i
^!p
da. t-
+ p^
dat
(154)
+ A F + *9?- 0 .
d&i d&i
We now define
3P
P<(t) = (155)
3af
_3/dP\ = d/ap\
3a,\dt / dt\3at/
S 4 = crPjC (156)
where (after putting a = a0 in Eq. [154] and making use of Eq. [152])
dt
Pt- + FtF , + PiF + Qs = 0 (157)
P.CV) = 0 (158)
436 ELEMENTS OF MODERN CONTROL THEORY
and
(159)
d&i _ a-to
Thus, having the sensitivity function, Eq. (156), the increment in the
performance function due to prescribed increments 5at in the system
parameter is obtained from Eq. ( 130) as
AJ = E S/5a; . (160)
Remark: For the situation analyzed above, the calculation of the per¬
formance index sensitivity function is conceptually simple.
However, it is readily apparent that, in order to obtain numeri¬
cal results, the use of a computer is virtually mandatory. The
theory may also be generalized to include nonlinear systems.
In this case, it can be shown!9) that quantities J*(t, x, a) and
J*(t, x, a0) of Eq. (129) each satisfy a type of Hamilton Jacobi
equation. However, as may be expected, the computational
aspects become overwhelming.
We may also note that the sensitivity function defined in
this section is not the only one that could be used. Rohrer
and SobraK2^) suggest the use of a relative sensitivity function.
This is defined as follows.
At the plant parameter, a, the relative sensitivity for the
control, u(t), is defined as the difference between the actual
value of the performance index and the value that would be
obtained if the control were the optimal for the plant param¬
eters, a (divided by the optimal performance index for nor¬
malization).
x = Ax (Al)
x(0) = c (A2)
A u i = \i uf
(A3)
= 1,2,-
and
M = [U] us . u,] .
Vl
v2
M"1 =
437
438 ELEMENTS OF MODERN CONTROL THEORY
M-1M = I
it follows that*
X= X Ui (A6)
i= 1
Vkx = Vi X U V*U;
i=l
n n n n n
X vAx = X b X v4 u£ = X it = Xf*
^=1 i=l &=1 i=l /c=l
£* = VfcX (A7)
and, therefore
n
X = Xv,xu«. (A8)
i= 1
Ax = A X V,-x Uj .
*In the language of formal matrix theory, the set of vectors juj) is said to form a basis, and the
set of vectors |v,-j is the reciprocal basis.
SENSITIVITY ANALYSIS 439
Ax = ^2 \ Uj Vix
i=l
which leads to
A = £ XiUi Vi (A9)
n
(A 10)
X= a*(t)Ui
1=1
C = J2 a*(0)Ui
i=1
or
on = \i a i . (A12)
440 ELEMENTS OF MODERN CONTROL THEORY
X = X eXi<Vi C u,-
i= 1
or, equivalently
(A 14)
x = X eXi' u,-v,-c .
i=l
x = e^'c (A 15)
n
(A 16)
i=l
REFERENCES
G(s)= K(s+ a) .
(s + b)(s + c)
G(s)= — -(s-+1) .
s2 + 3s + 3.25
Construct the root locus for this system. Verify that for K = 1.0,
there are closed-loop poles at —2 ± j 0.5. Suppose now that the
open-loop poles are shifted to —1.75 ±j 0.85 and the zero shifts to
—1.20. Using the appropriate sensitivity relations, calculate the
new positions of the closed-loop poles.
_ 4_
G(s)
s(s + a)(s + 5)
The nominal value of the parameter, a, is 1.0. For this case, the
closed-loop poles are located at —0.40 ± j 0.80 and —5.1. Manu¬
facturing tolerances on the equipment require that permissible
limits be specified for the parameter, a. If it is required that
deviations in the parameter, a, should be in the direction of
increasing relative damping factor for the complex closed-loop poles,
should the permissible variation in a be positive or negative? Sub¬
stantiate your answer.
0 1 0
A = 0 0 1
-20 -29 -10
The wide scope of the Wiener theory, however, stimulated fresh ap¬
proaches to related problems. Press and Houbolt(49) showed how some
of Wiener’s equations could be applied to gust loads on airplanes in the
sense of determining output power spectra from random inputs expressed
in power spectral form.
Perhaps the most significant new result since Wiener was obtained by
Kalman/13) who expressed Wiener’s formulation in state transition con¬
cepts in the time domain (rather than frequency) and thereby simplified
the mathematical structure immensely; more important, he greatly en¬
larged the area of application. Most of the engineering literature in the
last three or four years in stochastic control has been expressed in
Kalman’s format. The method enables one to treat problems of noise
and disturbance minimization in a highly systematic manner; it has been
used to design optimal filters and control functions as well as to identify
unknown system parameters. Bryson and Johansen(29) extended Kal¬
man’s technique to include the case of “colored noise.”
The theory of Wiener-Kalman filtering (as it is now called) has become
virtually classical in a very short time. The basic results are firmly and
rigorously established, and the theory has assumed a prominent role in
modern control system design. Recent studies have demonstrated that
an intimate relationship exists between the Wiener-Kalman theory and
related statistical optimization methods such as linear regression/53)
maximum likelihood/54) and dynamic programming.!25) A summary of
these ideas is contained in a stimulating paper by Smith/55) A judicious
blend of advances in related fields will no doubt further enlarge the
scope and application of stochastic control concepts.
IT II o (2)
F(oo) = 1 (3)
d F(x)
f«- a*
Since F(x) is nondecreasing, it follows that
Also
STOCHASTIC EFFECTS 449
f f(x)
^ — ao
dx = F(x) - F(— oo) = F(x) (6)
Mean:
E(R) -- m (8)
Mean Square:
E(R2) (9)
Variance:
E[(R — m)2] = a2 = v — m2
(10)
= / (x — m)2f(x) dx .
J — oo
Note that
ao = 1 Mo = 1
oti — m Mi = 0
Oi2 = v M2 = <72
7 2
a1 = ct2 — a\
450 ELEMENTS OF MODERN CONTROL THEORY
Binomial:
n! (13)
f(x) p‘(l - p)n~x .
x!(n — x) !
f(x) is the probability that an event will occur x times in n trials, where
p is the probability that the event will occur in a given trial.
Here n and x are positive integers such that 0 < x < n, and p is a real
number between zero and one. It follows readily that
m = np (14)
Poisson:
Ax
-A
f(x) xl e
(16)
x = 0, 1, 2,
A > 0 .
In this case
m = cr2 = A . (17)
Gaussian:
(x — m)2 (18)
f(x)
2 a2
STOCHASTIC EFFECTS 451
F(xi, x2, . , xn) = Prob (Ri < xi, R2 < x2, . (19)
. j Rn ^ X„] .
In words, F(xx, x2, . , x„) is the probability that R< takes on a value
equal to or less than x, for i = 1, 2, • • • •, n. All the inequalities must be
satisfied simultaneously.
The joint probability density function is defined by
Mean:
m, = E(R») . (21)
i = j ; variance
an E[(R, - m,)(Ry - my)] (22)
i ^ j ; covariance
Correlation:
1 — — (24)
f(xi, x2, . , x,) = [(2x)»det V]-1'2 exp[— —(x - m)^V-(x - m)].
Moments:
(26)
Central Moments:
(27)
«io = E(Ri) = mi
«oi = E(R2) = m2
= a 02 — aoi2 = variance of R2
0T2 — Mil
(Til — M20
<r22 — M 02
STOCHASTIC EFFECTS 453
The quantity
P12 — 0'12(o’ll<722)1,2
= AUi(m20
M02)^^2
where
f(x)
= 0.9545 if b = 2a
= 0.9973 if b = 3a .
In other words, the smaller the value of a, the greater the tendency of
the random variable to assume a value close to the mean.
Time
It follows that
Prob [x, < x(ti) < Xi + dx^ = fi (xi, ti) dxi . (30)
Similarly, the probability that simultaneously x(ti) < xj and x(t2) < x2
is given by the second probability distribution function
F2(xi, ti ; x2, t2) = Prob [x(ti) < Xi; x(t2) < x2] . (31)
Similarly
F,(xi, ti; x2, t2; x3, t3) = Prob [x(ti) < xj; x(t2) < x2.; x(t3) < x3
(33)
fn-i(xi, tj ; /oo
fn(Xl, tl,
— OO
xK, t„)dx„
The first term on the right-hand side of this expression plays a funda¬
mental role in stochastic control theory, and it is accorded a separate
name and symbol.
= Prob [x(ti) < xi, . , x(tm) < xm; y(tj;) < yi,
. , y(t*) < yn]
(38)
E[x(ti) y(t2)] f
j — 00
f x(ti)y(ts)
J — CO
f^Cxi,
ti; yi, t2) dxi dyi .
If x {(t) } and {y(t) } are statistically independent— that is, if the value of
x(t) does not depend in any way on the value of y(t)-then
where {x(t)| and (y(t)} are two given random processes. In this case, we
find, directly from the definitions
<pzz(
tl, t2) = t2) + (Pxy(tl, t2) + t2) + tl, t2) . (41)
if one or both of the random processes has a zero mean value. In the
particular case when Eqs. (40) and (42) hold, then Eq. (41) reduces to
<Pzz(
tl, t2) = <pxx(tl,t2) + <Pyy(tl,t2) • (43)
E[x(0)] = E[x(t)] .
<Pxx{
tl + T, t2 T" t) = <Pxx(
tl, t2)
<Pxx(t>l,
t2) — PxxiO, t2 tj) .
IVv(t) = “PatyCO,
t2 tl) . (45)
Note that
(51)
E[x(t)] = lim —
2T
[1 x(t)dt
J _ rp
E[x2(t)] - lim —
T—>» 2T
f% x2(t)
J _ rp
dt (52)
1 fT (53)
- lim — /x(t) x(t + r) dt .
r->« 2T J-T
460 ELEMENTS OF MODERN CONTROL THEORY
1 rT (54)
E[x*(t)] = Urn-L x2(t) dt ^ '
2T J -t
which represents the mean square* of x(t), is finite for nearly all random
= 0 otherwise. (55)
/oo
Xr(t)e-,“‘
- oo
dt
(56)
(57)
:r(t)-±.[ Xr(&»)e,'a* da> .
Xr(co)|2 (58)
: lim
2T
is called the power spectral density, f We will show that this quantity
exists and is finite if the mean square value ofx(t) is finite.
We have
f |Xr(co)|2daj
= f Xf(co)Xy(—oj)do
" — oo " — 00
/oo — 0O
XT(co)dw / X]p(t)e'"‘dt
*
/*oo
— 00
/oo — CD V
/*oo
xj-(t) dt / Xr(w)e'“‘ dco
- 00
/oo
xr*(t)
- CO
dt .
Therefore
lim— f lXr-(-^2
dw= lim— [ xj(t)dt = lim— [ x2(t)
dt.
r-*.2w•'-« 2T r-.. 2T•'-» r-»«2T
J-r
This last quantity is the mean square value of x(t), which is finite by
assumption. Consequently, the power spectral density defined by Eq.
(58) exists, since its integral converges.
It may be shown that the autocorrelation function is the Fourier
transform of the power spectral density. To do this, define
Cr(r)= — f
2T J-t
x(t)x(t+ r) dt
lim[ Cr(r)e~;“T
T - %º
oo J — 00
dr
= lim —
r-- 2T
/OO /•oo
e_)<JT
—00
dr / xr(t) xr(t + r) dt
j —00
= lim —
2T
/OO
dt /i»0O
-
[xr(t)eJ’“(][xT(t
00 " — oo
+ r)e->'“(,+T)]
d7
STOCHASTIC EFFECTS 463
= lim —
T—*« 2T
/OO
- CO
poo
xT(t)eia‘ dt / xr(ti)e_J'“‘i d^
J — ao
= lim|Xr(a,)l2
2T
= G**(w)
by Eq. (58).
This means that
(59)
G**(co) = / rxx(r) e )lut dr
and
(60)
e'“T do .
Since both ria:(r) and GIX(co) are real-valued even functions of their
respective arguments, we may write
(61)
Gxx(co) — 21 r xx(r) cos cor dr
Jo
1 f*3 (62)
r«(r) = — / Gxx(co) cos cor dco .
•7r " n
|Xr(co)|»|Xr(co)|2etc.
2ttT
00
or
(Ref. 48)
The Fourier transform pair, Eqs. (56) and (57), is apparently the most
widely adopted.* This form will be used consistently in this chapter.
There even seems to be some inconsistency by the same author. Laning
and Battin (Ref. 5, p. 123) use
F(w)= —U f f(t) dt
to define the Fourier transform, while in deriving the Wiener-Khinchin
equations (Ref. 5, p. 132), they write
/oo
0
h(u) x(t — u) du .
fT roo t* oo
= lim
/ T
J — J'
*0
/ h(u) x(t — u) du
J
/ h(v) x(t + r — v) dv dt .
0
1 f
Tm(r)
= r /”h(u)
h<v)lim — J x(t —u) x(t + r —v) dt du dv
lim — I x( t — u) x(t + r — v) dt
r-..2T J-T
1 (T
= lim — / x(ti) x(ti + u + r - v) dti = ria:(u — v + r)
9T J _
2T
/OO /i»00
v j0
h(u) h(v) rM(u — v + r) du dv .
Taking the Fourier transform of both sides and making use of Eqs.
(56) and (59), we find
/OOJ/'CO
/»
C O
J h(u)h(v)rM(u—v -(-r)dudv e~} dt
roo noo r oo
/co
h(t) e-“ dt
S = a + jcj
while
by Eq. (59).
We have, finally
= |H(j«)|*Gx,(«) . (63)
o ij = riI(riy) — m2 (67)
where rty = t,- — ty. The Gaussian random process is completely deter¬
mined from its joint density function f2 (xx, t»-; x2 ty) because the auto¬
correlation function is expressed in terms of this density function by
/oo/I*
co
x(ty)
—00 j - <30
x(ty) f2(xj, t,; x2, ty) dxx dx2 . /Z"0\
The Gaussian random process has the important property that it
remains Gaussian after passing through a linear filter.
Consider now a single noise record of the type shown in Fig. 4.3. This
is assumed to be a typical member of an ensemble representing a station¬
ary, ergodic, random process. We are interested in calculating the mean
and autocorrelation function. One may proceed as follows. Divide the
trace of Fig. 4.3 into n equally spaced points a distance, d, apart. With
the kth point is associated an ordinate, a*, that may be positive or nega¬
tive. The mean is simply determined by using the discrete version of Eq.
(51), viz.
(69)
E[x(t)] = - a* •
0 i=l
where r = kd. The curve of r**(r) vs t is plotted in Fig. 4.4. The general
shape of this curve is typical of a wide variety of random phenomena of
interest, such as vacuum tube noise, radar fading records, and atmospheric
turbulence. It may be very closely approximated by
where A, k, and c are positive constants. The power spectral density may
be obtained from Eq. (61), viz.
lim - ke~ft|T| =0 r ^ 0
k—
»°°L2 J
and
k r°
lim - / e-A,T|dr = 1 .
k—>00 2 •'-CO
STOCHASTIC EFFECTS 469
c
d
8
oi
<D
.52
Z
n>
Tb
c
rn
a)
.1
tL
ci
470 ELEMENTS OF MODERN CONTROL THEORY
r»(r) = — e~A,T'
it follows that
where 5(r) is the Dirac delta function.* This last relation simply states
that successive values of the random function are completely uncorre¬
lated, no matter how small the interval between successive samples. This
is, of course, a mathematical abstraction that is never realized completely
in practice. Nevertheless, this concept is a highly useful theoretical tool
and closely approximates many important random phenomena (e.g.,
Brownian motion). The idea of a purely random process, often called
white noise, may be arrived at in many ways. Some of these are dis¬
cussed in Appendix A. However, in general, we will say that a random
process is white if its autocorrelation function is given by
where B is a positive constant. The power spectral density for Eq. (73)
is obtained by direct application of Eq. (62), viz.
*See Appendix A.
STOCHASTIC EFFECTS 471
Gxx(“)
a. 3C2- K2 < 0
G
XX
(W)
2 2
b. 3C -K >0
G„(«) = B . (74)
a. The signal, f,(t), and the noise, f„(t), are each members of a stationary
random process.
b. The device used to operate on f,(t) is a physically realizable linear
filter.*
*A linear filter is said to be physically realizable if its transfer function has no poles in the right-
half plane or on the imaginary axis.
STOCHASTIC EFFECTS 473
(76)
fo(t)
= j h(r)f,-£t
- t)dr
where h(r) is the weighting function* of the filter. Since this filter is to
be physically realizable, we must have
(78)
fo(t) = h(r) f,(t - r) dr .
what follows, it will be more convenient to use Eq. (78) than Eq.
In
(76). If we denote the desired signal by fd(t), then the rms error is given
by
(79)
fd(t)p dt .
fd(t) = f.(t)
fd(t) = f.(t + a)
See Appendix B.
474 ELEMENTS OF MODERN CONTROL THEORY
By substituting Eq. (78) in Eq. (79) and making use of the fact* that
1 fT (80)
ro6(r) lim — /
r->«°2T ^-T
f0(t) f6(t + r) dt
we obtain ^
/OO
h(r)dr Jpoo
h(cr)
Ta(r—a)d<r
(81)
- 2 / h(r) rM(r) dr + r*,(0) •
Gras(tu)
H(s) Hrf(s) (83)
Gss(co) + Gnn(w)
where
H(s) = £[h(t)]
and G„(g>) and Gnn(w) are the power spectral densities for f,(t) and f„(t),
respectively.
The solution of Eq. (82), which takes account of condition Eq. (77),
is summarized here as follows.*
a. Given the power spectral densities, G„(w) and G»»(w), for the signal
and noise, respectively, replace « by s/j and form the sum
hi(t) = 0 t < 0
Hx(s) (90)
H(s)
Gi(s) '
* Details of the analysis leading to the result given here may be found in Refs. 5 and 6. The
method as outlined here appears to be substantially simpler than that given in most references.
fThat is, those terms having poles in the left-half plane only.
476 ELEMENTS OF MODERN CONTROL THEORY
Example 1: Given
1
Gss(co)
1 + CO2
Gnn(<o) = k2
1 + k2 - k2 s2
Gij(s) + k2
1 — s2
= G,(s) Gx(-s)
where
\T + k2 + ks
Gi(s)
1 + s
Now
1 1 - s
Hd0Hs)
1 - s2 Vl + k2 - ks
Ai A2
ea“
S + 1 vl + k2 - ks
where
_ 1_
Ai
k + X/1 + k2
k
A2
k + x/1 + k2 '
Only the first term inside the bracket represents a physically realizable
filter. Taking the Laplace transform of this yields
STOCHASTIC EFFECTS 477
hi(t) = 0 t < 0
= Aie~<'+a) t > 0 .
Consequently
Hi(s) = <£[Aie~((+a)]
/
Aie~tt
s 4- 1 '
We have, finally
H(s) - SW . _ Aier" .
Gi(s) ks + Vl + k2
Gss(o>) 6 _ 1_
ir (“2 +qT
2 (a>2+ 2)
G
r (o>4+ 4)
( 2 (2 - s2)
G«(s) = —
vr (f - S’) 7T (4 + S4)
We have therefore
Hi<°>(s) = Hd(s)
G,(-s)
(s2 — 2s T- 2)ea8
HiW(s) = —
2tt
(s + i)(s2 -2.264s + 2.031)
hx(t) = 0 t < 0
(V) X09444
e_1'6('+o) t > 0 .
Therefore
2.004 e_1-8a
Hi(s)
V 7T
(S + |)
and finally
Hi(s)
H(s)
Gx(s)
Gss(w) = - —
4 + w2
25
Gnft(co)
25 + co2
G„(s) —
4 — s2
25
Gn^(s) =
25 - s2
and
r, , ,
G;,(s) = -
1 ,h 25
4 - s2 25 - s2
'5.0990
(s+ 2.1926)1
|"5.0990
(2,1926
- s)l
. (s+ 2)(s+ 5) J L (2- s)(5- s) J
5.0990(s + 2.1926)
GM =
(s + 2) (s + 5)
Hj^Cs) - 1 %X (2 S)(5 S)
(4 - s2) 5.0990(2.1926 - s)
= 0.1961
[Hggi
L(s+2)
+ °-6!9-6 1
(2.1926—s) J
hi(t) = 0 t < 0
= 0.3274 e_2‘
0.3274
Hi(s) =
s + 2
and finally
0.3274 (s + 2) (s + 5)
(s + 2) X 5.0990(s+ 2.1926)
0.0642(s + 5)
(s + 2.1926)
480 ELEMENTS OF MODERN CONTROL THEORY
*The power spectral density gives the same information, since these are related by Eqs. (61) and
(62).
STOCHASTIC EFFECTS 481
We shall discuss these in turn. The first premise involves the concept
of a “shaping filter which has been considered in the previous section
in a somewhat disguised form. By definition, a shaping filter for an
arbitrary random process with a power spectral density, Gyy(u), has the
482 ELEMENTS OF MODERN CONTROL THEORY
property that a white noise input generates a random process having the
same power spectral density. Referring to Eq. (63), and letting G**(w) = k2
(i.e., white noise), we see that the shaping filter, Y(j«), is defined by
Gw(co)= k2|Y(jco)|2
. (91)
Example 4:
1
G „(&))
1 + w2
G„(s)
Yi(s) = — L. .
1+ s
Example 5:
25 + a)2
G»s(co)
4 + w4
25 - s2 5 + s 5 — s
Gss(s)
4 + s4 Ls2+ 2s + 2. Ls2- 2s + 2J
and
5 + s
Yi(s)
s2 + 2s + 2
Example 6:
169 + o>2
Gss(co)
w4 + 238w2 + 1692
G„(s)
13 + s 1_1CO co 1- 1
Kalman’s second premise involves the use of the concepts of state and
state transition. Thus, instead of expressing the transfer function in terms
of the complex variable, s, it is used merely to relate the output-input
function in state transition terms in the time domain. In general, the
shaping filter, Yi(s), may be written as
/3n_iSl
V /• N 0(s) i=0 ^
Yi(s) = — = - ,m > n . (93)
^?(s) m
^ *m—jS
i= o
x = Ax + b^> (94)
/
J - OO
tij x2, t2j . , xnj tn) dx„
= ^(*0 = (95)
hi—i(xi, ti ; x2, t2; , xrt—i, tn_i)
where f*( ) has the meaning defined by Eqs. (32) and (33). In other
words, ^(x*) is the probability that x(t„) takes on a value less than or
equal to x», given that x(ti), x(t2), . , x(t,_i) have taken on the
values, xi, x2; ••• xn_i, respectively. The conditional mean or expectation
is defined by
(96)
/ x(t„) d'k„(xm)
.
V - CO
Suppose now that we are given a signal x(t) and a noise n(t). Only the
sum y(t) = x(t) + n(t) can be observed. Assume that we have observed
and know the values of y(ti), y(t2), . y(t„_i) exactly. What can be
inferred about the value of x(t)? If we know the conditional probability
distribution
then this conveys all the information derived by the measurements y(tx),
. y(tn-i). Any statistical estimate of x(ti), denoted by X(ti), will
then be some function of ^(xn). In general, the value of X(tn) will be
different from the actual (but unknown) value of x(t„). A rational
method of determining some optimal estimate, X(tn), is to define a func¬
tion of the estimation error, e, and try to minimize this. Let a loss
function be defined by
L(e) = L( —e)
L(0) = 0 (98)
where
e = x(t„) - X(t„) .
Theorem I: Assume that L(e) satisfies Eq. (98) and that the condi¬
tional distribution function, 'k (x„), defined by Eq. (97), is such that
for all x„(1), xre(2) < xn and 0 < X < 1, where xn = E [x*].
Then the estimate X(tR) of x(t„), which minimizes E[L(e)], is the
conditional expectation
This, then, forms the basis for formulating the problem in the following
manner.
486 ELEMENTS OF MODERN CONTROL THEORY
I. Continuous Case
E[v(t) wr(r)] = 0
(103)
where 8 is the Dirac delta function and Q(t) and R(t) are symmetric non¬
negative definite matrices continuously differentiable in t.
It is assumed that the measurement of z(t) starts at some fixed instant,
t0 (which may be — °o), at which time P(t0) = E[x(t0) xr(t0)] is known.
The optimal estimation problem is then formulated as follows. Given
the known values of z(r) in the interval t0 < r < ti, find an estimate
x(ti|t) of x(ti) that minimizes the function, E(«2), where
€ = x(ti) - x(ti|t) .
If ti < t this is called the smoothing problem. If ti = t this is called
the filtering problem, while for ti > t we have the prediction problem.
where <t>(t+ 1, t) and r(t + 1, t) are transition matrices and the other
symbols have the same meaning as in the continuous case, except that in
Eqs. (101)-(103) the numbers t and r are integers. In other words, w(t)
and v(t) are constant during the sampling interval (which in the present
case has been normalized to unity).
The optimal estimation problem is as stated in the continuous case
except that instead of being given z(r) in the interval t0 < r < t, we are
given the sequence of measurements, z(0), z(l), . , z(t). As before,
we are also given the initial covariance of x(0); namely, P(0) = E[x(0)
xr(0)].
Given now the first and second statistical moments, how does one
obtain Eqs. (99) and (104)? This is largely an unsolved problem, and
we are compelled to start with these models and consider the question
of how to obtain them as a separate problem. There is one exception.
If the statistics of the message are known in the form of a power spectral
density or an autocorrelation function (of the stationary type), then
one may obtain Eq. (99) from the shaping filter for the message in the
manner described by Eqs. (92)—(94).
Having stated the basic problem, how does one obtain the solution?
As noted earlier, Kalman showed that the optimal estimate may be
obtained by calculating conditional probabilities; the details of the pro¬
cedure may be found in Refs. (13) and (14). The analysis relies heavily
on the methods of abstract probability theory, which are not readily
accessible to control engineers. It has been shown by GreensiteG6) that
Kalman’s solution may be derived from first principles in dynamic
488 ELEMENTS OF MODERN CONTROL THEORY
x(t) = X,
z(t) = z,
etc. Also
4>(t + 1, t) ss <f>(
r(t + l, t) = Tf
M(t) = M,.
We note that
< t+1
E(x0) = M (110)
and covariance
But since
“I T
1 -1
- 4>(_!
x,_i)lj
=f,exp
£ - (x, -
2
x,_i)r U (_i (x, - $,_i xt_i)
where
u, = r, Q( rv (112)
and
gX!+l|Z([X0, • , X£_f_l
| Zq> • , z,J = 0t exp {“ 5<x”
_ m)1
where $t depends only on the sequence, z0, . ,zt, and known con¬
stants.
Since the random effects are white and Gaussian, the least-square
estimate and the maximum likelihood!1 9) estimate are identical/20) The
problem is then reduced to one of choosing the sequence x0, . , x, that
maximizes the quantity in the braces of Eq. (113) or minimizing the
quantity
+ E (115)
(zi — MiX,)r Ri_1(zi ~ MiXi)
where
x( = L t V< . (121)
(122)
where
&fThis particular form of b(x<) is suggested by the fact that the criterion function Eq. (104) is a
quadratic form, and the state equations are linear.
492 ELEMENTS OF MODERN CONTROL THEORY
where
Ci = m X m matrix (nonsingular)
C2 = m X n matrix
Ci = n X n matrix (nonsingular)
C4 = n X m matrix
We now seek to obtain the physical significance of the matrix A<. For
t = —1, we find, from Eq. (115)
fo(x0) — Xo Ao xo — 2b0r x0 + C( .
Ao = Po"1
b0 = Po~V
c0 = Po~V
x(t —
|—1 t) = 4>(t + 1, t) x(t | t — 1)
P(t + 1 | t) = <t>(t
+ 1, t){P(t | t — 1) — P(t | t — l)Mr(t)
[M(t) P(t | t — l)Mr(t) + R(t)]-1 M(t) P(t [ t — 1)} (133)
where
Figure 4.6 depicts both the dynamic system of Eqs. (104) and (105),
i.e., the message, and the optimal filter described by Eqs. (131 )—( 133).
If the unit interval for the discrete case solution is allowed to approach
zero, we obtain the solution for the continuous case, Eqs. (99) and (100),
viz.
Figure 4.7 is a block diagram for the message and optimal filter for the
continuous case.
STOCHASTIC EFFECTS 495
>
4.6.
of
Case.
Discrete
Filter
Optimal
Figure
Model
and
Message
the
496 ELEMENTS OF MODERN CONTROL THEORY
4.7.
Figure
of
Optim
Mode
and
Messa
the
Filter
Contin
Case.
STOCHASTIC EFFECTS 497
= P t_2Mr(2M P mM t + I)"1
= P0Mr(tMP<>Mr + I)"1 •
2 X*—
2+ 1 [3x+ M_1(v,+ vt—i
+ vt—
2)]
t+ 1 t + 1
t - j - , 1
- x,_,- H- (j + l)x + M-i E v,-
t + 1 t + 1 i= t— j -1
STOCHASTIC EFFECTS 499
X(+l
1
t + 1
[ (t + l)x + E M-‘
1=0
(147)
t+ i
1
E M-1
t + 1
1 *^
X«+1= X H- — E M-1 V,-.
t + lH
P, = i(MrM)_1
b
.
is readily found to be
500 ELEMENTS OF MODERN CONTROL THEORY
Expressed in the time domain, in the form analogous to Eq. (99), this
becomes
x = —x + w
E[w(t)wr(r)] = 5(t — t)
z = x + v
where the power spectral density of v has been given as G^w) = k2;
consequently*
E[v(t)vr(r) = k2 5(t— r) .
Figure 4.8 is the schematic of this system. The solution for the optimal
filter is given by Eqs. ( 136)-( 138), where, in the present case
A(t) = 1 Q(t) = 1
B(t) = 1 R(t) = k2 .
M(t) = 1
dx(t|t)
— dt — x(t|t) + K[z(t) —x(t|t)l
P
K =
R
P '
0 =
-2P-i + Q.
The quantity dP/dt has been set equal to zero in Eq. (138), since we
are seeking the steady-state (time invariant) form of the optimal filter.
From the above equations, we find that this filter has the form shown
in Fig. 4.9. The corresponding transfer function is
x(t|t) _ K .
z s+l + K
Xi = x2
X2 = Wi
zi = hn xi -f vi
z2 = h22 x2 T v2 .
0 1 0
A = B =
O O _1_
Ji o
M =
_0 h.22_
qn 0 rn 0
Q = R =
o o _0 T22
Figure 4.11 is the schematic for this filter. Note that this is time
varying, since, in the general case, only finite data is available. A time
invariant form results when t —> » , since, for this case, P becomes a
constant (corresponding to the classical assumption that an infinitely
long past record of the signal is available).
To solve the problem in the case of finite data one must be given the
quantity* P(0), which is written in the form
<711
(0) <712(0)
P(0) =
(72l(0) (722(0)_
hn <7ii
kn =
Til
111 (712
k2i —
Til
I22 <712
ki2 —
T22
122 <722
k22 —
r22
hn =1 Qn = 1 <7n(0) = 1
h22 = 2 Tn = 16 <712(0) = 0
r22 — 1 <722(0) = 0 .
0.12
0.10
0.08
cq
0.06
0.04
0.02
The solution for the variance and optimal gains is shown in Figs. 4.12
and 4.13, respectively. Kalman!14) shows that undei mild restrictions, a
steady state is reached that is equivalent to the time invariant optimal
filter. This steady state is apparent in the present case. Note that the
initial values of the time varying parameters of the filter are strongly
dependent on the assumed values for <rt>. However, the final steady state
is independent of the assumed values, which is plausible, since with
increased information available, the initial assumptions play a vanishingly
decreasing role in determining the form of the filter. Note that the
506 ELEMENTS OF MODERN CONTROL THEORY
steady-state values for ktJ and could have been obtained directly from
the a equations by assuming = 0 and solving the resulting set of alge¬
braic equations. For high order systems, however, this is a difficult
process, and it is much simpler to integrate the variance equations until a
steady state is reached. The speed with which one converges to this
steady state is, of course, dependent on how good the initial estimate is.
The Kalman filter for the continuous case, as described by Eqs. ( 136)-
(138), is based on the fact that the measurement noise, v(t) in Eq. (100),
is white and nonzero. If indeed v(t) were identically zero (i.e., the
measurement were perfect), then the variance equation, (138), would be
singular, since R(t) would be a null matrix. Furthermore, the analysis
leading to Eqs. ( 136)—( 138) is based on the assumption that v(t) is white,
i.e., successive values for small intervals are essentially uncorrelated. The
case where v(t)has a power spectral density that is not a constant cannot
be treated within the framework of the theory thus far presented. This
gap has been filled in a recent paper by Bryson and Johansen, (29) which
treats the general case of measurements corrupted by colored noise, white
noise, no noise, or any combination thereof. The details of the general
procedure are quite lengthy, and we will accordingly describe only one
special, though important, case. The reader is referred to the afore¬
mentioned reference for a complete treatment.
It is assumed that the process (message) is described by
Xi — Fn xi + ui (148)
y = Ci xi + m (152)
STOCHASTIC EFFECTS 507
m = Am + Bu2 (153)
/
where A and B are p X p matrices, u2 is a p vector, and
H == Ci + Ci Fn — AC i (162)
*See Ref. 29 for the derivation. The result given here is a slightly generalized version of Example
1 in that paper.
508 ELEMENTS OF MODERN CONTROL THEORY
The discontinuities at the initial time are due essentially to the fact
that an initial estimate and an exact measurement are simultaneously
available at the start.
We illustrate the application of these ideas in the following.
The shaping filters for the message and noise are found to be
1
Gi«(s) =
s + 2
5
G1(*)(s) =
s + 5 .
Xi — — 2xi Ui
with
The measurement is
y = xi + m
where
Comparison of these equations with the set, Eqs. ( 148-1 55), indicates
the following equivalence.
Eii = —2 A = —5
Qi = 1 R = 5
Ci = 1 Q2 — 1
STOCHASTIC EFFECTS 509
R = 26.0
H = 3.0
Pi = 0.2232
K = 0.0642 .
The equations of the optimal filter are then given by Eqs. (158) and
(159), viz.
Xi(t|t) — x*(t[t) + Kv
x?(t|t) = —2xi(t|t) — K[—5y + 3xi(t|t)] .
xx(t|t) _ 0.0642(s + 5)
y ~ (s + 2.1926)
since the effect of poor initial estimates diminishes as more and more
measurements become available. The linearity stipulation is more funda¬
mental; usually this is satisfied by treating the estimation error (which
may be taken as the state variable) as a linear expansion about some refer¬
ence condition. The prospects are, therefore, very favorable for applying
the theory in order to obtain significant improvements in system per¬
formance for a wide variety of realistic situations in aerospace guidance
and control. A sampling of the literature in this area is contained in Refs.
30 through 36.
In what follows, we will discuss three problems which are fairly
typical of present applications of the theory.
The vehicle equations of motion then take the following form. (37)
m M«(X - X.)
= fi(X, Y, Z)
rm3 A.3
m»(Y - Ym)
Y = -
r3
(167)
Z = -
r3 1+J(!) ( ^Um(Z Z m)
(168)
M»(Z — Z„)
- f.(X, Y, Z)
r1 m ^ A.3
512 ELEMENTS OF MODERN CONTROL THEORY
where
r = (X2 + Y2 + Z2)1'2
J = 1.6246 X 10-3 .
+ ^ (Z—ZB)+ higher
O/j
order
terms
with similar equations for Y and Z.
If we now define a vector, x, whose components are
Xi = X - xfi
X2 = Y - yr
X3 = Z - Z*
x4 = X - x« = Xi (169)
X8 = Y - Yb = X2
X6 = Z - Zr = Xs
STOCHASTIC EFFECTS 513
x = A(t)x (170)
0 I
A(t) =
_ A,(t); o
where
afi dU dfi
dX dY dZ
Ai(t) df2 df2 dft
ax dY dZ
df3 df3 dt.
_dX dY dZ.
1=3X3 unit matrix
All the partial derivatives are evaluated along the reference trajectory.
Since measurements are to be made at discrete intervals, it is con¬
venient to express Eq. (170) in the form
where <*>(t+ At, t) is the transition matrix!1 8) for the system in Eq.(170).
It is assumed that
E[x(t0)] = 0
P0 = E[x(to) Xr(t0)]
z
a = sin-1 — (172)
r
Y
P — sin 1 (173)
(X2 + Y2)
7 = sin-1 —• (174)
r
OC-m OCact4“ Vj
* @act -F V2
ym y act + V3
where the subscript “act” denotes actual value, and the v, represents
noise. It will be assumed that the noise has zero mean and is uncorre¬
lated between successive measuring instants, i.e., white.
We now define
Aftm @m ^nom
Ay m ' y m y nom
Earth
If we now let
A am = Aa + Vi (175)
A7„, = A7 + v3 . (177)
da da da
Aa dX dY dZ Xl
dp dp dp
AP = dX dY dZ x2
dy dy dy
_Ay_ _dX dY dZ. ,X3.
where v(t) is the noise vector whose components are vi, v2, v3; x is the
state vector defined by Eq. (169); and
Zi(t) A am
M(t) - Mi(t) • 0
where
da da da
dX dY dZ
dp dp dp
dX dY dZ
dy dy dy
dX dY dZ
necessary that the computation time on the computer be less than the
time interval, At, between successive observations.
Some results of a digital computer study of this problem are given by
Smith and others.!30)
<P= t + 56 . (182)
= e (183)
STOCHASTIC EFFECTS 519
e = 6c -Y er (184)
the first of which is an unknown constant vector (gyro bias) and the
second a zero mean random vector whose power spectral density is
known. Specifically, we assume that each component er» of er, corre¬
sponding to each of the three platform gyros, has a power spectral density
given by
(185)
r.(T) = (186)
E[e„-(t) 6„-(t)] = 0 i ^ j •
G„ <«>(«) = 2 (187)
is passed through the linear (shaping) filter having the transfer function
s + p
then the output power spectral density is equal to that of Eq. (185).
Accordingly, we may represent the statistical properties of er, by the
dynamic system
where
er = Her + W . (190)
6C= 0. (191)
X = Ax + Rw (192)
where
r i
X = «r
- .
o : I : I
A = o • H jO
o i o : o
B I
cos 7 COS a
Sr cos 7 sin a
sin 7
S, = ASP
1
A =
522 ELEMENTS OF MODERN CONTROL THEORY
or*
So = Sp + ASp
where
i/'2SaC7 — \J/yS
7
ASp — \pzCaCy + \pxsy (193)
\pv(iacy — xpxSacy
C7 Sa
s(y + A7) _ ST _
*Wc have used the abbreviations, Sa, ca, for sin a, cos a, etc.
STOCHASTIC EFFECTS 523
which reduces to
— Aa Gy Sa — Ay S7 Ca
Aa Ca cy — A7 Sa S7 (194)
Ay cy
Zl Aa
M8!^ -f V (195)
_z 2_ A7_
where
z 0 x -f v (196)
where $(t +1, t) is the transition matrix for the system, and
r(t + 1 , t) = / 4>(t +
Jt
1, r) Bdr .
524 ELEMENTS OF MODERN CONTROL THEORY
Equations (197) and (196) are in the form of Eqs. (104) and (105),
so that the optimal estimation equations are given by Eqs. (131 HI 33).
The optimal estimates thus obtained are used to correct the system in
some appropriate fashion. In the paper by Bona and Hutchinson!34) this
information is used to “reset” the inertial system to compensate for
gyro drift rates. Typical results obtained are shown in Figs. 4.17 and 4.18.
The upper curve in each figure indicates the accuracy obtainable by
conventional methods.* The runs were made with the system operated
as described above, employing optimal estimation followed by corrections
for a period of six hours, after which a conventional correction procedure
was used. The improvement in accuracy achieved by employing optimal
estimation procedures is quite dramatic.
*That is, the so-called “damped inertial mode.” Cf. Ref. 40, p. 146.
STOCHASTIC EFFECTS 525
feet)
(thousand
Longitude
Error
Latitude
true anomaly, <p, and the mean anomaly, 7, of the elliptic orbit. (See
Fig. 4.19.)
„ . 1 5 107 \
f(t) = l 2e - - e3 + 96
— e6 +
“ ' 4608
e7jsin 7(t)
(200)
47273
+ %+ e7 sin 7 7(f)
32256
Also
9ir
7(t) = — (T + t) (201)
T
where T is the (virtual) time from perigee passage to the injection point,
and
r = Ca3/2 (202)
which is simply Kepler’s third law. Here, “a” is the semimajor axis of
the ellipse, and C is a known constant.
Consequently, a(t) may be expressed as a function of five independent
parameters and time, t, viz.
E[v(t<)l = 0 (205)
where
Ae,- = e — e,-
Aa j ~~ a &i
Aco, = w — oji
A/3, = 0 - p{
STOCHASTIC EFFECTS 529
Zi = M,-Xi 4- Vf (213)
where
Zi = z(t<) = Aam(t 0
V,- = v(t,-)
a 1 X 5 matrix
“A e<
Aa,
x< = AT<
A 03i
M i -
R, = a2
P0 = E[x0x0r]
and
E[x0J = 0 .
x = f(x, r) (217)
where the t,- are, in general, not all equal; i.e., we are considering a
multipoint boundary value problem.
Let x(0)(r) be an initial approximation to the solution of Eq. (217).
The initial approximation may be any function that satisfies Eq. (218).
Then the (k + l)th approximation is determined from the kth approxima¬
tion by the linear differential equation
df .-(x<*>,t)
F..l‘>(r)
i; v '
= (220)
dXj
i, j = 1, 2,
and
|x(*+d—x(A:)|< e (222)
where
p(m)(To) = o
T > T0 .
x/*'+»(t) = E Ha<*+1>(t)
c,<A
+i>+ p/t +1)(r) .
q=l
Using the n boundary values given by Eq. (218), this may be written
as
where
E[w(t)] = w(t)
and*
Q(t) = 0 .
*The(hm-n) excess equations axe merely discarded, since they are redundant for present purposes.
STOCHASTIC EFFECTS 535
or, equivalently
We are given the power spectral densities for the signal and noise as
follows.
2 (T2ft
G zz(co) (242)
ft2 A O)2
Gnn(co) = k2 (243)
1
L(s) = (244)
Ts + 1 '
538 ELEMENTS OF MODERN CONTROL THEORY
E[e2(t)
] = ? = I\.(0)= —/ G„(»)
dco. (245)
Also, via Eqs. (63) and (241)
Gee(co)
= |L(jco)- l|2 G„(«) + IL(jto)|2G„(w). (246)
Combining the last two equations, we have
e2= — f°|L(jco)
- l|2G„(a)dco+ — [ |L(jco)|2
Gn„(w)
dco. (247)
Taking the first integral on the right-hand side of this equation, we have
— f |L(jco)—l|2Gii(co)
do;
2k L„
——f [L(jco)
—1][L(—
jw)—1]
G**(co)
dco
2k•Loo
j_ r r icoT i r_jMT_i 2^
2kloo L 1 + jcoTJ Ll - jcoTJ (/32
+ co2)
1 r _ 2jcr2 /3T2 co2dco_
27rjloo (1+ jcoT)03
+ jco)(l—jcoT)(/3
—jco)
This may be evaluated by the method given in Appendix D. The final
result is
o-2j3T
l|2G,*(co)dco
1 + T|8
2- /|L(jco)|2an(co)
2k Loo
dco
= 21
.
Therefore
o-2/3T
e-
(248)
1 + 0T
STOCHASTIC EFFECTS 539
k
T = (249)
<jy/20— k/3
a w (253)
The vehicle geometry is shown in Fig. 4.20; the symbols have the fol¬
lowing meaning.
I = vehicle moment of inertia
Ka = servoamplifier gain
m = vehicle mass
These equations are derived in Chapter One, “Short Period Dynamics,” in the companion volume,
Analysis and Design of Space Vehicle Flight Control Systems, Spartan Books, 1970..
540 ELEMENTS OF MODERN CONTROL THEORY
Tc = control thrust
W„ = wind velocity
a = angle of attack
6 = pitch angle
where ti is the distance from the ith mass to station j , and the summation
is taken over all discrete masses forward of station j.
By combining Eqs. (250)-(254), we obtain the transfer function
relating the output M,(t) to wind input W„(t) as follows.
where
where
M2(W„) W w(t2)
P(W„) = w. =
JX*(W„)_ _w„(to_
542 ELEMENTS OF MODERN CONTROL THEORY
X)
X2
X =
M<(W.)
= ^N P=1
Z W„W(t,) (261)
and
*«(W.) = i £ [Ww(p)(t,-)
- M,'(WJ][W„,(p)(ty)
- w(W„)] (262)
IN p=l
*N should be a fairly large number, perhaps a few hundred, for the results to be meaningful.
STOCHASTIC EFFECTS 543
Strictly speaking, this procedure is not correct, since the true means
and variances are given by
Mi(
Ww)=[. " - 00
f xi f(xi,x2,.
J - 00
, x„)dxxdx2. dxn (263)
and
=f. J — CO
[ [xf- m(Wv)][Xi
j — 00
- #i,-(W„)]
f(Xl,
x2,
. , xn) dxi . dx„ . (264)
However, for large values of N, the sample means and variances cal¬
culated from Eqs. (261) and (262) will be a close approximation to the
true values.
From Eq. (255), we find
t
Mr(My)
= E[My(tr)]= f g(tr —r) ftT(Ww)d
t (265)
^0
where
g(t) = £-'[G(s)] .
Equation (265) gives the mean value of the bending moment at station
j at time, tr.
The bending moment variance is calculated from
My(tr)
- Mr(My)
= f g(tr- t) W.(r)dr
= [ g(tr—t) fW.(r)
—Mr(W„)]
dr
Jo
544 ELEMENTS OF MODERN CONTROL THEORY
Similarly
a
t)[W.(t)
- HT(WW)}
dr [g(t.- f)[W..(f)
- Mf(W.)]
dr
*0
- dr dr .
o-r.(My)
= f J g(tr- t)g(t,- f)oT^(Ww)
drdr. (267)
Thus, the mean and variance of the bending moment response is given
in terms of the mean and variance of the wind by Eqs. (265) and (267).
Furthermore, since the wind statistics are Gaussian, so is the bending
moment response; viz.
My(t2) M.(My)
My = A(My) =
. My(t„) . _ Mn(My)_
STOCHASTIC EFFECTS 545
yi
ys
y=
yn j
Since there are structural design limits for any launch vehicle, it is
essential that one calculate the probability of exceeding a given value
of bending moment, M? , at various values of station j. Using the pre¬
vious results, we find that the probability of not exceeding the value
M* at any time of flight, tlt t2, . , tn, is given by
[yr - Mr(M,)]2(.
Prob [My(tr) <
2cr
rr2(My)j Jr
(269)
= 0.999 if a = 3 .
For the case under consideration, jur(My) and o>r(My) are available from
Eqs. (265) and (267), respectively; therefore, for a given M*, the appro¬
priate value of a is obtained from Eq. (270), thus permitting the evalua¬
tion of the required probability from Eq. (272).
The scheme herein presented, while having a certain rational and
intuitive appeal, has several drawbacks in practice. First of all, the
available evidence that the wind statistics satisfy a Gaussian distribution
is far from conclusive. Also, the validity of Eq. (265) is based on the
assumption that the system under consideration is linear and stationary,
which, of course, is not strictly true, except for very short time intervals.
Finally, the computations become extremely cumbersome for systems of
only moderate order. The method is useful, therefore, only for obtaining
crude estimates, which must be interpreted with caution.
In actual practice, the complete nonlinear time varying system dynamics
are used to simulate the vehicle on a computer, and a large number of
runs are made for different wind profiles. Time histories of critical param¬
eters, such as bending moment, angle of attack, and engine deflection,
then become available. Various criteria, such as not exceeding a limiting
value of a parameter 95 or 99 percent of the time, are used to evaluate
the system performance.
APPENDIX A
8 (x) = 0 x 0 (Al)
/ f 8 (x)dx = 1 . (A2)
" a
8 (x) = 0 x ^ 0
5 (x) = oo x = 0 .
0 < 6 < oo
1. 5 (x, e) = -«<x<£
= 0 otherwise .
547
548 ELEMENTS OF MODERN CONTROL THEORY
1
2. 5 (x, e) x > 0
e
= 0 x < 0 .
3. 5 (x e) = e (l/e)2 .
«V 7T
e sin2(x/f)
4. 5 (x, e) =
7T x2
/oo
g(x) 5(x - x0) dx = g(x0) (A6)
—00
where
h (t)
549
550 ELEMENTS OF MODERN CONTROL THEORY
where H(s) is the transfer function relating the output to the input.
Applying the Complex Multiplication Theorem* to Eq. (B3) yields
f0(t)= £~l[F0(s)]
= [ h(t- n) f»-(n)
dr,
Jo
(B4)
Via the change of variables, r = t — tu we find
Xi — e*1 = 0 (Cl)
Xi = x2 = fi
(C3)
x2 = e*1 s f2
with
xi(0) = 0 = a0i
(C4)
xi(l) = 0 = au .
Fuw(t) = 0 F./^r) = 1
(*)
(C5)
F„lS)(r) = e1' F2!W(t)= 0
u ,W(r) = 0
<k)
(C6)
u2(i)(r) = (I - x,(i))e‘
551
552 ELEMENTS OF MODERN CONTROL THEORY
o'
I_ X 0
(C7)
x<0)(r)
_x2(0)(r)_ _ 0 _
Then
F„(0)(r) = 0 F12(0)(t)= 1
F21(0>(t)= 1 F220)(t) = 0 .
Hn(1)(r) = H21(1)(r)
H.2(1>(t)= H22(1)(r)
H2i(r)(x)= Hn(1)(r)
H22(1)(t)= H12(1)(t)
Hn(1)(0) = H22^(0) = 1
Ht2(1)(0)= H21(,)(0)= 0 .
H„(1)(r)
=H22(1>(t)
=~(e-+e-)
(C8)
H12(1)(r)
=H21(1)(r)
=1(ex- e-x^)
.
Also,
fromEqs.
(225),
(C5),
and
(C6)
P,roW - p,“»
p 2U)(x)= p i(1)(r) + 1
P iO,(0) = p2(1)(0) = 0 .
STOCHASTIC EFFECTS 553
where H12(1)(r)
and Pi(1)(f)are given by Eqs. (C8) and (C9), respectively.
The calculations of successive iterations are straightforward. The
following table summarizes the results obtained after two iterations.
0 0 0 0 0
0.1 0 -0.04128 -0.04144 -0.04144
0.2 0 -0.07297 -0.07327 -0.07327
0.3 0 -0.09539 -0.09580 -0.09580
0.4 0 -0.10874 -0.10924 -0.10924
0.5 0 -0.11318 -0.11370 -0.11370
0.6 0 -0.10874 -0.10924 -0.10924
0.7 0 -0.09539 -0.09580 -0.09580
0.8 0 -0.07297 -0.07327 -0.07327
0.9 0 -0.04128 -0.04144 -0.04144
1.0 0 0 0 0
The last column is calculated from the closed form solution of Eq.
(Cl), which is known to be
Note that the order of the numerator is at least two less then the order
of the denominator. Also, it is required that the roots of h(x) be in the
upper-half plane. Any function, f(x), which contains only even powers of
x can be factored in the required form, since if x0 is a root of f(x), so is
(~x0).
The value of the integral of Eq. (Dl) is tabulated in Ref. 47 for n = 1,
2, . ,7.
The values of In for n = 1, 2, 3, 4 are given below
2 Bo Bi
h = -
2 Bo Bi
- a, b0 + a, b, -
I
2 Bo (Bo B3 — Bi B2)
554
STOCHASTIC EFFECTS 555
Mean:
m = E(xi)
M2 = E(x2)
Covariance:
Then
(El)
If we let
xi = E(xi|x2)
X Xi - Xi .
Then
Finally, if xi, x2, and x3 are Gaussian random vectors, and the pair
x2, x3are independent, then
Properties in Eqs. (El), (E2), and (E3) are derived in Ref. 57.
We now consider the discrete form of the optimal filter problem as
stated in Sect. 4.3 and repeated here for -convenience. Given the dynamic
system
The random vectors w(t) and v(t) are Gaussian with means and covari¬
ance given by
E[v(t) wr(r)] = 0 .
We are given the observed values, z(0), z(l), . , z(t). Kalman shows
that the optimal estimate of x(t + 1), given that z(0 ), z(l), . ,z(t) have
occurred, is merely the conditional mean
it follows from Eq. (E5) that z(0), z(l), . . z(t) is also a sequence of
Gaussian random variables with zero mean.
It is convenient to adopt the following notation, in a manner similar to
Eq. (E9).
= <t>(t+ 1, t) x(t 11 — 1)
which becomes
via Eq. (El 5). Applying Eq. (El), this may be written as
Finally
560 ELEMENTS OF MODERN CONTROL THEORY
Substituting Eqs. (E20) and (E21) in (El 8), and making use of Eq.
(El 7) yields
while
using the fact the x(t) w(t), and v(t) are all independent of each other.
Rearranging, simplifying, and eliminating K(t) via Eq. (E23), we obtain
REFERENCES
= 0, elsewhere.
2(72j
3
Gxx(w)
O)24-
where o-2and p are positive constants. Show that, for this case, the
autocorrelation function is
r xx(t) = cr2e~^lTl.
566 ELEMENTS OF MODERN CONTROL THEORY
4. It was stated in Sect. 4.1.4 that white noise is not physically realiz¬
able since it implies that successive values of the random function
are completely uncorrelated no matter how small the interval
between successive samples. To examine the physical impossibility
from another point of view, calculate the mean square value for a
white-noise signal and interpret the result.
7. Let x(t) represent the input signal to a stable filter whose transfer
function is H(s), and let y(t) denote the output of the filter. Ifx(t)
is white noise of unit amplitude, show that
Tri/r) = h(r)
where ILi/t) is the cross correlation function for the signals x(t) and
y(t), and h(r) is the weighting function of the filter. Consequently,
this is one way of determining the transfer function of a filter.
G‘a:x(a))
— 1 , |CO
| Ci)
o
otherwise
STOCHASTIC EFFECTS 567
= 0 , otherwise.
E[x2(0)] = 10 ft2
E[x2(0)] = 10 (ft/sec)2 .
This chapter is concerned primarily with the analytic tools for solving
optimization problems. The natural vehicle for this exposition is the
classical calculus of variations. Broadly speaking, the problem is formu¬
lated as follows. Assume a system whose dynamic properties are described
by a set of ordinary differential equations. Some type of initial and
terminal conditions are prescribed. One or more degrees of freedom are
available in the form of a control function that must be programmed such
that a stipulated function of the control and state variables is minimized
(or maximized).
In this fashion, we have the classical problem of Mayer* and the well-
established solution. Some qualifications are in order, however. First,
this “solution” is often in the form of a system of nonlinear differential
equations with two-point boundary conditions. The computational diffi¬
culties associated with this are far from trivial and often overwhelming.
Second, the classical formulation includes no constraints on the state or
control variables. This severely limits its usefulness, since, in the real
world, bounds on energy and magnitude of a control force are hard
realities. In some instances, a state variable must be contained within
prescribed bounds— such as limiting an angle of attack to ensure structural
integrity.
As a result of these limitations in the classical techniques, several new
approaches to the problem have been initiated in the last ten years. These
have led to the development of dynamic programming, the gradient
method, the maximum principle, and the generalized Newton-Raphson
technique. With this new body of theory, the range of problems that can
be resolved has been considerably enlarged. But as usual, few panaceas
are available.
Dynamic programming, which exhibits the greatest indifference to the
presence of “pathological” functions, is limited by computer storage
capacity.
The gradient method, one of the more powerful techniques, generally
yields only local extrema, primarily because small deviations from a
reference condition are analyzed.
The maximum principle, in which is included the effect of control
constraints, is a more elegant formulation of the classical Weierstrauss
condition in variational calculus. It is, however, completely identical to
the classical methods when the inequality constraints are included via the
Valentine condition.!")
*Also called the problem of Lagrange or Bolza, when expressed in slightly different fashion.
OPTIMAL CONTROL 571
^•(x, x, t) = 0 j = 1, 2, - , p(< n)
0)
x(t) = n vector
where t,- and tf are the initial and final times, respectively, and*
x, = x(t<)
X/ = x(t/)
d_
(') = dt (' ) .
J = [G(x,t)]/ i
= G(x, t) | (=%
ty ~ Cr(x,t)| t=<4- (3)
is minimized subject to the constraints, Eqs. (1) and (2).
The problem thus formulated is more general than is superficially ap¬
parent. For example in the problem of Lagrange, the function to be
minimized is
J = f L(x,x, t) dt . (4)
J t.
I
x„+1(t)
= J L(x,
x,t)dt (5)
with
XTl+l(tj) 0 (6)
J X„+l(t/) (7)
*The subscripts i and f are used exclusively to denote initial and final value, respectively, not the
components of a vector. Symbols other than i and f will be used to denote vector components.
574 ELEMENTS OF MODERN CONTROL THEORY
f
,] = [G(x, t)] + L(x, x, t) dt . (9)
'
t
the problem is put in the Mayer format, where J, defined by Eq. (3), is
minimized subject to the added differential constraint
<Pn+
1 = (Xi - K0(K2 - Xk) - xn+i2 = 0 . (12)
The Mayer format is, therefore, extremely general and finds wide
application in aerospace controls problems. A solution is obtained in the
following way. Form the augmented function
F=Zx^ (13)
j-i
where the \3 are time dependent functions called the Lagrange multipliers.
A necessary condition for the minimization of the criterion function
defined by Eq. (3) is that the Euler-Lagrange equations
d_ k = 1, 2, ••••, n (14)
dt
be satisfied.
The system of Eqs. (1) and (14) is subject to (2n + 2) boundary condi¬
tions. Of these, q are supplied by Eq. (2), while the remaining ones are
A dF . dF_
5G + A - X/t (15)
*-i dxk dxk
(17)
%
(18)
(19)
(20)
(21)
See Appendix B.
576 ELEMENTS OF MODERN CONTROL THEORY
(f-
\dxk
+ f)
dxk) t=tf
= 0. (22)
= 0 (24)
(25)
(28)
k=1 j=l C^Xfc
SXj
T - D = m(V + g) (29)
h = V (30)
T = a,* (31)
0 m . (32)
578 ELEMENTS OF MODERN CONTROL THEORY
Here
T ; - thrust
D s D (V, h) s drag
m = instantaneous mass
h = altitude
Y = velocity
g = gravity acceleration
Physical considerations require that the mass flow satisfy the inequality
Vi = h - V = 0 (34)
^ Y + g + (D ~ M = 0 (35)
m
v>8= m p —0 (36)
The last relation, which is in the form of Eq. (12), accounts for the
inequality constraint, Eq. (33). No special physical significance is
attached to the variable, a.
The problem variables are identified as follows.
xi = h
x2 = V
x„ = m
X4 = /3
X5 = a
OPTIMAL CONTROL 579
F —Xi(h—V) -j-X2^V
+ g+ m + 0)
(38)
+ \<[0(0m ~ 0) ~ a2]
/
and the Euler-Lagrange equations become
X2 dD
X, = (39)
m dh
X
X’ -“ —X
*' 4-
+ m W (40)
(0H — D)X2
Xa — (41)
nr
° =“(mX2
“ Xa)
+X4(0*~2/3) (42)
0 = a\a . (43)
Since F does not contain t explicitly, we have, from Eq, (23), com¬
bined with the Euler-Lagrange equations, (42) and (43)
^iV
—^g+ —
^x2+ X2
— 0=c (44)
hj =0 ti - 0
V, = 0 V/ = 0 (45)
m, - m0 m7 =
then J becomes*
J = — h/ . (46)
Since t, and h,are not prescribed, the transversality condition, Eq. (17),
yields
(47)
(48)
C = 0 (49)
— 1 + Xi/ = 0 . (50)
K = — X2 — X3 (51)
m
K = 0 and K = 0 . (52)
Taking the time derivative of Eq. (51) and combining with the Euler-
Lagrange equations, we obtain
dD \
m(D + mg) K = 0
^V+ D) K -f* Xi (V - M) D
(53)
+ ' v &v ~ mg i M
(V- M)
D+ (v^ - mgjn=0 (54)
must be satisfied. This, therefore, determines the variable thrust param¬
eter, (8 = — rii. In order to ascertain the physical significance of this
equation, consider a drag function given by
Ci V2
m
(Xi)_ = (Xi)+
(Xj)—= (x2)+
(56)
(X8)_ = (x8)+
(C)_ = (C)+ .
That is, the above parameters are continuous at the corner points of the
extremal arc.
Now since C = 0 and since V and m are also continuous, Eq. (44)
shows that 8 may be discontinuous provided that
—2A„[(5«)2 + m > 0 .
- 5^5 [(!“)!
+H 5 0
or, equivalently
K
> o . (59)
Pm - 2/3
p = Pm when K > 0
The last relation follows from the fact that for \4 = 0, a ^ 0 which
means that p may assume an intermediate value as shown by Eq. (37).
The quantity P (= — m) is in fact determined from Eq. (54).
It now appears that the solution is completely determinate. The six
equations, (34)-(36) and (39H41), together with the six boundary con¬
ditions
V4 = 0 V/ = 0
m, = 0 ni/ = mp
(62)
at Point © : K = 0, K < 0
at Point © : K = 0, K > 0 .
In this case, Eq. (61) indicates that Q < 0 at Point © and Q > 0 at
Point ©. Expressed in terms of Eq. (62)
According to Fig. 5.1, Vr < V2, so that the above two conditions are
incompatible. By similar arguments, we eliminate the Subarcs © - © - ©
OPTIMAL CONTROL 585
Figure 5.3. Possible Ambiguity in Thrust Program for General Drag Law.
X = f(x, u, t) (63)
Vi < uy < nj
(64)
j = 1, 2, - , m
J = Cr X, (65)
(66)
(67)
with
Xn+i(t,-) = 0 (68)
J = x„+i(t/) . (70)
Xn+i
= Vf d± (73)
U ’ **
^ X7!/^-i(t»/)
* (74)
-Xn(ti).
i:dt . /
(75)
Xn+1
= J‘if dt (77)
X,= -
*-i dx> (78)
j = 1, 2,
H = Z V; = W % (79)
J— 1
Our notation anticipates the interpretation of this quantity as the Lagrange multiplier.
590 ELEMENTS OF MODERN CONTROL THEORY
(80)
j = 1, 2, - , n
(81)
j = 1, 2, ••••, n .
Xy(ti) Xy{
(82)
i = 1, 2, - , n
Ay(tr) = -c,
(83)
j = 1, 2, - , n
dip
by (85)
dXj t—t
1
\ (t/) = - Oy - 7?by[x(t/)]
(86)
j = 1, 2,
t [*(t/)] = 0 (87)
x;'(t/) = Xy
(88)
j = 1, 2, - , q(< n)
X*(t/) = — c*
(89)
k = (q + 1), - ,n .
Xm-1 = 1
x„+i(ti) = 0 .
The final value of xn+i replaces ty in the criterion function, and the
added relation takes the form
»H-1
where I and K are scalar functions that are independent of u. Here it has
been assumed that the explicit dependence of the system equations or
criterion function on time, t, has been removed by defining an additional
state variable xn+i such that
Xn+l = 1 , xB+1(0) = 0 .
* Actually, the discussion that follows is applicable to any function of u as long as the form of Eq.
(92) remains.
OPTIMAL CONTROL 593
H* = I + u* K = Max(I + u K) = 0
u _ _ (93)
0 < t < t, .
I = I = I = •••• = 0
(95)
K = K = K = •••• = 0
dH(x, X, u*)
Xy = (97)
dx,
into Eq. (95), then it may happen that the optimal control in the singular
region reduces to a function involving the state variables only, viz.
n* = u*8(x).
*Here, and in the remainder of the chapter, the asterisk superscript will be used to denote tne
optimal value.
f The subscript s on u* is used to denote the optimal control in the singular region.
594 ELEMENTS OF MODERN CONTROL THEORY
<f>j= ~ fj = 0 (98)
then
A*=-LXy^-. j OX/t
(100)
Comparing this with Eq. (78), we see that the choice of notation is
indeed justified.
It should be emphasized that the maximum principle provides only a
set of necessary conditions, much the same way as the Euler-Lagrange
equations and the Legendre-Clebsch condition provide only necessary
conditions in the variational calculus. Sufficient conditions are hard to
come by (except in the case of linear systems). Nevertheless, this is
usually a mathematical luxury. In practical situations, physical con¬
siderations will generally confirm uniqueness and sufficiency.
OPTIMAL CONTROL 595
Xi = x2 = fi (101)
x, = (102)
x8
x8 = — 0 = ft . (103)
Xi(ti) = 0 ti = 0
x,(t<) = 0 x»(t/) = mp
where xi = h, x2 = V, x8 = rn .
We seek to maximize the final altitude. Consequently, the criterion
function to be minimized is
\i(t/) = 1. (106)
(107)
x3 dxi
X2 = — Xx + -~x2 (108)
x3 dx2
% V_ (0M - D) (109)
A 8 — - -- A2 .
X3
The hamiltonian is
H = XxX2+ X2 “ D) - gj - X,/3
(110)
=
[Xx (| +g)]+0 ~ ]•
x2~X2
This function is maximized if
when K > 0
(111)
0 = 0 when K < 0
where
TT X2M .
Iv — — Xg . (112)
x8
X2|u
— — x8 = o (113)
x8
/<3D D\
- x8\! + + - )x2= 0 (114)
Xix2—X2(— + gV - 0 . (115)
Xi = xj, = fi (117)
7 = prescribed constant.
It is required to determine the form of ui(t) and u2(t) such that the
system is brought to the equilibrium position, xi = x2 = 0, in minimum
time from an arbitrary initial position
xi(tf) = a
x2(t<) = b .
x8 =
and
x8 = 1 = f8
x»(ti) = 0 .
J = Xg(t/) .
The components of the costate vector, x, are given by Eq. (78), viz.
Xi = 0 (121)
X2 = — Xi + u22X2 (122)
X8 = 0 . (123)
x,(t,) = -1 .
OPTIMAL CONTROL 599
Xa(t) = -1
ui = — sgn X2 . (126)
H=~1+XlX2
+X2[(S)
"X2
(Ua
+1~)] (127)
a. If x2 > 0
then u2 = 1
b. If x2 < 0, then
u2 = 1 if |p| > 1
= 7- if |p| < 7 •
a. If x2 < 0
then u2 = 1
*sgn y = 1 if y > 0
= - 1 if y < 0.
600 ELEMENTS OF MODERN CONTROL THEORY
If x2 > 0, then
u2 = 1 if p > 1
= p <- IV
*4-» Q. IV r~H
= 7 ~o Al (128)
where
P 2x2 .
xx(t<) -- a xi(t/) - 0
x2(t.) - b x2(t/) = 0
obtained from Eq. (91), are sufficient to determine the optimal mand u2.
For systems of moderately high order, the corresponding computa¬
tional tasks are not at all trivial, even with computer assistance. In the
present case, however, the phase plane representation may be utilized to
advantage in order to exhibit the salient features of the solution.
We observe parenthetically that the fundamental difficulty in starting
a solution is that the sign of X2is not known initially. However, the form
of Xi and X2is known, since Eqs. (121) and (122) are readily integrated;
viz.
X, =
X2
= eA{/32
—ft3le~A
dt (129)
where
OPTIMAL CONTROL 601
and /8i and /32 are constants of integration. The significant information
available from an inspection of Eq. (129) is that X2 cannot change sign
more than once.
Now for purposes of phase plane representation, we express Eqs. (117)
and (1 18) as
dxx = - - - . (130)
u2(u2 x2 a Ui)
From condition ( 126), we know that uj = 1, while Eq. ( 128) shows that
u2 = 1 whenever|x2| < “
regardless of the sign of X2. Consequently, in the phase plane, only two
trajectories can enter the origin. One trajectory, Li (see Fig. 5.4), is
obtained from the solution of Eq. (130) with uj = u2 = 1. The other
trajectory, Lj , is given by the solution of (130) with uj = -1 and u2 = 1.
We solve Eq. (130) by working backward; that is, we take xi(0) = x2(0) =0.
The resulting equations are
a
xi = — x2 — aln
x2 + a
Xl — x2 + aln
a — X2
* a
satisfying this relation until the trajectory reaches Point ®, after which
= 1.
602 ELEMENTS OF MODERN CONTROL THEORY
Continuing the analysis along similar lines, we see that the phase
plane is divided into six semi-infinite regions as shown in Fig. 5.5. In
the half plane to the left of Li-Li , the control function, uj = -1 , while
in the other half plane, uj = 1. Since, as shown earlier, \2 can change
sign only once, a trajectory, once it hits L, — L[, will remain on it until
reaching the origin. Several typical trajectories from arbitrary initial
points are sketched in Fig. 5.5.
A complete solution for the case where
x2(t,) = 0 7 = 0.5
f = f(x) (131)
606 ELEMENTS OF MODERN CONTROL THEORY
where
df_
dxi
Vxf
(133)
df_
dXn _
dx = Kv»f (134)
df = K(V*f)2 . (135)
|f(x<r+l>)
_ f(x(r))| < e
(137a)
e = small preassigned positive constant.
The vector x(r)is then the one that minimizes f(x) to the desired degree
of accuracy. If x(0; is the exact minimizing vector, then
Vxf(x(0>) = 0 . (138)
(139)
(140)
(141)
(142)
*See Appendix B.
fSee Appendix B.
OPTIMAL CONTROL 609
However
ftf d
51 /Jt. (5x)rVxL - - (V,L)
dt
(144)
%
[ ]=^-v-
x
(t45)
we may write
r‘f
51 = / (5x)»-[L],dt. (146)
J t.
%
df 51
dx <=> dx(t)
V*f L .
5x = K[L], (147)
*See Appendix C.
610 ELEMENTS OF MODERN CONTROL THEORY
as the variation in x(t) that produces the greatest change in 51. Here, K is
a constant that is negative if it is desired to decrease I, and positive
otherwise. The gradient is evaluated along the nominal path x(1)(t). The
variation in I is expressed, therefore, as
(148)
V*f(x) = 0 (149)
(150)
since 5x is arbitrary between the end points, t< and tf. In other words, a
necessary condition for the extremum in function space is that the
generalized gradient vanish.
In order to relate this condition to the necessary conditions derived
via the conventional variational approach, we will solve the problem
using the methods of Sect. 5.1.1. To this end we define
(151)
J = Xn+i(t/) (152)
Xn+l(t<) = 0 . (154)
OPTIMAL CONTROL 611
F = A(xn+1 - L)
we find
dF dL
- = — X -
dXj dXj
dF _ _ a
dXj dx.
while
dF
dXn+l
dF
= 0 .
dxn+i
d / dL \ dL
-j- ( X ) —X = 0
dt \ dxy/ dxj
j = 1, 2,
X = 0 .
The last relation shows that A is a constant, which means that the
preceding equations become
dL
d_
dt dx; (155)
J = T 2,
These are the Euler-Lagrange equations for the problem at hand. Written
as
(156)
we find by comparison with Eq. (150) that the vanishing of the gradient
in function space is another way of expressing the Euler-Lagrange equa¬
tions.
For purposes of numerical computation, the evaluation of the gradient
employing the operator defined by Eq. (145) involves the calculation of
higher order derivatives. This is never an advisable numerical procedure,
and other computational devices must be sought. It turns out that the
use of Green’s functions leads to an efficient algorithm for the numerical
computation of the gradient. We turn to a discussion of the basic ideas
involved in this approach.
Consider the system described by
Xy = fy(X, U, t)
(157)
J = J[x(ty)] (158)
decrease, as the case may be) in the performance criterion, J. For 5u and
8x sufficiently small, we have, to first order
dfj
it (SXi> Z — <5
x* + J2 p- (160)
fc»=i
dXk £=\ oMf
(j = 1, 2, % n)
>
5x(tj) = 0 (162)
where
*See Appendix B.
614 ELEMENTS OF MODERN CONTROL THEORY
ah dh_ dii_
dxi dx2 dxn
af2 af.
dxx dx2
(163)
A(t) =
din di, dU
dxt dx2 dxn
af, df2
aux dU2
(164)
B(t) =
5x(t/) = f <t>(t;,
t) B(t)5u(t)dt (165)
where t*) is the state transition matrix for the system, Eq. (161). It
satisfies the relations
OPTIMAL CONTROL 615
Now /
which reduces to
where
=0 =M (170)
5J= J (5u)r
Br(t)
f>r(t
/,t)c dt (171)
X = —A(t)X (172)
which is the vector matrix equation, adjoint to the system, Eq. (161). If
we take the initial conditions on this to be
X(t/) = c (173)
where o is given by Eq. (170), then the solution of Eq. (172) is given by
where ddt, t*) is the state transition matrix for the system, Eq. (172), and
satisfies
l-¥(t,
dt
t*)= -A(t)¥(t,t*) (175)
Notice that Eq. (174) is integrated “backwards’’ from the given final
conditions. A(t) is, therefore, a known function (which, of course, de¬
pends implicitly on the reference trajectory).
Now it is known!155) from the theory of adjoint equations that
r*f
8J= J (8u)T
BT(t)X(t)dt . (179)
If we define
H = f7’A (180)
then
m
= V„H = BrX . (181)
du
5.J = (182)
OPTIMAL CONTROL 617
This now has the form of an inner product in function space, and, in
the light of the previous discussions, we interpret (dH/du) as the gradient
in function space with respect to the control vector, u(t). Since the latter
is arbitrary, the condition for the vanishing of 5,J(which thereby ensures
the J is an extremal) is that the gradient vanish; viz.
It will be shown later (Sect. 5.1.5) that this condition is precisely the
maximum principle discussed in the previous section. Thus, H, as defined
by Eq. (180), is indeed the hamiltonian of the system, and A is the vector
Lagrange multiplier. We note also that the quantities, dH/du^, are the
Green’s functions (sometimes called “influence functions”) for the sys¬
tem; these indicate how small perturbations, <5u,affect the performance
function, J.
It now follows that a maximum change in 6J is obtained when
Su = KV„H (184)
5J = K (185)
(186)
*We recall that for some vector, b, the notation ba means brb.
ELEMENTS OF MODERN CONTROL THEORY
618
5J =f' (Su)Tdt+IX
|\*- j' (6u)2
dtj (187)
= J (5u)r - - ]c
~ ldt+ ^ '
For an extremal 5J, the first variations must vanish; i.e.
S(«J) -I'M [@ - H * - °
which leads to
(188)
1. Terminal:
0 (191)
/3 = p + 1, p + 2, - , q .
2. State Variable:
k = 1, 2, - , r .
3. Control:
vt< H (193)
t =1,2, - ,m .
The final time, tf, is generally free (unspecified), and the initial condi¬
tions, xy(t») , may be specified or left open to optimization.
The simplest type of constraint, as far as the gradient method is con¬
cerned, is the inequality constraint, Eq. (193), on the control variables.
These are accounted for in the computational sequence, as follows.!55)
Assume that the nominal control vector unom(t) satisfies Eq. (193) for
all t.* The gradient technique then generates a new control
*Since the numerical procedure must of necessity deal with discrete instants of time, u"om(t)
satisfies Eq. (193) at all discrete time instants involved in the computation.
620 ELEMENTS OF MODERN CONTROL THEORY
vl < + <5u^(tjt)< m
{l = 1, 2, ••••, m)
V^t*) = h if
and
x„+*(t) G*(x, u, t) dt
(194)
k = 1 2 r
with
= g k~ if gfc > 0
k = 1,2, r
and add
= G*(x, U, t)
(196)
(k = 1, 2, r)
Xn+fc(t,) 0
(197)
k = 1, 2, - , r .
OPTIMAL CONTROL 621
x„+*(t /) — 0
(198)
x = f(x, u, t) (199)
x,(t.) = a,
(200)
j = 1, 2, ••••, s(< n)
where the (n — s) initial conditions, which are not specified, are left open
to optimization. At the final time, t/; the constraints
= p+1, P + 2,
t = 1, 2, - , m .
such that the restraints, Eqs. (201)-(203), are satisfied, given the initial
conditions, Eq. (200).
The problem thus formulated is of wide generality. Many realistic
optimization problems in aerospace guidance and control can be ex¬
pressed in this format.
In presenting the solution to the problem, we will limit the discussion
to the main results. The complete derivations are quite lengthy, requiring
many digressions that are beyond the scope of the present discussion.
For a detailed treatment, the reader should consult the appropriate
references.^, 42,55, 75, 76,103,135, 153)
Suppose now that we arbitrarily select a nominal u(t) that satisfies the
constraint in Eq. (203) and that thereby yields a nominal trajectory, x(t),
satisfying the initial conditions of Eq. (200) but not necessarily all the
terminal constraints, Eqs. (201) and (202).
The problem is now taken as one of minimizing the function*
Q
The values of J and <py are, of course, obtained from the nominal trajectory.
OPTIMAL CONTROL 623
where
K0 = —1 if J is to be maximized
= + 1 if J is to be minimized
The basic idea here is that if the I\7 are large (compared with K0), any
control, u(t), that seriously violates terminal conditions of Eqs. (201) and
(202) will give rise to large values of K7^72, with the result that the next
iteration on u(t) will be biased in favor of satisfying the terminal con¬
straints rather than optimizing J. Consequently, the more closely the
terminal constraints are satisfied, the smaller the value of K7^72 (what¬
ever the size of K7), with the result that subsequent iterations on u(t)
become biased in favor of optimizing J. This notion, which is intuitively
plausible, is also mathematically legitimate. 056, 157)
Now since the final time, tf, is free, one of the \py = 0 may be singled
out as a stopping condition and is, therefore, not included in the sum¬
mation term of Eq. ( 205). Call this stopping condition
H = f*\ (207)
where X satisfies
X = —A(t)X (208)
with A(t) given by Eq. (163), we take, for the initial conditions on X
w^(t)[<5u^(t)]2dt (210)
e = prescribed constant
5u^(t) =
_l _ dH
2/uw^(t) duf (212)
1 = 1, 2, • • • •, m
_L_ (213)
Vw^(t)
(214)
(215)
V
e dH
5u^ = (216)
du(
OPTIMAL CONTROL 625
dP = eV (217)
(218)
j = s + 1, s + 2, - , n .
It is noted that the \y( t<) are the Green’s functions for the initial con¬
ditions.
The general computation procedure is now summarized as follows.
8. Calculate P from Eq. (205), and check whether this is less than
the values of P from the previous iteration. If not, reduce the
size of e and return to Step 6.
rj2< <j
I'Pa]< Pa
a = 1, 2, ' • • •, P
i0< o
0 = 1, 2,
h = V sin y (221)
R = V cos 7 (222)
r0 + h /
lh = —(8 (223)
with
T = V£/3 (224)
g = go (225)
ro + h /
The control variables are /3 and a, i.e., the thrust magnitude and
orientation.*
*Short period dynamics are neglected; therefore the thrust angle and angle of attack are equivalent.
OPTIMAL CONTROL 629
V(t.) = 0
7(t<) = tt/2 V(t,) = Vr
R(t<) = 0 7(t/) = 7t (230)
h(t,) = 0 h(t/) = hr
m(t<) = m7
t< = 0 t / = free
Q, = h(t/) — hr = 0 (231)
h * V(t,) - Vr = 0 (232)
^2 = 7(t/) - 7r = 0 . (233)
The inequality constraints, Eqs. (228) and (229), involve state variables.
We, therefore, define
x6 = Gj dt (234)
*7 = G2 dt (235)
630 ELEMENTS OF MODERN CONTROL THEORY
where
Gi = 0 if gi < 0
(236)
= gi2 if gi > 0
G2 = 0 if g2 < 0
(237)
= g22 if g2 > 0
and
T + L sin a — D cos a T
gi - - la (238)
mg
L cos a + D sin a T
g2 GiV • (239)
mg
x6 = Gi (240)
x7 = G2 (241)
x6(t<) = 0 (242)
x7(tf) = 0 . (243)
In this fashion, the state variable inequality constraints, Eqs. (228) and
(229), are transformed to the terminal constraints.
= xe(t/) = 0 (244)
- x7(t,) = 0 (245)
x2 = 7 u2 = /3
OPTIMAL CONTROL 631
x8 = k
x4 = R
x6 = m
+
Xi cos x2 (247)
2 cos <pe = t2
r0 + x8
x8 = Xi sin x2 = f8 (248)
(249)
(250)
(251)
(252)
where
Gi = 0 if gi < 0
= gi2 if gi > 0
G2 == 0 if g2 < 0
= g22 if g2 > o
^4 = X7(t/) = 0 (256)
stopping condition
= Xg(t/) — hT = 0 (257)
xi(t.) = 0 xs(ti) = mi
x2(t<) = tt/2 XjCtt) = 0 (258)
x3(ti) = 0 xj(ti) = 0.
x4(ti) = 0
J = X„(t/) (259)
often the case with the appearance of a new idea, a flood of papers dealing
with extensions, ramifications, and applications has resulted. The theory
has helped influence the development of such diverse areas as manage¬
ment science, information theory, sequential analysis, optimum filtering,
adaptive control, learning theory, optimal trajectories, and variational
calculus. The passage of time, however, has had a sobering effect on
some of the overly optimistic claims made in the early stages of the
theory’s development. Space limitations preclude discussion of all facets
of the theory save those immediately relevant. We shall content ourselves
with the barest outline of the main ideas and discuss a variety of appli¬
cations that hopefully will clarify the method and exhibit its utility. In
the next section we will show how the basic premise of dynamic program¬
ming includes, as a special case, each of the optimization techniques thus
far considered.
Of fundamental concern in what follows will be the concept of a
multistage decision process. At any one stage in this process, one may
make a decision (select a control function), following which the next
stage is reached. Successive stages are related by a known transformation.
Each stage is characterized by a state vector, and each decision results in
some cost. These ideas may be made explicit as follows.
Let the state of a dynamic process be characterized by a state vector,
x. Suppose that successive states are related by
and incur a cost, g(c, u(1)), with (N — 1) stages left in the process. If we
proceed in an optimal fashion from this new state, the cost incurred is
WAr_x[T(c,u(1))], by definition. By way of the principle of optimality, we
see that for some specific value of u(1) we incur a cost
Consider now the function, Wx(c), i.e., the optimal value of J with
only one stage to go. It is easy to see that
*For definiteness, we consider the problem of maximizing J. The argument, of course, holds
equally well for minimization.
OPTIMAL CONTROL 635
X = X2 + u (267)
where x and u are both scalars, determine u(t) such that the function
(268)
is a minimum; the final time, tf, is given. The control function u(t) must
satisfy the inequality constraint
dx = (x2 + u) dt (270)
t / = N dt . (272)
A = dt
and this represents the “'initial condition for the recurrence relation in
Eq. (274).
We now examine the computational sequence in some detail. To do
this, we must decide on the range of interest for the state variables and
the number of values of u in the range -- 2 < u < 2. For present pur-
poses, we will consider a painfully simplified version wherein
/
c Wi(c) u
-5 0.3 -2
-4 0.3 -1
-3 0.2 -1
-2 0.1 -1
-1 0 -1
0 0 0
1 0 1
2 0.1 1
3 0.2 1
4 0.3 1
5 0.3 2
W2(3) = 0.50; u = 1 .
C Wjv(c) U
J = g[x(t/)] . (276)
In this case, we define
W„(c)= MinW*_![T(c,
u)] (2?7)
In other words, with zero stages to go, the optimal W0(c) is merely the
value g[x(t/)], where c replaces x(t/).
For the functional equations thus far considered, it was possible to
write explicit initial conditions; that is, the number of stages in the
process was known beforehand. This is not always the case. To examine
this situation, consider the problem of driving the solution of
±1 = xs
xi(0) = Ci
x2(0) = c2
OPTIMAL CONTROL 639
a < ui < b .
A = dt
W(ci, c2) = the time required to reach the equilibrium region (xi2 T
x22) < e, starting in state (ci, c2), and using an optimal
policy
also for a range of values of cx and c2, storing the improved policy ui(1).
It can be shownO6) that
|W(,)(cj,c2) - W(#-1)(Cl,
c2)|< v
(280)
dh = VA (281)
A = dt
We have to take account of the fact that the control function satisfies
the relation
0 = —m (282)
where
00 = 0
and that
_ (dm)*
Pk — - (285)
A
0k = ~k (286)
m(0) = m0 + mF(0)!
where
642 ELEMENTS OF MODERN CONTROL THEORY
mf = fuel mass
m,(0) = N .
/ (3ic[i — D
WN(V, h) = MaxlVA+ War-k V
Hk 1
+
\m0 + N ;)a,h+vaJJ. (287)
y + ~ D
\m0 + N
h + VA .
\ g <1
i
W N—k -d + > o
_ \m0 -j- N /
initial condition, W0(V, h), is easily obtained since, with zero stages to
go, the altitude acquired is simply the solution of the system
V =
D(V, h)
- g(h)
m0
h = V
W0(V, h) = hr(V, h)
V2
w„(V, h) = -L .
2g
n = 2000 ft/sec
= 25 slugs/sec
A = dt = 0.1 sec
N = 9000 .
The solution time for this problem would be on the order of a few
minutes on a computer of the 7090 type.
h = V sin 6 (289)
where
V = velocity
D = drag
T = thrust (constant)
m = mass (constant)
' r
mg
g = gravity acceleration
h - altitude
d = angle of thrust vector with respect to horizontal reference.
dV =
T - D(V, h)
— g sin 6 A (290)
m
dh = VA sin 6 (291)
A = dt
and defining
T - D
W(Y, h) = A + Min V +
6
( m
?)a,
— g sin 9 JA, h + VA sin 9 . (292)
A =
K £> (293)
ex* - °)
w = mg .
646 ELEMENTS OF MODERN CONTROL THEORY
Imagine now that the h — V plane is subdivided into grids with incre¬
ments dh and dV. We suppose that the mass point representing the
airplane moves in discrete steps from one node to the next, either straight
up (V = constant) or horizontally to the right (h = constant). At each
node point, a decision must be made whether to proceed vertically up
or horizontally to the right. The criterion, of course, is that the grid
from the current (initial) V and h to the prescribed VT and hrbe tra¬
versed in minimum time. The “costs” associated with the respective
alternatives are
_ wdh for constant V
~ V(T - D)
(294)
wdV
- tor constant h
g(T - D)
via Eq. (293).
Let us note at the outset that the insensitive pragmatist might be
tempted to enumerate all possible paths and then select the minimum
time path by direct comparison. In the case of a p by q grid, the total
number of possible paths under the conditions stipulated is
(p + q) !
(p - 1) ! (q + 1) ! '
wdh
A. + W[V, h + dh]
V(T - D)
V(T - D)
OPTIMAL CONTROL 647
(i.e., this is the time consumed in going from h to h + dh) and arrive at
the new grid point, (V, h + dh). Proceeding optimally thereafter, the
cost is W(V, h + dh). This is alternative A. Alternative B is arrived at
by similar reasoning. Finally W(V, h) is taken as the smaller of the two
alternatives A and B. Since the terminal values of velocity and altitude,
Vr and hr, are specified, we see that
x wdV
W(Vr - dV, hr) (296)
g[T - D(V r, hr)]
wdh
W(Vr, hr - dh) (297)
V r[T - D(V r, h r)J '
x = V cos 6
x = horizontal distance
In the methods presented thus far, only the main results have been
stated, together with various examples illustrating their application. The
development of detailed derivations in each case would require a docu¬
ment the size of a textbook rather than a chapter. Nevertheless, the
omission of derivations constitutes a pedagogic deficiency that precludes
complete comprehension. Furthermore, a superficial examination would
seem to indicate that the four optimization techniques considered thus far
are basically unrelated. It may, in fact, be shown that not only are these
648 ELEMENTS OF MODERN CONTROL THEORY
techniques intimately related, but that they all fall out as special cases
from a more general point of view—namely, they are all necessary conse¬
quences of the Principle of Optimality. We proceed to demonstrate this
in the following way.
Let the system dynamics be given by
x = f(x, u, « (298)
x(t,) = £ (299)
where, as before, x and u are the state and control vectors of dimension
n and r, respectively. It is required to choose u(t)such that a function of
the terminal state
iA(x, t) = 0 . (302)
(304)
aw _ / aw \ (305)
at Vdx )
(306)
(307)
X = - — (308)
dx
(309)
(310)
*See Appendix C,
OPTIMAL CONTROL 651
u = u(x, X, t) . (311)
W - YV(x, t) . (312)
Finally, eliminating W between Eqs. (311) and (312) yields the opti¬
mal u as follows.
u = u(x, t)
Noting that
(313)
652 ELEMENTS OF MODERN CONTROL THEORY
But
d\ _ / dX\ f dX (314)
dt \ dx / dt
Also
In other words
(315)
dX
+ 0 . (316)
dt
It is not difficult to show that this is the matrix form of the Euler-
Lagrange equations.
For this purpose, consider the augmented function defined by Eq.
(13), viz.
F = X*(x* - fO •
fc-1
We obtain directly
— - - t X,
dXj k=x \Xj
j = 1, 2, % ,n .
_d
dt
j = 1, 2,
may be expressed as
*;+ i: - o
k=i dXj
j = 1, 2,
Now by defining
H = \rf
we have
~ xT
(™)
V dx J
f = Min -Xrf - Max Xrf Max H
u u u
dx _ aH
dt ax
dX dH
dt d x
xy(t/) = 7 i
W(x, t,) = J
dW = Z — dxy + — dt = 0 (317)
i dx,- at
and
#- £ 4t dx, + * dt - 0 . (318)
j dx, at
Here
j - P + 1, p + 2, . , n
dxa = 0.
»“*/
dW = dXy = 0
j = p + 1, p + 2, • • • n .
Since this must hold for all arbitrary variations, dx>, it follows that at
the end point
/ a.J \
dW
= Xy(t/) =
dJ Ut hi
dXj dx, ( df N
\ dXj
)
(319)
V at ;
*"•/
j = p + 1, p + 2, . ,n .
Therefore
656 ELEMENTS OF MODERN CONTROL THEORY
fd J N
) dt
dW dJ
Adxa (320)
dt at ( d* N\ dt
\ dxa ,1
Note that
\n -1 = 0.
,=</
d_
dt
results are useful not only because they solve specific problems, but also
because they may be taken as first approximations to more complex
problems. This theme will be further developed in Sect. 5.3.
x = Ax + Bu (321)
V = Gx (322)
x(0) = £ (323)
where
The problem is usually stated in the following form. Given the system,
Eqs. (321) and (322), with the initial condition in Eq. (323), determine
the control vector, u(t), such that the performance index, Eq. (324), is a
minimum.*
+ XT(ty) Mx(t,)
Note that W(x, t) is a function of the current state and the current
time. Applying the Principle of Optimality, we find that W(x, t) satisfies
the recurrence relation
W(x, t) = Min
u
^xrGrQGx
+urRu^
A+W^x
+dx,
t +A^(326)
A = dt .
W^x
+dx,
t +A^=W(x,
t)+ dW (x, t)
dx
dx
dW
dt
dW (x, t) dW
W (x, t) + (Ax + Bu) A + — A
dx dt
dW
dt
= Min
u
( XT’ QGx -I- ur M?y (Ax + Ru) . (327)
Equation (327) is the Hamilton-J acobi equation for the system which
could also be derived from the maximum principle. Despite its formida¬
ble appearance, Eq. (327) has a closed-form solution that may be readily
obtained as follows. From Eq. (327), we write the two equations
au
^xrGrQGx
+urRu (Ax -f- Bu) = 0 . (330)
Now Eq. (329) is satisfied only for that value of u obtained by solving
Eq. (330). Thus u, when substituted in Eq. (329), yields an equation
that is completely equivalent to Eq. (327). Performing the indicated
operations in Eq. (330), we find
_a_ = o
au
aw
2Ru + Br — =0
ax
or, finally
u
*
- — R_1 Bt (331)
2
form
dW
= 2 P(t) x
dx
aw
— = xr P(t) X
2 PA = (PA + A^ P) -+%(PA - AT P)
where
Now let
a = Ab
&T = AT
and
P(t/) = M. (336)
P = ZY-1 . (339)
V = FV (340)
where
662 ELEMENTS OF MODERN CONTROL THEORY
Z
V a
Y
—GT QG
F
-BR-1 Br A
we find
where V0(0) replaces V2(0)Vr1(0). The matrix V0(0) is evaluated using the
boundary condition of Eq. (336).
Equation (342) is the solution of the matrix Riccati equation, (335).
Having P(t), we evaluate the optimal control vector, u*(t), from Eq.
(331), making use of Eq. (333), viz.
u*(t) = —K x(t)
(345)
Iv = R_1 Br P, a constant matrix.
x = g(x, t) + u (346)
x = g(x, t) (348)
|*|| = ^ ^ = o
1 11 llxll IJxll
which leads to
xT g(x, t) = 0 . (350)
The results that follow were first obtained by Athans.(9°) His deriva¬
tions, however, relied heavily on function space methods and were not in
the “mainstream” of optimal control theory.
A simpler and more direct development is possible by application of
the basic concepts of dynamic programming, as follows.
Define
W(x) = Min
u
A + W x + (g + u)A ]}' (351)
aw(x)
W[x + (g + u)A] = W(x) + (g + u)A + 0(A2)
dx
0 = Min
u
1 + (S?)' (g + u) (352)
(353)
(354)
(355)
aw l
x
dx 7
x(t)
u*(t) = — (357)
x(t)||
(um) dt (359)
u*(t)
HI . x(t) (360)
U ||x(t)||
x(0) = c
if = prescribed.
OPTIMAL CONTROL 667
The optimal trajectory. x*(t), is the solution of Eq. (346), with u*Ct)
given by Eq. (360). Note that now the optimal control depends explicitly
on the initial state and the prescribed time, t/, whereas previously the
optimal control was a function of the current state of the system only.
This last case, the minimization of the integral, Eq. (359), is a type of
minimum energy problem.
The above results are quite significant, since closed-form solutions for
nonlinear optimization problems are hard to come by. More important,
several cases of practical interest in aerospace control are indeed character¬
ized by norm-invariant properties. The latter is often a consequence of
conservation of momentum. In particular, two classes of norm-invariant
systems are the following.
||u(t)|| < 7
||u(t)|| < y
11 xi = (E — I3) x2 x3 + T! (365)
13 x3 = (Ii — E) xi x2 + T3 (367)
where Ti, T2, and T3 are the components of the torque vector T.
ELEMENTS OF MODERN CONTROL THEORY
668
yi = Ii xi
y2 = I2 x2 / (368)
y8 = Is xa '
yi = | a )y2y8+ Tj (369)
Quantities yi, y2, and y3 are the components of the angular momentum
vector, y. It is easy to show that when Tt = T2 = Ts = 0
i ||y(t)||= 0.
dt
(372)
In fact, in the present case, this is merely a statement of the Principle of
Conservation of Angular Momentum. Now if the constraint on the
torque vector is of the form
the theory of Sect. 5.2.2 is directly applicable. Therefore, the control law
y(t)
'J'* (374)
||y(t)||
will bring the system from an arbitrary angular momentum to zero
angular momentum in minimum time. This corresponds to stopping the
tumbling motions of the body in the shortest possible time.
Equation (374) shows that the torque vector takes on its maximum
absolute value and is directed opposite to the angular momentum vector.
OPTIMAL CONTROL 669
Aui cos u2 ,
Xi = = II (375)
X3
Aui sin u2
x2 = - g = f2 (376)
X3
XS = — aUl = f3 (377)
X4 = Xi - f4 (378)
X5 = x2 = f6 . (379)
Here
x4(t) - range
xs(t) = height
m(t)
x3(t) = nondimensional mass =
m(0)
T
ui(t) - nondimensional thrust = tjt
i v
A =
m(0)
a =
m(0)ju
functions can be obtained, and the complete solution requires that one
solve a set of differential equations with prescribed initial and terminal
boundary conditions.
The development that follows is based on the work of Isaevd64) and
includes the results of many other investigators as special cases.
We formulate the general problem as follows. Given the initial condi¬
tions
x*(0) = ay
(382)
i = 1, 2, ',5
find the control (ui, u2) that transfers the system described by Eqs.
(37 5)-(379) from the given initial state to a prescribed final state such
that the function
J = £ Cy
Xy(ty) (383)
j=i
L — (384)
fc=i dx.j
— X4 (385)
— x6 (386)
X4 = 0 (388)
672 ELEMENTS OF MODERN CONTROL THEORY
H = Z Xyfy
j-1
or, in expanded form
H = ux
-(
X3 \
Xxcos U2 -T X2sin u2 J — a\, )- - X2 g
(390)
+ X4 Xj + Xg x2 .
H = ux
-( Xx cos u2 +
)-
X2 sin u2 ) — a\
1
X
(391)
Ux = 1 > 0
< 0 (392)
u2 = tan-1 — (393)
Xx
and minimized by
ux = 1 $2 > 0
= 0 <t>2< 0 (394)
u2 = tan-1 —
X,
, Xl 2T
= — tan 1- (395)
X2 2
OPTIMAL CONTROL 673
where
X4 — x40 (398)
X6 = X50 (399)
and, in turn
then after substituting Eqs. (394) and (395) into the set of Eqs. (375)-
(379), (385M389), we obtain
.* AAi Sg *^2
(404)
Xl x8(Xi2
+ X22)1'2
.* A\2 Sg ^2
x, = — - -- — g (405)
X8(Xj2+ X22)1'2
X3 = — a Sg <t>2 (406)
X4 = Xi (407)
Xg = x2 (408)
v< II 1 (409)
•/<
w II 1
(410)
. A(Xi2 + X22)1'2Sg $2
A8 (411)
X82
v? II 0 (412)
v<0 II O (413)
4>2= — + Xt)
X8(Xi2+ X22)1'2
^ = —- - - (414)
Xg(t2 + 01 t + ft)1'2
where the /3, are constants. Since x8(t) is a bounded monotonic function,
OPTIMAL CONTROL 675
it follows that <t>2can vanish for only one value of t, say t0. Therefore,
by Rolle’s Theorem, $2 can vanish at no more than two points.
It follows that the optimal trajectory can have no more than two
powered phases.
We now consider several special cases.
Given
Xy(0) = a,
(415)
j = 1, 2,
x*(t/) = 0
x5(t/) = b6
maximize J = x^t,) .
Therefore, ci = 1
c2 — c8 — C4 = Cj = 0 .
Xi(t/) — —1 = X10
(417)
^(t/) = 0 = X40.
2. Maximum Altitude
Given
xy(0) = ay
(419)
j = 1, 2,
Xs(ty) = b8
t/ given (420)
x4(t,) = b4
maximize J = x5(ty) .
Therefore, c8 = 1
Ci = C2 = c8 = c4 = 0
and
Xl(ty) = 0
\*(ty) = 0 (421)
x#(t /) — —1 .
. 1
u2 (t) = — — = constant. (422)
X4o
3. Maximum Range
Given
xy(0) = a,-
(423)
OPTIMAL CONTROL 677
x*(t/) = bg
if given (424)
x*(.t/) = 0
maximize J = x4(t/)
'
C4 = 1
Cl = C2 = Cg = Cj = 0
Xi(ty) = 0 = Xio
X4(t/) — 1 = x40.
4. Minimum Fuel
Given
xy(0) = a>
(427)
j = 1, 2, -,5
X4(ty) = b4
tf given (428)
X8(ty) = 0
maximize J = —x8(ty)
678 ELEMENTS OF MODERN CONTROL THEORY
Ci = C2 = C4 = C6 = 0
Ai(t/) — 0 = Aio
Remark: It has been shown that the optimal control for this problem is
characterized by the following properties.
. 1 The thrust magnitude is either zero or maximum (bang-
bang).
2. The thrust inclination is, in the general case, a bilinear
function of time. For specialized cases, it may be a con¬
stant or a linear function of time.
3. The optimal trajectory contains no more than two power¬
ed phases.
In order to obtain the complete solution (exact form of the
optimal trajectory), it is necessary to solve a set of nonlinear
differential equations with two point boundary conditions.
This is not a trivial computational task. Some promising tech¬
niques for doing this efficiently have been developed in recent
years; a description of these is contained in Appendix A.
It should also be noted that in the above analysis there was
the implicit assumption that a solution exists. While this may
be true in most cases of practical interest, it must be recognized
that not all combinations of conditions give a meaningful or
well-defined problem. Generally, however, if there exists a con¬
trol that satisfies the given boundary conditions, one may pro¬
ceed to calculate an optimal control with the assurance that this
is physically and mathematically meaningful.
OPTIMAL CONTROL 679
(431)
680 ELEMENTS OF MODERN CONTROL THEORY
where
in = instantaneous mass
h(0) = xio
(432)
h(0) = x2o
hit/) = 0
(433)
h(t/) = 0
rii
in
h= - c^ (thmi)
- g. (436)
OPTIMAL CONTROL 681
c/n^ = ft(0)
- gt,.
m(0)
Solving for m(t,)
h(0) ~ gt,
m(t,) = m(0) exp (437)
c
J = m(0) 1 exp
h = xi
Xi = x2
x8 = m
u = m .
Xi = x2 = fi (438)
c u — g = ff2
x2= — — (439)
X3
Xs = U = f3 (440)
682 ELEMENTS OF MODERN CONTROL THEORY
min
x« = f dt
* 0
X4 = 1 (443)
X4(0) = 0 . (444)
H=Xi
x2—X2
^ +X8
u-fX4 (445)
Xi = 0 (446)
x2 = -Xr (447)
x8 = X2cu (448)
" ~^r
X4 = 0 . (449)
OPTIMAL CONTROL 683
(450)
— 0 when $ > 0
0X2 '
<t> = X, (451)
x8
(452)
x8
x2(t) gt + Xm (453)
0 < t ^ 11 .
*A somewhat lengthy analysis shows that the singularity condition <t>= 0 is incompatible with
the physical constraints of the problem. See Ref. 52.
ELEMENTS OF MODERN CONTROL THEORY
684
xi(tr) = x2(tr) = u .
(456)
+ X20 1 T T x10 .
x2o = (457)
Xl'0
= - 2M
a
in(l
\
- 4- tT)
x3n /
- ct
r - i
2
gtr^ (458)
If we eliminate tr between these two equations, we obtain a relation
of the form F(x10, x20) = 0, which determines the altitude and altitude
rate at which one switches to full thrust. A fairly simple form for F is
obtained by using the approximation
Oft t 1 / atjA
X30 2 \x3'0/
which, for
m(tr)
> 0.75 (459)
X30
where
Ca - gx3'0
X30
b=-(—)•
2 \X30 /
Since the ratio of maximum thrust to initial mass must be greater than
g, we see that a > 0. Also, since x^ is positive, it is apparent that the
only value of tT that is physically meaningful is
tt — (462)
F^x10,
x20^
=^xio
+x2'0
+2a/j/^-• (463)
Physical considerations dictate that xi0 > 0 and x20 < 0, since altitude
is positive and the vehicle velocity is in the negative xi direction. Further¬
more, the inequality (459), together with the relation
0 < t tr
leads to
X30
0 < tr < (464)
4a
686 ELEMENTS OF MODERN CONTROL THEORY
(465)
(466)
A plot of the switching function, Eq. (463), for the range of values
given by Eqs. (465) and (466) is shown in Fig. 5.13.
Assume now that the vehicle is at some initial altitude, xi0, and with an
initial altitude rate, x2o- The free fall trajectory is readily obtained from
Xl = X10
1_ (467)
2g
The fact that a complete solution is available for the problem of opti¬
mal control of a linear system with a quadratic performance index (Sect.
5.2.1) has motivated the effort to relate pragmatic design requirements
to this format. As a matter of fact, a far from trivial problem is that of
formulating a meaningful optimality criterion for a launch vehicle auto¬
pilot. One usually requires that the control system stabilize the unstable
airframe. In addition, it is desired that the deviations from desired atti¬
tude be small, and that the angle of attack be small in order to reduce
bending loads, engine angle deflections, etc. With the exception of
stability, none of these automatically results from the quadratic per¬
formance criterion.
Several recent studies have attempted to relate meaningful performance
requirements in booster autopilot systems to the quadratic performance
criterion. We shall consider in detail the papers by Tyler and Tuteur/13)
Fisher/14) Bailey/122) and Tyler. )136) Rather than discuss these indi¬
vidually, we will take a unified point of view. The results of particular
studies will be shown to be special cases of a general approach. Among
the problems investigated are the following.
x = Ax + Bu (468)
y = Gx (469)
ELEMENTS OF MODERN CONTROL THEORY
688
which are merely Eqs. (321) and (322) repeated here for convenience.
For the performance index, we take the following version of Eq. (324).
J = lim[ (xTGrQGx
+ UrRu) dt (470)
*/-»«Jo \ /
u*(t) = —Kx(t)
(471)
Iv = R-1 BT P
where R, B, and P are constant matrices, the latter obtained as the solu¬
tion of the steady-state Riccati equation
This is the simplest case; the optimal control, Eq. (471), is merely a
constant matrix times the current state of the system (feedback principle).
If we take for the performance criterion
J= (xtCTr
QGx+ urRu^dt (473)
then the form of the optimal control is still given by Eq. (471) except
that P, and, therefore, K, are now time varying, with P obtained as the
solution of
<3W
X =
dx
X = — 2Px
1 dW
X = Px (476)
2 dx
A = Px + Px (477)
where we have used Eqs. (468) and (471). Furthermore, Eqs. (468) and
(471) show that the optimal trajectory is described by
in terms of the costate vector, Eq. (476). The last two relations may be
written as
x A -BR"1 Br x
(480)
A -GrQG — AT X
690 ELEMENTS OF MODERN CONTROL THEORY
(Is - A) BR-1 Br
= 0 (481)
G^QG (Is + A)
consist of the eigenvalues of the optimal system and their mirror images
about the imaginary axis in the s plane, which belong to the adjoint
system.
We note further that the optimal trajectory is also described by
x* = (A — BK) x (482)
using Eqs. (468) and (471). The characteristic Eq. (482), together with
its adjoint, is, therefore, given by
Thus, the roots of Eq. (481) and (483) must be identical. This fact
will be used later to establish a relationship between the eigenvalues of
the optimal system and its feedback gains.
We turn now to a discussion of Merriam’s method. The performance
criterion is taken as
=!0 (y°~y)®(yo
~y)+(ud~u)r(u°- dt . (484)
W(x,
t)=Min
| - y^Q^yD
-
T
+
R^uc
— dt (485)
OPTIMAL CONTROL 691
. J1 rp
dW
at
= Min
u
\Yd—yj Q^yc
—y^+ R^uD
—
(486)
+
(f ) (Ax
+Bu)
The optimal control is found to be
u* = uD
1
— — R-1
2
Br
(") (487)
we find
<9\V n n n
— = P - 2 22 Pa Xa + 22 22 P/37 X7 (489)
at a=l /3=1 7=1
dW o , o V
— = ~ 2 pa + 2 2^ P/3a X/3• (490)
dxa 0=1
Substituting Eqs. (487M490) in Eq. (486), and assuming that p^7 = p7/3
leads to
n n
The quantities F, Fa, and F^7 depend only on the components of the
matrices A, B, Q, R, and G. In order for Eq. (491) to be satisfied, each of
692 ELEMENTS OF MODERN CONTROL THEORY
- P = F (492)
- Pa = Fa (493)
~ P/3-y
= \ (%
F|37
+ F7jSJ (494)
Yj 1)2, 1 • * *, n
W(x, t/) = 0
Problem 1:
g = gravity acceleration
I„ = moment of inertia of vehicle about pitch axis
lc = distance from mass center of vehicle to engine swivel point
(a = distance from mass center of vehicle to center of pressure
La = aerodynamic load per unit angle of attack
m = mass of vehicle '
Mi = generalized mass of first bending mode
qi = generalized displacement of first bending mode
T0 = control thrust
U0 = forward velocity of vehicle
a = angle of attack
5 = thrust angle deflection
6 = attitude angle
dD = steady-state attitude angle
Me = T/Tc I„
or
Rate Gyro:
Position Gyro:
Accelerometer:
(501)
Here, KS) K0, Ka, and Kp are the gains associated with the respective
sensors; aT is the thrust acceleration, and the a(1) quantities are the
normalized bending mode slopes whose subscripts indicate the location
of the sensor along the vehicle.
If we define
xi = 0 x4 = qi
x2 = d x5 = qi (503)
Xg = a u = 5
x = Ax + Bu
with
0 1 0 0 0
0 0 Ma 0 0
A = g COS0o/Uo 1 —La/mU0 0 0
0 0 0 0 1
0 0 0 —coi2 — |iWi
0
Me
B = Tc/mU0
0
—T0/Mi_
This is equivalent to the system, Eqs. (468) and (469), with G = 1, the
unit matrix.
Taking a performance index of the form in Eq. (470), the optimal
control is given by Eq. (471), viz.
71
(505)
1, 2 m
We note, for example, that two rate gyros may be viewed as two
different sensors if their KB and <j0(1)are different. Thus, for the system,
Eqs. (496H498), the use of one position gyro and two rate gyros and
two accelerometers (having different gains) will provide five independent
measurements that are linear combinations of all the state variables.
If now we use
to denote the control function employing the sensors (the c* are as yet
undetermined constants), then Eqs. (504) and (506) will be equivalent if
696 ELEMENTS OF MODERN CONTROL THEORY
n m
22 k» x, = 22 c» z» 22 c» 22 x, . (507)
»=1 i— 1 1 ?•=!
This leads to
brc = -K «% (508)
b = m X n matrix
— = 2(bbr c + bKr) = 0
dc
or
Problem 2:
x = Ax + Bu (511)
y = Gx . (512)
OPTIMAL CONTROL 697
i = Ltj . (513)
J
/ 7[(y- Ly
)*%
Qtf- Ly)
+urRuJ
dt. (514)
and
R = Br Gr QGB + R (517)
d = Q - QGBR-1 Br Gr Q (519)
6 = GA - LG . (520)
where
" 0 1 ~0
A = B =
_1 1
<1 1
1
B21 _
1 0
G =
0 1
Qn 0
Q = R = R11, a scalar
0 Q22
0 1
L =
— L21 — L2:
We then find
An = 0 A12 = 1
A22 — — A21 —
B212Q22 (L22 — A22)
B212Q22 + Ru
OPTIMAL CONTROL 699
\w Q222
22
B212Q22 + R11
Qn — Qn
B21 Q 22
K, [(L21 — A21)(L22 — A22)]
B212Q22 + R11
K„
while
Q22—»0
and, in turn
Kr re 0 .
u(t) re — Kmx
Xl
' si [(L,‘
_A“)(L>'
_A”)] x2
(524)
=—
—[(l21
—A2i^Xi
+(l22
—A22^x2J
Substituting this in Eq. (511)
x = Ax + Bu
= (A - BKm)x
But
0 1 0
(A - BKm) = [(L21 — A2i)(L22 — A 22)]
—a22
B2i B21
- A2i
0 1
%L2i l2
PLANT
(a)
(b)
PLANT
+ G(s) M(s)
M(s) for K -»
R(s)
Problem 3:
r‘f
J = J [(yo - y)r Q(y*>- y) + (u*, - u)r R(uj> - u)] dt (525)
(526)
OPTIMAL CONTROL 703
(527)
Here
xi = Y
x2 = Y
x8 = ^
x4 = i
u = 5
The pertinent system matrices become
1
(Tc + Lp)
m Uo m
1 0 0 0
A. =
hJi o o
h Uo
0 0 1
T,
m
0
B =
Tc 4
I,
0
704 ELEMENTS OF MODERN CONTROL THEORY
u = u0 ———F-(pi+ 23P0iX0
'j + -r( p34-23P^3
X(3>)1
(528)
RuLm\ 0=1 / lz\ 0=1 / J
where R = Ru is a scalar, since u is a scalar. The p’s are obtained from
Eqs. (492)-(494), which in the present case take the form*
• B 2
~~%Pl = Qll Y o 4~ All Pl + P2 + A31 P3 — —— Pl P11
Kn
(529)
B31Bn ( , \ B312
~ —- - 1 Pll P3 4“ Pl Pl3 ) — —- Pl3 P3
R11 \ / Rn
_ a — n v _ ®n2 „ B31Bn ( ^
P2 — Li22 J-D — p —‘ Pl Pl2 — - p - ( Pl2 P3 4“ P23 Pl J
(530)
B3i2
P23 P3
Rn
—P3—Q33
\pD
4*P4— Bn2
PlPl3—^31 ( P3 P13 4* Pl P; 3)
R11 Rn \
(531)
Bn2
P3 P33
Rn
B 2
P>4= Q44 'pD 4" Pl A14 4- P3 A34 - 2- pj p14
Rn
(532)
B31 Bn / . \ B312
r; - l P3 Pl4 4- Pl P34 1 — —- P3 P34
R11 \ / Rn
(533)
2 B31 Bn B3i2 _ ,
- - - P11 P13 ~ —— P132
rtn R11
B 2
Pl3 = An P13 + P23 + A31 P33 + P14 — - Pn P13
' R11
(535)
B31 Bn
Ru
+ Pn P33 ) B^
Rn
P33
B 2
— P14 = All Pl4 + P24 + A31 P34 + Al4 Pll + A34 P13 — —— Pn Pl4
Rn
(536)
B31 Bn / \ B312
Pl4 Pl3 + Pll P34 P13 P34
Ru V / Ru
Bn2 2 B31 Bn
— P22 — Q22 — —— Pl22 — P12 P23 — •" - P23 (537)
Kn R11 R11
. _ Bn2 B31 Bn / , \
P23 — P24 — —— P12 P13 — ——- I P13 P23 + P12 P33 1
R11 R11 V /
(538)
>31
P23 P33
Ri
B31I Bn/
P24 = A14 P12 + A34 P23 — —— P12 Pl4
Bn2
Rn R II V
P23 Pl4 + Pl2 P34 )
(539)
P23 P34
Ri
Bn2
— P44 = Q44 + 2 Aj4 Pl4 + 2 A34 P34 Pir
Ri
(542)
2 B31 Bn
Pl4 P34
B312
d 2
R11 TrTP34
The weighting matrix Q has been assumed diagonal.
Q11 0
Qa
Q = Q33
Q44_
and the subscripted A and B terms are the components of the respective
matrices.
The set of Eqs. (530M542), when solved with the initial conditions
Pa(t/) = Pa0(t/) = 0
(543)
«, 0 = 1, 2, 3, 4
yield the p’s which, when substituted in Eq. (528), give a control function
of the form
Problem 4:
x = Ax + Bu (545)
y = Gx (546)
OPTIMAL CONTROL 707
x = (A - BK)x . (549)
A - BR-1 Br
(550)
Gr QG - Ar
Is - A BR-1 Br
= 0 . (552)
GT QG Is + AT
Since Eqs. (551) and (552) are the characteristic equations for the
same system, they must have the same roots. Therefore, by equating
coefficients of like powers of s, we obtain a set of equations that relates
a = m X m matrix (nonsingular)
P = m X n matrix
7 = n X m matrix
8 = n X n matrix
M = scalar
I< = i X i unit matrix.
The first of these is proven in Bodewig’s book!1 72) (p. 217), and the
second is due to Plotkin.073) Applying Eq. (553) to (552) yields
Furthermore
is the matrix transfer function* for the system, Eqs. (545) and (546).
Also
G = 1 X n matrix
B = n X 1 matrix
Q = Qn = scalar
R == Rn = scalar.
where
Y(s)_ TM ~ N(s) _
kfr
M (3~8J)
(563)
U(s) D(s) fr ( 8“ P»)
r < n
and z, and p< are the zeros and poles, respectively, of the open-loop
system, with k a known factor.
Defining
a = - s2
Mr = - z/
Vi = - Pi2
0 = 1 +
k2 Qn U (a~w) = T(s) T ( —s) . (564)
Rn
n y®- Vi'j
Thus, standard root locus techniques can be used with k2 Qn/Rn as
the variable gain. A given value of this variable gain yields pairs of roots
12,, each pair representing a point in the s plane, together with its mirror
OPTIMAL CONTROL 711
image about the imaginary axis; i.e., a closed-loop pole for the optimal
and adjoint system. If these closed-loop roots are acceptable from a
dynamic response point of view, then the corresponding Qu and Ru are
used to determine the optimal gain matrix, K.
Note that the optimal system contains only those roots in the left-hand
s plane. It can be shown that the optimal system obtained by this pro¬
cedure is always stable. Therefore, the mirror image poles belong to the
adjoint of the optimal system. '
In simple situations, the components of K may be directly related to
Q11/R11. For the second order case described by
+
_ X2 _ _ ~b —a _ _ x2 _ _ c _
Xi
y = [1 0]
X*
= (s2 + as + b)(s2 — as + b) + ^ c2 .
Ru
Equating coefficients of like powers of s and solving for Ki and Iv2 yields
Ki = b i/i + Or <r
c c V + Rub2
K2 = - a , ./a* + 2-Ki
c y c2 c
K = [Ki K*].
h = — V -sin 7 (565)
v,
V = g sin 7 -
D (566)
m
g cos 7 V cos t L
7 = - V - (567)
te -T h mV
where
D = drag force
g = gravity acceleration
h = altitude above earth’s surface
L = lift force
m = mass of vehicle
rE = radius of earth
V = velocity of vehicle
7 = flight path angle.
(568)
(L2 + D2)1'2
(569)
m
(570)
r‘ (v + d2) (572)
X6
Jt m2
with
xi(t<) = 0 (573)
xs(t<) = 0 (574)
x* = K4 p1'2 Vs (575)
.6 L2 4- D2 (576)
m2
X! = h rs = Ki
X to II g = Ka
II
-> m = K3
P = Kr e*6*1 (578)
(579)
L = — p X22 K10 Cl
D = — p Xa" K10 CB
(580)
OPTIMAL CONTROL 715
where Ki0 is a reference area and Ct and CD are the lift and drag coeffi¬
cients. The latter are approximated by
where Kn, Ki2, and Kis are appropriate /constants for the lift drag polar
and u is the angle of attack, which is taken as the control variable.
Combining all the above relations, the state equations for the system
become
Xi = — x2 sin x3 = h (583)
cos x3 x2 cos x8
x3 K>
x2 Kj xi
_ K62Kiq Kn e KeXj
6 1 x2 sin u cos u = fa (585)
K3
+2Ki2
Knsin2
u+Ki32
sin4
uJ=f6 (587)
xi(t.) = h0 xx(t/) = h /
x2(t.) = Vo x2(t/) = V/
x3(ti) = To Xg(t/) = y, (588)
x4(t.) = 0
x6(ti) = 0
716 ELEMENTS OF MODERN CONTROL THEORY
This problem was analyzed by Payne,!1 67) using the maximum princi¬
ple. We form the hamiltonian
H = Z Xi ft (589)
l
K6’Iv, Kiq^
Xl — X;
[( k3 ^eK&Xl
x22
^Ki2
+K13
sin2
u)]
f x2cos x3 / K62K6 Kio Kn \^ e216*1
Ki x2
K3
sin u cos u
]
-[( k4 k6 k6
)ei I Ktxl x^3
]
-,[( 2 K64 K6 K 7 -^^-10
k3
-
j 2^ga:i4 ( 17
1 e e Ax24 I Kn
2*9
sin2 u
9
cos2 u
X2 = Xi sin x3
+X22 Kl°
) e*6*1
x2
L\
(Kl2
+K13
sin2
u)JKa /
X3 — X] x2 cos x3 — X2 K2 cos x3 X;
[ K22 si
sin
x2
x3
(592)
x2 sin x3 "1
" Ki+ XlJ
See Sect. 5.1.2.
OPTIMAL CONTROL 717
X4 = 0 (593)
X6 = 0. (594)
dH (x, X, u)'
(595)
du
Max H (x, X, u)
(596)
u
X4(t,) = - 1 (597)
feet/sec)
feet)
of
(thousands
ALTITUDE
VELOCITY
(deg)
ANGLE
FLIGHT
-ft/sec)
HEATING
(lb
RATE
(g’s)
ACC
(deg
CON
Figure 5.22. Optimal Trajectories, Control Function.
OPTIMAL CONTROL 721
Figure 5.23 shows how the total heat and acceleration effects along
the optimal trajectory vary with parameter K7. This plot can be used to
determine the minimum amount of additional heating that the vehicle
must absorb in order to achieve a specified reduction in acceleration
effects. For example, to reduce the acceleration as shown in Fig. 5.21,
it is necessary that the vehicle absorb about 10 percent additional heat
during the re-entry maneuver.
The basic data for the problem is given below.
Kj = 2.09 X 107 ft
K2 = 32.2 ft/sec2
K3 = 250 lb sec2/ft
Iv4 = 1.0 X 10-4 (lb)1/2 sec
Kj = 0.052 (lb)1/2 sec/ft2
K* = -4.26 X 10-6 ft-1
K10 = 66.5 ft2
Ku — 1.2
Ki2 = 0.274
K13 = 1.8
Initial conditions
Xl(ti) 400,000 ft
Final conditions
Xl(t/) 250,000 ft
Xa(t/) 0 deg.
APPENDIX A
where y (t) and h(t) are n vectors and A(t) is an n X n matrix. It is required
to determine y(t) in the interval (0, t/) such that the boundary conditions
y>(0) = a j (A2)
j = 1, 2, — , r
723
724 ELEMENTS OF MODERN CONTROL THEORY
y>(t/) = by (A3)
j = r + 1, n
are satisfied.
A simple and direct method for obtaining y(t) is the following. (14?)
Integrate the homogeneous version of Eq. (Al)
u = A(t) u (A4)
u,(m)(0) = 0 j ^ r + m
(A5)
= 1 j = r -f m .
Vy(0) = ay , j = 1, 2, •••, r
(A8)
= o, j = (r + 1), • • •, n .
j = (r + 1), • • •, n
The distinctive feature of this method is that all integrations are per¬
formed with initial conditions; it is not necessary to “guess” at final
values. The solution is exact (within roundoff error) after a finite
number of operations.
x = f(x, t) (A1 2)
where x and f are n vectors, and the boundary conditions are given by
x/(0) = a, (A 13)
j = 1, 2, •••, r
Xy(t/) = by (A 14)
j = (r + 1), •••, n .
Let x(0)(t) be an initial guess for the solution of Eq. (A12). Successive
improvements are obtained from
where f (*>==f(x<*>, t), and F(4> is the Jacobian matrix whose ijth com¬
ponent is given by
t) (A 16)
dXy
where
g(xt)
Xt+l = Xi (A 19)
g'(Xi)
x0 = initial guess.
*A function <p(x) is strictly convex in an interval (a,b) if, for xi, x2, X, with a < xi < x2< b
and 0 < X < li, we have
If <?" (x) exists, then p(x) is strictly convex if <?" (x) > 0.
OPTIMAL CONTROL 727
t = as
dx
= x
ds
then we have
Xj = af ,(x, as) .
a' = 0 .
Thus, the new state vector of the system is of dimension (n + 1), and
we proceed as before. Note that one makes an initial guess on a (in x(0)(t)).
This is, therefore, constant during each iteration, but it varies from one
iteration to the next. Thus the determination of tf becomes an integral
part of the quasilinearization procedure.
APPENDIX B
F = F(y,y, t) (Bl)
Y = y + ei](t)
(B2)
e = a constant.
The change er?(t) in y(t) is called the variation of y(t) and is denoted by 5y
5y = e)j(t) . (B3)
5y = Y - y (B4)
5y = eh • (B5)
it(Sy>
= % (B6)
(B7)
728
OPTIMAL CONTROL 729
Taking a Taylor expansion in which only linear terms are retained, this
reduces to
dF dF
AF = eri + eil •
dF dF (B8)
SF- ^ iy + » * %
For a complete analogy with the definition of a differential, one
might perhaps anticipate the definition
dF dF dF
5F * Ty Sy+ w s* + Tt st
However, t is not varied, so that 5t = 0. Hence the analogy is indeed
complete.
It is easy to verify from the above definitions that the variation opera¬
tor obeys the same laws as the differentiation operator. Thus
/ Fj \ F2 8 Fi — Fi 5 F2
5\ f2) “ F22 (BIO)
etc.
a-b = 0 (C3)
730
OPTIMAL CONTROL
731
and that for any two vectors, a and b, the following relations hold.
d&i
then*
i (b'Aa)
=A*b (Cl 1)
d
77" (brAa) = Aa . (Cl 2)
OD
In particular
d
— (arAa) = 2Aa . (C13)
dy
da,i
dy
da,
dy_
dan
d / du \T / dv \T
s(u'v)=(dr)v +(dr)u- <C14)
/ du \ _
dUi
Here I — 1 is an n X m matrix whose (ij)111component is —
dF / dF \ / d^ \
dz \y d (p /f y\ dz )
d«p
^ imi
= ini= Ti dt
4 (ara)_l/2 2 a<a<
z i=-i
which reduces to
11 n aTd (C15)
W " Hall•
Consider now two functions, f(t<) and gft,), which are defined for
discrete values of t<, in an interval, ti, t2, - , t„. Suppose we define
OPTIMAL CONTROL 733
where
At,' — t<4-1 ft .
The quantity, (f, g), is defined as the inner product of the functions,
f(t) and g(t), in function space. Loosely speaking, we say that Eq.
(C 16) gives the inner product of two vectors in an infinite dimensional
Euclidean space.
The norm of f(t) is defined by
llfll= VP • (Cl 7)
(fr,f.)= / "h(t)
f.(t)dt= 0, r ^ s (C23)
734 ELEMENTS OF MODERN CONTROL THEORY
REFERENCES
Xi = -Xj + u .
(b — ae 7)
u = - — — — e‘ .
sinh tf
Xj = x2
x2 - — ax2 — bxi + cu
where
P12 - ~ C2 (1 ~ t)
a
P22 —
c-
|u| < M
u = M sgn \i
ELEMENTS OF MODERN CONTROL THEORY
748
where
Ai — Xi + 2xi , Ai(t/) — 0 .
Xi = x2
x2 = u
|u| < 1 •
Show that the control, u, which drives the system from an arbitrary
initial state to xi(t/) = x2(t/) = 0, in minimum time can be expressed
in terms of the current state of the system; in other words, the
optimal control is closed loop. (Hint: Derive the equation of
motion in the xi x2 plane and find the expression for the switching
curve.)
is a minimum. The final time, tf, is given. This is one type of fuel
optimal problem. Using the maximum principle, determine the
form of the optimal control in terms of the costate variables. Is it
possible for the optimal control to have the value zero during the
control interval?
differential equations
Xi = a cos u
x2 = b sin u — g
X3 = Xi
x4 = x2 .
Here xx, x2, x3, and x4 denote horizontal velocity, vertical velocity,
horizontal displacement, and vertical displacement, respectively.
These equations result from the assumption of flat earth, constant
thrust, constant mass, and negligible aerodynamic effects. Thus a,
b, and g are positive constants.
Given are the initial conditions
xi(0) = £x x3(0) = £3
x2(0) = £2 x4(0) = |4
and the terminal condition, x4 (t/) = h, where t/, the terminal time,
is prescribed. It is required to determine the form of the thrust
angle, u(t), such that the final horizontal velocity is a maximum.
Using the maximum principle, find the form of the optimal u in
terms of the costate variables.
12. Show how the optimal control, u, for Problem 1 1, could be deter¬
mined via dynamic programming methods. Derive the necessary
recurrence relation together with whatever initial conditions are
required to obtain a solution.
13. List and explain the advantages and disadvantages of dynamic pro¬
gramming vis a vis the maximum principle
a. theoretical
b. computational
c. in practical design problems.
Many schemes have been devised to cope with these problems. They
have been called “adaptive” since the control system parameters vary
(adapt) as a function of flight environment. A discussion of these con¬
cepts forms the subject matter for the present chapter.
A discussion of adaptive control cannot proceed without at least a
tentative definition. One widely accepted is that an adaptive control
system can monitor its own performance, evaluate the effectiveness of
its response, and then modify its parameters automatically to achieve
desired performance. When applied to a launch vehicle autopilot this
implies that the adaptive control system will identify an unstable bending
mode frequency, and activate some suitable compensation device to
ADAPTIVE CONTROL 753
stabilize the system. This compensation may take the form of cancel¬
lation, phase stabilization, or gain stabilization. The cancellation tech¬
nique, as typified by the model-reference or virtual slope methods, for
example, makes the sensor output behave as if the sensor were located on
an antinode (point of zero bending mode slope). Phase and gain stabili¬
zation techniques depend on an accurate identification of the bending
mode frequency since these modes are highly tuned (very low relative
damping factor). In some cases signal Shaping by conventional filtering
is inadequate, since the signals to be separated do not have a sufficiently
large frequency separation. The digital adaptive filter and frequency
independent signal processing techniques have been developed to cope
with this problem.
One thing is certain. Adaptive control has not suffered from a lack of
attention in the literature. Neither has Air Force nor NASA funding
been lacking for pursuing investigations of particular concepts. Yet
today, some ten years after adaptive control began to be studied in¬
tensively, very few adaptive systems have reached the flight test stage
(these mainly on aircraft), and none are operational. A partial answer
is that so-called conventional methods have been refined considerably,
and the point at which adaptive control becomes mandatory has never
been clearly defined. There is a considerable (and understandable) reser¬
vation in committing oneself to complicated and exotic control schemes
where multimillion dollar vehicles are concerned if their need has not
been clearly established.
At present, conventional and straightforward control methods appear
adequate. However, if past experience is any indication, one will as¬
suredly be confronted with problems that strain the capabilities of
current technologies. When this occurs, advanced schemes presently in
the “drawing board” stage will be evaluated under actual operating
conditions and thus serve both to refine and enlarge these concepts
presently called “adaptive control.”
The following sections describe the major bending mode adaptive
techniques. They also provide a critical evaluation of the virtues and
limitations of each type. To date, none of these schemes has been flight
tested, although many of them have been thoroughly simulated (via
computer) with varying degrees of complexity. Consequently, while
different government agencies and aerospace contractors have their pref¬
erences, there is no universal agreement that any one type represents a
definitive solution to the problem.
In order to simplify the presentation, and provide a common frame¬
work for analysis and comparison of the different methods, each will be
754 ELEMENTS OF MODERN CONTROL THEORY
(1)
At the desired operating point the slope of the ISE as a function of the
ADAPTIVE CONTROL 755
0P„aP,r
.,*P.(/(E),dt)-° <2>
If the limits of integration are independent of Pn and the integral of the
partial derivative of the function exists, the partial differentiation may
be carried out under the integral sign, resulting in the error quantity
(EQ). dt = E dt
(3)
= / W£(t)
(E)dt.
The error quantity is the integral of the error weighted by a function
such that the error quantity is indicative of the state of the system. The
756 ELEMENTS OF MODERN CONTROL THEORY
sign and magnitude of the error quantity indicate the direction and
amount that the variable parameter should be changed. A simple mech¬
anization results if the change of the parameter is made proportional to
the error quantity. The weighted error is then proportional to a time
rate of change of the parameter, which will result in setting the parameter
to the desired operating point. The integration of Eq. (3) may then be
carried out in the parameter adjustment device.
The weighting function is determined by performing a straightforward
partial differentiation of the differential equation for the error as a
function of the input quantities, or by considering the change in the
parameter as a disturbance entering the system after the parameter. The
error weighting function can be generated by taking the signal at the
input to the variable parameter and feeding it through a filter having the
same performance function as the system, cascaded with a filter that is
the reciprocal of some of the forward path components. It is not possible
to obtain a signal that exactly satisfies the equations, since the ill-defined
or unknown characteristics of the system are what lead to the require¬
ment for the adaptive system. An approximation to Eq. (3) is obtained
by substitution of the performance function of the model for the per¬
formance function of the system. In cases where a system model does
not exist explicitly, Eq. (3) may be approximated by using an approxi¬
mation for the dynamic characteristics of the optimum system.
The validity of this approximation depends upon the accuracy of the
approximation of the system by the model and weighting function
filter. In the neighborhood of optimum response, this approximation is
good. For system parameter settings far removed from optimum the
approximation is poor; however, if the algebraic sign of the weighting
function is correct, proper system operation is obtained. Even though
the weighted error has the incorrect sign instantaneously, satisfactory
results are obtained if the error quantity has the proper sign over the
evaluation period. The derivation of these equations is based on the
assumption of the constant coefficient linear system. When the param¬
eters are varied during the response to input signals, these equations are
in error by perturbation terms.
Because of the nonlinear nature of the system, operation can best be
studied by analog simulation. Simulation is also required to evaluate the
use of approximate dynamics in the weighting function filter. It is to be
expected that the operating state selected by the adaptive system will
not be the optimum operating state predicted by the exact Eq. (2), but
will have some error due to the approximations. It has been found that
these errors are usually small and the characteristics of the weighting
function filter usually are not critical.
ADAPTIVE CONTROL 757
where q(1) is the generalized coordinate of the first bending mode.* The
control system to be described is then adaptive only with respect to this
mode.
We must now choose the parameter to be varied in order to minimize
Eq. (4). For this purpose, KezeE4) takes the “effective bending” sensed
by the rate gyros. Such a signal may be obtained, for example, by
taking the difference between the outputs from two gyros located at
different positions along the vehicle; thus
{6a + crG8(1)
q<1J)- {6a + <rG,(1)
q»>) = <r*(1)q<J>.
^/[q<1>?dt
- hh* [q0,1‘
dt=2/q<1,( =0%<5)
The system is now made adaptive by setting the rate of change of
<jfi(1)proportional to
(6)
(s + Kc) 8 = Kc 8C (9)
— Or + <r<r(1)
Q(1)+ Ks (Or + ^oh1I)q(1))
where
<Tna)—c0r1) — cr0}l)
and
<reRW= crG(1>
G„(s)
q(1)
=- ^ (s2
- dc (12)
where
ka kc tc
[Kfl S (o-G(1>
+ 0>(1))+ cto^] (s2 — Ha) . (13)
(i)
Taking the partial derivative of Eq. (12) with respect to <rj , we find
„ , s aq(1> Ka Kc Kb Tc ,
Go(s) T- (jy - - (s2 - Ma) s q(1) = 0
dai
or
K4 Kc Tc (s2 — Mq)
G, (s) = (15)
M(1) Go (s)
and if the input is Kscrtf(1)q(n, we see that the output of the weighting
filter is the quantity
760 ELEMENTS OF MODERN CONTROL THEORY
i.e., proportional to the error quantity, Eq. (6). The signal is used to
adjust the sign and magnitude of the gain X. A steady-state value is
achieved when the multiplier output is zero, which indicates that the
first bending mode signal has been essentially eliminated from the
“blended” rate gyro feedback.
Some typical results of a computer simulation for this system are
shown in Fig. 6.3. The input 9C was a square pulse in each of the three
cases shown. In the first case, with the adaptive loop open, it is apparent
that the bending mode signal is poorly damped. The improvement in
response for the next two cases (where the adaptive system is employed)
is quite evident.
A PITCH
ANGLE
signals, and a “clean” rigid body rate signal is obtained from a model of
the rigid body dynamics. A similar procedure is applied to the attitude
displacement information.
In accordance with this control philosophy, it would appear that one
could generate “model” attitude and rate information from a configura¬
tion of the form shown in Fig. 6.5. However, it is not possible for Ha to
operate only on computed information, since a closed-loop pole-zero pair
would appear in the right-half s plane, and these would cancel only if
Ha = Ha exactly. Therefore, the configuration shown in Fig. 6.4 is used
where h% is acted upon by a combination of computed and actual body
attitudes.
A crucial factor in the model-reference approach is the required ac¬
curacy in H* and ju* , of which the latter is the more critical. Computer
traces for attitude rate in response to step input commands in 9C are
shown in Figs. 6.6 and 6.7 for ju* 15 percent less than and 15 percent
greater than Ha, respectively. These traces indicate (and analytical studies
verify) that the tolerance on h1 (with respect to Ha) should be heavily
weighted toward the positive (h% > Ha)-
In Fig. 6.8 are shown the rate response traces due to a step wind
disturbance for both model feedback and conventional systems, each
ADAPTIVE CONTROL 763
t (sec)
with the same system parameters. The very poor response for the model
feedback system was to be expected, since actual body motion is not
generated solely by engine deflection. In the configuration of Fig. 6.4,
the engine responds only to the heavily filtered real body loop.
764 ELEMENTS OF MODERN CONTROL THEORY
0 1 2 3 4 5
t (sec)
Since for a realistic launch vehicle the wind inputs are often the
predominant inputs to the system, the model-reference configuration of
Fig. 6.4 is clearly inadequate. Various modifications are proposed by
Tutt and Waymeyer/5) of which all are based essentially on some form
of angle-of-attack sensing. The problem is to obtain some sort of dis¬
turbance information without reintroducing the bending previously re¬
jected.
To do this we may adopt the same control philosophy used in the
attitude and rate loops; namely, to use an actual value and a “model-
reference” value for angle of attack, which is combined and used as an
additional feedback loop for the control system. The system would
then take the form shown in Fig. 6.9. The schematic for the a model is
derived from Eqs: (Al) and (A2) of Appendix A, with
T„ fa W
N« = N„ oc M —
m U0 m Uo Uo‘
MODEL
t (sec)
Figure 6.10. Comparison of Response to Wind Inputs.
ADAPTIVE CONTROL 767
The essential idea for the gyro blending scheme is motivated by the
following considerations. Suppose two rate gyros located at different
points along the vehicle are used. The output of the rear gyro is*
(19)
Kfl )% (20)
(1 — K) <T(JB(1)
T" K<JOFl) — 0 . (22)
are passed through bandpass filters designed to reject all signals outside
the anticipated frequency range of the first bending mode. The signal
appearing at the input of the integrator is then
rigid body mode and the E< with the bending modes. It will again be
assumed that only the first bending mode is significant.
The adaptive system described here generates a phasor whose ampli¬
tude and phase angle are identical to Ex, and subtracts this from the
feedback control signal, thereby giving a signal free of first bending mode
information. The method is illustrated schematically in Fig. 6.13. In
Fig. 6.14 are shown the frequency response characteristics of the adjust¬
able bandpass filter. To analyze the operation of the system, assume that
is equal to w(1),the first bending mode frequency. Then the output of
the adjustable bandpass filter is a signal whose amplitude and phase are
equal to that of the first bending mode signal, with all other frequencies
sharply attenuated. Therefore, subtracting this from e'F gives a signal
essentially free of first bending mode information.
The function of the adaptive controller is to track the bending mode
frequency and set u, equal to u>(1)at all times. This may be accomplished
in the manner shown in Fig. 6.15. In general, wF will not be equal to co(1),
which means that the bending phasor, after passing through the adjust¬
able bandpass filter, will incur a phase shift, A^x (see Fig. 6.14). This
signal, after then passing through the secondary bandpass filter, will
incur another phase shift approximately equal to A^x if u, is not too far
from w'11(this means essentially that ax'1' is such that operation is along
the nearly linear portion of the phase shift curve in Fig. 6.14). Thus, A^x
is a measure of the deviation of wF from a>(1).
FROM OUTPUT
TO ADJUSTABLE
OF ADJUSTABLE
BANDPASS FILTER
BANDPASS FILTER
LIMITER
NO. 1
SECONDARY
BANDPASS FILTER
2£ s
LIMITER
MULTIPLIER
NO. 2
^F S +^F
T
Figure 6.15. Detail of Adaptive Controller.
772 ELEMENTS OF MODERN CONTROL THEORY
A sin w(1) t
and
If A^i is small such that sin A^i & Afa, this last relation may be written as
- +5 deg - — —
v/ \ju Imff
H - 1
sec
— -- - - - V
o _
- +5 -
%
- ^10 rad/sec -
°F - 0-
• -10 •
%-20 % XU
k = 250
- +0.05 ft -
(1).
-0.05 -
k = 400
- 20 RAD/SEC %
1
sec
k = 150
- +5°-
- 20 RAD/SEC
V(s)
= G(s)
U(s)
k
12 a» 8s'
i=0
(24)
12 bi s*
<-0
776 ELEMENTS OF MODERN CONTROL THEORY
IT (s - zi)
i=0
= K -
n Cs- p<)
The output v(t) and its first (n — 1) derivatives have the initial values
d4 v(t)
tv - v,u (o)
i = 0, 1, 2, (n - 1) .
H bi s4 12 b<s4
t=0
u(t) = 2Z Ri t4 (26)
«=0
or
Substituting this into Eq. (25) and taking a partial fraction expansion,
we find
?+>*+l
D„, t”’-1
+ E (29)
W>=1
(w - 1) ! '
Here the second summation extends over the complex poles in the
upper half s plane. For present purposes it is presumed that the form of
v(t)as shown by Eq. (29) is known. However, the parameters
K* and p, x = 1, 2, • • •, y
K, = |K,| /Kx
> x = (y + 1), • • (n - g)
and p, = —ax + i/3, ;
D„ w = 1, 2, • • •, (g + h + 1)
Vrr (t N- M+l)
Vm (t N-M+'i)
Vm (W)
*One may without loss of generality assume that bn = 1 in Eq. (24), in which case the number of
unknown parameters becomes 2n — g -f- h.
778 ELEMENTS OF MODERN CONTROL THEORY
~ (F • F) = 0 (32)
X = 1, 2, • • •, (n - g)
^;<F-F)=0 (33)
x = 1, 2, • • - , (n - g)
m. (F %F> = 0 (34)
w = 1, 2, • • •, (g + h + 1)
P* = Pxat + Apx
where the pIJV, KIJV,DWJVare initial estimates and the A( ) quantities are
assumed small. The value of v(t) at some generic timet* is then given by
ADAPTIVE CONTROL 779
(36)
where the asterisk signifies that the nominal (subscript N) parameters are
used. Using Eqs. (35) and (36), the set of Eqs. (32)-(34) is linear and
may be solved for Ap*, AKX, and AD„, thereby determining updated
values for px, Kx, and D„.
Once the present values of px, Kx, and Dro are known, Eq. (29) com¬
pletely determines the signal v(t). Furthermore, Eq. (29) determines this
signal as a sum of its individual components. Any component or any
group of components can then be separated from the signal v(t) by
computing in real time a partial sum which includes only the desired
components. This separation of the signal components can best be done
on a digital computer. The output of the digital computation can then
be converted, in line, to an “analog” form and used to replace the actual
v(t) for the purposes of control. By this process it is possible to restrict
the composition of the control signal to only the desired modes of
response; for example, the rigid body modes of a missile. All other
modes, such as the elastic modes, can be filtered out simply by leaving
them out of the summation of Eq. (29) as carried out on the computer.
Clearly this filtering operation should be effective for filtering any
modes for which the roots of the characteristic equation, px, are separated
by a large distance in the s plane. It is not required that the frequencies
be different; only the total distance counts. In fact, two modes of the
same frequency are effectively filtered if their damping is quite different.
The digital instrumentation of the operation described in the pre¬
ceding section is sketched in Fig. 6.19. An analog-to-digital converter
samples and digitizes the measured continuous signal v(t) every T seconds.
The present sample is v(NT) where N is the total number of samples
from the beginning of the operation. A fixed number M of the most
recent samples is stored in the digital computer. When a new sample
arrives, it is put at the first address, the other samples are shifted to the
left by one place, and the oldest sample in the memory is dropped. An
equal number of samples produced by the computer, which represents the
present fitted curve, is stored in the block below. These latter samples are
computed from the present best estimates of the parameters px, K*, and
ELEMENTS OF MODERN CONTROL THEORY
780
D„ (denoted by pw, KXNand Dw) and jointly represent the sampled form
over the last r = MT seconds of the v(t) or v(t; px, Kx, Dw) signal of Eqs.
(29)—(36). The difference of signals v(t) and vm(t) over the last t = MT
seconds
<t> = v — Vm (38)
with coordinates
<J>i= <t>[(N- M + i) T]
= v[(N - M + i) T]
- v„[(N - M + i) T] . (39)
of
Schemat
6.19.
Figure
Filter.
Adaptiv
Digital
782 ELEMENTS OF MODERN CONTROL THEORY
o
33
a
o
3
<
%§
>
c
3
&
4
<
*§b
5
c
.o
o
—
a
D.
<
©
<N
d
a>
fc-
I
h
ADAPTIVE CONTROL 783
where the constants A and B replace |KX| and /%. The total input signal
is
* Actually it can be shown that the desired portion of v(t) can be readily identified even if the
undesired portion, v"(t), is not negligible, providing that the relative damping of the latter is
small compared to v'(t). See Ref. 14.
784 ELEMENTS OF MODERN CONTROL THEORY
Vm
Vm (W—if+z)
(43)
_ vm (tw) _
M
d
X F(W+i) = 0 (44)
d(AA) i=l
d(AB)
f) F2(t*_*+i)
= 0. (45)
i= 1
These two equations, which are linear in AA and AB, may be readily
solved for AA and AB, giving the new estimate
A^ = A* + isA (46)
B*+i = B* + AB . (47)
open loop and attenuate with the small structural damping. In this
latter case the digital filter simply prevents closed-loop excitation of the
bending modes, which might make them unstable.
The operation of a typical launch vehicle autopilot employing the
configuration of Fig. 6.20 is shown in Fig. 6.22, which illustrates the
response to a one-degree step command. It is apparent that there is firm
control of the vehicle as long as there is a component in the error corre¬
sponding to the predominant rigid body modes; that is, as long as there
is a transient. As the transient signal dies out, however, the digital filter
output fades out and the control loop for the airframe opens. This
loss of control is especially objectionable when the airframe is aero-
dynamically unstable.
This feature represents the major problem in the application of the
digital adaptive filter to launch vehicle control systems. The difficulty is
that the digital adaptive filter in effect closes the loop for the rigid body
transient response but keeps the loop open for the body elastic oscil¬
lations and all other signals. This open-loop operation is the desired
condition as far as the body elastic oscillation is concerned. However,
there are desired signals, such as guidance commands and feedback
signals resulting from wind shear, that must be transmitted just as the
rigid body transients are transmitted. The frequency band of these
signals is usually significantly lower than the frequency of the rigid
body transient. In order for the control system to respond to these
desired signals, the signal rejected by the digital adaptive filter (residual)
may be passed through a secondary filter, as illustrated in Fig. 6.23. This
786 ELEMENTS OF MODERN CONTROL THEORY
CO
UJ
UJ
OC
'O
ll]
Q
W
a:
o
w
Q
TIME (sec)
filter will have a low cutoff frequency so that it will adequately pass
slowly varying command signals yet suppress the elastic oscillations.
Analysis indicates that a conventional linear filter may be suitable for
this secondary filtering operation. The combination of the digital
ADAPTIVE CONTROL 787
DIGITALADAPTIVEFILTER
I- 1FITTED
Figure 6.23. Block Diagram of Vehicle Control System with the Digital Adaptive
Filter and Secondary Linear Filter.
adaptive filter and secondary filter will then transmit the desired com¬
ponent of the total error signal but will reject body bending oscillations.
With a secondary filter added in the manner shown in Fig. 6.23, the
system whose response is illustrated in Fig. 6.22 now exhibits the response
shown in Fig. 6.24. It is apparent that there is no longer any loss of
control. For best operation the secondary filter is cut in when the digital
filter output reaches a predetermined level (at approximately two sec¬
onds in Fig. 6.24). This cut in time is not critical; there would be similar
results with a ± 50 percent tolerance on this value.
The response shown in Fig. 6.24 combines the speed and accuracy of
the digital filter control shown in Fig. 6.22 with the steady-state stability
of the secondary filter. A small amplitude residual oscillation with the
low predominant frequency of the sluggish secondary filter can be ob¬
served in Fig. 6.24, but the overall response is still nearly ideal.
In Figs. 6.22 and 6.24 the system, including the bending modes, was
initially at rest at time t = 0 when the step input was applied. Figure
6.25 shows the system response when the three bending modes are
already oscillating with a rather high amplitude at the time of applying
788 ELEMENTS OF MODERN CONTROL THEORY
DEGRE
0 5 1.0 1.5 2.0 2.5 3.0 3 5
I
....
ml
:i i :
' ;i igl tb ;[ l;i:
0.8
*4r
35
it::
liii
titii ft
1
du JliL
si |Hj ii ii SSI
m 1 Iliii f
r tt
£ Si
I S It;
1 HH j:t.
uT n:
• • •*
liii
Hi:
did. jf m: Y % Fm fr~ rfff P
5a .
1 tf8 j!i
* - i
»+J
ttr
1 1 H1SB IS
.... — -
0
Pji L
::::
i m hi;
4—
Tr
jitt
Hi 2 HI i{I] hit IfIi
—
-0.8
ttr
I §gjf i fH: d; • tSf| tjf Tf j|: liii Hjf frtf frfi
-1.6 1i i | 11§ mm
m lip!Hii
Us U} Ii iiliiilil
Siti
HHiiiii:
ii j
f
1 IIi jjj jl
1 Jl |
11 i ]
s
c) 0.5 1.0 i .5 2 .n 2 !.5 3 .0 3.
TIME (sec)
Figure 6.24. Basic Transient Response as in Fig. 6.22 but Control Transferred to
Secondary Filters at t = 1.96 Sec.
the step input command. It can be observed in Fig. 6.25 that the digital
filter output curve regains its clear damped sinusoidal form after the first
quarter cycle, although in the total error, v(t) = 1 - 0(t), the rigid body
ADAPTIVE CONTROL 789
-0.8
-1.6
1.5 2.0
TIME (sec)
Figure 6.25. Basic Transient Response as in Fig. 6.22, but with Applied Step.
data are insufficient for positive identification of the rigid body mode;
consequently, some of the bending signal component is seen in the
digital filter output during this period. Comparison of the digital filter
output in Figs. 6.22 and 6.24 reveals that the presence of body bending
oscillations does not cause significant deterioration in the rigid body
response. The bending modes still oscillate essentially with open-loop
frequency and damping. Different amplitude selections of the initial
bending oscillations will have varying effects on the late part of the sys¬
tem response, but the transient remains essentially unaffected as long as
the bending amplitude does not much exceed half the amplitude of the
step command.
What would happen in the case shown in Fig. 6.25 if no digital filtering
were incorporated in the control system is revealed by Fig. 6.26. Com¬
parison of Figs. 6.25 and 6.26 makes it possible to appreciate the effect
of the digital filter with signal decomposition in stabilizing the bending
modes.
Instead of using a secondary filter, another possible solution to the
loss-of-control problem can be the periodic restarting of the fitting
process. This simply consists of periodically resetting the length of the
fitting interval to M = 1 sample in the computer. Then the computer
raises this number one sample at a time until Mmax, the ultimate length of
the fitting interval, is reached. Such restarting will assure the continuous
supply of small transients because there always will be some small error
at the instant of restart. This initial error is then eliminated by the control
action. The resulting sequence of successive small transients will assure
continued stability. It also assures the stepwise following of any con¬
tinuous control signals dc that may be present. Discontinuities in the
command dc or its lower derivatives must be monitored. Such discon¬
tinuities mark the start of new transients and should occasion the
switching to M = 1 and the ensuing development of a new fitting
cycle.
The discussion in the preceding paragraphs assumes that the length of
the fitting interval is stationary; that is, the length of fitting interval does
not change from one sample to the next. This type of operation is only
justified if the entire fitting interval is filled with a signal branch of a
transient. Consequently, the length of the fitting interval must be made
variable when new transients start. Immediately after the start of a new
transient, the fitting interval will be only one sample. Then the fitting
interval will gradually increase until it reaches the length of the desired
stationary fitting interval. This transition is referred to as the fade-in
period.
ADAPTIVE CONTROL 791
DEGRE
0 0.5 1.0 1.5 2.0
TIME (sec)
a. The curve fitting process does not start until the fourth sample, thus
allowing a new unfiltered excitation to the bending mode.
b. The phase relationship between the bending frequency and the input
is less favorable to curve fitting as each succeeding step is applied.
Note that the succeeding steps of Fig. 6.31 do not produce the same trend.
Figure 6.32 shows the response to a ramp input. Since the system
used here has only one integration in the open loop, a steady-state error
794 ELEMENTS OF MODERN CONTROL THEORY
0.20
0.15
0.10
0.05
-0.05
-0.10
0 8 16 24 32 40 48 56
will develop in the rigid body output response. This is clearly observed
in the figure. Also apparent is the fact that the rigid body output
ADAPTIVE CONTROL 795
LEGEND
- INPUT COMMAND
AMPLITUDE
SIGNAL
ERROR
COMMAND
AND
CURVE,
FITTED
OUTPUT,
BODY
RIGID \
\
A
\\ / /
\j
/
\ /
\ /
\/
y ^ — • /
/V--
V
/ \
u
<y V
/
21 -
0
-
10 J-
-
L
- -
30 4ii
- -
50
-
60
-
70 75
Figure 6.31. Response to an Alternating Step Input with a Half -Period of Twelve Samples.
Thus, while the Digital Adaptive Filter system has provided an elegant
solution to the main problem, its application to launch vehicle auto¬
pilots is frustrated by difficulties that present no problem in conventional
approaches. Whether some bright control engineer will be blessed with
new insight to revitalize this method remains to be seen.
A simple and highly effective way to gain stabilize the bending modes
in a feedback control system is to employ notch filters whose resonant
W
Q
P
H
P
<
P
<
£
O
oo
a
O
cd
£
C
U
<
H
(X
>*
Q
O
CQ
a
o
5
Figure 6.33. Control System Response with Variation in Sampling Rate per
Rigid Body Cycle.
ADAPTIVE CONTROL 799
AMPLITUDE
SIGNAL
ERROR
COMMAND
AND
CURVE,
FITTED
OUTPUT,
BODY
RIGID
Figure 6.34. Control System Response with Body Bending Frequency Equal to Rigid
Body Frequency.
RATE GYRO
NO. 1
1 1 1
VEHICLE 1 1 16
ACTUATOR
DYNAMICS
<XH-
RATE
NO.
GYRO
2
LP 1 1 1
SPECTRAL
FILTERS
DISPLACEMENT
GYRO
frequencies are set equal to the bending mode frequencies. There are
two fundamental difficulties in this approach; namely the bending mode
frequencies are not known a priori with sufficient accuracy, and they
vary with flight time due to variation in vehicle inertial properties.
It would appear that if one could track the bending mode frequencies
automatically during flight and use this information to vary the resonant
frequency of a notch filter accordingly, the problem would be solved.
This approach has indeed been studied extensively and the methods
proposed differ only in the means whereby the bending mode frequencies
are identified and in the manner of mechanizing the notch filter. The
two most promising techniques are discussed below.
T T (48)
fr(t) = f(t), - - < t — = 0, otherwise.
where
(50)
(51)
If we let
*The mathematically sophisticated reader will recognize the following presentation as a heuristic
development of the basic results of Fourier Integral theory.
802 ELEMENTS OF MODERN CONTROL THEORY
2irn
A 2 00 (52)
f r(t) = — + — X] cos cont + B„ sin co„ t)
i t „=1
f T/2
An = I f(t) COSCOrt,
t dt (53)
J -T/2
fT/2
B n= f(t) sin con t dt . (54)
•'-T/2
ADAPTIVE CONTROL 803
Actin = w,+i — wn
‘27r(n + 1) 27rn
T Y~
2ir y
~ Y '
\ 1 x°°' (55)
fr(t) = — H— (A, cos w„ t + B, sin w, t) Aw, .
TJ T n=J
,
Aq Ao AcOq
T 2ir
Ao cos coot
Acoq
2ir
\ 1 00 (56)
f„(t) = — — + - 23 (A, cos w, t + B„ sin u, t) Aw, .
T7T »= 0
where
Equation (57) is one form of the Fourier Integral. The functions A(«)
and B(to) are the cosine and sine transforms, or Fourier transforms, of the
function f(t).
Note that in a Fourier series we get contributions from integral values
of wt/2 7T. In a Fourier integral there are contributions from every real
value of co.
For present purposes the importance of the representation in Eq. (57)
is that the total squared amplitude of f(t)at the frequency co4is
the other hand, when the bending mode stability is poor, the
bending energy is high and the bending frequencies are well
identified by the spectral system. The performance variation
is thus in the proper direction.
The same study181 also noted that the third mode identifi¬
cation was, in general, poorer than either the first or the
second mode identification. This was attributed to several
possible causes: '
a. There is an attenuation of high frequencies due to system
lags, and thus the third mode energy is in general lower
than the first and second mode energy.
b. The spectral filter accuracy decreases as the number of
samples per tuned frequency period decreases, thus the
accuracy of the higher frequency spectral filters is not as
high as the lower frequency spectral filters.
c. The actual frequency being identified is the closed-loop
bending frequency, which may shift more radically for
the third bending mode.
The general question of how close the lowest bending mode
frequency can be to the rigid body frequency without having
the notch filter phase lag deteriorate rigid body response to
unacceptable proportions has not been investigated.
2
C0j>
(61)
ei = A sin cot
COy 2 (62)
- TT- j + WT
AK sin cot
— + tan /3
w
(63)
1 + — tan 0
u
where
co0= UT(0)
0=2 n + sin2 —
)
n = < >
7r
and < > denotes the maximum integer of the argument; e.g., <23/4> = 5.
It is noted that the system is inherently stable (in this simplified
version). Furthermore, the manner in which the input signal amplitude,
integrator gain, and initial setting of wT influence the speed with which
uiT approaches « is indicated explicitly.
A means whereby the signal frequency is tracked and the notch filter
resonant frequency set automatically is shown in the schematic of Fig.
6.40.
The demodulator is half wave and phase sensitive with a transistor
switch performing the demodulation function. To describe the operation,
we assume that e<„ is composed of a single signal of fixed frequency.
Now since e'in and ei differ in phase only by the contribution of the
(Krs2 + 1) term, we see that e,'„ is either in phase or 180° out of phase
with ei. Assuming that the resonant frequency of the notch filter is above
the frequency of the input signal ein, we note that ei„ and ei are in phase.
Consequently, the demodulator output is positive, which, in turn, acts to
increase KT, thereby causing the resonant frequency of the notch filter
to decrease. If the notch filter resonant frequency is below the input
signal frequency, then e,;n and ei are 180° out of phase and the reverse
occurs. When the two frequencies are equal, then the phase difference
ADAPTIVE CONTROL 809
between e,;„ and ei is 90° and the average value of ec is zero and eK nulls
to a fixed value that is a measure of the input signal frequency.
It is easy to see that, in the steady state, the overall transfer function
is
ei _ Krs2 + 1
&in K7.S2 -|- T2S-j- 1
where
1
Kr =
t32s2 + s2
(raS + l)2 '
This has the general form of a complex lag-lead network. The impor¬
tant feature here is that after the bending mode frequency has been
tracked the filter may be used for phase stabilization of the bending
mode rather than gain stabilization, which is achieved by the use of a
notch filter.
With phase stabilization the control system provides the proper gain
and phase characteristics at the bending mode frequency to obtain a
closed-loop damping of the mode greater than the open-loop damping.
Gain stabilization, on the other hand, provides enough attenuation at
the bending mode frequency to ensure system stability regardless of
ADAPTIVE CONTROL 811
e
1
R
7
vW
/
ek
b. OUTPUT SIGNAL, e
-
— - — —
'
in
c. INTEGRATOR (XJTPUT, e,
k
NOTCH FILTER
I_ _ _ _ _ ,_ _ _ _
— = - — + “p2_ . (65)
e<n s2 + 2£p <jivs + Up2
Si s2
I Closed Open High Pass
II Open Closed Low Pass
III Closed Closed All Pass
ADAPTIVE CONTROL 813
AMP
RAT
(deg)
(dB)
PHA NORMALIZED
FREQUENCY,u/cjp
ELEMENTS OF MODERN CONTROL THEORY
814
AM
RAT
(deg
(dB)
PHA
0.1 0.2 0.4 0.6 0.8 1 2 4 6 8 10
NORMALIZED
FREQUENCY,
<j/(jp
Figure 6.45. Frequency Response of Notch Filter, Eq. (66).
ADAPTIVE CONTROL 815
II
III
er — C sin (coat —
F <p)
(67)
0 < y? < 90° .
816 ELEMENTS OF MODERN CONTROL THEORY
1 (68)
= —AC cos ^ .
When = «0, that is, when <p= 90°, we see that the average value of
ed is zero, indicating that in the steady state the resonant frequency of
the filter is equal to the frequency of the input signal.
After expressing the output of the chopper demodulator in a Fourier
series, the fundamental (d.c.) component is found to be
2A COS ^
7T
(69)
fund
ed (70)
fund
Remark: Many studies have shown that the adaptive tracking filter con¬
cept is feasible and practical. Among the objections that may
be leveled against it are that the instrumentation is relatively
complex and, therefore, involves a compromise with reliability.
There are also lower limits in the ratio of allowable first
bending mode to rigid body frequency. Nevertheless, this
approach is a prime candidate for adaptive control of highly
flexible launch vehicles.
0 — Or 4" q(j) .
;
(71)
a a qlj>
i=i
where
01 % O' 10 0*11 G In Or
$2 021 • • • * 0*271 q CD
=
q(2)
(72)
. '
where
o+o = 1 for i = 1, 2, • • •, n + 1 .
© = crq (73)
q = o--1© (74)
where
Zio z* z. z n+1,0
Zn z, z* z n+ 1,1
z. z. z n+1,2 (75)
Zi z. z- z n+\,n
D = det [c]
(76)
If we define
(77)
n+l
(78)
X Pi(<r)Oi.
i=l
n+l
where 08-is the output of the rate gyro located at station i, and P<(cr) is as
defined by Eq. (77).
In the case of n = 2, for example (which presupposes that bending
modes higher than the second may be neglected), we find that
X P.fr) 0.
where
ADAPTIVE CONTROL 821
Zso
Ptto
D -5( 011022 012 021
)
D = ((72
1032 02203l) (oil O'32 — 01203l) T (oil 022 — (71202])
Remark: Subject to the stipulation that bending mode data are known,
the above method is an elegant and attractive means of ob¬
taining a theoretically perfect filtering of bending mode signals,
which is in no way affected by the frequency separation be¬
tween the desired and parasitic signals. Extra sensors are the
only additional equipment required and there is no need for
“exotic” signal processing.
In application to real vehicles there are two obvious limita¬
tions: (1) there are inherent limits in the accuracy of bending
mode data, and (2) the bending mode properties vary with
flight time, due to varying properties of the vehicle as fuel is
expended.
Figure 6.49. Block Diagram of System for Processing Attitude and Attitude Rate Signals.
822 ELEMENTS OF MODERN CONTROL THEORY
VEHICLE DYNAMICS
6r = He 5 + Ha a (Al)
(s + Kc)5 = Kc 8C (A4)
6e — 6c 6? (A6)
W
a — — T aw (A7)
U0
Ww
(A8)
a* ~ ~ u; •
Bending displacement
The vehicle geometry and sign conventions are shown in Figs. 6. A 1 and
6.A2. Typical normalized mode shapes are shown in Fig. 6.A3. Nomen¬
clature is defined as follows.
Sloshing and engine inertial effects have been neglected. The engine
actuator has been represented by a' simple first order lag and gyro
dynamics have been neglected. This is a valid approximation for studying
the lower frequency elastic modes.
The above model is adequate for exhibiting the salient features of the
adaptive methods considered in this chapter.
In studying the stability properties of the system one may also assume
that
Kr 0
and
a 6r .
REFERENCES
C(s)
G(s) .
R(s)
B(s)
C(b)
M(s) .
R(s)
df = (Rjj s -f- 1) da .
a. positive.
b. negative.
/
Use Eqs. (7)—( 10) as the description of the system with the rate
feedback as shown in Fig. 6.12. Take for the system parameters
Ma = 1 Kc = 15
Me = 4 Kx = 3
r(i) = o Tc = 340,000
M(1) = 1600
x = Ax + Bu
832 ELEMENTS OF MODERN CONTROL THEORY
C(s) = jA(b)
where C(s), R(s), B(s), and E(s) represent the output, input, feed¬
back, and error signals, respectively. The quantities KA, Kfl, and
Meare constants.
Show that the closed-loop transfer function is
C(s) _ _ w2
_ (a)
R(s) s2 -f 2 f oi s + co2
where
u* = MeIvA
f = 2 V^mTKx•
ADAPTIVE CONTROL 833
One of the more exciting concepts that has arisen in recent years is that
of learning control. The basic idea is due to Rosenblatt and his co¬
workers at Cornell University, and stems from an attempt to model the
functioning of the human brain by a network of threshold logic units/4)
The initial experiments produced quite dramatic results and the tempta¬
tion to characterize this model in anthropomorphic terms was irresistible.
However, the passage of time has had a sobering effect on some of the
overly optimistic claims and predictions which were made in the early
days. It was found that the device which they constructed (and called
the Perceptron) was nothing more than a pattern classifier whose esoteric
mysteries were dispelled when viewed in the light of switching theory.
The idea of learning control nevertheless exerted a mesmerizing influence
on researchers in automatic control theory. A good summary of recent
developments in the field is contained in Refs. 23 through 25, which also
include extensive bibliographies.
As far as application to space vehicle control systems is concerned,
significant progress has been made mainly in one area, that of closed-loop
optimal control. It will be recalled from Chapter Five that, in general,
the algorithms of optimal control theory generate trajectories which are
open- loop; that is, they are a function of the initial state of the system.
In practical situations, we usually require closed- loop control, i.e., a
control which is a function of the current state of the system.
An attempt to achieve closed-loop optimal control via the concepts of
learning theory exploits the so-called “generalizing” feature of threshold
logic networks.
Basic studies in this area have been made by Smithd) and Mendel. (2)
The material in this chapter is based on an investigation of a space
vehicle reorientation problem conducted by the author/22)
In the following sections, we restrict ourselves to the time optimal
problem where it is known that the control is bang-bang. All of the tools
834
LEARNING CONTROL SYSTEMS 835
needed to solve the problem are developed from first principles. Our
problem is one of designing a time optimal controller in which the
controller output is a function of the current state of the system, i.e.,
closed loop.
H = -1 + Xr (h + Bu) (3)
where
m
X (4)
dz
'
X1 »
xi fa
PATTERN
•
CLASSIFIER
•
•
Xd- - %º
PATTERN
"% Occasionally we shall write Xi to denote a specific class of pattern vectors, and X?1 for the
jth member of this class.
LEARNING CONTROL SYSTEMS 837
j, k = 1, 2, • R
j 7^ k .
Clearly, the patterns are linearly separable if, and only if, there exists a
hyperplane such that all the points (patterns) belonging to category A
are on one side of the hyperplane, and all the points (patterns) belonging
to category B are on the other side of this hyperplane. If we define a d
dimensional vector W, whose components are wi, w2, . , wd, then the
equation of the hyperplane takes the form
W •X = — wd+i . (9)
belong to category B. Since Eq. (9) and set (10) comprise the complete
training set, we have
N = L + M (12)
Y aU) %OL> 0
(13)
j = 1, 2, ') L
Ybw %a < 0
(14)
k = 1, 2, - , M .
If sets (10) and (11) are linearly separable, then a vector, 5, satisfying
(13) and (14) can always be found. These weights are then used in the
TLU of Fig. 7.3, which then acts as the pattern classifier. We turn to a
discussion of methods for calculatings.
When the patterns are presented cyclically, each pass through of the
training set is called an iteration.
Before training begins, the TLU weights (i.e.,«) are set to any arbitrary
value, say zero. If the TLU correctly identifies a pattern, no adjustments
are made in a. Suppose now that the TLU responds incorrectly to an
augmented pattern vector Y. Then the current a is changed to a new
value, «, as follows. '
a = a T" pY
if Y belongs to category A
(15)
a1 = a — pY
if Y belongs to category B
where p is a scalar greater than zero.
If p is constant at each trial, this is known as the fixed increment rule.
As long as the patterns in the training set are linearly separable, the
above procedure will generate an a, which will correctly identify all the
patterns in the training set after a finite number of iterations.
One of the earliest proofs of convergence of the above procedure was
given by Block. (4) Many other investigators have derived alternate proofs.
There exist variations in the above procedure whereby p is modified
in the course of the iteration in order to speed up the convergence. A
full discussion of these is contained in Nilsson’s book. (3)
D = [ Y,<*> ]t % (16)
- [ Y6(1) ]r
- [ Y(,(2) ]*•
- [ Y»<*>]r
842 ELEMENTS OF MODERN CONTROL THEORY
D« > 0.
The condition
*Also called generalized inverse. See Ref. 1 1 An elementary discussion of this topic is contained
in Appendix A.
LEARNING CONTROL SYSTEMS 843
0 1
0 0
y/> = Y (2) -
0 1A ~ 0
1 1
1 1
0 1
y,(3) = 1
li
0
1 1
0 0 1 0
1 0 1 1
v ,(2) - v (3) - v (4) _
* B — I B — 1 B —
1 1 1 0
1 1 1 1
-
-
_ _
— “ — —
Wi 3
W2 -2
w3 -3
w4 1
_ — _ _
Q Denotes Category B
Table
1.Summary
7.
of
Correction
Error
Procedure
ELEMENTS OF MODERN CONTROL THEORY
846
C
Procedure
Correction
Error
of
1.Summary
7.
Table
( ontinued)
LEARNING CONTROL SYSTEMS 847
0 0 0 1
1 0 0 1
1 0 1 1
1 1 0 1
0 -1 -1 -1
0 0 -1
1 -1 -1 -1
0 -1 0 -1
-1 1 1 1 1 1 -1 1
-1 -1 -1 1 -1 1 -1 -1
4 -1 -1 1 -1 -1 -1 -1 1
2 1 0 0 0 -1 1 -1
If we now take
/3r(o) = [1 1 1]
1
-1
«(0)
-1
.5
Consequently
.5
1.5
.5
.5
Da(o)
1.5
.5
.5
.5
ELEMENTS OF MODERN CONTROL THEORY
848
Since all the components of this vector are positive, «(o) is a solution
vector.
Thus, the Ho-Kashyap method obtained a solution with only one
iteration. .
Finally, we seek to obtain the solution vector, a, by the algorithm of
Eqs. (18) and (19). For starting values we take
p — 1
a(o) - 0
7r = [1 1 1 1 1 1 1 1] •
ar(l) [4 -4 -4 0]
[Da(l)]r -
[0 4 0 0 0 4 4 4]
zr(l) [-1 3 -1 -1 —
1 3 3 3]
<xT(2) -
[8 -4 -4 4]
[Da(2)]r =
[4 12 8 8 4 0 -4 0]
zr(2) =
[3 11 7 7 3 - 1 -5 - 1]
«r(3) =
[6 -8 -8 -2]
[Da(3)]r =
[-2 4 -4 -4 18 10 12 1
zr(3) _
[-3 3 -5 -5 17 9 11 9]
ar(4) =
[10 -6 -6 4]
[Da(4)]r =
[4 14 8 8 8 2 -2 2]
zr(4) =
[3 13 7 7 7 1 -3 1]
«r( 5) =
[8 -8 -8 2]
[Da(5)F =
[2 10 2 2 14 6 6 6] .
WX = - wd+i . (26)
If there are many points in the pattern set for which f is small, then it
seems plausible that new points (i.e., not in the training set) which are
near training set points having a small t have a significant probability of
being incorrectly classified by the TLU. For this reason, it appears
desirable to orient the separating hyperplane such that the minimum (
850 ELEMENTS OF MODERN CONTROL THEORY
C(N, d) (28)
><o
%¡
(0 - inrry, ^
Cover has shown that:(H) “If each of the linearly separable dichot¬
omies of N patterns in d space is equally probable, then the probability,
A (N, d), that a new pattern X is ambiguous relative to a random linearly
separable dichotomy is
A(N, d) gs — —- . (31)
N - d
N
(32)
(N - 2d)2 '
For this case we find that a new pattern has a probability A (4000, 80)
« 0.020 of being ambiguous relative to a random dichotomy. Of course,
in actual practice the random stipulations of Cover’s result are not
strictly met. Nevertheless, this serves to substantiate in some degree the
extraordinary capability of a TLU to correctly identify new patterns,
when it has been “trained” on only a small fraction of the patterns
likely to be encountered.
Note that when a new pattern is equally likely to be on either side of
the separating hyperplane, the probability of misclassification is 1/2
A (N, dj.
7.3.1 Coding
where a < and b, are known numbers. We may quantize this signal into
dj intervals of length, A<, in which case we have
interval
number z i in the interval Xir
1 ai to [aj + Aj [0 . 0 1]
2 [ai -f- A,-] to [a,- + 2Aj] [0 . Oil]
By constructing the pattern vectors in this way, we are assured that they
constitute a linearly independent set. *
There are otner ways of constructing linearly independent pattern
vector sets. For instance, using the so-called “one out of n” code, we have
interval
T
number X<
1 [0 •••01]
2 [0 • -010]
3 [0 •0100]
4 [0 0 1 0 0 0]
di [1 0 0].
di components
implies
Xj — X2 — — — Xm — 0.
854 ELEMENTS OF MODERN CONTROL THEORY
The above codes are not the most efficient in terms of using a mini¬
mum number of components in the pattern vector.
For example, if d< < 2", where M is a positive integer, then a simple
binary code would use only M components in the pattern vector. Thus,
if d<= 12, we could write the following.
interval interval
number X/ number x/
1 [0001] 7 [0111]
2 [0010] 8 [1000]
3 [0011] 9 [1001]
4 [0100] 10 [10101
5 [0101] 11 [1011]
6 [0110] 12 [1100]
However, the above pattern vectors are not linearly independent. Since
the property of linear independence is a fundamental requirement for the
TLU realizability theorems discussed in the next section, this type of
code will be employed exclusively in the remainder of the chapter.
Thus, the process of coding is nothing more than a transformation
from state space to pattern space. Physically this is accomplished by an
encoder. The scheme just described may be mechanized in the manner
shown in Fig. 7.6. This particular arrangement uses simple summing and
threshold elements and is, therefore, structurally analogous to a TLU.
= f(z) > 0
= f(z) < 0 .
LEARNING CONTROL SYSTEMS 855
ii(zi) = Xi- Wi
(35)
i = 1, 2, M
and
(i.e., there are no cross product terms), then each fj(zt) in Eq. (35) can
exactly match the corresponding f<(z,-) except for quantization errors.”
The scheme is illustrated in Fig. 7.7. If we use the method of Sect. 7.3. 1
for generating the linearly independent pattern vectors for zt-, and if z<
then is quantized into ds intervals, the corresponding pattern vector X<
will have d, components. We let F,(3) denote the value of fi(zt) at the
midpoint of the jth interval. We also define the matrix
[x^r '
[X.W] r
(38)
[X i (d^]r
W, = Pr1 F, (39)
where F, is the column vector whose components are F,(», j = 1,2. - , d*.
The inverse of Pt- always exists, since the row vectors are linearly
independent. We note that P, is in fact a d, X d< matrix of the form
all 1
zeros 1
P, = (40)
1 all
1 ones
all -1 1
zeros 1
Pr1 -1 (41)
-1 1 all
1 zeros
in the interval —1.0 < z1; < 2.5 using the TLU approximation technique
described above. Talcing d, = 7, which means that z< is quantized in
increments of 0.5, the computations and coding may be summarized as
follows.
1 -1 to -.5 -.247 [0 0 0 0 0 0 1]
2 -.5 to 0 .322 [0 0 0 0 0 1 1]
3 0 to .5 1.102 [0 0 0 0 1 1 1]
4 .5 to 1.0 2.290 [0 0 0 1 1 1 11
5 1 to 1.5 4.191 [0 0 1 1 1 1 1]
6 1.5 to 2.0 7.289 [0 1 1 1 1 1 1]
7 2.0 to 2.5 12.376 fl 1 1 1 1 1 1]
858 ELEMENTS OF MODERN CONTROL THEORY
From Eqs. (39) and (41), the weight vector, Wi, is found to be
(i «V
set of training pattern vectors, and hope that “new” pattern vectors have
a high probability of being correctly identified. It will be recalied from
Sect. 7.2.6 that, with numbers typical of practical situations, a training
set constituting only 2 1/2 percent of the patterns in state space is suf¬
ficient to classify new patterns correctly with about 98 percent proba¬
bility. A TLU network constructed in this way could thus be expected
to be very effective and well within the usual engineering accuracy.
Such a procedure, however, implicitly assumes that a dichotomization
via TLU is possible for the switching surface, the form of which is not
known to begin with. In short, we must ascertain those properties of the
switching surface (the exact form of which is unknown) which ensure
that it is indeed “TLU realizable.” Significant results in this area have
been obtained by Berkovec.(16d7) They are broadly similar to those
derived by Smith (discussed earlier in this section), but are somewhat
more general. They are based on the concept of projectability, which may
be defined as:
z = Az + Bu
u < |u0|
* Actually Bercovec specifies a “1 out of n” code. However, it is easy to show that any linearly
independent code satisfies the requirements of the theorem.
LEARNING CONTROL SYSTEMS 861
a. Quantize each component, z < of the state vector, z, into d,- intervals.
M
b. Construct a linearly independent code, X,, for each z*. Define the
total pattern vector, X, by
Xr = [Xjr X2r . XMr) .
interval
number, Mi X/
1 [0 0 0 0 0 0 0 1]
2 [0 0 0 0 0 0 1 1]
3 [0 0 0 0 0 1 1 1]
4 [0 0 0 0 1 1 1 1]
i\ 5 [0 0 0 1 1 1 1 1]
6 [0 0 1 1 1 1 1 1]
7 [0 1 1 1 1 1 1 1]
8 [1 1 1 1 1 1 1 1]
1 [0 0 0 0 0 1]
2 [0 0 0 0 1 1]
3 [0 0 0 1 1 1]
z2 4 [0 0 1 1 1 1]
5 [0 1 1 1 1 1]
6 [1 1 1 1 1 1]
1 [0 0 0 0 11
2 [0 0 0 1 1]
Z3 3 [0 0 1 1 1]
4 [0 1 1 1 1]
5 [1 1 1 1 1]
The total pattern vector associated with the hypercube with an interval
number set of (2, 5, 4), say, would be
Xr = [0 00000110111110111 1],
Wr = [Wir W/ . Wmt] .
^ A(45,20)= 0.322
which is little better than tossing a coin, even though the training set
contained 45 percent of the squares in the complete state space.
(yA(1,000,90) 0.049 .
22 wf xf + wd+i
i=i
which is the linear discriminant function of Eq. (7), and the output
voltage, E0, would be the i„ of Fig. 7.3.
Many other mechanizations are possible. The study by Tanaka^S) con¬
tains a detailed description and analysis of the above scheme together
with five other possible TLU mechanizations, namely:
iR °
where lvv, lzz, and lyz are the moments and product of inertia, respectively,
about the indicated axes. 9 and 4/, the pitch and yaw angles, respectively,
and Ly and Lz denote the control torques about the indicated axes. The
control torques satisfy a saturation constraint of the form
|Ly(t)| < L0
(46)
|L,(t)| < L0 .
zi = 6 , 2,3 - 'P
(47)
z2 = 6 , z4 = i
Zi - z2 (48)
z3 = z4 (50)
z4 = a3 Ly -f- a2 L2 (51)
where
813 Ity^/(J^-yy
Izz -1
-yz ) • (54)
4(t)
Uy(t) = (55)
.. L.(t)
U,(t) = T (56)
-L'o
T = (L„a01/21 (57)
ELEMENTS OF MODERN CONTROL THEORY
868
Z\ = z2 (58)
z2 = U!/ “b bi u. (59)
z3 = z4 (60)
z4 = bi uv + b2 uz (61)
3* VI (62)
where
bl lyz/bz (64)
b2 (65)
z' = Az + Bu (66)
where
zT = [Zi z2 z3 z4]
ur = [u„ u«]
0 10 0
OOOO
0 0 0 1
0 0 0 0
0 0
B “ oo
bi b2
LEARNING CONTROL SYSTEMS 869
M < (67)
We now formulate the main problem as follows. Given system Eq. (66)
and the control function constraint, Eq. (67), determine the control
function, u(t), which satisfies Eq. (67), And which transfers the system
from the initial state
z(0) = £ (68)
z (Tt) = 0 (69)
G = [B : AB j A2B : A3B]
is the rank 4. *
Thus, we know that there exists a control which can accomplish the
reorientation in finite time.
By invoking the maximum principle, it is found that the optimal
control must minimize the hamiltonian
where
(71)
As long as b2 9^ bA
ELEMENTS OF MODERN CONTROL THEORY
870
But
x; = 0 (75)
x; = -Xx (76)
X3 = 0 (77)
X4 = -X3 (78)
and
X3 = X30 (81)
The actual computation of u*(t) now requires that we solve a two point
boundary value problem. In this respect, one of the more powerful
techniques available in the present case is Neustadt’s method. (6) How¬
ever, this requires that the system be “normal.” Loosely speaking,
normality implies that the system is controllable with respect to each
component of the control vector. Using a simple testU) for normality,
we conclude that since the matrices
u = -Kz (85)
where
ELEMENTS OF MODERN CONTROL THEORY
872
ki k2 0 0
K =
0 0 k3 Ic,
z' = (A - BK) z
(86)
= Fz
where
0 1 0 0
— ki — k2 — b]k3 — bik4
(87)
0 0 0 1
ai = k2 + b2k4
We assume that
In this case, Routh’s criterion shows that Eq. (88) has no roots in the
R-H plane if each of thekf are positive and if the inequality
a42a4
aia2 — a3 > - (91)
B<3
is satisfied.
LEARNING CONTROL SYSTEMS 873
If k, = 0, then*
ai — k2
For this case, inequality (91) is automatically satisfied if ki, k2, and k3
are positive. In other words, the system is stable for all positive values of
k<.
We now write Eq. (88) as
where
02 = k3 (b* - bi2)
ft2 =
(ki b2k3)
/ 4 ktk3 (b2 - bt2)
(ki — b2k3)2 T 4 k]k3b2
A2 =
(ki + b2k3)
i - i/7T 4 kxk3 (b2 - bi2)
(ki — b2k3)2 + 4 kik3b2
It is apparent that when b2 — bf2 > 0, and ki and k3 are positive, then
the quantity inside the radical is always positive, and less than one. Thus,
The root locus for this case is shown in Fig. 7.13. Once can now pro¬
ceed along conventional lines to obtain a prescribed type of transient
response for the system. However, for a more precise analysis, one must
take account of the saturation constraint on u (via describing functions,
say). Since this is a routine (though cumbersome) procedure, it will not
be further pursued here.
One fact should be noted, however. Because of the nonlinearity in¬
troduced by the control saturation, the characteristics of the time response
will differ for different inputs or initial conditions. In a sense, one can
say that the conventional system can be “optimized” only for one (or a
small class of) inputs. This limitation does not exist with TLU optimal
controller.
I
m
B# = (BrB)~l Br . (Al)
C# = Cr(CCr)-1 (A2)
X = (BrB)-1 BrG
(A3)
= B#G .
Y = Cr(CCr)-1 H
(A4)
= C#H .
I, = BB# (A5)
875
876 ELEMENTS OF MODERN CONTROL THEORY
I* = C#C . (A6)
A = BC (A7)
A# = C# B# . (A8)
q = 1 - p . (B2)
and let*
(B4)
B,(k)
- t ( Nr
) P'<!-'
where the notation
ion(b) denotes the binomial coefficient, viz.
/ a\ _ a!
a|_ (B5)
\ b/ b! (a — b)!
k + 1
A„(k) (k) PAqAr-*+1 k + 1 - (N + 1) p
(B6)
A^k) (B7)
1 - Mk) < l + v~2
877
878 ELEMENTS OF MODERN CONTROL THEORY
where
k - Np (B8)
11 VpqN
Thus, for the special case of p = q = 1/2 and for tj sufficiently large, it
follows that
k + 1
sewo k + 1 - | (N+ 1)_
(B9)
Using the fact that the binomial coefficient satisfies the relation
O)-(.-b)
Eq. (B9) may be written as
N + 1 -d
IdH(f) L2
1
+ 1) -d J
(BIO)
where
d = N - k . (B 11)
y ( N \ ~ ( N\ (N ~ d) (B 12)
h\ i ;~Vd^(N-2d)
Using this approximation for the C function of Eq. (28), we find that
A(N, d), as given by Eq. (30), becomes*
A(N, d) (B 13)
while
N - 2d (B 15)
V~ Vn
1 N
(B 16)
rf (N - 2d)2 '
REFERENCES
1 0
0 0
Y^(2)
0 1
1 1
0 0
1 1
(3) YY4)
1 0
1 1
1 0
0 0
Y*<2> =
1 0
0 1
1 1
1 1
Yb«> =
0 0
1 1
— _
= 1 , N < d .
C = 2(d + 1) .
f (z) = |sin z
in the range 0 < z < Stt/2. Use a quantization level of ten equal
intervals.
INDEX
883
884 INDEX
Lag
Feedback systems, 13 compensation, 97
Filter pure (see Transport lag)
compensation, 97, 99-101 Lagrange
Kalman, 481, 497, 499 multiplier, 589
notch, 810 problem, 573
tracking, 805 Laplace transform, 190
Flexible launch vehicle, 817 Launch vehicle dynamics, 178, 823
Focus Lead compensator, 100
stable, 229 Legendre-Clebsch conditions, 577
unstable, 229 Limit cycle, 232
Forced systems (nonlinear), 357 Linear
Fourier discriminant function, 838
series, 247 separability, 843
transform, 463-64 systems (in optimal control), 653
Frequency Linearly independent
response, 13, 91, 93 codes, 853
tracking, 805 vectors, 853
Function space, 730 Linearly separable, 838
Lyapunov function, 304
Lur’e method, 31 1
Gain
margin, 22-24
sensitivity, 380 Matrix
Gaussian distribution, 450-51, 466 companion, 123, 126
Gradient method, 603 canonical, 126
Green’s function, 617 modal, 125
sensitivity, 404
system, 121
Hamilton-Jacobi equation, 659 transition, 124
Hamiltonian, 589 Vandermonde, 126
Hidden oscillations, 147 Maximum principle, 586
Ho-Kashyap method, 842-43 Mayer problem, 572
Hold circuit, 72 Mean, 451, 456
Hysteresis, 256, 257, 258, 259, 262 Model-reference, 754
Modified Z transform, 72
Moments (probability), 452
Ideal relay, 340
Impulse
function (see Delta function) Nichols plot, 29
response (see Weighting function) Nodes, 229
Indirect control, 313 Noise
Inner product, 730 colored, 506
Invariance, white, 470
absolute, 169 Nonlinearities
conditional, 169 describing function, 245
INDEX 885
TL
3280
9780876715536
02/11/2019 15:42-2