Adaptive Control For Distributed Leader-Following
Adaptive Control For Distributed Leader-Following
A Thesis
Presented to
the Faculty of the School of Engineering and Applied Science
University of Virginia
In Partial Fulfillment
of the Requirements for the Degree
Master of Science in Electrical Engineering
by
Ge Song
May 2015
APPROVAL SHEET
This thesis is submitted in partial fulfillment of the
requirements for the degree of
Master of Science in Electrical Engineering
Ge Song, Author
This thesis has been read and approved by the examining committee:
May, 2015
Acknowledgments
First, I would like to gratefully thank my advisor, Professor Gang Tao. Without
his patient guidance and persistent encouragement, I would never completed this
thesis. His deep knowledge and insight on control theory helps me to solve problems
from one to another. It is a great privilege to have studied under him through the
past two years.
Second, I would also like to extend my gratitude to my research group members
and friends at the University of Virginia who share ideas with me.
Finally, I would like to sincerely thank my parents for their help and love. Without
their continuous and unconditional support during my time in University of Virginia,
I would never got the opportunity to be the person I am today.
i
Abstract
1 Introduction 1
1.1 Research Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Literature Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3 Thesis Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2 Background 6
2.1 Control System Models . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.2 Signal Measures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.3 System Stability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.4 Classical Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
2.5 Adaptive Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
2.6 Graph Theory and Communication Topology . . . . . . . . . . . . . . 16
ii
iii
3.1 Follower state (solid) vs. leader state (dotted) with fixed gain control. 54
3.2 Follower state (solid) vs. leader state (dotted) with adaptive gain control. 54
3.3 Tracking errors with adaptive control vs. fixed gain control. . . . . . 55
3.4 Parameter errors ki (t) − ki∗ with adaptive control (i = 1, 2, 3). . . . . 55
3.5 Follower state (solid) vs. leader state (dotted) without disturbance
rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
3.6 Follower state (solid) vs. leader state (dotted) with adaptive distur-
bance rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
3.7 Tracking errors with adaptive disturbance rejection vs. without dis-
turbance rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
3.8 Follower state (solid) vs. leader state (dotted) without disturbance
rejection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
iv
v
3.9 Follower state (solid) vs. leader state (dotted) with adaptive distur-
bance rejection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
3.10 Tracking errors with adaptive disturbance rejection vs. without dis-
turbance rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
3.11 Follower state (solid) vs. leader state (dotted) with fixed gain control. 62
3.12 Follower state (solid) vs. leader state (dotted) with adaptive gain control. 62
3.13 Tracking errors with adaptive control vs. fixed gain control. . . . . . 63
3.14 Parameter errors Ki (t) − Ki∗ with adaptive control (i = 1, 2, 3). . . . . 63
3.15 Follower state (solid) vs. leader state (dotted) without disturbance
rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
3.16 Follower state (solid) vs. leader state (dotted) with adaptive distur-
bance rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
3.17 Tracking errors with adaptive disturbance rejection vs. without dis-
turbance rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
3.18 Follower state (solid) vs. leader state (dotted) without disturbance
rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
3.19 Follower state (solid) vs. leader state (dotted) with adaptive distur-
bance rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
3.20 Tracking errors with adaptive disturbance rejection vs. without dis-
turbance rejection. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
3.21 Follower state (solid) vs. leader state (dotted) with adaptive gain con-
trol (LDU decomposition). . . . . . . . . . . . . . . . . . . . . . . . . 71
3.22 Follower state (solid) vs. leader state (dotted) with adaptive gain con-
trol (LDU decomposition). . . . . . . . . . . . . . . . . . . . . . . . . 71
3.23 Tracking errors with adaptive control vs. fixed gain control (LDU
decomposition). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
vi
4.1 Case I: Interaction graph of two followers and one leader aaaaa aaaaaaaa 79
4.2 Case II: Interaction graph of two followers and one leader . . . . . . . 80
4.3 Case III: Interaction graph of two followers and one leader . . . . . . 80
4.4 Case IV: Interaction graph of two followers and one leader . . . . . . 80
4.5 A basic structure exists in directed graphs . . . . . . . . . . . . . . . 80
4.6 Interaction graph after adding one follower on follower v1 . . . . . . . 81
4.7 Interaction graph after adding one follower on follower v1 . . . . . . . 82
4.8 Interaction graph after adding one follower for follower vk . . . . . . . 83
4.9 Interaction graph of three follower agents and one leader . . . . . . . 94
4.10 State trajectories of three follower states xi and leader state x0 vs.
time(sec) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
4.11 Tracking errors between three follower states xi and leader state x0 vs.
time(sec) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
4.12 Interaction graph of five follower agents and one leader . . . . . . . . 96
4.13 State trajectories of five follower states xi and leader state x0 vs.
time(sec) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
4.14 Tracking errors between five follower states xi and leader state x0 vs.
time(sec) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
4.15 State trajectories of three follower states xi and leader state x0 with
disturbance rejection vs. time(sec) . . . . . . . . . . . . . . . . . . . 99
4.16 Tracking error between follower states xi and leader state x0 vs. time(sec) 99
Chapter 1
Introduction
1
2
With the development of sensor networks, control of multi-agent systems has been
emerging and has drawn lots of attention. Since the main objective in distributed
control of multi-agent systems is to enable a group of agents to perform a special
task, distributed control is also referred as the cooperative control. There are several
typical problems existing in the cooperative control including the consensus problem
[9, 37, 45], the formation problem [4, 8, 34] and the flocking problem [36, 39, 41].
Among those three typical cooperative control problems, the consensus problem
is the most important one because the consensus algorithm is basic and fundamen-
tal [26]. In particular, consensus means to reach an agreement on a certain quantity
of interest, namely their position or velocities [37]. As the most fundamental and
important control problem in cooperative control, Jadbabaie et al. [13] considered
such the leader-following consensus problem and proved that if all the agents were
jointly connected with their leader, their states would coverage to that of the leader as
time goes on. The controllabilty of the leader-following dynamic network was studied
in [14, 22, 46]. The leader-following consensus problem can be classified into two dif-
ferent kinds. The first one is the consensus regulation problem, which is also referred
as leaderless problem (behavior-based) [20, 27, 31, 35]. The other one is the consen-
sus tracking problem, also known as leader-following consensus(synchronization). For
leader-following consensus problems, there should exist an active leader whose states
keep changing in the multi-agent system [11, 21, 40, 50]. In another word, leader-
following consensus means that all follower agents eventually reach an agreement on
the state or output of a preassigned leader, which specifies a desired objective for all
3
with [28], [24] solved the consensus problem for a group of followers with different
dynamics under an external disturbances with an active leader which model can be
also different with the followers. Plant parameters of the followers in [24] are in
uncertainty. The distributed control protocol in [24] designed by model reference
adaptive control method and also uses the local information of an individual agent.
However, although the author said that some part of the leader’s parameter can be
unknown, this statement is not convinced. Motivations of the aforementioned research
are clear but in practical application sometimes conditions of those researches are
unsatisfied.
Motivated by the above observations, in this thesis we study the adaptive leader-
following consensus problem of a group of linear dynamics agents guided by an active
real leader with external disturbances under directed graphs. Different from [28], sys-
tem parameter matrices in this thesis can be completely different with each other, and
also the system parameter matrices of the leader are supposed to be different from all
the followers. Comparing to [24], dynamic parameters can be completely unknown
including dynamic parameters of all followers and the leader. This assumption is
meaningful because in the real engineering environment, sometimes it is difficult to get
accurate dynamic parameters. In the meanwhile, like many papers [12, 17–19, 23, 44],
this thesis also considers the external disturbances acting on the real application.
Hence, we develop a new protocol structure based on the principle of adaptive con-
trol method in order to solve the consensus leader-following problem with an external
disturbances such that all follower agents can track the active leader with a desired
reference signal. Then, the performance of a close-loop system is analyzed to verify
the designed control protocol with the corresponding adaptive laws. Finally, sev-
eral simulation results will be displayed to verify the effectiveness of our proposed
theoretical protocol.
5
Background
Before starting to discuss the adaptive control of multi-agent system, some basic
background about control system and multi-agent system needs to be presented. In
this chapter, topics about control system including control system modeling, system
stability, classical control and adaptive control are presented first. Topics following
the background of control system are some basic background about topology.
6
7
Fi (y (p) (t), . . . , y (1) (t), y(t), u(p) (t), . . . , u(1) (t), u(t), t) = 0, t ≥ t0 , (2.1)
di y(t) di u(t)
i = 1, 2, . . . , l. y (i) (t) and u(i) (t) denote the ith time derivatives dti
and dti
of
y(t) and u(t), with a common notation ẏ(t) = y (1) (t), u̇(t) = u(1) (t) and ÿ(t) =
y (2) (t), ü(t) = u(2) (t). A special form of Fi depends on a special system under consid-
eration.
A n-th order dynamic system can be described by a group of n interactive first-
order differential equations
the non-linear systems to the linear system on (x0 , u0 ) by Taylor expansion,1 which
is
δ ẋ = Aδx + Bδu
with
∂f1 ∂f1 ∂f1 ∂f1
∂x1 ... ∂xn ∂u1 ... ∂un
∂f .
..
∂f .
..
A= |(x ,u ) = .. . , B= |(x ,u ) = .. . ,
∂x 0 0 ∂u 0 0
∂fn ∂fn ∂fn ∂fn
∂x1
... ∂xn ∂u1
... ∂un
|(x0 ,u0 ) |(x0 ,u0 )
∂h1 ∂h1 ∂h1 ∂h1
∂x1 ... ∂xn ∂u1 ... ∂un
∂h .
..
∂h .
..
C= |(x ,u ) = .. . ,D = |(x ,u ) = .. . .
∂x 0 0 ∂u 0 0
∂hn ∂hn ∂hn ∂hn
∂x1
... ∂xn ∂u1
... ∂un
|(x0 ,u0 ) |(x0 ,u0 )
.
Since linearization always be made in a small range, in oder to denote conveniently
usually people express linearized system (2.4) as
ẋ = Ax + Bu
y = Cx + Du (2.5)
(2.5) is called the “linearized system” and the method used in this linearization pro-
cess is called ‘Lyapunov linearization method”. Lyapunov linearization method is
concerned with the local stability of a nonlinear system. It is a formalization of the
intuition that a nonlinear system method should behave similarly to its linearized
1
Often, (x0 , u0 ) is an equilibrium point which means f (x0 , u0 ) = 0
9
approximation for small range motions. Because all physical systems are inherently
nonlinear, Lyapunov’s linearizion method serves as the fundamental justification of
using linear control techniques in practice. More details about Lyapunov linearization
method can be found in literature [38].
ẋ = A(t)x + B(t)u
ẋ = Ax + Bu
y = Cx + Du, t ≥ t0 , (2.7)
where A ∈ Rn×n , B ∈ Rn×M , C ∈ Rq×n , and D ∈ Rq×M are all constant system
10
Z t
A(t−t0 )
x(t) = e x0 + eA(t−τ ) Bu(τ )dτ (2.8)
t0
Kp = LD∗ U (2.9)
for some M × M unit(i.e., with all diagonal elements being 1) lower triangular matrix
L and unit upper triangular matrix U , and
∗ ∆2 ∆M
D = diag {d∗1 , d∗2 , . . . , d∗M } = diag ∆1 , ,..., . (2.10)
∆1 ∆M −1
In order to have a unique measurement for vector signals, a new concept called
“norm” is defined. Consider a vector signal x(t) = [x1 (t), . . . , xn (t)]T ∈ Rn . x(t) =
[x1 (t), . . . , xn (t)]T is a vector at any t, and is a vector function as t changes. Vector
norms can measure vectors, and while function norms can measure vector functions.
Since in control systems there exist many signal vectors that need to be measured,
it is imperative to introduce signal norms before we develop a adaptive control scheme.
The L1 , L2 , and L∞ norms are defined below as
Z ∞
kx(·)k1 = (|x1 (t)| + · · · + |xn (t)|)dt, (2.14)
0
sZ
∞
kx(·)k2 = x21 (t) + · · · + xn1 (t), (2.15)
0
Remark 2.1 By definition of L∞ , we can conclude that x(t) ∈ L∞ iff x(t) is bounded,
i.e., x(t) ∈ L∞ and x(t) is bounded are equivalent.
Stability definitions Since all systems to be discussed in this thesis are LTI sys-
tems, we will introduce the definition of stability for LTI [2]
Definition 2.2. The response of ẋ(t) = Ax(t) is marginally stable or stable in the
sense of Lyapunov if every finite initial state x0 excites a bounded response. It is
asymptotically stable if every finite state excites a bounded response which, in addition,
approaches 0 as t → ∞.
Usually, we do not use the definition to check the stability of a LTI system.
Theorem 2.1 can help us to check the stability of a LTI system more quickly.
Theorem 2.1. The equation ẋ(t) = Ax(t) is marginally stable if and only if all eigen-
values of A have zero or negative real parts and those with zero real parts are simple
roots of the minimal polynomial of A. The equation ẋ(t) = Ax(t) is asymptotically
stable if and only if all eigenvalues of A have negative real parts.
Also in the Lyapunov sense, we can check the stability of matrix A by Lyapunov
theorem
Theorem 2.1 and Theorem 2.2 are theorems which we usually used to check the
stability of close-looped system by classical control. But for adaptive control, The-
orem 2.1 and Theorem 2.2 would not work because in adaptive control there exist
14
uncertainty on the dynamics model, we have no access to get a set of accurate system
parameters. So we introduce a new method called Lyapunov direct method to help
us check the system stability when applying adaptive control.
Theorem 2.3. (Lyapunov direct method) If in some Ball B(h) there exists a positive
definite function V (x, t) with V̇ ≤ 0, then the equilibrium state xe = 0 of the the
system ẋ = Ax is stable. If, in addition, V (x, t) is decrescent, then it is uniformly
stable.
The first thing we need to know about the classical control is the feedback is
pervasive. Feedback is a very crucial method to stabilize the unstable system stable
[1]. Usually, the output y(t) is fed back and compared with the input u(t). The block
diagram of the most classical feedback control system is shown in Fig 2.1. System
transfer function can be computed by block diagrams like Fig 2.1. Remind that in Fig
2.1, k is a constant through the whole control process. Whether control parameter k
can update or not is one of a big differences with adaptive control and the classical
control.
With the state space representation it is convenient to feedback the state variables
x(t) to the control signal u(t) instead of an input signal y(t). With this configuration
each state variable can be adjusted by a gain vector K to give the desired closed loop
15
u(t) y(t)
k G(s)
−
Figure 2.1: Block diagram of a system with output feedback and an adjustable pream-
plifier gain k.
poles. A typical control system represented with the state space representation uti-
lizing state feedback is displayed in Figure 2.2 [29] where double error lines represent
vector signals.
Unlike classical control systems, adaptive control system has capability to payload
system uncertainties including unknown system structures, system parameters and
other environmental uncertainties. An adaptive controller structure is designed based
on a known parameter case, however, the value of the adaptive controller parameter is
updated by parameter estimator instantly. Combining an adaptive controller with an
parameter estimators, adaptive control can achieve a desired performance by dealing
with variety kind of uncertainties.
There are two approaches to adaptive control design. The first one is direct adap-
−K
Figure 2.2: State space representation of a plant with state feedback. Reproduced
from [29].
16
tive control. For direct adaptive control, the parameter estimator estimates controller
parameters online directly according to the plant information of input/output. Plant
parameters are parametrized implicitly in terms of a set of parameters of a nominal
controller. The second approach is indirected adaptive control and is characterized
by estimating the parameters of the plant first. Then based on some design equation
to calculate control parameters to achieve the desired performance. Direct adaptive
control will be employed in this thesis.
Definition 2.3. By a graph G we mean a finite set V(G) = (vi , . . . , vn ), whose ele-
ments are called nodes or vertexes, together with set E(G ⊂ V × V), whose elements
are called edges. An edge is therefore an ordered pair of distinct vertexes.
Definition 2.4. A graph is called direct graph if and only if for all (vi , vj ) ∈ E(G),
the edge (vj , vj ) ∈ E(G), then the graph is said to be undirected. Otherwise, it is called
a directed graph.
Definition 2.5. An edge (vi , vj ) is said to incoming with respect to vj and outgoing
with respect to vi ) and can be represented as an arrow with vertex vi as its tail and
vertex as its head.
2 3 4
0 1 5
Figure 2.3: An illustrative of a multi-agent system with five followers and one leader
aaaaaa
aaaaaa
aaaaaa
aaaaaa
aaaaaa
aaaaaa
aaaaaa
aaaaaa
aaaaaa
aaaaaa
aaaaaa
Chapter 3
The final objective for this thesis is to design a distributed adaptive control scheme
of leader-following consensus problem for multi-agent system. However, before going
that far, it is important for us to know how adaptive control works for systems in-
cluding one follower and one leader. Fully understand this basic problem is important
for us to develop a more complex controller for multi-agent systems in Chapter 4.
For state tracking control problems, there are two different common cases we are
interested in. One is the control problem with a single input, the other one is the
control problem with multiple inputs. Adaptive control of single input is easier and
more fundamental than multiple input cases and adaptive control of multiple input
systems are more significant and practical. In this chapter, both of these two control
problems will be discussed.
18
19
Case I: Single input systems without disturbances In this case, d(t) = 0 and
dm (t) = 0. There exist no disturbance in the multi-agent system. The dynamic of
the follower and the leader can be re-presented as
Case II: Single input systems with a disturbance acting on the followers
In this case, d(t) 6= 0 and dm (t) = 0. The dynamic plant of the follower is
q
X
d(t) = d0 + dβ fβ (t) (3.6)
β=1
where d0 and dβ are unknown constant and fβ (t) is a bunch of known continuous
bounded functions for β = 1, 2, . . . , q for some q ≥ 0. The dynamic of the given leader
remains as (3.4).
Case III: Single input systems with disturbances acting on both the leader
and the followers In this case, d(t) 6= 0 and dm (t) 6= 0. When d(t) and dm (t) are
not equal zero, the follower plant is (3.5). The leader dynamic is
where
qm
X
dm (t) = dm0 + dmβ fmβ (t) (3.8)
β=1
with dm0 and dmβ are unknown constant and fmβ (t) are some known continuous
bounded functions for β = 1, 2, . . . , p for some p ≥ 0.
The control objective is to design a bounded state feedback control signal u(t) to
make the follower system state x(t) bounded and tracking the leader system xm (t)
asymptotically, i.e., limt→∞ (x(t) − xm (t)) = 0.
21
In section 3.1.1, the control input u(t) and the leader reference signal um (t) are
scalars, and parameters b and bm are vectors. However, in practical engineering
applications, there are many multiple inputs systems with u ∈ Rm , m 6= 1.
Consider a follower plant expressed as (3.9), and the leader is given as (3.10)
Case II: Multiple inputs systems with a disturbance acting on the follower
systems d(t) 6= 0 and dm (t) = 0. The dynamic plant of the follower is
22
q
X
d(t) = d0 + dβ fβ (t) (3.14)
β=1
where d0 ∈ Rp and dβ ∈ Rp are unknown constant vectors and fβ (t) are a bunch of
known continuous bounded functions for β = 1, 2, . . . , q for some q ≥ 0. The given
leader remains as
Case III: Single input systems with disturbance acting on both leader and
follower d(t) 6= 0 and dm (t) 6= 0. the follower plant is (3.13). The dynamic of the
given leader is
where
q
X
dm (t) = dm0 + dmβ fmβ (t) (3.17)
β=1
with dm0 ∈ Rm and dmβ ∈ Rm are unknown constant and fmβ (t) are some known
continuous bounded functions for β = 1, 2, . . . , q for some p ≥ 0.
The control objective is to design a bounded control signal u(t) to make the
follower system state x(t) bounded and track the leader system state xm (t) asymp-
totically.
23
Input Systems
In this section, we will solve the single input system leader-following state tracking
problem. The single input system without disturbance will be solved first which in-
cludes the basic principle of the leader-following state tracking problem. Disturbance
rejection is developed in Section 3.2.4.
(A3.1) all the eigenvalues of Am are in the open left-half complex plane;
(A3.3) there exist two constant vectors k1∗ ∈ Rn and k3∗ ∈ Rn and a non-zero constant
scalars k2∗ ∈ R, k4∗ ∈ R such that the following equations are satisfied:
Ae = A + bk1∗T , bm = bk2∗
Assumptions (A3.1) and (A3.2) are for a stable, well-defined reference system with
a bounded output ym (t). Assumption (A3.3) is the so-called matching condition such
that if the parameters of Am , bm , A and b are known and (3.18) is satisfied then the
control law
u∗ (t) = k1∗T (x(t) − xm (t)) + k2∗ um (t) + k3∗T xm (t) (3.19)
whose state vector x(t) belongs to L∞ , and the tracking error e(t) = x(t) − xm (t)
satisfies:
ė(t) = Ae e(t), e(0) = x(0) − xm (0) (3.21)
which indicates that limt→∞ e(t) = 0 exponentially. So that with the nominal con-
troller (3.19) , the desired control objective can be achieved.
In our problem, the parameters of A, b, Am and bm are unknown, so (3.19) can not
be used for control. In this case an adaptive controller which has the same structure
of (3.19) is to be used, whose structure is given as
u(t) = k1T (t)(x(t) − xm (t)) + k2 (t)um (t) + k3T (t)xm (t) (3.22)
25
where k1 (t), k2 (t) and k3 (t) are the estimates of k1∗ , k2∗ and k3∗ respectively. The de-
sign task now is to choose adaptive laws update these estimates so that the control
objective is still achievable even all the four parameters are unknown.
To be specific, the adaptive laws to update the control parameters are proposed
as
where P ∈ Rn×n is a positive definite matrix satisfying ATe P + P Ae = −Q, for any
chosen Q ∈ Rn×n being constant and Q = QT > 0. Γ ∈ Rn×n , Ψ ∈ Rn×n are constant
matrices, and Γ = ΓT > 0, Ψ = ΨT > 0, γ > 0 is a constant scalar. k1 (0), k2 (0) and
k3 (0) are arbitrary.
In summary, according to the previous development, we now present the following
result.
Theorem 3.1. The adaptive controller (3.22), with the adaptive laws (3.23), (3.24)
and (3.25), applied to the plant (3.3) guarantees that all closed-loop signals are bounded
and the tracking error e(t) = x(t) − xm (t) goes to zero as t goes to infinity, i.e.,
limt→∞ e(t) = 0.
26
= (k1 (t) − k1∗ )T (x(t) − xm (t)) + (k2 (t) − k2∗ )um (t)
and then we derive the adaptive control based on the tracking error equation
ė(t) = Ae e(t) + b(k1 (t) − k1∗ )T e(t) + b(k2 (t) − k2∗ )um (t) + b(k3 (t) − k3∗ )T xm (t)
1 T 1 1 T
= Ae e(t) + be k̃ (t)e(t) + ∗ k̃2 (t)um (t) + ∗ k̃3 (t)xm (t) (3.27)
k4∗ 1 k4 k4
where
k̃1 (t) = k1 (t) − k1∗ , k̃2 (t) = k2 (t) − k2∗ , k̃3 (t) = k3 (t) − k3∗ (3.28)
ec (t) = (eT (t), k̃1T (t), k̃2 (t), k̃3T (t))T ∈ R3n+1 . (3.29)
1 T −1 1 1
V (ec ) = eT P e + ∗ 1
k̃ Γ k̃1 + ∗ k̃22 γ −1 + ∗ k̃3T Ψ−1 k̃3 (3.30)
|k4 | |k4 | |k4 |
d
V̇ = V (ec )
dt
T T T
∂V (ec ) ∂V (ec ) ˙ ∂V (ec ) ˙ ∂V (ec )
= ė(t) + k̃1 (t) + k̃2 (t) + k̃˙ 3 (t)
∂e ∂ k̃1 ∂ k̃2 ∂ k̃3
2 2 2
= 2eT (t)P ė(t) + ∗ k̃1T (t)Γ−1 k̃˙ 1 (t) + + ∗ k̃2 (t)γ −1 k̃˙ 2 (t) + ∗ k̃3T (t)Ψ−1 k̃˙ 3 (t)
|k4 | |k4 | |k4 |
(3.32)
2 T 2
V̇ = −eT (t)Qe(t) + eT (t)P be k̃ (t)e(t) + eT (t)P be ∗ k̃2 (t)um (t)
∗ 1
k4 k4
2 2 2
+ eT (t)P be ∗ k̃3T (t)xm (t) + ∗ k̃1T Γ−1 k̃˙ 1 (t) + ∗ k̃2 (t)γ −1 k̃˙ 2 (t)
k4 |k4 | |k4 |
2
+ ∗ k̃3T (t)Ψ−1 k̃˙ 3 (t) (3.33)
|k4 |
where qm > 0 is the minimum eigenvalue of Q. From here on, the desired properties
of the proposed adaptive laws are obvious:
(i) V > 0 and V̇ ≤ 0 implies that the equilibrium state ec = 0 of the closed-
28
loop system consisting of (3.27), (3.23), (3.24) and (3.25) is uniformly stable
and its solution ec (t) is uniformly bounded, which gives the boundedness of
x(t), k1 (t), k2 (t) and k3 (t), and in turn of the boundedness of ė(t) because of
(3.27);
(iii) with e(t) ∈ L2 ∩ L∞ and ė(t) ∈ L∞ , applying Barbalat lemma, we conclude that
limt→∞ e(t) = 0. 5
From the results demonstrated above, all the properties mentioned in Theorem 3.1
are proved.
In Section 3.1.1 Case II has a disturbance acting on the follower, and Case III has
disturbances acting on the leader and the followers respectively. To reject the effect
of unknown disturbances so that the desired system performance can be achieved,
certain matching conditions should be satisfied and additional compensation term
should be introduced to the controller (3.22).
Design for the disturbance acting on the follower system(Case II) When
Assumptions (A3.1)-(A3.4) are satisfied, the ideal control law (3.19) is modified as
u(t) = k1∗ (x(t) − xm (t)) + k2∗ um (t) + k3∗ xm (t) + k5∗ (t), (3.35)
Pq
where k5∗ = −d(t) = −d0 − β=1 dβ fβ (t), which leads to the desired closed-loop
system
ẋ(t) = Ae (x(t) − xm (t)) + Am xm (t) + bm um (t) (3.36)
29
an in turn the desired tracking error system ė(t) = Ae e(t) making limt→∞ e(t) = 0,
as Ae is stable.
When the parameters A, b, Am , bm and the value of d(t) are unknown, the adaptive
version of the controller (3.35) is used:
u(t) = k1T (t)(x(t) − xm (t)) + k2 (t)um (t) + k3T (t)xm (t) + k5 (t), (3.37)
q
X
k5 (t) = k50 (t) + k5β (t)fβ (t) (3.38)
β=1
∗
with k5β (t) being the estimate of k5β = −dβ , β = 0, 1, 2, · · · , q.
To develop adaptive laws for ki (t) for i = 1, 2, 3 and k5β (t), β = 0, 1, 2, . . . , q, we
first derive an system error equation in terms of the tracking error e(t) = x(t) − xm (t)
and the parameter errors
Let k̃5 (t) = k5 (t) − k5∗ . Using (3.37),(3.38) and (3.39), we obtain
ẋ(t) = Ax(t) + b(k1T (t)(x(t) − xm (t)) + k2 (t)um (t) + k3T (t)xm (t) + k5 (t))
1 T
ė(t) = Ae e(t) + be (k̃ (t)(x(t) − xm (t)) + k̃2 (t)um (t) + k̃3T xm (t) + k̃5 (t)). (3.41)
k4∗ 1
where Γ1 = ΓT1 > 0, γ2 > 0, Ψ3 = ΨT3 > 0 and γ5β > 0, β = 0, 1, 2, · · · , q are
adaptation gains, and P = P T > 0 satisfies (3.31).
To analysis the close-loop system stability, choose a positive definite function as
Lyapunov candidate function which is
q
1 1 2 −1 1 1 X 2 −1
V = e P e + ∗ k̃1T Γ−1
T
1 k̃1 + k̃ γ + ∗ k̃3T Ψ−1
∗ 2 2 3 k̃3 + k̃ γ (3.47)
|k4 | |k4 | |k4 | |k4∗ | β=0 5β 5β
Taking time derivative and apply the adaptive laws (3.42) - (3.46) into its derivative.
The time derivative of (3.47)
2 T 2
V̇ = −eT (t)Qe(t) + eT (t)P be k̃ (t)e(t) + eT (t)P be ∗ k̃2 (t)um (t)
∗ 1
k4 k4
q
T 2 T T 2 X 2
+ e (t)P be ∗ k̃3 (t)xm (t) + e (t)P be ∗ k̃5β (t) + eT (t)P be ∗ k̃5β (t)fβ (t)
k4 k4 β=1
k4
2 T −1 ˙ 2 2
+ ∗ 1
k̃ Γ k̃1 (t) + ∗ k̃2 γ −1 k̃˙ 2 (t) + ∗ k̃3T Ψ−1 k̃˙ 3 (t)
|k4 | |k4 | |k4 |
q
2 −1 ˙
X 2
+ ∗ k̃50 γ50 k̃50 (t) + k̃ γ −1 k̃˙ (t)
∗ 50 5β 5β
(3.48)
|k4 | β=1
|k4 |
31
from which we can obtain the desired system properties, i.e., we can also have the
result of Theorem 3.1. 5
Remark 3.1 The disturbance d(t) in (3.3) is matched to the control input u(t), that
is, both act on the plant dynamics through the same vector b. If they are not matched,
for example, for the plant
where b and bd ∈ Rn are linear independent, the above adaptive design may not be
able to ensure asymptotic state tracking.
Design for disturbance acting on both leader and follower systems (Case
III) For the purpose of designing an adaptive control scheme to address the dis-
turbance acting on both leader and follower systems, we assume all the Assumptions
(A3.1)-(A3.4) aforementioned are satisfied. In order to reject the disturbance so that
the desired system performance can be achieved, the ideal controller with disturbance
compensator is chosen as:
u(t) = k1∗T (x(t) − xm (t)) + k2∗ um (t) + k3∗T xm (t) + k5∗ (t) (3.51)
which has the same form with the ideal controller chosen in the last paragraph.
However, when d(t) and dm (t) present in (3.6) and (3.7) respectively at the same
32
time, we re-define
with
∗
k50 = k2∗ dm0 − d0 , k5mβ
∗
= k2∗ dmβ , k5β
∗
= −dβ . (3.53)
The ideal controller (3.51) with the re-defined k5∗ leads (3.5) to a desired system
closed-loop system
In result, we obtain limt→∞ e(t) = 0, since ė(t) = ẋ(t) − ẋm (t) = Ae e(t) and Ae is
stable.
When the parameters A, b, Am , bm and the value of disturbance d(t) are unknown,
we use the adaptive version of controller (3.50) as
u(t) = k1T (t)(x(t) − xm (t)) + k2 (t)um (t) + k3T (t)xm (t) + k5 (t) (3.55)
where
qm q
X X
k5 (t) = k50 (t) + k5mβ (t)fmβ (t) + k5β (t)fβ (t) (3.56)
β=1 β=1
k5mβ (t) for β = 1, 2, . . . , p, derive the tracking error e(t) = x(t) − xm (t) first and the
parameter errors
Let k̃5 (t) = k5 (t) − k5∗ . Substituting (3.50), (3.56) and (3.57) into (3.54), we obtain
1 T
ẋ(t) = Ae (x(t) − xm (t)) + be (k̃ (t)(x(t) − xm (t)) + k̃2 (t)um (t) + k̃3T xm (t) + k̃5 (t))
k4∗ 1
+ Am xm (t) + bm um (t) + bm dm (t) (3.58)
1 T
ė(t) = Ae e(t) + be (k̃ (t)(x(t) − xm (t)) + k̃2 (t)um (t) + k̃3T xm (t) + k̃5 (t)). (3.59)
k4∗ 1
definite function as
q
1 T −1 1 1 1 X 2 −1
V = eT P e + k̃ Γ k̃1 + ∗ k̃22 γ2−1 + ∗ k̃3T Ψ−1
∗ 1 1 3 k̃3 + k̃ γ
|k4 | |k4 | |k4 | |k4∗ | β=0 5β 5β
qm
1 X 2 −1
+ ∗ k̃5mβ δ5mβ (3.66)
|k4 | β=1
has the negative semidefinite time derivative after substituting (3.60) -(3.65)
from which, we can obtain all the desired system properties, i.e, the control objective
is achieved. 5
tiple Inputs
Recall that in Section 3.1.2, the control objective for the multiple inputs systems
is stated as to design a control input u(t) to make the follower system state x(t)
bounded and to track the leader state xm (t) asymptotically. In order to achieve this
control objective, several design conditions are presented as follows.
(A3.5) all the eigenvalues of Am are in the open left-half complex plane;
35
(A3.7) there exist four parameter matrices K1∗ ∈ Rn×p , K2∗ ∈ Rp×m , K3∗ ∈ Rn×p and
K4∗ ∈ Rp×p such that the following equations are satisfied:
(A3.8) there is a known matrix S ∈ Rp×p such that K4∗ S is symmetric and positive
definite: Ms = K4∗ S = (K4∗ S)T = S T K4∗T > 0.
If the parameter of Am , Bm , A and B are known, (3.68) is satisfied, then the ideal
control law
u∗ (t) = K1∗T (x(t) − xm (t)) + K2∗ um (t) + K3∗T xm (t) (3.69)
whose state vector x(t) belongs to L∞ , i.e., x(t) is bounded. The tracking error
e(t) = x(t) − xm (t) satisfies:
When the parameters of A, b, Am and bm are unknown, update (3.69) to the adap-
tive version,
u(t) = K1T (t)(x(t) − xm (t)) + K2 (t)um (t) + K3T (t)xm (t) (3.72)
where K1 (t), K2 (t) and K3 (t) are the estimates of K1∗ , K2∗ and K3∗ respectively.
Choose the adaptive laws (3.73) - (3.75) to update K1 (t), K2 (t) and K3 (t):
Theorem 3.2. The adaptive controller (3.72) with the adaptive laws (3.73), (3.74)
and (3.75), applied to the system (3.11) guarantees that all closed-loop signals are
bounded and the tracking error e(t) = x(t) − xm (t) goes to zero as t goes to ∞.
In order to prove the Theorem 3.2, in another word, to prove that the closed-loop
system (3.11) with state feedback adaptive controller (3.72) is stable, we first define
the parameter error as
K̃1 (t) = K1 (t) − K1∗ , K̃2 (t) = K2 (t) − K2∗ , K̃3 (t) = K3 (t) − K3∗ . (3.76)
37
Substituting (3.12) in ė(t) = x(t) − xm (t), we have the tracking error equation
ė(t) = Ae e(t) + Be K4∗−1 K̃1T (t)e(t) + K4∗−1 K̃2 (t)um (t) + K4∗−1 K̃3T (t)xm (t) . (3.78)
Now K1 (t), K2 (t) and K3 (t) are matrices so the state vector of the closed-loop
system is
n
X n
X n
X
T
V (ec ) = e P e + T
k̃1i Ms−1 k̃1i + T
k̃2j Ms−1 k̃2j + T
k̃3q Ms−1 k̃3q (3.81)
i=1 j=1 q=1
V (ec ) = eT P e + tr[K̃1 Ms−1 K̃1T ] + tr[K̃2T Ms−1 K̃2 ] + tr[K̃3 Ms−1 K̃3T ] (3.82)
T n T m
!T
∂V (ec ) ∂V (ec ) ˙ ∂V (ec )
k̃˙ 2j (t)
X X
V̇ = ė(t) + k̃1i (t) +
∂e i=1
∂ k̃ 1i j=1
∂ k̃2j
n
!T
∂V (ec )
k̃˙ 3q (t)
X
+
q=1 ∂ k̃3q
n n
(t)Ms−1 k̃˙ 1i (t) (t)Ms−1 k̃˙ 2j (t)+
X X
T T T
= 2e (t)P ė(t) + 2 k̃1i +2 k̃2j
i=1 j=1
n
(t)Ms−1 k̃˙ 3q (t)
X
T
2 k̃3q (3.83)
q=1
= 2eT (t)P ė(t) + tr[K̃1 (t)Ms−1 K̃˙ 1T (t)] + tr[K̃2T (t)Ms−1 K̃˙ 2 (t)] + tr[K̃3 (t)Ms−1 K̃˙ 3T (t)]
V̇ = −eT (t)Qe(t) + 2eT (t)P Be K4∗−1 K̃1T (t)e(t) + 2eT (t)P Be K4∗−1 K̃2 (t)um (t)
+ 2eT (t)P Be K4∗−1 K̃3T (t)xm (t) + 2 tr[K̃1 (t)Ms−1 K̃˙ 1T (t)] + 2 tr[K̃2T (t)Ms−1 K̃˙ 2 (t)]
Using the definition Ms = K4∗ S = MsT > 0 and the properties that tr[M1 M2 ] =
tr[M2 M1 ], tr[M3 ] = tr[M3T ] for any matrices M1 , M2 and M3 of appropriate dimen-
sions, we obtain
= tr[eT (t)P Be K4∗−1 K̃1T (t)e(t)] = tr[eT (t)P Be SMs−1 K̃1T (t)e(t)]
= tr[e(t)eT (t)P Be SMs−1 K̃1T (t)] = tr[K̃1 (t)Ms−1 S T BeT P e(t)eT (t)] (3.85)
= tr[eT (t)P Be K4∗−1 K̃2 (t)um (t)] = tr[eT (t)P Be SMs−1 K̃2 (t)um (t)]
= tr[um (t)eT (t)P Be SMs−1 K̃2 (t)] = tr[K̃2T (t)Ms−1 S T BeT P e(t)uTm (t)] (3.86)
= tr[eT (t)P Be K4∗−1 K̃3T (t)xm (t)] = tr[eT (t)P Be SMs−1 K̃3T (t)xm (t)]
= tr[xm (t)eT (t)P Be SMs−1 K̃3T (t)] = tr[K̃3 (t)Ms−1 S T BeT P e(t)xTm (t)] (3.87)
Apply adaptive laws (3.73), (3.74) and (3.75) into (3.84), combing with the facts
(3.86) and (3.87), we have,
To reject the effect of unknown disturbances so that the desired system perfor-
mance can be achieved, certain matching conditions should be satisfied and additional
compensation is introduced to the controller (3.69).
Case II and Case III in Section 3.1.2 are developed in the later two paragraphs
respectively.
Design for disturbance acting on follower systems (Case II) When Assump-
tions (A3.5)-(A3.8) are satisfied, the ideal control law (3.69) is modified as
u(t) = K1∗ (x(t) − xm (t)) + K2∗ um (t) + K3∗ xm (t) + k5∗ (t), (3.89)
ẋ(t) = (A + BK1∗T )(x(t) − xm (t)) + (A + BK3∗T )xm (t) + BK2∗ um (t) + Bk5∗ + Bd(t)
an in turn the desired tracking error system ė(t) = Ae e(t) making limt→∞ e(t) = 0,
as Ae is stable.
When the parameters of A, B, Am , Bm and the value of d(t) are unknown, update
the controller (3.89) to an adaptive version which is
u(t) = K1 (t)t (x(t) − xm (t)) + K2 (t)um (t) + K3T (t)xm (t) + k5 (t), (3.91)
41
q
X
k5 (t) = k50 (t) + k5β (t)fβ (t) (3.92)
β=1
∗
with k5β (t) being the estimate of k5β = −dβ for β = 0, 1, . . . , q.
To develop adaptive laws for Ki (t) for i = 1, 2, 3, and k5β (t) for β = 0, 1, 2, · · · , q,
firstly we derive an error equation in terms of the tracking error e(t) = x(t) − xm (t)
and the parameter errors
Substituting (3.89) into (3.13) with the definition of parameter errors, then we
obtain
ė(t) = Ae e(t) + Be K4∗−1 (K̃1T (t)(x(t) − xm (t)) + K̃2 (t)um (t) + K̃3T xm (t) + k̃5 (t)) (3.94)
Applying adaptive control (3.91) with adaptive laws (3.95) - (3.99), (3.13) should
have all the desired the properties. In order to show that the controller (3.91) with
adaptive laws (3.95) - (3.99) works and can drive e(t) = x(t) − xm (t) bounded and
stable which are the desired properties of trajectories of (3.13). Choose a positive
definite function
V̇ = −eT (t)Qe(t) + 2eT (t)P Be K4∗−1 K̃1T (t)e(t) + 2eT (t)P Be K4∗−1 K̃2 (t)um (t)
+ 2eT (t)P Be K4∗−1 K̃3T (t)xm (t) + 2eT (t)P Be K4∗−1 k̃50
T
(t)
which is negative semidefinite. From (3.102) we can obtain the desired system prop-
erties. 5
structure as:
u∗ (t) = K1∗T (x(t) − xm (t)) + K2∗ um (t) + K3∗T xm (t) + k5∗ (t). (3.103)
Being different with Case II, when d(t) and dm (t) is present in (3.13) and (3.16)
respectively, we re-define
with
∗
k50 = K2∗ dm0 − d0 , k5mβ
∗
= K2∗ dmβ , k5β
∗
= −dβ (3.105)
For the ideal controller, the re-defined k5∗ leads to a desired closed-loop system
In result, we obtain limt→∞ e(t) = 0, since ė(t) = ẋ − ẋm = Ae e(t) and Ae is stable.
When the parameters A, B, Am , Bm and the value of disturbance d(t) are unknown,
we use the adaptive version of controller (3.103) which is
where
qm q
X X
k5 (t) = k50 (t) + k5mβ (t)f5mβ (t) + k5β (t)fβ (t) (3.108)
β=1 β=1
Let k̃5 (t) = k5 (t) − k5∗ (t). Substituting (3.103), (3.104) and (3.109) into (3.13), we
obtain
ė(t) = Ae e(t) + Be K4∗−1 K̃1T (t)(x(t) − xm (t)) + K̃2 (t)um (t) + K̃3T xm (t) + k̃5 (t) .
(3.111)
Based on (3.111), we choose the adaptive laws as
K̃˙ 5β = K̇5β
T
(t) = −S T BeT P e(t)fβT (t), β = 1, 2, . . . , q. (3.113)
S ∈ Rp×p satisfies Assumption (A3.8). K1 (t), K2 (t) and K3 (t) are still in (3.95)
-(3.97).
Use the Lyapunov direct method to analysis the stability of the closed-loop system
(3.13) with adaptive controller (3.103) and adaptive laws (3.95) - (3.97) and (3.112)
-(3.114). Choose a positive definite function
as a measurement of the closed-loop error system. Then get the time derivative of
(3.118) which is the negative semidefinite when adaptive laws (3.95) - (3.97) and
(3.112) -(3.114) are substituted.
The adaptive design of previous section for the p-inputs plant (3.11) needs As-
sumption (A3.8): Ms = K4∗ S = (K4∗ S)T = S T K4∗T > 0 for some known matrix
S ∈ Rp×p . This S matrix is analogous to the sign of k4∗ for the case p = 1; however,
46
the knowledge of such an S matrix is more difficult to obtain then that of the sign
of k4∗ . Relaxation of such knowledge for multivariable adaptive control is important.
In this section, we present an adaptive control scheme for the plant (3.11), using
different knowledge for the gain matrix K4∗ . This adaptive design employs a modified
controller parametrization based on an LDU decomposition of the gain matrix K4∗−1 .
Actually, with LDU decomposition we can address all those six different cases
mentioned in Section 3.1.1 and Section 3.1.2. In this section we will solve Case I with
multi-input systems as an example. The basic principle of adaptive control scheme
and stability analysis for the other two cases are the same.
Kp = LD∗ U (3.117)
for some M × M unit(i.e., with all diagonal elements being 1) lower triangular matrix
L and unit upper triangular matrix U , and
∗ ∆2 ∆M
D = diag {d∗1 , d∗2 , . . . , d∗M } = diag ∆1 , ,..., . (3.118)
∆1 ∆M −1
for some p × p unit (i.e., with all diagonal elements being 1) lower triangular matrix
L and unit upper triangular matrix U , and
∗ ∆2 ∆p
d∗1 , d∗2 , . . . , d∗p
D = diag = diag ∆1 , ,..., (3.120)
∆1 ∆p−1
(A3.9) All leading principle minors ∆i of the matrix K4∗−1 are nonzero, and their
signs, signs[d∗i ], i = 1, 2, . . . , p, are known;
Controller structure Using (3.68) and (3.117), we express the plant (3.9) as
+ Be LD∗ (u(t) − (I − U )u(t) − U K1∗T (x(t) − xm (t)) − U K2∗ um (t) − U K3∗T xm (t))
+ Be LD∗ (u(t) − (I − U )u(t) − U K1∗T (x(t) − xm (t)) − U K2∗ um (t) − U K3∗T xm (t))
(3.123)
where
Φ∗0 = I − U, Φ∗T ∗T ∗ ∗ ∗T ∗T
1 = U K1 , Φ2 = U K2 , Φ3 = U K3 (3.124)
where Φ0 (t), Φ1 (t), Φ2 (t) and Φ3 (t) are the estimates of Φ∗0 , Φ∗1 , Φ∗2 and Φ∗3 , respec-
tively; in particular, the parameter matrix Φ0 has the same special upper triangular
form as that of Φ∗0 = I − U , that is,
0 φ12 φ13 · · · φ1p
0 0 φ23 · · · φ2p
. . .. .. ..
Φ0 = . . (3.126)
. . . . .
0 0
· · · 0 φp−1p
0 ··· ··· 0 0
49
This special form ensures that the control signals u(t) is implementable from (3.125)
without singularity, that is,
up−1 (t) = φp−1p up (t) + [ΦT1 (x(t) − xm (t)) + Φ2 um (t) + ΦT3 xm (t)]p−1 ,
up−2 (t) = φp−2p−1 up−1 (t) + φp−2p u( t) + [ΦT1 (x(t) − xm (t)) + Φ2 um (t) + ΦT3 xm (t)]p−1 ,
..
.
p
X
u2 (t) = φ2i ui (t) + [Φ0 u(t) + ΦT1 (x(t) − xm (t)) + Φ2 um (t) + ΦT3 xm (t)]2 ,
i=3
p
X
u1 (t) = φ1i ui (t) + [ΦT1 (x(t) − xm (t)) + Φ2 um (t) + ΦT3 xm (t)]1 , (3.127)
i=2
Adaptive laws To derive at a compact and exact expression of the control law
(3.125) with the special parameter structure (3.126), we let ΦT1i (t) be the ith row of
ΦT1 (t), ΦT2i (t) be the ith row of ΦT2 (t) and ΦT3i (t) be the ith row of ΦT3 (t), i = 1, 2, . . . , p,
and define
θ1 (t) = [φ12 (t), φ13 (t), . . . , φ1p (t), ΦT11 (t), ΦT21 (t), ΦT31 (t)]T ∈ R2n+2p−1
θ2 (t) = [φ23 (t), φ24 (t), . . . , φ2p (t), ΦT12 (t), ΦT22 (t), ΦT32 (t)]T ∈ R2n+2p−2
..
.
θp−2 (t) = [φp−2p−1 (t), φp−2p (t), ΦT1p−2 (t), ΦT2p−2 (t), ΦT3p−2 (t)]T ∈ R2n+p+2
θp−1 (t) = [φp−1p (t), ΦT1p−1 (t), ΦT2p−1 (t), ΦT3p−1 (t)]T ∈ R2n+p+1
which are the estimates of the corresponding θi∗ from the rows Φ∗0 , Φ∗1 , Φ∗2 and Φ∗3 in
(3.124). It follows from (3.123),(3.125) and (3.128) that
θ̃1T (t)ω1 (t)
T
θ̃2 (t)ω2 (t)
∗
..
ė(t) = Ae e(t) + Be LD . ,
(3.129)
T
θ̃p−1 (t)ωp−1 (t)
T
θ̃p (t)ωp (t)
ω1 (t) = [u2 (t), u3 (t), . . . , up (t), eT (t), rT (t), xTm (t)]T ∈ R2n+2p−1
ω2 (t) = [u3 (t), u4 (t), . . . , up (t), eT (t), rT (t), xTm (t)]T ∈ R2n+2p−2
..
.
ωp−2 (t) = [up−1 (t), up (t), eT (t), rT (t), xTm (t)]T ∈ R2n+p+2
where ēi (t) is the ith component of eT (t)P Be L with P = P T > 0, satisfying (3.31),
Γi = ΓTi > 0, and sign[d∗i ] is from Assumption A3.9.
51
p
X
T
V (e, θ̃i , i = 1, 2, . . . , p) = e P e + |d∗i | θ̃iT Γ−1
i (t)θ̃i (3.132)
i=1
M
X
T
V̇ = 2e (t)P ė(t) + 2 |d∗i | θ̃iT (t)Γ−1
i θ̇i (t)
i=1
XM M
X
T
= 2e (t)P ė(t) + 2 ēi (t)d∗i θ̃iT (t)ωi (t) +2 |d∗i | θ̃iT (t)Γ−1
i θ̇i (t)
i=1 i=1
To verify the adaptive laws, simulation studies are displayed below. To simplify
the problem and make the result easy to follow, choose n = 2 for all systems below.
Leader parameters are given for simulation study.
In this section, simulation results for all cases including single input systems and
multiple inputs systems are shown. Section 3.4.1 displays simulation results of single
input systems. Section 3.4.2 displays simulation results of multiple inputs systems.
52
This section shows simulation results of single inputs systems. Simulation results
of three different cases classifying by the disturbance performance in section 3.1.1 are
shown in order. Through the simulation result,we can verify that the adaptive control
laws works well.
Case I: Singe input system without disturbances For numerical study, pa-
rameter matrices in (3.3) and (3.4) are selected as: (n = 2):
1 1 0 1 1 0
Am = , bm = A = ,b =
−11 −6 8 −1 0 2
Based on the matching condition (3.8), we can easily obtain the ideal value of
k1∗ , k2∗ and k3∗ .
k1∗T = [−4, −2], k2∗ = 4, /, k3∗T = [−5, −3].
This set of ki (i = 1, 2, 3) is the values that would make up the nominal controller
(3.19). However, this nominal controller maybe unknown in real some experiences.
Therefore, the estimates k1 (t), k2 (t) and k3 (t) of k1∗ , k2∗ and k3∗ will be determined from
adaptive laws and used in the adaptive controller (3.22).
For the purpose of the simulation, the ideal gains were calculated. They are
used to show how a fixed gain controller cannot handle the parameter unknown case.
Since k1 (0), k2 (0) and k3 (0) can be chosen arbitrary, here we choose ki (0) = 0.5ki∗ , (i =
53
1, 2, 3). r = sin(t) + cos(t). The initial leader and follower state are xm = [0, 0]T and
x = [0, 0]T , respectively.
The simulation result with a fixed gain controller for the state tracking of x is
shown in figure 3.1. Likewise, the simulation result with adaptive controller for x is
shown in figure 3.2. Tracking errors in both fixed gain control and adaptive control
are shown in figure 3.3. Parameter errors ki (t) − ki∗ (t) are shown in figure 3.4.
Case II: Single input systems with disturbance acting on follower systems
For numerical study, parameter matrices in (3.5) and (3.4) are selected as: (n = 2):
1 1 0 1 1 0
Am = , bm = , A = , b =
−11 −6 8 −1 0 2
Ae , be are chosen as
1 1 0
Ae = , be = ,
−9 −4 4
−1
−2
−3
0 5 10 15 20 25 30
Follower state x1 and leader state xm1 vs. time(sec)
−1
−2
−3
0 5 10 15 20 25 30
Follower state x2 and leader state xm2 vs. time(sec)
Figure 3.1: Follower state (solid) vs. leader state (dotted) with fixed gain control.
−1
−2
0 5 10 15 20 25 30
Follower state x1 and leader state xm1 vs. time(sec)
−1
−2
−3
0 5 10 15 20 25 30
Follower state x2 and leader state xm2 vs. time(sec)
Figure 3.2: Follower state (solid) vs. leader state (dotted) with adaptive gain control.
55
0.6
e1(t)
e2(t)
0.4
0.2
−0.2
−0.4
0 5 10 15 20 25 30
Tracking error e(t) with adaptive control vs. time(sec)
0.8
e (t)
1
0.6 e2(t)
0.4
0.2
−0.2
−0.4
−0.6
0 5 10 15 20 25 30
Tracking error e(t) with fixed gain control vs. time(sec)
Figure 3.3: Tracking errors with adaptive control vs. fixed gain control.
2
K (t)
1,1
K (t)
1 1,2
−1
0 5 10 15 20 25 30
Parameter errors k − k* vs. time(sec)
1,i 1,i
−1.8
K2(t)
−1.9
−2
−2.1
−2.2
0 5 10 15 20 25 30
Parameter error k2 − k*2 vs. time(sec)
3
K3,1(t)
2.5 K3,2(t)
1.5
1
0 5 10 15 20 25 30
Parameter errors k3,i − k*3,i vs. time(sec)
Figure 3.4: Parameter errors ki (t) − ki∗ with adaptive control (i = 1, 2, 3).
56
and k3 (0) can be chosen arbitrary, here we choose ki (0) = 0.8ki∗ , (i = 1, 2, 3). k50 (0) =
5 × 0.8 = 4, k5β (0) = 4 × 0.8 = 3.6. r = sin(t) + cos(t). The initial leader and follower
state are xm = [1, 0]T and x = [−1, 5]T , respectively.
The simulation result with a fixed gain controller for the state tracking of x is
shown in figure 3.5. Likewise, the simulation result with an adaptive controller for
the state tracking of x is shown in figure 3.6. Tracking errors in both fixed gain
control and adaptive control are shown in figure 3.7.
Case III: Single input system with disturbance acting on leader and fol-
lower systems For numerical study, parameter matrices in (3.5) and (3.7) are
selected as: (n = 2):
1 1 0 1 1 0
Am = , bm = A = ,b = ,
−11 −6 8 −1 0 2
−1
−2
−3
0 2 4 6 8 10 12 14 16 18 20
Follower state x1 and leader state xm1 vs. time(sec)
−1
−2
−3
0 2 4 6 8 10 12 14 16 18 20
Follower state x2 and leader state xm2 vs. time(sec)
Figure 3.5: Follower state (solid) vs. leader state (dotted) without disturbance rejec-
tion.
1.5
0.5
−0.5
−1
−1.5
−2
0 2 4 6 8 10 12 14 16 18 20
Follower state x1 and leader state xm1 vs. time(sec)
−1
−2
−3
0 2 4 6 8 10 12 14 16 18 20
Follower state x2 and leader state xm2 vs. time(sec)
Figure 3.6: Follower state (solid) vs. leader state (dotted) with adaptive disturbance
rejection.
58
5
e1(t)
4 e2(t)
3
−1
−2
0 2 4 6 8 10 12 14 16 18 20
Tracking error e with adaptive conrol vs. time(sec)
5
e1(t)
4 e2(t)
3
−1
−2
0 2 4 6 8 10 12 14 16 18 20
Tracking error e with fixed gain conrol vs. time(sec)
Figure 3.7: Tracking errors with adaptive disturbance rejection vs. without distur-
bance rejection.
10
2
0 5 10 15 20 25 30
Follower state x1 and leader state xm1 vs. time(sec)
−5
−10
−15
0 5 10 15 20 25 30
Follower state x and leader state x vs. time(sec)
2 m2
Figure 3.8: Follower state (solid) vs. leader state (dotted) without disturbance rejec-
tion
10
3
0 5 10 15 20 25 30
Follower state x1 and leader state xm1 vs. time(sec)
10
−5
−10
−15
0 5 10 15 20 25 30
Follower state x2 and leader state xm2 vs. time(sec)
Figure 3.9: Follower state (solid) vs. leader state (dotted) with adaptive disturbance
rejection
60
6
e1(t)
e2(t)
4
−2
−4
0 5 10 15 20 25 30
Tracking error e with an adaptive control vs. time(sec)
4
e1(t)
3
e2(t)
2
−1
−2
−3
−4
0 5 10 15 20 25 30
Tracking error e with fixed gain control vs. time(sec)
Figure 3.10: Tracking errors with adaptive disturbance rejection vs. without distur-
bance rejection.
This section shows simulation results of single inputs systems. Simulation results
of three different cases classifying by the disturbance performance in section 3.2.1 are
shown in order. Through the simulation result,we can verify that the adaptive control
laws works well.
1 1 2 0 2 1 1 2 0
Am = , Bm = ,A = ,B = .
−11 −6 0 2 0 −1 0 0 2
61
Based on the matching condition (3.68), we can easily obtain the ideal value of
K1∗ , K2∗ and K3∗ .
0 0 1 0 1 0 0
K1∗T = ∗
, K2 =
∗T
, K3 = .
−4 −2 0 1 0 −5 −3
For the purpose of the simulation, the ideal gains were calculated. They are used
to show how a fixed gain controller cannot handle the parameter unknown case. Since
K1 (0), K2 (0) and K3 (0) can be chosen arbitrary, here we choose Ki (0) = 0.5Ki∗ , (i =
1, 2, 3). r = [sin(t), cos(t), sin(t) + cos(t)]T . The initial leader and follower state are
xm = [0, 0]T and x = [0, 0]T , respectively.
The simulation result with a fixed gain controller for the state tracking of x is
shown in figure 3.11. Likewise, the simulation result with adaptive controller for x is
shown in figure 3.12. Tracking errors in both fixed gain control and adaptive control
are shown in figure 3.13. Parameter errors Ki (t) − Ki∗ (t) are shown in figure 3.14.
Case II: Multiple inputs systems with disturbance acting on follower sys-
tems For numerical study, parameter matrices in (3.13) and (3.15) are selected as:
62
−5
0 5 10 15 20 25 30 35 40
Follower state x1 and leader state xm1 vs. time(sec)
10
−5
−10
0 5 10 15 20 25 30 35 40
Follower state x and leader state x vs. time(sec)
2 m2
Figure 3.11: Follower state (solid) vs. leader state (dotted) with fixed gain control.
−5
0 5 10 15 20 25 30 35 40
Follower state x1 and leader state xm1 vs. time(sec)
10
−5
−10
0 5 10 15 20 25 30 35 40
Follower state x2 and leader state xm2 vs. time(sec)
Figure 3.12: Follower state (solid) vs. leader state (dotted) with adaptive gain control.
63
2
e1(t)
1.5 e2(t)
0.5
−0.5
−1
0 5 10 15 20 25 30 35 40
Tracking error e with adaptive control vs. time(sec)
2
e1(t)
1 e2(t)
−1
−2
−3
0 5 10 15 20 25 30 35 40
Tracking error w with fixed gain control vs. time(sec)
Figure 3.13: Tracking errors with adaptive control vs. fixed gain control.
−5
−10
−15
0 20 40 60 80 100 120 140 160 180 200
Parameter error K − K* vs. time(sec)
1 1
−2
−4
0 20 40 60 80 100 120 140 160 180 200
Parameter error K2 − K*2 vs. time(sec)
−5
−10
−15
0 20 40 60 80 100 120 140 160 180 200
Parameter error K3 − K*3 vs. time(sec)
Figure 3.14: Parameter errors Ki (t) − Ki∗ with adaptive control (i = 1, 2, 3).
64
(n = 2):
0 1 1 −2 −2 3 0 1
Am = , Bm = ,A = ,B = ,
−1 −3 4 1 1 −2 1 2
Ae and Be are
1 1 1 −2
Ae = , Be = .
−9 −4 4 1
Based on the matching condition (3.68), we can easily obtain the ideal value of
K1∗ , K2∗ and K3∗ .
−16 2 2 5 −6 3
K1∗T = ∗
, K2 =
∗T
, K3 = ,
3 −2 1 −2 2 −2
∗
According to the equation k5β = −d(t), k5∗ = [5, 5]T + [−1, −2]T sin(5t).
For the purpose of the simulation, the ideal gains were calculated. They are used
to show how a fixed gain controller cannot handle the parameter unknown case. Since
K1 (0), K2 (0) and K3 (0) can be chosen arbitrary, here we choose Ki (0) = 0.9ki∗ , (i =
∗ ∗
1, 2, 3). k50 = [−4.5, −4.5]T , k5β = [−0.8, 1.8]T . r = [sin(t), cos(t)]T . The initial
leader and follower state are xm = [1, 1]T and x = [2, 3]T , respectively.
The simulation result without adaptive disturbance rejection for the state tracking
of x is shown in figure 3.15. Likewise, the simulation result with adaptive disturbance
65
rejection for x is shown in figure 3.16. Tracking errors corresponding with disturbance
rejection and non-disturbance rejection are shown in figure 3.17.
Remark 3.3 To show the effectiveness of the adaptive compensator, figure 3.15 is the
result making by an adaptive controller without an adaptive disturbance compensator
term, i.e., we adopt the adaptive controller (3.72) rather than (3.91). Figure 3.16 is
created by the adaptive disturbance compensator (3.91).
Case III: Multiple inputs systems with disturbance acting on leader and
follower systems For numerical study, parameter matrices in (3.13) and (3.16) are
selected as: (n = 2):
0 1 1 −2 −2 3 0 1
Am = , Bm = ,A = ,B = ,
−1 −3 4 1 1 −2 1 2
Ae and Be are
1 1 1 −2
Ae = , B
e = .
−9 −4 4 1
Based on the matching condition (3.45), we can easily obtain the ideal value of
66
−2
−4
−6
0 10 20 30 40 50 60 70 80 90 100
Follower state x1 and leader state xm1 vs. time(sec)
−1
−2
−3
0 10 20 30 40 50 60 70 80 90 100
Follower state x and leader state x vs. time(sec)
2 m2
Figure 3.15: Follower state (solid) vs. leader state (dotted) without disturbance
rejection.
−2
−4
−6
0 10 20 30 40 50 60 70 80 90 100
Follower state x1 and leader state xm1 vs. time(sec)
−1
−2
−3
0 10 20 30 40 50 60 70 80 90 100
Follower state x2 and leader state xm2 vs. time(sec)
Figure 3.16: Follower state (solid) vs. leader state (dotted) with adaptive disturbance
rejection.
67
1.5
e1(t)
e2(t)
1
0.5
−0.5
0 10 20 30 40 50 60 70 80 90 100
Tracking error e with adaptive disturbance rejection vs. time(sec)
1
e1(t)
0.8 e2(t)
0.6
0.4
0.2
−0.2
−0.4
0 10 20 30 40 50 60 70 80 90 100
Tracking error e without adaptive disturbance rejection vs. time(sec)
Figure 3.17: Tracking errors with adaptive disturbance rejection vs. without distur-
bance rejection.
−16 2 2 5 −6 3
K1∗T = ∗
, K2 =
∗T
, K3 = .
3 −2 1 −2 2 −2
The simulation result without adaptive disturbance rejection for the state track-
ing of x is shown in figure 3.18 (i.e., created by (3.72)). Likewise, the simulation
result with adaptive disturbance rejection for x is shown in figure 3.19 (i.e., cre-
ated by (3.107)). Tracking errors corresponding with disturbance rejection and non-
disturbance rejection are shown in figure 3.20.
1 1 2 0 1 1 2 0
Am = , Bm = ,A = ,B = ,
−11 6 0 2 −1 0 0 2
Ae and Be are
1 1 1 −2
Ae = , Be = .
−9 −4 4 1
Based on the matching condition (3.68), we can easily obtain the ideal value of
Φ∗0 , Φ∗1 , Φ∗2 and Φ∗3 .
0 −2 ∗T −8 −4 ∗ 1 2 ∗T −10 −6
Φ∗0 = Φ1 = Φ2 = Φ3 =
0 0 −4 −2 0 1 −5 −3
For the purpose of the simulation, the ideal gains were calculated. They are used
to show how a fixed gain controller cannot handle the parameter unknown case. Since
θ1 (0), θ2 (0) and θ3 (0) can be chosen arbitrary, here we choose θi (0) = 0.8θi∗ , (i = 1, 2).
r = [sin(t) cos(t)]T . The initial leader and follower state are xm = [2.5, 1]T and
69
−2
−4
−6
−8
0 5 10 15 20 25 30
Follower state x1 and leader state xm1 vs. time(sec)
12
10
0
0 5 10 15 20 25 30
Follower state x and leader state x vs. time(sec)
2 m2
Figure 3.18: Follower state (solid) vs. leader state (dotted) without disturbance
rejection.
−2
−4
−6
−8
0 5 10 15 20 25 30
Follower state x1 and leader state xm1 vs. time(sec)
12
10
0
0 5 10 15 20 25 30
Follower state x2 and leader state xm2 vs. time(sec)
Figure 3.19: Follower state (solid) vs. leader state (dotted) with adaptive disturbance
rejection.
70
4
e1(t)
2 e2(t)
−2
−4
−6
−8
0 5 10 15 20 25 30
Tracking error e with adaptive disturbance rejection vs. time(sec)
6
e1(t)
4 e2(t)
−2
−4
0 5 10 15 20 25 30
Tracking error e without adaptive disturbance rejection vs. time(sec)
Figure 3.20: Tracking errors with adaptive disturbance rejection vs. without distur-
bance rejection.
3.5 Summary
In this chapter basic adaptive control theory relevant to this research is intro-
duced. Since this is a state tracking problem, in multi-agent system we also called
this kind of problem as ”leader-following consensus” problem. Leader-following con-
sensus problems presented in this chapter are categorized into six cases in total in-
cluding three single input cases and three multiple inputs cases. Stability analyses are
71
−1
−2
0 5 10 15 20 25
Follower state x1 and leader state xm1 vs. time(sec)
−2
−4
−6
0 5 10 15 20 25
Follower state x and leader state x vs. time(sec)
2 m2
Figure 3.21: Follower state (solid) vs. leader state (dotted) with adaptive gain control
(LDU decomposition).
−1
−2
0 5 10 15 20 25
Follower state x1 and leader state xm1 vs. time(sec)
−2
−4
−6
0 5 10 15 20 25
Follower state x2 and leader state xm2 vs. time(sec)
Figure 3.22: Follower state (solid) vs. leader state (dotted) with adaptive gain control
(LDU decomposition).
72
2
e1(t)
e2(t)
1
−1
−2
−3
0 5 10 15 20 25
Tracking error e with adaptive control vs. time(sec)
6
e1(t)
e2(t)
4
−2
−4
0 5 10 15 20 25
Tracking error e with fixed gain control vs. time(sec)
Figure 3.23: Tracking errors with adaptive control vs. fixed gain control (LDU de-
composition).
given for each case to show with the adaptive controllers and the corresponding adap-
tive laws, closed-loop systems in all six cases have obtained the desired properties.
Also, in the last subsection in Section 3.3 which focuses on the multiple input sys-
tems, LDU decomposition is applied. LDU decomposition is used because in practice
some parameters are hard to get, so we need another functional adaptive controller
structure with different adaptive laws. Remind that with LDU decomposition, both
disturbances-free cases and cases with disturbances can be solved. However, we just
show the disturbance-free case to show the basic principle. Many other literatures
show more details which are listed in Section 3.3.5. In Section 3.4, simulation results
are shown in order to verify that the adaptive controllers with corresponding adaptive
laws work well. Simulation results show the difference between adaptive control and
fixed gain control directly.
Chapter 4
Adaptive Leader-Following
Consensus for Multiple Agents
73
74
Consider a set of N unknown follower agents. The dynamic system of the ith
follow agent can be expressed as
where xi (t) ∈ Rn is the state of the ith agent, ui (t) ∈ Rpi is the control input(in
multi-agent system, we also call it control protocol), All parameter matrices Ai ∈
Rn×n , Bi ∈ Rn×pi , i = 1, . . . , N in (4.1) are unknown. The dynamics of the leader is
given by
where x0 ∈ Rn is the state of the leader, u0 (t) ∈ Rm is the bounded reference signal
and A0 ∈ Rn×n , B0 ∈ Rn×m are unknown constant matrices.
The control objective is to design a distributed control protocol using local infor-
mation for each follower to drive all follower consensus on their states with a given
leader, i.e., make all the followers track the give leader on states eventually. Such a
control problem is defined as leader-following consensus problem. The control objec-
tive can be described mathematically as
The above equation means if the control objective is achieved, all states errors between
follower agents and the leader should go to zero as t goes to infinity.
75
The interaction graphs among N + 1 agents including one leader and N followers
are denoted by a directed graph G = (V, E) with a set of vertices V and a set of
directed edges (ordered pairs of vertices) E ⊆ V × V. A vertex represents an agent
and the directed edge (vj , vi ) denotes that agent vi can obtain the information from
agent vj including both the state information and the input information, but not vice
versa. In this case, we say that vj is one of the neighbors of vi (although the link
between them is directed). Define a neighborhood set Ni for the follower agent vi
such that Ni = {vj ∈ V : (vj , vi ) ∈ E} for i = 1, . . . , N . v0 denotes the leader in the
multi-agent system. It also could be a neighbor of the follower agent vi if vi can get
the information from v0 (i.e. (v0 , vi ) ∈ E). A directed path is a sequence of ordered
edges of the form (vi1 , vi2 ), (vi2 , vi3 ), . . . in a directed graph, where vij ∈ V. Define
an indegree (the number of head endpoints adjacent to a vertex) matrix D such that
D = diag{n1 , n2 , . . . , nN } = diag{deg(v1 ), deg(v2 ), . . . , deg(vN )} where deg(·) denotes
the indegree of an agent. This indegree matrix D represents the total number of the
neighbors of the follower agent vi , i.e., the total agent number in the neighborhood
set Ni from i = 1 to N .
To describe the relationship between the leader and followers, we denote B =
diag {µi } if the agent i can get the information from the leader directly then µi = 1
otherwise µi = 0. Denote Q = L + B. matrix Q is a matrix which can fully described
the N + 1 order graph including N followers and one leader. It is obviously that
elements qii on diagonal are the total number of agents in the neighbor sets Ni of ith
agent respectively. In order to make the equation more clear, we denote ni equals
qii to represent the total agent number in neighbor set Ni . More examples show in
section 4.5.
76
Assumption 4.1. For each agent vi , there should exist at least one directed path
(v0 , v1 ), (v1 , v2 ), . . . , (vi−1 , vi ) which starts from the leader and ends at the agent vi .
Assumption 4.2. Directed graph G is simple, which means that it has no loops, and
no multiple arcs (arcs with same starting and ending nodes).
1 0
∗ 1
.
(4.4)
.
.. ..
∗ ∗ 1
N ×N
To make adaptive control design work, several design conditions should be sat-
isfied. Design conditions are presented in Section 4.3.1 and a distributed adaptive
control protocol with adaptive laws is presented in Section 4.3.2.
To meet the control objective and solve the leader-following consensus problem,
we need several design conditions to be satisfied.
(A4.1) all the eigenvalues of A0 are in the open left-half complex plane;
∗ ∗
(A4.3) there exist two matrices K1i ∈ Rn×pi and K4i ∈ Rpi ×pi for each follower agent
vi , which satisfy the following equations
∗T ∗
Ae = Ai + Bi K1i , Be = Bi K4i . (4.5)
where Ae ∈ Rn×n is a stable and known matrix, such that ATe P + P Ae = −Q <
0, Q = QT > 0 is satisfied, and Be ∈ Rn×p is also a known matrix.
∗
(A4.4) there exist two matrices K2i0 ∈ Rpi ×m , K3i0
∗
∈ Rn×pi such that
∗T ∗
A0 = Ai + Bi K3i0 , B0 = Bi K2i0 (4.6)
are satisfied, if the leader is one of the neighbors of the follower agent vi (i.e.,
78
∗T ∗
Aj = Ai + Bi K3ij , Bj = Bi K2ij (4.7)
∗ ∗
should be satisfied for some K3ij ∈ Rn×pi and K2ij ∈ Rpi ×pj for each pair of
(vj , vi ) ∈ E (j 6= 0).
(A4.5) there is a known matrix Si ∈ Rpi ×pi for each follower i such that K4i
∗
Si is
∗ ∗ ∗T
symmetric and positive definite: Ms = K4i Si = (K4i Si )T = SiT K4i > 0.
Assumptions (A4.1) and (A4.2) are for a stable, well-defined reference system
(A0 , B0 ) with a bounded output y0 (t). Assumptions (A4.3) and (A4.4) are called the
matching conditions. Assumption (A4.4) indicates that we classify the N followers
into two groups. In the first group, each follower has a direct access to obtain the
∗
information of the leader, thus for each pair of (v0 , vi ), there exist two matrices K2i0
∗
and K3i0 satisfying (4.6). Followers in the other group have no direct accesses to the
∗ ∗
leader, thus there exist two matrices K2ij and K3ij for each pair of (vj , vi ), j 6= 0. In
summary, for each directed edges (ordered pair) (vj , vi ) in the multi-agent system, no
∗ ∗
matter vj is the leader or a follower, there exist a set of corresponding K2ij and K3ij
(0 ≤ j ≤ N ).
Suppose all the parameters Ai , Bi for i = 0, 1, . . . , N are known and Assumptions
(4.5) - (4.7) satisfied, an ideal distributed adaptive protocol which can achieve the
control objective is designed as
1 X
u∗i (t) = ∗T ∗ ∗T
K1i (xi (t) − xj (t)) + K2ij uj (t) + K3ij xj (t) (4.8)
ni v ∈N
j i
We substitute (4.8) into (4.1), the ith closed-loop subsystem (4.1) becomes
1 X
ẋi (t) = Ai xi (t) + Bi ∗
(K ∗T (xi (t) − xj (t)) + K2ij ∗T
uj (t) + K3ij xj (t)) (4.9)
ni v ∈N 1ij
j i
1 X
= (Ae (xi (t) − xj (t)) + Aj xj (t) + Bj uj (t))
ni v ∈N
j i
To begin with a control design, define a local tracking error for each agent here in
this multi-agent case as follow
1 X
ei (t) = xi (t) − xj (t). (4.10)
ni v ∈N
j i
Lemma 4.1. Under Assumptions (A4.1) and (A4,2), if limt→∞ ei (t) = 0 holds, then
limt→∞ (xi (t) − x0 (t)) = 0 holds for all i = 1, . . . , N .
Proof : First we use a one leader and two followers case to verify the Lemma 4.1. For
a one leader two followers directed graph, there are four formation possible cases in
total.
Case I: In figure 4.1, since we have known that e1 (t) = x1 (t) − x0 (t) → 0 and
e2 (t) = x2 (t) − x1 (t) as t → ∞, then obviously we can get x1 (t) − x0 (t) → 0 and
x2 (t) − x0 (t) → 0 as t → ∞.
0 1 2
Figure 4.1: Case I: Interaction graph of two followers and one leader
aaaaa
aaaaaaaa
80
1 2
Figure 4.2: Case II: Interaction graph of two followers and one leader
Case II: In figure 4.2, we have known that e1 (t) = x1 (t) − x0 (t) → 0 and e2 (t) =
x2 (t)−x0 (t) → 0 as t → ∞, then obviously we can get x1 (t) → x0 (t) and x2 (t) → x0 (t)
as t → ∞.
Case III: In figure 4.3, since we have known that e1 (t) = x1 (t) − 21 x0 (t) − 12 x2 (t) → 0
1 2
Figure 4.3: Case III: Interaction graph of two followers and one leader
and e2 (t) = x2 (t) − x0 (t) as t → ∞, then obviously we can get x1 (t) − x0 (t) → 0 and
x2 (t) − x0 (t) → 0 as t → ∞.
Case IV: Case IV is similar with Case III. We can also verify Lemma 4.1 quickly.
1 2
Figure 4.4: Case IV: Interaction graph of two followers and one leader
After the verification, we will start our proof. Under Assumption 4.1, for each follower,
there exists at least one directed path from the leader to the follower, which indicates
that each directed graph with N followers and one leader in this thesis consists of at
least one basic structure which is like figure 4.5. In figure 4.5, j = 1, . . . , N .
0 1 2 ··· j-1 j
We assume that for all followers, limt→∞ ei (t) = 0, i = 1, . . . , j, j < N in figure 4.5.
Each directed graph may has several basic structures with different orders (i.e., j can
be different). Under the condition that limt→∞ ei (t) = 0 holds for i = 1, . . . , j, j < N ,
if we can prove that all followers in figure 4.5 can achieve the global tracking, i.e.,
limt→∞ (xi (t)−x0 (t)) = 0 no matter how the basic structure changes, we can conclude
that the Lemma 4.1 is true.
We have two different methods to make figure 4.5 more complex. The first method
is to add neighbors on follower v1 in figure 4.5 (i.e., the follower who can get the
information from the leader directly). The second method is to add neighbors on
follower v2 to vj (i.e., followers who can not get information from the leader directly).
Before we making changes on figure 4.5, we assume that all followers in figure 4.5
have already achieved the global tracking which means limt→∞ (xi (t) − x0 (t)) = 0, i =
1, . . . , j.
First method: there is only one possibility to add neighbors on follower v1 un-
der Assumption 4.1 and Assumption 4.2, which is to add a new follower outside into
figure 4.5 (maybe a follower from other basic structures in the same directed graph).
Figure 4.6 shows the interaction graph when we add a new follower vj+1 into figure
4.5. According to the condition of Lemma 4.1, we have ej+1 (t) = xj+1 (t) − x0 (t) → 0
0 1 2 ··· j-1 j
j+1
there are more neighbors adding on follower vi , the situations will be the same. aaaa
aaaa
Second method: Add followers on the follower who does not have the access to get
the information from the leader directedly. aaaa
Case I: Add the leader as a neighbor of follower vk (k 6= 0 and k < N , see figure
4.7). Here we use mathematic induction to prove that the lemma is true under Case
I by the second method.
First, when adding the leader as a neighbor of follower vk , since all limt→∞ ei (t) =
0, we have limt→∞ ek (t) = limt→∞ (xk (t) − 12 x0 (t) − 12 xk−1 (t)) = 0. Since we as-
sume that all the followers have already achieved the global tracking before we start
adding neighbors, we have limt→∞ (xk−1 (t) − x0 (t)) = 0. Thus, obviously we have
limt→∞ (xk (t) − x0 (t)) = 0.
Then we assume when we adding n followers on the follower vk including the
leader, we can get
1 X
lim ei (t) = lim (xk (t) − xk−1 (t)) = 0
t→∞ t→∞ n v ∈N
k−1 kn
according to limt→∞ (xi (t) − x0 (t)) = 0. Nkm denotes the neighbors of vk when it has
m neighbors (m¡N).
83
Under the situation when adding n followers on follower vk , we add the (n + 1)th
follower va on the follower vk including the leader, we have
1 X
lim ek (t) = lim xk (t) − ( xk−1 (t) + xa (t)) = 0
t→∞ t→∞ n + 1 v ∈N
k−1 m+1
1 X
lim ek (t) = lim (xk (t) − ( xk−1 (t) + xa (t))) = 0
t→∞ t→∞ nk+1 vk−1 ∈Nk
we have,
lim (xk (t) − x0 (t)) = 0
t→∞
Then,
lim (xi (t) − x0 (t)) = 0, i = 1, . . . , j
t→∞
Case II: Add other followers as neighbors of follower vk (k 6= 0 and k < N , see
figure 4.8).
Figure 4.8: Interaction graph after adding one follower for follower vk
We still use mathematic induction to prove that the lemma is true under Case II
by the second method.
84
In summary, we have proved that the global tracking can be achieved if we known
that all limt→∞ ei (t) = 0 whatever changes we make on a basic structure in directed
graph. Also the directed graphs under Assumption (A4.1) and (A4.2) consist of
several different basic structures. Thus we can prove that Lemma 4.1 is true. 5
From Lemma 4.1, we conclude that as long as all the local tracking errors ei (t)
for i = 1, . . . , N go to zero as t → ∞, the global tracking will be achieved.
Substituting (4.9) into the derivative of the local tracking error (4.10), we obtain,
1 X
ėi (t) = ẋi (t) − ẋj (t)
ni v ∈N
j i
1 X 1 X
= Ae (xi (t) − xj (t)) + Aj xj (t) + Bj uj (t) − Aj xj (t) + Bj uj (t)
ni v ∈N ni v ∈N
j i j i
1 X
= Ae (xi (t) − xj (t))
ni v ∈N
j i
1 X
= Ae xi (t) − xj (t) = Ae ei (t) (4.11)
ni v ∈N
j i
In our problem, the leader parameter matrices A0 , B0 and the followers parameter
matrices Ai and Bi (i = 1, . . . , N ) are unknown, so the ideal control protocol (4.8)
will no longer work. An adaptive version protocol of (4.8) is proposed as follows,
1 X T T
ui (t) = K (t)(xi (t) − xj (t)) + K2ij (t)uj (t) + K3ij (t)xj (t) (4.12)
ni v ∈N 1i
j i
T 1 T T
K̇1ij (t) = − Si Be P ei (t)(xi (t) − xj (t))T (4.13)
ni
1
K̇2ij (t) = − SiT BeT P ei (t)uTj (t) (4.14)
ni
T 1
K̇3ij (t) = − SiT BeT P ei (t)xTj (t), (vj ∈ Ni , i = 1, 2, . . . , N ) (4.15)
ni
∗ ∗ ∗
where K1ij (t), K2ij (t) and K3ij (t) are the estimates of K1i , K2ij and K3ij respectively,
and Si satisfies Assumption (A7) for i = 1, . . . , N ) and P = P T > 0 satisfying
ATe P + P Ae = −Q < 0, Q = QT > 0. K1ij (0), K2ij (0) and K3ij (0) can be chosen
arbitrarily.
∗
Remark 4.2 For each follower agent vi , the ideal controller parameter K1i is
∗
unique. However, according to (4.12), for each different vj ∈ Ni , the estimate K1i s
are different. In order to distinguish these different estimate values, we denote the
∗
estimate K1i as K1ij (t).
We now present the main result in solving the leader-following consensus tracking
problem for multi-agent systems as follows.
Theorem 4.1. The distributed adaptive controller (4.12), with the adaptive laws
86
T T ∗T ∗ T T ∗T
K̃1ij (t) = K1ij (t) − K1i , K̃2ij (t) = K2ij (t) − K2ij , K̃3ij (t) = K3ij (t) − K3ij , (4.16)
Substituting the distributed adaptive protocol (4.12) into the multi-agent subsystem
(4.1) leads to
1 X
ėi (t) = ẋi (t) − ẋj (t)
ni v ∈N
j i
1 X
= Ae (xi (t) − xj (t))
ni v ∈N
j i
1 X ∗−1
T T
+ Be K4i K̃1i (t)(xi (t) − xj (t)) + K̃2ij (t)uj (t) + K̃3ij (t)xj (t)
ni v ∈N
j i
= Ae ei (t) (4.18)
1 X ∗−1
T T
+ Be K4i K̃1i (t)(xi (t) − xj (t)) + K̃2ij (t)uj (t) + K̃3ij (t)xj (t)
ni v ∈N
j i
87
and
X X X
V2i = tr[K̃1i Ms−1 K̃1i
T
]+ T
tr[K̃2ij Ms−1 K̃2ij ] + tr[K̃3ij Ms−1 K̃3ij
T
] (4.21)
vj ∈Ni vj ∈Ni vj ∈Ni
2 ∗−1
X
+ eTi (t)P Be K4i K̃2ij (t)uj (t)
ni v ∈N j i
2 T ∗−1
X
T
+ e (t)P Be K4i K̃3ij (t)xj (t) (4.22)
ni i v j ∈Ni
tr[K̃1i Ms−1 K̃˙ 1i Ms−1 K̃˙ 2ij ]+2 tr[K̃3ij Ms−1 K̃˙ 3ij
X X X
T T T
V̇2i = 2 ]+2 tr[K̃2ij ] (4.23)
vj ∈Ni vj ∈Ni vj ∈Ni
88
1 X
=− tr[K̃1i Ms−1 SiT BeT P ei (t)(xi (t) − xj (t))]
ni v ∈N
j i
1 X
=− tr[K̃1i K4∗−1 BeT P ei (t)(xi (t) − xj (t))]
ni v ∈N
j i
1 X
=− tr[eTi (t)P Be K4∗−1 K̃1i
T
(xi (t) − xj (t))]
ni v ∈N
j i
1 T X
=− ei (t)P Be K4∗−1 T
K̃1i (xi (t) − xj (t)) (4.24)
ni v ∈N j i
Similarly, it is obtained
1
Ms−1 K̃˙ 2ij ] = − eTi (t)P Be K4i
X X
T ∗−1
tr[K̃2ij K̃2ij (t)uj (t) (4.25)
vj ∈Ni
ni v ∈N j i
1
tr[K̃3ij Ms−1 K̃˙ 3ij
X X
T ∗−1
] = − eTi (t)P Be K4i T
K̃3ij (t)xj (t) (4.26)
vj ∈Ni
ni v ∈N j i
2 T ∗−1
X
T
V̇2i = − ei (t)P Be K4i K̃1i (t)(xi (t) − xj (t))
ni v ∈N j i
2 T ∗−1
X
− e (t)P Be K4i K̃2ij (t)uj (t)
ni i v j ∈Ni
2 T ∗−1
X
T
− ei (t)P Be K4i K̃3ij (t)xj (t) (4.27)
ni v ∈N j i
V̇i = V̇1i + V̇2i = eTi (t)(P Ae + ATe P )ei (t) = −eTi (t)Qei (t) ≤ 0 (4.28)
89
N
X N
X N
X
V̇ = V̇i = eTi (t)Qei (t) ≤ −qm kei (t)k22 ≤ 0 (4.29)
i=1 i=1 i=1
• V > 0 and V̇ ≤ 0 implies that the equilibrium state eic = 0 of the closed-loop
system consisting of (4.13) and (4.15) for i = 1, . . . , N is uniformly stable and
its solution eic (t) is uniformly bounded, which gives the boundedness of all xi (t),
K1ij (t), K2ij and K3ij in the multi-agent system, and in turn the boundedness
of ėi (t) for i = 1, . . . , N because of (4.18);
case and chose an appropriate expression for that new term according to which part
does the disturbance act on.
However, in this section, there are N + 1 followers, for each follower i it could be
followed by other agents as well. In order to develop the problem more conveniently,
we re-state the followers structure.
Now consider the dynamic system of all the followers becomes as
where
qi
X
di (t) = di0 + diβ fiβ (t) (4.31)
β=1
with di0 , diβ ∈ Rpi being some unknown constants and fiβ(t) being some known
bounded continuous functions, β = 1, 2, . . . , qi and some qi > 0. The given leader is
(4.2).
To cancel the disturbance, the nominal control protocol (4.8) is modified as,
1 X
u∗i (t) = ∗T ∗ ∗T ∗
K1i (t)(xi (t) − xj (t)) + K2ij (t)uj (t) + K3ij (t)xj (t) + k5ij (t) , (4.32)
ni v ∈N
j i
∗
where the compensation signal k5ij (t) is
∗ ∗
k5ij = K2ij dj (t) − di (t)
qj qi
X X
∗ ∗
= (K2ij dj0 − di0 ) + K2ij djβ fjβ (t) − diβ fiβ (t)
β=1 β=1
qj qi
X X
∗ ∗ ∗
= k50ij + k5jβ fjβ (t) + k5iβ fiβ (t) (4.33)
β=1 β=1
91
∗ ∗
k50ij = K2ij dj0 − di0
∗ ∗
k5jβ = K2ij djβ
∗
k5iβ = −diβ (4.34)
The ideal disturbance compensator (4.32) leads the subsystem (4.1) to a closed-
loop result
1 X ∗T ∗T
ẋi (t) = ((Ai + Bi K1i )(xi (t) − xj (t)) + (Ai + Bi K3ij )xm (t)
ni v ∈N
j i
qj qi
X X
∗ ∗
+ BK2ij uj (t) + Bi K2ij dj0 + Bi K2∗ djβ fjβ (t) − Bi diβ fiβ (t)
β=1 β=1
+ Bi di (t) − Bi di0 )
1 X
= (Ae (xi (t) − xj (t)) + Aj xj (t) + Bj uj (t) + Bj dj (t)) (4.35)
ni v ∈N
j i
1 T T
ui (t) = K1i (t)(xi (t) − xj (t)) + K2ij (t)uj (t) + K3ij (t)xj (t) + k5ij (t) (4.36)
ni
where
qj qi
X X
k5ij (t) = k50ij (t) + k5jβ (t)fjβ (t) + k5iβ (t)fiβ (t) (4.37)
β=1 β=1
∗ ∗
with k50ij (t) being the estimate of k50ij and k5jβ (t) being the estimate of k5jβ . k5iβ (t)
∗ ∗
is the estimate of the k5iβ . k5ij (t) is the estimate of k5ij (t).
92
The adaptive law for K1i (t), K2ij (t) and K3ij (t) are still given in (4.13), (4.14)
and (4.15). The estimate k5ij is developed as
1 T T
k̇50ij (t) = − S B P ei (t), (4.38)
ni i e
1
k̇5jβ (t) = − SiT BeT P ei (t)fjβ
T
(t), β = 1, . . . , qj , (4.39)
ni
1
k̇5iβ (t) = − SiT BeT P ei (t)fiβ
T
(t), β = 1, . . . , qi . (4.40)
ni
Theorem 4.2. The distributed adaptive controller (4.32), with the adaptive laws
(4.13)-(4.15) and (4.38) - (4.40), applied to multi-agent systems (4.1) guarantees that
all closed-loop signals are bounded and global tracking are achieved: limt→∞ (xi (t) −
x0 (t)) = 0.
Proof: the stability of the closed-loop system, first denote the parameter errors for
k5ij (t),
∗
k̃50ij (t) = k50ij − k50ij (t)
∗
k̃5jβ (t) = k5jβ − k5jβ (t)
∗
k̃5iβ (t) = k5iβ − k5iβ (t) (4.41)
1 X ∗−1 T
ėi (t) = Ae ei (t) + Be K4i K̃1i (t)(xi (t) − xj (t)) + K̃2ij (t)uj (t)
ni v ∈N
j i
qj qi
X X
T
+ K̃3ij (t)xj (t) + k̃50ij (t) + k̃5jβ (t)fjβ (t) + k̃5iβ (t)fiβ (t) (4.42)
β=1 β=1
multi-agent system
N
X
V̄ = V̄i (4.43)
i=1
with
V̄i = Vi + Vai (4.44)
where
X X X
Vai = T
tr[k̃50ij Ms−1 k̃50ij ] + T
tr[k̃5jβ Ms−1 k̃5jβ ] + T
tr[k̃5iβ Ms−1 k̃5iβ ] (4.45)
vj ∈Ni vj ∈Ni vj ∈Ni
and Vi is in (4.19).
Being similar with (4.24), we obtain
1 T
Ms−1 k̃˙ 50ij ] = −
X X
T ∗−1
tr[k̃50ij ei (t)P Be K4i k̃50ij (t) (4.46)
vj ∈Ni
ni v ∈N j i
1
tr[k̃5jβ Ms−1 k̃˙ 5jβ
X X
T ∗−1
] = − eTi (t)P Be K4i T
K̃5jβ (t)fjβ (t) (4.47)
vj ∈Ni
ni v j ∈Ni
1 T
tr[k̃5iβ Ms−1 k̃˙ 5iβ
X X
T ∗−1 T
]=− ei (t)P Be K4i k̃5iβ (t)fiβ (t) (4.48)
vj ∈Ni
ni v ∈N j i
N
X
V̇ = − eTi (t)Qei (t), Q = QT > 0 (4.49)
i=1
(the details are similar to the proof of Theorem 1 and are omitted here). From this
expression of V̇ , we can similarly obtain the results of Theorem 2. 5
94
−2 3 0 1 0 2 −1 2
A1 = , B
1 = , A
2 = , B
2 = ,
1 −2 1 2 −2 1 0 1
−1 0 1 0 0 1 1 −2
A3 = , B3 = , A0 = , B0 = ,
−1 −2 3 1 −1 −3 4 1
1 1 1 −2
Ae = , Be = , (4.50)
−9 −4 4 1
1 2
3
Figure 4.9: Interaction graph of three follower agents and one leader
1 0 0
The corresponding matrix Q of Fig 4.1 is 0 1 0.
−1 −1 2
Case II: One leader and five followers without disturbance Consider a multi-
agent systems consisting of five followers and one leader. This case is verified that a
follower can achieve the consensus performance no matter the path length from its
neighbors to the leader are the same or not.
The interaction relationships between the followers and the leader is shown in
Figure 4.4. Leader-following consensus performance is shown in Figure 4.5. Tracking
errors are shown in Figure 4.6. In this case, agents 0-3 have the same dynamic models
with agents 0-3 in Case I. Also Ae and Be are the same. And
1 3 −1 2 −1 −4 1 0
A4 = , B4 = , A5 = , B5 = , (4.51)
1 1 0 1 3 −2 3 1
.
96
6
x
1,1
x2,1
4
x3,1
x4,1
2
−2
−4
0 5 10 15 20 25 30 35 40 45 50
10
x1,2
8 x2,2
6 x
3,2
x
4,2
4
−2
−4
0 5 10 15 20 25 30 35 40 45 50
Figure 4.10: State trajectories of three follower states xi and leader state x0 vs.
time(sec)
1
x −x
1,1 0,1
x2,1 −x0,1
0
x3,1 − x0,1
−1
−2
−3
−4
0 5 10 15 20 25 30 35 40 45 50
8
x −x
1,2 0,2
6 x2,2 − x0,2
x3,2 − x0,2
4
−2
−4
0 5 10 15 20 25 30 35 40 45 50
Figure 4.11: Tracking errors between three follower states xi and leader state x0 vs.
time(sec)
97
2 3 4
0 1 5
Figure 4.12: Interaction graph of five follower agents and one leader
1 0 0 0 0
0 1 0 0 0
The corresponding Q of Fig 4.4 =
−1 −1 2 0 0
−1 0 −1 2 0
−1 0 0 0 1
Case III: One leader and three followers with disturbance Multi-agent sys-
tem in this case is a system which has disturbances acting on multi-agent system in
Case I in Section 4.6. Parameters matrices of each agents including all the followers
and the leader are shown in (4.50). Disturbances acting on agent 1-agent 3 are
respectively. u0 = [sin(t), cos(t)]T . K1ij (0), K2ij (0), K3ij (0) and k5ij (0) are 0.85 times
their ideal value.
Use adaptive disturbance compensator (4.4) we can get the desired state trajecto-
ries. Leader-following consensus performance is shown in Figure 4.7. Tracking errors
are shown in Figure 4.8.
98
6 x
0,1
4 x1,1
x2,1
2
x3,1
0 x4,1
x5,1
−2
−4
−6
0 10 20 30 40 50 60 70 80
5
x0,2
x1,2
x2,2
x3,2
0 x
4,2
x5,2
−5
0 10 20 30 40 50 60 70 80
Figure 4.13: State trajectories of five follower states xi and leader state x0 vs.
time(sec)
4
x1,1 − x0,1
x2,1 − x0,1
2 x −x
3,1 0,1
x4,1 − x0,1
0 x5,1 − x0,1
−2
−4
0 10 20 30 40 50 60 70 80
3
x1,2 − x0,2
2 x2,2 −x0,2
x3,2 − x0,2
1
x4,2 −x0,2
0 x5,2 − x0,2
−1
−2
−3
0 10 20 30 40 50 60 70 80
Figure 4.14: Tracking errors between five follower states xi and leader state x0 vs.
time(sec)
99
8
x0,1
6
x1,1
4 x2,1
2 x3,1
−2
−4
−6
0 5 10 15 20 25 30 35 40 45 50
8
x0,2
6 x1,2
x2,2
4
x3,2
2
−2
−4
0 5 10 15 20 25 30 35 40 45 50
Figure 4.15: State trajectories of three follower states xi and leader state x0 with
disturbance rejection vs. time(sec)
.
1
−1
x1,1 − x0,1
−2 x2,1 − x0,1
x3,1 − x0,1
−3
0 5 10 15 20 25 30 35 40 45 50
8
x1,2−x0,2
6 x2,2 − x0,2
x3,2 − x0,2
4
−2
−4
0 5 10 15 20 25 30 35 40 45 50
Figure 4.16: Tracking error between follower states xi and leader state x0 vs. time(sec)
Chapter 5
5.1 Conclusions
In this thesis, in order to solve the distributed leader-following problems for the
multi-agent systems, we first solved the one leader-one follower problem with and
without the disturbances in Chapter 3. Stability analyses have shown that the pro-
posed controllers in Chapter 3 can achieve the desired properties and the simulation
studies also verify the capability of the proposed adaptive control scheme, i.e., all
agents can track the prescribed leader eventually. Based on the basic idea used in
Chapter 3, we have developed two solutions to the distributed adaptive control and
disturbance compensation problems for multi-input multi-agent systems in Chapter 4.
Our study has shown that the desired closed-loop system stability and tracking prop-
erties can be achieved by control adaptation based on a complete parameterization of
the leader and the followers system parameters as well as the disturbance parameters.
Simulation results for multi-agent systems have also confirmed the capability of the
proposed adaptive control.
100
101
Adaptive control for distributed multi-agent coordination with leader and followers
parametric uncertainties, as discussed and analyzed in this thesis, is a new research
focus. However, there are still many challenges need to be overcame. In this section,
we will introduce several potential extensions to our research.
The control framework discussed in Chapter 3 assumes that all of the states of
each agent are available for feedback control design and exchange in a communication
network. However, in real application, not all of the states can be accessible for
measurement and exchange through a communication network. It is both practically
and theoretically important to relax the requirement for full state measurement or
exchange for multi-agent systems. Therefore,a state observer is needed to achieve
the state tracking. Also this useful extension results can be applied to the multi-
agent case in Chapter 4. It is worthwhile investigating the estimation of relative state
estimation with appropriate estimation accuracy as well as the stability of closed-loop
multi-agent systems with estimated information.
At the meanwhile, for some practical multi-agent systems, links between different
agents have different weights. It is worth and important to figure out how to make
the multi-agent consensus well based on those different weights. Compared to the
weighted case, in our thesis, we assume that every edge in the communicate graph
have the same weight. This challenge need to be solved for many real practical
applications.
Bibliography
[1] Bernstein, D. S., “A student’s guide to classical control,” IEEE Control Systems
Magazine, vol. 17, no. 4, pp. 96-100, 1997.
[2] Chen, Chi-Tsong, Linear System Theory and Design, Oxford University Press
Inc., 1995.
[3] Davison, E., “The robust control of a servomechanism problem for linear time-
invariant multivariable systems,” IEEE Transactions on Automatic Control, vol.
21, no. 1, pp. 25-34, 1976.
[4] Fax, J. A. and R. M. Murray, “Information flow and cooperative control of vehicle
formations,” IEEE Transactions on Automatic Control, vol. 49, no. 9, pp. 1465-
1476, 2004.
[6] Godsil, C. and G. Royle, Algebraic Graph Theory (Vol. 207), Springer, New York,
2001.
[7] Gross, J. L. and J. Yellen, Handbook of Graph Theory, CRC Press, Boca Raton,
FL, 2004.
102
103
[9] Hong, Y., J. Hu and L. Gao, “Tracking control for multi-agent consensus with
an active leader and variable topology,” Automatica, vol. 42, no. 7, pp. 1177-1182,
2006.
[10] Hong, Y., G. Chen and L. Bushnell, “Distributed observers design for leader-
following control of multi-agent networks,” Automatica, vol. 44, no. 3, pp. 846-850,
2008.
[11] Hong, Y., X. Wang and Z. Jiang, “Distributed output regulation of leaderfollower
multi-agent systems,” International Journal of Robust Nonlinear Control, vol. 23,
no. 1, pp. 48-66, 2013.
[13] Jadbabaie, A., J. Lin and A. S. Morse, “Coordination of groups of mobile au-
tonomous agents using nearest neighbor rules,” IEEE Transactions on Automatic
Control, vol. 46, no. 6, pp. 988-1001, 2003.
[15] Khalil, H. K., Nonlinear Systems, Prentice Hall Inc., 3rd Ed., Upper Saddle
River, NJ, 2002.
104
[16] Lavretsky, E. and K. A. Wise, Robust and Adaptive Control with Aerospace Ap-
plication, Springer, London, United Kingdom, 2013.
[17] Li, T. and J. Zhang, “Mean square average consensus under measurement noises
and fixed topologies: necessary and sufficient conditions,” Automatica, vol. 45, no.
8, pp. 1929-1936, 2009.
[18] Li, Z., Z. Duan and L. Huang, “H∞ control of networked multi-agent systems,”
Journal of Systems Science and Complexity, vol. 22, no. 1, pp. 35-48, 2009.
[19] Lin, P., Y. Jia, “Robust H∞ consensus analysis of a class of second-order multi-
agent systems with uncertainty,” IET Control Theory and Applications, vol. 4, no.
3, pp. 487-498, 2010.
[20] Li, Z., Z. Duan and G. Chen, “Dynamic consensus of linear multi-agent systems”,
IET Control Theory and Applications, vol. 5, no. 1, pp. 19-28, 2011.
[21] Li, Z., X. Liua, W. Ren and L. Xie, “Distributed tracking control for linear
multi-agent systems with a leader of bounded unknown input,” IEEE Transactions
on Automatic Control, vol. 58, no. 2, pp. 518523, 2013.
[22] Liu, B., T. Chu, L. Wang and G. Xie, “Controllability of a leader-follower dy-
namic network with switching topology,” IEEE Transactions on Automatic Control,
vol. 53, no. 4, pp. 1009-1013, 2008.
[23] Liu, Y. and Y. Jia, “H∞ consensus control of multi-agent systems with switching
topology: a dynamic output feedback protocol,” International Journal of Control,
vol. 83, no. 3, pp. 527537, 2010.
105
[25] Luo, J., Adaptive Control for Distributed Multi-Agent Coordination, Ph.D. The-
sis, University of Connecticut, 2013.
[26] Meng, Z., W. Ren, Y. Cao and Z. You, “Leaderless and leader-Following con-
sensus with communication and input delays under a directed network topology,”
IEEE Transactions on Systems, Man and Cybernetics-Part B: Cybernetics, vol. 41,
no, 1, pp. 75-88, 2011.
[29] Nise, N. S., Control Systems Engineering, John Wiley and Sons Inc., 6th Ed.,
Hoboken, NJ, 2011.
[30] Peng, Z., D. Wang, H. Zhang, G. Sun and H. Wang, “Distributed model refer-
ence adaptive control for cooperative tracking of uncertain dynamical multi-agent
systems, IET Control Theory and Applications, vol. 7, no. 8, pp. 1079-1087, 2013.
[31] Ren, W. and R. W. Beard, “Consensus seeking in multi-agent systems under dy-
namically changing interaction topologies, IEEE Transactions on Automatic Con-
trol, vol. 50, no. 5, pp. 655-661, 2005.
106
[33] Ren W., “On consensus algorithms for double-integrator dynamics,” IEEE
Transactions on Automatic Control, vol. 53, no. 6, pp. 1503-1509, 2008.
[36] Olfati-Saber, R., “Flocking for multi-agent dynamic systems: Algorithms and
theory.” IEEE Transactions on Automatic Control, vol. 51, no. 3, pp. 401-420,
2006.
[38] Slotine, J. J. E. and W. Li, Applied Nonlinear Control, Prentice Hall Inc., Upper
Saddle River, NJ, 1991.
[39] Su, H., X. Wang and Z. Lin. “Flocking of multi-agents with a virtual leader,”
IEEE Transactions on Automatic Control, vol. 54, no. 2, pp. 293-307, 2009.
107
[41] Tanner, H. G., A. Jadbabaie and G. J. Pappas, “Flocking in fixed and switching
networks,” IEEE Transactions on Automatic Control, vol. 52, no. 5, pp. 863-868,
2007.
[42] Tao, G., Adaptive Control Design and Analysis, John Wiley and Sons Inc., New
York, NY, 2003.
[44] Xiao, L., S. Boyd and S. Kim, “Distributed average consensus with least-mean-
square deviation,” Journal of Parallel and Distributed Computing, vol. 67, no. 1,
pp. 33-46, 2007.
[46] Wang, J., D. Cheng and X. Hu, “Consensus of multi-agent linear dynamic sys-
tems,” Asian Journal of Control, vol. 10, no. 2, pp.144-155, 2008.
[47] Weiss G., Multiagent systems: Intelligent robotics and autonomous agents, MIT
Press, Cambridge, MA, 2013.