Detection Filter Invariant Zero
Detection Filter Invariant Zero
Original citation:
Sumislawska, M. (2012) Fault detection and diagnosis and unknown input reconstruction
based on parity equations concept. Unpublished PhD thesis. Coventry: Coventry University
Copyright © and Moral Rights are retained by the author(s) and/ or other copyright
owners. A copy can be downloaded for personal non-commercial research or study,
without prior permission or charge. This item cannot be reproduced or quoted extensively
from without first obtaining permission in writing from the copyright holder(s). The
content must not be changed in any way or sold commercially in any format or medium
without the formal permission of the copyright holders.
http://curve.coventry.ac.uk/open
Fault detection and diagnosis
and unknown input
reconstruction based on parity
equations concept
Malgorzata Sumislawska
MSc Teleinformatics, MSc Control Engineering
October 2012
There are two main threads of this thesis, namely, an unknown (unmeasurable) input
reconstruction and fault detection and diagnosis. The developed methods are in the
form of parity equations, i.e. finite impulse response filters of the available input and
output measurements.
In the first thread the design of parity equations for the purpose of an unknown
input reconstruction of linear, time-invariant, discrete-time, stochastic systems is taken
into consideration. An underlying assumption is that both measurable system inputs
as well as the outputs can be subjected to noise, which leads to an errors-in-variables
framework. The main contribution of the scheme is accommodation of the Lagrange
multiplier method in order to minimise the influence of the noise on the unknown
input estimate. Two potential applications of the novel input reconstruction method
are proposed, which are a control enhancement of a hot strip steel rolling mill and an
estimation of a pollutant level in a river.
Furthermore, initial research is conducted in the field of the unknown input recon-
struction for a class of nonlinear systems, namely, Hammerstein-Wiener systems, where
a linear dynamic block is preceded and followed by a static nonlinear function. Many
man-made as well as naturally occurring systems can be accurately described using
Hammerstein-Wiener models. However, it is considered that not much attention has
been paid to Hammerstein-Wiener systems in the errors-in-variables framework and in
this thesis it is aimed to narrow this gap.
The second thread considers a problem of robust (disturbance decoupled) fault de-
tection as well as fault isolation and identification. Unmeasurable external stimuli,
parameter variations or discrepancies between the system and the model act as distur-
bances, which can obstruct the fault detection process and lead to false alarms. Thus,
a fault detection filter needs to be decoupled from the disturbances. In this thesis
the right eigenstructure assignment method used for the robust fault detection filter
design is extended to systems with unstable invariant zeros. Another contribution re-
gards the design of robust parity equations of any arbitrary order using both left and
right eigenstructure assignment. Furthermore, a parity equation-based fault isolation
and identification filter is designed which provides an estimate of the fault. A simple
method for the calculation of thresholds whose violation indicates a fault occurrence is
also proposed for the errors-in-variables framework.
i
Acknowledgements
ii
Contents
Page
Abstract i
Acknowledgements ii
Contents iii
Nomenclature vii
Abbreviations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
Notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
2 Review 9
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.2 Linear system representation . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.2.1 Polynomial representation . . . . . . . . . . . . . . . . . . . . . . . . 10
2.2.2 State-space representation . . . . . . . . . . . . . . . . . . . . . . . . 10
2.2.3 Polynomial and state-space representations of stochastic systems . 10
2.2.4 Invariant zeros . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.2.5 Properties of a linear system in geometric theory . . . . . . . . . . 12
2.3 Block oriented models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
2.4 Unknown input reconstruction . . . . . . . . . . . . . . . . . . . . . . . . . . 18
iii
CONTENTS
iv
CONTENTS
v
CONTENTS
References 180
Appendices 193
vi
Nomenclature
Abbreviations
AO-PE-UIO-HW adaptive order parity equations-based unknown input observer for Hammerstein-
Wiener systems
ARMAX . . . . . . . autoregressive model with moving average and exogenous input
ARX . . . . . . . . . . . autoregressive model with exogenous input
cf. . . . . . . . . . . . . . . confer (compare)
DC . . . . . . . . . . . . . direct current
DRFDF . . . . . . . . deadbeat robust fault detection filter
e.g. . . . . . . . . . . . . . exempli gratia (for example)
EIV . . . . . . . . . . . . errors-in-variables
INPEST . . . . . . . . input estimation (input reconstruction method)
LMI . . . . . . . . . . . . linear matrix inequalities
MIMO . . . . . . . . . . multiple-input multiple-output
MVU . . . . . . . . . . . minimum variance unbiased (MVU) state and input estimator
NMSS . . . . . . . . . . non-minimum state-space
OE . . . . . . . . . . . . . output error model
PE . . . . . . . . . . . . . parity equations
PE-UIO . . . . . . . . parity equations-based unknown input observer
PE-UIO-HW . . . parity equations-based unknown input observer for Hammerstein-Wiener systems
PIP . . . . . . . . . . . . proportional integral plus
PIP-LQ . . . . . . . . proportional integral plus linear quadratic
RFDF . . . . . . . . . . robust fault detection filter
SISO . . . . . . . . . . . single-input single-output
SVF . . . . . . . . . . . . state variable feedback
Notation
Latin variables
vii
Nomenclature
viii
Nomenclature
href (t) . . . . . . . . . exit gauge reference signal (in steel rolling mill model)
H(t) . . . . . . . . . . . input gauge (in steel rolling mill model)
H .............. feedforward matrix of unknown input in state-space model
H′ . . . . . . . . . . . . . . auxiliary matrix
J, J1 , J2 . . . . . . . . . gain matrices
l ................ stroke length of the piston (in steel rolling mill model)
lj , lj∗ . . . . . . . . . . . . transposes of left eigenvectors of filter state transition matrix
ix
Nomenclature
rq . . . . . . . . . . . . . . number of rows of Q
R .............. covariance matrix of ζ(t) or an auxiliary matrix (depending on context)
R̃ . . . . . . . . . . . . . . covariance matrix (used in MVU)
x
Nomenclature
z(t) . . . . . . . . . . . . position of the hydraulic piston (in steel rolling mill model)
zi . . . . . . . . . . . . . . system zero
zi (t) . . . . . . . . . . . . state estimate
zref (t) . . . . . . . . . hydraulic piston position reference signal (in steel rolling mill model)
Greek variables
xi
Nomenclature
∈ ............... element in
∪ ............... union of two sets or subspaces
/ ............... difference of two sets or subspaces
∩ ............... intersection of two sets or subspaces
⊂ ............... a subset of
⊆ ............... a subset or equal to
⊕ ............... direct sum
Rm×n . . . . . . . . . . . m × n dimensional space of real numbers
arg min f (x) . . . . value of x that minimises f (x)
rank(M ) . . . . . . . rank of matrix M
E {⋅} . . . . . . . . . . . . expected value operator
MT . . . . . . . . . . . . transpose of matrix M
M −1 . . . . . . . . . . . . inverse of matrix M
M ............. Moore-Penrose pseudo inverse of matrix M
M . . . . . . . . . . . . . nullspace of matrix M
V . . . . . . . . . . . . . . orthogonal completion of subspace V
Im{A} . . . . . . . . . . image of A
Ker{A} . . . . . . . . . kernel of A
round(q) . . . . . . . rounding of the scalar q to the nearest natural number
span{A} . . . . . . . . subspace spanned by A
sumrow (A) . . . . . column vector whose elements are sums of the appropriate rows A
var(e(t)) . . . . . . . variance of e(t)
z −1 . . . . . . . . . . . . . backward shift operator
xii
List of Algorithms
2.1 MVU . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
3.1 PE-UIO . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
3.2 PE-UIO for single-output OE . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
3.3 PE-UIO for multiple-output OE . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
3.4 Two stage PE-UIO . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
3.5 Generalised two stage PE-UIO . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
4.1 PE-UIO-HW . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
4.2 AO-PE-UIO-HW . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
5.1 RFDF via right eigenstructure assignment, q = 1, q1 = 1 . . . . . . . . . . . . . 114
5.2 RFDF via right eigenstructure assignment, q = 1, q1 ≥ 1 . . . . . . . . . . . . . 117
5.3 RFDF via right eigenstructure assignment, q ≥ 1, qi ≥ 0 . . . . . . . . . . . . . 119
5.4 Robust PE via right eigenstructure assignment . . . . . . . . . . . . . . . . . . 139
5.5 Robust PE via left eigenstructure assignment . . . . . . . . . . . . . . . . . . . 143
6.1 Fault isolation using directional PE . . . . . . . . . . . . . . . . . . . . . . . . . 154
6.2 Fault isolation and identification via diagonal PE . . . . . . . . . . . . . . . . 159
xiii
Chapter 1
1
1. Introduction, motivation and outline of approach
Fault diagnosis schemes based on the so-called Beard-Jones fault detection filter have
been in use up to this day in various industrial processes, e.g. chemical batch reac-
tors (Pierria, Paviglianiti, Caccavale & Mattei 2008), satellite attitude control systems
(Wang, Jiang & Shi 2008), or gas turbines (Gao, Breikin & Wang 2007).
Another important method which is used up to this day are parity equations (PE),
developed independently by Mironovski (1979) and Chow & Willsky (1984). Chan,
Hua & Hong-Yue (2006) applied PE for fault diagnosis of DC motors. A recent work
of Berriri, Naouar & Slama-Belkhodja (2011) presents a parity space approach for
diagnosis of a current sensor electrical system.
Fault detection and diagnosis algorithms are used practically in every industry
(Isermann 2005). They are not only used to detect abrupt malfunctions, but also to
signal wear and tear of machine parts, hence indicating when particular parts should
be replaced and facilitate the maintenance process.
The other topic, which is explored in this thesis, is an unknown (unmeasurable)
input reconstruction. Early contributions to this subject can be found in (Dorato 1969,
Sain & Massey 1969, Moylan 1977). Approximate input reconstruction has been used in
(Fu, Yan, Santillo, Palanthandalam-Madapusi & Bernstein 2009, Fu, Kirtikar, Zattoni,
Palanthandalam-Madapusi & Bernstein 2009) for diagnosing aircraft control surfaces.
Rocha-Cozatl, Moreno & Vande Wouwer (2012) utilised a continuous-discrete unknown
input observer in order estimate unknown variables in phytoplanktonic cultures. Recent
work of Czop (2011) presents reconstruction of the passenger vehicle wheel vertical
movement under ride conditions.
2
1. Introduction, motivation and outline of approach
unmeasured
inputs
?
measured measured
inputs outputs
System
1. Is there a malfunction in the system? If the answer is ‘yes’, a fault has been
detected. Thus, the first stage of the process is called fault detection.
3. By how much the component is faulty? Determining the magnitude of the fault is
denoted as fault identification, i.e. defining the quantity by which the particular
system parameter deviated from its acceptable/nominal value.
In the literature fault isolation and identification are often denoted as fault diagnosis.
Faults are usually modelled as extra inputs to the system. Thus, a fault diagnosis
process can be understood as unknown input estimation.
3
1. Introduction, motivation and outline of approach
completeness, at the beginning of Chapters 2–7 the nomenclature used in the par-
ticular chapter is provided. Additionally, Chapters 3–7 start with an indication of a
preliminary reading from specific sections of this thesis. Outlines of following chapters
are provided in Subsection 1.3.2.
Chapter 4: This chapter builds on the algorithm developed in Chapter 3. The un-
known input reconstruction method is extended to a class of nonlinear systems,
namely Wiener-Hammerstein systems, where a linear dynamic block is preceded
and followed by a memoryless nonlinear function. Similarly, as in Chapter 3,
an EIV framework is considered. Due to nonlinearities, the impact of the noise
on the unknown input estimate depends on the values of the known input and
output themselves, and is changing over time. Therefore, an adaptation scheme
is devised, which allows adjustment of the order of the parity space (and, con-
sequently, the filter bandwidth) based on the change of the measured input and
output.
4
1. Introduction, motivation and outline of approach
Chapter 5: There are two main outcomes of this chapter: firstly, the robust fault
detection filter based on right eigenstructure is extended to systems with unstable
invariant zeros, which extends the applicability of the aforementioned scheme. It
is also demonstrated that the devised algorithm is computationally simpler than
that of Chen & Speyer (2006a). Then, a robust PE of user-defined order is
designed using right and left eigenstructure assignment. In order to obtain an
open-loop solution (i.e. equivalent to PE) a finite time convergent state observer
is utilised. The disturbance decoupling property of the novel scheme is proven
algebraically and its efficacy is shown using a numerical example.
Chapter 8: In this chapter concluding remarks are given and proposals for further
work are stated.
1.4 Contributions
Contributions of the author are listed in descending order with respect to their consid-
ered relative significance.
5
1. Introduction, motivation and outline of approach
Chapter 2
Review
Chapter 3 Chapter 5
Parity equations−based Robust fault detection
unknown input reconstruction via eigenstructure assignment
for linear stochastic systems
Chapter 4 Chapter 6
Parity equations−based Fault isolation
unknown input reconstruction via diagonal PE
for Hammerstein−Wiener systems
Chapter 7
Potential applications
[1] Sumislawska, M., Burnham, K. J., Larkowski, T., Design of unknown input
estimator of a linear system based on parity equations. In Proc. of the
17th International Conference on Systems Science, pages 81–90, Wroclaw,
Poland, September 2010
[2] Sumislawska, M., Larkowski, T., Burnham, K. J., Design of unknown input
reconstruction algorithm in presence of measurement noise. In Proc. of
the 8th European ACD2010 Workshop on Advanced Control and Diagnosis,
pages 213–216, Ferrara, Italy, November 2010
[3] Sumislawska, M., Larkowski, T., Burnham, K. J., Design of unknown input
reconstruction filter based on parity equations for EIV case. In Proc. of the
18th IFAC World Congress, pages 4272–4277, Milan, Italy, 2011
[4] Sumislawska, M., Larkowski, T., Burnham, K. J., Parity equations-based
unknown input estimator for multiple-input multiple-output linear systems.
Systems Science, 36(3):49–56, 2010
6
1. Introduction, motivation and outline of approach
[5] Sumislawska, M., Larkowski, T. and Burnham, K. J., Unknown input recon-
struction observer for Hammerstein-Wiener systems in the errors-in-variables
framework. In Proc. of the 16th IFAC Symposium on System Identification,
pages 1377–1382, Brussels, Belgium, 2012.
[6] Sumislawska, M., Burnham, K. J., Hearns, G., Larkowski, T., Reeve, P.
J., Parity equation-based friction compensation applied to a rolling mill. In
Proc. of the UKACC International Conference on Control, pages 1043–1048,
Coventry, UK, September 2010
4. Robust fault detection based on eigenstructure assignment: Chen & Patton (1999)
have shown that a special case of the robust fault detection filter based on left
eigenstructure assignment is equivalent to first order PE. Also up to date the link
between the right eigenstructure assignment-based robust fault detection filter
and PE has not been derived. The main contribution of [7] and [8] is the design
of robust PE of an arbitrary order using, respectively, right and left eigenstructure
assignment.
[7] Sumislawska, M., Larkowski, T., Burnham, K. J., Design of parity equations
using right eigenstructure assignment. In Proc. of the 21st International
Conference on Systems Engineering, pages 367–370, Las Vegas, USA, August
2011
[8] Sumislawska, M., Larkowski, T. and Burnham, K. J., Design of robust par-
ity equations of user-defined order using left eigenstructure assignment. In
Proc. of the 9th European Workshop on Advanced Control and Diagnosis,
Budapest, Hungary, November 2011
7
1. Introduction, motivation and outline of approach
zeros are unstable. A solution is proposed in this thesis which allows the design
of stable robust fault detection filters for systems with unstable invariant zeros.
8
Chapter 2
Review
2.1 Introduction
The purpose of this chapter is to familiarise the reader with the notation and the
background knowledge used to develop algorithms in the next chapters. Firstly, the
notation of a linear discrete-time time-invariant stochastic system representation and a
class of nonlinear systems, namely, block-oriented systems, is provided in, respectively,
Section 2.2 and Section 2.3. In Section 2.4 the problem of an unknown input recon-
struction is presented. Two methods for a reconstruction of the unknown input signal
are presented, which are further used for a benchmark comparison with the algorithms
developed in Chapter 3. Section 2.5 provides the nomenclature used in the subject of
fault diagnosis as well as describes open- and closed-loop fault detection. Furthermore,
in Section 2.6 the problem of a robust (disturbance decoupled) fault detection is re-
viewed. A closely related fault isolation and identification is presented in Section 2.7.
Concluding remarks are given in Section 2.8.
9
2. Review
na nb
y(t) = − ∑ ai y(t − i) + ∑ bj u(t − j) (2.1)
i=1 j=0
The terms u(t) ∈ Rp and y(t) ∈ Rm refer to, respectively, the input and output vectors,
whilst na and nb , with na ≥ nb , are the orders of the auto-regressive and exogenous
parameters, respectively, and ai ∈ Rm×m and bi ∈ Rm×p are coefficient matrices.
where x(t) ∈ Rn is the state vector, whilst A ∈ Rn×n , B ∈ Rn×p , C ∈ Rm×n , and D ∈ Rm×p .
A notation (A, B, C, D) is used to refer to the system (2.2). In the case when D = 0,
the system (2.2) is denoted as (A, B, C). System (2.1) can be described by an observer
canonical form of a state-space model (Ljung 1999, Yiua & Wang 2007), where matrices
A, B, C, and D are:
⎡ ⎤
⎢ b1 − a 1 b0 ⎥
⎢ ⎥
⎡ −a
⎢ 0 ⋯ 0 ⎤
⎥
⎢
⎢ b2 − a 2 b0 ⎥
⎥
⎥ ⎢ ⎥
I
⎢
0 I ⋯ 0 ⎥ ⎢ ⎥
1
⎢ −a
⎢ ⎥ ⎢ ⋮ ⎥
⎥ ⎢ ⎥
2
⎢
⎢
A=⎢ ⋮ ⋮ ⋮ ⋱ ⋮ ⎥
⎥ B=⎢
⎢ bnb − anb b0 ⎥
⎥ C = [ I 0 ⋯ 0 ] D = b0 (2.3)
⎢ ⎥ ⎢ ⎥
⎢ −ana −1
⎢ 0 0 ⋯ I ⎥
⎥
⎢
⎢ −anb +1 b0 ⎥⎥
⎢ ⎥ ⎢ ⎥
⎢ −ana
⎣ 0 0 ⋯ 0 ⎥
⎦
⎢
⎢ ⋮ ⎥
⎥
⎢ ⎥
⎢ −ana b0 ⎥
⎣ ⎦
10
2. Review
where Π ∈ Rn×m , Ω ∈ Rm×m . The terms u0 (t) ∈ Rp and y(t) ∈ Rm refer to, respectively,
the input and output vectors. The term e(t) ∈ Rm is a column vector of m zero-mean,
white, Gaussian, independent and identically distributed (i.i.d.) noise sequences. The
term ũ(t) ∈ Rp is a vector of white, zero-mean, Gaussian i.i.d. noise sequences, which is
uncorrelated with e(t). Equation (2.4) is a generalised representation of a linear system
and can be simplified in more specific cases, some of which are given below.
A MIMO auto-regressive model with a moving average and exogenous input (ARMAX)
is given by, see (Ljung 1999, Yiua & Wang 2007):
na nb nc
y(t) = − ∑ ai y(t − i) + ∑ bj u(t − j) + ∑ ck e(t − k) (2.5)
i=1 j=0 k=0
where u(t) and y(t) are, respectively, the input and output vectors of the system and
e(t) is a vector of white, zero-mean, Gaussian, i.i.d. noise sequences. The terms na ,
nb , nc , with na ≥ nb and na ≥ nc , are the orders of the auto-regressive, exogenous and
moving average parameters, respectively, and ai ∈ Rm×m , bi ∈ Rm×p and ci ∈ Rm×m are
coefficient matrices. The last component of the right-hand side of (2.5) refers to the
moving average (coloured) process noise of the system.
The state-space system matrices (2.2) for the ARMAX model (2.5) in the observer
canonical form are given by:
⎡ ⎤ ⎡ ⎤
⎢ b1 − a 1 b0 ⎥ ⎢ c 1 − a1 c 0 ⎥
⎢ ⎥ ⎢ ⎥
⎡ −a 0 ⋯ 0 ⎤ ⎢ ⎥ ⎢ c 2 − a2 c 0 ⎥
⎢ ⎥ ⎢ b2 − a 2 b0 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
I
⎢ −a 0 I ⋯ 0 ⎥ ⎢ ⎥ ⎢ ⎥
1
⎢ ⎥ ⎢ ⋮ ⎥ ⎢ ⋮ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
2
A=⎢
⎢ ⋮ ⋮ ⋮ ⋱ ⋮ ⎥
⎥ B=⎢
⎢
⎥
bnb − anb b0 ⎥ Π = ⎢
⎢ c nc − a nc c 0 ⎥
⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ −ana −1 0 0 ⋯ I ⎥ ⎢ ⎥ ⎢ −anc +1 c0 ⎥
(2.6)
⎢ ⎥ ⎢ −anb +1 b0 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ −ana 0 0 ⋯ 0 ⎥ ⎢ ⎥ ⎢ ⎥
⎣ ⎦ ⎢ ⋮ ⎥ ⎢ ⋮ ⎥
⎢ ⎥ ⎢ ⎥
⎢ −ana b0 ⎥ ⎢ −ana c0 ⎥
⎣ ⎦ ⎣ ⎦
C =[ I 0 ⋯ 0 ] D = b0 Ω = c0
The ARMAX model assumes that the input u(t) is known exactly (there is no noise
present on the input variable), hence ũ(t) = 0 and u(t) = u0 (t). Note that an autore-
gressive model with an exogenous input (ARX) is obtained from an ARMAX model by
setting ci , i = 1, ⋯, nc , to zero.
An OE model assumes, that there is no process noise present in the system, however
the noise-free output y0 (t) is subjected to zero-mean, white, Gaussian measurement
11
2. Review
na nb
y0 (t) = − ∑ ai y0 (t − i) + ∑ bj u(t − j)
i=1 j=0 (2.7)
y(t) = y0 (t) + e(t)
This case can be modelled by the system representation (2.2), where matrices A, B, C,
and D are all given as in the ARMAX case. The matrix Π is null, and Ω is diagonal.
Also, there is no noise present on the input variable, hence ũ(t) = 0.
In the EIV framework, see, for example, (Söderström 2007), all measured variables, i.e.
the inputs and outputs of the system, are affected by zero-mean, white, Gaussian, i.i.d.
measurement noise sequences. This can be represented by (2.2), where ũ(t) ≠ 0, Π = 0,
and Ω is diagonal.
⎡ ⎤
⎢ zi I − A −B ⎥
⎢
P (zi ) = ⎢ ⎥
⎥
⎢
(2.8)
−D ⎥
⎣ ⎦
C
loses its rank (MacFarlane & Karcanias 1976). Two vectors are associated with an
invariant zero: the invariant zero state direction v ∈ Rn and the invariant zero input
direction g ∈ Rp , which conform the following equation (El-Ghezawi, Billings & Zinober
1983, Patel 1985, Patel & Munro 1982):
⎡ ⎤⎡ ⎤
⎢ zi I − A −B ⎥ ⎢ v ⎥
⎢ ⎥⎢ ⎥ = 0
⎢ ⎥⎢ ⎥
⎢ −D ⎥ ⎢ g ⎥
(2.9)
⎣ ⎦⎣ ⎦
C
12
2. Review
Ker{A} = {x ∈ Rm ∶ Ax = 0} (2.12)
Denote an n-dimensional space over the field of real numbers as X . Consider a matrix
V ∈ Rn×k , where k ≤ n. Denote V = Im{V }; then V is a k-dimensional subspace of the
space X . (Note that X = Im{I}.) Consider the following subspaces V, Y, and Z of
vector spaces Rn and Rm and a matrix A ∈ Rm×n . For completeness, basic operations
on subspaces are given below (Halmos 1958, Basile & Marro 2002):
1. Sum:
Z = V + Y ∶= {z ∶ z = v + y, v ∈ V, y ∈ Y} (2.13)
2. Intersection:
Z = V ∩ Y ∶= {z ∶ z ∈ V, z ∈ Y} (2.14)
3. Direct sum:
Z = V ⊕ Y ∶= {z ∶ z = v + y, v ∈ V, y ∈ Y, V ∩ Y = 0} (2.15)
4. Linear transformation:
Y = AV ∶= {y ∶ y = Av, v ∈ V} (2.16)
6. Orthogonal completion1 :
Y = V ∶= {y ∶ ⟨v, y⟩ = 0, v ∈ V} (2.18)
1
In the coordinate-free subspace algebra a product of vectors v and y is usually denoted as ⟨v, y⟩.
Note that this refers to v T y in the linear matrix algebra.
13
2. Review
Invariant subspaces
AV = V X (2.19)
Note that if X is diagonal, then columns of V are the eigenvectors of A. The invariance
has a physical meaning in the linear systems theory. Consider an autoregressive system
described by the following equation:
It holds that if x(t) ∈ V, where V is A-invariant, then x(t + 1) ∈ V. This means that if
the system is initialised with x(0) ∈ V, then the state vector will remain within V.
Using the notation B = Im{B}, the term ⟨A∣B⟩ = B + AB + ⋯ + An−1 B is the infimal
A-invariant subspace containing B, i.e. the reachable subspace of (A, B). This means
that the state trajectory of the system (2.21) driven by the input u(t) can be anywhere
within the reachable subspace of (A, B). Furthermore, because ⟨A∣B⟩ is A-invariant,
the state driven by u(t) cannot leave the reachable subspace of (A, B). Note that, the
reachable subspace of (A, B) can be defined as:
where:
R = [ B AB ⋯ An−1 B ] (2.23)
x(t + 1) = Ax(t)
(2.24)
y(t) = Cx(t)
Denote the kernel of the matrix C as K = Ker{C}, i.e. CK = 0. Then the unobservable
subspace of the system (2.24) is defined as the supremal A-invariant subspace contained
14
2. Review
in K (Massoumnia 1986):
Note that if the state vector x(t) ∈ K then y(t) = Cx(t) = 0. Because ⟨K∣A⟩ ⊆ K, it
holds that y(t) = Cx(t) = 0 for any x(t) ∈ ⟨K∣A⟩. Furthermore, due to the fact that
⟨K∣A⟩ is A-invariant it holds that x(t + 1) ∈ ⟨K∣A⟩ if x(t) ∈ ⟨K∣A⟩, i.e. the state vector
stays within ⟨K∣A⟩. Thus, if the system (2.24) is initialised with x(0) ∈ ⟨K∣A⟩, then the
state vector remains within the unobservable subspace of (A, C) and the output y(t)
remains zero.
Now consider the relation between ⟨K∣A⟩ and the system observability matrix O:
⎡ ⎤
⎢ ⎥
⎢ ⎥
C
⎢ CA ⎥
⎢ ⎥
O=⎢ ⎥
⎢ ⎥
(2.26)
⎢ ⋮ ⎥
⎢ ⎥
⎢ CAn−1 ⎥
⎣ ⎦
(V ∩ Z) = V + Z (2.27)
Hence:
⟨K∣A⟩ = K + (A−1 K) + ⋯ + (A−n+1 K)
(2.28)
(2.30)
+ Im{(An−1 )T C T } = Im{[ C T AT C T ⋯ (An−1 )T C T ]}
15
2. Review
The concept of controlled invariants was introduced in (Basile & Marro 1969, Wohnam
& Morse 1970), whilst the concept of conditioned invariance was introduced in (Basile
& Marro 1969). Consider a pair (A, B). A subspace V is an (A, B)-controlled invariant
if:
AV ⊆ V + B (2.32)
This means that there exists such a matrix K that the input u(t) = Kx(t) keeps the
state of the system (2.21) within V, i.e. there exists such a K that (A + BK)V ⊆ V. A
dual of the controlled invariant is a conditioned invariant. Consider a pair (A, C). A
subspace S is said to be an (A, C)-conditioned invariant if:
A(S ∩ K) ⊆ S (2.33)
The (A, C)-conditioned invariance means that there exists such a matrix K that (A −
KC)S ⊆ S. For more properties of controlled and conditioned invariants the reader is
referred to (Basile & Marro 2002).
Invariant zeros
Denote the minimal (A, C)-conditioned invariant containing Im{B} as S0 and use the
notation V0 for the maximal (A, B)-controlled invariant contained in ker{C}. Consider
a matrix V1 such that Im{V1 } ∩ V0 = Im{V1 }, Im{V1 } + V0 = V0 , and Im{V1 } ∩ S0 = 0.
Invariant zeros of (A, B, C) are the eigenvalues of the matrix M1 , which fulfils the
following equation, see (Basile & Marro 2010):
⎡ ⎤
⎢ M1 ⎥
[ V1 ⎢
−B ] ⎢ ⎥ = AV1
⎥
⎢ M2 ⎥
(2.35)
⎣ ⎦
16
2. Review
Note that if the invariant zeros of (A, B, C) are distinct, the matrix M1 is diagonalisable,
and J is diagonal, equation (2.38) is equivalent to (2.9), where the diagonal elements of
J are the invariant zeros of (A, B, C), whilst columns of V1 Vm−11 and columns of M2 Vm−11
are, respectively, invariant zeros state and input directions. Therefore, if the invariant
zeros of (A, B, C) are distinct, (2.9) is equivalent to (2.35), where D = 0. Nevertheless,
as opposed to (2.9), Equation (2.35) can be used to determine the number of all of the
invariant zeros of the system, including the repeated ones.
In the case when the system contains a feedthrough term, i.e. D ≠ 0 Basile &
Marro (2002) proposed some manipulations to represent the quadruple (A, B, C, D)
with a triple. However, the geometric approach is used in this thesis to analyse/design
fault detection and isolation filters, where the considered transfer functions between
disturbances or faults and the output of the system do not contain any feedthrough
term. Thus, for more details on invariant zeros of systems with a feedthrough term the
reader is referred to (Basile & Marro 2002).
17
2. Review
u(t) y(t)
G1 (z) N1 (⋅) H1 (z) ... Gk (z) Nk (⋅) Hk (z)
18
2. Review
where u(t) is a vector of unknown inputs (the number of unknown inputs is lower or
equal to the number of outputs), whilst ξ(t) and ζ(t) are vectors of white, zero-mean,
Gaussian, i.i.d. noise sequences.
It is assumed throughout this thesis that the noise distribution is Gaussian (i.e.
normally distributed) and zero-mean. Whilst in practice it is known that noise is not
necessary Gaussian, it is commonly accepted that a Gaussian assumption is appropriate;
offering an approach which, although may no longer be optimal, would be consistent
and generalisable, and is applicable to a wide range of situations.
1. Initialisation
3. Measurement update
19
2. Review
4. Time update
The MVU requires the knowledge of the covariance matrices of the noise sequences,
Q̃ and R, which are the tuning parameters for the algorithm.
Remark 2.1. If the system (2.39) is SISO, the MVU is equivalent to a naive system
inversion.
The gain of the filter (2.43) is equal to the reciprocal of the gain of the system, see
(Gillijns & De Moor 2007b). Denote any arbitrary zero of the filter (2.43) as zmvu , then
the following formula holds, cf. (2.9):
⎡ ⎤⎡ ⎤ ⎡ ⎤
⎢ zmvu I − A + M GC −M G ⎥ ⎢ ⎥ ⎢ ⎥
⎢ ⎥⎢ χ ⎥ = ⎢ 0 ⎥
⎢ ⎥⎢ ⎥ ⎢ ⎥
⎢ −M ⎥ ⎢ κ ⎥ ⎢ 0 ⎥
(2.44)
⎣ −M C ⎦⎣ ⎦ ⎣ ⎦
where a column vector χ denotes the zero state direction and a scalar κ refers to the
input zero direction of zmvu (El-Ghezawi et al. 1983). Consequently:
20
2. Review
Analogously, it can be demonstrated that the poles of (2.43) are equivalent to the
zeros the the system (2.39). Thus, the MVU behaves as a naive inversion in the case
when (2.39) is a SISO system.
y(t) + ŷ(t − τ )
controller model
_
(2.46)
where z(t) is an integral of the tracking error:
where g is the controller gain. The INPEST method makes use of the PIP linear
quadratic (PIP-LQ) design, which minimises the cost function defined as:
∞
J = ∑ {xT (t)Qx(t) + rû2 (t)} (2.49)
t=0
21
2. Review
where:
Q = diag([ qy ⋯ qy qu ⋯ qu qe ]) (2.50)
where ν is a tuning parameter. The reason for penalising the derivative of the recon-
structed input signal is that in the case of noisy measurements of y(t) the algorithm
would tend to amplify the effects of the disturbance if the rate of change on û(t) was
not penalised. Note that the only tuning parameter of the INPEST method is the
reconstructed input derivative weighting ν.
∑t (û(t) − u(t))2
R2T = × 100%
∑t u2 (t)
(2.52)
where u(t) and û(t) refer to, respectively, the unknown input and its estimate. Note
that in an ideal case, i.e. when û(t) = u(t), R2T = 0 and it increases as the discrepancy
between the original and the estimated input increases.
22
2. Review
2.5.1 Nomenclature
The fault detection and diagnosis terminology used throughout this thesis is given
below (Isermann & Balle 1997):
Fault detection: A binary decision, whether a fault is present in the system. Due to
the fact that there is always a certain level of noise in the system a need arises
to distinguish, whether the residual deviates from zero due to the noise or due
to presence of a fault. This is achieved by setting a threshold whose violation
indicates the presence of a fault.
Fault isolation: Determination of the component which deviates from the accept-
able/usual/standard condition.
Fault diagnosis: Includes fault isolation and identification, i.e. determination of the
source of the fault and the fault magnitude.
23
2. Review
where x(t) ∈ Rn is the system state vector, u(t) ∈ Rp and y(t) ∈ Rm are, respectively,
the system input and output, and µ(t) ∈ Rk is a fault signal. Matrices A, B, C,
D, and F are constant and have appropriate dimensions. The aim of fault detection
is to define the time instances t, when µ(t) ≠ 0. When both the input and output
measurements are available, so-called process-model-based fault detection methods are
used for the purpose of residual generation (Isermann & Balle 1997, Simani, Fantuzzi
& Patton 2002). These, in particular, are:
2. PE
In this thesis two fault detection methods are considered, which are the state observers
and the PE.
A schematic illustration of a state observer is presented in Fig. 2.5 and given by the
following set of equations (Patton 1997, Chen & Patton 1999, Simani et al. 2002):
where K is the observer gain matrix, whilst Q is an arbitrary matrix. (The matrix
Q plays an important role in robust fault detection which is described in Section 2.6;
in this section, however, without loss of generality it is assumed that Q is an identity
matrix.) Consider the state estimation error ξ(t) = x̂(t) − x(t). Then the residual is
governed by:
24
2. Review
µ(t)
x(t + 1) = Ax(t) + Bu(t) + F µ(t) y(t)
y(t) = Cx(t) + Du(t)
u(t)
B
+ x̂(t + 1) −1 x̂(t)
C
+ ŷ(t) +
_
r(t)
z + Q
+ +
PE are widely used for the purpose of fault detection and isolation, see, for example,
(Chow & Willsky 1984, Gertler & Singer 1990, Li & Shah 2002). A schematic illus-
tration of PE is presented in Fig. 2.6. As opposed to the observer-based (closed-loop)
fault detection filters, PE have an open-loop structure.
µ(t)
x(t + 1) = Ax(t) + Bu(t) + F µ(t) y(t)
y(t) = Cx(t) + Du(t)
u(t)
delay delay
_
WQ + W
r(t)
Consider a state-space representation of the system (2.2). The stacked vector of the
system output y(t) is defined as:
Y (t) = [ y T (t − s) y T (t − s + 1) ⋯ y T (t) ]
T
(2.57)
25
2. Review
where the term s denotes the order of the parity space. Analogously, one can construct
a stacked vector of u(t), which is denoted as U (t). Using this notation the system
defined by (2.53) in a fault-free case can be expressed in the form of:
⎡ C ⎤
⎢ ⎥
⎢ ⎥
⎢ CA ⎥
⎢ ⎥
Γ=⎢ ⎥ ∈ R(s+1)m×n
⎢ ⋮ ⎥
(2.59)
⎢ ⎥
⎢ ⎥
⎢ CAs ⎥
⎣ ⎦
⎡ 0 ⎤
⎢ ⋯ ⎥
⎢ ⎥
D 0
⎢ CB 0 ⎥
⎢ ⋯ ⎥
⎢ ⎥
D
⎢
Q = ⎢ CAB 0 ⎥
⎥∈R
(s+1)m×(s+1)p
⋯ (2.60)
⎢ ⎥
CB
⎢ ⋮ ⎥
⎢ ⋮ ⋮ ⋱ ⎥
⎢ ⎥
⎢ CA B CA B D ⎥
⎣ ⋯ ⎦
s−1 s−2
For the purpose of elimination of the unknown state vector from (2.58), a matrix
W ∈ Rl×(s+1)m , l ≥ 1, is defined, which belongs to the left nullspace of Γ, i.e.
WΓ = 0 (2.61)
deviates from zero if a fault occurs in the system. It is worth noting that, because the
residual response to fault is open-loop, the residual is correlated only with the last s + 1
samples of the fault signal.
26
2. Review
2. QCE = 0
It is assumed that E is full column rank. Denote the ith column of E as Ei , which
is also a right eigenvector of (A − KC) corresponding to the desired eigenvalue λi . In
order to satisfy the decoupling condition 1, it should hold that:
The procedure for finding such a gain matrix K, for which the decoupling conditions
are fulfilled, has been proposed by Chen & Patton (1999). Using the notation:
27
2. Review
KCE = Aλ (2.68)
A′ = A − Aλ (CE) C
(2.69)
C ′ = (I − CE(CE) )C
28
2. Review
where vi and liT are, respectively, the right and left eigenvectors of (A − KC) corre-
sponding to the eigenvalue λi , see (Patton & Chen 1991b, Patton & Chen 1992, Chen
& Patton 1999). Denote left and right eigenvector matrices, respectively, as:
⎡ l1T ⎤
⎢ ⎥
⎢ ⎥
⎢ T ⎥
⎢ l2 ⎥
L=⎢ ⎥ V = [ v1 v2 ⋯ vn ]
⎢ ⋮ ⎥
(2.73)
⎢ ⎥
⎢ ⎥
⎢ lnT ⎥
⎣ ⎦
It is known that the left eigenvector liT is orthogonal to the right eigenvector vj if i ≠ j,
cf. (Patton & Chen 1991b, Chen & Patton 1999). Therefore, if the vectors liT and vi
are appropriately scaled:
LV = I (2.74)
and hence:
L = V −1 (2.75)
The transfer function between the disturbance and the residual can be expressed
as:
n QCvi liT E
Grd (z) = ∑ (2.76)
i=1 z − λi
Hence, Grd (z) vanishes if and only if for i = 1, ⋯, n:
Therefore, the first step for the disturbance decoupling is to find the matrix Q, such that
QCE = 0, see (Patton & Chen 1991b, Chen & Patton 1999). Consequently, sufficient
conditions for the disturbance decoupling using the left eigenstructure assignment are:
1. QCE = 0
29
2. Review
The proof of the above conditions can be found in (Chen & Patton 1999).
Assignability condition
where liT is the ith row of QC. The above expression can be reformulated as:
Consequently:
liT = −liT KC(A − λi I)−1 (2.81)
Note that liT ∈ R1×n and K ∈ Rn×m . Therefore, liT K ∈ R1×m , whilst the matrix C(A −
λi I)−1 ∈ Rm×n . This means that by premultiplying the matrix C(A − λi I)−1 by a row
vector liT K a linear combination of rows of C(A − λi I)−1 is obtained. Consequently,
a solution to (2.79) exists for the desired λi if and only if the vector liT lies in a
row subspace spanned by C(λi I − A)−1 , i.e. li lies in a column subspace spanned by
(λi I −AT )−1 C T , see (Chen & Patton 1999). Therefore, li must be equal to its projection
on the subspace Im{(λi I −AT )−1 C T }. Denote (λi I −AT )−1 C T as P (λi ). The projection
of li onto Im{(λi I − AT )−1 C T } is given by:
where:
wi∗ = [P (λi )T P (λi )]−1 P (λi )T li for i = 1, ⋯, rq (2.83)
In the case when li∗ = li , the left eigenvector li is assignable. Otherwise a complete dis-
turbance decoupling using the left eigenstructure assignment is not possible. Consider
the following equation:
(li∗ )T (A − KC) = λi (li∗ )T (2.84)
it holds that:
li∗ = −(λi I − AT )−1 C T K T li∗ (2.85)
30
2. Review
where:
L = [ l1∗ ⋯ lr∗q lrq +1 ⋯ ln ] (2.89)
and:
W = [ w1∗ ⋯ wr∗q wrq +1 ⋯ wn ] (2.90)
This occurs when eigenvalues of (A − KC) assigned to the columns of QC are equal
zero. Then, the z-form of the residual is, cf. (2.54):
where u(z), y(z), and r(z) are the z-transform forms of, respectively, u(t), y(t), and
z(t). Note that:
Hence, QC(zI −A+KC)−1 = z −1 QC. Therefore, the computational form of the residual
vector r(z) can be rewritten as:
⎡ ⎤ ⎡ ⎤
⎢ y(t) ⎥ ⎢ ⎥
r(t) = [ Q −QCK ] ⎢
⎢
⎥ − [ QD −QC(B − QD) ] ⎢ u(t)
⎥ ⎢
⎥
⎥
⎢ y(t − 1) ⎥ ⎢ u(t − 1) ⎥
(2.95)
⎣ ⎦ ⎣ ⎦
31
2. Review
The above scheme is also referred as a deadbeat robust fault detection filter (DRFDF),
see (Chen & Patton 1999).
Structured residual set: Each fault yields certain residuals deviate from zero, whereas
other residuals remain zero. This can be interpreted as the fault µi (t) causing the
residual vector to lie in a certain subspace of the residual space, see Fig 2.7(a).
An example of a structured residual set is presented in Table 2.1.
Fixed direction residuals: Presence of the fault µi (t) yields the residual to lie in a
fixed direction, see Fig 2.7(b). Residual directions do not need to be linearly in-
dependent. However, multiple faults cannot be detected unless residual directions
are linearly independent.
Diagonal residual set: A combination of the two above, i.e. the fault µi (t) causes
the residual ri (t) to deviate from zero, whilst the remaining residuals are equal
to zero. A diagonal residual set can be used to isolate multiple faults. Note that
a diagonal residual set can be obtained from a set of linearly independent fixed
direction residuals by a similarity transformation (change of basis).
r3 (t)
r3 (t)
µ2 (t)
µ3 (t)
µ2 (t)
µ3 (t)
µ1 (t)
r2 (t) r2 (t)
µ1 (t)
r1 (t) r1 (t)
(a) Structured residual set (b) Directional residual set
32
2. Review
Table 2.1: Example of a structured residual set. The entry 1 in the ith row (cor-
responding to the fault µi (t)) and the j th column (corresponding to the
residual rj (t)) denotes that µi (t) yields the residual rj (t) deviate from
zero. Note that multiple faults cannot be isolated using this residual set.
that the fault isolation schemes developed by them are often referred to as ‘Beard-Jones
fault detection filters’. In this thesis the term ‘fault detection’ refers to the process of
determining a fault occurrence, whereas the schemes proposed by Beard (1971) and
Jones (1973) are, for sake of consistency, referred to as ‘fault isolation filters’.
Consider the system described by equation (2.53). Denote each column of the matrix
F as Fi and each corresponding fault signal as µi (t). Let δi be the smallest non-negative
integer such that CAδi Fi ≠ 0. Then the term fi = Aδi Fi is further referred to as the
fault direction (Massoumnia 1986, Chen & Speyer 2006a). Consider the following fault
isolation filter:
The residual lies in the direction Cfi when the fault µi occurs
In order for the filter design to be feasible, the following conditions must be fulfilled
(Chow & Willsky 1984, Massoumnia 1986, Chen & Speyer 2006b):
Assumption 1. ensures that all the filter eigenvalues can be arbitrarily specified, whereas
Assumption 2. allows for the faults to be isolated, i.e. yields residuals caused by
different faults lie in different directions.
2
Constraint of no repeated eigenvalues is often imposed for clarity of analysis and derivation pro-
cesses (Chen & Patton 1999, Chen & Speyer 2006a)
33
2. Review
Different solutions have been proposed to design a fault diagnostic observer. Mas-
soumnia (1986) represented the Beard-Jones fault isolation filter in a geometric do-
main and, furthermore, added a solution for the filter design when the output re-
sponse to fault has invariant zeros. White & Speyer (1986) reformulated the Beard-
Jones fault isolation filter to an eigenstructure assignment problem. Then, Chen &
Speyer (2006b) used the spectral theory to design a Beard-Jones fault isolation filter.
Furthermore, a design of the filter has been presented using eigenstructure assign-
ment (Chen & Speyer 2006a) and linear matrix inequalities (LMI), see e.g. (Chen &
Nagarajaiah 2007).
A fault isolation filter design using the right eigenstructure assignment developed by
Chen & Speyer (2006a) is presented here. Eigenvalues of (A − KC) can be arbitrarily
specified and for each column of F ni eigenvalues, denoted λj , are allocated to (A −
(i)
(i)
KC), corresponding to the eigenvectors wj , j = 1, 2, ⋯, ni :
(A − KC)wj = λj wj
(i) (i) (i)
(2.97)
It is demonstrated in (Chen & Speyer 2006a) that the number of assignable eigenvalues,
ni , depends on the rank of the observability matrix of (Ci , Ai ):
⎡ ⎤
⎢ ⎥
⎢ ⎥
Ci
⎢ CA ⎥
⎢ ⎥
Oi = ⎢ ⎥
i i
⎢ ⎥
(2.98)
⎢ ⋮ ⎥
⎢ ⎥
⎢ Ci Ain−1 ⎥
⎣ ⎦
ni = n − rank(Oi ) (2.99)
KCfi = (A − λj )wj
(i) (i)
(2.101)
34
2. Review
(i) (i)
wj = Θi βj (2.102)
⎡ (i) ⎤
mulated as:
⎢ β̄j ⎥
= [ Θ̄i fi ] ⎢
⎢
⎥ = Θ̄i β̄ (i) + fi
⎥
(i)
⎢ 1 ⎥
wj (2.103)
⎣ ⎦
j
where:
Im{Θ̄i } = Im{[ Fi AFi ⋯ Aδi −1 Fi ]} ⊕ Vi (2.104)
After subtracting the last row of (2.106) from the others, the following expression is
obtained:
⎡ (A − λ1 )Θ̄i ⎤
−(A − λni )Θ̄i
⎢ ⎥
(i) (i)
⋯
⎢ ⎥
0
⎢ (A − λ2 )Θ̄i ⋯ −(A − λni )Θ̄i ⎥
⎢ ⎥
(i) (i)
⎢ ⎥
0
⎢ ⎥
⎢ ⋮ ⋮ ⋱ ⋮ ⎥
⎢ ⎥
⎢ ⋯ (A − λni −1 )Θ̄i − (A − λni )Θ̄i ⎥
⎣ ⎦
(i) (i)
0 0
⎡ β̄1 ⎤ ⎡ ⎤
(2.107)
⎢ ⎥ ⎢ (λ1 − λni )fi ⎥
(i) (i) (i)
⎢ ⎥ ⎢ ⎥
⎢ β̄2 ⎥ ⎢ (λ2 − λni )fi ⎥
⎢ ⎥ ⎢ ⎥
(i) (i) (i)
⎢ ⎥=⎢ ⎥
⎢ ⋮ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⋯ ⎥
⎢ (i) ⎥ ⎢ ⎥
⎢ β̄q1 +1 ⎥ ⎢ (λni −1 − λni )fi ⎥
⎣ ⎦ ⎣ ⎦
(i) (i)
Denote the matrix at the left hand side of (2.107) as Ãi . Then the coefficient vectors
(i)
β̄j , j = 1, 2, ⋯, ni are calculated using a pseudoinverse of Ãi :
35
2. Review
KCf = [ λ(1)
1 w1
(1) (2) (2)
λ1 w 1 ⋯ λ1 w 1 ]
(q) (q)
(2.110)
36
2. Review
tion and diagnosis. Both closed-loop (observer-based) and open-loop (PE-based) fault
detection/isolation filters have been discussed and appropriate algorithms selected from
the literature have been presented.
37
Chapter 3
Nomenclature
38
3. Parity equations-based unknown input reconstruction for linear stochastic systems
39
3. Parity equations-based unknown input reconstruction for linear stochastic systems
3.1 Introduction
In the literature the problem of the unknown (unmeasurable) input estimation is solved
either by a system inversion or by a joint state and input estimation. Early contribu-
tions to the inversion of multiple-input multiple-output (MIMO) deterministic systems
have been presented by Dorato (1969) and Sain & Massey (1969), however their ap-
proaches did not ensure stability of the inverted systems. Moylan (1977) provided a
stable inversion algorithm for minimum-phase systems, whilst Antsaklis (1978) devel-
oped a straightforward state feedback-based method, which allows to assign poles of
the inverted system. This latter method is however limited to the systems with stable
zeros.
Over the last decade a geometric approach to an unknown input reconstruction has
gained considerable interest, see e.g. (Edelmayer 2005). Kirtikar, Palanthandalam-
Madapusi, Zattoni & Bernstein (2009) proposed an unknown input reconstruction
scheme for minimum phase systems. An exhaustive solution to an unknown-state
unknown-input reconstruction for both minimum-phase and nonminimum-phase sys-
tems has relatively recently been developed by Marro & Zattoni (2010). Nevertheless,
this approach does not consider the effects of measurement noise.
Another approach to the unknown input estimation for deterministic systems is
based on state observers. The Luenberger state observer, see (Luenberger 1964), has
been extended to the class of systems with both, known and unknown system inputs, see
for example (Hou & Müller 1992, Darouach & Zasadzinski 1997). The work of Fernando
& Trinh (2006) presents a joint input and state observer based on a descriptor approach.
40
3. Parity equations-based unknown input reconstruction for linear stochastic systems
When dealing with stochastic systems Kalman filter-based approaches have gained
an interest, see, for example, (Hsieh 2000, Floquet & Barbot 2006). Gillijns & De Moor
(2007a) combined the state observer proposed by Darouach & Zasadzinski (1997) and
the unknown input estimator of Hsieh (2000) creating a joint state and unknown input
observer, which is optimal in the minimum variance sense. This approach has subse-
quently been extended to the case of a linear system with a direct feedthrough term, see
(Gillijns & De Moor 2007b). Palanthandalam-Madapusi & Bernstein (2007) introduced
concept of a state and input observability, i.e. they provided a scheme, which allows
to determine, if both the unknown input and the state can be derived from the output
measurements. Keller & Sauter (2010) proposed a variable geometric Kalman filter,
where the statistical effect of each unknown input is tested before deriving the state
estimate. In the recent work of Ghahremani & Kamwa (2011) an extended Kalman
filter with unknown inputs has been developed and applied to state estimation of a
synchronous machine in a power system.
In this chapter a novel approach to the unknown input reconstruction for MIMO
discrete-time stochastic systems is presented. The parity equation-based unknown in-
put observer (PE-UIO) utilises a parity equations (PE) concept for the unknown input
reconstruction. The design freedom is used to minimise the effect of stochastic distur-
bances on the unknown input estimate. For this purpose a Lagrange multiplier method
is utilised. The proposed method is suitable for both minimum and nonminimum-
phase systems, which is an important result, because unstable zeros may result from
a discretisation of a continuous-time system. The PE-UIO has been originally devel-
oped for single-input single-output (SISO) output error (OE) systems in (Sumislawska,
Burnham & Larkowski 2010). The algorithm has been subsequently extended to the
errors-in-variables (EIV) framework in (Sumislawska, Larkowski & Burnham 2010b).
The analysis of the PE-UIO in frequency domain has been provided in (Sumislawska,
Larkowski & Burnham 2011a). In (Sumislawska, Larkowski & Burnham 2010a) the
scheme has been extended to a MIMO case and a potential application to a steel
rolling mill has been described. In this chapter the PE-UIO is extended to a coloured
process noise case. A generalised form of the algorithm is provided, where the output is
subjected to coloured noise (accounting for measurement and process noise), whilst the
input is affected by white measurement noise. An extension of the PE-UIO algorithm
for the cases when systems zero is close or equal to unity is also provided.
This chapter is organised as follows: in Section 3.2 the problem of the unknown
input reconstruction is stated. Subsequently, in Section 3.3 the PE-UIO is presented.
Then, in Section 3.5, the limitation of the scheme in the case when the system has
zeros close or equal to unity is discussed and an extension, which tackles this problem,
is provided. The proposed algorithms are demonstrated on tutorial examples in Sec-
tion 3.7. Finally, in Section 3.8, the efficacy of the proposed methods is compared with
two existing methods, namely, the minimum variance unbiased (MVU) joint state and
41
3. Parity equations-based unknown input reconstruction for linear stochastic systems
e(t)
v(t)
linear y(t)
u0 (t) system
ũ(t) u(t)
input estimator, see (Gillijns & De Moor 2007b), and the input estimation (INPEST)
method of Young & Sumislawska (2012).
42
3. Parity equations-based unknown input reconstruction for linear stochastic systems
OE: An OE case can be modelled by the system representation (3.1), where matrices
A, B, C, D and also G and H are all given as in the ARMAX case. The matrix
Π is null and Ω is diagonal. Also there is no noise present on the input variable,
hence ũ(t) = 0. The PE-UIO algorithm for a SISO OE case has been developed
in (Sumislawska, Burnham & Larkowski 2010).
EIV framework: The EIV framework, see, for example, (Söderström 2007), can be
represented by (3.1), where ũ(t) ≠ 0, Π = 0, and Ω is diagonal. The PE-UIO algo-
rithm for a SISO case in the EIV framework has been presented in (Sumislawska,
Larkowski & Burnham 2010b, Sumislawska et al. 2011a).
Y (t) = [ y T (t − s) y T (t − s + 1) ⋯ y T (t) ]
T
(3.2)
where the term s denotes the order of the parity space. Analogously, one can construct
stacked vectors of v(t), u(t), u0 (t), ũ(t) and e(t) which are denoted, respectively, as
V (t), U (t), U0 (t), Ũ (t) and E(t). Using this notation the system defined by (3.1) can
be expressed in the form of:
43
3. Parity equations-based unknown input reconstruction for linear stochastic systems
By rearranging the measured (known) variables to the right-hand side of (3.4) and the
unknowns to the left-hand side, the following parity relation is obtained, cf. (Li &
Shah 2002):
W T V (t) + W ΞE(t) − W QŨ (t) = W Y (t) − W QU (t) (3.5)
In the next subsection the PE are used in order to derive the PE-UIO.
Selection of a sufficiently large s would lead (3.6) to be a set of equations with an explicit
solution or an overdetermined set of equations. Nevertheless, in practice, precision of
the solution to (3.6) can still be seriously affected by noise. The algorithm proposed here
provides an on-line approximation of the unknown input, simultaneously minimising
unwanted effects of noise.
It is proposed to calculate the value of the unknown input as:
where τ is an estimation lag (estimation delay) and it accounts for the fact that the un-
known input may not be reconstructed instantenously. Therefore, at the time instance
t the estimate of v(t − τ ) is obtained. The estimation delay τ is defined further in this
section. In the noise-free case, v̂(t − τ ) is simply:
Therefore, based on the assumption that the unknown input is varying relatively slowly
(see Subsection 3.5.1), its estimate can be calculated as a linear combination of the
sequence v(t − s), v(t − s + 1), ⋯, v(t), i.e.
44
3. Parity equations-based unknown input reconstruction for linear stochastic systems
where the α parameters are dependent on the choice of the vector W , such that:
W T = [ αs αs−1 ⋯ α0 ]
T
(3.11)
One can note that (3.10) represents a moving average finite impulse response filter with
the gain being given by the sum of the α parameters, i.e. the sum of elements of the
vector W T . Thus, it is suggested that W should be selected in such a way, that the
sum of elements of the vector W T is equal unity. Furthermore, it is anticipated that
the choice of the order of the parity space s, as well as the vector W , both influence
the estimation lag τ in the estimate of the unknown input (due to the moving average
filtering property of the unknown input estimator). The estimation lag is defined as the
centre of gravity of the moving average filter rounded to the nearest natural number
and is calculated via:
∑ αi i
τ = round ( ) (3.12)
∑ αi
In the following subsection an algorithm for the selection of the optimal vector W
is derived based on the Lagrange multiplier method.
where the vector coefficients wξi and wqi , i = 1, 2, ⋯, s + 1, are constructed from the
appropriate elements of the vectors W Ξ and W Q, respectively. (In the case when
p = m = 1, i.e. u(t) and y(t) are scalars, wξi and wqi refer to the ith elements of vectors
W Ξ and W Q, respectively.) Note, that in (3.14) the estimate of the unknown input
is affected by two coloured noise sequences. However, by a careful choice of W , the
degrading effect of these disturbances can be minimised.
Furthermore, the influence of measurement noise on the unknown input estimate
can be reduced by minimising the variance of the term W ΞE(t) − W QŨ (t), i.e.:
where Σũ = E{Ũ (t)Ũ T (t)}, Σe = E{E(t)E T (t)}, and Σũe = E{Ũ (t)E T (t)} = 0. Conse-
45
3. Parity equations-based unknown input reconstruction for linear stochastic systems
2. W Γ = 0.
Note, that the condition 1 is sufficient to ensure unity gain, because E{e(t)} = 0 and
E{ũ(t)} = 0.
The cost function (3.16) can be minimised by making use of the Lagrange multiplier
method, see, for example, (Bertsekas 1982). Denote the rows of Γ by γ1 , γ2 , ..., γk ,
where:
Γ = [ γ1T ⋯ γkT ]
T
γ2T (3.17)
The vector W is a linear combination of rows of Γ , which ensures that the constraint
2 is satisfied, i.e.
k
W = ∑ pi γ i (3.18)
i=1
Hence, the cost function (3.16) can be reformulated as a function of the parameter
vector P = [ p1 p2 ⋯ pk ] :
T
k ⎛k ⎞
f (P ) = (∑ pi γi ) Σ ∑ pj γjT
⎝j=1 ⎠
(3.19)
i=1
where
Σ = ΞΣe ΞT + QΣũ QT (3.20)
where the operator sumrow (A) denotes a column vector whose elements are sums of
the appropriate rows of an arbitrary matrix A. (In the case of a row vector q, the term
sumrow (q) is simply a scalar being a sum of elements of the vector q, whilst, if q is a
column vector, sumrow (q) = q.)
The solution to the Lagrange minimisation problem is given by, see (Bertsekas 1982):
∇f (P ) = λ∇g(P ) (3.22)
46
3. Parity equations-based unknown input reconstruction for linear stochastic systems
Hence, the partial derivative of f (P ) with respect to the ith element of the vector
P (denoted as pi ) is given by:
∂f (P )
=p1 γi Σγ1T + p2 γi Σγ2T + ⋯ + pi γi ΣγiT + ⋯ + pk γi ΣγkT +
∂pi (3.24)
p1 γ1 ΣγiT + p2 γ2 ΣγiT + ⋯ + pi γi ΣγiT + ⋯ + pk γk ΣγiT
⎡ ⎤ ⎡ γ Σγ T γ Σγ T ⋯ γ1 ΣγkT ⎤
⎢ ⎥ ⎢ 1 1 ⎥
∂f (P )
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ γ Σγ T γ Σγ T T ⎥
∂p1 1 2
⎢ ⎥ ⎢ 2 1 ⋯ γ2 Σγk ⎥
∂f (P )
⎢ ⎥=⎢ ⎥P+
2
⎢ ⎥ ⎢ ⎥
∂p2 2
⎢ ⋮ ⎥ ⎢ ⋮ ⋮ ⋱ ⋮ ⎥
⎢ ⎥ ⎢ T ⎥
⎢ ⎥ ⎢ γk Σγ1T γk Σγ2T ⋯ γk Σγk ⎥
⎣ ⎦ ⎣ ⎦
∂f (P )
∂pk
⎡ γ Σγ T γ Σγ T ⋯ γk Σγ1T ⎤
(3.25)
⎢ 1 1 ⎥
⎢ ⎥
⎢ γ Σγ T γ Σγ T T ⎥
2 1
⎢ 1 2 ⋯ γk Σγ2 ⎥
⎢ ⎥ P = (∇f (P ))T
2
⎢ ⎥
2
⎢ ⋮ ⋮ ⋱ ⋮ ⎥
⎢ T ⎥
⎢ γ1 Σγ T γ2 Σγ T ⋯ γk Σγk ⎥
⎣ k k ⎦
k
g(P ) = sumrow (W T ) − 1 = ∑ sumrow (pi γi T ) − 1 (3.27)
i=1
∂g(P )
= sumrow (γi T ) (3.28)
∂pi
47
3. Parity equations-based unknown input reconstruction for linear stochastic systems
and
ψ = sumrow (Γ T ) (3.31)
the solution to the Lagrange optimisation problem (3.22) can be rewritten as:
SP = λψ (3.32)
P = λS −1 ψ (3.33)
PTψ − 1 = 0 (3.34)
λ (S −1 ψ) ψ − 1 = 0
T
(3.35)
λ = ((S −1 ψ) ψ)
T −1
(3.36)
Consequently, the algorithm for calculating the optimal vector W and estimation of
the unknown input is summarised as follows:
1. Select the order of the parity space s ≥ n and build matrices Γ, Q, T , and Ξ.
2. Obtain Γ .
3. Compute Σ as:
48
3. Parity equations-based unknown input reconstruction for linear stochastic systems
λ = ((S −1 ψ) ψ)
T −1
(3.37d)
P = λS −1 ψ (3.37e)
W = P T Γ (3.37f)
∑ αi i
τ = round ( ) (3.37g)
∑ αi
where:
W T = [ αs αs−1 ⋯ α0 ]
T
It should be noted that, due to the fact that the arg min f (P ) needs to be found,
cf. (3.19), the function f (P ) can be scaled by an arbitrary number. Therefore, the
covariance matrices of ũ(t) and e(t) do not require to be known explicitly. It is sufficient
to know only the ratio between the variances of the noise sequences and scale Σũ and
Σe accordingly.
49
3. Parity equations-based unknown input reconstruction for linear stochastic systems
Single-output OE
Σ = Σe = var(e(t))I (3.38)
Since the objective is to find the minimum of the cost function f (P ), it can be scaled
by any arbitrary number. Therefore, for sake of simplicity, the term var(e(t)) can be
omitted. Consequently, the cost function f (P ) becomes, cf. (3.19):
k ⎛k ⎞
f (P ) = (∑ pi γi ) ∑ pj γjT
⎝j=1 ⎠
(3.39)
i=1
One can select Γ such that its rows are orthonormal, i.e. Γ (Γ ) = I. Therefore, the
T
∂f (P )
= 2pi (3.41)
∂pi
Consequently:
(∇f (P ))T = 2P (3.42)
Incorporating (3.29) and (3.42) into (3.22) the solution to the Lagrange optimisation
problem is calculated as:
P = λsumrow (Γ ) (3.43)
50
3. Parity equations-based unknown input reconstruction for linear stochastic systems
1. Select the order of the parity space s ≥ n and build matrices Γ, Q, T , and Ξ.
Multiple output OE
Consider the system (3.1), where m > 1, ũ(t) = 0, Π = 0 and Ω is diagonal. In such
a case Ξ and consequently Σ, cf. (3.20), are diagonal matrices. Assume that Γ is
selected, such that its rows are orthonormal vectors, i.e. Γ (Γ ) = I. Then the cost
T
k
f (P ) = ∑ βi p2i (3.47)
i=1
where βi denotes the ith element of the diagonal of Σ. Therefore, the partial derivative
of f (P ) is calculated as:
∂f (P )
= 2βi pi (3.48)
∂pi
Consequently, the gradient of f (P ) is:
Therefore, the Lagrange optimisation problem can be reformulated as, cf. (3.43):
51
3. Parity equations-based unknown input reconstruction for linear stochastic systems
1. Select the order of the parity space s ≥ n and build matrices Γ, Q, T , and Ξ.
Denote y(t), u0 (t), and v(t) in the z-domain, respectively, as Y (z), U0 (z), and V (z).
Consequently, equation (3.4) in the noise-free case can be reformulated as the following
relation:
W (z)Y (z) = WQ (z)U0 (z) + WT (z)V (z) (3.55)
where terms W (z), WQ (z) and WT (z) are appropriate polynomial vectors of the z-
variable with parameters defined by vectors W , W Q, and W T , respectively. Therefore,
in the noise-free case, the relationship between the unknown input and its estimate in
52
3. Parity equations-based unknown input reconstruction for linear stochastic systems
In the case when noise is present in the system, equation (3.56) becomes:
where WΞ (z) and E(z) refer to, respectively, the appropriate polynomial vector of the
z-variable with parameters defined by the vector W Ξ and the variable e(t) in z-domain,
whilst Ũ (z) denotes the z-domain representation of ũ(t).
In the case when p = m = 1 the transfer functions corresponding to u0 (t) and v(t)
are given, respectively, by:
WQ (z)
Gu (z) =
W (z)
WT (z)
(3.58)
Gv (z) =
W (z)
where Gu (z) defines the relationship between U0 (z) and the output, whereas Gv (z)
describes the relationship between V (z) and Y (z), cf. (3.54). In the case when s = n,
the left nullspace of Γ is a row vector Γ = W (it is assumed here that the system (3.1)
is observable) and the degree of the polynomial W (z) is equal to the order of the
system. Hence, one can deduce from (3.58) that the roots of the polynomial W (z) are
eigenvalues of the matrix A (i.e. poles of both Gv (z) and Gu (z)). Denote the set of
poles and zeros of Gv (z) by Pv and Zv , respectively. Analogously, refer to Pu and Zu
as, respectively, poles and zeros of Gu (z). Then, it is true that the roots of W (z) are
Pv ∪ Pu , the roots of WQ (z) are defined by the set Zu ∪ (Pv /Pu ), whilst roots of WT (z)
are Zv ∪ (Pu /Pv ).
If the order of the parity space is higher than that of the system, i.e. s > n, then
the set of equations (3.58) must still be fulfilled. This means, that W (z), WQ (z) and
WT (z) have common s − n roots (a zero-pole cancellation occurs, hence both
WQ (z)
W (z)
WT (z)
and W (z) remain unaltered). The choice of those additional s − n zeros influences
the properties of the noise filtration of the filter (3.13). Hence, the problem of finding
the optimal vector W can be reformulated as a filter zeros assignment problem. The
unknown input reconstruction is possible when the bandwidth of the unknown input is
narrower than that of WT (z), whilst the ability of the PE-UIO to filter ũ(t) and e(t)
depends on the frequency response of both, i.e. WQ (z) and W (z).
53
3. Parity equations-based unknown input reconstruction for linear stochastic systems
z − z0
Gv (z) = G′v (z) (3.59)
z
54
3. Parity equations-based unknown input reconstruction for linear stochastic systems
where H ′ and G′ are the appropriately modified matrices H and G, respectively. The
matrix T ′ is calculated by replacing G and H in T by, respectively, G′ and H ′ . Subse-
quently, (3.5) can be reformulated as:
v̂ ′ (t − τ ) = W T ′ V ′ (t) (3.65)
Subsequently, the unknown input estimate can be calculated via, cf. (3.61):
Note that this scheme is applicable only to systems with ∣z0 ∣ ≤ 1. Otherwise, (3.66)
becomes unstable.
In the noisy case the term v̂ ′ (t) is given by:
where ǫ(t) accounts for the disturbance introduced by e(t) and ũ(t), i.e.:
Hence, it follows from equations (3.66) and (3.67), that the estimate of v(t − τ ) is
affected by the error term ǫ∗ (t), whose relation to ǫ(t) is given by:
where terms E ∗ (t) and Ũ ∗ (t) are built from the current and previous values of e∗ (t)
and ũ∗ (t), respectively, cf. (3.2). It is required to minimise the variance of the term
55
3. Parity equations-based unknown input reconstruction for linear stochastic systems
where Σũ∗ = E{Ũ ∗ (t)(Ũ ∗ (t))T }, Σe∗ = E{E ∗ (t)E ∗ T (t)} and Σũ∗ e∗ = E{Ũ ∗ (t)E ∗ T (t)} =
0. Hence, the function to be minimised is given by:
In order to calculate (3.73), first, the terms Σe∗ and Σũ∗ need to be obtained. The
signal e∗ (t) can be described by a function of its previous values, cf. (3.70):
Therefore, by recalling that e(t) is assumed white, the expected value of e∗ (t)e∗ (t − i)
is calculated as:
which is a sum of a geometric series and in the case when ∣z0 ∣ < 1 it can be simplified
to:
z0i
E{e∗ (t)e∗ (t − i)} = E{e2 (t)} (3.76)
1 − z02
Analogously, by recalling that ũ(t) is assumed white, the expected value of ũ∗ (t)ũ∗ (t−1)
can be derived as:
z0i
E{ũ∗ (t)(ũ∗ (t − i))T } = E{ũ(t)ũT (t)} (3.77)
1 − z02
(Note that e(t) is a scalar, whilst ũ(t) is, in general, a vector.) In the case when ∣z∣ = 1
the sum of the geometric series (3.75) is infinite. Therefore, to cope with such a case
it is proposed to replace z0 in (3.76) and (3.77) by a value smaller than unity in order
to indicate that e∗ (t) and ũ∗ (t) are not white.
The matrices Σe∗ and Σũ∗ are built by filling their entries by the appropriate values
of, respectively, E{e∗ (t)e∗ (t − i)} and E{ũ∗ (t)ũ∗ (t − i)}. For convenience, a new term
is introduced, cf. (3.20):
Σ∗ = ΞΣe∗ ΞT + QΣũ∗ QT (3.78)
56
3. Parity equations-based unknown input reconstruction for linear stochastic systems
s−n+1 ⎛s−n+1 ⎞
f (P ) = ( ∑ pi γiT ) Σ∗ ∑ pj γ j
⎝ j=1 ⎠
(3.79)
i=1
The solution to this constrained optimisation problem is solved analogously to the one in
Section 3.3. Therefore, the algorithm for calculating the unknown input is summarised
as follows:
1. Select the order of the parity space s ≥ n and build matrices Γ, Q, T ′ , and Ξ.
2. Obtain Γ .
λ = ((S −1 ψ) ψ)
T −1
(3.81c)
P = λS −1 ψ (3.81d)
W = P T Γ (3.81e)
57
3. Parity equations-based unknown input reconstruction for linear stochastic systems
∑ αi′ i
τ = round ( ) , for i = 0, ⋯, s (3.81f)
∑ αi′
where the coefficients of the polynomial WT ′ (z) are appropriate elements of the vector
W T ′ . Consequently, the unknown input estimate in the z-domain is, see (3.57):
It can be deduced that the use of the two stage PE-UIO is advisable if z0 is a positive
real number lower or equal unity. Firstly, if the single stage PE-UIO is used in such
a case, the presence of z0 in WT (z) will cause an overshoot in the step response of
z
the input estimation filter, which may me undesirable. Secondly, the factor z−z0 , for
0 ≤ z0 ≤ 1, reduces the bandwidth of the noise affecting the input estimate, cf. (3.83).
On the other hand the use of the single stage PE-UIO may be preferred over its two
stage version if z0 > 0 is relatively close to zero, and the phase lead caused by the
presence of z0 is desirable, e.g. in an on-line application, when the fast response of the
filter is required. It is not recommended to use the two stage PE-UIO in noisy systems
z
when z0 is lower than zero, due to highpass properties of z−z0 , which would cause an
amplification of noise effect on the unknown input estimate.
58
3. Parity equations-based unknown input reconstruction for linear stochastic systems
Consequently:
Gv (z)V (z) = Mv (z)V ′ (z) (3.87)
where H ′ and G′ are the appropriately modified matrices H and G, respectively. The
matrix T ′ is calculated by replacing G and H in T by, respectively, G′ and H ′ . Subse-
quently, in the first stage of the algorithm, the term V ′ (z) is estimated as:
The unknown input is calculated in the second stage of the algorithm as:
59
3. Parity equations-based unknown input reconstruction for linear stochastic systems
V̂ (z) = WT′ (z)V (z) + W (z)Ξ Nv−1 (z)E(z) − WQ (z)Nv−1 (z)Ũ (z) (3.93)
The term E ∗ (z) is denoted in time domain as e∗ (t) and, subsequently, the stacked
vector of e∗ (t) is E ∗ (t). Analogous notation is used for Nv−1 Ũ (z), i.e. Ũ ∗ (z), ũ∗ (t),
and Ũ ∗ (t).
Analogously to (3.72), the variance of the following term must be minimised:
where Σũ∗ = E{Ũ ∗ (t)(Ũ ∗ (t))T }, Σe∗ = E{E ∗ (t)E ∗ T (t)} and Σũ∗ e∗ = E{Ũ ∗ (t)E ∗ T (t)} =
0. Hence, the function to be minimised is given by:
f (W ) = W Σ∗ W T (3.96)
where:
Σ∗ = ΞΣe∗ ΞT + QΣũ∗ QT (3.97)
The covariance matrices Σe∗ and Σũ∗ depend on zi , i = 1, ⋯, k and variances of e(t) and
ũ(t) and should be calculated for each case individually. Finally, the generalised two
stage PE-UIO is summarised as follows:
3. Select the order of the parity space s ≥ n and build matrices Γ, Q, T ′ , and Ξ.
4. Obtain Γ .
60
3. Parity equations-based unknown input reconstruction for linear stochastic systems
λ = ((S −1 ψ) ψ)
T −1
(3.98c)
P = λS −1 ψ (3.98d)
where z −1 is a backwards shift operator, i.e. z −1 y(t) = y(t − 1), G(z −1 ) denotes the
transfer function between any of the system input (either v(t) or u(t)) and the output,
61
3. Parity equations-based unknown input reconstruction for linear stochastic systems
bn + bn−1 z −1 + ⋯ + b0 z −n
G(z −1 ) = (3.100)
an + an−1 z −1 + ⋯ + a0 z −n
bn z n + bn−1 z n−1 + ⋯ + b0
G(z) = (3.101)
an zn + an−1 z n−1 + ⋯ + a0
3.7.1 PE-UIO
This subsection presents a step-by-step design of the PE-UIO algorithm for two different
cases, namely, an OE and an ARMAX case in the EIV framework. The importance of
the tuning parameter s on the frequency response of the filter is also explained.
z + 0.01
Gu (z) =
(z − 0.9)(z − 0.85)
(z + 1.95)(z − 0.2)
(3.102)
Gv (z) =
(z − 0.9)(z − 0.85)
⎡ ⎤ ⎡ ⎤ ⎡ ⎤
⎢ 1.750 1 ⎥ ⎢ 1.00 ⎥ ⎢ 3.500 ⎥
⎢
A=⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎥ B=⎢ ⎥ G=⎢ ⎥
⎢ −0.765 0 ⎥ ⎢ 0.01 ⎥ ⎢ −1.155 ⎥
⎣ ⎦ ⎣ ⎦ ⎣ ⎦ (3.103)
C =[ 1 0 ] D=0 H =1
It is assumed that the output of the system is affected by a white, zero-mean, Gaussian
noise sequence of the variance var(e(t)) = 1. Since the considered case is single-output
OE, Algorithm 3.3 is used for the unknown input reconstruction.
The parity space order is chosen to be s = 4, hence the extended observability matrix
⎡ ⎤
is:
⎢ ⎥
⎢ ⎥
1.0000 0
⎢ ⎥
⎢ ⎥
⎢ ⎥
1.7500 1
Γ=⎢
⎢ 2.2975 1.7500 ⎥
⎥
⎢ ⎥
(3.104)
⎢ 2.6819 2.2975 ⎥
⎢ ⎥
⎢ ⎥
⎢ 2.9357 2.6819 ⎥
⎣ ⎦
62
3. Parity equations-based unknown input reconstruction for linear stochastic systems
⎡ 0 0 ⎤
⎢ ⎥
⎢ ⎥
0 0 0
⎢ 0 0 ⎥
⎢ ⎥
⎢ ⎥
1 0 0
Q=⎢
⎢ 0 0 ⎥
⎥
⎢ ⎥
1.7600 1 0 (3.106)
⎢ 0 0 ⎥
⎢ 2.3150 1.7600 1 ⎥
⎢ ⎥
⎢ 2.7049 2.3150 1.7600 1 0 ⎥
⎣ ⎦
⎡ −0.1608 ⎤
⎢ ⎥
⎢ ⎥
P =⎢ ⎥
⎢ −0.1199 ⎥
⎢ ⎥
(3.108)
⎢ 0.0083 ⎥
⎣ ⎦
63
3. Parity equations-based unknown input reconstruction for linear stochastic systems
Example 3.2. Design of the PE-UIO for an ARMAX system in the EIV
framework
Consider an ARMAX system, whose A, B, C, D, G, and H matrices are the same as
in Example 3.1. The moving average coloured noise parameters are c0 = 1, c1 = 0.3,
and c2 = 0.1. It is assumed that the input measurement is affected by a white, zero-
mean, Gaussian noise ũ(t) of the variance equal to the variance of e(t). Thus the noise
distribution matrix is given by:
⎡ ⎤
⎢ 2.050 ⎥
⎢
Π=⎢ ⎥
⎥
⎢ −0.665 ⎥
(3.112)
⎣ ⎦
⎡ 0 ⎤
⎢ ⎥
⎢ ⎥
1 0 0 0
⎢ 0 ⎥
⎢ ⎥
⎢ ⎥
2.0500 1 0 0
Ξ=⎢
⎢ 0 ⎥
⎥
⎢ ⎥
2.9225 2.0500 1 0 (3.113)
⎢ 0 ⎥
⎢ 3.5461 2.9225 2.0500 1 ⎥
⎢ ⎥
⎢ 3.5461 2.9225 2.05 1 ⎥
⎣ 3.9700 ⎦
Note that the variances of ũ(t) and e(t) are unknown, however for the purpose of
finding the optimal filter parameters only the ratio between those variances is needed,
which is equal to one. Substituting unity for the variances of both ũ(t) and e(t), the
term Σ is calculated using (3.37a):
⎡ 3.9700 ⎤
⎢ ⎥
⎢ ⎥
1.0000 2.0500 2.9225 3.5461
⎢ 14.3895 ⎥
⎢ ⎥
⎢ ⎥
2.0500 6.2025 9.8011 12.5071
Σ=⎢
⎢ 28.8699 ⎥
⎥
⎢ ⎥
2.9225 9.8011 17.8411 24.2391 (3.114)
⎢ 44.5789 ⎥
⎢ 3.5461 12.5071 24.2391 35.7753 ⎥
⎢ ⎥
⎢ 58.8525 ⎥
⎣ 3.9700 14.3895 28.8699 44.5789 ⎦
⎡ −3.9894 ⎤
⎢ ⎥
⎢ ⎥
ψ=⎢
⎢ −2.9748 ⎥
⎥
⎢ ⎥
(3.116)
⎢ 0.2060 ⎥
⎣ ⎦
64
3. Parity equations-based unknown input reconstruction for linear stochastic systems
Using (3.37d), the Lagrange multiplier is calculated as λ = 3.1271 and the parameter
vector P is computed from (3.37e) as:
⎡ −0.0477 ⎤
⎢ ⎥
⎢ ⎥
⎢
P = ⎢ −0.2749 ⎥
⎥
⎢ ⎥
(3.117)
⎢ −0.0383 ⎥
⎣ ⎦
The estimation lag is equal to 1. Therefore, the unknown input estimate v̂(t − 1) is
calculated via:
The order of the parity space s is a tuning parameter of the PE-UIO algorithm. It is
anticipated that an increase of s will lead to a reduction of the impact of disturbances
on the unknown input estimate. At the same time it is expected that an increase
of the order of parity space will yield a reduction of the filter bandwidth, which will
result in the input reconstruction filter being sluggish. This phenomenon can be seen
in Fig. 3.2, where the frequency responses of the polynomial filters WT (z), WQ (z), and
W (z), cf. (3.55), for three different cases of s are compared. This effect is also visible
in Fig. 3.3, where the reconstructed input signals are compared for different values of
parity space orders. The system from Example 3.1 is considered in this experiment.
Whilst for s = 4 the unknown input estimate is noisy (i.e. the noise filtering is rather
poor in this case), for s = 15 the filter does not reproduce high frequency oscillations
of the input. The PE-UIO with s = 7 seems to be the optimal setting for the given
example.
65
3. Parity equations-based unknown input reconstruction for linear stochastic systems
W (z)
0
s=4
−20 s=7
s=15
−40
−60 −1 0
10 10
WQ (z)
0
Magnitude [dB]
−20
−40
−60 −1 0
10 10
WT (z)
0
−20
−40
−60 −1 0
10 10
Frequency [rad/s]
Figure 3.2: Frequency responses of W (z), WQ (z) and WT (z) for different values of
the parity space order s
15 true
Unknown input estimate
s=4
s=7
10
s=15
5
−5
−10
−15
4750 4800 4850 4900 4950
Time [samples]
z − 0.1
Gu (z) =
(z − 0.9)(z − 0.8)
(z + 1.2)(z − 0.95)
(3.120)
Gv (z) =
(z − 0.9)(z − 0.8)
66
3. Parity equations-based unknown input reconstruction for linear stochastic systems
⎡ ⎤ ⎡ ⎤ ⎡ ⎤
⎢ 1.70 1 ⎥ ⎢ 1 ⎥ ⎢ 1.95 ⎥
⎢
A=⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎥ B=⎢ ⎥ G=⎢ ⎥
⎢ −0.72 0 ⎥ ⎢ −0.1 ⎥ ⎢ −1.86 ⎥
⎣ ⎦ ⎣ ⎦ ⎣ ⎦ (3.121)
C =[ 1 0 ] D=0 H =1
It is assumed that the output of the system is subjected to white, zero-mean, Gaussian
measurement noise (OE case) of the variance var(e(t)) = 2.7. After elimination of the
zero at 0.95, the corresponding modified matrices H ′ and G′ are built such that H ′ = 1
⎡ ⎤
and:
⎢ 2.90 ⎥
⎢
G =⎢ ⎥
⎥
′
⎢ −0.72 ⎥
(3.122)
⎣ ⎦
The covariance of e∗ (t) is calculated as, cf. (3.76):
0.95i
E{e∗ (t)e∗ (t − i)} = var(e(t)) (3.123)
1 − 0.952
Consequently:
Fig. 3.4 compares step responses of WT ′ (t) for the two stage PE-UIO with s = 5
and WT (t) for two cases of the standard PE-UIO (s = 5 and s = 15). It can be noted
that the system zero at 0.95 causes a large overshoot in the case when the standard
PE-UIO is used (for both s = 5 and s = 15). This results in a significant distortion of the
unknown input estimate, which can be seen in Fig. 3.5, where the time-domain result of
67
3. Parity equations-based unknown input reconstruction for linear stochastic systems
0
0 5 10 15
Time [samples]
Figure 3.4: Step responses of WT ′ (t) for the two stage PE-UIO and WT (t) for two
cases of the standard PE-UIO
true
60
standard PE−UIO, s=5
Unknown input estimate
20
−20
Figure 3.5: Comparison of the unknown input estimate using the two stage PE-UIO
and the standard PE-UIO, z0 = 0.95. Distortion of the unknown input
estimate caused by the phase lead can be seen in the case of the standard
PE-UIO.
the unknown input reconstruction in a noise-free case is shown. The advantage of the
two stage PE-UIO in this particular case can be also seen in Fig. 3.6, where frequency
responses of W (z), WQ (z), WT (t), and WT ′ (t) are presented.
Example 3.4. Comparison of the standard and the two stage PE-UIO in an
on-line application
In this example a situation is presented, where the phase lead introduced by the PE-
UIO is advantageous. Consider an on-line application, where an estimation delay is
crucial for the system performance, e.g. where the reconstructed input is utilised by
a feedback controller. Bearing in mind that at the time instance t the delayed input
68
3. Parity equations-based unknown input reconstruction for linear stochastic systems
W (z)
−20
standard PE−UIO s=5
−40 standard PE−UIO s=15
two−stage PE−UIO s=5
−60 −1 0
10 10
WQ (z)
20
Magnitude [dB]
0
−20
−40
−60 −1 0
10 10
WT (z) and WT′ (z)
20
0
−20
−40
−60 −1 0
10 10
Frequency [rad/s]
Figure 3.6: Frequency responses of WT ′ (t) for the two stage PE-UIO and WT (t) for
two cases of the standard PE-UIO
(In contrary, in an off-line situation or when the estimation delay is not crucial the
difference between v̂(t − τ ) and v(t − τ ) is taken into consideration.)
The system used in this example is given by the equation:
(z − 0.3)(z + 1.8)
Gv (z) =
(z − 0.8)(z − 0.9)
(3.128)
The zero at z0 = 0.3 causes a phase lead (and consequently an overshoot of the step
response) of WT (z) when the standard PE-UIO is used. However, it is expected that
the zero at 0.3 will reduce the impact of the estimation lag caused by the zero at −1.8.
(Due for the fact that 0.3 lies relatively far from unity, it is not expected to cause
such as damaging distortion in the step response of WT (z) as shown in Example 3.3.)
Consequently, a faster response is anticipated when using the PE-UIO instead of the two
stage PE-UIO with z0 = 0.3, which may be particularly desired in on-line applications.
In Fig. 3.7, for completeness, step responses of WT (z), for the standard PE-UIO,
69
3. Parity equations-based unknown input reconstruction for linear stochastic systems
and WT ′ (z), for the two stage PE-UIO, are compared (the parity space order is in both
cases s = 2).
2
standard PE−UIO
two−stage PE−UIO
0.5
0
0 0.5 1 1.5 2 2.5 3
Time [samples]
Sample time responses of the unknown input estimates using the two algorithms
are presented in Fig. 3.8. It is assumed that the output of the system is subjected
2
5300 5305 5310 5315 5320 5325 5330 5335 5340
Time [samples]
Figure 3.8: On-line unknown input estimation using the standard PE-UIO and the
two stage PE-UIO. In both cases order of parity space s = 2, variance of
noise var(e(t)) = 14e-4.
to low level OE noise (var(e(t)) = 14e-4). It can be noted that the input estimate,
when using the standard PE-UIO, yields a smaller estimation delay compared to the
70
3. Parity equations-based unknown input reconstruction for linear stochastic systems
two stage PE-UIO (τ = 0 in the case of the PE-UIO and τ = 1 for the two stage
PE-UIO), which is due to the fact that, when the standard PE-UIO is used, the lead
caused by the zero at 0.3 partially compensates for the lag caused by the zero at −1.8.
Nevertheless, the two stage PE-UIO has superior noise filtering properties (in terms
of the bandwidth of W (z)), what can be observed in Fig. 3.9. The efficacy of the
W (z)
10
standard PE−UIO
0
two−stage PE−UIO
−10
−20
−30
−40 −1 0
10 10
WT (z)
5
0
−5
−10
−15 −1 0
10 10
Frequency [rad/s]
two algorithms for different levels of noise and different orders of the parity space s
are compared in Table 3.1. It can be noted that for relatively low levels of noise the
standard PE-UIO preforms better in terms of the on-line input estimation error (3.127)
than the two stage PE-UIO, due to the lag compensation. However, as the noise level
increases, the on-line input estimation error variance increases more slowly when the
two stage PE-UIO is utilised, which is due to superior noise filtering properties of the
two stage PE-UIO. As the order of the parity space, i.e. s, is increased, the level of the
OE noise, for which both the standard PE-UIO and the two stage PE-UIO perform the
same in terms of the variance of the on-line input estimation error, is also increased.
71
3. Parity equations-based unknown input reconstruction for linear stochastic systems
Table 3.1: Comparison the standard PE-UIO and the two stage PE-UIO in terms
of the variance of the on-line unknown input reconstruction error, i.e.
ǫon−line (t), for different levels of noise and different orders of the parity
space. The term ‘% std dev’ refers to the percentage value of the ratio
between the output meausrement noise e(t) and the system output y0 (t)
in terms of the standard deviation.
namely, the PE-UIO the MVU is compared. The following single-input two-output
ARX model is considered:
⎡ ⎤ ⎡ ⎤
⎢ ⎥ ⎢ ⎥
y(t) = ⎢ ⎥ y(t − 1) + ⎢ 0.76 0 ⎥ y(t − 2)+
−1.75 0
⎢ ⎥ ⎢ ⎥
⎢ −1.75 ⎥ ⎢ 0 0.765 ⎥
⎣ ⎦ ⎣ ⎦
0
⎡ ⎤ ⎡ ⎤ ⎡ ⎤ (3.129)
⎢ 1⎥ ⎢ ⎥ ⎢ ⎥
⎢ ⎥ v(t) + ⎢ 1.3 ⎥ v(t − 1) + ⎢ 2.4 0 ⎥ e(t − 1)
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ 1⎥ ⎢ −0.3 ⎥ ⎢ 0 0.7 ⎥
⎣ ⎦ ⎣ ⎦ ⎣ ⎦
which corresponds to the system (3.1), whose matrices are given by:
⎡ 1.75 1 0 ⎤ ⎡ 3.05 ⎤ ⎡ ⎤
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
0 2.4 0
⎢ 0 ⎥ ⎢ ⎥ ⎢ 0.7 ⎥
⎢ 0 1 ⎥ ⎢ 1.45 ⎥ ⎢ ⎥
A=⎢ ⎥ G=⎢ ⎥ Π=⎢ ⎥
1.75 0
⎢ −0.76 ⎥ ⎢ ⎥ ⎢ 0 ⎥
⎢ 0 0 ⎥ ⎢ −0.76 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
0 0
⎢ 0 ⎥ ⎢ ⎥ ⎢ 0 ⎥
⎣ 0 0 ⎦ (3.130)
−0.765 ⎣ −0.765 ⎦ ⎣ 0 ⎦
⎡ ⎤ ⎡ ⎤ ⎡ ⎤
⎢ 1 0 0 0 ⎥ ⎢ 1 ⎥ ⎢ 0 0 ⎥
⎢
C =⎢ ⎥ H =⎢ ⎥ Ω=⎢ ⎥
⎥ ⎢ ⎥ ⎢ ⎥
⎢ 0 1 0 0 ⎥ ⎢ 1 ⎥ ⎢ 0 0 ⎥
⎣ ⎦ ⎣ ⎦ ⎣ ⎦
Note, that the model (3.129) corresponds to the model (2.39), whose A, B, G, and H
matrices are as in (3.130), whilst ζ(t) = 0 and:
⎡ ⎤
⎢ ⎥
⎢ ⎥
2.4 0
⎢ 0.7 ⎥
⎢ ⎥
ξ(t) = ⎢ ⎥ e(t)
0
⎢ 0 ⎥
(3.131)
⎢ ⎥
⎢ ⎥
0
⎢ 0 ⎥
⎣ 0 ⎦
72
3. Parity equations-based unknown input reconstruction for linear stochastic systems
Consequently, the noise covariance matrices used by the MVU are R = 0 and:
⎡ 0 0 ⎤
⎢ ⎥
⎢ ⎥
5.76 0
⎢ 0.49 0 0 ⎥
⎢ ⎥
Q̃ = var(e(t)) ⎢ ⎥
0
⎢ 0 0 ⎥
(3.132)
⎢ ⎥
⎢ ⎥
0 0
⎢ 0 0 ⎥
⎣ 0 0 ⎦
The algorithms are compared for tree different levels of var(e(t)), namely, 1, 8, and
0.001, which correspond to, respectively, 4.8 %, 13.6 %, and 0.152 % of noise on each
output by means of the standard deviation. In the experiment the MVU is compared
with the PE-UIO designed with different values of s. A Monte-Carlo simulation with
100 runs is carried-out in order to provide reliable results, which are presented in Ta-
ble 3.2. The MVU ensures the minimum variance of the estimation error resulting from
var(e(t)) 1 8 0.001
s τ R2T [%] τ R2T [%] τ R2T [%]
2 0 2.2153 0 17.6747 0 0.0091202
3 1 1.7598 1 9.4901 1 0.6559958
4 1 0.9371 1 6.0884 1 0.2019992
5 2 1.5533 2 5.4152 2 1.0017866
6 2 1.5191 2 4.6083 2 1.0776268
7 3 2.4092 3 4.9843 3 2.0409051
MVU 0 1.9986 0 15.7028 0 0.0019627
the disturbances. Therefore, achieving lower R2T than that of MVU and ensuring at the
same time τ = 0 is not feasible, what can be seen in the simulation results. However, the
major advantage of the PE-UIO is the ability to adjust the filter bandwidth by selecting
the tuning parameter s. By choice of an optimal s, the R2T is reduced approximately
2 and 4 times for, respectively, var(e(t)) = 1 and var(e(t)) = 8 compared to the MVU.
The results show that for a low level of noise (var(e(t)) = 0.001) the MVU performs
better than the PE-UIO. This is due to the fact that the PE-UIO provides an estimate
of the unknown input, cf. (3.10) and (3.11).
73
3. Parity equations-based unknown input reconstruction for linear stochastic systems
The output of the system is subjected to white, zero-mean, Gaussian noise of unity
variance. Note that in the SISO case, the MVU resembles a naive inversion, cf. Re-
mark 2.1. Results of 100-run Monte-Carlo simulation are presented in Table 3.3, whilst
samples of the estimated input signals are presented in Fig. 3.10. It can be noted
Error
2
0
−2
1.13 1.132 1.134 1.136 1.138 1.14 1.142 1.144 1.146 1.148 1.15 1.152
4
Unknown input estimate x 10
true input
10 MVU
PE−UIO
2 stage PE−UIO
INPEST
−5
−10
1.13 1.132 1.134 1.136 1.138 1.14 1.142 1.144 1.146 1.148 1.15
Time [samples] x 10
4
that the standard PE-UIO, the two stage PE-UIO, and the INPEST provide compa-
rable results, whereas the MVU seems to give inferior results in terms of R2T . This is
due to the relatively high bandwidth of the MVU, which results in the lowest possible
estimation lag (in this case τ = 0). The other examined algorithms can be tuned to
reduce the reconstruction filter bandwidth (by increasing ν in the case of the INPEST
method and s in the case of the standard and the two stage PE-UIO), which also yields
an inherent estimation delay (τ > 0).
Example 3.7. Comparison of the INPEST and the two stage PE-UIO
In this example the efficacy of the two stage PE-UIO and the INPEST is compared in
the case when the output response to unknown input contains a zero close to unity.
74
3. Parity equations-based unknown input reconstruction for linear stochastic systems
Table 3.3: Results of comparison of various input reconstruction methods. Noise vari-
ance var(e(t)) = 1
The system has two zeros at: −4.3023 and 0.9735. Note, that (3.134) is non-minimum
phase, however both the two stage PE-UIO and the INPEST methods can cope with the
zero at −4.3023. The zero at z0 = 0.9735 needs to be eliminated from the parity equation
in the two stage PE-UIO. The output of the system is subjected to white, zero-mean,
Gaussian noise of the variance 0.003, which means that the standard deviation of the
output measurement noise is equal to approximately 6% of the standard deviation of
the output. The INPEST method is optimised for ν = 0.004, which results in q̊e = 0.2635
and τ̊ = 7. The results of the input reconstruction are presented in Fig. 3.11, whereas
the efficacy in terms of R2T and τ is compared in Table 3.4. It can be noted that both
Table 3.4: Comparison of INPEST and two stage PE-UIO. System has zero at 0.9735
75
3. Parity equations-based unknown input reconstruction for linear stochastic systems
Error
0.4
0.2
0
−0.2
−0.4
0 50 100 150 200 250 300 350 400
Unknown input estimate
4.5 true
4 PE−UIO
INPEST
3.5
3
2.5
2
1.5
1
0.5
0
0 50 100 150 200 250 300 350 400
Time [samples]
Figure 3.11: Comparison of unknown input reconstruction efficacy of the two stage
PE-UIO and the INPEST; in both cases τ = 7. The system has zeros at
−4.3023 and 0.9735.
methods, namely, the Kalman filter-based MVU and the INPEST method which is
based on a closed loop control concept.
The main advantage of the PE-UIO is its simplicity; the filter parameters are cal-
culated once at the beginning of the reconstruction process. The method is fast as it
utilises two moving average filters. The only tuning parameter of the PE-UIO is the
order of the parity space s. By altering it, the bandwidth of the input reconstruction
filter is shaped. This property allows the designer to tune the algorithm for different
levels of noise. It should be noted that by reduction of the filter bandwidth (hence
improvement of the noise filtering properties of the scheme) an estimation lag is intro-
duced. Similar property has been observed in the INPEST method, whereas the MVU
does not allow for introduction of an estimation lag in order to reduce the impact of
the noise (in terms of a bandwidth reduction). Furthermore, the PE-UIO is suitable
for non-minimum phase systems.
The two versions of the PE-UIO algorithm have been compared. The two stage PE-
UIO allows to eliminate selected system zeros from the PE, which changes the response
of the input reconstruction filter. This is particularly desirable, when the system zeros
lie close to unity, which results in a large overshoot in the step response of the standard
PE-UIO. The design of the two stage PE-UIO allows the elimination of this overshoot
and hence a distortion of the unknown input estimate. Furthermore, the two stage
PE-UIO provides better (in terms of bandwidth) noise filtering properties. On the
76
3. Parity equations-based unknown input reconstruction for linear stochastic systems
other hand it has been shown using a numerical example that the phase lead caused
by the system zero when using the standard PE-UIO may be desirable. This might be
the case in an on-line application when the estimation delay is crucial. The phase lead,
when using the standard PE-UIO, result in a reduced estimation delay compared to
the two stage PE-UIO.
The comparison of both PE-UIO methods, the INPEST, and the MVU revealed
comparable efficacy of the PE-UIO and the INPEST. Both algorithms have the pos-
sibility to shape the filter bandwidth (by introducing an inherent delay) by selection
of a single tuning parameter (ν in the case of the INPEST and s in the case of the
PE-UIO). The MVU does not have such a possibility of shaping the bandwidth of the
filter to this extent as in the case of the PE-UIO or the INPEST.
Further work aims towards an extension of the algorithms to systems with multiple
unmeasurable inputs. Although the proposed algorithms are generally applicable for
nonminimum-phase systems, a solution for systems, whose nonmiminum-phase zero is
close to unity, still remains an open question.
77
Chapter 4
Nomenclature
78
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
79
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
Preliminary reading: Sections 2.2, 2.3, 2.4, Subsection 2.5.2, Sections 3.2 and 3.3.
4.1 Introduction
Block oriented models are convenient for modelling nonlinear systems. Their relatively
simple structure of a linear dynamic block interconnected with nonlinear memoryless
function(s) provides a powerful tool for an approximation of a large class of nonlin-
ear systems, see (Pearson & Pottmann 2000, Pearson 2003). Block oriented models
have been used for modelling such phenomena as, for instance: infant EEG (electroen-
cephalogram) seizures (Celka & Colditz 2002), a radio frequency amplifier (Crama &
Rolain 2002), a glucose-insulin process in diabetes type I patient (Bhattacharjee, Sen-
gupta & Sutradhar 2010), ionospheric dynamics (Palanthandalam-Madapusi, Ridley &
Bernstein 2005) or human operator dynamics (Tervo & Manninen 2010). Furthermore,
such models are also used for control purposes, see, for example, (Anbumani, Patnaik
& Sarma 1981, Fruzzetti, Palazoglu & McDonald 1997, De-Feng, Li & Guo-Shi 2010),
and fault detection (Korbicz, Koscielny, Kowalczuk & Cholewa 2003, Lajic, Blanke &
Nielsen 2009).
A two-input single-output Hammerstein-Wiener model is considered, i.e. the linear
dynamic block is preceded and followed by nonlinear static functions. (In the case of a
Hammerstein model a linear block is preceded by a static nonlinear function, whereas
in the case of a Wiener model the order of these elements is reversed.) A problem
of the reconstruction of the unknown/unmeasurable input to the system is taken into
consideration. Up to date, only a limited number of publications are available on this
subject. Szabo, Gaspar & Bokor (2005) proposed an inversion of Wiener systems using
a geometric method based on the assumption that the static nonlinearity transforming
the output is invertible, whilst Ibnkahla (2002) used neural networks for Hammerstein
system inversion.
The algorithm presented here extends the approach developed in Chapter 3 to a
Hammerstein-Wiener case. An EIV framework, see (Söderström 2007), is considered,
i.e. all the measured signals are affected by white, Gaussian, zero-mean and mutually
uncorrelated measurement noise sequences. The theory described in Sections 4.2–4.3
has been presented in (Sumislawska, Larkowski & Burnham 2012).
This chapter is organised as follows: in Section 4.2, for completeness, the idea of
80
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
block oriented models is presented and the Hammerstein-Wiener model, for which the
unknown input reconstruction algorithm is designed, is defined. Then, the PE-UIO
method for Hammerstein-Wiener systems (PE-UIO-HW) is described in Section 4.3.
Furthermore, in Section 4.4 the PE-UIO-HW is extended to an adaptive version, which
accounts for changes in noise levels. Finally, conclusions are provided in Section 4.5.
where ϕ(⋅) is a static nonlinearity transforming the first system input u0 (t) into an in-
accessible signal ū0 (t) which serves as the first input to the linear block. It is assumed
that the second input v(t) is fed directly (without a nonlinear transformation) to the
linear block, which is described by a state-space model, where A ∈ Rn×n , B ∈ Rn×1 ,
C ∈ R1×n , D ∈ R1×1 , G ∈ Rn×1 and H ∈ R1×1 . The term ȳ0 (t) refers to the output of
the linear part of the system, which is then transformed by the memoryless function
η(⋅) into the overall system output y0 (t). Since the EIV case is considered, all mea-
sured variables, which are u(t) and y(t), are affected by white, Gaussian, zero-mean,
and mutually uncorrelated measurement noise sequences denoted by ũ(t) and ỹ(t), re-
spectively. Noise sequences are postulated to be uncorrelated with the noise-free but
unmeasured system input and output, denoted as u0 (t) and y0 (t), respectively. It is
assumed here that η(⋅) is strictly monotonic, hence its inverse exists. Note that (4.1)
represents a Hammerstein or a Wiener model if, respectively, η(⋅) or ϕ(⋅) is an identity
function.
Similarly as in Chapter 3, the objective of the proposed scheme is to estimate the
unknown input v(t), simultaneously minimising the effect of the measurement noise on
the unknown input estimate. It is assumed that the model of the system is known and
that v(t) is varying relatively slowly, cf. Subsection 3.5.1.
81
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
v(t)
where ϕ(U0 (t)) is a vector whose elements are ϕ(u0 (t−s)), ϕ(u0 (t−s+1)), ⋯, ϕ(u0 (t)).
Analogously, the function η(Ȳ0 (t)) is defined.
The linear part of the system, defined by (4.2b), can be represented by the following
parity relation, cf. Subsection 3.3.1.
where η −1 (⋅) denotes an inverse of η(⋅). Due to the fact that y0 (t) and u0 (t) are
inaccessible, the parity relation (4.4) can be approximated by the measured values of
82
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
where ξ(t) accounts for an overall error resulting from the presence of measurement
noise. (Note that ξ(t) depends also on the current values of u(t) and y(t) due to the
nonlinearities in the system.) By rearranging the measured (known) variables to the
right-hand side and the unknowns to the left-hand side, the following parity equation
is obtained, cf. (Li & Shah 2002):
In the case of noisy input and output measurements the unknown input estimate is
affected by an error, cf. (4.6):
resulting from both the input and output measurement uncertainties, which can be
deduced to be given by:
83
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
and
ȳ˜(t) = η −1 (y(t)) − η −1 (y0 (t)) (4.14)
are white and mutually uncorrelated (as ũ(t) and ỹ(t) are white and mutually uncor-
˜(t), further
related), which is demonstrated further in this Section. The variance of ū
˜(t)), is time varying and depends on ϕ(u(t)), u(t) and the variance
referred to as var(ū
of ũ(t), denoted as var(ũ). Analogously, the variance of ȳ˜(t), i.e. var(ȳ˜(t)), is depen-
dent on var(ỹ) and the current values of η(y(t)) and y(t). The expression ϕ(u0 (t))
can be approximated using a first order Taylor expansion at u(t):
∂η −1 (y(t))
ȳ˜(t) ≈ ỹ(t) (4.17)
∂y(t)
Note that:
˜(t − i)ū
˜(t − j)} ≈ E{
∂ϕ(u(t − i)) ∂ϕ(u(t − j))
E{ū ũ(t − i) ũ(t − j)}
∂u(t) ∂u(t)
} × 0 = 0, for i ≠ j
∂ϕ(u(t − i)) ∂ϕ(u(t − j))
= E{
∂u(t) ∂u(t)
Hence, the sequence ū˜(t) is white. Analogously, it can be demonstrated that ȳ˜(t) is
˜(t) and ȳ˜(t) are mutually uncorrelated.
white as well as ū
˜(t) and ȳ˜(t) can be approximated, respectively, as:
The variances of ū
2
˜(t)) ≈ ( ) var(ũ)
∂ϕ(u(t))
var(ū
∂u(t)
(4.19)
∂η −1 (y(t))
2
var(ȳ˜(t)) ≈ ( ) var(ỹ)
∂y(t)
84
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
˜(t)) and var(ȳ˜(t)) is not constant, i.e. the impact of either input
ratio between var(ū
or output measurement noise on the unknown input estimation error can be prevailing,
depending on the system operating point. Therefore, the unknown input reconstruction
filter should adapt to these changes.
The aim of the PE-UIO for Hammerstein-Wiener systems is to select such a vector
W that the variance of the error term ξ(t) is minimised, i.e.
˜ (t)Ū
where Σū˜ = E{Ū ˜ T (t)}, Σ = E{Ȳ˜ (t)Ȳ˜ T (t)}, Σ = E{Ū
˜ (t)Ȳ˜ T (t)}. The term Σ
ȳ˜ ˜ȳ˜
ū ˜
ū
is calculated via, cf. (4.19):
⎡ var(ū
˜(t − s)) ⎤
⎢ ⎥
⎢ ⎥
⋯ 0 0
⎢ ⎥
⎢ ⎥
Σū˜ = ⎢ ⎥
⋮ ⋱ ⋮ ⋮
⎢ ⎥
(4.21)
⎢ ˜(t − 1)) ⎥
⎢ ⎥
0 ⋯ var(ū 0
⎢ var(ū(t)) ⎥
⎣ 0 ⋯ 0 ˜ ⎦
Analogously, the expression Σȳ˜ is obtained by replacing the terms var(ū˜(⋅)) in (4.21)
by var(ȳ˜(⋅)). Due to the fact that ũ(t) and ỹ(t) are mutually uncorrelated, Σū˜ỹ = 0.
For convenience, an expression Σ is introduced, which is equal to:
f (W ) = W ΣW T (4.23)
2. W Γ = 0
The solution to the constrained optimisation problem has been solved using the La-
grange multiplier method in Chapter 3.
˜(t)) and var(ȳ˜(t)) is
Note that due to the fact that the ratio of the variances var(ū
changing over the time, cf. (4.19), as opposed to the linear case in Chapter 3, the vector
W needs to be updated at each time step, i.e. the elements of W are time varying. This
may eventually result in an unnecessary jitter of the estimation lag τ . This happens if
∑ αi i
the mantissa of ∑ αi
, cf. (3.11) and (3.12), is close to 0.5 and in some time instances it
exceeds 0.5, whilst in the other is lower than 0.5. Thus, it is suggested to calculate τ
only once at the beginning of the input reconstruction process. Finally, the algorithm
for calculating the optimal vector W is summarised as follows:
85
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
1. Select the order of the parity space s ≥ n and build matrices Γ, Q and T .
2. Obtain Γ .
for t = 1 ∶ N
4. Compute Σ using:
λ = ((S −1 ψ) ψ)
T −1
(4.24d)
P = λS −1 ψ (4.24e)
if t = 1
∑ αi i
τ = round ( ) (4.24g)
∑ αi
where:
W T = [ αs αs−1 ⋯ α0 ]
T
86
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
end
end
Therefore, it can be noted that the PE-UIO-HW algorithm minimises the sum of φu (t)
and φy (t).
The accuracy of the unknown input estimation alters over the time, as var(ξ(t)) is
changing. Based on the assumption of a Gaussian distribution of ũ(t) and ỹ(t) it can
be assumed that the distribution of ξ(t) can be approximated with a Gaussian curve
with the variance of var(ξ(t)). Consequently, confidence bounds of the unknown input
estimate can be approximated using Gaussian distribution tables as multiplicities of
the standard deviation of ξ(t).
⎡ ⎤ ⎡ ⎤ ⎡ ⎤
⎢ 0 −0.56 ⎥ ⎢ −0.1200 ⎥ ⎢ ⎥
A=⎢⎢ ⎥ ⎢ ⎥ G = ⎢ 0.0055 ⎥
⎥ B=⎢ ⎥ ⎢ ⎥
⎢ 1 1.5 ⎥ ⎢ 0.4125 ⎥ ⎢ 0.0963 ⎥
(4.27)
⎣ ⎦ ⎣ ⎦ ⎣ ⎦
C =[ 0 1 ] D = 0.125 H = 0.025
87
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
The memoryless input and output nonlinearities are arbitrarily selected as:
Fig. 4.2 depicts functions ϕ(⋅) and η(⋅) where it is observed that they are both mono-
1.5
ū0 (t)
0.5
u0 (t)
0 0.2 0.4 0.6 0.8 1 1.2
80
60
y0 (t)
40
20
ȳ0 (t)
0 2 4 6 8 10
tonic and strictly increasing. This means that the impact of the input measurement
noise on the unknown input estimate is expected to be relatively low for low values of
u(t) (as the gradient of ϕ(u(t)) is small for low values of u(t)). On the other hand, this
impact will be relatively high for large values of u(t) (as the gradient of ϕ(u(t)) is large
for high values of u(t)). Due to the fact that the scheme utilises an inversion of η(⋅), an
opposite situation is expected according to the output measurement noise. Low values
of the output are expected to yield a significant impact of the output measurement
error on the accuracy of the unknown input estimate.
The known input and output signals as well as ū0 (t) and ȳ0 (t) are presented in
Fig. 4.3. For the first 1000 samples of the simulation y0 (t) is relatively high and, as
the slope of η(⋅) becomes steeper for higher values of ȳ0 (t), it is anticipated that the
inversion of the noisy measurement y(t) for the first 1000 samples will significantly
reduce the impact of the output measurement noise. After 1000 samples both u0 (t)
and y0 (t) decrease, which results in a higher vulnerability of the input reconstruction
process to the output measurement noise, cf. the slope of η(⋅) for the relatively low
values of the output. The input and output measurements are subjected to white,
Gaussian, zero-mean, and mutually uncorrelated noise sequences, whose variances are,
88
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
2
u0 (t)
u0 (t), ū0 (t) ū0 (t)
1
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
y0 (t)
y0 (t), ȳ0 (t)
10
2 ȳ0 (t)
0
10
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Figure 4.3: Input and output of the considered system (grey solid curve) compared
with input and output of linear block (dashed-dotted curve)
respectively, var(ũ(t)) = 0.002 and var(ỹ(t)) = 0.5. As a result of the inversion of η(⋅),
needed for the calculation of ȳ0 (t), the ratio of standard deviations of ȳ˜(t) to ȳ0 (t) is
2.2 %. However, as expected, the impact of the measurement noise on the accuracy of
the estimate of ȳ0 (t) changes over time. For the period between 100 and 900 samples
the standard deviation of ȳ˜(t) is equal to 1.4 % of the standard deviation of ȳ0 (t).
Whereas for the period between 1100 and 1900 samples this ratio is 12.3 %. This
can be interpreted that the impact of the measurement noise decreases over 8 times
after 1000 samples. The order of the parity space has been selected as 12, which gives
τ = 6 samples. The unknown input estimate with 95% confidence bounds is presented
˜(t)) and φu (t) are compared, whilst
in Fig. 4.4. In the upper subfigure of Fig. 4.5 var(ū
the middle subfigure of Fig. 4.5 compares var(ȳ˜(t)) and φy (t). The lower subfigure
of Fig. 4.5 presents the optimisation effect by comparing the sum of var(ū˜(t)) and
var(ȳ˜(t)) with the sum of φu (t) and φy (t). During the first 1000 samples the input
measurement noise has a larger influence on the unknown input estimation error in
comparison to the output measurement noise. One can note that for the first 300
samples the effect of the output measurement noise is actually amplified (as a result
of the minimisation of the joint impact of the input and output measurement noise).
˜(t), it has a negligible effect on the input estimation
However, due to a relatively large ū
error. After 1000 samples of the simulation the situation changes. The effect of the
input measurement noise becomes less significant, whereas the term ȳ˜(t) increases as
it depends strongly on the value of the output.
89
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
4.5
confidence bounds
true input
input estimate
4
Unknown input estimate
3.5
2.5
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Figure 4.4: Unknown input estimation for Hammerstein-Wiener system in the EIV
framework
˜(t))
var(ū
−3
φu (t)
10
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Influence of output measurement error
var(ȳ˜(t))
−2 φy (t)
10
−4
10
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Overall influence of measurement error
˜(t)) + var(ȳ˜(t))
−1
10
var(ū
φu (t) + φy (t)
−2
10
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Figure 4.5: Adaptive minimisation of the effect of measurement noise on the input
estimate
90
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
at each time sample. The system from the previous example is used for the simulation.
In Fig. 4.6 those two distributions have been compared as functions of time. Further-
more, in Fig. 4.7 both theoretical and experimental distributions of ξ(t) are presented
for four different time instances. It can be noted that the experimentally obtained dis-
tribution of ξ(t) matches the theoretical Gaussian distribution with the variance given
by equation (4.20). In Fig. 4.8 values of φu (t) and φy (t) as functions of time for a single
simulation run have been compared with functions of time of mean values of, respec-
˜ (t)Ū
tively, W QŪ ˜ T (t)QT W T and W Ȳ˜ (t)Ȳ˜ T (t)W T from the Monte-Carlo simulation.
It can be noted that the values of φu (t) and φy (t) calculated using (4.26) match the
experimental data.
91
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
t = 400, φu (t) = 8.3e-3, φy (t) = 3.5e-4 t = 1060, φu (t) = 1.9e-3, φy (t) = 2.3e-3
5 8
histogram of ξ(t)
4 Gaussian6 curve
3
4
2
2
1
0 0
−0.4 −0.2 0 0.2 0.4 −0.2 −0.1 0 0.1 0.2
t = 1280, φu (t) = 1.3e-3, φy (t) = 8.0e-4 t = 1800, φu (t) = 1.7e-3, φy (t) = 2.6e-3
10 8
8
6
6
4
4
2
2
0 0
−0.2 −0.1 0 0.1 0.2 −0.2 −0.1 0 0.1 0.2 0.3
The linear block of the considered system is given by (4.28), whilst the Hammerstein
and Wiener nonlinearities are:
ū0 (t) =
10
−5
1 + e−0.4u0 (t) (4.30)
y0 (t) =
bi
+ ci
1 + e i ȳ0 (t)
−a
where ai , bi , and ci are the coefficients of the Wiener nonlinearity η(⋅). The experi-
ment has been performed for three different Wiener nonlinearities (i = 1, 2, 3), whose
coefficients are given in Table 4.1. The Hammerstein nonlinearity as well as the three
considered Wiener nonlinearities, denoted as η1 (⋅), η2 (⋅), and η3 (⋅), are presented in
Fig. 4.9.
Both known and unknown inputs to the system, u0 (t) and v(t), are the same as
in Example 4.1. The upper subfigure of Fig. 4.10 presents u0 (t) and ū0 (t). Due to
the fact that the Hammerstein nonlinearity at the operating point is negligible ū0 (t)
92
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
φu (t)
0.015
obtained from MC
calculated (single run)
0.01
0.005
0
0 200 400 600 800 1000 1200 1400 1600 1800 2000
0
φy (t)
10
−2
10
−4
10
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Figure 4.8: Comparison of calculated and experimental values of φu (t) and φy (t).
Black dashed curves present theoretical values of φu (t) and φy (t) as func-
tions of time (calculated using (4.26) for a single simulation run). Grey
curves are experimentally obtained values of φu (t) and φy (t) from 10000
runs of Monte-Carlo simulation.
i ai bi ci
1 0.25 24 −12
2 0.2 26 −13
3 0.1 43 −21.5
is very close to u0 (t) (the considered system is virtually a Wiener system). The lower
subfigure of Fig. 4.10 shows ȳ0 (t) and the corresponding y0 (t) for three different Wiener
nonlinearities. It is anticipated that the accuracy of the unknown input estimation using
the linear PE-UIO will depend on the severity of the Wiener nonlinearity, i.e. the best
accuracy is expected for η3 (⋅), whilst it is anticipated that η1 (⋅) will result is the most
distorted unknown input estimate. The measured input and the output of the system
are subjected to white, zero-mean, Gaussian, mutually uncorrelated sequences with the
variances, respectively, var(ũ(t)) = 0.002 and var(ỹ(t)) = 0.003.
For each case of a nonlinear system (i.e. a system with different Wiener nonlinearity)
a linear model is obtained using the least squares technique in order to estimate the
unknown input using the PE-UIO with s = 12 samples. A Monte-Carlo simulation with
100 runs is carried out, whose results in terms of R2T are compared with results of the
PE-UIO-HW and presented in Table 4.2 . Sample plots of the unknown input estimate
93
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
1
ū0 (t)
0
−1
u0 (t)
−0.5 0 0.5 1 1.5
Wiener nonlinearities
10
y0 (t)
5 η1 (⋅)
η2 (⋅)
0 η3 (⋅)
−2 0 2 4 6 8 10 12
ȳ0 (t)
1
u0 (t), ū0 (t)
u0 (t)
0.5 ū0 (t)
−0.5
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples] ȳ0 (t)
y0 (t) (η1 (⋅))
y0 (t) (η2 (⋅))
y0 (t), ȳ0 (t)
10
y0 (t) (η3 (⋅))
8
6
4
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Figure 4.10: The upper subfigure shows the input of the considered system (grey solid
curve) compared with the input of linear block (dashed-dotted curve).
The lower subfigure presents the output of the linear dynamic block
(grey solid curve) compared with the output of the system for different
Wiener nonlinearities (black dashed, dashed-dotted and dotted curves).
for the considered models are plotted in Fig. 4.11. As expected the distortion in the
unknown input estimate using the linear PE-UIO is least when the Wiener nonlinearity
is given by η3 (⋅), whilst for η1 (⋅) the reconstructed signal is least accurate.
94
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
η1 (⋅)
5
true input
4 PE−UIO−HW
3 PE−UIO
2
Unknown input estimate
0 200 400 600 800 1000 1200 1400 1600 1800 2000
η2 (⋅)
5
4
3
2
0 200 400 600 800 1000 1200 1400 1600 1800 2000
η3 (⋅)
4
2
200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Table 4.2: Comparison of efficacy (in terms of the mean value of R2T [%] from a Monte-
Carlo simulation with 100 runs) of the linear PE-UIO and the PE-UIO-HW
for three different Wiener nonlinearities
95
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
φu
˜(t))
ρu = (4.31a)
var(ū
φy
var(ȳ˜(t))
ρy = (4.31b)
are not equal due to the presence of the matrix Q in (4.26a) and hence its influence
on (4.31a). This means that the impact of the change of s(t) will be different for the
input and the output measurement noise. It is proposed to create a two-dimensional
˜(t)) and var(ȳ˜(t)). Fur-
map, which assigns the value of s(t) for each couple of var(ū
˜(t)) and var(ȳ˜(t)) are calculated based on the current
thermore, as the values of var(ū
values of the measured input and output signals (affected by noise), cf. (4.19), the order
˜(t)) and var(ȳ˜(t))
of the parity space s(t) selected based on the current values of var(ū
may jitter unnecessarily. In order to avoid this problem, it is proposed to use local
˜(t)) and var(ȳ˜(t)) defined as:
mean values of var(ū
t+t2
˜(t)) = ∑ (var(ū
˜(i))
1
var(ū (4.32a)
t1 + t2 + 1 i=t−t1
t+t2
var(ȳ˜(t)) = ∑ (var(ȳ˜(i))
1
(4.32b)
t1 + t2 + 1 i=t−t1
where τ (t) is time varying, due to the alternating value of s(t). (Note that the notation
τ (t), W (t), and Q(t) has been used instead of τ , W , and Q in order to indicate that
the estimation lag τ , the vector W , and the matrix Q as well as sizes of W and Q are
time varying.) This would eventually lead to difficulties, such as some time instances
of the unknown input would be omitted, and some of them estimated more than once.
Therefore, a logic must be implemented, which copes with the variable estimation lag.
A difficulty may arise in two situations:
96
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
In the first case a particular time instance of the unknown input estimate is calculated
twice. In such a case from the two values of the unknown input estimate sample the
one should be selected, which is less affected by the noise. The fact that τ (t) increases,
means an increase of the noise influence, i.e. var(ū ˜(t)) or var(ȳ˜(t)) has increased.
Therefore, the impact of the measurement noise on the unknown input estimate has
also increased. Consequently, it can be deduced that the previously calculated value of
the unknown input estimate is less affected by noise.
In the second case, the situation is opposite, i.e. some time instances of v̂(t) will
be omitted. It is proposed to use W (t − 1) and Q(t − 1) to calculate the missing values
of the unknown input estimate.
Incorporating this logic into Algorithm 4.1 the adaptive order PE-UIO-HW (AO-
PE-UIO-HW) is obtained:
for t = 1 ∶ N
end
v̂(t − τ (t)) = W (t)η −1 (Y (t)) − W (t)Q(t)ϕ(U (t)) (4.36)
else
● Do nothing
end
97
˜(t)) and var(ȳ˜(t)). The row of the table is
Table 4.3: The table assigns value of the parity space order s(t) based on the values of var(ū
˜(t))} ≤ uu , whereas the column is chosen such that yd < log {var(ȳ˜(t))} ≤ yu
selected such that ud < log {var(ū
-2.6 -2.5 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
log {var(ū
-2.5 -2.3 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31
-2.3 -2.1 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32
-2.1 -2.0 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33
-2.0 -1.8 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34
-1.8 -1.7 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35
-1.7 -1.6 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36
-1.6 -1.4 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37
-1.4 -1.1 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38
systems
-1.1 -0.7 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39
98
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
2
u0 (t)
u0 (t), ū0 (t)
ū0 (t)
1
0 200 400 600 800 1000 1200 1400 1600 1800 2000
2 y0 (t)
y0 (t), ȳ0 (t)
10
ȳ0 (t)
0
10
0 200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Figure 4.12: Input and output of the considered system (grey solid curve) compared
with input and output of linear block (dashed-dotted curve)
The unknown input in this example is slightly lower than that in Example 4.1, which
yields lower values of y0 (t) compared to Example 4.1. Therefore, it is anticipated that
the effect of measurement noise on the unknown input reconstruction process will be
more significant than in Example 4.1 (especially when y0 (t) is very low between 1600
and 1800 sample).
As a result of the inversion of η(⋅), needed for the calculation of ȳ0 (t), the ratio of
99
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
standard deviations of ȳ˜(t) to ȳ0 (t) is 4.5 %. However, this ratio changes over the time.
For the period between 100 and 800 samples the standard deviation of ȳ˜(t) is equal
to 1.7 % of the standard deviation of ȳ0 (t). Whereas for the period between 1100 and
1600 samples this ratio is 27.0 %. In the extreme case of the period between 1600 and
1800 samples this ratio is equal to 81.1 %. Such a large deviation of the measurement
noise impact requires adaptivity of the unknown input reconstruction scheme. The
term var(ȳ˜(t)) has been calculated with t1 = 2τ (t − 1) + 1 and t2 = 0, cf. (4.32).
4
confidence bounds
true input
Unknown input estimate
2.5
1.5
1
200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
−5
10
200 400 600 800 1000 1200 1400 1600 1800 2000
Parity space order
30
s(t)
20
10
200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Figure 4.14: Adaptive minimisation of the effect of measurement output noise on the
input estimate
The unknown input estimate with 95 % confidence bounds is presented in Fig. 4.13.
The parity space order varies according to Table 4.3 and as a function of time is
100
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
presented in the lower subfigure of Fig. 4.14. The upper subfigure of Fig. 4.14 presents
var(ȳ˜(t)) and φy (t). As expected, the parity space order is low in the first half of
the simulation, when the impact of the measurement noise is low. Such a small s(t)
ensures a high bandwidth of the filter, and therefore even high frequency components
of v(t) are reconstructed, cf. Fig. 4.13. In the second half of the simulation, when the
impact of the measurement noise becomes more significant, the order of the parity space
increases. Furthermore, a higher parity space order yields stronger noise attenuation
(in terms of the ratio between var(ȳ˜(t)) and φy (t)), which can be seen in the upper
subfigure of Fig. 4.14.
A Monte-Carlo simulation with 100 runs has been carried out to compare the per-
formance of the AO-PE-UIO-HW and the PE-UIO-HW with a constant parity space
order for two cases of s. The aim of this experiment is to quantify the improvement of
the unknown input reconstruction process when the AO-PE-UIO-HW is used instead
of the PE-UIO-HW. Results in terms of the R2T are compared in Table 4.4. It can be
noted that by varying the parity space order an improvement of the accuracy of the
algorithm has been achieved. However, it needs to be remembered that the adaptive
algorithm needs more computational power.
PE-UIO-HW AO-PE-UIO-HW
sample s = 10 s = 23 s = 26 variable s
100:1990 0.0305 0.0320 0.0438 0.0197
100:1000 6.5e-4 0.0023 0.0033 5.7e-4
1000:1990 0.0600 0.0609 0.0837 0.0378
101
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
to reduce. Similarly as in the previous example the parity space order has been ob-
˜(t)) and var(ȳ˜(t)) have been calculated
tained using Table 4.3, whereas terms var(ū
using t1 = 2τ (t − 1) + 1 and t2 = 0.
4
confidence bounds
true input
Unknown input estimate
2.5
1.5
1
200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Figure 4.15: Unknown input estimation for a Hammerstein-Wiener system in the EIV
framework using AO-PE-UIO-HW
The unknown input estimate is presented in Fig. 4.15. The values var(ū ˜(t)) and
φu (t)) are depicted in the upper subfigure of Fig. 4.16, whilst var(ȳ˜(t)) and φy (t)), are
shown in the middle subfigure of Fig. 4.16. The lower subfigure of Fig. 4.16 presents
the parity space order s(t) as a function of time, which is compared with the s(t) from
the previous example. Note that the only difference between Examples 4.4 and 4.5 is
presence of the input measurement noise. The term var(ū˜(t)) is relatively large for
the first half of the simulation, whereas it becomes negligible after 1000 samples. This
influence of the input measurement noise can be noticed by comparing the values of
s(t) for the two considered examples. The presence of the input measurement noise
causes an increase of s(t) by approximately 5 samples compared to the OE case during
the first half of the simulation. After the first 1000 samples, as the impact of the input
measurement noise on the unknown input estimate becomes negligible, s(t) is similar
for both the OE (Example 4.4) and the EIV (Example 4.5) cases.
102
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
−2
10
−4
10
200 400 600 800 1000 1200 1400 1600 1800 2000
−2
Influence of input measurement error
10
−3
˜(t))
10
var(ū
φu (t)
200 400 600 800 1000 1200 1400 1600 1800 2000
Parity space order
Example 4.5
25
Example 4.4
20
15
10
200 400 600 800 1000 1200 1400 1600 1800 2000
Time [samples]
Figure 4.16: Upper and middle subfigures demonstrate adaptive minimisation of the
effect of measurement output noise on the input estimate in Example 4.5.
Lower subfigure compares values of parity space order s(t) as functions
of time in Examples 4.4 (OE) and 4.5 (EIV).
mutually uncorrelated noise sequences. The calculated values of the known input and
the output of the linear block are affected by measurement noise and the impact of
the EIV disturbance sequences depends on the values of the known input and out-
put themselves. This is due to the nonlinearities preceding and following the linear
block. Consequently, the impact of the measurement noise on the unknown input es-
timate changes over time. Therefore, the filter parameters are calculated at each time
instance.
In the first of the proposed algorithms, the PE-UIO-HW, the order of the parity
space, and thus the estimation delay, remains constant, whilst the filter parameters
vary over time.
Due to the fact that the impact of the measurement noise on the unknown input
estimate may vary significantly, a further extension to the scheme is proposed, where the
order of the parity space, s(t), is time varying. The parity space order is selected based
on values of the input and output measurement noise impact coefficients, calculated
as functions of noise variances and measured signals. The variable parity space order
allows the adjustment of the bandwidth of the unknown input reconstruction filter
103
4. Parity equations-based unknown input reconstruction for Hammerstein-Wiener
systems
to the changing impact of the measurement noise on the unknown input estimate.
The variation of s(t) imposes a variable estimation lag, τ (t). Consequently, a logic is
implemented, which resolves the problem of time varying τ (t), resulting in a smooth
estimate of the unknown input.
The proposed schemes, since inherently adaptive, require at each discrete time
step a non negligible computational effort. The future work, therefore, aims towards
an optimisation of the computational procedure. It is also intended to extend the
algorithm to the multivariable case. Furthermore, block oriented models in the EIV
framework are a new topic in the literature, for which effective identification schemes are
required. Although it has been assumed that the unknown input is fed directly to the
linear block, the algorithm can be easily extended to the case, when the unknown input
is transformed by a nonlinear memoryless function, and afterwards, the transformed
unknown input is fed to the linear dynamic block (based on the assumption that the
static nonlinearity is invertible).
104
Chapter 5
Nomenclature
105
5. Robust fault detection via eigenstructure assignment
106
5. Robust fault detection via eigenstructure assignment
5.1 Introduction
Increasing complexity of industrial systems leads to a growing demand for system fault
diagnosis. Furthermore, system uncertainties (disturbances), such as modelling errors,
parameter variations or unmeasurable external stimuli, obstruct the fault detection
process, leading to false alarms. Therefore, a need arises for robust, i.e. disturbance
decoupled, fault detection schemes. In this chapter robust fault detection is consid-
ered. This means that the residual generator is sensitive to faults but insensitive to
disturbances.
Frank & Wünnenberg (1989), Duan & Patton (2001), and Edelmayer (2005) pre-
sented robust fault detection schemes based on unknown input observers. LMI have
been also used for the robust fault detection (Chen & Nagarajaiah 2007, Ding, Zhong,
Bingyong & Zhang 2001). Zhong, Ding, Lam & Wang (2003) proposed an LMI ap-
proach to design a robust fault detection filter for uncertain linear time-invariant sys-
tems. Patton and Chen (Patton & Chen 1991b, Chen & Patton 1999) used the left
and right eigenstructure assignment techniques for the purpose of disturbance decou-
pling. Furthermore, the equivalence between the left eigenstructure assignment-based
robust fault detection filter and the first order PE has been demonstrated by Patton &
Chen (1991a, 1991b, 1991c). Also the problem of the robust fault detection via eigen-
structure assignment has been of the topic of the research of Park & Rizzoni (1994)
and Shen & Hsu (1998). Douglas & Speyer (1995) proposed an algorithm for pre-
venting ill-conditioning when using left eigenstructure assignment. A novel method for
left eigenstructure assignment has been proposed in (Kowalczuk & Suchomski 2005).
Patton & Liu (1994) presented a robust control design method using eigenstructure
assignment, genetic algorithms and a gradient-based optimisation. A reconfigurable
control scheme has been presented in (Ashari, Sedigh & Yazdanpanah 2005a, Ashari,
Sedigh & Yazdanpanah 2005b).
Eigenstructure assignment has been used in various industrial applications. A re-
view of applications has been presented in (Isermann & Balle 1997). Robust fault
detection filters based on eigenstructure assignment have been used in a rolling mill
(Gu & Poon 2003), a jet engine (Patton & Chen 1992), an automotive engine (Shen &
Hsu 1998), an advanced vehicle control systems (Douglas, Speyer, Mingori, Chen, Mal-
ladi & Chung 1996), a single-shaft gas turbine (Fantuzzi, Simani & Beghelli 2001), a
flexible manipulator (Tan & Habib 2006), an inverted pendulum (Tan & Habib 2004), a
vehicle health monitoring system (Ng, Chen & Speyer 2006), and a longitudinal motion
of an unmanned aircraft model (Siahi, Sadrnia & Darabi 2009). Luenberger state ob-
servers using a fixed-structure H∞ optimization have been applied to fault detection of a
107
5. Robust fault detection via eigenstructure assignment
108
5. Robust fault detection via eigenstructure assignment
where x(t) ∈ Rn is the system state vector, u(t) ∈ Rp and y(t) ∈ Rm are, respectively,
the system input and output, d(t) ∈ Rq denotes a disturbance vector, whilst µ(t) ∈ Rr
is a fault signal. Matrices A, B, C, D, E, and F are constant and have appropriate
dimensions. It is assumed that (C, A) is an observable pair and the matrix E is of full
column rank.
Consider the robust fault detection filter described in Subsection 2.6.1 applied to the
system (5.1). Tan et al. (2006a) observed that the invariant zeros of the system
(A, Ei , C) are unobservable modes of the pair (C ′ , A′ ), cf. (2.69). As a result, if the
invariant zero is unstable, the pair (C ′ , A′ ) is not detectable. Consequently, in order to
ensure stability of the fault detection filter (2.54) in the case, when the triple (A, Ei , C)
has an unstable invariant zero, the design procedure needs to be altered, which is
proposed in Subsections 5.2.2, 5.2.3, and 5.2.4.
Lemma 5.1. Denote invariant zeros of (A, Ei , C) as z1 , z2 , ⋯, zqi . Then zeros of (A, Ei , C)
fulfil the following recursive set of equations:
⎡ ⎤⎡ ⎤
⎢ zj I − A −vj−1 ⎥ ⎢ vj ⎥
⎢ ⎥⎢ ⎥=0
⎢ ⎥⎢ ⎥
⎢ ⎥ ⎢ gj ⎥
(5.2)
⎣ ⎦⎣ ⎦
C 0
⎡ 0 ⎤
⎢ ⎥
⎢ ⎥
z1 −g2 0 ⋯
⎢ ⋯ 0 ⎥
⎢ ⎥
A [ v 1 v 2 ⋯ v qi ] − [ v 1 v 2 ⋯ v qi ]⎢ ⎥
0 z2 −g3
⎢ ⋱ ⋮ ⎥
⎢ ⎥
⎢ ⎥
⋮ ⋮ ⋮ (5.3)
⎢ ⋯ zqi ⎥
⎣ 0 0 0 ⎦
+ Ei [ g 1 0 ⋯ 0 ] = 0
Lemma 5.2. If the pair (A, C) is observable, scalars gj , j = 1, ⋯, qi in (5.2) are non-
zero.
⎡ ⎤
⎢ zj I − A ⎥
⎢ ⎥ vj = 0
⎢ ⎥
⎢ ⎥
(5.4)
⎣ ⎦
C
which means that zj is an unobservable mode of the system (A, Ei , C). However, the
pair (C, A) is observable, hence it does not have unobservable modes. Consequently,
gj ≠ 0.
109
5. Robust fault detection via eigenstructure assignment
Rank condition
is not fulfilled is considered here and a solution is proposed, which allows to relax the
strict rank condition and, consequently, apply the robust fault detection filter presented
in Subsection 2.6.1. Denote δi , i = 1, ⋯, q the lowest number for which:
CAδi Ei ≠ 0 (5.6)
(A − λi I)Ei (5.7)
The rank condition (5.5) is fulfilled if and only if the ith column of Aλ , i = q0 + 1, ⋯, q,
is zero. This occurs if and only if Ei is a right eigenvector of A corresponding to the
eigenvalue λi . This would, however, mean that Ei belongs to the unobservable subspace
of the pair (C, A). Thus, as it has been assumed that (C, A) is observable, i.e. (C, A)
does not have any unobservable subspace, Ei is not a right eigenvector of A and the
rank condition (5.5) is not fulfilled. In order to relax this condition it is proposed to
slightly alter the robust fault detection filter design presented in Subsection 2.6.1, by
replacing E with the following matrix:
e = [ e1 e2 ⋯ eq ] (5.8)
where:
ei = Aδi Ei (5.9)
From the definition of δi , cf. (5.6), it holds that Cei ≠ 0 for i = 1, ⋯, q. Therefore,
columns of Aλ do not require to be equal to zero.
Theorem 5.1. Invariant zeros of the triple (A, ei , C) are equal to the invariant zeros
of (A, Ei , C) plus δi zero-valued invariant zeros.
⎡ ⎤ ⎡ j−1 ⎤
⎢ −A −Aj Ei ⎥ ⎢ A Ei ⎥
⎢ ⎥⎢ ⎥=0
⎢ ⎥⎢ ⎥
⎢ C ⎥ ⎢ −1 ⎥
(5.10)
⎣ ⎦⎣ ⎦
0
Therefore, from Lemma 5.1 it follows, that invariant zeros of (A, ei , C) are z1 = z2 =
⋯ = zδi = 0, whilst the corresponding vectors v0 = Aδi Ei = ei , v1 = Aδi −1 Ei ,⋯, vδi = Ei ,
and g1 = g2 = ⋯ = gδi = −1.
110
5. Robust fault detection via eigenstructure assignment
Consequently, replacing the matrix E with e in the filter described in Subsection 2.6.1
results in the pair (C ′ , A′ ) having zero-valued unobservable modes. It may be, however,
desired to set the eigenvalues of (A − KC) to different numbers that zero. In such a
situation the algorithms presented in the following subsections may be used.
Positive values of δi indicate a delay between the disturbance and the system output.
Thus, the system (5.1) can be reformulated as:
d∗ (t) = [ d1 (t − δ1 ) d2 (t − δ2 ) ⋯ dq (t − δq ) ]
T
(5.12)
⎡ ⎤⎡ ⎤
⎢ ze I − A −e ⎥ ⎢ v ⎥
⎢ ⎥⎢ ⎥ = 0
⎢ ⎥⎢ ⎥
⎢ 0 ⎥⎢ g ⎥
(5.13)
⎣ ⎦⎣ ⎦
C
where v and g are the invariant zero state and input directions, respectively. Utilising
a similar solution to that in (Massoumnia 1986) a vector ē is created, such that:
ē = [ e v ] (5.14)
The aim of the scheme is to create such a filter that the state trajectory yielded by the
disturbance d(t) remains in the subspace Im{ē}, as opposed to the algorithm presented
in (Chen & Patton 1999), where the state trajectory of the disturbance d(t) remains in
the one-dimensional subspace Im{e}. In order for the solution to this problem to exist
the subspace Im{ē} must be (C, A)-invariant, i.e. there must exist such a gain matrix
K that Im{(A − KC)ē} ⊆ Im{ē}, see (Halmos 1958, Basile & Marro 2002). This means
that there exists such a matrix X, that, cf. (2.19):
Note that, if (A, e, C) has no invariant zeros, then v = ∅, and, consequently, ē = E and
X = λ1 , where λ1 is the desired eigenvalue of (A − KC) corresponding to the vector E.
The necessary and sufficient conditions for disturbance decoupling are:
111
5. Robust fault detection via eigenstructure assignment
2. QCe = 0
From (5.15) it holds that the columns of ē are linear combinations of eigenvectors of
the matrix (A − KC):
ē = Ve Ψ (5.16)
where columns of Ve are the first two eigenvectors of (A − KC) and Ψ is an appropriate
matrix. Because (A − KC) is allocated distinct eigenvalues, rank{Ve } = 2, i.e. columns
of Ve are linearly independent. Also columns of ē are linearly independent (El-Ghezawi
et al. 1983). Consequently, matrix Ψ is of full rank. Furthermore:
(A − KC)Ve = Ve Λe (5.17)
where Λe is a diagonal matrix, whose diagonal elements are user defined eigenvalues
corresponding to the columns of Ve . By postmultiplying both sides of (5.17) by Ψ, the
following equation is obtained:
(A − KC)Ve Ψ = Ve Λe Ψ (5.18)
Ve Λe Ψ = Ve ΨX (5.19)
This yields:
Λe Ψ = ΨX (5.20)
Therefore, recalling that Ψ is of full rank, the matrix X can be defined as:
X = Ψ−1 Λe Ψ (5.21)
where columns of Ψ−1 are right eigenvectors of X, whilst diagonal elements of Λe are
its corresponding eigenvalues. Consequently, it can be noted that the eigenvalues of X
are equal to the eigenvalues of (A − KC) corresponding to the columns of Ve , i.e. the
linear combinations of columns of ē.
From (5.15) it follows that:
Denote Aē − ēX as Ae . The necessary and sufficient conditions to assign all columns of
ē as linear combinations of the right eigenvectors of (A − KC) are:
⎡ ⎤
⎢ Ae ⎥
⎢
(i) rank(C ē) = rank(⎢ ⎥)
⎥
⎢ C ē ⎥
⎣ ⎦
112
5. Robust fault detection via eigenstructure assignment
A′ = A − Ae (C ē) C
(5.23)
C ′ = (I − C ē(C ē) ) C
Consequently:
K ′ (I − Ce(Ce) ) Ce = K ′ C ′ wi = 0 (5.26)
⎡ ⎤
As a result it holds that:
⎢ λi I − A′ ⎥
⎢ ⎥ wi = 0
⎢ ⎥
⎢ ⎥
(5.28)
⎣ ⎦
C′
which means that the diagonal elements of Λe are unobservable modes of the pair
(C ′ , A′ ) and only remaining n − 2 eigenvalues can be allocated by K ′ .
Av = ze v − ge (5.29a)
Cv = 0 (5.29b)
(A − KC)v = Av (5.30)
113
5. Robust fault detection via eigenstructure assignment
Consequently, from (5.29a) it can be deduced that x12 = −g and x22 = ze . Knowing that
X has the same eigenvalues as Λe , x11 and x21 are calculated as:
x11 = λ1 + λ2 − ze (5.33a)
(ze − λ1 )(ze − λ2 )
x21 = − (5.33b)
g
Rank condition
Recall the rank condition (i) in Subsection 5.2.1 and (5.29b). It can be deduced that:
and:
Ae = [ Ae − x11 e − x21 v Av − ze v + ge ] (5.35)
From (5.29a), it holds that the second column of Ae is equal to zero. Therefore, using
the notation:
A∗e = Ae − x11 e − x21 v (5.36)
⎡ ⎤ ⎡ ∗ ⎤
it holds that:
⎢ Ae ⎥ ⎢ ⎥
rank(⎢
⎢
⎥) = rank(⎢ Ae ⎥)
⎥ ⎢ ⎥
⎢ C ē ⎥ ⎢ Ce ⎥
(5.37)
⎣ ⎦ ⎣ ⎦
Hence, the assignability condition can be reformulated as:
⎡ ∗ ⎤
⎢ Ae ⎥
(i) rank(Ce) = rank(⎢
⎢
⎥)
⎥
⎢ Ce ⎥
⎣ ⎦
(ii) (C ′ , A′ ) is a detectable pair, where:
A′ = A − A∗e (Ce) C
(5.38)
C ′ = (I − Ce(Ce) ) C
The algorithm for the design of a robust fault detection filter (RFDF) using right
eigenstructure assignment is summarised below.
114
5. Robust fault detection via eigenstructure assignment
⎡ ⎤⎡ ⎤
⎢ ze I − A −e ⎥ ⎢ v ⎥
⎢ ⎥⎢ ⎥ = 0
⎢ ⎥⎢ ⎥
⎢ 0 ⎥⎢ g ⎥
(5.39a)
⎣ ⎦⎣ ⎦
C
x11 = λ1 + λ2 − ze (5.39b)
(ze − λ1 )(ze − λ2 )
x21 = − (5.39c)
g
7. Obtain:
⎡ ⎤⎡ ⎤
zero zi such that:
⎢ zi I − A −vi−1 ⎥ ⎢ vi ⎥
⎢ ⎥⎢ ⎥=0
⎢ ⎥⎢ ⎥
⎢ ⎥ ⎢ gi ⎥
(5.40)
⎣ ⎦⎣ ⎦
C 0
where v0 refers to the vector e. It follows from Lemma 5.2 that gi are non-zero, hence,
vectors vi , i = 1, 2, ⋯, q1 can be scaled such that gi = −1 for i = 1, 2, ⋯, q1 . The aim of the
algorithm is to force the state trajectory governed by the disturbance d(t) to remain
within the subspace Im{ē} defined as:
ē = [ e v1 ⋯ vq1 ] (5.41)
115
5. Robust fault detection via eigenstructure assignment
which requires ē to be a (C, A)-invariant subspace, i.e. there exist such a gain matrix
K that, cf. (5.15):
KC ē = Aē − ēX (5.42)
Due to the fact that Cvi = 0 for i = 1, ⋯, q1 , all columns of Ae = Aē − ēX, except of the
first one, must be equal to zero for the solution of (5.42) to exist. Denote the elements
of X as xij , then the ith column of Ae is given by:
x1i v0 + x2i v1 + ⋯ + (xi−1,i − 1)vi−2 + (xi,i − zi−1 )vi−1 + ⋯ + xq1 +1 vq1 = 0 (5.45)
⎡ x 0 ⎤
⎢ ⎥
⎢ ⎥
1 0 0 ⋯ 0
⎢ x 0 ⎥
11
⎢ ⎥
⎢ ⎥
z1 1 0 ⋯ 0
⎢ x 0 ⎥
21
⎢ ⎥
X =⎢ ⎥
31 0 z2 1 ⋯ 0
⎢ ⋮ ⎥
(5.47)
⎢ ⎥
⎢ ⎥
⋮ ⋮ ⋮ ⋮ ⋱ ⋮
⎢ xq1 1,1 0 0 1 ⎥
⎢ 0 ⋯ zq1 −1 ⎥
⎢ ⎥
⎢ xq1 +1,1 0 0 z q1 ⎥
⎣ 0 ⋯ 0 ⎦
The first column of X is chosen such that the eigenvalues of X are equal to the desired
eigenvalues of (A − KC) corresponding to the linear combinations of columns of ē,
cf. (5.20). Consequently, (for derivation details see Appendix A) the first column of X
given by:
⎡ x ⎤
⎢ ⎥
⎢ ⎥
⎢ x ⎥
11
⎢ ⎥ ˜−1 ˜
⎢ ⎥ = Ã B̃
12
⎢ ⎥
(5.48)
⎢ ⎥
⎢ ⎥
⋮
⎢ xq1 +1,1 ⎥
⎣ ⎦
116
5. Robust fault detection via eigenstructure assignment
q
Øj,k = (−1)k−1 ∏(zl − λj )
i
(5.49)
l=k
˜ ∈ Rq1 +1 , denoted as B̃
whilst the j th element of the vector B̃ ˜ is:
j
q
˜ = λ i (z − λ )
B̃ j j∏ l j (5.50)
l=1
3. Denote e as v0 and obtain invariant zeros of the pair (A, e, C) and corre-
sponding vectors vi , for i = 1, ⋯, q1
⎡ ⎤⎡ ⎤
⎢ zi I − A −vi−1 ⎥ ⎢ vi ⎥
⎢ ⎥⎢ ⎥=0
⎢ ⎥⎢ ⎥
⎢ ⎥ ⎢ −1 ⎥
(5.51a)
⎣ ⎦⎣ ⎦
C 0
5. Calculate coefficients x11 , ⋯, xq1 +1,1 using (5.48), (5.49), and (5.50)
7. Obtain:
9. Calculate K
K = A∗e (Ce) + K ′ (I − Ce(Ce) ) (5.51e)
117
5. Robust fault detection via eigenstructure assignment
where Ω is the set of all invariant zeros of (A, e, C), whilst Ωi denotes the set of the
invariant zeros of (A, ei , C), see (Massoumnia 1986). The aim of the algorithm is to
ensure that the state trajectory driven by di (t) remains in the subspace Im{ēi }, where:
Therefore, the necessary and sufficient conditions for the robust fault detection are:
1. For each column of the matrix e it holds that Im{e¯i } is an invariant subspace of
(A − KC)
2. QCe = 0
Analogously to the case where e is a column vector, cf. Subsection 5.2.2, the matrix
A∗e is built, such that:
A∗e = [ A∗(1)
e
∗(2)
Ae
∗(q)
⋯ Ae ] (5.54)
where:
= (A − x11 I)ei − x21 v1 − x31 v2 − ⋯ − xqi +1,1 vq(i)
(i) (i) (i) (i) (i) (i)
A∗(i)
e i
(5.55)
The matrices A′ and C ′ are built as in (5.51c) and (5.51d). Note that if the (A, ei , C)
has no invariant zeros then it holds that δi = 0 and ei = Ei , see Lemma 5.2, and:
A∗(i)
e = (A − λi I)ei (5.56)
Hence, if the system has no invariant zeros and δi = 0 for i = 1, ⋯, q the algorithm
presented here is equivalent to that of Chen & Patton (1999). The necessary conditions
for the solution of the robust fault detection filter to exist are:
⎡ ∗ ⎤
⎢ Ae ⎥
(i) rank(Ce) = rank(⎢ ⎢
⎥)
⎥
⎢ Ce ⎥
⎣ ⎦
(ii) (C ′ , A′ ) is a detectable pair, where:
A′ = A − A∗e (Ce) C
(5.57)
C ′ = (I − Ce(Ce) ) C
118
5. Robust fault detection via eigenstructure assignment
⎡ x(i) 0 ⎤
⎢ ⎥
⎢ ⎥
1 0 0 ⋯ 0
⎢ x(i) 0 ⎥
11
⎢ ⎥
(i)
⎢ ⎥
z1 1 0 ⋯ 0
⎢ x (i) 0 ⎥
21
⎢ ⎥
(i)
X (i) = ⎢ 31 ⎥
0 z2 1 ⋯ 0
⎢ ⋮ ⎥
(5.58)
⎢ ⎥
⎢ (i) ⎥
⋮ ⋮ ⋮ ⋮ ⋱ ⋮
⎢ x 1 ⎥
⎢ q1 1,1 ⎥
(i)
0 0 0 ⋯ zq1 −1
⎢ (i) (i) ⎥
⎢ x z q1 ⎥
⎣ q1 +1,1 0 0 0 ⋯ 0 ⎦
q
Øj,k = (−1)k−1 ∏(zl − λj )
i
(i) (i) (i)
(5.60)
l=k
˜ ∈ Rq1 +1 , denoted as B̃
whilst the j th element of the vector B̃ ˜ is:
j
q
˜ (i) = λ(i) i (z (i) − λ(i) )
B̃ j j ∏ l j (5.61)
l=1
The generalised form of the algorithm for disturbance decoupled fault detection
filter is given below.
3. For each column of e obtain invariant zeros of the triple (A, ei , C), denoted
(i) (i)
as zj , and corresponding vectors vj , for j = 1, ⋯, qi
⎡ (i) ⎤ ⎡ (i) ⎤
⎢ zj I − A −vj−1 ⎥ ⎢ vj ⎥
⎢ ⎥⎢ ⎥=0
(i)
⎢ ⎥⎢ ⎥
⎢ ⎥ ⎢ −1 ⎥
(5.62a)
⎣ C 0 ⎦⎣ ⎦
(i)
where v0 denotes ei .
119
5. Robust fault detection via eigenstructure assignment
A∗e = [ A∗(1)
e
∗(2)
Ae
∗(q)
⋯ Ae ] (5.62b)
where
7. Obtain:
9. Calculate K
K = A∗e (Ce) + K ′ (I − Ce(Ce) ) (5.62f)
(i)
Cvj = 0 (5.63a)
QCei = 0 (5.63b)
120
5. Robust fault detection via eigenstructure assignment
(i)
QCwj = 0 (5.64a)
(Iλj − A + KC) wj = 0
(i) (i)
(5.64b)
Without loss of generality assume that F is a column vector. Using the notation
ē = [ ē1 ē2 ⋯ ēq ], the matrix F can be expressed as a sum of its orthogonal
projections on Im{ē} and the orthogonal completion of Im{ē}
This means that the necessary condition for the robust fault detection filter to exist
is that the dimension of the unobservable subspace of (QC, A − KC) is lower that n
(otherwise no fault can be detected as the whole state space is unobservable for the
fault detection filter).
Invariant zeros shape the response of the residual to a fault. In some situations, e.g.
fault identification, it may be desirable to influence not only its poles, but also zeros.
121
5. Robust fault detection via eigenstructure assignment
⎡ ′ ⎤⎡ ⎤
⎢ A − K ′ C ′ − zf I F ⎥ ⎢ v f ⎥
⎢ ⎥⎢ ⎥=0
⎢ ⎥⎢ ⎥
⎢ 0 ⎥ ⎢ gf ⎥
(5.68)
⎣ ⎦⎣ ⎦
QC
where vf and gf are the invariant zero state and input directions, respectively. Note
that Im{vf } ⊂ Ker{QC}, i.e. the state direction vf belongs to the right nullspace of QC.
The matrix Q which fulfils the condition QCe = 0 can be defined as (Basilevsky 1983):
Q = Q0 (I − Ce(Ce) ) (5.69)
where Q0 is an arbitrary matrix. Note that rows of Q are linear combinations of rows
of (I − Ce(Ce) ). Therefore, if rank(Q0 ) = rank (I − Ce(Ce) ) = n − rank(Ce), then
rank(Q) = n − rank(Ce). Hence, it follows, that if rank(Q0 ) = rank (I − Ce(Ce) ),
then the subspace spanned by the rows of Q is the subspace spanned by the rows of
(I − Ce(Ce) ). Furthermore, it holds that:
QC = Q0 (I − Ce(Ce) ) C = Q0 C ′ (5.70)
This shows that the invariant zero of the residual response to fault, defined by (5.68),
does not depend on choice of K ′ , i.e. zf does not depend on the choice of the eigenvalues
of (A − KC), which are not corresponding to the linear combinations of ēi , i = 1, ⋯, q.
As a result (5.68) can be rewritten as:
A′ vf − zf vf + F gf = 0 (5.72a)
QCvf = 0 (5.72b)
Now it will be demonstrated that the invariant zeros of the residual response to
a fault do not depend on the choice of eigenvalues of (A − KC) corresponding to the
linear combinations of ēi , i = 1, ⋯, q. Consider the following change of basis using the
following orthonormal matrix:
T = [ T 1 T2 ] (5.73)
where:
Im{T1 } = Im{[ ē1 ē2 ⋯ ēq ]} (5.74)
122
5. Robust fault detection via eigenstructure assignment
T T A′ T T T vf − zf T T vf + gf T T F = 0 (5.75a)
QCT T T vf = 0 (5.75b)
Knowing vf′ one can calculate vf which fulfils vf′ = T2T vf by solving dim{T2 } equations
with n unknowns. Hence the obtained solution has n−dim{T2 } parameters. The second
part of (5.76), i.e.
consists of n − dim{T2 } equations, from which remaining parameters can be found and
the vector vf calculated. Therefore, the existence of zf depends on A3 = T2T A′ T2 .
Recall (5.62b) and (5.62c):
where:
Ω = [ ∑qi=0
1 (1)
vi q2
∑i=0
(2)
vi
qq
⋯ ∑i=0 vi ]
(q)
(5.80)
Recall that:
A′ = A − Ae(Ce) C + Ω(Ce) C (5.81)
123
5. Robust fault detection via eigenstructure assignment
Therefore, the matrix A3 and hence the invariant zeros of the residual response to a
fault do not depend on the choice of the eigenvalues of (A − KC).
Note that the condition (ii) in (Chen & Speyer 2006a) has been specified as:
where Vi is the subspace spanned by invariant zero state directions of (A, Fi , C).
Chen & Speyer (2006a) explicitly indicate that each eigenvector of (A − KC) corre-
sponding to fi contains the vector fi . Although it is not explicitly said Algorithm 5.3
is characterised by the same property.
Lemma 5.3. Each eigenvector of (A − KC) obtained using Algorithm 5.3 correspond-
ing to linear combination of e¯i contains ei .
′(i) (i)
to linear combination of ē, denoted as wj , is a linear combination of vectors vk , k =
1, 2, ⋯, qi but not ei :
(i) (i) (i)
= α1 v1 + α2 v2 + ⋯ + αq(i) vq(i)
′(i)
wj i i
(5.86)
124
5. Robust fault detection via eigenstructure assignment
− α1 g1 ei + (α1 z1 − α2 g2 − λj ) v1 + (α2 z2 − α3 g3 − λj ) v2 + ⋯+
(i) (i) (i) (i) (i) (i) (i) (i) (i) (i) (i) (i)
(5.90)
(αqi −1 zqi −1 − αq(i) gqi − λj ) vqi −1 + (αq(i) − λj ) vq(i)
(i) (i) (i) (i) (i)
i i
zq(i)
i i
=0
Because columns of ēi are linearly independent, the above equation holds if and only
if:
(i) gi = 0
(or)
(i) (i) (i)
(ii) α1 = α2 = ⋯ = αqi = 0
Assumption (i) does not hold as the system is observable, see Lemma 5.2. Assumption
(ii) would mean that the eigenvector of (A−KC) corresponding to a linear combination
of columns of ēi is equal zero. Consequently, eigenvectors of (A − KC) corresponding
to linear combination of e¯i must contain ei . Due to the fact that eigenvectors can be
(i)
arbitrarily scaled, the coefficient α0 can be set to unity and the rest of the coefficients
can be scaled accordingly.
whilst the solution to the algorithm of Chen & Speyer (2006a) is:
KCf = Aw (5.92)
where
Aw = [ (A − λ(1) (1)
1 I)w1 (A − λ1 I)w1
(2) (2)
⋯ (A − λ1 I)w1 ]
(q) (q)
(5.93)
125
5. Robust fault detection via eigenstructure assignment
The main difference between both algorithms is the calculation of the coefficients
needed to obtain columns of A∗e (Aw ). The algorithm presented in Subsection 2.7.1
requires to calculate qi (qi + 1) of β̄j coefficients in order to calculate one column of
Aw . This is done by a pseudoinverse of a matrix of the dimension nqi × qi (qi + 1).
Furthermore, the β̄j coefficients are linearly dependent, cf. (B.9), and not all of them
are needed to compute a column of Aw . On the other hand, Algorithm 5.3 requires qi +1
coefficients to obtain any column of A∗e , which are calculated by solving a set of qi + 1
linear equations, which requires an inverse of a matrix of the dimension (qi +1)×(qi +1),
which is computationally less demanding compared to the algorithm of Chen & Speyer
(2006a).
⎡ 0 ⎤ ⎡ −0.8000 ⎤ ⎡ 1 ⎤
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
1 0 0
⎢ 0 ⎥ ⎢ 1.4000 ⎥ ⎢ 1.1 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
A=⎢ ⎥ E=⎢ ⎥ F =⎢ ⎥
0 1 0
⎢ 0 ⎥ ⎢ 1.2000 ⎥ ⎢ 0 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ (5.94)
0 0 1
⎢ −0.1155 −0.7985 −2.06 −2.35 ⎥ ⎢ 3.7725 ⎥ ⎢ 0 ⎥
⎣ ⎦ ⎣ ⎦ ⎣ ⎦
⎡ ⎤
⎢ −0.8165 0.5266 −0.2367 0 ⎥
⎢
C =⎢ ⎥
⎥
⎢ −0.4082 −0.2367 0.8816 0 ⎥
⎣ ⎦
The triple (A, E, C) has one invariant zero at ze = 1.2, whose input direction is g = 1
and the zero state direction is given by:
v=[ 1 2 1 0 ]
T
(5.95)
Also CE ≠ 0, thus δ1 = 0 and e = E. The aim of the algorithm is to limit the state
trajectory governed by d(t) to the subspace Im{[ E v ]}. Eigenvalues of (A − KC)
corresponding to linear combinations of E and v are selected to be 0.45 and 0.65. This
corresponds to x11 = −0.1, cf. (5.39b), and x21 = 0.4125, cf. (5.39c). Consequently, the
matrix A∗e is calculated as, cf. (5.39d):
⎡ 0.9075 ⎤
⎢ ⎥
⎢ ⎥
⎢ ⎥
⎢ ⎥
A∗e = ⎢ ⎥
0.5150
⎢ 3.4800 ⎥
(5.96)
⎢ ⎥
⎢ ⎥
⎢ −11.9856 ⎥
⎣ ⎦
126
5. Robust fault detection via eigenstructure assignment
⎡ 0.5186 ⎤
⎢ ⎥
⎢ ⎥
0.8704 −0.2593 0
⎢ 0.2943 −0.0736 ⎥
′ ⎢ ⎥
A =⎢ ⎥
0.8529 0
⎢ 1.9886 −0.4971 −0.9943 ⎥
⎢ ⎥
⎢ ⎥
1
⎢ −6.9644 ⎥
(5.97)
⎣ 0.9137 1.3645 −2.35 ⎦
⎡ ⎤
⎢ ⎥
C′ = ⎢ ⎥
−0.1843 0.3685 −0.5528 0
⎢ ⎥
⎢ 0.1936 −0.3872 0 ⎥
⎣ ⎦
0.5807
⎡ ∗ ⎤
It is noted that:
⎢ Ae ⎥
rank(CE) = rank(⎢
⎢
⎥) = 1
⎥
⎢ CE ⎥
(5.98)
⎣ ⎦
and (C ′ , A′ ) is a detectable pair, i.e. its unobservable modes, 0.45 and 0.65, lie within
the unit circle. Therefore, the solution for the stable filter design exists. The remaining
eigenvalues of (A − KC) are chosen to be 0.35 and 0.4 and consequently the matrix K ′
is given by:
⎡ −0.5936 4.0532 ⎤
⎢ ⎥
⎢ ⎥
⎢ 0.7442 2.6206 ⎥
′ ⎢ ⎥
K =⎢ ⎥
⎢ 0.6739 −7.8011 ⎥
(5.99)
⎢ ⎥
⎢ ⎥
⎢ −0.9526 2.0485 ⎥
⎣ ⎦
Finally, the gain matrix K is obtained as:
⎡ −1.8760 2.8324 ⎤
⎢ ⎥
⎢ ⎥
⎢ −0.7107 ⎥
⎢ ⎥
K =⎢ ⎥
1.2356
⎢ 5.8663 −2.8584 ⎥
(5.100)
⎢ ⎥
⎢ ⎥
⎢ −7.1590 −3.8596 ⎥
⎣ ⎦
Note that the fault distribution matrix F can be represented as a sum of:
⎡ 0.3889 ⎤ ⎡ 0.6111 ⎤
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ 0.0528 ⎥ ⎢ 1.0472 ⎥
⎢ ⎥ ⎢ ⎥
F (Ē ) = ⎢ ⎥ and F (Ē) = ⎢ ⎥
⎢ −0.4944 ⎥ ⎢ 0.4944 ⎥
(5.101)
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ 0.2201 ⎥ ⎢ −0.2201 ⎥
⎣ ⎦ ⎣ ⎦
where F (Ē) ∈ Im{Ē}, whilst the vector F (Ē ) is orthogonal to the subspace Im{Ē}.
Furthermore, the generalised angle between and F and the subspace Im{Ē} is given
by:
F T F (Ē) 180o
arccos ( ) = 26.7o
∣∣F ∣∣2 ∣∣F (Ē)∣∣2
(5.102)
π
Consequently, as the fault input direction F ∉ Im{Ē}, the fault occurrence can be
detected by the robust fault detection filter. The simulation results are presented in
Fig. 5.1. As expected the fault detection filter is insensitive to disturbances, whereas
127
5. Robust fault detection via eigenstructure assignment
Disturbance
0.2
0.1
d(t)
−0.1
−0.2
0 10 20 30 40 50 60 70 80 90 100
Fault detection
1.5
µ(t), r(t)
0.5
0
µ(t)
r(t)
−0.5
0 10 20 30 40 50 60 70 80 90 100
Time [samples]
Figure 5.1: Robust fault detection, q = 1, the triple (A, E, C) has single invariant zero.
Upper subfigure presents disturbances, whilst lower subfigure demon-
strates robust fault detection process. It can be noted that residual,
r(t), is insensitive to disturbances.
it is sensitive to fault.
⎡ 0 ⎤ ⎡ −2.4400 ⎤
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
1 0 0 0
⎢ 0 ⎥ ⎢ 0.6200 ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
0 1 0 0
⎢
A=⎢ 0 ⎥ E = ⎢ 1.5600 ⎥
⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
0 0 1 0
⎢ 0 ⎥ ⎢ −3.4816 ⎥
⎢ 0 0 0 1 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ (5.103)
⎢ −0.0751 −0.6345 −2.1375 −3.5875 −3 ⎥ ⎢ 18.3643 ⎥
⎣ ⎦ ⎣ ⎦
⎡ ⎤ ⎡ ⎤T
⎢ 0.0948 −0.4811 0.8675 0 −0.0837 ⎥ ⎢ 1 −1 ⎥
⎢
C =⎢ ⎥ F =⎢
1 1 0.3 ⎥
⎥ ⎢ ⎥
⎢ −0 ⎥ ⎢ 2 ⎥
⎣ ⎦ ⎣ ⎦
0 0 1 0 3.4 2 0 0
The triple (A, E, C) has two invariant zeros, namely z1 = 1.3 and z2 = 1.2, and:
⎡ 0.8000 ⎤ ⎡ 1 ⎤
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ −1.4000 ⎥ ⎢ 2 ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
v1 = ⎢ −1.2000 ⎥ v2 = ⎢
⎢ ⎥
⎢ 1 ⎥
⎥
⎢ ⎥ ⎢ ⎥
(5.104)
⎢ 0.0000 ⎥ ⎢ 0 ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ −3.4816 ⎥ ⎢ 0 ⎥
⎣ ⎦ ⎣ ⎦
128
5. Robust fault detection via eigenstructure assignment
Thus, the first column of the matrix X is given by, see (5.48):
⎡ −1.6000 ⎤
⎢ ⎥
⎢ ⎥
X =⎢
⎢
⎥
⎥
⎢ ⎥
−2.7075 (5.106)
⎢ −0.7268 ⎥
⎣ ⎦
⎡ −0.3912 ⎤T
⎢ ⎥
⎢ ⎥
⎢ 0.2150 ⎥
⎢ ⎥
∗ ⎢ ⎥
⎢
Ae = ⎢ −3.5079 ⎥⎥
⎢ ⎥
(5.107)
⎢ 12.7937 ⎥
⎢ ⎥
⎢ ⎥
⎢ −26.1909 ⎥
⎣ ⎦
⎡ 0.0019 ⎤
⎢ ⎥
⎢ ⎥
−0.0021 1.0106 −0.0192 −0.1078
⎢ −0.0010 ⎥
⎢ ⎥
⎢ ⎥
0.0012 −0.0058 1.0105 0.0593
′ ⎢
A =⎢ 0.0166 ⎥
⎥
⎢ ⎥
−0.0188 0.0954 −0.1720 0.0331
⎢ 0.9394 ⎥
⎢ 0.0686 −0.3480 0.6275 3.5263 ⎥
⎢ ⎥
(5.108)
⎢ −2.8760 ⎥
⎣ −0.2154 0.0779 −3.4220 −10.8064 ⎦
⎡ ⎤
⎢
′ ⎢ −0.0804 ⎥
C =⎢
0.0910 −0.4617 0.8324 −0.1968 ⎥
⎥
⎢ 0.0165 ⎥
⎣ ⎦
−0.0187 0.0947 −0.1708 0.0404
⎡ ∗ ⎤
⎢ Ae ⎥
rank(CE) = rank(⎢
⎢
⎥) = 1
⎥
⎢ CE ⎥
(5.109)
⎣ ⎦
and (C ′ , A′ ) is a detectable pair, hence the solution to the robust fault detection prob-
lem exists. The remaining eigenvalues of (A − KC) are selected to be 0.45 and 0.4 and,
129
5. Robust fault detection via eigenstructure assignment
⎡ 2.6891 −0.4392 ⎤
⎢ ⎥
⎢ ⎥
⎢ 3.1731 −0.7126 ⎥
⎢ ⎥
⎢ ⎥
⎢
K = ⎢ 0.1983 0.9669 ⎥
⎥
⎢ ⎥
(5.110)
⎢ −0.7233 −3.5263 ⎥
⎢ ⎥
⎢ ⎥
⎢ 1.4808 ⎥
⎣ 7.2189 ⎦
The orthogonal projections of F on the subspace Im{Ē} and the orthogonal com-
pletion of Im{Ē} are given by:
One can calculate the generalised angles between, respectively, F1 and F2 and their
projections on the subspace Im{Ē}, cf. (5.102), which are, respectively, 88.4264o and
4.2009o . Note that the F1 is ‘almost orthogonal’ to the subspace Im{Ē}, whilst F2
‘almost lies’ in the subspace Im{Ē}, i.e. the angle between F2 and Im{Ē} is low. Since
the norm of F1 (Ē ), i.e. 2.02, is significantly larger than the norm of F2 (Ē ), i.e. 0.32,
it is expected that the steady state gain of the residual response to the fault µ1 (t) will
be larger than the one of µ2 (t). The responses of the residual to both faults in the
z-domain are given by:
where r(z), µ1 (z), and µ2 (z) are, respectively, z-domain representations of r(t), µ1 (t),
and µ2 (t). The responses of the residual to both faults are presented in Fig. 5.2. It can
be noted that the residual is insensitive to disturbances. The steady state gain of the
residual response to fault µ1 (t) is larger than the one of the residual response to fault
µ2 (t).
130
5. Robust fault detection via eigenstructure assignment
Disturbance
0.2
0.1
d(t)
−0.1
−0.2
0 10 20 30 40 50 60 70 80 90 100
Fault detection
µ1 (t)
1.5
µ2 (t)
µ1,2 (t), r(t)
1
r(t)
0.5
−0.5
0 10 20 30 40 50 60 70 80 90 100
Time [samples]
Figure 5.2: Robust fault detection, q = 1, the triple (A, E, C) has two invariant zeros
at 1.2 and 1.3. Upper subfigure presents disturbances, whilst lower sub-
figure demonstrates robust fault detection process, i.e. residual, r(t), is
insensitive to disturbances.
⎡ 0 ⎤ ⎡ −0.8 0 ⎤
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
0 1 0 0
⎢ ⎥ ⎢ 1.4 1 ⎥
⎢ 0 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
0 0 1 0
A=⎢⎢
⎥
0 ⎥ E=⎢⎢ 1.2 0 ⎥
⎥
⎢ ⎥ ⎢ ⎥
0 0 0 1
⎢ ⎥ ⎢ 0 1 ⎥
⎢ 0 0 0 0 1 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ −0.0751 −0.6345 −2.1375 −3.5875 −3 ⎥ ⎢ 3.4816 1 ⎥
(5.113)
⎣ ⎦ ⎣ ⎦
⎡ 0.8675 0 −0.0837 ⎤
⎢ ⎥ ⎡ ⎤T
⎢ ⎥ ⎢ 1 1 0 0 0 ⎥
0.0948 −0.4811
C =⎢
⎢
⎥
⎥ F =⎢
⎢
⎥
⎥
⎢ ⎥ ⎢ 0 0 1 1 0 ⎥
0 0 0 1 0
⎢ 0.3929 ⎥ ⎣ ⎦
⎣ 0.8375 −0.3069 −0.2238 0 ⎦
Note that CE1 = 0 and δ1 = 1, whilst δ2 = 0. The invariant zero of (A, E1 , C) is 1.2 and
e1 is given by:
e1 = AE = [ 1.4 1.2 0 3.4816 −13.8381 ]
T
(5.114)
whilst e2 = E2 . Note that the triple (A, e1 , C) has two invariant zeros z1 = 0 and z2 = 1.2
= − [ 1 2 1 0 0 ]. Eigenvalues
(1) (1) (1)
and the corresponding v0 = e1 , v 1 = E1 and v2
of X (1) are selected to be 0.4, 0.5, and 0.6, whilst the eigenvalue corresponding to E2
is 0.7. Then matrices Ø(1) and B̃
˜ (1) are calculated as:
131
5. Robust fault detection via eigenstructure assignment
⎡ 0.300 ⎤
⎢ ⎥
⎢ ⎥
X (1) = ⎢
⎢
⎥
⎥
⎢ ⎥
−0.380 (5.116)
⎢ 0.336 ⎥
⎣ ⎦
⎡ ⎤
⎢ ⎥
⎢ ⎥
0.1400 1
⎢ −0.5000 −0.7 ⎥
⎢ ⎥
∗ ⎢ ⎢
⎥
⎥
Ae = ⎢ ⎥
⎢ ⎥
3.6016 1 (5.117)
⎢ −14.8826 0.3 ⎥
⎢ ⎥
⎢ ⎥
⎢ 33.6320 −7.922 ⎥
⎣ ⎦
⎡ ∗ ⎤
Note that:
⎢ Ae ⎥
rank(CE) = rank(⎢
⎢
⎥) = 2
⎥
⎢ CE ⎥
(5.118)
⎣ ⎦
and (C ′ , A′ ) is a detectable pair, hence the solution to the robust fault detection prob-
lem exists. The remaining eigenvalue of (A − KC) is chosen to be 0.3. As a result the
gain matrix K is obtained as:
⎡ −0.7707 0.2586⎤
⎢ ⎥
⎢ ⎥
0.5424
⎢ 0.6203 −0.3448 −0.0556⎥
⎢ ⎥
⎢ ⎥
⎢
K = ⎢ −0.6264 0.6775 −0.3647⎥⎥
⎢ ⎥
(5.119)
⎢ −1.3397 −0.6720 ⎥
⎢ 2.5006 ⎥
⎢ ⎥
⎢ 6.5303 −3.4705 −8.8598⎥
⎣ ⎦
⎡ 0.4397 ⎤ ⎡ 0.4397 ⎤
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ −0.5181 ⎥ ⎢ −0.5181 ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥
F1 (Ē ) = −0.4143 ⎢ 0.5965 ⎥ F2 (Ē ) = 1.0108 ⎢ 0.5965 ⎥
⎢
⎥
⎢ ⎥ ⎢ ⎥
(5.120)
⎢ 0.4143 ⎥ ⎢ 0.4143 ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ 0.1038 ⎥ ⎢ 0.1038 ⎥
⎣ ⎦ ⎣ ⎦
As a result the transfer functions between, respectively, µ1 (t) and µ2 (t) and the residual
132
5. Robust fault detection via eigenstructure assignment
r(z) −0.29
µ1 (z) (z − 0.3)
=
(5.121)
r(z) 0.70
µ2 (z) (z − 0.3)
=
The results of the simulation are presented in Fig. 5.3. It can be seen that transient
Disturbance
d1 (t)
0.5
d2 (t)
d(t)
−0.5
0 10 20 30 40 50 60 70 80 90 100
Fault detection
1
µ1 (t)
µ2 (t)
µ1,2 (t), r(t)
0.5 r(t)
0
−0.5
0 10 20 30 40 50 60 70 80 90 100
Time [samples]
Figure 5.3: Robust fault detection, q = 2, the triple (A, E, C) has two invariant ze-
ros. Upper subfigure presents disturbances, whilst lower subfigure demon-
strates robust fault detection process. Trajectory of disturbances remains
within a 4-dimensional subspace of 5-dimensional state space of fault de-
tection filter, thus only one-dimensional subspace is left for fault detec-
tion. As a result the responses of residual to different faults differ only
by steady state gain, cf. Remarks 5.1 and 5.2.
behaviour of the residual responses to both faults is the same, but their steady state
gains differ.
133
5. Robust fault detection via eigenstructure assignment
state observer. This means that the state estimate converges within a finite time (as
opposed to the asymptotic observer in Subsection 2.6.1 and in the previous section).
Consequently, due to a finite impulse response of the state observer, the proposed
scheme is equivalent to PE. The material presented in this section is extension to
(Sumislawska, Larkowski & Burnham 2011b), where a design of robust PE for systems,
which do not contain any invariant zeros, has been proposed.
d(t) f (t)
u(t) y(t)
where:
Aci = A − Ki C (5.124)
134
5. Robust fault detection via eigenstructure assignment
The ith state estimation error ξi (t) = zi (t) − x(t) is then governed by:
Therefore, the state estimation error as a function of time and initial conditions is
defined by:
ξi (t) = Atci ξi (0) (5.126)
ξi (t − τ ) = At−τ
ci ξi (0) (5.127)
In order to obtain the correct state estimate the term Aτci x(t−τ ) needs to be eliminated
from the expression (5.129). Note that there are two unknowns in (5.129), namely,
x(t) and x(t − τ ), and two equations, i.e. two Luenberger-type observers, therefore an
explicit solution to the set of equations can be found. By combining the two observer,
the following is obtained, cf. (Engel & Kreisselmeier 2002):
where:
⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤
⎢ Ac1 0 ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥
Ac = ⎢
⎢
⎥ K = ⎢ K1 ⎥ G = ⎢ B − K1 D
⎥ ⎢ ⎥ ⎢
⎥ z = ⎢ z1 ⎥
⎥ ⎢ ⎥
⎢ 0 Ac2 ⎥ ⎢ K2 ⎥ ⎢ B − K2 D ⎥ ⎢ z2 ⎥
(5.131)
⎣ ⎦ ⎣ ⎦ ⎣ ⎦ ⎣ ⎦
The gain matrix J should be chosen in such a way, the the terms Aτci x(t − τ ), i = 1, 2
are eliminated from (5.130b), i.e.
⎡ ⎤
⎢ x(t) − Aτc1 x(t − τ ) ⎥
(z(t) − Aτc z(t − τ )) =J⎢
⎢
⎥ = x(t)
⎥
⎢ x(t) − Aτc2 x(t − τ ) ⎥
J (5.132)
⎣ ⎦
⎡ ⎤
⎢ I Aτc1 ⎥
J⎢
⎢
⎥=[ I 0 ]
⎥
⎢ I Aτc2 ⎥
(5.133)
⎣ ⎦
135
5. Robust fault detection via eigenstructure assignment
Consequently, the gain matrix J is calculated as, cf. (Engel & Kreisselmeier 2002, Raff,
Menold, Ebenbauer & Allgöwer 2005):
⎡ ⎤−1
⎢ I Aτc1 ⎥
J = [ I 0 ]⎢
⎢
⎥
⎥
⎢ I Aτc2 ⎥
(5.134)
⎣ ⎦
Note that the solution of the expression (5.134) exists if the observer transition matrices
Ac1 and Ac2 have distinct eigenvalues. Therefore, if the system (A, E, C) contains an
invariant zero, Algorithm 5.3 should be used in order to ensure distinct eigenvalues of
Ac1 and Ac2 .
1. QCe = 0
2. For each column of the matrix E it holds that Im{e¯i } is an invariant subspace of
Ac1 and Ac2
ξi (t + 1) = Aci ξi (t) + F µ(t) + Ed(t) = Aci ξi (t) + F µ(t) + ed∗ (t) (5.135)
t−1 t−1
ξi (t) = Atci ξi (0) + ∑ Ajci F µ(t − j) + ∑ Ajci ed∗ (t − j) (5.136)
j=0 j=0
136
5. Robust fault detection via eigenstructure assignment
t−1 t−1
zi (t) − Aτci z(t − τ ) = x(t) + Atci ξi (0) + ∑ Ajci F µ(t − j) + ∑ Ajci ed∗ (t − j)−
j=0 j=0
t−τ −1 t−τ −1
Aτci x(t − τ ) − Aτci At−τ
ci ξi (0) − Aci ∑ Aci F µ(t − τ − j) − Aci ∑ Aci ed (t − τ − j)
τ j τ j ∗
j=0 j=0
t−1 t−1 t−1
(5.137)
= x(t) − Aτci x(t − τ ) + ∑ Aci F µ(t − j) + ∑ Aci ed (t − j) − ∑ Aci F µ(t − j)−
j j ∗ j
j=0 j=0 j=τ
t−1 τ −1 τ −1
∑ Aci ed (t − j) = x(t) − Aci x(t − τ ) + ∑ Aci F µ(t − j) + ∑ Aci ed (t − j)
j ∗ τ j j ∗
j=τ j=0 j=0
Therefore, one can obtain the state estimate x̂(t), see (5.133):
⎡ τ −1 j ⎤ ⎡ τ −1 j ∗ ⎤
⎢ ∑j=0 Ac1 F µ(t − j) ⎥ ⎢ ⎥
x̂(t) = x(t) + J ⎢ ⎥ + J ⎢ ∑j=0 Ac1 ed (t − j) ⎥
⎢ τ −1 j ⎥ ⎢ τ −1 j ∗ ⎥
⎢ ∑j=0 Ac2 F µ(t − j) ⎥ ⎢ ∑j=0 Ac2 ed (t − j) ⎥
(5.138)
⎣ ⎦ ⎣ ⎦
⎡ τ −1 j ∗ ⎤
⎢ ∑j=0 Ac1 ed (t − j) ⎥
r(t) = QCJ ⎢
⎢ τ −1 j ∗
⎥
⎥
⎢ ∑j=0 Ac2 ed (t − j) ⎥
(5.139)
⎣ ⎦
J = [ J1 J2 ] (5.140)
where J1 , J2 ∈ Rn×n . Incorporating (5.140) into (5.134), the following relationships are
obtained:
⎡ τ −1 j ∗ ⎤
⎢ ∑j=0 Ac1 ed (t − j) ⎥
r(t) = QC [ J1 J2 ] ⎢
⎢ τ −1 j ∗
⎥
⎥
⎢ ∑j=0 Ac2 ed (t − j) ⎥
⎣ ⎦ (5.142)
τ −1 τ −1
= QCJ1 ∑ Ajc1 ed∗ (t − j) + QCJ2 ∑ Ajc2 ed∗ (t − j)
j=0 j=0
137
5. Robust fault detection via eigenstructure assignment
PV ⊆ V
(5.143)
RV ⊆ V
Therefore:
P V + RV = (P + R)V ⊆ V (5.144)
Thus in the fault-free case the residual, cf. (5.139), is equal to zero.
2 τ −1 2 τ −1
x̂(t) = ∑ Ji ∑ Ajci (B − Ki D)u(t − j − 1) + ∑ Ji ∑ Ajci Ki y(t − j − 1) (5.148)
i=1 j=0 i=1 j=0
(The derivation of the above equation is analogous to (5.137).) Therefore, the residual
generator can be described by the following parity relation:
where:
Y (t) = [y T (t − τ ) y T (t − τ + 1) ⋯ y T (t)]
T
(5.150)
U (t) = [uT (t − τ ) uT (t − τ + 1) ⋯ uT (t)]
T
138
5. Robust fault detection via eigenstructure assignment
and:
ei = Aδi Ei (5.152a)
4. For each column of E obtain invariant zeros of the triple (A, ei , C), denoted
(i) (i)
as zj , and corresponding vectors vj , for j = 1, ⋯, qi , such that:
⎡ (i) ⎤ ⎡ (i) ⎤
⎢ zj I − A vj−1 ⎥ ⎢ vj ⎥
⎢ ⎥⎢ ⎥=0
(i)
⎢ ⎥⎢ ⎥
⎢ ⎥ ⎢ −1 ⎥
(5.152b)
⎣ C 0 ⎦⎣ ⎦
(i)
where v0 denotes ei .
(i) (i)
5. Calculate coefficients x11 , ⋯, xqi +1,1 using (5.59–5.61)
A∗e = [ A∗(1)
e
∗(2)
Ae
∗(q)
⋯ Ae ] (5.152c)
where
139
5. Robust fault detection via eigenstructure assignment
7. Obtain:
9. Calculate K1
K1 = A∗e (CE) + K1′ [I − CE(CE) ] (5.152g)
⎡ 0 3 4 ⎤ ⎡ 1 0 ⎤ ⎡ 1 ⎤ ⎡ 1 ⎤
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎡ ⎤
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ 0 1 0 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢
A = ⎢ 1 2 3 ⎥ B = ⎢ 0 0 ⎥ E = ⎢ 2 ⎥ F = ⎢ −0.5 ⎥ C = ⎢ ⎥ (5.153)
⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ 0 0 1 ⎥
⎢ 0 2 5 ⎥ ⎢ 0 1 ⎥ ⎢ −1 ⎥ ⎢ 0.5 ⎥ ⎣ ⎦
⎣ ⎦ ⎣ ⎦ ⎣ ⎦ ⎣ ⎦
and the matrix D is null. The eigenvalues of the matrix Ac1 are chosen to be λAc1 =
{0.9, 0.925, 0.95}, whereas the eigenvalues of Ac2 are λAc2 = {0.965, 0.975, 0.995}. (Note
that eigenvalues of Ac1 and Ac2 are close to unity.) Gain matrices K1′ and K2′ have
140
5. Robust fault detection via eigenstructure assignment
been selected to minimise Frobenius norms of K1 and K2 and are given by:
Algorithm 5.4 is used to calculate Wu and Wy for parity space orders equal to τ = 5
and τ = 15 samples. The impulse and step responses of the residual r(t) to the fault
µ(t) for the two aforementioned parity space orders are compared in Fig. 5.5. It is
worth noting that the response of the residual to fault is strongly dependent on the
chosen parity space order. This is due to the fact that the eigenvalues of Ac1 and Ac2
are selected to be close to unity, therefore the two asymptotic state observers which are
combined to create the finite time convergent observers are relatively slow in comparison
to the chosen parity space orders.
0.3 τ = 15
0.2
0.1
0
0 2 4 6 8 10 12 14 16 18 20
Step response of residual to fault
1
0.5
0
0 2 4 6 8 10 12 14 16 18 20
Time [samples]
Figure 5.5: Comparison of step and impulse responses of the residual to fault for
different cases of τ . Eigenvalues of (A − K1 C) and (A − K2 C) are, respec-
tively, λAc1 = {0.9, 0.925, 0.95} and λAc2 = {0.965, 0.975, 0.995}.
In the second simulation ‘fast’ eigenvalues are taken into consideration, i.e. λ∗Ac =
{0.33, 0.25, 0.3} and λ∗Ac = {0.15, 0.20, 0.35}, and compared with ‘slow’ eigenvalues
1
2
from the previous experiment for parity space τ = 15. The gain matrices K1∗ and K2∗
are:
⎡ 2.7472 3.8243 ⎤ ⎡ 2.7472 3.8243 ⎤
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
∗ ⎢ ⎥ ∗ ⎢
K1 = ⎢ 0.7943 0.2487 ⎥ K2 = ⎢ 0.7943 0.2487 ⎥
⎥
⎢ ⎥ ⎢ ⎥
(5.155)
⎢ 2.3278 5.3257 ⎥ ⎢ 2.3278 5.3257 ⎥
⎣ ⎦ ⎣ ⎦
Step and impulse responses of the residual to the fault for two different sets of eigen-
141
5. Robust fault detection via eigenstructure assignment
values (‘slow’, i.e. λAc1 and λAc2 , versus ‘fast’, i.e. λ∗Ac and λ∗Ac ) are compared in
1 2
Fig. 5.6. As expected, the response of the residual to the fault is faster when the
eigenvalues of Ac1 and Ac2 are closer to the origin. The experiment also revealed that,
in the case of ‘fast’ eigenvalues, the increase of the parity space order has a negligible
influence on the residual response to the fault.
0.1
0
0 2 4 6 8 10 12 14 16 18 20
Step response of residual to fault
1
0.5
0
0 2 4 6 8 10 12 14 16 18 20
Time [samples]
Figure 5.6: Comparison of step and impulse responses of the residual to fault
for different choices of eigenvalues of Ac1 and Ac2 . Slow eigenvalues:
λAc1 = {0.9, 0.925, 0.95} and λAc2 = {0.965, 0.975, 0.995}. Fast eigenval-
ues: λ∗Ac = {0.33, 0.25, 0.3} and λ∗Ac = {0.15, 0.20, 0.35}.
1 2
142
5. Robust fault detection via eigenstructure assignment
1. QCE = 0
3. Compute P (λi ) as
P (λi ) = (λi I − AT )−1 C T (5.156a)
6. Obtain L and W as
7. Calculate K1 via
K = −(L−1 )T W T (5.156g)
143
5. Robust fault detection via eigenstructure assignment
Proof. Assume that two arbitrary matrices P and R have a common eigenvector x
corresponding to eigenvalues λp and λr , respectively:
P x = λp x Rx = λr x (5.157)
P = V ΛV −1 (5.159)
P i = V Λi V −1 (5.160)
which means that P and its ith power have common eigenvectors. Consequently, if
all rows of QC are left eigenvectors of Ac1 and Ac2 , then all rows of QC are also left
eigenvectors of J1 and J2 .
The residual in the fault-free case can be described by, see (5.142):
2 τ
r(t) = ∑ ∑ Pl,k d(t − k) (5.161)
l=1 k=1
where:
Pl,k = QCJl Ak−1
cl E (5.162)
144
5. Robust fault detection via eigenstructure assignment
Denote the ith left eigenvector, right eigenvector, and the corresponding eigenvalue of
the matrix Ak−1
cl as, respectively, (liAc )T , viAc , and λA
i . Analogously, use the notation
c
(liJ )T , viJ , and λJi for the left eigenvector, right eigenvector, and the corresponding
eigenvalue of the matrix Jl . The first rq left eigenvectors of Ak−1 k−1
c1 , Ac2 , J1 , and J2 are
rows of QC. Then Pl,k can be expressed as:
rq n
cl E + QC ∑ λi vi (li ) Acl E
Pl,k = QC ∑ λJi viJ (liJ )T Ak−1 J J J T k−1
(5.163)
i=1 i=rq +1
The second element of the above expression is equal to zero because QCviJ = 0, for
i = rq + 1, ⋯, n (because rows of QC are first left eigenvectors of Jl ). Subsequently:
rq rq
i λj vj (lj ) vi (li ) E
Pl,k = QC ∑ ∑ λA c J J J T Ac Ac T
i=1 j=1
n rq (5.164)
∑ ∑ λi c λj vj (lj ) vi c (li c ) E
A J J J T A A T
+ QC
i=rq +1 j=1
The first element of the above expression is zero because (liAc )T is equal to ith row of
QC and QCE = 0. The second element is equal to zero because (ljJ )T viAc = (ljJ )T viJ = 0,
for i = rq + 1, ⋯, n, j = 1, ⋯, rq . This shows, therefore, that the residual is null if there is
no fault present in the system.
⎡ 0.25 0 0 ⎤ ⎡ 0 ⎤ ⎡ 1 ⎤ ⎡ 1.0 ⎤
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎡ ⎤
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ 1 1 0 ⎥
⎢
A=⎢ 0 ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢
0 ⎥ B = ⎢ 0 ⎥ E = ⎢ 1 ⎥ F = ⎢ 0.1 ⎥ C = ⎢ ⎥ (5.165)
⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ 0 1 1 ⎥
0.5
⎢ 0 ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎣ ⎦
⎣ 0 0.375 ⎦ ⎣ 1 ⎦ ⎣ 0 ⎦ ⎣ 1.0 ⎦
145
5. Robust fault detection via eigenstructure assignment
⎡ 1.3793 0 ⎤
⎢ ⎥
⎢ ⎥
P (λ1 ) = ⎢
⎢
⎥
⎥
⎢ ⎥
2.1053 2.1053 (5.167)
⎢ 0 ⎥
⎣ 1.6667 ⎦
and corresponding w1∗ = [ 0.7250 −1.2 ] , cf. (5.156c). It holds that P (λ1 )w1∗ = l1 ,
T
hence a complete decoupling can be achieved. The remaining eigenvalues of Ac1 are
λ2 = 0.985 and λ3 = 0.995 and the corresponding matrices P (λ2 ) and P (λ3 ) are obtained
using (5.156a):
⎡ 1.3605 0 ⎤ ⎡ 1.3423 0 ⎤
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
P (λ2 ) = ⎢
⎢
⎥
⎥ P (λ3 ) = ⎢
⎢
⎥
⎥
⎢ ⎥ ⎢ ⎥
2.0619 2.0619 2.0202 2.0202 (5.168)
⎢ 0 ⎥ ⎢ 0 ⎥
⎣ 1.6393 ⎦ ⎣ 1.6129 ⎦
Vectors w2 and w3 can be freely chosen and they are selected to be w2 = [ 1 0 ] and
T
⎡ 1.3605 ⎤ ⎡ 0 ⎤
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
l2 = ⎢
⎢
⎥
⎥ l3 = ⎢
⎢ 2.0202
⎥
⎥
⎢ ⎥ ⎢ ⎥
2.0619 (5.169)
⎢ 0 ⎥ ⎢ 1.6129 ⎥
⎣ ⎦ ⎣ ⎦
Consequently, matrices L and W are obtained via, respectively, (5.156e) and (5.156e)
⎡ 1 1.3605 0 ⎤
as:
⎢ ⎥ ⎡ ⎤
⎢ ⎥ ⎢ 0.725 1 0 ⎥
L=⎢ ⎥
⎢ −1 2.0619 2.0202 ⎥ W =⎢
⎢
⎥
⎥
⎢ ⎥ ⎢ −1.200 0 1 ⎥
(5.170)
⎢ −2 0 1.6129 ⎥ ⎣ ⎦
⎣ ⎦
and the gain matrix K1 calculated using (5.156g) is:
⎡ 0.7203 −5.8212 ⎤
⎢ ⎥
⎢ ⎥
K1 = ⎢
⎢
⎥
⎥
⎢ ⎥
−0.9603 3.8412 (5.171)
⎢ 1.2028 −5.4312 ⎥
⎣ ⎦
Analogously, the gain matrix K2 is calculated. Vectors w2 and w3 are selected such
that the Frobenius norm of K2 is minimised and K2 is given by:
⎡ −1.6037 −0.5727 ⎤
⎢ ⎥
⎢ ⎥
K2 = ⎢
⎢
⎥
⎥
⎢ ⎥
0.7037 0.1227 (5.172)
⎢ −0.8287 −0.8727 ⎥
⎣ ⎦
Then, the matrices J1 and J2 and the PE coefficient vectors Wu and Wy are cal-
146
5. Robust fault detection via eigenstructure assignment
−0.2
2 4 6 8 10 12 14 16
Step response of residual to fault
1
0.5
2 4 6 8 10 12 14 16
Time [samples]
Figure 5.7: Comparison of responses of residual to fault for different cases of par-
ity space order. Eigenvalues of Ac1 and Ac2 are, respectively, λAc1 =
{0.9, 0.925, 0.95} and λAc2 = {0.965, 0.975, 0.995}.
culated using, respectively, (5.156h), (5.156i), and (5.151) for the parity space orders
equal to τ = 5 and τ = 14 samples. Impulse and step responses of the residual r(t) to
the fault µ(t) for the two aforementioned values of τ are compared in Fig. 5.7. Simi-
larly, as in Example 5.4 a strong influence of the parity space order τ on the residual
response to fault can be observed (due to the fact that the eigenvalues of Ac1 and Ac2
are close to unity). Also the impulse response of the residual to the fault decays almost
linearly. In the second simulation ‘fast’ eigenvalues are taken into consideration, i.e.
λ∗Ac = {0.2, 0.3, 0.4} and λ∗Ac = {0.8, 0.7, 0.6}, and compared with ‘slow’ eigenvalues
1 2
from the previous experiment for the parity space order τ = 8. The gain matrices K1∗
and K2∗ of the ‘fast’ filter are:
Step and impulse responses of the residual to the fault for two different sets of eigen-
values (‘slow’, i.e. λAc1 and λAc2 , versus ‘fast’, i.e. λ∗Ac and λ∗Ac ) are compared in
1 2
Fig. 5.8. Similarly to the experiment in Example 5.4, the response of the residual to
the fault is faster when the eigenvalues of Ac1 and Ac2 are closer to the origin. Also
in the case of ‘fast’ eigenvalues, the increase of the parity space order has negligible
influence on the residual response to the fault.
Choice of a low parity space order or selection of eigenvalues close to the origin leads
to a fast reaction of the residual to a fault. However, in practice there is always noise
present in the system. Therefore, it may be required to increase the parity space order
147
5. Robust fault detection via eigenstructure assignment
1 2 3 4 5 6 7 8 9 10
Step response of residual to fault
1.5
0.5
−0.5
1 2 3 4 5 6 7 8 9 10
Time [samples]
Figure 5.8: Influence of eigenvalues of matrices Ac1 and Ac2 on step and impulse
response of residual to fault for τ = 8. Slow eigenvalues: λAc1 =
{0.975, 0.985, 0.995} and λAc2 = {0.9, 0.925, 0.95}. Fast eigenvalues:
λ∗Ac = {0.2, 0.3, 0.4} and λ∗Ac = {0.8, 0.7, 0.6}.
1 2
for the purpose of minimising the effects of noise on the residual generator. In this ex-
periment the scheme proposed here is compared with the DRFDF, see Subsection 2.6.4,
which is equivalent to the PE, whose Wu and Wy coefficients are:
Wu = [ 0 2 ]
(5.174)
Wy = [ 1 −2 −0.25 0.75 ]
Algorithm 5.5 has been designed using the ‘slow’ eigenvalues set, i.e. λAc1 and λAc2 for
two cases of the time delay, τ = 3 and τ = 8. The PE coefficient vectors for τ = 3 are:
Fig. 5.9 shows the efficacy of the developed algorithm and the DRFDF, when the
system output is affected by an additive, white, Gaussian, zero-mean measurement noise
with variance equal to 0.01. The DRFDF yields a fast reaction of the residual to a fault,
148
5. Robust fault detection via eigenstructure assignment
1.5
0.5 f(t)
DRFDF
0 T =3
d
T =8
d
−0.5
40 50 60 70 80 90 100
Time [samples]
Figure 5.9: Fault detection in the case when the output is subjected to measurement
noise
however it is relatively sensitive to noise. The filter obtained using Algorithm 5.5 allows
for an increase of the parity space order, hence a reduction of the residual sensitivity to
noise. Nevertheless, a high parity space order results in slow response of the residual
to the fault.
149
5. Robust fault detection via eigenstructure assignment
150
Chapter 6
Nomenclature
151
6. Fault isolation via diagonal PE
6.1 Introduction
Directional residuals have been used in various industrial applications, such as induc-
tion motor drives (Campos-Delgado 2011), a class of linear networked control systems
(Chabir, Sauter & Keller 2009), neuro-fuzzy diagnosis of AC motors (Alexandru 2003,
Alexandru & Popescu 2004), and engine fault detection and isolation (Dutka, Javahe-
rian & Grimble 2009). Also a structured residual set has been applied for fault diagnosis
of many industrial systems, such as: a two non-interacting tank system (Bhattacharjee
& Roy 2010), a heat exchanger (Fagarasan & St. Iliescu 2008), an aircraft (Fravolini,
Brunori, Campa, Napolitano & La Cava 2009), and a Tennessee Eastman process ex-
ample (Xie, Zhang & Wang 2006, Ye, Shi & Liang 2011).
152
6. Fault isolation via diagonal PE
Furthermore, Patton & Chen (1997) proposed a scheme for condition monitoring
and fault diagnoisis of a seawater pumping system in operation at the Nuclear Electric
Heysham 2 power station. Simultaneous sensor and actuator fault diagnosis on a water
treatment system has been presented in (Fragkoulis, Roux & Dahhou 2009). Lia &
Jengb (2010) demonstrated a fault detection isolation and identification filter for a
nonisothermal continuous stirred tank reactor.
This chapter is an extension of the Algorithm 5.4 to a fault diagnosis scheme. In
Section 6.3 a fault isolation algorithm utilising a directional residual set is devised,
whilst in Section 6.4 the scheme is extended to fault identification.
where x(t) ∈ Rn is the system state vector, u(t) ∈ Rp and y(t) ∈ Rm are, respectively, the
system input and output, and µ(t) ∈ Rr is a fault signal. Matrices A, B, C, D, and F
are constant and have appropriate dimensions. It is assumed that (C, A) is observable
and F is of full column rank. The aim of this chapter is to provide algorithms for fault
isolation an identification.
153
6. Fault isolation via diagonal PE
d(t) µ(t)
u(t) y(t)
x̂(t)
+ ŷ(t) + r(t) µ̂(t)
C + _ residual
evaluation
where fi = Aδi Fi and δi is the smallest number for which CAδi Fi ≠ 0. Condition 1
allows all eigenvalues of (A − K1 C) and (A − K2 C) to be arbitrarily specified, whilst
condition 2 ensures that residual vectors yielded by different faults lie in different di-
rections.
1. For each column of F obtain δi , which is the smallest number for which
CAδi Fi ≠ 0 and compute zero directions fi = Aδi Fi
(i)
and corresponding directions vj , for j = 1, ⋯, ri , where ri is the number of
invariant zeros of the triple (A, fi , C), such that
⎡ (i) ⎤ ⎡ (i) ⎤
⎢ zj I − A vj−1 ⎥ ⎢ vj ⎥
⎢ ⎥⎢ ⎥=0
(i)
⎢ ⎥⎢ ⎥
⎢ ⎥ ⎢ −1 ⎥
(6.2a)
⎣ C 0 ⎦⎣ ⎦
(i)
where v0 denotes fi .
r
Øj,k = (−1)k−1 ∏(zl − λj )
i
(i) (i) (i)
(6.2b)
l=k
r
˜ (i) = λ i (z (i) − λ(i) )
B̃ j j∏ l j (6.2c)
l=1
154
6. Fault isolation via diagonal PE
(i) (i)
5. For i = 1, ⋯, r calculate coefficients x11 , ⋯, xri +1,1 which fulfil the condition
A∗e = [ A∗(1)
e
∗(2)
Ae
∗(r)
⋯ Ae ] (6.2e)
where
7. Obtain:
9. Calculate K1 as
10. Repeat steps 2 to 9 for the gain matrix K2 and obtain Ac1 = A − K2 C
155
6. Fault isolation via diagonal PE
Remark 6.1. The residual ri (t) obtained using Algorithm 6.1 lies in the direction Cfi .
Demonstration. The residual calculated using Algorithm (6.1) is, cf. (5.139):
⎡ τ −1 j ⎤
⎢ ∑j=0 Ac1 F µ(t − j) ⎥
r(t) = C [ J1 J2 ] ⎢
⎢ τ −1 j
⎥
⎥
⎢ ∑j=0 Ac2 F µ(t − j) ⎥
⎣ ⎦ (6.3)
τ −1 τ −1
= CJ1 ∑ Ajc1 F µ(t − j) + QCJ2 ∑ Ajc2 F µ(t − j)
j=0 j=0
(i)
Recall that Cvj = 0 for j = 1, ⋯, rl , i = 1, ⋯, r. Consequently:
Thus, a design of the fault isolation filter using Algorithm 6.1 ensures that the residual
driven by the fault signal µi (t) lies in the direction Cfi .
where γ1 , γ2 , ⋯, γr are scalar coefficients. The parameter γi deviating from zero indicates
a presence of the fault µi (t) and its value depends on the magnitude of the fault.
Consequently, a change of coordinates is defined as:
T = [ T 1 T2 ] (6.7)
156
6. Fault isolation via diagonal PE
Remark 6.2. An occurrence of µi (t) causes r̊i (t) deviate from zero, whilst the re-
maining components of r̊(t) are zero.
r̊(t) = [ γ1 γ2 ⋯ γr 0 ⋯ 0 ]
T
(6.11)
Thus r̊i (t) deviates from zero only when the fault µi (t) occurs.
where:
and:
M (t) = [ µT (t − τ ) µT (t − τ + 1) ⋯ µT (t) ]
T
(6.14)
It has been shown that the fault direction driven by µi (t) is Cfi . Consequently, the
directional residual, r(t), can be formulated as:
τ r
r(t) = ∑ ∑ αj Cfi µi (t − j)
(i)
(6.15)
j=0 i=1
157
6. Fault isolation via diagonal PE
(i)
where αj are scalar coefficients. Then, the term r̊(t) can be reformulated as:
Note that r̊i (t) is the residual vector r(t) represented by the basis T , and, consequently,
r̊i (t) can be reformulated as, cf. (6.15):
Therefore, the steady state gain of the response of r̊i (t) to the fault µi (t) is equal to
(i)
∑τj=0 αj , which is the sum of elements of the ith row of Ω. Consider a diagonal matrix:
The first r elements of the vector µ̂(t) are estimates of µ(t), whereas remaining m − r
elements of µ̂(t) should be equal zero and may be treated as control variables (if they
deviate from zero it indicates there is a fault or a disturbance that is not covered by
the model). Consequently, the PE for fault isolation and identification is given by:
where
The algorithm for fault isolation and identification via diagonal PE is summarised
below.
158
6. Fault isolation via diagonal PE
Ω = T −1 Wµ (6.23a)
where noise-free input and output, u0 (t) and y0 (t) are affected by white, Gaussian,
zero-mean, mutually uncorrelated noise sequences ũ(t) and ỹ(t), respectively. The
terms u(t) and y(t) are measured values of the input and output. The term µ̂(t) is
159
6. Fault isolation via diagonal PE
calculated as:
µ̂(t) = W̊y Y (t) + W̊u U (t) = W̊y (Y0 (t) + Ỹ (t)) + W̊u (U0 (t) + Ũ (t)) (6.25)
It should be noted that each column of µ̂(t) is affected by the measurement noise in
different level. Therefore, thresholds calculation should be based on the expected values
of the variance each element of µ̂(t) in the fault-free case. The covariance matrix of
µ̂(t), denoted as Σµ̂ , in fault-free case is given by:
where Σỹ = E{Ỹ (t)Ỹ T (t)} and Σũ = E{Ũ (t)Ũ T (t)}. In the EIV framework with no
process noise, cf. (6.24), Σỹ and Σũ are diagonal matrices. Based on the assumption
that the measurement noise sequences are white, Gaussian, zero-mean, and mutually
uncorrelated, the threshold which µ̂i (t) needs to violate for the fault to be noticed
should be calculated as an appropriate multiplicity of the standard deviation of µ̂i (t)
in a fault-free case, i.e. an appropriate multiplicity of the square root of the ith diagonal
element of Σµ̂ .
⎡ ⎤ ⎡ 0 ⎤ ⎡ 1 1 0 ⎤
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
0 1 0 0 0
⎢ ⎥ ⎢ ⎥ ⎢ 0 0 ⎥
⎢ 0 ⎥ ⎢ 0 ⎥ ⎢ 1 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
0 0 1 0
A=⎢⎢ 0 ⎥ B=⎢
⎥
⎢ 0 ⎥ F = ⎢ 0 −1 0 ⎥
⎥ ⎢
⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
0 0 0 1
⎢ 1 ⎥ ⎢ 0 ⎥ ⎢ 0 ⎥
⎢ 0 0 0 0 ⎥ ⎢ ⎥ ⎢ 2 1 ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ −4.85 3.5 ⎥ ⎢ 1 ⎥ ⎢ 0 ⎥ (6.28)
⎣ 0.1512 −1.1274 3.325 ⎦ ⎣ ⎦ ⎣ 0 2 ⎦
⎡ 1 −1 0 0 0 ⎤ ⎡ ⎤
⎢ ⎥ ⎢ 0 ⎥
⎢ ⎥ ⎢ ⎥
⎢ 0 1 1 0 ⎥ ⎢ 0 ⎥
⎢ ⎥ ⎢ ⎥
C =⎢ ⎥ D=⎢ ⎥
0
⎢ ⎥ ⎢ 0 ⎥
⎢ 0 1 0 1 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
0
⎢ ⎥ ⎢ 0 ⎥
⎣ 0 0 0 0 1 ⎦ ⎣ ⎦
The first output is affected by white, Gaussian, zero-mean noise with the variance equal
to σy21 = 0.01, whilst the remaining three are subjected to white, Gaussian, zero-mean
noise sequences with the variances of σy22 = σy23 = σy24 = 0.0001. Output measurement
160
6. Fault isolation via diagonal PE
f1 = AF1 = [ 1 0 0 0 −0.9762 ]
T
(6.29)
⎡ 1 ⎤ ⎡ 1 ⎤ ⎡ 0 ⎤
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ 0 ⎥ ⎢ ⎥ ⎢ 1 ⎥
⎢ ⎥ ⎢ 1 ⎥ ⎢ ⎥
Cf1 = ⎢ ⎥ Cf2 = ⎢ ⎥ Cf3 = ⎢ ⎥
⎢ −0.9762 ⎥ ⎢ ⎥ ⎢ 2 ⎥
(6.30)
⎢ ⎥ ⎢ −1 ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥ ⎢ ⎥
⎢ −0.9762 ⎥ ⎢ 0 ⎥ ⎢ 2 ⎥
⎣ ⎦ ⎣ ⎦ ⎣ ⎦
Note that rank([ Cf1 Cf2 Cf3 ]) = 3, hence the possibility for a complete fault
isolation exists.
Eigenvalues of (A − K1 C) are chosen to be 0.22, 0.77, 0.66, 0.55, 0.44. Note that
the triple (A, f1 , C) has an invariant zero at z1
(1)
= 0, whilst the state and input zero
= −1. Eigenvalues of (A − K1 C) corre-
(1) (1)
directions are, respectively, v1 = F1 and g1
sponding to a linear combination of columns of f¯1 = [ f1 F1 ] are 0.22 and 0.44. As
a result, the first column of the matrix X (1) is given by, cf. (6.2b), (6.2c), (6.2d):
⎡ ⎤
⎢ 0.6600 ⎥
⎢
=⎢ ⎥
⎥
(1)
⎢ −0.0968 ⎥
X (6.31)
⎣ ⎦
⎡ −0.5632 −0.5500 ⎤
⎢ ⎥
⎢ ⎥
0
⎢ 0 ⎥
⎢ ⎥
∗ ⎢ ⎥
−1 0
⎢
A =⎢ 0 ⎥
⎥
⎢ ⎥
2.5500 1 (6.34)
⎢ −0.9762 −1.1000 1.34 ⎥
⎢ ⎥
⎢ ⎥
⎢ −2.7157 −12.8738 0.83 ⎥
⎣ ⎦
The remaining eigenvalue of (A′ − K1′ C ′ ) is chosen to be 0.77. Subsequently, the gain
161
6. Fault isolation via diagonal PE
The eigenvalues of (A − K2 C) are selected to be 0.95, 0.85, 0.18, 0.33, 0.47 and, as
a consequence, K2 is calculated as:
The order of the parity space is chosen as τ = 8 and matrices Wu and Wy are
calculated using (6.2l) and (6.2m), respectively. Then T is obtained using (6.7):
⎡ 1 0.3072 ⎤
⎢ ⎥
⎢ ⎥
1 0
⎢ 0 1 1 −0.6294 ⎥
⎢ ⎥
T =⎢ ⎥
⎢ −0.9762 −1 2 −0.3222 ⎥
(6.37)
⎢ ⎥
⎢ ⎥
⎢ −0.9762 0 2 0.6369 ⎥
⎣ ⎦
⎡
⎢
⎢
0 0 0 2.3e − 2 0 0
⎢
⎢
Ω = T Wµ = ⎢
0 3.2e − 2 0 0 1.3e − 1 0
⎢
−1
⎢
⎢
0 0 4.3e − 3 0 0 3e − 2
⎢
⎣ 0 0 0 0 0 0
(6.38)
1 0 0 0 0 0 ⎤
⎥
⎥
1.8e − 1 0 0
0 1 0 0 0 0 ⎥
⎥
⎥
0 3.8e − 1 0
1.8e − 1 0 0 1 0 0 0 ⎥
⎥
⎥
0 0
0 0 0 0 0 0 0 0 0 ⎥
⎦
The first three diagonal elements of Ξ are equal to sum of appropriate rows of Ω. The
162
6. Fault isolation via diagonal PE
⎡ 0 ⎤
⎢ ⎥
⎢ ⎥
1.2004 0 0
⎢ 0 ⎥
⎢ ⎥
Ξ=⎢ ⎥
0 1.5386 0
⎢ 1.2118 0 ⎥
(6.39)
⎢ ⎥
⎢ ⎥
0 0
⎢ 1 ⎥
⎣ 0 0 0 ⎦
Consequently, W̊u and W̊y are calculated using, respectively, (6.23b) and (6.23c) and
the fault estimate is obtained using (6.23d).
Measurement noise
Recall that output measurements are affected by white, zero-mean, Gaussian, and
mutually uncorrelated noise sequences with variances of σy21 = 0.01 and σy22 = σy23 =
σy24 = 0.0001, whilst τ = 8. Hence, the output measurement covariance matrix is
a 36th order ((τ + 1)m = 9 × 4 = 36) diagonal matrix, whose diagonal is the vector
[ σy21 σy22 σy23 σy24 ] repeated τ + 1 = 9 times. Consequently, Σµ̂ is calculated as,
cf. (6.27):
⎡ 0.0190 −0.0119 0.0048 ⎤
⎢ ⎥
⎢ ⎥
0.0129
⎢ −0.0119 0.0190 −0.0118 −0.0034 ⎥
⎢ ⎥
Σµ̂ = ⎢ ⎥
⎢ 0.0129 −0.0118 0.0045 ⎥
(6.40)
⎢ ⎥
⎢ ⎥
0.0118
⎢ 0.0048 −0.0034 ⎥
⎣ 0.0045 0.0038 ⎦
Thus, the variances of the consecutive elements of µ̂(t) in a fault-free case are, respec-
tively, 0.0190, 0.0190, 0.0118, and 0.0038. The threshold of each element of µ̂(t) is
selected as its standard deviation in a fault-free case multiplied by 3.1 (which results in
the confidence bound of 0.999, assuming a Gaussian distribution of the measurement
noise) and are given by, respectively, 0.4272, 0.3557, 0.3363, and 0.1915. Results of the
simulation are presented in Fig. 6.2. The filter identifies abrupt faults µ1 (t) and µ2 (t),
as well as an incipient fault µ3 (t). After the 100th sample an unmodelled fault occurs.
This is indicated by µ̂4 (t) violating the threshold.
6.7 Conclusions
The fault isolation and identification filter devised in this chapter utilises a diagonal
residual set. Therefore, multiple faults can be isolated and identified. Furthermore, if
the number of linearly independent outputs exceeds the number of modelled faults, a
new variable has been introduced, which indicates occurrence of an unmodelled fault.
This provides a signal to stop the process and investigate the source of a possibly
dangerous fault. Also a straightforward method to calculate residual thresholds, whose
violation indicates a fault, has been proposed for the EIV framework. A simulation
study has demonstrated promising results for both diagnosis of abrupt and incipient
163
6. Fault isolation via diagonal PE
f2 (t)
8
6
µ̂2 (t)
4
2
0
10 20 30 40 50 60 70 80 90 100 110 120
0
−2
−4
f3 (t)
µ̂3 (t)
−6
−8
10 20 30 40 50 60 70 80 90 100 110 120
−2
µ̂4 (t)
−4
10 20 30 40 50 60 70 80 90 100 110 120
Time [samples]
Figure 6.2: Fault identification using diagonal PE. At the 100th time sample unmod-
elled fault occurs causing µ̂4 (t) deviate from zero.
faults.
Further work aims to extend the scheme to fault isolation and identification of
multidimensional faults.
164
Chapter 7
Potential applications
Nomenclature
Ap . . . . . . . . . . . . . piston area
C .............. compensation variable
d1 , d2 , d2 . . . . . . . . damping coefficients
Fc (t) . . . . . . . . . . . Coulomb friction
Ff ric (t) . . . . . . . . total friction force
Fh (t) . . . . . . . . . . . hydraulic force
Fr (t) . . . . . . . . . . . roll force
Fs (t) . . . . . . . . . . . Stribeck friction (stiction)
Fv (t) . . . . . . . . . . . viscous friction
G1 , G2 . . . . . . . . . . controller gains
Gv (z) . . . . . . . . . . transfer function
k1 . . . . . . . . . . . . . . spring constant
k2 . . . . . . . . . . . . . . spring constant
k3 . . . . . . . . . . . . . . steel strip spring constant
Kc . . . . . . . . . . . . . hydraulic oil compressibility coefficient
Kp . . . . . . . . . . . . . proportional gain of hydraulic piston controller
l ................ stroke length of the piston
h(t) . . . . . . . . . . . . exit gauge
ĥ(t) . . . . . . . . . . . . estimate of exit gauge
href (t) . . . . . . . . . exit gauge reference signal
H(t) . . . . . . . . . . . input gauge
m1 . . . . . . . . . . . . . mass of hydraulic piston
m2 . . . . . . . . . . . . . mass of backup roll
m3 . . . . . . . . . . . . . mass of work roll
M .............. mill modulus (spring constant)
M̂ . . . . . . . . . . . . . . estimated mill modulus
p(t) . . . . . . . . . . . . hydraulic pressure
qf (t) . . . . . . . . . . . fluid flow to hydraulic capsule
v(t) . . . . . . . . . . . . unknown input
v̂(t) . . . . . . . . . . . . unknown input estimate
vc , v1 , v2 . . . . . . . . auxiliary coefficients
y(t) . . . . . . . . . . . . system output
z(t) . . . . . . . . . . . . position of the hydraulic piston
165
7. Potential applications
7.1 Introduction
Algorithms developed in Chapter 3 are evaluated using two practical examples. In
Section 7.2 the PE-UIO (see Algorithm 3.1) is used to improve control performance
of a steel rolling mill. Furthermore, the two stage PE-UIO (Algorithm 3.4) is used to
estimate the concentration of river pollutant. Conclusions and a critical appraisal of
practical use of the developed algorithms are presented in Section 7.4.
166
7. Potential applications
work roll
backup roll
Control scheme
Harsh temperature conditions close to the rolling mill stand render direct measurement
of the exit gauge impossible (Yildiz et al. 2009), hence a need arises to estimate the
exit gauge from the measured value of the roll force and the mill modulus, i.e. mill
sensitivity to force, see (Yildiz et al. 2009). Due to the fact that the roll force is
inaccessible for measurements, the force in the hydraulic actuator capsule is measured.
Therefore the exit gauge is estimated via:
Fh (t)
C
ĥ(t) = z(t) + (7.1)
M̂
167
7. Potential applications
where ĥ(t) denotes the estimated exit gauge, M̂ is the estimated value of the mill
modulus M (i.e. mill sensitivity to force), z(t) denotes the hydraulic piston position
and Fh (t) corresponds to the value of the force measured in the hydraulic actuator
capsule, noting that Fh (t) ≈ Fr (t), where Fr (t) denotes the roll force. In fact it is
the difference between the measured Fh (t) and the actual Fr (t) which constitutes the
unknown input. The actual exit gauge, denoted h(t), is is given by:
Fr (t)
C
h(t) = z(t) + (7.2)
M
In order to improve the robustness of the control loop a compensation variable C < 1
is introduced.
The control scheme of the rolling mill is presented in Fig. 7.2. The controller gains
G1 and G2 are given by:
G1 = 1 + (1 − C)
k3 M̂ + k3
, G2 = (7.3)
M̂ M̂ + k3 + Ck3
where k3 denotes the steel strip sensitivity to force (strip modulus). The term Kp
denotes the proportional gain of the hydraulic piston position controller and defines
a relation between the piston position error and the fluid flow, denoted qf (t), to the
hydraulic actuator capsule.
href (t) + + zref (t) + qf (t) Fh (t) h(t)
G1 G2 Kp
_ _ Actuator Stack z(t)
+
C
M̂
+
ĥ(t) +
Figure 7.2: Control loop, href (t) – exit gauge reference signal, zref (t) – piston po-
sition reference signal, Fh (t) – hydraulic force, qf (t) – flow of hydraulic
fluid
The stack of rolls (i.e. backup and work rolls with a steel strip between them, further
referred to as the stack) is modelled by making use of a classical mass-spring-damper
model representation, see Fig. 7.3. Due to the symmetrical construction of the stack,
only the upper backup and work rolls are taken into consideration. The values of the
model parameters are given in Table 7.1. In the further analysis the damper denoted
168
7. Potential applications
m1
z(t)
d1
k1
m2
d2 k2
m3
h(t)
d3
k3
Friction model
169
7. Potential applications
where the term Ff ric (t) denotes the total frictional force, whilst Fc (t), Fv (t) and Fs (t)
refer to the Coulomb, viscous and Stribeck friction, respectively. The Coulomb friction
is modelled as:
Fc (t) = −µc sign(ż(t))(1 − e∣ )
ż(t)
∣
vc (7.5)
where the term µc denotes the Coulomb friction level, whilst the exponential term is
introduced in order to avoid a zero-crossing discontinuity. The element related to the
viscous friction is modelled as a linear function of the hydraulic piston velocity, denoted
ż(t):
Fv (t) = −µv ż(t) (7.6)
where the term µv denotes the viscous damping of the frictional force. The Stribeck
friction (stiction) model is given by:
where the term µs determines the magnitude of the static friction, whilst v1 and v2 are
utilised to shape the stiction model.
Dependency between the fluid flow, denoted qf (t), into the capsule and the pressure
denoted p(t) acting on the piston area denoted Ap is represented by the following linear
relation, see e.g. (Jelali & Kroll 2003):
∫ qf (t)dt − Ap l
p(t) = Kc (7.8)
Ap l
where l denotes the stroke length of the piston, and the term Kc corresponds to the
hydraulic oil compressibility coefficient. Therefore, the force denoted Fh (t) acting on
the hydraulic piston is given by:
The values of the hydraulic actuator model parameters are given in Table 7.2.
170
7. Potential applications
Recall that due to the fact that the roll force Fr (t) is inaccessible, the hydraulic force
Fh (t) is measured. Subsequently, the force measurement is affected by the friction force
and the parasitic dynamics of the stack. Due to the fact that the bandwidth of the
parasitic dynamics of the stack exceeds the sampling frequency, it can be assumed that
the frictional force has the most significant contribution to the roll force measurement
error. Therefore, the difference between the roll force Fr (t) and the force measured
in the hydraulic capsule Fh (t) is considered to be the unknown input to the system,
further referred to as v(t), where v(t) = C
(Fr (t) − Fh (t)). (The factor C
is used to
M̂ M̂
ensure to ensure numerical scalability.) Hence, the matrices G and H of the system
(3.1) are:
H =[ 1 0 ]
T
G = 0, (7.11)
Fh (t) − v̂(t)
C
ĥ(t) = z(t) + (7.12)
M̂
where v̂(t) is the correction term (i.e. the estimated force the measurement error v(t),
corresponding to the unknown input obtained from the PE-UIO).
Engineering knowledge and past experience of technicians with the plant indicate
that it is reasonable to assume that the piston position and hydraulic force measure-
ments are affected by white, zero-mean, Gaussian, mutually uncorrelated noise se-
quences, whose standard deviations are, respectively, 0.1µm and 1000N. The PE-UIO
algorithm with s = 4 samples is used to obtain v̂(t).
171
7. Potential applications
0.03
no compensation
with compensation
0.02
Gauge error [mm]
0.01
−0.01
−0.02
−0.03
10 11 12 13 14 15 16 17 18 19 20
Time [s]
5
x 10
2
estimated
1.5 true
1
Friction force [N]
0.5
−0.5
−1
−1.5
−2
15 15.5 16 16.5 17 17.5 18 18.5 19 19.5 20
Time [s]
Fig. 7.4 presents the simulated results for the cases of no compensation and with the
unknown input compensation applied. The grey dashed line corresponds to a simulated
reconstruction of a typical limit cycle condition found in practice. The black solid line
corresponds to the compensated case and clearly indicates that the limit cycle ampli-
tude is significantly reduced, implying potential for improved product quality. Fig. 7.5
shows the actual unknown input and the estimated unknown input corresponding to
the simulated condition in Fig. 7.4. The PE-UIO accurately estimates the friction
force affecting the exit gauge and a subsequent feedback compensation that utilises the
estimated unknown input results in a significant improvement in control.
172
7. Potential applications
schematic illustration of the experiment is depicted in Fig. 7.6. A tracer has been
poured into the river at the point (1). Two tracer concentration sensors have been
placed downstream, at points (2) and (3), whose readings are denoted, respectively,
as v(t) and y(t). A linear model of the relation between v(t) and y(t), where v(t) is
the input to the system, whilst y(t) is the output, has been developed in (Young &
Sumislawska 2012). The input v(t), further referred to as upstream tracer concentra-
tion, and the output y(t), the tracer concentration measured downstream, are plotted
as the grey line and black line, respectively, in Fig. 7.7. The aim of this simulation is
to use the two stage PE-UIO (Algorithm 3.4) scheme to estimate the input v(t) based
on output measurements and the knowledge of the system model. The system can be
5
v(t)
y(t)
v(t) and y(t) [mg/l]
0
0 100 200 300 400 500 600 700 800 900
Time [hours]
approximated by a linear second order model, with time constants of 17.4 and 83.7
hours, i.e. a stiff system. The discrete time model of the two-hourly sampled system is
given by (Young & Sumislawska 2012):
Gv (z) =
0.017591(z + 4.302)(z − 0.9735)
(z − 0.9764)(z − 0.8916)
(7.13)
Note that the same model has been used in Example 3.7 to obtain the unknown in-
put of the simulated system (in contrast to this example, where real input and output
measurements are used). The model is nonminimum-phase. Although the standard
PE-UIO can cope with the zero at −4.302, the zero at 0.9735 requires the two stage
173
7. Potential applications
PE-UIO to be used. For the design of the unknown input reconstruction it has been
assumed that the measurements are affected by white, zero-mean, Gaussian, mutually
uncorrelated measurement noise. The parity space order has been set to 15 samples,
which leads to an estimation time lag of 7 samples. The result of an unknown input
estimation using the two stage PE-UIO is compared with the input reconstructed us-
ing the INPEST, see Fig. 7.8. The parameters of the INPEST method are τ̊ = 7 and
q̊e = 0.8 obtained for λ = 0.001, see (Young & Sumislawska 2012). It is noted that both
Error
0.5
0
−0.5
−1
0 50 100 150 200 250 300 350 400
Unknown input estimate [mg/l]
5
true
PE−UIO
4 INPEST
0
0 50 100 150 200 250 300 350 400
Time [samples]
methods give similar results. Both methods detect a rise of v(t) in approximately the
33th sample, whereas the measured input starts rising at approximately the 42th sam-
ple. Furthermore, after the 200th sample the estimation errors of the PE-UIO and the
INPEST are virtually the same. It is believed that the input reconstruction discrep-
ancies are caused mainly by the modelling inaccuracy, presumably caused by system
nonlinearities. This hypothesis is supported by Fig. 7.9, which compares the measured
output with the model output. The simulated output starts rising approximately 10
samples after the rise of the measured output. Furthermore, the observed characteris-
tic ‘bumps’ of the measured output between 100 and 150 samples result in the input
reconstruction error pattern visible in the upper subfigure of Fig. 7.8.
174
7. Potential applications
4
measured
Dye concentration [mg/l] simulated
3
0
0 50 100 150 200 250 300 350 400 450
Time [samples]
and the compensation significantly reduced the amplitude of limit cycles. The high
frequency and low amplitude oscillations, which may be observed after enhancement of
the control, are probably the result of unknown input estimation delay. In the industrial
plant, however, unmeasured variations of the input gauge H occur, which should be
treated as a disturbance. Thus, before application to an actual plant, the possibility of
disturbance decoupling in the PE-UIO needs to be explored.
The two stage PE-UIO has estimated the tracer concentration in the river accu-
rately. Further improvement could possibly be achieved, if, instead of the assumption
of a white measurement noise, a coloured process noise model, which can explain dis-
crepancy between modelled and simulated output (model mismatch), is assumed.
175
Chapter 8
8.1 Conclusions
This thesis presents novel developments in the fields of unknown input reconstruction
and fault detection, isolation and identification. The developed algorithms are applica-
ble to time-invariant, discrete-time systems. Most of the research is devoted to linear
systems, except for the unknown input reconstruction method presented in Chapter 4,
which has been designed for a class of nonlinear systems, namely, Hammerstein-Wiener
systems. Two potential applications for the algorithms developed in Chapter 3 have
been proposed and promising results demonstrated via simulation studies.
This section is divided into three logical parts. Subsection 8.1.1 summarises the
development of unknown input reconstruction schemes. In Subsection 8.1.2 fault de-
tection and diagnosis algorithms are concluded. The main contributions of this thesis
are summarised in Subsection 8.1.3.
176
8. Conclusions & further work
reconstruction filter (it is, however, required to know the noise model). In the case of
OE systems the design procedure can be simplified, which has also been demonstrated.
The only tuning parameter for the PE-UIO is the parity space order. By increasing it,
the bandwidth of the filter is reduced, and, consequently, noise filtering properties are
improved. On the other hand, reduced bandwidth causes an estimation lag. Thus, the
trade-off between noise filtering and estimation lag as well as an a’priori knowledge of
the bandwidth of the reconstructed signal needs to be taken into consideration. The
algorithm is applicable to both minimum-phase and nonminimum-phase systems.
The drawback of the PE-UIO is that it may produce a distorted unknown input
estimate, if a zero of the system transfer function to an unknown input is unity (a system
with a derivative term) or lies close to unity. To tackle this problem an extension to
the PE-UIO, i.e. a two stage PE-UIO, has been proposed. The two stage PE-UIO is
applicable to systems, whose minimum-phase zeros lie close to unity or its zeros are
equal unity. It has been demonstrated that the two stage algorithm has superior noise
filtering properties compared to the standard PE-UIO, however, it may introduce larger
estimation lag. Both, the standard and the two stage input reconstruction algorithms,
are computationally simple. The filter parameters need to be calculated only once
before the filter is applied to the system.
Both, the standard (single stage) PE-UIO and the two stage PE-UIO, have been
compared with two other methods found in the literature: a Kalman filter-based MVU
and the INPEST method, based on linear quadratic control. A simulation study has
revealed superior noise filtering properties of the PE-UIO compared to the MVU. This
is due to the adjustable bandwidth of the PE-UIO (which, however, causes the trade-
off between the noise filtering and estimation delay). Furthermore, the MVU in the
case of a single output system resembles a naive inversion, thus it cannot be used
for unknown input reconstruction of single output nonminimum-phase systems. The
INPEST method has shown comparable results to those of the PE-UIO (both single
stage and the two stage).
Potential industrial applications of the proposed unknown input reconstruction
schemes have been demonstrated via simulation studies in Chapter 7. The PE-UIO
has been used to improve the control performance of a steel rolling mill, by recon-
struction of a parasitic friction force. The two stage PE-UIO has been proposed in a
hydrological application in order to estimate the level of pollutant in a river.
In Chapter 4 the PE-UIO has been extended to Hammerstein-Wiener systems, i.e.
systems which can be modelled as a linear dynamic block preceded and followed by
a static nonlinearity. The algorithm has been developed for a system with a single
measurable input, single output, and a single unknown input to be reconstructed in
an EIV framework, where both measured input and output are subjected to white,
Gaussian, zero-mean mutually uncorrelated noise sequences. As the system operating
point changes, the influence of the input and output noise on the unknown input es-
177
8. Conclusions & further work
timate varies. Thus, the algorithm needs to adapt to these changes. Two versions of
the scheme are proposed. In the first version the parity space order remains constant,
whilst the filter parameters vary at each time sample. In the second version the parity
space order varies according to the system operating point. Furthermore, assuming a
Gaussian distribution of the measurement noise, a method for computation of the confi-
dence bounds has been proposed. The simulation study has demonstrated applicability
of proposed algorithms to the particular class of nonlinear systems. It has also been
shown that for relatively mild nonlinearities a linear algorithm can be used in order to
reduce the computational effort.
178
8. Conclusions & further work
8.1.3 Contributions
The main contributions of this thesis are briefly summarised in order of importance as
follows:
3. Application of the novel PE-based unknown input reconstruction method for en-
hancement of a control loop in a single stand of a rolling mill and for a hydrological
application (Chapter 7).
4. Use of right and left eigenstructure assignment to develop robust fault detection
PE of user defined order (Chapter 5).
5. Extension of the robust fault detection filter via right eigenstructure assignment
to systems with unstable invariant zeros (Chapter 5).
● Up to date the PE-UIO can be applied to systems where a single unknown input
needs to be reconstructed. Thus, an extension of the algorithm to systems with
multiple unmeasurable inputs could be considered.
● The single stage PE-UIO produces a distorted unknown input estimate when a
zero of the system response to an unknown input lies close to unity. The two
stage PE-UIO copes with such a situation if the problematic zero lies inside the
unit circle. However, the problem remains open for the cases when a system
nonminimum-phase zero lies close to unity.
The following aspects of the input reconstruction scheme for Hammerstein-Wiener sys-
tems could be taken into consideration:
179
8. Conclusions & further work
Additional research on the topic of fault detection and diagnosis could include:
● Robust fault detection for stochastic systems. Design freedom of both robust
PE and a robust asymptotic filter could be used to minimise the influence of the
noise of the residual. Also calculation of thresholds, whose violation indicates the
presence of faults has not been discussed for the robust fault detection filter (it has
been discussed only for the fault isolation and identification filter in Chapter 6).
● Exploring applicability of the robust fault detection scheme to systems where the
disturbance direction vectors ei combine with each other to create new invariant
zeros.
● Extension of the proposed robust fault detection and diagnosis methods to non-
linear systems; particularly, Hammerstein-Wiener and bilinear systems could be
considered.
● Whilst two of the developed algorithms have been applied to practical applica-
tions, it would be desirable to evaluate the other methods developed within this
thesis to real world applications to assess their potential benefits.
180
References
Alexandru, M. (2003), ‘Analysis of induction motor fault diagnosis with fuzzy neural
network’, Applied Artificial Intelligence: An International Journal 17(2), 105–133.
Basile, G. & Marro, G. (1969), ‘Controlled and conditioned invariant subspaces in linear
system theory’, Journal of Optimisation Theory and Applications 33(5), 305–315.
Basile, G. & Marro, G. (2002), Controlled and Conditioned Invariants in Linear Sys-
tems Theory, Department of Electronics, Systems and Computer Science, Univer-
sity of Bologna, Italy.
181
REFERENCES
Berriri, H., Naouar, M. W. & Slama-Belkhodja, I. (2011), Parity space approach for
current sensor fault detection and isolation in electrical systems, in ‘Proceedings
of the 8th International Multi-Conference on Systems, Signals & Devices’, Sousse,
Tunisia.
Bhattacharjee, N. & Roy, B. (2010), Fault detection and isolation of a two non-
interacting tanks system using partial PCA, in ‘Proceedings of 2010 International
Conference on Industrial Electronics, Control & Robotics (IECR)’, Rourkela, In-
dia.
Campos-Delgado, D. (2011), ‘An observer-based diagnosis scheme for single and si-
multaneous open-switch faults in induction motor drives’, IEEE Transactions on
Industrial Electronics 58(2), 671–679.
Celka, P. & Colditz, P. (2002), ‘Nonlinear nonstationary Wiener model of infant EEG
seizures’, IEEE Transactions on Biomedical Engineering 49(6), 556–564.
Chabir, K., Sauter, D. & Keller, J. (2009), Design of fault isolation filter under network
induced delay, in ‘Proceedings of 2009 IEEE Conference on Control Applications,
(CCA) & Intelligent Control, (ISIC)’.
Chan, C., Hua, S. & Hong-Yue, Z. (2006), ‘Application of fully decoupled parity equa-
tion in fault detection and identification of dc motors’, IEEE Transactions on
Industrial Electronics 53(4), 1277–1284.
Chen, J. & Patton, R. J. (1999), Robust Model-Based Fault Diagnosis for Dynamic
Systems, Kluver Academic Publishers, Boston.
Chen, R. H. & Speyer, J. L. (2006a), Detection filter analysis and design using
eigenstructure assignment, in ‘Proceedings of the American Control Conference
(ACC’06)’, Minneapolis, Minnesota USA.
182
REFERENCES
Chen, R. & Speyer, J. (2006b), Generalization of the detection filter using spectral the-
ory, in ‘Proceedings of the American Control Conference (ACC’06)’, Minneapolis,
Minnesota USA.
Chen, R. & Speyer, J. (2007), Spectral analysis and design of detection filter for
multiple-dimensional faults, in ‘Proceedings of the American Control Conference
(ACC’07)’, New York City, USA.
Chow, E. & Willsky, A. (1984), ‘Analytical redundancy and the design of robust failure
detection systems’, IEEE Transactions on Automatic Control 29(7), 603–614.
De-Feng, H., Li, Y. & Guo-Shi, Y. (2010), A pole placement-based NMPC algorithm
of constrained Hammerstein systems, in ‘Proceedings of 29th Chinese Control
Conference (CCC 2010)’, Beijing, China.
Dever, J. A. (1960), ‘Control apparatus’, United States Patent, No. 2,931,763, 5 May.
Ding, S. X., Zhong, M., Bingyong, T. & Zhang, P. (2001), An LMI approach to the
design of fault detection filter for time-delay LTI systems with unknown inputs,
in ‘Proceedings of the American Control Conference (ACC’01)’, Arlington, VA,
USA.
183
REFERENCES
Dorato, P. (1969), ‘On the inverse of linear dynamical systems’, IEEE Transactions on
Systems Science and Cybernetics 5(1), 43–48.
Douglas, R. K., Speyer, J. L., Mingori, D. L., Chen, R. H., Malladi, D. P. & Chung,
W. H. (1996), ‘Fault detection and identification with application to advanced ve-
hicle control systems: Final report’, Research Reports, California Partners for Ad-
vanced Transit and Highways (PATH), Institute of Transportation Studies (UCB),
UC Berkeley.
Douglas, R. & Speyer, J. (1995), Robust fault detection filter design, in ‘Proceedings of
the American Control Conference (ACC’95)’, Vol. 1, Seattle, WA, USA, pp. 91–96.
Duan, G. & Patton, R. J. (2001), ‘Robust fault detection using Luenberger-type un-
known input observers – a parametric approach’, International Journal of Systems
Science 32(4), 533–540.
Dutka, A., Javaherian, H. & Grimble, M. J. (2009), Model-based engine fault detection
and isolation, in ‘Proceedings of the American Control Conference (ACC’09)’, St.
Louis, MO, USA.
El-Ghezawi, O., Billings, S. & Zinober, A. (1983), Variable-structure systems and sys-
tem zeros, in ‘IEE Proceedings Control Theory and Applications’, Vol. 130, pp. 1–
5.
Fagarasan, I. & St. Iliescu, S. (2008), Parity equations for fault detection and isolation,
in ‘Proceedings of IEEE International Conference on Automation, Quality and
Testing, Robotics, (AQTR 2008)’, Cluj-Napoca, Romania.
Fantuzzi, C., Simani, S. & Beghelli, S. (2001), Robust fault diagnosis of dynamic pro-
cesses using parametric identification with eigenstructure assignment approach, in
‘Proceedings of the 40th IEEE Conference on Decision and Control’, Orlando, FL,
USA.
Floquet, T. & Barbot, J. (2006), ‘State and unknown input estimation for linear
discrete-time systems’, Automatica 42(11), 1883–1889.
184
REFERENCES
Fragkoulis, D., Roux, G. & Dahhou, B. (2009), A global scheme for multiple and
simultaneous faults in system actuators and sensors, in ‘Proceedings of the 6th In-
ternational Multi-Conference on Systems, Signals and Devices (SSD ’09)’, Djerba,
Tunisia.
Fravolini, M., Brunori, V., Campa, G., Napolitano, M. & La Cava, M. (2009), ‘Struc-
tural analysis approach for the generation of structured residuals for aircraft FDI’,
IEEE Transactions on Aerospace and Electronic Systems 45(4), 1466–1482.
Fruzzetti, K., Palazoglu, A. & McDonald, K. (1997), ‘Nonlinear model predictive con-
trol using Hammerstein models’, Journal of ProcessControl 7(1), 31–44.
Fu, H., Kirtikar, S., Zattoni, E., Palanthandalam-Madapusi, H. & Bernstein, D. (2009),
Approximate input reconstruction for diagnosing aircraft control surfaces, in ‘Pro-
ceedings of the AIAA Guidance, Navigation, and Control Conference’, Chicago,
Illinois.
Fu, H., Yan, J., Santillo, M. A., Palanthandalam-Madapusi, H. & Bernstein, D. (2009),
Fault detection for aircraft control surfaces using approximate input reconstruc-
tion, in ‘Proceedings of the American Control Conference (ACC’09)’, St. Louis,
MO, USA.
Gao, Z., Breikin, T. & Wang, H. (2007), ‘High-gain estimator and fault-tolerant design
with application to a gas turbine dynamic system’, IEEE Transactions on Control
Systems Technology 15(4), 740–753.
Garrick, B. J., Gekler, W. C., Goldfisher, L., Karcher, R. H., Shimizu, B. & Wilson,
J. H. (1967), Reliability analysis of nuclear power plant protective system, Tech-
nical Report HN-190 AEC Research & Development Report, Holmes & Narver,
Inc., Nuclear Division.
Gertler, J. & Kunver, M. (1995), ‘Optimal residual decoupling for fault diagnosis’,
International Journal of Control 61(2), 395–421.
Gertler, J. & Singer, D. (1990), ‘A new structurel framework for parity equation-based
failure detection and isolation’, Automatica 26(2), 381–388.
185
REFERENCES
Gillijns, S. & De Moor, B. (2007a), ‘Unbiased minimum variance input and state esti-
mation for linear discrete-time systems’, Automatica 43(1), 111–116.
Gillijns, S. & De Moor, B. (2007b), ‘Unbiased minimum variance input and state es-
timation for linear discrete-time systems with direct feedthrough’, Automatica
43(5), 934–937.
Halton, H. (1963), ‘Design philosophy of an automatic checkout and launch system for
a drone’, IEEE Transactions on Aerospace 1(2), 538–546.
Hou, M. & Müller, P. (1992), ‘Design of observers for linear systems with unknown
inputs’, IEEE Transactions on Automatic Control 37(6), 871–875.
Hsieh, C. (2000), ‘Robust two-stage Kalman filters for systems with unknown inputs’,
IEEE Transactions on Automatic Control 45(12), 2374–2378.
Ibnkahla, M. (2002), ‘Natural gradient learning neural networks for adaptive inversion
of Hammerstein systems’, IEEE Signal Processing Letters 9(10), 315–317.
Isermann, R. & Balle, P. (1997), ‘Trends in the application of model-based fault detec-
tion and diagnosis of technical processes’, Control Engineering Practice 5(5), 709–
719.
Janis, J. P. (1963), ‘Checkout methods for space vehicle subsystems’, IEEE Transac-
tions on Aerospace 1(2), 547–549.
Jirauch, D. H. (1967), ‘Software design techniques for automatic checkout’, IEEE Trans-
action on Aerospace and Electronic Systems AES-3(6), 934–940.
186
REFERENCES
Kaufmann, R. H. & Finison, H. J. (1952), D-C Power Systems for Aircrafts, John
Wiley & Sons, Inc., New York.
Keller, J. & Sauter, D. (2010), A variable geometric state filtering for stochastic linear
systems subject to intermittent unknown inputs, in ‘Proceedings of the Conference
on Control and Fault-Tolerant Systems (SysTol)’, pp. 558–563.
Korbicz, J., Koscielny, J. M., Kowalczuk, Z. & Cholewa, W., eds (2003), Fault Diag-
nosis: Models, Artificial Intelligence, Applications, Springer.
Lajic, Z., Blanke, M. & Nielsen, U. D. (2009), Fault detection for shipboard monitor-
ing Volterra kernel and Hammerstein model approaches, in ‘Proceedings of IFAC
Symposium on Fault Detection, Supervision and Safety of Technical Processes’,
pp. 24–29.
Lee, F. (1962), ‘An automatic self-checking and fault-locating method’, IRE Transac-
tions on Electronic Computers EC-11(5), 649–654.
Li, W. & Shah, S. (2002), ‘Structured residual vector-based approach to sensor fault
detection and isolation’, Journal of Process Control 12, 429–443.
Lia, C.-C. & Jengb, J.-C. (2010), ‘Multiple sensor fault diagnosis for dynamic processes’,
ISA Transactions 48(4), 415–432.
Ljung, L. (1999), System Identification - Theory for the User, PTR Prentice Hall In-
formation and System Sciences Series, 2nd edn, Prentice Hall, New Jersey.
Luenberger, D. G. (1964), ‘Observing the state of linear systems’, IEEE Trans. Mil.
Electr. MIL-8, 70–80.
MacFarlane, A. & Karcanias, N. (1976), ‘Poles and zeros of linear multivariable systems:
A survey of the algebraic, geometric and complex variable theory’, International
Journal of Control 24, 33–74.
187
REFERENCES
Ng, H., Chen, R. & Speyer, J. (2006), ‘A vehicle health monitoring system evaluated
experimentally on a passenger vehicle’, IEEE Transactions on Control Systems
Technology 14(5), 854–870.
Park, J. & Rizzoni, G. (1994), ‘An eigenstructure assignment algorithm for the design
of fault detection filters’, IEEE Transactions on Automatic Control 39(7), 1521–
1524.
Patel, R. V. & Munro, N. (1982), Multivariable System Theory and Design, Pergamon
Press, Inc.
Patton, R. & Chen, J. (1997), ‘Observer-based fault detection and isolation: Robustness
and applications’, Control Engineering Practice 5(5), 671–682.
188
REFERENCES
Patton, R. J. & Chen, J. (1991a), A parity space approach to robust fault detection
using eigenstructure assignment, in ‘Proceedings of European Control Conference
ECC91’, Grenoble, France.
Patton, R. J. & Chen, J. (1991b), Robust fault diagnosis using eigenstructure assign-
ment: A tutorial consideration and some new results, in ‘Proceedings of the 30th
Conference on Decision and Control’, pp. 2242–2247.
Patton, R. J. & Chen, J. (1991c), A robust parity space approach to fault diagno-
sis based on optimal eigenstructure assignment, in ‘International Conference on
Control 1991. Control ’91’, Edinburgh , UK, pp. 1056–1061.
Patton, R. J. & Chen, J. (1992), ‘Robust fault detection of jet engine sensor systems
using eigenstructure assignment’, Journal of Guidance, Control, and Dynamics
15(6), 1491–1497.
Patton, R. & Liu, G. (1994), Robust control design via eigenstructure assignment, ge-
netic algorithms and gradient-based optimisation, in ‘IEE Proceedings on Control
Theory and Applications’, Vol. 141, pp. 202–208.
Pierria, F., Paviglianiti, G., Caccavale, F. & Mattei, M. (2008), ‘Observer-based sensor
fault detection and isolation for chemical batch reactors’, Engineering Applications
of Artificial Intelligence 21, 1204–1216.
Raff, T., Menold, P., Ebenbauer, C. & Allgöwer, F. (2005), A finite time functional ob-
server for linear systems, in ‘Procedings of the 44th IEEE Conference on Decision
and Control and the European Control Conference 2005’, Seville, Spain.
189
REFERENCES
Shen, L. & Hsu, P. (1998), ‘Robust design of fault isolation observers’, Automatica
34(11), 1421–1429.
Siahi, M., Sadrnia, M. A. & Darabi, A. (2009), ‘A new method for observer design using
eigenstructure assignment and its application on fault detection and isolation’,
World Applied Sciences Journal 6(1), 100–104.
Simani, S., Fantuzzi, C. & Patton, R. J. (2002), Model-based fault diagnosis in dynamic
systems using identification techniques, Springer-Verlag, London, UK.
Sumislawska, M. (2009), Prediction of Limit Cycles in Hot Strip Mill Gauge Con-
trol, M.Sc. thesis, Control Theory and Applications Centre, Coventry University,
Coventry, UK.
Sumislawska, M., Burnham, K. J., Hearns, G., Larkowski, T. M. & Reeve, P. J. (2010),
Parity equation-based friction compensation applied to a rolling mill, in ‘Proced-
ings of the UKACC International Conference on Control 2010’, Coventry, UK,
pp. 1043–1048.
190
REFERENCES
Szabo, Z., Gaspar, P. & Bokor, J. (2005), Reference tracking for Wiener systems using
dynamic inversion, in ‘Proceedings of the 2005 IEEE International Symposium on,
Mediterrean Conference on Control and Automation Intelligent Control’, Limassol,
Cyprus.
Tan, C. P., Edwards, C. & Kuang, Y. C. (2006a), Robust sensor fault reconstruction
using a reduced order linear observer, in ‘9th International Conference on Control,
Automation, Robotics and Vision (ICARCV ’06)’, Singapore.
Tan, C. P., Edwards, C. & Kuang, Y. C. (2006b), Robust sensor fault reconstruc-
tion using right eigenstructure assignment, in ‘Third IEEE International Work-
shop on Electronic Design, Test and Applications (DELTA 2006)’, Kuala Lumpur,
Malaysia.
Tan, C. P. & Habib, M. K. (2004), Robust sensor fault reconstruction for an inverted
pendulum using right eigenstructure assignment, in ‘Proceedings of the 2004 IEEE
International Conference on Control Applications’, Taipei, Taiwan.
Tervo, K. & Manninen, A. (2010), Analysis of model orders in human dynamics iden-
tification using linear polynomial and Hammerstein-Wiener structures, in ‘Pro-
ceedings of 2010 International Conference on Networking, Sensing and Control
(ICNSC)’, Chicago, IL, USA.
191
REFERENCES
Wang, J., Jiang, B. & Shi, P. (2008), ‘Adaptive observer-based fault diagnosis for
satellite attitude control system’, International Journal of Innovative Computing,
Information and Control 4(8), 1921–1929.
White, J. E. & Speyer, J. L. (1986), Detection filter design: Spectral theory and algo-
rithms, in ‘Proceedings of the American Control Conference (ACC’86)’, Seattle,
WA, USA.
Wohnam, W. M. & Morse, A. (1970), ‘Decoupling and pole assignment in linear mul-
tivariable systems: A geometric approach’, SIAM Journal of Control 8(1), 1–18.
Xie, L., Zhang, J. & Wang, S. (2006), ‘Investigation of dynamic multivariate chemical
process monitoring’, Chinese Journal of Chemical Engineering 14(5), 559–568.
Ye, L., Shi, X. & Liang, J. (2011), ‘A multi-level approach for complex fault iso-
lation based on structured residuals’, Chinese Journal of Chemical Engineering
19(3), 462–472.
Yildiz, S. K., Forbes, J. F., Huang, B., Zhang, J., Wang, F., Vaculik, V. & Dudzic, M.
(2009), ‘Dynamic modelling and simulation of a hot strip finishing mill’, Applied
Mathematical Modelling 33, 3208 – 3225.
Yiua, J. C. & Wang, S. (2007), ‘Multiple ARMAX modeling scheme for forecast-
ing air conditioning system performance’, Energy Conversion and Management
48(8), 2276–2285.
Young, P. C., Behzadi, M. A., Wang, C. L. & Chotai, A. (1987), ‘Direct digital and
adaptive control by input-output state variable feedback’, International Journal
of Control 46, 1861–1881.
Zhong, M., Ding, S. X., Lam, J. & Wang, H. (2003), ‘An LMI approach to design robust
fault detection filter for uncertain LTI systems’, Automatica 39(3), 543–550.
192
Appendices
193
Appendix A
⎡ x 0 ⎤
⎢ ⎥
⎢ ⎥
1 0 0 ⋯ 0
⎢ x 0 ⎥
11
⎢ ⎥
⎢ ⎥
z1 1 0 ⋯ 0
⎢ x 0 ⎥
21
⎢ ⎥
X =⎢ ⎥
31 0 z2 1 ⋯ 0
⎢ ⋮ ⎥
(A.1)
⎢ ⎥
⎢ ⎥
⋮ ⋮ ⋮ ⋮ ⋱ ⋮
⎢ xq1 1,1 0 0 1 ⎥
⎢ 0 ⋯ zq1 −1 ⎥
⎢ ⎥
⎢ xq1 +1,1 0 0 z q1 ⎥
⎣ 0 ⋯ 0 ⎦
Eigenvalues corresponding to the linear combinations of columns of ē, λj , j = 1, ⋯, q1 ,
must fulfil the equation:
det(λj I − X) = 0 (A.2)
i.e.
⎡ x −λ ⎤
⎢ 11 ⎥
⎢ ⎥
1 0 0 ⋯ 0 0
⎢ x ⎥
j
⎢ ⎥
⎢ ⎥
z1 − λj 1 0 ⋯ 0 0
⎢ x ⎥
21
⎢ ⎥
det ⎢ ⎥
31 0 z2 − λj 1 ⋯ 0 0
⎢ ⎥
⎢ ⎥
⎢ ⎥
⋮ ⋮ ⋮ ⋮ ⋱ ⋮ ⋮
⎢ xq1 1,1 ⎥
⎢ 0 0 0 ⋯ zq1 −1 − λj 1 ⎥
⎢ ⎥
⎢ xq1 +1,1 z q1 − λj ⎥
⎣ 0 0 0 ⋯ 0 ⎦
⎡ ⎤
⎢ x11 − λj xq1 +1,1 ⎥
⎢ ⎥
x21 x31 ⋯ xq1 1,1 (A.3)
⎢ ⎥
⎢ ⎥
⎢ ⎥
1 z1 − λj 0 ⋯ 0 0
⎢ ⎥
⎢ ⎥
⎢ ⎥
0 1 z2 − λj ⋯ 0 0
⎢
= det ⎢ ⎥=0
⎥
⎢ ⎥
0 0 1 ⋯ 0 0
⎢ ⎥
⎢ ⎥
⎢ ⎥
⋮ ⋮ ⋮ ⋱ ⋮ ⋮
⎢ ⎥
⎢ 0 0 0 ⋯ zq1 −1 − λj 0 ⎥
⎢ ⎥
⎢ z q1 − λj ⎥
⎣ 0 0 0 ⋯ 1 ⎦
194
A. Calculation of parameters xij
⎢ 2 ⎥
⎢ ⎥
⋯ 0 0
⎢ ⎥
j
⎢ ⎥
− det ⎢ ⎥
0 1 ⋯ 0 0
⎢ ⎥
⎢ ⎥
⎢ ⎥
⋮ ⋮ ⋱ ⋮ ⋮
⎢ ⎥
⎢ 0 0 ⋯ zq1 −1 − λj 0 ⎥
⎢ ⎥
⎢ z q1 − λj ⎥
⎣ 0 0 ⋯ 1 ⎦
The determinant of a lower triangular matrix is equal to the product of its diagonal
elements, hence the first element of (A.4) is calculated via:
⎡ z −λ ⎤
⎢ 1 ⎥
⎢ ⎥
0 ⋯ 0 0
⎢ ⎥
j
⎢ ⎥
⎢ ⎥
1 z2 − λj ⋯ 0 0
⎢ ⎥
⎢ ⎥
(x11 − λj )det ⎢ ⎥
0 1 ⋯ 0 0
⎢ ⎥
⎢ ⎥
⎢ ⎥
⋮ ⋮ ⋱ ⋮ ⋮ (A.5)
⎢ ⎥
⎢ 0 0 ⋯ zq1 −1 − λj 0 ⎥
⎢ ⎥
⎢ z q1 − λj ⎥
⎣ 0 0 ⋯ 1 ⎦
= (x11 − λj )(z1 − λj )(z2 − λj )⋯(zq1 − λj )
⎡ xq1 +1,1 ⎤
⎢ ⎥
⎢ ⎥
x21 x31 ⋯ xq1 1,1
⎢ ⎥
⎢ ⎥
⎢ ⎥
1 z2 − λj ⋯ 0 0
⎢ ⎥
⎢ ⎥
− det ⎢ ⎥=
0 1 ⋯ 0 0
⎢ ⎥
(A.6)
⎢ ⎥
⎢ ⎥
⋮ ⋮ ⋱ ⋮ ⋮
⎢ ⎥
⎢ 0 0 ⋯ zq1 −1 − λj 0 ⎥
⎢ ⎥
⎢ z q1 − λj ⎥
⎣ 0 0 ⋯ 1 ⎦
195
A. Calculation of parameters xij
⎡ z −λ ⎤
⎢ 2 ⎥
⎢ ⎥
0 ⋯ 0 0
⎢ ⎥
j
⎢ ⎥
⎢ ⎥
1 z3 − λj ⋯ 0 0
⎢ ⎥
⎢ ⎥
= −x12 det ⎢ ⎥
0 1 ⋯ 0 0
⎢ ⎥
⎢ ⎥
⎢ ⎥
⋮ ⋮ ⋱ ⋮ ⋮
⎢ ⎥
⎢ 0 0 ⋯ zq1 −1 − λj 0 ⎥
⎢ ⎥
⎢ z q1 − λj ⎥
⎣ 0 0 ⋯ 1 ⎦
⎡ x xq1 +1,1 ⎤
⎢ 31 ⎥
⎢ ⎥
x41 ⋯ xq1 ,1
⎢ 1 ⎥
⎢ ⎥
⎢ ⎥
z3 − λj ⋯ 0 0
⎢ 0 ⎥
⎢ ⎥
+ det ⎢ ⎥
1 ⋯ 0 0
⎢ ⋮ ⎥
⎢ ⎥
⎢ ⎥
⋮ ⋱ ⋮ ⋮
⎢ 0 ⎥
⎢ 0 ⋯ zq1 −1 − λj 0 ⎥
⎢ ⎥
⎢ 0 z q1 − λ j ⎥
⎣ 0 ⋯ 1 ⎦
⎡ ⎤ ⎡ λ ∏q1 (z − λ ) ⎤
(A.8)
⎢ ⎥ ⎢ 1 k=1 k ⎥
⎢ ⎥ ⎢ ⎥
x11
⎢ ⎥ ⎢ λ ∏q1 (z − λ ) ⎥
1
⎢ ⎥ ⎢ 2 k=1 k ⎥
⎢ ⎥=⎢ ⎥
x21 2
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ ⎥
⋮ ⋮
⎢ ⎥ ⎢ λq1 ∏ (zk − λq1 ) ⎥
⎣ ⎦ ⎣ ⎦
q 1
xq1 +1,1 k=1
⎡ x ⎤
⎢ ⎥
⎢ ⎥
⎢ x ⎥
11
⎢ ⎥
⎢ ⎥ = (Ã) ˜
˜ −1 B̃
12
⎢ ⎥
(A.9)
⎢ ⎥
⎢ ⎥
⋮
⎢ xq1 +1,1 ⎥
⎣ ⎦
q1
Øj,k = (−1)k−1 ∏(zl − λj ) (A.10)
l=k
196
A. Calculation of parameters xij
˜ ∈ Rq1 +1 , denoted as B̃
whilst the j th element of the vector B̃ ˜ is:
j
q1
j ∏ (zl − λj )
˜ =λ
B̃ (A.11)
j
l=1
197
Appendix B
(i)
For the sake of brevity, the superscript is omitted. Matrices A∗e and Aw are equal if
their appropriate columns are equal. From equation (2.103) it holds that eigenvectors
of (A − KC) are given by:
(i) (i)
where β̄j is a column vector of qi parameters. The parameters β̄j conform to (2.107),
which can be reformulated as:
it follows that:
βk,1 Av1 + βk,2 Av2 + ⋯ + βk,qi Avqi − λk βk,1 v1 − λk βk,2 v2 − ⋯ − λk βk,qi vqi −
βqi ,1 Av1 − βqi ,2 Av2 − ⋯ − βqi ,qi Avqi + λqi +1 βqi ,1 v1 + λqi +1 βqi ,2 v2 + ⋯ + λqi +1 βqi ,qi vqi(B.4)
= (λk − λqi +1 )ei
198
B. Proof of Lemma 5.4
Then:
− (λk − βk,1 )ei + (βk,1 z1 + βk,2 − λk βk,1 )v1 + (βk,2 z2 v2 + βk,3 − λk βk,2 )v2
+ ⋯ + (βk,qi −1 zqi −1 + βk,qi − λk βk,qi −1 )vqi −1 + (βk,qi zqi − λk βk,qi )vqi =
(B.7)
− (λqi +1 − βqi ,1 )ei + (βqi ,1 z1 + βqi ,2 − λqi +1 βqi ,1 )v1 + (βqi ,2 z2 + βqi ,3 − λqi +1 βqi ,2 )v2
+ ⋯ + (βqi ,qi −1 zqi −1 + βqi ,qi − λqi +1 βqi ,qi −1 )vqi −1 + (βqi ,qi zqi − λqi +1 βk,qi )vqi
Due to the fact that ei , v1 , v2 , ⋯, vqi are linearly independent, it holds that:
λk − βk,1 = λj − βj,1
βk,1 z1 + βk,2 − λk βk,1 = βj,1 z1 + βj,2 − λj βj,1
βk,2 z2 v2 + βk,3 − λk βk,2 = βj,2 z2 + βj,3 − λj βj,2
(B.9)
⋮
βk,qi −1 zqi −1 + βk,qi − λk βk,qi −1 = βj,qi −1 zqi −1 + βj,qi − λj βj,qi −1
βk,qi zqi − λk βk,qi = βj,qi zqi − λj βk,qi
x11 = λj − βj,1
x21 = −βj,1 z1 − βj,2 + λj βj,1
x31 = −βj,2 z2 − βj,3 + λj βj,2
(B.10)
⋮
xqi ,1 = −βj,qi −1 zqi −1 − βj,qi + λj βj,qi −1
xqi +1,1 = −βj,qi zqi + λj βj,qi
199
B. Proof of Lemma 5.4
200
Appendix C
⋯ Cf1 α0
(1) (2)
Cf2 α0 ⋯ Cfr α0
(r)
]
M (t) = [ µ1 (t − τ ) µ2 (t − τ ) ⋯ µk (t − τ ) µ1 (t − τ + 1) µ2 (t − τ + 1) ⋯
(C.3)
µk (t − τ + 1) ⋯ µ1 (t) µ2 (t) ⋯ µk (t) ]
T
It follows from (C.1), (C.2), and (C.3) that (C.4) is equivalent to:
Thus:
r̊(t) = ΩM (t) (C.6)
201