Resilient Control in Cyber-Physical Systems: Countering Uncertainty, Constraints, and Adversarial Behavior
Resilient Control in Cyber-Physical Systems: Countering Uncertainty, Constraints, and Adversarial Behavior
R
in Systems and Control
Resilient Control in Cyber-Physical
Systems
Countering Uncertainty, Constraints, and
Adversarial Behavior
Suggested Citation: Sean Weerakkody, Omur Ozel, Yilin Mo and Bruno Sinopoli (2020),
“Resilient Control in Cyber-Physical Systems”, Foundations and Trends
R
in Systems and
Control: Vol. 7, No. 1-2, pp 1–252. DOI: 10.1561/2600000018.
Sean Weerakkody
Carnegie Mellon University
[email protected]
Omur Ozel
George Washington University
[email protected]
Yilin Mo
Tsinghua University
[email protected]
Bruno Sinopoli
Carnegie Mellon University
[email protected]
This article may be used only for the purpose of research, teaching,
and/or private study. Commercial use or systematic downloading
(by robots or other automatic processes) is prohibited without ex-
plicit Publisher approval.
Boston — Delft
Contents
1 Introduction 3
1.1 Applications . . . . . . . . . . . . . . . . . . . . . . . . . 10
2 Modeling 21
2.1 System Identification . . . . . . . . . . . . . . . . . . . . 21
2.2 Challenges in Modeling Cyber-Physical Systems . . . . . . 24
2.3 State Space Models . . . . . . . . . . . . . . . . . . . . . 25
2.4 Hybrid Systems . . . . . . . . . . . . . . . . . . . . . . . 29
6 Conclusions 229
References 232
Resilient Control in Cyber-Physical
Systems
Sean Weerakkody 1 , Omur Ozel2 , Yilin Mo3 and Bruno Sinopoli4
1 Carnegie Mellon University; [email protected]
2 George Washington University; [email protected]
3 Tsinghua University; [email protected]
4 Carnegie Mellon University; [email protected]
ABSTRACT
Cyber-Physical Systems (CPS), the amalgamation of sophis-
ticated sensing, communication, and computing technolo-
gies, applied to physical spaces, have become intrinsically
linked to society’s critical infrastructures. Indeed, CPS find
applications in energy delivery systems, intelligent trans-
portation, smart buildings and health care. Within these
systems, technological advances have enabled mankind to
improve their ability to both accurately monitor large scale
systems and precisely manipulate their behavior in order
to achieve complex local and global objectives. Nonetheless,
the opportunities created by CPS are met with significant
burdens and challenges, threatening the resilience of these
systems to both benign failures and malicious attacks.
In this monograph, we provide a comprehensive survey of
intelligent tools for analysis and design that take fundamen-
tal steps towards achieving resilient operation in CPS. Here,
we investigate the challenges of achieving reliable control
and estimation over networks, particularly in the face of
uncertainty and resource constraints. Additionally, we ex-
amine the threat of bad actors, formulating realistic models
Sean Weerakkody, Omur Ozel, Yilin Mo and Bruno Sinopoli (2020), “Resilient
Control in Cyber-Physical Systems”, Foundations and Trends
R
in Systems and
Control: Vol. 7, No. 1-2, pp 1–252. DOI: 10.1561/2600000018.
2
3
4 Introduction
frequently and with less delay. Additionally, sensing devices are in many
cases cheap and economically viable. The availability of cheap and
accurate sensing allows the designer to better understand physical pro-
cesses by obtaining larger numbers of spatial and temporal samples. An
example in this regard is the increased presence of phasor measurement
units (PMUs) in the power grid (Abur and Exposito, 2004). We note
that modern PMU technology has significantly changed the operation of
the electricity grid. In particular, the high sampling rates and accuracy
of voltage phasor measurements have changed state estimation from a
static problem to a dynamic problem.
In addition to monitoring physical processes, it is typically desirable
to physically manipulate a system to achieve some objective. In a
waste management system, a relevant task would be to treat and
purify the wastewater. Alternatively, in smart buildings we wish to
regulate the environment (i.e. using HVAC systems) in an energy efficient
manner. Cyber-physical systems allow us in many cases to automate this
process using computing technologies. The intelligent control of physical
systems is generally a time sensitive task. Thus, a key to incorporating
CPS is improvement in the processing speed of our computers. Today,
programmable logic controllers (PLCs) and microcontrollers are able
to quickly process sensory information and automatically implement
an intelligent algorithm for control. The speed at which this can be
done has allowed humans to explore new frontiers. As an example, the
ability to safely incorporate safe driving cars to transportation systems
is in part a result of the vast computational abilities of the embedded
systems in today’s vehicles.
Finally, a sophisticated communication infrastructure allows oper-
ators to control cyber-physical systems remotely while also enabling
them to reliably control large scale systems. Many systems have transi-
tioned from wired to wireless communication technologies, which allows
for ease of maintenance and installation, lower costs, as well as au-
tomation in geographically disparate systems. As an example, wireless
communication technologies play a major role in supervisory control
and data acquisition (SCADA) systems, see, e.g., Cardenas et al. (2009).
A SCADA system is a hierarchical system, which enables the supervi-
sory management of a control system. The lowest layer consists of field
5
devices such as sensors and actuators, which directly interact with the
physical environment. Remote terminal units (RTUs) and PLCs are
often used to implement autonomous local control. These units typically
interface with both field devices such as pumps, valves, and switches
as well as a centralized supervisory control layer which monitors the
system. SCADA systems are regularly seen in the smart grid as well
as water distribution and waste management systems. Communication
technologies allows RTUs to interface with human operators at SCADA
systems in real time. This allows operators to make high level control
decisions remotely in a timely fashion. This capability is especially
important when monitoring at the supervisory layer raises an alarm,
which requires immediate operator attention. Communication technolo-
gies not only allow devices and components to interface with central
operators, but also each other. Local communication among field devices
can enable distributed control. Here, autonomous controllers/agents
share information and act to achieve a larger task. Distributed control
can be used to achieve formations in aerial vehicles and platoons in
ground transport.
Unexpected challenges arise when accounting for the tight inter-
action of computing elements with the physical plant in CPS. Unlike
normal IT infrastructures, the operations of CPS are often safety critical
(Lee, 2008; Rajkumar et al., 2010; Giani et al., 2008). For example,
malfunctioning teleoperated robots in surgery may harm or possibly
kill patients. Likewise, blackouts on the electricity grid may disrupt
vital services. Thus, operators are obligated to ensure these systems
perform resiliently. Complicating the matter is the time sensitive nature
of CPS. To ensure that the dynamics of a physical process are well
regulated, CPS must be monitored and acted upon frequently. In this
monograph, we aspire to identify significant challenges, which hinder the
successful operation of cyber-physical systems. To this end, we consider
several proposed tools and methodologies aimed towards addressing
these fundamental problems.
First, in section 2 we consider the problem of modeling CPS.In
control systems, an accurate numerical representation of a plant is often
a crucial component to developing intelligent algorithms for automation,
with provable mathematical properties. These models can be developed
6 Introduction
from first principles. For instance, Newton’s laws can be used to de-
scribe the dynamics of vehicles while Maxwell’s equations can be used
to derive dynamical equations associated with mathematical genera-
tors. Alternatively, we can utilize big data and in particular system
identification/machine learning techniques to obtain effective models of
our systems. We briefly discuss system identification in subsection 2.1.
Cyber-physical systems pose a particular challenge due to the inherent
diversity of the systems being considered (Derler et al., 2012). They
not only contain a physical plant, which needs to be modeled like a
traditional control system, but also have heterogeneous hardware and
software systems which enable computing and data transfer. The chal-
lenges of modeling CPS are detailed in subsection 2.2. We then look
at specific classes of models. In addition to examining traditional state
space, LTI, and stochastic systems in subsection 2.3, we will address
modeling CPS through a brief discussion of hybrid systems in subsection
2.4.
Even with a precise and accurate model of CPS, operators must
account for sources of uncertainty and how they impact subsequent
analysis and design. As an example, in section 3 we will study net-
worked control systems, focusing on achieving feedback control over
stochastic, resource constrained, communication networks. While tran-
sitioning from wired to wireless communication technologies can reduce
costs and improve efficiency, reliability may be sacrificed. Packets con-
taining sensory or control data may be delayed or dropped over the
communication network. In a cyber-physical system, the availability of
real-time data is often essential for correct and reliable operation. Sen-
sor packet drops leads to inadequate monitoring and feedback control.
Input packet drops prevent corrective commands from being delivered
to the plant. As communication failures can significantly affect the
functionality of CPS, operators must carefully model and account for
their presence through robust analysis and design. We will discuss the
design of robust feedback controllers in CPS with sensor and input
drops respectively in subsection 3.1 and 3.2. In these cases, we will
additionally arrive at fundamental conditions on network reliability,
which allow the aforementioned algorithms to successfully stabilize CPS.
7
1.1 Applications
methods for device placement and drug delivery. We expand upon these
topics below.
Traditionally, intelligent sensing and actuation has found applica-
tions in scenarios that involve wearable devices and implantable devices
such as pacemakers and defibrillators. In particular, mobile monitoring
of vital signals and physical activities obviate the need of doctors to be
physically present to diagnose the health of individual patients. Schirner
et al. (2013) suggests that embedded sensors which measure human cog-
nitive activity are enablers of human in the loop CPS. Specifically, the
development of human and machine interfaces can improve interactions
with assistive robots, which perform actions for the benefit of a person
with a disability and allow for enhancements in intelligent prostheses,
restoring function to amputees.
Similarly, there is now also growing interest towards in-body and on-
body sensor networks that can measure activity and athletic performance
based on body state indicators such as heart and breathing rate, blood-
sugar level and skin temperature. In this respect, developing energy
harvesting technologies (such as RF energy harvesting or thermoelectric
energy harvesting using body heat) enable battery free operation and
ease of implementation in various types of body sensor applications. RF
energy harvesting is a well known technique for increasing the lifetime
of implantable devices (Ho et al., 2014). In addition, thermoelectric
generators, kinetic harvesters and solar technology are also being used
in body sensor networks to harvest energy in wireless bio-sensor devices
(Mitcheson, 2010).
It is argued by Lee and Sokolsky (2010) that monitoring and control
in medicine could greatly benefit from newly developed cyber-physical
technologies. Real time embedded closed-loop control could facilitate
immediate diagnostic evaluation of vital signals and make constant care
possible. For example, Lee and Sokolsky (2010) discuss how intelligent
coordination between x-ray machines and ventilators during an operation
can save patient lives. Specifically, to currently obtain good images
(without patient motion) during an operation, a ventilator must be
paused, thus preventing lung movement. However, patients have died in
cases where the ventilator would not restart. An intelligent alternative
involving precise control would be to enable automatic coordination
14 Introduction
between the x-ray and the ventilator. The x-ray would take images
when it detects the end of a breathing cycle. As a result, the ventilator
does not need to be turned off.
Additionally, the use of computing, sensing, and communication
technologies can reduce humans erros. Cyber-physical technologies
promise to minimize human mistakes by automating various medical
tasks both in clinical scenarios and operation room practice. For instance,
Lee and Sokolsky (2010) consider patient-controlled analgesia and argues
that it can benefit from feedback control. In this process, infusion pumps
are commonly used to deliver opioids for pain management before and
after surgery. Current technological safeguards suchs as drug libraries
and programmable limits can be insufficient in safely addressing pain
management. The authors propose a closed-loop control system with
a supervisor to monitor patient data for the early signs of respiratory
failure. The automated supervisor can stop infusions and sound an alarm
in case of an adverse event. We also remark the role of nanorobots in
the development of new drug delivery methods, see, e.g., Douglas et al.
(2012). This technology promises to deliver drugs to a targeted region in
the body and hence minimize the risks and possible side effects caused
by its use.
Unfortunately, without proper care, cyber physical technologies can
negatively impact the security and reliability of medical devices. First,
medical devices may be subject to a significant failure risk with poten-
tially catastrophic impacts on patients. Alemzadeh et al. (2012) argue
that faulty monitoring devices could cause serious injury and death.
An over reliance on autonomous monitoring and treatment in a faulty
scenario could result in harm to a patient, which could have otherwise
been prevented with a doctor in the loop. In addition, the dependence
of cyber-physical systems on information technology make them more
vulnerable to cyber attacks. Alemzadeh et al. (2013) report that tele-
operated robots are vulnerable to malicious adversaries. In particular,
this work considers attackers who install malware to strategically affect
robots during surgery. To detect and mitigate such attacks, Alemzadeh
et al. (2013) devises a model-based analysis framework using the dy-
namics of the surgical robot. This framework is utilized to determine if
a command is trustworthy before execution.
1.1. Applications 15
et al., 2015; Hoh et al., 2006). Hoh et al. (2006) for instance notes that
location monitoring services in next generation traffic systems can allow
drivers to be be tracked. Privacy can be corrupted by eavesdroppers
on the network, attackers who install spyware, or malicious insiders
with access to a traffic monitoring server. Significant information can
be gleaned from tracking a user. As noted by Hoh et al. (2006) ,
one can learn about the health of a driver if they frequently visit a
doctor/specialist or political leanings from visits to activist organizations.
Perhaps more worrisome is the home identification of particular drivers.
As such, the privacy of transportation data requires significant attention.
21
22 Modeling
tion has been enhanced by tools from machine learning which offers a
plethora of tools and algorithms to efficiently process data and learn
models. For more information regarding the interactions between system
identification and machine learning, see Ljung et al. (2011)
The steps for system identfication described in this subsection often
allow designers to obtain models of physical processes. However, CPS
are comprised not only of physical systems, but also cyber technologies.
An in depth discussion of identifying models for computational and
networking systems in CPS is out of the scope of this monograph (see,
e.g., Lee (2006) for a more detailed discussion). In the next subsection,
we examine challenges that arise when attempting to model software
and communication systems along with physical processes in a CPS.
Here, the jumps in the discrete variable q are induced by the events
described by g. For each q, the system acts as a LTI system.
A well known example of a hybrid system is the hysteresis loop
represented by the following equations:
(
1 + u, x≤∆
ẋ = . (2.15)
−1 + u, x > −∆
33
34 Networked Control Systems
ỹk = Γk yk ,
Γk = diag(γk,1 , . . . , γk,m ),
γk,1
Sensor 1
γk,2
Sensor 2
γk,3
System Sensor 3 Estimator
..
.
γk,m
Sensor m
free, especially for low-cost devices. For example, for the TCP protocol,
two-way communication between the sender and the receiver is required.
Hence, in this section, we shall mainly investigate the impact of packet
drops on CPS in the absence of more sophisticated communication
protocols such as TCP. The CPS designer can then deside whether the
impact of packet drops is tolerable.
For the single sensor case, the random process {γk } is the packet
drop process and is typically assumed to be an IID Bernouli process
with
Pr(γk = 1) = γ̄.
Furthermore, the packet drop process {γk } is assumed to be independent
from the noise {vk }, {wk } and initial condition x0 . It is worth noticing
that other packet drop model exists, e.g. Markovian model, and some
results discussed in this subsection can be applied to a more general
packet drop process. It is further assumed, that the estimator can
distinguish a valid packet from a dropped packet. This can usually
be achieved via an error-detecting code. In other words, the estimator
knows Γk at time k. Therefore, the information available to the estimator
at time k can be written as
Ik = {Γ0 y0 , Γ0 , . . . , Γk yk , Γk } . (3.2)
36 Networked Control Systems
The goal of state estimation is to infer the current state xk given the
information set Ik . Here we use the MMSE (minimum mean squared
error) estimator. Define
• Initialization:
• Prediction:
• Correction:
h(X) , AXAT + Q,
−1
g0 (X) , X −1 + C T R−1 C = X − XC T (CXC T + R)−1 CX.
P∞ = g(P∞ ).
Combining (3.8) with the fact that g(X) ≤ h(X) and h and g are
monotonically increasing, we can prove the following lemma:
Such a sequence γk:0 ∗ can be found using the following branch and
bound method:
Algorithm 1: Branch and Bound Method
Data: N ≥ P∞
Result: γk:0∗ = (γ ∗ , . . . , γ ∗ )
k 0
Initialization: t ← k;
while t ≥ 0 do
if Pk|k (γk∗ , . . . , γt+1
∗ , 0, 1, . . . , 1) ≤ N then
∗
γt ← 0;
else
γt∗ ← 1;
end
t ← t − 1;
end
∗ , the probability can be then
For the IID case, once we derive γk:0
recursively calculated as
γ̄ + (1 − γ̄)Pr(γ ∗ if γk∗ = 0
∗ k−1:0 . γk−1:0 )
Pr(γk:0 . γk:0 )= ,
γ̄Pr(γk−1:0 . γ ∗ ) if γk∗ = 1
k−1:0
Assuming that {γk } are IID, then γk+1 would be independent from Pk|k ,
since Pk|k is a function γk , . . . , γ1 . Taking the expectation on both sides
of (3.9), we get
Sk ≤ E[Pk|k ] ≤ Vk , ∀k.
Now using the fact that both h and g are monotonically increasing, we
get
With the upper and lower bounds provided by Theorem 3.6, we can
easily derive upper and lower bounds for the critical value γc .
5
We use the notation supk Ak = +∞ when the sequence Ak ≥ 0 is not bounded;
i.e., there is no matrix M ≥ 0 such that Ak ≤ M, ∀k.
44 Networked Control Systems
γl ≤ γc ≤ γu .
2. If rank(C) = 1, then
1
γu = 1 − Ql u 2
,
i=1 |ρi |
It is worth noticing that, for general systems, the upper and lower
bounds provided by Theorem 3.8 are not tight. This is mainly due to
the “eigen-cycle” of the A matrix, which is a set of eigenvalues with
the same absolute value. If an eigen-cycle exists, then it is possible that
even if (A, C) is observable, (Aτ , C) is not observable for some τ . To
see this, consider the following system:
" #
0 2 h i
A= , C = 1 0 , Q = I, R = 1. (3.11)
2 0
The A matrix has two eigenvalues, 2 and −2, which form an eigen-cycle.
It is worth noticing that (A2 , C) is not observable. To see why this loss
of observability causes a problem, we illustrate the information flow
of the system in Figure 3.2. Intuitively speaking, if all measurements
made at the even (or odd) times are lost, then the estimator cannot
“reconstruct” the system state. In other words, we need at least one
measurement made at the odd time and one measurement made at
3.1. Estimation in Lossy Networks 45
the even time to have a bounded estimation of the system state. This
restriction will deteriorate the performance of our estimator.
The following theorem provides a quantitative result on the impact of
eigencycles on the critical value for diagonalizable second-order systems:
2. rank(C) = 2.
Notice that if rank(C) = 1 and |ρ1 | = |ρ2 |, then the lower bound
and upper bound of the critical value given by Theorem 3.8 are:
y1 y2 y3 y4 y5
Theorem 3.9 clearly shows that neither the lower nor the upper bound
is tight in general.
For general linear systems with diagonalizable A, we use linear
transformations to put A into the diagonal standard form. The system
can then be decomposed into subsystems with eigenvalues on the same
eigen-cycle. To this end, we first define an equivalence relation on C,
such that z and z 0 are equivalent if and only if there exists q ∈ N+ , such
that z q = (z 0 )q . A necessary and sufficient condition for z and z 0 to be
equivalent is that z and z 0 have the same absolute value and the angle
between them are a rational fraction of π. The proof that the relation
defined by us is indeed an equivalence relation is left to the readers.
Next we partition the eigenvalues of A into g equivalent classes.
As a result, A can be diagonalized as diag(A1 , . . . , Ag ), where each
Ai ∈ Cni ×ni is a diagonal matrix containing all eigenvalues in one
equivalent class. Hence, we can write Ai as
Ai = αi diag(exp(jφi,1 ), . . . , exp(jφi,ni )).
We can split the corresponding C matrix into
h i
C = C1 , . . . , C g ,
and it is unbounded if
The packet loss rate is chosen to be 0.25 for both systems. The reader
can verify that both systems are detectable. However, system 2 is degen-
erate since (A2 , C) is not detectable, while system 1 is non-degenerate.
Figure 3.3 illustrates the tail distribution of Pk . Notice that for both
systems, the tail follows a power decay law. However, the tail converges
much quicker for non-degenerate system.
100
Pr(tr(Pk|k ) ≥ x)
10−3
10−6
10−9
10−12
100 101 102 103 104 105 106 107 108
x
Figure 3.3: The Tail Distribution of tr(Pk|k ). The systems are chosen so that
A = diag(2, −2) and Q = I. The dashed line corresponds to C = R = I and the
solid line corresponds to C = [1 1] and R = 1.
uak = νk uk ,
Pr(νk = 1) = ν̄.
Controller
Figure 3.4: Diagram of a Control System where the Control Input Goes Through
a Lossy Network.
Gk , {y0 , y1 , . . . , yk , u0 , . . . , uk−1 } .
• Initialization:
• Prediction:
• Correction:
For the TCP case, we know that f (xk |Ik ) is a Gaussian distribu-
tion with mean x̂k|k and a deterministic covariance Pk|k , where Pk|k is
deterministic. Hence, instead of defining Vk on all possible probability
distributions, we only need to consider the value of Vk on Gaussian
3.2. Control in Lossy Networks 53
and the optimal control law is uk = −(U + B T Sk+1 B)−1 B T Sk+1 Ax̂k|k .
6
We can omit Pk|k since it is deterministic and independent from the control
input u0 , . . . , uN .
54 Networked Control Systems
For the TCP case, one could also consider the infinite horizon LQG
problem, where
JN
J∞ = lim .
N →∞ N
The following theorem provides the optimal solution for the infinite
horizon LQG problem:
then
A = 1.5, B = C = Q = R = U = W = 1.
Figure 3.5 shows the optimal control gain L defined in (3.26) versus
the packet drop rate 1 − ν̄. Notice that (3.25) does not admit a solution
when the packet drop rate 1 − ν̄ ≥ 1/A2 = 4.9. One can see that as the
packet drop rate increases, the control gain L converges to A. Intuitively
speaking, this means that the controller is trying to drive the state xk
back to 0 in one step, instead of accomplishing it over multiple steps,
since future control packets may not be successfully delivered.
Figure 3.6 shows the tail distribution of |xk | for packet drop rates
of 1/4 and 4/9 respectively. For 1 − ν̄ = 1/4, the optimal control gain
from Theorem 3.14 is used. For 1 − ν̄ = 4/9, we use the limiting gain
L = A.
1.5
1.4
1.3
L
1.2
1.1
Figure 3.5: The optimal LQG gain L versus the packet drop rate. The system is
chosen so that A = 1.5 and B = C = Q = R = U = W = 1.
3.2. Control in Lossy Networks 57
100
10−1
10−2
Pr(|xk | ≥ x)
10−3
10−4
10−5
10−6
φ−
k,i αk,i
φk,i = P2k −1 , (3.29)
i=0φ−k,i αk,i
1
− T T −1 −
with αk,i = exp − (yk − Cξk,i ) (R + CPk|k−1 C ) (yk − Cξk,i ) ,
2
(3.30)
− −
ξk,i = ξk,i + Pk|k−1 C T (CPk|k−1 C T + R)−1 (yk − Cξk,i ), (3.31)
Pk|k = (Pk|k−1 + C T R−1 C)−1 , (3.32)
and
φ− −
k+1,2i = (1 − ν̄)φk,i , φk+1,2i+1 = ν̄φk,i , (3.33)
− −
ξk+1,2i = Aξk,i , ξk+1,2i+1 = Aξk,2i + Buk , (3.34)
T
Pk+1|k = APk|k A + Q, (3.35)
φ− −
0,0 = 1, ξ0,0 = x̄0 , P0|−1 = Σ0 .
Before proving the theorem, we would like to point out for each xk ,
the number of Gaussian distributions in the mixture is 2k . Conceptually,
this is due to the fact that there are 2k possible realizations of the
packet drop process ν0 , . . . , νk−1 . Therefore, a Kalman filter is built for
each possible realization. In fact, the update equation of ξk,i is exactly
the same as the Kalman filter for xk , assuming that νj = ij , where ij
is the jth digit in the binary representation of i. Similarly, Pk|k is the
corresponding estimation error. φk,i can be interpreted as how likely
this particular sequence of packet drop process ν0 , ν1 , . . . , νk−1 happens,
given our observation y0 , . . . , yk . It is not difficult to see that φk,i is
P k
positive and 2i=0−1 φk,i = 1.
The following lemma, which can be verified easily, is used to simplify
the proof:
3.2. Control in Lossy Networks 59
with
− T −1 −
ςk,i = (xk − ξk,i ) Pk|k−1 (xk − ξk,i ) + (yk − Cxk )T R−1 (yk − Cxk )
−1
= (xk − ξk,i )T Pk|k (xk − ξk,i )
− T −
+ (yk − Cξk,i ) (R + CPk|k−1 C T )−1 (yk − Cξk,i ).
60 Networked Control Systems
where
Z
f (yk+1 |Ik− )Vk+1 (f (xk+1 |yk+1 , Ik− ))dyk+1
Rm
2k+1
X−1 φ−
k+1,i αk+1,i
Z
= q Vk+1 (φk+1 , ξk+1 )dyk ,
Rm i=0 (2π)m |CPk+1|k C T + R|
3.2. Control in Lossy Networks 61
with
φ− −
k+1,2i = (1 − ν̄)φk,i , φk+1,2i+1 = ν̄φk,i ,
− −
ξk+1,2i = Aξk,i , ξk+1,2i+1 = Aξk,2i + Buk ,
φ−k+1,i αk+1,i
φk+1,i = P2k+1 −1 − ,
i=0 φ k+1,i α k+1,i
− −
ξk+1,i = ξk+1,i + Pk+1|k C T (CPk+1|k C T + R)−1 (yk+1 − Cξk+1,i ),
As a result,
Z
− −
f (yN |IN −1 )VN (f (xN |yN , IN −1 ))dyN = tr(PN |N W )
Rm
N −1
2X
αN,i
Z
+ φ−
N,i q T
ξN,i W ξN,i dyN .
i=0 Rm (2π)m |CPN |N −1 C T + R|
γi ≥ γj , if j is a child of i. (3.39)
σk ← σk {s};
S
end
To prove a performance bound for the greedy algorithm, we need a
definition of submodular functions:
Theorem 3.18. Suppose that the optimal value of the following opti-
mization problem is ν∗
Suppose that the solution from the greedy algorithm is Γk (σk ) and
the corresponding log det Pk|k = ν. If R is diagonal, then the following
inequality holds:
H~γk ≤ b,
γk,i = 0 or 1.
(3.42)
In both Problem (3.41) and (3.42), the objective function and the
constraints are convex except for the binary constraints on γk,i . There-
fore, one way to relax the problem into a convex optimization problem
is to change the binary constraints γk,i = 0 or 1 to 0 ≤ γk,i ≤ 1. Notice
that the relaxation will increase the feasible space of γk and thus the
optimal value of the relaxed problem will be a lower bound for optimal
value of the original problem.
On the other hand, the optimal solution ~γk of the relaxed problem
is not necessarily a binary vector. One can quantize the solution to
get a binary vector (Joshi and Boyd, 2009). Alternatively, one can
adopt a stochastic scheduling strategy, where each sensor will transmit
with certain probability and ~γk can be interpreted as the transmitting
probability. For more details, please refer to Mo et al. (2011b).
Next we consider the case where R is not diagonal, which is consid-
ered in the multi-step sensor scheduling problem Rigtorp (2010). To
this end, the following lemmas are needed:
Lemma 3.19. Define the function L(K, X) as
L(K, X) = (I − KC)h(X)(I − KC)T + KRK T .
3.3. Designing Sensor Network with Resource Constraints 69
g(X, Γ) ≤ L(KΓ, X)
= (KΓ − K∗ (X))(Ch(X)C T + R)(KΓ − K∗ (X))T + g(X).
minimize f (P ) (3.43a)
Γ, K, P
subject to P ≥ L(K, Pk−1|k−1 ), (3.43b)
H~γk ≤ b, (3.43c)
kKi k0 ≤ nγk,i , (3.43d)
γk,i = 0 or 1, (3.43e)
which implies that the optimal value of (3.38) is no greater than the
optimal value of (3.43), due to the non-decreasing property of f . As a
result, the optimal value of (3.38) equals the optimal value of (3.43).
T
1 X
J (P0|−1 , σ) , lim sup f (Pk|k ), (3.44)
T →∞ T k=0
3.3. Designing Sensor Network with Resource Constraints 71
where Pk|k satisfies the recursive equation. Pk|k = g(Pk−1|k−1 , Γk (σk )).
It is clear that Pk|k is a function of both the initial condition P0|−1 and
the schedule σ and can be denoted as Pk|k (P0|−1 , σ). However, we will
write it as Pk|k for short if there is no confusion. We will assume that
the function f satisfies the following properties:
1. f is non-decreasing on the set of positive semidefinite matrices.
3. For any > 0, there exists a δ, such that for any X > 0
f ((1 + δ)X) ≤ (1 + )f (X). (3.46)
Remark 3.3. Notice that both the trace and maximum eigenvalue of X
satisfy (3.45). However, the log determinant fails to satisfy (3.45) since
the log determinant of diag(ρ, 1/ρ) is always 1. Equation (3.46) implies
that f is actually continuous.
J (P0|−1 , σ) can be seen as the average estimation error. Moreover,
let us define the average communication rate of sensor i as
N
1 X
ratei (σ) , lim sup 1s ∈I ,
N →∞ N k=1 i k
We further define feasible schedules:
Definition 3.3. A schedule σ is called feasible if for any initial condition
P0|−1 , J (P0|−1 , σ) is bounded (by a function of P0|−1 ).
Before continuing on, we need the following lemma by Mo et al.
(2014a) to establish another contraction property of g:
Lemma 3.22. For all ρ ≥ 0, X > 0, the following inequalities hold,
g(X, Γ) ≤ g((1 + ρ))X, Γ) ≤ (1 + ρ)g(X, Γ). (3.47)
Furthermore, if AXAT ≤ αQ, then
ρ
g(X, Γ) ≤ g((1 + ρ)X, Γ) ≤ 1 + g(X, Γ). (3.48)
1+α
72 Networked Control Systems
Pk+1|k+1 (Y, σ) = g(Pk|k (Y, σ), Γk+1 ) ≤ g((1 + ρk )Pk|k (X, σ), Γk+1 )
≤ (1 + ρk )g(Pk|k (X, σ), Γk+1 ) = (1 + ρk )Pk+1|k+1 (X, σ).
AXAT ≤ αQ.
Since we assume (3.45) holds and that Q > 0, we can always find such
an α. Now from the definition of J (X, σ), the following inequality holds
infinitely often (i.e. for an infinite number of integers k):
for sufficiently large k. Hence, J (Y, σ) ≤ J (X, σ), which finishes the
proof.
and
ratei (σp ) ≤ ratei (σ) + εi , i = 1, . . . , m. (3.51)
X = g(X, Γ).
On the other hand, if (A, ΓC) is not detectable, then we define P∗ (Γ) =
∞.
If f is continuous, then it is easy to prove that for the period 1
schedule, J = f (P∗ (Γ)). Similar to the schedule problem (3.38), the
periodic scheduling problem can be formulated as
1. P∗ (I) ≤ P .
and
" #
Q−1 − S + C T R−1 C Q−1 A
−1 ≥ 0. (3.54)
T
A Q S + AT Q−1 A
minimize f (P ) (3.56a)
Γ
subject to P∗ (Γ) ≤ P, (3.56b)
H~γ ≤ b, (3.56c)
γi = 0 or 1, (3.56d)
Measurement update:
where
h i−1
Kk = Pk|k−1 C T CPk|k−1 C T + R + (1 − γk )Y −1 , (3.65)
with initial condition
x̂0|−1 = x̄0 , P0|−1 = Σ0 . (3.66)
Proof. We prove the theorem by induction. Since x0 is Gaussian, (3.66)
holds. We first consider the measurement update step. Assume that
xk conditioning on Ik−1 is Gaussian with mean x̂k|k−1 and covariance
Pk|k−1 . We consider two cases depending on whether the estimator
receives yk .
1. γk = 0:
If γk = 0, then the estimator does not receive yk . Consider the
joint conditional pdf of xk and yk ,
f (xk , yk |Ik ) = f (xk , yk |γk = 0, Ik−1 )
Pr(γk = 0|xk , yk , Ik−1 )f (xk , yk |Ik−1 )
=
Pr(γk = 0|Ik−1 ) (3.67)
Pr(γk = 0|yk )f (xk , yk |Ik−1 )
= .
Pr(γk = 0|Ik−1 )
The second equality follows the Bayes’ theorem and the last one
holds since γk is conditionally independent with (Ik−1 , xk ) given
yk . Let us define the covariance of [xTk , ykT ]T given Ik−1 as
" #
Pk|k−1 Pk|k−1 C T
Φk , (3.68)
CPk|k−1 CPk|k−1 C T + R
From the definition of ϕ, we have
!
1 1
Pr(γk = 0|yk ) = Pr exp(− ykT Y yk )≥ζk yk = exp(− ykT Y yk ).
2 2
(3.69)
From (3.67), (3.68), and (3.69), we have
1
f (xk , yk |Ik ) = αk exp(− θk ),
2
3.3. Designing Sensor Network with Resource Constraints 79
where
1
αk = p
Pr(γk = 0|Ik−1 ) det(Φk )(2π)m+n
and
" #T " #
x − x̂k|k−1 xk − x̂k|k−1
θk = k Φ−1
k + ykT Y yk . (3.70)
yk − ŷk|k−1 yk − ŷk|k−1
where
and " #
Θxx,k Θxy,k
Θk = ,
ΘTxy,k Θyy,k
with
2. γk = 1:
If γk = 1, then the estimator receives yk . Hence
minimize f (P (Y )) (3.75a)
Y ≥0
subject to rate ≤ rate (3.75b)
minimize f (P ) (3.76a)
Y ≥0
subject to P (Y ) ≤ P, (3.76b)
tr ΠY ≤ −2 log(1 − rate) (3.76c)
In Figure 3.7, we plot the apriori mean squared error in the state
estimate as a function of the average communication rate, where each
data point is obtained over a run of 10,000 trials. We consider 4 main
designs. We first consider a random design where for each sensor at
each time step, the probability of transmission is λavg . We also consider
a stochastic design where each sensor communicates at the same rate,
and an optimized design. In the optimized design, the communication
rate for each sensor is chosen so as to minimize an objective function
related to the total communication rate subject to an upper bound on
the covariance of the estimation error.
Finally, for comparison we include a deterministic trigger defined
(i)
where sensor i transmits if kyk k > δ (i) and does not transmit otherwise.
δ (i) is chosen so sensor i communicates at the same rate as sensor i in the
optimized stochastic trigger. A sub-optimal estimator is incorporated
here where a posteriori estimates and error covariances are obtained
using a Kalman filter for just the received measurements. Also shown
are upper and lower bounds for the un-optimized approach.
In Figure 3.8, we plot the percent improvement of the stochastic and
deterministic designs relative to the random design with regards to the
mean squared error plotted in Figure 3.7. An un-optimized stochastic
design provides as much as 15% improvement, a deterministic design
offers as much as 20% improvement, and the optimized stochastic design
offers as much as 30% improvement.
1.5
Stochastic Un-Optimized
1.4
1.4 Stochastic Optimized
Random Offline
1.3 Deteministic Trigger
Upper Bound Un-Optimized X
1.2
1.2 Lower Bound X
X=
Mean Squared Error
1.1
11
0.9
0.8
0.8
0.7
0.6
0.6
0.5
00 0.1
0.2
0.2 0.3
0.4
0.4 0.5
0.6
0.6 0.7
0.8
0.8
λavg , Communication Rate
0.9
11
Communication Rate
avg
Figure 3.7: Mean square error (Kelvin2 ) for random, deterministic, stochastic, and
stochastic optimized strategies vs λavg , the communication rate
35
Stochastic Un-Optimized
3030 Stochastic Optimized
Deterministic Trigger
25
% Improvement Compared to Random
2020
15
1010
5
00
-5
00 0.1
0.2
0.2 0.3
0.4
0.4 0.5
0.6
0.6 0.7
0.8
0.8
λavg , Communication Rate
0.9
11
Communication Rate
avg
Theorem 3.28. Consider the control system ẋ(t) = Ax(t) + Bu(t), with
controller u = Lx. Assume (A + BL) is Hurwitz, and event trigger
(3.86) is implemented. Then the inter-event times ti+1 − ti , i ∈ N are
bounded below by τ which satisfies,
φ(τ, 0) = σ, (3.87)
ke(t)k2 ≤ σkx(t)k2 .
Theorem 3.30. Consider the control system (3.96) with model-based es-
timator (3.97). Assume there exists a solution to the Lyapunov equation
(3.98). Moreover, assume k(Ã + B̃L)T P + P (Ã + B̃L)k2 ≤ δ < λm (Q)
and kB̃k2 ≤ β. Consider the following trigger with update instants ti
implicitly defined by the relation
ση
keq (ti )k2 ≥ kq(x(ti ))k2 (3.101)
γ+1
where η = (λm (Q) − δ)/b, 0 < σ < σ 0 < 1. Then,
We note that there are examinations and results which account for
delays, quantization, and model uncertainty. For more information, see
Garcia and Antsaklis (2013).
Output Based and Decentralized Control: Typically, an operator
does not have access to the entire state during measurements. Attempts
to use similar triggers in this scenario can result in undesired Zeno
behavior. Additionally, we may need to consider multiple sensing nodes.
In such cases, evaluating a trigger may need to be a decentralized
process. Here, each sensing node must make its own decision in regards
to whether a measurement is sent. In this scenario, we model our control
system as follows
Since v̂(0) = v(0), we know tj0 = 0. Here the error is e(t) = v̂(t) − v(t).
We construct eIj (t) and vIj (t) by taking elements of e(t) and v(t) that
belong to Ij = {i ∈ {1, · · · , m+p}|γji = 1}. The structure of this trigger
ensures that for all nodes j ∈ {1, · · · , N } and all time t ≥ 0,
keIj (t)k22 = σj kvIj (t)k22 + j . (3.108)
94 Networked Control Systems
for the ideal system ẋ = (A + BL)x. In the self triggered case, with
w = 0, we may instead desire the relaxed performance objective
where 0 < λ < λ0 is some desired slower rate of decay. Note that
implicitly defining a self trigger based on (3.117) is difficult as it re-
quires a continuous check across t. Instead, we will consider checks at
discrete instants of time, separated by a period ∆s . In the absence of a
disturbance, and for ti + n∆s ≤ ti+1 , we define
n−1
n
Ajd Bd L,
X
x(ti + n∆s ) = R(n)x(ti ), R(n) = Ad + (3.118)
j=0
Z ∆s
Ad = eA∆s , Bd = eA(∆s −τ ) Bdτ. (3.119)
0
Let tmin be the minimum time between updates, which will be implicitly
defined by our choice of λ, and let tmax be the maximum time, which
3.4. Event-based control 97
Theorem 3.31. Consider the system (3.115) with controller (3.81) and
self trigger (3.120). The resulting system is exponentially input to state
stable. That is there exists constants λ > 0, κ > 0, and γ > 0 such that
for any essentially bounded function w and x(0) ∈ Rn , we have
We also let F̄11 = F11 (h), F̄12 = F12 (h), F̄21 = F21 (h), F̄22 = F22 (h).
For small enough sampling period h, it can be shown that F11 (τ ) is
invertible for all τ ∈ [0, h] and there exists a matrix S̄ that satisfies
−1
S̄ S̄ T = −F̄11 F̄12 . Given this we have the following result from Heemels
et al. (2013).
The threat of attacks create new challenges for achieving reliable per-
formance in CPS. We note that CPS have become targets for malicious
adversaries. In this section, we will see that achieving security in CPS
requires tools which extend beyond what is offered in state of the art
software and cyber security. In particular, we will consider a science
of cyber-physical system security which combines tools from both cy-
ber security and system theory to defend against adversarial behavior.
We discuss realistic and intelligent attack models from an adversarial
perspective and then offer mechanisms for defenders to recognize and
respond to such malicious behavior in order to achieve resilience.
More specifically, the rest of the section is summarized as follows.
First, in subsection 4.1, we discuss the motivation for investigating
cyber-physical system security. Next, in subsection 4.2, we consider
possible attacks on cyber-physical systems. Here, we classify adversaries
by their knowledge, disclosure resources, and disruption resource before
placing a significant focus on undetectable integrity attacks. Then, in
subsection 4.3, we begin pursuing defensive strategies. In particular,
we discuss how the robust structural design of control systems can
prevent classes of stealthy attacks. Later, in subsection 4.4, we review
103
104 Secure Cyber-Physical Systems
4.1 Motivation
which can interface with both field devices and the monitoring layer.
For instance, in SCADA systems, remote terminal units often allow for
dial up access and may not even require authentication. An attacker
can also take the initiative to introduce vulnerabilities to CPS devices
by targeting supply chains. If production is not performed securely,
adversaries can install backdoors in components, which can later be
leveraged to compromise the CPS.
Beyond attempting to access CPS through a network, an attacker
can simply attempt to target the physical plant itself. In many cases,
due to the scale of CPS it is impossible to physically monitor and protect
all devices and components. As an example, it is often the case that
substations as well as smart meters and PMUs are left unattended in
the electricity grid. Likewise, it is impractical to guard all the sensors,
pumps, and valves in a water distribution system or traffic lights and
vehicles in a transportation system. The defender must also account for
the actions of malicious insiders. Malicious insiders can leverage their
understanding of a CPS and their access to the system in order to target
the infrastructure. A notable example in this regard is the Maroochy
Shire incident (Slay and Miller, 2007) where a former employee was
able to hack a SCADA system performing waste management, causing
millions of gallons of sewage to leak. A malicious insider’s actions can
be amplified if intelligent access control policies are not implemented.
Estimation
We obtain a minimum mean squared error estimate by using a Kalman
filter as follows:
x̂k+1|k = Ax̂k|k + Buk , x̂k|k = x̂k|k−1 + Kk zk , (4.2)
T T −1
zk = yk − C x̂k|k−1 , Kk = Pk|k−1 C (CPk|k−1 C + R) , (4.3)
T
Pk|k = Pk|k−1 − Kk CPk|k−1 , Pk+1|k = APk|k A + Q, (4.4)
where we define
x̂k|k , E[xk |y0:k ], x̂k|k−1 , E[xk |y0:k−1 ], (4.5)
Pk|k , E[ek eTk |y0:k ], ek , xk − x̂k|k , (4.6)
Pk|k−1 , E[ek|k−1 eTk|k−1 |y0:k−1 ], ek|k−1 , xk − x̂k|k−1 . (4.7)
We observe that Pk|k−1 and Kk converge to unique matrices, which we
define as P and K respectively. Assuming that the system has been
running for a long time, we assume Pk|k−1 = P and Kk = K for all k.
Thus, here Σ = P .
For resilient operation of CPS, it is imperative for the operator
to recognize attacks as they occur. One possible method to achieved
resilience in attack detection is through control algorithm design. Here,
the defender uses the information Ik that includes model information
M = {A, B, C, Q, R, P, K}, the set of prior outputs, the set of prior
inputs, and an initial state estimate x̂−1|0 = x̄0|−1 . That is, Ik =
{M, y0:k , u0:k−1 , x̂−1|0 }.
4.2. Attacks on CPS 113
Control
A defender’s goal in CPS is to design a resilient controller C, which
under normal operation, allows the defender to successfully meet control
objectives. Under attack, the defender aims to achieve graceful degra-
dation of performance. Mathematically, a controller C is a sequence of
functions {Uk } which take a defender’s information Ik and obtains a
control input as uk = Uk (Ik ).
When considering the LQG controller, we will typical assume that this
convergence has occurred.
The likelihood ratio test is the most powerful test at size ᾱk where
ᾱk = Pr(Λ(y0:k ) > τ |H0 ). In particular, it is the statistical test which
maximizes the probability of detection βk for a given probability of false
alarm αk = ᾱk .
has two design parameters, the threshold τ and the window size WD .
The threshold captures tradeoffs between the probability of detection
and the probability of false alarm. Specifically, increasing the threshold
simultaneously reduces the probability of false alarm and detection.
The window size captures tradeoffs between time to detection and av-
erage detection performance. A larger window likely improves average
detection performance, as there is more evidence of an attacker’s ma-
nipulation. However, a large window also increases time to detection
since it will take a longer time to ignore older "normal" measurements
when computing a detection statistic.
Of course the preceding subsection provides only a brief overview of
attack detection in CPS. For a more detailed survey, see Giraldo et al.
(2018).
System Knowledge
An attacker’s system knowledge refers to his/her a priori understanding
of the dynamics of the system. Increased familiarity with a CPS can be
leveraged by an attacker to construct more sophisticated and effective
attacks. In the context of the model presented in subsection 4.2.1, an
attacker’s system knowledge may consist of some imperfect estimate of
the system model M̂, the controller C,ˆ and the detector D̂. Denoting
the adversaries information at time k as Ika , we observe that
ˆ D̂} ⊂ I a .
{M̂, C, (4.14)
k
4.2. Attacks on CPS 117
Disclosure Resources
Disclosure resources refer to the collection of data an adversary gathers
during the operation of a CPS. In our model, the attacker can potentially
observe a subset of control inputs and a subset of sensor measurements.
It can often be assumed that an attacker will only be able to observe
some constant fixed subset of inputs and outputs. We define the set
of readable inputs and outputs respectively as Sur , {i1 , · · · , ip0 } and
Syr , {j1 , · · · , jm0 }. Here j ∈ Sur and l ∈ Syr implies an attack can read
the jth entry of uk and lth entry of yk for all k.
Thus at each time step k an attacker is able to read Υy yk and
0 0
Υu uk−1 , where Υu ∈ {0, 1}p ×p , Υy ∈ {0, 1}m ×m . Moreover, Υu and Υy
are defined entrywise as
Υu (s, t) = 1is =t , Υy (s, t) = 1js =t . (4.15)
Here, Υu (s, t) refers to the entry of Υu at the sth row and tth column
and 1 is the indicator function. Thus, we observe that the attacker’s
information Ik grows as follows.
a
Ik−1 ∪ {Υy yk , Υu uk−1 } ⊂ Ika . (4.16)
Disclosure resources enable an attacker to improve upon prior knowl-
edge of a system’s model and controller, which in turn can be used
to construct an intelligent attack. The collection of data in a CPS by
passive observation is referred to as an eavesdropping attack. In certain
cases, disclosure resources can be directly used to construct an attack
without any need for additional data processing. One instance is a replay
attack, described later in this section.
Disruption Resources
An attacker’s disruption resources describe an attacker’s ability to act
on and corrupt a CPS under consideration. We examine three methods
through which an attacker can disrupt a CPS: integrity attacks, denial
of services attacks, and topology attacks. During an integrity attack, an
attacker is able to modify a subset of the defender’s inputs and outputs.
The dynamics of the attacked system are given by
xk+1 = Axk + Buk + B a uak + wk , yk = Cxk + Da dak + vk . (4.17)
118 Secure Cyber-Physical Systems
The proof follows from the definition of a stealthy attack and the
linearity of the system. Details are left to the reader. Note that the
stealthiness of an attacker’s inputs is independent of the defender’s
control strategy in the deterministic case. We next consider attacks that
are stealthy for all k ≥ 0. We define a perfect attack as follows.
In other words, the set of perfect attacks is the set of all attacks in
deterministic systems with known initial state that are stealthy for all
time k. We can relate perfect attacks to the fundamental property of
left invertibility.
where
2 0 4 −3 2 1
0 3 −2 1 1 −2 1 0 0 0
A= , B = , C = 0 1 −1 0 .
3 −5 3 0 0 1
0 0 −2 2
0 −1 −1 4 1 0
4.2. Attacks on CPS 123
Assume that the first actuator and first sensor are being attacked. In
h iT h iT
this case we have B a = 2 1 0 1 and Da = 1 0 0 . Moreover
m∗ = p∗ = 1.
From Corollary 4.4, we can see that there is no perfect attack. Indeed,
we have that for all but finitely many λ ∈ C.
λ−2 −4 −2 0
0 3
0
λ−3 2 −1 −1 0
−3 5 λ−3 0 0 0
rank 0
1 1 λ − 4 −1 0 = 6 = n + m∗ + p∗ .
1 0 0 0 0 1
−1
0 1 0 0 0
0 0 −2 2 0 0
(4.31)
However, when λ = 1 or λ = 5, the rank of the preceding matrix pencil
drops to 5. As such, while there is no perfect attack, there does exist a
zero dynamics attack.
Before continuing, we remark that if B a and Da each have full
column rank (as currently constructed), then strong observability will
imply the left invertibility of a system. We now wish to assess the impact
of zero dynamics attacks. The true impact of the attack is dependent
on the control strategy Uk . For our purposes, we assume the defender’s
goal is to stabilize the system at 0. This can be accomplished even if x0
is unknown if (A, B) is stabilizable and (A, C) is detectable by using
state feedback and a stable observer.
Assume yk (x0 , u0:k−1 , ua0:k−1 , da0:k ) = yk (x00 , u0:k−1 , 0, 0) for all k ≥ 0.
Let xk (x0 , u0:k−1 , ua0:k−1 , da0:k ) denote the state xk generated by (4.21)
as a function of the initial state, the defender’s input, and the attacker’s
inputs. Under attack we observe that xk = xk (x0 , u0:k−1 , ua0:k−1 , da0:k ).
The defender, however has designed his feedback control inputs u0:k−1
so that he stabilizes a system with initial state x00 .
In this case, we make the assumption that
lim xk (x00 , u0:k−1 , 0, 0) = 0. (4.32)
k→∞
By the linearity of the system we see that
xk = xk (x00 , u0:k−1 , 0, 0) + xk (x0 − x00 , 0, ua0:k−1 , da0:k ). (4.33)
124 Secure Cyber-Physical Systems
∆zk , zk (e0|−1 , v0:k , w0:k−1 , ua0:k−1 , da0:k ) − zk (e0|−1 , v0:k , w0:k−1 , 0, 0).
(4.37)
We arrive at the ensuing result.
From the stability of the Kalman filter the bias on the residue ∆zk
asymptotically approaches 0. In this case, we see that an attacker will be
asymptotically stealthy against a χ2 detector so that limk→∞ βk −αk = 0.
The prior result also applies to alternative continuous residue based
detectors with finite memory. We will soon demonstrate that small values
of ∆zk fundamentally lead to poor detection performance. Using the
same rationale as in the deterministic case, the impact of a zero dynamics
attack on the state xk in a stochastic system can be characterized using
the state δxk in (4.30). Consequently, in many scenarios, a zero dynamics
attack allows an attacker to surreptitiously destabilize a control system.
We conclude our study of zero dynamics attacks, by relating such
attacks to the class of unidentifiable attacks in control systems. We
assume an adversary is unable to insert their own actuators. Suppose an
attacker targets actuators Kua = {δ1 , · · · , δp∗ } ⊂ {1, · · · , p} and sensors
Kya = {η1 , · · · , ηm∗ } ⊂ {p + 1, · · · , p + m}. To write the corresponding
B a and Da uniquely as a function of their attack set we, without loss
of generality,
h assume all attacki sets are given in ascending order. Here,
B (Ku ) = Bδ1 · · · Bδp∗ where Bδi is the δi th column of B. Da (Kya )
a a
δxk+1 = Aδxk + B a (Ku ∪ Ku0 )ũak , 0 = Cδxk + Da (Ky ∪ Ky0 )d˜ak . (4.40)
The input sequence is nonzero since K0 6= K and all sensors and actuators
are attacked. Thus, there exists a zero dynamics attack on a set of 2q or
fewer actuators or sensors. Now suppose there is a zero dynamics attack
on a set of 2q or fewer nodes K∗ . Assume, without loss of generality that
all nodes are attacked. In addition, without loss of generality assume
K∗ = K ∪ K0 where K = Ku ∪ Ky , K0 = Ku0 ∪ Ky0 , Ku0 , Ku ⊂ {1, · · · , p} ,
and Ky0 , Ky ⊂ {p + 1, · · · , p + m}. Moreover, without loss of generality,
assume K ≤ q, K0 ≤ q, K ∩ K0 = ∅, and |K0 | ≤ |K|. We know there
exists a zero dynamics attack {uak },{ūak },{dak },{d¯ak }, with each node
being attacked satisfying
Thus, for all k ≥ 0, we have an attack sequence {ūak },{dak } targeting all
sensors and actuators in K satisfying
A, [B a (Ku ) 0n×|Ky | ], C, [0m×|Ku | Da (Ky )] is strongly observable and
left invertible.
This result follows immediately from Theorem 4.5 and Theorem
4.10.
Example 4.3. Consider the following control system
xk+1 = Axk + Buk , yk = Cxk , (4.43)
where
1 2 1 1 0 1 0 0
A = −2 3 1 B = 0 1 , C = 0 1 0 . (4.44)
3 −3 4 0 0 0 0 1
We would like to determine how many corrupted nodes can be identified.
By construction, since there are only 3 sensors, we know that there are
attacks on 3 nodes which can not be detected. Indeed, if all 3 sensors
are attacked, then for any eigenvalue λ(A) of A, we have
" #!
λ(A)I − A 0
rank < 6. (4.45)
C I
Next, it can be shown that for all feasible attacks K of size 2, and for
all λ ∈ C,
" #!
λI − A −B a (Ku ) 0n×m∗
rank = 5. (4.46)
C 0m×p∗ Da (Ky )
As a result, the attack system is strongly observable and left invertible
for any attack of size 2. Consequently, attacks of size 1 (a single corrupted
actuator or sensor) can be identified. Note, that as considered thus far,
the number of attack subsystems we have to potentially evaluate is
combinatorial. Simpler, graphical conditions for system verification will
be discussed in the next subsection.
We note that if B is not injective, this provides a path for an
adversary to generate unidentifiable attacks. For instance, if redundant
actuators are used and one or more are compromised, it would be
impossible for a defender to determine which if any actuators are secure.
While redundancy could compromise the ability to identify attacks, it
does not affect the ability to perform resilient estimation.
130 Secure Cyber-Physical Systems
moreover assume that inputs {dak } and {uak } are chosen independently
of the systems true inputs and outputs. Thus, they are independent of
the stochastic processes {wk } and {vk }.
In this case, we again assume the attacker wishes to remain stealthy.
As such the attacker designs his input sequence so that
q
∆zkT (CP C T + R)−1 ∆zk ≤ B, ∀k≥0 (4.48)
where we note that the probability of false alarm αk for the χ2 detector
is a constant in k.
Then there exists some 0 < δ < 1 and a passive detector which satisfies
1
βk ≥ δ, ∀ k ≥ 0, lim sup − log(αk ) > (4.52)
k→∞ k+1
Moreover, suppose yk collected under attack is ergodic and that
−1
1 NX
lim ∆zkT (CP C T + R)−1 ∆zk ≤ . (4.53)
N →∞ 2N
k=0
∆ek , ek (e0|−1 , v0:k , w0:k−1 , ua0:k−1 , da0:k ) − ek (e0|−1 , v0:k , w0:k−1 , 0, 0),
(4.55)
where ek is written as a function of the initial apriori estimation error,
the noise realizations, and the attacker’s inputs. It can be shown that
xk (x0 , x̂0|−1 , v0:k , w0:k−1 , ua0:k−1 , da0:k ) − xk (x0 , x̂0|−1 , v0:k , w0:k−1 ),
x̂k|k (x0 , x̂0|−1 , v0:k , w0:k−1 , ua0:k−1 , da0:k ) − x̂k|k (x0 , x̂0|−1 , v0:k , w0:k−1 ).
Here, xk and x̂k|k are written as a function of the initial state, the initial
apriori state estimate, the noise realizations, and the attack inputs. For
aesthetic purposes, we omit the attack inputs in the second term of
each expression, corresponding to the absence of an attack.
It can be shown that ∆xk is bounded if and only if ∆ek is bounded
when (4.48) is satisfied. Moreover, by similar methods the a priori
estimation error bias ∆ek|k−1 will be bounded if and only if ∆xk is
bounded. We now obtain the following result from Mo and Sinopoli
(2010), which characterizes when an attacker can destabilize ∆ek in the
sensor attack case.
Theorem 4.15 (Mo and Sinopoli (2010)). Consider a false data injection
attack where B a = 0. There exists a feasible attack input sequence
satisfying (4.48), which destabilizes ∆ek so that lim supk→∞ k∆ek k2 =
∞ if and only if A has unstable eigenvalue λ with eigenvector v, which
satisfies
1. Cv ∈ Im(Da ),
1. Cv ∈ Im(Da ),
134 Secure Cyber-Physical Systems
2. v is an eigenvector of (A + B a La ).
1 1
h i h i
C̄ = P̄ − 2 0 CA , D̄ = P̄ − 2 CB a Da . (4.60)
Without loss of generality, it is assumed uak = 0, k ≤ −1 and dak =
0, k ≤ 0. We let ua∞ ,= {ua0 , da1 , ua1 , da2 , ua2 , · · · }.
Definition 4.10. An attacker’s actions ua∞ are feasible if (4.48) holds
for B = 1.
" #
∆xk
Definition 4.11. The reachable region Rk of is defined as
∆ek
" #
∆xk (ua∞ )
2n
Rk = {x̄k ∈ R , x̄k = for some feasible ua∞ }. (4.61)
∆ek (ua∞ )
The union of all Rk is defined as
∞
R = ∪ Rk . (4.62)
k=0
Thus, R characterizes all biases an attacker can inflict on the system.
We next provide a recursive definition of these reachability sets. In
order to do this, we must define the reach set, Rch, and the one step
set, Pre, operators.
Definition 4.12. Given a set S ⊂ R2n we define
Rch(S) , {x̄+ ∈ R2n : ∃ζ ∈ Rp∗ +m∗ , x̄ ∈ S (4.63)
s.t. Āx̄ + B̄ζ = x̄+ , kC̄ x̄ + D̄ζk2 ≤ 1},
Pre(S) , {x̄ ∈ R2n : ∃ζ ∈ Rp∗ +m∗ , (4.64)
s.t. Āx̄ + B̄ζ ∈ S, kC̄ x̄ + D̄ζk2 ≤ 1}.
The reach set of S characterizes the set of states that can be reached
while the current state is in S while ensuring the attacker is stealthy
at the current time step. Note that the reach set does not account
for stealthiness at future time steps. The one step set of S describes
the states that can be driven to a state in S in one time step, while
remaining stealthy at the current time step. Using the Rch and Pre
operators, the reachable set Rk can be recursively computed.
136 Secure Cyber-Physical Systems
Theorem 4.18 (Mo and Sinopoli (2016)). The reachability set Rk can
be computed as
Replay Attack
During a replay attack, an attacker attempts to fool the defender into
believing a system is operating normally, by sending a prior sequence
of sensor measurements, collected during normal system operation. The
attacker follows the following steps.
1. (s)he records a long sequence of outputs y−T 0 :−T where (T 0 > T >
0). As a result, Υy = I.
k
T
zj0 (CP C T + R)−1 (zj0 + 2C(A − AKC)j ζ0 )
X
j=k−WD +1
T
+ C(A − AKC)j ζ0 (CP C T + R)−1 C(A − AKC)j ζ0 (4.70)
lim βk − αk = 0.
k→∞
System Knowledge
Zero Dynamics
Attack
Covert
False Data
Attack
Injection
Attack
Eavesdropping
Attack
Disclosure
Resources
DoS Attack
Replay Attack
Disruption
Resources
F = {F ⊂ V : |F | ≤ q}. (4.81)
We define the graph GFa = (VFa , EFa ) of a DCS when a set F of agents/
sensors is compromised.
where ula (k) is an input from node ual at time k. If xi is secure then
ula (k) = 0. We define yia (k) as the output of yi at time k under attack.
If (ual , yi ) ∪ (xj , yi ) ⊂ EFa , then
If yi is secure then in (4.83), ula (k) = 0. Concatening xia (k), yia (k), and
uia (k) into xa (k), ya (k), and ua (k), we have :
with BFa (i, j) , 1(uaj ,xi )∈EU a ,X , DFa (i, j) , 1(uaj ,yi )∈EU a ,Y . We assume the
F F
attacker knows (A, BFa , C, DFa ). The estimator policy remains unchanged
during an attack.
Definition 4.13. ([A], [BFa ], [C], [DFa ]) is structurally left invertible (or
strongly observable) if an admissible realization of (A, BFa , C, DFa ) is
left invertible (or strongly observable).
3. ∆0 ⊂ Vess (UFa , Y)
where ∆0 = {x ∈ X |ρ(x ∪ UFa , Y) = ρ(UFa , Y)}.
As a result, structural left invertibility for a fixed attack strategy
requires the existence of a |UFa |-linking from the set of attack inputs to
the set of outputs. Moreover, stronger conditions are required to achieve
strong observability.
150 Secure Cyber-Physical Systems
0 0 ∗ ∗ 0 0
From Theorem 4.22, the system is structurally strongly observable and
left invertible.
First, the edges (u2 , y1 ), (u1 , x2 ), (x2 , y2 ), (x1 , x1 ), (x3 , x3 ), (x4 , x4 )
constitute a maximal matching so that θ(X ∪ UFa , X ∪ Y) = n + q = 6.
Next, each agent or state has a path to the outputs. In fact each
state has a directed edge to a sensor (x1 , y1 ), (x2 , y2 ), (x3 , y2 ), (x4 , y3 ).
Finally, we note that ∆0 = ∅. Indeed ρ(x1 ∪UFa , Y) = ρ(x2 ∪UFa , Y) =
ρ(x3 ∪ UFa , Y) = ρ(x4 ∪ UFa , Y) = 3 while ρ(UFa , Y) = 2.
We are careful to remark that just because this system is structurally
strongly observable and left invertible does not mean all valid realizations
are strongly observable and left invertible. In fact the system considered
in Example 4.2 is a valid realization which we have shown is vulnerable
to a zero dynamics attack. Nonetheless, almost all valid realization
except a set of zero Lesbesgue measure will be strongly observable and
left invertible. In particular, a structurally strongly observable and left
invertible system with independent system parameters (system matrix
entries) will almost surely be strongly observable and left invertible.
B a
We make the assumption that is injective. This means the attack
Da
inputs need not be dedicated attack inputs. We attempt to make this
assumption structurally
" # with θ(U a , X ∪ Y) = |U a |. This is a necessary
Ba
condition for to be injective and it is a sufficient condition almost
Da
surely.
This assumption can be made without loss of generality, as the effect
of an attack on the state of a system where the input matrices are not
injective is equivalent to the effect of an alternative attack on a system
where the input matrices are injective.
152 Secure Cyber-Physical Systems
Corollary 4.25. For all feasible sets of attacks, assume that we have
|U a | ≤ q and θ(U a , X ∪ Y) = |U a |. Then, a system is structurally
left invertible for all feasible attack vectors if for all xi ∈ X , the
minimum vertex separator Si between (xi , o) in g(G) has size |Si | ≥ q.
Moreover, the system is structurally left invertible and structurally
strongly observable if
The proof is similar to the proofs of Theorem 4.23 and 4.24 and
is left to the reader. The conditions obtained to stop attackers with
dedicated inputs remain sufficient for security even when the attacker
has the freedom to inject inputs that are not dedicated. We note that in
practice, it may be difficult for a defender to verify the first condition
of Theorem 4.24 as the problem appears combinatorial. However, the
problem is greatly simplified if each agent has self loops.
Lemma 4.27. Suppose each agent has a self loop. Then, [A], [C] is not
discreetly attackable only if the out degree of each node xi ∈ X satisfies
|NxOi | ≥ q + 2.
1
case complexity of O(n(2|V|) 2 (|E 0 | + |V| − 1)) where V and E 0 are
associated with matrices [Ā], [C].
Joint Design: We now wish to minimize both sensing and communi-
cation costs through our choice of communication links and dedicated
sensor placement. Suppose the cost of a communication link is α1 ≥ 0
and the cost of a sensor is α2 ≥ 0. We wish to solve the following
problem:
Figure 4.4(a) with self loops abstracted away. If [Ā](u, v) is not a fixed
zero, there exists an edge (xv , xu ). Suppose the goal is to design an
optimal communication network which prevents all perfect attacks when
q = 2 and all zero dynamics attacks when q = 1. Recalling Algorithm 3,
we start with the digraph associated with [Ā], and for each of the state
vertices xi we keep enough outgoing agent neighbors to ensure the size
of the minimum vertex separator between (xi , o) is q + 1 (to ensure the
system is not discreetly attackable) or q (to ensure the system is not
perfectly attackable). Figures 4.4(b), 4.4(c), and 4.4(d) show the results
of these iterations.
Figure 4.4: Process of Algorithm 3, starting with the constraint matrix in (a).
4.3. Robust Structural Design of Control Systems 157
In subsection 4.2, we have seen that there are instances when passive
detection techniques provably fail, for instance during zero dynamics
attacks, false data injection attacks and replay attacks. One possible
method to counter stealthy attack scenarios is to consider offline robust
system design to reduce the space of stealthy attacks as considered in
subsection 4.3. In this subsection, we consider a different method for
attack detection. We consider how a defender can intelligently change
his policy online in order to better distinguish between normal and
attack scenarios. We refer to this process as active detection.
In active detection, the defender aims to authenticate the physical
dynamics of the system. In particular, the defender changes his policy
by introducing a physical secret into the system that is unknown to the
adversary. This physical secret enables a challenge response mechanism
in the CPS. Here, an attacker who targets a CPS must do so in a
manner that is consistent with a secret embedded in the dynamics
of the system to remain undetected. This serves as a response to the
defender’s challenge. Since the attacker is presumably unaware of the
160 Secure Cyber-Physical Systems
uk Plant yk
A/ack Strategy
C
P
(independent
of secret)
Communica*on Network
uk uk
z 1
uk 1
S
x̂k State
Controller yk
Estimator
x
Failure Detector
Failure Detector
zk Ac$ve Detec$on:
Physical Secret is
introduced via
Recognizes a9ack Possible Detection Methods: Dynamics or Input
that doesn’t use Input: Physical Watermarking
secret Dynamics: Moving Target
4.4. Active Detection of Attacks on CPS 161
Physical Watermarking
a) Output with Optimal Input b) Add Watermark uk
Faulty/Attack Normal
linear controller
u∗k = Lx̂k|k , (4.93)
where the control gain matrix L is defined as
−1
L , − B T SB + U B T SA, (4.94)
where P is the state state apriori state estimation error covariance and
K is the steady state Kalman gain. Due to watermarking, the LQG
cost increases to J = J ∗ + ∆J where from Mo and Sinopoli (2009), we
have
∆J = tr ((U + B T SB)J ) (4.97)
Note that cost of control increases linearly with the covariance of the
watermark. Thus, increasing the randomness of the watermark degrades
performance. However, this randomness also is the key mechanism
164 Secure Cyber-Physical Systems
subject to U − B T J B = AUAT ,
tr ((U + B T SB)J ) ≤ δ,
Such a design balances the need to ensure that watermarking leads
to sufficient detectability without significantly affecting system perfor-
mance.
T
Theorem 4.31 (Mo et al. (2015)). Let sym(X) = X+X
2 . The additional
LQG cost ∆J due to the additive watermark ∆uk is:
∞
( " #)
X
d
∆J = tr U Γ(0) + 2U sym L (A + BL) BΓ(1 + d)
d=0
h i
T
+ tr (W + L U L)Θ1 , (4.103)
where ∞
X h i
Θ1 , 2 sym (A + BL)d L1 (Γ(d)) − L1 (Γ(0)),
d=0
and L1 : Cp×p → Cn×n is a linear operator defined as
∞
X
L1 (X) = (A + BL)i BXB T ((A + BL)i )T
i=0
= (A + BL)L1 (X)(A + BL)T + BXB T . (4.104)
maximize tr(Σ̄P̄ −1 )
Γ(d)∈G(ρ̄)
subject to ∆J ≤ δ, (4.111)
Both the objective function and ∆J are linear with respect to the
autocovariance functions Γ(d). Unfortunately, there are infinitely many
optimization variables Γ(d). Moreover, it is unclear how we can guarantee
that Γ(d) ∈ G(ρ̄). To address this we make the following additional
assumption.
Assumption 1: Γ̃(d) = ρ̄−|d| Γ(d) is an autocovariance function.
Intuitively if ρ(Aω ) is marginally smaller than ρ̄, Assumption 1
can be more easily satisfied. If ρ̄ = 1, the space is not constricted by
Assumption 1 and in fact one will be able to optimize over all stationary
Gaussian watermarks. Given this additional assumption, an equivalent
formulation can be obtained by converting the optimization problem to
the frequency domain using Bochner’s theorem. In particular, according
to Bochner’s theorem Γ(d) can be taken in the following form
Z 1
2
Γ(d) = exp(2πωdj)dν(ω) (4.112)
− 21
subject to H ≥ 0, 0 ≤ ω ≤ 0.5
F1 (ω, H) ≤ δ,
(4.114)
with Ψ = (1 − ρ̄2 )I. Here, Re(h) and Im(h) refer to the the real and
imaginary potions of h respectively. Thus, the dimension of the hidden
state is always 2, and the resulting stationary Gaussian watermark is a
noisy sinusoid.
The attacker can choose uak arbitrarily to perturb the system along the
controllable subspace (A, B). To perfectly avoid detection, the attacker
chooses dak as follows.
xak+1 = Axak +Buk +wka , dak = Cxak +vka −Cxk −vk , xa0 = x0 , (4.119)
where vka ∼ N (0, R), wka ∼ N (0, Q) are IID and independent processes.
Such an attack remains effective even if x0 is unknown to the attacker.
In both these cases, the attacker not only uses his or her access to the
channels, but also a detailed understanding of the system model. We
endeavor to limit the effectiveness of an adversary by removing their
system knowledge.
The matrices are taken as IID random variables which are independent
of the sensor and process noise processes with distribution
A1,k , A2,k , Bk , Ck+1 ∼ fA1,k ,A2,k ,Bk ,Ck+1 (A1 , A2 , B, C). (4.123)
where
" # " #
Q̃ Q̃12 R̃ R̃12
Q= > 0, R= > 0. (4.125)
Q̃T12 Q T
R̃12 R
Since the moving target system is linear and the noises remain
Gaussian, we can use a Kalman filter to still perform state estimation.
4.4. Active Detection of Attacks on CPS 173
Remark 4.6. While the system introduced above involves IID matrices
A1,k , A2,k , Bk , Ck+1 , the moving target design can still be effective in
other scenarios. For instance, the dynamics need not be linear as long
as the defender can accurately model the system. Moreover, the system
parameters can evolve at multiple time scales. In this case, the longer
the target remains in place, the easier it is for the adversary to identify
the system.
Observe that the adversary can exactly subtract his influence from
measurements yk due to his knowledge of the system model. However,
the adversary should be unable to completely subtract his bias from
the extraneous sensors ỹk .
Define ȳka , [ỹkaT ykaT ]T , x̄k , [x̃Tk xTk ]T , w̄k , [w̃kT wkT ]T , v̄k ,
[ṽk vkT ]T , and ȳk , [ỹkT ykT ]T . The adversary’s observations can be
T
We show that the pdf can be recursively computed at each step. Letting
ζk+1 = {x̄k+1 , x̄ak+1 , Ck+1 } we have
a
f (ζk+1 |Ik+1 ) = f (ζk+1 |Ika , ȳka , ȳk+1 , d¯ak , uak , uk ),
= f (ζk+1 |Ika , ȳk+1 , uak , uk ),
f (ȳk+1 |Ika , ζk+1 )f (ζk+1 |Ika , uk , uak )
= . (4.137)
f (ȳk+1 |Ika , uk , uak )
178 Secure Cyber-Physical Systems
The inputs uk , uak and d¯ak are incorporated into the definition of Fk , while
uncertainty in the model (A1,k , A2,k , Bk , Ck+1 ) can be incorporated in
the process noise ωk . It can be shown that the following posterior
Cramer-Rao lower bound holds
h i
c
Efk+1 ē0:k+1 ēT0:k+1 |ȳ1:k ≥ I −1 (ζ0:k+1 |ȳ1:k ), (4.147)
where
Observe that (4.147) gives us an expected lower bound for the error
matrix associated with the entire state history ζ0:k+1 with knowledge of
measurements ȳ1:k . This expectation is taken over the state history as
well the measurement ȳk+1 so that ζ̂0:k+1 is a function of the measure-
ment ȳk+1 . Observe that unlike the traditional Cramer-Rao bound which
is limited to unbiased estimators, the Bayesian Cramer-Rao bound here
considers both biased and unbiased estimators ζ̂. While the lower bound
given here applies to the entire state history ζ0:k+1 , in practice we care
about estimating a lower bound on the current state ζk+1 . Nonetheless,
one can show that
h i
c
Efk+1 ēk+1 ēTk+1 |ȳ1:k ≥ I −1 (ζk+1 |ȳ1:k ), (4.151)
where
h i
Dk11 = Efk+1
c −4ζζkk log f (ζk+1 |ζk , ȳ1:k ) ,
h i
ζ
Dk12 = Efk+1
c −4ζk+1
k
log f (ζk+1 |ζk , ȳ1:k ) = (Dk21 )T ,
h i
ζ
Dk22 = Efk+1
c −4ζk+1
k+1
log f (ζk+1 |ζk , ȳ1:k )f (ȳk+1 |ζk+1 ) .
In addition,
−1
IA (ζk |ȳ1:k ) = Ek22 − Ek21 Ek11 Ek12 , (4.153)
where
h i
ζ
Ek11 = Ef (ζ0:k |ȳ1:k ) −4ζ0:k−1
0:k−1
log f (ζ0:k |ȳ1:k ) ,
h i
Ek12 = Ef (ζ0:k |ȳ1:k ) −4ζζk0:k−1 log f (ζ0:k |ȳ1:k ) = (Ek21 )T ,
h i
Ek22 = Ef (ζ0:k |ȳ1:k ) −4ζζkk log f (ζ0:k |ȳ1:k ) .
We observe that it is still difficult to obtain matrices Ek11 , Ek12 , Ek21 , Ek22
so Zuo et al. (2011) introduces the following approximate recursion.
h i−1
T
IA (ζk |ȳ1:k ) ≈ Sk22 − Sk12 Sk11 + IA (ζk−1 |ȳ1:k−1 ) Sk12 , (4.154)
where
h i
ζ
Sk11 = Ef (ζ0:k |ȳ1:k ) −4ζk−1
k−1
log f (ζk |ζk−1 , ȳ1:k−1 ) ,
h i
Sk12 = Ef (ζ0:k |ȳ1:k ) −4ζζkk−1 log f (ζk |ζk−1 , ȳ1:k−1 ) ,
h i
Sk22 = Ef (ζ0:k |ȳ1:k ) −4ζζkk log f (ζk |ζk−1 , ȳ1:k−1 )f (ȳk |ζk ) .
Example 4.14. We test the moving target on the quadruple tank pro-
cess, a four state system Johansson (2000). The goal is to control the
water level of two of four tanks using two pumps. Two sensors mea-
sure water heights. We use an LQG controller with weights following
suggestions in Grebeck (1998). Q and R are created by generating a
matrix from a uniform distribution, multiplying it by its transpose, and
dividing by 100.
4 extra states and 2 extra outputs are added. The time varying ma-
trices A1,k , A2,k , Bk , Ck+1 are somewhat sparse (50% of entries nonzero).
The non-zero elements follow a multivariate Gaussian distribution with
means generated from U (−0.5, 0.5). The covariances of the nonzero pa-
rameters are created by generating a matrix from a uniform distribution,
multiplying it by its transpose, and dividing by 100.
4.4. Active Detection of Attacks on CPS 183
(a) Full Knowledge: Det. Stat vs (b) Stochastic knowledge: Det. Stat
Time vs Time
(c) Mean absolute height deviation (d) Mean absolute height deviation
(cm): Full Knowledge (cm): Stochastic Knowledge
to a controller can detect replay attacks while Hoehn and Zhang (2016a)
propose using modulation matrices to remove an attacker’s system
knowledge and prevent zero dynamics and covert attacks.
The topic of physical watermarking has been investigated in detail
in the literature. Beyond the seminal work in Mo and Sinopoli (2009)
and extensions to stationary Gaussian watermarks Mo et al. (2015),
several other areas have been explored. For instance, in addition to the
Neyman Pearson Detector and χ2 detector, alternative detectors have
been explored. In particular, Satchidanandan and Kumar (2017) and
Hespanhol et al. (2017) provide asymptotic detectors, which guarantee
that zero average distortion power is injected into sensor measurements.
Moreover, Mo et al. (2014b) and Chabukswar et al. (2011) propose
the use of correlation detectors, which can be advantageous when the
4.4. Active Detection of Attacks on CPS 185
The ability to detect attacks, a topic of major focus in subsections 4.3 and
4.4, is a necessary component to achieving resiliency in CPS. However,
the ultimate goal is to design systems and architectures which can
maintain performance in the presence of malicious behavior. To design
viable, resilient feedback control laws in the presence of an attacker, it
is important for the defender to have an understanding of the attacked
state. Indeed having a reliable estimate will allow a defender to better
understand the portions of a system that have been compromised and
design attack specific solutions to counter an adversary’s actions. The
nature of these attack specific solutions is system dependent and out of
scope for this monograph. Nonetheless, in order to begin the process
of achieving resilient system performance, this section investigates the
problem of resilient estimation.
To begin, we consider deterministic systems and consider integrity
attacks on sensors and actuators. In this scenario, we allow the errors
4.5. Resilient Estimation in CPS 187
Sensor Attacks
Consider the resilient estimation problem under adversarial attacks at
the sensors only. In particular, the objective here is to estimate the
initial system state x0 from the corrupted observations y0:T −1 . In this
case, we assume B a = 0. Moreover, without loss of generality we are
able to assume uk = 0 as well since the impact of the control input on
188 Secure Cyber-Physical Systems
Theorem 4.35 (Fawzi et al. (2014)). For any integer T > 0, the initial
state is recoverable under q sensor attacks after T steps if and only
if there does not exist x0 , x̄0 ∈ Rn with x0 6= x̄0 , and da0 , ..., daT −1 ∈
4.5. Resilient Estimation in CPS 189
Im(Da (K)) and d¯a0 , ..., d¯aT −1 ∈ Im(Da (K̄)) with |K| ≤ q, |K̄| ≤ q such
that CAk x0 + Da (K)dak = CAk x̄0 + Da (K̄)d¯ak for all k ∈ {0, ..., T − 1}.
Theorem 4.35 illustrates the close relationship between attack iden-
tifiability and state recovery. In particular, an initial state is not recover-
able if there exists an alternative initial state and a possible alternative
attack sequence on q or fewer sensors, which can lead to the observed
measurement sequence. The condition in Theorem 4.35 is also a stan-
dard extension for the property of observability to the case when q
attacks are present. In the following theorem, we provide a simpler
characterization of the recoverability of the initial state under q sensor
attacks.
Theorem 4.36 (Fawzi et al. (2014)). Define the support of a vector
v ∈ Rl as supp(v) = {i ∈ {1, · · · , l}, v(i) 6= 0}, where v(i) is the ith
entry of v. For any integer T > 0, the initial state is recoverable under
q sensor attacks after receiving measurements y0:T −1 if and only if for
any ν ∈ Rn − {0}, |supp(Cν) ∪ supp(CAν) ∪ . . . ∪ supp(CAT −1 ν)| > 2q.
Example 4.15. Let
0 3 1 −1 2 1 0 0 0 0
2 −3 0 0 1 0 1 0 0 0
A=
9 0 3 1 0 , C = 0
0 1 0 0. (4.161)
0 0 1 1 0 0 0 0 1 0
1 2 −1 4 2 0 0 0 0 1
We want to know how many sensor attacks q we can withstand while
still recovering x0 given measurements y0:4 . From Theorem 4.36, we
know that 2q < m. As such, q ≤ 2. We next observe that for any
row C i , the system (A, C i ) is observable. Consequently, |supp(Cν) ∪
supp(CAν) ∪ . . . ∪ supp(CA4 ν)| = 5 for all ν =6 0. As such, q = 2. Up
to 2 sensor attacks can be tolerated with resilient state estimation.
We next consider the estimator that recovers the initial state x0
under the condition in Theorem 4.36.
min |K|
x̂0 ,K
Note that the estimator finds the smallest set of attacked outputs K
which can explain the observed behavior, while simultaneously deter-
mining the initial state. If the initial state is correctly recovered, the
set of malicious sensors is easily determined by comparing the expected
outputs as derived by the true initial state to the received outputs.
Given the output sequence y0:T −1 , the preceding estimator is able to
correctly determine the initial state if the system can correct q sensor
attacks, and q or fewer sensor attacks have occurred.
Theorem 4.37 (Fawzi et al. (2014)). Whenever the initial state is recov-
erable under q sensor attacks after T time steps, solving problem (4.162)
recovers the initial state if q or fewer sensor attacks have occurred.
Despite recovering the initial state x0 , the estimator (4.162) intro-
duces high computational complexity since the optimization is combi-
natorial. A well known approach to tackle the computational burden of
this problem is to relax the zero norm by the one norm. This approach
finds its motivations from recent advances in compressive sensing. Let
us consider the `1 estimator that minimizes the sum of `r norm of errors
m
X
minimize kD̂i kr (4.163)
x̂0 ,dˆa
0:T −1 i=1
subject to D̂i = dˆa0 (i), · · · , dˆaT −1 (i)
yk = CAk x̂0 + dˆak , 0 ≤ k ≤ T − 1
First, we let
Finally, we let yk (x0 , u0:k−1 , B a (Ku )ua0:k−1 , Da (Ky )da0:k ) be the output
yk due to the initial state, the defender’s inputs, and the attacker’s
inputs. We can now provide an equivalent characterization for being
able to recover a sequence of states.
192 Secure Cyber-Physical Systems
yk = C x̂k + dˆak , 0 ≤ k ≤ T − 1
x̂k+1 = Ax̂k + B(uk + ûak ), 0 ≤ k ≤ T − 2.
Again dˆak (i)) and ûak (i) are the ith entries of dˆak and ûak respectively. Thus,
kD̂i kr = 0 would imply the ith sensor is unattacked while kÛj kr = 0
would imply the jthe actuator has not been tampered with. λ controls
the relative weight of sensor and actuator attacks. The prior optimiza-
tion problem (4.166) is convex, introducing an `1 relaxation of the `0
optimization problem (4.165). While in general, this problem (4.166)
may not be able to recover and estimate and perform identification in
the presence of a maximum number of q sensor and actuator attacks as
defined by Proposition 4.39 (and obtained in theory by (4.165)), em-
pirical results obtained by Fawzi et al. (2014) indicate that for a small
enough number of malicious nodes, optimal estimation and identification
in the presence of malicious behavior can be performed.
We denote `n∞
as the space of infinite sequences of n-dimensional vectors
with bounded infinity norm. We will write `∞ when there is no confusion
on dimension of the vector. For any matrix A ∈ Rm×n . We denote its
induced norm as
kAxk∞ X
kAk∞ = sup = max |aij |.
x6=0 kxk∞ i
j
We assume that the defender knows both ε and q. However, the set
of compromised sensors K is unknown. Note here that we will often
consider q as a design parameter of our system. In particular, we will aim
to design estimators that can tolerate up to q sensor attacks. Increasing
q will increase system resilience at the cost of computational complexity
4.5. Resilient Estimation in CPS 197
ek , xk − x̂k , (4.178)
= xk − hk (y0:k−1 ). (4.179)
Lemma 4.42 (Nakahira and Mo (2015)). There does not exist a resilient
estimator for system (4.177) if there exist infinite sequences x, x0 , w,
w0 , Da (K)da , Da (K0 )da0 , y and y 0 of proper dimension, which satisfy
ykI , P I yk .
L , {I ⊂ S : |I| = m − q}.
4.5. Resilient Estimation in CPS 199
These local errors and residues are associated with the following linear
operators
" #
A + K(I)C I B + K(I)DI
EI (K(I)) , , (4.184)
I 0
" #
A + K(I)C I B + K(I)DI
GI (K(I)) , . (4.185)
CI DI
from time 0 to k if (4.187) holds at time k. Note that the sensors from
a valid estimator may still be under attack. However, the resulting
outputs could still have been generated by a valid noise sequence. One
can potentially design better local detectors to check if there exist
compromised sensors in the index set I. However, the local detector
based on (4.187) is sufficient for us to design a resilient estimator.
We further define the set Lk as
Lk , {I ∈ S : (4.187) holds at time k}. (4.188)
We will then fuse all the valid local estimates x̂k (I) at time k to
generate the state estimate x̂k . As a primary goal is to minimize the
infinite norm of the estimation error, we will use the following equation
to compute each entry of x̂k :
1
x̂ik = min x̂ik (I) + max x̂ik (I) . (4.189)
2 I∈Lk I∈Lk
Remark 4.9. Observe that when I only contains secure sensors, then
minimizing the infinite norm of the local estimation error ek (I) is
equivalent to minimizing kEI (K(I))k1 . The second term on the RHS of
(4.190) exists since the estimator does not know which local estimate
can be trusted at the beginning.
Since we are able to achieve a resilient estimator when every sub-
system with m − 2q sensors is detectable, we arrive at the following
result.
Corollary 4.45 (Nakahira and Mo (2015)). A necessary and sufficient
condition for the existence of a resilient estimator is that (A, C J ) is
detectable for any index set J ⊂ S with cardinality m − 2q.
Example 4.16. We illustrate the estimator and its performance by
means of a numerical example. We take the matrices that define the
system as:
1 h i h i
A = 1, C = 1 , B = 1 0 0 0 , D = 0 I .
1
Additionally, we assume that ε = 1 and q = 1. First, we design a
linear estimator described in (4.171) for nonadversarial scenarios.
h Wei
utilize an estimator with the following symmetric gain K = θ θ θ ,
where θ ∈ R. The `1 norm of E(K) is (1 − |1 + 3θ|)−1 (1 + |3θ|) where
|1+3θ| < 1 in order to ensure the stability of the estimator. The optimal
θ, which minimizes kE(K)k1 , is given by θ = −1/3. As a result, when
no attacker is present
1h i
x̂k+1 = x̂k + 1 1 1 (yk − C x̂k ). (4.191)
3
In an adversarial scenario we additionally need the gains K(1, 2),
K(2, 3), K(3, 1). Once again, we consider the following symmetric gains:
h i
K(1, 2) = K(2, 3) = K(3, 1) = µ µ .
We observe that α(J ) = 2 for J = {1}, {2}, {3} and
1 + |2µ| 1 + |2µ|
kEI (K(I))k1 = , kGI (K(I))k1 = 1 + .
1 − |1 + 2µ| 1 − |1 + 2µ|
202 Secure Cyber-Physical Systems
The value of µ that minimizes the right hand side of (4.192) is µ∗ = −0.5
and the corresponding upper bound is 8.
We compare the nonadversarial estimator (4.191) and the resilient
estimator. Here, wk is generated from a uniform distribution on the
set kwk k∞ ≤ 1. The attacker targets only the first sensor with a bias
h iT
increasing in time: Da dak = 0.5k 0 0 . The trajectories for the
estimation error of the estimator (4.191) and the resilient estimator are
plotted in Figure 4.11 and Figure 4.12, respectively.
fusion. The teal line and black line terminate at time 4 and time 12,
respectively. These times correspond to the instants in which detection
of a violation of (4.187) occurs by local detectors. As a consequence,
sensor 1 is identified as a sensor under attack. Note that the error
for the estimator (4.191) grows linearly and hence, it is unbounded.
On the other hand, our resilient estimator detects that the index sets
{1, 2} and {1, 3} contain the compromised sensor and hence discard the
corresponding local estimates. As a result, the estimation error remains
bounded.
Fi 1n = Ki , V −1 Fi J = J V −1 Fi . (4.198)
Thus, this error is stable since (A − KCA) is stable. Note the estimation
error ik can be decomposed as the sum of an error φik due to system
noise and ψki due to an attack.
We concatenate ik , φik , ϕik to obtain ˜k , φ̃k , ϕ̃k respectively. Let W
f ,
limk→∞ Cov(ϕk ) which can be obtained by solving a Lyapunov equation.
The optimal Kalman estimate x̂k|k can be additionally found by solving
the following least squares problem
206 Secure Cyber-Physical Systems
1 H f−1
minimize ˇ W ˇk (4.203)
x̌k ,ˇ
k 2 k
ζ̂ 1 G1
.k .
subject to . = . x̌k − ˇk .
. .
ζ̂km Gm
1 H f−1
minimize φ̌ W φ̌k + γkϕ̌k k1 (4.204)
x̌sk ,φ̌k ,ϕ̌k 2 k
subject to ζ̂ki = Gi x̌sk − φ̌ik − ϕ̌ik , ∀i ∈ {1, · · · , m},
The proof is omitted here, but a similar result and procedures can
be found in Han et al. (2019)
A = diag{1, 1, −2}, B = I3 , Q = I3 ,
C1 1 0 0
C2 1 1 −1
C=
C3 = 1
2 , R = I5 .
1
C
4 1 −1 −0.5
C5 −0.5 1 1
Figure 4.13: The normalized MSE of the secure estimator v.s. different choices of
γ.
210
5.1. Data Privacy 211
about when and where a user travels, while data in a medical CPS can
include sensitive health information about specific patients. This section
aims to provide an initial discussion about how to strike a balance
between the utility achieved in algorithms that leverage big data in CPS
and the need to protect critical information from adversarial actors.
For the rest of the section, we first consider the notion of data
privacy, with an emphasis on differential privacy (subsection 5.1). We
then consider an example of average consensus and design a differentially
private mechanism to guarantee the privacy of the initial conditions
of consensus (subsection 5.2). Due to the limitation of the differential
privacy mechanism, e.g., it cannot achieve the exact average consensus,
we consider a different concept, inference privacy, and propose an
inferentially private mechanism that can achieve exact average consensus.
Finally, we give a brief introduction on cryptography based privacy
(subsection 5.3) before providing additional references for further reading
(subsection 5.4).
Definition 5.1. Two data sets d and d0 are called δ-adjacent if they
differ at no more than one entry and the difference is no greater than δ.
kd − d0 k0 ≤ 1, kd − d0 k1 ≤ δ.
212 Privacy
It is worth noticing that other metrics for data privacy exist. For
example, -identifiability ensures that for all neighboring d and d0 ,
Pr(d|u)
≤ exp().
Pr(d0 |u)
-mutual information privacy guarantees that the mutual information
between d and u is less than . However, both identifiability and mutual
information privacy require the prior distribution of the data base, while
differential privacy does not. For more detailed discussions on other
privacy metrics, please refer to Sun and Tay (2017).
As an example of differential privacy, let us consider a data base
d ∈ Rn with n entries, such that entry i represents the monthly power
consumption of household i. We would like to publish the average power
consumption q(d) = 1T d/n.
We next define our privacy mechanism. To this end, we say a
random variable w follows a Laplacian distribution with parameter b if
its probability density function satisfies:
1 |w|
f (w) = exp − .
2b b
We will write w ∼ Lap(b) for short.
In order to ensure -differential privacy, we can choose our mechanism
to be
p(u|d)
1T d 1T d0 u
n n
Figure 5.1: The probability density function of the output of the Laplacian mecha-
nism M (d) defined in (5.1).
f −1 (R) = {x : f (x) ∈ R} .
the set of edges. (i, j) ∈ E if and only if agent i and j can communicate
directly with each other. The neighborhood of agent i is defined as
Suppose that each agent has an initial scalar state x0,i . At each itera-
tion, agent i will communicate with its neighbors and update its state
according to the following equation:
X
xk+1,i = aii xk,i + aij xk,j . (5.2)
j∈N (i)
x+
k,i = xk,i + wk,i . (5.4)
Define
wk , [wk,1 , . . . , wk,n ]T ∈ Rn , x+ + + T n
k , [xk,1 , . . . , xk,n ] ∈ R . (5.6)
We can write (5.4) and (5.5) in matrix form as
xk+1 = Ax+
k = A(xk + wk ). (5.7)
Let us define the adjacency relationship on the initial conditions:
Definition 5.2. x0 and x00 are δ-adjacent if only one entry of x00 is
different from x0 , and the difference is no greater than δ.
Let us denote the infinite sequence x+ as (x+ + +
0 , x1 , . . .). Clearly x is
a function of both the noise sequence w = (w0 , w1 , . . .) and the initial
condition x0 . Hence, we could write it as x+ (x0 , w). To ensure the
-differential privacy, one would need to design the noise sequence w,
such that
Pr(x+ (x0 , w) ∈ R) ≤ exp()Pr(x+ (x00 , w) ∈ R), (5.8)
for any measurable set R and any δ-adjacent initial condition x0 and
x00 .
5.2. Differential and Inference Privacy in Average Consensus 217
Notice that this adding one-shot noise scheme is optimal in the sense
that it provides the best accuracy within a certain privacy requirement.
For more details, please refer to Nozari et al. (2017).
We further provide an impossibility result on differentially private
average consensus:
Theorem 5.2. Suppose that the consensus algorithm is -differential
private, i.e., (5.8) holds, then xk cannot converge to x̄ in probability for
any initial condition x0 .
where ρ is defined as
!1/k
E kzk k22
ρ , lim sup 2 , (5.11)
k→∞ z0 6=0 kz0 k2
with zk = xk − x̄.
Proof. Since the RHS of (5.10) is strictly less than 1, we only need to
prove (5.10), since it implies mean square convergence. By (5.7),
k−1
X
x k = Ak x 0 + Ak−t wt
t=0
k−2
X
= Ak x0 + Aϕk−1 vk−1 + ϕt Ak−t−1 (A − I)vt .
t=0
Define matrix A to be
A , A − 11T /n.
Ak (A − I) = Ak (A − I), (5.12)
Ak − 11T /n = Ak (I − 11T /n). (5.13)
220 Privacy
Since {vk } are IID Gaussian vectors with zero mean and covariance
σ 2 I, the mean square error can be written as
h i
E zkT zk = z0T A2k z0 + σ 2 tr(A2 )ϕ2k−2 (5.14)
k−2
X h i
+ σ2 ϕ2t tr A2k−2t−2 (A − I)2 ) .
t=0
The last inequality is true due to the fact that for all t,
ϕ2t λi2k−2t−2 (λi − 1)2 ≤ [max(ϕ, |λ2 |, |λn |)]2k−2 (λn − 1)2 .
where
Yk+ = VYk V. (5.22)
Proof. Notice that since the noise is Gaussian, the maximum likelihood
estimator of x0 given Ik is the unbiased estimator with the smallest
covariance, which is denoted as Pk .
It is easy to see that since agent n receives more information as k
increases, Pk is monotonically non-increasing. We then need to establish
that Pk converges to P , which is quite technical. Readers can refer to
Mo and Murray (2017) for the full details of the proof.
For the detailed proof of the theorem, please refer to Mo and Murray
(2017). Notice that the condition is local in the sense that the privacy
of the initial condition of agent i can only be breached by its neighbors.
If we compare the differential privacy based approach and inference
privacy based approach, we see that differential privacy provides much
stronger privacy guarantees, due to the following reasons:
5.2. Differential and Inference Privacy in Average Consensus 223
1. Inference privacy assumes that the agent can only access the
messages from itself and its neighbors y(k), while differential
privacy does not.
2. To prevent a privacy breach, inference privacy requires that no
super-neighbors exists. However, this is not required by differential
privacy.
However, inference privacy mechanisms can achieve exact average
consensus, which is its main advantage over differential privacy mecha-
nisms.
It is also worth noticing that other inference privacy metrics exist,
e.g., -information privacy, -average information leakage. For a more
detailed discussion, please refer to Sun and Tay (2017).
Example 5.1. We consider the following network consisting of 5 agents,
whose topology is illustrated in Figure 5.2. We assume that the following
A matrix is used:
2 1 0 0 1
1 2 1 0 0
1
A= 0 1 2 0 1 .
4
0 0 0 3 1
1 0 1 1 1
We first implement the privacy preserving consensus protocol proposed
1 2
4 5 3
where b(k) = ϕk , with ϕ = 0.9. From Figure 5.3, it can be seen that
although consensus is achieved, the final result is not the original average,
which may not be desirable for certain applications.
x1 (k)
x2 (k)
x3 (k)
0 x4 (k)
xi (k)
x5 (k)
−2
0 10 20 30 40 50
k
Figure 5.3: The trajectory of each state xi (k) when using the privacy preserving
consensus protocol proposed by Huang et al. (2012). The blue, red, green, yellow and
purple lines correspond to x1 (k), x2 (k), x3 (k), x4 (k), x5 (k) respectively. The black
dashed line corresponds to the average value of the initial x(0).
Figure 5.4 illustrates the trajectory of xi (k), when the noise follows
(5.9). It is worth noticing that all xi (k)’s converge to the true average
of the initial condition x(0).
However, the first algorithm, although it does not converge to the
exact average, can potentially provide more privacy guarantees. For the
example discussed, the first algorithm can preserve the privacy of agent
4. On the other hand, we prove in Theorem 5.5 that the initial condition
of agent 4 will be leaked to agent 5 using the second algorithm, as 5 is a
super-neighbor of 4. Therefore, there is a trade-off between privacy and
the accuracy of the consensus. To verify that, Figure 5.5 shows Pii (k)
of the maximum likelihood estimate of agent 4 and the asymptotic Pii
derived by Theorem 5.4, when using the second algorithm. P33 (k) is
omitted since it equals P11 (k) due to symmetry. Notice that both P11
and P22 are greater than 0. As a result, agent 5 cannot infer the exact
initial condition of agent 1 or agent 2. On the other hand, P44 = 0.
Therefore, the initial condition of agent 4 is not private to agent 5.
5.3. Cryptography Based Privacy 225
2
x1 (k)
x2 (k)
1 x3 (k)
x4 (k)
xi (k)
x5 (k)
0
−1
0 10 20 30 40 50
k
Figure 5.4: The trajectory of each state xi (k). The blue, red, green, yellow and
purple lines correspond to x1 (k), x2 (k), x3 (k), x4 (k), x5 (k) respectively. The black
dashed line corresponds to the average value of the initial x(0).
15 P11 (k)
P11
P22 (k)
0
5 10 15 20
k
Figure 5.5: Pii (k) v.s. k. The blue solid and dashed line correspond to P11 (k) and P11
respectively. The red solid and dashed line correspond to P22 (k) and P22 respectively.
The black solid and dashed line correspond to P44 (k) and P44 respectively.
comparison of inference privacy and data privacy in Sun and Tay (2017)
and Pin Calmon and Fawaz (2012). Liao et al. (2018) considers applying
inference privacy to hypothesis testing, which can also be interesting
for readers.
Finally we consider cryptography based approach to achieve pri-
vacy. At the time of writing, there are few results on applying fully
homomorphic encryption to cyber-physical systems, due to its com-
putational complexity. However, for certain applications, partially ho-
momorphic encryption has been sufficient. It has been shown in the
literature that additively homomorphic encryption can be used for
privacy preserving distributed optimization (Shoukry et al., 2016) or
average consensus (Ruan et al., 2019).
6
Conclusions
229
230 Conclusions
232
References 233
Amin, S., A. A. Cárdenas, and S. Sastry. 2009. “Safe and Secure Net-
worked Control Systems under Denial-of-Service Attacks”. In: Inter-
national Workshop on Hybrid Systems: Computation and Control.
Vol. 5469. Springer. 31–45.
Amin, S., X. Litrico, S. Sastry, and A. M. Bayen. 2013. “Cyber security
of water SCADA systems – Part I: Analysis and experimentation of
stealthy deception attacks”. IEEE Transactions on Control Systems
Technology. 21(5): 1963–1970.
Amoozadeh, M., A. Raghuramu, C.-N. Chuah, D. Ghosal, H. M. Zhang,
J. Rowe, and K. Levitt. 2015. “Security vulnerabilities of connected
vehicle streams and their impact on cooperative driving”. IEEE
Communications Magazine. 53(6): 126–132.
Anta, A. and P. Tabuada. 2010. “To sample or not to sample: Self-
triggered control for nonlinear systems”. IEEE Transactions on
Automatic Control. 55(9): 2030–2042.
Arulampalam, M. S., S. Maskell, N. Gordon, and T. Clapp. 2002. “A
Tutorial on Particle Filters for Online Nonlinear/Non-Gaussian
Bayesian Tracking”. IEEE Transactions on Signal Processing. 50(2):
174–188.
Bai, C. Z., V. Gupta, and F. Pasqualetti. 2017a. “On Kalman Filtering
with Compromised Sensors: Attack Stealthiness and Performance
Bounds”. IEEE Transactions on Automatic Control. 62(12): 6641–
6648.
Bai, C. Z., F. Pasqualetti, and V. Gupta. 2015. “Security in stochastic
control systems: Fundamental limitations and performance bounds”.
In: American Control Conference. IEEE. 195–200.
Bai, C.-Z., F. Pasqualetti, and V. Gupta. 2017b. “Data injection at-
tacks in stochastic control systems: Detectability and performance
tradeoffs”. Automatica. 82: 251–260.
Bi, S. and Y. J. Zhang. 2014. “Graphical methods for defense against
false-data injection attacks on power system state estimation”. IEEE
Transactions on Smart Grid. 5(3): 1216–1227.
Bobba, R. B., K. M. Rogers, Q. Wang, H. Khurana, K. Nahrstedt,
and T. J. Overbye. 2010. “Detecting false data injection attacks
on DC state estimation”. In: Workshop on Secure Control Systems,
CPSWEEK.
234 References
Han, D., Y. Mo, and L. Xie. 2019. “Convex optimization based state
estimation against sparse integrity attacks”. IEEE Transactions on
Automatic Control. 64(6): 2383–2395.
Han, D., J. Wu, H. Zhang, and L. Shi. 2017a. “Optimal sensor scheduling
for multiple linear dynamical systems”. Automatica. 75(Jan.): 260–
270.
Han, S., U. Topcu, and G. J. Pappas. 2017b. “Differentially Private
Distributed Constrained Optimization”. IEEE Transactions on Au-
tomatic Control. 62(1): 50–64.
He, L., D. Han, X. Wang, and L. Shi. 2013. “Optimal linear state
estimation over a packet-dropping network using linear temporal
coding”. Automatica. 49(4): 1075–1082.
Heemels, W. H., M. Donkers, and A. R. Teel. 2013. “Periodic event-
triggered control for linear systems”. IEEE Transactions on Auto-
matic Control. 58(4): 847–861.
Heemels, W., K. H. Johansson, and P. Tabuada. 2012. “An introduction
to event-triggered and self-triggered control”. In: Conference on
Decision and Control. IEEE. 3270–3285.
Henzinger, T., B. Horowitz, and C. Kirsch. 2001. “Giotto: A time-
triggered language for embedded programming”. In: Embedded soft-
ware. Springer. 166–184.
Hespanhol, P., M. Porter, R. Vasudevan, and A. Aswani. 2017. “Dynamic
Watermarking for General LTI Systems”. In: Conference on Decision
and Control. IEEE. 1834–1839.
Ho, J. S., A. J. Yeh, E. Neofytou, S. Kim, Y. Tanabe, B. Patlolla,
R. E. Beygui, and A. S. Y. Poon. 2014. “Wireless power transfer to
deep-tissue microimplants”. Proceedings of the National Academy
of Sciences. 111(22): 7974–7979.
Hoehn, A. and P. Zhang. 2016a. “Detection of covert attacks and zero
dynamics attacks in cyber-physical systems”. In: American Control
Conference. IEEE. 302–307.
Hoehn, A. and P. Zhang. 2016b. “Detection of replay attacks in cyber-
physical systems”. In: American Control Conference. IEEE. 290–
295.
References 239
Lee, C., H. Shim, and Y. Eun. 2015. “Secure and robust state estimation
under sensor attacks, measurement noises, and process disturbances:
Observer-based combinatorial approach”. In: European Control Con-
ference. IEEE. 1872–1877.
Lee, E. A. 2006. “Cyber-physical systems-are computing foundations
adequate”. In: NSF Workshop On Cyber-Physical Systems: Research
Motivation, Techniques and Roadmap. Vol. 2.
Lee, E. A. 2008. “Cyber physical systems: Design challenges”. In: In-
ternational Symposium on Object Oriented Real-Time Distributed
Computing. IEEE. 363–369.
Lee, I. and O. Sokolsky. 2010. “Medical Cyber Physical Systems”. In:
Design Automation Conference. ACM/IEEE. 743–748.
Liao, J., L. Sankar, V. Y. F. Tan, and F. du Pin Calmon. 2018. “Hy-
pothesis Testing Under Mutual Information Privacy Constraints
in the High Privacy Regime”. IEEE Transactions on Information
Forensics and Security. 13(4): 1058–1071.
Liu, L., M. Esmalifalak, Q. Ding, V. A. Emesih, and Z. Han. 2014.
“Detecting false data injection attacks on power grid by sparse
optimization”. IEEE Transactions on Smart Grid. 5(2): 612–621.
Liu, X., S. Weerakkody, and B. Sinopoli. 2016. “Sensor placement for
reliable observability: a structured systems approach”. In: Conference
on Decision and Control. IEEE. 5414–5421.
Liu, X., Y. Mo, and E. Garone. 2017. “Secure Dynamic State Estimation
by Decomposing Kalman Filter”. IFAC-PapersOnLine. 50(1): 7351–
7356.
Liu, Y., P. Ning, and M. K. Reiter. 2011. “False data injection attacks
against state estimation in electric power grids”. ACM Transactions
on Information and System Security. 14(1): 13.
Ljung, L. 1998. System Identification: Theory for the User. Pearson
Education.
Ljung, L., H. Hjalmarsson, and H. Ohlsson. 2011. “Four encounters
with system identification”. European Journal of Control. 17(5-6):
449–471.
Mazo, M. and P. Tabuada. 2009. “Input-to-state stability of self-triggered
control systems”. In: Conference on Decision and Control. IEEE.
928–933.
242 References
Wang, W., L. Ying, and J. Zhang. 2016. “On the Relation Between Iden-
tifiability, Differential Privacy, and Mutual-Information Privacy”.
IEEE Transactions on Information Theory. 62(9): 5018–5029.
Weerakkody, S., X. Liu, and B. Sinopoli. 2017a. “Robust Structural
Analysis and Design of Distributed Control Systems to Prevent Zero
Dynamics Attacks”. In: Conference on Decision and Control. IEEE.
Weerakkody, S., X. Liu, S. H. Son, and B. Sinopoli. 2016a. “A graph
theoretic characterization of perfect attackability and detection in
Distributed Control Systems”. In: American Control Conference.
IEEE. 1171–1178.
Weerakkody, S., X. Liu, S. H. Son, and B. Sinopoli. 2017b. “A Graph
Theoretic Characterization of Perfect Attackability for Secure Design
of Distributed Control Systems”. IEEE Transactions on Control of
Network Systems. 4(1): 60–70.
Weerakkody, S., Y. Mo, and B. Sinopoli. 2014. “Detecting Integrity
Attacks on Control Systems using Robust Physical Watermarking”.
In: Conference on Decision and Control. IEEE. 3757–3764.
Weerakkody, S., Y. Mo, B. Sinopoli, D. Han, and L. Shi. 2016b. “Multi-
Sensor Scheduling for State Estimation With Event-Based, Stochas-
tic Triggers”. IEEE Transactions on Automatic Control. 61(9): 2695–
2701.
Weerakkody, S., O. Ozel, P. Griffioen, and B. Sinopoli. 2017c. “Active
detection for exposing intelligent attacks in control systems”. In:
Conference on Control Technology and Applications. IEEE. 1306–
1312.
Weerakkody, S., O. Ozel, and B. Sinopoli. 2017d. “A Bernoulli-Gaussian
watermark design for detecting integrity attacks in control systems”.
In: Allerton Conference on Communication, Control and Computing.
IEEE.
Weerakkody, S. and B. Sinopoli. 2015. “Detecting integrity attacks on
control systems using a moving target approach”. In: Conference on
Decision and Control. IEEE. 5820–5826.
Weerakkody, S. and B. Sinopoli. 2016. “A moving target approach
for identifying malicious sensors in control systems”. In: Allerton
Conference on Communication, Control, and Computing. IEEE.
1149–1156.
References 251