AI - Recent Trends and Applications
AI - Recent Trends and Applications
Edited by
S. Kanimozhi Suguna, M. Dhivya, and Sara Paiva
First edition published 2021
by CRC Press
6000 Broken Sound Parkway NW, Suite 300, Boca Raton, FL 33487-2742
The right of S. Kanimozhi Suguna, M. Dhivya, and Sara Paiva to be identified as the authors of the editorial
material, and of the authors for their individual chapters, has been asserted in accordance with sections 77 and
78 of the Copyright, Designs and Patents Act 1988.
Reasonable efforts have been made to publish reliable data and information, but the author and publisher cannot
assume responsibility for the validity of all materials or the consequences of their use. The authors and publish-
ers have attempted to trace the copyright holders of all material reproduced in this publication and apologize to
copyright holders if permission to publish in this form has not been obtained. If any copyright material has not
been acknowledged please write and let us know so we may rectify in any future reprint.
Except as permitted under U.S. Copyright Law, no part of this book may be reprinted, reproduced, transmit-
ted, or utilized in any form by any electronic, mechanical, or other means, now known or hereafter invented,
including photocopying, microfilming, and recording, or in any information storage or retrieval system, without
written permission from the publishers.
For permission to photocopy or use material electronically from this work, access www.copyright.com or con-
tact the Copyright Clearance Center, Inc. (CCC), 222 Rosewood Drive, Danvers, MA 01923, 978-750-8400. For
works that are not available on CCC please contact mpkbookspermissions@tandf.co.uk
Trademark notice: Product or corporate names may be trademarks or registered trademarks and are used only
for identification and explanation without intent to infringe.
Typeset in Times
by Deanta Global Publishing Services, Chennai, India
Contents
Preface......................................................................................................................vii
Acknowledgements....................................................................................................ix
Editor biographies......................................................................................................xi
List of Contributors................................................................................................. xiii
v
vi Contents
vii
Acknowledgements
We would like to recognize CRC Press for providing us with the opportunity and
the professionalism shown by teams along the process. Namely, we would like to
thank Ms. Erin Harris and Keith Emmanual Arnold for all the support and leadership
during the course of publishing this book.
A word of appreciation and recognition to all the authors who have contributed
to this book. Also, a special recognition and our gratitude to all the reviewers who
made an extraordinary effort to enhance the quality of the final chapters of this
book.
Kanimozhi Suguna S has a great privilege to thank the Vice Chancellor,
Dean – School of Computing and other fellow colleagues of SASTRA Deemed
University, Thanjavur, for their assistance in and encouragement in throughout the
progress of the publication. She would like to extend her gratitude to her husband,
Shyam Sundar M, son Dharun Krishna S, and parents Gunavathi K and Subramanian
G. She would also like to thank her brother Arun Prasad S and all his family mem-
bers for their continuous support.
Dhivya wishes to thank Dr. Mohan Manghnani, Chairman, New Horizon Edu
cational Institutions and Dr. Manjunatha, Principal, New Horizon College of
Engineering, for their whole-hearted cooperation and great encouragement in all
endeavours. She also wishes to thank her parents for their continuous support and
motivation.
With great respect and honour, Sara Paiva would like to thank the Instituto
Politécnico de Viana do Castelo, Portugal and all her colleagues. Special thanks
go to her husband Rogério Paiva, children Diana and Leonardo Paiva, and also her
parents for all their support.
Obviously, a big word of appreciation goes to our families, our main source of
inspiration and strength, namely, in these difficult times we have lived in the last
year. Dedicated to The Almighty – The Supreme Power That Drives Us.
ix
Editor biographies
S. Kanimozhi Suguna received her B.Sc., M.C.A., M.Phil., and Ph.D., in 2005,
2008, 2009, and 2015, respectively. She has seven years of teaching experience and
is working as Assistant Professor in the Department of CSE, School of Computing,
SASTRA University, Thanjavur, India since 2017. Her research interests include
image processing, data mining, web mining, and wireless sensor networks. She
is an editorial member in IGI: Global and Editorial Advisory Board in various
International Conferences. She is a member of IEEE and CSI. She has published 4
book chapters, 12 international journals, and 5 conference presentations.
M. Dhivya received her B.E., M.E., and Ph.D. degree in Electrical Engineering in
2006, 2008, and 2013, respectively. At present, she is working with New Horizon
College of Engineering, Bangalore. She has worked as Associate Professor in the
Department of Electronics and Communication Engineering at Dr. N.G.P. Institute
of Technology. She has received a grant of Rs. 4,00,000 from Texas Instruments
for organizing four-day faculty empowerment programme on “Analog, Power,
Embedded Systems and Wireless (IoT).” She has guided 30 postgraduate scholars
in several disciplines: Applied Electronics, Electric Drives and Embedded Control,
Power Electronics and Drives and Control, and Instrumentation. Her research inter-
ests are wireless sensor networks, embedded systems, real-time systems, optimiza-
tion techniques, and machine learning.
xi
Contributors
Alper Yılmaz Indu Manimaran
Department of Electrical and Electronics Deaprtment of Computer Science
Bursa Technical University California State University
Turkey Long Beach, California
xiii
xiv Contributors
CONTENTS
1.1 Literature Review..............................................................................................3
1.1.1 Simulation Methodologies Applied in Wireless Communication
Systems (WCS)......................................................................................3
1.1.1.1 Simulation of WCS................................................................. 3
1.1.1.2 Discrete Event Simulation....................................................... 4
1.1.1.3 Event Scheduling....................................................................7
1.1.2 Channel Assignment in WCS.............................................................. 10
1.1.3 Multi-Agent Systems in WCS.............................................................. 12
1
2 Artificial Intelligence (AI)
1.1 LITERATURE REVIEW
1.1.1 Simulation Methodologies Applied in Wireless
Communication Systems (WCS)
1.1.1.1 Simulation of WCS
A real wireless network’s efficiency and behaviour can be tested using simulation
systems without the need for field experiments and prototype creation. The simula-
tion solutions give the opportunity to grow to a desired wireless network channel
allocation schemes, network architectures, etc. The simulation software development
approach becomes a very critical issue influencing the resulting network model and
efficiency, due to the complexity of real wireless networks. A big challenge for wire-
less network simulation is the discovery of a way to tackle the actual actions of the
network and not just speed up execution time using parallel machines. The simula-
tion model and environment structure affect the performance of simulated wireless
networks, and for this reason the design and development of such systems is studied
thoroughly. Modern simulation tools provide network engineers with the opportu-
nity to develop and test wireless communication systems at low cost very quickly.
There are three major simulation techniques (Chaturvedi, A., et al. 2001): discrete
event simulation (DES), system dynamics, and multi-agents. The most widely known
simulation tools are based on the DES concept and use various model architectures
to implement. A more accurate and reliable simulation environment can be developed
with the help of efficient model architectures (Chaturvedi, A., et al. 2001; Liu, W.,
et al. 1996; Zeng, X., et al. 1998; Bajaj, L., et al. 1999; Kelly, O.E., et al. 2000; Liu, et al.
2001; Boukerche, A., et al. 2001; Bononi, L. and D’Angelo, G., 2003) to speed up
execution times. Rapid development of parallel systems with multiple processors has
contributed to more effective simulation execution. On the other hand, the program-
ming technology, and especially the multithreading systems, provides an alternative
approach to simulation model implementation. Inside the simulation model, the actual
network activities must be realistically modelled. In the case of cellular networks,
such operations are the essential network functions such as new call entry, realloca-
tion (handoff) (Krishna, S., et al. 2016), device movement, and call termination. The
simulation model will include sub-models for three types of basic components:
• Facilities by network
• Network operating parameters (e.g. number of cells, base station locations,
allocation channel schemes, etc.)
4 Artificial Intelligence (AI)
The sequential simulation is a very time-consuming method due to the high complex-
ity of the models used (Liu, W., et al. 1996) and thus the parallelization is a critical
issue particularly for wireless networks of large scale (Liu, W., et al. 1996; Zeng, X.,
et al. 1998; Bajaj, L., et al. 1999; Kelly, O.E., et al. 1999; Liu, et al. 2001; Boukerche,
A., et al. 2001; Bononi, L. and D’Angelo, G. 2003). The key goal of parallel simula-
tion is the minimization of the execution time (Zeng, X., et al. 1998; Bajaj, L., et al.
1999). In the process of parallelization, sets of network entities have to be mapped to
multiple processors to achieve load balance (Zeng, X., et al. 1998). Multiple-processor
load distribution is a key issue (Liu, W., et al. 1996; Zeng, X., et al. 1998; Bajaj, L., et
al. 1999). Several features, procedures, and entities, such as geographical area (Kelly,
O.E., et al. 1999; Boukerche, A., et al. 2001), radio channels (Kelly, O.E., et al. 1999;
Boukerche, A., et al. 2001), interference calculations (Kelly, O.E., et al. 1999), mobile
hosts (MH) (Boukerche, A., et al. 2001), and network cells (Boukerche, A., et al.
2001) can be parallelized in parallel implementations. There are important problems
that need to be tackled effectively in order to achieve optimum parallelization, such as
synchronization of processors (Zeng, X., et al. 1998; Bajaj, L., et al. 1999; Boukerche,
A., et al. 2001), load balancing control (Liu, W., et al. 1996; Zeng, X., et al. 1998), and
cycle mapping to processors (Liu, W., et al. 1996; Zeng, X., et al. 1998). The above-
mentioned studies have resulted in two basic points for further investigation:
Sequential DES systems are the most common in science community. In such systems,
in a three-step cycle, the scheduling mechanism can be analyzed:
ns-2, ns-3 (Fall, K., and Varadhan, K., 2015), widely accepted by the scientific com-
munity (Kurkowski, S., et al 2005), use this scheduling mechanism for event execu-
tion. In ns-2, ns-3 (Fall, K., and Varadhan, K., 2015), the scheduler selects the next
earliest incident, executes it until ending it up, and returns to run the next event of
highest priority or with least time stamp among the outstanding events. In ns-2, ns-3
(Fall, K., and Varadhan, K., 2015), only one operation can be executed at any given
time, and so it is a single-threaded simulator. If two or more events are scheduled to
take place (to execute) at the “same time,” their execution is performed in a first come
first scheduled scheme – first dispatched, based on the time stamp of each incidence
and thus, the simulation model’s adaptability to the actual network behaviour, includ-
ing all relevant network physical activities, is strongly based on the scheduling algo-
rithm. On the other hand, parallel implementation of DES systems was introduced in
the literature primarily to achieve speed up of execution. Due to the fact that the main
goal is load distribution and speed-up execution, the parallel DES systems do not
modify the principle of the event scheduling found in sequential DES. The system per-
formance is thus a critical point in terms of execution time (Liu, W., et al 1996; Zeng,
X., et al 1998; Bajaj, L., et al 1999; Kelly, O.E, et al 1999; Liu, et al. 2001; Boukerche,
A., et al 2001; Bononi, L. and D’Angelo, G., 2003). Parallel efficiency plays a major
role in the performance of the entire system (Liu, W., et al. 1996; Zeng, X., et al. 1998;
Kelly, O.E, et al. 1999; Boukerche, A., et al. 2001) and therefore synchronization of
multiple processors is important (Liu, W., et al. 1996; Zeng, X., et al. 1998).
Physical time is the time in the physical (real) system; simulation time is the mod-
elled physical time in the simulation system; and wall-clock time is the simulation
program’s runtime.
The definition of simulation time is given in Fujimoto (2000) as follows:
«Simulation time is defined as a totally ordered set of values in which each value
represents an instant of time in the modelled physical system. Furthermore,
for any two simulation time values T1 representing physical time P1, and
6 Artificial Intelligence (AI)
T2 representing P2, if T1 < T2, then P1 occurs before P2, and (T2 − T1) is
equal to (P2 −P1) K for some constant K. If T1 < T2 then T1 is said to occur
before T2 and if T1>T2, then, obviously T1 happened after T2.»
according to Kelly et al. (1999), and significant speedup is achieved using up to eight
processors. For the effective simulation of wireless and mobile networks, a cluster
of workstations is used (Boukerche, A., et al. 2001). Quite impressive speedup is
accomplished using up to 16 processors based on Mobile Host partitioning. For par-
allel DES, a networked cluster of PCs use a set of physical execution units (Bononi,
L. and D’Angelo, G., 2003).
1.1.1.3 Event Scheduling
1.1.1.3.1 The Importance of Event Scheduling
Events, being the main entities of a DES system, express the physical activities of
a real-world wireless network. With respect to the aforementioned network facili-
ties, each service could be considered an incidence for a particular mobile user
(MU). During simulation time, an incidence generator creates events as, for instance,
new call arrivals in the system. Such an event is marked with a time stamp (within
simulation time) that specifies the starting execution point in time. A critical entity
within the simulation system, namely, the scheduler, is executing by selecting the
next earliest event, based on its associated time stamp, running it until ending it up,
and returning to execute the next event expressed by the next least time stamp. The
scheduling system is a standard tool within the actual wireless network for the event
service incidence sequence control. When an operation is not executed, the pending
operation set (PES) (Goh, R.S.M. and Thng, I., L-J, 2003; Tan, K.L., and Thng, L.-J.,
2000; Siangsukone, T., et al. 2003) is the set of all non-simulated and processed
tasks generated during simulation time. Consequently, PES corresponds to a priority
queue controlling the stream of events simulation based on the contemporary mini-
mum time stamp, that is the task of highest priority (Goh, R.S.M. and Thng, I., L-J,
2003). The selected scheduling method imposes how the actual network activities
that occurred can be reflected in the simulation model in a realistic way. Therefore,
scheduling could be viewed as a generalized mapping method within the simulation
time of the DES algorithm for the real network events-activities execution.
from the common desktop calendar of one-page entry every day. Each incident is
scheduled on the relevant calendar page. Each event’s schedule timing defines its
priority. If an event on the calendar is enqueued, then this event is scheduled for
future running. The earliest event within the calendar is dequeued by searching it
within the relevant page of the current date and removing it from that page (Brown,
R. 1988).
A CQ scheme is considered to consist of an array of lists. Each such list queue
contains future incidences. According to the CQ principle, a large queue of N events
is partitioned to M shorter queues called Buckets. Each bucket is associated with a
specific time range corresponding to future events executions. Any incidence bound
to an occurrence time t(e) is associated with the mth bucket in year y (y = 0, 1, 2, …)
if and only if
(
ë )
t (e) Î é ( yM + m ) d , ( yM + m + 1) d ù (1.1)
û
In order to find the associated relevant bucket number m(e) where an incidence e will
occur at time t(e), the following type is involved:
ê t ( e) ú
m (e) = ê ú mod M (1.2)
ë d û
Consider that N = 10, δ = 1, M = 8, and t(e) = 3.52 (Figure 1.1) for a new incidence e.
Using Equation 1.2, the relevant bucket number for the incidence e is m(e) = 3.
An analytical study of the CQ scheme can be found in Erickson et al. (1994).
FIGURE 1.2 Sample of a Sequential DES operation (events/facilities: NC – new call arrival,
RC – call reallocation/handoff, MC – movement of user running a call, FC – terminated call).
The generator of the incidences creates events with corresponding time signs. The
dequeued incidences are running either in rising priority order or in increasing time
stamp order (Figure 1.3). The final processes running is sequential oversimulation
time except for events of nearly identical time stamps.
FIGURE 1.3 Event running following the dequeuing order of the processing unit (events/
facilities: NC – new call arrival, RC – call reallocation/handoff, MC – movement of user run-
ning a call, FC – terminated call).
Many channel allocation schemes were widely invested in literature with a goal
of maximizing reuse of frequency. Channel allocation schemes are classified into
three strategies: fixed channel allocation (FCA) (Zhang, M. and Yum, T.S., 1989;
Lai, W.K. and Coghill, G.C., 1996; MacDonald, V.H., 1979; Elnoubi, S.M., et al
1982; Xu, Z. and Mirchandani, P.B., 1982); dynamic channel allocation (DCA)
(Zhang, M. and Yum, T.S., 1989; Cimini L.J. and Foschini, G.J., 1992; Cox, D.C.
and Reudink, D.O., 1973; Del Re, E, Fantacci, R, and Giambene, G. 1996; Kahwa,
TJ, and Geor-gans, ND. 1978; Papazoglou, PM, and Karras, DA. 2016; Papazoglou,
PM, Karras, DA, and Papademetriou, RC. 2008a; Papazoglou, PM, Kar-ras, DA,
and Papademetriou, RC. 2008b; Papazoglou, PM, Karras, DA, and Papademetriou,
RC. 2009; Papazoglou, PM, Karras, DA, and Papademetriou, RC. 2011; Papazoglou,
PM, Karras, DA, and Papademetriou, RC. 2016; Sivarajan, KN, McEliece, RJ, and
Ketchum, JW. 1990; Subramania Sharma, T, and Thazhuthaveetil, MJ. 2001) and
HCA (Hybrid Channel Allocation).
12 Artificial Intelligence (AI)
FIGURE 1.5 Excerpt of different CQs based on network spatial partitioning (events/facili-
ties: NC – new call arrival, RC – call reallocation/handoff, MC – movement of user running
a call, FC – terminated call).
networks is provided in Iraqi and Boutaba (2000). The call dropping likelihood is
small based on the proposed MAS (Iraqi, Y. and Boutaba, R., 2000), although the
wireless network makes a high average usage of the bandwidth. According to Iraqi
and Boutaba (2000), the final decision to accept a call is based on neighbouring cells
participating. Therefore, every cell or BS has an agent working. A cooperative agree-
ment in a MAS is defined in Jiang and Jiang (2008) to help real-time load balancing
of a mobile cellular network. Within the current MAS, agents are used (a) for market
representation of different service providers and (b) for network operators handling
the radio capital of different network regions. According to Jiang and Jiang (2008),
the coordination of agents is done by messages, and negotiations are focused on the
final agreements. Negotiation is an intelligent control strategy in the above study to
dynamically change the shape and size of the cells and to manage the traffic load
over the network.
To suggest a distributed channel allocation scheme using intelligent software
agents, a detailed simulation model for wireless cellular networks was developed
by Bodanese (2000). In the above analysis, smart collaborative software agents
offer BSs autonomy, enhance network robustness, enable network resource negotia-
tion, and improve resource allocation. To this end, several aspects of infrastructure
and operation of the cellular network have been modelled. According to Bodanese
(2000), the basic features implemented in the model are call set-up, handoffs based
on signal strength and defined threshold, forced call termination emulation (due
to insufficient signal), Mobile Station self-location based on signal strength from
nearest BSs, Mobile Stations with arbitrary trajectories, cell boundaries, number of
Mobile Stations per unit. Four specific aspects of the configuration and function of
the entire network were modelled in Bodanese (2000). Those aspects include (a) the
network model defining the network layout (cell structure, BSs, MSs), (b) the mobile
station (signal calculations, call generation inter-arrival time, call length), (b1) call
server (SNR measurements), (b2) handoff, (b3) call dropping, (b4) call termination,
(b5) filter (sending packets from the radio link to the server), (b6) trajectory track-
ing, which is responsible for Mobile Station movement (random cyclical, driving on
a highway, etc.), (c) BS which includes SNR measurements, handoff (monitoring the
received voice packets), etc., (d) MTSO (Mobile Telephone Switching Office) which
offers services such as HLR (Home Location Register), VLR (Visitor Location
Register), etc., (d1) keeps track of each active MS. All the implemented network
models in Bodanese (2000) do not support multimedia real-time services. A novel
modelling methodology for supporting wireless network services is also described
in the current report, based on MAS technology.
links or dedicated leased lines to the Base Station Controller (BSC). The BSC pro-
vides logic for the management of the base stations’ radio resources under its control.
As a MU moves from cell to cell, it is also responsible for transferring an ongoing
call from one base station to another. A mobile terminal must first obtain a channel
from the BS in order to establish a communication with a BS. A channel consists of
a pair of signal transmission frequencies between the BS and MU.
In order to pick every SL correctly, multiple variables need to be taken into account.
Those characteristics include:
Most approaches to simulation models and applications have allowed us using gen-
eral-purpose languages, general SLs, and packages for special-purpose simulations.
Languages of general use have the following principal characteristics:
16 Artificial Intelligence (AI)
• Bottom cost
• Are available
• No supplementary training is required
• Every model starts from the ground up
• Hard to check
• Reusability of code is minimal
• The business process period is a long one
General SLs deliver uniform modelling features, shorter development cycle, and
simple verification. These languages have higher costs, restricted portability, and
require more instruction.
Languages such as Simula (Nierstrasz, O., 1989), GPSS (Karian, Z.A., and
Dudewicz, E.J., 1998), SimScript (Russell, E.C., 1999), Parsec (Subramania, S.,
2001; Zeng, X., et al. 1998; Takai, M., 1999; Reddy, D., et al. 2006; Abar, S., et al.
2017), modsim (West, J. and Mullarney, A., 1998), slam (Miller, J.O., et al. 1988),
GlomoSim (Zeng, X., et al. 1998), and other special purposes of high-flexibility
simulation devices. Other software tools such as MATLAB have plenty of scientific
libraries and capabilities for the creation of portable script code but suffer from low-
performance execution. A large-scale simulation model includes a scalable network
environment with advanced features such as portability, internetworking, and high
network activity adaptability.
The invention and development of new languages for simulation introduces
important drawbacks:
• The new languages are specific to the domain and rarely adapted by the
scientific community.
• The respective libraries require designers to adapt their applications to spe-
cific needs.
• Designers can’t achieve high simulation kernel adaptability in their
applications.
Java language is the most appropriate to develop scalable, portable, and high-perfor-
mance network applications, and is adopted by the majority of the scientific com-
munity. On the other hand, Java supports multiple threads that mean manipulation of
the concurrent tasks. This functionality is very critical for supporting network events
at the same time.
The basic parameter for mesh building is the fineness, which means the distance
between the spots.
1.2.1.1 Operational Parameters
A set of parameters determine the conditions of network operation, such as signal
propagation, signal rates, and MU behaviour. The parameter basics are: CNR (car-
rier-to-noise ratio) on cell edge (dB), CNIR (carrier-to-noise plus interference ratio)
threshold (dB), average call arrival rate (calls/hour, lambda), average call hold time
(seconds), simulation time, cell channels, path loss factor (alpha), standard shadow-
ing deviation (sigma), cell mesh fineness, etc.
the new calls result from a random or a Poisson distribution with regard to a pre-
defined regular pattern.
1 - x2 / 2
f ( x) = e (1.4)
2p
In the MU movement procedure, firstly a Gaussian number is generated, e.g. x1 and
the corresponding f(x1) whenx1 Î éë -0.5, 0.5ùû . If another Gaussian number, e.g. x2,
where x2 ³ f ( x1 ) , a MU move is generated. The new position of any MU is maxi-
mum two cells distance from the initiated cell.
1.2.2.2 Channel Allocation
1.2.2.2.1 Channel Allocation Criteria
Any channel allocation procedure is successful only if the following criteria are met:
• Channel availability
• Carrier strength (between MU and BS)
• CNR (signal purity)
• Signal to noise plus interference ratio CNIR (interference from other con-
nected MUs)
Large-Scale Multi-Agent Systems Simulation 19
(a) Check if the maximum MU capacity in the cell neighbour has been reached.
(b) Calculate a random MU position in the mesh.
(c) Place the new MU according to the cell’s BS position and mesh spot.
(d) Calculate the signal strength between BS and new MU in the call-initiated
cell.
Firstly, the shadow attenuation (Molisch, A., 2012; Lee, W.C.Y., 1995) is derived.
Next
s ×n
sh = 10 10 (1.5)
where σ is the shadowing standard deviation and n is the associated number from the
normal distribution. The distance attenuation dw can be derived from the shadow
attenuation and distance between MU and BS. Through MU and BS, the CNR could
be computed (Molisch, A., 2012; Lee, W.C.Y., 1995):
cnedge
cn = 10 10 × dw (1.6)
(e) Calculate interference among the new MU and other MUs that use the same
channel.
(f) Check if the CNIR ratio is acceptable according to a predefined threshold.
(g) If CNIR is acceptable, establish the new call and update UR, otherwise use
any alternative selected DCA variation.
1.2.2.3 Traffic Generation
When a new call arrival occurs with regard to Poisson distribution, a random number
x is generated from Poisson distribution and the corresponding P(x) is calculated as
follows:
e-l × l x
P( x ) = (1.7)
x!
where λ represents the new call arrival rate. This rate is analogous to the time of the
day. A fixed set of simulation time steps represents a whole day. Each day is divided
into five zones according to traffic conditions. Table 1.1 illustrates the five zones with
corresponding λ parameter. Zone 1 represents the hours 12–9 a.m. (λ = 1) and zone
3 the hours 1–4 p.m. (λ = 5).
20 Artificial Intelligence (AI)
TABLE 1.1
Daily Zones
Zone Percentage
Zone (Approximately) λ Description
1 34% 1 Least busy zone
2 21% 4
3 12% 5 Most busy zone
4 17% 3
5 17% 2
The fundamental layer (layer 1) embodies the structure of the cellular network,
where the basic events (new call, call rearrangement, user movement, call termina-
tion) take place. Layer 2 is made up of four agents that implement the synchronized
events and describe the behaviour of the network. The control agent (main agent) that
exists in the third layer activates the system with corresponding event agents. The set
of four agents is recurring for every individual MU (Figure 1.8).
The main agent includes a clock logic which is responsible for the layer 2 agents’
activation and ensures that the additional procedures are triggered in the correct
order. The agents communicate for the activation status through layers 2 and 3 and
exchange information (messages) etc. Event agents are independent entities that react
with the appropriate environment and are coordinated by the control agent. The cel-
lular network environment includes parameters such as information about the signal
power, user characteristics, network status, etc. Every agent is implemented as a Java
thread and is defined in relation to layered architecture by state, actions, and position.
Agents in layer 2 are concurrently executed according to signals from the control
node. The simulation time changes only when the layer 2 agents have completed the
concurrent acts. Layer 1 uses shared data objects that are protected from the event
agents in order to simultaneously access (read or write). Each event agent keeps the
22 Artificial Intelligence (AI)
control agent informed about the status of its execution. The control agent is active,
while the time for simulation is not over and so the rest of the agents. This agent has
a sequential step-value clock (e.g. 1,2,3, etc.). A corresponding action is turned on
in each value. All the acts (agents and supplementary procedures) are carried out as
threads. The requisite supplementary activities (preparation tasks) are triggered in
the first phase of the clock, while the other agents and procedures are disabled. In
the second step of the clock, event agents are turned on while other procedures are
disabled. Control agent prevents simultaneous activation within simulation time of
the event agents, initial and final simulation procedures.
Moreover, let’s assume that the probable discrete states of the wireless network set-
ting up can be defined by the set E as
where the memberships of E signify the network performance, L signifies low level,
and H the high level. The pairs are linked to blocking and dropping probability,
respectively.
On the other hand, it is adopted that each agent has a set of possible interactions
with this set-up. These interactions change the environmental status and are deter-
mined by the set
Ac = {a k 0 , a k1,¼} (1.10)
In the case of network set-up, the matching actions are determined by the set
where INP is the action “Increase Priority,” DeP represents the action “Decrease
Priority,” and, finally, DN is the action “Do Nothing.”
The external network set-up changes its state with regard to the above-mentioned
activities. A sequence of actions results in a sequence of environmental state changes.
Thus, a run r of an agent within the environment can be defined as
a0 a1 a2 a3 a u-1
r : e0 ® e1 ® e2 ® e3 ® ® eu (1.12)
24 Artificial Intelligence (AI)
A state transformer function has been made known by Fagin et al. (1995) in order to
depict the effect of an agent activity on the network set-up:
t : R Ac ® g ( E ) (1.14)
t ( r ) = Æ (1.15)
Env = E, e0 ,t (1.16)
where Ε is the state set, eo is the initial state, and τ is the transformer function. For
modelling agents, it is adopted that an agent characterizes a function for mapping
runs to activities and thus:
Ag : R E ® Ac (1.17)
In other words, an agent decides about actions (to select an appropriate action to
perform) based on the past events of the system. For expressing the entire system
(agents, environment), a suitable set is developed:
Finally, the sequence (e 0 ,a 0 , e1,a1, e 2 ,… )expresses a run of an agent Ag (in the net-
work set-up) Env = E, e0 ,t , if:
a 0 = Ag ( e0 ) (1.19)
eu Î t ( e0 , a 0 ,..., a u -1 ) (1.20)
a u = Ag ( e0 , a 0 ,..., eu ) (1.21)
e1 = { x , y } , e2 = { x , y} , e3 = { x, y } , e4 = { x, y} (1.25)
Network behaviour is evaluated through two basic statistical metrics which are the
blocking and dropping likelihoods. Thus, (1.24) and (1.25) will be expressed in terms
of the above metrics as follows:
ì p1 if e = e1 or e = e2 ü
see(e) = í ý (1.28)
î p2 if e = e3 or e = e4 þ
According to (1.26) and (1.27), the “perceive function” of the NCA will have two
observations in its range P1 and P2 that indicate if the blocking possibility is accept-
able or not. The behaviour of the “perceive” function can be outlined as follows:
ì P1 if e = e1 or e = e2 bad ü
see(e) = í ý (1.29)
î P2 if e = e3 or e = e4 good þ
ì P1 if e = e1 or e = e3 bad ü
see(e) = í ý (1.30)
î P2 if e = e2 or e = e4 good þ
Large-Scale Multi-Agent Systems Simulation 27
With two given environmental states e Î E and e¢ Î E , then e~e’ can be written only
if see(e) = see(e’). An agent has perfect perception if the different environment states
are equal to distinct observations. In such case
~ = E (1.31)
In contrast with (1.15), when the perception of an agent does not exist,
~ = 1 (1.32)
function assigns a real number to each outcome, indicating how good the outcome
is for the selected agent. In other words, this utility function defines the preference
ordering over all the outcomes.
Let ω and ω’ be two possible outcomes in Ω set (where Ω = {ω1,ω2,…}), with
utility functions that give
ui (w ) ³ ui (w¢ ) (1.33)
1.2.4.6 Multi-Agent Encounters
In a large-scale environment with more than one agent, except the environment
itself, the agent encounters also have to be modelled. Assume that there are two pos-
sible actions by two agents such as “C”, which means “Cooperate”, and “D”, which
means “Defect” and let the action set Ac = {C,D}. Based on the above, the environ-
ment behaviour can be modelled with the function
t : Aci ´ Ac j ® W (1.34)
where Aci and Acj represent the two agents i and j, respectively.
As an example, let the environment function
t ( D, D ) = w1, t ( D, C ) = w2 , t ( C , D ) = w3 , t ( C , C ) = w4 (1.35)
The two possible actions for NCA and RCA are the Cooperation (CR) and Competition
(CT) and so the action set becomes Ac = {CP, CT}. Now the expression (1.35) is as follows:
ui ( D, D) = 1 ui ( D, C ) = 1 ui (C , D) = 4 ui (C , C ) = 4
(1.39)
u j ( D, D ) = 1 u j ( D, C ) = 4 u j (C , D) = 1 u j (C , C ) = 4
Large-Scale Multi-Agent Systems Simulation 29
Combining now (1.36) and (1.38) for NCA and RCA, the outcomes can be expressed
as follows:
uNC (CT , CT ) = 1 uNC (CT , CR) = 1 uNC (CR, CT ) = 4 uNC (CR, CR) = 4
(1.40)
uRC ( CT , CT ) = 1 uRC (CT , CR) = 4 uRC (CR, CT ) = 1 uRC (CR, CR) = 4
The only possible agent actions for the given outcomes can be rewritten (see (1.39)):
C , C ³i C , D >i D, C ³i D, D (1.41)
In the same way, the possible NCA, RCA actions based on (1.20a) are
It is obvious from (1.39), (1.40), and (1.41) which action will be selected from each
agent, and thus each agent knows exactly what to do.
analysis in Liu and Layland (1973) and deadline monotonic analysis in Audsley et
al. (1992). These two improvements were incorporated into the scheduling theory
of fixed priorities (Audsley, et al. 1995). Some very important characteristics of the
RTSs found in Buttazzo (2011) are (a) repetitive contact with the environment, (b)
multiple actions concurrently transitioning rapidly to events and requiring a high
degree of competition, (c) competition for shared resources, (d) actions caused exter-
nally by events (within the system) or after time development, (e) must be stable in
overload, and (f) must be maintainable and of scalable open architecture.
as ns-2/ns-3, are focused on the scheduling of the CQ form, which reflects the serial
logic. Within this conceptual framework the execution sequence of each generated
event is defined from the associated time stamps. CQ retains information about
activities for future executions. Such time stamps can also be used as goals. The
computational multi-processing approach can be implemented when the network
events are addressed as synchronized. In a simultaneous model, another MU is mov-
ing or attempting for reallocation (handoff) when a MU is under network processing.
Thus, the execution of the mission (event) has to be incomplete to accommodate the
waiting synchronized MUs. This reasoning leads to the inter-leaving scheme that
can normally be enforced using a multi-threading framework. Figure 1.16 demon-
strates the queuing models for the two approaches (consecutive and concurrent).
In a concurrent set-up, events can be viewed as multiple processes that must be
facilitated by the available computing unit (Figure 1.17).
and an analysis of schedulability (Fidge, C.J. 2002), reflect events behaviour that
could be analyzed as follows:
Any event e, with e Î {NC , RC , MC , FC}has a release time r:
re = ( n × Te ) - J e (1.43)
where n is the nth event, Te is the event period, and Je is the associated Jitter. Je is
computed as
C = (n + 1) × Ce (1.45)
Let Ri be the response time for process i, Ci be the worst-case calculation time, and Ii
be the interference, due to higher priority processes. The response time for process i is
Ri = Ci + I i (1.46)
Assuming that all priorities are unique, the interference for process i, is
é Ri ù
Ii = å êê T úú C (1.47)
jÎhp ( i ) j
j
where hp(i) is the set of all processes that have greater priority than i.
From Equations 1.46) and 1.47, the worst-case response time of process i can be
computed through the following formula:
é Ri ù
Ri = Ci + å êê T úúC (1.48)
jÎhp ( i ) j
j
Equation 1.48, can be solved iteratively (Audsley et al. 1993), commencing with the
initial estimate for Ri equal to 0. Knowing the xth value, the (x + 1) value can be
approached as follows:
é Rix ù
Rix +1 = Ci + å êê T
jÎhp ( i )
úC j (1.49)
j ú
The main and useful conclusion derived from Equation 1.49 is process i schedulabil-
ity, which means that the deadlines required can be met. For safe conclusions, the
above equation must be re-evaluated to see if it converges to an Ri value such that
Ri < Di (Di = the given process i time limit). Instead of the existence of specialized
Large-Scale Multi-Agent Systems Simulation 33
tools to evaluate the above equations, such tools do not necessarily give the pro-
grammer any insight into the dynamic conduct of each process (Fidge, C.J., 2002).
Simulation is the most effective method for demonstrating the timing actions of the
processes in consideration (Fidge, C.J., 2002).
The event goals must be set for conducting full calculations. Let PFC, PRC, PMC,
and PNC be the priorities for the event types FC, RC, MC, and NC with PFC > PRC,
PRC > PMC, PMC > PNC, respectively. The corresponding hp(j) sets involving the
definitions above can be defined as follows:
hp ( FC ) = {Æ} (1.50)
hp ( RC ) = {FC} (1.51)
Equation 1.49 can be solved iteratively (Audsley et al. 1993). Thus, the response time
for next event n + 1 can be calculated as follows:
é Ren ù
Ren +1 = C + å ê ú × C j (1.54)
T
jÎhp ( e ) ê j ú
Formulating the above equations to network procedures, the resulted response times
are as follows:
Assuming that Re0 = 0 ,
é RFC
n ù
n +1
RFC = CFC + å ê
jÎhp ( FC ) ê
Tj ú
ú × C j Þ RFC = CFC (1.55)
Since hp ( FC ) = {Æ}
n +1 é Rn ù
RRC = C RC + ê RC ú × CFC (1.56)
ê TFC ú
n +1 é Rn ù é Rn ù
RMC = C MC + ê MC ú × C RC + ê MC ú × CFC (1.57)
ê TRC ú ê TFC ú
n +1 é Rn ù é Rn ù é Rn ù
RNC = C NC + ê NC ú × C MC + ê NC ú × C RC + ê NC ú ×CFC (1.58)
ê TMC ú ê TRC ú ê TFC ú
Based on the above rules, a pre-emptive event scheduling can be applied.
34 Artificial Intelligence (AI)
All the above algorithms assume sequential search only within the neighbouring
cells between the current user and available channels.
The following proposed algorithm assigns channels inside neighbouring cells
simultaneously, provided that N agents/users emerge at a given time.
• Artificial intelligence-based variation controlled and good CNR DCA for
concurrent network assignment
Therefore, the chromosome comprises the possible channels (out of the available
ones) to be assigned to every agent/user. If there are no available channels for all N
agents/users we put zeros in the corresponding positions in the chromosome. An
exhaustive search of all possible combinations of C in total available channels in
æC ö
groups of N agents/users, ç ÷ , requires factorial time increase as it is known and
èNø
makes impossible the task in real time. Genetic algorithms in this case can provide
an affordable near optimal solution.
The fitness function used is comprised of the following parts:
all cells C ( j )
1. å åCNR (i, j ) should be maximized. This means that for all channels
j =1 i =1
C(j) of cell j assigned, let’s say, to C(j) agents/users out of N and for all cells
the total CNR should be maximized.
åå ( congestion (i ) - congestion ( j ))
2
2. should be minimized for all cell
i j
pairs (i,j).
3. Each available channel ch(i) assigned to agent i should be used only once.
å
4. ch ( i, j ) £ C ( j ) , that is, all channels of cell j assigned to agents/users i
i
1.2.7 Implementation Architectures
We have implemented and tested three different approaches of agent-oriented archi-
tectures using Java threads. Each approach is characterized by the parallelization
grade that influences the simulation efficiency and the reliability related to network
behaviour.
1.2.7.1 Conventional Model
The conventional model consists of one thread and the event processing is based on
the CQ mechanism (Figure 1.23). According to this model, any generated event is
served through a three-step procedure which includes enqueue, dequeue, and execu-
tion. Based on the type of the highest priority event, the corresponding network
service is activated.
1.2.7.2 Concurrent Models
As we mentioned before, in a real network, processes take place simultaneously. A
model that supports concurrent architectures approaches in a more reliable way the
real network. We have developed and tested two architectural models with different
levels of concurrency. These models rely on the mentioned layered agent architecture
using Java multi-threading techniques for the implementation.
1.2.7.2.1 Basic Concurrency
The concurrent version of the model must prevent simultaneous access to the user
registry where critical data for the MUs are stored. Figure 1.24 illustrates the
structure of the implemented architecture. Level 0 represents the control agent
that synchronizes the execution of the concurrent tasks (New call, reallocation,
etc.) in combination with the thread scheduler and administers the simulation
time. Concurrent tasks are activated within the level 1 and share the connection
establishment methods. The main network processes like new call arrival, real-
location check, user movement, etc. execute common tasks such as access to user
registry and connection establishment. Connection establishment and user reg-
istry access (read, write) are implemented with three mutual excluded methods
(level 2).
These methods belong to the same thread. The main agents act simultaneously
and try to finish for each MU in each cell. A new connection based on reallo-
cation, user movement, or new call arrival requires signal calculations and other
DCA operations. These actions are implemented using the try_connection method.
All the threads work simultaneously except the above method and access to user
registry that is used by one thread at a time preventing the data corruption of user
registry.
40 Artificial Intelligence (AI)
1.2.7.2.3 Thread Usage
The main implementation concept is based on thread logic. Each user/network
operation is implemented as a thread. The Java programs with regard to the above-
mentioned concurrent architectures consist of four basic threads (call termination,
reallocation, new call arrival, user move) and three additional (clock synchroniza-
tion, initial/final procedures). As the concurrency grade increases, the Java threads
become more competitive for accessing the user table (read/write tasks). Thus, the
possibility for deadlocks is rising. In order to face effectively the deadlock conditions,
synchronized blocks instead of synchronized methods can be used. Synchronized
blocks set locks for shorter periods than synchronized methods and so the deadlock
possibilities decrease. Table 1.2 shows the threads and synchronized methods/blocks
that are used in the proposed architectures.
Large-Scale Multi-Agent Systems Simulation 41
TABLE 1.2
Thread Usage by Architectural Model
Concurrency with Multiple
Concurrency Architecture Basic Concurrency Synchronized Blocks
No. of threads 7 7
Thread tasks Clock, initial loop procedures, final loop procedures, call termination,
reallocation check, new call arrival, user movement
No. of synchronized methods 3 0
Synchronized methods Read User Registry –
Write User Registry
Try_connection
No. of synchronized blocks 0 18 (Read/Write User registry)
In the first approach, the most of the access points for user registry exist within
the try_connection method that is synchronized. Due to that synchronization, the
competition for user registry access is low. Increasing concurrency, the try_connec-
tion is concurrent among the four main threads and so only the access methods
(read/write to user registry) are synchronized. With this approach, we achieve high
concurrency but the deadlocks seem to be a difficult situation. Using synchronized
blocks in contrast with synchronized methods, we faced successfully the deadlock
conditions. Extending the above models for avoiding deadlocks, conditional synchro-
nization can be applied. With conditional synchronization, the user registry access
is synchronized only when two or more threads try to access data for the same MU.
where k is the selected pair and i is the i-element of each set. The difference seta[i]
− setb[i] is calculated only if the seta[i] and setb[i] values are not outliers; otherwise,
the data pairs are rejected. We then compute the expected value Mk for all possible
such successive pairs:
å BPD éëi ùû
k
M k [] = i =1
(1.62)
n
Third, we calculate periodicity by measuring the mean value between the mean Mk.
When the difference between elements tends to be zero between pairs of different days,
the pairs tend to be equal; so the factor of periodicity tends to be zero for the given
pair. When more pairs have a low periodicity factor, then the mean factor value of more
pairs is also low. Smaller periodicity factor means higher periodicity and simulation
model behaviour is more stable in simulation time as opposed to the regular model.
1.3.4 Results Accuracy
Outliers and other common statistical metrics are used to measure the accuracy
of results. Outliers are unusual data values which may distort the results obtained
Large-Scale Multi-Agent Systems Simulation 43
(Barnett, V. and Lewis, T., 1994). When out of an agreed area, we describe a value
as outlier. This region’s boundaries are defined as follows:
L = m ± 0.95stdX (1.63)
where stdX is the standard deviation of the likelihood of blocking or link dropping,
and m is the mean simulation time value.
To investigate how the system design and implementation affects the accuracy and
reliability of the test, we employed several overall statistical metrics with regard to
blocking and dropping probabilities throughout all simulation runs such as expected
value of standard deviation, mean value, variance, etc.
mathematical model (e.g. for signal propagation) since it is based on how the chan-
nels express the clients (MUs).
Both models were initially tested for event-generation with deterministic data.
Figure 1.26 shows how some of the five events (NC1, NC2, RC1, MC1, RC2) per-
form. The scheduler restores the event with the least time stamp (highest priority)
for subsequent execution from the queue. When the event interleaving system man-
ages the same events queue (Figure 1.27), the execution sequence remains identical,
while the scheduler time slice (time slice width – TSW) is greater than or equal to
the event computational time (ECT). In other words, if TSW > ECT, then one slice
of time in the predetermined sequence is enough to complete execution of each gen-
erated event. And the two conceptual models yield similar results. When the TSW
is less than the ECT for finishing execution of the event, or the TSWs are assigned
asymmetrically to active threads, the results are totally different from the traditional
CQ approach. This is due to the competitiveness of running threads, associated with
individual MUs, for specific resources management and the relevant scheme of chan-
nel allocation at different time units.
1.4 EXPERIMENTAL RESULTS
1.4.1 Indicative Results Based on Five Days of Network Operation
The simulation results are based on unbalanced and balanced DCA schemes where
the wireless network consists of 19 cells with capacity of 50 users per cell and 32
channels per cell. Figures 1.28–1.30 show a typical blocking probability graph for
the given network using the daily traffic model (five days of network operation) and
based on the proposed architectures.
0.45
0.4
Blocking probability (%)
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Simulation time
0.45
0.4
Blocking probability (%)
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Simulation time
0.45
0.4
Blocking probability (%)
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000
Simulation time
0.206
0.204
Blocking Probability(%)
0.202
0.196
0.194
0.192
0.19
0.188
0 5 10 15
Executions (x2)
FIGURE 1.31 Proposed balanced DCA and best CNR based on GAs, blocking probability.
0.186
Blocking Probability(%)
0.184
0.182
0.18
0.178
0.176
0.174
0 5 10 15
Executions (x2)
0.206
0.204
Blocking Probability(%)
0.202
0.196
0.194
0.192
0.19
0.188
0 5 10 15
Executions (x2)
0.186
Blocking Probability(%)
0.184
0.182
0.18
0.178
0.176
0.174
0 5 10 15
Executions (x2)
FIGURE 1.34 Proposed balanced DCA and best CNR based on GAs, dropping probability
variance.
Large-Scale Multi-Agent Systems Simulation 49
0.076
0.075
Blocking Probability(%)
0.074
0.073
0.072
0.071
0.07
0.069
Conventional model
0.068 Basic concurrency
Concurrent, Multiple Synch Blocks
0.067
0 5 10 15
Executions (x2)
0.026
0.024
0.022
0.02
Variance
0.018
0.016
0.014
0.008
0 5 10 15
Executions (x2)
FIGURE 1.36 Proposed balanced DCA and best CNR based on GAs, dropping probability
periodicity.
50 Artificial Intelligence (AI)
0.7
Blocking Probability (%)
0.65
0.6
0.55
0.5
0.45
Event Interleaving
Calendar Queue
0.4
0 5 10 15
Executions (x2)
FIGURE 1.37 CQ blocking probability and Event Interleaving techniques (MC runs, refer-
ence model of one cell).
0.85
Dropping Probability (%)
0.8
0.75
0.7
Event Interleaving
Calendar Queue
0.65
0 5 10 15
Executions (x2)
FIGURE 1.38 CQ dropping probability and Event Interleaving techniques (MC runs, refer-
ence model of one cell).
Large-Scale Multi-Agent Systems Simulation 51
15
Delay (sec)
10
0
0 5 10 15 20 25 30
Executions (x2)
FIGURE 1.39 Delay calculations between events arrival and departure time units (equal
arrival time units).
16
PEI-NCDD
PEI-RCDD
14 PEI-MCDD
PEI-FCDD
FEI-NCDD
12 FEI-RCDD
FEI-MCDD
Delay (sec)
FEI-FCDD
10
2
0 5 10 15 20 25 30
Executions (x2)
FIGURE 1.40 Delay calculations between events arrival and departure time units (aperi-
odic and random arrival times).
52 Artificial Intelligence (AI)
20
Delay (sec)
15
10
0
0 5 10 15 20 25 30
Executions (x2)
FIGURE 1.41 Delay measurements between events departure times (equal arrival times).
25
20
15
10
0
0 5 10 15 20 25 30
Executions (x2)
FIGURE 1.42 Delay measurements between events departure times (aperiodic arrival
times).
Large-Scale Multi-Agent Systems Simulation 53
based clearly on the real-time scheduling concept. On the other hand, a variation of
PEI called Fair Event Interleaving (FEI) is also tested. In FEI, each event is executed
in turn based on priority settings, and the system gives equal computational periods
for each event under processing.
REFERENCES
Abar, S, Theodoropoulos, GK, and Lemarinier, P. 2017. Agent based modelling and simula-
tion tools: A review of the state-of-art software. Computer Science Review, 24, 13–33.
Audsley, NC, Burns, A, Richardson, MF, and Wellings, AJ. 1992. Deadline monotonic sched-
uling theory. In Proceedings 18th IFAC/IFIP Workshop on Real-Time Programming
(WRTP’92), June.
Audsley, N, Burns, A, Richardson, M, Tindell, K, and Wellings, A. 1993. Applying new
scheduling theory to static priority pre-emptive scheduling. Software Engineering
Journal, 8(5), 284–292.
Audsley, NC, Burns, A, Davis, RI, Tindell, KW, and Wellings, AJ. 1995. Fixed priority pre-
emptive scheduling: An historical perspective. Real-Time Systems, 8, 173–198.
Bajaj, L, Takai, M, Ahuja, R, and Bagrodia, R. 1999. Simulation of large-scale heteroge-
neous communication systems. In Proceedings of IEEE Military Communications
Conference (MILCOM’99).
Baker, Mark, Rajkumar Buyya, and Dan Hyde. “Cluster computing: A high-performance
contender.” arXiv preprint cs/0009020 (2000), pp. 79–83, https://arxiv.org/ftp/cs/paper
s/0009/0 009020.pdf
Balakrishnan, V, et al. 1997. A framework for performance analysis of parallel discrete event
simulators. In Proceedings of the 29th Conference on Winter Simulation.
54 Artificial Intelligence (AI)
Lai, WK, and Coghill, GC. 1996. Channel assignment through evolutionary optimization.
IEEE Transactions on Vehicular Technology, 45(1), 91–96.
Lee, WCY. 1995. Mobile Cellular Telecommunications. McGraw-Hill Professional.
ISBN:978-0-07-038089-9, https://dl.acm.org/doi/ book/10.5555/555761
Liu, CL, and Layland, JW. 1973. Scheduling algorithms for multiprogramming in a hard real-
time environment. Journal of the ACM, 20(1), 46–61.
Liu, W, Chiang, CC, Wu, HK, Jha, V, Gerla, M, and Bagrodia, R. 1996. Parallel simula-
tion environment for mobile wireless networks. In Proceedings of the 1996 Winter
Simulation Conference, WSC’96, 605–612.
Liu, J, Perrone, LF, Nicol, DM, Liljenstam, M, Elliott, C, and Pearson, D, 2001, June.
Simulation Modeling of Large-Scale ad-hoc Sensor Networks. In European Simulation
Interoperability Workshop (Vol. 200, No. 1).
Locke, CD. 1992. Software architecture for hard real-time applications: Cyclic executives vs.
fixed priority executives. The Journal of Real-Time Systems, 4, 37–53.
MacDonald, VH. 1979. The cellular concepts. The Bell System Technical, Journal, 58,
15–42.
Maes, P. 1995. Artificial life meets entertainment: Life like autonomous agents.
Communications of the ACM, 38(11), 108–114.
McCormack, WM, and Sargent, RG. 1981. Analysis of future event-set algorithms for dis-
crete event simulation. Communications of the ACM, 24(12), 801–812.
Mehta, S, Kwak, KS, and Najnin, S. 2010. Network and System Simulation Tools for Next
Generation Networks: A Case Study. Modelling, Simulation and Identification, Azah
Mohamed (Ed.), ISBN: 978-953-307-136-7, InTech, Available from: http://www.inte
chopen.com/ books/modelling--simulation-andidentification /network-and-system-simu
lation-tools-for-wireless-networks-a-case-study
Miller, J, Weyrich, O, and Suen, D. 1988. A software engineering oriented compari-son of
simulation languages. In Proceedings of the 1988 Eastern Simulation Conferences:
Tools for the Simulationists. Orlando, FL, April.
Misra, J. 1986. Distributed discrete-event simulation, ACM. Computing Surveys, (CSUR),
18(1), 39-65, March.1
Molisch, AF. 2012. Wireless Communications, Vol. 34. John Wiley & Sons, UK, 2nd edition,
ISBN: 1118355687, 9781118355688
Nahas, M. 2014. Studying the impact of scheduler implementation on task jitter in real-
time resource-constrained embedded systems. Journal of Embedded Systems, 2(3),
39–52.
Naoumov, V, and Gross, T. 2003. Simulation of Large Ad Hoc Networks, MSWiM’03. In
Proceedings of the 6th ACM international workshop on Modeling analysis and simula-
tion of wireless and mobile systems (pp. 50–57), September, San Diego, CA.
Nierstrasz, O. 1989. A survey of object-oriented concepts. In W Kim and F Lochovsk, editors,
Object- Oriented Concepts, Databases and Applications. ACM Press, Addison Wesley,
3–21.
Norman, T, and Long, D. 1995. Goal creation in motivated agents. In Proceedings of the
Workshop on Agent Theories, Architectures, and Languages on Intelligent Agents.
Papazoglou, PM, and Karras, DA. 2016. A conceptual multi-agent modeling of dynamic
scheduling in wireless sensor networks. Lecture Notes in Electrical Engineering, 348,
385–398. Publisher Springer-Verlag, ISSN:1876-1100.
Papazoglou, PM, Karras, DA, and Papademetriou, RC. 2008a. An improved multi-agent
simulation methodology for modelling and evaluating wireless communication sys-
tems resource allocation algorithms. Journal of Universal Computer Science, 14(7),
1061–1079.
Large-Scale Multi-Agent Systems Simulation 57
Papazoglou, PM, Karras, DA, and Papademetriou, RC. 2008b. On a new generation of event
scheduling algorithms and evaluation techniques for efficient simulation modelling
of large scale cellular networks bandwidth management based on multitasking the-
ory. WSEAS Transactions on Communications, 7(10), 1024–1034, Publisher: World
Scientific and Engineering Academy and Society (WSEAS) Press, ISSN: 1109-2742.
Papazoglou, PM, Karras, DA, and Papademetriou, RC. 2009. On the efficient implementa-
tion of a high performance multi-agent simulation system for modeling cellular com-
munications involving a novel event scheduling. International Journal of Simulation
Systems, Science & Technology, Special Issue on: Performance Engineering, 10(1).
Papazoglou, PM, Karras, DA, and Papademetriou, RC. 2011. Evaluating novel DCA varia-
tions for efficient channel assignment in cellular communications through a generic
Java simulation system. In Proceedings of 2011 Intern. Conference on Computer
Science and Network Technology. Harbin, December 24–26.
Papazoglou, PM, Karras, DA, and Papademetriou, RC. 2016. On integrating natural comput-
ing based optimization with channel assignment mining and decision making towards
efficient spectrum reuse in cellular networks modelled through multi-agent system
schemes. Lecture Notes in Electrical Engineering, 348, 783–798. Publisher Springer-
Verlag, ISSN:1876-1100.
Pincus, R. (1995). Barnett, V., and Lewis T.: Outliers in Statistical Data. J. Wiley & Sons
1994, XVII. 582 pp.,£ 49.95. Biometrical Journal, 37(2), 256–256.
Pont, MJ, Kurian, S, Wang, H, and Phatrapornnant, T. 2007. Selecting an appropriate sched-
uler for use with time-triggered embedded systems. In EuroPLoP, 595–618.
Reddy, D, et al. 2006. Measuring and explaining differences in wireless simulation models.
In 14th IEEE International Symposium on Modeling, Analysis, and Simulation. IEEE.
Russell, EC. 1999. Building Simulation Models with Simscript II.5. CACI Products Co.,
CACI Products Company, 3333 North Torrey Pines Court La Jolla, California 92037,
https://www.cosc.brocku.ca /Offerings/4P94/sii5/zbuildin.pdf
Russell, S, and Norvig, P. 2003. Artificial Intelligence: A Modern Approach, 2nd ed. Upper
Saddle River, NJ: Pearson Education.
Schriber, TJ, and Brunner, DT. 1997. Inside discrete-event simulation software: How it works
and why it matters. In Proceedings of the 1997 Winter Simulation Conference.
Siangsukone, T, Aswakul, C, and Wuttisittikulkij, L. 2003. Study of optimised bucket widths
in calendar queue for discrete event simulator. In Thailand’s Electrical Engineering
Conference (EECON-26).
Siow, R, Goh, M, and Thng, L-J. 2004. DSplay: An efficient dynamic priority queue structure
for discrete event simulation. In Simtect Simulation Conference.
Sivarajan, KN, McEliece, RJ, and Ketchum, JW. 1990. Dynamic channel assignment in cel-
lular radio. In IEEE 40th Vehicular Technology Conference, 631–637.
Smith, DC, Cypher, A, and Spohrer, J. 1994. KidSim: Programming agents without a pro-
gramming language. Communications of the ACM, 37(7), 55–67.
Splunter, S, Wijngaards, N, and Brazier, F. 2003. Structuring agents for adaptation. In E
Alonso et al. editors. Adaptive Agents and Multi-Agent Systems, LNAI, Vol. 2636, 174–
186, Heidelberg: Springer, Berlin.
Sprunt, B. 1990. Aperiodic Task Scheduling for Real-Time Systems. PhD dissertation.
Carnegie Mellon University.
Subramania Sharma, T, and Thazhuthaveetil, MJ. 2001. TWLinuX: Operating system sup-
port for optimistic parallel discrete event simulation. High Performance Computing–
HiPC 2001, 262.
Sycara, K. 1998. Multi-agent systems. Artificial Intelligence Magazine, 19(2), American
Association for Artificial Intelligence, 0738-4602-1998.
58 Artificial Intelligence (AI)
Takai, M, Bagrodia, R, Lee, A, and Gerla, M (1999, August). Impact of channel models on
simulation of large scale wireless networks. In Proceedings of the 2nd ACM interna-
tional workshop on Modeling, analysis and simulation of wireless and mobile systems
(pp. 7–14).
Tan, KL, and Thng, L-J. 2000. Snoopy calendar queue. In Proceedings of the 2000 Winter
Simulation Conference.
Tolk, A. 2012. Verification and validation. Engineering Principles of Combat Modeling and
Distributed Simulation, pp. 263–294, published by John Wiley and Sons, NJ, USA,
ISBN: 978-0-470-87429-5
Tropper, C. 2002. Parallel discrete-event simulation applications. Journal of Parallel and
Distributed Computing, Volume 62, Issue 3, March 2002, 327–335, Elsevier
West, J, and Mullarney, A. 1998. ModSim: A language for distributed simulation. In
Proceedings SCS Multi Conference on Distributed Simulation. San Diego, CA,
February, 155–159.
Whitaker, P. 2001. The Simulation of Synchronous Reactive Systems in Ptolemy II. Electronics
Research Laboratory, College of Engineering, University of California, https://ptolemy
.berkeley.edu/publications/papers/01/sr/sr.pdf
Wooldridge, M. 2009. An Introduction to Multi-Agent Systems, 2nd edition. John Wiley &
Sons, UK, ISBN: 978-0-470-51946-2
Wooldridge, M, and Jennings, NR. 1995. Intelligent agents: theory and practice. The
Knowledge Engineering Review, 10(2), 115–152.
Xu, Z, and Mirchandani, PB. 1982. Virtually fixed channel assignment for cellular radio-
telephone systems: A model and evaluation. In IEEE International Conference on
Communications, ICC’92, Chicago, Vol. 2, 1037–1041.
Zeng, X, Bagrodia, R, and Gerla, M. 1998. GloMoSim: A library for parallel simulation of
large-scale wireless networks. In Proceedings of the 12th Workshop on Parallel and
Distributed Simulations.
Zhang, M, and Yum, TS. 1989. Comparisons of channel assignment strategies in cellular
mobile telephone systems. IEEE Transactions on Vehicular Technology, 38(4), 211–215.
Zhang, X, et al. 2001. A proposed approach to sophisticated negotiation. In AAAI Fall
Symposium on Negotiation Methods for Autonomous Cooperative Systems. November.
2 Why Do Users React
Let’s Find Out
Differently to Applications
Infused with AI Algorithms?
Jayden Khakurel, Indu Manimaran, and Jari Porras
CONTENTS
2.1 Introduction..................................................................................................... 59
2.2 Related Work and Hypothesis Formulation..................................................... 61
2.2.1 Excitement........................................................................................... 62
2.2.2 Anger................................................................................................... 62
2.2.3 Desire................................................................................................... 62
2.2.4 Happiness............................................................................................. 63
2.2.5 Relax.................................................................................................... 63
2.3 Methodology.................................................................................................... 63
2.3.1 Participants..........................................................................................64
2.3.2 Procedure.............................................................................................64
2.4 Findings........................................................................................................... 65
2.4.1 Descriptive Statistics and Hypothesis Testing Outcomes.................... 65
2.4.2 Qualitative Feedback...........................................................................66
2.5 Discussions......................................................................................................66
2.6 Limitations and Future Work.......................................................................... 68
2.7 Conclusion....................................................................................................... 68
References................................................................................................................. 68
2.1 INTRODUCTION
In recent years, researchers and application developers have increasingly begun
to infuse applications with machine learning or natural language processing algo-
rithms that offer robust empirical performance to tackle real human problems and
enhance the quality of life in various domains (Khakurel et al. 2018; Inkpen et
al. 2019). Russell, Moskowitz, and Raglin (2017) point out, “Humans’ interaction
with information will only increase in the future, and this interaction will likely be
facilitated by artificial intelligent proxies” (p. 33). This supports the idea that appli-
cations infused with artificial intelligence (AI) algorithms will become ubiquitous
in our lifetimes, and humans will both interact and integrate with these programs.
59
60 Artificial Intelligence (AI)
library (Revelle 2017). The results expand the existing research on the influence of
DSEs on users’ interaction behaviours, providing informal guidelines that can be
used to minimize unpredictable behaviours and improve user engagement with apps
or technologies.
The remainder of this chapter is structured as follows. Section 2.2 provides related
work and hypothesis formulation. Section 2.3 then provides the methodology applied
in the study. We discuss the findings in Section 2.4, and finally, we present the dis-
cussion in Section 2.5 and promising avenues for further research in Section 2.6.
2.2.1 Excitement
Excitement refers to an individual’s response to a situation which increases the
adequacy; it supplements the routine modes of responses which appear inadequate
(Stratton 1928). We propose that when a player interacts with opponents, there
appear to be surprises (i.e., don’t know what to expect in the next move), which usu-
ally excite the other player more than when interacting with precise and pragmatic
games, which are more dry in nature unless the players are playing for the sake
of playing them, addicted to, or wanting to win the games. Such excitement may
cause other players to cry if the excitation is strong or can result in wild laughter.
Therefore, we hypothesize that playing with a human as opponent will lead to excite-
ment, whereas playing against AI-infused apps will lead to less excitement, affecting
a user’s interaction behaviour. H1: Playing with AI-infused apps will lead to less
excitement and thus generate different behavioural responses.
2.2.2 Anger
Williams (2017) and Mill et al. (2018) refer to the term “anger” as a multifaceted
construct that entails a negative activation and can be perceived as a threat to an
individual’s emotional well-being. In line with Charlton (2009) and Hadlington
and Scase (2018), we propose that a poor game app experience with an AI app,
including the app’s self-efficacy and related anxiety, can be a detrimental fac-
tor that may frustrate users, triggering an anger experience. We hypothesize that
playing against AI gaming apps triggers a stronger anger experience than when
playing against a human opponent, which reflects what Mill et al. (2018) refer to
as the “appraisal of an anger-provoking or frustrating situation triggers the anger
experience which, in turn, generates different behavioral responses” (p. 739). In
contrast, we propose that while playing with a human, individuals will use more
coping strategies, and anger will hence be less intense because both players could
solve the matter with a wide range of discussion. H2: Playing with AI-infused
apps will lead to frustration among users, triggering anger experiences and thus
generating different behavioural responses.
2.2.3 Desire
Zalta (Schroeder 2017) refers to desire as a state of mind that is commonly associ-
ated with a number of different effects. Furthermore, the author also points out that
a person with a desire tends to act in certain ways, feel in certain ways, and think in
certain ways. In the context of the current study, during a game experience, players
may have the desire to win the game, regardless of who they are playing against.
Previous studies have shown that when it comes to desire in game experiences, indi-
viduals take relative gains more vigorously than personal gains because they either
have the desire to defeat, or create a flow of experiences during the game, or do not
like receiving less than others (Messick and Thorngate 1967; Meyer-Parlapanis et al.
2017). However, we propose that when playing against AI-infused apps, the desire to
Let’s Find Out 63
win is heightened more because of the collision between two perceptual factors, i.e.,
time pressure and achieving the highest score. Therefore, we hypothesize that play-
ing with an AI game app increases the desire to win compared with playing against
a human opponent, leading the user to behave differently. H3: Increased feelings
of the desire “to win” while playing against AI-infused apps will evoke different
behavioural responses.
2.2.4 Happiness
Previous studies have conceptualized happiness as the positive emotions formed due
to general interaction between internal (endogenic) and external (exogenic) factors,
which are associated with and precede numerous successful outcomes (Lyubomirsky
et al. 2005; Dfarhud et al. 2014; Harmon-Jones et al. 2016). In the context of the
current study, we propose individuals are less happy during and after the game
experience with the AI because of (i) social disengagement (Baym et al. 2004), (ii)
presence of parameters related to game addiction (i.e., salience, mood modification,
tolerance, withdrawal, conflict, and relapse [Griffiths 2008; Hull et al. 2013]), and
(iii) frustrating experiences (i.e., error messages, dropped network connections, long
download times, and hard-to-find features [Ceaparu et al. 2004]). In line with Mehl
et al. (2010), in face-to-face interactions (i.e., conversations while having the game
experience), excitement may actually make individuals happier than the actual game
experience with the AI app. Therefore, we hypothesize that interpersonal communi-
cation and other DSE entities, namely excitement, may cause individuals to remain
happier while playing against a human opponent. H4: The interpersonal communica-
tion and other DSE entities resulting from the dyadic game experience will influence
happiness, generating more positive behavioural responses than the game experience
with an AI app.
2.2.5 Relax
Previous studies (Smith 2007) have organized relaxation states into four groups:
basic relaxation, core mindfulness, positive energy, and transcendence. These relax-
ation states are essential in creating a relaxation response (Benson et al. 1974). In the
context of the current study, we propose that a face-to-face game experience will let
players disclose personal items, build impressions, and compare values (Baym et al.
2004), which will eventually lead to basic relaxation. Therefore, we hypothesize that
playing a game against a human opponent makes a player more relaxed than play-
ing against an AI app, leading the user to behave differently. H5: Having a human
opponent during the game experience makes players more relaxed than having an
AI opponent.
2.3 METHODOLOGY
The following sections elaborate on the methodological approach used in the current
study.
64 Artificial Intelligence (AI)
2.3.1 Participants
A total of 26 participants (19 males, 7 females; mean age: 25.6 years; standard devia-
tion: 2.5; with adequate knowledge about technology; 12 participants were outgoing
and enjoyed being with a lot of people; 15 participants enjoyed more solitude) had
mobile devices using an Android operating system (OS) were voluntarily recruited
at the university by word-of-mouth. The sample size is sufficient based on the rec-
ommendation from Macefield (2009), where the author states that a group size of
3–20 participants is typically valid, with 5–10 participants demonstrating a sensible
baseline range in experimental studies. All of the participants were presented with
an informed consent form detailing participants’ right to confidentiality, risks, data
storage, the use of anonymized data, the voluntary nature of participation, and that
no health-related data would be collected. No incentives were provided in exchange
for participation.
2.3.2 Procedure
We adopted the complete counterbalancing technique to control the order effect,
which is generally recommended for conditions (K!) less than 4, K ≤ 4 (Allen Mike
2017). For instance, in this study, there are two conditions (k! 2), condition A: apps
game experience, and condition B: dyadic game experience. Using all possible
orders, two different combinations – AB, BA – were generated. Two groups (a total
of 26 participants) were formulated in which group 1 (14 participants) interacted
with conditions A and then B. Figure 2.1 shows the complete counterbalancing for
an experiment with two conditions adopted in this study.
During the first session, group 1 (14 participants) interacted with condition A,
where participants were asked to download an android-based game app based on the
traditional game “rock–paper–scissors”; this app used machine learning algorithms.
The participants were instructed to play exactly 30 rounds. While playing, players
made a hand gesture for rock, paper, or scissors before the timer ran out. The win-
ner of the round was decided based on the standard rules. The players then reported
the DSEs that they experienced while interacting with the game through the DEQ
adopted from Harmon-Jones et al. (2016). In a survey, the participants were asked to
rate on a 7-point Likert scale (1 = not at all to 7 = an extreme amount): “While playing
with a mobile game with integrated AI algorithms, to what extent did you experience
2.4 FINDINGS
2.4.1 Descriptive Statistics and Hypothesis Testing Outcomes
The results indicate that participants remained excited (M = 6.54, SD = 0.859),
had desire (M = 6.62, SD = 0.571), were happy (M = 5.77, SD = 0.704), and were
relaxed (M = 6.50, SD = 1.273) while playing against a human opponent compared
with the AI app. The results also indicate that individuals were angrier (M = 2.58,
SD = 1.79) while playing against the AI compared with a human opponent. To
establish the main hypotheses, the sub-hypotheses formulated in Section 2.3 were
tested, they are summarized in Appendix A. The results indicate that all the sub-
hypotheses are correct. The results show DSEs, namely excitement (p < 0.001;
r = 0.74), happiness (p < 0.001; r = 0.59), and relaxation (p < 0.001; r = 0.71), were
felt during the dyadic experience, whereas there were increased feelings of the
desire “to win” when playing against the AI-infused app. However, anger caused
because of frustration (p < 0.05; r = 0.38) was found to have medium effect on
the app’s game experience. The research findings support that either a lack or
66 Artificial Intelligence (AI)
2.4.2 Qualitative Feedback
We further analyzed our qualitative data that measured the overall experience after
playing the games to see what essential challenges affected the emotions of users
while interacting with the AI app. Many participants commented that the interac-
tion between humans was more fun, more competitive, and quicker than with the
AI app. For example, P6 commented, “Human is more fun definitely, but AI could
be an alternative when you want to play and no one is around.” The comment shows
that users may use the mobile app as an alternative method when lonely. Similarly,
P7 stated, “Human was more fun, AI takes more time to recognize the gesture, so I
feel that communication is easier and fun with humans.” Further, P12 commented,
“Human is more fun as the game is much quicker and competitive.” Consequently,
some participants also commented that AI game lack a certain level of intensity.
For example, P8 commented, “Playing with human is more fun. Rock, paper, scis-
sors should be more intense, and I didn’t feel the intensity. Good work on the app
though!” Here, further research should be conducted on why users like the highly
intense AI games and to what level the game should be intensified.
Further, out of 26 participants, 7 expressed that they had challenges regarding
gesture detection when using the app, which caused frustration. For example P4
commented, “The app didn’t recognize my hand gesture in a high percentage of
cases. I wasn’t sure if I should hold my hand vertically or horizontally.” P13 stated,
“The app recognizes rock all the time. It never recognizes scissors!” P3 noted, “Out
of 10 rounds only two rounds were recognized properly … makes me feel frustrated.”
Regarding enriching the user experience, the need for interactive visual feedback in
the form of either graphics or emojis was stated. For example, P12 said, “The app
should give some graphical feedback or emoji signs to make it more interactive.”
2.5 DISCUSSIONS
In light of concerns about why users behave differently with AI-infused apps/systems,
the current study performed two tests. From a theoretical point of view, the results
indicate that DSEs are important, and AI-based apps should have an ability to evoke
and include emotions such as happiness, excitements, and relaxation at the same level
as they have during dyadic interaction to develop mutual emotional attachments. The
finding is in line with the proposal presented by LaGrandeur (2015), who states that
inducing emotions is important to create safer and more attractive AIs, allowing both
humans and AI to develop mutual emotional attachments. Brave and Nass (2002) also
state that “any interface that ignores a user’s emotional state or fails to manifest the
appropriate emotion can dramatically impede performance and risks being perceived
as cold, socially inept, untrustworthy, and incompetent” (p. 82). Further, supporting
previous research (Thüring and Mahlke 2007), this study further reveals that poor
app experience triggers the anger experience among users that leads them to behave
Let’s Find Out 67
differently. Based on the current study’s findings, we suggest the below informal guide-
lines for application developers seeking to design better user experiences and mini-
mize unpredictable behaviours, such as anger experience. These guidelines may not
adequately address all AI-infused apps; however, they can be taken into consideration.
Improving the detection rate of gesture: although the model performed well dur-
ing development, the participants reported that they were frustrated while playing
because sometimes the app failed to detect their hand gestures. Thus, we recommend
developers to (i) use a public data set to train the model offline and update the appli-
cations more frequently to improve gesture detection n, (ii) use images that are taken
in different lighting conditions and with different backgrounds, (iii) to quantify the
detection performance, use Detection Error Trade-off (DET) curves and miss rate
versus False Positive Per Window (FPPW) as proposed by Lahiani and Neji (2018),
and (iv) apply the gesture recognition method based on a convolutional neural net-
work (CNN) and deep convolution generative adversarial networks (DCGAN) pro-
posed by Fang et al. (2019), which can train the model using fewer samples and
achieve better gesture classification and detection effects.
Considering interference time, response time, and progress indication with the
AI-infused app to improve the user interaction: some participants reported that they
were impatient because of the response time when interacting with the app. AI-infused
apps are built with (i) a set of powerful algorithms within the app itself that can be used
to collect the data sets, train the model, and display the results or (ii) can use different
algorithms on the data set on the server side to analyze patterns in data and make pre-
dictions, displaying the results on the app by synchronizing them between the app and
the server. In the latter scenario, there might be delays in the response because of mobile
computational power and server response time because of slow database queries, librar-
ies, and resource central processing unit (CPU) starvation, and slow application logic,
which may increase users’ cognitive load and affect their performance (Alnanih and
Ormandjieva 2016). Therefore, we recommend application developers (i) consider run-
ning the longer running operation as the background task; (ii) test the interference time;
and (iii) test the response time on apps utilizing different time frames, as mentioned by
Nielsen (1993). Furthermore, in order to reduce uncertainty among users, we recom-
mend developers to provide an accurate estimation of the waiting times using either a
determinate or an indeterminate progress indicator (Luo 2017).
Automatic recognition of emotional states and visualizing in the form of graphi-
cal feedback or digital pictograms/emojis to develop human–AI emotional attach-
ments: some participants were more interested in having graphics or having digital
pictograms such as emoji as a way to receive feedback. This may be because apps
that use emojis as a feedback method communicate a positive effect, specifically
joy (Riordan 2017), or the graphical feedback makes users less frustrated and more
engaged compared with textual feedback (Rieber et al. 1996). Stark and Crawford
(2015) point out, “Emojis can act as an emotional coping strategy and a novel form
of creative expression, even if, in both cases, working within real limits” (p. 1).
Therefore, the form of feedback should manifest the appropriate DSEs with a new
algorithm that describes the relationship between cognition and emotion better based
on behavioural science and neuroscience (Wang et al. 2016). Aiming at inducing
68 Artificial Intelligence (AI)
and evoking the emotions on apps through graphical feedback or digital pictograms/
emojis, developers could refer to the work of Gao et al. (2012) on how tactile behav-
iour on the screen can reflect users’ emotional state.
2.7 CONCLUSION
In recent years, AI agency has been applied in forefront applications to simulate
the intelligent behaviour and critical thinking to perform a range of activities that
require human intelligence, such as decision-making, learning, sensing, and chal-
lenging humans through games. However, previous studies have shown that there is a
need to understand the constraints of a user’s interaction behaviours with AI-infused
apps. We formulated and tested a hypothesis using the perspective of DSEs. The con-
tributions of the current chapter are twofold. First, this research provides new knowl-
edge about the influence of DSEs on users’ interaction behaviours with AI-infused
apps. Second, it provides an informal guideline on how to minimize unpredictable
user interaction behaviours and improve their acceptance of these services or tech-
nologies in the future. In conclusion, the present study suggests DSEs should be an
indispensable part of AI-infused apps (i) for individuals to react the same way as
they act during human–human interactions and (ii) to minimize their adverse effects
to enrich human–AI relationships.
Appendix A: https://doi.org/10.5281/zenodo.3361476
Acknowledgements. Thank you to Antti Knutas for your contribution during the
data analysis phase and all the reviewers for their valuable comments and consider-
able time and effort. The first author would like to thank INVEST Research Flagship
funded by the Academy of Finland Flagship Programme (decision number: 320162).
REFERENCES
Abdi H (2007) The Bonferonni and Šidák Corrections for Multiple Comparisons. In: Salkind
N (ed) Encyclopedia of Measurement and Statistics. SAGE Publications Inc., Thousand
Oaks, CA.
Let’s Find Out 69
Allen M (2017) The sage encyclopedia of communication research methods (Vols. 1–4).
SAGE Publications, Inc, Thousand Oaks, CA. doi: 10.4135/9781483381411
Alnanih R, Ormandjieva O (2016) Mapping HCI Principles to Design Quality of Mobile User
Interfaces in Healthcare Applications. Procedia Comput Sci 94:75–82. ISSN 1877-
0509, https://doi.org/10.1016/j.procs.2016.08.014
Amershi S, Weld D, Vorvoreanu M, et al (2019) Guidelines for Human-AI Interaction. In:
CHI Conference on Human Factors in Computing Systems Proceedings (CHI 2019).
Angie AD, Connelly S, Waples EP, Kligyte V (2011) The Influence of Discrete Emotions on
Judgement and Decision-Making: A Meta-Analytic Review. Cogn Emot 25:1393–1422.
https://doi.org/10.1080/02699931.2010.550751
Baym NK, Zhang YB, Lin MC (2004) Social Interactions Across Media: Interpersonal
Communication on the Internet, Telephone and Face-To-Face. New Media Soc. https://
doi.org/10.1177/1461444804041438
Benson H, Beary JF, Carol MP (1974) The Relaxation Response. Psychiatry. https://doi.org
/10.1080/00332747.1974.11023785
Bergmann R, Ludbrook J (2000) Different Outcomes of the Wilcoxon—Mann—Whitney
Test from Different Statistics Packages. Am Stat 54:72–77. https://doi.org/10.1080/0
0031305.2000.10474513
Brave S, Nass C (2002) Emotion in Human–Computer Interaction. In: Julie A. J. L. (ed)
The Human-Computer Interaction Handbook. Erlbaum Associates Inc, Hillsdale, NJ,
81–96. doi: 10.1201/9781410615862.ch4
Ceaparu I, Lazar J, Bessiere K, et al (2004) Determining Causes and Severity of End-User
Frustration. Int J Hum Comput Interact 17:333–356. https://doi.org/10.1207/s15327
590ijhc1703_3
Charlton JP (2009) The Determinants and Expression of Computer-Related Anger. Comput
Human Behav 25:1213–1221. https://doi.org/10.1016/j.chb.2009.07.001
Cohen S (1994) Perceived Stress Scale. Psychology 1–3. http://www.mindgarden.com/prod
ucts/pss.htm.
Corener Paul (2013) The Fine Points of Feelings : Why Discrete Emotions Matter and How
to Reveal Them. Quirks. https://emotiveanalytics.com/the-fine-points-of-feeling-why-
discrete-emotions-matter-and-how-to-reveal-them/.
de Freitas JS, Gudwin R, Queiroz J (2005) Emotion in Artificial Intelligence and Artificial
Life Research: Facing Problems. In: Panayiotopoulos T, Gratch J, Aylett R, Ballin D,
Olivier P, Rist T (eds) Intelligent Virtual Agents. IVA 2005. Lecture Notes in Computer
Science, vol 3661. Heidelberg, Springer, Berlin. https://doi.org/10.1007/11550617_ 52.
Dfarhud D, Malmir M, Khanahmadi M (2014) Happiness & Health: The Biological Factors-
Systematic Review Article. Iran J Public Health 43:1468–1477.
Fang W, Ding Y, Zhang F, Sheng J (2019) Gesture Recognition Based on CNN and DCGAN
for Calculation and Text Output. IEEE Access 7:28230–28237. https://doi.org/10.1109/
ACCESS.2019.2901930
Feng W, Tu R, Lu T, Zhou Z (2019) Understanding Forced Adoption of Self-Service
Technology: The Impacts of Users’ Psychological Reactance. Behav Inf Technol
38(8):820–832. doi: 10.1080/0144929X.2018.1557745
Gao Y, Bianchi-Berthouze N, Meng H (2012) What Does Touch Tell Us about Emotions
in Touchscreen-Based Gameplay? ACM Trans Comput Interact. https://doi.org/10.1145
/2395131.2395138
Griffiths MD (2008) Diagnosis and Management of Video Game Addiction. New Dir Addict
Treat Prev 12:27–41.
Hadlington L, Scase MO (2018) End-user Frustrations and Failures in Digital Technology:
Exploring the Role of Fear of Missing Out, Internet Addiction and Personality. Heliyon.
https://doi.org/10.1016/j.heliyon.2018.e0 0872
70 Artificial Intelligence (AI)
R Core Team (2017) R: A Language and Environment for Statistical Computing. https://www
.r-project.org/
Revelle W (2017) Psych: Procedures for Psychological, Psychometric, and Personality
Research. https://cran.r-project.org/package=psych
Rieber LP, Smith M, Al-Ghafry S, et al (1996) The Role of Meaning in Interpreting Graphical
and Textual Feedback During a Computer-Based Simulation. Comput Educ 27:45–58.
https://doi.org/10.1016/0360 -1315(96)00005-X
Riordan MA (2017) Emojis as Tools for Emotion Work: Communicating Affect in Text
Messages. J Lang Soc Psychol 36:549–567. https://doi.org/10.1177/0261927X17704238
Russell S, Moskowitz IS, Raglin A (2017) Human Information Interaction, Artificial
Intelligence, and Errors. In: Autonomy and Artificial Intelligence: A Threat or Savior?
Springer International Publishing, Cham, pp. 71–101. doi: 10.1007/978-3-319-59719-5_4
Rzepka C, Berger B (2018) User Interaction with AI-enabled Systems: A Systematic Review
of IS Research. ICIS Proc no. December:1–17.
Schroeder T (2017) Desire. In: Edward N Zalta (ed) The Stanford Encyclopedia of Philosophy,
Summer 201. Metaphysics Research Lab, Stanford University. https://plato.stanford.ed
u/archives/sum2017/entr ies/desire/
Smith JC (2007) The New Psychology of Relaxation and Renewal. Biofeedback.
Stark L, Crawford K (2015) The Conservatism of Emoji: Work, Affect, and Communication.
Soc Media Soc. https://doi.org/10.1177/2056305115604853
Stratton GM (1928) The Function of Emotion as Shown Particularly in Excitement. Psychol
Rev 35:351–366. https://doi.org/10.1037/ h0071406
Thüring M, Mahlke S (2007) Usability, Aesthetics and Emotions in Human-Technology
Interaction. Int J Psychol 42:253–264. https://doi.org/10.1080/00207590701396674
Tofan D, Galster M, Lytra I, et al (2016) Empirical Evaluation of a Process to Increase
Consensus in Group Architectural Decision Making. Inf Softw Technol 72:31–47. https
://doi.org/ https://doi.org/10.1016/j.infsof.2015.12.002
University of Stirling (2017) Can a Computer Learn Game Strategy as it Plays? Dep Comput
Sci Maths http://www.cs.stir.ac.uk/~kms/schools/rps/index.php. Accessed 28 Nov.
2018.
Väänänen K, Pohjola H, Ahtinen A (2019) Exploring the User Experience of Artificial
Intelligence Applications : User Survey and Human-AI Relationship Model. In: CHI′19
Workshop on Where is the Human? Bridging the Gap Between AI and HCI. Glasgow,
p. 5.
Wang Z, Xie L, Lu T (2016) Research Progress of Artificial Psychology and Artificial
Emotion in China. CAAI Trans Intell Technol. https://doi.org/10.1016/j.trit.2016.11.003
Williams R (2017) Anger as a Basic Emotion and Its Role in Personality Building and
Pathological Growth: The Neuroscientific, Developmental and Clinical Perspectives.
Front Psychol 8. https://doi.org/10.3389/fpsyg.2017.01950
Wohlin C, Runeson P, Höst M, et al (2012) Experimentation in Software Engineering.
Springer Berlin Heidelberg, Berlin.
Yannakakis GN, Togelius J (2018) Artificial Intelligence and Games. Springer International
Publishing, Cham.
3 AI vs. Machine Learning
vs. Deep Learning
R. Lalitha
CONTENTS
3.1 Introduction: Background and Driving Forces................................................ 73
3.2 Overview of Artificial Intelligence.................................................................. 74
3.3 Steps to Implement Artificial Intelligence Algorithms................................... 74
3.4 When/Where/How/Why to Use Artificial Intelligence?................................. 74
3.5 Examples for Artificial Intelligence Applications........................................... 74
3.6 Overview of Machine Learning...................................................................... 75
3.7 Steps to Implement Machine Learning Algorithms........................................ 75
3.8 When/Where/How/Why to Use Machine Learning?...................................... 75
3.9 Examples for Machine Learning Applications................................................ 76
3.10 Overview of Deep Learning............................................................................ 76
3.11 Steps to Implement Deep Learning Algorithms............................................. 76
3.12 When/Where/How/Why to Use Deep Learning?........................................... 76
3.13 Examples for Deep Learning Applications..................................................... 77
3.14 Comparisons of Artificial Intelligence, Deep Learning, and Machine
Learning.......................................................................................................... 77
3.15 Summary......................................................................................................... 79
73
74 Artificial Intelligence (AI)
reminders for the users to complete a task in time, to answer questions given
by the user, and to order items online.
• With the help of AI, smart home devices can be controlled and used eas-
ily. Smart voice assistant and smart assistant for old age people are a few
examples.
• AI can be used to diagnose disease in human bodies. For example, earlier
cancer detection with AI is possible; radiology assistant with AI is also
possible.
• Process automation in factories with AI.
learning can be used to adapt to the changes in data by itself and can predict the
desired result easily. As the data size grows enormously in any domain, com-
putational algorithms are needed to derive meaningful insights from the data.
Hence, machine learning algorithms are needed and essential to provide precise
information to the user.
TABLE 3.1
AI vs. ML vs. DL: Based on Classification
S.
No. Artificial Intelligence Machine Learning Deep Learning
1 Meaning: Meaning: Meaning:
It makes the software It is a subset of artificial It is also called deep
think intelligently. It is intelligence. It relies on structured learning or
done by studying how patterns. The system uses hierarchical learning. It is
human brains think, statistical models to perform a based on artificial neural
learn, and decide to solve specific task without any networks. It is widely used
a problem. AI systems explicit instructions. It is for feature detection. It can
behave like humans widely used to make be referred as the procedure
without any fatigue, predictions and decisions to implement machine
emotion, and limitations learning
2 Types: Types: Types:
• Weak AI • Supervised learning • Unsupervised
• Strong AI • Unsupervised learning pre-trained networks
• Super-intelligence AI • Reinforcement • Convolutional neural
• Learning networks
• Recurrent neural
networks
• Recursive neural
networks
78 Artificial Intelligence (AI)
TABLE 3.2
AI vs. ML vs. DL: Based on Characteristics
S.
No. Artificial Intelligence Machine Learning Deep Learning
1 Key components: Key components: Key components:
The AI algorithms mainly The algorithms depend on The deep learning networks
depend on heuristics and structured data. Based on rely on the layers of artificial
the study of human the structured data, the neural networks
intelligence algorithms classify and
predict the required
information
2 Benefits: Benefits: Benefits:
• Error reduction is easy • Enormous volumes of • Possible to extract more
• Easy-to-handle data can be processed features in an incremental
repetitive tasks • Easy to identify manner
• Availability at all time patterns in data • Domain expertise is not
• Applicable to all fields • No human intervention needed
• Automation is easier • Easy to handle multiple • As the scalability of data
• Permanent memory is varieties of data increases, the performance
available will become better
TABLE 3.3
AI vs. ML vs. DL: Based on Performance Measures
S.
No. Artificial Intelligence Machine Learning Deep Learning
1 Performance: Performance: Performance:
The performance of AI algorithms is In machine learning, The performance
measured by comparing the observed performance measures measures are used to
values and inferred values in the data are used to evaluate the evaluate the deep
input. The intelligent agents are the learning algorithms. learning methods and
autonomous entities which direct their The performance models.
activities to achieve the goal. The metrics in machine The performance of
performance measure in AI is done learning are as follows: deep learning
through the intelligent agents 1. Classification algorithms can be
In AI, the different types of intelligent accuracy measured through the
agents are as follows: 2. Logarithmic loss following parameters:
1. Simple reflex agent 3. Confusion matrix 1. Programmability
2. Model-based reflex agent 4. Area under a curve 2. Latency
3. Goal-based agent 5. Mean absolute 3. Accuracy
4. Utility-based agent error 4. Size of the model
5. Learning agent 6. Mean squared error 5. Throughput
The performance measure is the 6. Efficiency
criterion that measures the success of 7. Rate of learning
an agent
AI vs. Machine Learning vs. Deep Learning 79
TABLE 3.4
AI vs. ML vs. DL: Based on Its Workflow
S.
No. Artificial Intelligence Machine Learning Deep Learning
1 Workflow: Workflow: Workflow:
1. Transform the real • Provide input data • Provide the training data
intelligence into the • Analyze the input • Identify the neural network
system • Identify the model
2. Process and train the patterns • Configure the model with the
system • Make the future learning process
3. Deploy and execute prediction • Train the model
the system • Generate feedback • Interpret the results
TABLE 3.5
AI vs. ML vs. DL: Based on Objectives and Limitations
1 Objective: Objective: Objective:
The main objective is to simulate The main objective is to Deep learning has been
human intelligence in a system and design the system for introduced as a new area of
to make the system to work smart learning by itself and machine learning to bring
and to solve complex problems for making predictions machine learning algorithms
closer to artificial intelligence
2 Limitations: Limitations: Limitations:
• Issues and challenges in • Flaws in input data • Need for a very large
integration may lead to amount of data
• Implementation is erroneous output • Training the system is
time-consuming • Bad input may complex and
• Challenges in interoperability bring down the time-consuming
with cross-platforms reliability of the • Requires more graphical
• Difficult to interpret the results output processing units and
machines
Table 3.4 shows the comparisons between AI, ML, and DL based on workflow.
Table 3.5 shows the objectives and limitations of AI, ML, and DL.
Table 3.6 shows the futuristic issues in AI, ML, and DL.
Table 3.7 shows the differences between AI, ML, and DL based on future scope
and tools.
3.15 SUMMARY
This chapter has provided an overview of artificial intelligence, machine learn-
ing, and deep learning and has provided solutions for when, where, how, and why
80 Artificial Intelligence (AI)
TABLE 3.6
AI vs. ML vs. DL: Based on Futuristic Issues
1 Futuristic issues: Futuristic issues: Futuristic issues:
• When it gives imperfect • Ethical issues • Availability of very
solutions, it leads to a huge loss • End-user satisfaction large data set for
in business environment • Variation in norms training
• Imperfect solutions lead to chaos and rules in different • Time consumption in
and will have a great social impact regions and countries deep learning networks
• More human intervention and • False correlation • Data overfitting occurs,
computation is needed to produce leads to when the number of
a perfect AI solution misinterpretation parameters exceeds the
• Availability of skilled resources • Time-consuming to number of observations
for developing solution with AI build correct • Information privacy
• End-user training cost and time structured data • Latency occurs when
• Data protection • Reusability and retraining is needed
• Compatibility with devices and integration for new data and
systems information
• Upgrading the algorithm • Leads to unstable
periodically and its implementation conditions when there
in AI-based system is needed to is variation in input data
produce reliable results
TABLE 3.7
AI vs. ML vs. DL: Based on Future Scope, Tools, and Areas of Application
1 Areas of application: Areas of application: Areas of application:
Examples: Examples: Examples:
• Medical diagnosis • Social media services • Speech recognition
• Aviation • Email spam and malware • Image recognition
• Robots filtering • Natural language processing
• Workspace communication • Virtual personal assistant • Self-driving cars
2 Future scope: Future scope: Future scope:
• It may spread across all • It will be widely used in • Wider scope for
spans of daily life digital marketing development of more deep
• It will be applied more for • It has more scope in the learning tools and standards
language translations field of education • Development of more
• It will be applied in all • It will have a greater simplified programming
branches of engineering impact on social media, frameworks
and for automation search engines, and • Deployment of transfer
• Development of expert predictions learning concept through
systems reusable components
3 Sample tools needed for Sample tools needed for Sample tools needed for
development: development: development:
• TensorFlow • Weka • Caffe
• Keras • PyTorch • Torch
• PyTorch • TensorFlow • DeepLearning4J
• Theano • KNIME • Cuda
AI vs. Machine Learning vs. Deep Learning 81
these techniques can be used. Artificial intelligence vs. machine learning vs. deep
learning is compared based on their meaning, types, examples, benefits, limita-
tions, futuristic issues, tools, future scope, areas of application, etc. Basic features
of artificial intelligence, machine learning, and deep learning are summarized in
this chapter.
4 Ethical Reasoning and
AI and Big Data
Responsibility
Sweta Saraff
CONTENTS
4.1 Introduction..................................................................................................... 83
4.2 Ethics Reasoning in Artificial Intelligence.....................................................84
4.3 Ethical Responsibility in AI ........................................................................... 88
References.................................................................................................................90
4.1 INTRODUCTION
Do we have the capabilities to understand the infinite potentialities of Nature? This
question was contemplated years ago by Sri Aurobindo (1993).
“One might ask whether science itself has arrived at any ultimate truth; on the con-
trary, ultimate truth even on the physical plane seems to recede as science advances.
Science started on the assumption that the ultimate truth must be physical and objec-
tive – and the objective ultimate (or even less than that) would explain all subjective
phenomena.” Sri Aurobindo (1972)
83
84 Artificial Intelligence (AI)
to solve the problem if they are unique or novel. Rubinstein(1998) opined that people
choose the best alternative by integrating desirability with feasibility.
Economists believe that people try to make decisions based on their satisfaction
of wants or pursue a maximum satisfaction model. This model analyses the value
of choice for its efficacy in a particular situation and also in accordance with their
personal preferences, past experiences, etc. According to Oliveira (2007), one can
analyze the “Expected Utility Theory” either analytically or synthetically. When one
adopts the analytical view, they will observe the available choices and then decide
their outcomes, whereas in a synthetic method, they will ascertain the utility first
and then find out means to achieve it.
Based on these contemplations, the following pertinent questions arise – Is ratio-
nality overpowering ethics? Are we reasoning ethically? What strategies are we uti-
lizing in differentiating right from wrong? Does science have the capacity to reason
ethically? Can machines or robots reason like humans? Ethics, morality, and values
are both subjective and culture based. It does not follow the “one size fits all” rule.
Do these advanced algorithms follow policies of fairness above all? Are we careful
about the specificity and sensitivity measures taken by the predictive tools? It evokes
different opinions from utilitarian (consequential) as well as deontological (rule)
school of ethics. People are held accountable for their doings, and similar responsi-
bility needs to be fixed for decisions taken by automated machines. The engineers
and the software experts need to take the ethical responsibility for the outcomes.
It is quite expected that automated vehicles (AV) such as self-driven (SD) cars
would be facing such a dilemma in real-life situations where they must be pro-
grammed to take a quick decision. Which ethical rule will they follow? Would they
harm an innocent pedestrian to save the car speeding from the other side? To come
to an acceptable answer is quite difficult for AI engineers and scientists. We can also
assume that if we present this question to people from different geographical loca-
tions, ethnicity, culture, education, work types and levels, value orientation, age, and
gender, there will be disparities in responses.
Traditionalists may differ from pluralists in their outlook about ethical beliefs.
Their reasoning style is more consequential and dependent upon the probable out-
come. Some may consider saving more people over one due to higher order reasoning
and future orientation, whereas people with emotional attitude may be overwhelmed
by the thought of killing an innocent man. This dilemma still remains unanswered.
First we must understand that the philosophy of ethics is based on inductive reason-
ing. Here, the order and straightforward rules of logic do not apply. There is hardly
an agreement between nations regarding following the same laws and moral code of
conduct for delivering justice. What ethical principles or design will then be imbibed
by an automatic vehicle? Would one nation allow such self-driven cars which do not
fully adhere to their legal systems?
The requirement of ethical reasoning is not limited to self-driven cars. Artificial
intelligence today can predict stock prices, regulate capital markets, admissions, and
tax evasions, and prepare data on a range of requirements like weather forecasts,
cyclones, pollution levels, and epidemics. They are useful in maintaining applica-
tions in educational institutions, jobs, etc. AI has supported advancements and ease
86 Artificial Intelligence (AI)
of living at most levels to the extent of pushing us to dependency and thinking like
machines without emotions. Based on social media, AI can predict an individual’s
preferences, expected behaviour, or even reactions, intentions, beliefs, friends, etc.
Are these predictions on expected human opinion or behaviour reliable? Humans
by nature are dynamic, subjective thinkers who act according to situations. They
present different behaviours based on mood, situations, current wants, and expected
utility. The argument remains – Can a machine accurately predict the behavioural
patterns and emotional sensitivity of humans?
Another ethical challenge is taking informed consent before collection of data or
any information. Data collected from naïve internet users without their knowledge or
prior permission creates an environment of distrust and violation of personal space.
The predictive technology used by Google and Facebook provides us with segre-
gated information based on our search history, creating suggestiveness, which may
be misleading. They sell this information to different companies which are dealing
with products ranging from apparels, processed food, and travel agencies, making
people vulnerable to different frauds.
They can predict locations travelled, home, school, and office addresses, shopping
behaviour, suggestive friends list, chances of your partner’s infidelity, investments,
choice of candidates in an election, etc. All personal data are accessible easily over
the internet. Just like a detective, AI can draw up a conclusion about the future
behaviour, interests, or probable tendencies of an individual. This tool combines
various factors to produce a single score for a person, and this prediction is used
as a guide to take a future course of action. Does it follow the nuances of ethical
reasoning? Predictions can be a brilliant tool, but are they adaptive? Does this intel-
ligence conform to the principles of ethical sensitivity and awareness? Does it reason
ethically?
Descriptive ethics, normative ethics, and applied ethics are recognized as major
domains of ethical philosophy. Descriptive ethics take an empirical view of the
beliefs of the people. This domain defines the principles behind the moral reason-
ing for differentiating right actions from wrong actions. The theorists working on
descriptive ethics look into the values in which people repose their faith and the basic
features of virtuous conduct. The importance of descriptive ethics lies in the fact
that they are different from normative ethics. It believes that the ideals and values
of various cultures are dynamic in nature. They are always evolving with develop-
ment of societies and mankind. The theorists opine that ethical requirements of the
current generation may be different from their predecessors. So descriptive theorists
try to systematically inquire by collecting information and observations from dif-
ferent fields like basic sciences, applied sciences, and social sciences. Colby and
Kohlberg’s (2011) work on cognitive moral development is an example of descriptive
ethics. This field of ethics deals largely with describing and predicting human behav-
iour (Donaldson & Dunfee, 1994). A framework on ethics may include three steps
or major components (Figure 4.1): identifying the sensitivity or awareness, making a
moral reasoning, and engaging in moral conduct.
Normative ethics studies how people should act or what course of action can be
considered ethical (Hoffe, 1989). Normative ethical theories can be observed from
AI and Big Data 87
three perspectives: virtue ethics, deontology, and utilitarianism. Virtue ethics lays
emphasis on individual character, which includes how they are acquired, nurtured,
and applied in real-life scenarios. It is based on a set of stable dispositions where
decisions are not taken based on suitability in a particular context. Traits like hon-
esty, truthfulness, and care are attributes of the personality and not just part of a
habit. It includes choices, interests, values, attitudes, courage, and temperaments.
Virtue ethics differs from deontology as it focuses on the inherent disposition of an
individual rather than his or her adherence to an established set of rules. A virtuous
action is representative of an inner sense of morality and “a way of being” leading to
consistency in action and sensibilities.
Deontology ethics are well-defined or rule-based ethics (Waller, 2005). The main
tenet of the theory lies in engaging in a behaviour when the action is morally justifi-
able rather than considering its consequences. Deontology is different from conse-
quentialism as it believes in moral action rather than its effects (Flew, 1979). It deals
with personal and professional duties and legal rights. Such ethics may contrast with
personal or cultural values, but people have a moral obligation to follow a certain
code of rules. The premise of utilitarian philosophy of ethics is grounded in select-
ing an action based on its perceived consequences. The main thought is to maximize
88 Artificial Intelligence (AI)
4.3 ETHICAL RESPONSIBILITY IN AI
With power comes great responsibility, and systems dependent on machine learn-
ing and artificial intelligence are no exception (Dignum, 2017). According to
Dignum (2018),
Whatever their level of autonomy and social awareness and their ability to learn, AI
systems are artefacts, constructed by people to fulfil some goals. Theories, methods,
algorithms are needed to integrate societal, legal and moral values into technologi-
cal developments in AI, at all stages of development (analysis, design, construction,
deployment and evaluation).
AI and Big Data 89
Both ethical reasoning and responsibility form the edifice on which future moral robots
must be built. The primary goal is to train robots to behave according to the programmed
code of conduct. However, can they explain the reasons behind their conduct? So, this
responsibility of training and mentoring ethical robots lies on the software programmer.
The next question is whether the programmer is equipped with moral reasoning? Does
he understand the nuances of virtue, justice, and legality? How to solve the ambiguities
of an ethical dilemma and make the most fair and honest choice?
A framework of ethical responsibility envisaged on the teachings of Vedanta is
relevant for most of the cultures, societies, and overall development of humanity.
1. Delivering one’s duty without any doubt and lethargy is expected from all.
The world we live in will be better if each individual performs even the
smallest task assigned to them. There is no use talking about reforming the
world without caring for the pressing problems of the moment. The imme-
diate crisis, the task that lies ready to hand, must engage the individual’s
attention, and if he does it well, he will indirectly help in the betterment of
the world. Every man has a certain place to occupy and a certain function
to fulfil in the social economy.
2. Belief in respecting dignity of all mankind above self-interest and personal
desires is of paramount importance. Justice and benevolence are the pri-
mary attributes required to fulfil the ideals of social service. The laws made
for the benefit of society, if given more importance than the existence of
humanity, will lead to extremism by creating conflicts between duties. The
laws serve as the guiding path to deliver justice without any biasness. But
man has no call to act in blind obedience to them. The need is not confor-
mity to rules, but conformity to the law of reason. When the rules come
into conflict, we must fall back on the supreme commandment, and ask
ourselves which course is most conducive to the realization of reason in the
world. We must serve as moral rational beings, with acts expressive of the
pivotal purpose.
3. Radhakrishnan (1914) said: “The inner spirit is more important than out-
ward conformity to law. An action is good, not because of its external con-
sequences, but on account of its inner will. Virtue is a mode of being and
not of doing. It is not something to be found, but a function or an exercise of
the will.”
4. Reasoning credits us with accountability towards ideals of justice, human-
ity, and righteousness which must be delivered without any prejudices. If
each individual is given the power to play according to their conviction, the
universal goal of social balance and harmony would be destroyed. Logical
judgement with moral intent must not be sacrificed to fulfil the norms of
social convention. Myopic traditions and values are regressive and must not
be favoured at the cost of knowledge and growth. We must build institutions
which deliver harmony and progress with equal veracity.
5. No act should be done with a selfish interest or for the gratification of incli-
nations. An individual, if engaged in moral reasoning, serves humanity
90 Artificial Intelligence (AI)
with sensibilities and empathy. The highest ideal in life must be to serve
society by exercising logical judgement even in the moment of crisis.
6. Character, which is the habit of will, is not determined but self-created.
The rational self, if consciously agree to the immoral attributes or accept
them as its stable disposition, then such characteristics are selfish in nature.
Human beings cannot justify a mistake by blaming the genetic lineage. It is
his wilful desire to accept or reject. He must control the acquired behaviour
and reason with honesty and integrity.
7. Visionary idealism which lacks focus and intelligent direction is ineffec-
tive. If private interests are permitted, then there are chances of disrup-
tion in the society. A man cannot stop and debate within himself at every
point as to what his duty is. He must fulfil his duty with enthusiasm and
eagerness.
8. “Know the self to be sitting in the chariot, the body to be the chariot, the
intellect (buddhi) the charioteer, and the mind the reins” – Radhakrishnan
(1914). One must have control over senses for a steady and sharp intellect.
9. Knowledge should not be acquired for the purpose of completing a course
but must enable us to take a right perspective and a pragmatic frame of
reference about other things and our place in the world.
10. We must do the right whether the right is done or not. “Devoted each in his
own work, man attains perfection” Radhakrishnan (1914). Every man is
required to contribute to the national strength his quota of earnest work. It
is by the endorsement of this philosophy of work that the world as a whole
will flourish.
Bonnemains, Saurel, and Tessier (2018) in their paper on “Embedded
ethics: some technical and ethical challenges” propose development of
“formal tools” that are descriptive of different circumstances and “models”
of ethical philosophies which have the potential to reason not only auto-
matically but also ethically, giving proper justification. Today’s millennials
nurture the myth of a utopian world maybe, but such a world looks imper-
fect, even if it is the mightiest. An imperfect world, which is harmonious
and hopeful of future growth, is more blissful and happier.
REFERENCES
Allen, C, Smit, I, Wallach, W. Artificial morality: Top-down, bottom-up, and hybrid
approaches. Ethics and Information Technology, 7 (2005): 149–155.
Aurobindo Sri. The Life Divine. First Edn. Pondicherry: Sri Aurobindo Ashram; 1993. The
Divine Life; p. 1034. 1939-40, Tenth impression 1993.
Aurobindo Sri, Mother. On Science. Pondicherry: Sri Aurobindo Ashram; 1972. Drawbacks
and Limitations; p. 12.
Bohm, D. Science, spirituality, and the present world crisis. ReVision, 15(4) (1993): 147–152.
Bonnemains, V, Saurel, C, Tessier, C. Embedded ethics: Some technical and ethical chal-
lenges. Ethics and Information Technology, 20(1) (2018): 41–58.
Chakraborty, S.K. Values and Ethics for Organizations: Theory and Practice. New Delhi:
Oxford University Press, 1998, 153–171.
AI and Big Data 91
CONTENTS
5.1 Introduction..................................................................................................... 93
5.2 Liquid Level Measurement in Dynamic Environments..................................94
5.2.1 Influence of Temperature.....................................................................94
5.2.2 Influence of Inclination........................................................................ 95
5.2.3 Influence of Sloshes.............................................................................96
5.3 Sensor Design..................................................................................................97
5.3.1 Fibre Bragg Grating Sensor.................................................................97
5.3.2 Cantilever Beam.................................................................................. 98
5.3.3 Float Sensor.........................................................................................99
5.3.4 System and Working Principle.......................................................... 100
5.4 Introducing Neural Networks for Accurate Level Prediction....................... 101
5.4.1 Sampling of Sensor Output................................................................ 101
5.4.2 Artificial Neural Networks................................................................ 102
5.4.3 Activation Function........................................................................... 103
5.5 Wavelet Neural Network................................................................................ 103
5.5.1 Training of WNN.............................................................................. 104
5.6 Results............................................................................................................ 107
5.7 Conclusion..................................................................................................... 108
References............................................................................................................... 109
5.1 INTRODUCTION
A liquid level is an essential parameter to measure in almost every process system such
as the food and beverage industry, petrochemical plants, water reservoirs, and auto-
motive systems. Level measurements based on the principle of differential pressure,
vibrating wire, capacitive magnetic floating, and ultrasonic are conventionally used in
the industries. In automotive, mainly variable resistive, capacitive, float, and ultrasonic
sensors are used for measuring the level of the fuel tank. Resistive float-type sensors
93
94 Artificial Intelligence (AI)
are widely used as automotive fuel indicators more than other conventional sensors.
Many conventional level measurement techniques are accurate only when the liquid
contained in the fuel tank is stationary. With the presence of dynamic environments,
such as variation in temperature, an inclination of the fuel tank, and liquid slosh, most
of the methods fail to produce accurate level measurements which are very essential
[1]. The conventional sensors are incapable of compensating the effects of temperature
variations, inclination, or slosh while measuring the liquid level. The level measure-
ment in such environments requires the best effective system with high sensitivity,
which will result in highly accurate level measurement.
The proposed method of liquid measurement integrates the neural network
approach along with the measurement sensor for improving the system accuracy. As
the neural network is a computing system that predicts the output accurately by con-
sidering the past input data, the ability of a neural network is employed to improve
the accuracy of the proposed system in the dynamic environments [1, 2]. Another
level measurement technique with optical fibre has been tremendously increasing for
the past few years due to its advantages such as being unaffected from electromag-
netic interference, having low signal loss and distortion, and being lightweight. Thus,
designing a system with fibre Bragg grating (FBG), a type of optical fibre, will fur-
ther enhance the performance of the measurement system. The proposed level mea-
surement sensor based on FBG-embedded cantilever beam and a float combination
[3], integrated with the machine learning approach, specifically wavelet neural net-
work, contributes a highly accurate and sensitive system apt for level measurements
in dynamic environments [4]. The proposed design of the liquid level measurement
system is specifically formulated for the automobile fuel tank which is affected by
the dynamic environment when the vehicle undergoes different acceleration.
5.2.1 Influence of Temperature
The temperature has a great impact on the variation of liquid properties, notably den-
sity and viscosity. The movement of molecules in the liquid varies according to the
temperature variations. The volume of the liquid in the fuel tank seems to be increased
or decreased as a result of low or high densities, respectively (density is the ratio of mass
and volume). Viscosity is based on the cohesive forces between the molecules of the
Parameter Estimation in Dynamic Environments 95
TABLE 5.1
Dynamic Viscosity, Kinematic Viscosity, and Density Variations Due to
Temperature Change of Engine Oil SAE 15W-40 [5]
Temperature Dynamic Viscosity Kinematic Viscosity Density
S. No. (°C) (mPa·s) (mm2/s) (g/cm3)
1 0 1328.0 1489.4 0.8916
2 10 582.95 658.60 0.8851
3 20 287.23 326.87 0.8787
4 30 155.31 178.01 0.8725
5 40 91.057 105.10 0.8663
6 50 57.172 66.464 0.8602
7 60 38.071 44.585 0.8539
8 70 26.576 31.350 0.8477
9 80 19.358 23.006 0.8414
10 90 14.588 17.467 0.8352
11 100 11.316 13.648 0.8291
liquid. As temperature increases, the cohesive force between the molecules decreases
and results in a reduction of viscosity; in the same way, a decrease in temperature
causes an increase in cohesive force between molecules and viscosity increases. In
short, both density and viscosity have an inverse relation to temperature.
As the fuel level measurement is considered, the temperature dependency of the
engine oil SAE 15W-40 is given in Table 5.1 [5]. The variation in dynamic viscos-
ity, kinematic viscosity, and density over temperature from 0 to 100°C is mentioned.
It is clear from the table that these three parameters are decreasing concerning the
increase in temperature of the oil. In Figures 5.1 and 5.2, the three parameters are
plotted to analyze the effect of temperature on them.
The viscosity values have a remarkably high dependency on temperature as it
reduces rapidly to very low values up to 50°C, and for further increase in tempera-
ture, the rate of decrease in viscosity is found to be lesser. The density decreases
linearly as the temperature of the fuel increases. From the graphs, it is clear that
the consideration of temperature variation of liquid in the level measurement sys-
tem is significant. Due to the presence of environmental temperatures, the defects
in the coolant circulation system, and excessive loads, the engine temperature will
increase which results in the variation of liquid density and viscosity. As the proper-
ties of liquid vary because of the temperature, the liquid volume in the tank will also
vary accordingly. Thus, the fuel level measurement will give errors in level readings
unless the disturbances are compensated properly.
5.2.2 Influence of Inclination
The automobile is movable systems that pass through different geographical varia-
tions such as elevations, slopes, and irregularities in the road. In such cases, the
96 Artificial Intelligence (AI)
FIGURE 5.1 Dynamic and kinematic viscosity variations of engine oil SAE 15W-40 over
temperature.
FIGURE 5.2 Density variations of engine oil SAE 15W-40 over temperature.
baseline of the sensor will be in inclination with the normal horizontal line and the
liquid level will also get displaced in such a way that one side has the highest-level
reading while the other has the lowest. If in case the position of the sensor is at the
centre of the fuel tank, the level reading is at 45° inclination with the horizontal line
and 15° inclination with the horizontal line, thus considering the inclination is sig-
nificant in the level measurement system of a fuel tank.
5.2.3 Influence of Sloshes
Another main disturbance to be considered is the slosh of the liquid produced
inside the automobile fuel tank during the movement of the vehicle with variable
accelerations [6]. It is a highly unpredicted and undesired disturbance which seri-
ously affects the accuracy of the liquid level measurement system. The sloshes of
Parameter Estimation in Dynamic Environments 97
the liquid vary the liquid level dynamically with respect to the different accelera-
tion of the vehicle, which is difficult to compensate with the conventional liquid
measurement system. Therefore, the size and shape of the fuel tank, acceleration
of the vehicle, and inclination of the tank position are the parameters that vary the
slosh intensity.
5.3 SENSOR DESIGN
5.3.1 F ibre Bragg Grating Sensor
An optical fibre can be transformed into an FBG by exposing the core to intensive
laser light for introducing a periodic modulation in the refractive index. The broad-
band light signal passing through the FBG fibre gets refracted and reflected due
to the gratings. Because of the reflection and transmission of different wavelength
signals, interference occurs. The Bragg condition as given in Equation 5.1 must be
satisfied by the wavelength for the reflection of the optical signal:
lB = 2nL (5.1)
where λB is the Bragg wavelength, n is the refractive index of the optical fibre,
and Λ is the grating period. By changing the parameters n and Λ, the Bragg
wavelength can be varied. The grating length is also one factor that affects the
Bragg wavelength. Figure 5.3 shows the reflected signal from FBG looks like only
a single wavelength is reflecting, so it resembles peak signals. FBGs are passive,
robust, have less size, high sensitivity, and precision, which make the sensor suit-
able in many optical applications such as telecommunication fields and sensor
technologies, where wavelength selection is required [7]. The Bragg wavelength
change is described as
DlB
= C1 + C2 DT (5.2)
lB
where λB represents the Bragg wavelength under the unstrained condition of the FBG,
ΔλB is the Bragg wavelength shift due to the presence of the strain ε and temperature
variation ΔT, C1 = 0.78 × 10 −6 and C2 = 6.67 × 10 −6°C. For longer Bragg wavelengths,
the changes in wavelength occur due to the applied strain and temperature [3, 7, 8].
The axial strain and temperature are the physical parameters that the FBG sen-
sor can measure directly. To measure displacement by the FBG, the displacement
must be converted to the axial strain experienced on FBG by any transducer. The
basic block diagram presenting the working of the FBG sensor for measurement of
displacement is shown in Figure 5.4. The cantilever structure of the sensing element
gets deformed and induces a strain according to the external displacement. This
strain is experienced by the FBG, which converts the displacement into variations in
wavelength, bandwidth, light intensity, and finally to voltage using an optical spec-
trum analyser (OSA) connected to FBG circuitry. Calibrations can be done with this
information concerning the displacement for the measurement.
98 Artificial Intelligence (AI)
5.3.2 Cantilever Beam
A cantilever beam is a beam that is fastened at one end and left free at the other end;
it means one end is fixed to support and connected, while the other end is left without
any support [9, 10]. When pressure or force is exerted on the free end of the beam,
the beam carries this load to the support, i.e., fixed end, from where the moment of
force and the shear stress can be managed. The tendency of an applied force to turn
or twist the beam is the moment of force, while the stress applied parallel to the
beam is the shear stress.
Due to the support on the cantilever beam’s fixed end, the free end can carry
a specific weight which effects in bending of the beam instead of breaking down
Parameter Estimation in Dynamic Environments 99
due to the shear stress. Without external bracing or support, cantilever construction
allows for overhanging structure, a stainless-steel cantilever beam is used in the
described methodology of design. The cantilever beam deflection for the applied
external force in the free end is illustrated in Figures 5.5 and 5.6.
5.3.3 Float Sensor
Float is a structure that is suspended or freely placed in the liquid surface in such a
way that it will float fully or partially over the liquid surface. To follow the variations
in liquid level, floats and float switches use the buoyancy principle which depends on
FIGURE 5.6 Cantilever beam deflection under load at the fixed end.
100 Artificial Intelligence (AI)
the liquid level and its density. Float structure is made up of plastics or PVC mate-
rial having lesser density than the liquid whose level must be measured so that they
can float over the liquid surface. Most of the floats are hollow spherical or capsule-
shaped filled with air with lesser density than the fluid used in the applications.
For reed switch activation, commonly stainless-steel magnetic floats are used, and
they are tubed magnetic floats with a hollow tubed connection for suspending from
the top. In the measurement of the liquid level, with the consideration of strength,
corrosion resistance, and buoyancy, the magnetic floats are suitable in a wide range
of industries. The floats are made up of welding two symmetrical half-shells together
in an airtight manner and the perfection in the process decides the strength and
durability of the float. The welding should be done uniformly in such a way that it
produces a smoothly finished seam, which is like the surface of the rest of the float.
In normal cases, i.e., when FBGs are not strained or no change in the liquid level,
the refractive index does not experience any change. The increase or decrease in the
liquid level forces the cantilever beam to deflect up or down, respectively, resulting
in a change in the refractive index due to strain variations experienced by the optical
fibre. This forms a wavelength shift in the back-reflected light, and the wavelength
shift is detected by an optical spectrum analyser connected with a computer. The
OSA results are stored for further processing, analyzing, and calculations of the
liquid level [11]. This proposed design of the measurement system is used to improve
the sensitivity and accuracy of the liquid level measurement system. In the proposed
setup, the two FBGs are welded to both the surfaces of an elongated cantilever beam,
in a way to provide temperature compensation [12, 13]. The FBG sensors provide
high measurement accuracy and resolution and can measure strain experienced by
the cantilever beam at the point where it is embedded.
[14]. Ts is the sampling interval showing the time between two sample points of the
signal, which is equal to the reciprocal of sampling frequency:
Ts = 1/fs
network weights are adjusted and updated by comparing the output with the target to pro-
duce errors, till the error minimizes or network output becomes equivalent to the target.
For high accuracy, more input and target pairs are required for training the network [18].
5.4.3 Activation Function
The activation function used by the neuron decides the output produced by the neu-
ral network. To deliver the output signal, the activation function considers both the
inputs and the adjusted values of the weights [19]. This transfer function may come
under one of the categories mentioned below:
• Linear (ramp): The output signal is proportional to all inputs and corre-
sponding weights of the neuron.
• Threshold: The output will always be any one of the mentioned levels, com-
paring whether the total input given to a neuron is greater or less than the
specified threshold value.
• Sigmoid: The output has no linear relationship between the inputs but var-
ies continuously. This function resembles real biological neurons compared
to linear and threshold transfer functions.
• Wavelet: The output has no linear relationship between the inputs but var-
ies continuously. Different wavelets can be used as the activation function.
æ - x2 ö
( )
ç 2 ÷ø
f ( t ) = 1 - x2 eè (5.3)
weights and bias values. The non-linear and complex relationships can be modelled
by the ANNs and it is applicable in many real-life practical systems. The response of
unseen data can be inferred by the ANN by learning from the initial input and output
relationships, i.e., an accurate prediction of output is possible from random inputs.
The neural networks can learn even the hidden complex relationship between input
and output by proper training to provide better results. All the mentioned features
introduced the neural network-based machine learning approach in industrial, tech-
nical, image processing, banking, finance, and many other fields as a very powerful
tool for the prediction of unknown, unexpected, and future outcomes.
Among the machine learning approaches, the WNN is faster and accurate, giving
better results compared to the back propagation network (BPN) and support vector
machine (SVM) techniques [4]. Training the programmed WNN algorithm with real-
time data provides true, faultless level readings. Thus, the WNN is useful in different
areas where exact true outputs are necessary, particularly for fuel level measurement in
the dynamic environments where the disturbance behaviour is unknown.
5.5.1 Training of WNN
Neural networks can undergo training to perform a particular task. Many engineering
tools are available for training neural networks. One of the powerful tools for train-
ing, analyzing, and simulating the neural network is the MATLAB software. The
training procedure modifies the weights and bias weights of a network by following a
training algorithm learning rule. Supervised learning and unsupervised learning are
Parameter Estimation in Dynamic Environments 105
two broad classifications of learning rules for data clustering. Supervised learning
always needs the support of a teacher for supervising and guiding the output accord-
ing to what desired response has to be given for each input signal. A training set is
provided with the learning rule for gaining proper network behaviour. While the
inputs are given to the neural network, comparisons are made between the network
outputs and the target values. To get the network outputs close to the target values,
the weights and biases of the network are updated using the learning rules [15].
Unsupervised learning gets trained by the available information only, and this
method never needs an external teacher. It is a self-organization technique because it
self-organizes the data given to the neural network and finds out their common proper-
ties. According to the network inputs, the weights and biases are adjusted and modi-
fied. In unsupervised learning, there is no need for target outputs and a majority of the
unsupervised learning algorithms are based on clustering operations. In the case of
clustering, the algorithm categorizes the input patterns into several finite classes. For
applications such as vector quantization, unsupervised learning is useful [15].
The Training Algorithm for WNN (Forward Calculations) [21]
Vj ( n ) = åWjk ( n ) * Xk ( n ) (5.4)
k =0
f (Vj ( n ) - bj ( n ) )
f a, b (Vj ( n ) ) = (5.5)
aj ( n )
Sum of input neurons to output neuron is
The output of the neuron is calculated by passing V(n) through the non-linear activa-
tion function, obtaining
Learning Algorithm
At time n, the instantaneous sum of squared error is
1
E ( n ) = e2 ( n )
2
1
= éë y ( n ) - d ( n ) ùû (5.8)
2
2
d(n) The desired response at time n
η Learning rate
The method of steepest descent used to minimize the above cost function. Weight
between the hidden layer neuron j and input layer neuron k can be updated according to
¶E ( n )
DWjk ( n + 1) = -h * + m * DWjk ( n )
¶Wjk ( n )
Xk (n)
(
= h * e(n)* s (V (n)* Wij (n)* f a, b Vj ( n ) * ) ai(n)
+ m * DWjk (n) (5.9)
Weight between output layer neuron i and hidden layer neuron j can be updated
according to
¶E ( n )
DWij ( n + 1) = -h * + m * DWij ( n )
¶Wij ( n )
( )
= h * e ( n ) * d (V ( n ) * Wij ( n ) *f a, b Vj ( n ) + m * DWjk ( n ) (5.10)
¶E ( n )
Dbj ( n + 1) = -h * + m * Dbj ( n )
¶bj ( n )
1
(
= -h * e ( n ) * s (V ( n ) * Wij ( n ) * f a, b Vj ( n ) * ) aj ( n )
+ m * Dbj ( n ) (5.11)
¶E ( n )
Daj ( n + 1) = -h * + m * Daj ( n )
¶aj ( n )
Vj ( n ) - bj ( n )
( )
= -h *e ( n ) * d (V ( n ) * Wij ( n ) * f a, b Vj ( n ) * + m * Daj ( n ) (5.12)
( aj ( n ) )
2
Parameter Estimation in Dynamic Environments 107
The training and target data have been loaded to the network model once the training
is done. The neural network parameters such as learning rate, the maximum number of
epochs, and training function are fixed beforehand to call the training function. The trains
are nothing but the training function having training vectors, network objects, and target
vectors as parameters. The training steps involved are shown as a flowchart in Figure 5.12.
5.6 RESULTS
After calculating the samples from the mathematical model, the training data in dis-
turbances such as variations in temperature and inclination of the vehicles are saved
108 Artificial Intelligence (AI)
FIGURE 5.13 Final output: the red line shows the actual level and the blue line shows the
output of WNN.
as individual data. The saved data are loaded to the wavelet neural network and clas-
sified efficiently. After training and validation of the wavelet neural network model,
the proposed system is tested for its accuracy even in the dynamic environments.
The testing inputs are given randomly with the disturbances and dynamic variations
with respect to time. The neural network is tested to predict the level output and it is
compared with the known true values of fuel level of the tank.
The output of the neural network is given in Figure 5.13. The red line shows the
true level and the blue line shows the predicted level output, which are matching
with trained output a at highest accuracy. The MATLAB simulation and valida-
tion results carried out based on the given algorithm shows that the wavelet neu-
ral network predicted accurate results, which are exactly matching with the true
values of the level of the fuel tank at dynamic variations due to temperature and
inclinations.
5.7 CONCLUSION
WNN-based signal classification and signal processing method integrated with
FBG-embedded cantilever sensor was used for accurately predicting the level of
the fuel tank in the automotive system under dynamic conditions. Extensive train-
ing was done to ascertain an effective configuration for the WNN-based liquid level
measurement system. The WNN parameters selection and the signal pre-processing
configurations were all based on training with the values from the model. While
comparing with the existing level measurement methods, the results obtained from
the WNN-based FBG-embedded cantilever beam sensor system were having higher
sensitivity and higher accuracy in the dynamic environments.
Based on the accuracy, the WNN-based system is suitable to use in vehicles
used for racing competitions where vehicles are exposed to high manoeuvres. The
Parameter Estimation in Dynamic Environments 109
drivers of the automotive systems can monitor the dashboard for predicting the fuel
level accurately without any errors due to variations in accelerations, temperature,
and inclination. In conclusion, the proposed measurement system along with wave-
let neural network method for signal processing and classification is effective and
highly accurate in the determination of fuel level using FBG-embedded cantilever
sensor in dynamic environments.
REFERENCES
1. Edin Terzic, Romesh Nagarajah and Muhammad Alamgir, “A neural network approach
to fluid quantity measurement in dynamic environments”, Mechatronics, vol. 21, no. 1,
pp. 145–155, Feb. 2011.
2. Edin Terzic, Jenny Terzic, Romesh Nagarajah and Muhammad Alamgir, A Neural
Network Approach to Fluid Quantity Measurement in Dynamic Environments.
London: Springer-Verlag, 2012.
3. Dong-Sheng Xu, Hua-Bei Liu and Wei-Li Luo, “Development of a novel settlement
monitoring system using fibre-optic liquid-level transducers with automatic tempera-
ture compensation”, IEEE Transactions on Instrumentation and Measurement, vol. 67,
no. 9, pp. 2214–2222, Sept. 2018.
4. Bashir Shokouh Saljoughi and Ardesir Hezarkhani, “A comparative analysis of arti-
ficial neural network (ANN), wavelet neural network (WNN) and support vector
machine (SVM) data-driven models to mineral potential mapping for copper mineral-
izations in the Shahr-e-Babak region, Kerman, Iran”, Applied Geomatics, vol. 10, no. 3,
pp. 229–256, Sept. 2018.
5. https://wik i.anton-paar.com/in-en/engine-oil/
6. Mengmeng Han, Jian Dai and Kok Keng Ang, “Hydrodynamic aspects of moving vehi-
cle with sloshing tanks”, MATEC Web of Conferences, 2018. https://doi.org/10.1051/
matec conf/201821115002
7. Ilian Haggmark and Michael Fokine, “Fiber Bragg gratings in temperature and strain
sensors”, Laser Physics Group, May 2014.
8. Thulasi M Santhi and S Sathiya, “Performance enhanced liquid level sensing system
for dynamic environments”, IEEE 5th International Conference for Convergence in
Technology (I2CT), March 2019.
9. B Vasuki, S Sathiya and K Suresh, “A new piezoelectric laminated cantilever resonance
based hydraulic pump”, 2013 IEEE Sensors Applications Symposium Proceedings, vol.
978, no.1, pp. 4673–4637, April 2013.
10. S Sathiya and B Vasuki, “Higher resonant mode effect on the performance of piezo
actuated 2-DOF rectangular cantilever shaped resonators (2-DOF RCR) for liquid vis-
cosity and density sensing”, Microsystem Technologies, vol. 23, no. 7, pp. 2431–2445,
July 2017.
11. Hua-Fu Pei, Jian-Hua Yin, Hong-Hu Zhu, Cheng-Yu Hong, Wei Jin and Dong-Sheng
Xu, “Monitoring of lateral displacements of a slope using a series of special fibre Bragg
grating-based in-place inclinometers”, Measurement Science and Technology, vol. 23,
Issue 2, article id. 025007, 8pp, 2012.
12. Wenlong Liu, Yongxing Guo, Li Xiong and Yi Kuang, “Fiber Bragg grating based
displacement sensors: state of the art and trends”, Emerald Insight, vol. 39, no. 1, pp.
87–98, 2019.
13. T Guo et al. “Temperature-insensitive fiber Bragg grating liquid-level sensor based
on bending cantilever beam,” IEEE Photonics Technology Letters, vol. 17, no. 11, pp.
2400–2402, Nov. 2005.
110 Artificial Intelligence (AI)
14. M Pharr and G Humphreys. Sampling and Reconstruction. Physically Based Rendering:
From Theory to Implementation. Amsterdam: Elsevier/Morgan Kaufmann, 2004, pp.
279–367.
15. Neural Networks. http://www.doc.ic.ac.uk/^nd/sur prise_96/journal /vol4/cs11/report
.htm l
16. BD Ripley, “Statistical aspects of neural networks”, In OE Barndorff-Nielsen, JL
Jensen and WS Kendall (Eds.), Networks and Chaos—Statistical and Probabilistic
Aspects. London: Chapman and Hall, 1993, pp. 40–123.
17. H Demuth, M Beale and M Hagan, “Neural network toolbox 5 users guide”, MathWorks,
2007.
18. R Rojas, Neural Networks—A Systematic Introduction. New York: Springer, 1996.
19. U von Luxburg and G Rätsch (Eds.), Advanced Lectures on Machine Learning: ML
Summer Schools, Canberra, Australia, February 2–14, 2003. Tübingen, August 4–16,
2003, (Rev. lectures/Olivier Bousquet). Berlin, New York: Springer.
20. Ping Zhou, Chenyu Wang, Mingjie Li, Hong Wang, Yongjian Wu and Tianyou Chai,
“Modeling error PDF optimization based wavelet neural network modeling of dynamic
system and its application in blast furnace iron making”, Neurocomputing, vol. 285, pp.
167–175, April 2018.
21. MR Mosavi, “Wavelet neural network for corrections prediction in single-frequency
GPS users”, Neural Processing Letters, vol. 33, no. 2, pp. 137–150, April 2011.
6 Computer Vision
Concepts and
Applications
Bettina O’Brien and V. Uma
CONTENTS
6.1 Introduction................................................................................................... 112
6.1.1 Evolution of Computer Vision........................................................... 113
6.2 Feature Extraction......................................................................................... 113
6.2.1 Types of Features............................................................................... 114
6.2.2 Feature Extraction Methods.............................................................. 114
6.2.2.1 I. Low-Level Features......................................................... 114
6.2.2.2 Texture Estimator................................................................ 115
6.2.2.3 Colour Histogram............................................................... 115
6.2.2.4 Colour Descriptor............................................................... 116
6.3 Object Detection............................................................................................ 120
6.3.1 Image Classification........................................................................... 120
6.3.1.1 Classification and Localization........................................... 120
6.3.2 Image Segmentation.......................................................................... 120
6.3.2.1 Semantic Segmentation....................................................... 120
6.3.2.2 Demerits of Sliding Window.............................................. 121
6.3.2.3 Instance Segmentation........................................................ 121
6.3.3 Region-based Methods...................................................................... 121
6.3.3.1 Region Proposal.................................................................. 121
6.3.3.2 Region-based Convolutional Neural Network(R-CNN)..... 121
6.3.3.3 Fast Region-based Convolutional Neural Network............. 122
6.3.3.4 Faster Region-based Convolutional Neural Network......... 122
6.3.4 Alternative Methods.......................................................................... 122
6.3.4.1 HOG Features..................................................................... 122
6.3.4.2 You Only Look Once (YOLO)........................................... 122
6.3.4.3 Demerits of YOLO.............................................................. 122
6.4 Computer Vision Hardware, Software, and Services.................................... 123
6.4.1 Computer Vision Hardware............................................................... 124
6.4.2 Software Libraries and Tools............................................................. 124
6.4.3 Computer Vision Services................................................................. 124
6.5 Applications of Computer Vision.................................................................. 125
6.5.1 Healthcare.......................................................................................... 125
111
112 Artificial Intelligence (AI)
6.1 INTRODUCTION
Artificial intelligence (AI), which is the source of subfields like machine learning (ML),
deep learning, computer vision, etc., is a technology that defines intelligence as knowl-
edge acquisition and representation, where knowledge is acquired and applied to a prob-
lem. The objective of any AI problem [1] would be to elevate the success rate and not
the accuracy, whereas an ML problem would aim at utilizing the model to increase its
accuracy but does not concentrate on its success. In general, artificial intelligence is the
ability to engineer intelligent machines and programs, while machine learning is the
ability to learn without being explicitly programmed. AI pertains to decision-making
where the goal is to simulate natural intelligence in machines. But machine learning
learns from data to maximize the performance of the machine based on a specific task.
Deep learning [2, 3], a subset of machine learning, makes the computation of mul-
tilayered neural networks feasible, working with large amounts of data. Algorithms
with single-layered neurons worked just fine, but the weights remained the same
during the training phase, which didn’t have a very good impact on the accuracy.
So, algorithms were designed such that adjustment of weights made changes which
reduced the error. This, in turn, concentrates on delivering high accuracy for task-
specific applications such as object detection [1, 4], image segmentation [5, 6],
speech recognition [7], feature extraction [8, 9], language translation, and other
recent upgrades in technology. These image-related applications significantly co-
relate with another domain of AI: computer vision.
Computer vision [10, 11] is a simple phenomenon that aims at making computers
view and sense its surroundings. It has been the source for almost all computations
which implicate visual content, especially in the form of digital images. Machines
simply illuminate images as a series of pixels with their own set of colours [4] and
values. Hence, it helps to replicate how a human brain reads an image. This is the
basic way of how intelligence will be modelled in a machine. It ultimately aims at
extracting information from pixels. This field of computer science, which initially
worked based on statistics alone, is now switched to deep neural networks [7]. Here,
both the objective and accuracy of achieving the goal are taken care of.
Image processing [2] is another term that is often thought to resemble computer
vision, but it is not. The commonality is that both subfields of AI require images to
work with, but the method in which these images are used is varied. Image process-
ing, as the name implies, processes images by adding in functionalities like smooth-
ing, sharpening, adding contrast, and stretching, or performs other transformations
in images and the output would be an image. But computer vision aims to identify
images and interpret information to provide an output in terms of the image size,
Computer Vision Concepts and Applications 113
colour intensity, etc. It also tries to classify images, identify instances from an image,
etc. Trying to classify images or associate the information between images to differ-
entiate between individuals are applications that involve deep learning techniques.
Image processing uses probabilistic methods like the hidden Markov model, editing,
restoration techniques, and filtering. These methods use neural networks to incorpo-
rate more learning and to model intelligence.
Computer vision has a non-complete list of applications that tends to grow over
the years. It has a dramatic impact on almost all sectors like retail, healthcare, finan-
cial services, and so on. In general, computer vision [10, 11] is applied in automatic
inspection, identification systems, detecting events [12], controlling processes, mod-
elling or recreating objects or environments, navigation, etc. Section 6.2 gives a
detailed explanation of feature extraction and the techniques used to select the neces-
sary information from the image. In Section 6.3, methods used to identify and locate
objects and instances in images are dealt with. In Section 6.4, some of the important
hardware, software, and services in computer vision are discussed. In Section 6.5,
the applications in the field of healthcare, self-driving cars, and automatic target
detection are dealt with. A case study of how computer vision and deep learning
methods influence autonomous robotic path planning around a dynamic environ-
ment with obstacles is explained.
6.2 FEATURE EXTRACTION
Feature extraction is one of the indispensable preconditions for processing images.
In the field of machine learning, feature extraction [8] learns and identifies patterns
from images. It can be interpreted as the transformation from a set of raw data into
derived values called features. A feature [9] represents a function of characteristics
or measurements specifying the assessable property of an image. These features
must be both instructive and non-repetitive. Instead of using the entire data as a
whole, relevant information from features is extracted to perform the expected task.
The extracted features that are rich in the needful information are further used for
feature selection and classification tasks. Alternatively, feature extraction seems to
be a significant form of dimensionality reduction. In the problem of image classi-
fication, the extracted features should contain the relevant information required to
differentiate between classes.
114 Artificial Intelligence (AI)
6.2.1 Types of Features
The outcome of feature extraction would be a set of features referred to as the feature
vector, which is a representation of images. A feature is of good quality when it con-
tains discriminative information that can differentiate between objects. The types of
features can be classified as follows:
1. Local features are features derived from the results of the edge detection or
image segmentation process. They are individual results or features gained
from processing different partitions of an image.
2. Global features are features calculated from the overall image or a well-
ordered sub-section of an image.
3. Pixel-level features are calculated at each pixel in terms of colour and
texture. Colour holds features such as prominence, colour spread, and co-
occurrence, whereas texture holds features like contrast, sharpness, and
change in intensity.
4. Object-level features are calculated in terms of shape, size, texture, and
spatial distribution. Shape and size hold features like perimeter, bending
energy, Fourier shape descriptor, model dimensions, area, and solidity. The
texture holds the same features as pixel-level features. Spatial distribution
holds features like edge connectedness, length of edge, compactness, and
distance among peer neighbours.
5. Domain-specific features are conceptual features that are application
dependent; e.g., character recognition pertains to the formation of letters
alone, whereas face identification deals with facial features alone.
6. Semantic-level features are based on a subset of low-level features that
obtains a high-level concept. The inference or relation is identified between
two or more low-level features.
i. Texture
These features are described in the tone and structure of a texture. Tone defines the
properties of texel’s pixel intensity, and structure defines the spatial representation or
relation between texels. Structurally, the texture is defined as a set of elementary texels in
a regular pattern, and statistically, it can be defined as a measure of intensity arrangement
in any particular region of an image. These measurement sets define the feature vector.
6.2.2.2 Texture Estimator
• The range is the simplest operator which calculates the difference between
maximum and minimum intensity values of neighbours in an image. Range
estimates texture by converting the original image into a brighter image.
The texture is read as the brightness value.
• Variance calculates the sum of squares of the differences between the
intensities of the central pixel and its neighbours, i.e. it simply calculates
the variance between neighbour regions.
ii. Colour
Colour is the primary image perception in understanding how
human brains perceive an image. It is a robust descriptor which sim-
plifies object detection and image segmentation problems. The colours
[4] that are perceived from an image are made up of light of different
wavelengths. A colour space depicts the colour representing its inten-
sity values. Colour space models can either be hardware-oriented based
on a three-colour stimulus like RGB or user-oriented.
6.2.2.3 Colour Histogram
A colour can be described in various shades (e.g., green colour being in shades of
dark or light green). But for the machine to interpret variations between shades of
the same colour, the probability of pixel intensities needs to be defined. Colour his-
tograms define a set of bins, where each bin holds the probability of pixels of a par-
ticular colour. A vector representation of a histogram is Equation 6.1:
{ }
H H éë0 ùû, H éë 1ùû ,¼, H éëi ùû ,¼, H éë N ùû (6.1)
116 Artificial Intelligence (AI)
Here, each H[i] in Equation 6.1 represents the number of pixels in the colour i in an
image and N represents the total number of bins. Histograms need to be normalized
if two different sized images are to be compared. The normalized vector representa-
tion is given in Equation 6.2 and the reduced form is given in Equation 6.3:
{ }
H ¢ H ¢ ëé0 ùû , H ¢ éë1ùû ,¼, H ¢ éëi ùû ,¼, H ¢ éë N ùû (6.2)
P tends to the number of pixels in the image. The range of intensity values for a
greyscale image can be given in the range I(u,v), denoting the pixel intensities which
belong to [0, k − 1], where k represents the number of colours. For example, for an
8-bit image, k = 28 = 256 colours. So, in general, colour histograms deal with prob-
lems like over- and underexposure, brightness, contrast, and dynamic range.
6.2.2.4 Colour Descriptor
Descriptors provide a way to relate pixels in an image. A colour descriptor is catego-
rized under the general information descriptors. Since colour is the most significant
feature of any visual content, it has to be described exclusively. For any image, the
colour description can be performed using the following methods:
To trace the relation between colours among a group of images, the following meth-
ods are used:
åa x = å å a a s
i =1
i i i j ij (6.4)
where σ denotes the covariance between the ith and jth variables.
The variable which contributes less information along the axes
would be dropped off. In terms of matrix representation, a¢Ca,
where C represents the covariance matrix and a represents the vec-
tor of variable weights, PCA finds the weight vector a that mini-
mizes a¢Ca with a constraint defined in Equation 6.5:
åa 2
i = a¢a = 1 (6.5)
åx (6.6)
1
mi =
Ni xÎwi
åy (6.7)
1
mi¢ =
Ni yÎwi
å w x (6.8)
1
mi¢ = T
Ni yÎwi
118 Artificial Intelligence (AI)
mi¢ = wT mi (6.9)
I. High-Level Features
High-level representations of an image are well-built features inclined to be
seemingly hidden. But minimizing the bridge between low-level and high-
level features is significant in emphasizing object detection problems. The
low-level local features after detection can be combined to classify images.
So, to learn in deep about these hidden features, one requires time and
computational power. This is where the most important segment of deep
learning is used. Deep neural networks use several layers to learn high-level
representations of an image.
a. Template Matching
Template matching is a high-level method that tries to identify areas
of an input image using a given image pattern or any patch. So given
an image, this method works by sliding the template over the whole
image to detect edges of instances. The matching approach completely
FIGURE 6.1 Distance between means and class separability of x and y spaces.
Computer Vision Concepts and Applications 119
n m
dist ( I , t , a, b ) = åå ( I ( a + i, b + j ) - t (i, j ))
2
(6.10)
i =1 j =1
where
(a,b) denotes the top corner coordinates of the template t
I denotes the greyscale image with a grey-value template t of size n*m
Correlation is the measure of degree where two variables correlate with each
other in terms of general behaviour if not directly through values. Correlation is
given by Equation 6.11:
å (u - u¢) ( v - v¢)
N
i i
C= i =1
(6.11)
å (u - u¢) å ( v - v¢)
N 2 N 2
i i
i =1 i =1
where
u and v denote the template grey-level image and source image, respectively
u′ and v′ denote the average grey-image level in the template image and source
image, respectively
N denotes the image template size in terms of pixels and is given by
N = columns * row
C takes values that range between ±1, where greater values represent stronger
relationships between images
b. Hough Transform
Hough transform is a method for calculating global feature repre-
sentation, where it gradually builds classes by working with the local
features with the number of classes initially being unknown. Hough
transform is, in particular, used to detect shapes such as lines, circles,
120 Artificial Intelligence (AI)
6.3 OBJECT DETECTION
Object detection [9] is the principal challenge in computer vision. It is known to rap-
idly grow to give improved results in terms of accuracy and testing time when used
with machine learning [1, 13] and deep learning [14] approaches. One such latest
software system that implements object detection algorithms using deep learning
concepts is the Detectron, developed by Facebook. Nowadays, there are more deep
learning pre-trained models [3, 7] used for object detection like Convolutional
Neural Network (CNN), Region-based Convolutional Neural Network (RCNN),
You Only Look Once (YOLO), Fast R-CNN, etc. The basic structure for computer
vision is CNN implemented using deep learning. CNN combines image classifica-
tion, object classification, and localization. For the past few years, CNN has proved
to be a reliable method for object detection and classification due to its speed and
accuracy.
6.3.1 Image Classification
To classify images, the image given as input goes through a deep convolutional net-
work, and this network gives a feature vector to the fully connected layers that pro-
vide different class scores. Finally, the output delivers the content of the entire image
as a whole. This is the basic possible task in computer vision. There are various types
of other computer vision tasks that detect spatial pixels inside the image. These are
achieved through a convolutional neural network.
6.3.2 Image Segmentation
6.3.2.1 Semantic Segmentation
In semantic segmentation [6], the output image is the decision of the category of
every pixel in that image. It is similar to image classification. Rather than assigning
a single category labelled to the entire image, the output of this segmentation [5]
produces a category label for each pixel of the input image. Semantic segmentation
does not differentiate instances, it only labels the category of that pixel. When two
similar objects are close together, the semantic model does not distinguish between
Computer Vision Concepts and Applications 121
the objects. Instead, it labels the whole mass of pixels. This is a small drawback for
semantic segmentation that can be fixed using instance segmentation. One potential
approach to attack semantic segmentation is through classification. The sliding win-
dow method can be approached to segment an image semantically.
For a sliding window, the input image is cropped into many small images and
each cropped image is treated as a classification problem. This will help to find
the category of the central pixel of the cropped image. The sliding window is com-
putationally expensive. To label a pixel, the full image needs to be cropped into
smaller images and this would be expensive to run. A model that would be even
more efficient than semantic segmentation is a fully convolutional network. Rather
than extracting single patches from the image and classifying it separately, it can be
imagined as a whole stack of convolutional layers.
6.3.2.3 Instance Segmentation
Given an input image, the objective is to identify and detect the location and identi-
ties of objects in the image, which is, perhaps, similar to detecting objects [16] but,
instead of predicting a bounding box for each object, the whole segmentation mask
has to be predicted so that a particular instance from the image would be captured.
6.3.3 Region-based Methods
6.3.3.1 Region Proposal
This method does not use deep learning as such, but it is mostly based on a tradi-
tional approach to computer vision. This region proposal network uses the basic
signal processing technique to generate several proposals. For a given input image,
the region proposal network gives an output of thousand boxes as to where the object
might be present. These are relatively fast to run.
images to predict categories. Besides, R-CNN also predicts a regression for each of
the input region proposals. The main disadvantage of R-CNN is that it is computa-
tionally expensive, and the runtime is slow.
6.3.3.3 Fast Region-based Convolutional Neural Network
Problems faced with R-CNN can be fixed by Fast R-CNN. Both methods are almost
the same but rather than each RoI being processed separately, the entire image is sent
through the convolutional network all at once. It gives a feature map that effects high
resolution that corresponds to the entire image. If there are fully connected layers down-
stream, those layers will be expecting fixed-size input. So, reshaping the cropped images
from the convolutional feature map is done using the RoI pooling layer, which is similar
to max pooling. Once the images are wrapped from a feature map, classification scores
can be predicted by running those images through fully connected layers. In terms of
speed, Fast R-CNN is ten times faster than the SPP net (which is in between R-CNN and
Fast R-CNN) and R-CNN. In terms of test time, Fast R-CNN is superfast since the test-
time is dominated by computing region proposals. So, Fast R-CNN ends up being bottle-
necked by computing these region proposals. This problem is solved by faster R-CNN.
6.3.3.4 Faster Region-based Convolutional Neural Network
In Faster R-CNN, self-generated region proposals are created. It runs the entire input
image through convolutional layers to get a feature map representing the entire image
in high resolution. There is a separate region proposal network that works above those
convolutional features, which in turn predicts its region proposals inside the network.
It looks like Fast R-CNN when the region proposals are predicted. Now the image is
cropped from those region proposals and passed to the network for classification.
6.3.4 Alternative Methods
6.3.4.1 HOG Features
The HOG is an important technique that can be used for object detection, mainly in
self-driving cars. It is a feature descriptor, powerful when used with a support vector
machine (SVM) that works for object detection. To determine whether the object is
found or not, each computed HOG descriptor is fed to the SVM classifier. By chang-
ing the image scale, HOG solves the scaling issues. HOG represents a single vector
which describes a segment of an image.
6.3.4.2 You Only Look Once (YOLO)
YOLO algorithm is a better solution to create accurate boxes around the images and is
also faster than CNN. YOLO is just a modified version of CNN. In YOLO, the image is
divided into multiple grids and for each grid, classification and localization algorithm
is implemented. There are multiple versions for YOLO models. The latest version is
YOLO9000 which is faster, stronger, and better. It is trained on 9000 classes.
6.3.4.3 Demerits of YOLO
• Detecting one object multiple times can be confusing and time-consuming
to correct them. This can be solved by Non-Max suppression. This helps
Computer Vision Concepts and Applications 123
to remove boundary boxes that have a low probability that is close to high-
probability boxes.
• No possibility to detect multiple objects in one grid. With the use of Anchor
box, each object is assigned to each anchor box in a single grid to detect
multiple objects. This can also be solved by selecting a small grid, but when
objects are too close to each other this method can fail.
a. OpenCV
OpenCV is the most fundamental multi-platform open-source library
that has in-built techniques and algorithms to implement image-related
problems. It works well for a smaller image data set. But for larger data
sets, since OpenCV does not support GPU, it uses CUDA.
b. MATLAB
Due to the ability of prototyping, MATLAB is the most used research
tool for building computer vision applications. The ease of debugging using
MATLAB is higher. But its demerit is that it is a pay-and-use platform.
c. TensorFlow
TensorFlow is an important tool that is gaining attention nowadays. It
provides a way to integrate deep learning with computer vision. It also
has APIs that can be used for other applications like classification, detec-
tion, and so on. One demerit of TensorFlow is that it consumes a lot of
resources.
d. CUDA
CUDA is a platform developed by NVIDIA to enable increased comput-
ing performance in GPUs. The toolkit includes a primitive library that uses
image and video functions. Memory distribution and power consumption is
one downside to this platform.
b. Amazon Rekognition
Rekognition is a deep learning-based service developed by Amazon to
analyse images and videos. The basic technique supported by this service
is object detection and segmentation. This service provides the best analysis
for facial recognition as well as sentimental analysis.
c. Microsoft Azure Computer Vision API
This service invented by Microsoft provides the same functionalities
with images and videos. Besides, it provides handwriting recognition.
6.5.1 Healthcare
Computer vision combined with machine learning has made adequate progress in
the field of healthcare [2]. The possible areas in healthcare where technology can
directly intervene are medical imaging analysis, monitoring of patients using gad-
gets, predictive analysis of diseases, and much more. Medical imaging is the process
of making visual representations of the inner body organs for analysis. Such analysis
requires feature extraction techniques [17], which is one of the key concerns. Other
disciplines like electroencephalography (EEG) and electrocardiography (ECG) pro-
vide graphical representations of the body organs, whereas on computer vision the
concept of image co-registration is used. Image co-registration is the transformation
of multiple images or photographs from different sources and time into one coordi-
nate system. Co-registration implies the integration of information from different
images into a single image for analysis. This concept has a significant impact in the
field of healthcare as well as in satellites.
A fundamental machine learning algorithm for visual perception is the polyhar-
monic extreme learning machine (PELM) algorithm. This algorithm visually learns
the image details after which it tries to classify images with a particular disease, based
on the learned parameters. This is where machine learning and deep learning are used
along with computer vision for disease prediction. One such important accomplishment
in the healthcare field is the prediction of breast cancer [18] using deep learning.
6.5.2 Augmented Reality
Augmented reality can be discerned as adding digital information to a real-
world environment. AR just creates additional graphics and sounds to an existing
126 Artificial Intelligence (AI)
environment, whereas virtual reality (VR) virtually creates a wholly artificial envi-
ronment of the real world. AR tends to change the perception of the reality observed
from a physical space, which completely involves computer vision techniques. There
are various types of AR [19], under which generic digital augmentation is one. This
type visualizes multiple 3D objects in its environment. Figure 6.4 gives a general
outline of how an AR-based system works and how the hardware, software, and
servers integrate to visualize virtual objects in a captured real-world environment.
To randomly include objects into a physical space, a combination of three tech-
niques is required.
temporal and spatial recognition of the vision, and vision fusion from various sen-
sors. Stereo vision [1, 7] involves the perception of the environment from sensors,
which can be categorized as direct and mediated perception. Perception [20] can be
taken directly from the working space regarding the texture, 3D shape, etc., without
any knowledge of object categories. The basic methodology of how object detection
and classification with localization are performed for a given image has been dis-
cussed in the previous section.
LSTM, the next action towards the goal will be learned. Finally, the model generates
a path from the start to the target.
BIBLIOGRAPHY
1. K. Grauman and B. Leibe, Visual Object Recognition (Synthesis Lectures on Artificial
Intelligence and Machine Learning). Morgan and Claypool, 2011.
2. Muhammad Imran Razzak, Saeeda Naz, and Ahmad Zaib, “Deep learning for medical
image processing: Overview, challenges and the future,” in Classification in BioAapps,
pp. 323–350. Springer: Cham, 2018.
3. N. Buduma and N. Locascio, Fundamentals of Deep Learning: Designing Next-
Generation Machine Intelligence Algorithms. O’Reilly Media Inc., 2017.
4. F.S. Khan, R.M. Anwer, J. Van De Weijer, A.D. Bagdanov, M. Vanrell, and A.M.
Lopez, “Color attributes for object detection,” in 2012 IEEE Conference on Computer
Vision and Pattern Recognition, pp. 3306–3313. IEEE, 2012.
5. Y.-J. Zhang, Advances in Image and Video Segmentation. IGI Global, 2006.
6. M. Seyedhosseini and T. Tasdizen, “Semantic image segmentation with contextual hier-
archical models,” IEEE Transactions on Pattern Analysis and Machine Intelligence
38(5): 951–964, 2015.
7. M. Alam, M.D. Samad, L. Vidyaratne, A. Glandon, and K.M. Iftekharuddin, “Survey on
deep neural networks in speech and vision systems,” arXiv Preprint ArXiv:1908.07656,
2019.
8. I. Guyon, S. Gunn, M. Nikravesh, and L.A. Zadeh, Feature Extraction: Foundations
and Applications, vol. 207. Springer, 2008.
9. Y. Zhang, Y. Chen, C. Huang, and M. Gao, “Object detection network based on feature
fusion and attention mechanism,” Future Internet 11(1): 9, 2019.
10. R. Szeliski, Computer Vision: Algorithms and Applications. Springer Science &
Business Media, 2010.
Computer Vision Concepts and Applications 129
11. S.J. Prince, Computer Vision: Models, Learning, and Inference. Cambridge University
Press, 2012.
12. A.T. Mousavian, M. FiŠer, J. KoŠecká, A. Wahid, and J. Davidson, “Visual representa-
tions for a semantic target driven navigation,” in 2019 International Conference on
Robotics and Automation (ICRA), pp. 8846–8852. IEEE, 2019.
13. J. Stilgoe, “Machine learning, social learning and the governance of self-driving cars,”
Social Studies of Science 48(1): 25–56, 2018.
14. C.C. Aggarwal, Neural Networks and Deep Learning. Springer International
Publishing, 2018.
15. T. Taketomi, H. Uchiyama, and S. Ikeda, “Visual SLAM algorithms: A survey from
2010 to 2016,” IPSJ Transactions on Computer Vision and Applications 9(1): 16, 2017.
16. L. Xu, Y. Li, Y. Sun, L. Song, and S. Jin, “Leaf instance segmentation and count-
ing based on deep object detection and segmentation networks,” in 2018 Joint 10th
International Conference on Soft Computing and Intelligent Systems (SCIS) and 19th
International Symposium on Advanced Intelligent Systems (ISIS), pp. 180–185. IEEE,
2018.
17. J. Gao, Y. Yang, P. Lin, and D.S. Park, “Computer vision in healthcare applications,”
Journal of Healthcare Engineering, 2018, p. 5157020.
18. D. Wang, A. Khosla, R. Gargeya, H. Irshad, and A.H. Beck, “Deep learning for identi-
fying metastatic breast cancer,” arXiv Preprint ArXiv:1606.05718, 2016.
19. R. Silva, J.C. Oliveira, and G.A. Giraldi, “Introduction to augmented reality,” National
Laboratory for Scientific Computation 11, 2003.
20. O.I. Abiodun, A. Jantan, A.E. Omolara, K.V. Dada, N.A. Mohamed, H. Arshad, K.V.
Dada, N.A. Mohamed, and H. Arshad, “State-of-the-art in artificial neural network
applications: A survey,” Heliyon 4(11): e00938, 2018.
7 Generative Adversarial
Network
Concepts, Variants,
and Applications
K. Rakesh and V. Uma
CONTENTS
7.1 Introduction................................................................................................... 132
7.2 Overview........................................................................................................ 133
7.2.1 Deep Learning .................................................................................. 133
7.2.2 Deep Generative Models .................................................................. 133
7.2.3 Generative Adversarial Networks ..................................................... 134
7.3 GAN Architecture ........................................................................................ 135
7.3.1 General Structure............................................................................... 135
7.3.2 Adversarial Process........................................................................... 136
7.3.3 Background Mathematics.................................................................. 137
7.4 GAN Variations............................................................................................. 138
7.4.1 Overview............................................................................................ 138
7.4.2 Techniques......................................................................................... 138
7.4.2.1 Architecture-based Variant Class ...................................... 138
7.4.2.2 Formulation-based Variant Class ....................................... 139
7.5 Applications .................................................................................................. 141
7.5.1 Image Generation and Prediction...................................................... 141
7.5.2 Image Translation.............................................................................. 142
7.5.3 Image Editing.................................................................................... 143
7.5.4 3D Object Generation........................................................................ 143
7.5.5 Video Manipulation........................................................................... 143
7.5.6 Audio Generation and Translation..................................................... 144
7.5.7 Medical Image Processing................................................................. 145
7.6 Conclusion and Future Directions................................................................. 145
Bibliography........................................................................................................... 145
131
132 Artificial Intelligence (AI)
7.1 INTRODUCTION
Humans always prefer automation of tasks. The invention of computers and com-
puter programming has made it possible to automate some tasks by creating a series
of instructions, referred to as algorithm. As we approach more and more complex
problems, algorithms start to demand a certain degree of intelligence that is similar
to human cognitive functions. This kind of intelligence that is induced to a machine
preparing them to mimic our intelligence is termed as artificial intelligence (AI). A
subfield of AI that is responsible for most of the recent advances of many fields is
machine learning (ML). ML algorithms are statistical models built by finding pat-
terns and inferences from the data that is provided. One class of ML algorithms that
learns a significant amount of cognitive tasks is deep learning (DL) [1].
Deep learning comprises multiple layers of artificial neural network (ANN), a
computing system that learns low-level to high-level features of the input data. DL is
the technology behind many of the current awesome products and services like voice
assistants, recommendation systems, self-driving cars, medical image analysis, com-
puter vision systems, and many more [2]. The discussion on the basics of deep learning
and deep neural network (DNN) is made in Section 7.2.1. Though we made signifi-
cant progress in recognition and prediction tasks, the creative applications of AI are
demanding more attention and in recent years more research momentum is noticed
in this application. The research works on generative tasks have many applications in
areas like content writing, film industry, game development, advertisement, etc.
The class of deep learning networks that is capable of generating data from the
input is known as deep generative networks [3]. Over the years, many generative
models have been proposed. All these different approaches for generative modelling
have been extensively discussed in Section 7.2.2. Out of these models, generative
adversarial networks (GAN) is the most popular model used in generation-based
research works and applications. It provides state-of-the-art results and there pre-
vails a debate on social and cultural aspects of its applications.
GAN can mimic the distribution of a data set, such that new data of the same
distribution can be generated, which will be similar to the input data [4]. The GAN
and its architecture are detailed in Section 7.3. Their architecture can be changed
accordingly to our needs or use cases. This flexibility has provided a lot of variations
while dealing with the distribution. Even the distribution of two different data sets
has been utilized to produce some interesting results. The overview of the variations
in GAN and the techniques used in these variations are discussed in Section 7.4.
The various research directions on GAN show several types of GAN that can
be used for different architecture and functions. These diverse types of GAN have
made it possible to use them in a variety of applications. Some of the interesting
applications include generating photo-realistic images, converting an object in an
image to another object, changing features of a person, generating missing areas in
an image, image editing, converting a text to image, future scene prediction, creating
3D objects, medical image generation, and many more. Many of these applications
and their impact on the industry have been discussed in Section 7.5. The future
directions of GAN are discussed in Section 7.6.
Generative Adversarial Network 133
7.2 OVERVIEW
7.2.1 Deep Learning
Machine learning algorithms have been sufficient for basic classification tasks like
spam detection. But for solving complex tasks like image classification, a huge num-
ber of parameters have to be understood by the algorithm. This is not possible in
classical machine learning algorithms. Then, there evolved a particular type of ML
algorithm that was modelled after the human brain known as the artificial neural
network. ANN comprises many connection units called neurons. Weight and bias
values are set as parameters to connections that prevail between neurons in one layer
to every other neuron of the adjacent layer. These parameter values are automatically
figured out using algorithms like backpropagation as the network progresses along
with the data samples. ANN shows significant results than classical ML algorithms.
More and more complex problems have been solved by adding more layers and pro-
gressing deep into ANN. These deep structures are termed as deep learning.
Deep learning is capable of learning even from unsupervised data which makes
it learn features automatically without much human effort. Deep learning has pro-
gressed over the years with the rise of graphical processing units (GPUs) and a huge
increase in data and research over different ANN architectures and techniques.
Many standard deep learning architectures have been proposed for various appli-
cations. DL architectures that are mostly applied in industries include deep neural
networks, convolutional neural networks (CNN), recurrent neural networks (RNN),
deep belief networks (DBN), AutoEncoders, and GAN.
Each network performs better in different applications. DNN is used for basic
prediction and classification tasks of structured data like medical records. These data
sets have rows of data and many parameters that are tuned to provide the required
results. CNN performs better at image processing, object recognition, and image
segmentation. RNN processes sequential data like time-series data and paragraphs
of text for prediction of the upcoming data. Autoencoders are good at capturing the
essential data from the input, and it is used in data compression and image process-
ing. DBN is used in image clustering, recognition, and generation. At present, GAN
is used for the generation of images and videos with high accuracy.
Explicit PDF generative models that use approximation are Variational AutoEncoder
(VAE) and Boltzmann machine.
In implicitly defined functions, the density function is not mentioned and the
model will internally figure out a way to draw samples from the same distribution
provided. Some implicit PDF generative models are generative stochastic networks
(GSN) and GAN.
The explicitly defined model requires defining the density function which is hard
to be established. Because the density function should be able to model the whole
distribution, defining such a function is not possible. In the explicitly defined models
that have been mentioned, a tractable density function is modelled or approximation
of a density function is utilized. The tractable density function is a density function
designed in such a way that it can model most of the distribution. But, for a com-
plex distribution like distributions in Nature, the distribution of speech, and natural
image, designing a parametric function that can capture the distribution perfectly
is difficult. The problem is that it cannot handle the vast distribution of some data
sets and this limits the usability of these models. The models that approximate the
density function learn the distribution using two approaches. One by approximating
an intractable density function by placing a lower bound in log-likelihood and maxi-
mizing the bound. Another approach uses the Markov chain to use an estimation of
the density function or its gradient. Both of these approaches include approximation
which does not guarantee to capture the complex distribution. These complications
led to the design of implicit density function which is mostly automatically learned
by the model itself. The GAN is the only implicit density function that shows promis-
ing results.
A lot of research has been done to track and improve these promising models.
Most recent researches show results of GAN to be state-of-the-art in generative mod-
elling. GAN has various applications when compared to other types of generative
models.
7.3 GAN ARCHITECTURE
eneral Structure
7.3.1 G
GAN is a combination of two neural networks: the generator model and the discrimi-
nator model. The architecture of vanilla GAN is represented in Figure 7.1.
The generator model tries to generate data similar to the distribution of training
data set and the discriminator network checks whether the generated data is agree-
ably realistic.
In the vanilla GAN which was first introduced by Ian Goodfellow in 2014 [5], a
high-dimensional input noise is generated randomly and given as input to the gen-
erator. Any distribution can be used to generate these random noise values. The
generator network has randomly assigned weights and biases as parameters. Both
the weight and bias are numerical values that are subjected to mathematical opera-
tions along the process. The generator network processes the input noise with its
parameters and produces output with the required dimension of the output image.
These processes depend on the functions that are used in the neural network. The
generated output will be mostly noisy and does not have any similarity with the
distribution of our training data set. The data set can be an image, video, text, audio,
or even 3D objects. Then, the generated output and the real data are passed into the
discriminator. The discriminator classifies the real and fake images after processing
it with its parameters, randomly assigned weights, and biases. From the output of
the discriminator, the generator loss and discriminator loss are calculated and then
backpropagation is performed. It helps the generator and discriminator to update
their randomly assigned weights.
The generation, discrimination, and backpropagation are repeated many times
with new samples of input data; this is referred to as training of GAN. Throughout
the training, the generator model gets better at generating more realistic data and the
discriminator improves in classifying generated and real data [6, 7]. Both the genera-
tor model and the discriminator model converge at a point, after which both do not
improve. Now, the discriminator model is discarded and the generator alone is used
for generating the samples. The discriminator model was used as a tool for improv-
ing the generator and not used again in the generation of samples. However, the dis-
criminator is a good classifier and sometimes used for validation of other models [4].
7.3.2 Adversarial Process
The generator and discriminator networks compete against each other to improve
each other in their respective tasks. This adversarial technique was inspired by game
theory, a mathematical modelling study of the interaction between two rational deci-
sion-makers. The interaction is performed based on the zero-sum game, in which the
participant either wins or loses. Here, when the discriminator successfully identifies
the real and fake data, the generator is penalized and vice versa when it fails to iden-
tify. GAN is mainly based on the Nash equilibrium to balance the generator and the
discriminator. Nash equilibrium is a solution state when many players are playing
a game and each knows the other player’s strategies and makes decisions without
changing his or her strategy. The generator and discriminator know the strategies
of each other but make the best decisions by not changing their strategy, making
the models to compete with each other and achieve the Nash equilibrium state and
converge at a point.
Both the generator and discriminator models are modelled as differentiable func-
tions. A value function is defined using the generator and discriminator functions in
the form of a two-player minimax game and then optimized throughout the training.
If the discriminator is completely optimized before starting to optimize the gen-
erator, then the generator will not get better. So, k-steps of discriminator optimiza-
tions are performed before one step of optimizing the generator. Then, the generator
model tries to increase the chances of the discriminator making mistakes.
Algorithm
Step 1: Iterate through the number of iterations.
Step 2: Iterate through the discriminator for a fixed
number of times.
Step 2a: Sample noise and minibatch of n examples from
given examples.
Step 2b: Calculate the error and gradients.
Step 2c: Update the discriminator weights.
Step 3: Sample minibatch of n noise samples.
Step 4: The generator is provided with the noise sample
and output is generated.
Step 5: Output is sent to the discriminator and loss value
is calculated.
Step 5: The weights of the Generator are updated using the
loss value from Discriminator.
Generative Adversarial Network 137
7.3.3 Background Mathematics
Every data set has a distribution of data. Every model tries to understand the distri-
bution of input data and captures relevant information from the distribution using
probability estimation. The divergence of distribution between two data sets is pro-
vided in Figure 7.2. The generative models capture joint probability and discrimina-
tive models capture conditional probability. The generator and discriminator models
are mathematically differentiable functions G and D, respectively. A value function
is defined using these differentiable functions [5].
min G max D V ( D, G )
( ( ))
V ( D, G ) = E x ~ pdata ( x ) éë log D ( x ) ùû + Ez ~ pz ( z ) é log 1 - D G ( z ) ù
ë û
The whole training process is designed in such a way that D tries to maximize the
probability of labelling the generated and input data correctly. Log D(x) is the loga-
rithmic probability of discriminator predicting the real data label. The generator
( ( ))
tries to minimize log 1- D G ( z ) , the log probability of discriminator predicting
generated images label correctly. E denotes the expectation of a random variable.
In Equation 7.1, there are two parts. The first part handles the real data sample
and the latter part after the plus symbol deals with the generated samples. But, prac-
tically, the equation may not fare well as the discriminator can reject generated sam-
ples easily with high confidence. To change this in the equation, the generator
function is changed. Rather than decreasing the log probability of discriminator
making correct predictions, the generator increases the log-probability of discrimi-
( )
nator making mistakes é log D G ( z ) ù . The value function is optimized by using
ë û
backpropagation. This eliminates the use of a Markov chain which adds a huge
advantage over other generative models.
The value function is based on Jensen–Shannon divergence (JSD), which mea-
sures the similarity between two distributions. JSD is based on Kullback–Leibler
138 Artificial Intelligence (AI)
divergence (KLD). KLD measures the difference of one distribution to another one.
The difference in the distance between the two distributions can be estimated by
KLD. Minimizing KLD can minimize the distance between the distributions. Log-
likelihood is the estimator that helps to minimize KLD.
The GAN internally finds the density function (PDF) of the input data distribu-
tion. A sample from the density function provides a relative likelihood. Likelihood
expresses different plausible parametric values of the sample. The generator of GAN
tries to model data that have a similar distribution of input data using the density
function. At first, the generated distribution will be different from the real input dis-
tribution. This difference is then reduced with minimizing KLD by maximizing the
log-likelihood of the distributions.
7.4 GAN VARIATIONS
7.4.1 Overview
The vanilla GAN invented in 2014 by Ian Goodfellow was popular mainly because
of the adversarial process. Many different variations of GAN which improve its
functionality, accuracy, and applications are proposed over the years. The varia-
tions may possess changes in architecture or formulations. Every variation provides
a variety of applications for GAN. The number of GAN variant papers proposed
has exponentially increased each year. The architectural variation includes changes
by including concepts and components of other deep learning architectures. These
changes can be an increase in the number of generators or discriminators, the way
they are connected, and the introduction of other deep learning components in the
discriminator or generator. The variations are mainly classified into architecture-
based variants and formulation-based variants [8, 9].
7.4.2 Techniques
7.4.2.1 Architecture-based Variant Class
The architectural variant class has two subclasses: condition-based variants and
structure-based variants.
I. Condition-based Variants
Condition-based variants involve conditional input that changes the
output according to it. This class of variants started with the condi-
tional GAN (CGAN) proposed in 2014 [10]. It is based on vanilla GAN
with additional conditional input to generator and discriminator. This
extracts features based on the modelling of the conditional input. The
InfoGAN proposed in 2016 provides the condition as latent information
to the generator. Auxiliary Classifier GAN (ACGAN) is an extension
of CGAN that adds conditional input only to the generator. All these
GANs include modifying the objective function according to the condi-
tion structure.
Generative Adversarial Network 139
modifying the formulation. The original GAN was unstable and hard to train. It had
many issues in training and convergence. Many proposed variants use these alterna-
tive objective functions to solve these issues.
I. Divergence-based Modification
Original GAN used JSD for the objective function. f-GAN and
Wasserstein GAN (WGAN) improved stability of GAN training by chang-
ing JSD to other divergence formulations. f-GAN utilizes f-divergence for
the objective function. WGAN formulated the objective function using
Wasserstein distance. WGAN claims to be more stable than GAN, uses
JSD, and also performs weight clipping to avoid mode collapse, a condition
where data generated by the generator are not diverse or even produces the
same sample for different inputs.
II. Loss Function-based Modifications
In some GANs, loss function is tuned to achieve desired results. Least-
square GAN (LSGAN) uses the least-square loss function and f-GAN uses
gradient-based loss function. Discover cross-domain relationship GAN
(DiscoGAN) [13] attempts to automatically discover relationships and fea-
tures between different domains. This is achieved by modifying the objec-
tive function with reconstruction loss. Cycle consistency GAN (CycleGAN)
[14] introduces an alternative loss function, named cycle-consistency loss.
It helps CycleGAN to minimize the distribution of data between two
domains. StarGAN aims at the multi-domain image-to-image translation.
It uses domain classification loss along with the reconstruction loss.
III. Technique-based Modifications
Another technique to optimize the objective function was performed
in energy-based GAN (EBGAN), where discriminator is formulated as an
energy function. The mode collapse problem was also handled in unrolled
GAN (UGAN) by using unrolled optimization in the objective function.
IV. Regularization
GAN has to converge properly to achieve good results. The generator
and discriminator model training has to be stable to achieve convergence.
The unstable training process could lead to mode collapse. The regulariza-
tion technique is a trick that could lead to the stabilization of the training
process. Many different kinds of regularization are used to solve this prob-
lem. Regularization techniques like mode regularization, Lipschitz regu-
larization, Jacobian regularization, and orthogonal regularization are used
in variants like MRGAN, LSGAN, JRGAN, and BigGAN, respectively
[15]. WGAN-LP uses a penalty function that uses Lipschitz constraint.
V. Normalization
Normalization is another trick performed to scale the values in GAN. It
also aids in the stabilization of GAN training and convergence. The nor-
malization techniques like gradient penalty and spectral normalization are
widely used [15]. The gradient penalty is used in WGAN-GP, DRAGAN,
and BWGAN. WGAN-GP uses the gradient penalty in WGAN. Along with
Generative Adversarial Network 141
the gradient penalty, DRAGAN uses deep regret analytics and Banach
Wasserstein GAN (BWGAN) uses dual normalization in Banach space.
Spectral Normalization is used in Spectral normalization GAN (SNGAN)
and Self-attention GAN (SAGAN) [16]. The attention method is also used
along with spectral normalization in SAGAN.
7.5 APPLICATIONS
7.5.1 Image Generation and Prediction
The creation of visual content remains challenging for humans. The generation of
new images can be useful in many areas where more visual data is needed. More
data can provide more possibilities and opportunities. In the case of DL, huge data
sets are required to train models. These data can be hard to collect and pre-pro-
cess. Data augmentation is a process done to increase the data set by performing
transformations like cropping, scaling, padding, and flipping to the existing images.
The augmented images are extremely useful in creating machine learning models
for medical and geological image classification, where collecting medical images
is hard, and augmented images serve the purpose of training these models. These
problems can be tackled by GAN. They have the potential to create many images
similar to the input image.
All the GAN variants are capable of data augmentation, depending on the type
of data they process. DCGAN [11] and StyleGAN [17] are famous for the genera-
tion of human faces from scratch. Figure 7.3 provides an example of the image
generation. BigGAN [18] is capable of generating photo-realistic images. Images
of objects are also generated which can be used in websites, clothing lines, media,
and many more areas.
7.5.2 Image Translation
Image translation is the process of converting an image into another modified image
or other forms of data like text, audio, etc. Image-to-image translation leads to many
fascinating applications. Image translation includes the conversion of objects from
one domain to objects of another domain. For example, images of a horse can be con-
verted into images of zebra and vice versa. Many neural networks like CycleGAN
[14], Pix2Pix [12], and BigGAN [18] have demonstrated this using various data sets.
The conversion between day and night images, seasons, colourization of black and
white images, and many more have been demonstrated using Pix2Pix. Even alter-
ing human features like hair colour and facial expression has been translated using
StyleGAN [17] and DiscoGAN [13]. An example of converting label image from
CMP facades data set [19] to real-world buildings using Pix2Pix [12] is provided in
Figure 7.4. Even the conversion of faces into cartoon or anime characters has been
demonstrated. Image super-resolution is another application where GAN shines.
Image super-resolution is the problem where fewer dimension images are con-
verted into images with higher resolution. Variants like Super Resolution GAN
(SRGAN) [20] and Enhanced Super Resolution GAN (ESRGAN) [21] have achieved
remarkable results in image super-resolution tasks. Style transfer is another problem
where the style or feature of one image is transferred to the input image. The image
which has to be referred for the features or style is processed as conditional input
and the main input is the image where style has to be applied. Many variants under
conditional GAN class are capable of solving this problem with good accuracy.
Image translation into textual data has been achieved in DCGAN, StackGAN
[22], AttnGAN, GAN-INT-CLS, and OP-GAN. Image inpainting can also be per-
formed in GAN. In some images, certain parts of the pictures are missing or hid-
den by overlapping objects. Image inpainting is a process where these images are
translated into images that are filled with accurate details in missing regions. Image
inpainting can also be used to remove watermarks in images.
7.5.3 Image Editing
Image editing is the technique of manually providing input to the model that alters
the image accordingly. This technique enables us to paint objects and scenes like
trees, architectural structures, sea, and sky into the image. The variants like IGAN,
GP-GAN, and GauGAN are capable of image editing. This is under development
as the technique has a long way to follow, even though the results seem stunning
and natural in some cases. Unlike other models, GauGAN [23] has achieved some
remarkable quality in images. Figure 7.5 describes a made-up example of image
editing using GAN.
7.5.5 Video Manipulation
Video generation has been an interesting area of work for humans. Creative video
content creation is an important profession and many industries have been built upon
this platform. Video content has to be created by using video capture devices or
graphically creating each scene through software. In each case, the creator needs
144 Artificial Intelligence (AI)
plenty of time and resources to complete the task. The generation and manipulation
of video content using GAN can make it easy for the creators.
Video is nothing but a sequence of pictures, the same as images. Video can be
created, manipulated, and predicted. Generation of video can be done using the
GAN models used for image generation and style transfer. Video manipulation has
created a great discussion in society, as there occurs a possibility of changing faces
and expressions in video content. Face swapping GAN (FSGAN) [26] is a recent
addition to the class of GANs that can manipulate videos. Few research papers have
proposed works on video prediction. Video prediction guesses the next frame and it
can evolve machines into a better content creator.
The conversion of audio content to the image has been demonstrated in Wav2Pix
architecture.
BIBLIOGRAPHY
1. Francois, Chollet, Deep Learning with Python Manning Publications, 2017.
2. Goodfellow, Ian, Yoshua Bengio, and Aaron Courville, Deep Learning. MIT Press,
2016.
3. Nielsen, Michael A., Neural Networks and Deep Learning, Vol. 25. Determination
Press, 2015.
4. Langr, Jakub, and Vladimir Bok, GANs in Action Manning Publications, 2018.
146 Artificial Intelligence (AI)
25. Yu, Chong, and Young Wang, “3D-scene-GAN: Three-dimensional scene reconstruc-
tion with generative adversarial networks,” in ICLR Workshop, ICLR, 2018.
26. Nirkin, Yuval, Yosi Keller, and Tal Hassner, “Fsgan: Subject agnostic face swapping
and reenactment,” in Proceedings of the IEEE International Conference on Computer
Vision, IEEE, 7184–7193, 2019.
27. Donahue, Chris, Julian McAuley, and Miller Puckette, “Adversarial audio synthesis,”
arXiv Preprint ArXiv:1802.04208, 2018.
28. Dong, Hao-Wen, et al., “Musegan: Multi-track sequential generative adversarial
networks for symbolic music generation and accompaniment,” in arXiv preprint
arXiv:1709.06298, 2018.
29. Ma, Jiayi, et al., “FusionGAN: A generative adversarial network for infrared and visible
image fusion,” Information Fusion, Elsevier, 48: 11–26, 2019.
30. Schlegl, Thomas, et al., “Unsupervised anomaly detection with generative adversarial
networks to guide marker discovery,” in International Conference on Information
Processing in Medical Imaging. Springer, 146–157, 2017.
31. Armanious, Karim, et al., “Med GAN: Medical image translation using GANs,”
Clinical Orthopaedics and Related Research (CoRR), Elsevier, 79, p.101684, 2018.
8 Detection and
Classification of
Power Quality
Disturbances in Smart
Grids Using Artificial
Intelligence Methods
Gökay Bayrak and Alper Yılmaz
CONTENTS
8.1 Introduction................................................................................................... 149
8.1.1 Signal Processing (SP)-based PQD Detection Methods.................... 151
8.1.2 Artificial Intelligent (AI) Methods for PQD Detection..................... 151
8.2 Wavelet Transform (WT)-based PQD Detection Methods........................... 153
8.2.1 Wavelet Transform (WT)................................................................... 154
8.2.2 Proposed DWT-based PQD Detection Method................................. 155
8.3 AI-based PQD Classification Methods.......................................................... 155
8.3.1 Deep Learning Structures................................................................. 158
8.3.1.1 SAE-based Methods .......................................................... 158
8.3.1.2 DNN-based Methods ......................................................... 160
8.3.1.3 DBN (Deep belief network)-based Methods...................... 160
8.3.1.4 CNN-based Methods.......................................................... 160
8.3.2 Proposed Deep Learning and WT-based Hybrid PQD
Classification Method........................................................................ 161
8.4 Results............................................................................................................ 165
8.5 Conclusion..................................................................................................... 167
References............................................................................................................... 168
8.1 INTRODUCTION
Integration of DG units to the grid improves the distribution system performance,
reliability, stability, security, power quality (PQ), and voltage profile. Also, DG-based
microgrid should have fewer losses, high-degree voltage support, and a reliable
power quality disturbance (PQD) detection system [1]. The interconnection of DG
149
150 Artificial Intelligence (AI)
with a grid may reduce the cost, losses, complexity, a load of lines, interdependence,
and inefficiencies associated with low-carbon emission in microgrids [2].
The DG-based power plants must adapt the defined electrical standards to provide
reliable grid operation and grid stability in cases of any fault or disturbances at the
point of common coupling (PCC). The PQDs [3], islanding detection and protection,
and electromagnetic interference are critical issues to adapt DG systems to the grid.
According to Ref. [4], this fault and disturbance conditions costs in the European
Union (EU) are predicted to be high.
DGs inclusion in the system has a lot of positive effects, but it creates technical
difficulties facing the integration of DG. These difficulties are voltage regulation,
increased distribution losses, harmonic control and harmonic injection problem,
islanding condition control, and the sensitivity of existing protection schemes.
The conventional power system is designed to have a unidirectional flow of power
from the high-voltage to the low-voltage side. Today, power distribution systems
with bidirectional power flows have been created due to grid-integrated DG systems,
which have increased in numbers rapidly. As a result, classical analysis, operation,
and design methods cannot manage this complex grid structure. DG integration into
the main grid is an essential issue if considering this aspect. The DG systems are
expected not to operate out of the defined threshold values of the related standards’
voltage and frequency. In a microgrid, unintentional islanding conditions cause sev-
eral damages for both the system operators and the DG unit side. An incorrect detec-
tion in islanding conditions with conventional methods also leads to switching of
DGs and economic losses. The voltage, frequency, and current balance of DG sys-
tems operating with the grid are also essential parameters for microgrids.
Figure 8.1 shows the general structure of a DG system. There are some technical
electrical criteria to switch DG systems from the grid, and these conditions must be
considered to provide the sustainable operation, reliability, and high-power quality
of the grid. IEEE 929-2000 defines some technical electrical rules for microgrids to
provide a reliable power system [5].
methods for detection of PQDs are fuzzy expert system-based classifiers, neural net-
work (NN)-based classifiers, decision tree (DT) and random forest (RF) classifiers,
support vector machine (SVM)-based classifiers, and deep learning (DL)-based clas-
sifiers. Figure 8.3 shows a classification of AI-based PQD detection methods used
in smart grids. The classifier is an algorithm that takes raw data or extracted futures
data as an input and makes a decision about the normal operating state of the system.
Artificial neural network (ANN) has been used for fault detection in PV systems and
wind turbines in Refs. [18, 19]. A SVM is a classification method that is used suc-
cessfully in many pattern recognition problems. It is also originally a proposed and
defined classifier to distinguish two classes from each other [20]. A multiclass SVM
method that performs multiple PQDs is proposed in Ref. [21]. A DT-based algorithm
developed for the classification of complex PQDs is proposed in Ref. [22]. Adaptive
probabilistic neural networks (APNN) and probabilistic neural networks used the
classification of PQDs [23, 24].
Complex learning networks gradually replace the few layers used in regression
and classification problems with developing information technologies. DL algorithms
can be considered as multilayers of ANNs. In the study where the DL algorithm was
used for the first time in islanding detection is Ref. [25]. In another study [26], a new
approach based on single spectrum analysis, curvelet transform, and deep convolu-
tional neural network (DCNN) has been performed to realize the classification of
PQDs. A novel DL method proposed using raw data for the classification of PQD
Power Quality Disturbances in Smart Grids 153
[27]. This method recognizes automatic feature extraction and selection using any
filter or signal processing method. The proposed algorithm is compared to other DL
network structures – long short-term memory (LSTM) network, ResNet50, stacked
auto-encoder (SAE), and gated recurrent unit network – in this study. Results show
that the proposed DCNN-based method has higher accuracy and less training cost.
In Refs. [28, 29], 1D time series PQD data converted 2D image data and then 2D
convolutional neural network (CNN) performed the classification of PQDs. The data
type for this method is not suitable because PQDs are 1D time series signals. While
there is a significant correlation between the pixels on the horizontal and vertical
axes in the image data, there is only one direction correlation in the PQD signals.
Results are compared to other AI-based networks but noise independence, training
time, and model size of the proposed network are unspecified.
process using a short-time Fourier transform (STFT) [31]. STFT has a fixed window
size, and it does not contain decomposed low- and high-frequency components simul-
taneously [32, 33]. WT-based methods have flexible time-frequency representation,
locally decompose discontinuities, and sudden changes in signal in high-grade deriva-
tives where other signal processing methods fail to detect PQDs.
1 æ t -b ö
y s ,b = yç ÷ (8.1)
s è s ø
¥
1 æ t -b ö
CWT ( s, b ) =
s ò f ( t )y çè
-¥
s ÷ø
dt (8.2)
CWT method is used for limited real-time processes because of intensive computa-
tion. DWT is designed by using filter banks to remove the disadvantages caused by
the computational load in CWT. DWT is performed to signal as in Equation 8.4,
using g j,k(n) main wavelet function (Equation 8.3):
DWT ( j, k ) = ååS ( n ) g
nÎZ kÎZ
j, k (n), g j , k Î Z , j Î N , k Î Z (8.3)
-j
(
g j ,k ( n ) = a0 2 g a0 j n - kb0 (8.4) )
The relation between main wavelet function and S(n) can be expressed as in Equation 8.5:
S (n) = ååd
nÎZ kÎZ
j ,k g j ,k ( n ) (8.5)
The DWT applied signal is observed approximation coefficients (cAn) and detail
coefficients (cDn). Equation 8.6 gives the formula of the wavelet coefficients:
cA éë n ùû = åS ( k ) g éë2n - k ùû ;
k
cD éë n ùû = åS ( k ) h éë2n - k ùû (8.6)
k
Power Quality Disturbances in Smart Grids 155
The UWT method uses filter banks like DWT, but the coefficient lengths are the
same as the original signal [17]. Detail and approximation coefficients are upsam-
pled using multi-resolution analysis (MRA) by the “á trous” algorithm in the UWT
method [35]. UWT can reduce oscillations and noise, and it is useful for selecting
threshold values rather than DWT [36]. UWT detail and approach coefficients can
be calculated by using Equation 8.7.
( ) åc
cUj éë k ùû = h0( ) * cUj -1 éë k ùû =
j U
j -1 é k + 2 j n]h0 ( k ) ù
ë û
n
( ) åc
wUj éë k ùû = h0( ) * cUj -1 éë k ùû =
j U
j -1 é k + 2 j n]h1 ( k ) ù (8.7)
ë û
n
FIGURE 8.4 MATLAB/Simulink model of developed electrical power distribution system [30].
Artificial Intelligence (AI)
Power Quality Disturbances in Smart Grids 157
TABLE 8.1
PQD Events Obtained from Developed Model
PQD Event Causes of Event
Voltage sag Short-circuit faults, switching on a large load
Voltage swell Short-circuit faults, switching of large capacitors
Voltage interruption Short-circuit faults, component failures, CB tripping
Harmonics Non-linear power electronic-based drivers, arc welders
Oscillatory transient Switching of large capacitor bank and inductive loads
Notch Non-linear power electronic-based drivers
In the training phase with ANN, PQD signals or extracted futures are given as
input to the system, and the weight coefficients of each node in the model are deter-
mined. In the test phase, the unknown signal is provided as an input to the trained
ANN system, and the class of test signal is decided according to the value calculated
in the output node. As a result of each output layer, the probability of input signal
belonging to that class is obtained.
SVM classifier, which is a classification method used successfully in many pat-
tern recognition problems, was originally a proposed and defined classifier to distin-
guish the two classes. Multiple classifications are performed with the help of kernel
function. The purpose of kernel functions is to move the features that cannot be
separated linearly in the existing space into a higher dimensional space and to make
the features linearly separable in this new high dimensional space. Decision tree
(DT) is a supervised learning method, and it has tree-shaped architecture. DT estab-
lishes the relationship between the class and the attributes. Random forest (RF) is
a further supervised learning method, and it has multiple tree-shaped architectures.
RF classifier improves the generalization performance of DT.
FIGURE 8.6 PQD detection using the proposed DWT-based method. (a) Voltage sag. (b)
Voltage swell. (c) Voltage interruption. (d) Oscillatory transient.
8.3.1.1 SAE-based Methods
AE structures provide better results than other AI methods in extracting features
from data, removing unnecessary information, and compressing useful information.
AE consists of an encoder and decoder network. Represented features are defined as
Power Quality Disturbances in Smart Grids 159
Equation 8.8 using the SAE structure where σ is the activation function and θ is the
training parameters:
( )
hi = f0 ( xi ) = s f wT × xi + b (8.8)
( )
x̂i = gq ¢ (hi ) = s g w¢T × hi + b¢ (8.9)
8.3.1.2 DNN-based Methods
DNN algorithms can be considered as multilayers of ANNs.
8.3.1.4 CNN-based Methods
CNN is capable of capturing the shift-variant properties of input data [38]. CNN-
based PQD classification methods are learned features from the raw data without
using any signal processing methods compared to other DL structures. Different
Power Quality Disturbances in Smart Grids 161
from traditional methods, CNN methods were performed for detection and classifi-
cation of PQDs without using an additional method of pre-processing, feature extrac-
tion, and classification. Besides, the number of training parameters in classification
algorithms is reduced using CNN.
The signal processing and classification-based hybrid method and DCNN-based
method are shown in Figure 8.7. DCNN has stacked units to extract features. Stacked
units consist of pooling layers, convolutional layers, and batch-normalization layers.
FIGURE 8.7 (a) Signal processing and classification-based hybrid method. (b) DCNN-
based PQD classification method [27].
the multilayer FDNN. At the last stage, PQDs were classified using the softmax
activation function. Figure 8.8 shows the flow diagram of the proposed method. The
classified power quality events by proposed FDNN-based method are as follows:
Most of the available literature has gone into using a mathematical model or simula-
tion to build data sets due to the difficulty of obtaining real signals related to PQD. In
this study, the integral-based mathematical model was used to generate PQD signals
Power Quality Disturbances in Smart Grids 163
[39, 40]. Train, validation, or test data sets can be generated randomly by the soft-
ware created based on the equations specified in PQD parameters, possible values in
Table 8.2. Mathematical definitions of five of the most common PQDs mentioned in
IEEE standards are given in Table 8.2. Other PQDs were generated with the help of
an integral-based mathematical model.
A five-level decomposition process was performed using the high-pass and low-
pass filters with daubachies4 mother wavelet, and then the energy values of the detail
164
TABLE 8.2
Mathematical Model of PQDs [39]
Class PQD Signals Equations Threshold Parameters
C1 Nominal 49.8 £ f £ 50.2, -p £ j £ p
x ( t ) = sin ( 2p ft - j )
C2 Sag T £ t2 - t1 £ 9T
x ( t ) = é1 - a u ( t - t1 ) - u ( t - t2 ) ù sin ( 2p ft - j )
( )
ë û
0.1 £ a < 0.9
C3 Swell T £ t2 - t1 £ 9T
x ( t ) = é1 + b u ( t - t1 ) - u ( t - t2 ) ù sin ( 2p ft - j )
( )
ë û
0.1 £ b < 0.8
C5 Transient 300 Hz £ fn £ 900 Hz;
x ( t ) = sin ( 2p ft - j ) + b exp - ( t - t1 ) / t sin 2p fn ( t - t1 ) - J ((u ( t - t2 ) - (u ( t - t I ) )
( ( ) )
8 ms £ t £ 40 ms; - p £ J £ p x
0.5T £ t2 - t1 £ 3T
C7 Harmonics 7
n = {3, 5, 7} ; 0.05 £ a n £ 0.15
x ( t ) = sin ( 2p ft - j ) + n
åa sin ( np ft - J ) n
n =3
-p £ Jn ,Jn¢ ,Jn¢¢ £ p
n¢ = {3, 5, 7} ; 0.05 £ a n¢ £ 0.15
n¢¢ = {1, 3, 5}
a n¢¢ = 1 n¢¢ = 1; 0.05 £ a n¢¢ £ 0.15 n¢¢ = {3, 5}
Artificial Intelligence (AI)
Power Quality Disturbances in Smart Grids 165
Ej = åD
n =1
jn j = 1,¼, l (8.10)
8.4 RESULTS
The performance scores of FDNN are given in Table 8.3. FDNN classifier has the
best score with a training accuracy (Train Acc) of 100% and with a test accuracy
TABLE 8.3
Detailed Performance Report for FDNN Classifier Test Set
Classes Precision Recall F1 Score Support
C1 95.17% 99.11% 97.10% 338
C2 100% 97.20% 98.76% 321
C3 99.64% 97.89% 98.55% 285
C4 99.00% 100% 99.50% 298
C5 100% 100% 100% 301
C6 100% 100% 100% 298
C7 99.66% 99.66% 99.66% 292
C8 99.06% 98.13% 98.59% 321
C9 99.64% 99.64% 99.64% 290
Accuracy 99.09% 2744
Macro avg 99.17% 99.11% 99.13% 2744
166 Artificial Intelligence (AI)
TABLE 8.4
Performance Comparison of Existing Methods
Method No. of Futures Number of PQDs Accuracy
Reference [23] 5 11 98.2%
Reference [41] 5 14 97.28%
Reference [42] 19 9 99.66%
Reference [43] 10 5 98.51%
Reference [44] 45 22 99.09%
Reference [45] 2 10 94.2%
Reference [21] 6 9 97.22%
Proposed method 5 9 99.13%
(Test Acc) of 99.13%. Training and test set were randomly distributed, and the test
set was determined to be approximately 30%.
The training lasted 200 Epoch – about 608 sec. Learning curves are shown in
Figure 8.10.
As shown in Table 8.4, the PQDs classification accuracy corresponding to the WT
and FDNN-based algorithm is higher than all the other investigated methods.
8.5 CONCLUSION
This chapter presents SP-based and AI-based PQD detection and classification meth-
ods used in smart grids. In Section 8.1, SP-based PQD detection methods used in
smart grids are classified, and then wavelet transform (WT)-based detection meth-
ods are presented. Current WT-based methods were introduced for the detection
of PQDs. Some PQD events are investigated in MATLAB Simulink by using the
developed discrete WT method.
The classification of PQDs with AI-based methods are also discussed in
the last section. Deep neural networks are presented in the chapter to show the
effectiveness of AI-based methods. A hybrid method based on DWT and feed-
forward deep neural network (FDNN) is proposed for the classification of PQDs
occurring in grid-connected PV systems. In the developed process, the nominal
and PQD signals were subjected to five-level DWT decomposition, and then the
energy values of the detail coefficients were calculated. The calculated energy
values constitute the inputs of the multilayer FDNN. At the last stage, PQDs
were classified using the softmax activation function. The results show that the
proposed hybrid WT and FDNN methods have 99.13% accuracy in classifying
PQDs.
The proposed detection and classification methods for detecting PQDs are suit-
able for smart grids, and the obtained results in this chapter will contribute to select-
ing the proper detection or classification methods of PQDs using AI-based methods
in smart grids.
168 Artificial Intelligence (AI)
REFERENCES
1. Biswas, S., Goswami, S. K., & Chatterjee, A. (2012). Optimum distributed generation
placement with voltage sag effect minimization. Energy Conversion and Management,
53(1), 163–174.
2. Hung, D. Q., Mithulananthan, N., & Bansal, R. C. (2010). Analytical expressions for DG
allocation in primary distribution networks. IEEE Transactions on Energy Conversion,
25(3), 814–820.
3. Bollen, Math H. J. (2000). Understanding power quality problems. In Voltage Sags and
Interruptions. IEEE Press.
4. Chapman, David. (2001). The cost of poor power quality. Power Quality Application
Guide 1–4.
5. IEEE. Std. 929 (2000). IEEE Recommended Practice for Utility Interface of Photovoltaic
(PV) Systems. Institute of Electrical and Electronics Engineers, Inc., New York.
6. F II I (1993). IEEE Recommended Practices and Requirements for Harmonic Control
in Electrical Power Systems. New York.
7. IEEE (2014). Std. IEEE Recommended Practices and Requirements for Harmonic
Control in Electrical Power Systems, 519.
8. Chen, C. I., & Chen, Y. C. (2014). Intelligent identification of voltage variation
events based on IEEE std 1159–2009 for SCADA of distributed energy system. IEEE
Transactions on Industrial Electronics, 62(4), 2604–2611.
9. Mahela, O. P., Shaik, A. G., & Gupta, N. (2015). A critical review of detection and
classification of power quality events. Renewable and Sustainable Energy Reviews, 41,
495–505.
10. Khokhar, S., Zin, A. A. B. M., Mokhtar, A. S. B., & Pesaran, M. (2015). A comprehen-
sive overview on signal processing and artificial intelligence techniques applications
in classification of power quality disturbances. Renewable and Sustainable Energy
Reviews, 51, 1650–1663.
11. Latran, M. B., & Teke, A. (2015). A novel wavelet transform based voltage sag/swell detec-
tion algorithm. International Journal of Electrical Power and Energy Systems, 71, 131–139.
12. Karegar, H. K., & Sobhani, B. (2012). Wavelet transform method for islanding detec-
tion of wind turbines. Renewable Energy, 38(1), 94–106.
13. Mohammadnian, Y., Amraee, T., & Soroudi, A. (2019). Fault detection in distribution
networks in presence of distributed generations using a data mining–driven wavelet
transform. IET Smart Grid, 2(2), 163–171.
14. Bayrak, G. (2018). Wavelet transform-based fault detection method for hydrogen
energy-based distributed generators. International Journal of Hydrogen Energy,
43(44), 20293–20308.
15. Deokar, S. A., & Waghmare, L. M. (2014). Integrated DWT–FFT approach for detection
and classification of power quality disturbances. International Journal of Electrical
Power and Energy Systems, 61, 594–605.
16. Lopez-Ramirez, M., Cabal-Yepez, E., Ledesma-Carrillo, L. M., Miranda-Vidales,
H., Rodriguez-Donate, C., & Lizarraga-Morales, R. A. (2018). FPGA-based online
PQD detection and classification through DWT, mathematical morphology, and SVD.
Energies, 11(4), 769.
17. Yılmaz, A., & Bayrak, G. (2019). A real-time UWT-based intelligent fault detection
method for PV-based microgrids. Electric Power Systems Research, 177. http://www
.ncbi.nlm.nih.gov/pubmed/105984.
18. Dhimish, M., Holmes, V., Mehrdadi, B., & Dales, M. (2018). Comparing Mamdani
Sugeno fuzzy logic and RBF ANN network for PV fault detection. Renewable Energy,
117, 257–274.
Power Quality Disturbances in Smart Grids 169
19. Bangalore, P., & Patriksson, M. (2018). Analysis of SCADA data for early fault detec-
tion, with application to the maintenance management of wind turbines. Renewable
Energy, 115, 521–532.
20. Zhang, L., Zhou, W., & Jiao, L. (2004). Wavelet support vector machine. IEEE
Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 34(1), 34–39.
21. Thirumala, K., Pal, S., Jain, T., & Umarikar, A. C. (2019). A classification method for
multiple power quality disturbances using EWT based adaptive filtering and multiclass
SVM. Neurocomputing, 334, 265–274.
22. Puliyadi Kubendran, A. K., & Loganathan, A. K. (2017). Detection and classification of
complex power quality disturbances using S‐transform amplitude matrix–based deci-
sion tree for different noise levels. International Transactions on Electrical Energy
Systems, 27(4), e2286.
23. Lee, C. Y., & Shen, Y. X. (2011). Optimal feature selection for power-quality distur-
bances classification. IEEE Transactions on Power Delivery, 26(4), 2342–2351.
24. Khokhar, S., Zin, A. A. M., Memon, A. P., & Mokhtar, A. S. (2017). A new optimal fea-
ture selection algorithm for classification of power quality disturbances using discrete
wavelet transform and probabilistic neural network. Measurement, 95, 246–259.
25. Kong, X., Xu, X., Yan, Z., Chen, S., Yang, H., & Han, D. (2018). Deep learning
hybrid method for islanding detection in distributed generation. Applied Energy, 210,
776–785.
26. Liu, H., Hussain, F., Shen, Y., Arif, S., Nazir, A., & Abubakar, M. (2018). Complex
power quality disturbances classification via curvelet transform and deep learning.
Electric Power Systems Research, 163, 1–9.
27. Wang, S., & Chen, H. (2019). A novel deep learning method for the classification of
power quality disturbances using deep convolutional neural network. Applied Energy,
235, 1126–1140.
28. Balouji, E., & Salor, O. (2017). Classification of power quality events using deep learn-
ing on event images. In 2017 3rd International Conference on Pattern Recognition and
Image Analysis (IPRIA), 216–221. IEEE.
29. Mohan, N., Soman, K. P., & Vinayakumar, R. (2017). Deep power: Deep learning
architectures for power quality disturbances classification. In 2017 International
Conference on Technological Advancements in Power and Energy (TAP Energy), 1–6.
IEEE.
30. Bayrak, G., & Yilmaz, A. (2019). Assessment of power quality disturbances for grid
integration of PV power plants. Sakarya University Journal of Science, 23(1), 1–1.
31. Heydt, G. T., Fjeld, P. S., Liu, C. C., Pierce, D., Tu, L., & Hensley, G. (1999). Applications
of the windowed FFT to electric power quality assessment. IEEE Transactions on
Power Delivery, 14(4), 1411–1416.
32. Cohen, L. (1995). Time-Frequency Analysis. Prentice-Hall.
33. Samantaray, S. R., Samui, A., & Babu, B. C. (2011). Time-frequency transform-based
islanding detection in distributed generation. IET Renewable Power Generation, 5(6),
431–438.
34. Strang, G., & Nguyen, T. (1996). Wavelets and Filter Banks. Wellesley-Cambridge
Press.
35. Shensa, M. J. (1992). The discrete wavelet transform: Wedding the a trous and Mallat
algorithms. IEEE Transactions on Signal Processing, 40(10), 2464–2482.
36. Zafar, T., & Morsi, W. G. (2013). Power quality and the un-decimated wavelet trans-
form: An analytic approach for time-varying disturbances. Electric Power Systems
Research, 96, 201–210.
37. Hinton, G. E., & Salakhutdinov, R. R. (2006). Reducing the dimensionality of data with
neural networks. Science, 313(5786), 504–507.
170 Artificial Intelligence (AI)
38. Lei, Y., Yang, B., Jiang, X., Jia, F., Li, N., & Nandi, A. K. (2020). Applications of
machine learning to machine fault diagnosis: A review and roadmap. Mechanical
Systems and Signal Processing, 138. http://www.ncbi.nlm.nih.gov/pubmed/106587.
39. Igual, R., Medrano, C., Arcega, F. J., & Mantescu, G. (2018). Integral mathemati-
cal model of power quality disturbances. In 2018 18th International Conference on
Harmonics and Quality of Power (ICHQP), 1–6. IEEE.
40. Igual, R., Medrano, C., Arcega, F. J., & Mantescu, G. (2017). Mathematical Model of
Power Quality Disturbances. Mendeley Data, 1.
41. Thirumala, K., Prasad, M. S., Jain, T., & Umarikar, A. C. (2016). Tunable-Q wavelet
transform and dual multiclass SVM for online automatic detection of power quality
disturbances. IEEE Transactions on Smart Grid, 9(4), 3018–3028.
42. Abdoos, A. A., Mianaei, P. K., & Ghadikolaei, M. R. (2016). Combined VMD-SVM
based feature selection method for classification of power quality events. Applied Soft
Computing, 38, 637–646.
43. Erişti, H., & Demir, Y. (2012). Automatic classification of power quality events and
disturbances using wavelet transform and support vector machines. IET Generation,
Transmission and Distribution, 6(10), 968–976.
44. Dalai, S., Dey, D., Chatterjee, B., Chakravorti, S., & Bhattacharya, K. (2014). Cross-
spectrum analysis-based scheme for multiple power quality disturbance sensing device.
IEEE Sensors Journal, 15(7), 3989–3997.
45. Borrás, M. D., Bravo, J. C., & Montaño, J. C. (2016). Disturbance ratio for optimal
multi-event classification in power distribution networks. IEEE Transactions on
Industrial Electronics, 63(5), 3117–3124.
9 Robust Design of
Artificial Neural Network
Methodology to Solve
the Inverse Kinematics of
a Manipulator of 6 DOF
Ma. del Rosario Martínez-Blanco,
Teodoro Ibarra-Pérez, Fernando Olivera-Domingo,
and José Manuel Ortiz-Rodríguez
CONTENTS
9.1 Introduction................................................................................................... 171
9.1.1 Kinematics of Robotic Manipulators................................................. 173
9.1.2 Artificial Neural Networks................................................................ 174
9.1.3 Inverse Kinematics Solution with Artificial Neural Networks......... 178
9.1.4 Robust Design of Artificial Neural Networks................................... 179
9.2 Robust Design of Artificial Neural Networks Methodology......................... 180
9.3 Kinematics Analysis of Robotic Manipulator Called Ketzal........................ 184
9.3.1 Data Set Description.......................................................................... 189
9.3.2 Description of Reduction Data Filter Algorithm............................... 191
9.3.3 Data Set Analysis of Training and Test............................................. 194
9.3.4 Planning and Experimentation Stage................................................ 195
9.3.5 Analysis and Confirmation Stage...................................................... 196
9.4 Conclusions and Discussions......................................................................... 203
Future Scope...........................................................................................................205
Acknowledgements................................................................................................. 205
References............................................................................................................... 205
9.1 INTRODUCTION
With recent advances in electronics, mechanics, computer science, and robotics,
robotic manipulators have become the main focus of interest for the development
of industrial applications [1]. Among the multiple articular configurations available,
generally, the manipulators that have more degrees of freedom (DOF) offer greater
171
172 Artificial Intelligence (AI)
control flexibility for complicated tasks; so the scientific community has taken an
interest in this type of robotic manipulators during the last decades [2–4].
During the last years, several investigations have been carried out in the field of
robotics due to its high impact in various areas such as space exploration [5], indus-
trial [6], military [7], and medical applications [8], among others. The use of these
manipulators has contributed considerably to the development and application of
new methods and technologies applied to the field of robotics, mainly in the control
of movement in real time [9].
In the field of robotics manipulators, the main challenge of motion control is to
find a precise and reliable solution for inverse kinematics. The calculation necessary
for the resolution of inverse kinematics requires expensive processing and is compu-
tationally complex [10].
In general, the solutions of the inverse kinematics of a robotic manipulator
are based on three methods: the geometric, iterative, and analytical or algebraic
methods. The singularities and uncertainties, or configurations where manipulator
mobility is reduced, are the main complications in the aforementioned methods for
the control of movement in robotic manipulators. For example, for the algebraic
method, closed-form solutions are not guaranteed and such solutions must exist
for the first three joints if the geometric method is used. In the case of an itera-
tive method, it converges to a single solution that depends on the initial point; so
it requires high-performance hardware without guaranteeing the accuracy of the
calculations [3, 11].
Therefore, the traditional methods used in inverse kinematics problems with geo-
metric, iterative, and algebraic approaches are sometimes complex and generally
unsuitable for multiple joint configurations that can present robotic manipulators
nowadays. As a consequence, various approaches based on artificial neural networks
(ANNs) have been proposed due to the great advantages they present in their parallel
distribution, non-linear mapping, the ability to learn through examples, and the high
performance in the ability to generalization, among others [4, 12].
Generally speaking, the application of ANNs has reduced the error in con-
trolling movement in real time, allowing increasing accuracy [8]. A well-trained
neural network can work with very fast response times, so it is ideal for real-time
applications, compared to other conventional methods where the response time
is longer [13].
Most of the research in ANNs has focused on specific applications of model
development and training algorithms to improve the convergence and accuracy of
the results obtained; however, finding the parameters to achieve adequate learning in
the training of neural networks remains a difficult and complex task [9, 10, 14–17].
In general, the quality in the development and training of ANNs is highly reliable;
however, structural parameters, just like the number of hidden layers and neurons
per hidden layer, represent an important role in the accuracy of the expected results
and these are usually proposed due to the previous experience of the researcher in
trial and error procedures, consuming time and resources without guaranteeing that
the optimal configuration of the parameters to achieve a better performance in the
neural network is obtained [3, 10, 18–21].
Robust Design of Artificial Neural Networks 173
In this work, the use of a robust design methodology based on the fractional fac-
torial design of experiments is proposed to get the optimal parameters of an ANN
architecture for the calculation of the robotic manipulator of six DOF.
The chapter is organized as follows: First, the introduction is addressed, where
the bases of the kinematics of robotic manipulators are presented, as well as those
of the artificial neural networks, to approach the solution of the inverse kinemat-
ics solution with artificial neural networks. The robust design of artificial neural
networks applied in this chapter is also addressed in the introduction. In the second
section, the proposed methodology and its different phases are presented. The third
section describes Ketzal – the robotic manipulator used, performing a kinematic
analysis of it, which allows obtaining a data set to be used in training. The data set
description and the reduction data filter algorithm used are also presented, followed
by data set analysis of training and test. Finally, the different phases of the proposed
methodology are addressed in Planning and Experimentation Stage and Analysis
and Confirmation Stage subsections. In the fourth section, the conclusions are ana-
lyzed and finally, in the fifth section, the future work is presented.
and orientation of its end, so the complexity in the solution to solve this problem is
mainly due to the manipulator geometry, being able to obtain multiple solutions and
singularities that do not always correspond to physical solutions [12, 24].
In this work, the direct kinematics model applied to a six DOF manipulator was
developed, obtaining a set of data on position, orientation, and joint values cor-
responding to the robotic manipulator’s workspace. A filter was used to decrease
the volume of the data and the methodology of Robust Design of Artificial Neural
Networks (RDANN) was applied to obtain the optimal parameters of an ANN archi-
tecture to solve the problem of inverse kinematics for the proposed manipulator.
signals captured by the dendrite that come from another neuron, where the weights
Wi represent the intensity of the synapse that connects two neurons that may be posi-
tive (excitatory) or negative (inhibitory). θ is the transfer function or threshold that
the neuron must overcome to enter the activation state. Finally, the accumulated sum
in the node from all the input signals and multiplied by the synaptic weights go to the
output signal through the transfer function or activation function [28].
In essence, an artificial neuron mimics the behaviour of a biological neuron, that
is, it receives varying signals from other neighbouring neurons and can easily pro-
cess the information from the other neurons to let it pass to another neuron with a
certain level of energy [29].
The signals that enter the neuron are weighted by the multiplication of weights,
which represents a process similar to the strength of the synaptic connection in a
biological neuron; the weighting is added by a neural node and the output is calcu-
lated as the sum of the weighted inputs plus a value b called bias. The output of the
neuronal node is used as input to a transfer function that responds to the artificial
neuron [30]. The output signal of a neuron net can be represented as described below.
Figure 9.3 shows the simplification of an artificial neuron model, where it can be
observed that the inputs to a neuron, xj (j = 1, 2, …, n), are represented by the input
array component, the synaptic weights are represented by the matrix Wi,j (j = 1, 2, …,
n) and b represents the bias. The response of the artificial neuron to the input signals
can be described mathematically as follows:
æ n
ö
yi = f (net ) = f ç
ç
è
å
i =1
x j wij + b ÷ (9.1)
÷
ø
where yi is the value obtained at the end of the artificial neuron and f is the transfer
function used. Usually, a neuron can have one or more inputs.
A model of neurons with multiple inputs is shown in Figure 9.4. In this diagram,
the inputs x1, x2, …, xi are shown, which are multiplied by the corresponding weights
w1,1, w1,2, …, wi,j belonging to the synaptic weight matrix. The subscripts of the
matrix represent the neurons involved in the link, where the first subscript represents
the target neuron and the second represents the source of the input signal to the target
neuron [14]. For example, w2,5 indicates that this weight is the connection from the
fifth input and the second neuron.
To obtain the optimal factors, for the NTB case, Taguchi developed a
two-phase optimization procedure. For both other cases, Taguchi recom-
mends the direct minimization of the expected loss.
In dynamic characteristics the relationship
is used to evaluate the S/N ratio, and MSE represents the average of the
squared errors, that is, the difference between the measured response and
what is estimated.
Confirmation test in optimal conditions. The main objectives in the
4.
design of parameters are to reduce the design variation of the processes or
products to different environments or conditions. At this stage, a confirma-
tion experiment is performed using the optimal design conditions.
Currently, the ANNs can be trained to solve complex and difficult prob-
lems for conventional computers and humans, including various fields such
as pattern recognition and speech, vision, or control systems. In the field of
robotics, ANNs have been successfully applied in recent decades [13, 48],
mainly in the solution to the problem of the inverse kinematics in robotic
manipulators with a high number of joints [1–3, 49].
for the multiple articular configurations that robotic manipulators may present due to
the high degree of precision required in the manipulator positioning [50].
In geometric and algebraic methods, a closed solution can only be guaranteed as
long as the manipulator meets some special conditions, such as the spherical wrist
[51]. For their part, iterative methods tend to be too slow and do not meet the require-
ment for implementation in real-time control systems [49].
This problem has been addressed by researchers during the last decades, who
have proposed various methods of solution through the use of ANNs, so they con-
stitute one of the most appropriate methods in solving this problem. An important
problem in the calculation of the inverse kinematics using ANNs is the determina-
tion of the parameters of the network. The choice of basic ANN parameters deter-
mines success in ANN training [9, 15, 52].
However, there are no rules to identify the optimal selection of these parameters.
Generally, the traditional trial and error technique produces a low capacity for gen-
eralization and low performance, consuming a great deal of time and computational
resources [3].
To overcome this disadvantage easily and efficiently, the RDANN methodology is
used. This methodology, proposed in Ref. [14] and shown in Figure 9.7, describes a
systematic process to find the optimal parameters of an ANNBP architecture applied
to the inverse kinematics solution of a six DOF robotic manipulator.
According to Figure 9.8, the steps to obtain the optimal ANN parameters are
described below:
Planning stage
1.
In this stage, the objective function and the design and noise variables
are identified.
(a) The objective function. The requirements of the problem allow defin-
ing the objective function. The objective function in this research is the
prediction or classification of errors between the expected values and
the output values of the ANNBP in the testing stage. The performance
or value of the mean square quadratic error (MSE) of the ANNBP is
used to achieve this goal, which is expressed mathematically as follows:
N
å (q )
1 2
MSE = i
predicted
- qiDeduced (9.3)
N i =1
TABLE 9.1
Design Variables and Their Levels
Design Variables Level 1 Level 2 Level 3
A L1 L2 L3
B L1 L2 L3
C L1 L2 L3
D L1 L2 L3
Robust Design of Artificial Neural Networks 183
TABLE 9.2
Noise Variables and Their Levels
Noise Variables Level 1 Level 2
U Set 1 Set 2
V 9:1 8:2
W Tr-1/Ts-1 Tr-2/Tst-2
TABLE 9.3
OA with L9(34) and L4(32) Configuration
Trial No. A B C D S1 S2 S3 S4 Mean S/N
1 1 1 1 1
2 1 2 2 2
3 1 3 3 3
4 2 1 2 3
5 2 2 3 1
6 2 3 1 2
7 3 1 3 2
8 3 2 1 3
9 3 3 2 1
184 Artificial Intelligence (AI)
Analysis Stage
3.
The S/N ratio considers the mean and the variation in the measured
responses allowing a quantitative evaluation of the design parameters. The
unit of measure is a decibel and the formula used is
where MSD is the measure of the mean square deviation in the perfor-
mance of the neural network.
Since it is desirable to have a higher signal and less noise, the highest S/N
ratio indicates the best design parameters. A statistical program (JMP) was
used at this stage to determine the optimal parameters of the network.
Confirmation Stage
4.
The robust measurement value is predicted by the optimal design condi-
tion. A confirmation experiment is performed by calculating the perfor-
mance robustness measurement and verifying if the value of the robust
measurement is close to the value deduced.
Once the aspects involved in each of the stages of the RDANN methodology were
identified, the training data set regarding the position, orientation, and articular val-
ues of the robotic manipulator was generated, based on the analysis of the direct
kinematics of the six DOF robotic manipulator, to subsequently perform the 36
workouts described in the proposed OA.
é nx ox ax px ù
ê ú
éR 0
6 P ù ê ny
6
0
oy ay py ú
T60 = ê ú=ê (9.5)
ë0 1û nz oz az pz ú
ê ú
êë 0 0 0 1 úû
where R60 is the 3 × 3 rotation matrix which contains the orientation array [n o a] of
the final effector, and P60 is the position array [p] of the final effector in the reference
coordinate system.
The Denavit–Hartenberg (DH) method was used to analyze the direct kinematics
of the Ketzal robot by implementing four basic transformations that depend exclu-
sively on the geometric characteristics of the links [55]. The DH parameters of the
Ketzal robot are shown in Table 9.4.
These transformations consist of a succession of rotations and translations that
allow the reference system of element i to be related to the system of element i − 1,
where the homogeneous transformation matrix for each joint is given by the follow-
ing equation [10].
i -1
Ai = Rot z,q iTrans x ,diTrans x ,a i Rot z,a i (9.6)
Thus,
éci -si ca i si sa i ai ci ù
ê ú
si ci ca i -ci sa i ai si ú
i -1
Ai = ê (9.7)
ê0 sa i ca i di ú
ê ú
ë0 0 0 1 û
TABLE 9.4
DH Parameters of the Ketzal Robot
Link Offset (cm) Joint Angle (rad) Link Length (cm) Twist Angle (rad)
d1 = 20.2 q1 = q1 a1 = 0 p
a1 =
2
d2 = 0 q 2 = q2 a2 = 16 a2 = 0
d3 = 0 p a3 = 0 p
q3 = q3 + a3 =
2 2
d4 = 19.5 q 4 = q4 a4 = 0 p
a4 = -
2
d5 = 0 q 5 = q5 a5 = 0 p
a5 =
2
d6 = 6.715 q6 = q6 a6 = 0 a6 = 0
186 Artificial Intelligence (AI)
where i is the number of the link, αi is the joint rotation, ai is the length of the link, di
is the displacement of the link, ci = cos(ϴi), and si = sin(ϴi); therefore, ϴi is the rota-
tion angle of the joint.
The matrix that describes the location of the final system with respect to a refer-
ence system located at the base of the robot, also known as the homogeneous trans-
formation matrix, is obtained by making the product of the six matrices obtained
from (9.5) [10].
T60 = 0 A1 × 1 A2 × 2 A3 × 3 A4 × 4 A5 × 5 A6 (9.8)
éc1 0 s1 0 ù
ê ú
s1 0 -c1 0 ú
0
A1 = ê (9.9)
ê0 1 0 a0 + a1 ú
ê ú
ë0 0 0 1 û
éc2 -s2 0 a2 * c2 ù
ê ú
s2 c2 0 a2 * s2 ú
1
A2 = ê (9.10)
ê0 0 1 0 ú
ê ú
ë0 0 0 1 û
é -s3 0 c3 0ù
ê ú
c3 0 s3 0ú
2
A3 = ê (9.11)
ê 0 1 0 0ú
ê ú
ë 0 0 0 1û
é c4 0 - s4 0 ù
ê ú
s4 0 c4 0 ú
3
A4 = ê (9.12)
ê0 -1 0 a3 + a4 ú
ê ú
ë0 0 0 1 û
éc5 0 s5 0ù
ê ú
s5 0 -c5 0ú
4
A5 = ê (9.13)
ê0 1 0 0ú
ê ú
ë0 0 0 1û
éc6 -s1 0 0 ù
ê ú
s6 c6 0 0 ú
5
A6 = ê (9.14)
ê0 0 1 a5 + a6 ú
ê ú
ë0 0 0 1 û
Robust Design of Artificial Neural Networks 187
é nx ox ax px ù
ê ú
ny oy ay py ú
T06 = ê (9.15)
ê nz oz az pz ú
ê ú
ë0 0 0 1û
where
nz = c2 × (c3 × (c4 × c5 × c6 - s4 × s6 ) - c6 × s3 × s5 )
(9.18)
- (c3 × c6 × s5 + (c4 × c5 × c6 - s4 × s6 ) × s3 ) × s2
é0 0 1 42.215ù
ê ú
0 -1 0 0 ú
T06 = ê (9.28)
ê1 0 0 20.2 ú
ê ú
ë0 0 0 1 û
Where the initial position of the final effector of the Ketzal robot was px = 42.215
cm, py = 0 cm, and pz = 20.2 cm and the orientation obtained was nx = 0, ox = 0, ax = 1,
ny = 0, oy = −1, ay = 0, nz = 1, oz = 0, and az = 0. This confirms that the data obtained
coincide with the initial position and orientation of the robot arm with the calculation
of direct kinematics.
The simulation of the graphic representation of the Ketzal robot was carried out
with the Robotics Toolbox for MATLAB software [56]. Figure 9.9 shows the initial
position and orientation of the final effector of the robotic manipulator.
TABLE 9.5
Angular Ranges in the Joints of the
Ketzal Robot
rad θ1 θ2 θ3 θ4 θ5 θ6
Minimum 0 0 2π 0 2π 0
Maximum 2π π p 2π p 2π
2 2
190 Artificial Intelligence (AI)
ay, az}, and finally the elements ar13ar18 correspond to the array of articular values
[ϴ] = {ϴ1, ϴ2, ϴ3, ϴ4, ϴ5, ϴ6}.
The spatial resolution used in this work was 17 × 17 × 17 × 17 × 17 × 17 × 17 with
a total of 24 137 569 data with 18 variables. The jump values where set to Δϴ1 = π/8,
Δϴ2 = π/16, Δϴ3 = π/16, Δϴ4 = π/8, Δϴ5 = π/16, and Δϴ6 = π/8. The 18 variables con-
sidered in the data set generated correspond to the position array [p], orientation
array [n o a], and joint value array [ϴ].
To guarantee better performance in the knowledge extraction algorithms, the data
pre-processing stage is fundamental, since the efficiency in the results of perfor-
mance and knowledge extraction depends largely on the quality of the data, and low
quality regularly leads to low quality in the knowledge extracted [57].
The techniques generally used in data pre-processing are divided into two areas:
data preparation and data reduction. Data preparation is generally an obligatory
technique and consists of a series of algorithms whose purpose is to prepare the data
so that the knowledge extraction algorithms can be executed without difficulties dur-
ing processing. This area includes data transformation, normalization, cleaning, and
probably the recovery of lost values [40].
On the other hand, data reduction techniques are not always mandatory, but
sometimes they help to obtain better results, where the main objective in this type
of techniques is to obtain a reduced amount of the original data, keeping as much
as possible the integrity of the information in the data, highlighting the methods of
feature selection, instance selection, grouping, compaction, and sampling [60, 61].
Theoretically, obtaining a complete data set will allow for a broader understand-
ing of the problem; however, if the population is too large, the processing time
required will be much more expensive, so it would be unfeasible.
In practice, sampling is one of the most appropriate processes to solve this prob-
lem due to the advantages obtained in the performance of time and cost of process-
ing, even better results can be obtained from the knowledge extracted compared to
the processing of data sets without sampling [57].
Therefore, data sampling can be widely applied in various engineering fields such
as statistics, machine learning, and data mining [59].
Systematic sampling, either alone or in combination with any other technique, is
one of the most used methods because of its simplicity and ease of use; in its simplest
form, it is also known as linear systematic sampling (LSS) and was introduced in
1944 in Ref. [62].
The LSS method can be seen as a sampling technique by cluster in its simple
form, where the procedure is to divide the population N into k groups of n elements
each. Taking only one group of the sample accordingly, each of the k groups has a
probability of 1/k to be selected; in this case, the application of systematic sampling
is feasible because N = k∙n [63].
LSS allows all units of the data set to have the same opportunities to be selected,
for example, if you have an initial set of size N from which you want to select a new
set of size n, a random number is selected for sampling between 1 and k to select
the instance k of the data set where k represents the sampling interval. From the
selected instance, the kth item is selected forward up to the n instances proposals.
Robust Design of Artificial Neural Networks 191
This procedure ensures that each unit has the same possibility of being included in
the sample [64].
The sampling interval constant is usually taken as the integer closest to N/n,
known as the inverse of the sampling fraction. This method has two advantages
mainly. Firstly, the selection of the first unit is determined by a sampling interval or
sampling period chosen, guaranteeing a random sample among a set of samples of
a given interval. Secondly, systematic samples are distributed in a good way among
the population, guaranteeing a representative sample of the entire original data set,
that is to say, there is less risk that some or a large part of the population will not be
represented, maintaining a constant and uniform distribution between the data [57].
Although it is one of the most common and simple methods, it has two main
drawbacks. Firstly, the sampling variance cannot be taken impartially based on the
only systematic sample taken. Secondly, when population size N cannot be divided
equally by the desired sample size n, systematic sampling cannot be performed, that
is, when N is not an integer multiple of the desired sample size n where N ¹ n∙k.
In this case, k is not an integer value and an inefficient sampling effect can prob-
ably be produced, that is, if at any time the characteristics of the population were
periodic and they coincide with a selected sampling interval, the representativeness
of the sample could be biased [63].
In this work, the data set proposed N1 is not an integer multiple of the desired size
n due to the spatial resolution used in generating the data set. It is clear that in this
case the criteria for applying the method LSS is not met, where N = n∙k.
To solve this problem, the reduction data filter (RDF) algorithm was designed
based on the LSS method in such a way that it allows obtaining a representative dis-
tribution of the original data set, minimizing the risk that some or part of the popula-
tion is not represented and that allows obtaining a constant and uniform distribution
for training and test data sets.
1. The values of the original population size N and the desired sample size n
are set.
192 Artificial Intelligence (AI)
According to the flowchart shown in Figure 9.11, in the first section the constant
n = 1000, which represents the amount of desired samples, is initialized. The vari-
able N represents the number of samples for the data set previously generated by
Equations 9.16–9.27.
The variable k represents the sampling interval, also known as the inverse of the
sampling fraction with a value of k = 24 137 569. The start variable represents the
random value between 1 and k. The index variable is the decimal value of the index
in the instance, while the index_r variable is the rounded-up integer value of the
index in the instance.
In the loop section of the iterative control structure for, the instance selection is
made from the kth element forward until reaching the 1000 instances.
In this process, the index variable accumulates the values of the indexes with
decimals and the value of the sampling interval k. Finally, the p variable calculates
the percentage of reduction applied by the proposed algorithm.
Robust Design of Artificial Neural Networks 193
FIGURE 9.11 Dispersion matrix of the position data set D: (a) Before filtering (b) After
filtering.
194 Artificial Intelligence (AI)
FIGURE 9.12 Distribution analysis of position and orientation data set A: (a) Before filter-
ing (b) After filtering.
FIGURE 9.13 Distribution analysis of input data set with training and test: (a) Position (b)
Orientation.
OA cross-configuration was used in the experimental stage with L9(34) and L4(32),
where 36 different ANN architectures were trained and tested, as shown in Table 9.8.
The analysis of variance (ANOVA) was used with the JMP statistical program to
obtain the signal-to-noise ratio.
FIGURE 9.14 Distribution analysis of output data set with training and test.
197
198 Artificial Intelligence (AI)
TABLE 9.6
Design Variables and Their Levels
Design Variables Level 1 Level 2 Level 3
A 6 9 12
B 0 3 0
C 0.1 0.2 0.3
D 0.01 0.1 0.2
TABLE 9.7
Noise Variables and Their Levels
Noise Variables Level 1 Level 2
U Set 1 Set 2
V 9:1 8:2
W Tr-1/Ts-1 Tr-2/Tst-2
TABLE 9.8
OA with L9(34) and L4(32) Configuration
Trial
No. S1 S2 S3 S4 Mean S/N
1 0.34028409 0.33660998 0.33523831 0.33120539 0.3358344 −39.042614
2 0.34294367 0.34775807 0.34162374 0.34110013 0.3433564 −41.071318
3 0.33780647 0.33966576 0.34055964 0.33274657 0.3376946 −39.708489
4 0.31766289 0.32279780 0.32111454 0.31413515 0.3189276 −38.378724
5 0.33082154 0.32892629 0.32892339 0.31509698 0.3259421 −33.015696
6 0.32950029 0.31278398 0.31908650 0.31464315 0.3190035 −32.598799
7 0.30203761 0.29094929 0.30443676 0.29524616 0.2981675 −33.657940
8 0.32171345 0.31260463 0.32582416 0.30687483 0.3167543 −31.330813
9 0.30484374 0.30907752 0.31075605 0.29823255 0.3057275 −34.659736
FIGURE 9.15 S/N analysis for the determination of optimal network parameters: (a) Best
topology with normal profile (b) Best topology with desirability profile (c) Best topology with
maximizing desirability profile.
200 Artificial Intelligence (AI)
TABLE 9.9
Best Values to Design the ANN
with Normal, Desirability and
Maximized Profiles
Trial No. A B C D
1 12 0 0.01 0.1
2 12 0 0.01 0.1
3 9 3 0.01 0.1
the design of the network. The cube on the left describes that the best average value
of the MSE obtained is given by 12 neurons in the first layer, a moment equal to 0.01,
and a learning rate equal to 0.1. However, an increase in neurons in the second layer
can improve the S/N ratio but tends to decrease the yield by approximately 3% above
the MSE average.
On the other hand, in the cube on the right can be seen that 12 neurons in the first
layer, a moment equal to 0.01, and a learning rate equal to 0.3 are sufficient to obtain
the best performance. However, there is a tendency that coincides with the cube on
the left in including a second hidden layer to improve the S/N ratio, but the network
performance tends to decrease by 3% above the MSE average.
From the results obtained in the JMP statistical analysis, three trainings were
carried out with the parameters obtained in each of the profiles, and tests were per-
formed to validate the data obtained with the designed ANN. In the final validation,
the statistical correlation and chi-square tests were performed for the best and worst
prediction, as shown in Figures 9.17 and 9.18, respectively.
The cross-validation method was used with three data sets A, B, and C with data
sizes of 600, 1000, and 1500 elements with 18 variables each, respectively. Each
data set was divided into four subsets of the same size, as shown in Figure 9.19. The
validation set in each training was used to estimate the generalization error, that is,
Robust Design of Artificial Neural Networks 201
FIGURE 9.17 Best joints predicted and correlation test: (a) Best joints predicted (b)
Correlation test.
202 Artificial Intelligence (AI)
FIGURE 9.18 Worst joints predicted and correlation test: (a) Worst joints predicted (b)
Correlation test.
Robust Design of Artificial Neural Networks 203
TABLE 9.10
Results of the Fourfold Cross-Validation
MSE
Fold
Number A B C
1 0.30089341 0.31660459 0.33148271
2 0.32114314 0.32018527 0.33704913
3 0.30725039 0.32869244 0.34501222
4 0.31404578 0.32941529 0.34597482
Average 0.31083318 0.32372439 0.33987972
Std Deviation 0.00872260 0.00633206 0.00688028
the incorrect classification rate of the model with data different from those used in
the training process.
Table 9.10 shows the average MSE value and the standard deviation obtained dur-
ing the workouts performed for each of the sets proposed during the validation of the
neural network using the cross-validation method.
The average value of MSE obtained, and highlighted in Table 9.10 for data set B,
is approximately equal to the value obtained in the confirmation step. However, as it
is shown in Table 9.10, the average MSE value obtained during training with data set
A is approximately 4% less than that obtained with set B and the average MSE value
is approximately 5% higher than that obtained with the set B.
The use of the RDANN methodology allowed solving problems in the design
of the ANN critical parameters. The density of the data used was 80% for training
data and 20% for the test. The best architecture obtained in this work was 12:12:6, a
momentum = 0.01, a learning rate = 0.1, trainrp training algorithm and a MSE = 1E-4.
a great architecture and stop the training of the network in a certain time to avoid
overfitting the data.
The initialization of the weights significantly affects the performance of the
ANN because a large amount of noise is introduced into the training data and
has an impact on the effects produced at the start, specifically on the joints ϴ1,
ϴ2 , ϴ4, and ϴ6. In this case, the negative value does not make physical sense
for the joints, because they cannot make movements outside the mechanically
realizable range.
FUTURE SCOPE
The robust design methodology can be used to find a better artificial neural net-
work configuration with high performance and generalization capability, reducing
the time spent determining the optimal architecture of the artificial neural network
compared to the trial and error approach. In the field of robotics, it is possible to
direct future work towards the use of the robust design of artificial neural networks
methodology presented here for the solution of the inverse kinematics of robots that
do not have an algebraic solution or even to find the direct kinematics of parallel
robots, which is very complex, unlike in serial manipulators.
ACKNOWLEDGEMENTS
This work was partially supported by CONACYT – Becas Nacionales de Posgrado
con la Industria under contract 431101/640582 and paid salary license program
for research IPN-COTEBAL under assignments CPE/COTEBAL/14/2018, CPE/
COTEBAL/33/2019, and CPE/COTEBAL/23/2020. This work was supported
by OMADS S.A. of C.V., an enterprise dedicated to innovation and technological
development.
REFERENCES
1. S. Li, Y. Zhang, and L. Jin, “Kinematic Control of Redundant Manipulators Using
Neural Networks,” IEEE Transactions on Neural Networks and Learning Systems, vol.
28, no. 10, pp. 2243–2254, Oct. 2017, doi:10.1109/TNNLS.2016.2574363.
2. Z. Zhou, H. Guo, Y. Wang, Z. Zhu, J. Wu, and X. Liu, “Inverse Kinematics Solution for
Robotic Manipulator Based on Extreme Learning Machine and Sequential Mutation
Genetic Algorithm,” International Journal of Advanced Robotic Systems, vol. 15, no.
4, p. 1729881418792992, Jul. 2018, doi:10.1177/1729881418792992.
3. G. Jiang, M. Luo, K. Bai, and S. Chen, “A Precise Positioning Method for a Puncture
Robot Based on a PSO-Optimized BP Neural Network Algorithm,” Applied Sciences,
vol. 7, no. 10, p. 969, 2017, doi:10.3390/app7100969.
4. R. Köker, “Reliability-Based Approach to the Inverse Kinematics Solution of Robots
Using Elman’s Networks,” Engineering Applications of Artificial Intelligence, vol. 18,
no. 6, pp. 685–693, Sep. 2005, doi:10.1016/j.engappai.2005.01.004.
5. M. Ono, M. Pavone, Y. Kuwata, and J. Balaram, “Chance-Constrained Dynamic
Programming with Application to Risk-Aware Robotic Space Exploration,” Autonomous
Robots, vol. 39, no. 4, pp. 555–571, Dec. 2015, doi:10.1007/s10514-015-9467-7.
206 Artificial Intelligence (AI)
40. S. García, J. Luengo, and F. Herrera, Data Preprocessing in Data Mining, vol. 72.
Springer International Publishing, 2015.
41. S. S. Mahapatra, and A. Patnaik, “Parametric Optimization of Wire Electrical Discharge
Machining (WEDM) Process Using Taguchi Method,” Parametric Optimization of
Wire Electrical Discharge Machining (WEDM) Process Using Taguchi Method, no.
Journal, Electronic, 2006.
42. R. K. Roy, A Primer on the Taguchi Method, Second Edition. Society of Manufacturing
Engineers, 2010.
43. M. Ibrahim, N. Zulikha, Z. Abidin, N. R. Roshidi, N. A. Rejab, and Mohd Faizal
Johari, “Design of an Artificial Neural Network Pattern Recognition Scheme Using
Full Factorial Experiment,” Applied Mechanics and Materials, 465–466, pp. 1149–
1154, Dec. 2013, doi:10.4028/www.scientifi c.net /AMM.465-466.1149.
44. J. Limon-Romero, D. Tlapa, Y. Baez-Lopez, A. Maldonado-Macias, and L. Rivera-
Cadavid, “Application of the Taguchi Method to Improve a Medical Device Cutting
Process,” The International Journal of Advanced Manufacturing Technology, vol. 87,
no. 9–12, pp. 3569–3577, Dec. 2016, doi:10.1007/s00170-016-8623-3.
45. T. Y. Lin, and C. H. Tseng, “Optimum Design for Artificial Neural Networks: An
Example in a Bicycle Derailleur System,” Engineering Applications of Artificial
Intelligence, vol. 13, no. 1, pp. 3–14, Feb. 2000, doi:10.1016/S0952-1976(99)00045-7.
46. M. S. Packianather, and P. R. Drake, “Modelling Neural Network Performance through
Response Surface Methodology for Classifying Wood Veneer Defects,” Proceedings of
the Institution of Mechanical Engineers, vol. 218, no. 4, pp. 459–466, Apr. 2004.
47. T. Y. Lin, and C. H. Tseng, “Optimum Design for Artificial Neural Networks: An
Example in a Bicycle Derailleur System,” Engineering Applications of Artificial
Intelligence, vol. 13, no. 1, pp. 3–14, 2000, https://doi.org/10.1016/S0952-1976(99)00
045-7.
48. K. Vahab, “An Optimized Artificial Neural Network for Human-Force Estimation:
Consequences for Rehabilitation Robotics,” Industrial Robot: An International
Journal, vol. 45, no. 3, pp. 416–423, Jan. 2018, doi:10.1108/IR-10-2017-0190.
49. S. Kucuk, and Z. Bingul, “Inverse Kinematics Solutions for Industrial Robot
Manipulators with Offset Wrists,” Applied Mathematical Modelling, vol. 38, no. 7, pp.
1983–1999, 2014, doi:https://doi.org/10.1016/j.apm.2013.10.014.
50. L. Zhang, and N. Xiao, “A Novel Artificial Bee Colony Algorithm for Inverse
Kinematics Calculation of 7-DOF Serial Manipulators,” Soft Computing, vol. 23, no.
10, pp. 3269–3277, May 2019, doi:10.1007/s00500-017-2975-y.
51. D. L. Pieper, “The Kinematics of Manipulators under Computer Control,” Physiol.D.,
Stanford University, Ann Arbor, 1969.
52. M. Tarokh, and M. Kim, “Inverse Kinematics of 7-DOF Robots and Limbs by
Decomposition and Approximation,” IEEE Transactions on Robotics, vol. 23, no. 3,
pp. 595–600, Jun. 2007, doi:10.1109/TRO.2007.898983.
53. S. Tejomurtula, and S. Kak, “Inverse Kinematics in Robotics Using Neural
Networks,” Information Sciences, vol. 116, no. 2, pp. 147–164, Jan. 1999, doi:10.1016/
S0020-0255(98)10098-1.
54. A. Larrañaga, “3D Printable Robotic Arm,” GitHub, 2018. https://github.com/AngelLM
(accessed Sep. 18, 2019).
55. J. Denavit, and R. S. Hartenberg, “A Kinematic Notation for Lower-Pair Mechanisms
Based on Matrices.,” Transactions of the ASME Journal of Applied Mechanics, vol. 22,
pp. 215–221, 1955.
56. P. Corke, Robotics, Vision and Control: Fundamental Algorithms in MATLAB, vol. 73.
Springer, 2011.
Robust Design of Artificial Neural Networks 209
57. H. Liu, and H. Motoda, Instance Selection and Construction for Data Mining. Springer
US, 2001.
58. P. R. Krishnaiah and C. R. Rao, Eds., Sampling. Elsevier Science Publisher, 1988.
59. B. Gu, F. Hu, and H. Liu, “Sampling and Its Application in Data Mining,” Technical
Report TRA6/00, Department of Computer Science, National University of Singapur,
2000.
60. H. Liu, and H. Motoda, “On Issues of Instance Selection,” Data Mining and Knowledge
Discovery, vol. 6, no. 2, pp. 115–130, Apr. 2002, doi:10.1023/A:1014056429969.
61. H. Brighton, and C. Mellish, “Advances in Instance Selection for Instance-Based
Learning Algorithms,” Data Mining and Knowledge Discovery, vol. 6, no. 2, pp. 153–
172, Apr. 2002, doi:10.1023/A:1014043630878.
62. W. G. Madow, and L. H. Madow, “On the Theory of Systematic Sampling, I,” The
Annals of Mathematical Statistics, vol. 15, no. 1, pp. 1–24, 1944.
63. S. A. Mostafa and I. A. Ahmad, “Remainder Linear Systematic Sampling with Multiple
Random Starts,” Journal of Statistical Theory and Practice, vol. 10, no. 4, pp. 824–
851, Oct. 2016, doi:10.1080/15598608.2016.1231094.
64. L. H. Madow, “Systematic Sampling and its Relation to Other Sampling Designs,”
Journal of the American Statistical Association, vol. 41, no. 234, pp. 204–217, Jun.
1946, doi:10.1080/01621459.1946.10501864.
10 Generative Adversarial
Network and Its
Applications
A. Vijayalakshmi and Deepa V. Jose
CONTENTS
10.1 Introduction................................................................................................... 211
10.2 Discriminative Learning vs. Generative Learning........................................ 212
10.3 Deep Generative Model................................................................................. 213
10.4 Variational Auto Encoders............................................................................ 214
10.5 Generative Adversarial Network................................................................... 215
10.6 Architecture of Generative Adversarial Network.......................................... 216
10.7 Variations of GAN Architectures.................................................................. 217
10.7.1 Fully Connected GAN (FCGAN)..................................................... 217
10.7.2 Laplacian Pyramid of Adversarial Networks (LAPGAN)................ 217
10.7.3 Deep Convolutional GAN (DCGAN)............................................... 217
10.7.4 Conditional GAN............................................................................... 217
10.7.5 Least-Square GAN............................................................................ 218
10.7.6 Auxiliary Classifier GAN.................................................................. 219
10.7.7 InfoGAN............................................................................................ 220
10.8 Applications of GAN..................................................................................... 220
10.8.1 Image generation............................................................................... 220
10.8.2 Image Translation.............................................................................. 221
10.8.3 Anomaly Detection............................................................................ 221
10.9 Conclusion..................................................................................................... 221
References............................................................................................................... 222
10.1 INTRODUCTION
Generative adversarial networks (GAN) is a deep learning unsupervised machine
learning technique that was introduced in 2014 by Ian J. Goodfellow and co-authors
[1]. The objective of GAN is generating data from scratch wherein it consists of two
neural networks that generate new data points from random uniform distribution
to get an accurate prediction. GANs are modelled as a methodology to generative
modelling that uses deep learning techniques like convolutional neural networks. In
order to generate new samples, the model automatically learns the patterns in the
input image data. GANs consist of two models as represented in Figure 10.1: genera-
tor and discriminator.
211
212 Artificial Intelligence (AI)
The generator is a neural network that generates new data instances whose objec-
tive is to fool the discriminator. The discriminator evaluates the instances generated
by the generator for authenticity of the instances so as not to be fooled by the genera-
tor. The generator feeds as the input a random vector at each training step, thereby
preventing the model from producing the same sample every single time. The dis-
criminator receives the synthetic sample produced by the generator and outputs the
probability that the input image is a real training sample or a fake sample that is
generated by the generator predicting the authenticity of the input image.
Table 10.1 describes the steps taken by a generative adversarial network based on
the diagram depicted in Figure 10.1.
GAN is a deep learning method where two neural networks compete with each
other. Deep learning is a subset of machine learning based on artificial neural net-
works wherein there is a learning involved by improving its own computer algorithms
[2]. Neural network consists of layers made up of different numbers of nodes which
are connected to nodes of adjacent layers. Each node within the layer is assigned
a weight and the node with heavier weight will exert more effect on the next layer
of nodes. Deep learning is used in discovering hierarchical models that epitomize
probability distribution on data that is involved in artificial intelligence applications.
TABLE 10.1
Steps taken by GAN
• A generator takes random numbers as input and returns an image.
• The image so generated is fed to the discriminator together with the actual images from the data set.
• The real and fake images are taken by the discriminator and returns a probability
• The discriminator is in a feedback loop with the training set of images.
• The generator is in a feedback loop with the discriminator.
Generative Adversarial Network 213
on the observed data and quality of data rather than depending on the probabilistic
distribution. When provided with a good amount of data, discriminative models tend
to give better results. Figure 10.2 describes the process of a discriminative model.
Generative model understands the distribution of data and performs the task of
classification. Generative model creates similar data as the training data after learn-
ing the distribution of data. Figure 10.3 shows the process of generative modelling.
The training data consists of observations with many features for a problem of
image generation. Generative models can generate new sets of features that look
similar to the one that is created using the original data. In order to accomplish this,
the model should include a random element that builds a probabilistic distribution
model which affects each of the individual samples generated by the model.
the model using ideas from deep learning. Different types of generative models are
shown in Figure 10.4.
Explicit probabilistic model provides explicit parametric specification of the dis-
tribution of data that have controllable functions. Figure 10.5 shows a simple explicit
probabilistic model.
The explicit probabilistic model for Figure 10.4 is
( ) ( ) ( )
p x, z a = p x z p z a
ë ( )
min G max D V ( D, G ) = E x ~ pdata éë log D( x ) ùû + Ez ~ pz ( z ) é log 1 - D G ( Z ) ù
û
Generative Adversarial Network 217
10.7.4 Conditional GAN
GAN framework is extended to conditional setting [9] in which both generator and
discriminator class is conditioned with extra auxiliary information like class labels
218 Artificial Intelligence (AI)
by feeding this information as an extra layer. This model provides better representa-
tion for multimodal data generation. Figure 10.11 illustrates the structure of a condi-
tional adversarial network.
10.7.5 Least-Square GAN
GAN works with the basic concept of simultaneously training the discriminator and
generator where the function of the generator lies in generating fake image samples
that look real. There are innumerable works showing the application of GAN in gen-
eration of new image samples but the quality of such generated images is limited.
GAN uses sigmoid cross-entropy loss function in the discriminator. LSGAN is pro-
posed based on the assumption that the sigmoid loss function will lead to vanishing
gradient when updating the generator [10]. Least-square GAN (LSGAN) transforms
the GAN through L2 loss function as a replacement for log loss function and hence
we could say that LSGAN uses the least square loss function for the discriminator.
In this method labels for fake and real data, a and b, are used as a–b coding scheme
is used in the discriminator. The objective function can be written as
1 1
( ( ) )
E x ~ pdata ( x ) éê( D( x ) - b ) ùú + Ez ~ pz ( z ) éê D G ( Z ) - a ùú
2 2
min D VLSGAN ( D) =
2 ë û 2 ë û
Generative Adversarial Network 219
1
( ( ) )
Ez ~ pz ( z ) éê D G ( Z ) - c ùú
2
min G VLSGAN (G ) =
2 ë û
In the above example, c represents the value of fake data that the generator and dis-
criminator want to believe. The advantage of LSGAN is that unlike regular GAN,
LSGAN produce superior quality images and they are more stable in the learning
procedure.
ë ( û ) ë (
LS = E é log P S = real Xreal ù + E é log P S = fake X fake ù
û )
ë ( û ) ë (
LC = E é log P C = c Xreal ù + E é log P C = c X fake ù
û )
220 Artificial Intelligence (AI)
The discriminator is trained so that LS + LC are maximized, and the generator is
trained so that the value of LC-LS is maximized. This method provided a new metric
for image discriminability. The result of the work proves to be more discriminable
than that of a model that generates images with low resolution.
10.7.7 InfoGAN
InfoGAN incorporates the information theory for the transformation of noise in an
image to latent codes that have predictable consequences in the output [12]. In this
method, the generator is split into two parts: traditional noise vector and latent code
vector. Further, the mutual information between the codes is maximized. This algo-
rithm is implemented by including a regularization term − λ I(c;G(z,c) − to the objec-
tive function of GAN.
( )
min G max D VI ( D, G ) = V ( D, G ) - l I c; G ( z, c )
Lambda is the regularization term, I(c;G(z,c) is the mutual information between the
latent code c and G(z,c), which is the generator output. InfoGAN is unsupervised and
it learns interpretable and untangled information on images. The advantage of this
method is that it adds negligible computation cost on top of GAN and hence makes
the whole system very easy to train the images.
10.8 APPLICATIONS OF GAN
The ability of GAN to generate immense new samples on demand led to its use
in a variety of real-time applications. The currently existing GAN applications can
be segregated into three broad categories: image generation, image translation, and
anomaly detection. This section gives an overview of the various applications under
these three main categories.
10.8.1 Image generation
The major applications related to this category belong to the generation of image data
sets of handwritten digits, photographs of small objects, and human faces based on the
MNIST, CIFAR-10, and Toronto Face Database [1]. Generation of quite realistic human
faces is described in Ref. [13], which was later adopted to generate images of other
objects. Realistic photographs [14] is another commonly used and widely accepted
application. It seemed to be so perfect and was unable to differentiate between the syn-
thetically generated and the real-time ones. Creation of impressive cartoon characters
[15], especially characters from the Japanese comic books, was extensively used.
Similarly, another fancy application is the photos to emojis [16] where the human
faces or signboards, etc. will be converted to various emojis. Face frontal view gen-
eration [17] found immense use in human face identification and verification applica-
tions. Generating new human poses [18] seems to be useful in textile and modelling
industries.
Generative Adversarial Network 221
10.8.2 Image Translation
Another quite useful application of GAN is the image translation and editing. In
many scenarios, we need to convert images into different image formats or perform
various operations on it to extract information. Image-to-Image translation [19] is
used in many areas like creation of Google maps from satellite images, converting
sketches or black and white images to colour, etc.
Image-to-image translation includes producing another engineered rendition of a
given picture with a particular adjustment, for example, making an interpretation of
a late spring scene to winter. Training a model for picture to-picture interpretation
commonly requires an enormous data set of combined models. These data sets can
be troublesome and costly to make, and at times infeasible. The CycleGAN [20] is
a technique that does the automatic training of the image to-image interpretation
models without combined models. The models are prepared in an unsupervised way,
utilizing an assortment of pictures from the source and target domain that should not
be connected in any way. This procedure is an effective method for accomplishing
outwardly amazing outcomes on a scope of a variety of use areas.
Similarly, text-to-image translation [21] and semantic image-to-photo translation [22]
are quite useful in many scientific and learning applications. Photograph editing is another
vibrant application for reconstructing photographs [23]. This also helps in enhancing spe-
cific features and adding or removing features. To visualize the faces based on age, face
ageing [17] application is useful. Blending the features from different images is possible
with photo blending [24]. Besides, all the best quality view of images can be assured
through the high resolution using super resolution [25] methods. If we want to fill some
areas of images which were removed for some reason, it is possible with photo inpaint-
ing [26]. This will retain the originality of the image. Clothing translation [27] is now a
popular application installed in cloth marts where the customer can view how the attire
looks on them. In fact, video prediction [28], especially in sports and entertainment, is
another application related to this area. We can also see a variety of 3D object generation
[29] currently used in design applications to augment the user experience.
10.8.3 Anomaly Detection
Enhancing the security for comfortable living is always the need of the hour, espe-
cially in this era of smart manufacturing. To ensure security and prevent intrusion
from any sort through predictions is currently in practice in many scenarios wherein
we are completely connected to the internet. Security analysis [30] can be achieved
through GANs. Similarly, intrusion detection [31–33] is another area where GANs
are of much help where models can be devised to predict the chances of intrusions by
monitoring the continuous data from the actuators and the sensors.
10.9 CONCLUSION
The enormous attraction received by the generative adversarial network is not
only because of the ability of the model to learn deep mappings but also because
222 Artificial Intelligence (AI)
of its capability to make use of unlabelled image data. This chapter gives a brief
introduction on GAN and its variations with possible applications. It is very much
clear that the influence of deep neural networks offers enormous openings for novel
applications.
REFERENCES
1. I. J. Goodfellow et al., “Generative Adversarial Networks,” ArXiv14062661 Cs Stat,
Jun. 2014, Accessed: Dec. 16, 2020. [Online]. Available: http://arxiv.org/abs/1406
.2661.
2. X. Dong, J. Wu, and L. Zhou, “How deep learning works –The geometry of deep learn-
ing,” ArXiv171010784 Cs Stat, Oct. 2017, Accessed: Dec. 16, 2020. [Online]. Available:
http://arxiv.org/abs/1710.10784.
3. R. Salakhutdinov, “Learning Deep Generative Models,” Annu. Rev. Stat. Its Appl., vol.
2, no. 1, pp. 361–385, Apr. 2015, doi: 10.1146/annurev-statistics-010814-020120.
4. C. Doersch, “Tutorial on Variational Autoencoders,” ArXiv160605908 Cs Stat, Aug.
2016, Accessed: Dec. 16, 2020. [Online]. Available: http://arxiv.org/abs/1606.05908.
5. R. Wei, C. Garcia, A. El-Sayed, V. Peterson, and A. Mahmood, “Variations in
Variational Autoencoders - A Comparative Evaluation,” IEEE Access, vol. 8, pp.
153651–153670, 2020, doi: 10.1109/ACCESS.2020.3018151.
6. K. Wang, C. Gou, Y. Duan, Y. Lin, X. Zheng, and F.-Y. Wang, “Generative Adversarial
Networks: Introduction and Outlook,” IEEECAA J. Autom. Sin., vol. 4, no. 4, pp. 588–
598, 2017, doi: 10.1109/JAS.2017.7510583.
7. A. Creswell, T. White, V. Dumoulin, K. Arulkumaran, B. Sengupta, and A. A. Bharath,
“Generative Adversarial Networks: An Overview,” IEEE Signal Process. Mag., vol. 35,
no. 1, pp. 53–65, Jan. 2018, doi: 10.1109/MSP.2017.2765202.
8. Z. Wang, Q. She, and T. E. Ward, “Generative Adversarial Networks in Computer
Vision: A Survey and Taxonomy,” ArXiv190601529 Cs, Jun. 2020, Accessed: Dec. 16,
2020. [Online]. Available: http://arxiv.org/abs/1906.01529.
9. M. Mirza and S. Osindero, “Conditional Generative Adversarial Nets,” ArXiv14111784
Cs Stat, Nov. 2014, Accessed: Dec. 16, 2020. [Online]. Available: http://arxiv.org/abs
/1411.1784.
10. X. Mao, Q. Li, H. Xie, R. Y. K. Lau, Z. Wang, and S. P. Smolley, “Least Squares
Generative Adversarial Networks,” p. 9.
11. A. Odena, C. Olah, and J. Shlens, “Conditional Image Synthesis With Auxiliary
Classifier GANs,” ArXiv161009585 Cs Stat, Jul. 2017, Accessed: Dec. 16, 2020.
[Online]. Available: http://arxiv.org/abs/1610.09585.
12. X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, and P. Abbeel, “InfoGAN:
Interpretable Representation Learning by Information Maximizing Generative
Adversarial Nets,” ArXiv160603657 Cs Stat, Jun. 2016, Accessed: Dec. 16, 2020.
[Online]. Available: http://arxiv.org/abs/1606.03657.
13. T. Karras, T. Aila, S. Laine, and J. Lehtinen, “Progressive Growing of GANs for
Improved Quality, Stability, and Variation,” ArXiv171010196 Cs Stat, Feb. 2018,
Accessed: Dec. 16, 2020. [Online]. Available: http://arxiv.org/abs/1710.10196.
14. A. Brock, J. Donahue, and K. Simonyan, “Large Scale GAN Training for High Fidelity
Natural Image Synthesis,” ArXiv180911096 Cs Stat, Feb. 2019, Accessed: Dec. 16,
2020. [Online]. Available: http://arxiv.org/abs/1809.11096.
15. Y. Jin, J. Zhang, M. Li, Y. Tian, H. Zhu, and Z. Fang, “Towards the Automatic Anime
Characters Creation with Generative Adversarial Networks,” ArXiv170805509 Cs, Aug.
2017, Accessed: Dec. 16, 2020. [Online]. Available: http://arxiv.org/abs/1708.05509.
Generative Adversarial Network 223
CONTENTS
11.1 Introduction................................................................................................... 225
11.1.1 Artificial Intelligence......................................................................... 225
11.1.2 Artificial Intelligence and Its Roles in the Environment................... 226
11.2 Technological Solutions................................................................................. 226
11.2.1 Autonomous and Connected Electric Vehicles................................. 226
11.2.2 Conservation Biology........................................................................ 227
11.2.3 Next-Generation Weather and Climate Prediction............................ 229
11.2.4 Smart Earth........................................................................................ 231
11.3 AI in the Monitoring Environment................................................................ 232
11.3.1 Monitoring Soil.................................................................................. 232
11.3.2 Monitoring Water.............................................................................. 233
11.3.3 Monitoring Air................................................................................... 235
11.4 Risks of Artificial Intelligence...................................................................... 235
11.4.1 Bias.................................................................................................... 235
11.4.2 Liability............................................................................................. 235
11.4.3 ASI (Artificial Superintelligence)...................................................... 236
11.5 Conclusion and Future .................................................................................. 236
Acknowledgement.................................................................................................. 237
References............................................................................................................... 237
11.1 INTRODUCTION
11.1.1 Artificial Intelligence
In 1956, artificial intelligence (AI) was established as an educational field of study
[1]. In the field of computer applications and science, artificial intelligence can be
described as the intelligence that is indicated by machines, as compared to the
225
226 Artificial Intelligence (AI)
natural intelligence manifested in humans [2]. The idea was based on the assump-
tion that human intelligence “can be described so accurately that a machine can be
trained to simulate it” [3]. The human mind is capable of conjuring logical reasons
and explanations for events that occur beyond the measure of their control. Through
logic and emotions, some problems may be straightforward for humans to solve, but
rather difficult to computationally reciprocate. For this reason, there are two classes
of models: functionalist and structuralist. Where the functionalist data correlates
data to the equivalent obtained computationally, a structural model casually imper-
sonates the cardinal functioning of the mind [4]. Artificial intelligence is broken into
three domains: mundane, expert, and formal tasks. Each of these three domains, in
turn, contains the fields artificial intelligence plays a role in. Formal tasks are known
for the role they play in the mathematical and gaming fields. Mathematics is not
exclusive to arithmetic but includes geometry, calculus, and logical math as well.
Artificial intelligence has been used to verify existing theorems as well as theorems
being worked on. Expert tasks include artificial intelligence applications in diagnosis
in the medical field, analysis of scientific data, analysis of financial data as well as
in engineering. In engineering, artificial intelligence is applied to error probing and
analysis, manufacturing processes, efficiency testing, and monitoring all manufac-
tured items and processes. Mundane tasks describe the role of artificial intelligence
in language processing, perception, logic, and reasoning. Perception includes speech
and voice applications as well as computer vision [5].
11.2 TECHNOLOGICAL SOLUTIONS
11.2.1 Autonomous and Connected Electric Vehicles
The automotive industry is on the verge of the age of automated vehicles. Connected
and Autonomous Electric Vehicles (CAEV) works on three underlying technolo-
gies: Connected Vehicles (CV), Autonomous Vehicles (AV), and Electric Vehicles
(EV). The amalgamation of facilities offered by CAEVs plays a significant function.
CAEVs can provide services for individuals and products shipped, such as self-driv-
ing, advanced communications, and greatly enhanced mobility. Vehicle knowledge
AI in Environmental Science 227
and networking are part of an evolving response to the current problems of travel in
the world. These advance swiftly to provide fully controlled and partly autonomous
vehicles. By 2030, roads are going to be dominated by these vehicles, generating
significant advantages such as improvised equity, safer roads, and transport access.
Thus, a considerable reduction of greenhouse gases and congestion of traffic can be
expected.
The Society of Automotive Engineers states that there are six levels of driving
autonomy from level 0 (no automation) to level 5 (self-driving). Intermediate levels
(levels 1, 2, 3, and 4) are supposed to gain positive advantages, such as safer roads.
The automation level 5 is necessary to achieve the maximum economic benefit of
AV. That’s when cars don’t need drivers. An autonomous vehicle must be adaptive,
technical, and networking to its environment. Another significant consideration is
that energy is required to reduce GHG emissions as desired. An electricity source
should be incorporated in a vehicle to grid [V2G], a grid to the vehicle [G2V] to
coordinate its charging and unload operations and autonomously reduce its energy
usage. Many advantages can be seen as on-demand technologies such as mobility as
a service, dissipation of traffic congestion as a service, electricity storage as a service,
cellular charging as a service, sensing as a service, computation as a service, and
content distribution as a service. Such programs may be used by citizens, municipal
governments, and non-governmental organizations, and even by corporations. Such
services include a portal for accepting demands for facilities, prices, bill process-
ing, and vehicle assignment. Yet, in comparison to current systems for the software
administration of automotive clouds and V2G, this framework handles heterogeneous
vehicle service distribution. Therefore, the platform has to use appropriate strategic
planning and customized selection plans by integrating service characterizations and
accurate pricing specifications to achieve effective management [7].
11.2.2 Conservation Biology
The revolutionary advances in artificial intelligence have unlocked the ability to
quickly process a variety of signals, accurately identify risks, and provide real-time
alerts to the conservationists. Several AI models have been developed to take in
these signals, such as images, video, and audio, and use it to monitor the threat
faced by different wildlife populations. Machine learning (ML), a subset of AI, is
consistently being employed to develop algorithms for predicting the risk of extinc-
tion of several species [8], assess the overall footprint of fisheries [9], and identify
animals using sensor data obtained from biodiversity hotspots [10]. Initiatives by
tech giants such as Microsoft’s “AI for Earth” and Google’s “AI for Social Good”
are also contributing greatly to wildlife and biodiversity conservation. Deep learn-
ing (DL), yet another subset of AI, is highly effective in identifying patterns. Unlike
machine learning, deep learning can work on unlabelled data sets, such as video,
audio, and images, and hence is becoming a powerful tool in conservation biology.
For example, a fully trained deep learning model can be fed hours of signals col-
lected from the field to recognize unique species of organisms with remarkable accu-
racy. The data can also be collected from open-source repositories such as Google
228 Artificial Intelligence (AI)
Earth for satellite data [11] and Move Bank for animal tracking data [12] and fed
into ML or DL algorithms. In 2019, Google along with several organizations such
as Conservation International, developed “Wildlife Insights,” an AI-powered cloud-
based platform to store and process the millions of photos taken by biologists and
conservationists to streamline the conservation monitoring process. With the help
of AI, Wildlife Insights can label up to 3.6 million photos every hour, several times
faster than human experts [13].
Many projects are being conducted around the world using a combination of
AI and signals from the field to address specific problems. For example, Microsoft
AI for Earth and Gramener has developed a deep learning algorithm to assist the
Nisqually Foundation in the United States to identify salmon species from underwa-
ter camera trap videos. The algorithm was able to process several hours of video to
first identify the frames with salmon in it, and then subsequently identify the type of
salmon in those frames [14].
Automated browser-based tools for detecting and counting animals in images and
identifying individual animal species are also developed. One such tool is Wildbook, a
software framework produced by Wild Me in Portland, Oregon. Wildbook uses a com-
bination of neural networks and computer vision for animal detection and identification.
The team also uses AI to gather pictures from YouTube and Twitter of a particular ani-
mal. Researchers and scientists can start projects for a particular species by submitting
many manually annotated images to train the algorithm. Some examples of these proj-
ects include that for whale sharks (Rhincodon typus), manta rays (Manta birostris and
M. alfredi), Iberian lynxes (Lynx pardinus), and giraffes (Giraffa sp.) (Figure 11.1) [15].
Another example is “iNaturalist” one of the world’s most powerful citizen sci-
ence platforms. It was recently improved by the collaboration between the California
Academy of Sciences (the Academy) and the National Geographic Society.
Crowdsourced images, which can be clicked even with the simplest of cameras,
can be submitted on to this platform to help scientists all over the world track the
response of various species to climate change and devise solutions for the same. The
iNaturalist AI has learned to recognize more than 24,000 types of plants, animals,
and fungi by analyzing more than 65 million images submitted by ordinary citi-
zens. The network also shares its findings with the Global Biodiversity Information
Facility (GBIF), enabling scientists’ access to verified data [16].
AI is especially helpful in providing quick solutions for high-risk situations such
as poaching and climate change. For example, in the African subcontinent, monitor-
ing elephant population can be cumbersome due to its very dispersed population
across a large surface area. Here, AI can be used along with aerial images generated
by organizations such as “Save the Elephants” to identify concentrations of elephants
most susceptible to poaching [17].
The new tools presented by AI along with the rapid influx of data generated have
presented new opportunities for conservation biologists all over the world. AI has
enabled us to process large amounts of data in very less amount of time with high
accuracy, sometimes even better than human experts. It saves time and effort for
these experts, enabling them to focus on devising solutions based on the information
generated by the AI algorithms.
The huge data sets needed and the Earth atmosphere’s intrinsic unpredict-
ability make forecasting upcoming events very difficult indeed. Existing com-
puter systems are needed to make decisions on several phenomena on a massive
scale.
These include issues such as how the Sun heats the Earth’s atmosphere, how pres-
sure changes affect wind patterns, and how temperature-changing processes (ice to
water to vapor) impact the flow of energy through the atmosphere.
We also need to remember the rotation of the Earth in space which helps to churn
the atmosphere all day long. Some tiny shifts in one aspect will affect the potential
events profoundly.
AI may be used to find correlations and create a valid hypothesis, generalizing the
data, utilizing computer-generated mathematical programs, and analytical problem-
solving techniques on broad results sets.
Given the inherent difficulty of weather forecasting, scientists are now using
AI for climate prediction to achieve improved and precise performance. By using
deep mathematical machine learning, AI may learn to predict the future from past
weather data.
The Numerical Weather Prediction (NWP) is an example. This model analyses
and examines large satellite data sets as well as other sensor data sets to produce
short-term weather predictions and long-term temperature projections.
Many companies currently also invest heavily in forecasting AI environment.
IBM, for instance, recently acquired the Weather Service, integrating its data with
its in-house Watson AI growth.
This has led to the development of IBM’s Deep Thunder which provides hyper-
local weather forecasts to customers within a resolution of 0.2–1.2 miles [20].
AI in Environmental Science 231
11.2.4 Smart Earth
While AI offers significant potential to solve the environmental challenges con-
fronting the Earth left undirected, it also can accelerate deterioration of the climate.
This research focuses on leveraging AI systems today and providing maximum
significant effect in pressing environmental challenges. This proposes strategies
by which AI can effectively reinvent mainstream industries and processes tackling
climate change, provide security for food and water, preserve ecosystems, and
foster human well-being. This problem seems closely related to the ongoing issue
of how to guarantee AI isn’t detrimental to humans. The main aim for creating
“safe” AI is proving that it is value-aligned – the vision of a positive scope based
on the ideals of humanity, offering secure implementation of the technology to
humanity. This implies that checks and balances built to ensure that emerging AI
processes stay “secure” will include the well-being of the natural environment as
a fundamental aspect [21].
In the 1940s, the first concrete steps were taken toward artificial intelligence. AI
is extensively used in our daily lives and caused a revolutionary vision consisting of
six factors:
• Big data: Machines give exposure where huge quantities of data, either
organized (in databases and spreadsheets) or unstructured (such as text,
audio, video, and images), can be accessed. All this evidence records our
experiences and increases the understanding of the world for humans. Big
data will only get larger, as trillions of sensors are installed in devices,
products, shoes, autonomous vehicles, and elsewhere. AI-assisted analy-
sis of this knowledge helps us to use this evidence to uncover historical
patterns, to forecast more accurately, to render decisions more efficient,
and more.
• Power consumption: Emerging developments such as cloud computing
and graphics processing units have become simpler and easier for compli-
cated AI-enabled applications to handle large volumes of data by parallel
processing.
• Globally linked: Social networking sites have radically transformed the
way people communicate. Such expanded openness has facilitated informa-
tion exchange and enabled knowledge sharing, leading to the development
of “collective intelligence,” also comprising open platforms that involve
devising AI software and app sharing.
• Open-source applications and information: Open-source applications and
information facilitate the modernization and implementation of AI, as seen
in the success of open-source machine learning frameworks and platforms
such as TensorFlow, Caffe2, PyTorch, and Parl.ai.
• Enhanced algorithms: Scientists also achieved progress in various areas of
AI, especially in “deep learning,” involving layers of neural networks, con-
structed in a way that is inspired by the approach of the human brain to pro-
cessing information. Another recent research field is “extreme learning,” in
232 Artificial Intelligence (AI)
The combination of these variables helped to push AI from in vitro (in research labo-
ratories) to in vivo (in daily lives). Known businesses, as well as start-ups, will now
be exploring developments and innovations in AI. Many people have begun exploit-
ing AI technologies to traverse towns, shop online, find tips for movies, screen out
unwelcome emails, or share a journey to work, whether they know it or not.
Therefore, AI is being considered for its extensive potential. Fifty-four percent
acknowledged making substantial investments in AI in a 2017 PwC study of global
executives, while a shortage of technical expertise remains a major concern. The
realized prices are expected to take off as companies continue to invest in software,
technology management, assets, and AI-enabled innovations: rising from $1.4 bil-
lion in annual sales from AI-enabled applications in 2016 to $59.8 billion by 2025,
according to one research study [22].
11.3.1 Monitoring Soil
The agriculture industry consumes more than 70% of the world’s water usage, annu-
ally. Small farms produce 80% of the food source for the world population, so ensur-
ing the quality and safety of water supplied is essentially critical. IBM’s AgroPad is
a real-time enabled prototype that is being used for chemical analysis of soil or water
samples using AI (Figure 11.3) [6]. It is a business card-sized paper device with a
microfluidic chip that performs the analysis of the sample of water or soil and results
are obtained in seconds [6]. Machine learning and vision algorithms are used to
translate the measured colour of the composition and intensity into concentrations of
chemical constituents present in the sample [6]. This is comparatively more reliable
than human vision-based tests alone. The parameters that can be measured are pH,
AI in Environmental Science 233
FIGURE 11.3 IBM’s AgroPad to analyze soil and water samples [24].
chlorine, magnesium, nitrogen dioxide, and aluminium [6]. This technology could
revolutionize digital agriculture.
AI-enabled agriculture bots are helping farmers in protecting their crops from
weeds and also combat the labour challenge. Monitoring of weed and timely spray-
ing can be possible by incorporating computer vision.
Two research students have devised a smart farming system that works on the
principles of AI and IoT (Internet of Things). The device called E-parirakshak has
been designed to monitor agricultural fields based on factors like fertility, soil tem-
perature, and moisture, controlling water levels, blowers, and sprinklers from a
remote location. The data is pushed to the cloud from where it can be retrieved and
analyzed later. It also helps in the detection of diseases in crops and updates the
farmers in real time.
11.3.2 Monitoring Water
Plankton is considered to be biological and natural sensors of aquatic health.
These are a collection of organisms that serve as the basis of the oceanic food
chain and the primary source of protein for billions of people. These organisms
include bacteria, protozoa, archaea, algae, and other floating organisms. These
organisms can alter their behaviour even after slight changes in the quality of the
water are detected. Therefore, IBM researchers have developed small, autonomous
microscopes in water for plankton monitoring, determining different species, and
tracking their movement. [6]. These findings can be further used for their response
to changes from temperature to oil spills to run-offs. The microscope consists of
an imager chip to capture the shadow of plankton as it swims over, creating a digi-
tal image of its health [6].
234 Artificial Intelligence (AI)
The Ocean Cleanup Project based in the Netherlands is an initiative to extract plas-
tic pollution from oceans, mainly the Great Pacific Garbage Patch. The Interceptor
is a technology developed to avoid new plastic reaching the water, not only to clean
up the plastic (Figure 11.4). It is solar-powered, autonomous, and remote-sensing
enabled with scalability and cost-effectivity.
The very first Interceptor which is a prototype is currently deployed in the
Cengkareng Drain in Jakarta, Indonesia (Figure 11.4). The local government of
Malaysia has incorporated the Interceptor as an addition to their clean-up plans to
stop the flow of plastic into rivers using barriers. It has been deployed in the Klang
River which runs through Kuala Lumpur (Figure 11.5, left and right view).
FIGURE 11.5 The Interceptor deployed in Jakarta (left) and Klang River, Malaysia (right) [25].
AI in Environmental Science 235
11.3.3 Monitoring Air
Air pollution is of great concern in many cities around the world. It fluctuates greatly
based on certain parameters like weather, humidity, wind, and temperature [6].
Certain pollutants interact with water molecules to create smog and while winds
can disperse pollutants, they can also carry forward other pollutants with them. IBM
has launched an initiative called the Green Horizon that utilizes the concept of the
Internet of Things (IoT) and AI to predict air quality and bring down pollutant lev-
els [6]. These systems can make sense out of the huge amount of data and make
forecasts far more effectively. The Green Horizon initiative is making an impact in
Beijing and will move to other cities like Johannesburg and New Delhi [6].
Blue Sky Analytics is a big data and AI start-up with a mission to provide action-
able intelligence starting with air pollution [26]. In October 2019, it launched BreeZo,
an app that provides real-time and contextual air quality data and helps people mini-
mize exposure to air pollution. The company is also building an AI-enabled app
called Zuri, for mapping farm and forest fires, and Zorro, for monitoring industrial
emissions.
11.4.1 Bias
Biases are a huge problem in an everyday human’s life, and for a person with a bias
to feed biased data, and train a system to follow these biased data sets, we are essen-
tially curating a system that in turn logically produces results to match our sense of
bias. This could mean bias towards, for instance, ethnic minorities, religious minori-
ties, selective employment conditions, etc. [26].
11.4.2 Liability
Legal action has been a subject where artificial intelligence continually fails in,
because as humans, for an error in the actions of a machine with artificial intel-
ligence, who would we pass the blame to? Would it be the machine in its entirety or
the person who made it? The fact to be considered is, whether we are ready to accept
that in fields of high precision, medicine for instance, would humans be able to trust
a machine to replace himself and consider in retrospect the small margin of error it
could result in?
236 Artificial Intelligence (AI)
To combat such risks, the WEF states “the stability, the accountability and the integ-
rity of the AI system must be ensured for government and business.” More collab-
oration between public and private entities, technologists, policymakers, and also
philosophers and further innovation in science is required to avoid and consider the
possible danger of artificial intelligence.
ACKNOWLEDGEMENT
The authors listed in this chapter wish to express their appreciation to the RSST
trust Bangalore for their continuous support and encouragement. As a corresponding
author, I also express my sincere thanks to all other authors whose valuable contribu-
tion and important comments made this chapter to this form.
Conflict of Interest The authors listed in this chapter have no conflict of inter-
est as known best from our side. There was also no problem related to funding.
All authors have contributed equally with their valuable comments which made the
chapter to this form.
REFERENCES
1. Russell, S.J. and Norvig, P. 2010. Artificial Intelligence-A Modern Approach, Third
International Edition.
2. Poole, D.I., Goebel, R.G. and Mackworth, A.K. 1998. Computational intelligence (p.
142). New York: Oxford University Press.
3. McCarthy, J., Minsky, M.L., Rochester, N. and Shannon, C.E. 2006. A proposal for
the Dartmouth summer research project on artificial intelligence, august 31, 1955. AI
Magazine, 27(4), pp. 12–12.
4. Lieto, A., Lebiere, C. and Oltramari, A. 2018. The knowledge level in cognitive archi-
tectures: Current limitations and possible developments. Cognitive Systems Research,
48, pp. 39–55.
5. Tutorialspoint.com. 2020. Artificial intelligence - research areas - tutorialspoint. https
://www.tutorialspoint.com/artificial_ intelligence/a rtifi cial_ intelligence_ research _ ar
eas.htm [Accessed 16 Jan. 2020].
238 Artificial Intelligence (AI)
6. IBM Research - Energy and Environment. 2020. IBM research - energy and envi-
ronment. https://www.research.ibm.com/energy-and-environment / [Accessed 15 Jan.
2020].
7. Autonomous & Connected Vehicles. https://automotive.ricardo.com/ hybrid-electronic
-systems/autonomous-connected-vehicles [Accessed 21 Jan. 2020].
8. Darrah, S.E., Bland, L.M., Bachman, S.P., Clubbe, C.P. and Trias‐Blasi, A. 2017. Using
coarse‐scale species distribution data to predict extinction risk in plants. Diversity and
Distributions, 23(4), pp. 435–447.
9. Kroodsma, D.A., Mayorga, J., Hochberg, T., Miller, N.A., Boerder, K., Ferretti, F.,
Wilson, A., Bergman, B., White, T.D., Block, B.A. and Woods, P. 2018. Tracking the
global footprint of fisheries. Science, 359(6378), pp. 904–908.
10. Mac Aodha, O., Gibb, R., Barlow, K.E., Browning, E., Firman, M., Freeman, R., Harder,
B., Kinsey, L., Mead, G.R., Newson, S.E. and Pandourski, I. 2018. Bat detective—Deep
learning tools for bat acoustic signal detection. PLoS Computational Biology, 14(3).
https://doi.org/10.1371/journal.pcbi.1005995
11. Gorelick, N., Hancher, M., Dixon, M., Ilyushchenko, S., Thau, D. and Moore, R. 2017.
Google earth engine: Planetary-scale geospatial analysis for everyone. Remote Sensing
of Environment, 202, pp. 18–27.
12. Kranstauber, B., Cameron, A., Weinzerl, R., Fountain, T., Tilak, S., Wikelski, M.
and Kays, R. 2011. The movebank data model for animal tracking. Environmental
Modelling & Software, 26(6), pp. 834–835.
13. Google. 2020. Using AI to find where the wild things are. https://www.blog.google/p
roducts/ear th/ai-finds-where-the-wild-things-a re/ [Accessed 21 Jan. 2020].
14. partner.microsoft.com. 2020. Gramener case study. https://par tner.microsoft.com/en
-us/case-studies/gramener/ [Accessed 21 Jan. 2020].
15. Nature.com. 2020. AI empowers conservation biology. https://www.nature.com/article
s/d41586- 019- 00746-1 [Accessed 21 Jan. 2020].
16. National Geographic Society Newsroom. 2020. The California academy of sciences
and national geographic society join forces to enhance global wildlife observation
network iNaturalist: A growing community of iNaturalist users—and the artificial
intelligence they help power—help observe and monitor more than 165,000 species
around the world. https:// blog.nationalgeographic.org/2018/06/26/the-california-aca
demy-of-sciences-and-national-geographic-society-join-forces-to- en hance-global-w
ildlife-observation-network-inaturalist/ [Accessed 21 Jan. 2020].
17. Medium. 2020. How AI can save earth’s biodiversity. https://medium.com/@kesari/
how-ai-can-save-earths-biodiversity-94555d57dd28 [Accessed 21 Jan. 2020].
18. McGovern, A., Elmore, K.L., Gagne, D.J., Haupt, S.E., Karstens, C.D., Lagerquist, R.,
Smith, T. and Williams, J.K. 2017. Using artificial intelligence to improve real-time
decision-making for high-impact weather. Bulletin of the American Meteorological
Society, 98(10), pp. 2073–2090.
19. Cho, R., 2018. Artificial intelligence–a game-changer for climate change and the envi-
ronment. https://phys.org/news/2018–06-artificial-intelligence game-changer-climate.
html [Accessed 09 March 2020].
20. McFadden, C., Young, C., Lang, F., English, T. and Lang, F. 2020. AI might be the
future for weather forecasting. Interestingengineering.com. https://interestingengine
ering.com/ai-might-be-the-future-for-weather-forecasting [Accessed 21 Jan. 2020].
21. Bakker, K. and Ritts, M. 2018. Smart earth: A meta-review and implications for envi-
ronmental governance. Global Environmental Change, 52, pp. 201–211. [Accessed 09
March 2020].
22. Pwc.com. 2020. https://www.pwc.com/gx/en/news-room/docs/ai-for-the-earth.pdf
[Accessed 21 Jan. 2020].
AI in Environmental Science 239
CONTENTS
12.1 Introduction................................................................................................. 241
12.2 Transport Costs and Goods Characteristics in Logistics............................ 242
12.3 Logistics Challenges on Festive Days......................................................... 242
12.4 AI in Logistics............................................................................................. 243
12.5 Genetic Algorithms.....................................................................................244
12.6 Vehicle Routing Problem............................................................................. 245
12.7 Performance Measures................................................................................248
12.8 The Time-Dependent Vehicle Routing Problem with Time Windows.......248
12.9 Application of Artificial Intelligence in Route Planning and
Optimization............................................................................................... 249
12.10 Development of GA for Time-Dependent Vehicle Routing Problem
with Time Windows.................................................................................... 249
12.10.1 The GA-based Crossover for the Time-Dependent VRPTW....... 251
12.10.2 Evaluation of Fitness Function...................................................... 253
12.10.3 Mutation........................................................................................ 253
12.11 Comparison of Algorithms in Terms of Vehicles Utilized......................... 257
12.12 Conclusion...................................................................................................260
References............................................................................................................... 261
12.1 INTRODUCTION
Every manufacturing or service organization as well as e-commerce organization
makes use of warehouse storage and distribution to other stores or to customers using
transportation services of vehicles and carriers such as delivery trucks. Physical
distribution is the movement of goods by means of the outbound logistics process,
241
242 Artificial Intelligence (AI)
outward from the end of the assembly lines/shops of factories to the customers.
Supply-chain management comprises logistics as one of its fields and links logistics
with the firm’s other functions like engineering and production and links closely
with the communications network. It includes the procurement of raw materials and
components, upstream, transforming them into finished products, storing them, and
then transporting them to the customers efficiently and effectively. Logistics is the
process of transporting, storage, handling, and packaging of goods and materials,
from the beginning to the end of the manufacturing, sales, and finally the disposal
in order to satisfy customer requirements and also to add competitiveness to the
business.
companies these days are applying artificial intelligence (AI) to their logistics and
supply chain activities with the hope of maximizing their resources by minimizing
the time and cost of planning and executing the routes, the number of vehicles to be
used, and the time of delivery of the goods and services to the customers.
Artificial intelligence is the simulation of human intelligence processes by
machines, particularly computer systems. Artificial intelligence is extensively
applied for improving logistics experience by increasing reliability, reducing the
cost of transportation, faster processing, and deciding optimal routes for last-mile
operations. Given the intense competition in the industry, customer satisfaction
has become the main factor of competitive advantage and for business dominance;
e-commerce firms like Amazon and Flipkart are investing hundreds of millions of
dollars with a view to improve their delivery processes in terms of speed and effi-
ciency and to achieve next-day or even same-day delivery. Artificial intelligence is
the main technology used to realize these goals by finding the optimal route, calcu-
lating exact delivery times, and deciding the correct product mix in the warehouse in
order to minimize logistics expenses.
12.4 AI IN LOGISTICS
The efficiencies gained using AI have achieved the greatest visibility in the fields
of vehicle routing, network/route planning and optimization, and forecasting
demand. Innovations that have been achieved using machine learning in these
fields have led to the emergence of more agile businesses and also becoming more
dynamic. As a result, food delivery aggregators such as Foodpanda, Zomato,
Swiggy, and Uber are able to provide an excellent customer experience. AI can
be applied in order to achieve improved planning of capacity, route optimization,
dynamic pricing, optimizing resource allocation, and vehicles with agility and
nimbleness to the areas where demand is high in order to decrease the waiting
time of the customers.
The package carrier United Parcel Service (UPS) makes use of an application
called ORION (On-Road Integrated Optimisation and Navigation), which is a GPS
tool that used expansive fleet telematics and advanced algorithms to analyze vast
quantities of data that enables UPS’ drivers make deliveries on-time and at the same
time being cost-effective. UPS has also implemented projects to improve efficiencies
for its employees and to help reduce their environmental impact.
Dynamic route planning and optimization can be achieved based on traffic condi-
tions and other factors. A reduction of even 1 mile for a driver each day in one year
will lead to a savings of nearly $50 m for UPS. The deployment ORION will save
for UPS nearly 100 million miles per year, which amounts to consumption of nearly
10 million gallons of fuel less, annually. It leads eventually to a decrease in carbon
dioxide emissions by about 100,000 metric tonnes.
AI will eventually set a new standard of efficiency across supply-chain and logis-
tics processes in the months to come. The scenario is changing quickly, creating a
“new normal” in how global logistics companies manage data, run operations, and
serve customers, in a manner that’s automated, intelligent, and more efficient.
244 Artificial Intelligence (AI)
12.5 GENETIC ALGORITHMS
Meta-heuristics combine different mechanisms to find solutions of good quality.
In contrast to classical heuristics, they accept non-improving solutions during the
search to overcome local optima. Normally, meta-heuristics produce better results
than classical heuristics even though there is increased computation times (Gendreau
et al. 2002).
Genetic algorithm is a technique in AI and was first developed by J. Holland at the
University of Michigan in 1975. Solutions to a combinatorial problem are encoded
as chromosomes. The chromosomes are evaluated for their fitness by an evaluation
function and good properties of a generation of solutions are propagated to the fol-
lowing generations.
A genetic algorithm starts with a collection of chromosomes known as the ini-
tial population. Each of the chromosomes represents a solution to the given prob-
lem. There could be a random generation of the initial population, in which case it
will take more time for the algorithm to converge to the solution or some heuristic
method can be used for the generation of the initial population generated and since
the population is now closer to the solution, it would take less time to converge.
The GA’s next step is known as selection. This mechanism involves selecting
potential chromosomes as the parent chromosomes, which in turn depends on the
fitness values of the parent chromosomes and computed by using the fitness evalu-
ation function. The parent chromosomes thus selected will experience recombina-
tion by means of the crossover operations to generate the offspring chromosomes.
Mutation occurs in a small number of newly produced offspring chromosomes.
Mutation is carried out in order that a degree of randomness is introduced so that
the GA is preserved and prevented from getting converged to a local optimum.
The gene sequence undergoes a random swap in a mutation operation or randomly
undergoes a bit negation in the event of the offspring being encoded in bits. The
formation of new population is now made possible by substituting the parent chro-
mosomes with the offspring chromosomes.
The genetic algorithm will continue through this process until a stopping crite-
rion is met, which can be one of the following:
Genetic algorithms have been used extensively to obtain solution for the vehicle
routing problem with time windows (VRPTW). The genetic algorithm is based on
natural reproduction, selection, and evolution based on Darwin’s theory.
AI Solution for Optimizing E-Commerce 245
The vehicle routing problem is pictorially depicted in Figure 12.1. From this figure,
one can note that each vehicle originates at the central depot, visits a set of customer
or supplier nodes once, fulfil their demands or perform the pickups at the nodes,
and then returns to the depot once again. The nodes can be either supplier sites in
the case of e-commerce company supplier site pickups or customer sites where the
e-commerce company delivers to the customer sites. It can also be different manu-
facturers to whom the suppliers deliver raw materials, components, parts, and other
supplies of items. In this process, the number of customers served by all the vehicles
is equal to the total number of customers under consideration/in the network. The
various colours depicted in the figure represent the various routes of the vehicles to
fulfil customer demand. This output was generated in the Heuristic Lab software
environment, using a plugin developed in this research. If the number of vehicles
under consideration is only one, then that vehicle has to start its journey from the
depot and visit each customer node exactly once and return to the depot. In this
Most vehicle routing models assume constant travel times throughout the day.
In reality, however, travel times fluctuate because of predictable events such as
congestion during rush hours or from unpredictable events such as road accidents/
incidents, unexpected weather conditions etc. (Ichoua et al. 2003) and also uncer-
tainties like volatile demand for service. Many VRPs are based on the assump-
tion that there is a constant and deterministic travel times between the various
customer nodes and the central depot (Kok et al., 2012) or it is the same as the
distance travelled between customer nodes. But in an actual scenario, the vehicles
do not travel at a constant speed (due to factors like congestion, weather condi-
tions, etc.), which impacts the transportation cost because of increased fuel con-
sumption (Kuo 2010).
The TDTSP and the TDVRP were studied by Ehmke et al. (2012, 2016). The
time-dependent travel times were used on the basis of Floating Car Data (FCD)
from Stuttgart, Germany. The travel times so obtained were then converted into
planning data sets using data-mining techniques. The authors applied two differ-
ent kinds of travel-time planning sets. The time-dependent vehicle routing problem
with time windows (TDVRPTW) makes the basic and important assumption that
the travel times are a function of the present time. Traffic congestion affects travel
speeds and hence it affects the total time to complete the route and the number of
vehicles utilized and the total transportation cost can also be found. Travel time
between customers (nodes) and depot has been identified as one of the major fac-
tors that increases the adverse impacts of congestion. Traffic congestion also has a
direct effect on carriers’ cost structure and the relative weight of wages and overtime
expenses (Figliozzi, 2009).
TDVRPs satisfy a property which is known as the “non-passing property,”
and in simpler terms, the First-In First-Out (FIFO) property (Ichoua et al., 2003).
According to Doerfler (2017), supply-chain management professionals face a lot of
delivery challenges, especially in handling routing of the vehicles. She says that
according to Hani Mahmassani, the William A. Patterson Distinguished Chair in
Transportation and Director of the Northwestern University Transportation Center
in Evanston, Illinois, routing is a form of roulette because traffic congestion, road
construction, and weather conditions can adversely affect the time it takes to deliver
packages to their destination. A civil and environmental engineering professor, He
says “if you plan that a certain route should take two hours with travel time and it
takes twice as long, you lose your schedule synchronization quickly”. He also adds
that “if you plan for the worst case or even the 95th percentile of travel time, the extra
time may be unnecessary, and you’ve wasted the use of the vehicle. Thus, there is a
loss of productivity.”
The TDVRP takes the first aspect into account by assuming that travel times
depend on the time of the day. All of these uncertainties if not addressed properly
will result in an overall increase in logistics costs. Hence, vehicle routing is a com-
plex logistics management problem and represents a key class of problems to solve
in order to lower costs and optimize logistics resources. The main purpose of VRP is
to find optimal routes for multiple vehicles visiting a set of locations by minimizing
both time and cost.
248 Artificial Intelligence (AI)
12.7 PERFORMANCE MEASURES
The quality of solution obtained for the vehicle routing problem is judged in terms
of the number of vehicles utilized and also the total distance travelled. If a single
vehicle is used to satisfy the demands of all the customers, the cost of the number
of vehicles as well as the cost of transportation will be the minimum. If the total
demand exceeds the capacity of the vehicle, then the use of single vehicle becomes
infeasible, which necessitates the use of multiple vehicles. This then gives rise to the
multiple Traveling Salesman Problem (mTSP), which in other words is called the
vehicle routing problem. In this case, the objective of the problem is to minimize
the number of vehicles used and also the total distance travelled by the vehicle and
thereby the total costs of transportation.
When multiple vehicles are used to satisfy the demands of the customers, the
objectives are as follows:
• Minimize the total distance of travel of all the vehicles put together and
hence the total cost of transportation of satisfying the demands of all the
customers.
• Minimize the number of vehicles required to meet the demands of the
customers.
experiments confirm that the proposed procedure is effective to obtain very good
solutions to be performed in real-life environment.
To load and start the time-dependent VRP variant in HeuristicLab, a valid file
to load the time-dependent VRP data is required. So, the data from the Solomon
(CVRPTW) instances is used and the travel time matrix is added. The dimension of
the travel time matrix must match the number of cities. For the travel times, random
values between 0.5 and 1.0 were generated using random function in Excel.
So, the objectives of the research (Nanda Kumar and Panneerselvam, 2017) were
to minimize the total distance travelled and to minimize the total number of vehicles
used to cover all the customer nodes. The minimization of the total distance travelled
is treated as the primary measure and the corresponding number of vehicles utilized
to serve all the customer nodes is treated as the secondary measure.
In the genetic algorithm, a chromosome is a chain of integers (genes) and each of the
integers represents a customer node. This chromosome represents all the customers. The
customers on it are separated to several routes, each of them representing a sequence of
deliveries that must be covered by a vehicle. The chromosome consists of information on
the number of vehicles used to service the customer nodes, separately so as to minimize
the number of vehicles used to service the customer nodes. This information is called
as the vehicle-part. In the vehicle information, the number of vehicles is equal to the
number of routes in the customer chromosome. The number on each of the vehicle genes
represents the number of customers to service. The sum of these numbers in vehicle
information should be equal to the total number of customers. The application of the
RSIX crossover method to a pair of chromosomes yields a pair of offspring. The off-
spring chromosome is chosen based on the fitness value. Lower fitness value indicates a
desirable solution, since it will lead to less distance travelled. For example, in Figure 12.2,
there is an illustration of a possible solution of 3 vehicles for 8 customer nodes. There are
3 genes in the vehicle-part of chromosome, which means that the 8 customers are divided
into 3 routes. The 3 on the first gene in vehicle-part represents route 1 that services 3 cus-
tomer nodes: 1, 5, and 4. The 2 on the second gene means that route 2 services customers
2 and 8. The 3 on the third gene indicates that route 3 services customers 7, 6, and 3.
Vehicle travel times in the cities and urban areas vary due to various reasons and
factors, like congestions in traffic, accidents, road repairs, VIP movement, and delays
due to bad weather. If these travel time variations are ignored, when forming plans
for the routes needed for pickup and delivery of goods by vehicles from customers
and suppliers, one may end up in creating route plans that may make the vehicles end
up in urban heavy traffic congestions. As a result of this, vehicles may be spending
unnecessary time in traffic jams and get delayed, resulting in long waiting times
for the customers/suppliers and not having exact information regarding the times at
which the vehicles arrive actually. Due to this, the time windows during which the
demand or supply nodes have to be visited become difficult to be satisfied.
Consideration of time-dependent travel times and also the demand information
arising during real-time to solve the VRP can help in reducing the costs of travel that
occurs if the changes in traffic scenarios are ignored.
The total time taken for travel between two locations depends on the specific
time of departure. Hence, to take these external influences into consideration, the
VRPTW is extended and is studied as TDVRPTW. In this particular scenario, the
driving time that changes with the time of the day is suitably represented by a time-
dependent function.
The steps of the proposed genetic algorithm (SNRPGA) for the time-dependent
vehicle routing problem with time windows are presented below. Kumar and
Panneerselvam (2015) developed a crossover technique for the genetic algorithm,
named Random Sequence Insertion-based Crossover (RSIX).
A schematic view of the crossover operation (RSIX) used in this algorithm is
shown in Figure 12.3. A crossover operation is a major process which is applied on
two chromosomes from the current subpopulation to produce two offspring. This
operation modifies the arrangement of genes of the chromosomes considered and
the resultant chromosomes are called offspring. There are many methods for cross-
over operation according to different problems. “Random Sequence Insertion-Based
252 Artificial Intelligence (AI)
Crossover” method is proposed to perform the crossover operation between any two
chromosomes, which is explained in detail below.
The steps of the Random Sequence Insertion-based Crossover (RSIX) method for
the TDVRPTW are presented below.
Consider two parent chromosomes (P1 and P2) with seven genes in each of them
as shown in Figure 12.3a. The gene elements in each of the parent chromosomes are
from 1 to 7, but in some random order.
Step 1: The route chromosomes of the initial population are generated ran-
domly by using a random seed. The random seed is used to initialize the
new pseudo-random number generator for the time-dependent VRP prob-
lem instances. For the travel times, random values between 0.5 and 1.0 have
been generated.
Step 2: Two chromosomes from the chromosome pool are randomly chosen
as parents.
Step 3: Generate two crossover points, which will lead to three chromosome
segments in each chromosome as shown in Figure 12.3a.
Step 4: Copy the middle segment of the chromosome P1 in Figure 12.3a imme-
diately after the middle segment of the chromosome P2 in Figure 12.3a and
AI Solution for Optimizing E-Commerce 253
fs = å distance
r
s ,r
12.10.3 Mutation
After performing the crossover operation, mutation operation is performed to further
perturb the arrangement of the genes in each of the offspring. This operation selects
254 Artificial Intelligence (AI)
two gene positions randomly and interchanges them for a given probability. This
means that a random number is selected for each offspring. If this random number
has a lesser value, or equal in value to specified probability of mutation, chromosome
undergoes mutation. Else, mutation will not be carried out on that offspring.
Mutation helps to prevent the genetic algorithm from converging to a local opti-
mum. Other genetic algorithms parameters can also influence the GA efficiency.
Crossover probability which is specified in the GA determines the rate at which the
crossover occurs.
In this section, a genetic algorithm with the modified random sequence insertion-
based crossover (RSIX) method for the vehicle routing problem with time windows
(TDVRPTW) is proposed.
This crossover technique (RSIX) is applied to develop a new genetic algorithm
(SNRPGA) (Nanda Kumar and Panneerselvam, 2017)
The steps of the proposed genetic algorithm (SNRPGA) for the time-dependent
vehicle routing problem with time windows are presented below.
The algorithm developed by Nanda Kumar and Panneerselvam (2017) was compared
with some existing algorithms to find the superiority of their new algorithm. So, the
SNRPGA (Nanda Kumar and Panneerselvam 2017) was applied to the TDVRPTW
and was compared with another existing algorithm using a complete factorial experi-
ment with two factors: Problem Size (Factor A) and Algorithm (Factor B) with four
replications in each experimental combination.
In the first comparison of the new algorithm, SNRPGA with the existing algo-
rithm developed by Demir (2012) for the TDVRPTW in terms of the total distance
travelled, it was found by the authors that there was significant difference among
them. Further, through Duncan’s multiple rage test, they proved that the proposed
SNRPGA is superior to the existing algorithm developed by Demir (2012) in terms
of the total distance travelled.
The results of the factorial experiment in terms of the total distance travelled are
shown in Table 12.1. The application of ANOVA to the data given in Table 12.1 gives
the results as shown in Table 12.2.
From the ANOVA results shown in Table 12.2, one can infer that the factors
“Algorithm” and “Problem Size” have significant effects on the total distance trav-
elled. Since, there is a significant difference among the two algorithms compared in
terms of the total distance travelled, Duncan’s multiple range test is next conducted
to identify the best algorithm by arranging the algorithms in the descending order of
their mean total distance travelled from to right.
The standard error used in this test is computed as shown below using the mean
sum of squares of the interaction terms (Problem Size × Algorithm) and the number
of replications under each of the algorithms (24). One can notice the fact that the
mean sum of squares of the interaction term AB is used in estimating the standard
error (SE), because the F ratio for the factor “Algorithm” is obtained by dividing its
mean sum of squares by the mean sum of squares of the interaction term ABij (Nanda
Kumar and Panneerselvam, 2012):
SE = ( MSS AB ¸ n ) = ( 21194.060 ¸ 24 )
0.5 0.5
= 29.72
The least significant ranges (LSR) are calculated from the significant ranges of
Duncan’s multiple range tests table for α = 0.05 and 5 degrees of freedom as shown
in Table 12.3.
The treatment means for the Factor B (Algorithm) in terms of the total dis-
tance travelled are arranged in the descending order from left to right, as shown
in Figure 12.4.
256 Artificial Intelligence (AI)
TABLE 12.1
Results of the Factorial Experiment in Terms of the Total Distance
Travelled for TDVRPTW
Problem Class (Factor A) Algorithm (Factor B)
Class Replication of Class ALNS (Demir) TDVRPTW (SNRPGA2)
Random 1 1 R1 971 831
2 R1 932 675
3 R1 948 717
4 R1 1048 664
Clustered 1 1 C1 822 593
2 C1 826 567
3 C1 827 585
4 C1 827 580
Random Clustered 1 1 RC1 1207 867
2 RC1 1114 771
3 RC1 1258 847
4 RC1 1457 811
Random 2 1 R2 740 525
2 R2 701 688
3 R2 731 590
4 R2 794 506
Clustered 2 1 C2 585 422
2 C2 585 430
3 C2 586 432
4 C2 586 430
Random Clustered 2 1 RC2 777 762
2 RC2 783 573
3 RC2 923 732
4 RC2 962 606
TABLE 12.2
Analysis of Variance for Total Distance Travelled
Degrees F Ratio
Source of Sum of of Mean Sum Calculated F (α =
Variation Squares Freedom of Squares Ratio 0.05) Remarks
Algorithm (B) 697352.302 1 697352.302 146.138 4.12 Significant
Problem Size (A) 1313751.984 5 262750.397 55.062 2.47 Significant
Problem Size × 105970.302 5 21194.060 4.441 2.47 Significant
Algorithm (A × B)
Error 171787.094 36 4771.864
Total 2288861.682 47
AI Solution for Optimizing E-Commerce 257
TABLE 12.3
Duncan’s Multiple Range Tests
Significant Standard LSR = Significant Range ×
No. of treatments – 1 (j) Range Error Standard Error
2 2.872 29.72 85,356
FIGURE 12.4 Results of Duncan’s multiple range test for distance travelled.
From the Duncan’s multiple range test performed as shown in Figure 12.4, it
is also clear that the TDVRPTW is superior in performance when compared to
the existing algorithm used in this study for comparison, in terms of total distance
travelled.
12.11 COMPARISON OF ALGORITHMS IN
TERMS OF VEHICLES UTILIZED
In this section, a comparison is made among the proposed algorithm SNRPGA and
an existing algorithm for the TDVRPTW in terms of number of vehicles utilized
using a complete factorial experiment. The existing algorithm was an iterated local
search algorithm proposed by Hashimoto et al. (2008). The number of factors in the
experiment is 2, viz. Factor A (Problem Size) and Factor B (Algorithm). The number
of levels for the Factor A is 6, Random 1, Clustered 1, Random Clustered 1, Random
2, Clustered 2, and Random Clustered 2. The number of levels for the Factor B is 2,
viz. TDVRPTW and SNRPGA. The number of replications under each experimental
combination is 4. The results in terms of the number of vehicles utilized as per this
design are as shown in Table 12.4.
In this model, Factor A (Problem Size/Problem Class) is a random factor and the
Factor B (algorithm) is a fixed factor. Since Factor A is a random factor, the interaction
258 Artificial Intelligence (AI)
factor ABij is also a random factor. The replications are always random and the num-
ber of replications under each experimental combination is 4. The derivation of the
expected mean square (EMS) is given in Nanda Kumar and Panneerselvam (2012).
To test the effect of Ai as well as ABij, the respective F ratio is formed by dividing the
mean sum of squares of the respective component (Ai or ABij) by the mean sum of
squares of error. The F ratio of the component Bj is formed by dividing its mean sum
of squares by the mean sum of squares of ABij.
The results of ANOVA of the data given in Table 12.4 are shown in Table 12.5.
From the ANOVA results shown in Table 12.5, one can infer that the factors
“Problem Size,” “Algorithm,” and “Interaction of Problem Size and Algorithm” have
significant effects on the response variable “Number of Vehicles Utilized.” Since
there are significant differences among the algorithms, the best algorithm is obtained
TABLE 12.4
Results of Number of Vehicles Utilized for TDVRPTW Problem
Problem Class
(Factor A) Algorithm ((Factor B)
Replication of TDVRPTW SNRPGA2
Class (Hashimoto)
1. Random 1 1 9 9
2 14 9
3 13 10
4 10 9
2. Clustered 1 1 10 8
2 10 8
3 10 8
4 10 8
3. Random Clustered 1 1 10 10
2 11 11
3 11 11
4 14 12
4. Random 2 1 4 3
2 4 3
3 3 3
4 3 3
5. Clustered 2 1 3 3
2 3 3
3 3 3
4 3 3
6. Random Clustered 2 1 4 3
2 4 3
3 3 4
4 3 3
AI Solution for Optimizing E-Commerce 259
TABLE 12.5
Analysis of Variance for Number of Vehicles Utilized for TDVRPTW Problem
Degrees Mean F Ratio
Sum of of Sum of Calculated (α =
Source of Variation Squares Freedom Squares F Ratio 0.05) Remark
Problem Size (A) 551.604 5 110.321 260.430 2.47 Significant
Algorithm (B) 3.521 1 3.521 8.311 4.12 Significant
Problem Size × 5.604 5 1.121 2.646 2.47 Significant
Algorithm (A × B)
Error 15.250 36 0.424
Total 575.979 47
using Duncan’s multiple range test by arranging the algorithms in the descending
order of their mean number of vehicles utilized from left to right.
The standard error used in this test is computed as shown below using the mean
sum of squares of the interaction terms (Problem Size × Algorithm) and the number
of replications under each of the algorithms:
SE = ( MSS AB ¸ n ) = (1.121 ¸ 24 )
0.5 0.5
= 0.216
The least significant ranges (LSR) are calculated from the significant ranges of
Duncan’s multiple range tests table for α = 0.05 and 36 degrees of freedom as shown
in Table 12.6. The results of Duncan’s multiple range test are shown in Figure 12.5.
In this figure, the algorithms are arranged as per the descending order of their mean
number of vehicles utilized from left to right. From Figure 12.5, it is clear that there
is a significant difference between the two algorithms in terms of the mean num-
ber of vehicles utilized and further the proposed algorithm SNRPGA utilizes the
minimum mean number of vehicles compared to the other algorithm. Hence, the
proposed algorithm SNRPGA is superior to the existing algorithm considered in this
research for the TDVRPTW problem.
As already stated, any algorithm that is developed now is to be compared
with some existing algorithms to find the superiority of the new algorithm. So,
the SNRPGA is applied to the TDVRPTW and is compared with an existing
TABLE 12.6
Duncan’s Multiple Range Tests
Significant Standard LSR = Significant Range ×
No. of Treatments – 1 (j) Range Error Standard Error
2 2.872 0.216 0.6204
260 Artificial Intelligence (AI)
FIGURE 12.5 Results of Duncan’s multiple range test for number of vehicles used.
algorithm through a complete factorial experiment with two factors: Problem Size
(Factor A) and Algorithm (Factor B) with four replications in each experimental
combination.
In the first comparison of the proposed SNRPGA with an existing algorithm
developed by Demir (2012) for the TDVRPTW in terms of the total distance trav-
elled, it is found that there is significant difference among them. Further, through
Duncan’s multiple rage test, it is proved that the proposed SNRPGA is superior to the
existing algorithm developed by Demir (2012) in terms of the total distance travelled.
In the second comparison of the proposed SNRPGA with an existing algorithm
developed by Hashimoto et al. (2008) for the TDVRPTW in terms of the number of
vehicles utilized, it is found that there is significant difference among them. Further,
through Duncan’s multiple rage test, it is proved that the proposed SNRPGA is supe-
rior to the existing algorithm developed by Hashimoto et al. (2008) in terms of the
number of vehicles utilized.
In the second comparison of the proposed algorithm, SNRPGA with an existing
algorithm developed by Hashimoto et al. (2008) for the TDVRPTW in terms of the
number of vehicles utilized, it was found that there was significant difference among
them. Further, through Duncan’s multiple rage test, it was proved that the proposed
SNRPGA is superior to the existing algorithm developed by Hashimoto et al. (2008)
in terms of the number of vehicles utilized.
12.12 CONCLUSION
The time-dependent vehicle routing problem (TDVRP) is a class of vehicle rout-
ing problems, where the time to serve the customers vary along with the consid-
eration of the traffic conditions in the route. In order to collect the items from
various suppliers or deliver items to customers, the 3PL logistics transportation
service provider must visit all scheduled customers or suppliers during the differ-
ent hours specified by the supplier or the customer. Some customers or suppliers
AI Solution for Optimizing E-Commerce 261
may also need visits to happen during a certain time window. This variant of
the vehicle routing problem is known as the vehicle routing problem with time
windows. Time window constraints are difficult to meet given the real-time road
conditions like traffic congestions which makes the vehicles to wait long hours in
the traffic jams especially in urban environments. Peak hour traffic congestions
lead to long delays. Hence, in order to pick up or deliver items from the suppliers
and customers more efficiently, many challenges must be overcome and solved
among customers and suppliers, in a 3PL kind of arrangement, whereby the 3PL
vehicle visits the customers’ or suppliers’ sites to deliver or to pick up the items,
in an e-commerce or manufacturing or other set-ups, to satisfy the customers
and suppliers’ specific time windows and to minimize the travel times severely
impacted by heavy traffic conditions.
The vehicle routing problem with time windows is solved using the genetic algorithm
with multi-chromosome representation. The genetic algorithm is an AI technique that
finds near optimal solution to any combinatorial optimization problem which is of the
NP-hard type. It is later used for finding a (near) optimal solution to a VRPTW variant,
the TDVRPTW, which takes into consideration the various times of the day at which
the vehicle travels, since the travel time varies during various times of the day such that
the total distance travelled is minimized for the generated number of number of tours
for the vehicles. So, the objectives are to minimize the total distance travelled and to
minimize the total number of vehicles used to cover all the customer nodes. Hence,
in the logistics sector, genetic algorithms are reducing delivery times and decreasing
costs. In the logistics business, every mile and minute matter. Companies can use a
route planner based on genetic algorithms to map out optimal routes for deliveries.
This study can be useful for planning the supplier site pickups by e-commerce
companies, taking into consideration traffic conditions during different periods of
the day with time window requirements of the suppliers.
REFERENCES
Bell, J.E. and McMullen, P.R. “Ant colony optimization techniques for the vehicle routing
problem,” Advanced Engineering Informatics, vol. 18, no. 1, pp. 41–48, 2004.
Chang, Y.H. Logistical Management. Hwa-Tai Bookstore Ltd., Taiwan, 1998.
Clarke, G. and Wright, J. R. “Scheduling of vehicle routing problem from a central depot to
a number of delivery points,” Operations Research, vol. 12, no. 4, pp. 568–581, 1964.
doi:10.1287/opre.12.4.568.
Dantzig, G. and Ramser, J. “The truck dispatching problem,” Management Science, vol. 6,
pp. 80–91, 1959.
Demir, E. Models and Algorithms for the Pollution-Routing Problem and Its Variations. PhD
Thesis, University of Southampton, Southampton, 2012.
Doerfler, S. “Delivering last-mile options,” Inside Supply Management, vol. 28, no. 2, pp.
26–30, 2017.
Duygu, T., Jabali, O. and van Woensel, T. “A vehicle routing problem with flexible time win-
dows,” Computers & Operations Research, vol. 52, Part A, pp. 39–54, 2014.
Ehmke, J.F., Campbell, A.M. and Thomas, B.W. “Vehicle routing to minimize time-depen-
dent emissions in urban areas,” European Journal of Operational Research, vol. 251,
no. 2, pp. 478–494, 2016.
262 Artificial Intelligence (AI)
Ehmke, J.F., Steinert, A. and Mattfeld, D.C. “Advanced routing for city logistics service
providers based on time-dependent travel times,” Journal of Computational Science,
2012, doi:10.1016/j.jocs.2012.01.006.
Esmat, Z.-R. and Mirmohammadi, S.H. “Site dependent vehicle routing problem with soft
time window: modeling and solution approach,” Computers & Industrial Engineering,
vol. 90, pp. 177–185, December 2015.
Figliozzi, M.A. (2009) A Route Improvement Algorithm for the Vehicle Routing Problem
with Time Dependent Travel Times. Proceedings of the 88th Transportation Research
Board Annual Meeting, Washington DC, 11–15 January 2009, pp. 616–636.
Ganediwalla, S., Kumar, A., Vardhan Verma, A., Parasramka, T. and Jain, C. Retail
Technology – The Next Frontier. RedSeer, 2019. https://redseer.com/reports/retail-
technology-the-next-frontier/
Gendreau, M., Laporte, G. and Potvin, J.-Y. “Metaheuristics for the VRP,” in Toth, P. and Vigo,
D. eds, The Vehicle Routing Problem, SIAM Monographs on Discrete Mathematics
and Applications, Philadelphia, pp. 129–154, 2002.
Hashimoto, H., Yagiura, M. and Ibaraki, T. “An iterated local search algorithm for the time-
dependent vehicle routing problem with time windows,” Discrete Optimization, vol. 5,
pp. 434–456, 2008.
Ichoua, S., Gendreau, M. and Potvin, J.Y. “Vehicle dispatching with time-dependent travel
times,” European Journal of Operational Research, vol. 144, pp. 379–396, 2003.
Kok, A.L., Hans, E.W. and Schutten, J.M.J. “Vehicle routing under time-dependent travel
times: the impact of congestion avoidance,” Computers & Operations Research, vol.
39, no. 5, pp. 910–918, 2012.
Kumar, S. and Panneerselvam, R. “A time-dependent vehicle routing problem with time
windows for e-commerce supplier site pickups using genetic algorithm,” Intelligent
Information Management, vol. 7, pp. 181–194, 2015. doi:10.4236/iim.2015.74015.
Kuo, Y. “Using simulated annealing to minimize fuel consumption for the time-dependent
vehicle routing problem,” Computers & Industrial Engineering, vol. 59, no. 1, pp. 157–
165, 2010.
Meng, F., Ding, Y., Li, W. and Guo, R. “Customer-oriented vehicle routing problem with
environment consideration: two-phase optimization approach and heuristic solution,”
Mathematical Problems in Engineering, vol. 2019, 2019, Hindawi. https://doi.org/10
.1155/2019/1073609
Nanda Kumar, S. and Panneerselvam, R. “A survey on the vehicle routing problem and
its variants,” Intelligent Information Management, vol. 4, no. 3, pp. 66–74, 2012.
doi:10.4236/iim.2012.43010.
Nanda Kumar, S. and Panneerselvam, R. “Development of an efficient genetic algorithm for
the time dependent vehicle routing problem with time windows,” American Journal of
Operations Research, vol. 7, pp. 1–25, 2017. doi:10.4236/ajor.2017.71001.
Pisinger, D. and Ropke, S. “A general heuristic for vehicle routing problems,” Computers &
Operations Research, vol. 34, no. 8, pp. 2403–2435, 2007.
Schyns, M. “An ant colony system for responsive dynamic vehicle routing,” European Journal
of Operational Research, vol. 245, no. 3, pp. 704–718, 2015.
Yu, B., Yang, Z. and Yao, B. “An improved ant colony optimization for vehicle routing prob-
lem,” European Journal of Operational Research, vol. 196, no. 1, pp. 171–176, 2009.
13 Application of Machine
Learning for Fault
Detection and Energy
Efficiency Improvement
in HVAC Application
Umashankar Subramaniam, Sai Charan Bharadwaj,
Nabanita Dutta, and M. Venkateshkumar
CONTENTS
13.1 Introduction................................................................................................... 263
13.2 Sustainable Living in Buildings....................................................................264
13.3 Energy Scenario in Buildings........................................................................ 265
13.4 Digitalization and Artificial Intelligence for Energy Efficiency in
Buildings........................................................................................................ 269
13.4.1 Artificial Intelligence (AI)................................................................. 269
13.4.2 Internet of Things (IoT)..................................................................... 271
13.4.3 Machine Learning (ML).................................................................... 272
13.4.4 Influence of Digitalization on HVAC Systems.................................. 273
13.4.5 Energy Optimization and Scheduling............................................... 273
13.4.6 Predictive Maintenance and Fault Diagnosis.................................... 273
13.5 Conclusion..................................................................................................... 275
Conflict of Interest.................................................................................................. 277
References............................................................................................................... 277
13.1 INTRODUCTION
Sustainability is a continuous process of maintaining climate change where the
usage of natural resources for technological advancement has a neutral effect on
the ecosystem [1]. Whereas sustainable development refers to “comprehensive and
holistic approach of temporal process that will lead to an end point of sustainability.”
Sustainable living is a socio-economic challenge. United Nations Development
Programme (UNDP) has listed out 17 major sustainable development goals [2] and
this forms the basis for leaders and policymakers across the world as a visionary
263
264 Artificial Intelligence (AI)
framework for adopting and enforcing law of the land. Sustainable living describes
the human lifestyle of using natural resources in such a way that it has a neutral
effect on the ecology and ecosystem. The people who generally participate and prac-
tice a set style of living will often try to contemplate to reduce their carbon footprint
by every means: transport, energy, water, etc. It is often described as a form of liv-
ing that meets socio-economic, cultural, and ecological needs without jeopardizing
Nature for future generations.
Today, in India, pollution has become a vital issue. Globally, CO2 emission has
reached 37.1 billion tons, increasing at a rate of 2.7% compared to the previous year.
India is likely to add at least 6.3% on the estimate in 2017, i.e. 2.6 billion tons [3].
The buildings all over the world contribute significantly to carbon emissions and
are the major stakeholders in the end user energy consumption. The construction
sector accounts for 9.1% [4] of the country’s gross domestic product (GDP) and it is
contributing significantly to CO2 emissions. Globally, the building sector consumed
125 EJ in 2016; building and construction industry together accounted for 36% of the
total energy consumption. Now the onus lies on India to push the construction sector
towards green and sustainable buildings.
FIGURE 13.1 Per capita water consumption per day (in litres) in metros of India.
population is chiefly dependent upon groundwater[7]. Some experts say that at least
21 cities in India will reach zero groundwater level by 2020, which is quite alarm-
ing. In the following sections, we will look at various energy consumption scenarios
in the buildings worldwide as well as in the Indian context and how technological
advancements can achieve sustainable developmental goals and sustainable living
inside the buildings. Figure 13.1 illustrates the per capita water consumption per day
in major cities of India [8].
• Affordability
• Reliability
• Sustainability
The aforementioned are closely related to each other and require comprehensive
understanding and a line of action for energy policy. In the modern era of human
civilization with progressive industrialization, globalization, and digitalization and
its policies, it is expected that cities will add a population of 1.7 billion by 2040.
This is quite alarming for policymakers, engineers, think tanks, and people who
are concerned about meeting the unprecedented energy demand. Figure 13.2 shows
world energy consumption, and Figure 13.3 shows CO2 emissions by sector across
the world [9].
This astronomical growth of energy demand is dominated by emerging econo-
mies led by India and followed by other countries. Most of the people across the
world spend 80% of their daily life inside buildings. The building and construction
sector together account for 36% of global energy consumption (Figure 13.2). Forty
percent of total CO2 emissions which includes both direct and indirect emissions is
from buildings (Figure 13.3). Energy usage in buildings continues to grow at a rapid
pace. With the adoption of energy efficiency initiatives in the early millennium,
266 Artificial Intelligence (AI)
there has been a significant energy saving over the last decade and half; otherwise,
the energy usage would have been 31% more by 2022 [10].
Coming to the Indian context, industry consumes the highest energy with
around 40%. Buildings are the second largest consumers of end energy, fol-
lowed by transport and others (Figure 13.4). Figure 13.5 represents the percent-
age share of end user energy consumption in commercial buildings of India,
which illustrates that heating ventilation and air-conditioning (HVAC) systems
consume about 55% of total energy consumption. Figure 13.6 provides a com-
parison of end user energy consumption among various units inside the buildings
Energy usage in buildings is dominantly contributed by heating, cooling, light-
ing, pumping, and appliances, sometimes synonymously called HVAC systems
(Figures 13.5 and 13.6) [11].
Machine Learning for Fault Detection 267
Cooling: Energy usage for cooling the space inside the building grew three times
faster than any other end usage inside buildings from 1996 to 2016. Space cooling is
typically characterized by an electrical fan powered by an electrical source or an air-
conditioner. Cooling is a major social challenge as out of 2.8 billion living in the hot-
test parts of the globe, only 8% have access to cooling facilities (Figure 13.7) [12, 13].
This increase in energy demand creates great thrust for power generation as well
as distribution to meet the peak load demand of end energy usage as well as its
driving agents, resulting in increased emissions due to power generation. Electrical
energy is the driving force for the economic activity and development of infrastruc-
ture in the country. Energy demand is significantly influenced by the increase in
population and economic growth in the hottest demographic regions of the world:
268 Artificial Intelligence (AI)
India, China, Indonesia, and others. Household fans: Electrical fan is the simplest
form of cooling. It is estimated that 2.3 billion residential fans have been under usage
till 2016. Fifty-five percent of the global household has at least one fan in use. Fans
consume about 10% less energy than the air-conditioners. Fans will continue to play
a major role in the cooling demand in the developing countries as fans are much
more affordable than air air-conditioners [13]. Heating: It is the largest component of
energy uses in buildings, accounting for 36% of the total building energy consump-
tion according to the 2018 IEA Report [14]. Pumps provide space heating, cooling,
and hot water in buildings. Pumping is a predominant technology that has evolved
over the years and is being used in AC’s chillers and other systems. Heat pumps are
highly efficient and are commercially proven. Over the decades, the typical pumps
in the residential sector are
Machine Learning for Fault Detection 269
• Split or room AC
• Air water heater pump
• Water to water heater pump
• Ground source heat pumps
Lighting: Lighting applications in India accounts for about 59% of total energy
consumption in residential buildings and 26% in commercial buildings. With the
advent of compact fluorescent light (CFL), light-emitting diode (LED) lights, and
their market-penetrative investments, savings of over 6000 PJ have been achieved
in OCED countries. India is primarily leading the LED market. Sale of LEDs is
expected to exceed by 15% across the world. India has already distributed 50 mil-
lion LEDs until 2017.
In 2017, the lighting sector [15] had a demand of 2 EJ whereas the non-residential
sector had a demand of 4 EJ. With the policy implementation and effective invest-
ment into the market, LED account for 25% of the installed capacity.
Appliances: Appliances alone have consumed 30% of global fuel energy and
accounts for 17% of the world’s total end energy consumption [16]. As previously
mentioned, India is an emerging economy associated with meteoric urbanization and
increasing purchasing capacity of the individuals. These are some of the striking fac-
tors for mounting up the floor area of appliances. Energy consumption by appliances
has also swelled to a whopping level of 58% since 2000. It is estimated that appli-
ances will contribute 9% of the total energy in buildings with the implementation of
energy-efficient policies.
It is worth mentioning that India has taken a major step forward with its first
ever building energy code for residential buildings named Energy Conservation
Building Code (ECBC) in 2017 [17]. To sum up, the energy usage will now tend
to remain flat until 2040 despite the increase in the floor capacity of 60%, making
buildings 40% efficient. India has saved 6% additional energy since 2000. China
and India are the major energy consumers, with a stake of 82%, among the six
emerging economies in 2017; they contribute 5% of fossil fuel and 145 MT CO2
emissions (Figure 13.8)] [18].
especially intelligent. Integral philosophy and motto of this is “can a machine think
and behave like humans” [19].
This makes a loud statement that it is a technology of creating machines that
exhibit characteristic features like thinking, behaviour, sense, learn, explain, and
adhere. Thus, artificial intelligence is a multidimensional, cross-functional, and mul-
tidisciplinary subject involving computer science, physics, biology, psychology, lin-
guistics, arts, sociology, engineering, mathematics, and others, as depicted in Figure
13.10. AI systems will adopt the changes without explicitly being programmed,
rather than being specifically intended to perform a task (Figure 13.11)].
Machine Learning for Fault Detection 271
TABLE 13.1
Some Commonly Used Sensors Inside Buildings
S. No. Sensor Type/Name Comments
1 Temperature and RH Used for measuring temperature and humidity of room say DHT11
or LM35
2 CO2 and smoke sensor Useful in detection of gas leakage for home as well industry, e.g.
MQ2 smoke sensor
3 Water flow rate Measure water flow from the sensors, e.g. YF-S201
sensors
4 Motion detect sensors Used to detect the human activity, e.g. HC-SR501 PIR sensor
5 Acoustic sensor Used for sound detection, e.g. KY-038
6 Light sensors Deployed for luminosity and light measurement TSL2561 or LDR
8 HVAC and smart Used for measuring power and energy, e.g. smart energy meters
energy meters
9 Imaging Thermal IR- or CMOS-based sensors to detect and analyse
occupancy patterns
Machine learning [23] has been basically divided into three types: supervised
learning, unsupervised learning, and reinforcement learning. It is a form of cognitive
theory and science. This machine learning models can be used inside buildings for
predictive modelling of the energy and for fault detection of the various equipment
and appliances inside buildings.
The basic and simple flow block diagram of machine learning is shown in
Figure 13.13.
Pumps, which are an integral part of the buildings, also contribute significantly to
energy consumption as most of the time it is driven by an induction motor, which is
subjected to various faults (Figures 13.15 and 13.16) [27].
Data from vibration sensor smart flow rate meters and pressure sensors will
allow us to predict the pumping faults well in advance to save unnecessary excessive
energy usage and sudden collapse of the pump inside the buildings. Such a kind of
predictive fault diagnosis [28] can be extended to heating, cooling, and ventilation
FIGURE 13.16 Vibration sensor data collected to analyse real-time faults in pumps.
as well. If any fault happens in the motor or water pumping system, the trained data
of the motor or pumps will be fed to the simulation and predictive control software
which will classify the faults using training and testing data with the help of machine
learning software. Support vector machine algorithm is one of the machine learning
algorithms which will identify the faults in the system in very less time. So for fault
classification, SVM algorithm has been used which classified the faulty and no faulty
points from unclassified data (Figure 13.17).
Other interesting aspects of artificial intelligence are listed below:
To achieve this in greater efficacy in the system, we need to monitor each and every
facility management services and deployment of energy harvesting methods to pro-
vide power to sensors and AI-powered devices to conserve energy requirement to
carry out this service.
13.5 CONCLUSION
Buildings account for 36% of the total energy consumption and 28% of the total
carbon emissions. The construction sector in India accounts for 9% of the GDP.
On an average, an Indian citizen consumes around 125–150 litres of water per day
for his essential needs. To address this challenge, India needs to push towards the
green and sustainable buildings incorporating digital technologies like IoT in fusion
with data-driven machine learning models to conserve energy. There are numerous
applications that are being relied upon, including artificial intelligence. AI can be
implemented even in old buildings using retrofit methods. To achieve low carbon
276 Artificial Intelligence (AI)
emissions and higher energy efficiency, we need to adapt and implement new tech-
nologies that will conserve energy in buildings. Buildings provide a huge poten-
tial for energy savings and reducing the carbon emission and thereby achieving a
2°C reduction in global temperature. A 30% improvement in the global average of
the building energy intensity by 2030 is needed to meet the ambitious goals of the
Paris Accord 2015. Machine learning and deep learning technology have brought
remarkable solutions by enabling both continuous monitoring and ground-level fault
detection. The prediction control app-based device based on the hybrid technol-
ogy combined with machine learning, deep learning, and IoT, GPU-based human–
machine interface can be helpful to predict the fault condition before the system is
totally shut down. So whenever any abnormal condition is detected, the machine will
indicate through the alarm that the system is in danger, so it is possible for the system
operator to identify the faulty part easily and take necessary action. This predictive
controller GPU-based device is applicable not only for pumping system but also
for any heavy industry. Machine learning and deep learning technology have been
extensively applied in the biomedical sector, but in the industrial application and
mainly in the pumping system it is very rare. The predictive control hybrid model is
the new point of study where the researcher is planning to reduce the energy loss and
the time of the process and thus making the system flawless. This study is trying to
find out the possibilities of anomalies in the HVAC system and its solution by using
an AI technique.
Machine Learning for Fault Detection 277
CONFLICT OF INTEREST
The authors declare no conflict of interest.
REFERENCES
1. U. S. E. P. Agency, Learn About Sustainability, United States Environmental Protection
Agency, [Online]. Available: https://www.epa.gov/sustainability/learn-about-sustainabi
lity#what. [Accessed 16 February 2019].
2. U. N. D. Programme, Sustainable Development Goals, UNDP, [Online]. Available:
http://www.undp.org/content /undp/en/ home/susta inable - development-goals.html.
[Accessed 15 February 2019].
3. U. Goswami, “India’s CO2 Emissions Forecast to Increase by 6.3% This Year,”
Economic Times, 06 December 2018. [Online]. Available: https://economictimes.ind
iatimes.com /news/politics-and-nation /indias-co2-em issions-forecast-to-increase-by- 6
-3-this-yea r/articleshow/66963109.cms. [Accessed 13 February 2019].
4. C. India, India in Figures, MOSPI, 2018. [Online]. Available: http://www.mospi.gov
.in /sites/default /files/publication_ repor ts/ India_in_figures-2018_ rev.pdf. [Accessed 15
February 2019].
5. J. C. Howe, & M. Gerrard (2010). The Law of Green Buildings: Regulatory and Legal
Issues in Design, Construction, Operations, and Financing. American Bar Association.
6. O. &. T. Singh, “A Survey of Household Domestic Water Consumption Patterns in
Rural Semi-Arid Village, India,” GeoJournal, vol. 78, no. 5, pp. 777–790, 2013.
7. A. Shukla, “Alarming: 21 Indian Cities Will Run Out of Water by 2030,” Business
World, 19 June 2017. [Online]. Available: http://www.businessworld.in/article/
Alar m ing-21-In d ian- Citie s-Wil l-Run - Out- Of-Wa ter-B y-203 0 /19- 0 6-2017-12 0383/.
[Accessed 9 February 2019].
8. S. Mungara, “Five Major Metros Staring at Severe Water Crisis: Expert,” Times of
India, 5 October 2018. [Online]. Available: https://timesofi ndia.indiatimes.com/city/
hyderabad /five-major-metros-staring-at-severe-water-crisis-exper t /articleshow/660801
51.cms. [Accessed 8 February 2019].
9. B. D. a. J. D. Thibaut Abergel, UN Environment and International Energy Agency
(2017): Towards a Zero-Emission, Efficient, and Resilient Buildings, United Nations
Environment Programme, Paris, 2017.
10. C. India, Energy Statistics 2018, Ministry of Statistics and Programme Implementation,
New Delhi, 2018.
11. D. C. P. Commission, The Final Report of the Expert Group on Low Carbon Strategies
for Inclusive Growth, Planning Commission GOI, New Delhi, 2014.
12. 2. OECD/IEA, The Future of Cooling, OECD/IEA, Paris, 2018.
13. T. E. a. R. Institute, Improving Air Conditioners in India, TERI, New Delhi, 2018.
14. OECD/IEA, Energy Efficiency: Heating, IEA, [Online]. Available: https://www.iea
.org/topics/energyefficiency/ buildings/ heating/. [Accessed 1 February 2019].
15. OCED/IEA, Lighting: Tracking Clean Energy Progress, IEA, 2018. [Online]. Available:
https://www.iea.org/topics/energyefficiency/ buildings/ heating/. [Accessed 16 February
2019].
16. OCED/IEA, Energy Efficiency: Appliances, IEA, 2018. [Online]. Available: https://ww
w.iea.org/topics/energyefficiency/ buildings/appliances/. [Accessed 16 February 2019].
17. B. G. Ministry of Power, Energy Conservation Building Code, 2017. [Online]. Available:
https:// beeindia.gov.in/sites/default/files/ BEE_ ECBC%202017.pdf. [Accessed 16
February 2019].
278 Artificial Intelligence (AI)
18. OCED/IEA, Energy Efficiency: Buildings, IEA, 2018. [Online]. Available: https://www
.iea.org/topics/energyefficiency/ buildings/. [Accessed 16 February 2019].
19. T. Point, Artificial Intelligence - Overview, Tutorials Point, [Online]. Available: https
://www.tutorialspoint.com/artificial_intelligence/a rtifi cial_intelligence_overview.htm.
[Accessed 16 February 2019].
20. Wikipedia, Kevin Asthon, Wikipedia, [Online]. Available: https://en.wikipedia.org/
wiki/Kevin_ Ashton. [Accessed 16 February 2019].
21. K. Asthon, “That ‘Internet of Things’ Thing,” RFID Journal, pp. 97–114, 2009.
22. M. M. E. R. C. Jonathan Holdowsky, Inside the Internet of Things, Deloitte University
Press, UK, 2015.
23. S. a. S. B.-D. Shalev-Shwartz, Understanding Machine Learning: From Theory to
Algorithms, Cambridge University Press, Cambridge, 2014.
24. H. M. R. a. K. T. C. Zhou, “Artificial Intelligence Approach to Energy Management
and Control in the HVAC Process: An Evaluation, Development and Discussion,”
Developments in Chemical Engineering and Mineral Processing, pp. 42–51, 1993.
25. S. Salimi & A. Hammad (2019). Critical review and research roadmap of office build-
ing energy management based on occupancy monitoring. Energy and Buildings, 182,
214–241.
26. H. B. Gunay, W. Shen & G. Newsham (2019). Data analytics to improve building per-
formance: A critical review. Automation in Construction, 97, 96–109.
27. D. S. U. a. P. S. Nabanita, “Mathematical Models of Classification Algorithm of
Machine Learning,” in International Meeting on Advanced Technologies in Energy
and Electrical Engineering (IMAT3E’18), Fez, Morocco, 2018.
28. N. Dutta, “Centrifugal Pump Cavitation Detection Using Machine Learning Algorithm
Technique,” in 2018 IEEE International Conference on Environment and Electrical
Engineering and 2018 IEEE Industrial and Commercial Power Systems Europe
(EEEIC/I&CPS Europe), Palermo, 2018.
14 Smart City Using
Artificial Intelligence
Enabled by IoT
P. Srividya and Sindhu Rajendran
CONTENTS
14.1 Introduction................................................................................................. 279
14.2 Structure of a Smart City............................................................................280
14.3 Requirements to Build a Smart City........................................................... 281
14.4 Augmented Reality and Virtual Reality in Building a Smart City............. 281
14.5 Major Components Required to Build a Smart City................................... 282
14.6 Challenges in Building a Smart City.......................................................... 283
14.7 Technologies Involved in Building a Smart City........................................ 283
14.7.1 Integrating AI and IoT to Build a Smart City............................... 283
14.7.2 Blockchain in Building a Smart City............................................284
14.7.3 Big Data in Building Smart City..................................................284
14.7.4 Robotics in Building a Smart city.................................................284
14.8 Components of a Smart City.......................................................................284
14.8.1 Smart Energy................................................................................284
14.8.2 Smart Healthcare.......................................................................... 286
14.8.3 Smart Traffic Management........................................................... 287
14.8.4 Smart Parking............................................................................... 287
14.8.5 Smart Waste Management............................................................ 287
14.8.6 Smart Lighting.............................................................................. 288
14.8.7 Smart Governance........................................................................ 289
14.8.8 Smart Agriculture......................................................................... 289
14.9 Drawbacks in Implementing Smart Cities.................................................. 290
14.10 Conclusion................................................................................................... 291
References............................................................................................................... 291
14.1 INTRODUCTION
A city is said to be smart if the residents are able to access various services and
networks efficiently. The services must be sustainable, flexible, and easily acces-
sible. The goal is accomplished by the use of information, telecommunication, and
digital technologies. Traditional cities are transformed into smart cities by the use of
information and communication technology (ICT). Emerging technologies like IoT,
279
280 Artificial Intelligence (AI)
BD, blockchain, and robotics help to make a city smart. Developing a smart city is a
collaborative work that involves many government and private organizations, public
sectors, and citizens themselves.
The intention of building a smart city is to overcome the problems arising from
rapid urbanization and increasing population. Even though the initial cost involved in
developing a smart city is high, in the long run it helps in reducing energy consump-
tion, environmental pollution, city wastes, water consumption, traffic congestion,
and other problems. Different transportation forms, pollution management, Wi-Fi
systems, green energy, universal identification, and local commerce promotion are
some of the initiatives of building a smart city.
Some of the facilities that can be included in a smart city are public Wi-Fi sys-
tems, mobile apps for citizens to report congestion of roads, systems to manage and
monitor electricity and water usage, systems to intimate contamination of water
resources, systems to identify potholes in the roads, systems to manage parking sys-
tems, and many others.
A city is said to be smart if it possesses the following features [1]:
The infrastructure of smart city includes the physical infrastructure, installed sen-
sors, software, and firmware.
This chapter covers the various aspects of the smart city which include the struc-
ture of a smart city, requirements to build a smart city, various components involved,
challenges faced, and technologies involved in building a smart city. This chapter
also highlights the challenges faced in making a city smart.
• Smart things like sensors, actuators, cameras, etc. to gather the data.
• Tools to analyse the data collected by sensors.
• Cloud gateways to collect the data from IoT devices, store them, and for-
ward them to cloud in a secured manner.
• Data processors to collect the data, store it, and distribute it to control
applications.
• A data warehouse to clean and organize the collected data.
• Algorithms to automate city services and to improve the performance of
various control applications based on data analysis.
• Applications to send the data to the actuators.
• Applications to connect smart things and users.
A smart city built with AI powered by IoT can be used in many application areas
like smart environment, smart buildings, smart traffic management, smart grid,
smart farming, smart waste management, smart health sectors, smart security sys-
tems, smart transport, smart parking, drone delivery, smart postal services, smart
disaster and calamity managements, and many more.
14.8.1 Smart Energy
As energy is an important requirement for our day-to-day activities, hence conserv-
ing energy is a vital factor. Energies are classified as sustainable, renewable, and
Smart City Using Artificial Intelligence 285
grid. The energy from various sources is synchronized by the grid to deliver elec-
tricity of specified frequency and voltage without any fluctuations. The grid ensures
efficient, secured, and sustainable energy system by integrating the source of energy
to end users. Smart infrastructure is the crucial component in a smart energy system.
It is responsible for gathering the energy consumption information and for sharing
the unit rate information to the users.
Optimization of energy utilization can be achieved through smart storage and
smart metering. Smart energy storage can be obtained by using lithium-ion batter-
ies or fuel cells that have a longer life and prove to be effective in energy storage
and its efficient delivery. The smart energy meter records the energy consumed for
the specified time duration and sends the reading to the central office for monitor-
ing and billing.
14.8.2 Smart Healthcare
Healthcare sectors have many voids like the ratio of growing population to the num-
ber of medical practitioners is unmatchable, mistakes by the hospitals in diagnos-
ing diseases, patients receiving wrong treatments, patients not getting immediate
treatment in some emergency situations, and many more. To meet the demands of
the growing population with limited resources, the healthcare sectors must become
smart, intelligent, and efficient. Smart healthcare sector integrates the use of biosen-
sors, wearable devices, ICT, cloud computing, mobile apps, emergency services like
ambulance with the traditional healthcare [5].
Some of the smart healthcare applications include periodical intimation to the
nurses and doctors regarding the patient’s health condition, transferring the real-
time data of the patients to an expert doctor who resides in a far-off place for his
expert opinion, automatic intimation to the pharmacist in case of medicine require-
ment to a patient, and telemedicine. The idea of the smart healthcare sector is shown
in Figure 14.4.
14.8.4 Smart Parking
Due to narrow roads and increase in traffic, parking is always a problem, especially
during holidays which is a real struggle. IoT smart parking solutions can be imple-
mented using road surface sensors embedded in the parking area, which can deter-
mine whether the slots are occupied or free and provides a map for parking on a
real-time basis. A mobile app can also be developed to locate the available parking
slot and also pay for the slots, as shown in Figure 14.6. Smart parking helps with
congestion reduction and reduces carbon emissions.
14.8.6 Smart Lighting
Smart lighting emphasizes on less consumption of energy. Street lights are neces-
sary, but they consume a lot of energy. This can be reduced by the use of smart light-
ing. Since street lights are necessary, sensors are attached or there are Wi-Fi hotspots
which automatically adjust the brightness depending upon the presence of pedes-
trians and automobiles. It employs a mesh network in real time to trigger adjoining
lights and to create a safe circle of light around the localities. Figure 14.8 illustrates
smart lighting system.
Smart City Using Artificial Intelligence 289
14.8.7 Smart Governance
A smart city aims at providing sustainable high-quality amenities to the citizens.
This necessitates all the decision-makers to line up their goals and to work together
in planning all the activities. Smart use of the available resources demands the inter-
actions between different government agencies. Best results can be achieved when
citizens are also allowed to participate in urban planning activities. These changes
can be brought about by adopting smart governance. Smart governance involves the
use of ICT sensibly in order to improve decision-making. It involves the process of
refining the democratic process and changing the methods of delivering public ser-
vices. This is done by better collaboration among different stakeholders, including
government and citizens. Compliance towards the needs of people and improved
decision-making using data, evidence, and other resources are emphasized through
smart governance.
The following are the main aims of smart governance:
14.8.8 Smart Agriculture
The worldwide increase in population has increased the demand for food. In traditional
farming methods, farming is usually done based on predications. Many times it leads
to failure of crops. Due to this, farmers incur huge losses. For efficient cultivation, soil
290 Artificial Intelligence (AI)
moisture, soil quality, air quality, sufficient irrigation, and crop weeds, proper fertil-
ization plays an important role. If these parameters are monitored automatically, and
suitable information is passed on to the farmers, quality and quantity of cultivation
increase. This necessitates for smart agriculture that uses IoT, big data, GPS, and con-
nected devices. Smart agriculture assists in automated farming and also in collecting
and analyzing the data from the field by installing sensors, cameras, and actuators [7].
The analyzed data is then sent back to the farmer. This helps in growing high-quality
crops. Figure 14.9 illustrates smart agricultural system.
Usage of IoT in agriculture can help farmers in the following ways:
1. Monitoring weather conditions, crop development, and soil quality can help
in improving the cultivation.
2. Internal processes are better controlled and this reduces the risk of inferior
production and also allows for better distribution of the products.
3. If the crop yield is known in advance, planning to sell the product can be
made.
1. Building a smart city requires huge data acquisition. In this process, there
are chances that the information about the individual might get revealed.
2. Huge investment is required to build a smart city. This investment must be
borne by the tax payers. This will also raise the cost of living.
3. The whole process involves the usage of sensors and other IoT devices
placed at various places. These devices operate with battery. This causes a
problem when battery replacement is required.
Smart City Using Artificial Intelligence 291
14.10 CONCLUSION
With rapid population explosion in urban areas, the challenges faced by the cities
have drastically increased in an uncontrolled manner. Problems like resource scar-
city, traffic management, pollution level, and many more shoot up with a rising popu-
lation. To provide sustainable prosperity to inhabitants, cities now face economic,
political, and technological threats. This demands the establishment of smarter sys-
tems in cities to optimize the utility of the exhaustible resources and to provide
sustainable development without disturbing the environment. Thus, improvement of
cities powered by digital technologies is set at the highest priority worldwide by the
21st century. Even though the day-to-day living standards of the people in urban
areas is increased drastically due to digitization, the transformation of the traditional
city to smart city has barely begun.
In the development of smart cities, the roles and responsibilities of the participants
involved become significant. The key participants involved in building the smart city
include the government bodies that aim at transforming the lives of public, their
safety, and well-being; various private sectors that assist the government bodies by
funding, constructing, and managing the urban infrastructures; and the stakeholders.
The stakeholders include the citizens themselves and other non-profit organizations.
REFERENCES
1. https:// hub.packtpub.com/ how-ai-is-transforming-the-smar t-cities-iot-tutorial/ ‘
Hands-On Artificial Intelligence for IoT’.
2. Husam Rajab, Tibor Cinkelr, “IoT Based Smart Cities”, 2018 International Symposium
on Networks, Computers and Communications (ISNCC), June 2018.
3. Badis Hammi, Rida Khatoun, Sherali Zeadally, Achraf Fayad, Lyes Khoukhi, “Internet
of Things (IoT) Technologies for Smart Cities”, ET Research Journals, the Institution
of Engineering and Technology, 2015.
4. Subba Rao, Sri VidyaGarige, “IOT Based Smart Energy Meter Billing Monitoring and
Controlling the Loads”, International Journal of Innovative Technology and Exploring
Engineering (IJITEE), March 2019.
5. Rustem Dautov, Salvatore Distefano, Rajkumaar Buyya, “Hierarchical Data Fusion for
Smart Healthcare”, Journal of Bigdata, Feb 2019.
6. Sabeen Javaid, Ali Sufian, Saima Pervaiz, Mehak Tanveer, “Smart Traffic Management
System Using Internet of Things”, 2018 20th International Conference on Advanced
Communication Technology (ICACT), Feb 2018.
7. Sjaak Wolfert, Lan Ge, Cor Verdouw, Marc-Jeroen Bogaar, “Big Data in Smart Farming
– A Review”, Agricultural Systems, May 2017.
15 AI Emerging
Communication
and Computing
N. Girija and T. Bhuvaneswari
CONTENTS
15.1 Introduction................................................................................................. 293
15.2 Industrial Revolution 4.0............................................................................. 294
15.3 Stages of AI................................................................................................. 295
15.4 Classification of AI...................................................................................... 296
15.4.1 Artificial Narrow Intelligence....................................................... 296
15.4.2 Artificial General Intelligence...................................................... 297
15.4.3 Artificial Super Intelligence (ASI)................................................300
15.5 Machine Learning Algorithms....................................................................300
15.6 AI Emerging Communication..................................................................... 301
15.7 Conclusion................................................................................................... 303
Bibliography........................................................................................................... 305
15.1 INTRODUCTION
The current century is recognized as Industrial Revolution 4.0 (I.R. 4.0). I.R. 4.0,
which includes concepts such as Internet of Things (IoT), blockchain technology, and
artificial intelligence (AI), fulfils a substantial role and is recognized as a flourishing
model. This chapter presents the role of AI in emerging communication technology.
This emerging AI technology is affecting not only the communication and comput-
ing environment but also the emerging employment skill requirements in the indus-
try. There is no denying the fact that industry requirements and academic sectors
are interrelated. Therefore, in the coming years, it is inevitable that AI technology-
related programming languages, applications and, projects will be part of the student
curriculum. Deloitte is a UK origin freelancing auditing multinational company. Its
report says that robotics and artificial intelligence influence nanotechnology, drones,
sensor technology, and computer vision and hence are becoming vigorous in the
employment market. That is a reason why BRICS is focusing on vocational educa-
tion training and world skill development in their nations.
This chapter has two parts. The first part provides an introduction, history of
the Industrial Revolution, and the stages of AI. It also provides an overview of how
293
294 Artificial Intelligence (AI)
deep these technologies spread in the mortal world. In the world, every country rec-
ognizes the impact of the industrial revolution which enriching its communication
technologies areas like mobile networks, internet-based social media. AI develop-
ment is categorized into four stages. This chapter also discusses how each stage of
the AI-based algorithm has boosted and targeted the vulnerable brain and its sym-
pathetic expression.
The second part covers topics on AI-based technology diversification. In this part,
the AI technology is classified into artificial narrow intelligence (ANI), artificial
general intelligence (AGI), and artificial super intelligence (ASI) models. Each stage
of AI such as reactive AI, decision-making capability, emotional intelligence and
replication of the human brain, etc., does not allow anyone to guess the saturation
phase of AI. Each AI classification communication task with humane society is also
briefed in this chapter. The significance of machine learning and deep learning con-
cepts and their algorithms in AI are also discussed. The machine learning algorithms
such as the non-maximum suppression algorithm and YOLO have been implemented
in various industry sectors like the health sector and agriculture-based AI-based
applications. AI researchers have also focused on computer vision and hybrid think-
ing in AI communications. These are also discussed in this chapter.
In conclusion, emotional intelligence, ethics, and the law of machine–human
attitudes, and Way of the Future, i.e. Tech Church, are discussed. These combina-
tions based on technology design which will aid the human society reform are also
discussed.
to connect people in any distinct region in the world. Nevertheless, I.R. 4.0 is con-
verging into the replication of superhuman. These technologies are moving towards
concepts like smart cities, smart houses, and secured transactions like blockchain
and deep learning. AI technology is omnipresent not only in the modern work-
place but also in a remarkable part of the lifestyle of the community. Just as I.R.
3.0 introduced modern concepts like computer science, information technology, and
telecommunication, I.R. 4.0 seamlessly combines AI, machine learning, and deep
learning in all the fields ranging from medical to automobile.
15.3 STAGES OF AI
In the earlier stages, AI could perform tasks with human proficiency and functional-
ity. However, AI has limited capacity due to a lack of learning skills. Learning skills
develop with experience. For gaining experience, memory-based learning considered
representing an essential aspect. In this stage, AI response is based on the combina-
tion of various inputs groups. This stage is called reactive AI. In 1997, IBM devel-
oped Deep Blue as a reactive AI machine. Deep Blue beat the chess Grandmaster
Garry Kasparov. This incident was an eye-opener for the human community to turn
and explore AI machine proficiency.
In the second stage, AI continues with memory and decision-making. Storing a
massive volume of data needs memory. AI takes a decision based on an archive of
data stored in the memory. For the decision-making feature, the AI needs analyti-
cal skills. In this stage, AI is smarter in recognizing the thousands of images and
scanned objects in image processing, similar to Google Assistants like Chatbot.
Until the third stage, AI was still not considered equivalent to a human as human
beings learn various emotional aspects and can recognize other people’s emotions.
This feature in AI is called emotional intelligence (EI). Understanding the user’s
emotions and respond accordingly is very much needed for virtual assistants. An
example of an EI-based AI application is Siri. Siri is Apple’s voice-controlled per-
sonal assistant for Apple product users. Siri is a combination of AI and Natural
Language Processing (NLP). The three critical tasks of Siri are service tasks, aware-
ness of the situation, and communication interface.
The fourth stage is the replication of the human brain and self-awareness. AI
understands emotional intelligence but also like human primitive emotions. It barely
exhibits emotions like Sophia robot, which was developed by Hong Kong-based
Hanson Robotics research laboratory. Dr. David Hanson described Sophia as capable
of generating expressions of joy, grief, curiosity, confusion, mediation, sorrow, frus-
tration, among other feelings. It means that AI can now concentrate even more on
human muscle expression which is typically connected with the human brain.
The above-mentioned stages clearly state that in the future, most of the human
communications will be with AI-based support instead of actual human beings.
Each stage of AI releases physical stress to human society. The target of AI haunts
the human brain. Human brains include sentiments- and emotions-based compo-
nents. So, human brain characteristics might become more attached and depen-
dent on emotional intelligence technology. As human beings are most emotionally
296 Artificial Intelligence (AI)
attached to pets, likewise advanced features of Sophia model aims to enhance one
of the family members. This bondage constructs a fashionable relationship between
human–machine or human–robot (Figure 15.1).
15.4 CLASSIFICATION OF AI
AI technology is classified into three categories: Artificial Narrow Intelligence
(ANI), Artificial General Intelligence (AGI), and Artificial Super Intelligence (ASI)
(Figure 15.2).
Core AGI hypothesis: the creation and study of synthetic intelligence with sufficiently
broad (e.g. human-level) scope and strong generalization capability are at the bottom
qualitatively different from the creation and study of synthetic intelligence with a sig-
nificantly narrower scope and weaker generalization capability.
The upgrade from one specific field to another, i.e. from ANI to AGI, requires gen-
eral intelligence. That is the reason AGI aims to concentrate on machine learning.
The combination of AGI and machine learning can recognize and acquire any type
of intellectual task. This combination is called Strong AI.
AGI models use human cognitive skills to solve efficiently an unfamiliar task.
According to Stephen Hawking, mortal beings are slow biological evolutionary mod-
els. Suppose innovative Strong AI redesigns itself and constructs unlimited cognitive
298 Artificial Intelligence (AI)
FIGURE 15.3 Nationwide AGI project. Source: A Survey of Artificial General Intelligence Projects for Ethics, Risk, and Policy, 2017.
299
300 Artificial Intelligence (AI)
recognition and speech recognition are done using deep neural network and k-means
techniques (Figure 15.5).
Classification and regression represent the two prominent features of supervised
learning. The below-mentioned algorithms are the traditional algorithm techniques of
machine learning. For Artificial Narrow Intelligence-based AI technology execution,
the machine learning algorithms such as YOLO (You Only You Once) and Sliding
Window are used to supervise learning technique and are also applied in the image
processing. These types of algorithms are used in self-driving cars. Non-maximum
Suppression algorithm is used in the chessboard for object detection. In Figure 15.6,
the popular ML algorithms used in the real-time application are depicted.
TABLE 15.1
Summary of Classification of AI-based Application and Test
ANI AGI ASI
Application • Deep Blue • Google Lookout App • Amazon Alexa
• Google Assistant • Pixel Night Sight
• Google Translate
• Siri
• Alpha Go
Test • Loebner Prize • Multi-model
• Coffee Test Turning Test
• Robot College Student Test
• Employee Test
• Artificial Scientist Test
healthcare technology. Aware mostly of the Oil and Gas industries in remote spots,
Osprey’s intelligent visual monitoring is prominently used in the oil and gas industry
to detect leaks, safely monitor, and bolster security.
Worldwide, sustainable agriculture continues to be a crucial industry. Persisting
farms to be more profitable in business needs aerial phenotyping (“AP”) technologies,
which is used in Slant Range for crop development for successful breeders. The
above-discussed applications realistically are computer vision-based technology.
Undoubtedly, in a short time, computer vision apps will play a vital role in assuring
safety and risk awareness in routine life.
Based on searching continually and gaining information from the browser and
search engine, the machine learning understands the surroundings and also acts
accordingly. For browsing and seeking reliable information, any mechanical or elec-
tronic devices like the mobile are genuinely needed. The following promotion of
AI technology racks the vulnerable brain as a capable device, i.e. hybrid thinking.
Hybrid thinking has been intimately connected with the human brain. The neocor-
tex is a part of the human brain. The neocortex has multilayered prominent parts
involved in high-order brain functions like vision, audition and olfaction, cognition,
and spatial-visual intelligence, especially for imagination. The neocortex is bigger in
humans than in animals.
Nano-robotics adequately represent an imminent technology in the medical
nanotechnology field. Nanorobot or Nanobot is intimately connected to the neocor-
tex part of the human brain. Whenever a Nanobot allied, the ideal person instantly
thinks for specific information without traditionally using any electronic devices.
The Nanobot automatically connects to the cloud directly. This specific type of mod-
ern thinking is properly designated hybrid thinking. This intelligence technology
provides not only detailed information but also a solution. This modern superior
intelligence is addressed by non-biological intelligence. Ray Kurzweil, Director of
Engineering at Google, has authored the book How to Create a Mind. He correctly
stated in this book that for the next 10 years, hybrid thinking looks promising. He
also believes that in the next two decades non-biological intelligence would be more
potent than intelligent humans.
15.7 CONCLUSION
This chapter began with the description of I.R. 4.0 revolution. The key aim of I.R. 4.0
remains intellectually based automation, the implementation of which requires AI
technology. In the earlier stage, computer-based technology was adopted for accumu-
lating information, which was called a database management system for managing
304 Artificial Intelligence (AI)
databases. Besides, a database system was used for storing historical and pertinent
information needed for decision-making, which was called data warehousing and
data mining. Data analytics is most demanding by industry for not only storing and
decision but also for predicting. The demand of human society and industry is alter-
ing every industrial revolution based on technology development. This alteration is
explained in various stages of AI and the classification of AI topics. The machine
learning algorithm represents the core of AI application and AI emerging commu-
nication, and this chapter focuses on AI and ML with regard to various fields like
agriculture, healthcare, and nano-robotics.
To conclude, when human automation is emerging, then ethics and the law of auto-
mation and religious-based moral discipline are also compulsory. A crucial aspect
that enables human society to act is ethics and law. In 2017, Google DeepMind
gathered the group as “DeepMind Ethics & Society” (DMES). The DMES’ group
members are philosophers, economists, risk analysis experts, political advisors, and
AI technology experts. This group of experts designs the ethical procedure to be
followed by the AI. Verity Harding and Sean Legassick are co-leads of Ethics &
Society, who believe that AI technology must take the ethical and social impact
responsibility. DeepMind has identified the key ethical challenges for AI: privacy,
transparency, and fairness to economic impacts, governance, liability, and many
more.
The next aspect is law. Connecting humans and AI involves appropriate legal
regulations. Human faults are disciplined by legal decree; similarly, the Baltic
nations of Estonia has developed a constitutional status called “Kratt Law”, i.e.
algorithmic liability law for AI. Estonia is one country in which e-governance is
done through advanced AI applications. The abbreviated two letters AI instantly
flash at laymen as technology. But AI represents not only the emerging technology,
it is also considered as the next-century religion. The technology patronized in the
more rapid industrial revolutions. Way of the Future (WOTF) is a contemporary
AI religion’s tech church. WOTF religion focuses on the worship of a Godhead
based on artificial intelligence. Anthony Levandowski is a man behind self-driving
car technology in Uber and also part of Google light detecting and ranging (lidar)
engineering team. Levandowski worked as CEO in the start-up company Pronto
AI, which focused on developing the CoPilot project. The CoPilot project goal is
to develop Advanced Driver-Assistance Systems (ADAS) for truck drivers, which
will find lane detection like whether zebra lines exist, two lines whether white or
yellow are parallel, collision warning, and switching lanes for determining length
and width of the lane. This automated system is most useful for the entire globe
to prevent vehicles-caused accidents. He started the WOTF, and he believes super
intelligence is inevitable. Artificial Super Intelligence is much smarter than super-
human. Therefore, WOTF believes, it is clear to integrate human computational
power like a singular model that will be the most supreme on this planet. Anthony
Levandowski has confidence that God is immeasurable or can’t ensure or control,
but tech god (AI god) also has immeasurable smartness even though it is possible to
interact with tech god and recognize it is listening. Religion is a significant regula-
tory for the moral values of the humane society. In the same way, the forthcoming
AI Emerging Communication and Computing 305
era of AI religion has also influenced human society. Just like how god or spontane-
ous creations exist, the AI’s journey will also virtually exist from ANI to AGI to
ASI with more and more progress!
BIBLIOGRAPHY
1. https://www.bgp4.com/2019/04/01/the-3-types-of-ai-narrow-ani-general-agi-and-sup
er-asi / (15/10/2019)
2. https://www.britannica.com /event/ Industrial-Revolution
3. https://www.forbes.com /sites/cognitiveworld /2019/06/19/7-types-of-artificia l-intellig
ence/#7f0cc0cc233e
4. https://interestingengineering.com/t he-th ree-t ypes- of-a r tificial-intell igence-underst
anding-ai
5. https://www.hansonrobotics.com/news-meet-sophia-the -robot-who -laug hs-sm iles
-and-frowns-just-like-us/
6. https://www.pocket-lint.com /apps/news/apple/112346-what-is-si ri-apple-s-personal
-voice-assistant-explained
7. https://www.nuffieldfoundation.org/sites/default/fi les/files/ Ethical-and-Societal-Imp
lications-of-Data-and-AI-report-Nuffield-Foundat.pdf
8. https://futurism.com /artificial-intelligence-officially-granted-residency
9. https://www.searchenginejournal.com/yandex-artificial-intelligence-machine-learning
-algor ithms/332945/#close
10. https://towardsdatascience.com/ai-machine-lear ning-deep-learning-explained-simply
-7b553da5b960
11. https://9to5google.com/2019/03/18/deepmind-agi-control/
12. https://www.scientifi camer ican.com/a rticle/will-china-overtake-the-u-s-in-ar tificial-
intelligence-research /
13. https://www.mcki nsey.com/i ndust ries/technology-media-and-tele commu nicat ions/
our-insights/an-executives-guide-to-machine-learning
14. https://gmisummit.com /wp-content/uploads/2018/10/ What-is-Hybr id-Thinking.pdf
15. Burden, D, Baden, MS, (2019), Virtual Humans: Today and Tomorrow (1st Ed.),
Chapman & Hall/CRC. https://doi.org/10.1201/9781315151199
16. Chalmers, D, Appleyard, A, (2016), The Singularity: Could Artificial Intelligence
Really Out-Think Us (1st Ed.), Imprint Academic.
17. Goertzel, B, (2014), Artificial General Intelligence: Concept, State of the Art, and
Future Prospects, 5(1), 1–46, https://doi.org/10.2478/jagi-2014- 0001
18. Shabbir, J, Anwer, T, (2015), Artificial Intelligence and Its Role in Near Future, Journal
of Latex Class Files, 14(8), 1–11.
19. Baum, S, (2017), A Survey of Artificial General Intelligence Projects for Ethics, Risk,
and Policy. Global Catastrophic Risk Institute Working Paper 17-1. https://ssrn.com/
abstract=3070741.
20. Adriano, M, Althaus, D, Erhardt, J, Gloor, L, Hutter, A, Metzinger, T, (2015), Artificial
Intelligence. Opportunities and Risks. In: Policy Papers of the Effective, Altruism
Foundation (2), S.1–16. https://ea-foundation.org/files/ai-opportunities-and-risks.pdf
21. Watson, EN, (2019), The Super Moral Singularity—AI As a Fountain of Values,
Big Data and Cognitive Computing, 3(2), 23. https://www.mdpi.com /2504-2289/3
/2/23
22. Carlson, KW, (2019), Safe Artificial General Intelligence via Distributed Ledger
Technology, Big Data and Cognitive Computing, 3(3), 40. https://www.mdpi.com/2504
-2289/3/2/23
306 Artificial Intelligence (AI)
23. Bhasin, H, Bhati, S, et al, (2011), Application of Genetic Algorithms in Machine learn-
ing, International Journal of Computer Science and Information Technologies, 2(5),
2412–2415.
24. Wang, Yu, (2018), Reconfigurable Processor for deep learning in autonomous vehicles,
Special Issue 1- The Impact of Artificial Intelligence on Communication Networks and
Services, ITU Journal, 1(1), 9–21.
Index
A AR, see Augmented reality
Architecture-based variant class, 138–139
AAE, see Adversarial AutoEncoders Architectures variations of, GAN
Accurate level prediction, neural networks auxiliary generative adversarial networks
activation function, 103 (AC-GAN), 219–220
artificial neural networks (ANN), 102–103 conditional generative adversarial
sensor output sampling, 101–102 networks, 217–218
ACGAN, see Auxiliary Classifier GAN deep convolutional generative adversarial
Achieving learning, neural network, 174 networks (DCGAN), 217
ACO algorithm, 246 fully connected generative adversarial
Activation function, 103 networks (FCGAN), 217
Adaptive probabilistic neural networks InfoGAN, 220
(APNN), 152 Laplacian pyramid of adversarial networks
Advanced Driver-Assistance Systems (ADAS), 304 (LAPGAN), 217
Adversarial AutoEncoders (AAE), 139 least-square generative adversarial networks
Adversarial Learning and Inference (ALI), 139 (LSGAN), 218–219
Aerial phenotyping (AP), 302 variational auto encoder (VAE), 214–215
AGI, see Artificial General Intelligence Artificial General Intelligence (AGI), 297–299
AI-based PQD classification methods, 155, 157 Artificial intelligence (AI), 243–244, 269–271;
deep learning structures, 158 see also Individual entries
CNN-based methods, 160–161 agency, 68
deep belief network (DBN)-based applications, 74–75
methods, 160 classification, 296–300
DNN-based methods, 160 description, 225–226
SAE-based methods, 158–159 vs. DL, 77–80
proposed deep learning, 161–165 domain used, 74
WT-based hybrid PQD classification emerging communication, 301–303
method, 161–165 environmental role, 226
AI-infused apps, 67 ethical responsibility, 88–90
Air monitoring environment, 235 ethics reasoning, 84–88
Alexa, 74 implementation steps, 74
Algorithms comparison, vehicles utilized methods, PQD detection, 151–153
terms, 257–260 vs. ML, 77–80
ALI, see Adversarial Learning and Inference overview, 74
“Alpha GO” program, 297 risks of, 235
Amazon Alexa, 300 artificial superintelligence (ASI), 236
Amazon Rekognition, 125 bias, 235
Analysis of variance (ANOVA), 196 liability, 235
total distance travelled, 256 stages, 295–296
vehicles utilized, TDVRPTW, 259 Artificial Narrow Intelligence (ANI), 296–297
Anger, 62 Artificial neural networks (ANNs), 102–103, 132,
ANI, see Artificial Narrow Intelligence 152, 174–178
ANNBP, see Artificial Neural Propagation inverse kinematics solution, 178–179
Network optimal parameters, 181–184
ANNs, see Artificial neural networks robust design of, 179–180
AnoGAN, 145 Artificial Neural Propagation Network
ANOVA, see Analysis of variance (ANNBP), 177
AP, see Aerial phenotyping Artificial neuron model, 174
APNN, see Adaptive probabilistic neural with transfer function, 175
networks Artificial Scientist Test, 298
307
308 Index
Emotion, 61 F
Emotional intelligence (EI), 295
Encoder–decoder-based GAN, 139 Face swapping GAN (FSGAN), 144
Encoder–decoder components, 139 Factorial design of experiments, 179
Energy-based GAN (EBGAN), 139, 140 Fair Event Interleaving (FEI), 53
Energy Conservation Building Code False Positive Per Window (FPPW), 67
(ECBC), 269 Faster region-based convolutional neural
Energy efficiency, buildings network, 122
artificial intelligence (AI), 269–271 Fast Fourier transform (FFT) method, 151
digitalization influence, HVAC systems, 273 Fast region-based convolutional neural
internet of things (IoT), 271–272 network, 122
machine learning (ML), 272–273 FBG, see Fibre Bragg grating sensor
optimization and scheduling, 273 FBG-embedded cantilever beam, 94
predictive maintenance and fault diagnosis, FBG-embedded cantilever sensor, 109
273–275 FCA, see Fixed channel allocation
scenario, 265–269 FCGAN, see Fully connected GAN
sensors, 272 FDNN, see Feed-forward deep neural network
sustainability, 264–265 Feature, 75
Energy scenario structures, buildings Feature extraction, 113
appliances, 269 methods, 114
consumption, 265–267 colour descriptor, 116–120
cooling, 267 colour histogram, 115–116
heating, 268 low-level features, 114–115
household fans, 268 texture estimator, 115
HVAC systems, 266 types, 114
lighting, 269 Feed-forward deep neural network (FDNN), 161
Engine oil SAE 15W-40 over temperature classifier test set, performance report, 165
density variations, 96 model flowchart, 163
dynamic and kinematic viscosity structure, 165
variations, 96 FEI, see Fair Event Interleaving
Enhanced Super Resolution GAN FFT, see Fast Fourier transform method
(ESRGAN), 142 Fibre Bragg grating (FBG) sensor, 94, 97–98
Environmental science Fixed channel allocation (FCA), 11, 12
monitoring environment, 232 Float sensor, 99–100
air, 235 Formulation-based variant class, 139–141
soil, 232–233 Fourier transform (FT) methods, 153
water, 233–234 FPPW, see False Positive Per Window
technological solutions FSGAN, see Face swapping GAN
autonomous and connected electric FT, see Fourier transform methods
vehicles, 226–227 Fully connected GAN (FCGAN), 217
conservation biology, 227–229 FusionGAN, 140, 144
next-generation weather and climate
prediction, 229–230 G
smart earth, 231–232
ESRGAN, see Enhanced Super Resolution GAN GA, see Genetic algorithm
Estonia, 304 GAN, see Generative adversarial networks
Ethical responsibility, 88–90 GauGAN, 143
Ethics and law, 304 Gauss Surgical computer vision healthcare
Ethics reasoning, 84–88 technology, 301–302
components, 87 General AR-based system, 126
Event computational time (ECT), 44 Generative adversarial networks (GAN),
Excitement, 62 132–134, 215–216
Expected Utility Theory, 85 adversarial process, 136
Explicitly defined model, 134 applications of, 141–145, 220
Explicit PDF generative models, 134 anomaly detection, 221
Explicit probabilistic model, 214–215 image generation, 220
Index 311
H Java language, 16
Jensen–Shannon divergence (JSD), 137
Happiness, 63
Healthcare, 125 K
Heating ventilation and air-conditioning (HVAC)
application, see Energy efficiency, Ketzal robot manipulator, 182
buildings kinematics analysis, 184–188
Heuristic approach, 74 analysis and confirmation stage, 196,
High-level features, 114, 118 198–203
HLR, see Home Location Register data set description, 189–191
HOG features, 122 planning and experimentation stage,
Home Location Register (HLR), 14 195–196
Hough transform, 119–120 reduction data filter algorithm, 191–193
Hybrid approach, 88 training and test, data set analysis,
Hybrid thinking, 303 194–195
Hyperbolic sigmoid tangent transfer function, 204 KLD, see Kullback–Leibler divergence
Hypothesis formulation, 61 K-means techniques, 301
anger, 62 Kratt Law, 304
desire, 62–63 Kullback–Leibler divergence (KLD), 137–138
excitement, 62
happiness, 63 L
relax, 63
Hypothesis testing outcomes, 65–66 Label, 75
Laplacian pyramid of adversarial networks
I (LAPGAN), 139, 217
LDA, see Linear discriminant analysis
Image classification, 120 Learning algorithm, 106
and localization, 120 Learning curves, 166
312 Index
Least-square GAN (LSGAN), 140, 218–219 Microsoft Azure Computer Vision API, 125
Lee Se-Dol, 297 Mitchell, Tom, 75
Levandowski, Anthony, 304 ML, see Machine learning
Linear (ramp), 103 MLP, see Multilayer perceptron algorithm
Linear discriminant analysis (LDA), 117–118 Mobile Telephone Switching Office (MTSO), 14
Linear systematic sampling (LSS), 190 Mobile user (MU), 7, 10, 18
Liquid level, 93 MOH, see Massive obstetric haemorrhage
measurements, dynamic environments, 94 Monitoring environment, 232
inclination influence, 95–96 air, 235
sloshes influence, 96–97 soil, 232–233
temperature influence, 94–95 water, 233–234
Local feature detection, 114 Monte Carlo (MC) simulation method, 42
Loebner Prize competition, 298 MRA, see Multi-resolution analysis
Logistics; see also E-commerce logistics vehicle MSE, see Mean square quadratic error
routing MTSO, see Mobile Telephone Switching Office
artificial intelligence (AI), 243–244 mTSP, see Multiple Traveling Salesman Problem
challenges, 242–243 MU, see Mobile user
Long short-term memory (LSTM) network, 127 Multi-agent systems (MAS)
Loss function-based modification, 140 agent and, 12–13
Low-level features, 114–115 in WCS, 13–14
LSGAN, see Least-square GAN Multilayer perceptron (MLP) algorithm, 178
LSS, see Linear systematic sampling Multiple discriminator GAN (MD-GAN), 139
LSTM, see Long short-term memory network Multiple generator GAN (MGAN), 139
Multiple Traveling Salesman Problem (mTSP), 248
M Multi-resolution analysis (MRA), 155
MuseGAN, 144
Machine learning (ML), 73, 272–273 Mutation, genetic algorithm (GA)
vs. AI, 77–80 Duncan’s multiple range tests, 257
algorithms, 300–301 factorial experiment, 256
real-time application, 303 steps of the proposed genetic algorithm
applications, 76 (SNRPGA), 254–255
vs. DL, 77–80 variance analysis, 256
domain used, 75–76 Myopic traditions and values, 89
implementation steps, 75
overview, 75 N
Mann–Whitney U test, 65
Markov’s decision process, 127 Nano-robotics, 303
MAS, see Multi-agent systems National Rural Employment Guarantee Scheme
Massive obstetric haemorrhage (MOH), 301 (NREGS), 264
MATLAB/Simulink model, electrical power Nationwide AGI project, 299
distribution system, 156 Natural Language Processing (NLP), 295
MATLAB software, 16, 104, 124 NC agent (NCA), 22, 25
MC, see Monte Carlo simulation method Neocortex, 303
McCarthy, John, 74 Neural network approach, 94
MD-GAN, see Multiple discriminator GAN Next-generation weather and climate prediction,
Mean square quadratic error (MSE), 182 229–230
Measurement system, 102 NLP, see Natural Language Processing
MedGAN, 145 Noise variables, 182–183
Medical image processing, 145 Non-biological intelligence, 303
Medical imaging, 125 Non-maximum Suppression algorithm, 301
Meta-heuristics, 244 Non-Max suppression, 122
Methodological approach Non-verbal communications, 301
participants, 64 Normalization, 140–141
procedure, 64–65 Normative ethics, 86–87
Mexican hat wavelet function, 103 NREGS, see National Rural Employment
MGAN, see Multiple generator GAN Guarantee Scheme
Index 313
O Predictive technology, 86
Pre-emptive Event Interleaving (PEI) scheme, 45, 53
OA, see Orthogonal array Pre-emptive stationary priority scheduling
Object detection, 120, 126 (PSPS), 34
HOG features, 122 Principal component analysis (PCA), 116–117
image classification, 120 Probability density function (PDF), 133
classification and localization, 120 Progressive GAN (ProgGAN), 139
image segmentation Proposed deep learning, 161–165
instance segmentation, 121 Proposed DWT-based PQD detection method,
semantic segmentation, 120–121 155, 157, 158–161
sliding window demerits, 121 PSPS, see Pre-emptive stationary priority
region-based methods scheduling
faster region-based convolutional neural
network, 122
fast region-based convolutional neural Q
network, 122 Qualitative feedback, 66
region-based convolutional neural
network (R-CNN), 121–122
region proposal, 121 R
you only look once (YOLO), 122 Random forest (RF), 157
demerits, 122–123 Random Sequence Insertion-based Crossover
Objective function, 182 (RSIX), 250–254
Ocean Cleanup Project, 234 Rational decision-maker, 84
On-Road Integrated Optimisation and Navigation RBM, see Restricted Boltzmann machine
(ORION), 243 RC agent (RCA), 22, 23, 25
OpenCV, 124 R-CNN, see Region-based convolutional neural
Operating system (OS), 64 network
Optical spectrum analyser (OSA), 97, 100 RDANN, see Robust Design of Artificial Neural
Optimal ANN parameters Networks
analysis stage, 184 RDF, see Reduction data filter algorithm
confirmation stage, 184 Reactive AI, 295
experimentation stage, 183 Real-time system (RTS) and scheduling theory,
planning stage, 181–183 29–34
ORION, see On-Road Integrated Optimisation algorithms for implementing synchronized
and Navigation processes/events, 29–30
Orthogonal array (OA), 179 concurrent events, 30–31
OS, see Operating system pre-emptive stationary priority scheduling
OSA, see Optical spectrum analyser (PSPS), 34
Osprey’s intelligent visual monitoring, 302 process life span, 30
response time analysis, 31–33
P Recurrent neural networks (RNN), 133
Reduction data filter (RDF) algorithm, 191–192
PatchGAN, 139 Region-based convolutional neural network
PCA, see Principal component analysis (R-CNN), 121–122
PCC, see Point of common coupling Region-based methods
PDF, see Probability density function faster region-based convolutional neural
PEI, see Pre-emptive Event Interleaving scheme network, 122
PELM, see Polyharmonic extreme learning fast region-based convolutional neural
machine algorithm network, 122
Perception, 127 region-based convolutional neural network
Pixel-to-Pixel GAN (Pix2Pix), 139, 142 (R-CNN), 121–122
Point of common coupling (PCC), 150 region proposal, 121
Poisson distribution, 19 Region of interest (RoI), 115
Polyharmonic extreme learning machine (PELM) Region proposal, 121
algorithm, 125 Regularization, 140
Power quality disturbance (PQD), 150 Relax, 63
314 Index