Baddi Y. Big Data Intelligence For Smart Applications 2022
Baddi Y. Big Data Intelligence For Smart Applications 2022
Big Data
Intelligence
for Smart
Applications
Studies in Computational Intelligence
Volume 994
Series Editor
Janusz Kacprzyk, Polish Academy of Sciences, Warsaw, Poland
The series “Studies in Computational Intelligence” (SCI) publishes new develop-
ments and advances in the various areas of computational intelligence—quickly and
with a high quality. The intent is to cover the theory, applications, and design methods
of computational intelligence, as embedded in the fields of engineering, computer
science, physics and life sciences, as well as the methodologies behind them. The
series contains monographs, lecture notes and edited volumes in computational
intelligence spanning the areas of neural networks, connectionist systems, genetic
algorithms, evolutionary computation, artificial intelligence, cellular automata, self-
organizing systems, soft computing, fuzzy systems, and hybrid intelligent systems.
Of particular value to both the contributors and the readership are the short publica-
tion timeframe and the world-wide distribution, which enable both wide and rapid
dissemination of research output.
Indexed by SCOPUS, DBLP, WTI Frankfurt eG, zbMATH, SCImago.
All books published in the series are submitted for consideration in Web of Science.
© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature
Switzerland AG 2022
This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether
the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by similar
or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
For a large majority of companies, the value of data is no longer in doubt. Calculating
an application’s usage rates, measuring the time spent on social networks, or even
analyzing its usage are now some of the best practices accepted and implemented
by the most significant number. But while companies understand the importance of
collecting and analyzing their data, few know how to process and use it. However,
knowing how to use this data is a prerequisite for integrating Artificial Intelligence
into its products and services.
It is worth noting that technologies such as Artificial intelligence and Big Data
are evolving even faster. They are rapidly growing by holding great promise for
many sectors. The real revolutionary potential of these technologies mainly relies on
their convergence. Big Data and Artificial Intelligence are two technologies that are
inextricably linked, to the point that we can think about Big Data Intelligence. AI
has become ubiquitous in many industries where intelligent programs relying on big
data transform decision-making. Increased agility, smarter business processes, and
better productivity are the most likely benefits of this convergence.
Big data, which is still poorly exploited, is nevertheless the black gold of AI. Arti-
ficial Intelligence is the logical continuation of our data analysis methods and tech-
niques, an extension of Business Intelligence, followed by Big Data and Advanced
Analytics. The so-called intelligent machine needs a massive amount of data analyzed
and cross-referenced to draw innovative, even creative, capabilities close to the
human brain’s functioning.
Big Data Intelligence for Smart Applications comprises many state-of-the-art
contributions from scientists and practitioners working in big data intelligence and
smart applications. It aspires to provide a relevant reference for students, researchers,
engineers, and professionals working in this area or those interested in grasping its
diverse facets and exploring the latest advances in machine intelligence and data
analytics for sustainable future smart cities applications. More specifically, the book
contains 15 chapters. Accepted chapters describe and analyze various applications of
machine intelligence and big data for smart applications, such as smart cities, Internet
of Things, cybersercurity, vision systems, and healthcare, to mitigate COVID-19.
v
vi Preface
We want to take this opportunity and express our thanks to the contributors
of this volume and the editorial board for their tremendous efforts by reviewing
and providing interesting feedbacks. The editors would like to thank Dr. Thomas
Ditsinger (Springer, Editorial Director, Interdisciplinary Applied Sciences) and Prof.
Janusz Kacprzyk (Series Editor-in-Chief), and Ms. Swetha Divakar (Springer Project
Coordinator) for the editorial assistance and support to produce this important scien-
tific work. Without this collective effort, this book would not have been possible to
be completed.
vii
viii Contents
ix
x About the Editors
Big Data security, Recommendation Systems, and Cloud computing models for data
privacy.
Loai Tawalbeh completed his Ph.D. degree in Electrical & Computer Engineering
from Oregon State University in 2004, and M.Sc. in 2002 from the same university
with GPA 4/4. Dr. Tawalbeh is currently an Associate professor at the department
of Computing and Cyber Security at Texas A&M University-San Antonio. Before
that he was a visiting researcher at University of California-Santa Barbra. Since
2005 he taught/developed more than 25 courses in different disciplines of computer
engineering and science with focus on cyber security for the undergraduate/graduate
programs at: NewYork Institute of Technology (NYIT), DePaul’s University, and
Jordan University of Science and Technology. Dr. Tawalbeh won many research
grants and awards with over than 2 Million USD. He has over 80 research publications
in refereed international Journals and conferences.
Data Quality in the Era of Big Data:
A Global Review
Abstract Currently, Big Data is gaining wide adoption in the digital world as a new
technology able to manage and support the explosive growth of data. Indeed, data is
growing at a higher rate due to the variety of the data-generating adopted devices. In
addition to the volume aspect, the generated data are usually unstructured, inaccurate,
and incomplete, making its processing even more difficult. However, analyzing such
data can provide significant benefits to businesses if the quality of data is improved.
Facing the fact that value could only be extracted from high data quality, companies
using data in their business management focus more on the quality aspect of the
gathered data. Therefore, Big data quality has received a lot of interest from the
literature. Indeed, many researchers have attempted to address Big data quality issues
by suggesting novel approaches to assess and improve Big data quality. All these
researches inspire us to review the most relevant findings and outcomes reported in
this regard. Assuming that some review papers were already published for the same
purpose, we believe that researchers always need an update. It is worth noting that
all the published review papers are focused on a specific area of Big data quality.
Therefore, this paper aims to review all the big data quality aspects discussed in the
literature, including Big data characteristics, big data value chain, and big data quality
dimensions and metrics. Moreover, we will discuss how the quality aspect could be
employed in the different applications domains of Big data. Thus, this review paper
provides a global view of the current state of the art of the various aspects of Big
data quality and could be used to support future research.
Keywords Big Data Quality · Big Data Quality Approach · Big Data Value
Chain · Big Data Quality Dimensions · Big Data Characteristics
1 Introduction
With the digital transformation of companies, data has become the foundation on
which many organizations rely to differentiate themselves and develop a sustainable
competitive advantage. However, this data-driven strategy requires a solid infras-
tructure to collect, store, and large-scale process data. Furthermore, as the number
of devices generating data increases at a high rate, the volume of that induced data
increases. According to a study conducted by the research firm IDC, the world’s data
will grow from 23 zettabytes in 2017 to 175 zettabytes (175 billion terabytes) by 2025
(Reinsel et al. 2018). Thus, Big data has emerged as a new data ecosystem able to
support, process, and manage this explosive growth of data.
Big data may be defined as “the way we gather, store, manipulate, analyze, and
get insight from a fast-increasing heterogeneous data” (Taleb et al. 2018). Indeed,
big data is multi-source data. It is a mix of structured and unstructured data that
combines various data types, usually noisy and inaccurate. Hence, in big data, the
collected data is often poor-quality data that cannot be directly used, cleaned, and
transformed into useful information. For this reason, data quality has always been
a critical concern of companies believing that data is only helpful if it is of high
quality.
Indeed, data quality is a broad and multifaceted concept that can impact several
aspects of a company’s business. According to a survey conducted by TDWI
(Abdullah et al. 2015; TDWI 2021), the three most common problems caused by
low data quality are: the time lost to validate and fix data errors, the extra cost, and
the unreliability of the extracted data, so managers get less confident in analytics
systems and cannot trust the generated reports. Furthermore, Big data is highly used
by companies to analyze customer’s behavior and understand their insights. A survey
about the benefits of high-quality data conducted by TDWI has shown that the most
common service derived from high-quality data is customer satisfaction, a priority
for marketers (Abdullah et al. 2015). Indeed, understanding customer insights and
needs allows companies to focus their efforts on the right customer segment and gain
a competitive advantage. However, not ensuring a high quality of data in such a case
will negatively impact the company’s strategic planning. It may lead managers to
make the wrong decisions instead of the right ones.
Data quality was introduced long before big data emerged and was applied first
to traditional data stored in relational databases. However, big data have emerged
new quality issues that traditional data quality techniques cannot cope with. In Batini
et al. (2015), have shown that the evolution from data quality to big data quality is
related to the fact that big data involves various new data types, data sources, and
application domains. Thus, the concept of big data quality has been addressed by
many studies that have attempted to suggest new approaches to assess and improve
Data Quality in the Era of Big Data: A Global Review 3
big data quality. Therefore, this paper aims to review the most recent research focused
on Big data quality to explore the current state of the art and highlight some possible
future work directions. As we have mentioned earlier, the purpose of this paper is
to cover all Big data quality aspects addressed in the existing literature. Thus, each
section of this paper is intended for a specific Big data quality aspect.
The rest of this paper is organized as follows: Section 2 defines the Big data charac-
teristics introduced in the literature. Section 3 highlights big data quality dimensions
and metrics. Section 4 describes the different phases of the big data value chain and
presents the quality dimensions that should be considered in each step. In Sect. 5,
we present the adopted research methodology. Section 6 summarizes and provides a
classification of the most recent Big data quality improvement approaches followed
by the key findings that emerged out of the review. Section 7 describes big data
application domains where data quality is of critical concern. Finally, we discuss
some remaining Big data quality issues and highlight some possible future work
directions.
While it is true that the term big data refers to a massive amount of data that are
continuously increasing, this does not necessarily mean that big data is only about
data volume. Indeed, implementing a big data strategy involves storing, analyzing,
processing, and monitoring data. Therefore, numerous challenges are raised, such as
handling and visualizing data, integrating data from various data sources, securing
data, and so on. In Espinosa et al. (2019), have reviewed all Big data challenges
addressed in the existing literature and have shown that the high growth rate, the lack
of tools, and data management challenges are the most common Big data issues. This
variety in Big data issues is related to big data’s nature and its multiple characteristics,
a trending research topic.
The first Big data characteristics were introduced in 2001 by Doug Laney that have
defined big data in terms of the 3 V’s: Volume, Velocity, and Variety. Afterward, SAS
(Statistical Analysis System) has introduced two new characteristics, i.e., Variability
and complexity. Furthermore, Oracle has defined big data in 4 V’s, i.e., Volume,
Velocity, Variety, and Value. In 2014, Data Science Central, Kirk Born described big
data in 10 V’s, i.e., Volume, Variety, Velocity, Veracity, Validity, Value, Variability,
Venue, Vocabulary, Vagueness (Failed 2016a). Later, more V’s were suggested in
the literature, such as Vulnerability, Virality, Verbosity, etc. In 2018, the number
of proposed characteristics already exceeded 50. According to a study presented in
Dhamodharavadhani et al. (2018), this number will increase to reach 100 V’s by
2021.
It is worth noting that not all these characteristics are used at once. Instead, big
data characteristics were explicitly introduced to meet some application requirements
and are therefore only used in particular big data application domains.
4 W. Elouataoui et al.
In Pros (2013), the authors have highlighted the emerging Big data characteristics.
In addition, they have suggested the big data V’s that should be considered in some
big data application domains for better data analytics, such as social media, IoT, and
weather prediction. In Table 1, we explain the most common characteristics of Big
Data and highlight their impact on Big Data quality.
Ensuring the quality of data with such characteristics is a significant challenge.
The quality issues faced when dealing with big data are highly related to big data
V’s. Thus, data quality aspects should be reviewed and adapted to deal with the new
challenges raised by Big data characteristics. In the next section, we highlight how
data quality could be ensured in a Big data context and the quality dimensions that
should be considered.
Given the particular characteristics of big data, new challenges are introduced and
need to be explored. Indeed, the emergence of big data has raised a new challenge
related to understanding Big data quality requirements. Thus, before implementing
any quality approach, a deep understanding of the vital quality requirements must be
considered when managing big data solutions. In Failed (2016b), the authors have
suggested a novel approach to define Big data systems’ quality requirements. The
proposed method consists of associating each data characteristic with all quality
attributes, which leads to a global assessment of data quality without missing any
quality aspect. To better understand Big data quality requirements, we should talk
about its properties called Data Quality Dimensions (DQD’s).
Data quality dimensions (DQD’s) could be defined as “a set of data quality
attributes that represent a single aspect or construct of data quality” (Ramasamy and
Chowdhury 2020). The most common data quality dimensions (DQD’s) are Accu-
racy, Completeness, Consistency, Freshness, Uniqueness, and Validity. However,
more data quality dimensions were suggested in the literature to cover other data
quality aspects, such as security, traceability, navigation, and data decay.
Any data quality dimension can be quantified using metrics. Indeed, data quality
metrics allow measuring how data meet the quality requirements and provide a
concrete meaning to quality aspects. In Arolfo and Vaisman (2018), have presented
the mathematical definition of each quality dimension. They have defined data quality
metrics for dimension D as “a function that maps an entity to a value, such that this
value, typically between 0 and 1, indicates the quality of a piece of data regarding
the dimension D.”
It is worth noting that there is a high dependency between data quality dimensions
strongly related to each other. Thus, focusing only on one size without considering
its correlation with the different measurements may not be a practical approach for
supporting data quality. In Sidi et al. (2013), the authors have thrown light on the
existing relation between DQD’s. They have suggested a framework that allows
Data Quality in the Era of Big Data: A Global Review 5
Table 1 (continued)
Characteristic Elucidation Meaning Impacts on data quality
Visualization How is it challenging to It refers to the process of Visualizing a big scale of
develop a meaningful presenting the extracted data in a static or
visualization of data? information in a readable dynamic form may be
form, more comfortable challenging. In Agrawal
to understand et al. (2015), Agrawal
et al. have presented the
challenges and
opportunities of big data
visualization
Validity Is data suitable and Refers to the correctness Data Validity should be
accurate for the intended of the input data and its inspected before data
use? suitability for its processing as invalid data
intended purpose are useless and cannot
support decision making
Volatility How long could data be Refers to the duration of Outdated data are
kept and considered as usefulness and relevancy irrelevant and may bias
applicable? of data data analysis
Viability How relevant is the data? This process consists of “Big Data should have
“uncovering the latent, the capability to be live
hidden relationships and active forever, and
among these variables.“ able for developing, and
to “confirm a particular to produce more data
variable’s relevance when need. Khan et al.
before investing in the (2018)
creation of a
fully-featured model.
IRJET (2021)
Vocabulary What terminology is Refers to data Data readability and
used to describe data? terminology that availability are highly
describes data’s content related to the vocabulary
and structure and the wording used
Vagueness How much data is Confusion over the Unclear and
available? meaning of big data. It is non-understandable data
about its nature, content, could not be used
tools availability, effectively
tools selection, etc.
(Dhamodharavadhani
et al. 2018)
Venue What kind of platforms Various types of data Big data sources are not
and systems are used to arrived from different always credible and may
manage data? sources via different provide inaccurate
platforms like personnel information (e.g., social
systems and private & media)
public cloud
(IRJET-2021)
(continued)
Data Quality in the Era of Big Data: A Global Review 7
Table 1 (continued)
Characteristic Elucidation Meaning Impacts on data quality
Viscosity How difficult is it to Refers to the difficulty to Big Data Complexity
work with data? use or integrate the data may have a direct impact
(Failed 2018) on data usability and
effectiveness
The data value chain was introduced following the digital transformation as a process
that describes data flow within data-driven businesses and allows it to extract data’s
value. However, with the emergence of big data, the data value chain has shown severe
weaknesses in the face of the increasing volume, variety, and veracity of big data. It
thus has become inappropriate for Big data systems. Therefore, the big data value
chain (BDVC) was introduced as an extension of the data value chain that fits big
data requirements, addresses Big data characteristics, and allows the transformation
of raw data into valuable insights (Faroukhi et al. 2020).
Data Quality in the Era of Big Data: A Global Review 9
Fig. 2 Big data value chain model proposed by Faroukhi et al. (2020)
Many big data value chain models were suggested in the literature (Alaoui et al.
2019b; Failed 2013; Curry 2015). In the following, we describe the different phases
of the most exhaustive and most recent BDVC in the literature proposed by Faroukhi
et al. (2020). The BDVC can be divided into seven steps, as shown in Fig. 2. Finally,
we present the quality issues that should be addressed and the quality dimensions to
consider for each phase.
This first step refers to the process of producing data. The generated data may be
structured, semi-structured, or unstructured depending on the source type (human-
generated, approach mediated, machine-generated) (Arolfo and Vaisman 2018).
Then, data are gathered and recorded through an acquisition process.
This step consists of collecting data from various sources and storing it before
bringing any changes to the original data. As the gathered data could be of multiple
sources and types, it is essential to identify the integrity and variety of data sources
since each data type is processed differently.
Data sources’ accuracy must be considered in this phase, and the security restric-
tions must be identified (Failed 2014). In addition, the freshness of the collected data
must be considered, especially if the data source is interactive and dynamic, such as
social media data (Alaoui et al. 2019a).
The preprocessing phase is a crucial step in BDVC. It allows cleaning the data gath-
ered in the previous stage from inconsistencies and inaccurate values, thus trans-
forming the data into a valuable and practical format. One of the most common
issues that may be faced in this phase is noisy and incomplete data. This could be
resolved using clustering algorithms and some data filtering techniques (Zhang et al.
10 W. Elouataoui et al.
2017). In this phase, data quality is highly improved due to the multiple subphases
that this phase involves, such as data transformation, data integration, data cleaning,
and data reduction (Failed 2017):
• Data transformation: Refers to the process of converting the structure of data into
a more appropriate format to make it more valuable.
• Data integration: Consists of combining various types of data from multiple data
sources into a unified dataset, easy to access.
• Data cleaning: The process of ensuring that data is correct and free from anomalies
by removing the corrupted and duplicated values.
• Data reduction: This process identifies a model representing the original data in
the simplest possible form to reduce the required storage capacity. This process
includes multiple techniques, such as data noise reduction, data compression, and
other dimensionality reduction techniques (Failed 2017).
Many quality dimensions must be considered in this phase, such as completeness,
consistency, uniqueness, and validity (Alaoui et al. 2019b; Failed 2014).
After the preprocessing phase, data is stored to keep a clean copy of data for further
use. However, it is worth noting that the storage system could also affect data quality.
Indeed, if a Big data system requires real-time analysis, data timeliness could have
a high impact on the freshness and the relevancy of the stored data. In such a case,
quality aspects such as storage penalty, storage currency, and availability should be
considered in this phase (Alaoui et al. 2019b). Finally, after data storage, data can
be processed using the appropriate tools.
Data processing is the most critical stage of the whole process. It consists of analyzing
and manipulating data previously cleaned to identify the unknown correlations and
patterns and transform data into proper knowledge. Different big data analysis tech-
niques could be used in this phase, such as machine learning, deep learning, and data
mining.
One of the most common challenges faced in this stage is to ensure the reliability
of the obtained results and improve the prediction capabilities’ accuracy. This issue
could be addressed using more accurate algorithms and models such as Bayesian
classifiers, association rule mining, and decision trees (Zhang et al. 2017).
In this phase, it is difficult to identify the quality aspects that should be considered
since the data processing is positively related to the context of the analysis and the
intended use. However, some standard quality dimensions were suggested by many
Data Quality in the Era of Big Data: A Global Review 11
researchers to ensure a high quality of data during this phase, such as accuracy,
validity, and usability (Alaoui et al. 2019b; Failed 2014; Serhani et al. 2016). In
Failed (2014), the authors have suggested three general principles that should be
considered when assessing data quality:
• The system that performs the analysis should not influence the outcome.
• A version control system must be put in place.
• The quality assessment results must be checked against predefined quality criteria.
The analysis results are visualized in a readable form in this phase, easier to under-
stand, using visual elements such as graphs, maps, and dashboards. This would help
managers to explore data more efficiently and support them in their decision-making.
This phase has received less attention in the literature than the other phases of
BDVC. This stage’s quality aspects are more holistic and related to user satisfaction
and quality of data representation (Serhani et al. 2016). In Failed (2014), the authors
have introduced new dimensions to assess the quality of this last process, such as
complexity, clarity, and selectivity.
This last step consists of sharing and exposing insights and value generated
throughout each phase of the BDVC. For example, the extracted insights could be
used to sell specific services, shared with the public as open data, or shared internally
to improve organizational performance (Faroukhi et al. 2020).
Each phase of the big data value chain applies some changes to the data. Therefore,
it is essential to perform a continuous data quality assessment and monitor BDVC to
ensure that these changes are not affected. Thus, in the following figure, we present
the different data quality dimensions that should be considered in each phase of the
BDVC (Fig. 3).
Many approaches were suggested in the literature to improve the different DQD’s
throughout the BDVC. In the next section, we report the most recent quality
approaches introduced and classify these approaches based on multiple criteria.
5 Research Methodology
Fig. 3 The projection of the most common DQD’s on the different phases of the BDVC
• Planning the review: identifying the need for a review and preparing a review
proposal.
• Conducting the review: searching and selecting articles.
• Reporting the review: reporting the main findings and making conclusions.
To conduct the review, primary research was performed using first, broader
keywords such as “Big Data Quality,” “Big Data Quality Approach,” “Big Data
Quality Framework,” and then more detailed keywords such as “Big data Accuracy,”
“Big Data Consistency” and “Big data completeness.” The research was performed
on the following online libraries:
• IEEE Xplore (http://ieeexplore.org/)
• ACM digital library (http://dl.acm.org/)
• Science Direct (http://www.sciencedirect.com/)
• Springer (http://www.springer.com/gp/)
• Research Gate (https://www.researchgate.net/).
The papers were firstly filtered out based on their titles. Thus, papers with titles
that do not contain the selected keywords were excluded leaving 105 studies. Then,
we excluded irrelevant papers based on their abstracts, and 40 articles were kept for
our research. Finally, we included articles that were:
• Proposing approaches to assess or improve Big data quality.
• Proposing methods applied to one or different phases of the BDVC.
• In English.
• Available.
• Published in digital libraries.
Data Quality in the Era of Big Data: A Global Review 13
This section goes over the most recent and available researches that have suggested
novel approaches to assess and improve Big data quality throughout the different
phases of the BDVC. A summary of these studies is described in Table 2, followed by
the key findings that emerged out of the review. The selected papers were categorized
based on multiple criteria, namely:
The addressed BDVC phases: The most addressed phases of the BDVC are
the preprocessing and the processing phases. Indeed, few studies address the whole
BDVC. This is linked to the fact that most data processing is carried out during these
two phases, especially in big data systems.
The considered DQD’s: The most considered DQD’s are Accuracy, Complete-
ness, Consistency, and Timeliness. These quality dimensions are standard for any
data quality assessment and are of significant importance whether in a big data context
or not.
The addressed application domain: Out of the 22 studies reviewed, only seven
focus on a specific big data application domain. The most addressed big data domains
are social media and healthcare.
The used techniques: In a big data context, assessing the quality of the whole
data is costly in time, money, and effort. Thus, most of the selected studies use
data sampling for data quality assessment, which allows them only to analyze a
representative sample of data and generalize the obtained results to the entirety of
data. In addition, machine learning is yet another widely used technique to measure
data quality and extract valuable insights from data.
Based on the available studies, the reviewed approaches presented above lead
us to raise some challenging issues that need more research focus and which we
summarize in the following points:
Table 2 A summary of Big data quality approaches
14
References Year Phases DQD’s Domain Main idea and findings Used Techniques
Soni and Singh 2021 Preprocessing Accuracy, Reliability, – An approach that aims to reduce the Data sampling, Quality
(2021) Timeliness computation time and space could be measures
used when applying quality measures
to unstructured data. This approach
relies on data sampling and allows
measurement: a “believability factor,”
time is taken to execute the sampled
dataset, and the Mean Absolute Error
of the sampled dataset
Failed (2020) 2020 The whole BDVC Credibility and – It aims to improve data quality and A data governance
Accuracy especially data credibility using a model
new data governance model based on
data traceability that considers the
different ownership of data
Faroukhi et al. 2020 The whole BDVC Accuracy, – A unified approach that that extends Security Functions,
(2020) Credibility, BDVC with security and quality Orchestration process
Completeness, aspects. According to three
Readability, processes, the suggested framework
Freshness, was perceived: data quality, data
Consistency, Storage security, and orchestration. Thus, in
Penalty, Uniqueness, each phase of the BDVC, the quality
Conformity, aspects, the security functions, and
Referential Integrity, the orchestration processes to
Normalization consider were presented
(continued)
W. Elouataoui et al.
Table 2 (continued)
References Year Phases DQD’s Domain Main idea and findings Used Techniques
Faroukhi et al. 2020 The whole BDVC Shareability – The authors have suggested a generic Cloud computing and
(2021) model that includes data Analytical capabilities
monetization with the Big Data Value
Chain. The proposed model aims to
monetize and share insights and data
as a service at different levels of data
maturity
I. E. Alaoui et Y. 2019 The whole BDVC Accuracy, Freshness, Social media It aims to show the impact of the Using dynamics and
Gahi (2019) Consistency, Storage quality metrics on sentiment analysis contextual dictionaries
Penalty, Uniqueness, approaches. The method used based on hashtags,
Conformity, Integrity consists of grouping tweets into choosing the right
multiple classes using hashtags keywords for data
collect
A. Juneja et. and 2019 Preprocessing Accuracy, Traceability It consists of defining the data quality Auto-discovery quality
Das (2019) Completeness, Monitoring rules and performing data profiling. rules
Data Quality in the Era of Big Data: A Global Review
References Year Phases DQD’s Domain Main idea and findings Used Techniques
Xu et al. (2020) 2019 Preprocessing Accuracy – It aims to clean data from unreliable The local outlier factor
data using an incorrect data detection algorithm
method based on an extended
application of local outlier factor
(LOF)
Khaleel et and 2019 Acquisition and Completeness and – It aims to improve data quality by Machine learning,
Hamad (2019) Preprocessing Consistency converting unstructured data to Fragmentation
structured data using a metadata algorithms
approach, distributed data file system
(Fragmentation algorithm), and
quality assurance algorithms
Taleb et al. (2019) 2019 Preprocessing Accuracy, – A novel data quality profiling model Big data sampling and
Completeness, that inspects the different data quality data quality profiling
Timeliness dimensions and relies on a quality algorithms
profile repository that contains all
information related to data quality
such as data quality profile, data
quality dimensions, …
Arolfo and 2018 The whole BDVC Readability, Social Media · Data quality is improved when data Apache Kafka
Vaisman (2018) Completeness, is filtered Zookeper
Usefulness ·Data quality is affected by the
number of data properties considered
in the study
(continued)
W. Elouataoui et al.
Table 2 (continued)
References Year Phases DQD’s Domain Main idea and findings Used Techniques
Taleb et al. (2018) 2018 Preprocessing Accuracy, – It consists of assessing the quality of Data sampling and
Completeness unstructured data using a data quality profiling
repository where each data type is Data mining
linked to the appropriate DQD’s and
features extraction methods
Cappiello et al. 2018 Preprocessing Completeness, Transportation It provides a context-aware Data profiling
2018) Consistency, evaluation of data quality based on
Distinctness, data profile information and user
Timeliness requirements
Zan and Zhang 2018 Preprocessing Accuracy, Integrity, Healthcare It consists of performing a Analytic hierarchy
(2018) Normative, credibility test by removing process (AHP)
Consistency, unreliable data before embarking on
Timeliness the data quality assessment
Ardagna et al. 2018 Preprocessing Accuracy, – An adaptable approach that allows Data Sampling
Data Quality in the Era of Big Data: A Global Review
References Year Phases DQD’s Domain Main idea and findings Used Techniques
Wahyudi et al. 2018 The whole BDVC A generic approach – A generic process pattern model that Data access,
(2018) could be used as a reference for preprocessing, and
improving Big data quality. The analysis techniques
proposed model contains four main
steps (discover, access, exploit,
analyze). In addition, four variants of
the proposed pattern addressing
different quality issues were defined
to tackle data with variable quality
Failed (2017) 2017 Preprocessing Relevancy, Uniqueness – It aims to increase data quality and Machine learning
reduce data degradation in the
preprocessing phase using a new
algorithm that combines two feature
selection strategies: RELIEF
algorithm and mRMR algorithm
Taleb and Serhani 2017 Preprocessing Accuracy, – It consists of defining quality rules by Big data quality rules
(2017) Completeness, associating the quality scores that do discovery algorithm
Consistency not meet the quality requirements
with the appropriate preprocessing
activities. Then, the quality rules are
evaluated and optimized
Serhani et al. 2016 The whole BDVC Accuracy, Correctness, Healthcare It consists of performing a Data mining
(2016) Completeness data-driven quality assessment before Machine learning
and after the preprocessing phase
(continued)
W. Elouataoui et al.
Table 2 (continued)
References Year Phases DQD’s Domain Main idea and findings Used Techniques
Taleb et al. (2016) 2016 Preprocessing A generic approach – A data quality assessment approach Data sampling
consists of 4 main steps: data
sampling, data profiling, quality
evaluation using a quality assessment
algorithm, and data quality analysis.
The suggested method uses
representative data samples to reduce
the processing time
Taleb et al. (2015) 2015 Preprocessing A generic approach – A quality framework to support data Data profiling,
quality profile generation for the A preprocessing
preprocessing phase. The suggested algorithm
framework consists of the following
main components: a data quality
profile selection, a data quality
profile adapter, quality control, and a
Data Quality in the Era of Big Data: A Global Review
monitoring module
Al-Hajjar et al. 2015 Preprocessing Timeliness, usefulness, Social Media A framework identifies the issues of Social media data
(2015) availability, quality analysis of big data on social analytics techniques
consistency, accuracy media. It provides a mapping
between big data analysis techniques
and the satisfied quality control
factors for several social media
websites
19
20 W. Elouataoui et al.
• As we have mentioned earlier, the preprocessing and the processing phases have
received more attention in the literature than the other phases of the BDVC. This is
because these phases are the most critical phases of the BDVC. However, this does
not mean that the different phases could not impact the quality of the obtained
results. Therefore, the existing approaches should be extended to englobe all
BDVC phases to ensure that data quality is not degraded at any big data life cycle
stage.
• One of the most critical challenges to be addressed by researchers is exploring and
defining new quality dimensions related to big data rather than focusing only on
the standard quality dimensions such as accuracy, completeness, and consistency.
Thus, more DQD’s such as security, traceability, and clarity should be addressed in
the literature. Also, the existing dependencies between DQD’s should be highly
considered, as focusing on a specific data quality aspect may impact the other
DQD’s.
• Adopting a more business-oriented approach while assessing data quality is yet
another challenging issue to be addressed. Indeed, there is a Big lack of more
focused research that addresses Big data quality issues depending on the context in
which data is used. Thus, the proposed approaches should be adapted to meet the
quality requirements of the use of big data. This adaptation should be considered
while building the data quality assessment framework and should not be limited
to the experimentation phase.
• Addressing more big data V’s is another exciting research direction to explore.
Unfortunately, most of the reviewed approaches address only the most common
big data characteristics such as Volume and Variety.
• The existing quality assessment frameworks are usually implemented using data
sampling, data profiling, and machine learning techniques. However, the current
assessment algorithms address only the quality rules defined before the assessment
and could be extended to predict potential anomalies accurately.
As we have mentioned earlier, the BDVC must be redefined to address the partic-
ular characteristics of each Big data application domain. Therefore, the following
section presents some big data application domains addressed in the literature in a
Big data quality improvement context.
Today, many sectors like healthcare, manufacturing, and transportation use big data
analytics for data management. However, extracting value from big data remains a
challenge for most industries that use big data. As each application domain has its
characteristics, Big data quality issues are also related to the requirements of big data
use. Therefore, the standard quality approaches could only be used as a roadmap and
adapted to fit each sector’s characteristics. Many studies that have tackled big data
quality have focused on a specific application domain of big data.
Data Quality in the Era of Big Data: A Global Review 21
One of the most common big data application domains addressed in the literature
is social media. Indeed, many consumers use social media to share their thoughts
and emotions. Therefore, social media data has become the primary source of data
analysis techniques to understand customer insights, such as sentiment analysis and
opinion mining. However, the extracted knowledge from such kind of data is not
entirely reliable. Thus, these analysis techniques must be accompanied by a contin-
uous data quality assessment for better decision-making. In Faroukhi et al. (2021),
El Alaoui et al. have highlighted the essential Big data quality metrics that should be
considered in each phase of the BDVC. Also, they have shown that the quality aspect
of data directly impacts the reliability and performance of a sentiment analysis (SA)
approach.
Moreover, the authors have measured the impact of considering each data quality
metric on the accuracy of the SA model. The authors have used a quality-oriented
approach that uses dynamics and contextual dictionaries based on hashtags. The
suggested method was tested for the US 2016 election forecasting using Twitter
data. In the same context, Pääkkönen et al. (2017) have proposed a new reference
architecture for assessing social media data quality. The suggested approach extends
previous research. The main contribution of this study is the extension and evalu-
ation of the metadata management architecture. The proposed model consists of 3
components: quality management by defining quality rules, metadata management
consisting of creating metadata, and quality evaluation using quality metrics defined
by the quality rules.
Healthcare is yet another area where big data is widely used. Indeed, adopting
big data analysis in healthcare allows avoiding preventable diseases, supporting clin-
ical decision-making, and improving patient care. However, medical data is gener-
ally multi-source, unstructured, and heterogeneous. Thus, such kinds of data cannot
be directly analyzed and processed. In Zan and Zhang (2018), the authors have
suggested a 7-step methodology for assessing medical data quality. This study’s
main idea consists of performing a credibility test by removing unreliable data before
embarking on the data quality assessment. Ensuring the high credibility of data will
significantly affect the quality assessment process afterward, not only by improving
the quality of the studied data but also by reducing data volume, a highly valuated
factor in the Big data domain.
Moreover, the authors have presented an experimental simulation to check the
credibility of medical data. In Serhani et al. (2016), have suggested a novel process
to improve Big data quality and have assessed the proposed approach using a Sleep
Heart Health Study dataset. This evaluation consists of two parts: In the first part,
the authors have computed each attribute’s completeness and consistency values.
Then, they have visualized the improvement percentage of these metrics after the
preprocessing phase using graphical charts. A second part was intended to eval-
uate the processing phase’s quality based on some metrics such as accuracy and
processing time. Data processing was performed using many classification methods;
graphical charts were drawn to compare each classification model’s accuracy and
timely processing. Also, in Juddoo and George (2018), the authors have attempted
22 W. Elouataoui et al.
to define the most relevant quality dimensions in the healthcare domain using Latent
Semantic Analysis (LSA) to extract similarities between documents.
Banking is also one of the business domains where big data is making significant
changes. However, compromising between data quality and data security remains
a significant challenge in this sector. In Talha et al. (2019), have shed light on the
existing tradeoff between data quality and safety. The authors have analyzed some
points of convergence and divergence between data quality and data security in this
study. Indeed, data quality and data security share some common properties, such as
accuracy and integrity. However, there are some cases where data quality and data
security cannot coexist. For instance, data security tends to restrict access to data,
while improving data quality requires full access to information.
While multiple Big data quality approaches were suggested in the literature,
this paper’s big data application domains are not entirely explored and need more
research. In addition, other big data application domains such as education, manu-
facturing, and government have never been addressed in the literature in a Big data
quality improvement context.
8 Conclusion
Nowadays, big data is highly used for trade analytics by companies that use data to
make decisions about their business. However, big data analytics could not be of great
value if data is of poor quality. As the extracted data are usually unstructured and
inaccurate, a data quality assessment is required when processing big data. This paper
has reviewed the quality issues raised by big data V’s and how these issues could be
addressed using appropriate data quality dimensions throughout the whole BDVC.
The existing quality approaches suggested in the literature, and their correlation
with Big data application domain requirements are also discussed in this paper. This
review led us to identify the following open research issues and challenges that can
be further pursued in this area:
• The suggested quality approaches should be extended to address more big data
V’s.
• The standard quality dimensions should be redefined to fit Big data quality
requirements
• The quality assessment should not be limited to the preprocessing and the
processing stages and should englobe all BDVC phases.
• There is a Big lack of more focused research that address the characteristics of
the specific context of the use of big data
• Other big data application domains such as education, manufacturing, and
government should be addressed in the literature.
• New algorithms should be developed to discover the hidden anomalies and
troubleshoot data mismatches.
Data Quality in the Era of Big Data: A Global Review 23
References
N. Abdullah, S. A. Ismail, S. Yuhaniz, S. Mohd sam. Data Quality in Big Data: a Review, vol. 7,
pp. 16–27, Jan. 2015
R. Agrawal, A. Kadadi, X. Dai, F. Andres, Challenges and Opportunities with Big Data
Visualization. (2015). https://doi.org/10.1145/2857218.2857256
F. Arolfo, A. Vaisman, Data Quality in a Big Data Context: 22nd European Conference, ADBIS
2018, Budapest, Hungary, September 2–5, 2018, Proceedings, pp. 159–172 (2018). https://doi.
org/10.1007/978-3-319-98398-1_11
I.E. Alaoui, et Y. Gahi, The impact of big data quality on sentiment analysis approaches. Procedia
Comput. Sci. 160, 803–810, janv. 2019. https://doi.org/10.1016/j.procs.2019.11.007
D. Al-Hajjar, N. Jaafar, M. Al-Jadaan et, R. Alnutaifi, Framework for social media big data quality
analysis, in New Trends in Database and Information Systems II, ed. by N. Bassiliades, M.
Ivanovic, M. Kon-Popovska, Y. Manolopoulos, T. Palpanas, G. Trajcevski, et A. Vakali (Springer,
International Publishing, Cham, 2015), pp. 301–314. https://doi.org/10.1007/978-3-319-10518-
5_23
D. Ardagna, C. Cappiello, W. Samá, et M. Vitali, Context-aware data quality assessment for big
data. Future Gen. Comput. Syst. 89, 548–562, déc. 2018. https://doi.org/10.1016/j.future.2018.
07.014
C. Batini, A. Rula, M. Scannapieco, G. Viscusi, From data quality to big data quality. J. Database
Manag. 26(1), 60–82 (Jan. 2015). https://doi.org/10.4018/JDM.2015010103
E. Curry, The big data value chain: definitions, concepts, and theoretical approaches, in New Hori-
zons for a Data-Driven Economy: A Roadmap for Usage and Exploitation of Big Data in Europe
(2015). https://doi.org/10.1007/978-3-319-21569-3_3
C. Cappiello, W. Samá, M. Vitali, Quality awareness for a successful big data exploitation, in
Proceedings of the 22nd International Database Engineering & Applications Symposium on—
IDEAS 2018 (2018). https://doi.org/10.1145/3216122.3216124
S. Dhamodharavadhani, G. Rajasekaran, R. Ramalingam, Unlock Different V’s of Big Data for
Analytics (2018)
J. Espinosa, S. Kaisler, F. Armour, W. Money, Big Data Redux: New Issues and Challenges Moving
Forward. (2019). https://doi.org/10.24251/HICSS.2019.131
I. El Alaoui, Y. Gahi, R. Messoussi, Big data quality metrics for sentiment analysis approaches, in
Proceedings of the 2019 International Conference on Big Data Engineering (2019). https://doi.
org/10.1145/3341620.3341629
I. El Alaoui, Y. Gahi, R. Messoussi, Big Data Quality Metrics for Sentiment Analysis Approaches,
p. 43 (2019). https://doi.org/10.1145/3341620.3341629
A. Faroukhi, I. El Alaoui, Y. Gahi, A. Amine, Big data monetization throughout Big Data Value
Chain: a comprehensive review. J. Big Data 7, 3 (2020). https://doi.org/10.1186/s40537-019-
0281-5
A. Faroukhi, I. El Alaoui, Y. Gahi, et A. Amine, Big Data Value Chain: A Unified Approach for
Integrated Data Quality and Security, p. 8 (2020). https://doi.org/10.1109/ICECOCS50124.2020.
9314391
A.Z. Faroukhi, I. El Alaoui, Y. Gahi, et A. Amine, A novel approach for big data monetization as a
service, in Advances on Smart and Soft Computing (Singapore, 2021), pp. 153–165. https://doi.
org/10.1007/978-981-15-6048-4_14
IRJET-V4I957.pdf. Accessed 05 Apr. 2021. https://www.irjet.net/archives/V4/i9/IRJET-V4I957.
pdf
A. Juneja, et N.N. Das, Big data quality framework: pre-processing data in weather monitoring
application, in 2019 International Conference on Machine Learning, Big Data, Cloud and Parallel
Computing (COMITCon), févr., pp. 559–563 (2019). https://doi.org/10.1109/COMITCon.2019.
8862267.
S. Juddoo, C. George, Discovering the Most Important Data Quality Dimensions in Health Big Data
Using Latent Semantic Analysis. (2018). https://doi.org/10.1109/ICABCD.2018.8465129
24 W. Elouataoui et al.
G. Kapil, A. Agrawal, R.A. Khan, A study of big data characteristics,” in 2016 International
Conference on Communication and Electronics Systems (ICCES), Oct. 2016, pp. 1–4. https://doi.
org/10.1109/CESYS.2016.7889917
N. Khan, M. Alsaqer, H. Shah, G. Badsha, A.A. Abbasi, S. Salehian, The 10 Vs, Issues and
challenges of big data, in Proceedings of the 2018 International Conference on Big Data and
Education, New York, NY, USA, Mar. 2018, pp. 52–56. https://doi.org/10.1145/3206157.320
6166
M. Knight, What Is Big Data? DATAVERSITY, 05 Feb. 2018. https://www.dataversity.net/what-is-
big-data/. Accessed 05 Apr. 2021
M.Y. Khaleel, et M.M. Hamad, Data quality management for big data applications, in 2019 12th
International Conference on Developments in eSystems Engineering (DeSE), Oct. 2019, pp. 357–
362. https://doi.org/10.1109/DeSE.2019.00072
S.S.B.T. Lincy, N.S. Kumar, An enhanced preprocessing model for big data processing: a quality
framework, in 2017 International Conference on Innovations in Green Energy and Healthcare
Technologies (IGEHT), Mar. 2017, pp. 1–7. https://doi.org/10.1109/IGEHT.2017.8094109
J. Merino, I. Caballero, B. Rivas, M. Serrano, M. Piattini, A Data Quality in Use model for Big
Data. Future Gener. Comput. Syst. 63, 123–130 (Oct. 2016). https://doi.org/10.1016/j.future.
2015.11.024
G. Mylavarapu, J. P. Thomas, et K. A. Viswanathan, An Automated Big Data Accuracy Assessment
Tool, in 2019 IEEE 4th International Conference on Big Data Analytics (ICBDA), mars 2019,
p. 193-197. doi: https://doi.org/10.1109/ICBDA.2019.8713218.
I. Noorwali, D. Arruda, N.H. Madhavji, Understanding quality requirements in the context of big
data systems, in 2016 IEEE/ACM 2nd International Workshop on Big Data Software Engineering
(BIGDSE), May 2016, pp. 76–79. https://doi.org/10.1109/BIGDSE.2016.021
N. B. PROS, “The Missing Vs in Big Data: Viability and Value,” Wired, May 06, 2013. Accessed:
Apr. 05, 2021. [Online]. Available: https://www.wired.com/insights/2013/05/the-missing-vs-in-
big-data-viability-and-value/
P. Pääkkönen et, J. Jokitulppo, Quality management architecture for social media data. J. Big Data
4 (2017). https://doi.org/10.1186/s40537-017-0066-7
A. Ramasamy, S. Chowdhury, Big Data Quality Dimensions: A Systematic Literature Review, May
2020. https://doi.org/10.4301/S1807-1775202017003
D. Reinsel, J. Gantz, J. Rydning, The Digitization of the World from Edge to Core, p. 28 (2018)
F. Sidi, P. Hassany Shariat Panahy, L. Affendey, M.A. Jabar, H. Ibrahim, A. Mustapha, Data quality:
a survey of data quality dimensions, Aug. 2013. https://doi.org/10.1109/InfRKM.2012.6204995
R. Schmidt, M. Möhring, Strategic alignment of cloud-based architectures for big data, in 2013 17th
IEEE International Enterprise Distributed Object Computing Conference Workshops, Sep. 2013,
pp. 136–143. https://doi.org/10.1109/EDOCW.2013.22
M. Serhani, H. El Kassabi, I. Taleb, R. Nujum, An Hybrid Approach to Quality Evaluation across
Big Data Value Chain (2016). https://doi.org/10.1109/BigDataCongress.2016.65
S. Soni, A. Singh, Improving Data Quality using Big Data Framework: A Proposed Approach
(2021)
The Challenges of Data Quality and Data Quality Assessment in the Big Data Era. https://datasc
ience.codata.org/articles/https://doi.org/10.5334/dsj-2015-002/. Accessed 05 Apr. 2021
The UNECE Big Data Quality Task Team, A Suggested Framework for the Quality of Big Data,
Dec. 2014
I. Taleb, M. Serhani, R. Dssouli, Big Data Quality Assessment Model for Unstructured Data (2018).
https://doi.org/10.1109/INNOVATIONS.2018.8605945
TDWI Best Practices Report | Big Data Analytics, Transforming Data with Intelligence (2021).
https://tdwi.org/research/2011/09/best-practices-report-q4-big-data-analytics.aspx Accessed 05
Apr. 2021
D. Tranfield, D. Denyer, P. Smart, Towards a methodology for developing evidence-informed
management knowledge by means of systematic review. Br. J. Manag. 14(3), 207–222 (Sep. 2003).
https://doi.org/10.1111/1467-8551.00375
Data Quality in the Era of Big Data: A Global Review 25
I. Taleb, M. Serhani et, R. Dssouli, Big Data Quality: A Data Quality Profiling Model, pp. 61–77
(2019). https://doi.org/10.1007/978-3-030-23381-5_5.
I. Taleb, M.A. Serhani, Big data pre-processing: closing the data quality enforcement loop, in 2017
IEEE International Congress on Big Data (BigData Congress) (2017).https://doi.org/10.1109/
bigdatacongress.2017.73
I. Taleb, H.T.E. Kassabi, M.A. Serhani, R. Dssouli, et C. Bouhaddioui, Big data quality:
a quality dimensions evaluation, in 2016 Intl IEEE Conferences on Ubiquitous Intelli-
gence & Computing, Advanced and Trusted Computing, Scalable Computing and Commu-
nications, Cloud and Big Data Computing, Internet of People, and Smart World Congress
(UIC/ATC/ScalCom/CBDCom/IoP/SmartWorld), Toulouse, juill. 2016, pp. 759–765. https://doi.
org/10.1109/UIC-ATC-ScalCom-CBDCom-IoP-SmartWorld.2016.0122
I. Taleb, R. Dssouli, et M.A. Serhani, Big data pre-processing: a quality framework, in 2015 IEEE
International Congress on Big Data, New York City, NY, USA, Juin 2015, pp. 191–198. https://
doi.org/10.1109/BigDataCongress.2015.35
M. Talha, A.A. El kalam et, N. Elmarzouqi, Big data: tradeoff between data quality and data security.
Procedia Comput. Sci. 151, 916–922, Janv 2019. https://doi.org/10.1016/j.procs.2019.04.127
Y. Wand, R.Y. Wang, Anchoring data quality dimensions in ontological foundations. Commun.
ACM 39(11), 86–95 (Nov. 1996). https://doi.org/10.1145/240455.240479
A. Wahyudi, G. Kuk, et M. Janssen, A Process Pattern Model for Tackling and Improving Big Data
Quality. Inf. Syst. Front 20(3), 457–469, juin 2018. https://doi.org/10.1007/s10796-017-9822-7
X. Xu, Y. Lei, et Z. Li, An Incorrect Data Detection Method for Big Data Cleaning of Machinery
Condition Monitoring , IEEE Transactions on Industrial Electronics, vol. 67, no. 3, pp. 2326–
2336, Mar. 2020, https://doi.org/10.1109/TIE.2019.2903774.
S. Zan, X. Zhang, Medical data quality assessment model based on credibility analysis, in 2018
IEEE 4th Information Technology and Mechatronics Engineering Conference (ITOEC) (2018).
https://doi.org/10.1109/itoec.2018.8740576
G. Zhang, A data traceability method to improve data quality in a big data environment, in 2020 IEEE
Fifth International Conference on Data Science in Cyberspace (DSC), July 2020, pp. 290–294.
https://doi.org/10.1109/DSC50466.2020.00051
P. Zhang, F. Xiong, J. Gao, J. Wang, Data Quality in Big Data Processing: Issues, Solutions and
Open Problems, p. 7 (2017). https://doi.org/10.1109/UIC-ATC.2017.8397554.
Adversarial Machine Learning, Research
Trends and Applications
Izzat Alsmadi
1 Introduction
There is strong evidence that machine learning models are susceptible to attacks
that involve modification of input data intending to result in mis-classification. One
more vulnerable model than others is neural networks, but there are many other
categories of machine learning models that could fall prey to this kind of attack. To
demonstrate the attack, we can think of a machine learning system that takes some
input, which results in an accurate classification of that data input. On the other hand,
it is possible for some malicious actor to artificially construct an input that will result
in incorrectly classified input data when ingested by the machine learning example.
Generating these artificial input data or adversarial examples can be done in several
ways such as:
I. Alsmadi (B)
Texas A&M, San Antonio, TX 78224, USA
e-mail: [email protected]
Each method has its advantages and disadvantages when used in an actual attack.
For example, a fast gradient sign method is a simple and easy way to create adversarial
examples, but this method’s success rate is much lower than the others. After gen-
erating the adversarial examples, they can be used in two main forms of attacks that
target these machine learning systems: Black-box versus white-box attacks. Black
box attacks happen in a scenario where the malicious actor only has a small amount of
information about the specifics of the model and how it works. However, in white box
attacks, the attacker has full knowledge of the model and its parameters. There is one
particularly tricky property of an adversarial example, and that is its transferability.
Most adversarial examples can be transferred from one model to another, meaning
that a malicious actor can develop one adversarial example designed to target their
own test model. When it has been tested, use that example to target another model
in a black box scenario.
Adversarial examples are inputs for machine learning models that allow an
attacker to cause the target model to make a mistake. These examples are inten-
tionally designed to make the model output a wrong reading, and in the world of
machine learning at a huge scale, one exploitative example or sample can have a
cascading effect on the model and whatever entity uses them.
To guard against adversarial input attacks, what is known as adversarial training
is performed. Adversarial training is adding adversarial examples into the data set at
the time of training the model. By injecting this adversarial example into the training
set, we train the model to be resilient against this type of attack since it will know
how to handle this kind of input when it is introduced. There have been experiments
to show the effectiveness of ensemble training in improving the model’s resilience
against both black-box attacks and white-box attacks. Ensemble adversarial train-
ing has been shown to improve the resilience of the model from black-box attacks
drastically. While, in the case of white-box attacks, adversarial ensemble training
did not do much to improve the resilience of the model. The idea of transferred
attacks seems particularly difficult to handle. Still, with the experiments done with
ensemble training it was proven that with black box attacks it is possible to increase
the models resilience to transferred attacks. However, new attack types have been
suggested to increase the success rate of transferred attacks in a black box scenario.
On the other hand, with white box attack, ensemble training, although it slightly
increases resilience against transferred attacks, negatively affects overall resilience
to white-box attacks. In conclusion, although this type of attack is serious against
machine learning systems, with adequate training and especially ensemble training,
we can dramatically reduce the effectiveness of such attacks. Sadly, its a continuous
game between defenders that improve their algorithms, but attackers who will do
the same. For example, there have been some suggested attacks that are theoretically
robust against such training methods.
Adversarial Machine Learning, Research Trends and Applications 29
2 Literature Review
Adversarial examples are evolving in machine learning security because the problem
is ever-evolving and while can be addressed in the short term but cannot be solved in
a simple manner. It requires the defender to continue winning the ever-progressing
game with the attacker in order to get ahead of current adversarial examples. One of
the more obvious places where adversarial machine learning could be used with dis-
astrous results would be in applying self-driving cars (Kurakin et al. 2016). Having a
two-ton missile controlled by an algorithm under the influence of what is basically an
optical illusion isn’t good for the future of self-driving cars. A stop sign is interpreted
as a yield sign is a great adversarial model, and the self-driving software should abso-
lutely be secured against this. The one-step fast gradient sign method (FSGM) is a
30 I. Alsmadi
way of simply creating an input data ‘offset’ closely related to the first gradient of a
data point. This is a very simplistic methodology of creating adversarial input, which
typically results in less success overall in a machine learning attack vector, but the
nature of one-step adversarial generation permits the attack vector to be transitive to
almost any model with equally successful attack rates via black-box methods. This
approach increases the loss function of each step in the model. Alternative one-step
methods are available, which increases the probability of a false classification versus
the loss of information. Iterative methods of perturbation where data point gradients
are adjusted multiple times prior to being presented to the model for classification
are consistently more successful as an attack vector on the model being targeted as
they are tuned to the intricacies of that model. This effect subsequently makes their
capability to be transitive much less successful to other models due to their highly
tuned nature. Obviously, there are other iterative perturbation methods available.
The authors claimed that including one-step adversarial examples batched with the
clean training data improved the robustness of the final model being able to identify
one-step perturbations when presented in real life. The sacrifice was less than 1%
loss of accuracy for clean data classification but did not improve robustness against
iteratively created adversarial data. Along these same lines, increasing the model’s
depth further improved the robustness while reducing the accuracy discrepancy for
clean data. As this ratio approaches a value of 1, all adversarial data is being identified
and classified correctly.
Although it is easy to assume that adversarial examples would come from mali-
cious actors, there is mounting evidence that adversarial examples may occur nat-
urally, especially when inputs are from cameras or other sensors. Kurakin et al.
(2016) demonstrate this in their study by printing out both clean and slightly adver-
sarial images, taking their photo, cropping, and feeding them into a trained ImageNet
Inception classifier. This method of manually taking photos of printed images, with-
out control of lighting, camera angle, distance-to-page, is intentional and meant to
introduce the “nuisance” variability that blurs the difference between the base and
adversarial images. Overall, they found that even with trivial or slight perturbations,
about 2/3 of the images would be top-1 mis-classified and 1/3 would be top-5 mis-
classified.
Adversarial training is the concept of training a machine learning system using
examples of malicious data in order to make the system more resistant to such attacks
and more capable of functioning in a hostile environment. In Adversarial Machine
Learning at Scale, (Kurakin et al. 2016) describe a methodology to take adversarial
training beyond small data sets and scale to large models. They also observe that
“single-step” attacks are better for black-box attacks on systems and that adversarial
training improves a system’s performance against such attacks. In the base case,
adversarial training is the inclusion of attack data in a model’s training set. The
paper states that it has been shown that machine learning models are vulnerable to
attack during testing through the use of inputs that have been subtly manipulated to
influence the model. However, it has also been shown that if such adversarial inputs
are included in the training set, the model becomes more resistant to this type of
influence and performs more reliably at test time. Prior to this paper, research in this
Adversarial Machine Learning, Research Trends and Applications 31
space was restricted to relatively small data sets. This paper’s primary contribution is
to suggest methods for using adversarial training on much larger data sets. They use
a large data set from ImageNet as an example for this study. The paper explains in
detail the types of attack methods used in the experiment to generate adversarial data.
The methods included various ways to manipulate the image data to it. They include
adversarial examples in each training set they run, generating new examples to include
with each iteration. Ultimately, the researchers found that including adversarial data
in the training set brought the model’s accuracy in identifying attacks in tests up to
the level of tests on clean data. However, they found that the accuracy of the model
on clean data decreased slightly when adversarial data was included in the training.
Kurakin et al. (2016) observed how adversarial examples might be used to train
more robust models. Their idea of adversarial training, for the latter, is to inject
adversarial examples in the same batches as their normal counterparts at every step
of training, to allow models to train on the differences between the two simulta-
neously. They found that such a method significantly increased top-1 accuracy to
74% and top-5 accuracy to 92%, making said accuracy on par with accuracy on
clean images. They found this method to be a more reliable safeguard for dealing
with adversarial examples than structural considerations like choosing a more robust
model architecture. The most influential factor for having a robust enough model
is how many trainable parameters the architecture has. However, this consideration
falls apart if a model is too small or too large, and it may not be worth the effort to
look for a model within the “sweet” spot. Kurakin et al. (2016) focused on adver-
sarial inputs to a model and how to train a model to protect itself from adversarial
inputs. Specifically, the examples in this paper use deep learning neural nets for
image recognition. The adversarial inputs can be created through slight alterations
in the pixels in the image. Four different methods for generating these adversarial
inputs are presented: fast gradient sign method (FSGM), one-step target class meth-
ods, basic iterative method, and iterative least-likely class method. FGSM is the most
computationally efficient of the four but is least likely to be successful. One-step tar-
get class relies on the probability of a specific target class. The “step 1.1” variation
is based on the target class that is least likely to be predicted by the neural net, and
“step rnd” chooses a randomly selected target class. The remaining two methods
apply small changes using FSGM or step 1.1 multiple times to create a successively
better adversarial image on each iteration. To protect the model against adversarial
inputs, the supposition is that we can introduce these inputs into the training data to
teach the system to reject these inputs. The author chooses mini-batches of length
32, where 16 inputs are replaced with adversarial versions of the replaced inputs.
It was not expected, but when the authors ran their experiments, they saw a small
degradation in the accuracy of the model against the original dataset. The degrada-
tion was less than 1%. The biggest difference was in accuracy in the presence of
adversarial inputs in the test data. In that case, the model trained with adversarial
inputs performed much better than the model trained with clean input only. There
is also a discussion of “label leaking”. They had some ods of generating adversarial
inputs that would result in accuracy improving as more adversarial inputs were added
to the test data. This can be countered by choosing the one-step methods instead of
32 I. Alsmadi
the FSGM method. Another discovery reported is that deeper models performed bet-
ter with the adversarial training data. Finally, they examined how well the training
transfers when confronted with adversarial input generated with a different method.
The outcome was that FSGM trained models did best at rejecting adversarial input.
Huang et al. (2015) paper also examines training deep neural networks with adver-
sarial inputs to improve the accuracy of the network. This paper describes finding
adversarial inputs using a linear approximation of a softmax layer g(x) when working
with the formula that Szegedy et al. (2013) proposed. The paper describes in detail
the algorithms needed to train the network using optimized adversarial inputs. It also
discusses several reductions of formulas required to make them solvable. The paper
goes on to demonstrate how the methods presented will perturb an image. Examples
are shown where the original image is subtracted out of the perturbed image to show
the noise image that was generated by three different methods. One interesting find-
ing of the research is that more parameters in a model mean that the model is more
resistant to adversarial examples. It is not quite known why this is, given the black
box-like nature of models. However, a significant fraction of adversarial examples
that fool one model is able to fool a different model. This property is called “transfer-
ability” and is used to craft adversarial examples in the black box scenario (Yan et al.
2018). This paper’s results show that some of the transferability of the adversarial
examples allow for the same exploit to be used in multiple models and that changing
hyperparameters does not make much of a difference on the defensive side of the
model. In order to get around some of the adversarial examples, some results were
found that were quite interesting. One such finding was that deeper models are more
benefited by adversarial training and that models with many layers actually change
many of the aspects of the model’s architecture. The number of filters and layers in
the algorithms, like the number of iterations, had a diminishing marginal effect the
higher it went, but an optimal level of what the factor Rho, (multiplier for the num-
ber of filters from standard) is showed that robustness of the model proportionally
increases with the larger model size. This property of size increases theoretically
until the model reaches an accuracy ratio of 1, but no models in this experiment
were large enough to reach that precision. The main takeaways from the adversarial
research are that model capacity used in conjunction with adversarial training is able
to increase resilience to adversarial examples in machine learning.
Tramèr et al. (2017), paper indicated that models trained with adversarial data
generated from the clean training data teach the automated one-step generator to
be weak versus strong. This ultimately means the model behaves well when using
‘white-box’ attacks (those that know the features and model) but misbehave when
encountering black-box attacks. Subsequently, black-box one-step adversarial data
generation then showed an increased ability to be transitive between models directly
countering Kurakin et al. (2016) findings. They subsequently suggest that pertur-
bation creation be decoupled from the clean training data and leverage Ensemble
Adversarial Training methods where adversarial data is passed through multiple per-
turbation stages to minimize correlation to the clean input training data effectively.
This paper (Tramèr et al. 2017), presents an ensemble technique used for adversarial
training that is more robust against black-box attacks on machine learning models.
Adversarial Machine Learning, Research Trends and Applications 33
clean training data to ensure the model doesn’t identify with local minimums when
stabilizing weight assignments.
There is the constant threat of malicious actors identifying and acting on weak
points that may exploit a system or to render inoperable for all computer systems. This
is becoming especially concerning for machine learning models and systems, which
are often touted as achieving better, smarter results than existing “manual” heuristics
and able to handle previously unseen information. Recent studies, however, indicate
that the advantage of learning from unseen information may be used as a weak point
to exploit and manipulate machine learning systems. One such attack approach is
known as adversarial example training, where legitimate inputs are altered by adding
small, imperceptible, perturbations that lead a trained classifier to mis-classify the
altered input, while the input is still correctly classified by a human observer.
Papernot et al. (2016a) provide an example, where both would be classified as
stop signs by any individual. However, they forced their neural network to clas-
sify the image on the right as a yield sign by applying a precise alteration on it.
In their paper, they go on to propose a generalized “black-box” adversary attack
that departs from contemporary methodologies. Instead of constructing adversarial
examples using prior knowledge of a neural network and parameters or using some
collected training set to train another attack model, they propose a model that can
learn to perform a generalized attack from just using the original model’s output.
The adversary model collects a very small set of inputs representative of the entire
domain, then selects an architecture to be trained that is applicable to the domain,
and then trains said architecture by applying perturbations at certain components
of the images to determine which perturbations affect saliency of the most, while
being observably undistinguished. Using this methodology, they were able to gener-
ate adversarial examples that were misclassified by black-box neural networks and
caused MetaMind, Amazon and Google classifiers to achieve misclassification rates
of 84.24%, 96.19% and 84.24%, respectively.
Authors in Papernot et al. (2016a) went on to find a method to defensively reduce
the effectiveness of adversarial samples from 95 to 0.5% through distillation. Distil-
lation is a training method where knowledge from one neural network is transferred
to another neural network. In this paper, however, they propose to use knowledge
extracted from a neural network to improve its own resilience to adversarial samples.
Essentially, by training a model on natural images and then transfer learning the same
model on a dataset with adversarial examples, the model can be trained to be more
resilient and robust against future adversarial examples. When applied on DNN for
the CIFAR10 dataset, the model reduced the success of adversarial samples to 0.5%
while maintaining an accuracy comparable to previously applied models.
In addition to CNNs and DNNs, reinforcement learning models such as Deep
Q Networks (DQNs) can also be manipulated by adversarial examples. Common
algorithms like DQN, TRPO, and A3C are open to being exploited by adversarial
examples (Behzadan and Munir 2017). Almost all adversarial examples are indis-
tinguishable to a human but have a huge effect on a ML model. These can lead to
degraded performance even in the presence of perturbations too subtle to be per-
ceived by a human, causing an agent to perform actions it would otherwise never do
Adversarial Machine Learning, Research Trends and Applications 35
and cause the model to underperform. The two takeaways here are that it is possible
to train a model on adversarial data to prepare it for malicious attacks, and the more
sophisticated/deeper a model is, the more it will benefit from this approach. Given
the task’s nuanced nature, this makes intuitive sense, but seeing the successful results
is very encouraging. A second paper, Towards Deep Learning Models Resistant to
Adversarial Attacks, (Madry et al. 2017) expands upon this topic by conducting a
similar study. Here the researchers focus more on deep neural networks, but also
through the lens of image recognition. In this paper, the authors attempt, similarly
to the original, to create models that are resistant to adversarial input attacks. The
paper (Madry et al. 2017) investigates the apparent lack of robustness deep learning
models have towards attacks. This is apparently a hot topic of research these days
and is certainly a relevant one.
Boosting Adversarial Attacks with Momentum levy similar criticisms to the train-
ing methods of previous approaches (Dong et al. 2018). Namely that of poor “trans-
ferability” (i.e., poor generalization). The solution here is similar to the third paper
with a focus on an ensemble approach. Their group’s models are momentum-based
iterative approaches, which they claim can fool white-box attacks and black-box ones.
Dong et al. (2018) used similar attack vectors in attacking models they trained in their
research. This research aims to analyze the defensive hardening created by adver-
sarial training and develop strengthened black-box attack methods that can defeat
these models. Again focused on the image recognition space, the authors explain
the vulnerability neural networks and other models have to adversarial images, as
well as the transferability of attacks across multiple models. They state that different
machine learning algorithms tend to come up with the same decisions over time,
which means that the same adversarial image that fools one model is likely to fool
others. However, they find that adversarially trained models perform much better
against these attacks, as found in Kurakin et al. (2016).
In Akhtar and Mian (2018), the authors chose to focus more on adversarial attacks
against images. Basically, image NN models can be attacked by introducing a small
amount of adversarial input into an image. The change in the image is nearly imper-
ceptible to the human eye, but it significantly reduces the accuracy of the models.
Several different types of adversarial attacks are discussed. Examples include one
pixel attack, where one pixel is changed in the image to trick a classifier. Addi-
tionally, he discusses different methods for defense, such as the adversarial brute
force adversarial training that is explored in the main paper discussed in this exam
question.
It seems that a common technique for defending against adversarial example
attempting to use iterative, or gradient descent, based attacks is to “mask” the gradient
(Guo et al. 2018). This “masking” is sometimes intentional as part of the defense or
a byproduct of the adversarial training method. Tsipras et al. (2018) acknowledge all
of the research and efforts to attack and defend models using adversarial examples in
training. They discuss the difficulties in crafting truly robust models and the fact that
standard models are vulnerable to attack without adversarial training. The authors go
on to discuss the cost of defending models from adversarial attacks. The most obvious
cost they point out is the dramatic increase in training time required. However, beyond
36 I. Alsmadi
that, they also find that as more adversarial images are introduced into the training
data set, the model’s overall accuracy declines. They go on to discuss the opposing
goals of high general accuracy and adversarial robustness. They describe a trade-off
between security and correctness without making any definitive conclusions.
Cybersecurity is relying more and more on Machine Learning to identify and com-
bat attacks on computer systems. Machine Learning systems enable defenders to
respond quickly to newly discovered vulnerabilities and react to new attack vectors.
However, attackers are also learning how to exploit Machine Learning systems. Bar-
reno et al. (2010) presents a model for the evaluation of secure learning systems,
algorithms that perform well in securing computer applications under adversarial
conditions. Their goal is to define a taxonomy—a language that can be used to
describe learning systems that can be used for cybersecurity purposes in adversarial
conditions. In their research, (Biggio et al. 2012) leverage the paper (Barreno et al.
2010) taxonomy to describe an attack model in terms of goal, knowledge, and capa-
bility. In this research, the authors focus on the use of pattern recognition in machine
learning and how pattern recognition can be manipulated in an adversarial environ-
ment. Huang et al. (2015) paper discussed how to increase dataset robustness and
to prevent unwanted perturbations. The technique that they used was to introduce
adversarial examples during an intermediate step intentionally. The study is trying
to find a technique to handle strong adversarial examples from causing additional
model mis-classifications, especially associated with deep-learning image classifiers.
In their approach, they produced a min-max training procedure to increase robust-
ness. A common problem with previous models that used logistic regression without
regularization could not solve for linearly separable data, but this can be overcome
by using an adversary during training. They designed a new method for finding
adversarial examples based on linear approximation using the Jacobian matrix. Ulti-
mately, the authors claim that their algorithm for min-max formulation of adversarial
perturbation is better than previous works. They can produce better robustness and
maintain solid accuracy.
We investigated AML literature based on specific subjects. Our goal is to describe
current and future research trends in this area. We can see the following taxonomy
of AML attacks.
From the perspective of attacker knowledge of target machine learning models, there
are three categories of attacks on Machine Learning models, White Box, Black Box
and Grey Box attacks. White Box attacks are adversarial attacks in which the attacker
Adversarial Machine Learning, Research Trends and Applications 37
has intimate knowledge as to the inner workings of the model. This could range from
knowing the features and target values to understanding something more complex
such as the gradients weights in a deep learning model and understanding how to
manipulate those values through data injection just enough to change a prediction.
Black Box attackers are such that the adversary does not have knowledge of the
inner workings of the model. These are much more difficult to conduct but there are
theories, methodologies and some examples of attacks being orchestrated against a
white box of a different model and then those attacks being carried out against a
black box of a presumably different model. Overall, most adversarial attacks against
Natural Language Processing (NLP) models are being used for message or news
postings in order to spread misleading or hateful information as well as scenario’s
where spam defeating is desired or click baiting detection is needed.
There are several types of white box attacks that can be highly effective at manip-
ulating the output of a machine learning model. Data poisoning attacks, particularly
on sentiment analysis, are easy to accomplish and yield reliable results. This Adver-
sarial attack is done by accessing the training data within a model and changing the
target label from positive to negative or vice versa. The changes are very subtle only
being done on a limited number of data rows for a word or sentence that is very
similar, but not the same, as the values in which the attacker wants the model to be
manipulated (Wallace et al. 2021). For example, if a product review about a new
computer is negative containing words that would normally be classified as negative
in a sentiment analysis model are labeled as positive and review for a TV that has
similar wording and is truly negative will be labeled as positive. There are a few
ways to defend against this type of attack. The first and simplest method is to stop
training early before the weighting of the malicious changes to training data is sig-
nificant. This technique works but cost in the form of accuracy. Perplexity analysis, a
technique that analyzes and ranks sentence’s structure can also be used to determine
the validity of the sentences that are being used in the data and can help identify
poising data injected into the training set. Another example of a white-box attack is
the Hot Flip attack (Ebrahimi et al. 2017). This attack uses the forward and backward
propagation of a neural network architecture, over several iterations, to determine
the best possible character to flip to achieve the greatest change in gradient weight
towards the desired result. In most cases this is used to defeat ML NLP that specif-
ically looks for key words that would be flagged as vulgar, inappropriate, known to
be intentionally misleading or spam like in nature.
In this specific area, there are several challenges being addressed. There is the
challenge to the task of hand, of determining how to manipulate the models in an
effective manner that goes undetected. This challenge is magnified when it comes to
understanding how to manipulate a black-box model that does not have documented
interworking’s or provide access to directly manipulate the training. The greater
challenge is then devising strategies for defending against successful attacks. Often
defending is much more complex than the attack itself. Take for instance the cases
of transposition and substitution at the character level on key words. It would be an
overwhelming task to provide dictionaries that account for every possible permuta-
tion and if possible, would require significant compute resources. Compute resources
38 I. Alsmadi
are another challenge; it takes libraries that support GPU to efficiently train models
and that compute resource that is needed increases exponentially when considering
modifying algorithms beyond just training but to detect intentionally data poisoning.
These high-performance resources are also needed for adversarial attacks not only
for those trying to manipulate the ML systems but for those building them to test the
systems vulnerability to attacks so that hardening measures can be taken.
Machine learning (ML) models are vulnerable and sensitive to small modifications.
An attacker can tamper a model by adding just a few malicious perturbations in the
input data. Therefore, it’s been always an important defense goal, to develop such
techniques which make these models more robust and versatile. There have been
several studies so far in this area that focus on one of the two main aspects:
• Examples of what and how to change ML models with adversarial examples.
Example of papers include: (Ebrahimi et al. 2017; Alzantot et al. 2018; Gao et al.
2018; Liang et al. 2017).
• Defense against the adversarial attacks: The naïve method of defensing against the
adversarial attacks is to recognize and remove the unnoticeable perturbations e.g.
in computer vision, the denoising auto-encoders are used to eliminate the noise
which is added by the perturbations, however, this approach cannot directly be
used to the NLP tasks primarily because of below reasons:
– There are continuous pixels in the images whereas the input texts contain the
discrete tokens. Because of this nature of the text, another similar or semantically
matching word/token can replace a word/token which lowers the performance.
Therefore, the perturbations which have a similar look with other words or get
mixed up with the input text, cannot be easily identified. The other approaches
capture the non-natural differences between the nearby pixels.
– As sentences may contain words with a large vocabulary, it can be hard to
iterate through all the possible perturbed sentences. As a result, the already
present defense methods that heavily rely on the pixels, may not be applicable
to NLP tasks.
Wang et al. proposed a Synonym Encoding Machine (SEM) to defend against adver-
sarial perturbations (Wang et al. 2019). SEM defends against synonym substitution
adversarial attacks. Text-based adversarial attacks fall into two categories: the word-
level and character-level. First, word-level has defenses through the system’s training
data when compared against the synonym attack. Second, the character-level has a
proposed defense of using a downstream classifier that will act as the defense mecha-
nism for adversarial attacks and identify possible adversarial attack spelling mistakes.
The categorization used helped cage the defense efforts and made for statistically
significant results in defending against the synonym-styled attacks. As this is a small
Adversarial Machine Learning, Research Trends and Applications 39
subset of the natural language processing adversarial attacks, it shows the validity
in defense efforts being made to all caveats of this system. Zhou et al. proposed
a randomized smoothing method to improve the training model to better defend it
from adversarial substitution attacks (Zhou et al. 2019). The model would use data
interpretation as its systematic central defense and placed data as the centerpiece to a
holistic natural language processing defense mechanic. Rosenberg et al. evaluated 11
defense methods and found that certain ensembles and sequencing delivered different
amounts of accuracy (Rosenberg et al. 2019). The effectiveness was cross correlated
with the amount of effort needed for method implementation and as expected high
overhead lead to the highest rate, but a low effort method with sequence squeezing
was found to still produce results near 91% which shows room for growth and analy-
sis in these methods and defense strategies. Data manipulation is not a new technique,
and its employment in this research area shows machine learning modules need to be
aware of security concerns just like a traditional system. To be brief, all systems are
susceptible to attack, and manipulation and security techniques need to ensure these
systems’ confidentiality and integrity. They complete their intended functionality as
machine learning cannot guarantee its safety and data surety.
Many existing adversarial example generation algorithms require full access to target
model parameters (white-box attack), which makes them impractical for many real-
world attacks because in case of real-world attacks, there may not be full information
available about the model. Because of this, many models are vulnerable to adversarial
examples. In such cases, the inputs are crafted to make a targeted model fool into
outputting an incorrect result or class. We can create the adversarial examples using
white box technique, but the main problem is the discrete nature of the input text.
Research across multiple authors highlighted a trend of many subsystems being
addressed as vulnerable to adversarial attacks. Systems named were deep neural
networks, Dirichlet neighborhood ensemble, recurrent neural networks. Adversarial
40 I. Alsmadi
attacks were seen as credible attacks to all these methods under the natural language
processing lens, and independent efforts were made to defend the specific subsystem
when a combined effort to defend against adversarial attacks as a whole could prove
beneficial to a more significant percentage. The trend here to specialize in a sub-
topic and create a unique solution is a part of many researchers’ efforts. Still, in the
proposed solutions, it often was a stagnation of effects as many results proved to be
along the same line of thought. Combining efforts to fight adversarial attacks might
better native defense efforts with subcomponent analysis being done to prove the
efficacy of efforts made by the researchers in their perspective specialties. It may be
beneficial to work as a whole on problems than constantly race one another to the same
solution. Similar to the above reference point, private datasets and industry-standard
practices validate efforts and research consistency are difficult. The privatization of
efforts in the business environment means many datasets are set to be academic as
a whole. Any security defect discovered often is hidden from public purview not
to hurt the company’s reputation. The researchers used many academic datasets,
which often had relatively robust datasets. Still, it is easy to see the lost opportunity
when comparing these sets to actual data of deployed natural language processors.
Company integration with the academic environment has been a significant issue in
academia, and it persists as researchers are losing the ability to test and abuse real-
world datasets. This point of view is easy to see when the references repeatedly used
academic datasets and came to conclusions based on the data used. The significance
of a topic cannot be fully understood until a product is exposed to the greater public,
and in this area of review, we must stress test all defensive efforts.
A basic Generative Adversarial Network (GAN) model includes two main modules, a
generator, and a discriminator. The generator and discriminator are implicit function
expressions, usually implemented by deep neural networks (Creswell et al. 2018).
Applying GAN in Natural Language Processing (NLP) tasks such as text gen-
eration is challenging due to the discrete nature of the text. Consequently, it is not
straightforward to pass the gradients through the discrete output words of the gener-
ator (Haidar and Rezagholizadeh 2019).
As text input is discrete, text generators model the problem as a sequential decision
making process. In the model, the state is the previously generated characters, words,
or sentences. The action or prediction to make is the next character/word/sentence
to be generated. The generative net is a stochastic policy that maps current state to a
distribution over the action space. Natural Language Generation (NLG) techniques
allow the generation of natural language text based on a given context. NLG can
involve text generation based on predefined grammar such as the Dada Engine,
(Baki et al. 2017) or leverage deep learning neural networks such as RNN, (Yao
et al. 2017) for generating text. We will describe some of the popular approaches
that can be found in relevant literature in the scope of AML.
Adversarial Machine Learning, Research Trends and Applications 41
sal Perturbation, (Shafahi et al. 2020), and Stochastic Activation Pruning (Dhillon
et al. 2018).
As of today, training with a PGD adversary remains empirically robust (Wong
et al. 2020)
• Jacobian-based saliency map approach (JSMA)
JSMA is a gradient based white-box method that is proposed to use the gradient of
loss with each class labels with respect to every component of the input (Papernot
et al. 2016b). JSMA is useful for targeted miss-classification attacks (Chakraborty
et al. 2018).
• Accelerating Adversarial Training
The cost of adversarial training can be reduced by reusing adversarial examples
and merging the inner loop of a PGD and gradient updates of the model parameters
(Shafahi et al. 2019; Zhang et al. 2019)
• DAWNBench competition
As the field of adversarial examples continues to expand within NLP models better
systems will get put in place. However, until then the field faces many challenges,
chief among them being the closed source nature of research. As pointed out by
Morris et al. the need for researchers to publish there research openly is of great
importance, it will ensure less time is spent reproducing and reducing errors. Another
major challenge in AML text, unlike with image adversarial examples there is no one
standard to detect adversarial examples within text. This lack of a standardization
allows for many different ideas to exist and many different algorithms to continue to
pop-up. The issue with this is while many articles reviewed utilize the same datasets
and models the algorithms themselves can be flawed but that can’t be determined
due to lack of standardization. The third and final issue which I see is that of NLP
being a recently new area of adversarial attacks in regards to research. As more
research is conducted in the field their will be more advancements, this will be due to
new researchers comparing their work against these ground breaking articles. This
comparison will ensure algorithms which the greatest efficacy is selected and used
to train deep neural networks. This training will in the future ensure a higher degree
of defensive and identification of threats. This defense will allow for deep neural
networks and other machine learning networks to better serve applications while
protecting themselves.
7 Generation Models/Tasks/Applications
from generating text in particular are growing. We will discuss just a few in this
section.
For many applications that we use through our smart phones, or websites, next word
prediction (NWP, also called auto-completion) is a typical NLP application. From a
machine-learning perspective, NWP is a classical prediction problem where previous
and current text can be the pool to extract the prediction model features and other
parameters and the next word to predict is the target feature. Different algorithms are
proposed to approach NWP problem such as term frequencies, artificial intelligence,
n-grams, neural networks, etc.
One of the key issues in text generation is that there is no widely agreed-upon
automated metric for evaluating the text generated output. Text generation metrics
can be classified based on several categories. Here is a summary of categories and
metrics:
9 Memory-Based Models
Within these two algorithms there have been numerous contributions on improving
them. For example, there are models like Texar (Hu et al. 2018). Texar is a special
model for LSTM and GRU architectures that aims to transform any given input into
a natural language. There is also Grid LSTMs that’s goal is to allow hidden layers
at any step in the LSTM. With all, the results for each model vary based on the
implementation and dataset used to train each model. For example, in Chung et al.
(2014), both an LSTM and GRU model were trained on music and speech datasets
and it was found that the GRU was superior is terms of accuracy and less loss which
contradicts previous papers that state the LSTM was better in these areas. To really
sum up all the research surrounding this topic it comes down to having to run both
with the given datasets and see which one is better in terms of the trade-offs, are you
are willing to make it such that it has faster training times, more accuracy or less loss
at the cost of something else.
References
N. Akhtar, A. Mian, Threat of adversarial attacks on deep learning in computer vision: a survey.
IEEE Access 6, 14410–14430 (2018)
M. Alzantot, Y. Sharma, A. Elgohary, B.-J. Ho, M. Srivastava, K.-W. Chang, Generating natural
language adversarial examples (2018). arXiv:1804.07998
M. Arjovsky, S. Chintala, L. Bottou, Wasserstein gan (2017). arXiv:1701.07875
D. Bahdanau, P. Brakel, K. Xu, A. Goyal, R. Lowe, J. Pineau, A. Courville, Y. Bengio, An actor-critic
algorithm for sequence prediction (2016). arXiv:1607.07086
S. Baki, R. Verma, A. Mukherjee, O. Gnawali, Scaling and effectiveness of email masquerade
attacks: exploiting natural language generation, in Proceedings of the 2017 ACM on Asia Con-
ference on Computer and Communications Security (2017), pp. 469–482
S. Banerjee, A. Lavie, Meteor: an automatic metric for mt evaluation with improved correlation
with human judgments, in Proceedings of the acl Workshop on Intrinsic and Extrinsic Evaluation
Measures for Machine Translation and/or Summarization (2005), pp. 65–72
Adversarial Machine Learning, Research Trends and Applications 51
M. Barreno, B. Nelson, A.D. Joseph, J.D. Tygar, The security of machine learning. Mach. Learn.
81(2), 121–148 (2010)
V. Behzadan, A. Munir, Vulnerability of deep reinforcement learning to policy induction attacks,
in International Conference on Machine Learning and Data Mining in Pattern Recognition
(Springer, 2017), pp. 262–275
S. Bengio, O. Vinyals, N. Jaitly, N. Shazeer, Scheduled sampling for sequence prediction with
recurrent neural networks. Adv. Neural Inf. Proc. Syst. 28, 1171–1179 (2015)
A. Bennett, N. Kallus, T. Schnabel, Deep generalized method of moments for instrumental variable
analysis, in Advances in Neural Information Processing Systems (2019), pp. 3564–3574
B. Biggio, G. Fumera, G.L. Marcialis, F. Roli, Security of pattern recognition systems in adversarial
environments (2012)
J. Buckman, A. Roy, C. Raffel, I. Goodfellow, Thermometer encoding: one hot way to resist adver-
sarial examples, in International Conference on Learning Representations (2018)
Y. Burda, H. Edwards, A. Storkey, O. Klimov, Exploration by random network distillation (2018).
arXiv:1810.12894
P. Cai, X. Chen, P. Jin, H. Wang, T. Li, Distributional discrepancy: a metric for unconditional text
generation. Knowl.-Based Syst. 217, 106850 (2021)
A. Chakraborty, M. Alam, V. Dey, A. Chattopadhyay, D. Mukhopadhyay, Adversarial attacks and
defences: a survey (2018). arXiv:1810.00069
J. Chang, R. He, L. Wang, X. Zhao, T. Yang, R. Wang, A semi-supervised stable variational network
for promoting replier-consistency in dialogue generation, in Proceedings of the 2019 Conference
on Empirical Methods in Natural Language Processing and the 9th International Joint Conference
on Natural Language Processing (EMNLP-IJCNLP) (2019), pp. 1920–1930
T. Che, Y. Li, R. Zhang, R.D. Hjelm, W. Li, Y. Song, Y. Bengio, Maximum-likelihood augmented
discrete generative adversarial networks (2017). arXiv:1702.07983
L. Chen, S. Dai, C. Tao, H. Zhang, Z. Gan, D. Shen, Y. Zhang, G. Wang, R. Zhang, L. Carin,
Adversarial text generation via feature-mover’s distance, in Advances in Neural Information
Processing Systems (2018), pp. 4666–4677
J. Chung, C. Gulcehre, K.H. Cho, Y. Bengio, Empirical evaluation of gated recurrent neural networks
on sequence modeling (2014). arXiv:1412.3555
O. Cífka, A. Severyn, E. Alfonseca, K. Filippova, Eval all, trust a few, do wrong to none: comparing
sentence generation models (2018). arXiv:1804.07972
C. Coleman, D. Narayanan, D. Kang, T. Zhao, J. Zhang, L. Nardi, P. Bailis, K. Olukotun, C.
Ré, M. Zaharia, Dawnbench: an end-to-end deep learning benchmark and competition. Training
100(101), 102 (2017)
A. Creswell, T. White, V. Dumoulin, K. Arulkumaran, B. Sengupta, A.A. Bharath, Generative
adversarial networks: an overview. IEEE Signal Proc. Mag. 35, 53–65 (2018)
G.S. Dhillon, K. Azizzadenesheli, Z.C. Lipton, J. Bernstein, J. Kossaifi, A. Khanna, A. Anandkumar,
Stochastic activation pruning for robust adversarial defense (2018). arXiv:1803.01442
Y. Dong, F. Liao, T. Pang, H. Su, J. Zhu, X. Hu, J. Li, Boosting adversarial attacks with momentum,
in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018), pp.
9185–9193
W. Du, A.W. Black, Boosting dialog response generation, in Proceedings of the 57th Annual Meeting
of the Association for Computational Linguistics (2019), pp. 38–43
J. Ebrahimi, A. Rao, D. Lowd, D. Dou, Hotflip: white-box adversarial examples for text classification
(2017). arXiv:1712.06751
C. Esteban, S.L. Hyland, G. Rätsch, Real-valued (medical) time series generation with recurrent
conditional gans (2017). arXiv:1706.02633
J.A. Fodor, Z.W. Pylyshyn et al., Connectionism and cognitive architecture: a critical analysis.
Cognition 28(1–2), 3–71 (1988)
L. Fowl, M. Goldblum, A. Gupta, A. Sharaf, T. Goldstein, Random network distillation as a diversity
metric for both image and text generation (2020). arXiv:2010.06715
52 I. Alsmadi
J. Gao, J. Lanchantin, M.L. Soffa, Y. Qi, Black-box generation of adversarial text sequences to
evade deep learning classifiers, in 2018 IEEE Security and Privacy Workshops (SPW) (IEEE,
2018), pp. 50–56
N. Gao, H. Xue, W. Shao, S. Zhao, K.K. Qin, A. Prabowo, M.S. Rahaman, F.D. Salim, Generative
adversarial networks for spatio-temporal data: a survey (2020). arXiv:2008.08903
I.J. Goodfellow, J. Shlens, C. Szegedy, Explaining and harnessing adversarial examples (2014).
arXiv:1412.6572
I. Goodfellow, J. Pouget-Abadie, M. Mirza, X. Bing, D. Warde-Farley, S. Ozair, A. Courville, Y.
Bengio, Generative adversarial networks. Commun. ACM 63(11), 139–144 (2020)
I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, A.C. Courville, Improved training of wasserstein
gans, in Advances in Neural Information Processing Systems (2017), pp. 5767–5777
H. Guo, Generating text with deep reinforcement learning (2015). arXiv:1510.09202
J. Guo, S. Lu, H. Cai, W. Zhang, Y. Yu, J. Wang, Long text generation via adversarial training with
leaked information, in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32
(2018)
A. Haidar, M. Rezagholizadeh, Textkd-gan: text generation using knowledge distillation and gen-
erative adversarial networks, in Canadian Conference on Artificial Intelligence (Springer, 2019),
pp. 107–118
W. Hamilton, Z. Ying, J. Leskovec, Inductive representation learning on large graphs, in Advances
in neural information processing systems (2017), pp. 1024–1034
L.P. Hansen, Large sample properties of generalized method of moments estimators. Econ.: J. Econ.
Soc. 1029–1054 (1982)
M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, S. Hochreiter, Gans trained by a two time-scale
update rule converge to a local nash equilibrium, in Advances in Neural Information Processing
Systems (2017), pp. 6626–6637
G.E. Hinton, S. Osindero, Y.-W. Teh, A fast learning algorithm for deep belief nets. Neural Comput.
18(7), 1527–1554 (2006)
S. Hochreiter, J. Schmidhuber, Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Z. Hu, H. Shi, B. Tan, W. Wang, Z. Yang, T. Zhao, J. He, L. Qin, D. Wang, X. Ma, et al., Texar: a
modularized, versatile, and extensible toolkit for text generation (2018). arXiv:1809.00794
R. Huang, B. Xu, D. Schuurmans, C. Szepesvári, Learning with a strong adversary (2015).
arXiv:1511.03034
D. Jakubovitz, R. Giryes, Improving dnn robustness to adversarial attacks using jacobian regular-
ization, in Proceedings of the European Conference on Computer Vision (ECCV) (2018), pp.
514–529
A. Jalal, A. Ilyas, C. Daskalakis, A.G. Dimakis, The robust manifold defense: adversarial training
using generative models (2017). arXiv:1712.09196
E. Jang, S. Gu, B. Poole, Categorical reparameterization with gumbel-softmax (2016).
arXiv:1611.01144
H. Kannan, A. Kurakin, I. Goodfellow, Adversarial logit pairing (2018). arXiv:1803.06373
P. Keukeleire, Correspondence between perplexity scores and human evaluation of generated tv-
show scripts (2020)
D.P. Kingma, M. Welling, Auto-encoding variational bayes (2013). arXiv:1312.6114
A. Koochali, P. Schichtel, A. Dengel, S. Ahmed, Probabilistic forecasting of sensory data with
generative adversarial networks-forgan. IEEE Access 7, 63868–63880 (2019)
A. Kurakin, I. Goodfellow, S. Bengio, Adversarial machine learning at scale (2016).
arXiv:1611.01236
M.J. Kusner, J.M. Hernández-Lobato, Gans for sequences of discrete elements with the gumbel-
softmax distribution (2016). arXiv:1611.04051
M.J. Kusner, Y. Sun, N.I. Kolkin, K.Q. Weinberger, From word embeddings to document distances,
in Proceedings of the 32nd International Conference on International Conference on Machine
Learning, vol. 37, JMLR.org (2015), pp. 957–966
Adversarial Machine Learning, Research Trends and Applications 53
A.M. Lamb, A.G.A.P. Goyal, Y. Zhang, S. Zhang, A.C. Courville, Y. Bengio, Professor forcing:
a new algorithm for training recurrent networks, in Advances in Neural Information Processing
Systems (2016), pp. 4601–460
G. Lewis, V. Syrgkanis, Adversarial generalized method of moments (2018). arXiv:1803.07164
J. Li, M. Galley, C. Brockett, J. Gao, B. Dolan, A diversity-promoting objective function for neural
conversation models (2015). arXiv:1510.03055
J. Li, W. Monroe, A. Ritter, M. Galley, J. Gao, D. Jurafsky, Deep reinforcement learning for dialogue
generation (2016). arXiv:1606.01541
B. Liang, H. Li, M. Su, P. Bian, X. Li, W. Shi, Deep text classification can be fooled (2017).
arXiv:1704.08006
C.-Y. Lin, Rouge: a package for automatic evaluation of summaries, in Text Summarization Branches
Out (2004), pp. 74–81
J. Lin, C. Gan, S. Han, Defensive quantization: when efficiency meets robustness (2019).
arXiv:1904.08444
K. Lin, D. Li, X. He, Z. Zhang, M.-T. Sun, Adversarial ranking for language generation, in Advances
in Neural Information Processing Systems (2017), pp. 3155–3165
C.J. Maddison, A. Mnih, Y.W. Teh, The concrete distribution: a continuous relaxation of discrete
random variables (2016). arXiv:1611.00712
A. Madry, A. Makelov, L. Schmidt, D. Tsipras, A. Vladu, Towards deep learning models resistant
to adversarial attacks (2017). arXiv:1706.06083
R. Manzelli, V. Thakkar, A. Siahkamari, B. Kulis, An end to end model for automatic music genera-
tion: combining deep raw and symbolic audio networks, in Proceedings of the Musical Metacre-
ation Workshop at 9th International Conference on Computational Creativity, Salamanca, Spain
(2018)
X. Mao, Q. Li, H. Xie, R.Y.K. Lau, Z. Wang, S. Paul Smolley, Least squares generative adversarial
networks, in Proceedings of the IEEE International Conference on Computer Vision (2017), pp.
2794–2802
M. Mosbach, M. Andriushchenko, T. Trost, M. Hein, D. Klakow, Logit pairing methods can fool
gradient-based attacks (2018). arXiv:1810.12042
Y. Mroueh, T. Sercu, Fisher gan, in Advances in Neural Information Processing Systems (2017),
pp. 2513–2523
M. Nabeel, A. Riaz, W. Zhenyu, Cas-Gans: An Approach of Dialogue Policy Learning Based on
Gan and Rl Techniques (Int. J. Adv. Comput. Sci, Appl, 2019)
W. Nie, N. Narodytska, A. Patel, Relgan: relational generative adversarial networks for text gener-
ation, in International Conference on Learning Representations (2018)
S. Nowozin, B. Cseke, R. Tomioka, f-gan: training generative neural samplers using variational
divergence minimization, in Advances in Neural Information Processing Systems (2016), pp.
271–279
C. Olah, Understanding lstm networks (2015)
N. Papernot, P. McDaniel, X. Wu, S. Jha, A. Swami, Distillation as a defense to adversarial pertur-
bations against deep neural networks, in 2016 IEEE Symposium on Security and Privacy (IEEE,
2016a), pp. 582–597
N. Papernot, P. McDaniel, S. Jha, M. Fredrikson, Z.B. Celik, A. Swami, The limitations of deep
learning in adversarial settings, in 2016 IEEE European symposium on security and privacy
(EuroS&P) (IEEE, 2016b), pp. 372–387
K. Papineni, S. Roukos, T. Ward, W.-J. Zhu, Bleu: a method for automatic evaluation of machine
translation, in Proceedings of the 40th Annual Meeting of the Association for Computational
Linguistics (2002), pp. 311–318
K. Pearson, Asymmetrical frequency curves. Nature 48(1252), 615–616 (1893)
Q. Qian, M. Huang, H. Zhao, J. Xu, X. Zhu, Assigning personality/profile to a chatting machine
for coherent conversation generation, in IJCAI (2018), pp. 4279–4285
L.R. Rabiner, A tutorial on hidden markov models and selected applications in speech recognition.
Proc. IEEE 77(2), 257–286 (1989)
54 I. Alsmadi
Z. Yan, Y. Guo, C. Zhang, Deep defense: training dnns with improved adversarial robustness. Adv.
Neural Inf. Proc. Syst. 31, 419–428 (2018)
Z. Yang, W. Chen, F. Wang, B. Xu, Improving neural machine translation with conditional sequence
generative adversarial nets (2017). arXiv:1703.04887
S. Yang, X. Yu, Y. Zhou, Lstm and gru neural network performance comparison study: taking yelp
review dataset as an example, in 2020 International Workshop on Electronic Communication and
Artificial Intelligence (IWECAI) (IEEE, 2020), pp. 98–101
Y. Yang, G. Zhang, D. Katabi, Z. Xu, Me-net: towards effective adversarial robustness with matrix
estimation (2019). arXiv:1905.11971
Y. Yao, B. Viswanath, J. Cryan, H. Zheng, B.Y. Zhao, Automated crowdturfing attacks and defenses
in online review systems, in Proceedings of the 2017 ACM SIGSAC Conference on Computer
and Communications Security (2017), pp. 1143–1158
H. Yin, D. Li, X. Li, P. Li, Meta-cotgan: a meta cooperative training paradigm for improving
adversarial text generation, in AAAI (2020), pp. 9466–9473
F. Yu, Z. Xu, Y. Wang, C. Liu, X. Chen, Towards robust training of neural networks by regularizing
adversarial gradients (2018). arXiv:1805.09370
L. Yu, W. Zhang, J. Wang, Y. Yu, Seqgan: sequence generative adversarial nets with policy gradient,
in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31 (2017)
Y. Zhang, Z. Gan, K. Fan, Z. Chen, R. Henao, D. Shen, L. Carin, Adversarial feature matching for
text generation (2017). arXiv:1706.03850
Z. Zhang, S. Liu, M. Li, M. Zhou, E. Chen, Bidirectional generative adversarial networks for neural
machine translation, in Proceedings of the 22nd Conference on Computational Natural Language
Learning (2018), pp. 190–199
D. Zhang, T. Zhang, Y. Lu, Z. Zhu, B. Dong, You only propagate once: accelerating adversarial
training via maximal principle, in Advances in Neural Information Processing Systems (2019),
pp. 227–238
Z. Zhou, H. Guan, M.M. Bhat, J. Hsu, Fake news detection via nlp is vulnerable to adversarial
attacks (2019). arXiv:1901.09657
J. Zhu, R. Kaplan, J. Johnson, L. Fei-Fei, Hidden: hiding data with deep networks. In Proceedings
of the European conference on computer vision (ECCV) (2018a), pp. 657–672
Y. Zhu, S. Lu, L. Zheng, J. Guo, W. Zhang, J. Wang, Y. Yu, Texygen: a benchmarking platform
for text generation models, in The 41st International ACM SIGIR Conference on Research &
Development in Information Retrieval (2018b), pp. 1097–1100
Multi-agent Systems for Distributed Data
Mining Techniques: An Overview
Abstract The term “multi-agent systems” (MAS) refers to a mechanism that is used
to create goal-oriented autonomous agents in a shared environment and have commu-
nication and coordination capabilities. This goal-oriented mechanism supports
distributed data mining (DM) to implement various techniques for distributed
clustering, classification, and prediction. Different distributed DM (DDM) tech-
niques, MASs, the advantages of MAS-based DDM, and various MAS-based DDM
approaches proposed by researchers are reviewed in this study.
1 Introduction
Data mining (DM) is a process that is used to analyze data, determine data patterns,
and predict future trends from previously analyzed data. Training data for DM tech-
niques is useful only when data can be accessed by the running algorithm and there-
fore should be located physically or virtually in a single unit to increase computation
cost and maintain data privacy. Computation of data from different geographical
locations is a future perspective in DM and is referred to as distributed DM (DDM).
DM identifies hidden data patterns and categorizes them to obtain useful information.
The information is then collected in common areas, such as data warehouses, and
analyzed using DM techniques, including pattern matching, clustering, rule associ-
ation, regression, and classification (Chiang et al. 2011). DM algorithms are utilized
in business decision-making and other processes that require information to decrease
costs and increase revenues. Data are assumed to be centralized, memory-resident,
and static in traditional DM algorithms (Gan et al. 2017).
The collection of distributed data in a data warehouse is required in the application
of the classical knowledge discovery process in distributed environments for central
processing (Tsoumakas and Vlahavas 2009). This process, however, is usually either
ineffective or infeasible because of the following reasons:
60 M. H. Qasem et al.
1. High storage cost: A large data warehouse is required to store all data in a
central system.
2. High communication cost: The transfer of large data volumes over networks
requires considerable time and financial resources. Wireless network environ-
ments with limited bandwidth may encounter problems when transferring even
a small volume of data.
3. High computational cost: Mining a central data warehouse incurs a higher cost
than the sum of the cost of analyzing small parts of data.
4. Risk of exposure of private and sensitive data: DM applications handle sensi-
tive data, such as people’s medical and financial records. Central collection may
jeopardize the privacy of data.
The advancement of the Internet has led to the rapid generation of a large amount
of data, which are difficult to process even by supercomputers. Such data are stored
in multiple locations, requiring costly centralization. Data centralization is also
affected by bandwidth limitation and privacy concerns. DDM has gained considerable
attention to solve such problems (Liu et al. 2011).
DDM mines data sources regardless of their physical locations. This characteristic
is crucial, given that locally produced data may not be shared across the network
because of their large amount and privacy issues. As the Internet advances, the data
from businesses increase rapidly. Such data may not be shared or transferred because
of security and privacy issues, which involve legislation of particular locales and
system architecture (Moemeng et al. 2009). The decentralized architecture of DDM
can reach every networked business; hence, DDM has become a critical component
of knowledge-based systems. The business intelligence market is one of the fastest
growing and most profitable areas in the software industry, which consequently leads
to the popularization of DDM.
The need to extract information from decentralized data sources has given rise
to DDM. DM techniques must be dynamic because changes in the system can
influence its overall performance. DDM is a rapidly expanding area that aims to
find data patterns in an environment with distributed data and computation. Current
data analysis systems require centralized storage of data; nevertheless, the combina-
tion of computation with communication needs DDM environments that can utilize
distributed computation (Li et al. 2003).
DDM has evolved mainly due to factors such as privacy of sensitive data, trans-
mission cost, computation cost, and memory cost. DDM aims to extract useful infor-
mation from data located at heterogeneous sites. Distributed computing comprises
distributed sites that have computing units situated at individual heterogeneous
points. A decentralized mining strategy that makes the entire working system scalable
by distributing workload across heterogeneous sites is followed by DDM (Sawant
and Shah 2013). Figure 1 shows the typical DM architecture versus the DDM working
architecture used to construct a data warehouse.
The advances in computing and communication, including the Internet, intranets,
local area networks, and wireless networks, have resulted in many pervasive
distributed environments, which comprise various sources of large data and multiple
Multi-agent Systems for Distributed Data Mining Techniques … 61
computation nodes. Therefore, the analysis and management of these distributed data
sources require DDM technology that is designed for distributed applications (Ali
et al. 2018).
The DDM is applicable to the following conditions:
1. Systems have multiple independent data and computation sites and communi-
cate exclusively through messages.
2. Communication among sites is expensive.
3. Sites include resource constraints, such as battery power issues.
4. Sites involve privacy concerns.
Communication is assumed to be conducted solely by message passing. Several
DDM methods thus aim to minimize the number of messages sent. Load balance is
also proposed to prevent performance from being dominated by the time and space
usage of any individual site. Reference indicated that “building a monolithic database,
in order to perform non-distributed DM, may be infeasible or simply impossible” in
many applications. The transferring of considerable data may incur a high cost and
result in inefficient implementations.
The performance of DDM is affected by the following aspects:
1. Heterogeneous DM: Heterogeneous data lead to contradiction among the data
attributes. In case of heterogeneous data, a local data management model needs
to be integrated into a global model (Chen and Huang 2019)
2. Data consistency: Data from different sites are inconsistent. Modification of
the local data model affects the global data model and, consequently, the final
output produced after DM (Li et al. 2017).
3. Communication cost: Communication cost is dependent on network bandwidth
and the amount of transferred information, thereby requiring a cost model to be
built in DDM (Clifton et al. 2002).
62 M. H. Qasem et al.
and accumulated at the global level. Tsoumakas et al. (Adil et al. 2020) presented a
similar architecture.
One of the two assumptions in the DDM literature is about how data are distributed
across sites, namely, homogeneously (horizontally partitioned) and heterogeneously
(vertically partitioned). In both viewpoints, data tables at each site are partitioned into
a single global table. Classifiers for DDM include homogeneous and heterogeneous
approaches.
Vertical partitioning of data refers to the collection of different information on
the same set of entities or people by different sites or organizations; for example,
hospitals and insurance companies collect data on a set of people that can be jointly
linked. The data mined at such sites are similar (Ouda et al. 2012). In horizontal
partitioning, the same information on different entities or people is collected; for
example, supermarkets collect the transaction information of their clients. In this
case, the mined data are a combination of the data at the sites.
• Homogeneous Classifier
A global table is horizontally partitioned in this case. The tables at each site are
subsets of the global table and have similar attributes, as illustrated in Fig. 2. The
same set of attributes over all participating nodes is mined in homogeneous DDM.
This case occurs when databases are owned by the same organization (e.g., local
stores of a chain). Bandyopadhyay et al. (2006) proposed a technique based on the
principles of the k-means algorithm to cluster homogeneously distributed data in a
P2P environment, including sensor networks. The technique operates in a localized,
asynchronous manner by interacting with neighboring nodes. They theoretically
analyzed the algorithm and found that it binds errors in a distributed clustering
process, unlike the centralized approach, which requires all observed data to be
downloaded to a single site.
Li et al. (2017) identified dense regions as clusters and removed clutter at the
coarse level by using a proposed density-based clustering mechanism. The “can-
notlink” (CL) constraint was applied at the fine level, and overlapping clusters were
determined. A multisource homogeneous data-clustering model was also investi-
gated. The model poses CL constraints on the data from the same source. High
Fig. 2 Homogeneous
partitioned
64 M. H. Qasem et al.
level of clutter, misdetection, and unknown number of potential clusters affect the
dataset. Coarse-level clustering was based on density for fast computing, whereas
the fine-level one was based on CL-constrained distance to divide closely connected
clusters.
Montero-Manso et al. (2018) proposed a novel learning algorithm that combines
the outputs of classifiers trained at each node. Such outputs have different weights.
A distributed framework with training and test samples that belonged to the same
distribution was considered. The training instances spread across disjoint nodes. The
distributional distance between each node and the test set in the feature space was used
as a reference for the weights. Per-node weighting (pNW) and per-instance weighting
(pIW) were introduced. The same weight was assigned to all test instances at each
node in the former approach, whereas distinct weights were allowed for test instances
that are differently represented at the node in the latter approach. Experiments showed
that pIW was superior to pNW and standard unweighted approaches in classification
accuracy.
Amir et al. (2018) used the distributed associative memory tree (DASMET)
algorithm and presented a fully distributed pattern recognition system within P2P
networks to detect spam. Unlike server-based systems, the presented system is
cost-efficient and avoids a single point of failure. The algorithm can be scaled for
large and frequently updated datasets and is designed specifically for datasets with
similar existing patterns. DASMET used a relatively small amount of resources but
performed best in spam detection among all distribution methods.
• Heterogeneous Classifier
The table is vertically partitioned in this case. A collection of columns exist in each
site (sites do not have the same attributes), as illustrated in Fig. 3. However, a unique
identifier is present in each tuple at each site for matching. The global table viewpoint
is strictly conceptual. Physical realization and partitioning of such a table to form
tables at each site are unnecessarily assumed. Different sets of attributes in each
participating node are mined in heterogeneous DDM.
Stahl et al. (2012) proposed a pocket DM (PDM) that analyzes data streams
in mobile ad hoc distributed environments. Various applications can be run in such
environments given the advances in mobile devices, including smartphones and tablet
computers. Classification techniques for data streams for PDM were proposed for
adoption. An experiment verified that the application of classification techniques
for heterogeneous or homogeneous data streams to vertically partitioned data (data
Fig. 3 Heterogeneous
partitioned
Multi-agent Systems for Distributed Data Mining Techniques … 65
DDM is implemented by multiple methods with a unified goal. The roles of data
analysis at multiple distributed data sources are established. The contribution of
distributed methods is facilitated by result sharing. DDM can generally be classified
into low- and high-level DDM according to the level of information sharing. Low-
level DDM integrates the voting results of multiple independent decision-making
units, and high-level DDM combines learning results by meta-learning.
• Low-level DDM
Each method (i.e., decision-maker) is trained on the basis of its own data in low-level
DDM. The same task, which can be classifying, clustering, or analyzing an instance,
is given to all decision-makers. An output is acquired by combining all results. DBDS
(Januzaj et al. 2004) is a distributed clustering approach that is implemented on dense
clustering algorithms and operates locally. Cluster centers are produced locally at
each distributed source and transformed with a small number of data elements to
a decision-making center, which then recalculates the cluster centers in accordance
with the received centers and elements. Examples of such approach are bagging and
boosting, in which multiple classifiers operate on different datasets (Al Hwaitat et al.
2020). Bayesian classifiers are operated in a distributed environment; the local model
of distributed sources is averaged to obtain a global one (Adil et al. 2020).
Yan et al. (2019) discussed a distributed ensemble technique for mining health care
data under privacy constraints. A novel privacy-based distributed ensemble classifier
66 M. H. Qasem et al.
technique called adaptive privacy-based boosting was proposed to predict the model
for EHR data. Each distributed site was allowed to learn data distribution effectively
and share medical data without the need to reveal patients’ sensitive data. Minimal
computational complexity and communication cost were achieved accordingly.
• High-level DDM
The learned model of each source is shared with the global model to yield a
single learning model for mining input data in high-level DDM. This DDM type
is called metalearning or meta-DM. Various tools, such as JAM (Almaiah et al.
2020) and BODHI (Qasem et al. 2021), were proposed for classification. Da Silva
et al. (Almaiah) discussed distributed data clustering inferences and proposed the
kernel-based distributed clustering scheme algorithm. A local-level estimate was
established by a helper site for each local-level distributed site and then forwarded
to peer distributed sites. A global-level estimate was then determined by a local-
density-based clustering algorithm, and data confidentiality was achieved. Aouad
et al. (Almaiah et al.) described a lightweight clustering technique that minimizes
communication overhead by minimum variance formation of clusters.
Dong et al. (Yuan et al. 2019) introduced a clustering algorithm based on data
partitioning for unevenly distributed datasets. The proposed algorithm was used on
uneven datasets. They developed a fuzzy connectedness graph algorithm (PFHC)
on the basis of partitioning uneven datasets into similar datasets with equal density.
Clusters with equal density were obtained in the local level of distributed sites.
Several local clusters were integrated into global clusters, and uneven datasets were
mined efficiently.
As indicated above, several DDM techniques, such as distributed clustering,
distributed frequent pattern mining, and distributed classification, were provided
in the literature. Table 1 summarizes the DDM literature and highlights the
characteristics of such techniques.
3 Multi-agent System
Multi-agent systems (MASs) are a computerized environment that ensures that intel-
ligent communication and interactions among agents can achieve certain tasks by
integrating involved agents. Such systems are mostly created by the agent-based
model, which is an intelligent software that features states and behaviors. Protocols
that control the communication and interaction among the involved agents manage
the interaction in such MASs (Shen et al. 2006).
A MAS is a group of agents that each have their own problem-solving capa-
bilities and can interact with one another to achieve an overall goal (Obeid and
Moubaiddin 2010). Agents are specialized problem-solving entities that have clear
boundaries, and they agents fulfill a clearly defined purpose and demonstrate flexible
and proactive behavior (Obeid and Moubaiddin 2010).
Table 1 DDM literature
References Machine learning DDM approach Information sharing Application Algorithm Publish year
techniques
Bandyopadhyay et al. Clustering Homogenous Low-Level DDM Sensor Network K-Means 2006
(2006)
Stahl et al. (2012) Classification Homogenous Low-Level DDM Mobile Sensor Naive Bayes 2004
Heterogeneous
Iddianozie et al. (2019) Ensemble Heterogeneous High-Level DDM Street Networks Random Forests 2019
Li et al. (2017) Clustering Homogenous Low-Level DDM Sensor MODE Density-Based 2017
Clustering
Ranwa et al. (2019) Classification Heterogeneous High-Level DDM Nonrecurrent • Classification Tree 2019
Congestion NRC • Random Forest
• Bayesian Network
Montero-Manso et al. Classification Homogenous Low-Level DDM Intrusion Detection • Random Forest 2018
(2018) • Support Vector
Machine
• Logistic Regression
Amir et al. (2018) Classification Homogenous Low-Level DDM Image Spam Distributed Associative 2018
Detection Memory Tree
Multi-agent Systems for Distributed Data Mining Techniques …
Jeong et al. (2019) Clustering Heterogeneous Low-Level DDM UAV System Raft Algorithm 2019
Chen and Huang (2019) Clustering Heterogeneous High-Level DDM Sensorless Drive K 2019
Diagnosis -Means, SOM, And
Cartographic Spectra
Januzaj et al. (2004) Clustering Homogenous Low-Level DDM – K-Means 2004
Density-Based
Clustering
Chen and Huang (2019) Ensemble Homogenous Low-Level DDM Healthcare Adaboost 2019
(continued)
67
Table 1 (continued)
68
References Machine learning DDM approach Information sharing Application Algorithm Publish year
techniques
Al Hwaitat et al. (2020) Ensemble Homogenous Low-Level DDM – Decision Trees 2001
Khan et al. (2020) Classification Heterogeneous Low-Level DDM – Bayesian Networks 2004
Almaiah et al. (2020) Meta-Learning Homogenous High-Level DDM Fraud And Intrusion JAM 1997
Detection
Qasem et al. (2021) Regression Heterogeneous High-Level Ddm – CMD 1999
BODHI
Almaiah et al. Clustering Homogeneous High-Level DDM Internet Of Things LPDA 2007
Bandyopadhyay et al. Clustering Homogeneous High-Level DDM Attacks KDEC 2006
(2006)
Tong et al. (2018) Clustering Homogenous High-Level DDM Felony Incidents Of • Density Based 2018
New York City • Affinity Propagation
• Spectral
M. H. Qasem et al.
Multi-agent Systems for Distributed Data Mining Techniques … 69
The literature provides many definitions of agents due to their diverse application-
specific features. According to the authors in Russell et al. (1995), an agent is “a flex-
ible autonomous entity capable of perceiving the environment through the sensors
connected to it.” Other researchers have corroborated this definition (Jain and Srini-
vasan 2010). Another point of view was expressed in Ye et al.; according to the
authors, an agent is “an encapsulated computational system that is situated in some
environment and this is capable of flexible, autonomous action in that environment
in order to meet its design objective.”
The above definitions, among others, frame agents in the context of a specific
field of study. However, the concept of agents is generic and applicable to many
disciplines. Dorri et al. (2018) proposed a generalized definition that considers the
fundamental abilities and features of agents and defines them as an entity that is
placed in an environment and senses different parameters used to make a decision
on the basis of the entity’s goal. The entity then conducts the necessary action on the
environment on the basis of this decision.
Agents in MASs are autonomous in that they can function independently without
guidance. Such agents have control over their internal state and actions. An agent
can independently whether to perform a requested action. Agents exhibit flexible
problem-solving behavior to achieve their objectives, which they are designed to
fulfill. Autonomous agents can interact with other agents by using a specific commu-
nication language. Thus, they can take part in large social networks and respond to
changes and/or achieve goals with the assistance of other agents.
Agents typically operate in a dynamic, non-deterministic complex environment.
In MAS environments, no assumption exists with regard to global control and data
centralization and synchronization. Thus, the assumption is that agents in a MAS
operate with partial information or limited capabilities to solve the problem. Through
communication, agents share the information that they gathered for coordinating
their actions and increasing interoperation. The interactions between the agents can
be requests for information, particular services or an action to be performed by other
agents, and issues related to cooperation, coordination, and/or negotiation to arrange
related activities.
Agents must interact with one another to achieve their individual objectives.
However, such interaction is conceptualized as occurring at the knowledge level
(LiviuPenait 2005). Specifically, interaction can be conceived in terms of the goals
that should be followed, at a specified time, and by the assigned agent. Given that
agents are flexible problem solvers that have only partial control and knowledge
of the environment where they operate, interactions must be handled flexibly, and
agents must make runtime decisions as to whether or not to initiate interactions and
the nature and scope of these interactions. Through the following features, agents
can have broad applicability and solve complex tasks (Garcia et al. 2010):
• Sociability: Agents can share their knowledge and request information from other
agents so they can perform better in achieving their objective.
• Autonomy: Each agent is able to execute the decision-making process indepen-
dently and implement suitable actions.
70 M. H. Qasem et al.
• Proactivity: To predict possible future actions, each agent uses its history, sensed
parameters, and information of other agents, thereby enabling agents to take
effective actions that meet their objectives. From this ability, we can assume
that the same agent may resort to disparate actions when it is placed in different
environments.
Although agents working by themselves can take actions on the basis of autonomy,
their real benefit can be harnessed only when they collaborate with other agents.
Multiple agents that collaborate to solve a complex task are known as MASs
(McArthur et al. 2007).
Agent interactions and communications in MASs are controlled by the utilized
communication protocols. MAS protocols determine the external agent behavior
along with the internal agent structure without focusing on the internal behavior of
the agent. These specifications, along with the task implemented by the underlying
MAS, can determine the specification of the internal agent behavior. The existing
protocols for agent communication can be classified into two categories, namely,
centralized and decentralized protocols as illustrated in Fig. 4.
• Centralized protocols: In this protocol, the agent that is requesting to interact
with others, which is called an initiator, forwards its request to a center agent,
called broker. The center agent then selects the intended agent, which is called the
participant, on the basis of the specification given by the initiator and forwards
the initiator request accordingly.
The advantage of this type is the freedom of the broker to select a suitable
participant. However, the disadvantages include the need to record, manage, and
exchange the history of the participant to avoid an incorrect selection.
• Decentralized protocols: In decentralized protocols, the agent that is requesting
to directly interact with the participant obtains the address and status information
from the registry agent, which functions as a “yellow page” that contains no
information about the agent activities.
Through decision-making, or planning and control in MAS, an agent can accom-
plish its goals by determining the actions that it should perform. The decision-making
problem can either be episodic or sequential (Claes et al. 2011). The output of the
episodic problem is a single action, while the sequential problem produces a sequence
of actions or policy. The decision-making algorithm is evaluated on the basis of policy
optimality, search completeness, and time and space complexities.
This section presents the seven important features of MAS and discusses the different
categorizations that emerge when each feature is considered.
1. Heterogeneity: MAS can be divided into two categories, namely, homoge-
neous and heterogeneous, on the basis of agent heterogeneity. Homogeneous
MASs include agents that have the same characteristics and functionalities,
while heterogeneous MASs are composed of those with diverse features (Wen
et al. 2013).
2. Topology: Topology is the location and relations of agents. MAS topology can
be categorized as either static or dynamic. In a static topology, the position and
relations of an agent remain unchanged over the agent’s lifetime. In a dynamic
MAS topology, the position and relations of an agent change as the agent moves,
leaves or joins the MAS, or establishes new communications, such as relations,
with other agents (Li et al. 2014).
3. Data transmission frequency: Agents sense the environment and share the
data they sensed data with other agents either in a time-triggered or an event-
triggered way. In the time-triggered approach, the agent continuously senses the
environment, collects data, and transmits all newly sensed data to other agents
in pre-defined time intervals. In the event-triggered approach, the agent senses
the environment only when a particular event takes place. Subsequently, the
agent sends the collected data to other agents.
4. Agreement parameters: In certain MAS applications, agents need to agree on
particular parameters referred to as metrics. MASs are classified as first, second,
or higher order on the basis of the number of metrics. In the first order, agents
collaborate to agree on one metric (Du et al. 2013).
5. Mobility: Agents can be classified as static or mobile on the basis of their dynam-
icity. Static agents are always located in the same position in the environment.
72 M. H. Qasem et al.
By contrast, mobile agents can move around the environment. A mobile agent
can be hosted by other agents, that is, it uses the resources of other agents,
monitors them, or senses the environment from the position of other agents so
that it can perform actions (Ma et al. 2015).
6. Delay consideration: Agents might face multiple sources of delay when
performing tasks. Such delay sources include those communication media,
including wireless or wired, which are used by agents to exchange data, or
scheduling resources for each agent. MASs can be classified into two groups,
namely, with or without delay, depending on whether the delays are substantial
and relevant (Olfati-Saber and Murray 2004). The former considers the delay
sources, while the latter assumes that no delay sources exist. The latter instance
is simplified because no communication and processing delays exist. However,
most real-world applications always encounter non-trivial delays.
7. Leadership: We consider the existence of a leader, that is, an agent that defines
goals and tasks for the other agents on the basis of one global goal. MAS can be
categorized as either as leaderless or leader follow depending on the presence
or absence of such a leader (Du et al. 2013).
MASs have been applied to various fields, including cloud computing, robotics,
smart grid, modeling complex systems, and social network. In cloud computing, the
complexity in computer networks increases considerably as a result of the emer-
gence of new technologies and proliferation of Internet-connected devices. Agents
are primarily used to address this complexity given the broad range of applications
of MAS in various networks.
Bouchemal (2018) proposed an approach that is based on MASs and the cloud
computing concept to assist users, such as employees, managers, and customers,
in being permanently connected to the ERP system. The aforementioned author
proposed to assign a cloud manager agent at the cloud level to serve as a medi-
ator between the user and the ERP manager agents. The author initiated a prelim-
inary implementation by using the JADE-LEAP framework. Januário et al. (2018)
presented an architecture that is dependent on a hierarchical MAS to enhance
resilience, with particular attention paid to wireless sensor and actuator networks.
An IPv6 test bed that comprised several distributed devices was used to evaluate
the proposed framework; this evaluation involved the analysis of performance and
communication link health. The test bed results show the relevance of the proposed
approach.
Over the past two decades, the use of agents for robotics has been studied. Duan
et al. (2012) proposed an agent-based soccer robot to study such complexity. Agents
are grouped into teams. The agents in a team learn knowledge about the opponent
team and determine possible actions by interacting with the environment. These
agents then share the policies they have learned with other agents in their team.
Multi-agent Systems for Distributed Data Mining Techniques … 73
The popularity of social networks has exponentially increased with the rise in the
number of Internet users. A social network comprises various actors, such as users,
groups, and services (Jiang and Jiang 2014). The complexity of social networks orig-
inates from its dynamicity, that is, a large number of participants joining or leaving
the network or establishing new connections, and its broad range of applications and
services. MAS may potentially overcome the complexity of social networks. Gatti
et al. (2013) proposed an agent-based method for predicting user behavior, including
likes, posts, and follows, in social networks, such as Twitter. Such authors proposed
the use of multiple agents, such as actors, which are distributed throughout the social
network to gather a dataset of the user behavior. The agents then conduct topic and
sentiment classification on the data of each particular to create a user profile. Finally,
the user profile is input into a prediction system that predicts future user behavior,
including likes, topics, replies, posts, and shares.
In their work, Ma and Zhang (2014) considered a school as a social network. They
used MAS to help school managers finding the relations between the allocation of
funds to different school programs, such as athletic, academic, and cultural activities,
and school performance. In this approach, students, teachers, and school departments
are agents who collaboratively form the social network, that is, the school. A hierar-
chical structure is developed to organize the interactions of each student with other
students or teachers, the fund distribution policy from previous years, and the perfor-
mance of students in the current and past years. This structure is then input into a
learning function that appraises the relation between the fund distribution policy and
the academic performance of the students.
A MAS platform is a generic framework that provides the support required to perform
agent-based simulations. MAS systems are designed to make the creation and eval-
uation of multi-agent models easier. MASs have been used in different fields as
they provide logical, scientific, realistic, and efficient methods for deciphering and
solving challenges (Der et al. 2007). In addition to their ability to communicate
with one another and with their alternative behavior, MASs also feature self-healing
defense systems (LiviuPenait 2005; Charlton and Cattoni 2001). MASwill adapt to
situational variations and communicate and share information to achieve a common
goal without external interventions (LiviuPenait 2005).
During the last two decades, different Agent Platforms (APs) have been built by
the research community; these APs are either general-purpose or address a specific
domain of usage. Several platforms are no longer updated, although the introduction
of new versions of other platforms continues. In the meantime, a growing number of
new initiatives continue to be introduced by the agent-oriented research community.
The nature of these sites competes with that of their user base. Because of the large
number of platforms, people usually rely on word of mouth, past experiences, and
platform ads to decide the most appropriate one for their intent. Recently, however,
Multi-agent Systems for Distributed Data Mining Techniques … 75
people have become interested in using MASs to take advantage of agent technologies
that are increasingly dependent on solid survey products.
Several tools to produce agents have grown, and several more are being devel-
oped. Choosing an effective agent is a critical factor in agent growth. Java Agent
Development Environment (JADE) and Smart Python Agent Development Environ-
ment (SPADE) are the common modeling and assessment methods for evaluating
performance metrics in lectures, which are better suited for agent development.
There are also numerous agent structures, such as GAMA and PANGEA. Imple-
mentation of agents in this framework is based on requirements set by the Foun-
dation of Intelligent Physical Agents (FIPA) (FIPA 2002a). FIPA an association of
standards for the IEEE Computer Society, it is supporting agent-based technology
and the interoperability of its standards with other technologies. FIPA establishes
standards for computer software, interacting agents, and agent-based systems and
defines an agent as a computational process that implements the autonomous func-
tionality of an application. Agents are autonomous, that is, they are independent,
self-governing, and exercise control over their actions. Such agents are reactive and
respond to variations in the environment. Agents learn from previous experiences,
make impromptu tactical adjustments, and have the initiative to achieve their goals as
they are goal-directed, thereby demonstrating their proactivity and flexibility (Liviu-
Penait 2005). Also, agents are social, able to interact with humans and other agents,
mobile, and move easily from one machine to the other. Given their robust nature,
they can achieve continuity and vigorous fault recovery to meet their goals (Charlton
and Cattoni 2001). The FIPA reference model comprises the following (The FIPA
Specifications):
• Agent Management System: This system registers and deregister agents,
provides a directory of agent descriptions in alphabetical order, supervises and
controls access to the AP, and delivers life cycle management facilities.
• Directory Facilitator: This facilitator provides an indexed almanac to the services
given by other agents and registers and deregisters the agents.
• Agent Communication Channel: It facilitates inter-agent communication inter-
nally and to agents traversing through platforms.
• Internal Platform Message Transport: It ensures that message forwarding
services take place within a platform.
FIPA Open Source (FIPA-OS) is used to develop agents that conform to FIPA
benchmarks. Figure 5 shows the FIPA-OSAgent Platform. FIPA-OS has the following
abilities: maintenance of agent shells from various categories; message and conversa-
tion management; dynamic alignments to support multiple IPMTs; diagnostics and
virtualization tools; abstract interfaces; and software designs (FIPA 2002b). FIPA
Agent Communication Language (ACL) is used by the agents to communicate. The
four FIPA ACL components are conversation, ACL message, syntax, and semantics.
The message syntax is domain-independent and does not need to be defined for all
addressed domains.
The semantics/ontologies help define the names and data types to be used in
messages and their meanings. All involved participants interpret the messages in the
76 M. H. Qasem et al.
colloquy explicitly (Charlton et al. 2002). The FIPA standards have played a major
role in the development of MASs (Nicholas 1998).
Next, we discuss different agent frameworks:
A. Java Agent Development Framework (JADE)
JADE is a software framework that is fully implemented in the Java language.
This framework simplifies the implementation of MASs through middleware
that is compliant with FIPA specifications and by using a set of graphical tools
that support the debugging and deployment phases. A JADE-based system
can be distributed across machines, which do not require the same OS to be
shared, and a remote GUI can be used to control the configuration, which can
be changed during runtime by transferring agents from machine to machine as
necessary. The minimumsystem requirement for JADE is the Java 5 runtime
environment or the JDK (Jain and Srinivasan 2010).
Hudaib et al. (2017)proposed a semi-centralized protocol to address the
limitations of existing protocols and enable the development of a robust and
adaptable solution in a highly dynamic environment. The outcome of the JADE
framework implementation proves that the proposed protocol is highly capable.
The proposed protocol has better time and communication overhead under
various conditions than the Contract Net Protocol.
B. Smart Python Multi-agent Development Environment (SPADE)
SPADE is an instant messaging-based MAS platform written in Python. In
SPADE, agents can chat with other agents and humans. A SPADE agent
comprises a procedure for connecting to the platform, message disseminator,
and several behaviors to which the messages are transmitted. Each agent
establishes a platform link by using a unique Jabber identifier and valid
platform.
A SPADE agent has built-in jabber technology. Through this technology, an
agent who is connected to the SPADE platform is automatically registered and
remains active, constantly communicating with the agent platform.
Multi-agent Systems for Distributed Data Mining Techniques … 77
An agent can exhibit several behaviors, each of which is attached with a message
template. This template determines the type of message to be received by a
behavior.
Goryashchenko (2019) described the group formation problem in MASs. In
this study, SPADE MAS is selected after its scalability and performance for
several hundreds of agents were evaluated. A greedy algorithm for agent group
formation is proposed and implemented using SPADE MAS. The character-
istics of the implemented algorithm are studied. The results showed that the
performance of the proposed SPADE-based framework allows the implemen-
tation of additional complex algorithms for the formation of an agent group,
which may provide close-to-optimal results.
C. GAma Modeling Language(GAMA)
GAMA is a simulation platform that offers a complete modeling and simulation
development environment where modelers, field experts, and computer scien-
tists can build multi-agent simulations that are spatially explicit. The platform
enables capabilities related to the right combination of multilevel modeling,
3D visualization, and GIS data management. For support in rapid design and
prototyping, large models written in the GAML agent-oriented language can
be built in GAMA with an optional graphical modeling technique. Agents can
be instantiated by users from GIS data, files, or databases; large-scale simu-
lations reaching millions of agents can be run, and rich user interfaces (UIs)
can be designed to support in-depth inspections on agents, multiple multi-layer
2D/3D displays, and aspects, and user-controlled panels and actions (Kravari
et al. 2010).
A methodology was developed by Ruiz-Chavez et al. (2018) that identifies
the influencing factors of solid waste management in the Centro Histórico de
Quito. This method involves simulating with georeferenced multi-agents that
improve waste handling. A simulation model was built using GAMA with three
types of agents, namely, people, collectors, and waste containers, in an environ-
ment modeled with geographic information systems (GIS) layers. According
to the results, the complexity of waste handling increases with the population
in the area. Modifying the routers is insufficient. Thus, the number of collectors
must be increased.
D. EMERALD
EMERALD is a novel knowledge-based implementation framework wherein
trusted third-party reasoning services can be used for interoperable reasoning
among agents in the Semantic Web. The benefit of this framework is that each
agent can trade its position justification arguments even with agents that do
not follow the same rule logic or paradigm. EMERALD is fully Foundation of
Intelligent Physical Agents (FIPA)-compliant and established on top of JADE.
78 M. H. Qasem et al.
It is the only agent platform that supports reputation and trust mechanisms that
enable efficient decision-making and trustworthiness in MASs. Thus far, it has
been employed to examine how agents behave on behalf of their users in trading
and similar cases.
Kravari et al. (2010) took advantage of EMERALD to develop a MAS on top
of JADE. They used a reusable agent prototype for knowledge-customizable agent
behavior, trusted third-party reasoning services, and a reputation mechanism for
ensuring trust in the framework. The authors presented a use case scenario to show
the proposed framework’s viability.
Based on the MAS status (open or closed system) and communication type
(symmetric or asymmetric), Kravari et al. (2012) defined the space of potential
methods of interoperation for heterogeneous MASs. Their study showed how hetero-
geneous MASs can interoperate and eventually automate collaboration between
communities using one of these techniques. To exemplify the method, the authors
used two SW-enabled MASs, namely, Rule Responder and EMERALD, which help
user communities according to declarative SW and multi-agent standards, including
OWL, RDF, FIPA, and RuleML. This interoperation depends on high-quality rules
and facts and utilizes a declarative, knowledge-based method that allows information
agents to generate consistent and smart decisions. Finally, this work demonstrated the
interoperation’s added value by presenting multi-step interaction use cases between
agents from both MAS communities. Table 2 present a compersion between agent
frameworks.
MASs are applied in diverse fields to resolve and decipher challenges through
the technical, logical, efficient, and practical methods of the systems. MASs have
self-healing protection systems and the capacity to interact with one another and
their alternative actions. These intelligent agent systems can interact and exchange
knowledge to attain common objectives without external intervention as well asadjust
to situational changes. A MAS may involve a combination of agents and humans.
Agents in a MAS have only a local system view and are autonomous. No agent in
a MAS is granted supervisory or governing authority. Table 3 is summarizedMAS
literature.
MASs provides an architecture for distributed problem solving, and Distributed Data
Mining (DDM) algorithms focus on one category of such problem-solving tasks,
namely, distributed datamodeling and analysis. Here we present a viewpoint on DDM
algorithms in the MAS context and offer a general discussion on the relationship
between MASs and DDM..
MASs frequently handle complex applications that need distributingproblem-
solving. Meanwhile, DDM is a complex system focused on data mining processes
and resource distribution over networks. Scalability lies at the core of DDM systems.
Table 2 Comparative analysis of different agent frameworks
Framework Performance Stability Robustness Programming Operating Popularity Platform Standard Communication
language systems security compatibility
JADE Very High High High JAVA Any with The most Strong FIPA, CORBA ACL
JVM popular
SPADE Very High High High Python Any High Strong FIPA ACL
GAMA Good Good Good GMAL Mac OS X, Low Average FIPA, (GIS, 3D ACL
Windows, capabilities)
Linux
EMERALD High (on High High Java, JESS, Any with Low Strong FIPA, Semantic ACL
JADE) RuleML, Prolog JVM Web standards
Multi-agent Systems for Distributed Data Mining Techniques …
(plus use of
XML, RDF)
79
80
Given that system configurations may sometimes change, DDM system design looks
at many details regarding software engineering, such as extensibility, reusability, and
robustness. Therefore, the characteristics of agents are favorable for DDM systems
(Sawant and Shah 2013).
MAS is appropriate for distributed problem solving because it allows the creation
of autonomous, goal-oriented agents that operate in shared environments with coor-
dination and communication capabilities. This mechanism is beneficial for DDM in
implementing different distributed clustering, prediction, and classification methods.
For example, an agent may seek assistance from other agents in classifying instances
that cannot be easily classified locally. The agent communicates its beliefs or
outcomes, and the other agents decide whether using such beliefs will be advan-
tageous for classifying instances and adjusting their prior assumptions about each
data class.
In MASs, several agents mine and coordinate together. These systems are partic-
ularly appropriate for DDM because memory cost and computation time can be
decreased by having agents at the global level and deploying numerous agents to
each site of the distributed data. Agents can adjust to the faults and errors arising
in the entire system by cooperating with all distributed data sites. A MAS consists
of several agents that can perform tasks that a single agent cannot easily carry out.
MAS technology is classified under artificial intelligence; thus, its success depends
on the coordination of individual agent behaviors (Januário et al. 2018).
Agent technology in DDM improves scalability and efficiency by decreasing
network traffic, that is, only the code and not the data is shared among distributed
sites, thereby decreasing bandwidth and enabling efficient system scaling without
increasing system complexity (Gan et al. 2017). Agents adapt easily and react vigor-
ously to changing environments and operate asynchronously; that is, an agent discon-
nects from the network and automatically reconnects after completing its tasks (Gan
et al. 2017). Finally, agents are fault-tolerant; they bypass faults in distributed data
sites and shelter on reliable distributed data sites. MAS agents obtain the following
basic properties.
• Autonomy
• Decentralization
• Limited system view.
Security is a vital concern. Enterprises will not adopt systems that cannot secure
their data. The concern is also called data security and trust-based agents. However,
a rigid security model that aims to guarantee security can degrade the scalability of
systems. Agents that can travel through system networks provide alternative methods.
Also, scalability is an important concern in distributed systems. Added human inter-
ventions or highly complex mechanisms, which may compromise performance, are
needed to inform every unit in the system about any system configuration update,
such as during the joining of a new data site. Concerning this issue, collaborative
learning agents can exchange and disseminate information (about modifications in
system configuration in this case) with one another, thereby enabling the system to
adapt at the individual agent level. Furthermore, as previously discussed, these agents
82 M. H. Qasem et al.
DDM with MAS considers data mining a basic foundation and is improved with
agents; thus, this novel data-mining method will inherit all the powerful properties of
agents and acquire favorable features. Constructing a DDM with MAS involves three
key characteristics: dynamic system configuration, interoperability, and performance
aspects.
• Interoperability: This characteristic involves the collaboration of agents both
inside and outside a system, thus enabling new agents to integrate seamlessly into
the system. The system architecture should be flexible and open to dealing with
such interaction, which includes service directory, communication protocol, and
integration policy.
Communication protocol involves message encryption, encoding, and trans-
portation between agents. These operations are standardized by the FIPA and
are accessible to the public. Most agent platforms are FIPA-compliant, such as
SPADE and JADE, and thus support interoperability among them. Integration
policy stipulates the behavior of a system when an external component, such as a
data site or agent, seeks to leave or join.
The policy is known by all system components and is maintained by a version
that can be updated periodically. Document updating (e.g., policy) requires that
the present version of the document be known by all system components, thereby
resulting in system inconsistency. A service directory publishes system docu-
ments, such as system configuration and policy; all components may validate the
document version to guarantee system consistency. Further adoption may consider
how a service directory can be published, propagated, and maintained.
• Dynamic system configuration: Concerning interoperability, handling dynamic
system configurations is difficult because of the complex nature of mining algo-
rithms and planning. A mining task may include several data sources and agents,
Multi-agent Systems for Distributed Data Mining Techniques … 83
with the latter being equipped with an algorithm and dealing with datasets. Any
modification in data affects the mining task because an agent may still be running
the algorithm during the update. Therefore, the task priority should be whether it
can be restarted. The ideal algorithm is preemptive; it can pause for an update in
the dataset and continue the execution after the change is completed. Adding an
agent expecting to increase the speed of the mining task is likewise ideal because
the agents must share the instance of agent memory.
• Performance: Performance can be enhanced or degraded depending on the
major constraint posed by data distribution. Tasks can be executed in parallel
in distributed environments; however, parallelism leads to concurrency concerns.
Thus, ADDM systems must ensure acceptable (if not improved) system perfor-
mance. Other performance issues may be considered. For instance, when the
system enhances user experience, it may also optimize the mining performance.
An agent may represent a user with a profile. ADDM must determine the user
profile before data analysis. A user profile presents a pattern of queries; ADDM
must be able to manage query caches or some relevant performance improvement
methods.
In this manner, ADDM may ignore certain repetitive queries by replying with a
previous query result from the query cache. An agent may also describe a data source
with a profile. Heterogeneous data sources lead to various processing and query
specifications. A less powerful data source may impair the performance of the entire
MAS; thus, an ADDM must consider alternative data sources to acquire input data for
a query. Agents may share information with peers by initiating communications in the
MAS for a costor searching the service directory to identify alternative data sources.
This process is the extent of the cost–benefit analysis between communication and
query processing. As shown in Fig. 6, an ADDM system can be generalized into a
set of components. System activities may be simplified into request and response,
each involving distinct components.
The following are the basic components of an ADDM.
• Data: Data isthe foundation layer of interest. In distributed environments, data
can be hosted in different forms, such as data streams, online relational databases,
and web pages, where data have varied purposes.
• Communication: The system selects the associated resources from the directory
service, which keeps a list of data sources, data schemas, mining algorithms, and
data types, among others. Communication protocols, such as client–server and
peer-to-peer, may change by system execution.
• Presentation: The UI cooperates with the user by receiving and responding. It
simplifies complicated distributed systems into user-friendly messages, including
visual reporting tools and network diagrams.
When a user requests data mining through the UI, the following components are
involved.
• Query optimization: A query optimizer analyzes the request to identify the
mining task type and then selects resources appropriate for the request. The query
optimizer also decides whether the tasks can be parallelized, given that the data
are distributed and can be mined in parallel.
• Discovery plan: A planner allocates sub-tasks with related resources. Medi-
ating agents play an important role during this stage by coordinating numerous
computing units, given that mining sub-tasks are accomplished asynchronously,
and the results from such tasks.
When a mining task is executed, the following components are involved.
• Local knowledge discovery (KD): Mining may occur locally at each data site
depending on the individual implementation to transform data into patterns that
sufficiently represent the data and can be reasonably transferred over the network.
• KD: Also called mining, KD runs the algorithm as entailed by the task to acquire
knowledge from the specified data source.
• Knowledge consolidation: The knowledge obtained from different sources
should be normalized to offer a meaningful and compact mining result to the user.
This component includes a complex technique that incorporates patterns or knowl-
edge from distributed sites. Combining homogeneous knowledge or patterns is
probable but challenging for heterogeneous cases.
comprises the global system is distributed. The platform implements emotional and
musical genre classification and then gives context information about songs from
social networks, such as Twitter and Facebook. The selected techniques base on
meta-classifiers to conduct single- and multi-label classification and obtain superior
results. The proposed platform performed better in multi-label classification than did
platforms from previous works. Moghadam and Ravanmehr proposed a new DDM
method named multi-agent hierarchical data mining (MHDM) to classify meteo-
rology data gathered from various sites that were broadly distributed around Iran.
The REPTree classification algorithm is executed in local agents to obtain the needed
information from raw data kept in local databases, which are optimized to operate in
MAS. The researchers assessed the proposed method by executing it on 20 million
pieces of meteorology data. Experimental results indicated that MHDM can substan-
tially enhance performance compared with parallel and centralized methods for KD
in cases with large volumes of data.
A MAS-based clustering framework capable of enhancing the initial cluster
centers at each agent was recently developed. Chaimontree et al. developed a collabo-
rative clustering framework that presented better performance over non-collaborative
agent-based clustering. This framework retained the particularity of each agent and
enabled information exchange, including private and non-shareable data, among local
agents to improve the capabilities of autonomous agents.
A clustering system based on cooperative agents through a common and central-
ized ERP database was presented by Mesbahi et al. to enhance decision support
in ERP systems. The authors used a MAS paradigm to distribute the complexity
of the k-means algorithm to numerous autonomous entities called agents, whose
objective is to assemble observations on or records of similar object classes. This
process will assist business decision-makers in making sound decisions and giving
superior response time using the MAS. The JADE platform is a convenient means
of implementing the proposed architecture while offering a complete set of services
and having agents comply with the FIPA specifications. Qayumi and Norta proposed
a MAS-based business intelligence (BI) system model called BI-MAS with compre-
hensive designing steps as a running case. Given the complexity of BI generation
in distributed environments, the use of such a system is diverse because of inte-
grated DDM and MAS technologies. Incorporating these frameworks in the content
of BI systems poses difficulties during the design, analysis, and testing stages in the
development life cycle. To demonstrate the proposed approach, the authors first
considered an evaluation process of identifying appropriate agent-oriented tech-
niques. Second, the researchers adopted the chosen methodologies in designing
and analyzing concepts for BI-MAS life cycles. Finally, they demonstrate a novel
verification and validation approach for BI-MAS life cycles. Sun et al. examined
distributed deep learning as a MAS problem and designed a MAS approach for
distributed training of deep neural networks. In their approach, the agents assess the
utility of each action and then implement the best response strategy based on their
estimated utilities mixed with -random exploration. The proposed approach, called
Slim-DP, differs from the typical data parallelism in that it only communicates a
subset of the gradient or the global model. The experimental results demonstrated
86 M. H. Qasem et al.
that Slim-DP can decrease communication cost more and obtain better speedup than
the typical data parallelism and its quantization version without loss of accuracy.
Table 4 summarizes the advantages of the methods in the literature.
4 Conclusion
References
A.A. Ali, P. Vařacha, S. Krayem, P. Žáček, A. Urbanek, Distributed data mining systems: techniques,
approaches and algorithms, in MATEC Web of Conferences, vol. 210. (EDP Sciences, 2018),
p. 04038
A. Amir, B. Srinivasan, A.I. Khan, Distributed classification for image spam detection. Multimedia
Tools Appl. 77(11), 13249–13278 (2018)
M.A. Almaiah, A. Al-Khasawneh, Investigating the main determinants of mobile cloud computing
adoption in university campus. Educ. Inf. Technol. 25(4), 3087–3107 (2020)
M. Adil, R. Khan, M.A. Almaiah, M. Al-Zahrani, M. Zakarya, M.S. Amjad, R. Ahmed, MAC-
AODV based mutual authentication scheme for constraint oriented networks. IEEE Access 4(8),
44459–44469 (2020)
M. Adil, R. Khan, M.A. Almaiah, M. Binsawad, J. Ali, A. Al Saaidah, Q.T.H. Ta, An efficient
load balancing scheme of energy gauge nodes to maximize the lifespan of constraint oriented
networks. IEEE Access 8, 148510–148527 (2020)
M. Adil, M.A. Almaiah, A. Omar Alsayed, O. Almomani, An anonymous channel categorization
scheme of edge nodes to detect jamming attacks in wireless sensor networks. Sensors 20(8), 2311
(2020)
A.K. Al Hwaitat, M.A. Almaiah, O. Almomani, M. Al-Zahrani, R.M. Al-Sayed, R.M. Asaifi,
K.K. Adhim, A. Althunibat, A. Alsaaidah, Improved security particle swarm optimization (PSO)
algorithm to detect radio jamming attacks in mobile networks. Quintana 11(4), 614–624 (2020)
M. Adil, R. Khan, J. Ali, B.H. Roh, Q.T. Ta, M.A. Almaiah, An energy proficient load balancing
routing scheme for wireless sensor networks to maximize their lifespan in an operational
environment. IEEE Access 31(8), 163209–163224 (2020)
M.A. Almaiah, Z. Dawahdeh, O. Almomani, A. Alsaaidah, A. Al-khasawneh, S. Khawatreh, A
new hybrid text encryption approach over mobile ad hoc network. Int. J. Electr. Comput. Eng.
(IJECE) 10(6), 6461–6471 (2020)
M.A. Almaiah, A. Al-Zahrani, O. Almomani, A.K. Alhwaitat, Classification of cyber security threats
on mobile devices and applications. Artif. Intell. Blockchain Future Cybersecur. Appl. 107
M.A. Almaiah, A new scheme for detecting malicious attacks in wireless sensor networks based on
blockchain technology. Artif. Intell. Blockchain Future Cybersecur. Appl. 217
M.A. Almaiah, M. Al-Zahrani, Multilayer neural network based on MIMO and channel estimation
for impulsive noise environment in mobile wireless networks. Int. J. Adv. Trends Comput. Sci.
Eng. 9(1), 315–321 (2020)
M.A. Almaiah, M.M. Alamri, Proposing a new technical quality requirements for mobile learning
applications. J. Theore. Appl. Inf. Technol 96(19) (2018)
S. Bandyopadhyay, C. Giannella, U. Maulik, H. Kargupta, K. Liu, S. Datta, Clustering distributed
data streams in peer-to-peer environments. Inf. Sci. 176(14), 1952–1985 (2006)
N. Bouchemal, N. Bouchemal, Intelligent ERP based multi agent systems and cloud computing.
In International Conference on Machine Learning for Networking (Springer, Cham, Nov. 2018),
pp. 378–386
D. Chiang, C. Lin, M. Chen, The adaptive approach for storage assignment by mining data of
warehouse management system for distribution centres. Enterp. Inf. Syst. 5(2), 219–234 (2011)
C.Y. Chen, J.J. Huang, Double deep autoencoder for heterogeneous distributed clustering.
Information 10(4), 144 (2019)
C. Clifton, M. Kantarcioglou, X. Lin, M. Zhu, Tools for privacy preserving distributed data mining.
ACM SIGKDD Exp. 4(2) (2002)
A. Cuzzocrea, Models and algorithms for high-performance distributed data mining. Elsevier J.
Parallel Distrib. Comput. 73(93), 281–283 (2013)
R. Claes, T. Holvoet, D. Weyns, A decentralized approach for anticipatory vehicle routing using
delegate multiagent systems. IEEE Trans. Intell. Transp. Syst. 12(2), 364–373 (2011)
P. Charlton, R. Cattoni, Evaluating the deployment of FIPA standards when developing application
services. Int. J. Pattern Recogn. Artif. Intell. 15(03) (2001)
Multi-agent Systems for Distributed Data Mining Techniques … 89
P. Charlton, R. Cattoni, A. Potrich, E. Mamdani, Evaluating the FIPA standards and their role in
achieving cooperation. In Multi-Agent Systems (IEEE Xplore, Aug. 2002)
A. Dorri, S.S. Kanhere, R. Jurdak, Multi-agent systems: A survey. IEEE. Access 6, 28573–28593
(2018)
G. Dudek, M.R. Jenkin, E. Milios, D. Wilkes, A taxonomy for multi-agent robotics. Auton. Robot.
3(4), 375–397 (1996)
Y. Duan, B.X. Cui, X.H. Xu, A multi-agent reinforcement learning approach to robot soccer. Artif.
Intell. Rev. 38(3), 193–211 (2012)
R. Domínguez, S. Cannella, J.M. Framinan, Scope: a multi-agent system tool for supply chain
network analysis, in EUROCON 2015-International Conference on Computer as a Tool
(EUROCON), IEEE (IEEE, 2015), pp. 1–5
H. Du, S. Li, S. Ding, Bounded consensus algorithms for multiagent systems in directed networks.
Asian J. Control 15(1), 282–291 (2013)
FIPA, FIPA Abstract Architecture Specification, SC 00001L (2002a). http://www.fipa.org/specs/fip
a00001/SC00001l.pdf
FIPA, SC00067F (2002b). http://www.fipa.org/specs/fipa00067/SC00067F.pdf
I.E. Foukarakis, A.I. Kostaridis, C.G. Biniaris, D.I. Kaklamani, I.S. Venieris, Webmages: An Agent
Platform Based on Web Services
W. Gan, J.C.W. Lin, H.C. Chao, J. Zhan, Data mining in distributed environment: a survey. Wiley
Interdiscip. Rev.: Data Mining Knowl. Dis. 7(6), e1216 (2017)
A.P. Garcia, J. Oliver, D. Gosch, An intelligent agent-based distributed architecture for smart-grid
integrated network management, in 2010 IEEE 35th Conference on Local Computer Networks
(LCN) (IEEE, 2010), pp. 1013–1018
A. González-Briones, F. De La Prieta, M. Mohamad, S. Omatu, J. Corchado, Multi-agent systems
applications in energy optimization problems: a state-of-the-art review. Energies 11(8), 1928
(2018)
M. Gatti, P. Cavalin, S.B. Neto, C. Pinhanez, C. dos Santos, D. Gribel, A.P. Appel, Large-scale
multi-agent-based modeling and simulation of microblogging-based online social network, in
International Workshop on Multi-Agent Systems and Agent-Based Simulation (Springer, 2013),
pp. 17–33
A. Goryashchenko, Algorithm and application development for the agents group formation in
a multi-agent system using SPADE system, in Future of Information and Communication
Conference. (Springer, Cham, Mar. 2019), pp. 1136–1143
A. Hudaib, M.H. Qasem, N. Obeid, FIPA-Based semi-centralized protocol for negotiation,
in Proceedings of the Computational Methods in Systems and Software (Springer, Cham, Sept.
2017), pp. 135–149
D. Helbing, Agent-based modeling, in Social self-organization (Springer, 2012), pp. 25–70
C. Iddianozie, G. McArdle, A transfer learning paradigm for spatial networks, in Proceedings of
the 34th ACM/SIGAPP Symposium on Applied Computing (ACM, Apr. 2019), pp. 659–666
I.F. Ilyas, X. Chu, X. Trends in cleaning relational data: Consistency and deduplication. Found.
Trends® Databases 5(4), 281–393 (2015)
T. Ishida, H. Yukoi, Y. Kakazu, Self-organized norms of behaviour under interactions of selfish
agents, in IEEE SMC ’99 Conference Proceedings and IEEE Xplore Systems, Man and
Cybernetics, 1999, Aug. 2002
E. Januzaj, H.P. Kriegel, M, Pfeifle, Dbdc: density based distributed clustering, in International
Conference on Extending Database Technology. (Springer, Berlin, Heidelberg, Mar. 2004),
pp. 88–105
S. Jeong, U. Choi, J. Ahn, Distributed clustering algorithm for UAV systems, in AIAA Scitech 2019
Forum, p. 1795 (2019)
L.C. Jain, D. Srinivasan, Innovations in Multi-agent Systems and Applications (Springer, 2010)
F. Januário, A. Cardoso, P. Gil, Multi-Agent framework for resilience enhancement over a
WSAN, in 2018 15th International Conference on Electrical Engineering/Electronics, Computer,
Telecommunications and Information Technology (ECTI-CON) (IEEE, July 2018), pp. 110–113
90 M. H. Qasem et al.
Y. Jiang, J. Jiang, Understanding social networks from a multiagent perspective. IEEE Trans. Parallel
Distrib. Syst. 25(10), 2743–2759 (2014)
N.R. Jennings, K. Sycara, M. Wooldridge, A Roadmap of Agent Research and Development,
Springer: Autonomous Agents and Multi-Agent Systems, vol. 1, Issue 1, pp 7–38 (1998)
K. Kasemsap, Multifaceted applications of data mining, business intelligence, and knowledge
management, in Intelligent Systems: Concepts, Methodologies, Tools, and Applications, (IGI
Global, 2018), pp. 810–825
K. Kargupt, Chan, Distributed and parallel data mining: emergence, growth and future directions,
in Advances in Distributed Data Mining, ed. by H. Kargupta, P. Chan (AAAI Press, 1999),
pp. 407–416
M.N. Khan, H.U. Rahman, M.A. Almaiah, M.Z. Khan, A. Khan, M. Raza, M. Al-Zahrani, O.
Almomani, R. Khan, Improving energy efficiency with content-based adaptive and dynamic
scheduling in wireless sensor networks. IEEE Access 25(8), 176495–176520 (2020)
K. Kannan, K. Raja, A. Rajakumar, P.K. Nizar Banu, E-Business Decision Support System for
Online Shopping using MAS with Ontology and JADE Methodology (2019)
K. Kravari, E. Kontopoulos, N. Bassiliades,. EMERALD: a multi-agent system for knowledge-based
reasoning interoperability in the semantic web, in Hellenic Conference on Artificial Intelligence
(Springer, Berlin, Heidelberg, May 2010), pp. 173–182
K. Kravari, N. Bassiliades, H. Boley, Cross-community interoperation between knowledge-based
multi-agent systems: A study on EMERALD and Rule Responder. Expert Syst. Appl. 39(10),
9571–9587 (2012)
H. Li, L. Xu, J. Wang, Z. Mo, Feature space theory in data mining: transformations between
extensions and intensions in knowledge representation. Expert Syst. 20(2), 60–71 (2003)
B. Liu, S. Cao, W. He, Distributed data mining for e-business. Inf. Technol. Manag. 12(2), 67–79
(2011)
T. Li, F. De la Prieta Pintado, J.M. Corchado, J. Bajo, Multi-source homogeneous data clustering
for multi-target detection from cluttered background with misdetection. Appl. Soft Comput. 60,
436–446 (2017)
R. Lu, K. Heung, A.H. Lashkari, A.A. Ghorbani, A lightweight privacy-preserving data aggregation
scheme for fog computing-enhanced IoT. IEEE Access 5, 3302–3312 (2017)
O. Lopez Ortega, F. Castro Espinoza, O. Perez-Cortes, An Intelligent Multiagent System to Create
and Classify Fractal Music (Springer-Verlag GmbH Austria, Jan. 2018)
A. Louati, S. Elkosantini, S. Darmoul, H. Louati, Multi-agent preemptive longest queue first system
to manage the crossing of emergency vehicles at interrupted intersections. Eur. Transp. Res. Rev.
10(2), 52 (2018)
H. Li, C. Ming, S. Shen, W. Wong, Event-triggered control for multi-agent systems with randomly
occurring nonlinear dynamics and time-varying delay. J. Franklin Inst. 351(5), 2582–2599 (2014)
C. Moemeng, V. Gorodetsky, V., Z. Zuo, Y. Yang, C. Zhang, Agent-based distributed data mining:
a survey, in Data Mining and Multi-Agent Integration, (Springer, Boston, MA, 2009), pp. 47–58
R. Mendes, J.P. Vilela, Privacy-preserving data mining: methods, metrics, and applications. IEEE
Access 5, 10562–10582 (2017)
P. Montero-Manso, L. Morán-Fernández, V. Bolón-Canedo, J.A. Vilar, A. Alonso-Betanzos,
Distributed classification based on distances between probability distributions in feature space.
Inf. Sci. (2018)
L.S. Melo, R.F. Sampaio, R.P.S. Leão, G.C. Barroso, J.R. Bezerra, Python-based multi-agent
platform for application on power grids. Int. Trans. Electr. Energy Syst. e12012 (2019).
S.D. McArthur, E.M. Davidson, V.M. Catterson, A.L. Dimeas, N.D. Hatziargyriou, F. Ponci,
T. Funabashi, Multi-agent systems for power engineering applicationsâ Tpart i: concepts,
approaches, and technical challenges. IEEE Trans. Power Syst. 22(4), 1743–1752 (2007)
L. Ma, H. Min, S. Wang, Y. Liu, S. Liao, An overview of research in distributed attitude coordination
control. IEEE/CAA J. Automatica Sinica 2(2), 121–133 (2015)
Multi-agent Systems for Distributed Data Mining Techniques … 91
L. Ma, Y. Zhang, Hierarchical social network analysis using multiagent systems: a school system
case, in 2014 IEEE International Conference on Systems, Man and Cybernetics (SMC) (IEEE,
2014), pp. 1412–1419
L. Niu, N. Feng, Research on cooperation control of chassis multi-agent, in 2010 International
Conference on Computer, Mechatronics, Control and Electronic Engineering, vol. 2 (IEEE, Aug.
2010), pp. 464–467
M.A. Ouda, S.A. Salem, I.A. Ali, E.S.M. Saad, Privacy-preserving data mining (PPDM) method
for horizontally partitioned data. Int. J. Comput. Sci. 9(5), 339–347 (2012)
N. Obeid, A. Moubaiddin, A. Towards a formal model of knowledge sharing in complex systems, in
Smart Information and Knowledge Management (Springer, Berlin, Heidelberg, 2010), pp. 53–82
R. Olszewski, P. Pałka, A. Turek, B. Kietlińska, T. Płatkowski, M. Borkowski, Spatiotemporal
modeling of the smart city residents’ activity with multi-agent systems. Appl. Sci. 9(10), 2059
(2019)
R. Olfati-Saber, R.M. Murray, Consensus problems in networks of agents with switching topology
and time-delays. IEEE Trans. Autom. Control 49(9), 1520–1533 (2004)
A. Patel, W. Qi, C. Wills, A review and future research directions of secure and trustworthy mobile
agent-based e-marketplace systems. Inf. Manag. Comput. Secur. 18(3), 144–161 (2010)
L. Penait, S. Luke, Co-operative Multi-Agent Learning: The State of the Art, Springer Science +
Business Media, Netherlands: Autonomous Agents and Multi-Agent Systems, vol. 11, pp. 387–434
(2005)
M.H. Qasem, N. Obeid, A. Hudaib M.A Almaiah, A. Al-Zahrani, A. Al-khasawneh, Multi-Agent
System Combined with Distributed Data Mining for Mutual Collaboration Classification (IEEE
Access, 20 Apr. 2021)
A.M. Ranwa, F. Bilal, F., Q. Alejandro, Distributed Classification of Urban Congestion Using
VANET (2019). arXiv:1904.12685.
Russell, A.P. Norvig, Intelligence, “A modern approach”, vol. 25 (Artificial Intelligence. Prentice-
Hall, Egnlewood Cliffs, 1995), p. 27
Y. Rizk, M. Awad, E.W. Tunstel, Decision making in multiagent systems: a survey. IEEE Trans.
Cogn. Dev. Syst. 10(3), 514–529 (2018)
H. Rezaee, F. Abdollahi, Average consensus over high-order multiagent systems. IEEE Trans.
Autom. Control 60(11), 3047–3052 (2015)
Y. Ren, D. Fan, Q. Feng, Z. Wang, B. Sun, D. Yang, Agent-based restoration approach for reliability
with load balancing on smart grids. Appl. Energy 249, 46–57 (2019)
Z. Ruiz-Chavez, J. Salvador-Meneses, S. Díaz-Quilachamín, C. Mejía-Astudillo, (, October). Solid
Waste Management using Georeferenced Multi-agent Systems. In 2018 IEEE Third Ecuador
Technical Chapters Meeting (ETCM) (IEEE, Oct. 2018), pp. 1–6
V. Sawant, K. Shah, A review of distributed data mining using agents. Int. J. Adv. Technol. Eng.
Res. (IJATER) 3(5), 27–33 (2013)
F. Stahl, M.M. Gaber, P. Aldridge, D. May, H. Liu, M. Bramer, S.Y. Philip, Homogeneous and
heterogeneous distributed classification for pocket data mining, in Transactions on Large-Scale
Data-and Knowledge-Centered Systems V (Springer, Berlin, Heidelberg, 2012), pp. 183–205
S. Sharmila, S. Vijayarani, Association rule hiding using firefly optimization algorithm, In Interna-
tional Conference on Intelligent Systems Design and Applications (Springer, Cham, Dec. 2018),
pp. 699–708
W. Shen, et al. Applications of agent-based systems in intelligent manufacturing: an updated review.
Adv. Eng. Inf. 20.4, 415–431 (2006)
C.S. Shih, Cooperative Adaptive Control for Multi-Agent Systems (2018)
A.Q. Santos, R.M. Monaro, D.V. Coury, M. Oleskovicz, M., A new real-time multi-agent system
for under frequency load shedding in a smart grid context. Electric Power Syst. Res. 174, 105851
(2019)
S. Seng, K.K. Li, W.L. Chan, Z. Xiangjun, D. Xianzhong, Agent-based Self-healing Protection
System, in IEEE transactions on Power Delivery, vol. 21, Issue 02, Apr. 2006
92 M. H. Qasem et al.
Giuseppe Ciaburro
Keywords Time Series analysis · Deep learning · Smart Cities · Time Series
forecasting
1 Introduction
G. Ciaburro (B)
Department of Architecture and Industrial Design, Università degli Studi della Campania Luigi
Vanvitelli, 81031 Borgo San Lorenzo, Aversa, CE, Italy
e-mail: [email protected]
of this family of techniques is the fact that the machines, as the name suggests, can
automatically learn with experience to perform certain tasks, improving their perfor-
mance more and more over time (Ciaburro et al. 2020a). Specifically, in recent years
the use of methodologies based on Deep Learning has become widespread. It is a
sub-category of Machine Learning and indicates that family of methods belonging to
artificial intelligence that refers to algorithms inspired by the structure and function
of the brain, or artificial neural networks (Mohri et al. 2018). These architectures are
applied in different contexts, such as in Computer Vision, audio and spoken language
recognition, natural language processing and bioinformatics (Mohri et al. 2018).
In this chapter we will analyze and describe some methods based on Deep
Learning to capture the recurring structures from Times Series Data in order to be able
to elaborate forecast scenarios. The remainder of the chapter is structured as follows:
Sect. 2 introduces the Times Series, examining their characteristics and describes
some of the traditional statistical techniques used for their analysis. Section 3 intro-
duces deep learning-based methods for time series analysis. Section 4 reports the
results of a study on a model for predicting the noise levels produced near a road
infrastructure using a model based on recurrent neural networks. Finally, Sect. 5
summarizes the results obtained in applying these methods to real cases, highlighting
their merits, and listing their limits.
Time series modeling has been an area of growing interest for many disciplines
and in which much effort has been devoted to the development of new methods
and techniques. Its objective is to provide the researcher with the mathematical
representation of a time series, which allows to grasp in whole or in part the most
relevant characteristics of the real phenomenon, from the information contained in
the data (Granger and Newbold 2014). Although various models oriented to the
representation of time series have been proposed in the literature, their usefulness
depends on the degree of similarity between the dynamics of the series generation
process and the mathematical formulation of the model with which it is represented.
A time series is a stochastic process in which the set of observations are taken
sequentially over time, usually at equal intervals. Therefore, a time series model for
observed data is a specification of the joint distribution, possibly only the means and
covariances, of a sequence of random variables (Tsay 2005).
It is commonly accepted that many time series exhibit non-linear dynamic behav-
iors, the complexity of which makes it impossible to formulate a model based on phys-
ical or economic laws that adequately represent their evolution (Lütkepohl 2013).
The problem of formulating the model is aggravated by the presence of atypical
observations and structural changes, for which there are no mathematical models
that allow their representation in the non-linear case (Brockwell et al. 2016). One of
the characteristics that distinguishes time series from other types of statistical data
is the fact that, in general, the values of the series at different times are correlated.
96 G. Ciaburro
In this way, the fundamental problem in the analysis of time series is the study of
the correlation pattern of the values of the series and the elaboration of a statistical
model that explains its structure (Lütkepohl 2005).
Traditionally, time series have been considered as the sum of two contributions:
a deterministic component and a random disturbance contribution. We can therefore
represent a Time Series through Eq. (1):
Fig. 2 Components of a time series: a Trend component (T); b Seasonal Components (S); c Cyclical
component (C); d Irregular component (E)
98 G. Ciaburro
• Seasonal Components (S): Both cyclical and seasonal changes can be understood
as the undulating movement of the data above and below the trend line. But
the difference between one and the other has to do with the range of data being
analyzed. In the case of the cyclical components (Fig. 2c) we refer to processes that
occur more widely within the observation interval, while the seasonal components
occur with less separation over time. For example, when analyzing the data shown
in Fig. 1, a seasonal component (Fig. 2b) may refer to changes in the number
of passengers in the summer when more air travel is expected due to summer
holidays.
• Cyclical component (C): These are fluctuations with a periodicity of between one
and a half and up to ten years, depending on the definition of the cycle used.
They tend to be less frequent and less systematic than seasonal ones. A cyclical
component may present expansion or contraction over a longer period.
• Irregular component (E) (Fig. 2d): it relates to data that are very deviating from
the trend, as occurs when an aircraft is stopped. In this case, decisions need to be
made on how to handle these outliers, as including them in the calculations would
add a lot of noise to their description. For this reason, a value is usually estimated
to replace anomalous data.
The methods of time series analysis consider the fact that the data collected in
different periods may have some characteristics of autocorrelation, trend or season-
ality that must be taken into consideration to model the behavior of the observed
series (Baydogan and Runger 2016). The application of these methods serves two
purposes: to understand the forces of influence in the data and to discover the structure
that produced them.
The components that we have previously introduced can be combined in different
ways, the most common way of representing a time series is as a function of the
trend, seasonality, and irregularity components, that is through Eq. (2).
Yt = f (Tt , St , εt ) = Tt + St + εt , (2)
Yt = Tt + εt , (3)
Tt = α + β ∗ t, (4)
where α, β are constant, the latter is also called the trend component. The coefficient β
represents the average increase in Y per unit of time. Other authors prefer to call it the
trend of the parameter β, which represents the slope of the linear model. According
to the latter interpretation, the trend is the variation of the average level per unit of
time. The trend in Eq. (4) is a deterministic function of time and is known as the
global linear trend. If β is greater than zero there is a linear growth, while if β is less
than zero there is a decline over a long period. In the case where β = 0, it means
that there is no growth or decline for a long time in the series, and it is said to be
a trendless series. A convenient way to estimate α and β is with the least square’s
method (Ng and Perron 2005).
The exponential trend shape is used when the trend reflects an almost constant
percentage of growth. Since it poses a risk to assume that exponential growth will
continue indefinitely, then one can adapt the logistic trend form that provides S-like
trends.
In case a seasonal component is detected, it must be added to the model. This
addition can be made in different ways, the first and simplest is the additive model
which provides a simple sum between the different contributions, as already reported
in Eq. (2). Another model is the multiplicative one in which the three components
(trend, seasonality, and error) are multiplied by each other. Finally, a third model
provides a mixed mode between the two just seen, then we can multiply the trend
and seasonality components together, and then add the term due to the error to the
result. In all cases, the component due to the error must have the characteristics of
white noise, therefore an average of zero and constant variance.
An additive model is suitable, for example, when the seasonal component does
not depend on other components, such as for the trend, such as, it can be used for
series of time models that show constant seasonal variation. If, on the contrary, the
seasonality varies with the trend, such as, the time series shows an increasing or
decreasing seasonal variation, the most appropriate model is the multiplicative one:
However, the multiplicative model can be transformed into additive using logarithms.
100 G. Ciaburro
The moving average (MA) model models time series, based on the moving average
of their past terms. It is an alternative tool for estimating the trend and cyclical
components (Said and Dickey 1984). The moving average acts by reducing the
amplitude of the oscillations, so the greater the number of observations used for
calculating the average, the greater the flattening. Recall that the t-term moving
average is equal to the arithmetic mean of the series.
The MA model of order p is represented by Eq. (6).
the man. In this way it is possible to acquire greater independence and efficiency. In
fact, these machines are able to autonomously reach the goal for which they were
programmed by extracting recurring information from a set of data with the aid of
Machine Learning methods. However, it has been observed that some activities of
immediate resolution for humans are complicated for a machine if only the classical
methods of Machine Learning are applied. To address this problem, new types of
algorithms called Deep Learning have been developed, capable of providing the
machine with the ability to extract simple information from a set of data and then
combine them to obtain more elaborate information. The final result is achieved
by processing, in a hierarchical way, different levels of representation of the data
corresponding to characteristics, factors or concepts.
feature is the dynamism of the synaptic configuration: The number of synapses can
increase or decrease depending on the stimuli that the network receives.
A biological network receives external data and signals that are perceived through
the senses. These are processed into information through a massive number of inter-
connected neurons in a non-linear and variable structure in response to those external
data and stimuli. Likewise, artificial neural networks are non-linear structures of
statistical data organized as modeling tools. The ANNs receive external signals on
a layer of nodes, each of these input nodes is connected to various internal nodes
of the network which, typically, are organized on several levels so that each single
node can process the received signals by transmitting to subsequent levels the result
of his elaborations (Anitescu et al. 2019).
Typically, neural networks consist of three layers (Fig. 3):
• Input Layer: it is the first layer of the network and is the one that has the function
of receiving and processing the input signals, adapting them to the demands of
the neurons within the network.
• Hidden Layer: it is the intermediate layer that is in charge of the actual processing
process.
• Output Layer: it is the final layer where the processing results of the Hidden layer
are collected and adapted to the requests of the next block-level of the neural
network.
An artificial neuron represents the processing unit of the structure and its action
can therefore be represented by the following equation:
yi = f (xi ) (8)
104 G. Ciaburro
In Eq. (8), f is typically non-linear and is called the activation function. In Eq. (8)
the term xi is represented by Eq. (9).
xi = w ji + bi (9)
The term Big Data defines a large set of data in terms of volume and management
complexity that requires different processing methods and technologies than tradi-
tional software and computer architectures. To understand the difficulties inherent in
managing Big Data, we can think of a traditional database consisting of tables with
many records but a few dozen fields: In the case of Big Data, tools are required that
can manage the same number of records, but with thousands of fields. The problem
Time Series Data Analysis Using Deep Learning … 105
of Big Data management arises because the amount of data that modern infras-
tructures are able to generate has become really high: Web traffic, that generated
by modern smartphones, as well as the data detected by the countless quantities of
sensors installed for city monitoring represent only a few sources of data generation
(Tsai et al. 2015).
Huge volumes of heterogeneous data by source and format, which can be analyzed
in real time, constitute what are defined as Big Data. In recent years, the number of
data produced in the world has increased dramatically and companies are expected
to be able to produce Zettabytes of data by considering sources such as sensors, data,
financials, satellites, etc. The need for analysis on a single set of data is the cause of
the progressive increase in the size of the data sets. The goal is to extract patterns
and useful information compared to those that could be obtained by analyzing small
series as in the past. Another distinction between Big Data and traditional datasets is
that Big Data mainly includes data from heterogeneous sources, therefore they are
not formed only by structured data, such as databases, but also unstructured, such as
images, data collected by GPS, email, and information taken from social networks
(Rajaraman 2016).
To analyze huge amounts of data, different types of algorithms have been born
over the years. In addition to differing in the technologies used and the logic they
use to bring results, these algorithms are distinguished by the type of analysis they
conduct on the data. In particular, there are three types of analysis that can be carried
out on Big data:
• Descriptive analysis: It is the simplest type of analysis; these are studies that
aim to synthesize and describe raw data. These analyzes are conducted on large
databases with the aim of finding hidden correlations between the data, so that
useful information emerges for those conducting the analysis. In order to summa-
rize and clarify the dynamics and performance of pre-established metrics and
to obtain indications on the best way to approach future activities, it deals with
analyzing past events from different points of view. This type of analysis is of vital
importance for companies, which thanks to the results obtained are able to under-
stand what is happening in a specific domain of interest and make targeted choices
accordingly (Ristevski and Chen 2018). Although it is not necessary to use sophis-
ticated technologies to conduct these research, without the help of computational
power, humans would not be able to conduct these types of analyzes.
• Predictive analysis: It aims to predict what will happen in the future in a given
application field. To conduct these analyzes it is necessary to have previously
carried out a descriptive analysis on the data of interest, in order to have infor-
mation regarding the temporal evolution of the data and any correlations between
data sets. Subsequently, algorithms will be able to project this information in the
more or less near future and highlight what may be the trends and the results that
will be reached. It should be emphasized, however, that no statistical algorithm
is able to predict the future with accuracy, but it is still possible to arrive at very
good absolute probabilities which are often satisfactory for companies (Choi et al.
2018).
106 G. Ciaburro
• Prescriptive analysis: This analysis goes beyond the prediction of future results
that predictive analysis performs, automatically providing recommendations on
the best actions to take. This is possible thanks to the synthesis of data and the joint
use of mathematical sciences, business rules and Machine Learning technologies.
The similarity with predictive analysis is very high, both provide the user with
information about what the future is likely to be. The difference with the previous
type of analysis, however, is that the prescriptive analysis explains how to act to
achieve the desired goal. Prescriptive analytics are complex to administer, and
most companies are not using them yet. However, when implemented correctly,
they can have a big impact on how companies make decisions and, in doing so,
help them deliver the right products at the right time, thus optimizing the customer
experience.
These are algorithms that use various levels of non-linear units. These levels are
used in cascade to perform tasks that can be classified as problems of transforming
characteristics extracted from the data. Each level uses the output of the previous
level as an input (Sejnowski 2018). These algorithms fall within the broader class of
data representation learning algorithms within machine learning. They are formed
by multiple levels of representation that can be understood as different levels of
abstraction, capable of forming a hierarchy of concepts. The mechanisms by which
the human brain works are then simulated by these algorithms, in order to teach
machines not only to learn autonomously as in Machine Learning, but to do it in a
deeper way as the human brain would also do, whereby deep we mean on multiple
conceptual levels. Although the demand for enormous computational capabilities
may represent a limit, the scalability of Deep Learning to the increase in available
data and algorithms is what differentiates it from Machine Learning (Ciaburro 2020).
Deep Learning systems, in fact, improve their performance as data increases, while
Machine Learning applications, once a certain level of performance is reached, are
no longer scalable even by adding a greater number of examples and training data
to the neural network. To train a Deep Learning system usually the employees use
to label the data. For example, in the field of visual recognition, the bird meta tag
can be inserted into images containing a bird and, without explaining to the system
how to recognize it, the system itself, through multiple hierarchical levels, will guess
what characterizes a bird. (the wings, the beak, the feathers), and therefore to learn
to recognize it (Esteva et al. 2019).
In order for the final output to be satisfactory, huge amounts of data are needed,
but thinking about Big Data right away is a mistake, as at least in the initial part of
training, the system needs labeled and structured data, therefore the heterogeneity of
Big Data cannot be considered a solution. Unstructured data can be analyzed by a deep
learning model once it is formed and reached an acceptable level of accuracy, but not
for the initial training phase (Voulodimos et al. 2018). Deep Learning today is already
Time Series Data Analysis Using Deep Learning … 107
applied in various fields: self-driving cars, but also drones and robots used for the
delivery of packages or for emergency management are excellent examples. Finally,
speech recognition and synthesis for chatbots and service robots, facial recognition
for surveillance and predictive maintenance are among the most relevant emerging
applications (Ciaburro and Iannace 2020).
4.1 Introduction
Knowledge of the noise levels that characterize a given area represents a funda-
mental step for the description of the acoustic state and for the definition of remedi-
ation interventions, but also an important basis for territorial and urban planning and
programming. The possible objectives of a monitoring activity for the estimation of
the population index exposed to noise are such as to require a separate assessment
of the noise coming from the different sources (Ko et al. 2011). In fact, both in order
to define the mitigation interventions and to establish impacts on the population, the
distinction of the share of environmental noise to be attributed to the different sources
is indispensable.
It is evident that the main source of noise to which large sections of the population
are systematically exposed is the urban road network (Smith et al. 2017). The acoustic
state of the city can be defined through a representation of data relating to an existing
noise situation as a function of an acoustic descriptor, which indicates the exceeding
of relevant limit values, the number of people exposed, or the number of houses
exposed to certain values of the chosen acoustic descriptor (Can et al. 2008). But it is
equally useful to have models for forecasting the noise produced by vehicular traffic
starting from the data collected during the monitoring period (Cirianni and Leonardi
2012; Kumar et al. 2011).
The noise generated by road traffic varies according to the volume of traffic, the
type of vehicles that constitute it, the speed, the driving behavior. The resulting sound
field depends on numerous external conditions that affect its propagation.
By analyzing the traces of the noise emitted by vehicular traffic, it is possible to
identify trends that distinguish it as a data with time series characteristics (Murthy
et al. 2007).
Road traffic noise is caused by heavy vehicles, light vehicles, and motorcycles.
The noise produced by vehicles originates from various components: engine, air
resistance, tire rolling, accessory engines, as well as the actuation of the brakes
(Kulauzović et al. 2020). The engine is home to compressions, pops and decom-
pressions that produce a quantity of noise that is directly related to the number of
revolutions.
The rolling of the tires on the asphalt is a source of noise as a result of the trapping
and subsequent release of air from the cavities, as well as vibrations on the bodywork.
108 G. Ciaburro
The noise deriving from the air resistance is generally detected only at speeds above
200 km/h, therefore in a field that is extraneous to the normal flow of urban road
traffic. Finally, the action of the brakes that manifests itself through the friction
between the lining and the disc. If the pressure between the two elements is high, it
can cause the tire to drag on the asphalt; the combined action of the two phenomena
causes high noise levels. The noise produced by the motor vehicle engine is, at low
speeds, higher than that produced by the rolling of the tires on the asphalt. As the
speed increases, the rolling noise becomes more intense until it prevails over that
produced by the engine.
On the other hand, as far as heavy vehicles are concerned, the engine component
always predominates over the pneumatic component. It should also be considered
that particular aspect of urban noise consisting of the sound of sirens that inform the
activity of various public utility services (Ögren et al. 2018).
The models for predicting noise from vehicular traffic are elaborated starting from
a database of sound levels detected in standard conditions when single vehicles of
predefined types and at pre-established speeds pass (Steele 2001). In this work we will
apply a different approach, we will make predictions of the noise generated by road
traffic on the basis of the data detected by appropriate sensors using algorithms based
on recurrent neural networks. The data was collected in a measurement campaign
carried out near a road artery. The area in which the measurement was carried out is
characterized by intense human activity due to productive settlements and widespread
urbanization, with a population density of about 1500 inhabitants/sq. Km. These
activities contribute significantly to traffic volumes, particularly during peak hours.
The measurement methods, the position of the microphone and the acquisition
parameters are linked to the nature and purpose of the survey (ISO 2017). The
microphone was positioned in a place representative of the sound energy to which
the disturbed are exposed, at a distance of at least one meter from the facades of
the buildings exposed to the highest noise levels and at an altitude of 4 m from the
ground.
The monitoring was carried out over sufficiently long detection times, in order
to allow an overall view of the sound energy present in the investigated place. The
monitoring was performed with special instrumentation, with the microphone placed
in correspondence with the position occupied by the sensitive receptors.
The noise measurement was performed for a measurement time of not less than
one week. In this period, the A-weighted equivalent continuous level was measured
for each hour over the entire 24-h period using a mobile monitoring station. A mobile
station consists of devices equipped with a sufficient amount of memory to store the
acoustic monitoring data for several days, periodically downloading them manually
or with the aid of GSM transmitters.
Time Series Data Analysis Using Deep Learning … 109
These stations provide for the use of simplified outdoor microphones and an
autonomous power supply system that allows operation without being connected to
the mains. The instruments are normally placed inside mobile vehicles, specially set
up, or in suitable suitcases, to be used in short and medium-term monitoring. The
following instrumentation was used:
• Only 01 dB integrating sound level meter model of “Class 1”
• Larson Davis CAL 200 Calibrator
• Tripod
• Windproof headphones
The monitoring period lasted three weeks, which was sufficient to analyze the
daily and weekly trends.
Recurrent neural networks are one of the possible models for dealing with time
series. Its main advantage lies in the possibility of storing a representation of the
recent history of the series, which allows, unlike what happens with artificial neural
networks, that the output of a given input vector can vary according to the current
internal configuration of the network (Li et al. 2018). A recurrent neural network
(RNN) is a type of ANN that provides feedback connections, or loops or loops
within the network structure. This type of backward connection allows you to store
the status from one iteration to the next, using your output as an input for the next
step (Arras et al. 2017). The network receives inputs, produces an output, and returns
the latter back to itself (Fig. 4).
Fig. 4 Architecture of a Recurrent Neural Network with indication of the feedback connections
110 G. Ciaburro
To better understand the structure of an RNN, we can unfold the network which
gives us a feedforward version of the structure (Fig. 5). In this structure the weights
and biases of block G are shared, each output Xt depends on the processing by the
network of all inputs ai with i ∈ [0; t]. The number of blocks of the unfolded version
essentially depends on the length of the sequence to be analyzed (Lin et al. 2021).
Among the many RNN architectures developed over the years, Long Short-Term
Memory (LSTM) (Hochreiter and Schmidhuber 1997) has met with considerable
success and has proposed itself as a structure capable of reducing learning problems in
RNNs. The LSTMs have a structure that allows them to grasp long-term dependencies
in the data, thanks to the memory cell, capable of storing the state outside the normal
flow of the recurring network (Fischer and Krauss 2018). The state is calculated as the
sum between the previous state and the current input processing modulated through
masks so that only part of them is remembered. The network has, in fact, the ability
to remove, add or read the information contained therein at any time, respectively
through the gates.
In Fig. 6 are shown the equivalent sound pressure level detected near a road artery
for a period of three weeks.
Analyzing the graph of Fig. 6 we can see the typical recurring characteristics
of the time series, we can in fact clearly identify the periods of daily and weekly
observations. During a day, in fact, we can see that the noise levels rise in value
during peak periods when the population moves for work reasons, while they drop
significantly during the night. But this is not the only trend we can see, moving the
observation period from day to week we can see that noise levels drop significantly
over the weekend. In fact, both Saturdays and Sundays there are values that are
Time Series Data Analysis Using Deep Learning … 111
Fig. 6 Trend of vehicular traffic noise for an observation period of three weeks (number of
observations versus Leq (dBA))
significantly lower than working days. This demonstrates that the percentage of the
population that goes to work at the weekend is much lower.
From this first visual analysis we can deduce that road traffic noise, being depen-
dent on traffic flows, essentially depends on the volume of the population that uses
this means of transport to go to work. If we had extended the observation period to
one year, we could have seen that this descriptor is characterized by seasonal compo-
nents. In fact, in the summer period as a large part of the population moves to the
holiday resorts, then the cities record a considerable reduction in noise levels, which
instead increase precisely in the holiday resorts.
In order to be able to adequately manage a traffic plan in order to keep the popu-
lation’s exposure to noise within the limits imposed by the legislation, it is crucial to
have forecast models that can provide us with estimates of road traffic noise levels.
In this section I present a methodology for the elaboration of such forecast scenarios
based on algorithms that use recurrent neural networks. Specifically, a Long Short-
Term Memory (LSTM) is used, implemented with the use of the Python Keras library
(Hochreiter and Schmidhuber 1997).
To begin with, the data in our possession, which refer to a monitoring of the
noise levels introduced by vehicular traffic, have been appropriately separated into
two subsets that will be used for training and subsequently for testing the network.
Given a dataset of approximately 504 data records, we divide it into a 70:30 ratio,
meaning that 70% of the data (353) will be used in the network training phase,
while the remaining 30% (151) will be used for model testing. Usually, the data
subdivision is performed randomly in order to guarantee an effective network training
procedure. With data with time series characteristics this procedure is not feasible
as the sequence of values is important. A simple method we can use is to divide the
112 G. Ciaburro
dataset sequentially, first the first 353 values are used for the training phase, then the
next 151 values are used for the test.
Our goal is to use the noise levels detected in the vicinity of a road infrastructure,
to predict these values in the near future. We therefore need inputs and outputs to
train and test our network. It is clear that the input is represented by the data present
in the data set, hence the equivalent sound pressure levels. But this descriptor also
represents what we want to obtain as an output: Being sequential values, our output
will be the equivalent sound pressure level at time t + 1 compared to the value stored
at time t.
Previously, we said that a recurrent network has memory, and this is maintained by
fixing the so-called time step, this is the main reason why this design choice was made.
The time step is about the number of steps back in time that the backpropagation
uses to calculate the gradients for weight updates during the training phase. We then
define a function that provides a set of data and a time step, which then returns the
input and output data (Géron 2019).
At each level of a LSTM an input must be sent in the form of a 3D tensor with
these features:
• Batch: Number of noise levels detected
• TimeSteps: Equivalent to the amount of time steps performed on the recurring
neural network
• Features: they are simply the number of dimensions that we feed in each time step
Once trained, we use the model to make a prediction on vehicle noise levels Fig. 7.
From the analysis of the results, we can see that the model returns predictions that are
very close to the values detected in the monitoring phase. This result is guaranteed
both in the training phase and in the test phase. From a more precise control we can
see that the model underestimates the peak values of the noise, on the contrary we can
Fig. 7 Comparison of measured and predicted values from the LSTM-based model
Time Series Data Analysis Using Deep Learning … 113
see that the lowest noise values detected are instead overestimated. Therefore, the
difficulties encountered by the model are in predicting the limit values (Sherstinsky
2020).
To evaluate the performance of the model, we calculated the Root Mean Square
Error, which represents an accuracy measure based on the scale of the analyzed data
and is therefore not very suitable for comparing time series with data with different
scales. This measure metric is defined by the Eq. (10).
1 N
2
RMSE =
yi − y i (10)
N i=1
Here,
• yi is the actual value.
• ŷi is the forecasted value.
• N is the number of the sample.
Table 1 shows the RMSE values returned by the forecasts made with the model
based on the LSTMs.
The obtained RMSE values confirm that the forecasts obtained with the LSTM-
based model are very close to the data collected through monitoring. This is what
we had already deduced from the visual analysis of the data shown in Fig. 7.
5 Conclusions
In this chapter we have faced the problem of elaborating forecast models for data with
time series characteristics. A time series is a finite set of observations, ordered in time,
of a certain phenomenon that evolves randomly. It is a sequence of empirical data
ordered as a function of time, this allows us to represent them in a time diagram and
be able to understand any trends. Time series analysis models exploit forecasting
techniques based solely on the history of the variable of interest. They work by
capturing patterns in historical data and extrapolating them into the future. The
Times Series present recurring structures that can be captured through a careful and
precise analysis of its performance.
However, these structures are not easily identifiable, at least they are not simply
through a visual approach. The methods based on Machine Learning are able to
identify these recurring structures in a completely automatic way: These techniques
114 G. Ciaburro
References
A. Abanda, U. Mori, J.A. Lozano, A review on distance based time series classification. Data Min.
Knowl. Disc. 33(2), 378–412 (2019)
O.I. Abiodun, A. Jantan, A.E. Omolara, K.V. Dada, N.A. Mohamed, H. Arshad, State-of-the-art in
artificial neural network applications: a survey. Heliyon 4(11), e00938. (2018)
C. Anitescu, E. Atroshchenko, N. Alajlan, T. Rabczuk, Artificial neural network methods for the
solution of second order boundary value problems. Comput. Mater. Continua 59(1), 345–359
(2019)
L. Arras, G. Montavon, K.R. Müller, W. Samek, Explaining recurrent neural network predictions
in sentiment analysis (2017). arXiv:1706.07206
P.J. Brockwell, R.A. Brockwell, R.A. Davis, R.A. Davis, Introduction to Time Series and forecasting
(Springer, 2016)
M.G. Baydogan, G. Runger, Time series representation and similarity based on local autopatterns.
Data Min. Knowl. Disc. 30(2), 476–509 (2016)
C. Chatfield, H. Xing, The Analysis of Time Series: An Introduction with R (CRC Press, 2019)
Time Series Data Analysis Using Deep Learning … 115
M. Christ, N. Braun, J. Neuffer, A.W. Kempa-Liehr, Time series feature extraction on basis of
scalable hypothesis tests (tsfresh–a python package). Neurocomputing 307, 72–77 (2018)
G. Ciaburro, G. Iannace, J. Passaro, A. Bifulco, D. Marano, M. Guida, ... F. Branda, Artificial neural
network-based models for predicting the sound absorption coefficient of electrospun poly (vinyl
pyrrolidone)/silica composite. Appl. Acoustics 169, 107472 (2020)
M. Costa, A.L. Goldberger, C.K. Peng, Multiscale entropy analysis of complex physiologic time
series. Phys. Rev. Lett. 89(6), 068102 (2002)
G. Ciaburro, G. Iannace, M. Ali, A. Alabdulkarem, A. Nuhait, An artificial neural network approach
to modelling absorbent asphalts acoustic properties. J. King Saud Univ.-Eng. Sci. (2020)
G. Ciaburro, G. Iannace, Acoustic characterization of rooms using reverberation time estimation
based on supervised learning algorithm. Appl. Sci. 11(4), 1661 (2021)
T.M. Choi, S.W. Wallace, Y. Wang, Big data analytics in operations management. Prod. Oper.
Manag. 27(10), 1868–1883 (2018)
G. Ciaburro, Sound event detection in underground parking garage using convolutional neural
network. Big Data Cogn. Comput. 4(3), 20 (2020)
G. Ciaburro, G. Iannace, Improving smart cities safety using sound events detection based on deep
neural network algorithms, in Informatics, vol. 7, No. 3 (Multidisciplinary Digital Publishing
Institute, Sep. 2020) ,p. 23
A. Can, L. Leclercq, J. Lelong, J. Defrance, Capturing urban traffic noise dynamics through relevant
descriptors. Appl. Acoust. 69(12), 1270–1280 (2008)
F. Cirianni, G. Leonardi, Environmental modeling for traffic noise in urban area. Am. J. Environ.
Sci. 8(4), 345 (2012)
J. Durbin, S.J. Koopman, Time Series Analysis by State Space Methods (Oxford University Press,
2012)
Esteva, A. Robicquet, B. Ramsundar, V. Kuleshov, M. DePristo, K. Chou, ... J. Dean, A guide to
deep learning in healthcare. Nat. Med. 25(1), 24–29 (2019)
W.A. Fuller, Introduction to Statistical Time Series, vol. 428. (Wiley, 2009)
B.D. Fulcher, Feature-based Time-series Analysis (2017). arXiv:1709.08055
D. Folgado, M. Barandas, R. Matias, R. Martins, M. Carvalho, H. Gamboa, Time alignment
measurement for time series. Pattern Recogn. 81, 268–279 (2018)
T. Fischer, C. Krauss, Deep learning with long short-term memory networks for financial market
predictions. Eur. J. Oper. Res. 270(2), 654–669 (2018)
I. Goodfellow, Y. Bengio, A. Courville, Mach. Learn. Basics. Deep Learn. 1, 98–164 (2016)
C.W.J. Granger, P. Newbold, Forecasting Economic Time Series (Academic Press, 2014)
M. Gevrey, I. Dimopoulos, S. Lek, Review and comparison of methods to study the contribution of
variables in artificial neural network models. Ecol. Model. 160(3), 249–264 (2003)
A. Géron, Hands-on machine learning with Scikit-Learn, Keras, and TensorFlow: Concepts, tools,
and techniques to build intelligent systems. O’Reilly Media (2019)
J.D. Hamilton, Time Series Analysis (Princeton University Press, 2020)
Hannan, E. J. (2009). Multiple time series (Vol. 38). John Wiley & Sons.
M.H. Hassoun, Fundamentals of Artificial Neural Networks (MIT Press, 1995)
S. Hochreiter, J. Schmidhuber, Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
ISO 1996–2: 2017. Description, measurement and assessment of environmental noise. Part 2: deter-
mination of sound pressure levels. Switzerland: International Organization for Standardization
(2017)
M.I. Jordan, T.M. Mitchell, Machine learning: trends, perspectives, and prospects. Science
349(6245), 255–260 (2015)
J.H. Ko, S.I. Chang, M. Kim, J.B. Holt, J.C. Seong, Transportation noise and exposed population
of an urban area in the Republic of Korea. Environ. Int. 37(2), 328–334 (2011)
K. Kumar, V.K. Katiyar, M. Parida, K. Rawat, Mathematical modeling of road traffic noise
prediction. Int. J. Appl. Math Mech, 7(4), 21–28 (2011)
116 G. Ciaburro
B. Kulauzović, T. Pejanović Nosaka, J. Jamnik, Relationship between weight of the heavy trucks
and traffic noise pollution in the viewpoint of feasibility of fines for exceeded noise–a case study,
in Proceedings of 8th Transport Research Arena TRA (2020)
H. Lütkepohl, Introduction to Multiple Time Series Analysis (Springer Science & Business Media,
2013)
H. Lütkepohl, New Introduction to Multiple time Series Analysis (Springer Science & Business
Media, 2005)
J.C.W. Lin, Y. Shao, Y. Djenouri, U. Yun, ASRNN: a recurrent neural network with an attention
model for sequence labeling. Knowl.-Based Syst. 212, 106548 (2021)
S. Li, W. Li, C. Cook, C. Zhu, Y. Gao, Independently recurrent neural network (indrnn): Building a
longer and deeper rnn, in Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, pp. 5457–5466 (2018)
M. Mohri, A. Rostamizadeh, A., A. Talwalkar, Foundations of Machine Learning (MIT Press, 2018)
V.K. Murthy, A.K. Majumder, S.N. Khanal, D.P. Subedi, Assessment of traffic noise pollution in
Banepa, a semi urban town of Nepal. Kathmandu Univ. J. Sci. Eng. Technol. 3(2), 12–20 (2007)
S. Moritz, T. Bartz-Beielstein, imputeTS: time series missing value imputation in R. R J. 9(1), 207
(2017)
S. Ng, P. Perron, A note on the selection of time series models. Oxford Bull. Econ. stat. 67(1),
115–134 (2005)
M. Ögren, P. Molnár, L. Barregard, Road traffic noise abatement scenarios in Gothenburg 2015–
2035. Environ. Res. 164, 516–521 (2018)
H. Ren, B. Xu, Y. Wang, C. Yi, C. Huang, X. Kou, ... Q. Zhang, Time-series anomaly detection
service at Microsoft, in Proceedings of the 25th ACM SIGKDD International Conference on
Knowledge Discovery & Data Mining, pp. 3009–3017, July 2019
V. Rajaraman, Big data analytics. Resonance 21(8), 695–716 (2016)
B. Ristevski, M. Chen, Big data analytics in medicine and healthcare. J. integr. Bioinf. 15(3) (2018)
S.E. Said, D.A. Dickey, Testing for unit roots in autoregressive-moving average models of unknown
order. Biometrika 71(3), 599–607 (1984)
T.J. Sejnowski, The Deep Learning Revolution (Mit Press, 2018)
R.B. Smith, D. Fecht, J. Gulliver, S.D. Beevers, D. Dajnak, M. Blangiardo, ... M.B. Toledano,
Impact of London’s road traffic air and noise pollution on birth weight: retrospective population
based cohort study. Bmj, 359 (2017)
C. Steele, A critical review of some traffic noise prediction models. Appl. Acoust. 62(3), 271–287
(2001)
A. Sherstinsky, Fundamentals of recurrent neural network (RNN) and long short-term memory
(LSTM) network. Physica D: Nonlinear Phenomena, 404, 132306 (2020)
R.S. Tsay, Analysis of Financial Time Series, vol. 543 (Wiley, 2005)
The R Datasets Package (2021). https://stat.ethz.ch/R-manual/R-devel/library/datasets/html/00I
ndex.html Accessed 15 April 2021
C.W. Tsai, C.F. Lai, H.C. Chao, A.V. Vasilakos, Big data analytics: a survey. J. Big Data 2(1), 1–32
(2015)
A. Voulodimos, N. Doulamis, A. Doulamis, E. Protopapadakis, Deep learning for computer vision:
a brief review. Comput. Intell. Neurosci. (2018)
A.S. Weigend, Time Series Prediction: Forecasting the Future and Understanding the Past
(Routledge, 2018)
W.W. Wei, Time series analysis, in The Oxford Handbook of Quantitative Methods in Psychology,
vol. 2 (2006)
A Low-Cost IMU-Based Wearable
System for Precise Identification of Walk
Activity Using Deep Convolutional
Neural Network
1 Introduction
Rapid advancement in sensing technologies has ushered in a new, smart era. Coupled
with the strive for a better life, these smart solutions have been adapted as products of
daily use in almost every sector of our life, starting from education to healthcare. The
recognition of human activities from a wider aspect to an intrinsic level, lies at the
centre of such smart products. Simultaneously, as a research problem, the recognition
of human Activities of Daily Living (ADL) has been addressed for many years now.
This finds applications in elderly monitoring, indoor tracking, fall detection, assisted
living etc. Depending on the target activity set, the sensing can be performed using
ambient sensors or wearable devices individually, or a combination of both. While
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 117
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_5
118 A. Chakraborty and N. Mukherjee
the environment-deployed ambient sensing systems have been extensively used for
determining human activity related information, wearable systems perform on a per-
person basis, at a more intrinsic level.
One popular and frequently used product is a smart fitness tracking device, gener-
ally used for monitoring human walk activity. Recent studies show that such type of
wearable devices are projected to ship over ten times more in 2023, starting with less
than 30 million units in 2014 (Statista 2014). This reflects how people have become
more conscious about their bodily fitness related statistics. These fitness trackers
are generally worn on the wrist or as anklets, and are capable of alerting the users
about their vital physical statistics such as duration of walk, heart rate, step count
etc. With increasing dependence on such systems in recent times, researchers have
tried to determine their ineffectiveness in estimating fitness characteristics (Chen
et al. 2016). The development of miniature, smart health sensing wearable devices
is also a very popular field of research. Most of the human motion data captured
by such wearable devices are generated by the Inertial Measurement Unit (IMU)
based accelerometer and gyroscope sensors. Such sensors generate time-variant data
corresponding to the subjects’ linear acceleration and angular velocity in a 3-axis
co-ordinate system. Coupled with a micro-controller chip, these sensors can be used
to determine different sets of human ADLs with high precision. Their presence is also
noted in smartphones nowadays, and are utilized by multiple application programs
for physical fitness monitoring, similar to smart fitness devices.
Study of a recent, comprehensive survey reveals that though there are many pos-
sibilities such as chest, wrists, ankles etc., the selection of ideal deployment position
for wearable sensing systems is highly debated (Attal et al. 2015). Feasibility and
comfort are also factors to be considered, and these may be affected with prolonged
use of the wearable systems, as in the case of systems worn on the chest, waist etc.
A simpler and more comfortable solution is wrist-bands or anklets. As the activity
of walking generally involves a sequence of swing-step-swing action with legs, and
alternate swinging of hands, smart wrist-bands used for walk detection may be prone
to error in case of only hand swing without actual walking. Similarly, smartphones
when held in hand, often fail to differentiate between normal hand movement, peri-
odic hand swing and an actual walk activity with hand swing. As such, we may
consider human leg movement as a direct and better indicator of walking activity,
as also observed by Cleland et al. (2013) in their work. However, this understanding
may be proven wrong if a person decides to sit and imitate the leg movement of
actual walk, as shown in Fig. 1.
The current work takes up this challenge of developing a wearable, leg-worn
system in the form of a smart anklet that can be used to learn and differentiate
between a normal walk (swing-step-swing with forward movement using both legs)
and a mimicked walk activity which consists of repeated leg swing (swing-step-swing
of legs in sitting posture). The system utilizes state of the art learning algorithms
for enhancing the performance of activity recognition system, while reducing mis-
classification. This chapter is focused on the development of a smart wearable IMU-
based system that can efficiently distinguish between normal walk and mimicked
walk activities to eliminate the false detection of walk activity. It also contains detailed
A Low-Cost IMU-Based Wearable System for Precise Identification … 119
discussions on the concepts and general techniques that will be fruitful for students
and researchers working in the domain of Human Activity Recognition Systems
(HARS).
The rest of the chapter is organized as follows: Sect. 2 gives a study of the recent
research with respect to the wearable deployments and learning techniques used by
state-of-the-art systems; Sect. 3 illustrates the variation in characteristics of data for
the two activities, system setup, data acquisition, standardization and feature extrac-
tion processes, and preparation of final data corpus; Sect. 4 discusses the different
learning techniques used; the evaluation metrics and experimental results are illus-
trated and analyzed in Sect. 5; lastly the concluding remarks and future scope are
given in Sect. 6.
2 Related Works
The rapid progress in the field of Human Activity Recognition is a result of different
challenges that researchers have overcome for many years. This has been aided by the
rapid miniaturization and increasing availability of sensing devices. Another factor
that plays a very important role is the adaption of learning algorithms that can be
used to evaluate the performance of experimental setups. The systems that aim to
recognize different human activities of daily living can be broadly aggregated as
using wearable, or environment embedded devices. From a general understanding
of their working principles, the wearable systems are able to identify activities at a
more intrinsic level (drinking coffee etc.), while environment-embedded systems can
be used for identifying person-level activities in a multi-person scenario (standing,
sitting etc.). Also, the wearable systems have become easier to access and use, as even
all smartphones these days come with a wide array of sensors and can act as wearable
sensing devices too (Lane et al. 2010). The variation in types of activities detected
is also huge in the case of the former, as revealed by a study of the state-of-the-art
works.
The most common approach of physical activity recognition using wearable sys-
tems involves the use of Inertial Measurement Unit (IMU). The IMU sensors are
120 A. Chakraborty and N. Mukherjee
can smartly differentiate between the leg-swing and actual walk activities, is yet to
be met. Motivated by these observations, the proposed work contributes as follow:
• Our current work is, to the best of our knowledge, the only work where a IMU-
based smart anklet is designed with the challenge of recognizing a real walk activity
from leg swing (swing-step-swing of legs in sitting posture) as an alternative to
wrist worn activity monitoring systems,
• A systematic, comparative study of the performance of classification algorithms
such as deep convolutional neural networks and machine learning algorithms using
real and mimicked walk data acquired from the proposed system.
3 System Details
This section illustrates the implementation and working of the overall system along
with the computation framework used for proper identification of a real walk activity.
Each sub-section in the following discussion provides detailed descriptions for all
the steps followed in implementation of the proposed methodology namely, data
acquisition, data standardization, feature extraction, data processing and labelling,
followed by experimentation (training, testing) and performance evaluation. Figure 2
gives an overview of the implemented system model comprising the mentioned steps.
This system framework illustration follows the general structure implemented in
state-of-the-art wearable activity recognition systems (Pires et al. 2020; Avci et al.
2010; Chetty et al. 2015) etc.
The system developed during the course of this work consists of a portable, battery
powered, setup of inter-connected sensing devices controlled by a micro-controller
circuit. The miniature size of all the components used in developing this setup has
resulted in a light-weight, wearable module that can be worn as an anklet for activity
monitoring. All the off-the-shelf, low-cost components used in this wearable module
are mentioned in Table 1 along with their respective function and cost.
122 A. Chakraborty and N. Mukherjee
The smart wearable anklet used in this work captures the variation in activity data
retrieved from the 3-axis IMU module MPU-6050,1 which consists of the respective
vectors in x, y and z axis. The acquired data lies within ranges of +2g to −2g
and +250 deg/s to −250 deg/s respectively, with corresponding default sensitivity
scale factors in effect. These values are then logged in the SD card locally in CSV
format. The experiments have been conducted in a supervised manner in indoors
with 10 volunteers of varied physique. Every volunteer was asked to perform both
normal walking, and leg swing (swing-step-swing with legs while sitting on a chair)
activities, by wearing the smart anklet setup for specific durations upto 10 min. The
ground truth information for all the acquired data was recorded for future reference.
The sampling rate of the data is set 10 Hz, which gives us 10 data points every second.
Every such data point in the raw data is a tuple of the following 7 components: {Ts ,
A x , A y , A z , G x , G y , G z }, where,
Ts denotes the current data and time instance, at which,
A x gives the Linear Acceleration value along the x-axis,
A y gives the Linear Acceleration value along the y-axis,
A z gives the Linear Acceleration value along the z-axis,
G x gives the Angular Velocity value along the x-axis,
G y gives the Angular Velocity value along the y-axis,
G z gives the Angular Velocity value along the z-axis,
Figures 3 and 4 show the variations in a sample corpus of raw 3-axis accelerometer
and gyroscope data, for actual walk and leg swing activities respectively. While there
is a distinct variation in the patterns of accelerometer data for the two activities as
seen in Fig. 3, the gyroscope data for the two performed activities have more visual
similarity, but lie in different range of values, as evident in Fig. 4. This temporal
information, along with corresponding frequency domain features extracted in the
later sections, is utilized by the learning algorithms in the experiments, for precise
walk activity identification.
1 https://robu.in/product/mpu6050gyrosensor2accelerometer.
A Low-Cost IMU-Based Wearable System for Precise Identification … 123
Fig. 3 Illustration of 3-axis Accelerometer data for the two activities: walk (on the left) and leg-
swing (on the right)
In this section, a sequence of steps have been followed to process the acquired data
so that it can be used for the learning process. There has been extensive research
regarding the feature extraction process, and in here some of these most common
and relevant features are utilized, as discussed in the state-of-the-art review in Pires
et al. (2020). The different intermediate forms of the processed data, along with the
techniques used, are discussed below.
124 A. Chakraborty and N. Mukherjee
Fig. 4 Illustration of 3-axis Gyroscope data for the two activities: walk (on the left) and leg-swing
(on the right)
In Fig. 5, a sample distribution of acquired data is shown for the x-axes of accelerom-
eter (during actual walk activity) and gyroscope (for leg swing). From the histograms,
it can be determined that the distribution of data is gaussian in nature. However, these
normal distributions lie in different range of values for the two sensors, depending
on the activity type. As a first step, such variations in the distribution of 3-axes data
from both sensors is normalized.
A Low-Cost IMU-Based Wearable System for Precise Identification … 125
The main motive behind standardisation is to minimise the variability of data and
bring them to the same scale, where all the data points are re-scaled by shifting them
with the mean of the set of values and dividing by their standard deviation. This
widely used method is called z-score (Kreyszig 2009), and is calculated as:
xi − μ
z − scor e = (1)
σ
where, xi denotes the current data element, μ denotes the mean of the set of elements
and σ stands for the standard deviation. In this step, the complete corpus for both
the accelerometer and gyroscope sensors is standardized in this manner, denoted by:
{Ts , Ax , Ay , Az , G x , G y , G z }.
The standardised raw sensor data in a 7-tuple form is now processed, and from
the values of the IMU data along the 3 axes, we initially determine the resultant
acceleration (Rα ) and angular velocity (Rω ) values using the equations:
Rα = Ax 2 + Ay 2 + Az 2 (2)
and
Rω = G x 2 + G y 2 + G z 2 (3)
Now we have for every time instance Ts , the 6 acquired elements and 2 new features,
given as: {Ts , Ax , Ay , Az , G x , G y , G z , Rα , Rω }.
126 A. Chakraborty and N. Mukherjee
The simplest features that can be extracted from any set of continuous data are the
statistical features, namely mean, standard deviation and variance. For every row
of data, we calculate the mean, standard deviation and variance of the sensor data
and their resultant values. After this step, the corpus consists of the following 11
attributes: {Ts , Ax , Ay , Az , G x , G y , G z , Rα , Rω , μ, σ 2 , σ }.
The temporal data acquired by our wearable module captures details of actual walk
and leg swing activities, which consist of periodically repeated actions determined
by the variation in patterns in data points. This periodicity can be better analyzed by
using Fourier transformation (Bracewell and Bracewell 1986) on the signal We have
implemented the widely used Fast Fourier Transform (FFT) algorithm (Cooley and
Tukey 1965) for converting the acquired data from time-domain to frequency domain.
Such transformation breaks down the whole spectrum as a distribution of periodic
components. The generated FFT values are complex in nature, i.e. they contain a
real part and an imaginary part which correspond to the magnitude and phase of the
signal respectively, and both these values are used as features. Thus the FFT co-
efficients for the 3-axis accelerometer and gyroscope data and their resultant values,
generate 16 additional FFT features. For instance, the two FFT co-efficient features
for Accelerometer x-axis data are: {A x F F T , A y F F T }. The squares of absolute FFT
co-efficient values, generate another feature, namely the Power Spectral Density
(PSD) of the signal. It describes the frequency spectrum of the signal along with the
corresponding power components therein. These PSD values are also added to our
corpus as features. For instance, the two PSD features for Accelerometer x-axis data
are: {Ax P S D , Ay P S D }. Lastly, we calculate the corresponding FFT Frequencies for
the PSD values, as another set of features for our corpus. So, in this step, our data set
is enriched by another 32 features, along with every component in the element vector
from previous section. Figures 6 and 7 give the plots of resultant accelerometer and
gyroscope power features in terms of the inherent frequencies for the walk and leg
swing activities respectively, and highlight the difference between the two activities
in frequency domain.
3.3.6 Auto-Correlation
n−k
(yi − μ)(yi+k − μ)
Ak = i=1n (4)
i=1 (yi − μ)
2
The auto-correlation feature values for all the basic components in our data are
computed to get 8 new corresponding features in our corpus from Sect. 3.3.5,
namely: {Ax AutoCor , Ay AutoCor , Az AutoCor , G x AutoCor , G y AutoCor , G z AutoCor ,
Rα AutoCor , Rω AutoCor }.
Thus, our final processed corpus consists of 3 statistical, 32 FFT and 8 auto-
correlation component features along with the 3-axis standard values of acquired
data and their resultant values for two sensors, to give a total of 51 components in
every data point. This feature-rich corpus is now processed as described in the next
section.
The data points, after adding features as discussed in Sect. 3.3, are now processed
further before using the corpus for classification, as described below:
128 A. Chakraborty and N. Mukherjee
• Sliding Window and slicing: initially, we have used sliding windows such that
subsequent element vectors have a 50% overlap between themselves. Given that
labelled slices of data will be used during the learning process, this technique
ensures that time dependent variations are appropriately represented in every
instance of the processed data and in the trained model. Next, the corpus gen-
erated by the sliding window approach, is sliced into short, fixed sized windows
of 2.5 s each (or 25 data points for given sampling rate 10 Hz). This fixed width
windowing technique is widely used for sampling of activity data (Tran and Phan
2016; Anguita et al. 2012; Chetty et al. 2015). Each slice corresponds to one
particular activity, and contains the extracted feature attributes from Sect. 3.3.
• Labelling: as the experiments are performed with supervised data acquisition in
indoor environment, the corresponding labels of acquired data are already known
from ground truth information. In this step every individual slice is labelled appro-
priately as either 0 (for leg swing) or 1 (actual walk activity).
The final processed corpus consists of element vectors with 2551 components in
each (25 data points * 51 elements per data point * 2 (for 50% overlap) + 1 label).
This comprehensive corpus is used for learning and classification in the next section.
4 Learning Techniques
During the experiments, the processed data is classified in two phases. Initially, a
set of supervised machine learning algorithms have been used, to learn from sets of
training data, which then predicted the class of test data. Next, the processed data is
also fed as input to a deep convolutional learning network which attempts to identify
the test data labels by learning features from training set. The individual classifiers
are discussed below.
The Support Vector Machine (SVM) algorithm (Cortes and Vapnik 1995) trains a
model using labelled data to find an optimal plane of separation that can be used to
classify the new test data. In our case of a binary problem, the objective is to identify a
hyperplane which has maximum distance from data points of both the classes during
A Low-Cost IMU-Based Wearable System for Precise Identification … 129
training. In our work, we have used the SVM algorithm with a linear kernel as our
corpus is feature-rich and the data is linearly separable.
The Gaussian Naive Bayes (GNB) classifier algorithm (Hand and Yu 2001) uses
a concept of independent variables with the conditional probability model based
on Bayes theorem. For a given continuous data-set with normal distribution, the
conditional probability that a given event belongs to a certain class, is determined by
the classification algorithm. This probability is calculated for all test instances given
each class, repeated for all the classes. Finally the class with maximum probability
is predicted by the classifier.
4.4 DT Classifier
The Decision Tree classifier is based on a cascading tree structure (Breiman et al.
1984), where the leaf nodes correspond to the class labels, and all internal nodes
represent the tests on attributes. Beginning with the entire data at the root node, the
decision rules are applied at each node, to predict the outcome and generate the next
node. Here the Classification and Regression Tree (CART) algorithm (Breiman et al.
1984) based gini index cost function is utilized.
Apart from the aforementioned machine learning algorithms, we have also tested our
data corpus with deep learning techniques, namely Convolutional Neural Network
(CNN) (Fukushima 1980). Mostly used with 2-dimensional vectors such as images,
the CNN learning mechanism can also be utilized in training a classifier to recognise
human activity from sensor data. CNNs contain convolutional layers, pooling layers
etc. along with fully connected layers in the end, which can extract features from the
input data and map them to the respective classes/labels. The overall architecture of
the CNN used in our work is described below. The same is also shown in Fig. 8.
• First two layers: The first two layers are convolutional, each of which uses the given
kernel values to convolve the input (or previous layer’s output stream). Also, in
each layer the output is mapped by the Rectified Linear Unit (ReLu) activation
function (Glorot et al. 2011), given by r elu(x) = max(0, x) to re-enforce non-
linearity in the feature set after the convolution operations. For the first layer,
length/input height of every feature vector is 2550. A set of 32 feature detectors
(filters) with 3 kernels, learn the basic, local features from the data, and a output
130 A. Chakraborty and N. Mukherjee
neuron matrix of dimension 2548 × 32 is fed to the 2nd layer. This layer uses
64 new filters for learning other features to generate an output neuron of size
2546 × 64.
• Third Layer: this is a max-pooling layer, which eliminates the dependence on the
feature position in the vector using sliding window of height 2 across data, for
down-sampling and retaining only effective features (50% data discarded).
• Fourth Layer: The next layer is dropout (Hinton et al. 2016), which makes our
CNN model ignore certain neurons randomly (at a rate of 40%) during training
by assigning them zero weight. This ensures that there is a proper learning of the
model, less sensitivity to minute variations in data or specific neuron weights, thus
preventing overfitting.
• Fifth Layer: The flatten layer is used next to convert the previous output of dimen-
sion 1273 × 64, to a 1D vector such that it can be directly fed to the succeeding
fully connected network. The output of this layer has a set of 81472 neurons.
• Sixth Layer: The final section consists of two fully connected (dense) layers with an
intermediate normalization layer. The first dense layer takes the flattened feature
vector as input and applies the ReLu activation function on this input, with a 100
neurons as output. The result is normalized by scaling it to a mean of zero and
unit standard deviation in the intermediate layer. As ours is a two class problem,
the final dense layer takes the normalized feature vector and applies the sigmoid
activation function to predict the probability as output (between 0 and 1). These
probabilistic values along with the true labels are then used by the cost function
for model performance evaluation.
In this section, the different details pertaining to the experiments performed in this
work, have been highlighted. This includes the experimental setup, system perfor-
mance evaluation metrics and subsequent experimental results and observations.
The proposed work utilizes supervised learning, and the data instances are appro-
priately labelled based on volunteer and activity information gathered during data
acquisition. The feature enriched data corpus has been classified using the supervised
machine learning and deep learning algorithms described in Sect. 4. For evaluation
purposes, we have taken 10 fold cross validation to ensure that the learning process
involves proper representation of all data variations and to prevent holdout. The 10
random splits have been used for training and testing in a 70-30 ratio while maintain-
ing a proper representation of both classes in every set. All the classifiers have been
trained using each split’s training set and evaluated using the corresponding test sets.
132 A. Chakraborty and N. Mukherjee
For evaluating the capability of the system, we have used some standard metrics, as
discussed below.
• Accuracy—the proper recognition of each event, given by the formula:
TP +TN
Accuracy = (5)
T P + T N + FP + FN
• Precision—the correctly classified events out of all detected events, given by the
formula:
TP
Pr ecision = (6)
T P + FP
• Recall—the ratio of correctly detected events out of all true events, using the
formula:
TP
Recall = (7)
T P + FN
2 ∗ Pr ecision ∗ Recall
F1 − scor e = (8)
Pr ecision + Recall
where, True Positive (TP) means the model has correctly assigned the data to a class;
True Negative (TN) means the model has correctly determined that the data does not
belong to the class; False Positive (FP) means the model has wrongly assigned the
data to a class; False Negative (FN) means the model has incorrectly determined that
the data does not belong to the class.
With the use of discussed metrics, classification performance for all the classifiers
considering all the 10 random splits is shown in Figs. 9 and 10. Similarly, Figs. 11 and
12 illustrate the classifier-wise performance throughout the experiment. It is observed
from the two figures that deep CNN is consistently efficient in classifying all the splits
of actual walk and leg swing activity data accurately, with a maximum accuracy
of 97%, followed by kNN algorithm with a maximum accuracy of 87% and DT
classifier. The SVM classifier performs worse in terms of accuracy of identification,
while GNB is worse in terms of F1 scores i.e. in terms of precision and recall.
Figure 13a shows the distribution of mean accuracy for all the learning techniques
used for evaluation. Similarly, Fig. 13b shows the statistical distribution of F1 score
for all the learning algorithms used. From these two figures, we see that CNN outper-
forms all other learning approaches by a good margin, and has minimum deviation
among all. It is followed by kNN and DT. SVM and GNB classifiers rank lowest
A Low-Cost IMU-Based Wearable System for Precise Identification … 133
in terms of both accuracy and F1 score, where GNB also has the widest range of
deviation for both metrics. Table 2 contains accuracy and F1 scores in terms of mean,
standard deviation and maximum scores for all the classifiers considering all 10 splits.
It is evident that our system can differentiate a walk activity from leg swing in sitting
posture, with an accuracy of 95% using CNN. Similarly, the achieved F1 score of
97% reflects how the false positives and false negatives have been minimized by the
system. The performance of kNN shows mean accuracy and F1 score above 80%. We
also find that the GNB algorithm shows a highly varied performance, with maximum
accuracy of 90% and lowest at 64%. A similar trend is seen in the F1 scores for GNB.
SVM performs with moderate accuracy and F1 score below 75%.
A Low-Cost IMU-Based Wearable System for Precise Identification … 135
Considering all the splits of cross-validation, the three best identification instances
for each classifier are illustrated in Fig. 14a and b. This gives a general idea of the
system performance ad highlights that CNN is the ideal candidate as the proposed
systems classifier of choice.
In the case of CNN, one complete learning process for a particular split of data is
shown in Fig. 15a and b. In the first figure, we find that during training, the loss is
minimized from 1.2 to almost 0, whereas in testing the loss could be minimized
only upto 0.1, after which any training could not reduce the loss any further. Such
situations have been used as a stopping point for the training process. Consequently,
the second figure shows that the accuracy during testing had reached the maximum
value, any training beyond which does not lead to better recognition accuracy.
136 A. Chakraborty and N. Mukherjee
As we have not found any relevant state-of-the-art research that attempts to dis-
tinguish actual walk from leg swing activities, any direct comparison of system
performance is not feasible. However, a generalized discussion of the performance
of different classifiers in relevant related works aimed at physical activity recognition
is undertaken in this part of the discussion in the following points:
• Among the machine learning algorithms that have been used to classify the data,
kNN has outperformed all others and is closely followed by DT. This is consistent
with the relatively good performance of kNN (Jian and Chen 2015) and Decision
Tree (Fan et al. 2013) in detecting different physical activities in comparison to
SVM, NB, neural networks, as observed in Attal et al. (2015), Jatoba et al. (2008),
Liang and Wang (2017), Kose et al. (2012), Pirttikangas et al. (2006) etc. The poor
recognition accuracy of Naive Bayes is also seen in other works (Dernbach et al.
2012).
• Learning algorithms in general are known to perform better in data corpus of
limited size, and they generalize by learning from a fixed set of element vectors.
Consequently, the classification performance for the conventional algorithms such
as SVM, kNN, DT, GNB is lower than that of the CNN which benefits from
repeated error minimization and learning from the extensive element vectors in
the substantially large data corpus.
• Recent state-of-the-art works have used CNN to design models with a substan-
tially higher success rate than conventional learning algorithms (Zeng et al. 2014;
Yang et al. 2015; Panwar et al. 2017; Liang and Wang 2017), which resonate our
observation in the current work.
Thus, the proposed system utilizes a novel corpus of IMU sensor signals using differ-
ent classification techniques in a comparative study to determine system efficiency
in identifying a leg-swing activity from a real walk.
A Low-Cost IMU-Based Wearable System for Precise Identification … 137
References
N. Ahmed, J.I. Rafiq, M.R. Islam, Enhanced human activity recognition based on smartphone sensor
data using hybrid feature selection model. Sensors 20(1), 317 (2020). https://doi.org/10.3390/
s20010317
N.S. Altman, An introduction to kernel and nearest-neighbor nonparametric regression. Amer. Stat.
46(3), 175–185 (1992)
K. Altun, B. Barshan, Human activity recognition using inertial/magnetic sensor units, in Inter-
national Workshop on Human Behavior Understanding (Springer, 2010), pp. 38–51. https://doi.
org/10.1007/978-3-642-14715-9_5
D. Anguita, A. Ghio, L. Oneto, X. Parra, J.L. Reyes-Ortiz, Human activity recognition on smart-
phones using a multiclass hardware-friendly support vector machine, in International Workshop
on Ambient Assisted Living (Springer, 2012), pp. 216–223. https://doi.org/10.1007/978-3-642-
35395-6_30
F. Attal, S. Mohammed, M. Dedabrishvili, F. Chamroukhi, L. Oukhellou, Y. Amirat, Physical human
activity recognition using wearable sensors. Sensors 15(12), 31314–31338 (2015). https://doi.
org/10.3390/s151229858
A. Avci, S. Bosch, M. Marin-Perianu, R. Marin-Perianu, P. Havinga, Activity recognition using
inertial sensing for healthcare, wellbeing and sports applications: a survey, in 23th International
Conference on Architecture of Computing Systems 2010 (VDE, 2010), pp. 1–10
O. Banos, R. Garcia, J.A. Holgado-Terriza, M. Damas, H. Pomares, I. Rojas, A. Saez, C. Villa-
longa, Mhealthdroid: a novel framework for agile development of mobile health applications, in
International Workshop on Ambient Assisted Living (Springer, 2014), pp. 91–98. https://doi.org/
10.1007/978-3-319-13105-4_14
B. Barshan, M.C. Yüksek, Recognizing daily and sports activities in two open source machine
learning environments using body-worn sensor units. Comput. J. 57(11), 1649–1667 (2014).
https://doi.org/10.1093/comjnl/bxt075
138 A. Chakraborty and N. Mukherjee
Y. Jang, S. Shin, J.W. Lee, S. Kim, A preliminary study for portable walking distance measurement
system using ultrasonic sensors, in 2007 29th Annual International Conference of the IEEE
Engineering in Medicine and Biology Society (IEEE, 2007), pp. 5290–5293. https://doi.org/10.
1109/IEMBS.2007.4353535
L.C. Jatoba, U. Grossmann, C. Kunze, J. Ottenbacher, W. Stork, Context-aware mobile health
monitoring: evaluation of different pattern recognition methods for classification of physical
activity, in 2008 30th Annual International Conference of the IEEE Engineering in Medicine and
Biology Society (IEEE, 2008), pp. 5250–5253. https://doi.org/10.1109/IEMBS.2008.4650398
H. Jian, H. Chen, A portable fall detection and alerting system based on k-nn algorithm and remote
medicine. China Commun. 12(4), 23–31 (2015). https://doi.org/10.1109/CC.2015.7114066
M. Kose, O.D. Incel, C. Ersoy, Online human activity recognition on smart phones, in Workshop
on Mobile Sensing: From Smartphones and Wearables to Big Data, vol. 16 (2012), pp. 11–15
E. Kreyszig, Advanced Engineering Mathematics, 10th edn. (Wiley, New York, 2009)
V.S. Kumar, K.G. Acharya, B. Sandeep, T. Jayavignesh, A. Chaturvedi, Wearable sensor-based
human fall detection wireless system, in Wireless Communication Networks and Internet of
Things (Springer, 2019), pp. 217–234. https://doi.org/10.1007/978-981-10-8663-2_23
P. Kumari, L. Mathew, P. Syal, Increasing trend of wearables and multimodal interface for human
activity monitoring: a review. Biosens. Bioelectron. 90, 298–307 (2017). https://doi.org/10.1016/
j.bios.2016.12.001
J.R. Kwapisz, G.M. Weiss, S.A. Moore, Activity recognition using cell phone accelerometers. ACM
SIGKDD Explor. Newsl. 12(2), 74–82 (2011). https://doi.org/10.1145/1964897.1964918
N.D. Lane, M. Mohammod, M. Lin, X. Yang, H. Lu, S. Ali, A. Doryab, E. Berke, T. Choudhury,
A. Campbell, Bewell: a smartphone application to monitor, model and promote wellbeing, in 5th
International ICST Conference on Pervasive Computing Technologies for Healthcare (2011), pp.
23–26
N.D. Lane, E. Miluzzo, H. Lu, D. Peebles, T. Choudhury, A.T. Campbell, A survey of mobile
phone sensing. IEEE Commun. Mag. 48(9), 140–150 (2010). https://doi.org/10.1109/MCOM.
2010.5560598
G. Li, T. Liu, J. Yi, Wearable sensor system for detecting gait parameters of abnormal gaits: a
feasibility study. IEEE Sens. J. 18(10), 4234–4241 (2018). https://doi.org/10.1109/JSEN.2018.
2814994
X. Liang, G. Wang, A convolutional neural network for transportation mode detection based on
smartphone platform, in 2017 IEEE 14th International Conference on Mobile Ad Hoc and Sensor
Systems (MASS), (IEEE, 2017), pp. 338–342. https://doi.org/10.1109/MASS.2017.81
H.F. Nweke, Y.W. Teh, M.A. Al-Garadi, U.R. Alo, Deep learning algorithms for human activity
recognition using mobile and wearable sensor networks: state of the art and research challenges.
Expert Syst. Appl. 105, 233–261 (2018). https://doi.org/10.1016/j.eswa.2018.03.056
K. Ozcan, S. Velipasalar, Wearable camera-and accelerometer-based fall detection on portable
devices. IEEE Embed. Syst. Lett. 8(1), 6–9 (2015). https://doi.org/10.1109/LES.2015.2487241
M. Panwar, S.R. Dyuthi, K.C. Prakash, D. Biswas, A. Acharyya, K. Maharatna, A. Gautam, G.R.
Naik, Cnn based approach for activity recognition using a wrist-worn accelerometer, in 2017
39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society
(EMBC) (IEEE, 2017), pp. 2438–2441. https://doi.org/10.1109/EMBC.2017.8037349
I.M. Pires, G. Marques, N.M. Garcia, F. Flórez-Revuelta, M. Canavarro Teixeira, E. Zdravevski,
S. Spinsante, M. Coimbra, Pattern recognition techniques for the identification of activities of
daily living using a mobile device accelerometer. Electronics 9(3), 509 (2020). https://doi.org/
10.3390/electronics9030509
S. Pirttikangas, K. Fujinami, T. Nakajima, Feature selection and activity recognition from wearable
sensors, in International Symposium on Ubiquitious Computing Systems (Springer, 2006), pp.
516–527. https://doi.org/10.1007/11890348_39
A. Sano, A.J. Phillips, Z.Y. Amy, A.W. McHill, S. Taylor, N. Jaques, C.A. Czeisler, E.B. Klerman,
R.W. Picard, Recognizing academic performance, sleep quality, stress level, and mental health
using personality traits, wearable sensors and mobile phones, in 2015 IEEE 12th International
140 A. Chakraborty and N. Mukherjee
Conference on Wearable and Implantable Body Sensor Networks (BSN) (IEEE, 2015), pp. 1–6.
https://doi.org/10.1109/BSN.2015.7299420
M. Shoaib, S. Bosch, H. Scholten, P.J. Havinga, O.D. Incel, Towards detection of bad habits by
fusing smartphone and smartwatch sensors, in 2015 IEEE International Conference on Pervasive
Computing and Communication Workshops (PerCom Workshops) (IEEE, 2015), pp. 591–596.
https://doi.org/10.1109/PERCOMW.2015.7134104
H. Similä, M. Immonen, M. Ermes, Accelerometry-based assessment and detection of early signs of
balance deficits. Comput. Biol. Med. 85, 25–32 (2017). https://doi.org/10.1016/j.compbiomed.
2017.04.009
Statista, Forecast wearables unit shipments worldwide from 2014 to 2023, https://www.statista.
com/statistics/437871/wearables-worldwide-shipments/. Accessed 25 Jan 2020
D.N. Tran, D.D. Phan, Human activities recognition in android smartphone using support vector
machine, in 2016 7th International Conference on Intelligent Systems, Modelling and Simulation
(ISMS) (IEEE, 2016), pp. 64–68. https://doi.org/10.1109/ISMS.2016.51
P. van de Ven, H. O’Brien, J. Nelson, A. Clifford, Unobtrusive monitoring and identification of
fall accidents. Med. Eng. & Phys. 37(5), 499–504 (2015). https://doi.org/10.1016/j.medengphy.
2015.02.009
D. Yacchirema, J.S. de Puga, C. Palau, M. Esteve, Fall detection system for elderly people using iot
and big data. Procedia Comput. Sci. 130, 603–610 (2018). https://doi.org/10.1016/j.procs.2018.
04.110
J. Yang, M.N. Nguyen, P.P. San, X.L. Li, S. Krishnaswamy, Deep convolutional neural networks
on multichannel time series for human activity recognition, in Twenty-Fourth International Joint
Conference on Artificial Intelligence (AAAI Press, 2015)
M. Zeng, L.T. Nguyen, B. Yu, O.J. Mengshoel, J. Zhu, P. Wu, J. Zhang, Convolutional neural
networks for human activity recognition using mobile sensors, in 6th International Conference
on Mobile Computing, Applications and Services (IEEE, 2014), pp. 197–205. https://doi.org/10.
4108/icst.mobicase.2014.257786
Z. Zhang, S. Poslad, Improved use of foot force sensors and mobile phone gps for mobility activ-
ity recognition. IEEE Sens. J. 14(12), 4340–4347 (2014). https://doi.org/10.1109/JSEN.2014.
2331463
P. Zhang, X. Chen, X. Ma, Y. Wu, H. Jiang, D. Fang, Z. Tang, Y. Ma, Smartmtra: robust indoor
trajectory tracing using smartphones. IEEE Sens. J. 17(12), 3613–3624 (2017). https://doi.org/
10.1109/JSEN.2017.2692263
Facial Recognition Application with
Hyperparameter Optimisation
1 Introduction
Facial recognition is a term that is commonly used in areas within Big Data. Despite
its natural meaning of recognising a human’s face, it is connected to far more complex
processes. What appears self-evident to humans requires considerable training and
testing time for an algorithm. Moreover, it is not only important to recognise a
face, but to link it to a unique identity. A study by Sheehan and Nachman (2014)
found that human faces have evolved over time to become more individual. Hence,
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 141
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_6
142 H. M. Claus et al.
facial recognition has become more important in the past few years because “facial
recognition technology treats the face as an index of identity” (Gates 2011). This
can be observed in the recent development of implementing more closed circuit
television (CCTV) systems to identify criminals faster or to prevent attacks detected
through conspicuous behavioural patterns. Nevertheless, in order to implement such
a system, it has to be trained and tested before its application. The facial recognition
application uses an Artificial Neural Network (ANN) which takes its origin from
the human neural networks by having the “ability to learn by interacting with its
environment or with an information source” (Hassoun 1995). ANNs can be used for
several purposes which are not limited to image inputs. However this chapter will
focus on the training and testing of an ANN for a facial recognition application.
Biometric authentication is a widely known procedure to safely identify a person.
Like other biometric authentication techniques using fingerprints, the iris, voices, and
other unique traits (Bhatia et al. 2013), the face has a great significance. While a per-
son would have to be close enough to a camera, microphone or a similar authenticity
device in order to be identified, facial recognition has the advantage of not need-
ing the physical or near contact with a person. Cameras can be attached to public
buildings to process non-digital situations, or digital images can be evaluated by an
algorithm. Hence, the overall security can be enhanced significantly. This technology
makes it possible to identify several people at the same time, given the algorithm in
use has a high accuracy with a low error ratio. Public places, airports, governmental
buildings, hospitals, and every other place can be observed 24 h per day without
breaks. Thus, governments all around the world strive to enhance their security sys-
tems by implementing good facial recognition algorithms (Mann and Smith 2017).
Despite the fact, that the prevention of criminal behaviour has issues beyond the
authentication process because the legal and ethical aspects are still not defined yet
(Hornung et al. 2010), it is important to create an intelligent application that has
the best accuracy and the lowest error ratio that is possible. With such a great tool
comes a great responsibility to ensure that the identities detected are indeed correct.
For this reason, the ANN needs to be implemented in a way that guarantees the best
outcomes. This can be achieved by finding the best values for the hyperparameters
that control the training process of the ANN. There are several approaches to finding
these values and improving the recognition accuracies. This chapter will experiment
with two of these approaches and find the most efficient one, namely manual and
automated optimisation. The ANN will be optimised by tuning the hyperparameters
and training the ANN with each set of parameters to compare the accuracies.
This optimisation procedure is of great importance because it can improve the
ANN’s recognition accuracy just by the slightest change of values and therefore, this
can influence whether a criminal can be found or not. Thus, effective optimisation
of the hyperparameters and using a diverse set of training input is necessary in order
to safely implement a facial recognition application.
This chapter will document the optimisation of a developed Machine Learning
(ML) solution that will achieve the highest recognition accuracy with the minimal
recognition error by finding an optimal set of hyperparameters. Additionally, the two
Facial Recognition Application with Hyperparameter Optimisation 143
2 Problem Statement
An ANN is to be designed and trained that implements the optimal set of hyperpa-
rameters to recognise the faces of 30 different people from a database. The search for
the set of parameters is conducted using two different methods: 1. trial and error also
known as manual optimisation and 2. automated optimisation, such as Grid Search
and Random Search. The first part of this chapter discusses the manual optimisation
in detail. The second part continues with the utilisation of Grid Search and Ran-
dom Search. These methods will be compared and evaluated. The result is a set of
hyperparameters which achieves the highest recognition accuracy with the lowest
error.
3 Theoretical Framework
In order to be able to follow the upcoming experiments fully, the theoretical frame-
work needs to be established. Here, the most important concepts will be explained
and discussed.
In the following, the basic concepts of ANNs will be explained. An ANN has a clear
structure with a foundation of an input layer of neurons, a number of hidden layers of
neurons and a final layer of output neurons (Wang 2003). Furthermore, there are four
main elements of an ANN model: First, there are the connections between the layers
created by synapses, each link is characterised with a weight that will be multiplied
by the incoming signal. Second, the adder sums up all the products of the inputs
and the synapses’ weights. Third, there are several activation functions that can be
used to limit the outputs. Fourth, the external bias can decrease or increase the input
of the activation function (Csaji and Ten Eikelder 2001). Putting this information
into a context, it can be said that the inputs of each layer are the outputs of the
preceding layer only. Thus, the activation function is applied to the weighted sum
of the layer in order to obtain a new input for the next layer. This kind of network
is called a feed-forward ANN because the inputs are fed into one direction only,
namely forward.
144 H. M. Claus et al.
Additionally, there are two important methods that need to be mentioned here:
K-fold cross-validation (KCV) and Back-propagation (BP). They both have a signif-
icant impact on ANNs and will be discussed in the experiments conducted for this
documentation.
Starting with the KCV, it is an approach for “model selection and error estimation
of classifiers” (Anguita 2012). Furthermore, the dataset will be split into a specific
number of subsets which is determined by k. A number of subsets will be used to train
the ANN model, the rest will be the dataset for the testing phase in order to evaluate
the performance. An example can be seen in a script provided by Dr. Benham Kia,
where k = 4 and 3/4 are defined as the training set, while 1/4 defines the testing
set (Kia 2018). This separation of datasets is recommended in order to achieve a
high recognition accuracy that stays when new datasets are entered after the training
phase. The goal is to “maximize [the] predictive accuracy on the new data points”
(Dietterich 1995), without risking to train the model too good, so that it fits the noise
in the data and studies the different peculiarities of the training sets instead of finding
a general predictive rule (Dietterich 1995). This would result in an overfitted model
with too many parameters and, therefore, a high variance. The opposite case would
mean that the ANN model is too complex but does not have enough parameters to
train with, hence, the bias would be too high (Jabbar and Khan 2014), which is called
an underfitted ANN model. However, using KCV is said to prevent an ANN model
from overfitting (Lin et al. 2008) which is why it was utilised within the experiments
for this chapter.
3.3 Back-Propagation
Continuing with BP, this learning rule can be “employed for training nonlinear net-
works of arbitrary connectivity” (Rumelhart et al. 1995). It plays an important role
regarding real-world applications. The algorithm looks for the minimal values of
the error function using gradient descent (Rojas 1996). The BP “repeatedly adjusts
the weights of the connections in the network so as to minimize a measure of the
difference between the actual output vector of the [network] and the desired output
vector” (Rumelhart et al. 1986). To put it differently, after each training phase, the
BP goes back into the model and adjusts the ANN’s weights and biases in order to
minimise the cost function and have a better prediction in the next iteration.
Under these circumstances, using a ML algorithm seems like a justified approach
when creating a solution for face recognition. Nevertheless, a ML algorithm can only
work well if the parameters that are being used are the best ones. There is a great
number of different parameters that can change the outcome of a training phase for
Facial Recognition Application with Hyperparameter Optimisation 145
each classifier. The library MLPClassifier, which was used within conducting the
experiments, includes 23 parameters in total (scikit learn.org 2020). Each parameter
has at least two different values as an option. With atleast 22 3 options the dimensions
take on a great size. Using just trial and error as a method to find the optimal set of
parameters is time consuming. Hence, this process can be optimised automatically.
For this reason there are several optimisation algorithms that can be used. The two
algorithms that are in focus are Grid Search (GS) and Random Search (RS).
Starting with Grid Search, it is a simple approach that tries every possible combination
of parameters within a defined grid. First, the dimensions have to be set with specific
hyperparameters that want to be observed. Each dimension has a range of possible
values. As already mentioned, there are 23 parameters with different ranges, so the
search was limited to only a few parameters (Bergstra and Y. Bengio 2012).
However, Random Search chooses the values randomly from the configured space.
This makes it possible to use more values for each variable, because the chance of
choosing the same value twice is low, whereas GS can only work with the predefined
values (Bergstra and Y. Bengio 2012).
Moreover, face recognition has a complex training procedure that needs to be
observed attentively, in order to achieve a good accuracy that has a minimal error
ratio. Indeed, face recognition has many advantages, but they can be threatened by
superfluous training of the model and, therefore, risking overfitted or underfitted ANN
models. By tuning the hyperparameters efficiently, the accuracies can be improved
and a safe algorithm can be established.
4 Related Work
5 Methodology
This chapter is the documentation and evaluation of optimising the search of a specific
set of parameters used in a ML solution for a biometric recognition task. Therefore,
an ANN needed to be used. The programming platform Google Colab was used
intensely to conduct the experiments for this task. Moreover, a folder was provided
that included in total 1500 images of 30 different people, 50 images each in different
lighting conditions (Belhumeur and Kriegman 1997). This defined the testing and
training datasets for the ANN. There were two additional files provided, namely:
“process_yale_images” (Schetinin 2021) and “classify_yale” (Schetinin 2021). They
included the project scripts that created the basis of the ANN.
There are a few questions that need to be considered, in order to start with the
optimisation process:
• How much time can be put into the search and testing process?
• What level of complexity should the ANN apply? (Claesen and De Moor 2015)
• Which learning algorithm will be used?
• How many parameters should be defined through the search?
Facial Recognition Application with Hyperparameter Optimisation 147
After experimenting with these different strategies, specific patterns will show
and determine which strategy can be defined as the best one for this task. The fol-
lowing discussions will go deeper into the process of experimenting with manual
and automated optimisation and evaluate the results.
Facial Recognition Application with Hyperparameter Optimisation 149
6 Part 1
Part 1 discusses the manual search of a set of parameters that meets the requirements.
First, Google Drive had to be connected with the “process_yale_images” file, in order
to unzip the 1500 images in the folder. They were then prepared for the ANN and
put into a matrix with n = 1500 rows and h ∗ w columns with n being the number of
images and h = 77 and w = 68 being the height (h) and width(w) of the images in
pixels (Schetinin 2021). This is how the size of the input layer could be determined
with h ∗ w, 77 ∗ 68 = 5, 236. The 50 images of each person differ in quality of
lighting, where the first picture is the brightest, the last picture the darkest (Figs. 4,
5).
Consequently, functions from the ML sklearn library were imported (scikit
learn.org 2020). Using KCV, the dataset was split into k folds, k = 5 as a common
recommendation (Anguita 2012). Hence, 20% of the dataset was kept for the testing
phase. Furthermore, the principal component analysis (PCA) was applied using the
variable no f _ prin_components, which determines the number of principal com-
ponents (PCs). PCA is a technique “to reduce multidimensional data to lower dimen-
sions while retaining most of the information” (Karamizadeh et al. 2013). Hereby,
the eigenface method is used to extract the characteristic features of a face and lower
the dimensions of the images. Thus, eigenvectors are calculated, which are features
that characterise the variation between face images (Turk and Pentland 1991). By
comparing the Euclidian distance between the eigenvectors in the dataset, the small-
est distance must be found in order to identify the correct person. This means, that
the multidimensional images are reduced to a linear combination to improve the effi-
ciency of the algorithm. As a result, the PCA variable determines how much variance
there will be in a set of images (Karamizadeh et al. 2013).
After choosing a specific number of principal components, the number of hidden
neurons, NOHN, had to be chosen. Unlike the KCV, the procedure of finding the
optimal size of hidden neurons is not defined yet. However, there is a common
understanding, that an excessive use of hidden neurons will cause overfitting, as this
will create a more complex model than needed (Liu et al. 2007). To simplify the
process, one could argue that the number of hidden neurons should be between the
number of total input and output nodes and the number of hidden neurons should be
approximately 2/3 of the size of the input layer, plus the size of the output layer. For
this reason, this parameter can only be determined in combination with the KCV in
order to see at which point the performance of the algorithm stops improving. This
indicates that the training has started to “fit the noise in the training data” (Liu et al.
2007), hence, the model is overfitted. This procedure is a trial-and-error method with
the aim to find the best possible values. With this in mind, the number of hidden
neurons should be in the range between 30 − 5, 236, where 5, 236 is the number
of pixels in one image and 30 is the number of different people, hence the size of
the output layer. 2/3 of 5, 236 results in approximately 3, 508, adding 30, shows the
limit of 3, 538 hidden neurons. For simplifying the experiments, 3, 500 NOHN are
chosen.
The last lines of the script will be explained in the following to understand the
training process of the ANN. This is the script for the first experiments when searching
for the best parameters manually.
6.2 MLPClassifier
MLPClassifier is the function of the ML sklearn library (scikit learn.org 2020) that
trains the model iteratively. MLP is short for multi-layer perceptron which describes
the ANN, because it has multiple layers.
The first parameter hidden_layer _si zes passes the value that is consciously
choosen for the number of hidden neurons.
The second parameter solver passes the solver for weight optimisation, thus, in
which way the weights of the ANN should be adjusted. As already introduced, BP
uses gradient descent, and therefore, sgd stands for stochastic gradient descent. Thus,
this ANN uses BP.
The third parameter activation determines which activation function will be used
in the hidden layers. As this ANN should be a complex model, a non-linear activation
function is needed. In order to make the optimisation process more efficient, the
hyperbolic tangent tanh is chosen.
Facial Recognition Application with Hyperparameter Optimisation 151
The fourth parameter batch_si ze determines the size of mini batches for the
stochastic optimisation of the ANN. Mini batches are used to stabilise the training
process and to reduce the variance of the gradient estimate (Wang et al. 2019).
The fifth parameter ver bose decides whether progress messages should be printed
in the standard output. Here, it is set to true.
The sixth parameter early_stopping is set to true in order to stop the training
of the model when the validation score is not improving, for example to signalise a
potential overfitted model. In this case, it will automatically take 10% of the training
data as validation and terminate the training once there is no improvement after a
specific amount of iterations. In this case it is set to tol = 0.000100 for 10 consecutive
epochs.
Then, the function f it (X, y) is called to fit the model to the data matrix and target.
Here, X _train_ pca depicts the data where the principal components were already
applied. y_train describes the target data.
The training of the model starts by running this script. Once the training is com-
pleted, the testing dataset can be applied through the function pr edict (X ). Here, the
testing data matrix with the applied principal components will be used together with
the testing target data.
For these reasons, the goal of the experiments is to find the right combination of
both parameters, no f _ prin_components and nohn, and to create a good ANN for
this specific task.
A variety of experiments with different parameters was conducted to use the trial-
and-error method in order to find a good set of parameters that fit the requirements.
In the following, the experiments will be described in detail.
6.3 Experiments
In this part of the report, the different experiments conducted will be described
and discussed in detail. In total, there were four stages of experiments. Each set of
parameters is iterated 20 times, so that the resulting values could be representative
for the experiments. In general, it can be said, that the two wanted parameters, the
number of iterations (NOI), the testing accuracy (TA), the validation score (VS),
and the loss played the most important roles. Due to the early_stopping param-
eter, each run through of the script had a limited amount of iterations within the
ANN and stopped once the VS did not improve for 10 consecutive epochs. The
maximum amount of iterations within the model is 200, which is reached rarely. In
order to find a good solution, only sets of parameters were considered that achieved
equal to or higher than 90% TAs. The sets of parameters will be shown like this:
(no f _ prin_components, nohn).
It is important to mention that the conditions for the tables are not always com-
pletely the same. The ranges differ for each table. Hence, the overall average can
be biased because even the outer ranges with medium recognition accuracies count.
152 H. M. Claus et al.
This is just representing the overall results of a particular table in order to find a good
set of parameters in a great variety of experiments.
6.3.1 Stage I
In the first stage, there are four tables dedicated to a specific nohn value (NOHN)
respectively. The different NOHNs are 100, 150, 200, and 250, choosing to start with a
low number of hidden neurons. Simultaneously, the value for no f _ prin_components
increased by 10, initially starting at 70 principal components (PCs) and stopping at
450. The range is manually adjusted for a few tables, because after a certain amount
of PCs the iterations stopped automatically and overfitting had started. At the begin-
ning, only the number of iterations and the testing accuracy are documented. In the
following, each individual table will be analysed separately.
The table Nohn100 had 100 NOHNs and the range goes from 100 PCs to 250 PCs.
For each set of parameters, the script is run 20 times consecutively. The maximum
average testing accuracy (MATA) is at ca. 84.5%, the minimum at 79.9%, with an
overall average of 82.6%. Only one single iteration with 110 PCs had a 90% TA,
which is not enough for the wanted result. The set of parameters with the best average
accuracy is (200, 100).
For this reason, the NOHN is increased by 50 in the table Nohn150, using the
range from 70 PCs to 310 PCs. 12 sets of parameters showed 90% accuracy in at least
one iteration. Nevertheless, the MATA is at ca. 86.3%, which is only 1.8% better than
the MATA from the last table. Surprisingly, the overall average is at 84.3%, which is
also 1.8% better than the previous one. Hence, the set (170, 150) already achieved a
better result.
Increasing the NOHN even further to 200, the table Nohn200 was created. Here,
the range of the PCs starts at 70 PCs and ends after 450 PCs. Although the MATA
has reached 89.2% which already has an additional 2.9% accuracy compared to the
previous MATA, the overall average almost stays the same at 84.4%. This might
be caused by the wider range of PCs, even though 34 sets of parameters have 90%
accuracy in at least one iteration. The best performing set is: (200, 200) with 89.2%
TA.
The last table from this stage is called Nohn250 with 250 NOHN and a range of PC
from 100 to 400. In this table, out of 31 sets, 28 sets achieved an accuracy of 90% or
more in at least one iteration. However, the MATA is at 88.3% and the overall average
of all accuracies within the table are at 87.4%. In this case, the MATA is smaller than
the MATA from the previous table, but the overall average is higher. The difference
in the overall average could be the result of different PC ranges. Nevertheless, the
set with the best performance has the values (250, 250). Figure 6 shows a summary
of the best performances from this stage.
Facial Recognition Application with Hyperparameter Optimisation 153
Fig. 6 The best set of parameters from each table from stage I
In conclusion, four sets of parameters could be chosen that achieved the best
performance in their respective tables. The best testing accuracies were in a range
between 84% and 90%.
6.3.2 Stage II
The second stage introduced a radical approach to finding a good NOHN. The first
stage had shown that the values might be too small for the model to achieve a better
performance. In order to keep the trial-and-error method, this stage created new fixed
NOHN with varying PCs. This time, the NOHNs are 500, 1500, and 2500. Hence,
instead of increasing by 50, the NOHN increased by 1000. The tables concentrated on
increasing the PCs by 100 and finding good sets of parameters. Notably, at this stage
the VS (validation score) additional to the TA (testing accuracy) and the NOI (number
of iterations) was documented. In this case, the VS is initiated by the parameter
early_stopping from the script, as already explained before. This means, that 10%
of the 80% training data are used to tune the parameters and to find the optimal
NOHN to determine when to stop the algorithm in order to apply BP. Here, the
cross-validation already compares the performance of the training algorithm and
chooses the best results. Hence, the VS will indicate how good the performance of
a specific set of parameters is after it has been trained with the algorithm and the
training data.
Starting with the table PCA I, the NOHN is set to 500, whereas the PCs started at
100 and ended at 900. Again, each set of parameters is run 20 times in order to get a
statistically viable representation of the trained algorithm outcomes. In this table, the
MATA is at 89.8% with a connected average VS of 86.7% for the set of parameters
(300, 500). This accuracy is already better than all accuracies from the tables before.
Hence, the new approach shows progress. The maximum average validation score
(MAVS) is at 89.8%, but the connected average TA is at 89.2% for the set (200, 500),
which is a slightly lower result. Thereupon, it is important to note, that a high TA
outranks a high VS because the model is designed to have a good accuracy with new
data, hence, a good TA is what is looked for. Nevertheless, the second set would also
be the second best performing set in general, thus, these two sets can be noted as
the best performing sets of this table. Adding to this, it is interesting to see that after
154 H. M. Claus et al.
the MATA the accuracies of the next sets start decreasing with a local minimum at
62.7% with the set (900, 500) and an average VS of 28.4%. This indicates that this
set of parameters makes the model too complex and therefore, it creates unreliable
results that should not be used in the final algorithm.
Continuing with the table PCA II, the NOHN is defined as 1500 with PCs in a
range from 100 to 1200. The MATA is at 94.8% with an associated average VS of
91.3% with the set (300, 1500). Surprisingly, by keeping the same amount of PCs but
adding another 1000 hidden neurons, it keeps showing the best performance. Now,
the MAVS is at 94.5% with an associated average TA of 94.2% through the set (200,
1500), again it is in the same style as in the table before. In order to find a closer
solution, a sub-experiment was conducted with the middle value between 200 and
300 PCs, which resulted in the set of (250, 1500). This experiment did not achieve
a better TA, but it is close enough to the MATA with 94.7% (important to note that
the MATA is rounded to 94.8% in order to decrease the decimal places, but the table
shows 94.75% which means that the difference between the two TAs is only 0.05%
instead of 0.1%). However, the average VS is at 93.2%, which is less than the MAVS
but better than the average VS for the MATA. Hence, this sub-experiment showed an
approximation to an optimal set of parameters with the second best average TA and
the second best VS. At the same time, the set (1200, 1500) portrayed what happens,
when the NOHN is too high, because the majority of iterations stop at 12, which
means that the algorithm stopped learning after 2 iterations, which resulted in an
average VS of 2.6% and an average TA of 3.5%. Thus, the average results have a
deflection and can not be taken into the calculation. This portrays an example of a
potentially overfitted algorithm.
The last table for this stage, PCA III, had 2500 hidden neurons and started with
100 PCs increasing to 1100 PCs. Furthermore, the MATA is at 94.6%, which is lower
than the one from the last table, but this MATA appeared twice with two different
associated average VS. The set (300, 2500) with the MATA had an average VS of
93.2%, the set (400, 2500) had an average VS of 91.3%. Hence, the first set of
parameters had a better overall performance. In fact, it had the same number of PCs
as the two previous MATAs, namely 300. The MAVS of 96% is again achieved by
the set (200, 2500) with an associated average TA of 93.8%. Important to mention
is not only the fact, that the MAVS is again achieved by a set with 200 PCs, but also
the fact that every iteration of said set achieved a VS of 93% or higher, which seems
like a more reliable set of parameters than the set (400, 2500), where only half of
the iterations achieved a TA of 93% or higher. But repeating the behaviour of the
previous experiments shows that the higher the number of PCs, the lower the TAs
and VSs after reaching the MATA.
As can be seen, a pattern started to show. The sets with 300 PCs achieved the
highest average testing accuracy, whereas the sets with 200 PCs achieved the highest
average validation score.
Figure 7 shows final results taken out of all tables of the second stage. The table
PCA III depicts that increasing the NOHN unnecessarily does not improve the TA,
hence, the best performance is achieved with 1500 and 2500 hidden neurons. It is
important to mention that the number of iterations indicate when the cross-validation
Facial Recognition Application with Hyperparameter Optimisation 155
Fig. 7 The best set of parameters from each table from stage II regarding the MATA and the MAVS
with the result of another sub-experiment from PCA II
stopped the training process in order to prevent overfitting and rather apply BP to
improve the algorithm even more. The tables also show that with increasing the
number of PCs, the average number of iterations decrease, which only supports the
fact that the sets with high PCs only achieve the minimum TA and VS and, therefore,
might lead to overfitting. All in all, it can be said that the sets (250, 1500), (300,
1500), and (200, 2500) show the best performances.
In the previous stage it is shown that sets with 300 PCs achieved the best TA and the
sets with 200 PCs achieved the best VS. Hence, the middle (250) proved to be a good
approximation to achieve a good TA in combination with a good VS. For this reason,
the first approach in this stage is to see how the TAs improve when the number of
250 PCs stays the same but the NOHN increases instead. The second approach is
another radical experiment with a higher number of PCs (550) and again increasing
NOHNs. This stage documented the same values as the previous stage II.
The first table, Neuron I, has 250 PCs and starts with 50 hidden neurons with
increases following logarithmic scale increments due to previous findings up to
10,000 hidden neurons. As already stated, exceeding the overall number of 3500
is not the goal, but it is done nevertheless, because of experimental reasons. The
MATA here is achieved twice at 96.8% with an average VS of 96.1% for the set
(250, 5000) and an average VS of 96% for the set (250, 10000). It must be said that
the average NOI is at 51.2 for the first set and 43.2 for the second set, hence, the
learning process is faster than with lower NOHNs, i.e. at least 100 iterations for 50
to 500 hidden neurons. It is also interesting to see, that the experiment with the set
(250, 1500) is now conducted for a second time (first time is in the table PCA II).
Comparing both experiments, the following insights can be obtained: The TA for
156 H. M. Claus et al.
Fig. 8 The best set of parameters from each table from stage III regarding the MATA and the
MAVS with the result of another experiment from Neuron I to compare with the result from PCA
II
the first experiment is at 94.7% with a VS of 93.2% and an average NOI of 68.3.
The TA for the second experiment is at 95.1% with a VS of 92.9% and an average
NOI of 72.6. It is likely that the average values would not be exactly the same, but it
is clear that the difference is between 0.4% for the TA and 0.3% for the VS, which
is low. Thus, it can be said, that this set seems to be the most reliable one within
the acceptable range of NOHNs. The difference between the TAs of the set with the
MATA and the chosen set is at 1.7%. Now, one would have to choose whether to risk
overfitting but achieving a good accuracy or staying in the acceptable range and still
achieving 94–95% accuracy.
Looking at the table Neuron II, the number of PCs stays consistent with 550 and
the NOHN increases from 50 to 10,000, again in logarithmic scale increments. The
MATA is at 96% with an associated average VS of 85.5% through the set (550, 5000).
Surprisingly, the MAVS is only at 87.7% with an average TA of 95.9% through the
set (550, 10000). These sets have similar results, also compared to the results with
250 PCs from the previous table. Moreover, looking at the set (550, 1500), it can
be seen that the average TA of 92.1% is already decreasing, when compared to the
table PCA II, where it is shown that all accuracies after the set (300, 1500) were
decreasing. Hence, it is only natural, that the NOHN would have to be increased
in order to balance the constant decrease. But more hidden neurons would then
exceed the maximum. It can be stated, that the previous experiment in table Neuron
I achieved a better MATA.
Noting that the maximum NOHN is consciously exceeded, the following table
shows the best performances (Fig. 8).
In conclusion, increasing the NOHN radically changes the TA by 1.7%. Moreover,
the good accuracies were only achieved because the NOHN exceeded the maximum
of 3,500 by a great factor which made the model more complex than it has to be,
hence, it risks being overfitted. Furthermore, it could be seen that the results of the
set (250, 1500) only had a difference of 0.4/0.3% compared to the results of the first
experiment with the same set.
Facial Recognition Application with Hyperparameter Optimisation 157
6.3.4 Stage IV
Stage IV of the experiments is the last stage. It is not a necessary experiment given the
high number of hidden neurons it applies, it is rather an excursus on the limitations
of Machine Learning algorithms in general. Adding to this, the loss provided by the
loss function is now noted in order to see which parameters were able to minimise
the losses and still keep a good accuracy.
The table Top values I conducted experiments with two different NOHNs, 5000
and 10000, and with the number of PCs increasing from 100 to 600 by 100. Here,
the MATA is achieved twice at 95.3% with an associated average VS of 95.8% and
an average loss of 0.28 for the set (300, 5000) and with an average VS of 96.2%
and an average loss of 0.34 for the set (300, 10000). The MAVS is at 97.4% with
an associated average TA of 94.9% and an average loss of 0.49 for the set (200,
10000). Unfortunately, the stated TAs do not create a new maximum value, they
rather showed the losses when using complex models. Nevertheless, a new overall
maximum value for the validation score is achieved which is 1.3% higher than the
MAVS from table Neuron I.
Continuing with the table Top values II, the experiments depicted an approxima-
tion for the number of PCs in combination with 5000 hidden neurons. Here, a total
of 30 iterations were undertaken for each set of parameters. As already shown in
previous experiments, the range between 200 and 300 PCs seemed to have the best
performances without regarding the NOHN. In order to find a better set of parameters
with the best performing amount of hidden neurons (5000), several experiments were
conducted. Still, the MATA only achieved 96.3% with an average VS of 96.2% and
an average loss of 0.34 for the set (280, 5000), which is 0.5% less than the results of
the sets (250, 5000) and (250, 10000) from Neuron I. The MAVS is at 96.6% with
an average TA of 96.1% and an average loss 0.33 for the set (230, 5000).
Figure 9 summarises well that the best performances are achieved if the number
of PCs is between 200 and 300. Because of the automatic stop function in the script,
overfitting can not happen, it is only a risk for further experiments that do not include
said function.
Fig. 9 The best sets of parameters from each table from stage IV regarding the MATA and the
MAVS
158 H. M. Claus et al.
6.3.5 Conclusion
All things considered, the experiments not only showed the development of find-
ing good sets of parameters, they also showed the growth of understanding within
the limitations of an ANN. With each experiment certain values could be excluded
because the algorithms became either too complex or were not complex enough.
Fortunately, the function early_stopping prevented the algorithm from overfitting
and used 10% of the training set for cross-validation. After a few experiments a pat-
tern could be seen, which showed that the range of 200–300 PCs achieved the best
recognition accuracies. Now, only the specific amount of hidden neurons has to be
chosen.
6.4 Results
The experiments that were conducted, showed a great variety of results. The task was
to find the best set of parameters with the lowest error ratio and a good recognition
accuracy. The following table shows the best results from all stages. The loss depicts
the error ratio, hence, the lower the loss value, the lower the error ratio (Fig. 10).
Fig. 10 The the best sets of parameters from each table from each stage with their average TA,
VS, and loss
Facial Recognition Application with Hyperparameter Optimisation 159
There are three sets of parameters that need to be compared: 1. (200, 2500), 2.
(250, 5000), and 3. (300, 5000). The first set is still in the range of acceptable NOHNs,
where 3500 is the maximum. If 3500 is the official limit of hidden neurons, then the
first set would be the best one. If the number of hidden neurons is not as important,
then sets with higher NOHNs can be considered. In this case, the second set has
the overall best recognition accuracy of 96.8%. As stated several times, a function
prevents the algorithm from overfitting, hence, the risk is not as high. However, the
error ratio needs to be considered as well. According to the error ratio, the third set
would show the best performance.
Now, choosing the safer path with 1500 NOHN, an approximation is made, which
is shown in the table Approximation. Here, the pattern of 250–300 PCs showing the
best performances is taken and tested again. The NOHN stayed the same at 1500,
while the number of PCs were 250, 275, and 300. Adding to that, the loss is now
documented, as well, in order to be able to compare the results to the sets of parameters
from stage IV. Moreover, the experiments were conducted with two different testing
data sizes. The initial experiments had 20% of the overall data as testing data, which
is again the case for the first experiment in the table Approximation. The second
one has 33% of the overall data as testing data, which means that k = 3 in the KCV
process. This will indicate how well the parameters perform with a smaller training
data size, which fits the use of real-world applications more. Thus, lower accuracies
are expected because there is less data to use for training purposes (Fig. 11).
It is important to mention that the set (300, 1500) with 20% testing data is now
shown twice, one time in the table PCA II with the average TA of 94.8% and average
VS of 91.3%, and the second time now in the table Approximation with the average
TA of 96.4% and the average VS of 91.7%. The average values of these both exper-
iments result in 95.6% for the TA and 91.5% for the VS. Nevertheless, the set with
275 PCs shows the best recognition accuracy and is only 0.2% lower than the overall
MATA from the set (250, 5000). The average losses of 0.31 are too high, given the
fact that this value is the exact average value between 0.34 and 0.28 average losses.
The VS of 92.6% is lower compared to the other sets, but as already stated, the actual
testing accuracy outmatches the validation score, hence, the TA counts more than
the VS.
Looking at Fig. 12, the experiment with 33% testing data shows lower results,
given the fact that there is less data to train with. The testing accuracies still reach
Fig. 11 The results of the first experiment from the table Approximation with test_si ze = 0.2
160 H. M. Claus et al.
Fig. 12 The results of the first experiment from the table Approximation with test_si ze = 0.33
at least 92.7% with only 0.04-0.06 more losses than the average of the previous
experiment. Surprisingly, the TAs are very close with only 0.1% difference. But as
previous experiments have shown, more iterations can change the overall results and
potentially increase or decrease them. Adding to that, the lower accuracies could be
explained when putting the number of PCs and NOHNs into relation with the training
data size. If the training data size decreases from 80% to 66%, then the values for
the parameters would need to be adjusted in the same manner, in order to contain the
relative accuracies. For experimental purposes, the absolute values were compared
to show that the chosen values would still perform well enough with a larger testing
data size.
In conclusion, it can be said that the sets with values 250–300 PCs and 1500
hidden neurons show the best and most reliable performances and are still in the
limit of the acceptable complexity within the model. The exact accuracies can vary
with each new iteration, hence, it is hard to pick only one exact value. But the set
(275, 1500) seems to be a good approximation and leads to an average between 250
and 300 PCs.
6.5 Analysis
After finding a good approximation towards a set with a recognition accuracy over
90% and an error ratio of approximately 0.3, it needs to be shown why the model
behaved the way it did. The following Fig. 13 shows a graph designed in MATLAB
with the testing accuracy as a function of the number of PCs and the NOHN.
This figure shows that the function has the highest accuracy values between 200
and 300 PCs, increasing with the NOHN, as is proven by the tables in the spreadsheet.
Furthermore, Fig. 14 and Fig. 15 show the explicit maxima that can be seen in Fig. 13.
These figures prove that the pattern that is detected during the experiments indi-
cates that the number of PCs between 200 and 300 show the best performances.
Hence, the ANN should at least have 250 principal components. Figure 13 also dis-
plays that all accuracies between the PC values 100 and 800 achieve at least 90%
through 5000 hidden neurons.
Facial Recognition Application with Hyperparameter Optimisation 161
Fig. 13 The behaviour of the ANN with different values for PCs and hidden neurons in relation to
the testing accuracy
Fig. 14 The maxima with the number of PCs between 200 and 300 for 5000 hidden neuron
162 H. M. Claus et al.
Fig. 15 The maxima with the number of PCs between 200 and 300 for 500 hidden neurons
Fig. 16 The behaviour of the model with 250 PCs and increasing NOHNs
Fig. 17 The results of the third experiment from the table Approximation with test_size = 0.33 and
PC put into relation with test_size
of hidden neurons should be between the general amount of input and output nodes,
thus, 3500 in this case.
Moreover, one of the reasons why the accuracies with the bigger testing data size
were lower, besides the fact that there is a smaller amount of data to train with,
is the fact that the absolute values of the parameters were trained and tested with.
For the testing data size of 20%, 1200 images will be trained with. Here, it can be
assumed, that by choosing 33% to be testing data, the values would need to be adjusted
relatively. Now, only 1000 images will be used for training. This might also be the
reason, why the set (250, 1500) from the table Approximation achieved the lowest
value in the first experiment with 20% testing data and the best performance with
33% testing data. However, this hypothesis needs to be proven in another experiment.
Nevertheless, out of curiosity, a short excursus was undertaken, in order to see
how the relative parameters would perform. Two more experiments were added to
the table Approximation where the number of PCs was adjusted, so it was always in
the same relation to the training data as with the testing data size of 20%. The Fig. 17
shows the values for test_si ze = 0.33, hence 33% of the general dataset.
It can be seen that the TAs from Fig. 17 show at least 1% better accuracies than
from Fig. 12. By adjusting the number of PCs, the general accuracies improve, but
the losses become higher because there is less data to train with. Now the chance
of only training with, for instance, the lighter images or with the Caucasian looking
people, hence, leaving the darker images or the south-east Asian and African people
for the testing, are higher, which means that the algorithm could struggle with data
that it has not been trained with. Changing the variable test_si ze to 0.1, thus, having
only 10% as testing data, changes the outcomes again, because now the number of
PCs is adjusted and the sets are trained and tested with new values. In this case, there
is more data to train with, so higher accuracies are expected, which can be seen in
Fig. 18.
Here, the number of PCs has to be increased because the amount of training data
increased. When there are 300 PCs for 1200 training images (25%), then there must
be 251 PCs for 1005 training images (still 25%) and 338 PCs for 1350 training images
(also 25%). This was done with all PCs here to find the relative values. As can be
seen, the TAs were higher than for 33% testing data size, but lower than for 20%.
Hence, this only shows that all experiments from the several stages were conducted
with the optimal value for k.
164 H. M. Claus et al.
Fig. 18 The results of the fourth experiment from the table Approximation with test_size = 0.1
and PC put into relation with test_size
In the final analysis, the best performance can be achieved with the number of
principal components between 200 and 300. After reaching 1500 hidden neurons,
the model seems to slow down with the learning process and the accuracies are all
over 90% with increasing NOHNs.
7 Part 2
Optimising is a key aspect when using an ANN. In order to find a more efficient
approach for using the optimal set of hyperparameters, the process can be automated.
For Grid Search, a specific grid space had to be defined. It was recommended to use
a small number of dimensions, to prevent the grid from becoming too complex. The
chosen dimensions for this experiment were: 1. Number of hidden neurons (NOHN),
2. Activation function, 3. Solver, and 4. Batch size. The experiments conducted in
Part 1 had predefined values for the dimensions, where only the NOHN and PCs
were changed. Optimisation algorithms make it possible to broaden the dimensions,
hence, instead of using the same values as in Part 1, new values were added to
compare the outcomes. The values of the activation function included tanh, which
is the hyperbolic tan function, and r elu, which implements the rectified linear unit
function (scikit learn.org 2020a). The solver parameter has three possible values
of which two specific ones were used: sgd, the stochastic gradient descent, and
adam, a stochastic gradient-based optimiser (scikit learn.org 2020). The batch size
was provided values of 2n starting with 2 and closing with 512. The rest of the
parameters used their default values. Notably, the activation function’s default value
is r elu, and the solver’s default value is adam. This means, that in this experiment
Facial Recognition Application with Hyperparameter Optimisation 165
the default values are compared with other matching values, that were used in Part
1.
Grid Search was used on the MLPClassifier with the library GridSearchCV (scikit
learn.org 2020a). The challenge was to set a good range of values for the NOHN.
The fact that each added value made the process more complex, contributed to the
decision to start with small ranges and conduct several smaller experiments.
The results from Fig. 19 show the three best results with the best accuracies for
different numbers of PCs in the range that was established in Part 1. It is noticeable
that adam as a solver and tanh as an activation produce the best outcome. The batch
size was reduced from 256 from Part 1 to 8 and 4 which contributes to the NOHN
being so low. The very low loss is a benefit and already shows an improvement
when compared to Part 1 where the average loss was mostly at 0.3. Unfortunately,
the accuracy did not improve much when compared to the results of the manual
optimisation. From previous experiments, it could be seen that the optimal number
of PCs is between 200 and 300. This is why this was tested in Grid Search as well.
Figure 19 shows that 250 PCs achieved the best accuracy which is also the middle
of the recommended range. With this in mind, further experiments were conducted
by using extra hidden layers. The number of hidden layers so far has been 1. By
adding a second or even a third hidden layer the ANN becomes more complex.
Many experiments were started, but after 12 h of running time, the programme was
dropped due to Google Colab platform restrictions. Unfortunately, this meant that the
effort in order to find a better set of parameters through Grid Search became too time
consuming. By adding more values to the dimensions, the programme ran longer
and 10 h became the average time and the resulting accuracies were not satisfying.
At this point, the automated optimisation was not more efficient than the manual
optimisation. As a result, another optimisation algorithm was used.
Random Search chooses its values randomly and can therefore explore a wider range
of possible combinations with a lower chance of using the same values twice. Using
250 PCs to simplify the search, there were several parameters with different values
(scikit learn.org 2020b). These parameters with their possible values were defined:
166 H. M. Claus et al.
lb f gs solver, and tanh as an activation function. The fifth set of parameters from
the table seems like an outlier because of the low NOHN but still achieves a high
accuracy.
Grid Search and Random Search provided various sets of parameters. Notably, Ran-
dom Search only showed results that were chosen randomly, so it is not proven that
the provided values for the parameters are indeed the best ones. Thus, the experience
from Part 1 shows that changing just one value of a parameter manually can have a
significant impact. By choosing the two best sets of parameters from Random Search
that share the most values, new experiments were conducted manually. Because the
fourth and sixth set of parameters from Fig. 20 are similar and only differ signifi-
cantly in the size of mini batches, the search was extended. By trying out different
batch sizes it could be seen that 256 achieved the highest accuracy.
By training the ANN and testing the prediction accuracy, new values for the
NOHN could be identified (Fig. 21).
Although the NOHN ranges from 1750 to 2741 the difference in the accuracy
is very small. All accuracies that were achieved during this optimisation had an
average recognition accuracy between 98.1 and 98.2% and an average loss of 0.01.
The NOHN is still within the accepted range between 30 and 3,538. In order to choose
one set of parameters only, there are more iterations of the training process needed.
20 iterations for each set is enough to use it in this comparison but to distinguish and
identify one particular set that can be chosen for later uses as well, therefore, at least
several thousands of iterations are needed.
Fig. 21 The best results through manually optimising random search results
168 H. M. Claus et al.
8 Results
All methods that were used provided resulting sets of parameters that achieved at
least 96% accuracy. Important to mention is the fact that the aspect of time limited the
search for the optimal set of parameters for this face recognition task. Nevertheless,
the result of the optimisation of the outcomes from Random Search showed a variety
of possible NOHN that could be used within the ANN. 98.2% prediction accuracy
is a good outcome that can be used in several applications.
9 Analysis
Analysing the search for an optimal set of parameters showed that it matters which
optimisation method is used. A lot of time ran into the capture of the outcomes of
each iteration in Part 1. In total, 3,240 iterations were undertaken and documented
with an average of three values per iteration. Assuming that the average time of
one iteration and its documentation took 0.5 min, this results in 1,620 min or 27
full hours. The best accuracy achieved by manually looking for the parameters was
96.6%. Comparing this to the time that Grid Search needed, an average of 10 h to get
one good set of parameters where the best one so far was 96.9%, the total amount of
time needed is similar. For three good sets of parameters through Grid Search, it takes
30 h. There the argument of being more efficient only works when stating that the
time that the programme needs, can be used working on other tasks. But in order to
work on this chapter the actual results are needed, hence, Grid Search did not turn out
to be of benefit. Nevertheless, Random Search posed a new approach by displaying
different combinations of values of parameters with an average running time of 5 h.
By choosing the best results from Random Search and extending the optimisation by
adding manual approximations, a good set of parameters was found. Given that the
approximation took exactly one hour and finding six good sets of parameters, each
taking 5 h, the total time needed equals to 31 h. This is still the longest time needed,
but it also provided the best recognition accuracy and embedded two optimisation
methods. Provided that more time is available to conduct deeper experiments, an
even better recognition accuracy could be achieved.
Figure 22 shows the range of values for the hyperparameters that could be iden-
tified through these experiments for achieving the best recognition accuracy and the
lowest error ratio.
Grid Search might be able to provide the best values for the parameters eventually,
but the cost is too high to still be called an efficient optimisation. However, Random
Search does not guarantee to find the one best set of values, but there is a wider
range of better results in less iterations. And combining this with manually tuning
the already optimised search a good range can be defined as shown in Fig. 22.
Facial Recognition Application with Hyperparameter Optimisation 169
Recent studies have found that algorithms used for machine learning show biases
towards classes like race and gender (Buolamwini et al. 2018). If the training dataset
does not show a balanced representation of a variety of people, the use of that algo-
rithm can discriminate against minorities and reinforce disadvantages towards these
groups. After looking closely at the provided dataset (Belhumeur and Kriegman
1997), it proved the missing diversity of people. The dataset has a total of 30 differ-
ent people, of which only approximately 30% represent women, people of colour,
and people with south-east Asian descents together. Since 1997, when the dataset
was created, the demographics of society have changed, therefore, the dataset needs
to be changed, as well. It would be interesting to see, how the performance of the
algorithm would change, if the testing data was more diverse than the training data.
The performance would probably decrease, based on the fact, that the algorithm was
not trained well enough to identify a wider range of ethnicities and genders. Hence,
by choosing a training dataset, a greater representation of the world’s population must
be considered. Although a perfect training dataset does not exist, there are ways to
achieve better performances by using different databases, for instance the database
“Labeled Faces in the Wild” created by the University of Massachusetts (Univer-
sity of Massachusetts 2021). Equally important is the representation of different age
groups. A child or a young adult has different facial features than a middle-aged
or elderly person. The provided dataset showed people in a narrow range of ages,
the majority of which seemed middle-aged. Another experiment could also docu-
ment the behaviour of the algorithm when trained with datasets of people who have
birthmarks, facial hair, or injuries. For instance, the algorithm could have focused
its recognition on people’s eyebrows, thus, people who do not have eyebrows due
to medical conditions would not be identified at all. Identically, the algorithm could
have focused its recognition on the fact that every person has a set of two eyes, which
could potentially lead to a chimpanzee being identified as well. Moreover, another
experiment could analyse the performance of recognition if people wear facial acces-
sories, such as glasses, piercings or even makeup. As a last proposal, the database
“Labeled Faces in the Wild” (University of Massachusetts 2021) offers pictures of
people with various facial expressions and with different backgrounds. A potential
question would be: If a person closes their eyes, would the algorithm still be able to
170 H. M. Claus et al.
identify them? What happens, if people show their teeth? In order to answer these
questions, new experiments would have to be conducted. As a result, these issues can
only be noticed through a complex training and testing phase with a diverse dataset.
11 Conclusion
All things considered, the development and optimisation of a solution for a Machine
Learning task was a detailed process. In the same fashion as natural neural networks
learn, artificial neural networks learn by finding familiar patterns. The goal was
to find a set of parameters for the model to achieve the best recognition accuracy
with the lowest error ratio. Several experiments were conducted to display a wide
variety of solutions. The risk of overfitting the model could be ignored because
of the convenient script that stopped the training process before it could fit to the
peculiarities of the images. The first pattern that was noticed was the fact that sets of
parameters with numbers of principal components between 200 and 300 achieved the
best performances, once the model had achieved a level of complexity. By increasing
the complexity of the ANN, hence, by increasing the number of hidden neurons, the
testing accuracies were improved. Two different methods of tuning the values of the
hyperparameters were experimented with. The first part of this chapter dealt with the
manual optimisation of the values in favour of the recognition accuracy. The highest
accuracy achieved was 96.6% with an average loss of 0.3 and spending 27 h on it.
In the second part, two different optimisation algorithms were utilised, namely Grid
Search and Random Search. Grid Search provided a highest recognition accuracy
of 96.9% and an average error ratio of 0.01 and 30 h of work. Random Search used
the same amount of time but provided an accuracy of 98.1% with a loss of 0.01. By
manually optimising the values from Random Search, an accuracy of 98.2% could be
achieved. In other words, the combination of both automated optimisation followed
by manual optimisation for an approximation of a better accuracy turned out to be
the best solution.
For further use of the shown sets of parameters the risk of overfitting/underfitting
the model needs to be considered. Adding to this, the impact of biased datasets
was discussed with new proposals for experiments and databases. The experiments
showed that the accuracies can change with each iteration, hence, replicating the
experiments will never show the exact same values but they will likely be in the
same range. It would also be interesting to see how the performances would change
with bigger or smaller datasets and whether the relative values would show a similar
behaviour with different data sizes.
All in all, the empirical investigation of finding an acceptable set of parameters
to use in a specific Machine Learning algorithm proved to be an interesting but time
consuming challenge which provided a satisfying result. The requirements of the
task were met by creating a wide range of experiments and analysing the perfor-
mances of the algorithm precisely. This chapter proposed ideas for new experiments
Facial Recognition Application with Hyperparameter Optimisation 171
to understand the functions of this particular algorithm in more detail and to improve
further accuracies.
Acknowledgements The authors are grateful to Dr Schetinin from the University of Bedfordshire
for useful comments, guidance and support.
References
M.J. Sheehan, M.W. Nachman, Morphological and population genomic evidence that human faces
have evolved to signal individual identity, vol. 5 (2014)
K. Gates, Our Biometric Future: Facial Recognition Technology and the Culture of Surveillance
(NYU Press, 2011)
M.H. Hassoun, Foundamentals of Artificial Neural Networks (MIT Press, Cambridge, Mas-
sachusetts, 1995)
R. Bhatia et al., Biometrics and face recognition techniques. Int. J. Adv. Res. Comput. Sci. Softw.
Eng. 3, 93–99 (2013)
M. Mann, M. Smith, Automated facial recognition technology: Recent developments and
approaches to oversight (2017)
G. Hornung, M. Desoi, M. Pocs, Biometric systems in future preventive scenarios—legal issues and
challenges, in BIOSIG 2010: Biometrics and Electronic Signatures, Proceedings of the Special
Interest Group on Biometrics and Electronic Signatures (Bonn, Germany, 2010), pp. 83–94,
Gesellschaft fuer Informatik e.V
S.-C. Wang, Artificial neural network, in Interdisciplinary Computing in Java Programming. The
Springer International Series in Engineering and Computer Science, vol. 743 (2003), pp. 81–100
B. Csaji, H. Ten Eikelder, Approximation with artificial neural networks (2001)
D. Anguita, L. Ghelardoni, A. Ghio, L. Oneto, S Ridella, The “k" in k-fold cross validation, in 20th
European Symposium on Artificial Neural Networks, Computational Intelligence and Machine
Learning (ESANN) (2012), pp. 441–446
B. Kia, Session 14: K-fold cross validation (2018)
T. Dietterich, Overfitting and under computing in machine learning. ACM Comput. Surv. 27, 326–
327 (1995)
H. Jabbar, R. Khan, Methods to avoid over-fitting and under-fitting in supervised machine learning
(comparative study), in Computer Science, Communication & Instrumentation Devices (2014)
S.-W. Lin, K.-C. Ying, S.-C. Chen, Z.-J. Lee, Particle swarm optimization for parameter deter-
mination and feature selection of support vector machines. Expert. Syst. Appl. 35, 1817–1824
(2008)
D.E. Rumelhart, R. Durbin, R. Golden, Y. Chauvin, Backpropagation: Theory, Architectures, and
Applications (Lawrence Erlbaum, Hillsdale, New Jersey, 1995)
R. Rojas, The backpropagation algorithm, in Neural Networks. (Springer, Berlin, 1996)
D.E. Rumelhart, G.E. Hinton, R.J. Williams, Learning representations by back-propagating errors,
vol. 323 (1986), pp. 533–536
scikit learn.org, Mlpclassifier—scikit-learn 0.24.1 documentation (2020)
J. Bergstra, Y. Bengio, Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13,
281–305 (2012)
M. Claesen, B. De Moor, Hyperparameter search in machine learning, in The XI Metaheuristics
International Conference (2015)
J. Buolamwini, T. Gebru, S. Friedler, C. Wilson, Gender shades: intersectional accuracy disparities
in commercial gender classification, vol. 81 (2018), pp. 1–15
S. Sra, S. Nowozin, S. Wright, Optimization for Machine Learning (MIT Press, Cambridge, Mas-
sachusetts, 2012)
172 H. M. Claus et al.
H. M. K. K. M. B. Herath
Abstract The latest advances in the field of information and communication tech-
nology (ICT) are enabling organizations to evolve and expand in the age of “Big
Data”. The growth of big data and the development of Internet of Things (IoT) tech-
nology have aided the viability of modern smart city initiatives. The governments and
industries will use these technological advancements, as well as the widespread use
of ubiquitous computing, to address healthcare requirements in a variety of ways. The
novel COVID-19 and other major pandemic events are known for being unexpected,
unpredictable, and dangerous. With the rapid increase in coronavirus cases, big data
has the potential to facilitate the prediction of outbreaks. As we witnessed, COVID-19
has caused tremendous harm to humanity all over the globe. Owing to the knowl-
edge gained from the novel COVID-19, early pandemic prediction and responses are
crucial. Various approaches for predicting pandemics have been proposed, but none
have yet been developed based on people’s everyday behaviors and environmental
changes. The aim of this chapter is to develop a framework for pandemic prediction
in a smart city by utilizing big data intelligence provided by people and environ-
mental changes. The framework was tested using data from the novel COVID-19
virus, which was spread across Sri Lanka in 2020. Based on the experimental find-
ings, the proposed framework has the potential to predict pandemics in the notion of
the smart city.
1 Introduction
A pandemic is an infectious disease outbreak that has spread around a vast geographic
region, such as several continents or the entire world and affecting a large number
of people. Three major global pandemic outbreaks have occurred in the last few
H. M. K. K. M. B. Herath (B)
Faculty of Engineering Technology, The Open University of Sri Lanka, Nugegoda, Sri Lanka
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 173
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_7
174 H. M. K. K. M. B. Herath
Table 1 Comparison between Coronavirus disease 2019 (COVID-19), Severe Acute Respiratory
Syndrome (SARS), Middle East Respiratory Syndrome (MERS), and Ebola Virus Disease (EVD)
Description COVID-19 SARS MERS EVD
Incubation period (days) 2–14 2–10 2–14 2–21
Contagious during incubation Yes Yes Yes No
Vaccine as of 2020 No No No Yes
Infected body part Blood vessels Respiratory system All muscles
Main transmission Respiratory droplet secretions Body fluids
Shape and size in nm Spherical Spherical Spherical Filament
80–120 80–90 90–125 14,000 × 80 nm
Mortality rate % 2.65 14–15 34 Up to 90
Pronounced R-naught 2–2.5 3.1–4.2 <1 1.5–1.9
Internet-Assisted Data Intelligence for Pandemic Prediction … 175
Smart city prospects are very promising, and various edge computing manufac-
turers, such as IBM and Intel, are launching various projects to consolidate their guid-
ance in this field. Smart health, smart transportation, smart security, smart govern-
ment, smart buildings, smart grid, smart tourism, smart lifestyle, smart home, and
smart climate are known to be ten critical fields that will play key roles in creating a
smart community (Caragliu and Bo 2019). Big data in healthcare is problematic not
only because of its sheer scale but also because of the heterogeneous quality of the
data and the pace at which it must be handled. The integration of big data into smart
healthcare systems has ushered in a new age in electronic and mobile healthcare
that is bearing results in terms of cost-effectiveness and productivity gains. Rapid
advancements in ICT, the Internet of Things (Herath et al. 2021b), cloud computing,
and smart applications have allowed sharing of real-time data. The large volume
of various urban data allows for a more comprehensive and holistic view of urban
environments and real-time events. Cyber-physical systems, the Internet of Things,
big data, and cloud computing are the four primary tools used by software networks
for smart cities, according to Santana et al. (2017). Table 2 describes the different
smart city platforms and their available technologies for smart applications.
Big data refers to the massive evolution of structured, unstructured, and semi-
structured data. Through careful interpretation, businesses can gain greater perspec-
tives and make smarter decisions by processing them optimally. Big data analytics
can derive benefit by developing a modern wave of technologies and architecture that
can achieve high velocity, discover, and analyze large volumes and varieties of data
(Beyer and Laney 2012; Laney 2001; Ward and Barker 2013). The growth of big data
and the development of Internet of Things (IoT) technology have aided the viability
of smart city initiatives (Hashem et al. 2016). The IoT facilitates the convergence
of sensors, Radio-frequency identification (RFID), and Bluetooth in the real-world
setting using highly networked services, and big data offers the ability for cities to
gain useful information from a vast volume of data gathered from multiple sources.
176 H. M. K. K. M. B. Herath
A list of the various smart city applications is described in Table 3. In this chapter, the
author has mainly focused on smart healthcare and smart transportation technologies
in the IoT.
The author has recently identified a number of variables linked to the early iden-
tification of the COVID-19 pandemic (air pollution, hospital admission rate, death
cases, parking density). The latest pandemic is also having an effect on government
and medical facilities. As a result, there is also a need to identify potential pandemics
before they occur. One of the most critical technologies for the early prediction of a
pandemic scenario is big data. The aim of this chapter is to develop a framework by
enabling big data intelligence for future pandemic predictions in smart cities. The
rest of this chapter is categorized into four sections. In Sect. 2, a literature review is
carried out. Section 3 discussed the proposed framework architecture. Results and
observations are discussed in Sect. 4. Finally, Sect. 5 brings the proposed system to
a conclusion.
2 Literature Review
This section addresses related works in the fields of IoT, smart healthcare, and big
data-driven technologies in a brief manner. Google Scholar was used to reviewing
academic works from 2015 to 2020. The author has considered the factors that impact
the novel COVID-19 since this chapter primarily focuses on the big data-enabled
framework for pandemic prediction.
Remote sensing and the Hadoop ecosystem have recently provided effective and
low-cost solutions in a variety of domains, especially in healthcare. On the one hand,
sensing devices enable doctors and nurses to track patients remotely, whether in the
hospital or at home, and to diagnose medical emergencies in real-time. Hadoop-based
Internet-Assisted Data Intelligence for Pandemic Prediction … 177
tools, on the other hand, provide hospitals with cost-effective and fast data storage
and processing platforms, especially in extreme situations like virus contamination
or global pandemic situations. Noh and Lee (2015) have introduced an integrated
big data platform model which focuses on open-source S/W configuration.
Harb et al. (2020) have proposed an efficient big data analytics tool for real-time
patient tracking and decision-making, which would benefit both hospital and medical
personnel. They have proposed a framework that is made up of four layers such as
1: real-time patient surveillance, 2: real-time evaluation and data storage, 3: patient
classification and disease diagnosis, and 4: data retrieval and visualization based on
big data analytics and data processing techniques. The efficacy of their proposed plat-
form in terms of patient classification and disease diagnosis in healthcare applications
has been demonstrated in experimental findings.
Prospects of the modern healthcare industry have accelerated by combining
big data and smart systems. Pramanik et al. (2017) have addressed a systematic
assessment of various big data and smart system developments, as well as a crit-
ical examination of current advanced healthcare systems. They’ve also suggested a
big data-enabled smart healthcare system architecture, which provides theoretical
representations of intra and inter-organizational business models in the healthcare
sector.
Assist from sensors and IoT devices are important for monitoring the health condi-
tions of a person. Syed et al. (2019) have discussed healthcare services to diseased
and healthy populations through remote surveillance using intelligent algorithms,
software, and techniques, as well as fast analysis and expert intervention for effec-
tive treatment decisions. They have also presented an outline of various data analysis
algorithms and big data platforms for smart healthcare applications.
Jia et al. (2020) have presented a detailed study on characteristics of novel COVID-
19, as well as key big data capabilities, such as online public opinion analysis,
virus-host analysis, pandemic visualization, COVID-19 close contact inspection,
and pandemic prediction assessment.
The adoption of big data strategies to increase the quality of healthcare delivery has
become unavoidable due to the growing volume of data in the healthcare industry.
Aboudi and Benhlima (2018) have proposed an extensible big data model which
focused on both stream and batch computing to improve the efficiency of healthcare
services even further by providing real-time warnings and allowing precise forecasts
on patient health status.
The implementation and incorporation of smart city technologies can be supported
by smart city software platforms. Santana et al. (2017) have reviewed the state-of-the-
art of smart city software platforms and they have studied 23 proposals and classified
them into four groups based on the most often used supporting technology, as well
as practical and non-functional requirements: Big Data, Internet of Things, Cyber-
Physical Systems, and Cloud Computing. The author has studied existing smart city
system domains that have the ability to detect pandemic scenarios in a smart city by
studying Santana’s article. Table 4 shows the different domains of smart city systems
which are related to pandemic scenario prediction.
178 H. M. K. K. M. B. Herath
A comprehensive study and framework of the use of big data in the battle against
major public health outbreaks such as COVID-19 are also needed, which will serve as
a valuable guide for governments and medical institutions. The author has acquired
a clear understanding of existing big data-enabled technologies after completing the
literature review. Based on the knowledge and experience gained from the COVID-19
pandemic, the author has proposed a single framework to predict future pandemics
in a smart city. Section 3 of this chapter describes the methodology of the proposed
framework.
This section discusses the development of the proposed pandemic prediction frame-
work, which is based on big data intelligence and IoT technologies. After data was
collected from the big data sources, it was used to predict the spread of pandemics
and alert the relevant institution for future actions.
Big data obtained from various domains, such as vehicle parking density, air
quality standards, and healthcare-related data, was widely used in the proposed design
framework. The author has extracted a few domains that could be useful in predicting
a pandemic. These criteria were chosen based on the knowledge and experience of
several pandemics that have occurred in recent years. The big data repositories used
in the development of the proposed framework are described in Table 5.
Figure 1 shows the program flowchart of the proposed framework. The big data
collection procedure, as shown in the figure, separates data into three categories
(Healthcare data, PM2.5 index data, and parking data). From the A&P model, the
system was collected the dependent variables related to air quality and parking density
measures. The SEIR model was generated using information from hospital data. Input
variables DH , DA , and DT were sent to the Machine learning (ML) algorithm. The
ML process was used to predict the current state of the pandemic after the data was
collected from the SEIR and A&P models.
Internet-Assisted Data Intelligence for Pandemic Prediction … 179
Table 5 Domains and data repositories used for the proposed framework
Domain Data source/s Data uses
Healthcare Johns Hopkins University Center for COVID-19 confirmed, death, and
Systems Science and Engineering data recovered cases
repository (Dong et al. 2020)
Air pollution Colombo US Embassy’s air quality Air quality index data
monitor data repository (Embassy and
Air Pollution: Real-time Air Quality
Index (AQI). 2021)
Parking space Data.World data repositories Parking station ID, year/month, parking
(Data.World 2021) usage, and percentage of usage
Data was collected from the various sources mentioned in Table 5 which were
stored in big data repositories. The author has mostly centered on the data providers
relating to parking management, air quality, and healthcare-related data. According
to the latest pandemic experience, as a pandemic strikes, road traffic density has
decreased (available parking spaces have reduced). As a result, air pollution has
reduced. These criteria were chosen to identify a pandemic scenario. The SEIR
model, which is used to simulate pandemics, was developed using healthcare-related
data sources. The variable S(t), E(t), I(t), D(t), and R(t) were calculated from COVID-
19 data repositories which are available at (Johns Hopkins University Center for
Systems Science and Engineering data repository (Dong et al. 2020)).
180 H. M. K. K. M. B. Herath
The SIR model is well-known in the field of epidemiology, and it has yielded many
significant results. The susceptible, infected, and recovered people are represented
by the letters S, I, and R in the model (Smith and Moore 2004; McCluskey 2010; Zhu
and Ying 2014). The coronavirus disease, on the other hand, has a latent period in
which the individual is infected but not yet symptomatic. In the SEIR (Susceptible-
Exposed-Infected-Recovered) model, a new part E is added to define the symptomless
but infected process (Li et al. 2021). In this section, we proposed an SEIR model
for predicting pandemics based on the number of susceptible, exposed, infected, and
recovered individuals collected from big data repositories. Relationships between
each part of the SEIR model are shown in Fig. 2.
Differential equations represented in Eq. 1 describe the relationship of each
component of the SEIR model.
−r bS I r2 b2 S E
Ṡ = −
N N
r bS I r2 b2 S E
Ė = − aE +
N N
I˙ = a E − y I − u I (1)
Ṙ = y I
Ṅ = −u I
Ḋ = u I
were,
r: The number of contaminated contacts who are susceptible,
r 2 : The number of contacts who have been exposed to the virus,
b: Probability of disease transmission between a susceptible and infectious person,
b2 : Probability of disease transmission between a susceptible and an exposed
person,
a: The rate at which an infected person becomes infected after being exposed,
The total population (N(t)) is described using Eq. 2. As shown in the equa-
tion, the total population is a relationship between susceptible population (S(t)),
exposed population (E(t)), infected population (I(t)), recovered population (R(t)),
and the number of death cases (D(t)). Figure 3 illustrates the working principle of
the proposed SEIR algorithm.
After the acquisition of healthcare-related data from the big data repositories,
the “big data processing” function was performed to extract the model data for the
“parameter estimation” process. The SEIR model was generated using the parame-
ters extracted from the “parameter estimation” process. The model also used cloud
computing technologies to display real-time data and graphs related to the pandemic.
The development of an air quality and parking space measuring (A&P) algorithm to
predict possible pandemics in a smart city is described in this section. The author
has used big data repositories to derive dependent variables including the air quality
index and the data related to the parking space available in a parking station. The
algorithm was developed using the Python development environment. Following the
extraction of data, the system evaluated the value to determine the time span during
which the pandemic would have an effect on the city. We know from experience of
the latest pandemic that as people are exposed to the virus, the number of hospital
admissions rises. Therefore, People are placed in isolation at home or in a quarantine
area. As a result, traffic in the area reduced, and air quality improved. The author has
used the same phenomena to develop the algorithm. Figure 4 depicts the program
flow chart of the proposed A&P algorithm.
In the pre-processing phase, data normalization was performed and features were
identified. During the pandemics, parking spaces accessible at the parking station
were occupied at a higher percentile, according to public officials. In this study, more
than 80% of parking usages were selected as a potential danger in the area. It’s also
used as a reference for air quality measurements, with a 0–50 index. Following the
selection of features, an artificial neural network (ANN) was used to determine the
city’s pandemic status (P(V i , Ai )). For the activation of nodes in the ANN, the author
has used the ReLU activation function. After the ANN algorithm generates the data,
the pandemic level (Pi ) was estimated and the pandemic curve was plotted. The
author has developed a formula to calculate the pandemic stages as shown in Eq. 3.
Vi Ai
Pi = 0.78 − 0.63 − 2.32 + 0.51R0
n n
W her e,
Pi = pandemic level at ith day
Vi = rate o f parking usage at ith day (3)
Ai = rate o f air qualit y at ith day
R0 = basic r epr oduction number
i = day number
n = no o f days
Table 6 depicts the pandemic levels and the actions that must be taken to prevent the
spreading of the disease or virus. The author has identified three potential regions
of the developed pandemic curve in a smart city, as seen in the table. Where the
pandemic level is in the 0.8–1.3 range, it is considered a threat to the city, and
relevant institutions or government officials must act immediately. If Pi is in the
range of −0.3 to 0.0, there is no disease in the city, and no action is needed.
4 Experiment Results
The aim of this study is to develop a framework for predicting pandemics in smart
cities using big data from healthcare, air quality, and parking spaces. The author
used data obtained from the Johns Hopkins University Center for Systems Science
and Engineering (CSSE) data repositories, the US Embassy’s Colombo Air Quality
Monitor data repository, and parking station database accessible at Data.World to
evaluate the efficiency of the proposed framework. For the experimenting on the
proposed framework, the Python development environment was used.
184 H. M. K. K. M. B. Herath
For the extraction of confirmed, death, and recovered COVID-19 cases, the author has
used the COVID-19 upstream repository which is maintained by the Johns Hopkins
University Center for Systems Science and Engineering (CSSE). The COVID-19
status was generated after the data was collected. Figure 5 depicts the extracted data
from the repository by using the filtering keywords (LK) in the Python environment.
In Sri Lanka, Fig. 6 depicts the behavior of confirmed, death, and recovered
COVID-19 cases. The graph was generated in the Python environment using variables
extracted from the COVID-19 data repository.
The classic epidemic curve was obtained by plotting the values of the S, E, I, and
R compartments at each time step are shown in Fig. 7. After the data was collected
from the big data, it was processed for the S, E, I, and R models. The SEIR model was
generated based on the COVID-19 cases for the first 60 days in Sri Lanka. Initially,
the system was set to y = 0.12, and β = 0.3 for the total population (N) of 415 cases.
In the diagram, the population is completely vulnerable at the outset, with just a few
infected individuals. As the virus progresses, the number of infected people increases
Fig. 5 Result window of the processed data from big data repositories
Fig. 6 Result window of the acquired COVID-19 cases through big data repository
Internet-Assisted Data Intelligence for Pandemic Prediction … 185
until it reaches a peak, after which it starts to decrease as there are fewer and fewer
people to infect. When the R0 (basic reproduction number) value exceeds one (R0 >
1), the infection rate exceeds the recovery rate, and the infection spreads across the
population. R0 = 2.5 (R0 > 1) was calculated using the SEIR model, indicating that
the pandemic is spreading across the community.
The results and observations of the air quality-based pandemic prediction model
are discussed in this section. The proposed algorithm was developed in the Python
development environment. The author collected major data on air pollution from the
data repository of the Colombo US Embassy’s air quality monitor. Traffic congestion
on the roads reduces as people become isolated due to the pandemic. This would
lead to improved air quality than on a typical day. This phenomenon was used by the
author as one factor in predicting a pandemic in a city. Fine particulate matter (PM2.5 )
is an atmospheric pollutant that poses a health risk when concentrations are high.
When PM2.5 levels are high, the air becomes hazy and visibility is reduced. Data
on fine particulate matter (PM2.5 ) was used to design the algorithm. Figure 8 shows
the result window of data acquisition from data repositories in the air quality-based
pandemic prediction model.
Figure 9 depicts the data gathered during the last 17 months in Colombo. The lower
air quality index is shown in green, while the highest air quality index is shown in red.
During the COVID-19 period, Colombo, Sri Lanka, had lower air pollution due to
186 H. M. K. K. M. B. Herath
Fig. 9 Result window for last 17 months air quality index of the Colombo city
lower traffic effects, as seen in the illustration. From May to October 2020, people in
Sri Lanka were contaminated with COVID-19, and they were forced to stay at home
as a precaution. As a result of this occurrence, the city’s traffic density decreased,
and air pollution was decreased.
Figure 10 depicts the processed and analyzed data for the last 17 months in
Colombo city of Sri Lanka. Lower air quality indexes were also identified during the
COVID-19 pandemic, according to the data. The second potential pandemic threat
was observed in April 2021, as seen in the figure.
Internet-Assisted Data Intelligence for Pandemic Prediction … 187
This section outlines the findings and observations of vehicle parking availability
during the pandemic period. Figure 11 depicts the impact of parking capacity over
the past 11 months in 2020, based on the data extracted from the data repositories.
The results were produced using the Python programming language. According to
the test results, the highest percentile was registered from May to October, when
COVID-19 was at its peak in the city.
Figure 12 depicts the analysis of the experiment findings based on the data
collected. According to the test findings, the yellow color field is the threat zone.
The monthly average vehicle parking rate was found to be 38.31% (N mean = 552).
According to previous data, the author has observed that 80% of parking spaces were
not occupied on a typical day. Therefore, 80% was chosen as the reference value of
the model. As a result, the period from June to September 2020 was described as a
pandemic-affected period in the city. During the pandemic, the average percentage
of parking spaces was occupied by 88.75% (N mean = 1282). The city’s pandemic
period (Jun–Sep) was identified in the highlighted region (red) of the graph.
Figure 13 shows the pandemic curve (described in Eq. 3) generated by the frame-
work. According to the pandemic curve, the pandemic has arisen in the city when the
pandemic level reaches greater than or equivalent to one (Pi ≥ 1). In the illustration,
the danger zone is represented by red nodes.
The framework was tested using data obtained from the recent COVID-19 inci-
dent at Colombo, Sri Lanka. The proposed framework has the potential to predict
epidemics using big data, according to the test results. The mean absolute error
(MAE) of the proposed system was found to be 0.16. As a result, the system operates
at a higher degree of efficiency.
Internet-Assisted Data Intelligence for Pandemic Prediction … 189
5 Conclusion
Researchers, information technicians, and programmers from all over the world have
been experimenting with a variety of technological methods and developments in
order to improve pandemic detection and response using modern IoT-based tools.
With the rapid increase of infectious diseases, big data has the potential to facilitate
the prediction of outbreaks. The author of this chapter suggested the development of
a framework focused on big data related to healthcare, air quality, and the availability
of parking spaces in the city. Data from a recent COVID-19 incident in Colombo, Sri
Lanka, was used to test the framework. For the COVID-19 pandemic, each domain
has tested and confirmed the results. The mean absolute error (MAE) of the proposed
system was found to be 0.16. According to the findings of the experiments, the
proposed system has the potential to predict pandemics. The author’s next step is
to develop an algorithm that considers all geographical regions and compares the
efficacy of various smart city models.
References
M.A. Beyer, D. Laney, The Importance of ‘Big Data’: A Definition, Gartner, Stamford, CT, pp. 2014–
2018 (2012)
O. Borraz, P. Le Galès, Urban governance in Europe: the government of what? Pôle Sud 1, 137–151
(2010)
L. Benhlima, Big data management for healthcare systems: architecture, requirements, and
implementation. Adv. Bioinf. (2018)
A. Caragliu, C.F. Del Bo, Smart innovative cities: the impact of Smart City policies on urban
innovation. Technol. Forecast. Soc. Chang. 142, 373–383 (2019)
Colombo US Embassy Air Pollution: Real-time Air Quality Index (AQI) (2021). https://aqicn.org/
city/sri-lanka/colombo/us-embassy. Accessed 1 May 2021
E. Dong, H. Du, L. Gardner, An interactive web-based dashboard to track COVID-19 in real time.
Lancet. Infect. Dis 20(5), 533–534 (2020)
Data.World. (2021). https://data.world/home/. Accessed 1 May 2021
C. Drosten, S. Günther, W. Preiser, S. Van Der Werf, H.R. Brodt, S. Becker, H. Rabenau, M. Panning,
L. Kolesnikova, R.A. Fouchier, A. Berger, Identification of a novel coronavirus in patients with
severe acute respiratory syndrome. N. Engl. J. Med. 348(20), 1967–1976 (2003)
H.M.K.K.M.B. Herath, G.M.K.B. Karunasena, S.V.A.S.H. Ariyathunge, H.D.N.S. Priyankara,
B.G.D.A. Madhusanka, H.M.W.T. Herath, U.D.C. Nimanthi, Deep Learning Approach to
Recognition of Novel COVID-19 Using CT Scans and Digital Image Processing (2021)
H.M.K.K.M.B. Herath, G.M.K.B. Karunasena, H.M.W.T. Herath, Development of an IoT Based
Systems to Mitigate the Impact of COVID-19 Pandemic in Smart Cities, in Machine Intelligence
and Data Analytics for Sustainable Future Smart Cities (Springer, Cham, 2021), pp. 287–309
H.M.K.K.M.B. Herath, Internet of things (IoT) enable designs for identify and control the COVID-
19 pandemic, in Artificial Intelligence for COVID-19 (Springer, Cham, 2021)
I.A.T. Hashem, V. Chang, N.B. Anuar, K. Adewole, I. Yaqoob, A. Gani, H. Chiroma, The role of
big data in smart city. Int. J. Inf. Manage. 36(5), 748–758 (2016)
H. Harb, H. Mroue, A. Mansour, A. Nasser, E. Motta Cruz, A hadoop-based platform for patient
classification and disease diagnosis in healthcare applications. Sensors 20(7), 1931 (2020)
190 H. M. K. K. M. B. Herath
Q. Jia, Y. Guo, G. Wang, S.J. Barnes, Big data analytics in the fight against major public health
incidents (Including COVID-19): a conceptual framework. Int. J. Environ. Res. Public Health
17(17), 6161 (2020)
Lai et al. 2020 Apr 14Y. Lai, W. Yeung, L.A. Celi, Urban intelligence for pandemic response:
viewpoint. JMIR Publ. Health Surveill 6 (2), e18873 (2020). https://doi.org/10.2196/18873
D. Laney, 3D data management: Controlling data volume, velocity and variety. META Group Res.
Note 6(70), 1 (2001)
C. Li, J. Huang, Y.H. Chen, H. Zhao, A fuzzy susceptible-exposed-infected-recovered model based
on the confidence index. Int. J. Fuzzy Syst. 1–11 (2021)
C.C. McCluskey, Complete global stability for an SIR epidemic model with delay–distributed or
discrete. Nonlinear Anal. 11(1), 55–59 (2010)
K.S. Noh, D.S. Lee, Bigdata platform design and implementation model. Indian J. Sci. Technol.
8(18), 1 (2015)
M.I. Pramanik, R.Y. Lau, H. Demirkan, M.A.K. Azad, Smart health: big data enabled health
paradigm within smart cities. Expert Syst. Appl. 87, 370–383 (2017)
E.F.Z. Santana, A.P. Chaves, M.A. Gerosa, F. Kon, D.S. Milojicic, Software platforms for smart
cities: concepts, requirements, challenges, and a unified reference architecture. ACM Comput.
Surv. (CSUR) 50(6), 1–37 (2017)
L. Syed, S. Jabeen, S. Manimala, H.A. Elsayed, Data science algorithms and techniques for smart
healthcare using IoT and big data analytics, in Smart Techniques for a Smarter Planet (Springer,
Cham, 2019), pp. 211–241
D. Smith, L. Moore, The SIR model for spread of disease-the differential equation model.
Convergence (2004)
United Nations. 68% of the world population projected to live in urban areas by 2050, says UN.
2018 Revision of World Urbanisation Prospects (2018)
J.S. Ward, A. Barker, A. Undefined by data: a survey of big data definitions (2013). arXiv:1309.
5821.
K. Zhu, L. Ying, Information source detection in the SIR model: a sample-path-based approach.
IEEE/ACM Trans. Netw. 24(1), 408–421 (2014)
NHS Big Data Intelligence on Blockchain
Applications
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 191
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_8
192 X. Feng et al.
1 Introduction
2 Background of Blockchain
Blockchain was defined as a system used to make a digital record of all the occasions
a cryptocurrency (a digital currency such as bitcoin) was bought or sold, and that was
constantly growing as more blocks were added Cambridge (2021). In other words,
NHS Big Data Intelligence on Blockchain Applications 193
Blockchain technology had been widely applied in many areas. Blockchain tech-
nology was most simply defined as a decentralized, distributed ledger technology
that records the provenance of a digital asset (Builtin 2018). Due to the advantages
of Blockchain feature, applications were quite broad in the growth. Blockchain had a
nearly endless number of applications across almost every industry. For instance, the
ledger technology could be applied to track fraud in finance, securely sharing patient
medical records between healthcare professionals and even acted as a better way to
track intellectual property in business and music rights for artists. Metropolitan police
could also make use of Blockchain technology to secure cyber-bulling and cyber-
stalking forensics evidence (Feng and Short 2017). There were some of examples of
applications of this technology being discussed in Sect. 2.1.3. A combination of tech-
nologies of cryptography public and private key technology, decentralized network to
share distributed accounting technology and motivate mechanism of record, store and
integrity security service for network transactions formed Blockchain technology.
A schematic diagram of how Blockchain working being demonstrated in Fig. 1.
It demonstrated how Blockchain worked in principle. Basically, Blockchain was a
distributed digital transactions ledger. As soon as the transactions were recorded and
verified, it was almost impossible that anybody could alternate it. Everybody in the
Blockchain had the transactions record. That meant people could not change every
transactions records in the world, which made Blockchain immutability integration
characteristics (Zhao et al. 2021).
The principle of work was, whenever each transaction happened, this accounting
detail was sent to everyone in this Blockchain network nodes. Therefore, each
nodes kept a record of this transaction. After verified, the records were kept in this
Blockchain. When the transaction accounts (transaction bookkeeping) up to nearly
one megabyte, these transactions were packed into a block. A packed block contained
about 4000 transaction records. This block was chained to the previous Blockchain
and expect the next block.
Blockchain offered permanent and transparent transaction records. The whole
network had a consistent account record. Everybody was able to check and trace the
record. In Blockchain distributed database, account was not centralized control by any
one node, but controlled by all the nodes, joint maintenance and joint bookkeeping
by all.
Any single node could not assume it was a main part to alternate any record. In
order to change a record, which require to control the majority (more than 51%) of
the nodes. While Blockchain had unlimited nodes and increasing nodes all the time,
that made the attempt of making any change impossible.
The essence of Blockchain technology was an open accounting/recording system
which verify each other. This system record all the accounts happened all the
transactions or activities. Every transactions changes were recorded in the network
summary/total account. Every attendee all had a complete network summary. The
transparency characteristics guaranteed everybody could statistic all the attendee’s
records and the current situation independently (D’Aliessi 2016).
Due to all the Blockchain open and transparent to everyone, anyone could check
the source. People could trust this decentralized Blockchain system, and not to worry
if there were any hiding conspiracies inside there.
To summarize, since Blockchain technology allowed checking, tracing and storing
all types of information in its chain, these characteristics made Blockchain the most
security system on earth. Blockchain technology became more and more popular
and broadly used on many areas (Zhao et al. 2021).
Blockchain technology was not only to be able to create cryptocurrency, but also
to support personally identifiable information (Feng et al. 2021), peer review, e-
voting and other democratic decision and audit tracing trading, financial transaction,
culture, medical, science, business, logistics, messaging services and smart contract,
NHS Big Data Intelligence on Blockchain Applications 195
legal and justice system; as well as public health. For instance, NHS could store
patients’ medical records in Blockchain to make medical records location indepen-
dently accessed. Pharmaceutical industry could use Blockchain to verify medicine,
in order to prevent false, anti-fake and so on.
Zhao et al. (2021) had analyzed some details, Blockchain had four innova-
tive features if compare with conventional technologies: “peer-to-peer network,
cryptography, chaining-based data structure and PoW (proof of work)”.
If a Blockchain operated on a peer to-peer network where there was decentralized
control. This characteristic removed likelihood of any single point of fault in the
system to affect other nodes. That reduced any negative influence on compromise of
individual points, which improved the weakness of designed systems.
Cryptography is the foundation of any security system. Especially, cryptographic
hash offered one-way cryptographic hash encryption, was normally used broadly.
Furthermore, public key infrastructure cryptography algorithm were used to generate
and verify digital signature, that was ensuring accountability and non-repudiation for
all of digital transactions integration, for authentication of users.
Blockchain data structure, in a conventional Blockchain, was a single chain of
blocks, where every block had exactly one parent block, apart from the beginning
block in the chain, (which was referred to as genesis or parent block). This data
structure was the reason of Blockchain technology was referred. One block consisted
of multiple transactions. Typically, the block was one Megabyte per block maximum
size, (about 4000 transactions). This parameter would directly affect the throughput
of the Blockchain platform. By chaining the blocks together, any alteration of any
block, could be easily detected. Furthermore, due to big data redundancy existed in
the Blockchain system, it was almost impossible for any adversary to change any
transaction in the record on the Blockchain. This Blockchain was an implementation
of the secure distributed ledger. Each block had a predefined header structure. The
block header consists of several vital fields. One of them was the hash value of
the parent block that connected the new block to an existing Blockchain (D’Aliessi
2016).
PoW (proof of work) was the most important characteristic. Normally, traditional
consensus algorithms, which resorted to the use of election to control the current
membership of the blocks (including who should be the primary) and multiple rounds
of message exchanges to reach a consensus when a predefined condition was satisfied.
PoW converted the consensus issue into a competitive puzzle solving game and
whoever solves the puzzle first, gained to determine the next block and win a reward
as if the one had successfully mined a pot of treasure. This was the main reason
that participants who entered the proof of work competition were often referred to
as miners. A miner’s task was to assemble the next block as quickly as possible
and try out different random numbers that would make the hash value of the block
header smaller than the difficulty target value. Due to the nature of the cryptographic
hash, one cannot predict which random number would solve the question. Hence,
the only way to find the solution was to try many times until one was found. This
puzzle design is later formalized as a non-interactive zero-knowledge proof issue.
196 X. Feng et al.
These included memory security, thread security and process security, output and
input data security and AI security and big data management (Zhao et al. 2021).
In addition, Zhao et al. (2020) had applied Blockchain technology on Hierar-
chical Processing and Logging of Sensing Data and IoT Event. Besides, IBM also
had mainframe application on hybrid Blockchain technology. Another company that
using the unique technology was Ford, IBM had working together with Ford, plan
to track its raw materials transactions (IBM 2020).
The current healthcare system had not fully modernized. It relied heavily on the
interaction between patient and doctors and works on limited data. The limitation
aspect of the healthcare resulted in a healthcare system that kept taking advantage
of the data efficiently. Also, the current process of getting healthcare is long and
tedious at its best. To improve healthcare system effectively handling of the patient
is necessary (Feng et al. 2021).
Blockchain based health network held the complete medical history for each
patient, could be more efficient and secure solution (IBM 2020) as shown in next
section Fig. 2.
The Fig. 1 shown the flow of transactions in the Blockchain technology. Some
Blockchain application examples could be discussed next (Steemit 2017).
NHS clinical trial was a simple method of testing new drugs like Covid19 Coron-
avirus Vaccine and its effectiveness in a controlled environment. Not only that the
pharmaceutical industries needed to invest heavily in clinical trials. There might
appear counterfeit/frauds happened in clinical trials. Clinical trials produce many
big data. It includes statistics, reports, surveys, medical imagery, blood tests and so
on. Although anyone review the clinical reports could ensure that the results of the
trial were true. But that was not always the case.
Since Blockchain could act as the medium to facilitate clinical trials. As
Blockchain provided data integrity, it could act as a proof when the authenticity
of the documents needed to be verified. It required other nodes for the verifications
process. Overall, the distributed networks ensure that the data integrity should be
maintained and made no data could be modified without authorized access. The
simple idea of data integrity could change how clinical trials took place. Blockchain
would be a system with no loophole and Blockchain improve NHS healthcare posi-
tively. Blockchain used SHA-256 hash, one of cryptography algorithms to ensure that
198 X. Feng et al.
the data immutability by any third-party malicious act (Feng et al. 2021). Blockchain
implementation used examples included Amgen and Sanofi etc. (Iredale 2019).
Patient Data Management had been an outstanding issue. NHS health were facing
varied issues and health concerns. This meant that each patient was different which
made big data management of patients not easy. As each disease worked differently
for different patients, it was not possible to produce a clear logical structure or use
a common treatment strategy. If treatment worked on one patient, it did not mean
that it would work on everyone with the same syndrome. All these complex issues
made it necessary to had complete medical records for the patient. This would enable
personalized care with a focus on patient-centric treatment.
Another challenge that most of the medical doctor went through were lack of
information availability. This could result in patient treatment adding more cost
to overall healthcare. The patient big data were not safe as which should be. The
systems on which the data were stored were not completely secure. Doctors also use
insecure way to share information. For example, we could easily see doctors sharing
information through social media. This could easily lead to the patient PII data leak.
Moreover, the patients were never in charge of their data. Data ownership were also
an issue where the organizations mainly posed the patient’s big data without the
proper permission of the patient.
The solution would be, Blockchain provided a better trade-off to handle patient
data management. It provides a structured way to store data which could be accessed
by the right professionals. Patients’ data could be stored on the Blockchain and could
only be kept accessible to the patient, which the doctor who is handling the case.
Access could be revoked anytime which ensuring that the patient had full authority
on whose medical reports. That did not mean that not all PII data were inacces-
sible. Insensitive data could be available publicly which then could be accessed by
healthcare organization or stakeholders through API. This ensured proper collabora-
tion across different systems. Doctors could also request data whenever possible and
location independently (Feng et al. 2021). Patients could also share their data without
revealing their identity. This was especially favorable to the healthcare institutes as
they could use the anonymous data to improve their healthcare research and systems.
Patients always had control of their data, and they decided who had authentication
to access the data.
When combined with IoT big data, researches could constantly monitor the
patient’s condition such as heartbeat and other vital body functions. This would help
improve patients health care were taken care of. IoT with the help of Blockchain could
enable faster decision making and saving patients’ life in critical life-threatening
conditions and get the forensics evidence (Alexakos et al. 2020) for the multi devices.
To implement use of Blockchain as their solutions like digital health company
DokChain, the charitable foundation Patientory and so on (Iredale 2019).
NHS Big Data Intelligence on Blockchain Applications 199
n >3f (2.1)
where n is the total number of nodes in the Blockchain, and f is the number of
questionable nodes in the Blockchain.
A consistency and correctness could be achieved under the equation of (2.1)
condition. In fact, if more than 51% nodes in the Blockchain (the majority) were
good nodes, that Blockchain could perform its function well well as desired. Using
recurrence recursive algorithm, nested block took the largest vectors, consistency
and correctness were still achieved by PBFT (Practical Byzantine Fault Tolerance)
kept consistency and correctness. If the condition could not be satisfied, it would be
impossible to keep a consistency and correctness.
200 X. Feng et al.
For example, let f was the number of questionable nodes, and n was the total
computer nodes number in the system.
When n > 3f being satisfied, the integrity could be kept. There were consistency
and correctness achieved in that Blockchain.
For instance, let f = 2, n > 7, consistency and correctness would be kept. PBFT
recursion algorithm could make Blockchain network consistency and correctness
(Feng et al. 2021). DEA (Data Envelopment Analysis) was an optimization method
for measuring the efficiency and productivity of decision-making units with multiple
inputs/outputs (Ali et al. 2010) could be used on NHS data.
Blockchain was designed about 10 min to produce a block in the world. There were
21 million bitcoin in the world in total, which could be used as accounting reward for
mining blocks. As nobody could possibly to predict which member would solve the
question for the next new block, this unpredictability was essential for the integrity
of the network because if it is known a member could be the one, who decides on the
next new block, then this member could be corrupted by bribes or being threatened by
others to include questionable transactions that maybe led to some kind of challenges,
or to exclude legible transactions that might cause DDOS (distributed deny of service)
cyber-attack.
The algorithm mining principle would be:
a (character) string = Front block header + transactions + Time Stamp + random
number.
Hash = SHA256(SHA256(string)) = 000……00,000.…..; (a 256 digit string).
Then, this 256-digit hash was the newest block’s header.
Based on the n calculation.
The probability of success in PoW was, in binary, the probability of to get a 0,
was ½.
½ × ½ × ½ × … … × ½ = (½)n .
In order to adjust the value of number n could realize every 10 min to create a
new block (which contained several thousand transactions).
The complexity: if there were 10 000 mining machines in the world, each mining
machine calculation ability is 14 T per second.
1 T = 1012 .
That is, they could calculate 14 T (14 × 1012 ) Hash digest value computation in
every second = 1.4 × 1013 per second.
Therefore, in 10 min, the every mining machine computation could be
= 10 × 60 × 1.4 × 1013 .
= 600 s × 1.4 × 1013 at each mining machine.
NHS Big Data Intelligence on Blockchain Applications 201
For the 10 000 mining machine, the computation = 600 × 1000 × 1.4 × 1013 ;
which is about 8 × 1019 ;
that meant every 10 min, you could calculate 8 × 1.4 × 1019 times, in terms of
produce a new block.
If n = 66, the calculation probability is (2–1 )66 , that is an average computation =
(2) times, which is about 8 × 1019 .
66
By trace back to the whole Blockchain record to check if that transaction was reason-
able, that broadcast was accepted and confirmed/verified by this Blockchain network
then packed to a new block in that Blockchain. But, if which transaction were not
reasonable, that broadcast would be refused and not be verified by this Blockchain
networkintegrity. Once the block where the transaction was embedded in is deep
enough in the chain), the record becomes immutable, that is, it is impossible to
change the record without being detected.
3.2.1 Verification
3.2.2 Anti-Fake
The longest chain was the chosen one principle of Blockchain. This principle ensured
any modification attempt not feasible. When people in the network received a new
block, wait for the next new block. If a perpetrator tried to delete a transaction, change
this record, the anti-alteration method was, although everybody took the first received
new block into account, to carry on their work, until a newer block connected to the
Blockchain, that made the newer chain was longer than the others. So when the newer
one broadcasting to the network, everybody in this network known this was longest
chain. The others would stand in that line again, then continued the mining work.
The one did not have the longest chain failed that block, as the whole network was
looking for the longest chain. It was not realistic that you could success to against
all of the whole Blockchain network. Because everybody accepted the longest chain
principle in common sense, if a perpetrator wanted to change any record, he needed to
create another longer chain than the existed one that the whole world miner created,
which was impossible. (Except the perpetrator’s mining was better than everybody
in the world. Since if your chain was not longer than other, your alteration chain
wont be accepted.) The longer the time passed, the more robust of the chain (Feng
et al. 2021). Generally speaking, if after 6 blocks, your transaction record would be
anti-alteration.
A ten-year NHS Spine contract awarded to BT involved developing systems and soft-
ware to support more than 899,000 registered users. It has made transformational
healthcare applications available to approximately 1.3 million NHS healthcare staff
across England, providing care to circa 50 million UK citizens. In managing overall
software delivery lifecycle processes, BT used Rational Unified Process/Systems
Engineering (RUP/SE) processes. These were modified to create a methodology
that attained Capability Maturity Model Integration (CMMI) Level-3 accreditation.
204 X. Feng et al.
5 Evaluation
5.1 Discussion
When working with NHS, a security testing is required. Cyber security testing using
Blockchain technology was in demand, in order to ensure the NHS health system
guarantee patients safe and secure. During the past years Blockchain technology was
deployed in the health sectors to improve authenticity and security using the features
of decentralized structure and the distributed ledgers and the proof of work feature and
the hash function which make it very difficult to amend (Yang et al. 2017). Although
there is not many projects of Blockchain in the NHS, but during the COVID-19
pandemic, according to Reuters, two UK Hospital used the technology to trace and
transfer COVID-19 vaccine to monitor the supply and the sensitive temperature the
drug need to be save in (Meirino et al. 2019). In UK NHS, before accessing a medical
record, the user needed to be authenticated, using keyless infrastructure signature,
which is a mechanism to take protection of the integrity of the NHS record into
account for make sure it was not modified in any sense. The proposed implementation
and experimental finding results could be shown next (Hussein 2021).
Blockchain used hashing and publishing mechanism. Blockchain use keyless
signature infrastructure which consists of a sets aggregation server which create
a global hash tree. The Blockchain System will be deployed and tested by NHS,
and then we would run the above attacks and obtaining data and logs to be analyzed
(Laurence 2019).
NHS Big Data Intelligence on Blockchain Applications 205
Barclays was a second-largest bank in UK, with branches in Europe. Barclays and
companies used Blockchain technology. In reality, they were using Blockchain tech-
nology for streamlining fund transfers and Know-Your-Customer processes. To an
extent, they even filed for patents against these two features. They will shift from
their traditional paper-based records to fully digital and decentralized Vault platform.
Thus, their investors could now finally track their money in real-time, although with
latency.
Pfizer was another one of the large companies using Blockchain technology. Biogen
and Pfizer led organization Clinical Supply Blockchain Working Group (CSBWG),
just completed proof of concept for tracking records and managing the digital inven-
tory of pharmaceutical products. In reality, the group also comes with other players
such as GlaxoSmithKline, Merck, AstraZeneca, and Deloitte.
5.3 Evaluation
All these applications technology were Blockchain technology, which recorded all
the historic transactions and open to network transparently. Blockchain ledgers
contained time, transactions, and related activities. Although excessive delay in
206 X. Feng et al.
reaching consensus and limited throughput still outstanding, scientists have been
made effort to improve. Blockchain technology still grown in exponential increase
globally. To research Blockchain technology and master the trends impact to big data
intelligence for smart applications improve security in the future.
Although excessive delay in reaching consensus and limited throughput still
outstanding, scientists have been made effort to improve. Recently, Zhao et al.
proposed a method had hierarchical processing and logging of potentially large
amount of sensing big data with the Blockchain technology, which could drasti-
cally address the issue of limited throughput in Blockchain for IoT of the current
distributed ledgers (Zhao et al. 2020).
6.1 Summary
In this chapter, we had explored the main Blockchain application for NHS outstanding
security problems. Since this dealt with different operating systems and many
different fields, security issue was quite sophisticated in Blockchain technology appli-
cations. Blockchain technology possessed a number of advantages, such as, peer-to-
peer network; cryptography; chaining-based data structure; immutability and proof of
work that made it was well accepted in the society. Nevertheless, the majority security
problem in Blockchain application design, testing and management threat appear-
ance features had been discussed. The security potential was analyzed. Remedy and
problem-solving had been explored respectively. An investigation Blockchain tech-
nology utilization, testing for decentralized systems to date had been searched. The
attacking on authentication system also was looked at. Blockchain was a decentral-
ized, distribution multi-recorded system technology. Its application on AI, process
data for health system to realizing cyber security would be achievable and worthwhile.
Feng et al. (2021) had pointed out, because the advantages of public Blockchain with
cryptography, this technology could also serve for cyberstalking and cyberbullying
digital forensics evidence data storage security for law enforcement and society.
Blockchain technology might change IoT big data intelligence for smart application.
Currently IoT device were managed by supply companies. In the future, centralized
management could not support all the devices properly. Blockchain technology could
help IoT devices direct secure and reliable to communicating on Blockchain network
and avoid intermediary central control disadvantages. There was one more point, the
main difference between Blockchain and distributed ledger would be the internal
structure system. All Blockchain technologies were a form of the distributed ledger,
but not all the distributed ledger systems were Blockchain.
NHS Big Data Intelligence on Blockchain Applications 207
Acknowledgements We sincerely thank the anonymous reviewers and editors for their helpful
comments and suggestions on earlier draft of the manuscript.
References
Blockchain Technology, ACM Digital Library, March 2020 (2020), pp. 52–56. https://doi.org/10.
1145/3390566.3391672. Accessed 4 Mar 2021
IBM, IBM Blockchain. www.ibm.com/blockchain: www.ibm.com/blockchain. (2020). Accessed
16 Feb 2021
X. Feng, et al., A systematic approach of impact of GDPR in PII and privacy. J. IJESI 5 (2021).
ISSN 2319-6726
H. Yang, et al., A Blockchain-based approach to the secure sharing of healthcare data. NISK 2017
(2017). ISSN 1894-7735. http://www.nik.no/. Accessed 16 Feb 2021
IBM, IBM Blockchain (2020). www.ibm.com/blockchain: www.ibm.com/blockchain. Accessed 16
Feb 2021
Steemit, How smart contract works. Steem, White paper (2017). https://steem.com
X. Feng, et al., Artificial intelligence and Blockchain for future cyber security application, in
Accepted IEEE 5th International Workshop ACE-2021 (2021).
C. Alexakos, et al., Enabling digital forensics readiness for internet of vehicles, in 23rd Euro WP
on Transportation Meeting, EWGT 2020, Sep 2020, Paphos, Cyprus, Transportation Research
Procedia, vol. 52 (2021) pp. 339–346
G. Iredale, Blockchain for healthcare: use cases and applications (2019). https://101blockchains.
com/blockchain-for-healthcare. Accessed 25 June 2021, 4 Mar 2021
A. Takyar, “Use Cases of Blockchain in Healthcare” Leeway Hertz (2016). https://www.leeway
hertz.com/blockchain-in-healthcare. 15 June 2021
S. De Angelis, Assessing security and performances of consensus algorithms for permissioned
Blockchains. Southampton university (2018).
E. Ali et al., A semi-oriented radial measure for measuring the efficiency of decision making units
with negative data, using DEA. Eur. J. Oper. Res. 200(1), 297 (2010)
X. Feng, et al., Artificial intelligence and cyber security strategy, in IEEE International CyberSciTech
Conference, Athabasca University, Canada (2020)
J. Weisenthal, What’d you miss? Bloomberg Bundesbank together with Deutsche Boerse ledger
technology potential cheaper and faster than current settlement mechanisms. Bloomberg (2019).
https://www.bloomberg.com/podcasts/material_world1. Accessed 4 June 2021
IBM, Blockchain and the healthcare industry: how to get started. IBM (nd). https://www.ibm.com/
blogs/blockchain/category/blockchain-education/blockchain-explained/. Accessed 25 Feb 2021
BT, Smart city applications on the Blockchain. British Telecom. Springer-professional (nd).
Accessed 25 Feb 2021
Udemy, Blockchain Training (2021). http://www.udemy.com. Accessed 25 Feb 2021
NHS, Spine case study. NHS UK (2020). https://business.bt.com/solutions/resources/nhs-spine/.
Accessed 12 Feb 2021
C. Lewis, Can Blockchain help to deliver the NHS long-term plan? retrieved from the journal of
mhealth (2019). https://thejournalofmhealth.com/can-blockchain-help-to-deliver-the-nhs-long-
term-plan/. Accessed 10 Feb 2021
E. Liu, X. Feng, Trustworthiness in the patient centered health care system. Communications in
Computer and Information Science, vol. 426 (Springer, Berlin, 2014)
T. Laurence, Introduction to Blockchain technology the many faces of Blockchain technology in
the 21st century. Van Hearen Publishing (2019). ISBN: 978-9401804998
M.J. Meirino et al., Blockchain technology applications: a literature review. Braz. J. Oper. Prod.
Manag. 16(4), 672–684 (2019)
K. Hussein, Physical Blockchain design considerations, in Prepared for the 6th ACE WS (2021).
E. Eze, S. Zhang, X. Feng X et al., Mobile computing and IoT: radio spectrum requirement for
timely and reliable message delivery over internet of vehicles (Springer, Berlin, 2020)
V.J.B. Dhillon, Blockchain in healthcare: innovations that empower patients. Connect professionals
and improve care. Pristine Publishing (2019).
Depression Detection from Social Media
Using Twitter’s Tweet
Abstract Social media is using rapidly for expressing user’s opinions and feelings.
At present, ‘Depression’ is propagating in a cursory way which is the reason for the
increasing rate of suicides even more. Recently, the status on social media can give the
hints of user’s mental state along with the situation and activities that are happening
with them. Different research works had done for perceiving the mental health of any
user through social media which has impressive impact. That is done by analysing
expressed opinions, images, sentiments, linguistic style and other activities. Variant
studies had proposed variant methods or intelligent systems for detecting depression
or state of mental health through social media posts. In this chapter, the posted
tweets from users on Twitter will be considered in the sake of detecting depressions.
Six machine learning approaches named Multinomial Naive Bayes, Support Vector
Classifier, Decision Tree, Random Forest, K-Nearest Neighbor, Logistic Regression
had used to differentiate the depressed and non-depressed users. Finally, Support
Vector Classifier outperforms among all of the investigated and evaluated techniques
with 79.90% accuracy, 75.73% precision, 77.53% recall and 76.61% f1-factor. This
study can be a basement for the developers of intelligent systems in the area of users
mental conditions detection.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 209
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_9
210 R. J. Lia et al.
1 Introduction
Depression is a mental disorder which can impair many facets of human life. Depres-
sion has profound and varied impacts on both physical and mental health (Deshpande
and Rao 2017; Coello-Guilarte et al. 2019) and sometimes occurs anxiety or other
psychological disorders (Tadesse et al. 2019) such as mood disruption, uncertainty,
loss of interest, tiredness, and physical issues (Mustafa et al. 2020). It is also known
as a leading cause of disability and is a major contributor to the overall global burden
of disease (Burdisso et al. 2019; Cacheda et al. 2019). According to the World Health
Organization(WHO), more than 264 million people of all ages suffer from depression
globally and 70% of the patients would not even consult doctors, which may sustain
their condition to next echelons (AlSagri and Ykhlef 2020; Tong et al. 2019) and in
many countries, even fewer than 10% depressed people receive treatment (Gui et al.
2019). It is also entitled the second leading cause of death among 15–29-year-old’s
people (Tao et al. 2016; Singh et al. 2019; Yang et al. 2020; Wu et al. 2020). Depres-
sion is a conventional cause of committing suicide which is increased day by day
very rapidly. According to the recent report of World Health Organization(WHO),
the number of died people due to suicide is near to 800000 in every year, which means
one person in every 40 s. People who are facing less interaction with friends and fam-
ily use social media as a rule to share their feelings and opinions (Lin et al. 2020).
Often, the users express the unuttered feelings for feeling good which drive them on
the way of suicide. Some specific words are usually used by the depressed people for
expressing their mental stress such as pain, hurt or any other negative words. Through
this, the mental conditions of many depressed people had been revealed on different
social media (Leiva and Freire 2017). A humongous quantity of data is generated
in every second through various social media platforms like Facebook, Twitter, etc.
These all data from different platforms have a lot of relevant information for doing
behaviour analysis. Twitter’s tweet will be used in this research as Twitter is one of
the most visited social networking sites where an average of 58 million tweets are
generated per day (Lin et al. 2020) and there are over 271 million monthly active
users in Twitter (Nambisan et al. 2015). So, it becomes easy to get data for analysing
depression with tweets. There are eight basic emotions as features from Twitter posts
which helps to identify the type of any tweets even easily (Chen et al. 2018).
In this study, the main goal is to find out which users are depressed. Through
detecting those depressed users, there is a chance of decreasing the risk of suicide
and save them.
The rest of the chapter is organized as follows: This study has tried to find out
depressed twitter users through their posted tweets using six machine learning tech-
niques. In Sect. 2, have reviewed some related works of depression detection with
a table and description. In Sect. 3, the methodology of this chapter has shown in
Depression Detection from Social Media Using Twitters Tweet 211
figure and described elaborately. The experimental results have shown in Sect. 4 and
comparing the results the best algorithm for this topic has also declared. Section 5
mentioned about the limitations as well as talked about future research on it. Finally,
this paper ends with Sect. 6 with the conclusions.
2 Literature Review
Several research works had been performed till now for detecting depression through
Twitter’s tweets as depression is increasing in short order. Several research works
have been completed using deep learning algorithms (Guntuku et al. 2017) and
several worked with machine learning algorithms for getting the best outcome. Large-
scale passive monitoring of social media used for identifying depressed users or
any at-risk individuals and Linear Regression and Support Vector Machine(SVM)
algorithms were used for this research work (Almeida et al. 2017). A martingale
framework which is widely used for detection of changes in data stream settings
was implemented in the research work of Vioules et al. (2018). Several algorithms
such as Natural Language Processing (NLP), Naive Bayes (NB), Sequential Minimal
Optimization (SMO) with a poly kernel, C4.5 decision tree (J48), nearest neighbor
classifier (IB1), multinomial logistic regression, rule induction (Jrip), Random Forest,
and SMO with a Pearson VII universal kernel function (PUK) were used in the study
(Vioules et al. 2018).
Logistic Model Tree (LMT), Sequential Minimal Optimization (SMO) and Ran-
dom Forests were used in O’dea et al. (2015) and a multipronged approach was
designed in this paper that combines results obtained from both Information Retrieval
(IR) and Supervised Learning (SL) based systems for detecting depression. Text pro-
cessing had been used and explained elaborately in several work (Sohan et al. 2018;
Rahman et al. 2020, 2018), [?] and this study had gained knowledge from those.
Methodology used in Gaikar et al. (2019) this study was an automated computer
classifier that could replicate the accuracy of the human coders and this methodol-
ogy was implemented in this paper. Logistic Regression (LR) and Support Vector
Machine (SVM) were the techniques used in Gaikar et al. (2019) for the mentioned
methodology.
SVM with NB features (NBSVM) was applied for detecting any depression related
words or phrases from Tweets and also classified the type of depression (Lin et al.
2020). A system dubbed SenseMood was used in the paper of Morales et al. (2017).
SenseMood demonstrated that the users with depression can be efficiently detected
and analyzed by using the proposed system. A novel approach was proposed for
identifying users with or at risk of depression by incorporating measures of eight
basic emotions features from Twitter posts over time, including a temporal analysis
of these features and this approach provided the psychological state successfully
(Chen et al. 2018).
The proposed system of Lin et al. (2020) was a hybrid approach of Support vector
machine and Naïve Bayes classifier which works well not only with shorter snippets
212 R. J. Lia et al.
but also with longer snippets. The author of Deshpande and Rao (2017) had also used
the Support vector machine and Naive-Bayes classifier for depression detection and
got the result by presenting the primary classification metrics including F1-score,
accuracy and confusion matrix.
Information about some of the related literature are explained in Table 1. From
Table 1. We can acknowledge that most of the literature have worked with very less
number of algorithms in order to detect depression from social media.
In this study, we will concentrate on the used words in any tweet by any user and
six machine learning algorithms will be applied on the collected data for depression
detection in order to get the best result.
Table 1 (continued)
Author Title Year Methodology Algorithm
Stephen Wan b, Detecting 2015 Design and Support Vector
Philip J. suicidality implement an Machine(SVM),
Batterham c, on Twitter automated Logistic
Alison L. computer Regression(LR)
Calearc, Cecile classifier that could
Paris b, Helen replicate the
Christensen accuracy
of the human
coders
Jayesh Chavan, Depression 2019 Detect any SVM with NB
Kunal Indore, Detection and depression features
Mrunal Gaikar, Prevention related words or (NBSVM)
Rajashree System by phrases from
Shedge Analysing Tweets
Tweets and also classify
the
type of depression,
if detected.
Chenhao Lin, SenseMood: 2020 Design a system CNN, Bert
Pengwei Hu, Depression dubbed
Hui Su, Detection on SenseMood to
Shaochun Li Social Media demonstrate that
Jing Mei,Jie the users
Zhou, Henry with depression
Leung can be
efficiently detected
and analyzed by
using
proposed system
Xuetong Chen, What about 2018 Proposed a novel Naıve Bayes,
Martin D. Mood Swings? approach Sequential
Sykora, Thomas Identifying for identifying Minimal
W. Jackson, Depression on users with Optimization
Suzanne Elayan Twitter with or at risk of (SMO)
Temporal depression by
Measures of incorporating
Emotions measures of
eight basic
emotions as
features from
Twitter posts
over time,
including a
temporal analysis
of these
features
214 R. J. Lia et al.
3 Research Methodology
The proposed research methodology of this study provides the complete knowledge
about this research work. Through this framework, all detailed works of this study
had been shown clearly step by step which ensures that future works based on this
topic can take proper ideas from this work. From data collection to prediction’s result
all are there in this framework. There are several steps in the research methodology
of this study. They are:
1. Data Collection and Preparation
2. Data Preprocessing
3. Word Analysis
4. Tokenization
5. Extract Features
6. Depression Detection
7. Output
The purpose of this paper is to detect if any Twitter user is in depression or not
just like (Kumar et al. 2019) through following the proposed methodology in Fig. 1.
In the first component of this methodology, this study mainly focused on collecting
data and preparing the collected data which are the most important for starting the
work of this research. After preparing the required data, this study next did data
processing for making the data all ready for the consisted dataset. However, there
are several steps and processes to follow for preprocessing the data which will be
explained later. Then came word analysis which gave an idea of most used words
in the dataset. Next component named tokenization breaks the raw data into small
tokens consisting of words and sentences. For feature extraction of this study, TF-IDF
vectorization was used. Then splitted the data into testing data and training data for
further experiments. This study’s next component name was depression detection in
which several machine learning algorithms had been used for getting better result of
detecting depressed users of Twitter which was the main goal of this work alike (Islam
et al. 2018; Ziwei and Chua 2019). Then the last component completed with giving
the results of different algorithms and gave a clear conclusion of which algorithm is
the best for detecting depression.
Dataset collection and preparation is very essential for further work progression
of any research. In the research of Gaikar et al. (2019), all the tweets from 18th
February,2014 to 23rd April, 2014 which used English words or phrases that are
consistent with the vernacular of suicidal ideation were collected. In the training
phase of Lin et al. (2020), the labeled Dataset from kaggle were collected and used.
A total of 10,000 Tweets were collected from Twitter API for generating the training
Depression Detection from Social Media Using Twitters Tweet 215
and test dataset by the author of Deshpande and Rao (2017). The dataset used in
Tadesse et al. (2019) contained depression-indicative posts of 1293 and standard
posts of 548. The data used in Tao et al. (2016); Kumar et al. (2019) were also
collected from Twitter. 55,810 tweets from Twitter between Oct 2018 and Feb 2019
were collected for the research work of Singh et al. (2019). Following their shadow,
we also collected data from Twitter and will prepare the collected data for future
uses. The dataset used in this study was collected from kaggle.com with the title of
training.1600000.processed.noemoticon consisting 1600000 million tweets from 6th
April,2020 to 16th June,2020 and all of these datas were in English. Two different
datasets were splitted with 100000 datas in each dataset with two different names
from the collected dataset. The first splitted dataset’s data limitations were from
index 0 to index 100000 and datas from index 800000 to 900000 million were kept
in the second dataset. A new dataset had created concating those two splitted dataset
named as PrepareData which had 200000 datas. The new dataset was prepared for
216 R. J. Lia et al.
testing the work with new dataset for getting new or experimental output. So, from
collecting these 200000 tweets though the new dataset, this study had done further
research.
Preprocessing of data just started with the finalized dataset. Preprocessing was needed
and crucial before taking any next component as the collected data were not processed
at all. Lots of noises, missing values and unstable format are always contained in any
data. Before applying any algorithms or next components, this study had cleaned
the data and made it suitable for further work. Moreover, it’s impossible to apply
machine learning algorithms on any dataset without cleaning and formatting the
dataset. Besides this, the accuracy and efficiency of a machine learning model are
also increased with data preprocessing.
In the study of Lin et al. (2020), several preprocessing types had performen called
as Conversion from Uppercase to Lowercase, Apostrophe Lookup, Punctuation Han-
dling, Removal of URLs, Stop word removal and Lemmatization for removing noise
from the data and cleaning and simplifying the data. In the case of Tadesse et al.
(2019), all the URLs, punctuations and stop words were removed which could lead
to erratic results if stayed ignored. Before splitting the dataset of Islam et al. (2018),
the dataset was cleaned by removing quotes, extra spacing and symbols. The process
of cleaning data includes removing numeric and empty texts, URLs, mentions, hash-
tags, non-ASCII characters, stop-words† and punctuations in Rahman et al. (2020)
before predictions. Sixteen preprocessing techniques were experimented in Gun-
tuku et al. (2017) for giving ideas of choosing the right preprocessing technique for
cleaning the dataset. The study (Kumar et al. 2019) also removed different symbols
like exclamation marks, punctuations, digits, special characters etc. for cleaning the
dataset. However, three types of preprocessing had performed in this study’s dataset
named as:
1. Convert all texts into lowercase
2. Remove all punctuations
3. Remove all stopwords.
In the finalized dataset, some texts were in lowercase and at the same time some
were in uppercase. So, it creates confusion at the time of applying any algorithm for
any prediction. All the texts of the finalized dataset were converted into lowercase
in order to solve this problem like (Tao et al. 2016; Kumar et al. 2019) also did. The
study (Stephen and Prabu 2019) showed an example of the full process of converting
texts into lowercase for explaining.
There were lots of punctuations like comma, semicolon, colon, dash, hyphen in
the dataset which were not only unwanted but also problematic for this research as
punctuations look like just another character to Python. For removing punctuation,
the used programming language Python was shown what the punctuations look like
Depression Detection from Social Media Using Twitters Tweet 217
by accomplishing using the String package in Python (Singh et al. 2019). So, all
these unwanted punctuations were removed from the dataset.
Lastly, this dataset contained lots of stopwords like is, at, which, me, you etc.
which are not meaningful and all these stopwords should ignore as these words cas
create problems at the time of any word which is meaningful and expressed feeling.
So, all these stop words were removed from the dataset. This is done to reduce the
work for our algorithms (Singh et al. 2019). In order to remove stop-words from the
tweet, Deshpande and Rao (2017) used Nltk library has a set of stop-words which
can be used as a reference. The clean and formatted dataset had been obtained by
performing all these three types of preprocessing and had applied the clean dataset
further.
As the dataset of this study had been prepared and cleaned, so now it’s possible to
do word analysis of this dataset without any problem. However, if word analysis had
been performed before data preprocessing, it might give different output from the
output of word analysis of cleaned and formatted dataset as any stop word might take
place on the position of any important word. Word analysis is called the process of
learning more about any word in a meaningful way. Within word analysis, the most
used words were shown in large words and less used words were shown in small
words which can be witnessed in Fig. 1. So, though word analysis, the idea of the
most and less used words were cleared in this research work.
3.4 Tokenization
Tokenization breaks the raw data into small tokens consisting of words and sentences.
The tokens help easily in the sake of understanding the context or model developing.
There are two types of tokenization. One is called word tokenization and the other
is sentence tokenization. If text is splitted into words with any tokenization tech-
nique, then that is called word tokenization and if text is splitted into sentences with
any tokenization technique, then that is called sentence tokenization. In the case of
Deshpande and Rao (2017), the first column of the csv file containing the tweet was
extracted and later converted into individual tokens. In the research of Singh et al.
(2019), an example was shown for clearing the tokenization’s concept. In order to
filter the words, symbols and other elements called tokens, Tao et al. (2016) used
the TreebankWordTokenizer of Python Natural Language Toolkit (NLTK). However,
word tokenization was used in this study. There were 593588 numbers of tokens after
tokenization performed on the dataset of this research.
218 R. J. Lia et al.
Feature extraction is a general term for a method for developing blends of the factors
to get around these issues while depicting the information with adequate exactness.
The feature extraction method extracts the aspect (adjective) from the improved
dataset and the adjective is used to determine the polarity of a sentence. Unigram
model extracts the adjective and segregates it Deshpande and Rao (2017).
After feature extraction, the full dataset was splitted into Testing Data and Training
Data. 30% data were kept in Testing Data and rest 70% data were in Training Data.
After splitting the data into testing and training, now it’s time to work on those data.
The 30% testing data had been used in order to get accuracy, precision, recall, f1-
Depression Detection from Social Media Using Twitters Tweet 219
3.7 Output
Several machine learning algorithms were applied for depression detection from the
tweets in order to find out how many users are in depression and how many users
are not. After applying all those machine learning algorithms, this study saw that
different algorithms gave different accuracy. However, in spite of some poor output
of some algorithms, the bright side was that though that, this research work came to
a conclusion of which algorithm is the best for depression detection from tweets to
find out if any user is in depression or not in AlSagri and Ykhlef (2020), the study got
70% accuracy from Support Vector Machine and 71% accuracy from Naive Bayes.
79% accuracy was gained in Deshpande and Rao (2017) for depression detection.
However, in this study, it will be shown which algorithm is best in terms of pro-
viding accuracy, precision, recall and f1-score for getting the best machine learning
algorithm for depression detection.
There are several metrics which can be used in order to evaluate the machine learning
algorithms. However, for choosing any performance metric, the basic things need to
be remembered are the performance of the machine learning algorithms, dependency,
and importance of various characteristics which will influence the result of the algo-
rithms. Performance metrics used for this study were Confusion matrix, Accuracy,
Precision, Recall and F1-score.
Confusion Matrix
Confusion matrix is the most used and easiest way for measuring the performance
with two or more types of classes in any classification problem. Basically the con-
fusion matrix is a table which contains two dimensions called actual and predicted
which helps to get any prediction. There are four terms in the confusion matrix named
True Positives(TP), True Negatives (TN), False Positives (FP), False Negatives (FN).
F1-Score Formula=2*(Precision*Recall)/(Precision+Recall)
So, this study used all these performance metrics on all the machine learning
algorithms which were used in this research.
Six types of machine learning algorithms were used in this research work for finding
out which algorithm provides the best performance for detecting depression with
tweets. The finalized and preprocessed dataset was used in those algorithms after
splitting that dataset into 30% in testing and 70% in training data. On the basis of test
data, the predictions were performed and found out the accuracy, precision, recall
and f1-score for all of the machine learning algorithms. All the results of accuracy,
precision, recall and f1-score for all applied machine learning algorithms were judged
and compared among each other on account of declaring the best machine learning
algorithm.
Here goes the accuracy, precision, recall and f1-score of all applied machine
learning techniques-
In Fig. 2., there is a comparison graph for accuracy among all the applied algo-
rithms. From the graph, it’s clear that Support Vector Classifier had the best accuracy
among all the used machine learning algorithms which was 76.90% . At the same
Depression Detection from Social Media Using Twitters Tweet 221
time, Logistic Regression had only 0.38% less accuracy than Support Vector Classi-
fier. However, without Decision Tree, all other algorithms had almost approximate
accuracy as Decision Tree had the lowest accuracy. So, Fig. 2. indicated that among
all used machine learning algorithms in this research, Support Vector Classifier pro-
vided the highest accuracy where Decision Tree provided the lowest accuracy.
According to Fig. 3., the highest precision provided machine learning algorithm
was Multinomial Naive Bayes for this research work. Multinomial Naive Bayes
gave 79.96% precision which was quite effective for depression detection. On one
side, Multinomial Naive Bayes had the highest precision, while on the other side,
K-Nearest Neighbor had the lowest precision. Rest of the applied machine learning
algorithms provided between 72–76%precision which was also standard.
Figure 4. indicates the recall rate of all used machine learning algorithms through
a graph with data level. Through the graph, it shows that Support Vector Classifier
(SVC) resulted in the highest recall rate and K-Nearest Neighbor resulted in the
lowest recall rate. The recall rates of rest algorithms are also shown in Fig. 4.
F1-score for all the machine learning algorithms which were used in this study
had been picturized beautifully in Fig. 5. In the Fig. 5., Support Vector Classifier,
Decision Tree, Logistic Regression, Multinomial Naive Bayes, K-Nearest Neighbor
and Random Forest had gradually 76.61, 70.16, 76.09, 76.04, 28.12 and 75.32%
F1-Score rate. So, 76.61% is the highest F1-score which is Support Vector Classifier
and 28.12% is the lowest F1-score rate which is K-Nearest Neighbor.
In the previous component, accuracy, precision, recall and f1-factor of all applied
algorithms had been described elaborately. Though that description, it had been
stated that Support Vector Classifier (SVC) gave the highest accuracy, Multinomial
Naive Bayes(MNB) provided the best precision and Support Vector Classifier (SVC)
provided the highest recall and f1-factor both.
Depression Detection from Social Media Using Twitters Tweet 223
All applied algorithms provided accuracy, precision, recall and f1-factor had
shown in Table 2. Though Table 2., it’s clear that Support Vector Classifier (SVC)
had the highest accuracy, recall and f1-factor. At the same way, Multinomial Naive
Bayes(MNB) had the highest precision. However the precision rate of Support
Vect(SVC) is little less than Multinomial Naive Bayes(MNB) and Random For-
est(RF). So, Support Vector Classifier (SVC) is the best algorithm for depression
detection according to this thesis work (Fig. 6).
224 R. J. Lia et al.
This research had only worked with the data collected from Twitter. Some studies
had worked on anxious depression prediction (Rahman et al. 2020), depression level
detection. However, this study worked on depression detection based on Twitter’s
tweet. So, by this it’s clear that this study only worked for detecting if any Twitter user
is depressed or not but not for all social media users. This research also mentioned
that this work was only for English language. Moreover, this thesis only detected
depression of any user thought the tweet but couldn’t provide the status or level of
that depressed user. These are some limitations for this study.
In future, this study will be worked for identifying the types of depression in
depth so that it will be able to inform about the depressed person’s depression state.
Moreover, it will also work with the data from all other social media and in many
more languages.
6 Conclusions
This study presented a generic depression detection model with the use of six machine
learning algorithms. The objective of this study is depression detection from any
tweets which represent depression though any depressive word and reduce depression
& suicides out of depression. In this work, a methodology has been followed and
through following this methodology, any tool can be built for detecting depression
easily in future. Any other tweet which is pre-processed, it can be told if the data
indicates depression or not by following the methodology of this study. This study
also resulted that among all the applied machine learning algorithms, Support Vector
Classifier(SVC) had provided the best result for depression detection which means
if any other study wants to work on depression detection based on machine learning
algorithms, then that study can use Support Vector Classifier(SVC) without any doubt
for getting the best result for depression detection.
References
M. Deshpande, V. Rao, Depression detection using emotion artificial intelligence, in 2017 Interna-
tional Conference on Intelligent Sustainable Systems (ICISS) (IEEE, 2017), pp. 858–862
L. Coello-Guilarte, R.M. Ortega-Mendoza, L. Villaseñor-Pineda, M. Montes-y-Gómez, Crosslin-
gual depression detection in Twitter using Bilingual word alignments, in International Conference
of the Cross-Language Evaluation Forum for European Languages (Springer, Cham, 2019), pp.
49–61
M.M. Tadesse, H. Lin, B. Xu, L. Yang, Detection of depression-related posts in reddit social media
forum. IEEE Access 7, 44883–44893 (2019)
Depression Detection from Social Media Using Twitters Tweet 225
R.U. Mustafa, N. Ashraf, F.S. Ahmed, J. Ferzund, B. Shahzad, A. Gelbukh, A multiclass depres-
sion detection in social media based on sentiment analysis, in 17th International Conference on
Information Technology–New Generations (ITNG 2020) (Springer, Cham, 2020), pp. 659–662
S.G. Burdisso, M. Errecalde, M. Montes-y-Gómez, A text classification framework for simple and
effective early depression detection over social media streams. Expert. Syst. Appl. 133, 182–197
(2019)
F. Cacheda, D. Fernandez, F.J. Novoa, V. Carneiro, Early detection of depression: social network
analysis and random forest techniques. J. Med. Internet Res. 21(6), e12554 (2019)
H.S. AlSagri, M. Ykhlef, Machine learning-based approach for depression detection in twitter using
content and activity features. IEICE Trans. Inf. Syst. 103(8), 1825–1832 (2020)
L. Tong, Q. Zhang, A. Sadka, L. Li, H. Zhou, Inverse boosting pruning trees for depression detection
on Twitter (2019). arXiv:1906.00398
T. Gui, L. Zhu, Q. Zhang, M. Peng, X. Zhou, K. Ding, Z. Chen, Cooperative multimodal approach to
depression detection in Twitter, in Proceedings of the AAAI Conference on Artificial Intelligence,
vol. 33, no. 01 (2019), pp. 110–117
X. Tao, X. Zhou, J. Zhang, J. Yong, Sentiment analysis for depression detection on social networks,
in International Conference on Advanced Data Mining and Applications (Springer, Cham, 2016),
pp. 807–810
R. Singh, J. Du, Y. Zhang, H. Wang, Y. Miao, O.A. Sianaki, A. Ulhaq, A framework for early
detection of antisocial behavior on Twitter using natural language processing, in Conference on
Complex, Intelligent, and Software Intensive Systems (Springer, Cham, 2019), pp. 484–495
X. Yang, R. McEwen, L.R. Ong, M. Zihayat, A big data analytics framework for detecting user-level
depression from social networks. Int. J. Inf. Manag. 54, 102141 (2020)
M.Y. Wu, C.Y. Shen, E.T. Wang, A.L. Chen, A deep architecture for depression detection using
posting, behavior, and living environment data. J. Intell. Inf. Syst. 54(2), 225–244 (2020)
M.J. Vioules, B. Moulahi, J. Aé, S. Bringay, Detection of suicide-related posts in Twitter data
streams. IBM J. Res. Dev. 62(1), 1–7 (2018)
C. Lin, P. Hu, H. Su, S. Li, J. Mei, J. Zhou, H. Leung, Sensemood: depression detection on social
media, in Proceedings of the 2020 International Conference on Multimedia Retrieval (2020), pp.
407–411
V. Leiva, A. Freire, Towards suicide prevention: early detection of depression on social media, in
International Conference on Internet Science (Springer, Cham, 2017), pp. 428–436
P. Nambisan, Z. Luo, A. Kapoor, T.B. Patrick, R.A. Cisler, Social media, big data, and public health
informatics: ruminating behavior of depression revealed through twitter, in 2015 48th Hawaii
International Conference on System Sciences (IEEE, 2015), pp. 2906–2913
X. Chen, M.D. Sykora, T.W. Jackson, S. Elayan, What about mood swings: identifying depression
on twitter with temporal measures of emotions, in Companion Proceedings of the The Web
Conference 2018 (2018), pp. 1653–1660
S.C. Guntuku, D.B. Yaden, M.L. Kern, L.H. Ungar, J.C. Eichstaedt, Detecting depression and
mental illness on social media: an integrative review. Curr. Opin. Behav. Sci. 18, 43–49 (2017)
H. Almeida, A. Briand, M.J. Meurs, Detecting early risk of depression from social media user-
generated content, in CLEF (Working Notes) (2017)
B. O’dea, S. Wan, P.J. Batterham, A.L. Calear, C. Paris, H. Christensen, Detecting suicidality on
Twitter. Internet Interv. 2(2), 183–188 (2015)
M. Gaikar, J. Chavan, K. Indore, R. Shedge, Depression detection and prevention system by
analysing tweets, in Proceedings 2019: Conference on Technologies for Future Cities (CTFC)
(2019)
M. Morales, S. Scherer, R. Levitan, A cross-modal review of indicators for depression detection
systems, in Proceedings of the Fourth Workshop on Computational Linguistics and Clinical
Psychology–From Linguistic Signal to Clinical Reality (2017), pp. 1–12
C.S.A. Razak, M.A. Zulkarnain, S.H. Ab Hamid, N.B. Anuar, M.Z. Jali, H. Meon, Tweep: a system
development to detect depression in twitter posts, in Computational Science and Technology
(Springer, Singapore, 2020), pp. 543–552
226 R. J. Lia et al.
A. Kumar, A. Sharma, A. Arora, Anxious depression prediction in real-time social data, in Interna-
tional Conference on Advances in Engineering Science Management & Technology (ICAESMT)-
2019 (Uttaranchal University, Dehradun, India, 2019)
M.R. Islam, M.A. Kabir, A. Ahmed, A.R.M. Kamal, H. Wang, A. Ulhaq, Depression detection from
social network data using machine learning techniques. Health Inf. Sci. Syst. 6(1), 1–12 (2018)
B.Y. Ziwei, H.N. Chua, An application for classifying depression in tweets, in Proceedings of the
2nd International Conference on Computing and Big Data (2019), pp. 37–41
S. Symeonidis, D. Effrosynidis, A. Arampatzis, A comparative evaluation of pre-processing tech-
niques and their interactions for twitter sentiment analysis. Expert. Syst. Appl. 110, 298–310
(2018)
J.J. Stephen, P. Prabu, Detecting the magnitude of depression in twitter users using sentiment
analysis. Int. J. Electr. Comput. Eng. 9(4), 3247 (2019)
S.S.M.M. Rahman, K.B.M.B. Biplob, M.H. Rahman, K. Sarker, T. Islam, An investigation and
evaluation of N-Gram, TF-IDF and ensemble methods in sentiment classification, in International
Conference on Cyber Security and Computer Science (Springer, Cham, 2020), pp. 391–402
M.F. Sohan, S.S.M.M. Rahman, M.T.A. Munna, S.M. Allayear, M.H. Rahman, M.M. Rahman,
NStackSenti: evaluation of a multi-level approach for detecting the sentiment of users, in Inter-
national Conference on Next Generation Computing Technologies (Springer, Singapore, 2018),
pp. 38–48
M.M. Rahman, S.S.M.M. Rahman, S.M. Allayear, M.F.K. Patwary, M.T.A.A. Munna, Sentiment
analysis based approach for understanding the user satisfaction on android application, in Data
Engineering and Communication Technology (Springer, Singapore, 2020), pp. 397–407
S.S.M.M. Rahman, M.H. Rahman, K. Sarker, M.S. Rahman, N. Ahsan, M.M. Sarker, Supervised
ensemble machine learning aided performance evaluation of sentiment classification, in Journal
of Physics: Conference Series, vol. 1060, no. 1 (IOP Publishing, 2018), p. 012036
A Conceptual Analysis of IoT in
Healthcare
Abstract There are numerous health problems associated with human life. Usually,
they are not instantaneous. There are significant parameters based on which any
disease or health problem arises. Nowadays hospitals are using state of the art infras-
tructure to keep track of patients’ health conditions. It is possible using Wireless
Body Area Network (WBAN). This chapter proposes a machine learning model to
discover hidden facts and to predict the future behavior of patients based on current
health conditions. Further, a conceptual analysis is presented using a heart disease
dataset.
1 Introduction
M. A. Umer (B)
DHA Suffa University, and KIET Karachi, Karachi, Pakistan
e-mail: [email protected]
M. T. Jilani
Karachi Institute of Economics and Technology, Karachi, Pakistan
e-mail: [email protected]
A. Rafiq · S. A. Naz
DHA Suffa University, Karachi, Pakistan
e-mail: [email protected]
S. A. Naz
e-mail: [email protected]
K. N. Junejo
DNNae Inc., Karachi, Pakistan
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 227
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_10
228 M. A. Umer et al.
Fig. 1 Connectivity of
Wireless Body Area
Network over the Internet
continuously update the patients’ health parameters in the centralized system. The
patient does not need to be physically under supervision in a ward or hospital. The
mode of supervision is remote while visits are conducted virtually. The interface or
gadget at the patient’s end collects the data with the help of sensors like pulse rate,
heartbeat, blood pressure, oxygen level, temperature, sleep, etc. Thus patient’s care
workflow becomes automated and the data collected is viewable at the doctor’s end.
Mere collection from patient’s and viewing of data at doctor’s portal helps in analyz-
ing and drawing predictions about the health of patient and hence many calamities
can be avoided. But this process can become even more productive if analysis is
also automated and doctor receives the trend summary reports along with the actual
historical data. This historical data can become very bulky over the time and manual
A Conceptual Analysis of IoT in Healthcare 229
analysis becomes impossible. However, by using data mining and machine learning
techniques, regression functions can be constructed and hence trends can be found
predicting the possible future behavior of patients including heart attacks, stroke,
etc. The proposed model uses machine learning algorithms for this purpose.
WBAN is an active research area with great potential for step up in the field of
personal health care (Negra et al. (2016)). It consists of some strong sensing devices
which can sense vital signs as well as emotions like fear, happiness, anxiety level,
etc. These sensors can be rooted under patient’s skin or as detachable. Some of the
examples include Wearable devices like bands, gloves, etc. Basic working model of
communication of some devices is explained in Fig. 1. It shows a three step process.
Level 1 refers to the communication within WBAN which consists of some wireless
sensing devices and the main node of WBAN which may be patient’s smart phone.
In level 2, the main node of WBAN is connected with any personal device like laptop
etc. Last level is concerned with the connection of laptop with the internet and the
beyond. This chapter presents a model to use the information provided by WBAN
to predict the future behavior of patients. The following are some main research
questions that motivated the current study.
RQ1: Does the patient suffering from specific disease have some specific medical
traits?
RQ2: Can machine learning be used to predict the future behavior of patients?
Contributions: We answer the above research questions through following contribu-
tions (a) A model to predict the future behavior of patient (b) A conceptual analysis
of the proposed model.
Organization: Sect. 2 presents the related work in this domain. Section 3 describes
preliminaries related to WBAN. The proposed model is described in Sect. 4. A con-
ceptual analysis of proposed model is presented in Sect. 5. The discussion on research
questions stated earlier is provided in Sect. 6. Conclusion and possible future work
is discussed in Sect. 7.
2 Related Work
Machine Learning algorithms are being applied in health-care Ghassemi et al. (2018);
Qayyum et al. (2020), Cyber Security Ahmed et al. (2021); Umer et al. (2020),
Surveillance Hunain et al. (2021), Weather Forecasting Umer et al. (2021), and sev-
eral other domains. It has robust performance in predictive analytics. It has been
applied in healthcare analytics like in Chen et al. (2017) Convolutional Neural Net-
work was used for disease risk prediction. The latent factor model was used to
overcome the problem of missing data. Similarly, a study reported in Abdelaziz et al.
(2018) used the machine learning algorithms (Linear Regression and Neural Net-
works) to predict chronic kidney disease. Linear Regression was used to study the
factors influencing chronic kidney disease while Neural Network was used to pre-
dict the disease. Zhang et al. (2014), discusses a healthcare system in which Wireless
Sensor Network (WSN) was used to monitor the health condition of the patient. It has
230 M. A. Umer et al.
divided the architecture into three layers where layer 1 includes sensor node devices,
layer 2 includes the access network and the Internet while layer 3 includes the medical
server and the telemedicine. A study reported in Tennina et al. (2014) has proposed
an e-health system through which a patient at home could be tracked by the doctor
using WSN. In Yang and Logan (2006), the 1999 National Inpatient sample data sets
(which was used for Healthcare cost and utilization project) was analyzed using data
mining techniques. The results obtained by data mining techniques were compared
with the expert knowledge survey of 97 gastrointestinal tract surgeons. This analy-
sis showed that the comparison of these two results was consistent in some factors
which are highly associated with paraesophageal hernia. Moreover, the data mining
approach also revealed some other related disorders which were not known by the
expert at that time and neither reported in the literature. In Ferreira et al. (2012), data
mining techniques were used to improve the diagnosis of neonatal jaundice. To per-
form experiments, 227 healthy newborn infants with 35 or more weeks of gestation
were studied. For this purpose, more than 70 variables were collected and analyzed.
Machine learning algorithms like Decision tree (J48) and neural networks (multi-
layer perceptron) were applied in this study. The accuracy of prediction was 89%.
Naive Bayes (NB), Multi Layer Perceptron (MLP) and simple logistics algorithms
provided the best results.
Nexleaf Thomas (2020) offers wireless remote temperature monitoring system
helps to regulate the temperature inside the refrigerators to avoid vaccine spoiling.
Similarly ample work has also been done to ensure environment safety by emitting
less carbon dioxide. This process is controlled by a cloud-based remote monitoring
system. SystemOne Briodagh (2017) uses IoT services of Vodafone to endure real-
time diagnostic information to the concerned health professionals. In Pfizer taps
ibm for research collaboration to transform parkinson’s disease care (2016), the
consultant-patient communication-gap is bridged my making real-time modifications
in the prescribed amount of certain medicines on the basis of data collected by sensors
that collect the changes in patient’s behavior at a very minute and macro level. This
was implemented in patients having a brain disorder that leads to shaking, stiffness,
and difficulty with walking, balance, and coordination. The loop wearable device
in (2021) can be helpful in monitoring vital signs of COVID patients as it was
designed to remotely measure the heart rate, measuring oxygen saturation in blood
of patients with any chronic obstructive pulmonary disease. More applications of
IoT in healthcare can be found in Thomas (2020).
New experimental techniques were introduced in Moser et al. (1999) for Data
Mining Surveillance System (DMSS) using an electronic healthcare database to
monitor the emerging infections and antimicrobial resistance. Information provided
by DMSS can be a good indicator of an important shift in infection and antimicrobial
resistance patterns in ICU. In a study reported in Kaur et al. (2012), an initiative was
taken to increase the availability of abusive drug usage data to provide better drug
recovery services. Observational study of cases registered in the California Depart-
ment of Alcohol and Drug Program was utilized for this purpose. Initially, the Minitab
diagnostic tool was used to diagnose the above-mentioned cases. Further it provided
access to the Casemix databases to retrieve the hidden information using data mining
A Conceptual Analysis of IoT in Healthcare 231
tools. K-means clustering was used to determine the existence of patients admitted
and discharged on the usage of abusive substances between 2 and 3 consecutive
years. Data were classified into educated and non-educated classes for a categorized
race along with correlation age at the time of admission in the hospital. During the
analysis phase, it was made sure to include the treatments provided to patients during
their stay at the hospital and also the discharge status of the final medical diagnosis.
Results showed that the incidence rate of admission cases in the 45–49 age group
was on a rising edge. With more than 40% probability, patients who were acquiring
the maximum number of abusive substance had obtained post-graduate education
Kaur et al. (2012).
Data mining was used in Cao et al. (2008) to extract complex data from the major
clinical trial repository. This data was summarized and visualized to make informed
decisions about future cancer vaccine clinical trials. Data mining techniques were
used in Setoguchi et al. (2008) to determine in which situations analyses based on
various types of exposure propensity score (EPS) models produce unbiased and/or
efficient results. Recursive partitioning, logistic regression, and neural networks were
used in this study. Logistic regression provided the robust results while the neural
network provided the least numerically biased estimates.
WBAN consists of small wireless sensor nodes that are interconnected into/on/off a
human body (up to some cm) via Radio Frequency (RF) as shown in Fig. 2. It is used
for collecting both the medical and general purpose data from the body. Nowadays,
WBAN is playing a vital role in medical sciences, specifically for human healthcare
systems. It is a state of the art technology that has reshaped the medical practices
including diagnostics, investigations, prescriptions, etc. It provides a new mechanism
for medical practitioners and doctors to diagnose disease in an efficient manner
by providing quick and effective solution. Apart from medical, it has a significant
contribution in the field of entertainment and consumer electronics. Despite this, it is
an active research area for researchers in clinical advancements Rathee et al. (2014).
With the recent advancements in IoT, WBAN has acquired a crucial status in the
field of computation, diagnosis, and implementation. The diagnostic of health param-
eters requires power and energy sustainability, therefore some recent techniques like
fog computing are provisioned to overcome such issues in IoT based WBAN. We use
to cater millions of terabytes of data during the human health condition diagnostics,
therefore specific machine learning (ML) algorithms are required for the next level
optimization in fog computing techniques Amudha and Murali (2020).
Another important issue related to WBAN is “Coexistence”. In this problem,
WBAN has to deal with interference due to the same channel simultaneous occupancy
of multiple WBANs. A study reported in Sun et al. (2018) used the ML algorithms
namely decision tree (DT) and naïve Bayes classifier (NBC) to tackle the above
232 M. A. Umer et al.
4 Proposed Model
WBAN can be used to monitor the patients’ health condition. All this data gets
recorded in the system, therefore this data could be used to extract meaningful results
and to predict the behavior of patients in the future. To achieve this goal, this chapter
has proposed a model using machine learning techniques as described in Fig. 3. The
proposed model is further described in the following sub sections.
Data available in the system would be used to form data sets. Relevant attributes/
features would be selected from these data sets. This requires an understanding of
the dataset at the initial step. It helps in selecting the important attributes for training
the model. For this purpose, several statistical techniques can be applied to deeply
study the insights of data. It includes but not limited to finding the correlation among
A Conceptual Analysis of IoT in Healthcare 233
the attributes, finding the noise in the dataset using box plot, covariance, deviations,
central tendency, etc.
The selection of an algorithm does not depend only on the type of problem but also on
the type of dataset which is being used to train the model. There are several machine
learning algorithms which could be used for this purpose. For example, Neural net-
works can be used for this purpose. They have the capability of approximating a
function that depends on a large number of inputs. This particular system may have
to deal with a large number of inputs to cover every possible aspect of patients’ health
condition. Futhermore, Neural networks learn from the experience which could be
helpful in the prediction of a patients’ medical behavior.
Fuzzy logic has been used in conjunction with deep neural networks (DNN) in Leke
et al. (2015) to perform the missing data imputation tasks. The proposed model is
focused on critical issues related to patient’s health condition, therefore it is recom-
234 M. A. Umer et al.
mended to use DNN in conjuction with fuzzy logic to handle the missing values in
the dataset.
4.4 Implementation
To train the machine learning model, dataset must be divided into three sets i.e.
training set, validation set, and test set. At the begining, the model is trained using
the training dataset. The trained model is then evaluated on validation set to observe
the performance. This validation step is used to tune the model based on certain
parameters of algorithm. Futhermore, it also helps in avoiding the overfitting of the
model. This step is an iterative step as it is required to repeat the feature selection
process and data processing for making a good model. When the model gets the
desired accuracy, then it is tested on test data. This step gives the overall performance
of the model.
The study reported here has performed a conceptual analysis of the proposed model
using the heart disease dataset available at UCI repository Uci (2019). This dataset
consists of the following 14 attributes:
• Age
• Sex
• Chest Pain Type
• Resting Blood Pressure
• Serum Cholestrol
• Fasting Blood Sugar
• Resting Electrocardiographic Results
• Maximum Heart Rate Achieved
• Exercise Induced Angina
• ST Depression Induced By Exercise Relative to Rest
A Conceptual Analysis of IoT in Healthcare 235
5.1 Algorithms
Following machine learning algorithms were used to train the model on the heart
disease dataset:
• Logistic Regression (LR)
• Linear Discriminant Analysis (LDA)
• KNeighbors Classifier (KNN)
• Classification And Regression Tree Classifier (CART)
236 M. A. Umer et al.
Four iterations were run using the above-mentioned algorithms. The actual score
(accuracy) of these four iterations are shown in Fig. 7. In order to determine the
algorithms deviating on each iteration, mean and the standard deviation of each
algorithm was calculated, as shown in Figs. 8 and 9. It is evident from Fig. 9 that
there are three algorithms that are deviating from their scores in each iteration and
these are CART, MLP, and RF. Among which MLP was found to be a highly deviating
algorithm on each iteration with the same set of parameters configuration.
A Conceptual Analysis of IoT in Healthcare 237
Now we would discuss the research questions stated in the introductory section.
RQ1: Does the patient suffering from specific disease have some specific medical
traits?
RQ2: Can machine learning be used to predict the future behavior of patients?
The answer to both questions is yes, because the study reported in Sect. 5 used the
dataset consists of various medical related attributes to predict the heart disease. It
used the machine learning techniques to predict the heart disease. There are various
machine learning techniques which work differently. Some of which finds specific
patterns in the dataset and then classifies the new instance accordingly. As ML
algorithms were able to classify different heart disease, it means there are specific
medical traits in the patient that could be used to classify the patients’ disease. This
leads us to the answer of RQ1 stated earlier. ML algorithms finds the specific patterns
in the dataset of patients. This can lead to find out the changes in structural behavior
using certain attributes of patient, thus they are capable to predict the future behavior
of patient based on current health conditions.
This chapter has proposed a model to predict the patients’ medical behavior in future.
Success in the proposed model could be extremely beneficial for the human life and
opens a research direction for the scientist in the related fields. This chapter has
presented a conceptual analysis using a heart disease dataset. Complete Practical
A Conceptual Analysis of IoT in Healthcare 239
implementation of the proposed model using proper clinical data is still needed to be
done. Hopefully, novel aspects of the proposed model could lead to some significant
improvements in healthcare.
References
R. Negra, I. Jemili, A. Belghith, Wireless body area networks: applications and technologies. Pro-
cedia Comput. Sci. 83, 1274–1281 (2016)
M. Ghassemi, T. Naumann, P. Schulam, A.L. Beam, R. Ranganath, Opportunities in machine learn-
ing for healthcare (2018). arXiv:1806.00388
A. Qayyum, J. Qadir, M. Bilal, A. Al-Fuqaha, Secure and robust machine learning for healthcare:
a survey (2020). arXiv:2001.08103
C.M. Ahmed, M.A. Umer, B.S., S.B. Liyakkathali, M.T. Jilani, J. Zhou, Machine learning for cps
security: applications, challenges and recommendations, in Machine Intelligence and Big Data
Analytics for Cybersecurity Applications (Springer, 2021), pp. 397–421
M.A. Umer, A. Mathur, K.N. Junejo, S. Adepu, Generating invariants using design and data-centric
approaches for distributed attack detection. Int. J. Crit. Infrastruct. Prot. 28, 100341 (2020)
M. Hunain, T. Iqbal, M.A. Siyal, M.A. Umer, M.T. Jilani, A framework using artificial intelligence
for vision-based automated firearm detection and reporting in smart cities, in Artificial Intelligence
and Blockchain for Future Cybersecurity Applications (Springer, 2021), pp. 237–255
M.A. Umer, M.T. Jilani, K.N. Junejo, S.A. Naz, C.W. D’Silva, Role of machine learning in weather
related event predictions for a smart city, in Machine Intelligence and Data Analytics for Sus-
tainable Future Smart Cities (Springer, 2021), pp. 49–63
M. Chen, Y. Hao, K. Hwang, L. Wang, L. Wang, Disease prediction by machine learning over big
data from healthcare communities. IEEE Access 5, 8869–8879 (2017)
A. Abdelaziz, M. Elhoseny, A.S. Salama, A. Riad, A machine learning model for improving health-
care services on cloud computing environment. Measurement 119, 117–128 (2018)
Y. Zhang, L. Sun, H. Song, X. Cao, Ubiquitous WSN for healthcare: recent advances and future
prospects. IEEE Internet Things J. 1(4), 311–318 (2014)
S. Tennina, M. Di Renzo, E. Kartsakli, F. Graziosi, A.S. Lalos, A. Antonopoulos, P.V. Mekikis, L.
Alonso, Wsn4qol: a WSN-oriented healthcare system architecture. Int. J. Distrib. Sens. Netw.
10(5), 503417 (2014)
J. Yang, J. Logan, A data mining and survey study on diseases associated with paraesophageal hernia,
in AMIA Annual Symposium Proceedings, vol. 2006 (American Medical Informatics Association,
2006), p. 829
D. Ferreira, A. Oliveira, A. Freitas, Applying data mining techniques to improve diagnosis in
neonatal jaundice. BMC Med. Inf. Decis. Mak. 12(1), 143 (2012)
M. Thomas, 6 iot in healthcare applications leading to an improved industry (2020). https://builtin.
com/internet-things/iot-in-healthcare
K. Briodagh, Systemone and vodafone tackle medical challenge with iot (2017). https://www.
iotevolutionworld.com/iot/articles/432161-systemone-vodafone-tackle-medical-challenge-
with-iot.htm
Pfizer taps ibm for research collaboration to transform parkinson’s disease care (2016).
https://www.pfizer.com/news/press-release/press-release-detail/pfizer_taps_ibm_for_research_
collaboration_to_transform_parkinson_s_disease_care
The promise of clinical wearables for people with chronic conditions. https://spryhealth.com/the-
loop-monitoring-solution/. Accessed 30 June 2021
S.A. Moser, W.T. Jones, S.E. Brossette, Application of data mining to intensive care unit microbi-
ologic data. Emerg. Infect. Dis. 5(3), 454 (1999)
240 M. A. Umer et al.
H. Kaur, R. Chauhan, S.M. Aljunid, Data mining cluster analysis on the influence of health factors
in casemix data. BMC Health Serv. Res. 12(1), O3 (2012)
X. Cao, K.B. Maloney, V. Brusic, Data mining of cancer vaccine trials: a bird’s-eye view. Immunome
Res. 4(1), 7 (2008)
S. Setoguchi, S. Schneeweiss, M.A. Brookhart, R.J. Glynn, E.F. Cook, Evaluating uses of data
mining techniques in propensity score estimation: a simulation study. Pharmacoepidemiol. Drug
Saf. 17(6), 546–555 (2008)
D. Rathee, S. Rangi, S. Chakarvarti, V. Singh, Recent trends in wireless body area network (wban)
research and cognition based adaptive wban architecture for healthcare. Health Technol. 4(3),
239–244 (2014)
S. Amudha, M. Murali, Deep learning based energy efficient novel scheduling algorithms for body-
fog-cloud in smart hospital. J. Ambient Intell. Humaniz. Comput. 1–20 (2020)
Y. Sun, T. Chen, J. Wang, Y. Ji, A machine learning based method for coexistence state prediction in
multiple wireless body area networks, in EAI International Conference on Body Area Networks
(Springer, 2018), pp. 203–217
K.S. Kwak, S. Ullah, N. Ullah, An overview of ieee 802.15. 6 standard, in 3rd International Sym-
posium on Applied Sciences in Biomedical and Communication Technologies (ISABEL 2010)
(IEEE, 2010), pp. 1–6
C. Leke, T. Marwala, S. Paul, Proposition of a theoretical model for missing data imputation using
deep learning and evolutionary algorithms (2015). arXiv:1512.01362
Y. Liu, W.-k. Liao, A.N. Choudhary, J. Li, Parallel data mining algorithms for association rules and
clustering (2007)
Uci (machine learning repository). heart disease dataset (2019) https://archive.ics.uci.edu/ml/
datasets/Heart+Disease
Securing Big Data-Based Smart
Applications Using Blockchain
Technology
Abstract Nowadays, Big Data is the most salient paradigm. It has become a game-
changer in the current technology and has aroused various industries and research
communities worldwide. Big Data-based applications have shown much potential in
data-driven decisions that have changed all business operations, including cost opti-
mization, online reputation, and customer loyalty. Nevertheless, Big Data also brings
many security and privacy issues. For this reason, Big Data applications require accu-
rate methods to ensure reliable data storage, sharing, and decision-making. Among
these methods, the Blockchain is known as the clever combination of distributed
exchange, consensus, and cryptography mechanisms. It has been brought to the
forefront to resolve security challenges in various fields such as Healthcare, banking,
smart cities, etc. This paper first presents the considerable difficulties faced when
adopting Big Data that can be resolved using Blockchain. Afterward, we overview the
Blockchain technology by projecting its components, workflows, classification, and
related characteristics. Finally, we present the importance of combining Big Data and
Blockchain through reviewing the novel implementations proposed by researchers
in great domains.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 241
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_11
242 R. Benaich et al.
1 Introduction
With the increase of people using the Internet, the amount of produced data is practi-
cally exploding by the day. This explosion opens up incredible processing opportuni-
ties for organizations, such as data sharing and data-driven decisions (Faroukhi et al.
2020) and proposing customizable services. However, this noteworthy and extensive
usage also comes with the necessity of ensuring the security and protection of data
storage and transmissions (Gahi et al. 2016).
It is worth noting that this vast data, also known as Big data, is, most of the
time, very sensitive (Gahi et al. 2016) as it represents a direct projection of our day-
to-day activities. Big data is closely related to our personal life as it mainly relies
on our transactions, interactions, and pattern mining. As Big data becomes a grand
reality, it has to be clearly defined. Since the 2000s, many characteristics have been
associated with Big data (El Alaoui et al. 2019), such as Volume, Variety, Velocity,
Value, Veracity, Variability, and Visualization. Also known as 7Vs (El Alaoui and
Gahi 2019), see Fig. 1.
Because of these varied characteristics, ensuring security using traditional
methods becomes a complex task (Hu et al. 2014). For example, many conventional
security techniques cannot handle the scale and velocity required by Big Data-based
applications.
To overcome the scalability and velocity aspects, distributed approaches should
be used. These approaches are designed to run over large clusters, which are a set
of interconnected machines. Big Data systems’ common distribution aspect makes
Blockchain technology an ultimate solution for Big Data security challenges. It
allows coping with security issues related to sharing and storing Big Data by applying
cryptography, decentralization, consensus principles, and ensuring transaction trust.
Moreover, today, Blockchain technology has become an ideal fit for Big Data prob-
lems through its high capacity to prevent the significant malicious activities caused
by cybercriminals. The Blockchain not only restricts the intruders but also ensures
access to multiple nodes in the network. This safeguarding feature of Blockchain
strengthens and boosts the security level compared to many unique technologies
used in various enterprises such as the cloud or virtual private network, or any other
technology.
For this reason, many researchers have taken a keen interest in proposing attractive
and diverse Blockchain models and frameworks to share and securely store Big Data
in various sectors, namely Healthcare, Banking, IoT, gaming, and smart applications.
In this contribution, we focus on reviewing the recent studies that have tackled
Big Data and Blockchain and their alliance. Furthermore, we discuss the advantage
and the limits of the proposed models.
The following research questions will guide this contribution:
• What are the application domains of blockchain technology in the context of Big
data?
• How does Blockchain technology address the challenge of Big data security?
• What are the significant challenges remaining in the usage of Blockchain as a
secure solution for Big data intelligent applications?
This paper is organized as follows: in the second snd third sections, we present
the research methodology, and we overview the concept of Blockchain technology
by showing its components, workflows, and related characteristics. Afterward, in the
fourth section, we highlight exciting studies based on Blockchain brought to various
domains such as the healthcare field, smart cities, IoT, Vanets, 5G networks, Banking,
and Game theory. Then in Sect. 5, we discuss the importance of using Blockchain in
the context of Big Data. Finally, we conclude the paper and show some future works.
2 Research Methodology
To carry out this paper, we have used a systematic review methodology to iden-
tify and summarize the relevant research contributions that answer the above ques-
tions. According to (Tranfield et al. 2003) standards, we have conducted our review
to develop a review plan regarding the methods to be used and the main research
questions to be outlined and reports to be extracted.
This literature review highlighted the critical role of the outstanding Blockchain
technology in coping with Big data challenges in numerous domains.
We took three stages to perform this study:
1. First of all, we researched and identified multiple primary papers related to our
topic.
2. Secondly, we filtered the primary chosen papers.
3. Finally, we validated the filtered articles for data synthesis and analysis.
244 R. Benaich et al.
The primary studies were chosen based on the inclusion and exclusion criteria
applied to the reviewed publications, the study’s scope, and verifications done by
subject and title and reading the abstract and findings.
This research method enables the investigation to be focused on papers published
in renowned academic journals such as IEEE Xplore, Springer, Google Scholar,
Wiley online library, ACM digital library, web of science, and so forth.
In our research approach, we covered various papers and magazines that were
published in recent years. This allowed us to illustrate the progression of the thoughts.
Therefore, we selected particular keywords such as “Big Data,” “Blockchain,” “Secu-
rity,” and “Blockchain applications” to assure a superior approach in literature
research.
A considerable number of articles on the order of 300 were found through the
selection of source research. A relevant and rigorous study of cross-referenced and
redundant keyword combinations was used to refine this first bundle. A total of 180
articles were chosen, which were then subjected to a more in-depth study.
The filtered papers were carefully studied and assessed to ensure that they were
relevant to the subject of our research. The last group consisted of 100 articles.
• Merkle tree: also known as a binary hash tree, it is a tree in which each leaf node
is marked by the hash value of the block transaction data. Merkle trees are used
to effectively and securely encode Blockchain data.
An overview of the Blockchain workflow is shown in Fig. 2.
The procedure shown in Fig. 2 works as follows:
• The first step consists of initiating and broadcasting a transaction to the distributed
network via a node.
• After that, all Blockchain network nodes verify the transaction (the node that
performs the broadcast).
• More than one node can bundle different subgroups of the newly verified transac-
tions in their candidate constituencies and broadcast them over the entire network
in the third step.
• In the fourth step, some nodes or the entire network nodes validate blocks by
executing functions defined by consensus protocol.
• The last step consists of the attachment of the verified block to Blockchain, and
finally, all nodes update their local replica.
The Blockchain can be classified into the following categories depending on its
usage (Viriyasitavat and Hoonsopon 2019):
• Public: public Blockchains are decentralized and visible by anyone, and they
don’t have a specific owner, for example (Bitcoin, Ethereum).
• Private or Permissioned: they require some rules and privileges to manage and
control who can read and write to the Blockchain.
• Hybrid or Consortium: these Blockchains are public only to a specific group.
The consensus process is controlled by known, privileged servers using all parties’
rules.
It is also important to mention that Blockchain technology is acknowledged with
these main characteristics (Yaga et al. 2018):
• Decentralized: a key Blockchain feature. It means that Blockchain does not have
to depend on location nodes anymore. Data can be recorded, stored, and updated
for distribution.
• Immutable: every time a record of transactions is added, it’s permanent and
cannot be changed.
246 R. Benaich et al.
Researchers use Blockchain for different security aims in the context of Big Data.
In this section, we first emphasize the security aims by reviewing recent researches
in various domains. Then, we compare these studies by showing their benefits and
limits.
Healthcare is a critical and essential domain that has attracted massive attention from
researchers. As a result, several studies have proposed and enhanced Blockchain
models to cope with several healthcare challenges in the context of Big Data.
Blockchain technology redefines the data modeling and management embedded
in many health care systems (Liang et al. 2017b). This is mainly due to its adapt-
ability and differentiation capabilities, security, and sharing of information and
medical services unprecedentedly. One of the most crucial usages of Big Data and
Blockchain is securing and sharing personal health data obtained from wearable
devices such as smartwatches and activity trackers via the creation of a user-centric
health data sharing based on decentralized and permissioned Blockchain. This solu-
tion offers data security via the deployment of a channel formation scheme. Further-
more, it ensures the enhancement of identity management via the membership service
provided by Blockchain (Liang et al. 2017b). It also allows the prevention of any alter-
ation of the patients’ health data through applying a proof of integrity and validation
Securing Big Data-Based Smart Applications … 247
that is continuously reachable from a cloud database then linked to the Blockchain
network. Another potential of Blockchain that should not be overlooked is its capa-
bility to surmount vulnerabilities in the old sharing methods, such as the inappropriate
access of data and the integration of malicious programs.
Also, Blockchain is used in pharmacy (Clauson et al. 2018) and helps to control
and secure the medication return process. In this regard, the Blockchain has shown its
potential toward the product (medicines) tracing by allowing manufacturers, distrib-
utors, and pharmacists to submit tracing information in a shared ledger, with auto-
mated verification of critical data. Besides, Blockchain has demonstrated a potential
toward detection and response and enables public and private actors to report and
identify medicines suspected of imitation, illegal, or dangerous. Therefore, the safety
of medications increases without disclosing sensitive information.
It is also important to mention that the interest in the Blockchain has drastically
increased during the pandemic of COVID-19 (Abd-alrazaq et al. 2021). According to
the European Parliamentary Research Service, the Blockchain was considered among
the ten powerful technologies which have reduced COVID-19 challenges (Mihalis
Kritikos 2020). It is used to track public health information, particularly outbreaks of
infectious diseases. Combining both Blockchain and Big Data enables real-time data
processing, further preventing the spread of the disease to epidemic levels. It also
allows obtaining more accurate reporting and effective responses. However, these
effective responses are insufficient without safeguarding the patient’s information,
which requires a high level of security. Hence, by health information exchange based
on the Blockchain, the privacy of the patient is ensured. The aspect of securing
each patient’s medical data is obtained by deploying an electronic medical record
Blockchain that uses on and off-chain storage verification records (Ahir et al. 2020).
Blockchain is widely used in the banking sector and has many advantages in the Big
Data context. This innovation reduces the risk of fraud and prevents some potential
scams in all banking environments, such as financial and online markets (Hassani
et al. 2018) via monitoring and recording every change of the data within the blocks
for every transaction in real-time. Furthermore, blockchain technology is considered
a trusted network for the digital banking sector due to its encryption capabilities and
public key infrastructure. Additionally, the distributed ledger “Blockchain” offers
another aspect that should not be neglected, relying on the difficulty of operating the
newly added data by various units in real-time. This difficulty results in enhanced
privacy of the massive data transiting.
Blockchain also addresses various issues such as operational risk and admin-
istrative costs. Furthermore, integrating Blockchain and Big Data into the finan-
cial sector optimizes multiple processes, such as creating new financial services
and ensuring the integration of the performance and profitability of actors (micro-
credit, micro-payment transactions almost free of charge, and so forth. These benefits
248 R. Benaich et al.
In recent years, the notion of “smart city” has emerged as a new paradigm to enhance
citizen’s daily life, by providing personalized and adapted services. This paradigm
is based on personal data that are closely related to our daily basis. However, they
are several challenges related to security and storage in smart city systems. For this
reason, the idea of securing the data aroused the interest of many researchers.
Many studies have tackled the smart city field by proposing new methods and
practices. The major one is the combination of Blockchain and Big Data. This combi-
nation provides secure communication between physical devices in a heterogeneous
environment through real-time data monitoring. It also provides the quality of service
(QoS) by minimizing traffic rate fluctuations and the diversity of new devices (Alam).
Furthermore, Blockchains can be used to create a smart city system that enables
devices to securely and reliably transfer currency and data between all smart city
devices.
Also, the combination of Big Data and Blockchain is widely used in smart trans-
portation, especially in the automotive industry. This combination has revolution-
ized intelligent transportation systems (ITS) by building safe, dependant and self-
sufficient ITS ecosystems to provide services such as remote software-based vehicle
operation (Deepa et al. 2021).In the same context of smart transportation security,
it is noteworthy that (Wang and Qu 2019) have created a secure critical manage-
ment architecture based on Blockchain technology to provide network security. This
solution is based on using security managers to record vehicle departure data, encap-
sulate the blocks to transfer keys, and then deploy rekeying to the cars within the
secured domain. This framework provides a key management system for critical
transfers among security administrators in a heterogeneous vehicle network. Other
researchers have proposed an interesting system that ensure the security of vehicles
using Blockchain (Hîrţan et al. 2020). This system is based on an offline Blockchain
storage system. All confidential data gathered from users is kept and then shared
with the help of unique encryption keys applicable to a specific vehicle cluster. The
Securing Big Data-Based Smart Applications … 249
solution comprises two applications, the first one is installed on the client’s smart-
phone, and the second is installed on the server. It has shown the capacity to ensure
privacy policies sent to the client’s application to offer the precise transit routes.
Furthermore, the combination of Big Data and Blockchain allows managing the
security issues such as protocol vulnerabilities, privacy, eavesdropping, and attacks
on internet-connected devices and vehicle communication networks (Ismagilova
et al. 2020).
Today, most gaming platforms are hosted on centralized servers, and transactions
are frequently conducted on mobile phones or desktop computers that lack adequate
security (Jacob 2020). In fact, because of this centralization aspect, servers could
be easily hacked. Furthermore, this lack of server security could lead to enormous
damages such as data loss or interruption of services. Therefore, many studies have
tackled the combination of Blockchain and Big Data in gaming to cover these risks.
This combination has brought the potential to change the vulnerability of centralized
servers. Hackers will be unable to damage a decentralized Blockchain network since
(i) there is no server to destroy, (ii) nodes share the maintenance of distributed
databases, (iii) each node has complete information in the database. As a result,
players can safely store digital collectibles purchased in Blockchain-based games in
their crypto wallets (Aran Davies 2020).
Furthermore, due to the decentralization aspect of Blockchain, trust is established
between all nodes; players are unable to modify the data. This means that trust is
built between all players in the industry, from developers to actors (Gainsbury and
Blaszczynski 2017). Also, the Blockchain permit gamers to make their payments in
a secure way using digital crypto coins.
Besides the security aspect, the Blockchain also offers new features to the gaming
world, such as digital trade assets between players (Casino et al. 2019). The use
of intermediary sites obtains this ability of digital trading. In addition, to provide
user-friendly functions and interfaces for running shared data (Robin8 2019).
Another aspect of Blockchain technology resides in solving data instability in
peer-to-peer games by providing data authentication and permanent storage via a new
proof-of-play consensus methodology underpins this serverless turn-based strategy
game (Wu et al. 2020). This solution is based on three key steps: matchmaking,
gaming session, and gameplay.
The Internet of things (IoT) refers to the billions of physical devices deployed
worldwide connected to the Internet (Atzori et al. 2010). These devices collect and
250 R. Benaich et al.
share massive data. Thus, they allow conducting data analysis for better and faster
decision-making. However, there are many security issues, such as data leaks and
vulnerabilities concerning physical equipment; hackers can easily alter that.
For this reason, Blockchain was introduced to ensure privacy and to face chal-
lenges related to IoT security. Using Blockchain as the basis of devices reduces
the possibility of hacking by decreasing malicious programs and spyware (Banerjee
et al. 2018). In this context, the Blockchain was introduced to trace the history of the
firmware. If a corrupted firmware is identified, it will be driven to revert to its prior
version. Also, the Blockchain was brought to the front line to maintain the Refence
Integrity Metric (RIM) of the datasets by storing membership informations such as
the address, owner, and sharing policies.
Furthermore, Blockchain technology allows protecting IoT networks against
potential entry points of intruders without the need of a central authority. Therefore,
Blockchain is considered a leading solution for organizations involved in auditing,
tracing a supply chain, or securing connected streetlights in intelligent cities due to
the encryption capability of Blockchain, its dispersed storage, and its irrefutable,
tamper-proof record (Plummer and Writer 2018).
Besides the security aspect, Blockchain creates a marketplace that allows data
providers (IoT sensors) and data consumers (IoT application owners) to trade in
real-time via smart contracts and makes value from collected data (Jain 2021).
Another breakthrough of blockchain use in IoT is embodied in a reliable and
flexible solution designed for IoT services, specifically the adoption of drones in
various IoT scenarios, namely agriculture, delivery, and so forth. In fact, (Liang
et al. 2017b) have proposed a solution that consists of using a public Blockchain and
the standard cloud server in which the hashed data records collected from drones
are anchored to the Blockchain network. After that, a Blockchain receipt for each
data record stored in the cloud is generated to reduce the burden of moving drones
with limited battery and processing capability while gaining improved data security
assurance. In addition, this usage provides trust, data capability integrity audit, data
convenience, and scalability.
VANETs are one of the most brilliant inventions of the century. However, they
still include many challenges that could be resolved with Blockchain. Such as
data management and security in a vehicular environment. For example, the usage
of Blockchain inVANETs enables essential functionalities such as ensuring traffic
safety and parking space management through promoting decentralized system
management (Peng et al. 2020).
Furthermore, Blockchain played a pivotal role in resolving critical message
dissemination issues in VANETs via storing the history of vehicle trust level in
Blockchain and event messages (Shrestha et al. 2020). Thus, when a car experi-
ences an event, such as an accident, the event message with various parameters is
Securing Big Data-Based Smart Applications … 251
The novel features of 5G allow supporting new business models and services,
including mobile operators, businesses, call providers, government executives, and
infrastructure providers. However, they also face many challenges, such as data
mobility and network privacy (Nguyen et al. 2019).
For this reason, several studies have proposed various Blockchain models to
address 5G challenges in the context of Big Data. Moreover, blockchain char-
acteristics offer promising solutions to 5G’s challenges. One of these essen-
tial Blockchain features is decentralization, which allows for banishing external
authorities’ requirements in the 5G ecosystem.
In other words, decentralization of 5G networks enables avoiding single-point
failures and ensuring data availability (Nguyen et al. 2019). It is also important to
note that Blockchain offers high security for 5G networks due to smart contracts that
support the services of 5G, such as preservation of 5G resources against alteration
and data authentification (Nguyen et al. 2019).
252 R. Benaich et al.
5 Discussion
Through this paper, we highlighted how security challenges related to Big Data
could be resolved with Blockchain. Researchers have proposed Blockchain models
and frameworks to face many security challenges in diverse Healthcare, banking,
IoT, and so forth.
By reviewing these studies, we have noticed that Blockchain played a pivotal role
in securing and preserving data privacy through its decentralization, tamper-proof,
and immutability attributes. Moreover, far beyond, the combination of Big Data and
Blockchain has assured the security of the data by creating partitioning models based
on smart contracts. Accordingly, the data sharing models are generated without the
need for a reliable third party.
Besides the security aspect provided by Blockchain, many outstanding advantages
are emerging in the implementation of Blockchain claimed by several studies. For
instance, transparency and a remarkable reduction of the costs of data storage. This
significant minimization of the cost is achieved via the Blockchain storage capacity,
which allows storing a large amount of data for long periods instead of using shared
data holding platforms, which require many resources. In this process, Blockchain
relies on the automatic performance of transactions that are supervised by smart
contracts.
Moreover, due to the transparency and decentralization aspects of Blockchain
technology, users can look through the record of all transactions.
Despite the advances brought by the usage of both Blockchain and Big Data. A lot
of issues still unreliable and require more research. Scalability is an example of these
problems; it is known as a critical requirement of Big Data. Companies and businesses
Table 1 Summary of blockchain and big data applications
Field Reference The faced challenges Used methods Advantages Limits
Healthcare Liang et al. (2017b) The privacy issues Implementing an access control – Preservation of the Difficulty in combining
and vulnerabilities scheme by utilizing the Hyperledger integrity of health data both personal health
existing in personal Fabric membership service within each record data and medical data
health data and component, and a tree-based data – Protection and
storing system processing Validation of personal
health data
Wang et al. (2018) The challenge of A medical data sharing platform – The use of encryption – High cost of
sharing medical data based on permissioned Blockchains technology ensures centralized data
Smart contracts - Concurrent that users have storage
Byzantine Fault Tolerance- complete autonomy in – Lack of security and
Consensus mechanism their medical data protection of
– Smart contract personal privacy
technology is used to
Securing Big Data-Based Smart Applications …
be more easily
performed on big data
– Support to store large
data payloads with
Cassandra
Smart field Biswas and Overcoming security They are integrating blockchain – Improved reliability Interoperability issue
Muthukkumarasamy challenges in smart technology with smart devices – Better fault tolerance
(2016) cities capability
– Faster and efficient
operation
– Scalability
(continued)
255
Table 1 (continued)
256
architecture with Big on the IoT (IoT nodes, P2P network) communicate securely across all smart
data analytics to with other physical devices
increase connectivity devices in – The privacy issue
performance heterogeneous – Scalability issues
throughout the smart environments
cities – Big Data analytics
perform an
increasingly
significant role in
strategic planning
(continued)
257
Table 1 (continued)
258
(continued)
Table 1 (continued)
260
cannot efficiently handle and process the exploding volume of data without using
data scaling techniques. Moreover, the lack of standardization is another drawback
of combining Blockchain and Big Data. In other words, standardization denotes
a minimal degree of interoperability, such as (different Blockchain networks and
different consensus models).
The interoperability problem leads to increased resources used in transactions
generated by users in various blockchain networks. Also, without standardiza-
tion, sharing data and value between participants in divergent Blockchain networks
becomes complex or almost impossible. Furthermore, another limit of the adop-
tion of Blockchain in the service of Big Data is redundancy. This limit highlights
the fundamental issue of decentralization attributes brought by Blockchain. In other
words, redundancy means that every node in the Blockchain network must traverse
and process each intermediate node independently (Mudrakola 2018) to target the
objective node in the network. As a result, the redundancy inherent in Blockchain
technology has an impact on its efficiency.
Additionally, Blockchain faces another limit related to signature verification. The
employment of this signature verification in Blockchain technology means that each
transaction made in the Blockchain network should be verified and signed digitally
using private or public keys. Consequently, the computation of this signature verifi-
cation process becomes highly complex and takes a long time. This limit hampers
the alliance of Blockchain with Big Data.
6 Conclusion
The technology of Blockchain is considered a hot topic given its robust features
and good advantages. Today, Blockchain’s importance has extended far beyond the
financial sector and has shown notable improvements, especially toward security
concerns. On the other side, Big Data must be handled securely and accurately to
avoid any interruption or loss of the data. Therefore, Blockchain has been brought
to the front line in Big Data contexts. In this paper, we have presented the alliance
of Blockchain with Big Data to cope with Big Data storing and sharing security
issues. For this, we have overviewed the multiple advantages and limits of different
Blockchain researches that have tackled Big data issues in many sectors ranging from
health and technology (IoT, 5G, VANETs, Smart field) to entertainment concerns.
Despite the advancement brought by Blockchain technology, there are still plenty
of challenging issues concerning the use of Blockchain in the context of big data,
such as interoperability and scalability. However, distributed ledger technology is
still maturing. Moreover, it can be linked to new-age technologies, such as artificial
intelligence and IoT, to build platforms and infrastructures that ensure advanced data
privacy and security.
As future work, we propose a Blockchain model that copes with unsolved Big
data security issues such as scalability and standardization.
262 R. Benaich et al.
References
H. Hassani, X. Huang, E. Silva, Banking with blockchain-ed big data. J. Manag. Anal. 5(4), 256–275
(2018). https://doi.org/10.1080/23270012.2018.1528900
L.-A. Hîrţan, C. Dobre, H. González-Vélez, Blockchain-based reputation for intelligent transporta-
tion systems. Sensors 20(3), 791 (2020). https://doi.org/10.3390/s20030791
H. Hu, Y. Wen, T.-S. Chua, X. Li, Toward scalable systems for big data analytics: a technology
tutorial. IEEE Access 2, 652–687 (2014). https://doi.org/10.1109/ACCESS.2014.2332453
S. Iqbal, A. Malik, A.U. Rahman, A. Waqar, Blockchain-based reputation management for task
offloading in micro-level vehicular fog network. IEEE Access 8, 52968–52980 (2020). https://
doi.org/10.1109/ACCESS.2020.2979248
E. Ismagilova, L. Hughes, N.P. Rana, Y.K. Dwivedi, Security, privacy and risks within smart cities:
literature review and development of a smart city interaction framework. Inf. Syst. Front. (2020).
https://doi.org/10.1007/s10796-020-10044-1
N. Jacob, How blockchain is making digital gaming better, in Blockchain Pulse: IBM
Blockchain Blog (2020). https://www.ibm.com/blogs/blockchain/2020/02/how-blockchain-is-
making-digital-gaming-better/. Accessed 22 Mar 2021
S. Jain, Can blockchain accelerate internet of things (IoT) adoption, in Deloitte Switzer-
land (2021). https://www2.deloitte.com/ch/en/pages/innovation/articles/blockchain-accelerate-
iot-adoption.html. Accessed 25 Mar 2021
A.S. Khan, K. Balan, Y. Javed, S. Tarmizi, J. Abdullah, Secure trust-based blockchain architecture
to prevent attacks in VANET. Sensors 19(22), 4954 (2019). https://doi.org/10.3390/s19224954
S. Kim, Impacts of mobility on performance of blockchain in VANET. IEEE Access 7, 68646–68655
(2019). https://doi.org/10.1109/ACCESS.2019.2918411
X. Liang, J. Zhao, S. Shetty, D. Li, Towards data assurance and resilience in IoT using Blockchain,
in MILCOM 2017—2017 IEEE Military Communications Conference (MILCOM). (IEEE,
Baltimore, MD, 2017a) pp 261–266
X. Liang, J. Zhao, S. Shetty, J. Liu, D. Li, Integrating Blockchain for data sharing and collaboration in
mobile healthcare applications, in 2017 IEEE 28th Annual International Symposium on Personal,
Indoor, and Mobile Radio Communications (PIMRC). (IEEE, Montreal, QC, 2017b), pp. 1–5
Z. Lu, W. Liu, Q. Wang, G. Qu, Z. Liu, A privacy-preserving trust model based on blockchain
for VANETs. IEEE Access 6, 45655–45664 (2018). https://doi.org/10.1109/ACCESS.2018.286
4189
M. Kritikos, Ten technologies to fight coronavirus 28 (2020)
S. Mudrakola, Blockchain limitations: this revolutionary technology isn’t perfect—and here’s why,
in TechGenix (2018). https://techgenix.com/blockchain-limitations/. Accessed 28 Mar 2021
S. Nakamoto, Bitcoin: a peer-to-peer electronic cash system 9
D.C. Nguyen, P.N. Pathirana, M. Ding, A. Seneviratne, Blockchain for 5G and beyond networks: a
state of the art survey (2019). arXiv:191205062 [cs, eess, math]
C. Peng, C. Wu, L. Gao, J. Zhang, K.-L. Alvin Yau, Y. Ji, Blockchain for vehicular internet of
things: recent advances and open issues. Sensors (Basel) 20(18) (2020). https://doi.org/10.3390/
s20185079
L. Plummer, T. Writer, Blockchain, IoT & security, in Intel (2018). Accessed 21 Mar 2021
F. Rezaeibagha, Y. Mu, Efficient micropayment of cryptocurrency from blockchains. Comput. J.
62(4), 507–517 (2019). https://doi.org/10.1093/comjnl/bxy105
Robin8, Why is trading digital assets via Blockchain the best solution? in Medium
(2019). https://medium.com/@robin8/why-is-trading-digital-assets-via-blockchain-the-best-sol
ution-6897db75faff. Accessed 22 Mar 2021
School of Computer Science and Electronic Engineering, University of Essex, U.K., Dey S, A proof
of work: securing majority-attack in blockchain using machine learning and algorithmic game
theory. IJWMT 8(5), 1–9 (2018). https://doi.org/10.5815/ijwmt.2018.05.01
R. Shrestha, R. Bajracharya, A.P. Shrestha, S.Y. Nam, A new type of blockchain for secure message
exchange in VANET. Digit. Commun. Netw. 6(2), 177–186 (2020). https://doi.org/10.1016/j.
dcan.2019.04.003
264 R. Benaich et al.
H. Tan, I. Chung, Secure authentication and key management with blockchain in VANETs. IEEE
Access 8, 2482–2498 (2020). https://doi.org/10.1109/ACCESS.2019.2962387
D. Tranfield, D. Denyer, P. Smart, Towards a methodology for developing evidence-informed
management knowledge by means of systematic review. Br. J. Manag. 14(3), 207–222 (2003).
https://doi.org/10.1111/1467-8551.00375
W. Viriyasitavat, D. Hoonsopon, Blockchain characteristics and consensus in modern business
processes. J. Ind. Inf. Integr. 13, 32–39 (2019). https://doi.org/10.1016/j.jii.2018.07.004
R. Wang, W.-T. Tsai, J. He, C. Liu, Q. Li, E. Deng, A medical data sharing platform based on
permissioned blockchains, in Proceedings of the 2018 International Conference on Blockchain
Technology and Application—ICBTA 2018. (ACM Press, Xi’an, China, 2018), pp. 12–16
S. Wang, X. Qu, Blockchain applications in shipping, transportation, logistics, and supply chain
(2019), pp. 225–231
F. Wu, H.Y. Yuen, H.C.B. Chan, V.C.M. Leung, W. Cai, Infinity battle: a glance at how blockchain
techniques serve in a serverless gaming system, in Proceedings of the 28th ACM International
Conference on Multimedia. (ACM, Seattle WA USA, 2020), pp. 4559–4561
C. Xu et al. Making big data open in edges: a resource-efficient blockchain-based approach. IEEE
Trans. Parallel Distrib. Syst. 30(4), 870–882 (2018)
D. Yaga, P. Mell, N. Roby, K. Scarfone, Blockchain technology overview. (National Institute of
Standards and Technology, Gaithersburg, MD, 2018)
X. Zhang, X. Chen, Data security sharing and storage based on a consortium blockchain in a
vehicular ad-hoc network. IEEE Access 7, 58241–58254 (2019). https://doi.org/10.1109/ACC
ESS.2018.2890736
Overview of Blockchain-Based Privacy
Preserving Machine Learning for IoMT
Abstract The rapid growth of the internet of medical things (IoMT) in practical life
causes privacy concerns for data owners at the time of data analysis. There are many
methods available to examine the IoMT data in a privacy-preserving manner but a
study is needed to understand the best practical method for real life. So, this study will
not propose a novel method but will explore the state-of-the-art method to address
the most realistic method for privacy-preserving data analysis. This considered two
widely used technologies are Cryptography and Differential privacy. Performances
are considered in terms of accuracy and time consumptions for the latest published
research. Based on the analysis it is clear that both methods have flaws and can
be used depending on the constraints of domains. Moreover, Differential privacy is
more practical as it has almost similar time consumptions just like standard unsecured
methods.
1 Introduction
IoMT Joyia et al. (2017) is rapidly expanding in the healthcare sector, and enormous
volumes of data continuously gather from these devices. Diversified Machine Learn-
ing (ML) Goodfellow et al. (2016) algorithms are used to train these vast amounts of
data to build the prediction model. The data analyst conducts ML training on the data
collected from various data owners. However, data owners often disagree to share
R. U. Haque
School of Computer Science & Technology, University of Chinese Academy of Sciences,
Shijingshan District 100049, Beijing, China
e-mail: [email protected]
Institute of Automation Research and Engineering, Dhaka 1205, Bangladesh
A. S. M. T. Hasan (B)
Department of Computer Science and Engineering, University of Asia Pacific, Dhaka 1205,
Bangladesh
e-mail: [email protected]
Institute of Automation Research and Engineering (IARE), Dhaka 1205, Bangladesh
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 265
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_12
266 R. U. Haque and A. S. M. T. Hasan
their sensitive IoMTs data for training, as there are privacy issues. This study focuses
on the scenario where data analysts and data owners share their data in a Blockchain
platform.
ML method, where participants are unaware of each other sensitive information
in a Blockchain platform is known as Blockchain-based Privacy-Preserving Machine
Learning (BPPML) Al-Rubaie and Chang (2019). BPPML makes sure that sensitive
data and ML models are private to the owners and results are disclosed to the desig-
nated participants. Most of the studies focused on cryptography Haque et al. (2020);
Haque and Hasan (2021) and differential privacy Abadi et al. (2016). Blockchain
is consolidated in order to record the transaction’s information at the time of data
sharing and develop a private network. Homomorphic cryptography is often used in
BPPML in terms of the cryptographic method, where arithmetic operations are done
on encrypted data rather than the original text. On the other hand, in Differential
privacy-based BPPML methods, original data are added with noise before the ML
training. These are the two most popular methods often used for privacy-preserving
ML training. Each of these methods has advantages and limitations depending on
the use case.
The main focus of this article is to describe the best feasible method in terms of
accuracy and time complexity for BPPML in between cryptography and differential
privacy methods. Again, both BPPML techniques will be described to illustrate the
process of protecting data privacy. This study intends to help professionals in data
privacy, Blockchain, and machine learning fields by showing a comparative study
between Blockchain-based data privacy-preserving methods (i.e., cryptography and
differential privacy) on ML for IoMT data.
The rest of the paper is articulated as follows. Section 2 and 3 represent Related
works and Preliminaries. Section 4 represents System overview and Model construc-
tion and Sect. 5 represents Experimental setup and analysis of results. Finally, in
Sect. 6 we conclude our papers.
2 Related Works
Numerous research has focused on privacy issues. i.e., cryptographic Bost et al.
(2015), differential privacy Abadi et al. (2016), and privacy-preserving data publish-
ing Hasan et al. (2018), Hasan et al. (2018, 2017, 2016). Whereas cryptographic and
differential privacy is time-consuming and provides less data utility, respectively.
Again, ML training is not a concern of privacy-preserving data publishing. In addi-
tion, these methods cannot keep data owner and data analyst information at the time
of data sharing. Some recent solutions for protecting the data privacy of data owners
at the time of training a ML methods are secure k-nearest neighbor (k-nn) Haque et al.
(2020), secure support vector machine (SVM) Shen et al. (2019), and secure linear
regression (LR) Haque and Hasan (2021). All these methods consolidate Blockchain
to keep the transaction information of data owner and data analyst at the time of
ML training with encrypted IoMT data and achieve the best potential accuracy like
Overview of Blockchain-Based Privacy Preserving Machine Learning for IoMT 267
regular SVM, k-nn, and LR. However, secure SVM, secure k-nn, and secure LR need
several comparisons and calculations that result in higher space and time complexity
for analyzing the health data.
Research on cryptographic ML methods mostly focused on any specific domain
Sakuma and Kobayashi (2010), some do not consider Blockchain Saranya and
Manikandan (2013) and only a few sets of research utilizing Blockchain Zhu and
Li (2020). None of them are versatile as they are based on a specific setup and
address all the privacy concerns related to data integrity, authenticity, and privacy.
Some of the previous research discuss the flaws of PPML Al-Rubaie and Chang
(2019), which do not cover BPPML and also lack experimental analysis. This study
examines both cryptography and Differential privacy-based BPPML methods with
an experimental analysis and also suggest the best possible use for them. In this
study, Blockchain and a Cryptosystem (partially Homomorphic: Paillier) are applied
to address the above concerns when employing ML classifiers with IoMT data of the
owners. A distributed ledger is used to record the encrypted (Paillier) IoMT data of
each data owner. Encrypted data can be obtained from data owners by Data analysts
for employing the ML classifier. In the Privacy-Preserving DP-based ML method,
gaussian and laplacian noises are employed in this study. This study utilized two
security definitions to achieve security goals: modular sequential composition and
secure two-party computation Goldreich (2021); Canetti (2000).
3 Preliminaries
Machine Learning is a computer science domain where the system can learn without
any isolated programming Al-Rubaie and Chang (2019). ML methods’ prime goal
is to learn the process of achieving particular responsibilities by stereotyping from
data. Some of the tasks are prediction analysis and pattern recognition in data. ML
algorithms can be categorized depending on the learning style, such as supervised,
semisupervised, unsupervised learning. ML applications can be achieved in any
of the three mentioned above and do not rigidly belong to any methods, such as
recommender systems and dimensionality reduction. Figure 1 illustrates the three
types of ML methods.
268 R. U. Haque and A. S. M. T. Hasan
(a) (b)
(c) (d)
semisupervised unsupervised
Fig. 1 Machine Learning Tasks Overview. a Classification: determining a dividing line within the
two classes, “rectangle” and “circles”, b Regression: implementing a predictive design, c Clas-
sification: finding a separating dashed line among the three classes, “circles (black and white)”,
“rectangle” classes, where black circles are unsupervised data and white circles represent super-
vised data, d Clustering: classifying sample sets into several clusters
It uses labeled datasets. An output value (i.e., continuous value for regression and
class label for classification) is related to the individual feature vector. In the training
phase, that identified dataset is utilized in order to develop a model. That model can
forecast the label of the brand-new feature vectors at the testing phase. Classification
categorizes feature vectors into more than one class, and identifying the class of the
new sample is the main aim of the ML classification algorithm. For example, Fig. 1a
shows that a hyperplane is separating two classes. Various classification methods
can be utilized for categorization applications such as Logistic Regression, Neural
Networks, and Support Vector Machines, etc. If the label is dependent or a continuous
value in discrete value, then this situation is known as regression. The feature vectors
are known as independent variables. The main goal of the regression model is to fit a
forecasting model (i.e., line) with a labeled dataset in a way such that the gaps within
the line and the observed data points are lessened, which is shown in Fig. 1b.
Most real-world data are unlabeled as data labeling needs special devices and human
experts, making it costly. Scientific observation says that a tiny volume of labeled data
Overview of Blockchain-Based Privacy Preserving Machine Learning for IoMT 269
can significantly increase the training mechanism’s performance. Figure 1c shows the
process of training a model with the label and unlabeled data set.
It does not require any labeled dataset and each feature vector has no response vari-
able associated with it. This type of learning aims to find an internal structure from
the dataset. Figure 1d shows the working process of unsupervised learning. Cluster-
ing (K-means) is one of the most well-known unsupervised methods. It categorizes
feature vectors into various clusters. All similar samples should belong to a cluster,
and any distance measuring method (i.e., euclidian, manhattan, etc.) can be utilized
for the similarity test.
• M is the message space and C’s elements are all ciphertexts results, which are
encrypted by Enc pk .
• Decsk (σ (c1 , c2 )) = σ (m 1 , m 2 ) is held for any (m 1 , m 2 ) ∈ M, any c1 output by
Enc pk (m 1 ), and any c2 output by Enc pk (m 2 ).
Pr [K (D1 ) ∈ S]
≤ e
Pr [K (D2 ) ∈ S]
their raw data to its own smart contract, where the smart contract will apply Local
DP on the data set and send them to the smart contract of the data analyst. For better
privacy, participants can use private networks like Hyperladger.
3.4 Blockchain
System model, the threat model, and model construction are discussed in this section.
The employed systems aim to make sure privacy and protected sharing of data
between data analysts and data owners. Data analyst gets encrypted or DP-based
272 R. U. Haque and A. S. M. T. Hasan
intermediate
Interaction privet data sharing platform
Cryptographic
Model training
Computation
Analysis
privacy method
IoMT devices #1...#n
Fig. 2 System model: In this model data analysts want to train its ML model (with Cryptographic
and DP based privacy method) over the IoMT dataset of the data owners
IoMT data from each data owner. All the shared data are recorded in a distributed
ledger of Blockchain by forming transactions in order to keep authenticity. Data
analysts assemble its Ml model by acquiring recorded data from the public ledger
of Blockchain. Data analysts erect a protected method utilizing protected building
blocks (Secure Comparison and Secure Polynomial operation) for the Cryptographic
method and use the noise data for the DP-based method. At the time of employing
the secure ML model, it is important to have interactions among data owners and
data analysts in order to share results for the Cryptographic method. Figure 2 shows
the system model.
• IoMT Devices: are accountable for collecting and transferring important IoMT
data by wired and wireless networks.
• Data Owners: gather every data from the IoMT devices.
• Data Analyst: want to train a ML model upon the data-set gathered from multiple
Data Owners:.
In the employed model, participants do not esteem anyone, but themselves, this type
of participants can be formally termed as legitimate but inquisitive adversaries. To be
more specific, all participants in the system are legitimate in following the protocol
of the model but also strive to know the private data of other participants. The thread
model is discussed in this section. Again, the ML training protocol π , which is pre-
designed and data analyst is fair in obeying it. He has an interest in the private data
Overview of Blockchain-Based Privacy Preserving Machine Learning for IoMT 273
of other participants and also tries to gain further information from their data by
investigating. On the other hand, there might be an attempt from the data owner side
to recognize the model parameters of data analysts from the intermediate results.
This research assigned each data instance to the corresponding feature vectors, which
are preprocessed locally. To record encrypted or noisy datasets a transaction is main-
tained in the Blockchain. The employed transaction has two domains: inputs and
outputs. Input field has:
• Sender address
• Encrypted or DP-based data space
• Identity of source device (IoMT)
• Receiver address
• Encrypted or DP-based data space
• Identity of source device (IoMT)
This study employs the employed model in the Hyperledger Fabric platform. It is
known as a permissioned Blockchain platform. Sender and receiver’s addresses are
hash values. Paillier determines the encrypted data. The length of each entry is 128
bytes. The segment length is 4 bytes for the type of IoMT device. The sender node
assembling the transaction and broadcast it in the Peer to Peer (P2P) Angrish et al.
(2018) system of the Blockchain network. Operation’s correctness is validated by
the Miner nodes and transactions are packaged in a new block by a particular miner
node. Each block may record various transactions. Classical consensus protocols are
used to add blocks to the chain. i.e. the Proof of Work (PoW) mechanism Vukolić
(2015) or Byzantine Fault Tolerance (BFT) Stifter et al. (2019).
274 R. U. Haque and A. S. M. T. Hasan
This section presents the construction details of the employed k-nn model. Note
that we are just analysing proposing any novel method but analyzing the existing
Cryptographic and Differential Privacy based methods in order to identify the best
suitable algorithm for real-life. The goal these methods are to protect the privacy
of distinct data owner and data analyst while training an ML model over multiple
private datasets from various data owner.
Protected polynomial addition and subtraction are employed in order to train
the ML method in a cryptographic manner. Homomorphic property of Pailler is
derived as: [[m 1 + m 2 ]] = [[m 1 ]] × [[m 2 ]] mod N 2 and subtraction is derived
as: [[m 1 − m 2 ]] = [[m 1 ]] × [[m 2 ]]−1 mod N 2 .1 Secure polynomial multiplication
of Paillier is represented in Eq. 1.
δ
x = x + Lap( )
where, x noise data, x plain text, δ predefined sensitivity, privacy budget, and
function Laplace Lap( ).
Tools used for this study are: MacBook Pro (DDR3, 1600 MHz, 4 GB), 2.5 GHz,
processor: Intel Core i5, which acting as P and A concurrently. Secure Polynomeal
Operation and Comparison are developed in Google’s Collaboratory platform in
python 3 language. Breast Cancer Wisconsin data set (BCWD), Diabetes data set
(DD), and Heart disease data set (HDD) are the three health datasets are utilized in
this study Haque et al. (2020). BCWD represents breast mass and cell nuclei from
the images. Malignant and benign are used to label the dataset. On the other hand,
HDD and DD datasets have 9 numeric and 13 discrete attributes, respectively. For
1 [[m]]−1 represents the modular multiplicative inverse. It can calculate [[m]] × [[m]]−1 mod
N2 = 1 in Paillier. [[m]]−1 can be calculated by φ(N ) function, [[m]]−1 = [[m]]φ(N )−1 .
Overview of Blockchain-Based Privacy Preserving Machine Learning for IoMT 275
the employed system training and test, sets are divided into 80 and 20% respectively.
Table 1 illustrates the statistics of datasets.
Cryptosystems do not allow any floating-point operation so all the operations
must be done in whole numbers as a result format conversion from float to integer
become compulsory.2 One of the important aspects of asymmetric cryptosystem’s
security is the key length.
Three most popular metric, i.e., accuracy (2), precision (3) , recall (4).
t p + tn
accuracy = (2)
t p + tn + f p + f n
tp
pr ecision = (3)
tp + f p
tp
r ecall = (4)
t p + fn
Here, the positive or relevant classes has represented as t p . These classes are precisely
labeled. The negative or irrelevant class that are labeled correctly are represented as
f p . f n and tn represents the number of relevant but mislabeled and the number
mislabeled but irrelevant, respectively in the test result. The outcomes are shown in
Table 2.
Cryptographic methods provide robust privacy and also maintain the accuracy
of the dataset than any other algorithm. However, cryptographic methods’ main
limitations are that they are costly in terms of space and time complexity. These
methods work on integer formate, and in order to maintain that, lots of preprocessing
are required for formate conversion. Rakib Haque et al. (2020) applied BPPML
(k-nn) based on cryptographic method on three data sets containing less than 500
records and found that it took nearly one hour for each of the datasets. As a result,
in the case of a larger dataset, these methods become impractical. Table 2 shows the
performance comparison based on accuracy and time consumption between standard
k-nn and secure k-nn (i.e., Cryptographic and LDP). On the other hand, LDP-based
methods for BPPML can provide robust privacy based on privacy budget. The main
2 According to the global standard IEEE 754 representation of a floating-point binary number is D
is D = (−1)s × M × 2 E , where the sign bit is s, a significant number is M and exponential bit E.
276 R. U. Haque and A. S. M. T. Hasan
Table 2 Summary of performance for k-nn (t = 8). t stands for the threshold value of k-nn
Privacy budget of LDP
Measures Dataset Standard Cryptographic =0 =1 =3
Accuracy BCWD 96.96% 97.80% 93.55% 94.84% 96.54%
HDD 83.50% 82.33% 77.90% 79.50% 82.95%
DD 79.00% 78.00% 72.34% 76.12% 78.34%
Precision BCWD 96.54% 96.26% 87.84% 90.64% 95.94%
HDD 83.85% 82.30% 76.35% 79.15% 82.55%
DD 77.00% 76.00% 68.21% 72.54% 76.56%
Recall BCWD 96.85% 96.67% 89.33% 92.64% 95.85%
HDD 83.85% 82.66% 76.33% 79.55% 82.32%
DD 75.90% 75.10% 67.49% 71.75% 74.59%
Time BCWD 0.7 s 3357.2 s 2.1 s
HDD 0.4 s 2534 s 1.2 s
DD 0.2 s 3709 s 0.6 s
advantage of these kinds of methods is that they are not as expensive as cryptographic
methods in space and time complexity. A small amount of time is required to apply
LDP in the dataset, and the time consumption for model training is the same as the
original dataset. The main limitation is accuracy. LDP-based dataset never provides
the same accuracy as the original dataset because of the privacy budget. Less accuracy
confirms higher privacy, and higher privacy provides lesser accuracy. Moreover, DP
is more practical than cryptographic methods. Professional institutions like Apple,
Google, etc., already employed their own LDP methods.
6 Conclusion
Acknowledgements Authors thanks the school of computer science and technology of the Uni-
versity of Chinese Academy of Science, Beijing, China, and the Department of Computer Science
and Engineering of University of Asia Pacific, Dhaka, Bangladesh for their support towards this
study.
References
G.J. Joyia, R.M. Liaqat, A. Farooq, S. Rehman, Internet of Medical Things (IOMT): applications,
benefits and future challenges in healthcare domain. J. Commun. 12(4), 240–247 (2017)
I. Goodfellow, Y. Bengio, A. Courville, Machine learning basics. Deep Learn. 1, 98–164 (2016)
M. Al-Rubaie, J.M. Chang, Privacy-preserving machine learning: threats and solutions. IEEE Secur.
Priv. 17(2), 49–58 (2019)
R.U. Haque, A.S.M.T. Hasan, Q. Jiang, Q. Qu, Privacy-Preserving K-nearest neighbors training
over blockchain-based encrypted health data. Electronics 9, 2096 (2020). https://doi.org/10.3390/
electronics9122096
R.U. Haque, A.T. Hasan, Privacy-Preserving multivariant regression analysis over blockchain-based
encrypted IoMT data. Artif. Intell. Blockchain Future Cybersecur. Appl. 45
M. Abadi, A. Chu, I. Goodfellow, H.B. McMahan, I. Mironov, K. Talwar, L. Zhang, Deep learning
with differential privacy, in Proceedings of the 2016 ACM SIGSAC Conference on Computer and
Communications Security (ACM, New York, NY, USA, 2016), pp. 308–318
R. Bost, R.A. Popa, S. Tu, S. Goldwasser, Machine learning classification over encrypted data, in
NDSS, vol. 4324, p. 4325 (2015)
A.S.M. Hasan, Q. Qu, C. Li, L. Chen, Q. Jiang, An effective privacy architecture to preserve user
trajectories in reward-based LBS applications. ISPRS Int. J. Geo-Inf. 7(2), 53 (2018)
A.S.M. Hasan, Q. Jiang, H. Chen, S. Wang, A new approach to privacy-preserving multiple inde-
pendent data publishing. Appl. Sci. 8(5), 783 (2018)
A.S.M. Hasan, Q. Jiang, C. Li, An effective grouping method for privacy-preserving bike sharing
data publishing. Future Int. 9(4), 65 (2017)
A.T. Hasan, Q. Jiang, J. Luo, C. Li, L. Chen, An effective value swapping method for privacy
preserving data publishing. Secur. Commun. Netw. 9(16), 3219–3228 (2016)
M. Shen, X. Tang, L. Zhu, X. Du, M. Guizani, Privacy-preserving support vector machine training
over Blockchain-based encrypted IoT data in smart cities. IEEE Int. Things J. 6(5), 7702–7712
(2019)
J. Sakuma, S. Kobayashi, Large-scale k−means clustering with user-centric privacy-preservation.
Knowl. Inf. Syst. 25(2), 253–279 (2010)
C. Saranya, G. Manikandan, A study on normalization techniques for privacy preserving data
mining. Int. J. Eng. Technol. (IJET) 5(3), 2701–2704 (2013)
Y. Zhu, X. Li, Privacy-preserving k−means clustering with local synchronization in peer-to-peer
networks. Peer-to-Peer Netw. Appl. 13(6), 2272–2284 (2020)
O. Goldreich, Foundations of Cryptography: Volume 2, Basic Applications (Cambridge University
Press, 2009)
R. Canetti, Security and composition of multiparty cryptographic protocols. J. Cryptol. 13(1), 143–
202 (2000)
V. Aleksieva, H. Valchanov, A. Huliyan, Implementation of smart-contract, based on hyperledger
fabric Blockchain, in 2020 21st International Symposium on Electrical Apparatus and Technolo-
gies (SIELA), Bourgas, Bulgaria, pp. 1–4, (2020). https://doi.org/10.1109/SIELA49118.2020.
9167043
C. Dwork, Differential privacy, in Automata, Languages and Programming. ed. by M. Bugliesi, B.
Preneel, V. Sassone, I. Wegener (Springer, Berlin, Heidelberg, 2006), pp. 1–12
278 R. U. Haque and A. S. M. T. Hasan
Aitizaz Ali, Muhammad Fermi Pasha, Ong Huey Fang, Rahim Khan,
Mohammed Amin Almaiah, and Ahmad K. Al Hwaitat
Abstract In digital healthcare systems, the patients face significant problems iden-
tifying an optimal vacant and available slot for their appointment as the number of
patient requests directly correlates with the slot availability. Existing smart health-
care systems facilitate the patient to reserve a particular time slot and attain real-time
healthcare information. However, most of these systems need sensitive informa-
tion about patients, i.e., desired destination of the healthcare providers. Moreover,
existing systems utilize a centralized system which makes these systems vulnerable
to numerous intruders’ attacks and security breaches (particularly related to service
providers) and results in single-point failure of the entire system. In this paper,
Ring Signature for permissioned Block-chain-based Private Information Retrieval
scheme is proposed to improve privacy-preserving in the smart healthcare system.
Our proposed Scheme initially utilizes an improved multi-transaction mode consor-
tium block-chain constructed by different numbers of requests to the healthcare
providers to achieve maximized appointment offers based on availability, trans-
parency, and security. Our proposed Scheme is useful for information retrieval from
multiple domains. The simulation results verify that the proposed Scheme provides
a predominant performance in ensuring maximized patient privacy with reduced
computation and communication overheads.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 279
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_13
280 A. Ali et al.
1 Introduction
In the recent past, the rapid growth in digital healthcare in the big cities has intro-
duced the central issue for the patients to determine the vacant appointment space
for patient examination (Garson and Adams 2008). This process of identifying an
optimal space of booking in smart healthcare systems is considered to contribute to
about 30% of the traffic congestion (Griffiths 2005). Moreover, one report presented
that nearly millions of security breaches happened to patient health records each day.
This security breach is considered to be increased due to an unsecured access control
system (Gronwald and Zhao 2018). Hence, the network transaction time, jitter, and
traffic congestion need to be prevented by integrating technology that could aid them
in determining the suitable vacant slots in the smart healthcare system (Almaiah and
Al-Khasawneh 2020). In this context, the smart healthcare system is emerging as
a predominant solution for suitably identifying the appointment slot with the new
advancements happening in the Internet of Things (IoT) and wireless communica-
tions. IoT devices are deployed in every individual hospital in the smart healthcare
system, such that ultrasonic sensors are used to identify the availability of the appoint-
ment slot (Adil et al. 2020). Thus, it facilitates the healthcare service providers with
the status of occupancy associated with appointment spaces (Adil et al. 2020). The
healthcare service provider enables the option of verifying whether the appointment
space is available for online reservations, such that the patient can determine the
availability of the vacant appointment slots. Despite the benefits mentioned above,
implementing a smart healthcare system imposes a diversified number of challenges
that need to be handled before its wide implementation (Adil et al. 2020).
One of the major issues in the smart healthcare system is the privacy of information
associated with the patients (Khan et al. 2020). Most current approaches require
the patients to disclose sensitive information to the service providers about their
reservation times, destinations, and real identities (Al Hwaitat et al. 2020). This
disclosure of information by the patients aids the healthcare service provider to
perceive and interpret life patterns and daily activities such as income level, health
condition, home address, and work address depending on the data analysis attained
with the background information (Adil et al. 2020). The majority of the current smart
appointment systems are generally centralized and are identified to suffer from many
limitations (Almaiah et al. 2020). First, the current smart appointment systems are
vulnerable to data loss and security breaches when sensitive information such as
name, phone number, email address, and daily appointment information associated
with the patients are stored in the database (Qasem et al. 2021). Second, they are
highly and inherently prone to the problem of single-point failure. In addition, they
also have the risk of remote hijacking attacks and distributed denial of service (DDoS)
attacks that can convert the appointment services unavailable (nd).
In this paper, Ring Signature and Improved Multi-Transaction Mode Consor-
tium Blockchain-based Private Information Retrieval (RS-IMTMCB-PIR) Scheme is
proposed for Privacy-Preserving in Smart Appointment System. This RS-IMTMCB-
PIR Scheme is proposed for preventing the attacks that threaten appointment service
Big Data Based Smart Blockchain for Information … 281
2 Related Work
unprivileged users’ access and enhance privacy via a channel formation scheme. The
issue in this approach is the computational cost due to the complex cryptographic
mechanism. In their research work, Jain and Singh (2020) designed a clinical data
exchange system based truly on the blockchain and later developed a series of veri-
fication mechanisms for improved security and privacy. Li et al. (Ghanbari 2019)
suggested a healthcare framework for data privacy and integrity through novel algo-
rithms for memory management that help data management. Fan et al. (2020) devised
blockchain-based health information of patients through efficient consensus mech-
anisms to provide an efficient security and privacy framework. The problem with
this approach was a lack of flexibility and a cross-organizational approach. Pandey
et al. (2020) provide better security to record and share patient clinical information
using an attribute-based encryption scheme. In this approach, they integrated SC to
ensure reliability and to provide a facility to monitor the PHR sensitive data. Guo
et al. (2020) designed an attribute-based signature (ABS) method for clinical users
to manage medical health records truly based on blockchain. In this technique, the
objective was to achieve optimum privacy of the delivered model through a P2P tech-
nique for efficient privacy. Uddin et al. (2019) proposed a framework to monitor and
trace patient history-related data. This system relies on a node controlled by autho-
rized patients in the main module of the system; through this approach, the author
got better security of the system through various experiments and variation in data
sets. This approach has computational overhead due to its complex mechanism for
monitoring. Ghanbari et al. (2019) introduced a distributed attribute-based signature
scheme for medical systems founded on peer-to-peer ideas. They also developed P2P-
based records sharing protocol that support algorithms. But the main problem is that
this model uses a signature system that can’t predict collision and social engineering
attacks. Choo et al. (2020) highlighted recent research challenges and developed an
AC strategy for electronic health records (EHR) through a fine-grained access control
system. Pandey et al. (2020) Proposed an EHR system to provide security to clinical
records using distributed ledger technology (DLT). Further, the author has justified
an improved cross-organization sharing of health records using access control poli-
cies. Thakkar et al. (2020) designed a novel approach to measure the framework’s
efficiency using blockchain. Performance optimization related to caching and config-
uration authorization policy were achieved through this approach. Sukhwani (2019)
evaluated the performance parameters in a novel way for the hyperledger fabric
framework. Gorenflo et al. (2020) provided a detailed method for analyzing and eval-
uating the optimization for the provenance and performance of the blockchain frame-
work and developed a framework by configuring it to reduce input/output. But this
approach was considered complex computationally. They achieved enhanced perfor-
mance by reducing the computation time. Last but not least, Chen et al. (2019) proved
a novelty in designing the secure, searchable encryption (SSE) model for electronic
healthcare records using blockchain. The algorithm takes the index of healthcare
records as an input and then provides a related search using SSE. Comparative anal-
ysis of the state-of-the-art blockchain-based approaches for PHR systems is carried
out in Table 1 points using blockchain for the clinical environment. The approach
support is detecting security issues without affecting scalability. The problem with
Big Data Based Smart Blockchain for Information … 283
Table 1 Response time and the transaction time of the proposed approach
Number Proposed Medrec SHealth ECC-SS
of users framework
Time PR(i) Time PR(i) Time PR(i) Time PR(i)
Size of the 2 102 1 154 1 279 1 712 1
Block-10000 4 104 1.019 156 1.012 281 1.007 714 1.002
6 105 1.028 157 1.019 283 1.014 715 1.005
8 107 1.049 159 1.032 284 1.017 718 1.008
10 108 1.077 160 1.038 286 1.025 719 1.009
12 110 1.098 161 1.045 287 1.028 721 1.012
14 112 1.006 162 1.051 288 1.032 723 1.015
16 114 1.117 164 1.064 290 1.039 724 1.016
Irving et al. is that its security and computational cost are high, and its performance
is poor in case of security breaches such as collusion and phishing attack. Yazdinejad
et al. proposed a blockchain-based method to facilitate verification to 5G networks.
This method was based on software-defined networking (SDN). The main advantage
of this technique is that they eliminate the cost for Re-authentication when devices
usually exchange in between cells in 5G networks.
A Smart Appointment systems-based on Non-fungible token was proposed for
construction of appointment pool that introduces transparent platform for the patients
and the healthcare service providers in determining the available space with reduced
amount of time (Viale Pereira 2017; Manjula and Rajasekaran 2020; Niranjana-
murthy et al. 2019; Kumar and Mallick 2018). This appointment system included
non-fungible token for introducing maximized transparency into the complete system
(Hassanien et al.2018; Pavithran et al. 2020; Sultan et al. 2019; Karthikeyyan and
Velliangiri 2019). It was proposed as a method to derive added advantage of revenue
generated from the vacant space with the benefit of smart contracts. However, the
overheads of communication and computation were considered to be comparatively
compared to the conventional appointment approaches. A Consortium Blockchain-
based Smart Appointment and booking was proposed for achieving maximized avail-
ability, transparency and security by storing appointment slots inside theblockchain
nodes (Alotaibi 2019; Khan and Salah 2018; Wattenhofer 2016; Yu et al. 2019;
nd). It was proposed with the techniques of cloaking for hiding the locations of the
patients in order to protect their location privacy. The cloaking mechanism provided
the blockchain with the potential of validating and returning the available appoint-
ment offers that possibly exists the cloaked area. It provided the patients with the
option of selecting the optimal appointment offers such that direct reservation can
be attained. It was developed for preventing the time incurred during the search
of available appointment space performed a blockchain-based systematic review
of IoT application and its belonging security problems. Depending on a regulated,
assessed, and useful content analysis of the academic literature, they evaluated the
extended categorization of blockchain-depended IoT applications via diverse parts
284 A. Ali et al.
like smart health care, industrial, smart home, smart city, electronic business model,
ad-hoc vehicle networks, and supply chain management. Further, they demonstrated
the security difficulties of discussed IoT applications, basically restrictions of the
IoT technology. Given that, their study was not done systematically. Therefore, this
limitation has provided the context for this research.
Further, Elliptic Curve Cryptography (ECC)-based smart appointment Systems
(ECC-SMS) was proposed for protecting patients privacy based on the cumula-
tive prevention of attacks imposed over confidential information exchange (Hur
and Noh 2010). This ECC-SMS facilitated users’ privacy based on the adoption
of zero knowledge proofs integrated with the implementation of ECC algorithm.
The performance of this ECC-SMS was determined to reduce the network overhead
and execution based on their investigation attained with real world test-bed prepared
with the available hardware platform. An appointment space sharing mechanism
using blockchain was proposed with the name SHealth for sharing and publishing
the availability of appointment slots by the service providers without disclosing the
sensitive information of patients (Jalan 2018). This SHealth was developed with
incentives that are provided to the patients depending on the rates and the appoint-
ment time with anonymity. It provides data provenance using blockchain in order to
handle the privacy and security issues that are most possible during smart appoint-
ment of patients. The blockchain is used in this SHealth for enabling transaction
between the patient and doctor that share sensitive information with proofs of images
captured during appointment and leaving in a method of tamper-proof. The access
control mechanism provided by this SHealth prevented the inclusion of any third party
that has the possibility of data utilization. Bloom filters-based privacy preservation
retrieval Scheme (BF-PPRS) was proposed with the merits of specific application
navigation result in order to prevent attacks launched over the information shared by
the patients during appointment process (Pflanzner and Kertész2018). This BF-PPRS
included the cloud resources in identifying theavailable appointment spaces with
maximized anonymity. In specific, Bloom filters were incorporated for supporting
and retrieving the navigation results of the interaction of a specific application or
entity with the service providers. It facilitated the patients to access the cloud in esti-
mating the available appointment slots with anonymity in accessing the generated
query. It is considered to retrieve the encrypted navigation results effectively based on
the utilization of certain units and metrics. The results of this BF-PPRS were deter-
mined to facilitate potential appointment navigation with maximized probability of
retrieval with low communication and computation overhead.
Most of the researchers surveyed and reviewed the basic security problems of the
IoT. They categorized these problems depending on the low-level, intermediate-level,
and high-level layers of the IoT. They briefly discussed the procedures offered in the
literature to leverage the security of IoT at diverse stages. Additionally, a parametric
analysis of IoT attacks and the belonging feasible remedies have been supplied. They
considered the attack indications and mapped them to feasible solutions suggested in
the literature. They also discussed the way blockchain is able to be utilized for solving
some relevant IoT security issues. They outlined and identified security in IoT, but
the role of blockchain and the importance of its security in the IoT is discussed very
little.
Big Data Based Smart Blockchain for Information … 285
3 Proposed Framework
In this section, the complete details about the Improved multi-transaction mode
consortium blockchain constructed by appointment slot owners, improved private
information retrieval approach utilized for secretly retrieving appointment offers
of the improved multi-transaction mode consortium blockchain and Ring signature
potential anonymous authentication process during the online reservation of available
appointment slots reservation is presented.
Fig. 2 Part a The comparative analysis of our proposed Schemes communication overhead with
the benchmark part b Comparative analysis of proposed framework with the benchmark models
having number of rounds −9000
regulatory node system used in the design of MTMCB used in the proposed scheme.
The regulatory node system is included in the server-side of the healthcare service
providers of the smart appointment system for achieving the process of initializa-
tion, auditing, and transaction processing. However, the operation of initialization is
performed only once.
On the other hand, transaction nodes can represent the mobile devices, PCs, and
payment machines that could be used to check the availability of appointment space,
perform online reservations, and complete payment to the service providers for the
appointment space allocated to the Patients (Pandey et al. 2020). In this context, the
block name file related to the MTMCB blockchain represents the specific structure
of the system file managed by the regulatory node. In the block name file, the Name
of the block and the associated addresses of the complete set of patients performing a
transaction for online reservation of appointment space with the help of the block are
stored simultaneously. Figures 2 and 3 present the user set and block name set logical
storage structure. The performance of reading over the disk storage cannot satisfy the
requirements of the fast retrieval when the rapid retrieval frequency of the user file and
block name files are considered. When the system is initiated, the user and the block
name files are read, parsed semantically, and stored in the Redis cache. This Redis is
a NoSQL database technology that is developed based on the structure of memory
key value. The data can be cached in the memory, and master–slave replication can
be significantly realized by constructing the Redis database cluster. This realization
allows identifying the frequently accessed data from the memory to minimize the time
incurred in the data query. Redis maps the logical association between one or multiple
key-value pairs. Rather, the classical relational database maps the logical mode to
the table’s series. Thus, the selection concerning key structure is vital in the Redis
cache. A potential design of Redis cache is determined to enhance the query efficiency
Big Data Based Smart Blockchain for Information … 287
Fig. 3 a Proposed RS-IMTMCB-PIR-Response Time and delay with different number of blocks,
b ROC curve for our proposed blockchain based privacy preserving framework
with reduced memory overhead. However, the computational complexity of the Redis
cache-based private information retrieval may increase with a corresponding increase
in the number of users and transactions performed by the users. Hence, B+-tree is
integrated with Redis cache for handling computational overhead by reconstructing
the block name structure, and user set logical structure with the view to enhance
the response time and retrieval rate of the Private information in smart appointment
systems.
Admin Node.
Our proposed algorithm for admin node is mentioned below: It consist of three main
modules i.e. Input, Output and Initialization.
Input: Enrolment Certificate (E C ) requested from Certification Authority (C A ).
Output: Access to PHL , C HL and L HL transactions for all (PHL , C HL , L HL ) BN .
Initialization: N Admin should be valid node. N Admin can
Write/Read/Update/Remove nodes C ID , PID , L ID .
288 A. Ali et al.
1:procedure Admin(PID,CID,LID)
2: while (True) do
3: if (CIDis valid) then
4: Add_Clinician to the blockchain Network
5: Add_Clinician(BN,CID)
6: Grant_access(CID, UName, PK)
7: else
8: Not_exist(CID)
9: end if
10: if (PIDis valid) then
11: Add Patient to the blockchain Network
12: Add_Patient(BN,PID)
13: grant_access(PID, UName, PK)
14: else
15: Not_exist(PID)
16: end if
17: if (LIDis valid) then
18: Add Lab to the blockchain Network
19: Add_Lab(BN,LID)
20: grant_access(LID, UName, PK)
21: else
22: Not_exist(LID)
23: end if
24: end while
25: int N; {0 means bad behavior, 1 means good behavior}
The private information retrieval method is initiated on the MTMCB from the user
name (patient’s Name) and the block files associated with the user name for the
party to perform the appointment transaction. In order to achieve the Enhanced
Storage Model of Improved Redis Cache, the Name of the user is replaced with User
Transaction Identifier” as the key, and the combination of names corresponding to
block files is set to the value of the key. The logical storage structure of this User
Block Collection is depicted in Fig. 1.
Step 1: When the smart appointment system starts its operation, read and explore
function is imposed over “Block Name File” by initializing the “User Block Collec-
tion” in the Redis cache based on Name of the Block and User Transaction Identifier
existing in the block name file.
Step 2: Generate the B+-Tree index associated with “User Block Collection”
to accomplish the relation between the “User Block Collection” and the generated
index.
Step 3: Another new B+-Tree index is created to represent the user’s files as
presented in step 2 with the keyword as “Name of the User”, when the user collection
increases with low potential sequential retrieval process.
In this phase, information related to the Blockname is recorded in the “User Block
Collection” based on information derived from the transaction participants. The steps
involved in the construction of User Block Collection are presented as follows.
Step 1: Generate a block file for recording the current block contents when it is
stored in the block file, and it is named based on the blocks’ hash value.
Step 2: Choose a single participant of the transaction and identify its user trans-
action identifier from the B+-Tree index with the user’s name derived from the user
collection when one or more users participate in the appointment allocation process.
Step 3: Identify the key from the user block file based on the user transaction
identifier of the participating users. If the user transaction identifier exists, then the
value is extended. Further, the block’s name is written at the end of the value to sort
them based on the order of timestamp. On the other hand, if it does not exist, then
key-value metadata is appended. Finally, the key is the user transaction identifier,
file name, and updated B+-tree index.
Step 4: Until the complete set of users complete their processing and perform the
extension operation, repeat steps 2 and 3 for processing the successive transaction
user.
Once the user block collection is defined, then the algorithm of block-based private
information retrieval in the proposed framework completely prevents Block Name
File and tries to accomplish the relation between the “User Block Collection” and
“User Set.“ For instance, let us consider that a smart appointment system requires to
query the complete block files when two patients participate in the online reservation
of appointment space. In this scenario, the service provider can rapidly access all
block files associated with the participating patients (users), then the collection of
block file names related to the two patients addresses can be determined from the
290 A. Ali et al.
user block file as long as the associated addresses of associated patients are identified
through the user file.
Step 1: The retrieval method identifies the user transaction identifier from the
user’s name by the user set accessing from the name of one or more users participating
in the smart appointment system.
Step 2: Utilize B+-Tree to identify the key location that stores the user address
derived from the user block set for determining the complete records of block file
name associated with the users.
Step 3: Repeat Step 2 until the complete set of block file names relevant to the
transaction’s participants existing in the records are identified.
Step 4: A block file associated with the users is obtained as the target file output
of this query based on the block name file record set determined in steps 2 and 3,
respectively.
indexing played a vital role in eliminating the factors that delay the retrieval of
private information users when required by the service providers. The communi-
cation overhead of the proposed RS-IMTMCB-PIR in remote information retrieval
with a different number of blockchain nodes. In the first part of the investigation,
the proposed RS-IMTMCB-PIR and the benchmarked MBO-SMS, CB-SMS, and
ECC-SMS approaches are evaluated based on communication overhead in private
information retrieval with different parking numbers allocation available in each
cell and number of blockchain nodes. In this experiment, communication overhead
refers to the transmitted message size (in bytes) exchange between the driver and
the blockchain nodes in the retrieval phase and between the driver and the parking
lot service provider in the reservation phase. Illustrates the communication over-
head of the proposed framework in in private information retrieval with different
numbers of appointment allocation available in each cell. The communication over-
head of the proposed framework is considerably minimized even when the number
of parking allocations in the unit cell increases since it can handle the required
number of retrieval by storing in the B+-Tree indexing data structure that aids better
performance. Hence, the communication overhead of the proposed scheme in private
information retrieval with a different number of appointment allocation available in
each is reduced by 5.24, 6.84, and 7.54%, excellent to the benchmarked Medrec,
Shealth, and ECC-SS approaches. Demonstrates the communication overhead of
the proposed framework in private information retrieval with a different number of
blockchain nodes. Even when the numbers of blockchain nodes are increased, the
communication overhead of the proposed framework is considerably minimized as
Redis cache-based indexing played a vital role in eliminating the factors that delay
the retrieval of private information users when required by the service providers.
In Table 1 is shown clearly the response time and the transaction time of the
proposed approach and the bench mark framework.
In Fig. 2 is shown the simulation results of our proposed framework. It’s very
clear that for Blockchain size 10,000 the efficiency of our proposed framework is
better than the benchmark models (Table 2).
Further, Figs. 3, 4, 5 present the relative proportion of retrieval of our proposed
framework and the benchmarked Medrec, SHealth and ECC-SS approaches with
different users and block sizes of 2500, 5000, 7500 and 10,000, respectively. The
relative proportion of retrieval achieved by our proposed framework is considered to
be highly improved, even when the number of blocks in the blockchain gets increased
from 2,500 to 10,000.
In Figs. 2 and 3 is shown the simulation for the attribute of Blockchain having
size 7,500.
From Figs. 2 and 3 it’s very obvious that the efficiency of the proposed approach
is better than the bench mark model. The efficiency is calculated according to the
response time (RT ) and delay jitter (Dt ). As compared to the bench mark model the
efficiency of the proposed framework increases as the number of nodes increases up
to 7500. The limit for our proposed framework is up to 7500 nodes (Fig. 6).
In Figs. 4 and 5 are explained the simulations results of the number of blocks
and the communication overhead that are carried out during the transaction from
292 A. Ali et al.
Table 2 Retrieval time of proposed system with different users and block size-7500
Number Proposed Medrec SHealth ECC-SS
of users framework
Time PR(i) Time PR(i) Time PR(i) Time PR(i)
Size of the 2 94 1.000 135 1.000 248 1.000 612 1.000
Block-7500 4 96 1.021 137 1.014 250 1.008 615 1.004
6 97 1.032 139 1.029 253 1.02 616 1.011
8 99 1.053 140 1.037 255 1.028 619 1.014
10 100 1.064 142 1.051 258 1.04 621 1.017
12 102 1.085 143 1.059 259 1.044 623 1.018
14 104 1.006 145 1.074 261 1.052 624 1.019
16 105 1.117 147 1.088 262 1.056 625 1.017
Fig. 4 a Efficiency of the proposed system with different users and block size-7500, b Comparative
analysis of our proposed framework with the benchmark models
one node and to another nodes respectively. In Figs. 4 and 5 the number of nodes
that we have simulated ranges to 5000 respectively. From the representation of the
color i.e. Blue bar, green bar, yellow bar and red bar respectively represents different
framework and the overhead time. The blue color represents our proposed framework
which has least communication overhead as compared to the benchmark models.
Figure 7 represents the simulation of our proposed framework for the number of
nodes up to 5000. From this simulation diagram its very clear that the efficiency of
our proposed framework is more as compared to the benchmark modes. The number
of nodes are 5000 in this case.
Fig. 7 Efficieny of the proposed framework with the number of blockchain modes (5000)
294 A. Ali et al.
5 Conclusion
References
K. Garson, C. Adams, Security and privacy system architecture for an e-hospital environment, in
Proceedings of the 7th Symposium on Identity and Trust on the Internet (2008), pp. 122–130
N. Griffiths. Task delegation using experience-based multi-dimensional trust, in Proceedings of the
Fourth International Joint Conference on Autonomous Agents and Multiagent Systems (2005),
pp. 489–496
M. Gronwald, Y. Zhao, Crypto currencies and currency competition, in Has Hayek Been Too
Optimistic? (2018)
M.A. Almaiah, A. Al-Khasawneh, Investigating the main determinants of mobile cloud computing
adoption in university campus. Educ. Inf. Technol. 25(4), 3087–3107 (2020)
M. Adil, R. Khan, M.A. Almaiah, M. Al-Zahrani, M. Zakarya, M.S. Amjad, R. Ahmed, MAC-
AODV based mutual authentication scheme for constraint oriented networks. IEEE Access 4(8),
44459–44469 (2020)
M. Adil, R. Khan, M.A. Almaiah, M. Binsawad, J. Ali, A. Al Saaidah, Q.T.H. Ta, An efficient
load balancing scheme of energy gauge nodes to maximize the lifespan of constraint oriented
networks. IEEE Access 8, 148510–148527 (2020)
M. Adil, M.A. Almaiah, A. Omar Alsayed, O. Almomani, An anonymous channel categorization
scheme of edge nodes to detect jamming attacks in wireless sensor networks. Sensors. 20(8),
2311 (2020)
M.N. Khan, H.U. Rahman, M.A. Almaiah, M.Z. Khan, A. Khan, M. Raza, M. Al-Zahrani, O.
Almomani, R. Khan, Improving energy efficiency with content-based adaptive and dynamic
scheduling in wireless sensor networks. IEEE Access 25(8), 176495–176520 (2020)
Big Data Based Smart Blockchain for Information … 295
M. Adil, R. Khan, J. Ali, B.H. Roh, Q.T. Ta, M.A. Almaiah, An energy proficient load balancing
routing scheme for wireless sensor networks to maximize their lifespan in an operational
environment. IEEE Access 31(8), 163209–163224 (2020)
M.A. Almaiah, Z. Dawahdeh, O. Almomani, A. Alsaaidah, A. Al-khasawneh, S. Khawatreh, A
new hybrid text encryption approach over mobile ad hoc network. Int. J. Electr. Comput. Eng.
(IJECE) 10(6), 6461–6471 (2020)
M.H. Qasem, N. Obeid, A. Hudaib, M.A. Almaiah, A. Al-Zahrani, A. Al-khasawneh, Multi-agent
system combined with distributed data mining for mutual collaboration classification. IEEE
Access (2021). Accessed 20 Apr 2021
M.A. Almaiah, A. Al-Zahrani, O. Almomani, A.K. Alhwaitat, Classification of cyber security
threats on mobile devices and applications, in Artificial Intelligence and Blockchain for Future
Cybersecurity Applications vol. 107
M.A. Almaiah, A new scheme for detecting malicious attacks in wireless sensor networks based
on blockchain technology, in Artificial Intelligence and Blockchain for Future Cybersecurity
Applications, vol. 217
M.A. Almaiah, M.M. Alamri, Proposing a new technical quality requirements for mobile learning
applications. J. Theor. Appl. Inf. Technol. 96(19) (2018)
H. Halpin, M. Piekarska, Introduction to security and privacy on the Blockchain, in 2017 IEEE
European Symposium on Security and Privacy Workshops (EuroS&PW) (2017), pp. 1–3
J. Hur, D.K. Noh, Attribute-based access control with efficient revocation in data outsourcing
systems. IEEE Trans. Parallel Distrib. Syst. 22(7), 1214–1221 (2010)
R. Jalan, R.W.L. Szeto, S. Wu, Systems and methods for network access control. Accessed 2 Oct
2018. US Patent 10,091,237
T. Pflanzner, A. Kertész, A taxonomy and survey of IoT cloud applications. EAI Endorsed Transac-
tions on Internet of Things 3(12) (2018), p. Terjedelem: 14 p.-Azonosító: e2. Li, X., et al., Energy
consumption optimization forself-poweredIoT networks withnon-orthogonal multiple access. Int.
J. Commun. Syst. 33(1), e4174 (2020)
S.-L. Peng, S. Pal, L. Huang, Principles of Internet of Things (IoT) Ecosystem: InsightParadigm
(Springer, 2020)
A. Ghani, et al., Security and key management inIoT-basedwireless sensor networks: Anauthenti-
cation protocol using symmetric key. Int. J. Commun. Syst. 32(16), e4139 (2019)
P. Azad, et al., The role of structured and unstructured data managing mechanisms in theInternet of
things. Clust. Comput. 1–14 (2019)
Z. Ghanbari et al., Resource allocation mechanisms and approaches on the Internet of Things. Clust.
Comput. 22(4), 1253–1282 (2019)
K.-K.R. Choo, A. Dehghantanha, R.M. Parizi, Blockchain Cybersecurity, Trust and Privacy
(Springer, 2020)
Z. Lv, Security of Internet of Things edge devices (Software, Practice and Experience, 2020)
N. Dong, G. Bai, L.C. Huang, E.K.H. Lim, J.S. Dong, A blockchain-based decentralized booking
system. Knowl. Eng. Rev. 35 (2019)
H. Saini, et al., Innovations in Computer Science and Engineering: Proceedings of the Sixth
ICICSE2018, vol. 74 (Springer, 2019)
M. Niranjanamurthy, B. Nithya, S. Jagannatha, Analysis of Blockchain technology: pros, consand
SWOT. Clust. Comput. 22(6), 14743–14757 (2019)
N.M. Kumar, P.K. Mallick, Blockchain technology for security issues and challenges in IoT.
Procedia Comput. Sci. 132, 1815–1823 (2018)
A.E. Hassanien, N. Dey, S. Borra, Medical big data and internet of medical things: Advances,
challenges and applications (CRC Press, 2018)
D. Pavithran, et al., Towards building a blockchain framework for IoT. Clust. Comput. 1–15 (2020)
A. Sultan, M.A. Mushtaq, M. Abubakar, IOT security issues via blockchain: a review paper, in
Proceedings of the 2019 International Conference on Blockchain Technology (2019)
296 A. Ali et al.
P. Karthikeyyan, S. Velliangiri, Review of Blockchain based IoT application and its securityissues,
in 2019 2nd International Conference on Intelligent Computing, Instrumentation and Control
Technologies (ICICICT) (IEEE, 2019)
B. Alotaibi, Utilizing blockchain to overcome cyber security concerns in the internet of things: a
review. IEEE Sens. J. 19(23), 10953–10971 (2019)
R. Wattenhofer, The Science of the Blockchain (Create Space Independent Publishing Platform,
2016)
K. Gopalakrishnan, Security Vulnerabilities and Issues of Traditional Wireless Sensors Networksin
IoT, in Principles of Internet of Things (IoT)
K. Gopalakrishnan, Security vulnerabilities and issues of traditional wireless sensors networks in
IoT , in Principles of Internet of Things (IoT) Ecosystem: Insight Paradigm. (Springer, Cham,
2020), pp. 519–549
A.K. Al Hwaitat, M.A. Almaiah, O. Almomani, M. Al-Zahrani, R.M. Al-Sayed, R.M. Asaifi,
K.K. Adhim, A. Althunibat, A. Alsaaidah, Improved security particle swarm optimization (PSO)
algorithm to detect radio jamming attacks in mobile networks. Quintana 11(4), 614–624 (2020)
M.A. Almaiah, M. Al-Zahrani, Multilayer neural network based on MIMO and channel estimation
for impulsive noise environment in mobile wireless networks. Int. J. Adv. Trends Comput. Sci.
Eng. 9(1), 315–321 (2020)
A. Jain, T. Singh, Security challenges and solutions of IoT ecosystem, in Information and
Communication Technology for Sustainable Development. (Springer, 2020), pp. 259–270
M.A. Khan, K. Salah, IoT security: review, blockchain solutions, and open challenges. Futur. Gener.
Comput. Syst. 82, 395–411 (2018)
V. Manjula, R.T. Rajasekaran, Security vulnerabilities in traditional wireless sensor networks by an
intern in IoT, blockchain technology for data sharing in IoT, in Principles of Internet of Things
(IoT) Ecosystem: Insight Paradigm. (Springer, 2020), pp. 579–597
P. Pandey, S.C. Pandey, U. Kumar, Security issues of internet of things in health-caresector: an analyt-
ical approach, in Advancement of Machine Intelligence in Interactive Medical Image Analysis.
(Springer, 2020), pp. 307–329
G. Viale Pereira et al., Increasing collaboration and participation in smar city governance: across-
case analysis of smart city initiatives. Inf. Technol. Dev. 23(3), 526–553 (2017)
J.H. Yu, J. Kang, S. Park, Information availability and return volatility in the bitcoin market:
analyzing differences of user opinion and interest. Inf. Process. Manage. 56(3), 721–732 (2019)
Classification of Malicious and Benign
Binaries Using Visualization Technique
and Machine Learning Algorithms
Abstract Malware detection and classification are crucial with malware evolution
these days. More and more malware attacks are targeting various environments like
businesses and hospitals. Our main goal is to deliver an intelligent classifier as fast
as efficient able to classifier binaries into malicious or benign classes. In this work,
we consider a malware-benign classification using machine learning algorithms by
converting binary executable files into grayscale images. The concept of the visu-
alization gives powerful results in the malware family classification topics. In this
context, we adopt visualization technique for both benign and malicious samples.
So, a hybrid feature extraction method is implemented using local and global image
features by combining DAISY and HOG features. Then a comparative study of
machine learning algorithms leads us to a final efficient classifier that reaches an
accuracy of 97,36% using Random Forest classification algorithm.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 297
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_14
298 I. Ben Abdel Ouahab et al.
that, we notice beginners’ learners of hacking doing for fun malicious attacks. Others
are also having ideological background or because of grudge. Understanding, why
a hacker may target you, is the first step for defense. It help knows weaknesses and
vulnerabilities of any infrastructure.
Recently, cyberattacks are continuing to grow in sophistication and scale. Espe-
cially, in the period of the covid19 pandemic that has affected the surface of attacks
where cybercriminals are leading to probably a cyber-pandemic. In December 2020,
hackers broke into SolarWind’s systems and added a malware into the company’s
software system. The targeted system is being used by 33,000 customers (‘Inline
XBRL Viewer’ 2021). The malicious code created a backdoor to customer’s infor-
mation technology systems; moreover hackers installed more malware that helped
then spy on the SolarWind’s customers companies and organizations.
More recently, in February 2021, hackers are getting more inhuman by targeting
French hospitals in a critical sanitary period. A wave of ransomware attacks hit
two hospitals in less than week in France, prompting the transfer of some patients
to other facilities. The ransomware used is a crypto-virus called RYUK. Firstly,
the said ransomware attacks Dax hospital and then five days later hackers target
the Villefranche-sur-Saône hospital complex in France’s eastern Rhone. Meanwhile,
France’s cybersecurity agency has revealed that a weakness in software from the
company Centre on led to the defenses of several French companies being breached
over a period of three years.
New variants of malware continue to increase every month. In March 2021, AV-
TEST institutions have recorded 19.20 million of new malware (‘Malware Statistics
Trends Report | AV-TEST’ 2021). These variants are becoming more dangerous, and
making data protection a hard work. For instance, fake update is the new trend of
distributing a malware in a legitimate looking. The worst thing is that anti-malware
solutions are insufficient and can’t recognize new and zero-day malwares especially
if it presented in an update of a well-known provider name. Also, there are news
malware attacks. This type is frequently used with the covid19 pandemic where
people are very interesting in any Covid related news like vaccination and so on. In
addition, malware targets also IoT devices, since most of it does not have a strong
security measures, hackers find it easier to manipulate and access data in IoT devices
like monitoring systems and any smart devices.
In order to defend against malware attacks, we proposed a cybersecurity frame-
work (Abdel ouahab Ikram et al. 2020) which is an intelligent solution dedicated to
malware detection and analysis. The particularity of our framework lies in the use of
the visualization technique; convert PE files into grayscale images. It is composed
by 4 layers as shown in Fig. 1: Discovery, Identification, Reporting and Reaction. In
a previous works we developed the Identification layer using visualization technique
and machine learning classifiers to know the malware family and behaviors. In this
paper, we develop the first layer to know either the input PE is malicious or benign.
To do so, we use various machine learning algorithms and PE files in the form of
grayscale images. Once the malicious PE file is discovered, then we can know the
malware family using Identification Layer. After that, the framework is able to report
Classification of Malicious and Benign Binaries … 299
the particularity and malicious behavior of the given malware to help react effectively
against it.
More and more researchers are using Artificial Intelligence (AI) in different
domains and applications. We found the use of machine learning algorithms in the
healthcare domain and IoT. For instance, the diagnostic of breast cancer (Saoud et al.
2019), medical chatbots (Soufyane et al. 2021) and other smart applications dedi-
cated to prevent the Covid19 pandemic by monitoring social distance and facemask
detection (Abdel Ouahab et al. 2020). Also AI algorithms are also used in Agricul-
tural applications (Bakkali Yedri et al. 2021) to classify plants easily. The power of AI
is also been used in digital marketing (Ahsain and Kbir 2021), the detection of fake
social media profile (Elyusufi et al. 2020) and the detection of illegitimate bitcoin
transactions (Elbaghdadi et al. 2020). In cybersecurity both defenders and attackers
are using powerful deep learning and machine learning algorithms to develop their
tasks. We used machine learning algorithms to classify malwares into their corre-
sponding families using various techniques in other works to ensure high quality
defense framework (Ben Abdel Ouahab et al. 2019, 2020).
The remaining of this paper is organized as followed: Sect. 2 summarized some
related works regarding highly to techniques used for malware-benign classification
using various machine learning algorithms. Section 3 describes the proposed cyberse-
curity framework and the proposed method used to classify malware-benign samples.
In Sect. 4, we gave an overview of the methodology used including machine learning
algorithms, evaluation metrics, malware visualization technique and images descrip-
tors taken as features. Later in Sect. 5 we present experimentations and discussed
the results. Finally we conclude by giving our research perspectives.
300 I. Ben Abdel Ouahab et al.
2 Related Works
3 Methodology
Algorithm Definition
Gaussian Naïve Bayes (GNB) The Naive Bayes classifier works on the principle of conditional
probability, as given by the Bayes theorem. It is an easy to use
algorithm that only estimates the mean and the standard
deviation from training data. Naive Bayes is better suited for
categorical input variables than numerical variables
Decision Tree (DT) Decision tree creates the classifier model by building a decision
tree. It is a non-parametric method which doesn’t depend upon
probability distribution assumptions. Decision trees can handle
high dimensional data with a good accuracy. DT requires less
data processing from the user. It can be used for classification
and also for feature engineering like prediction of missing
values and variable selection
K Nearest Neighbor (KNN) KNN is a supervised machine learning algorithm used for
classification and regression problems. The principle of KNN is
that similar things exist in close proximity. Another important
thing is choosing the appropriate K value. So, to have the right
K value, we need to run the algorithm several times and reduces
the errors
Random Forest (RF) Random Forest is a supervised learning algorithm. It is a very
used algorithm because of it flexibility, versatility and usage. It
can be solve classification and regression problems and gives
good results using default hyper-parameters. The main limitation
of random forest is that a large number of trees can make the
algorithm too slow and ineffective for real-time predictions
Logistic Regression (LR) Logistic regression is a statistical based machine learning
technique. LR is a linear model for classification. Using LR
probabilities describing the possible outcomes of a single trial
are modeled using a Logistic function
Metric Definition
Accuracy The accuracy is a statistical measure of classifier. It is the proportion of correct
predictions, true positives and true negatives, among all data. Accuracy = (TP +
TN)/(TP + TN + FP + FN)
F1-score F1-score is a combination of recall and precision using the harmonic mean
Recall Recall refers to the true positive rate or sensitivity
Recall = TP / (TP + FN)
(continued)
Classification of Malicious and Benign Binaries … 303
(continued)
Metric Definition
Precision Precision = TP /(TP + FP)
ROC-AUC Receiver Operator Characteristic (ROC) curve is an evaluation metric for binary
classification problems
Malwares from the Malimg database are already converted from PE files to grayscale
images. However, the benign samples need to be converted using the same technique
by converting the binary data (bytes sequence) to an 8 bits unsigned integers vector
then to a 2D array in the range of [0, 255], so that we can visualize it as a grayscale
image. Steps of the process are detailed in the Fig. 2. By converting a binary to
images we can see visual similarities in samples belonging to the same family. And
we found dis-similarities in variants belonging to different families. The advantage
of the visualization technique is that new malware variants appear similar to variants
of same family. Contrary to signature based detection, that can’t detect new variants
of malwares because of the new hashes in each variant.
3.2 Database
Benign samples are collected from official sources; we chose various types of
executable files such as text editor, navigator, and others. After we download the
PE files, we did an online analysis to ensure that the version is benign. To do so we
used the VirusTotal online tool (‘VirusTotal’. 2021). So, all the 50 samples results
were 100% benign. Malicious samples are taken from the Malimg database (Nataraj
et al. 2011b) where we found 25 families. We took 4 samples per family and a total of
100 malware samples belonging to all families. The malicious samples are provided
304 I. Ben Abdel Ouahab et al.
• Allaple.L,
• Allaple.A,
• Yuner.A,
• Lolyda.AA 1,
• Lolyda.AA 2,
• Lolyda.AA 3,
• C2Lop.P,
• C2Lop.gen!g,
• Instantaccess,
• Swizzot.gen!I,
• Swizzor.gen!E,
• VB.AT,
• Fakerean,
• Alueron.gen!J,
• Malex.gen!J,
• Lolyda.AT,
• Adialer.C,
• Wintrim.BX,
• Dialplatform.B,
• Dontovo.A,
• Obfuscator.AD,
• Agent.FYI,
• Autorun.K,
• Rbot!gen and
• Skintrim.N.
Afterwards, we perform binary classifiers using many machine learning algo-
rithms to classify grayscale images into two classes labeled as follow: (0) Benign and
(1) Malicious. To split data into train and test parts we do cross-validation which split
randomly the samples. We use 25% for test and the remaining 75% for training each
classifier. Data distributes randomly while splitting data into train and test. Talking
with numbers, we train our models using 74 malicious sample and 38 benign. Then
for testing our model we use 26 malicious and 12 benign samples as shown in figure
(Fig. 4).
The DAISY local image descriptor is based on gradient orientation histograms similar
to the SIFT descriptor. It is formulated in a way that allows for fast dense extraction
which is useful for bag-of-features image representations. The DAISY descriptor is
widely used in many applications of object recognition using grayscale images. It
306 I. Ben Abdel Ouahab et al.
74
80
70
60 38
50
40
30 26
20 12
Train
10
0
Test
Benign
Malicious
shows rapid and efficiency so we are going to implement it in the malware Visio-
based approach. As indicated by his name, DAISY descriptor is similar to daisy,
which is constructed by some central-symmetrical circles, as shown in figure below.
The Histogram of Oriented Gradient (HOG) feature descriptor is popular and
gives excellent results especially in pedestrian detection applications. However, this
feature is never been applied to malware images. The idea of HOG features is that
local object appearance and shape can often be characterized rather well by the
distribution of the local intensity gradients or edge directions, even without precise
knowledge of the corresponding gradient or edge positions (Shu et al. 2011). The Hog
descriptor is characterized by its simplicity and rapidity in computation. In Fig. 5, we
show a sample of HOG and DAISY representation separately using 2 input images.
In our approach, we make use of global feature descriptors as well as local feature
descriptors simultaneously to represent each image, to improve recognition accuracy.
Inspired from this paper (Wilson and Arif 1702), that use this features combination
in object detection task. On fifteen scene categories dataset, the average accuracy of
their model was 76.4%.
First, we extract DAISY features (Tola et al. 2008) from the grayscale images
using the Scikit-image library.1 We use the K-means algorithm to quantize DAISY
features into K clusters to form visual words based on the standard “bag-of-visual-
words” concept. Then, we extract a standard HOG descriptor. Second, we make
two levels pooling scheme. Starting from the DAISY features, a sum pooling is
performed by processing frequency of each visual word in every single image. To
1 https://scikit-image.org/.
Classification of Malicious and Benign Binaries … 307
form the DAISY histogram feature we do the L2 normalization. After that, the HOG
global descriptor of each image is normalized as the previous one. The final feature
is done by a concatenation of the normalized DAISY and HOG histogram features
(Fig. 6).
4 Proposed Solution
This section is divided into two parts: first we present our proposed cybersecu-
rity framework where we’ll need the malware-benign classification task. Second,
we present the process of computing features using the hybrid feature extraction
technique.
The proposed cybersecurity framework aims to defend against malware attacks
using AI techniques. Recently, we are in the phase of benchmarking and testing
several techniques, in order to find the most appropriate one. We are looking for an
efficient malware classifier and we are focusing also in the processing time of the
classifier. As we mentioned above, the proposed approach is an intelligent system
dedicated for malware attacks. That we have initialized in a previous work (Abdel
ouahab Ikram et al. 2020). It is a layered architecture composed of 4 layers. The
main particularity of the proposed framework is the adoption of visualization tech-
nique in all the process, first to detect the malware and second for malware family
classification. The proposed cybersecurity framework is summarized in algorithm 1.
Algorithm 1: Proposed cybersecurity framework
In this paper, we focused on the first layer of the intelligent cybersecurity frame-
work. Our goal is to classify binaries into malicious or benign class using machine
learning and hybrid features. The first step is to convert the unknown binary into
grayscale image, and then we extracted HOG and DAISY features and combined
them. The resulting features are used as input to various machine learning algo-
rithms to do the classification task. Finally, we evaluated these classifiers to obtain a
final efficient one. In addition to the hybrid features, we also tested HOG and DAISY
features separately to analyses the performance with each classifier. As describe in
algorithm 2.
Algorithm 2: Malicious-Benign classification based on hybrid features
classifiers. In general, the results are very bad. The two top classifiers using only HOG
features are KNN and RF with an accuracy of 81.57% and 84.21% respectively. In
term of malware benign classification this could not be an acceptable accuracy. So, we
conclude from performing a machine learning based classifier using the visualization
technique that HOG features only are not suitable for malware images classification
and are not recommended to be used in such case.
The Table 3 represents results using only DAISY features for all the 5 classifiers
and by calculating all evaluation metrics. The performance of all classifier has been
improves comparing to the HOG features. The top classifier is Logistic Regression
based. The accuracy of this algorithm has moved from 68.42% (with HOG features)
to 97.36% using the DAISY features. Same thing with other classifiers. We can see
clearly that all evaluation metrics are improved including recall, precision, f1-score
and roc-auc values. So using only DAISY features gave satisfactory performances.
That we are going to improve even more using the proposed hybrid features.
The Table 4 present results of classifiers using hybrid features (HOG and DAISY).
In general, we can see that performances have been improved. In particular, using
decision tree algorithm the accuracy goes from 86.84 to 92.10% using the proposed
hybrid features. Also, the others evaluation metrics have clearly increased. In addi-
tion, random forest classifier shows also an important increase. For instance, the
accuracy goes from 94.73% using DAISY features to 97.36% using hybrid features.
Table 4 Using hybrid features (DAISY and HOG) for malware-benign classification
Model Train score Test accuracy F1-score Recall Precision ROC-AUC
Gaussian 1 0,921 0,8695 0,8333 0,909 0,8974
NB
Decision 1 0,921 0,8695 0,8333 0,909 0,8974
tree
K-Nearest 0,9553 0,9473 0,909 0,8333 1 0,9166
neighbor
Random 1 0,9736 0,9565 0,9166 1 0,9583
forest
Logistic 0,9553 0,9736 0,9565 0,9166 1 0,9583
regression
Fig. 7 Decision tree classifier using various features and evaluation metrics
312 I. Ben Abdel Ouahab et al.
0.95
0.9
0.85
0.8
0.75
0.7
Train score Accuracy F1-score Recall Precision ROC-AUC
Fig. 8 Random forest classifier using various features and evaluation metrics
and hybrid feature are also too close. However, the proposed hybrid features perform
better than to two others features separately. Finally, using hybrid features, metrics
have slightly improved for decision tree and random forest classifiers. But, the others
algorithms give almost the same values.
In Fig. 9, we can see clearly the use of daisy only features and hybrid features
give the same values of all evaluation metrics with K-Nearest Neighbor algorithm.
However, the use of HOG descriptor as feature for malware benign classification is
not suitable also with KNN.
In related works, we have seen various works about classification of benign and
malicious samples using many techniques (Table 1). Our results outperforms CNN
model using grayscale images in Azab and Khasawneh (2020), they obtain 95.5%
accuracy against 97.36% that we have obtained using hybrid features and machine
learning algorithm. Far from that, systems call and dynamic analysis based features
shows higher accuracy than Image based classification. However the advantages of
our approach are: 1/ safety, 2/ rapidity and 3/ efficiency. First, using image visualiza-
tion technique is considered as safe because we are not dealing with the malicious
file directly. Second, the classification of a new sample gives results in few seconds
which is very important in term of cybersecurity. Third, the classifier we trained
using machine learning algorithms gives high accuracy results.
Finally, in this work we tested 3 cases of image descriptors. First, HOG descriptor
as a feature which didn’t give a good result. So we conclude that HOG descriptor
is not recommended for images converted from binaries. Second, DAISY descriptor
Classification of Malicious and Benign Binaries … 313
0.9
0.8
0.7
0.6
0.5
Train score Accuracy F1-score Recall Precision ROC-AUC
is also tested with many machine learning classifiers. And third, a hybrid feature is
been used. We got this hybrid feature by combining DAISY and HOG descriptors as
described in section above (Algorithm 2).
6 Conclusion
Acknowledgements We acknowledge financial support for this research from the “Centre National
pour la Recherche Scientifique et Technique”, CNRST, Rabat, Morocco.
References
International Conference on Smart City Applications (Casablanca MA, 2019), pp. 110:1–110:6.
https://doi.org/10.1145/3368756.3369098
P. Vinod, V. Laxmi, M.S. Gaur, G. Chauhan, MOMENTUM: MetamOrphic malware exploration
techniques using MSA signatures, in 2012 International Conference on Innovations in Informa-
tion Technology (IIT) (2012), pp. 232–237. https://doi.org/10.1109/INNOVATIONS.2012.620
7739
S.M. Tabish, M.Z. Shafiq, M. Farooq, Malware detection using statistical analysis of byte-level
file content, in Proceedings of the ACM SIGKDD Workshop on CyberSecurity and Intelligence
Informatics (New York, NY, USA, 2009), pp. 23–31. https://doi.org/10.1145/1599272.1599278
B. Anderson, D. Quist, J. Neil, C. Storlie, T. Lane, Graph-based malware detection using dynamic
analysis. J Comput Virol 7(4), 247–258 (2011). https://doi.org/10.1007/s11416-011-0152-x
J. Singh, J. Singh, Assessment of supervised machine learning algorithms using dynamic API calls
for malware detection. Int. J. Comput. Appl. , 1–8 (2020). https://doi.org/10.1080/1206212X.
2020.1732641
S.I. Bae, G.B. Lee, E.G. Im, Ransomware detection using machine learning algorithms. Concurr.
Comput.: Pract. Exp. 32(18)(2020). https://doi.org/10.1002/cpe.5422
L. Nataraj, V. Yegneswaran, P. Porras, J. Zhang, A comparative assessment of malware classification
using binary texture analysis and dynamic analysis, in Proceedings of the 4th ACM Workshop on
Security and Artificial Intelligence—AISec ’11 (Chicago, Illinois, USA, 2011), p. 21. https://doi.
org/10.1145/2046684.2046689
A. Azab, M. Khasawneh, MSIC: malware spectrogram image classification. IEEE Access 8,
102007–102021 (2020). https://doi.org/10.1109/ACCESS.2020.2999320
‘VirusTotal’. https://www.virustotal.com/gui/. Accessed 04 Mar 2021
L. Nataraj, S. Karthikeyan, G. Jacob, B.S. Manjunath, Malware images: visualization and automatic
classification, in Proceedings of the 8th International Symposium on Visualization for Cyber
Security - VizSec ’11 (Pittsburgh, Pennsylvania, 2011), pp. 1–7. https://doi.org/10.1145/2016904.
2016908
C. Shu, X. Ding, C. Fang, Histogram of the oriented gradient for face recognition. Tsinghua Sci.
Technol. 16(2), 216–224 (2011). https://doi.org/10.1016/S1007-0214(11)70032-3
J. Wilson, M. Arif, Scene recognition by combining local and global image descriptors’, arXiv:
1702.06850 [cs], Feb 2017, Accessed 10 Feb 2021. http://arxiv.org/abs/1702.06850
E. Tola, V. Lepetit, P. Fua, A fast local descriptor for dense matching, in 2008 IEEE Conference on
Computer Vision and Pattern Recognition (2008), pp. 1–8. https://doi.org/10.1109/CVPR.2008.
4587673
FakeTouch: Machine Learning Based
Framework for Detecting Fake News
Abstract Fake news is any content or information that is false and often generated to
mislead its readers in believing something which is not true. Fake news has become
one of major threats that can harm someone’s reputation. It often circulates wrong or
made up information about various products, events, people or entity. The deliberate
making of such news is escalating drastically these days. Fake news deceives us
in taking wrong decisions. Therefore, Fake News Detection has attained immense
deal of interest from researchers all over the world. In this chapter, a machine learn-
ing approach has been proposed named FakeTouch starting with Natural Language
Processing based concept by applying text processing, cleaning and extraction tech-
niques. This approach aim to arrange the information to be “obeyed” into each
classification model for training and tuning parameters for every model to bring out
the optimized and best prediction to find out the Fake news. To evaluate the proposed
framework, three use cases with three different datasets has been developed during
this study. The proposed framework will also help to understand what amount of
data is responsible for detecting fake news, trying to stage the linguistic differences
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 317
Y. Baddi et al. (eds.), Big Data Intelligence for Smart Applications,
Studies in Computational Intelligence 994,
https://doi.org/10.1007/978-3-030-87954-9_15
318 A. B. Siddikk et al.
between fake and true articles providing a visualization of the results using different
visualization tools. This chapter also presents a comprehensive performance evalua-
tion to compare different well known machine learning classifiers like Support Vector
Machine, Naïve Bayes Method, Decision Tree Classifier, Random Forest, Logistic
Regression as well as to develop an ensemble method (Bagging & Boosting) like
XGBClassifier, Bagging Classifier of different combinations of classification models
to identify which will give the best optimal results for three part of datasets. As a
result, it has been found that with an appropriate set of features extracted from the
texts and the headlines, XGB classifier can effectively classify fake news with very
high detection rate. This framework also provides a strong baseline of an intelligent
anti-fake news detector.
1 Introduction
The term ‘fake news’ has become an alarming issue particularly the spreading of
factually incorrect and misleading articles published mostly for the aim of creating
money through page views. Researchers [8] at Indiana University found these two
sorts of information often go viral because “information overload and finite span
of users limit the capacity of social media to discriminate information on the idea
of quality.” Shlok et al. Mugdha et al. (2020) conducted a study considering the
fake news within American political Speech which was the topic of considerable
attention, particularly following the election of President Trump. In another study
Granik and Mesyura (2017) the authors developed a model based on naive Bayes
classifier to examine how this particular method works for this particular problem
given. Because of the availability of cheaper hardware and publicly available datasets,
machine learning algorithms have started to work much better for making intelligent
system in different areas. Shabani et al. Shabani and Sokhn (2018) have applied
Hybrid Machine-Crowd Approach as an advantageous solution to tackle the fake
news problem in general. Author focuses on distinguishing satire or parody and fab-
ricated content using the Fake vs Satire public dataset, this approach provides higher
accuracy at an acceptable cost and latency. Tanvir et al. Tanvir et al. (2019) experi-
mented several machine learning and deep learning approaches to detect fake news.
Different types of classification models including Support Vector Machine (SVM),
Naïve Bayes, Logistic Regression, Long short-term memory, Recurrent Neural Net-
work are implemented for that task. A combination of these classification models is
also tested to enhance further the accuracy of prediction. Those models are imple-
mented to prepare from the training dataset using k-fold (k = 2) cross-validation,
and then predict using the data set. Ahmad et al. Ahmad et al. (2020) proposed their
ability to require a choice relies totally on the sort of data consume. On the basis of
information digest, their worldview was shaped. There is an increasing evidence that
Fake Touch: Machine Learning Based Framework … 319
buyers have reacted absurdly to the news that later proved to be fake stated by the
authors. One recent case is the spread of the novel coronavirus where fake reports
cover the web about the origin, nature, and behavior of the virus. Things worsened
as more people examine the fake content online. Identifying such news online may
be a daunting task. An investigation based on the purpose of to let the developers
and researchers can understand further which model and technique is the best also
will help to implement the tools or models in future has been performed during this
study. In this chapter, several machine learning techniques are applied to evaluate
and improve the accuracy of the model performance. Thus, the major contribution of
this work is to identify the best outcome-based model using the desired framework.
Along with the machine learning algorithms, ensemble methods are also considered
during the experiments including Support Vector Machine, Naïve Bayes Method,
Decision Tree Classifier, Random Forest, Logistic Regression, XGBClassifier. On
one standard dataset divided in three way using a novel set of features and statisti-
cally validated the results by calculating accuracy and F1 scores. Then, those machine
learning algorithms which are out-performers are implemented for this task. A com-
bination of these classification models is also examined and evaluated to check the
enhancement of the accuracy to classify or identify the fake news. The techniques
are implemented to prepare from the training dataset using TF-IDF technique, and
then predict using the data set whether it is fake or real with the help of sci-kit learn
python package Tanvir et al. (2019). The rest of the chapter is organized as follows.
The implemented framework for fake news detection has been presented in Sect. 2.
Evaluation parameters of this approach are briefly explained in Sect. 3. The experi-
mental information including the result and analysis has been broadly discussed and
debated in Sect. 4. Finally, Sect. 5 concludes this chapter.
2 Methodology
The proposed and implemented methodology will be described in this section from
dataset collection to detection in details.
From background study, it has been found that most of the fake news threatens
not only for our society but also all over the world. Thus, this data includes the
information about the political news (most uses), social news as well as world news.
At the very beginning fake news dataset have been collected from Kaggle Bali et al.
(2019) where this dataset is related to the fake news spread during the time of the
2016 U.S. Presidential Election, IEEE Dataport and some of True news have been
collected using Web-scraping from relevant and trusted source such as CNN, BBC,
and The New York Times etc. Finally, fake and true data has been concatenated which
320 A. B. Siddikk et al.
are already collected from different sources. There is a total 44909 (https://doi.org/
10.6084/m9.figshare.13325198.v1) data in the dataset. Total dataset has been divided
into three case studies to identify whether the proposed approach is dependent on
the size of data or not to detect more efficiently. Total 44,909 of document has been
assign as CS-1, 33,681 documents have been assigned as CS-2 and 22,454 documents
have been assign as CS-3.
This dataset includes the Label class as F and T where ‘F’ stands for Fake news
also indicating Fake and ‘T’ stands for True news indicating Real or Fact which set
as a label. The dataset also includes one column for statements where all the data
including the links and mentions. There are 4 features namely Title, Text, Subject
and Date. Then have been labelled the data as a Fake news and a True news (Fig. 1).
Fake Touch: Machine Learning Based Framework … 321
TF-IDF Tanvir et al. (2019) is a mathematical measurement that speaks to the sig-
nificance of a term in a report among others in a bunch of records. It consolidates
two measurements, Term Frequency and Inverse Document Frequency. The initial
measurement considers the occasions that a term shows up in a record and gives
higher position. TF-IDF is widely used in Natural Language Processing Rahman et
al. (2018, 2020, 2020) for example sentiment analysis, classification and so on.
TF-IDF (implemented from sci-kit learn Shukla et al. (2021)) In Python, TF-IDF
is applied on text-data using sklearn. TfidfVectorizer from sklearn.feature extraction
at very first have imported for perform the work. After that, it have been setup
the vectorizer and then run the program fit and transform over it to compute the
TF-IDF score for the text Dataset using the TF-IDF algorithm. Finally, the sklearn
fit transform function have been applied the following fit and transform functions,
yielding the result of TF IDF.
322 A. B. Siddikk et al.
Following the Two Mathematical Formula have been compute the TF-IDF score
as follows (Fig. 2)
T otalnumber o f document
I D F(t) = log .
T otalnumber o f documentwithter mtinit
• Words Level TF-IDF: TF-IDF value of each term represented in a matrix format.
• N-gram Level TF-IDF: Combination of N termism spoke to by N-gram level. This
matrix portrays TF-IDF N-gram scores
• Character Level TF-IDF: TF-IDF values of the n- grams character level in corpus
represented in a matrix.
For this proposed framework 8 different types of machine learning algorithms have
been used. Decision Tree (DT), Support Vector Machine (SVM), Random Forest
(RF), K-Neighbours (KNN), Naïve Bayes (NB), Passive Aggressive Classifier (PAC),
Logistic Regression (LR) and Extrembaselinee Gradient Boosting (XGB) Bali et al.
(2019) have been implemented to fit and construct the model.Essentially, this model
was created based on our workflow, in which we performed many techniques using
the Python programming language and many useful Python packages to demonstrate
our structure. This method was created by taking a step-by-step approach from this
work. And this study found that this workflow produces the best results for detecting
false news ever. Using this method, researchers will continue to identify false news
in order to obtain the best estimation outcome.
3 Evaluation Metrics
Performance measures are based on data, and represent a clear story about whether
a representation or activity is achieving its objectives and if progress is being made
toward attaining policy or organizational goals. In order to assess the performance of
the proposed model, as evaluation matrices precision, recall, F1-score, true negative
rate, false positive rate and accuracy Siddikk et al. (2021) are used.
Fake Touch: Machine Learning Based Framework … 323
Fig. 2 FakeTouch Framework for training algorithms and classification of news articles
324 A. B. Siddikk et al.
Performance evaluation of a classifier in commonly done using the data in the con-
fusion matrix. A confusion matrix for the two-class problem given in Table 1.
True Positive (TP): The number of true positive examples is the number of news
articles, correctly classified as fake; False Positive (FP): The number of false positive
examples is the number of news articles incorrectly classified as fake; True Negative
(TN): The number of true negative examples is the number of news articles, correctly
classified as true; False Negative (FN): The number of false negative examples is the
number of news articles incorrectly classified as true.
The measure of the ability of the model to accurately identify the occurrence of a
positive class instance is determined by recall.
F1 Score is the weighted average of Precision and Recall. Therefore, this score
takes both false positives and false negatives into calculation. The true negative rate
(TNR), is the proportion of samples that test negative using the test in question
that are genuinely negative. A False Positive Rate is an accuracy metric that can be
measured on a subset of machine learning models. Accuracy is a measure of total
correctly identified samples out of all the samples.
TP TP
Recall = Pr ecision =
T P + FN T P + FP
TN FP
TNR = FPR =
FP + T N FP + T N
2 ∗ (Pr ecision ∗ Recall) TN +TN
F1 = Accuracy =
Pr ecision + Recall T P + T N + FP + FN
Fake Touch: Machine Learning Based Framework … 325
There are three type of case studies which have been experimented to evaluate the
assessment in this section. Here three case study will be described and evaluated
based on the multifarious machine learning techniques.
For the case study #1 have been used feature extraction technique TF-IDF. After
that, Confusion Matrix have been found using the selected Classification Model as
follows:
Tables 2, 3 and 4 strongly represents the confusion matrix for every model accord-
ing to Table 1 format.For example Table 2 represents data followed by the Table 1.
from top left represent True Positive(TP), top right represent False Negative(FN),
bottom left represents False Positive(FP) and bottom right represents True Nega-
tive(TN) for every confusion matrix.
Table 2 represents the confusion matrix for the assigned machine learning classi-
fiers where using the value of confusion matrix can find the evaluation metrics from
dataset one. From the mathematics, calculation, analysis and investigation, have been
found that eXtreme Gradient Boosting as classifier have achieved highest an accu-
racy among all classifier. For the precision, recall and f1-score Decision tree given
the highest score among all classifier. The accuracy, precession, recall and f1-score
have been presented Tables 6 and 7. In Fig. 3 represent the true negative rate and
false positive rate where have been found that in level 1, maximum 99.74% of TNR
attained from XGB individually with minimum 0.25 FPR. In case TNR, FPR given
the lowest rate for the KNN whereas XGB given a better performance individually
(Table 5).
326 A. B. Siddikk et al.
Table 2 Representation of Confusion Matrix; DT: Decision Tree; SVM: Support Vector Machine;
RF: Random Forest; KNN: k-Nearest Neighbors; MNB: Multinomial Naive Bayes; PAC: Passive
Aggressive Classifier; LR: Logistic Regression; XGB: eXtreme Gradient Boostingn Matrix
4361 34 4332 63
25 7304 28 7301
Confusion Matrix for DT ConfusionMatrix for SVC
4244 151 4257 34
28 7301 53 7276
Confusion Matrix for RF Confusion Matrix for LR
4374 21 3122 1273
19 7310 23 7306
Confusion Matrix for XGBC Confusion Matrix for MNB
4339 56 2824 1571
30 729 226 7103
Confusion Matrix for PAC Confusion Matrix for KNN
Table 3 Representation of Confusion Matrix; DT: Decision Tree; SVM: Support Vector Machine;
RF: Random Forest; KNN: k-Nearest Neighbors; MNB: Multinomial Naive Bayes; PAC: Passive
Aggressive Classifier; LR: Logistic Regression; XGB: eXtreme Gradient Boosting
3320 23 3282 61
18 5432 22 5428
Confusion Matrix for DT ConfusionMatrix for SVC
3204 139 3209 134
17 5433 31 5419
Confusion Matrix for RF Confusion Matrix for LR
3334 9 2256 1087
7 5443 13 5437
Confusion Matrix for XGBC Confusion Matrix for MNB
3277 66 2087 125
16 5434 6158 5292
Confusion Matrix for PAC Confusion Matrix for KNN
In this section case study #2 conduct with the feature extraction technique TF-IDF.
After trained the model Confusion Matrix have been found using the selected Clas-
sification Model as follows:
Confusion matrix based on selected classifier have been assigned in Table 3 where
using the value of confusion matrix can find out the accuracy, precession, recall, f1-
score etc. on dataset two. The accuracy, TNR and FPR have been visualized in Table
3 and Fig. 4. It has been found that eXtreme Gradient Boosting as classifier have
Fake Touch: Machine Learning Based Framework … 327
Table 4 Representation of Confusion Matrix; DT: Decision Tree; SVM: Support Vector Machine;
RF: Random Forest; KNN: k-Nearest Neighbours; MNB: Multinomial Naive Bayes; PAC: Passive
Aggressive Classifier; LR: Logistic Regression; XGB: eXtreme Gradient Boosting
2195 11 2153 53
10 3647 22 3635
Confusion Matrix for DT ConfusionMatrix for SVC
2135 71 2108 98
13 3644 28 3629
Confusion Matrix for RF Confusion Matrix for LR
2199 7 1449 757
7 3650 7 3650
Confusion Matrix for XGBC Confusion Matrix for MNB
2165 41 1336 870
26 3631 94 3563
Confusion Matrix for PAC Confusion Matrix for KNN
Fig. 3 TNR versus FPR Score among all classifiers for CS-1; FPR: False Positive Rate; TNR: True
Negative Rate
achieved highest an accuracy among all classifier also in terms of precision, recall
and f1-score. The accuracy, precision and recall have been presented Tables 6 and 7.
Figure 4 represents the TNR and FPR with, maximum 99.87% of TNR attained from
XGB individually which is little bit higher and minimum 0.12 FPR which is given
little bit high performance than dataset one. In case study 2 given little bit better
performance than the case study 1
328 A. B. Siddikk et al.
Table 5 Precision, Recall and F1-Score for Machine Learning algorithm; CS: Case Study; DT:
Decision Tree; SVM: Support Vector Machine; RF: Random Forest; KNN: k-Nearest Neighbours;
MNB: Multinomial Naive Bayes; PAC: Passive Aggressive Classifier; LR: Logistic Regression;
XGB: eXtreme Gradient Boosting
CS-1 CS-2
Classification
Precision Recall F1-Score Precision Recall F1-Score
Model
DT 99.43 99.22 99.53 99.46 99.31 99.38
SVM 99.35 98.56 98.95 99.33 98.17 98.74
RF 99.34 96.56 97.93 99.47 95.84 97.62
KNN 92.59 64.25 75.85 92.96 62.42 74.68
MNB 99.26 71.03 82.80 99.42 67.48 80.19
PAC 99.31 98.72 99.01 99.51 98.02 98.75
LR 98.77 96.86 97.90 99.04 95.99 97.49
XGB 98.63 98.49 98.55 99.79 99.73 99.75
CS-3
Classification
Precision Recall F1-Score
Model
DT 99.54 99.50 99.51
SVM 98.98 97.59 98.28
RF 99.39 96.78 98.06
KNN 93.42 60.56 73.48
MNB 99.51 65.68 79.13
PAC 98.81 98.14 98.23
LR 98.68 95.55 97.08
XGB 99.68 99.68 99.68
Case study #3 has been performed with feature extraction technique TF-IDF. After
extract the features, model have been trained and confusion matrices are obtained as
follows:
Similarly, the case study 1 and 2, Table 4 represents the confusion matrix for the
assign machine learning classifier where using the value of confusion matrix can find
out the accuracy, precession, recall, f1-score etc. on dataset three. It has been found
that eXtreme Gradient Boosting as classifier have achieved highest accuracy among
all classifier which are similar to the case study 1 and 2. The accuracy, precession,
recall and f1-score have been presented Tables 6 and 7. Figure 5 represents the true
negative rate and false positive rate have been found that in level 1 with maximum
99.80% of TNR attained from XGB individually which is a little bit higher from the
dataset one and also significantly higher from the dataset two and minimum 0.19 FPR
Fake Touch: Machine Learning Based Framework … 329
Table 6 Accuracy of word embedding model (TF-IDF) for machine learning algorithm
CS-1 CS-2 CS-3
Word embedding Classification Accuracy Accuracy Accuracy
model model
Term Frequency– DT 99.49 99.53 99.64
Inverse
Document
Frequency
(TF-IDF)
SVM 99.22 99.09 98.72
RF 98.47 98.23 98.57
KNN 84.67 83.92 83.56
MNB 88.95 87.49 86.97
PAC 99.27 99.09 98.86
LR 98.37 98.12 97.85
XGB 99.66 99.82 99.76
Fig. 4 TNR vs FPR Score among all classifiers for CS-2; FPR: False Positive Rate; TNR: True
Negative Rate
Fig. 5 TNR versus FPR score among all classifiers for CS-3; FPR: False Positive Rate; TNR: True
Negative Rate
which has given a little bit high performance than dataset one and less for the dataset
two. In case study 2 performs better performance than the case study 1. According
to the investigation, for the three case studies with machine learning-based models
have found that performance decrease as the scale of data increases and after a certain
period of data level performance increase gradually. In this case, the amount of data
extremely related for how accurately detect fake news. On the other hand, eXtreme
Gradient Boosting outperforms with higher accuracy which proves it to be the best
to detect the fake news accurately in every cases.
The Classification results for all machine learning models which have been obtained
for CS-1, CS-2, and CS-3 (Sect. 2.1) as Illustrates Fig. 6.
Fake Touch: Machine Learning Based Framework … 331
It has been noticed that classification models which investigated have given highest
performance for XGB classifier and provided less performance for KNN models for
the given three part of dataset (Sect. 2.1). But the important point is that the DT classi-
fication have been increased when this propose document gradually decrease. On the
other hand, SVM, KNN, MNB, PAC and LR classifier has been decrease their perfor-
mance. Using the word embedding model TF-IDF in this experiment have achieved
the score of 99.66, 99.82 and 99.76% (Table 7) highest accuracy respectively using
three amounts of dataset (Sect. 2.1) for XGB classifier among all classification mod-
els. Apart from that, lowest accuracy has found for KNN models. It’s been clearly
said that if someone implement the proposed model by applying on framework for
detect fake news, Boosting Technique will be help to find out the best estimate output
only of accurately detect Fake News. In addition, for the DT classifier will give the
highest performance if dataset gradually decrease. For the DT classifier, number of
documents depend on the estimate result. Model performance for detecting fake news
will be decrease while dataset gradually decrease for the classifier named SVM, RF,
KNN, MNB, PAC and LR. XGB classifier always provides the highest performance
of accuracy for any number of documents. Table 8 shows the comparison among
different proposed models with FakeTouch. Following is the configuration of the
computing environment for carrying out the experiment. The proposed Framework
shows comparatively best results and effectively (training accuracy, testing accuracy
also light-weight model for training). From the proposed Framework, this investiga-
tion achieved maximum accuracy of 99.66%. It’s also been claimed that the proposed
approach achieved better result with the real-world text based fake news dataset as
compared to existing works. Finally, the proposed approach strongly encourages all
researchers that is most significant framework to detect fake news.
332 A. B. Siddikk et al.
5 Conclusion
References
Y. Shukla, N. Yadav, A. Hari, A unique approach for detection of fake news using machine learning.
Int. J. Res. Appl. Sci. Eng. Technol. (IJRASET). ISSN, 2321-9653
D.M. Lazer, M.A. Baum, Y. Benkler, A.J. Berinsky, K.M. Greenhill, F. Menczer, J.L. Zittrain, The
science of fake news. Science 359(6380), 1094–1096 (2018)
V. Pérez-Rosas, B. Kleinberg, A. Lefevre, R. Mihalcea, Automatic detection of fake news (2017).
arXiv:1708.07104
K. Shu, S. Wang, H. Liu, Beyond news contents: The role of social context for fake news detection,
in Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining,
pp. 312–320 (2019)
X. Zhou, R. Zafarani, Fake news: a survey of research, detection methods, and opportunities (2018).
arXiv:1812.00315
T. Quandt, L. Frischlich, S. Boberg, T. Schatto-Eckrodt, Fake news. Int. Encycl. Journal. Stud. 1–6
(2019)
N. Ruchansky, S. Seo, Y. Liu, Csi: a hybrid deep model for fake news detection, in Proceedings of
the 2017 ACM on Conference on Information and Knowledge Management, pp. 797–806 (2017)
M. Potthast, J. Kiesel, K. Reinartz, J. Bevendorff, B. Stein, A stylometric inquiry into hyperpartisan
and fake news (2017). arXiv:1702.05638
X. Zhou, R. Zafarani, K. Shu, H. & Liu, Fake news: fundamental theories, detection strategies and
challenges, in Proceedings of the Twelfth ACM International Conference on Web Search and
Data Mining, pp. 836–837 (2019)
J. Albright, Welcome to the era of fake news. Media Commun. 5(2), 87–89 (2017)
E.C. Tandoc Jr., Z.W. Lim, R. Ling, Defining, “fake news” a typology of scholarly definitions. Digit.
Journal. 6(2), 137–153 (2018)
H. Karimi, P. Roy, S. Saba-Sadiya, J., J. Tang, Multi-source multi-class fake news detection, in
Proceedings of the 27th international conference on computational linguistics, pp. 1546–1557
(2018)
X. Zhou, R. Zafarani, A survey of fake news: fundamental theories, detection methods, and oppor-
tunities. ACM Comput. Surv. (CSUR) 53(5), 1–40 (2020)
K. Shu, X. Zhou, S. Wang, R. Zafarani, H. Liu, The role of user profiles for fake news detection, in
Proceedings of the 2019 IEEE/ACM International Conference on Advances in Social Networks
Analysis and Mining, pp. 436–439 (2019)
R. Oshikawa, J. Qian, W.Y. Wang, A survey on natural language processing for fake news detection
(2018). arXiv:1811.00770
X. Zhang, A.A. Ghorbani, An overview of online fake news: characterization, detection, and dis-
cussion. Inf. Proc. Manag. 57(2), 102025 (2020)
K. Shu, L. Cui, S. Wang, D. Lee, H. & Liu, Defend: explainable fake news detection, in Proceedings
of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining,
pp. 395–405 (2019)
R. Zellers, A. Holtzman, H. Rashkin, Y. Bisk, A. Farhadi, F. Roesner, Y. Choi, Defending against
neural fake news (2019). arXiv:1905.12616
K. Shu, S. Wang, H. Liu, Understanding user profiles on social media for fake news detection,
in 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), pp.
430–435 (IEEE, 2018)
H. Karimi, J. Tang, Learning hierarchical discourse-level structure for fake news detection (2019).
arXiv:1903.07389
R.K. Nielsen, L. Graves, “News you don’t believe”: audience perspectives on fake news (2017)
A. Kucharski, Study epidemiology of fake news. Nature 540(7634), 525–525 (2016)
S.S.M.M. Rahman, M.H. Rahman, K. Sarker, M.S. Rahman, N. Ahsan, M.M. Sarker, Supervised
ensemble machine learning aided performance evaluation of sentiment classification. J. Phys.:
Conf. Ser. 1060(1), 012036) (2018). IOP Publishing
S.S.M.M. Rahman, K.B.M.B. Biplob, M.H. Rahman, K. Sarker, T. Islam, An investigation and
evaluation of N-Gram, TF-IDF and ensemble methods in sentiment classification, in International
Conference on Cyber Security and Computer Science (Springer, Cham, 2020), pp. 391–402
334 A. B. Siddikk et al.
M.M. Rahman, S.S.M.M. Rahman, S.M. Allayear, M.F.K. Patwary, M.T.A. Munna, A sentiment
analysis based approach for understanding the user satisfaction on android application, in Data
Engineering and Communication Technology (Springer, Singapore, 2020)
A.B. Siddikk, M.F. Muntasir, R.J. Lia, S.S.M.M. Rahman, T. Islam, M. Alazab, Revisiting the
approaches, datasets and evaluation parameters to detect android malware: a comparative study
from state-of-art. Artif. Intell. Blockchain Future Cybersecur. Appl. 125 (2021)