Advanced Machine Intelligence and Signal Processing Deepak Gupta pdf download
Advanced Machine Intelligence and Signal Processing Deepak Gupta pdf download
https://ebookmeta.com/product/advanced-machine-intelligence-and-
signal-processing-deepak-gupta/
https://ebookmeta.com/product/coherence-in-signal-processing-and-
machine-learning-ramirez/
https://ebookmeta.com/product/eeg-signal-processing-and-machine-
learning-2nd-edition-saeid-sanei/
https://ebookmeta.com/product/advanced-sensing-in-image-
processing-and-iot-1st-edition-rashmi-gupta/
https://ebookmeta.com/product/teaching-of-science-2022nd-edition-
b-ed/
Savage King New Worlds The Crystal Kingdom Book 5 1st
Edition Milly Taiden Taiden Milly
https://ebookmeta.com/product/savage-king-new-worlds-the-crystal-
kingdom-book-5-1st-edition-milly-taiden-taiden-milly/
https://ebookmeta.com/product/lives-of-the-musicians-good-times-
bad-times-and-what-the-neighbors-thought-kathleen-krull/
https://ebookmeta.com/product/short-fiction-1st-edition-
voltairine-de-cleyre-2/
https://ebookmeta.com/product/good-night-mr-wodehouse-a-
novel-2nd-edition-sullivan-faith/
https://ebookmeta.com/product/hallowed-be-this-house-finding-
signs-of-heaven-in-your-home-thomas-howard/
Pocket Rough Guide Florence 4th Edition Rough Guides
https://ebookmeta.com/product/pocket-rough-guide-florence-4th-
edition-rough-guides/
Lecture Notes in Electrical Engineering 858
Deepak Gupta
Koj Sambyo
Mukesh Prasad
Sonali Agarwal Editors
Advanced
Machine
Intelligence
and Signal
Processing
Lecture Notes in Electrical Engineering
Volume 858
Series Editors
Leopoldo Angrisani, Department of Electrical and Information Technologies Engineering, University of Napoli
Federico II, Naples, Italy
Marco Arteaga, Departament de Control y Robótica, Universidad Nacional Autónoma de México, Coyoacán,
Mexico
Bijaya Ketan Panigrahi, Electrical Engineering, Indian Institute of Technology Delhi, New Delhi, Delhi, India
Samarjit Chakraborty, Fakultät für Elektrotechnik und Informationstechnik, TU München, Munich, Germany
Jiming Chen, Zhejiang University, Hangzhou, Zhejiang, China
Shanben Chen, Materials Science and Engineering, Shanghai Jiao Tong University, Shanghai, China
Tan Kay Chen, Department of Electrical and Computer Engineering, National University of Singapore,
Singapore, Singapore
Rüdiger Dillmann, Humanoids and Intelligent Systems Laboratory, Karlsruhe Institute for Technology,
Karlsruhe, Germany
Haibin Duan, Beijing University of Aeronautics and Astronautics, Beijing, China
Gianluigi Ferrari, Università di Parma, Parma, Italy
Manuel Ferre, Centre for Automation and Robotics CAR (UPM-CSIC), Universidad Politécnica de Madrid,
Madrid, Spain
Sandra Hirche, Department of Electrical Engineering and Information Science, Technische Universität
München, Munich, Germany
Faryar Jabbari, Department of Mechanical and Aerospace Engineering, University of California, Irvine, CA,
USA
Limin Jia, State Key Laboratory of Rail Traffic Control and Safety, Beijing Jiaotong University, Beijing, China
Janusz Kacprzyk, Systems Research Institute, Polish Academy of Sciences, Warsaw, Poland
Alaa Khamis, German University in Egypt El Tagamoa El Khames, New Cairo City, Egypt
Torsten Kroeger, Stanford University, Stanford, CA, USA
Yong Li, Hunan University, Changsha, Hunan, China
Qilian Liang, Department of Electrical Engineering, University of Texas at Arlington, Arlington, TX, USA
Ferran Martín, Departament d’Enginyeria Electrònica, Universitat Autònoma de Barcelona, Bellaterra,
Barcelona, Spain
Tan Cher Ming, College of Engineering, Nanyang Technological University, Singapore, Singapore
Wolfgang Minker, Institute of Information Technology, University of Ulm, Ulm, Germany
Pradeep Misra, Department of Electrical Engineering, Wright State University, Dayton, OH, USA
Sebastian Möller, Quality and Usability Laboratory, TU Berlin, Berlin, Germany
Subhas Mukhopadhyay, School of Engineering & Advanced Technology, Massey University,
Palmerston North, Manawatu-Wanganui, New Zealand
Cun-Zheng Ning, Electrical Engineering, Arizona State University, Tempe, AZ, USA
Toyoaki Nishida, Graduate School of Informatics, Kyoto University, Kyoto, Japan
Luca Oneto, Dept. of Informatics, Bioengg., Robotics, University of Genova, Genova, Genova, Italy
Federica Pascucci, Dipartimento di Ingegneria, Università degli Studi “Roma Tre”, Rome, Italy
Yong Qin, State Key Laboratory of Rail Traffic Control and Safety, Beijing Jiaotong University, Beijing, China
Gan Woon Seng, School of Electrical & Electronic Engineering, Nanyang Technological University,
Singapore, Singapore
Joachim Speidel, Institute of Telecommunications, Universität Stuttgart, Stuttgart, Germany
Germano Veiga, Campus da FEUP, INESC Porto, Porto, Portugal
Haitao Wu, Academy of Opto-electronics, Chinese Academy of Sciences, Beijing, China
Walter Zamboni, DIEM - Università degli studi di Salerno, Fisciano, Salerno, Italy
Junjie James Zhang, Charlotte, NC, USA
The book series Lecture Notes in Electrical Engineering (LNEE) publishes the
latest developments in Electrical Engineering - quickly, informally and in high
quality. While original research reported in proceedings and monographs has
traditionally formed the core of LNEE, we also encourage authors to submit books
devoted to supporting student education and professional training in the various
fields and applications areas of electrical engineering. The series cover classical and
emerging topics concerning:
• Communication Engineering, Information Theory and Networks
• Electronics Engineering and Microelectronics
• Signal, Image and Speech Processing
• Wireless and Mobile Communication
• Circuits and Systems
• Energy Systems, Power Electronics and Electrical Machines
• Electro-optical Engineering
• Instrumentation Engineering
• Avionics Engineering
• Control Systems
• Internet-of-Things and Cybersecurity
• Biomedical Devices, MEMS and NEMS
For general information about this book series, comments or suggestions, please
contact [email protected].
To submit a proposal or request further information, please contact the Publishing
Editor in your country:
China
Jasmine Dou, Editor ([email protected])
India, Japan, Rest of Asia
Swati Meherishi, Editorial Director ([email protected])
Southeast Asia, Australia, New Zealand
Ramesh Nath Premnath, Editor ([email protected])
USA, Canada:
Michael Luby, Senior Editor ([email protected])
All other Countries:
Leontina Di Cecco, Senior Editor ([email protected])
** This series is indexed by EI Compendex and Scopus databases. **
Advanced Machine
Intelligence and Signal
Processing
Editors
Deepak Gupta Koj Sambyo
Department of Computer Science Department of Computer Science
and Engineering and Engineering
National Institute of Technology Arunachal National Institute of Technology Arunachal
Pradesh (NITAP) Pradesh (NITAP)
Itanagar, Arunachal Pradesh, India Itanagar, Arunachal Pradesh, India
© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature
Singapore Pte Ltd. 2022
This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether
the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by similar
or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors, and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Singapore Pte Ltd.
The registered company address is: 152 Beach Road, #21-01/04 Gateway East, Singapore 189721,
Singapore
Preface
This book helps readers to interact with a selection of refereed papers presented
at the 3rd International Conference on Machine Intelligence and Signal Processing
(MISP-2021), National Institute of Technology Arunachal Pradesh, Jote, India, from
23 to 25 September 2021. This book coverage is concerned domains to explore
and discuss different aspects of data mining, artificial intelligence, optimization,
machine learning methods/algorithms, signal processing theory and methodologies,
and their applications. The significance, uniqueness, and technical excellence of all
contributions were considered. The technical programme committee was subjected
to a double-blind review process to ensure that the author names and affiliations were
unknown to them.
We appreciate the assistance of the advisory committee members, and we thank
all of the keynote speakers for sharing their knowledge and skills with us. Professor
Pinakeswar Mahanta, Director, NIT Arunachal Pradesh, deserves special gratitude
for his insightful advice and support. We would like to express our gratitude to the
organising committee as well as the many additional volunteers who helped make
this conference a success. We appreciate EasyChair’s assistance with the submission
and evaluation process. Finally, we would like to thank Springer for their enthusiastic
participation and prompt publication of the papers.
v
Contents
vii
viii Contents
Dr. Koj Sambyo received his Ph.D. in Computer Science and Engineering from
National Institute of Technology, Arunachal Pradesh, in 2017 and M.Tech. degree
in Computer Science and Engineering from Rajiv Gandhi University, Arunachal
Pradesh, India, in 2011. Currently, he is working as Assistant Professor in the Depart-
ment of Computer Science and Engineering in National Institute of Technology,
Arunachal Pradesh. His research activities mainly focused on cloud computing and
natural language processing. He is author of numerous international refereed journals
and in referred international conferences.
Dr. Mukesh Prasad is Senior Lecturer at the School of Computer Science in the
Faculty of Engineering and IT at UTS who has made substantial contributions to
the fields of machine learning, artificial intelligence, and the Internet of things.
Mukesh’s research interests include also big data, computer vision, brain computer
interface, and evolutionary computation. He is working also in the evolving and
increasingly important field of image processing, data analytics, and edge computing,
xiii
xiv About the Editors
which promise to pave the way for the evolution of new applications and services in
the areas of health care, biomedical, agriculture, smart cities, education, marketing,
and finance. His research has appeared in numerous prestigious journals, including
IEEE/ACM Transactions, and at conferences, and he has written more than 100
research papers. Mukesh started his academic career as Lecturer with UTS in 2017
and became Core Member of the University’s world-leading Australian Artificial
Intelligence Institute (AAII), which has a vision to develop theoretical foundations
and advanced technologies for AI and to drive progress in related areas. His research is
backed by industry experience, specifically in Taiwan, where he was Principal Engi-
neer (2016–2017) at the Taiwan Semiconductor Manufacturing Company (TSMC).
There, he developed new algorithms for image processing and pattern recognition
using machine learning techniques. He was also Postdoctoral Researcher leading
a big data and computer vision team at National Chiao Tung University, Taiwan
(2015). Mukesh received an M.S. degree from the School of Computer and Systems
Sciences at the Jawaharlal Nehru University in New Delhi, India (2009), and a Ph.D.
from the Department of Computer Science at the National Chiao Tung University in
Taiwan (2015).
Abstract Deep learning shown its potential in a variety of medical applications and
proved as a count on by people as a step ahead approach compared to traditional
machine learning models. Moreover, the other implementations of these models
such as the convolutional neural networks (CNNs) provide extensive applications
in the field of medicine, which usually involves processing and analysis of a large
dataset. This paper aims to create a CNN model which can solve the problem of
white blood cell subtyping which is a daunting one in clinical processing of blood.
The manual classification of white blood cells in laboratory is a time-consuming
process which gives rise to the need for an automated process to perform the task.
A CNN-based machine learning model is developed to classify the leukocytes into
their proper subtypes by performing tests on a dataset of around twelve thousand
images of leukocytes and their types, and a wide range of parameters is evaluated.
This model can automatically classify the white blood cells to save manual labor,
time and improve efficiency. Further, pretrained models like Inception-v3, VGGNet
and AlexNet are used for the classification, and their performance is compared and
analyzed.
1 Introduction
The primary step in the diagnosis of various illnesses is the detection and subtyping
of patient’s blood. Procedures that automatically perform this task have ground-
breaking utilities in the medical field. Leukocytes usually referred to as white blood
cells (WBCs) are the very important components of our internal functioning [1].
These are responsible for fighting with infection. WBCs are categorized into five
subtypes: neutrophils (60–70%), eosinophils (1–4%), basophils (0.5–1%), lympho-
cytes (20–40%) and monocytes (2–8%) [2]. The ability to record the count and
category of leukocytes and any changes in them acts as an indicator for different ill-
nesses. Increased levels of eosinophils and monocytes can be suggestive of bacterial
infestation. High levels of lymphocytes can point toward presence of diseases such as
leukemia (a type of blood cancer). On the other hand, less neutrophils could indicate
other ailments. Hence, creating a procedure for exact counting and classification of
leukocytes into their subtypes is considered as a prominent problem. The detection
and distinguishing of diverse WBCs is important due to its vast influence in clinical
functions [3].
The rest of the paper is organized as follows: Sect. 2 provides a literature review
about the past relevant work and provides a description about the motivation for
present work. Section 3 provides a background about CNN and its layers and certain
pretrained models. Section 4 provides the proposed methodology, Sect. 5 provides the
experiment analysis and results, and Sect. 6 provides the conclusion and direction in
this area of research.
Robust image processing algorithms have been applied for detecting the nuclei and
classifying WBCs in blood smear images based on features the nuclei. New image-
enhancing techniques are used to manage variations in illumination. Color variations
in nuclei have been managed by using TissueQuant method [4]. The devices that con-
duct blood tests detect WBCs based on traditional techniques like pre-processing,
segmentation, feature extraction, feature selection and classification. Moreover, a
computer-aided automated system has been proposed for easy identification and
location of WBC types in blood images. Additionally, region-based CNNs have
been for classifying the blood cells [5]. Artificial neural network is used to enhance
nucleus by a method called intensity maxima. Classification is done on the basis of
features extracted from segmented images [6]. In addition, the naive Bayes classifier
with Laplacian correction has been used for giving a robust and efficient method to
the problems involving multi-category classification of peripheral Leishman blood
stain images [7]. Principal component analysis and neural network are used for auto-
matic counting, segmentation and classification of WBCs [8]. Classification scheme
using color information and morphology has been proposed for isolating and classi-
fying of WBCs in manually prepared, wright-stained, peripheral blood smears from
whole-slide images [9]. Moreover, SVM classifier and neural network have been
implemented for classification of white blood cells. The segmentation and feature
extraction are also done for classification purpose. WBC segmentation is a two-step
process carried out on the HSV equivalent of the image, using k-means clustering
Leukocyte Subtyping Using Convolutional Neural Networks … 3
3 Background
A brief discussion of CNN and its layers is given in this section. The architectures
of Inception-v3, VGGNet, AlexNet are briefly discussed.
4 M. Sandhya et al.
This network consists of a sequence of layers, where every layer has a separate task.
Figure 1 shows the model diagram of CNN consisting of various layers.
Convolutional Layer This layer retrieves the features present in the input image.
A neural network is not aware about where the features will match exactly in the
image. Hence, it searches for them in the image with the help of filter. A filter
corresponds to a special feature. CNN implements convolution operations using a
filter which slips into the image and multiplies the filter value and the corresponding
pixel of the image. This continues for the leftover filters giving us a collection of
filtered images which is the final output. An auxiliary function known as ReLU
is carried out right after all convolutions. This is a function which is not linear in
nature and is done for each and every pixel to render some level of irregularity into
the network. It does a matrix multiplication, followed by element-wise addition for
changing the -ve pixels to zero in the feature map.
Pooling Layer This layer can be called downsampling whose job is to decrease the
size of all filtered images and retain the knowledge that is considered valuable. It
can be applied in different forms like average or max-pooling, etc. The outcome of
this layer has the exact same quantity of images, although they all consist of lesser
pixels than original. It is helpful for dealing with pre-processing step efficiently.
Fully Connected Layer This operation is performed following a series of convo-
lutional and pooling layers. The shape of convolutional features is transformed into
vector format that is then provisioned for a fully connected layer. The layers of con-
volutions are considered as the foundations for neural networks. They consider the
input to be one vector unlike two-dimensional arrays. They implicate a relationship
among the neurons of the layer before and the layer after them. The outcome as
gathered from the layers of convolutional and pooling contains various sophisticated
features that can be used by the fully connected layers for classifying the image to
the appropriate label based upon the learning data.
3.3 Inception-v3
3.4 VGGNet
This model is distinctly characterized by its simplistic structure which uses just 33
convolutions above one another in a fashion of making the model deeper, while the
maximum pooling provides a smaller magnitude. The network contains two fully
connected layers consisting of 4096 nodes. They are accompanied by a the softmax
classifier [16] (Fig. 3).
3.5 AlexNet
AlexNet is the CNN used particularly in the deep learning applications to computer
vision. It is famous for winning the ImageNet LSVRC competition held in 2012 by a
big margin. AlexNet is deeper with more filters per layer and more than one convo-
Leukocyte Subtyping Using Convolutional Neural Networks … 7
Transfer learning is the process of applying the information that is acquired while
working out a problem to a new problem which is co-related to the previous one.
Figure 5 shows the process of transfer learning. Deep CNNs can offer innovative
support for overcoming several challenges faced in classification. The lack of proper
training data is an extremely common issue in using deep CNN models that usually
require a big amount of data to perform well. Also, the collection of a big dataset
is a tedious process and more so now. Hence, the process of training is very costly.
The complicated network models usually take many days for training with the help
of multiple machines that tend to be exorbitant. Very few people are able to train
the CNN afresh because very rarely are able to gather a sufficing dataset. Hence,
the technique of transfer learning is presently used for overcoming the problem of
small dataset [18]. This technique is very much effective in solving the issue of a
lack of training data. With the help of weights of a model for continuous retrieval of
8 M. Sandhya et al.
distinctive features from the image, most of the issues are taken care of. This method
can be incorporated with the help of training classifier and fine-tuning.
Training Classifier A base network is trained on a database which is then re-purposed
to either learn features or shift them to a target network. This target network has to be
trained on a target database. This process usually works well if the features extracted
are suitable to both base and target jobs, not specific to a single job. The usual method
is truncating the final layer, i.e., softmax layer [18] of the model and replacing it with
a proprietary softmax layer which is pertinent to the task at hand. For example, the
ImageNet model is characterized by a softmax layer having 1000 classes. If the job is
to classify across ten classes, the updated softmax classifier of the pretrained model
will have ten classes rather than having 1000. Backpropagation is applied to the
neural cluster for regulation and tuning of pretrained model weights. The process
of cross-validation is carried out so as to improve the generalization ability of the
model.
Fine-Tuning In the process of fine-tuning, the plan is not just to replace and retrain
the classifier used. It also focuses on regulating the pretrained model weights with
the help of continual backpropagation. Every layer of the convolutional network can
be regulated, or some of the initial layers can be constant, and just the upper layers
can be regulated in order to avoid overfitting. This method is born out of the fact
that the preliminary features of a convolutional network are non-specific ones that
are useful to most of the tasks, but the later layers in the model are more relevant to
the image details for all the labels of the native data.
Softmax Classifier This uses cross-entropy loss function. It allows the computing
probabilities for all classes. It can be implied that the obtained scores given in Fig. 6
as unnormalized and then convert the normalized value of the correct class to be
high which is low for the equivalent value. The final loss is 1.04 using the natural
logarithm.
Leukocyte Subtyping Using Convolutional Neural Networks … 9
Dropout Layer This is a regularization technique for neural network models where
neurons are selected randomly selected and ignored during training. The aim of this
layer is to prevent overfitting. This implies that the contribution of removed neurons
is temporally removed for activating downstream neurons on the forward pass, and
no updations are performed to the weights on the backward pass.
4 Proposed Work
4.1 Dataset
The dataset used in this paper is BCCD dataset [19, 20] which is publicly avail-
able. The dataset contains 12,500 augmented images of blood cells. Approximately,
2500 images and 600 images for each of four different cell types are grouped into
four different folders according to blood cell type in train data and test data. The
images are accompanied by their corresponding cell labels (csv). The blood cell
types are eosinophil, lymphocyte, monocyte and neutrophil [21]. Training model
contains 2497 eosinophil images, 2483 lymphocyte images, 2478 monocyte images
and 2499 neutrophil images. Similarly, testing model contains 623 eosinophil images,
620 lymphocyte images, 620 monocyte images and 624 neutrophil images. Figure 7
shows a sample of blood subtypes in this dataset.
The layout of our simple CNN model is shown in Fig. 8. All the model details are
listed below:
Two Convolutional Layers
• First layer: Kernel size: 3 × 3, number of output filters: 32
• Second layer: Kernel size: 3 × 3, number of output filters: 64
• Activation function: ReLU
10 M. Sandhya et al.
• Stride: 1
• Input image size: 60 × 80 (3 channels).
Pooling Layer
• Pooling type: Maximize
• Pool size: 2 × 2
• Dropout: 0.25.
Hidden Layer
• Number of nodes: 128
Leukocyte Subtyping Using Convolutional Neural Networks … 11
The Inception-v3 pretrained model is combined with the fully connected structure
similar to the one in the simple CNN model 8 and trained with an input image size
of (60 * 80 * 3).
The original AlexNet architecture takes input image size as (227 * 227 * 3), and it is
used for classification of images which will be among 1000 labels, but in our case,
we are taking input image size as (60 * 60 * 3), and we need to classify the images
among only four classes. Hence, the above original architecture will not be apt for
direct use, but we need to make some changes in the above architecture. By following
the basic idea of it, a custom AlexNet architecture is implemented.
12 M. Sandhya et al.
Accuracy, precision, recall and F1-score are the four main evaluation parameters
used to measure performance of CNNs. In the light of our classification problem, the
true and false positives and negatives have been defined as follows:
• True Positive: A WBC accurately classified into either class—eosinophil, neu-
trophil, lymphocyte or monocyte—by the medical professional as well as the
model.
• False Positive: A WBC accurately classified by the model but not the medical
professional.
• False Negative: A WBC accurately classified by the medical professional but not
the model.
• True Negative: A leukocyte cell neither classified by medical professional nor
classified by the model.
Fig. 9 Accuracy curve, loss curve and confusion matrix over four classes using simple CNN model
Fig. 10 Accuracy curve, loss curve and confusion matrix over two classes using simple CNN model
The Inception-v3 model classified the leukocytes with an accuracy of 89.51%. The
evaluation scores of Inception-v3 are given in Table 3. The accuracy curve, loss curve
and confusion matrix are given in Fig. 11.
The VGGNet model classified the leukocytes with an accuracy of 80.26%. The other
recorded metrics were as follows:
14 M. Sandhya et al.
Fig. 11 Accuracy curve, loss curve and confusion matrix using Inception-v3
• Precision: 0.8170
• Recall: 0.7917
• F1-score: 0.7949.
Figure 12 shows the accuracy curve, loss curve and heatmap obtained using VGGNet
model to classify the WBCs.
The AlexNet model classified the leukocytes with a low accuracy of 64.7%. The
other metrics are as given below:
Leukocyte Subtyping Using Convolutional Neural Networks … 15
• Precision: 0.7019
• Recall: 0.6470
• F1-score: 0.6412.
Figure 13 shows the accuracy curve, loss curves and heatmap obtained using AlexNet
model to classify the WBCs.
The reason why AlexNet is showing less accuracy is because of its filter size in
the convolutional layers. In our custom case, the first convolutional layer contains
(5 * 5), but using a 3 × 3 filter expresses most information about the image across all
the channels while keeping the size of the convolutional layers consistent with the
size of the image.
The four different models used to classify the leukocytes gave varying results. It
was observed that Inception-v3 was the best performing model out of the four in
terms of accuracy, precision, recall and F1-score. It was followed by our simple
CNN model and VGGNet model. The AlexNet model fared low compared to the
other models. Table 4 shows the comparison of the models based of the various
performance metrics.
6 Conclusion
In this paper, features are extracted from blood cell images that were trained, tested
and validated. The use of convolutional neural networks (CNNs) in medical image
processing is immense as is proven in this study. The leukocytes are classified
into four subtypes, namely neutrophils, eosinophils, monocytes and lymphocytes
by building CNN models. Initially, a simple CNN model is used for classification.
The concept of transfer learning is employed in the later models. This is achieved
by using pretrained models such as Inception-v3, VGGNet and AlexNet. The results
of all the models are recorded by measuring the various performance metrics. These
results are then compared with each other and analyzed to rank the performance.
This gives an insight into how different models perform on the same dataset and also
suggests what models can perform better in a real-world scenario for classification
of leukocytes. The highest achieved accuracy of 89.51% proves that this method can
fare very well in real-time classification of white blood cells. Other nature-inspired
models can be used to reach the optimal value and use the appropriate number of
layers to improve the accuracy of the CNN models in leukocyte subtyping.
References
1. Doan, C.A.: The white blood cells in health and disease. Bull. N. Y. Acad. Med. 30(6), 415
2. Rezatofighi, S.H., Soltanian-Zadeh, H.: Automatic recognition of five types of white blood
cells in peripheral blood. Comput. Med. Imaging Graphics 35(4), 333–343 (2011)
3. Su, M.-C., Cheng, C.-Y., Wang, P.-C.: A neural-network-based approach to white blood cell
classification. Sci. World J. (2014)
4. Hegde, R.B., Prasad, K., Hebbar, H., Singh, B.M.K.: Development of a robust algorithm for
detection of nuclei and classification of white blood cells in peripheral blood smear images. J.
Med. Syst. 42(6), 110 (2018)
5. Kutlu, H., Avci, E., Özyurt, F.: White blood cells detection and classification based on regional
convolutional neural networks. Med. Hypotheses 135, 109472 (2020)
6. Manik, S., Saini, L.M., Vadera, N.: Counting and classification of white blood cell using
artificial neural network. In: 2016 IEEE 1st International Conference on Power Electronics,
Intelligent Control and Energy Systems, pp. 1–5. IEEE (2016)
7. Mathur, A., Tripathi, A.S., Kuse, M.: Scalable system for classification of white blood cells
from Leishman stained blood stain images. J. Pathol. Inform. (2013)
8. Nazlibilek, S., Karacor, D., Ercan, T., Sazli, M.H., Kalender, O., Ege, Y.: Automatic segmen-
tation, counting, size determination and classification of white blood cells. Measurement 55,
58–65 (2013)
9. Ramesh, N., Dangott, B., Salama, M.E., Tasdizen, T.: Isolation and two-step classification of
normal white blood cells in peripheral blood smears. J. Pathol. Inform. 3 (2012)
10. Sinha, N., Ramakrishnan, A.G.: Automation of differential blood count. In: Conference on
Convergent Technologies for Asia-Pacific Region, vol. 2, pp. 547–551. IEEE (2003)
11. Özyurt, F.: A fused CNN model for WBC detection with MRMR feature selection and extreme
learning machine. Soft Comput. 1–10(2019), 547–551 (2003)
12. Togacar, M., Ergen, B., Sertkaya, M.E.: Subclass separation of white blood cell images using
convolutional neural network models. Elektron. Elektrotech. 25(5), 63–68 (2019)
Leukocyte Subtyping Using Convolutional Neural Networks … 17
13. Sharma, M., Bhave, A., Janghel, R.R.: White blood cell classification using convolutional
neural network. In: Soft Computing and Signal Processing, pp. 135–143. Springer, Singapore
(2019)
14. Agarwal, C.C.: Neural Networks and Deep Learning. Springer, Cham
15. Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architec-
ture for computer vision. In: Proceedings of International Conference on Computer Vision and
Pattern Recognition (2016)
16. Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A.,
Khosla, A., Bernstein, M., Berg, A.C., Fei-Fei, L.: ImageNet large scale visual recognition
challenge. IJCV (2015)
17. Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional
neural networks. Advances in Neural Information Processing Systems, vol. 25 (NIPS 2012)
18. Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press, Cambridge, MA (2016)
19. https://github.com/Shenggan/BCCD_Dataset
20. https://www.kaggle.com/paultimothymooney/blood-cells
21. Özyurt, F.: A fused CNN model for WBC detection with MRMR feature selection and extreme
learning machine. Soft Comput. (2020)
Analysis of Fifteen Approaches
to Automated COVID-19 Detection Using
Radiography Images
1 Introduction
The remarkable increase in the spread of the novel coronavirus has put shocking pres-
sure on healthcare systems across the world. COVID-19 began initially as reporting
of pneumonia with unknown causes in Wuhan, Hubei province of China. This virus
spread uncontrollably throughout the world due to a lack of therapeutic medication,
vaccines, and lack prior medical knowledge. Almost every healthcare system in the
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 19
D. Gupta et al. (eds.), Advanced Machine Intelligence and Signal Processing, Lecture
Notes in Electrical Engineering 858, https://doi.org/10.1007/978-981-19-0840-8_2
20 K. Soni et al.
world was caught off-guard because of the rate of spread of the virus coupled with
the limited resources in hospitals such as beds, ventilators, and PPE kits.
The key to containing the virus spread by the infected individual and save lives
in this pandemic lies in preemptive detection of covid. The most widely used test
right now to detect COVID-19 is RT-PCR. Because of the low sensitivity of RT-PCR
tests, especially in mild cases, and most importantly, the lack of resources and time
to conduct the test on so many people, chest scans remain a vital way in detecting
early signs of COVID-19 in the patient, especially as X-ray facilities are readily
obtainable in healthcare systems of most countries across the globe. Studies proved
that visual abnormalities characteristic of a COVID-19 infection is present in chest
X-ray images.
Nonetheless, there are certain limitations to chest scans: time for image acquisi-
tion, cost for CT scans, unavailability in financially challenged areas.
At the same time, X-rays provide rapid triaging as they are cheaper, portable,
highly available for faster diagnosis, and pose less threat to the patient in the form
of radiation in comparison to CT scans. Due to the portable nature of X-rays, they
significantly reduce the risk of the spread of the virus via the transportation route of the
patient. The key factor in these radiological images to diagnose covid is the presence
of opacity-related findings. Ground glass (57%) and mixed attenuation (29%) are
the opacities most frequently reported [16]. During the early stages, ground glass is
minutely observed and may be difficult to see visually. Patchy or diffuse airspace
is other subtle abnormalities that are difficult to observe and usually interpreted by
a trained radiologist. It is said that CT and X-rays are successful methods to detect
COVID-19 coupled with RT-PCR. Radiology images are being widely used to detect
COVID-19 in patients in countries like Turkey where they are facing a paucity of
testing kits at the onset of the pandemic. In some cases, CT/X-ray images have shown
changes in the images of lungs even prior to COVID-19 symptoms’ onset. The biggest
tailback faced here is the requirement for healthcare professionals to spend their
time interpreting the radiography images. As such, automated AI techniques will
help radiologists correctly interpret CXR images to help detect COVID-19 cases
more rapidly and are vital, especially when the world finds itself without enough
healthcare personnel to fight this incredibly widespread pandemic, and how our
frontline workers find themselves massively overworked. Furthermore, a lack of
enough RT-PCR test kits in comparison to the number of growing cases means
further reliance on radiography imaging, especially when there is a cost of tests and
time to process the test results associated (Fig. 1).
As such, reliable and automated techniques that can interpret hundreds of images
with maximum accuracy in a short period are highly desired. Subtle abnormalities
which are present in CXR are difficult to spot even by trained radiologists. Consid-
ering the sheer number of reports to be reviewed and the limited number of trained
radiologists, human error due to exhaustion can play a factor in an incorrect diagnosis.
Automatic methods for identifying subtle abnormalities will significantly reduce the
mounting pressure on our healthcare system. Deep learning models can help us in
medical image analysis with significantly less time and human intervention to provide
a more precise diagnosis.
Analysis of Fifteen Approaches to Automated COVID-19 Detection … 21
Fig. 1 Chest X-ray images of (a) person with normal lung with no infection (b) person with
bacterial pneumonia (c) person with viral pneumonia (d) person with COVID-19
Paper organization
Section 2 of the paper elucidates related works. Section 3 gives an overview of the
fifteen approaches individually before using them is scrutinized and contrasted in
Sect. 4, the discussion and comparative analysis section. Section 5 concludes the
paper and discusses its future scope.
1.1 Motivation
The remarkable increase in the spread of the novel coronavirus has put shocking pres-
sure on healthcare systems across the world. COVID-19 began initially as reporting
of pneumonia with unknown causes in Wuhan, Hubei province of China. This virus
spread uncontrollably throughout the world due to a lack of therapeutic medication,
vaccines, and lack prior medical knowledge. Almost, every healthcare system in the
22 K. Soni et al.
world was caught off-guard because of the rate of spread of the virus coupled with
the limited resources in hospitals such as beds, ventilators, and PPE kits.
2 Related Work
ImageNets which were already pretrained CNN were used like GoogleNet, ResNet,
SqueezeNet, and VGG19. In [15], by Minaee et al., they pointed out that due to
the lack of a publicly available image dataset of X-ray, we cannot use CNN models
which are trained from scratch; fine-tuning of the final layer of the pretrained cate-
gory was done on the ImageNet database. Unlike traditional analysis, they used
end-to-end deep learning frameworks that predict COVID-19 directly from unpro-
cessed images without extracting features. Satpathy et al. in [17] used artificial intel-
ligence methods for mortality rate prediction in COVID-19 cases. Poonam et al. in
[18] analyzed factors that influence the spread of COVID-19. Sharma et al. in [19]
analyzed statistically the data of patient cases in Karnataka, India.
3 Overview
This section gives a brief introduction and overview of all fifteen papers individually.
Sara Hosseinzadeh et al. [6] collected publicly available CXR and available CT
images, and in the following step preprocessed the given dataset utilizing the usual
normalization strategies to better the nature of the input data information. Once the
images fed into were made, with the help of CNN, descriptors [quote] were fed in the
feature extraction step to extract the deep features of every input image. In training,
these functions were then inserted into ML classifiers like random forest, AdaBoost,
XGBoost, decision tree, bagging classifier, and LightGBM to determine if it was
COVID-19 or less case or control. Finally, the performance was evaluated on test
images. The best accuracy of 99% was achieved when the bagging tree classifier
along with the DenseNet-121 feature extractor was used. The next best accuracy
was that of a type of the feature extractor ResNet50 used with LightGBM whose
resulting accuracy was 98%. The authors used a dataset of publicly available X-
rays from DrJoseph Cohen, on the GitHub repository, from the aforementioned 117
chest X-rays and 20 CT images with a COVID-19 positive result. They also used
117 images of healthy patients with X-rays from the available Kaggle chest X-ray
images dataset (Pneumonia) and 20 images of healthy patients with CT images from
the Kaggle dataset. RSNA Pneumonia detection includes a record of positive and
normal cases. Sachin Sharma took CT scans of people with COVID-19 infection, [7]
next viral pneumonia, and normal healthy people and recorded them on a computer.
He then preprocessed the image, i.e., resizing and cropping the image, to extricate
the effective lung regions before taking the dataset into the analysis. The following
properties were selected: ground glass opacities and pleural effusion, as they are
distinguishable in CTS. So, he created separate folders for each of the three discrete
categories. To train the machine, specialized computer vision software analogical to
the Microsoft Azure residual neural network (ResNet) architecture was used. After
receiving the results, they are compared with the actual status (COVID-19 or other
viral pneumonia or a normal health case) of the patient to verify the accuracy of
the model. He achieved an accuracy of 91% for COVID-19 classification using this
technique. All images were collected from official databases of various hospitals
24 K. Soni et al.
in China, Italy, Moscow, and India. Approximately, 2200 images were collected
including 800 CT scans of COVID-19 infected patients, 600 CT images of rest
patients with viral pneumonia, and 800 CTs of healthy people.
Kadry et al. [8] performed a CT scan of the patient and acquired a three-
dimensional (3D) image of the lungs. As the 3D image is difficult to analyze, it
was converted into 2D sections for testing. A normal level of CTS/COVID-19 is
considered a test of the effectiveness of the proposed MLS. Chaotic bat algorithm
and Kapur’s entropy (CBA+KE) were used to improve the image quality of the
infected part as they used a three-level threshold. Next, a two-stage threshold filter
was used to divide the images into region of interests (ROI) and artifacts; then, a
feature extraction process was taken up to extract features from the initial image.
The dominant characteristics of each image type were selected using a statistical
test, and then, the selected characteristics are used for training, testing, and vali-
dating the classification system. In addition, the future blending method is also being
considered to increase the accuracy of the classification. This paper gave an accu-
racy of 89.80%. Normal CT scans were taken from LIDC-IDRI and RIDER-TCIA.
COVID-19 class images were acquired, warning from the Radiopaedia database and
reference images. Rachna Jain et al. learning algorithms PA image of CXR scans
for patients who contracted COVID-19 and vice-versa. After clearing the images
and applying data processing, CNN models based upon deep learning were used,
and their performance was analyzed and compared. Xception, ResNeXt, and Incep-
tion V3 models were compared, and their accuracy was examined. When analyzing
results, the Xception model provided the maximum accuracy of 97.97% for detecting
COVID-19 for CXRs. Mete et al. [10] examined the ability of 4004 ML techniques
to detect COVID-19 from CXRs as early as possible. This study considered compact
classifiers and deep learning approaches. In addition, a newer compact classifier, the
convolutional support estimator network (CSEN), was used for the same because it
is suitable for misclassifying data. The CSEN variant of models gives the highest
sensitivity level around >97%, whereas DenseNet-121 gives a decreased sensitivity
with higher specificity. Wang et al. [1] presented a human–machine-concerted design
strategy, and using DarwinAI, the Covid-Net CNN architecture was built in less than
7 days based on generator inquisitor pair (GSInquire 37) and was later thoroughly
audited (checking if the right classification is being made for the right reasons, etc.),
giving excellent positive predict value (PPV) and sensitivity, and the COVIDx dataset
was open sourced. Generative synthesis tool was used to provide granular insights
into neural network performance. The machine-carried design exploration was done
based on human-specific design requirements (sensitivity and PPV >80%), ensuring
high sensitivity as well as limiting false positives. Unique characteristics of the
network architecture, such as PEPX models of design, selective far communication,
and great diversity in architecture, are allowed for a computationally efficient yet
effective Covid-Net. A high accuracy of 93.3% was achieved. The dataset used was
COVIDx, consisting of CXRs from people. Jain et al. [2] carried out deep networks
in two phases involving two ResNet CNNs–ResNet50 and ResNet101. The former,
with 50 layers, to classify among viral, bacterial pneumonia, or normal, and if the
Analysis of Fifteen Approaches to Automated COVID-19 Detection … 25
image is classified as ‘viral’, it is then fed into the latter CNN to classify into COVID-
19 or other viruses (non-covid). Transfer learning was used on both CNNs, using
pre-trained CNNs on the ImageNet database, due to which the initial layers could
be made stagnant and only the final layers needed to be trained. Data augmenta-
tion techniques were employed to counter the class imbalance. The finding of the
optimal learning rate, as well as the overall implementation, was done in Python with
the assistance of the FastAi library. The dataset used consisted of chest X-rays of
normal healthy people, pictures of patients with bacterial pneumonia, and pictures
of viral pneumonia, from Cohen and Kaggle. Stage 1 model gave 93% accuracy,
and the Stage 2 model gave 97.77% accuracy. Abraham et al. [3] used multiple
pretrained networks, and the CFS algorithm is for feature selection in combination
with a forward selection-based search method; SSFS was exploited to get the best
subset of features. In combination with them, the Bayesnet classifier gave the best
accuracy. In combination with them, the Bayesnet classifier proved to be giving the
best results, especially with multi-CNNs. Dataset: Cohen et al. dataset of 560 CXR
images—453 COVID-19 and 107 non-COVID images. Chandra et al. [4] preferred
machine learning approaches over deep learning ones since, according to the authors,
they can work with lesser data as well as in a constrained environment, in contrast
with the ‘data-hungry deep learning algorithm. The given study uses eight first-order
statistical features (FOSF), which describes the complete image by using several
parameters such as the mean, variance, and roughness, 88 GLC matrix features and
8100 HOG features. The FOSF does not take care of the local information. As such,
the HOG and GLCM feature descriptors are utilized to do a thorough analysis of
texture. The GLCM describes the spatial correlation between intensities of pixels in
radio texture patterns in four directions, and the local information is encoded by HOG.
These statistical features can encode natural textures efficiently. For feature selection,
binary gray wolf optimization was used. Outputs from five ML approaches, SVM,
DT, KNN, NB, and ANN were considered, and the final prediction is the majority of
votes of the seven classifiers. The grid search algorithm selects the optimal hyperpa-
rameters while curtailing the losses of cross-validation. The dataset used was 2088
CXR images (696 each of normal, pneumonia, and COVID-19) for phase 1 and 258
images (86 images of each class) for phase 2. The accuracy achieved was 93.411%.
Arpan et al. [5] used a pretrained 121-layered DenseNet, CheXNet since the
visually identical inputs of the samples, the authors’ findings were this to be the
most accurate pretrained backbone to developing a model for identifying COVID-
19. The final layers were trained to adapt to COVID-19 detection, by replacing
the final 14 class classifier of CheXNet with a four-class classification layer. The
number of prediction classes was clubbed from 4 to 3 as that increased the model’s
overall precision. The dataset used was 5323 training CXR images, 654 test images,
and 37 validation images. An accuracy of 90.5% was achieved. Abbas et al. [14]
use DeTrac, by exploring the boundaries of its classes using a class decomposition
engine irregularities present in the image dataset were eliminated. They adapted their
previously proposed model to boost the efficacy of the model. The optimization was
done during the adoption and training of the previously trained ImageNet model. The
26 K. Soni et al.
dataset used here was the CXR images in the Japanese Society of Radiological Tech-
nology (JSRT) and the CXR images Cohen et al. (2020) COVID-19 images. Minaee
et al. [15] a dataset consisting of 5000 images was prepared by combining augmented
images which using based detection increased the factor of images available by 5 and
COVID-X-ray-5 k dataset. They used the transfer learning approach. They worked
on these four convolutional networks, DenseNet-121, ResNet50, SqueezeNet, and
ResNet18. The confusion matrix of each model precision-recall curve, character-
istic (ROC) curve, and average prediction, receiver operating was also presented.
Following the technique of Zeiler and Fergus, they also predict the infected region
of the chest. All results were obtained by fine-tuning four pre-trained CNN models
above. The dataset used was CXR images Cohen et al. (2020) COVID-19 images,
compilation CXR images Japanese Society of Radiological Technology (JSRT).
Ozturk et al. [11] worked on the Darknet-19 classifier model which forms the
basis of YOLO. Being inspired by DarkNet-19, they created the DarkCovidNet.
Unlike ResNets or ResNext, we need a less deep CNN which can identify subtle
details. Thus, they came up with DarkCovidNet which has 17 layers. Each CN
layer is added before a BatchNorm and LeakyReLU operations, where BatchNorm
standardizes the inputs as this decreases time and training, increasing stability. They
categorized the images into three classes—pneumonia, COVID-1, no findings. They
trained DarkCovidNet to classify the images into COVID-19, No-Findings. Dataset
Used: chest X-ray14, COVID-19 chest X-ray dataset, COVID-19 X-ray dataset,
compiled by researchers of University of Montreal. Khan et al. [13] worked on a
deep CNN model named CoroNet to identify COVID-19 by analyzing CXRs. This
model was constructed based on Xception architecture which was then trained end-
to-end and preemptively trained on the ImageNet dataset which was put together
by accumulating COVID-19 and pneumonia publicly in CXR available databases.
Three scenarios were implemented of the discussed model to detect COVID-19
present in the CXR. The key multi-class model is the first one (4-class CoroNet) that
was trained to classify CXRs films into the following of the four categories, namely
pneumonia-viral, pneumonia-bacterial, COVID-19, and normal. They have discussed
thee-class CoroNet (pneumonia, normal, and COVID-19) and binary CoroNet as
the alterations of the key multi-class model. Implementation of CoroNet is done
using Keras worked upon Tensorflow v2.0. CoroNet acquired an average accuracy—
96.6%, average accuracy of 89.6%, F-measure (F1-score)—95.6%, recall—98.25%,
and precision—93.17% which was achieved for the COVID-19 model.
Mohd et al. [12] independent sets for every training, validation, and testing phase
were used in their paper. The depth in this deep learning methodologies is significantly
numerous in visually recognizable applications. By using ResNet101, a 101-layer
CNN that they adopted for this study due to the advantage of the remaining learning
structure, which is known to be computationally less expensive than its counterpart
and also not sacrificing depth and thus maintaining the precision. The performance
achieved of a binary classified model with sensitivity—77.3%, specificity—71.8%,
and precision—71. 9%. Dataset used: chest X-ray14, COVID-19 chest X-ray dataset.
A dataset compiled by researchers of the University of Montreal were also used (Table
1).
Analysis of Fifteen Approaches to Automated COVID-19 Detection … 27
Table 1 (continued)
Paper No. Pros Cons
[5] • Infected regions were indicated • Significant computational cost
• Significant improvements over • Accuracy needs significant
Covid-Net were made improvement for the model to be
production ready
[6] • Faster Detection: With the method • Extracting features from X-ray analysis
described above, reports were is important when training models for
generated much faster, and a person ML because the performance of this
with COVID-19 could be easily model is directly linked to the quality
identified with greater accuracy of the extracted features. This means
• No complex data features are required that if the scans are not clear, there can
• Overcame overfitting issue be an error in the detection
• Increased generalization ability • Bagging was a slow learner but gave
• Deep CNN’s bag tree classification better accuracy and hence can delay
showed excellent COVID-19 data the release of the reports
classification despite the lack of data • Taken from scratch, the deep CNN
models are faster for computation and
less resource utilization, which is much
better than the proposed approach
[7] • The author turns to another technique • This technique got confused in a few
other than the traditional RT-PCR, cases, and hence, accuracy was not all
which requires kits that put a strain on that high
healthcare professionals, requiring
only easy-to-generate CT reports
[8] • Does not require operator assistance: • Accuracy could be improved for
The MLS method used in this production use
document is automated
• The MLS technology, despite of
orientation, it has better segmentation
of CTS
[9] • Gives the best performance and can be • The high accuracy of the XceptionNet
fruitfully utilized in the future was concerning due to overfitting
[10] • X-ray has less exposure to radiation • The author uses deep features with
and reduces the risk of spreading the CSENs but with lesser sensitivity
disease as compared to CT scans • The decision of methods is difficult to
• The CSEN uses deep learners which interpret since they are black-box
improved specificity techniques
• Faster technique than traditional
RT-PCR
[11] • It performed well in binary class • Poor results with low-quality images
classification with ARDS
• The model can be used to assist experts • The success rate is low in multi-class
classification
[12] • By an estimate, it is 258 times more • Because images were less in number
proficient than a radiologist during testing, the results presented
• With a GPU, an even better here are preliminary
performance can be achieved
(continued)
Analysis of Fifteen Approaches to Automated COVID-19 Detection … 29
Table 1 (continued)
Paper No. Pros Cons
[13] • Great recall (Sensitivity) and precision • Better accuracy can be achieved
(PPV) for covid cases
• The accuracy achieved for the
following cases: 4 classes—89.5%, 3
classes—94.59%, and binary class
classification—99%
[14] • Overcomes Lack of Datasets: Using • Usability scope
transfer learning and using already • Not deployable on handheld devices
gained insights from pretrained CNNs
for a specific medical imaging task
• DeTRac outperforms ResNet,
GoogleNet, ALexNet, SqueezeNet,
and VGG in the class decomposition
layer with a sensitivity of 100%
[15] • They can predict the infected region of • The dataset used has only 200 covid
the chest and visualize it using a heat images
map • Not conducted on a large scale
accuracy obtained could have been a cause of overfitting. Mete et al. [10] used CSENs
which resulted in high specificity, whereas it compromised on the sensitivity.
Wang et al. [1] were another impressive work. The network architecture was built
in <7 days using a human–machine collaborative design strategy, and very special
care was taken to keep the sensitivity and PPV high, which is crucial. With no
environmental constraints, this model can be deployed. As a matter of fact, Arpan
et al. [5], as per its authors, gave improved results over Covid-Net, and hence, with
enough resources, the state-of-the-art Covid-Net can be employed. The two-stage
architecture in [2] is a bit computationally costly as it uses two CNNs, and even
though stage 2 of this architecture has an excellent accuracy of 97.77%, the overall
model accuracy is dependent upon stage 1’s accuracy as well, which is 93%. The
overall model is excellent, and if the accuracy of stage 1 is improved, this could be the
preferred model in situations when there are no constraints on available resources. In
[3], the best permutation of using CFS, SFSS, CFG feature selection technique, and
Bayesnet classifier along with multiple pre-trained networks bode well for the model.
The accuracy of 91.15%, even though excellent, but some of the earlier mentioned
models did better. For instance, Chandra et al. [4] gives better accuracy despite using
Machine Learning techniques and hence lesser computational power than the model
in question. To overcome the lack of image datasets particular to our research, many
studies have used data augmentation techniques to produce transformed versions of
images (such as minute distortions and rotations.) and use transfer learning to better
train our CNN. One such paper that deals with is [14] using decompose, transfer,
and compose (DeTraC), since DeTrac transfers knowledge acquired in larger-scale
more generalized image recognition work to a domain-specific required calculation,
so we save computational time required to train a CNN from scratch. This model
is not handheld device friendly as it needs more comp multiple papers dealing with
multi-class classification (pneumonia, normal, and COVID-19) [1, 11] and [13].
Accuracy for multi-class classification has varied across papers since with some
models. In the paper [11], images of lungs with pneumonia are also considered in the
research afterward. The model ended up identifying COVID-19 patients classified in
pneumonia class. Even though COVID-19 being a kind of lung infection classified
in the model, the diagnosis is right, but the interpretation looks incorrect. So, some
patients will be classified in the pneumonia class because this probability of success
of the model in the problem of classifying binary classes is high compared to several
classes.
4.1 Discussion
The models proposed should help healthcare workers in drawing accurate and rela-
tively fast results. They require models which are not only accurate but portable too.
As we see in [1, 5, 14], these models require significant computational cost and are
not production ready. Except Sharma [7], we don’t see a platform or GUI which
makes operating the model easier even for people without a technical background
Analysis of Fifteen Approaches to Automated COVID-19 Detection … 31
5.1 Conclusion
Recently, a lot of researchers have come forward with several good approaches to
tackle the pandemic via automated COVID-19 detection. In this work, we have tried
to compare, contrast, and present a coherent idea about fifteen such approaches. Some
approaches stand out when a model needs to be worked in a constrained environment,
whereas some stand out despite requiring higher computational resources, for their
high sensitivity and positive predict value (PPV), both of which are extremely crucial
in our context; since a false positive can overburden, we already strained healthcare
systems using the authors false negative can result into spreading of the virus by the
misdiagnosed patient. We hope that this helps future researchers in coming up with
innovative ideas of their own.
The limitations that were faced in most of the studies can be dealt with more avail-
able patient data (both asymptomatic patients and symptomatic and) which will be
available in the future to help with more deep analysis. More disease-specific (cold,
pneumonia, COVID-19) features can be included to help the machine learn and
differentiate between these and give accurate, reliable results with high sensitivity.
To allow deployment on handheld devices, we need to increase efficiency by model
32 K. Soni et al.
pruning and quantization. With the ongoing pandemic, we hope to increase our avail-
able dataset which will further improve the research and experimental work and will
achieve more validation for the method with larger datasets. Moreover, with growing
research in this area, there will be many future studies worth considering. They can
be incorporated with the current work as well.
References
1. Wang, L., Lin, Z.Q., Wong, A.: Covid-net: a tailored deep convolutional neural network design
for detection of COVID-19 cases from chest x-ray images. Sci. Rep. 10(1), 1–12 (2020)
2. Jain, G., Mittal, D., Thakur, D., Mittal, M.: A deep learning approach to detect COVID-19
coronavirus with X-Ray images. Biocybernetics Biomed. Eng. 40(4), 1391–1405 (2020)
3. Abraham, B., Nair, M.: Computer-aided detection of COVID-19 from X-ray images using
multi-CNN and Bayesnet classifier. Biocybernetics Biomed. Eng. 40(4), 1436–1445 (2020)
4. Chandra, T., Verma, K., Singh, B., Jain, D., Netam, S.: Coronavirus disease (COVID-19)
detection in chest X-Ray images using majority voting-based classifier ensemble. Expert Syst.
Appl. 165, 113909 (2021)
5. Arpan, M., Surya, K., Harish, R., Krithika, R., Vinay, N., Subhashis, B., Chetan A.: CovidAID:
COVID-19 detection using chest X-ray. arXiv preprint (2020)
6. Sara Hosseinzadeh, K., Mike, W., Peyman Hosseinzadeh, K., Kevin, A.: Automatic detection of
coronavirus disease (COVID-19) in X-ray and CT images: a machine learning-based approach.
(n.d.) (2020)
7. Sharma, S.: Drawing insights from COVID-19-infected patients using CT scan images and
machine learning techniques: a study on 200 patients. Environ. Sci. Pollut. Res. 27(29), 37155–
37163 (2020)
8. Kadry, S., Rajinikanth, V., Rho, S., Raja, N.S.M., Rao, V.S., & Thanaraj, K.P.: Development of
a machine-learning system to classify lung CT scan images into normal/covid-19 class. arXiv:
2004.13122 (2020)
9. Jain, R., Gupta, M., Taneja, S., Hemanth, D.: Deep learning-based detection and analysis of
COVID-19 on chest X-ray images. Appl. Intell. 51(3), 1690–1700 (2020)
10. Mete, A., Aysen, D., Mehmet, Y., Serkan, K., Muhammad, E., Chowdhury, H., Khalid, H.,
Tahir, H., Rashid, M., Moncef, G.: Advance warning methodologies for covid-19 using chest
x-ray images. IEEE Access 9, 41052–41065 (2021)
11. Alqudah, A.M., Qazan, S., Alqudah, A.: Automated systems for detection of COVID-19 using
chest X-ray images and lightweight convolutional neural networks (2020)
12. Che A., Mohd, Z.M.Z., Hassan, R., Mohd Tamirn, M.I., Md Ali, M.A.: COVID-19 deep
learning prediction model using publicly available radiologist-adjudicated chest X-ray images
as training data: preliminary findings. Int. J. Biomed. Imag. (2020)
13. Agarwal, C., Khobahi, S., Schonfeld, D., Soltanalian, M.: CoroNet: a deep network architecture
for enhanced identification of COVID-19 from chest X-ray images. Med. Imag. Comput.-Aided
Diagnosis (2021)
14. Abbas, A., Abdelsamea, M.M., Medhat, M.: Classification of COVID-19 in chest X-ray images
using DeTraC deep convolutional neural network (2020)
15. Minaee, S., Kafieh, R., Sonka, M., Yazdani, S., Jamalipour Soufi, G.: Deep-COVID: predicting
COVID-19 from chest X-ray images using deep transfer learning. Med. Image Analysis 65,
101794 (2020)
16. Kong, W., Agarwal, P.P.: Chest imaging appearance of COVID-19 infection. Radiol. Cardio-
thoracic Imag. 2(1), e200028
17. Satpathy, S., Mangla, M., Sharma, N., Deshmukh, H., Mohanty, S.: Predicting mortality rate
and associated risks in COVID-19 patients. Spatial Inf. Res. 1–10 (2021)
Discovering Diverse Content Through
Random Scribd Documents
FOWLER, Japan, China, and India, 10s. 6d.
FRA ANGELICO. See Great Artists.
FRA BARTOLOMMEO, ALBERTINELLI, and ANDREA
DEL SARTO. See Great Artists.
FRANC, Maud Jeanne, Beatrice Melton, 4s.
—— Emily's Choice, n. ed. 5s.
—— Golden Gifts, 4s.
—— Hall's Vineyard, 4s.
—— Into the Light, 4s.
—— John's Wife, 4s.
—— Little Mercy, for better, for worse, 4s.
—— Marian, a Tale, n. ed. 5s.
—— Master of Ralston, 4s.
—— Minnie's Mission, a Temperance Tale, 4s.
—— No longer a Child, 4s.
—— Silken Cords and Iron Fetters, a Tale, 4s.
—— Two Sides to Every Question, 4s.
—— Vermont Vale, 5s.
A plainer edition is published at 2s. 6d.
France. See Foreign Countries.
FRANCIS, F., War, Waves, and Wanderings, 2 vols.
24s.
—— See also Low's Standard Series.
Frank's Ranche; or, My Holiday in the Rockies, n.
ed. 5s.
FRANKEL, Julius, Starch Glucose, &c., 18s.
FRASER, Bishop, Lancashire Life, n. ed. 12s. 6d.;
popular ed. 3s. 6d.
FREEMAN, J., Melbourne Life, lights and shadows,
6s.
FRENCH, F., Home Fairies and Heart Flowers,
illust. 24s.
French and English Birthday Book, by Kate D.
Clark, 7s. 6d.
French Revolution, Letters from Paris, translated,
10s. 6d.
Fresh Woods and Pastures New, by the Author of
"An Angler's Days," 5s., 1s. 6d., 1s.
FRIEZE, Duprè, Florentine Sculptor, 7s. 6d.
FRISWELL, J. H. See Gentle Life Series.
Froissart for Boys, by Lanier, new ed. 7s. 6d.
FROUDE, J. A. See Prime Ministers.
Gainsborough and Constable. See Great Artists.
GASPARIN, Sunny Fields and Shady Woods, 6s.
GEFFCKEN, British Empire, 7s. 6d.
Generation of Judges, n. e. 7s. 6d.
Gentle Life Series, edited by J. Hain Friswell, sm.
8vo. 6s. per vol.; calf extra, 10s. 6d. ea.; 16mo,
2s. 6d., except when price is given.
Gentle Life.
About in the World.
Like unto Christ.
Familiar Words, 6s.; also 3s. 6d.
Montaigne's Essays.
Sidney's Arcadia, 6s.
Gentle Life, second series.
Varia; readings, 10s. 6d.
Silent hour; essays.
Half-length Portraits.
Essays on English Writers.
Other People's Windows, 6s. & 2s. 6d.
A Man's Thoughts.