0% found this document useful (0 votes)
7 views

Breast Cancer Diagnosis in Mammography Images Using Deep Convolutional Neural Network-Based Transfer and Scratch Learning Approach

This research article presents a study on breast cancer diagnosis using mammography images through a Deep Convolutional Neural Network (Deep-CNN) model called MobileNetV2, focusing on transfer and scratch learning methods. The study evaluates the classification performance of mammography images into malignant or benign categories, achieving an accuracy of 0.994 with transfer learning and 0.96 with a scratch learning approach. The findings suggest that the proposed model demonstrates superior performance compared to existing models, making it suitable for implementation on low-cost embedded platforms.

Uploaded by

shanthinidevip01
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
7 views

Breast Cancer Diagnosis in Mammography Images Using Deep Convolutional Neural Network-Based Transfer and Scratch Learning Approach

This research article presents a study on breast cancer diagnosis using mammography images through a Deep Convolutional Neural Network (Deep-CNN) model called MobileNetV2, focusing on transfer and scratch learning methods. The study evaluates the classification performance of mammography images into malignant or benign categories, achieving an accuracy of 0.994 with transfer learning and 0.96 with a scratch learning approach. The findings suggest that the proposed model demonstrates superior performance compared to existing models, making it suitable for implementation on low-cost embedded platforms.

Uploaded by

shanthinidevip01
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

INDIAN JOURNAL OF SCIENCE AND TECHNOLOGY

RESEARCH ARTICLE

Breast Cancer Diagnosis in


Mammography Images Using Deep
Convolutional Neural Network-Based
Transfer and Scratch Learning
OPEN ACCESS
Received: 06-01-2023
Approach
Accepted: 01-04-2023
Published: 09-05-2023 Aarti Bokade1 , Ankit Shah2 ∗
1 Research Scholar, Ph.D., Gujarat Technological University, Ahmedabad, 380005, Gujarat,
India
2 Assistant Professor, Instrumentation & Control Engineering Department, L.D.College of
Citation: Bokade A, Shah A (2023) Engineering, Ahmedabad, 380015, Gujarat, India
Breast Cancer Diagnosis in
Mammography Images Using Deep
Convolutional Neural
Network-Based Transfer and Abstract
Scratch Learning Approach. Indian
Journal of Science and Technology Objectives: The study aims to utilize a Deep Convolutional Neural Net-
16(18): 1385-1394. https://doi.org/ work (Deep-CNN) model called MobileNetV2, which has low computational
10.17485/IJST/v16i18.39
requirements, to accomplish binary classification of mammography images.

Corresponding author. To achieve this objective, the study investigates two methods: transfer learn-
[email protected] ing and scratch learning. Methods: The proposed technique aims to classify
Funding: None
mammography images from the Digital Database of Screening Mammography
(DDSM) dataset into either malignant or benign categories using transfer learn-
Competing Interests: None
ing and scratch learning methods based on Deep-CNN. Before being fed into
Copyright: © 2023 Bokade & Shah. the Deep-CNN, the images’ contrast level is enhanced using the min-max con-
This is an open access article
distributed under the terms of the trast enhancement technique. MobileNetV2, a lightweight CNN architecture,
Creative Commons Attribution is used as a convolutional base with three transfer learning variants, includ-
License, which permits unrestricted
ing transfer learning without fine-tuning, transfer learning with fine-tuning,
use, distribution, and reproduction
in any medium, provided the and fixed feature extraction, to achieve binary classification. The study also
original author and source are attempts to develop a seven-layered CNN architecture to accomplish mam-
credited.
mography image classification through scratch learning. The trained breast
Published By Indian Society for cancer detection models’ classification performance, obtained through transfer
Education and Environment (iSee)
and scratch learning methods, is evaluated based on machine learning perfor-
ISSN mance metrics such as accuracy, precision, F1 Score, and area under the curve.
Print: 0974-6846
Findings: The study conducted three transfer learning variants and found that
Electronic: 0974-5645
combining MobileNetV2 with Random Forest classifier using the fixed feature
extraction approach produced the best results, with an accuracy of 0.994 and
a shorter training time of only 63.87 seconds compared to the other trans-
fer learning variants. On the other hand, a seven-layer CNN model developed
using scratch learning achieved a classification accuracy of 0.96, but it required
a longer training period of 7980 seconds. Novelty: The breast cancer detec-
tion model introduced in this study exhibited superior performance compared
to recently developed breast cancer detection models, as measured by accu-

https://www.indjst.org/ 1385
Bokade & Shah / Indian Journal of Science and Technology 2023;16(18):1385–1394

racy and Area Under the Curve (AUC). The study found that the model created
using the fixed feature extraction method showcases the effectiveness of utiliz-
ing MobilenetV2 and a Random Forest machine learning classifier in decreasing
the number of trainable network parameters and network training time. Con-
sequently, this presents an advantage for implementing the model on low-cost
embedded platforms with limited memory size.
Keywords: Breast Cancer (BC); Computer Aided Detection (CAD); Deep
Convolutional Neural Networks (DeepCNN); Transfer Learning (TL); Scratch
Learning (SL)

1 Introduction
Breast cancer (BC) is the most common cancer among the female population.
Early detection of breast cancer is crucial in saving lives since malignant tumors
develop rapidly and spread throughout the body. Screening mammography is the most
recommended imaging technique by radiologists to identify breast abnormalities (1) .
However, radiologists are constantly faced with a significant problem in accurately
diagnosing breast abnormalities due to the low occupancy of breast tissues on full
mammograms and the inadequate contrast of mammography images. As a result of the
success rate of Deep Learning (DL) algorithms in tackling computer vision challenges,
many researchers have developed Computer-Aided Detection (CAD) systems to
identify and forecast serious health issues from medical images. In recent years, many
Deep Convolutional Neural Networks (Deep-CNN) algorithms have been successfully
applied in breast cancer detection (BCD) (2,3) . Here, some of the recent BCD models
developed by Deep-CNNs are described.
Samriddha et. al. applied the ensemble transfer learning (TL) approach in BCD using
histopathology images. With an ensemble of GoogleNet, VGG11 and MobileNetV3, the
authors achieved maximum accuracy of 96.15% with 400x magnified images obtained
from Breakhis dataset (4) .
Wakili et al. and Voon et al. utilized a transfer learning (TL) approach with
Deep-CNN architecture for identifying breast cancer using histopathology images (5,6) .
Heenaye-Mamode Khan et al. employed the ResNet-50 architecture for multiclass
classification of mammography images using TL with a fine-tuning approach. Their
customized ResNet-50 achieved an 88% accuracy for classifying mammography images
into four different classes (7) . Mohamed E et al. developed a novel CNN pooling block
and integrated it with U-Net, Alexnet, GoogleNet, and ResNet18 architectures for the
segmentation and classification of thermogram dataset, achieving a global accuracy of
98.3% (8) .
Houssein et al. utilized ResNet-50 in combination with the marine predators
algorithm (MPA) to detect breast cancer from mammography images. By tuning the
hyperparameters of the MPA architecture, the model achieved an accuracy of 98.32%
with the DDSM dataset and 98.88% with the MIAS dataset (9) .
Hyo-Eun Kim et al. (10) utilized a TL approach with ResNet-34 pretrained model
architecture on a large-scale mammography dataset. From 2004 to 2016, the authors
collected 170,230 full-field digital mammography images (4 views) from five different
institutions: three in Seoul, South Korea, one in the United States, and one in the
United Kingdom. The algorithm was trained in two stages, namely patch-level training
(Stage-1) and image fine-tuning (Stage-2). Patch-level training was carried out to
extract low-high-level features, followed by fine-tuning approaches. Discriminative
learning methods were utilized in Stage I, whereas all semi-supervised learning methods
were used in Stage II. The performance of AI-based algorithms were compared with
radiologists using recall-based sensitivity, AUROC, and specificity. With three different

https://www.indjst.org/ 1386
Bokade & Shah / Indian Journal of Science and Technology 2023;16(18):1385–1394

validation sets, the overall AUROC evaluated was 0.959. The study showed that the performance of AI-based algorithms on
large-scale data was superior to that of radiologists in cancer detection.
Karin et al. (11) used InceptionResNet-v2 architecture to calculate a risk score for predicting the future occurrence of breast
cancer. They provided attributes related to mammographic images and radiation parameters to the network in the training
phase to calculate the mammographic risk score and the density score. The DNN model was able to predict which women were
susceptible to breast cancer in the future, with a focus on more aggressive tumors.
Qiyuan Hu et al. (12) developed a Computer-Aided Diagnosis (CADx) system for diagnosing breast cancer using
multiparametric magnetic resonance imaging (MRI). They collected 927 MR scans from 616 women and used a VGG-19
CNN model to extract features, followed by an SVM (Support Vector Machine) classifier to categorize the scans into benign or
malignant lesions. The performance of the DL model was compared with the DeLong test, which showed that the DL model
had increased diagnostic accuracy with an AUC of 95% in breast image interpretation.
Abhishek Das et al. (13) designed a stacked ensemble model using gene expression datasets and breast histopathology images
to identify breast cancer. The 1-D gene expression data was converted into 2-D deconstructed images using empirical wavelet
transform to perform 2-stage classification. Three CNNs served as the basic classifiers in the first stage, followed by a Multilayer
Perceptron in the 2nd stage. One of the three CNNs was trained with original images, whereas the other CNNs were trained
with deconstructed images. The proposed strategy achieved a 98.08% classification accuracy.
The transfer learning strategy proposed by Wessam M. Salama et al. (14) aimed to segment and classify mammography
images into benign and malignant categories. The authors employed a modified U-Net architecture for breast segmentation
in the first stage and utilized multiple CNNs including InceptionV3, DenseNet121, ResNet50, VGG16, and MobileNetV2 for
image classification in the second stage. The model was trained and validated on three datasets: CBIS-DDSM, MIAS, and
DDSM. In terms of performance with the DDSM dataset, the U-Net model and InceptionV3 outperformed other architectures,
achieving an accuracy of 98.87%, a sensitivity of 98.98%, AUC of 98.88%, a precision of 98.79%, and F1 score of 97.99%, with
a computational time of 1.2134 seconds.
Mohapatra et al. proposed a breast cancer detection model using VGG16, AlexNet, and ResNet-50 architectures. They applied
transfer learning with pretrained weights on ImageNet dataset. Among these architectures, AlexNet model outperformed the
others with an accuracy of 65% (15) .
Bibhu et. al. have created a computer-aided diagnosis (CAD) system for detecting breast cancer using the Wisconsin breast
cancer dataset (WBCD). The system employs a combination of PCA (principal component analysis) and NN (neural network)
algorithms. According to their simulation results, the system achieved an accuracy rate of 97%, surpassing the accuracy of other
existing state-of-the-art methods. This suggests that the integration of PCA and NN algorithms can be an effective strategy for
developing CAD systems for breast cancer detection (16) .
The BCD model proposed by Sahu et al. combined both statistical and machine learning methods, resulting in more precise
and advanced models for the prediction and diagnosis of breast cancer (17) . The authors also developed a BCD model and
evaluated various machine learning classifiers under different operational conditions and datasets. Their findings indicate that
SVM is a suitable choice for accurately identifying performance metrics, including sensitivity, accuracy, error, and specificity (18) .
In addition to implementing Deep CNN models on CPUs or GPUs, researchers have also investigated the deployment
of Deep CNN models on hardware platforms or in cloud environments. For instance, Shahirah Zahir et al. implemented a
Deep CNN-based breast cancer model on an IoT platform (19) . Similarly, Chowdhury et al. developed an Android-based cloud
environment for breast cancer identification (20) .
According to previous studies, the majority of existing Breast Cancer Diagnosis (BCD) models utilize a Deep-CNN based
transfer learning (TL) technique because developing a BCD model from scratch requires domain expertise in network layer
design and proper hyperparameter selection. Our paper aims to develop a Deep-CNN based BCD model using both TL
and scratch learning (SL) approaches. Additionally, we compare the performance of BCD models developed using TL and
SL approaches in terms of machine learning performance measures, training time, and total trainable parameters. We also
compare our obtained model with some existing BCD models to demonstrate its efficiency. In Section 2, we will describe the
methodology we used to develop the BCD model, while in Section 3, we will present and discuss the simulation results. Finally,
in Section 4, we will provide concluding remarks, future scope, and implications of our study.

2 Methodology
The proposed model for detecting breast cancer utilizes a Deep-CNN architecture and employs both TL and SL approaches. The
TL approach uses a pre-trained architecture, MobileNetV2, with three different variants of TL, namely TL without fine-tuning,
TL with fine-tuning, and fixed feature extraction approach. The suggested methodology also describes the development of a
BCD model with a 7-layered CNN from scratch. Figure 1 shows the proposed algorithmic methodology for the BCD model,

https://www.indjst.org/ 1387
Bokade & Shah / Indian Journal of Science and Technology 2023;16(18):1385–1394

along with a brief explanation of the procedures involved in developing it.

Fig 1. BCD (Breast Cancer Detection) Model Workflow

2.1 Dataset
The BCD model was developed using mammography images from the publicly available Digital Database for Screening
Mammography (DDSM) dataset (21) . A total of 5000 mammography images, consisting of 2500 images from the benign class
and 2500 images from the malignant class, were used for network training. Out of these, 20% of the images were reserved for
evaluating the trained model, while the remaining 80% were used to train the CNN architecture.

2.2 Image Preprocessing


Mammography images often suffer from poor contrast, which can limit the generalizability of a trained CNN model to new,
unseen images. Therefore, it is essential to select an appropriate image contrast enhancement technique to improve image
quality. In this study, we utilized the min-max contrast enhancement technique to enhance the contrast of all mammography
images after resizing them to 227x227x3. The image contrast enhancement process is demonstrated in Figure 2.

Fig 2. Image Contrast Enhancement Process

2.3 MobileNetV2
The MobileNetV2, which is a modified version of MobileNetV1, serves as the convolutional base in all three TL variants for
performing mammography image classification tasks. The low-powered and lightweight structure of MobileNetV2 makes it
suitable to deploy the trained model on a smart embedded platform with low memory and computation capabilities.
As shown in Figure 3, MobileNetV2 consists of three layers: the expansion layer, depth-wise convolution layer, and projection
layer. The residual bottleneck connection in the MobileNet architecture reduces the number of channels available at the

https://www.indjst.org/ 1388
Bokade & Shah / Indian Journal of Science and Technology 2023;16(18):1385–1394

projection layer. This reduction in the number of channels at the projection layer output is referred to as a ”bottleneck”. In the
standard MobileNet architecture, a 3x3 depth-wise convolutional layer extracts features from input channels, while a 1x1 point-
wise convolutional layer combines the feature maps generated by the depth-wise convolutional layer to efficiently reduce the
dimensionality of input channels. This feature makes depth-wise separable convolution filters faster than standard convolutional
filters and reduces the network training time.

Fig 3. MobileNetv2 Network schematic

2.4 Transfer Learning Approach


Transfer learning is an effective approach to leverage the features learned from one task to another related task, resulting in
improved model performance and reduced training time. This paper investigates the BCD model for mammography image
classification using three variations of the TL approach, which are graphically represented in Figure 4. The three variants of the
transfer learning approach used in this paper are as follows:

Fig 4. Transfer Learning (TL) strategies

2.4.1 Transfer Learning (Without Fine-tuning)


This variant utilizes the pre-trained MobileNetV2 architecture with a frozen convolutional base, meaning that no changes are
made to the layer parameters in the convolutional base. The last classification layer, which is a fully connected layer, is customized
to match the number of classification categories required for the task. Once the necessary changes are made to the classification
layer, the network is trained using DDSM training images.

2.4.2 Transfer Learning with fine-tuning


This approach involves fine-tuning specific layers in the convolutional base while replacing the last layers with a new fully
connected layer. The convolutional base is fine-tuned, and the fully connected (FC) layers are retrained based on the new

https://www.indjst.org/ 1389
Bokade & Shah / Indian Journal of Science and Technology 2023;16(18):1385–1394

learning task. In the present study, the last 23 layers of the convolutional base were trained with a learning rate of 0.1, and the
simulation results were obtained accordingly.

2.4.3 Fixed Feature Extraction approach


In this variant, the convolutional base of the Deep CNN model is kept frozen while only the last layer, the FC layer, is replaced
with a new classifier. In the proposed BCDM, the FC layer of the pre-trained CNN model architecture (MobileNetV2) is
substituted with a Random Forest (RF) machine learning classifier. This approach has proven effective in training the model
without the need for customizing the hyperparameters of MobileNetV2. The RF classifier is initialized with a suitable number
of Decision Trees (DTs) to estimate the accuracy based on the average of the predictions generated by all DTs.

2.5 Scratch Learning (SL) Approach


The SL approach provides developers with the flexibility to design network layers from scratch. In our study, we have created
a 7-layered CNN architecture using the SL approach. The structure of the 7-layered CNN model is illustrated in Figure 5. The
architecture consists of repeated convolutional layers, activation functions, and pooling layers. The final layer of the network is
a dense layer, followed by a dropout layer, which generates the classification score.

Fig 5. Structure of a 7-layered CNN Model

3 Results and Discussion


In this section, we present the results of the simulations conducted using three different TL approaches and the SL approach.
The model parameters used for each approach are summarized in Table 1. After training, the performance of each model is
evaluated based on machine learning performance measures such as accuracy, precision, F1-score, and AUC. The number
of trainable parameters and the network training time are also computed for each approach. Table 2 presents the results of the
four different approaches (Fixed Feature Extraction, Transfer Learning without fine-tuning, Transfer Learning with fine-tuning,
and Scratch Learning) used for mammography image classification. The total model parameters, trainable and non-trainable
parameters, training time, and performance metrics (accuracy, precision, F1 score, and AUC) are reported for each approach.
The corresponding confusion matrices and AUC for each approach are shown in Figures 6 and 7 respectively.
As demonstrated in Table 2, the FFE approach demands no trainable parameters, achieved impressive results when using
the RF classifier with MobileNetV2 as a convolutional base. The model achieved an accuracy of 0.994 and had a much shorter
network training time of 63.86 seconds compared to TL approaches. The FFE approach utilized transfer learning by using
pre-trained convolutional layers and adding a new classifier on top. This allowed the model to leverage the knowledge gained
from training on a large dataset, which is transferable to the task of mammography image classification. The addition of the
Random Forest classifier helped to improve the model’s performance by adding a decision-making layer that considers multiple
decision trees. This approach can help to reduce overfitting and improve model generalization Although TL with fine-tuning
had an accuracy of 1.0, the appropriate selection of hyperparameters is critical for achieving such results. On the other hand,
developing a scratch model with appropriate layers, filters, stride, and padding can lead to optimal performance, but this requires
domain expertise.

https://www.indjst.org/ 1390
Bokade & Shah / Indian Journal of Science and Technology 2023;16(18):1385–1394

Table 1. Model Parameters of different Deep CNN Strategies


Strategy Type Approach Changes in layer Model Parameters
Fixed Feature extraction Network weights=ImageNet The number of Decision
Transfer Learning (TL) Replacement of last FC layer Trees=50
with RF classifier
Transfer Learning (Without fine Training of Fully connected No. of classes=2
tuning) layer
Transfer learning (With fine- Training of last 23 layers Batch size=6, Epochs=25 Opti-
tuning) mizer=Adam Cross entropy=
categorical
Scratch learning (SL) 7 layered CNN 7 Blocks of Convolutional layer, Batch size=6, Epochs=25 Opti-
Maximum Pooling layer and mizer=rmsprop Cross entropy=
activation function Binary Dropout=0.25

Fig 6. Normalized Confusion Matrices obtained using different CNN strategies (0 for Benign and 1 for Malignant class)

Fig 7. ROC curves

https://www.indjst.org/ 1391
Bokade & Shah / Indian Journal of Science and Technology 2023;16(18):1385–1394

Table 2. Performance Metrics obtained with different Deep CNN Strategies


Strategy Approach Total Trainable Non- Training Accuracy Precision F1 AUC
Type Model Parame- trainable Time Score
Parame- ters Parame- (Sec)
ters ters
Transfer Fixed Feature extrac- 3228864 0 3228864 63.87 0.994 0.99 0.99 0.994
Learning tion (FFE)
(TL) Transfer Learning 2383426 125442 2257984 4743 0.69 0.75 0.67 0.75
(Without fine tuning)
Transfer learning 2383426 1485442 897984 4822 1.0 1.0 1.0 0.99
(With fine-tuning)
Scratch 7 layered CNN 805467 805467 0 7980 0.96 0.96 0.96 0.96
learning
(SL)

Table 3 provides a summary of the results obtained by different research studies that have employed deep learning-based
strategies for deep CNN based breast cancer detection tasks on different datasets.Trang et al. (22) have used transfer learning with
Xception as the pre-trained model and gradient boosting machine as the classifier for binary classification on a private dataset.
Their proposed approach achieves an accuracy of 84.5% and an AUC of 0.88. Huang et al. (23) have used transfer learning with
MobileNetV3 and bilinear structure for multiclass classification of whole slide images. They reported an accuracy of 0.88. Li
et al. (24) have used transfer learning with Resnet-v2 and a CNN for multiclass classification on the in breast dataset. Their
proposed approach achieves an accuracy of 70% and an AUC of 0.84. Finally, our paper proposed a transfer learning approach
using MobileNetV2 as the pre-trained model and RF classifier for binary classification on the DDSM dataset. Our approach
achieved an accuracy of 99.4% and an AUC of 0.99, which indicates that your approach performs exceptionally well on this
particular task.
Overall, Table 3 highlights the effectiveness of transfer learning techniques in achieving high accuracy and AUC for different
classification tasks on different datasets. However, it also suggests that the performance of these techniques can vary based on
the choice of pre-trained model, classifier, and dataset.

Table 3. Comparison of the proposed work with existing Breast Cancer Models
Year Author Strategy Network Dataset Application Accuracy AUC
Type
2023 Trang et. al. (22) TL Xception + Gradient Binary Classification 84.5 0.88
boosting machine
2022 Huang et. TL MobileNetV3+ Bilin- Whole slide Multiclass Classifica- 0.88 –
al. (23) ear Structure images tion
2020 Li et al. (24) TL Resnet-v2 + CNN INbreast Multiclass Classifica- 70 0.84
tion
Our paper TL MobileNetV2+RF DDSM Binary Classification 99.4 0.99
classifier

4 Conclusion
From the literature review and simulation results, it is clear that Deep CNNs have potential for CAD development in early-stage
breast cancer diagnosis and prediction without requiring extensive preprocessing and feature engineering. Transfer learning
makes it easier for developers to create application-specific models using pre-trained CNNs, and various transfer learning
approaches can improve the performance of DL architectures without relying heavily on large datasets. However, amending
existing model architectures and selecting appropriate hyperparameters remains challenging to create suitable models for
clinical applications. The lack of standard annotated medical datasets is an open area of research that poses a challenge to
supervised DL model optimization. Additionally, selecting suitable image pre-processing techniques is vital to enhancing model
performance by reducing the False Positive Rate (FPR) due to contrast variation and noise in mammography images. Network
hyperparameters, such as epochs, learning rate, and dropout rate, significantly influence reliable model performance. The

https://www.indjst.org/ 1392
Bokade & Shah / Indian Journal of Science and Technology 2023;16(18):1385–1394

systematic combination of supervised and unsupervised DL architectures can prevent models from overfitting and underfitting
issues. Training the network with a hybrid image dataset and clinical information may enhance network generalization ability.
The use of lightweight CNNs such as SqueezeNet, MobileNetV2, and Xception can enable developers to implement DL
models on smart platforms with low computational power, and ensembling DL architectures can enhance network capability
to perform accurate medical image analysis. Future research on breast cancer diagnosis can focus on customizing existing
model architectures and exploring cloud-based implementation for improved results and efficiency. The availability of open-
source mammography datasets and the implementation of ensemble models with simple architectures can further enhance the
potential of the BCD model.

Acknowledgement
We would like to thank Prof. Dr. M.K Shah, H.O.D. (Instrumentation & Control), VGEC-Chandkheda, Ahmedabad and Prof.
(Dr.) Yagnesh Shukla, Principal, NeoTech Technical Campus, Vadodara for their valuable guidance in this research work. The
research work is carried out on publicly available open-source data. We affirm that there is no financial assistance obtained to
conduct the research work presented in this study.

References
1) Mohapatra S, Muduly S, Mohanty S, Ravindra JVR, Mohanty SN. Evaluation of deep learning models for detecting breast cancer using histopathological
mammograms Images. Sustainable Operations and Computers. 2022;3:296–302. Available from: https://doi.org/10.1016/j.susoc.2022.06.001.
2) Mahoro E, Akhloufi MA. Applying Deep Learning for Breast Cancer Detection in Radiology. Current Oncology. 2022;29(11):8767–8793. Available from:
https://doi.org/10.3390/curroncol29110690.
3) Nasser M, Yusof UK. Deep Learning Based Methods for Breast Cancer Diagnosis: A Systematic Review and Future Direction. Diagnostics. 2023;13(1):161.
Available from: https://doi.org/10.3390/diagnostics13010161.
4) Majumdar S, Pramanik P, Sarkar R. Gamma function based ensemble of CNN models for breast cancer detection in histopathology images. Expert Systems
with Applications. 2023;213:119022. Available from: https://doi.org/10.1016/j.eswa.2022.119022.
5) Wakili MA, Shehu HA, Sharif MH, Sharif M, Umar A, Kusetogullari H, et al. Classification of Breast Cancer Histopathological Images Using DenseNet
and Transfer Learning. PMCID. 2022. Available from: https://doi.org/10.1155/2022/8904768.
6) Voon W, Hum YC, Tee YK, Yap WS, Salim MIM, Tan TS, et al. Performance analysis of seven Convolutional Neural Networks (CNNs) with transfer
learning for Invasive Ductal Carcinoma (IDC) grading in breast histopathological images. Scientific Reports. 2022;12(1):19200. Available from:
https://doi.org/10.1038/s41598-022-21848-3.
7) Khan MHM, Boodoo-Jahangeer N, Dullull W, Nathire S, Gao X, Sinha GR, et al. Multi- class classification of breast cancer abnormalities using Deep
Convolutional Neural Network (CNN). PLOS ONE. 2021;16(8):e0256500. Available from: https://doi.org/10.1371/journal.pone.0256500.
8) Mohamed E, Gaber T, Karam O, Rashed EA. A Novel CNN pooling layer for breast cancer segmentation and classification from thermograms. PMCID.
2022. Available from: https://doi.org/10.1371/journal.pone.0276523.
9) Houssein EH, Emam MM, Ali AA. An optimized deep learning architecture for breast cancer diagnosis based on improved marine predators algorithm.
Neural Computing and Applications. 2022;34(20):18015–18033. Available from: https://doi.org/10.1007/s00521-022-07445-5.
10) Kim HE, Heekim H. Changes in cancer detection and false-positive recall in mammography using artificial intelligence: a retrospective, multireader study.
2020. Available from: https://doi.org/10.1016/S2589-7500(20)30003-0.
11) Dembrower K, Liu Y, Azizpour H, Eklund M, Smith K, Lindholm P, et al. Comparison of a Deep Learning Risk Score and Standard Mammographic
Density Score for Breast Cancer Risk Prediction. Radiology. 2020;294(2):265–272. Available from: https://doi.org/10.1148/radiol.2019190872.
12) Hu Q, Whitney HM, Giger ML. A deep learning methodology for improved breast cancer diagnosis using multiparametric MRI. Scientific Reports.
2020;10(1). Available from: https://doi.org/10.1038/s41598-020-67441-4.
13) Das A, Mohanty MN, Mallick PK, Tiwari P, Muhammad K, Zhu H. Breast cancer detection using an ensemble deep learning method. Biomedical Signal
Processing and Control. 2021;70:103009. Available from: https://doi.org/10.1016/j.bspc.2021.103009.
14) Salama WM, Aly MH. Deep learning in mammography images segmentation and classification: Automated CNN approach. Alexandria Engineering
Journal. 2021;60(5):4701–4709. Available from: https://doi.org/10.1016/j.aej.2021.03.048.
15) Mohapatra S, Muduly S, Mohanty S, Ravindra JVR, Mohanty SN. Evaluation of deep learning models for detecting breast cancer using histopathological
mammograms Images. Sustainable Operations and Computers. 2022;3:296–302. Available from: https://doi.org/10.1016/j.susoc.2022.06.001.
16) Sahu B, Dash S, Mohanty SN, Rout SK. Ensemble Comparative Study for Diagnosis of Breast Cancer Datasets. International Journal of Engineering &
Technology. 2018;7(4.15):281. Available from: https://doi.org/10.14419/ijet.v7i4.15.23007.
17) Sahu B, Mohanty SN, Rout S. A Hybrid Approach for Breast Cancer Classification and Diagnosis. ICST Transactions on Scalable Information Systems.
2019;0(0):156086. Available from: http://dx.doi.org/10.4108/eai.19-12-2018.156086.
18) Sahu B, Panigrahi A. Efficient Role of Machine Learning Classifiers in the Prediction and Detection of Breast Cancer. SSRN Electronic Journal. 2020.
Available from: http://dx.doi.org/10.2139/ssrn.3545096.
19) Shahirahzahir A, Amir, Nik A, Zahri W, Ang C. Applying the Deep Learning Model on an IoT Board for Breast Cancer Detection based on
Histopathological Images. 5th International Conference on Electronic Design (ICED) 2020 Journal of Physics: Conference Series 1755. 2021;p. 12026. Available
from: https://doi.org/10.1088/1742-6596/1755/1/012026.
20) Chowdhury D, Das A, Dey A, Sarkar S, Dwivedi AD, Mukkamala RR, et al. ABCanDroid: A Cloud Integrated Android App for Noninvasive Early Breast
Cancer Detection Using Transfer Learning. Sensors. 2022;22(3):832. Available from: https://doi.org/10.3390/s22030832.
21) Lin TY, Huang ML. Dataset of Breast mammography images with Masses. 2020. Available from: https://doi.org/10.17632/ywsbh3ndr8.2.
22) Trang NTH, Long KQ, An PL, Dang TN. Development of an Artificial Intelligence-Based Breast Cancer Detection Model by Combining Mammograms
and Medical Health Records. Diagnostics. 2023;13(3):346. Available from: https://doi.org/10.3390/diagnostics13030346.

https://www.indjst.org/ 1393
Bokade & Shah / Indian Journal of Science and Technology 2023;16(18):1385–1394

23) Huang J, Mei L, Long M, Liu Y, Sun W, Li X, et al. BM-Net: CNN-Based MobileNet-V3 and Bilinear Structure for Breast Cancer Detection in Whole Slide
Images. Bioengineering. 2022;9(6):261. Available from: https://doi.org/10.3390/bioengineering9060261.
24) Li C, Xu J, Liu Q, Zhou Y, Mou L, Pu Z, et al. Multi-View Mammographic Density Classification by Dilated and Attention-Guided Residual Learning.
IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2021;18(3):1003–1013. Available from: https://doi.org/10.1109/TCBB.2020.
2970713.

https://www.indjst.org/ 1394

You might also like