A Novel Approach For Breast Cancer Detection Using Optimized Ensemble
A Novel Approach For Breast Cancer Detection Using Optimized Ensemble
A R T I C L E I N F O A B S T R A C T
Keywords: Breast cancer (BC) is a common and highly lethal ailment. It stands as the second leading contributor to cancer-
Breast cancer detection related deaths in women worldwide. The timely identification of this condition is of utmost importance in
Image processing mitigating mortality rates. This research paper presents a novel framework for the precise identification of BC,
Healthcare
utilising a combination of image and numerical data features with explainable Artificial Intelligence (XAI). The
Transfer learning
Ensemble learning
utilisation of the U-NET transfer learning model is employed for image-based prediction. Additionally, an
Deep convoluted features ensemble model is constructed by integrating characteristics from a customised convolutional neural network
(CNN) model with an ensemble comprising random forest (RF) and support vector machine (SVM). The exper
iments aim to evaluate the influence of original features compared to convoluted features. A comparative
analysis is carried out to assess the efficacy of various classifiers in accurately detecting BC, utilising the Wis
consin dataset. The model under consideration exhibits promising capabilities in enhancing BC diagnosis, with a
remarkable accuracy rate of 99.99%. The present study contributes to the advancement of BC diagnosis by
introducing a novel strategy based on machine learning and discussing the interpretation of the variables using
XAI. The primary objective of this approach is to get a notable level of precision, hence facilitating the early and
reliable identification of BC. Ultimately, the implementation of this approach is expected to enhance patient
outcomes.
1. Introduction critical in reducing the death rate associated with this condition. Ac
cording to Globocan 2018 data [5], breast cancer is found in one out of
Cancer is a worldwide flagellum, leading to millions of deaths per every four cases of cancer in women, making it the sixth most prevalent
year [1]. As per GLOBOCAN 2020 estimates of cancer incidence and cause of death globally. Breast cancer often develops in the breast tissue,
mortality produced by the International Agency for Research on Cancer especially the inner linings of milk ducts or lobules. Changes or muta
[2], approximately 2,261,419 new cases of BC were diagnosed world tions in Deoxyribonucleic acid (DNA) or Ribonucleic acid (RNA) cause
wide in 2020; there were 24.5% of cancer cases and 15.5% of cancer cancer cell formation. Air pollutants, bacteria, nuclear radiation, fungi,
deaths worldwide among women [3]. Saudi Health Council data in physical cell-level damage, viruses, parasites, high temperatures, water
dicates that BC accounted for 15.9% of all reported cancers among Saudi contaminants, electromagnetic radiation, dietary elements, free radi
citizens and 28.7% among women of all ages. According to the World cals, aging of DNA and RNA, and genetic changes are all factors that can
Health Organization (WHO), breast cancer affects approximately 2.3 lead to breast cancer. Most breast cancers are no special type (NST), also
million women each year, resulting in 685,000 deaths, accounting for called invasive ductal carcinomas (IDCs), followed by invasive lobular
13.6% of all female cancer-related fatalities [4]. Prompt detection is carcinomas (ILCs) [6], inflammatory breast cancer (IBC) [7], and
* Corresponding author.
E-mail addresses: [email protected] (R.M. Munshi), [email protected] (L. Cascone), [email protected] (N. Alturki), [email protected] (O. Saidani),
[email protected] (A. Alshardan), [email protected] (M. Umer).
https://doi.org/10.1016/j.imavis.2024.104910
Received 24 September 2023; Received in revised form 23 November 2023; Accepted 11 January 2024
Available online 20 January 2024
0262-8856/© 2024 Elsevier B.V. All rights reserved.
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
lobular breast cancer (LBC) [8]. The subsequent sections of this study are outlined as follows: Section
Breast cancer, a dangerous disease with high death rates, accounts 2 encompasses recent literature on breast cancer diagnosis and detec
for 2.5% of all fatalities, impacting one out of every thirty-nine women tion. Section 3 elaborates on the dataset, the proposed methodology, and
[9]. Early identification and treatment of breast cancer are critical since the machine learning classifiers employed. Section 4 comprises results
the disease can spread to other parts of the body if left untreated. Early and a comparative analysis. Discussions are presented in Section 5.
detection and treatment can increase the chance of survival by up to Lastly, Section 6 encapsulates the conclusion and outlines avenues for
80%. This emphasizes the critical need of early identification and future research.
treatment of breast cancer. Various procedures and methodologies, such
as screening examinations, self-assessments, and frequent visits with 2. Related work
healthcare specialists, help to detect breast cancer early [10].
Mammography is a popular and efficient approach for detecting breast Many approaches for BC detection and diagnosis have emerged last
cancer early. Several studies have confirmed its efficacy in detecting years. Recent research has primarily focused on identifying malig
early-stage breast cancer. Another often-used approach for identifying nancies through CAD and decision support systems. While the majority
breast cancer is a biopsy. Tissue is removed from the afflicted breast of investigations depend on individual models to attain dependable re
region and examined under a microscope to detect and categorize the sults, a handful explore the utilization of ensemble approaches. This
tumor during a biopsy [11]. Biopsy is an often-utilized approach for section presents an overview of these approaches.
identifying BC. During a biopsy, tissue is removed from the afflicted To achieve precise and accurate breast cancer diagnosis, Yadav and
region of the breast and examined under a microscope in order to di Jadhav [17] proposed a machine learning-driven system employing
agnose and categorise the tumour. Although present approaches help to thermal infrared imaging. The researchers evaluated various baseline
reduce breast cancer death rates, there is still room for improvement, models alongside transfer learning models such as VGG16 and Incep
particularly in terms of more simplified and automated diagnostic pro tionV3. Their experiments encompassed both augmented and non-
cesses. Data mining is a technique used to extract important and note augmented data scenarios. The findings reveal that the transfer
worthy insights from large data collections. It is recognized as a critical learning model InceptionV3 outperforms other models, achieving an
tool for the early detection of several ailments such as renal disease, accuracy rate of 93.1% without augmentation and 98.5% with
heart disease [12], diabetes [13], and cancer. Using data mining tech augmentation. In a separate study [18], researchers leveraged genetic
niques makes it easier to recognize patterns and trends in data, which programming to optimize features for automated breast cancer diag
aids in the early diagnosis and treatment of certain disorders. Its nosis. They experimented with nine machine learning classifiers. The
importance grows when it comes to recognizing illnesses like cancer, outcomes highlight genetic programming’s effectiveness in amalgam
when early identification improves survival chances dramatically. ating preprocessing and feature selection, leading to the identification of
Traditional cancer detection consists on three techniques: physical ex the best model. The AB classifier attained the highest accuracy score of
amination, pathology testing, and radiological imaging. These old ap 98.23%. In another research work, researchers proposed an approach
proaches, however, are time-consuming and prone to false negatives. for BC detection using extreme learning machine based on feature fusion
Machine learning approaches, on the other hand, are gaining popularity with CNN deep features. [19]
due to their superior results. They provide dependability, precision, and The authors devised an automated breast cancer detection system
speed, revolutionizing disease detection beyond traditional methods and employing various models like LR, KNN, SVM, and diverse CNN varia
giving more consistent findings across a wide range of ailments. tions [20]. During experimentation, the authors scrutinized zones of
Over the past decade, recent advances in machine learning have hostile ductal carcinoma tissue within entire slide images. Results
opened up new areas for the medical fields. Machine learning algorithms indicated that the CNN variant attained the highest accuracy at 87%,
require the capacity to comprehend predictions, which is referred to as surpassing the accuracy of machine learning models by 9%. This sug
explainable artificial intelligence (XAI). Justification of specific out gests that the proposed deep learning system significantly enhances
comes yields valuable, understandable insights, facilitating informed breast cancer detection accuracy. The researchers introduced a voting
decision-making and knowledge generation. Researcher applied XAI in classifier based on ensemble learning for breast cancer detection [21].
detecting breast cancer [14,15]. 1n a recent study authors applied ma Their findings demonstrated that the ensemble learning model out
chine learning with XAI for deting breast cancer in Indonesian patients performed individual machine learning models. A study proposed a
[16]. Due to the aforementioned benefits, this study proposes a machine breast tumor type detection approach and achieved 98.1% accuracy
learning-based approach leveraging XAI for detecting breast cancer with [22]. Another study utilized diverse-density mammograms in breast
the goal of achieving high accuracy. This study makes the following cancer detection, achieving an overall accuracy score of 88.1% [23].
contributions in this regard. Diverse approaches and techniques, including Breast Magnetic
Resonance Imaging (MRI), Mammography, breast ultrasound, biopsy,
• This study presents a complete framework for accurate breast cancer breast self-exam, and clinical breast exam, help in the timely detection
diagnosis that makes use of both image analysis and numerical data of cancer. Breast mammography and ultrasound are typical screening
characteristics. For image-based prognosis, the predictive model methods. [24]. MRI is often used in conjunction with mammography or
employs U-NET transfer learning. as a supplemental method for women with dense breasts or high-risk BC
• An ensemble strategy is proposed in this study to anticipate breast cancer. Ultrasound has played an important role in BC detection for
cancer among patients by merging features from a personalized many years thanks to its portability, availability, and their ability to
convolutional neural network (CNN) model. The proposed ensemble assess a variety of important aspects from multiple planes with high
model combines random forest (RF) and support vector machine resolution. These approaches focused on combining machine learning
(SVM) techniques, with the final prediction determined by a voting methods with various feature extraction techniques to enhance accuracy
process. in breast mass classification. Results showed high accuracy rates, with
• The experiments involve utilizing both the inherent features from the DLA-EABA achieving 97.2% accuracy. Aljuaid et al. proposed a transfer
dataset and those derived from the CNN model. The study aims to learning-based approach in [25] for breast cancer classification. The
assess the influence of the original features in comparison to the authors conducted the experiment in two ways: binary classification and
performance of models utilizing derived features. multi-class classification.
• Moreover, the efficacy of the proposed model is assessed by con Mangukiya et al. [26] conducted a study focused on efficient and
trasting its performance with state-of-the-art methodologies con accurate early breast cancer diagnosis employing multiple techniques.
cerning accuracy, precision, recall, and the F1 score. The authors employed a range of machine learning algorithms on a
2
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
dataset containing features with diverse units and magnitudes, requiring contains 32 features including ‘Texture SE’, ‘Texture Mean’, ‘Concave
standardization through standard scaling. Their findings highlighted Points Mean’, ‘Concave Points SE’, ‘ID’, ‘Area Worst’, ‘Smoothness
that, with standard scaling, the XGBoost machine learning algorithm Mean’, ‘Symmetry Worst’, ‘Compactness SE’, ‘Radius Mean’, ‘Texture
achieved an accuracy score of 98.24%. Similarly, [27] introduced a deep Worst’, ‘Concave Points Worst’, ‘Perimeter SE’, ‘Fractal Dimension SE’,
ensemble learning model for breast cancer detection using whole slide ‘Area Mean’, ‘Perimeter Worst’, ‘Fractal Dimension Mean’, ‘Compact
images. They utilized various deep-learning models and proposed an ness Worst’, ‘Compactness Mean’, ‘Radius Worst’, ‘Perimeter Mean’,
ensemble model. The outcomes demonstrated the superiority of the ‘Concavity SE’, ‘Smoothness SE’, ‘Fractal Dimension Worst’, ‘Concavity
ensemble model over other models, achieving an accuracy score of Mean’, ‘Smoothness Worst’, ‘Symmetry Mean’, ‘Symmetry SE’, ‘Area
86.21%. XAI has been used in many cancer detection tasks like ovarian SE’, ‘Radius SE’, ‘Concavity Worst’, ‘Diagnosis’ (target class). Within the
cancer analysis [28], cancer biomarkers evaluation [29], and cervical dataset, two distinct target classes exist: benign and malignant. Sample
cancer cell migration [30]. This research work is a continuation of our distribution indicates a composition of 45% in the malignant class and
previously published paper [31]. Muhammad Umer the co-author of 55% in the benign class. The dataset’s 32 features encompass various
both works helped us in implementing feature based analysis and pre classifications like numeric, nominal, and binary, among others.
diction of breast cancer in this research work and further extension of Notably, the target class represents a categorical attribute, while the rest
breast cancer to image based analysis too. of the attributes are numeric.
Numerous investigations have delved into employing machine
learning models to detect breast cancer, seeking to enhance classifica
tion accuracy and minimize pathological errors in automated diagnosis. 3.2. Machine learning models for breast cancer prediction
Table 1 provides an overview of select literature pertaining to breast
cancer detection through machine learning models. Machine learning classification involves supervised learning,
wherein the system assimilates knowledge from a particular dataset to
3. Materials and methods categorize new observations. The dataset can comprise binary or multi-
class classifications. This section addresses machine learning classifiers
This section covers the dataset utilized for breast cancer detection, employed in breast cancer detection. The implementation of these
outlines the proposed approach, and details the methodology steps. models is executed using the scikit-learn library within the Python
Additionally, it provides a concise overview of the machine learning environment.
classifiers incorporated in this research.
3.2.1. Random Forest
Random Forest (RF) stands as a prevalent ensemble learning tech
3.1. Dataset for experiments
nique in machine learning, extensively applied for classification and
regression tasks [33,34]. It operates by amalgamating multiple decision
This research applies supervised machine learning models to detect
trees, wherein each tree’s outputs contribute to the ultimate prediction.
breast cancer, aiming to assess their performance. The study proceeds
The core principle involves training numerous decision trees, each on a
through several stages, starting with the acquisition of the dataset [32].
distinct subset of the data, and subsequently amalgamating their pre
Specifically, the “Breast Cancer Wisconsin Dataset” is sourced from the
dictions to derive the final output.
UCI machine learning repository, accessible to the public. The dataset
3
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
Deep learning is an increasingly explored subject of artificial intel ot = σ (xt U o + ht− 1 W o + bo ) (2)
ligence that has shown tremendous potential in data modeling. The
( )
incorporation of automated techniques by medical specialists has shown ft = σ xt U f + ht− 1 W f + bf (3)
to be quite beneficial and helpful in illness diagnosis. Deep learning is a
popular approach for dealing with large amounts of data, eliminating where xt is the input sequence, ht− 1 is the preceding hidden state at
the need for manual feature extraction and finding significant use in current step t, it is the input gate, ot is the output gate and ft is the forget
4
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
Fig. 1. Architecture diagram of the CNN with voting classifier (RF + SVM) model.
5
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
Fig. 2. Workflow diagram of the proposed voting classifier (RF + SVM) model.
VC(RF + SVM) = argmax(g(x) ) (9) expansive datasets to effectively address new tasks. When applied to
brain tumor classification via image data, two frequently employed
To assess the performance of the proposed VC(RF + SVM) model, it
transfer learning models are U-Net and MobileNet.
undergoes testing on the’Breast Cancer Wisconsin Dataset’ in a two-
stage evaluation. Initially, breast cancer detection is executed using all
3.7.1. U-Net
32 features available in the dataset. In the subsequent experimental
U-Net, brought forth by Olaf Ronnenberg and associates in 2015,
stage, the dataset is preprocessed to employ convolutional features for
stands as a significant model in the realm of image segmentation.
machine learning models. The data is partitioned into two segments,
Originally tailored for biomedical image analysis, U-Net swiftly
with 70% allocated for training purposes and the remaining 30%
garnered acclaim due to its impressive accuracy and efficacy [45]. Its
reserved for testing. This strategy, recognized as the training–testing
notable strength lies in achieving remarkable results, even when con
split, serves as a common methodology in machine learning, facilitating
fronted with a scarcity of training data, a common challenge in medical
the evaluation of the model’s accuracy on new and unseen data.
image segmentation. The model’s architecture embodies two primary
pathways, resembling an auto-encoder. The initial pathway, termed the
3.7. Transfer learning models contracting or compressive path, functions as the encoder, utilizing a
standard deep CNN network. Conversely, the decoder or expanding path
Transfer learning stands as a widely embraced method in machine (also known as the up-sampling or synthesizing path in some literature)
learning and computer vision, utilizing pre-trained models from integrates deconvolutional and convolutional layers. The contracting
6
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
path reduces the resolution of input images, while the expanding path indicating the minimum and maximum values, respectively.
skillfully restores the original image quality and spatial structure via
TP + TN
optimized techniques, such as concatenated skip connections. Accuracy = (12)
TP + TN + FP + FN
As the network progresses, it assimilates spatial classification infor
mation, delivering dense predictions at higher resolutions along the TP
ascending path. It also enhances the output image’s resolution, leading Precision = (13)
TP + FP
to final processing through a convolutional layer, yielding a segmented
image mirroring the dimensions of the input image. This process Recall =
TP
(14)
maintains the constant shape of the input image throughout segmenta TP + FN
tion, ensuring precise representation. While classification remains
Precision × Recall
essential in medical image analysis, it lacks pixel-level contextual rep F1score = 2 × (15)
resentation, providing a single label for the entire image. U-Net, in Precision + Recall
conjunction with subsequent refinement techniques, was explicitly
4. Results
engineered to proficiently handle multi-dimensional tensors, accom
modating inputs with three or four dimensions. The network produces
A series of experiments were undertaken to identify breast cancer,
an output mirroring the input’s shape. Since its inception, U-Net has laid
employing machine learning models with both the original features and
a robust foundation for extensive exploration in medical image seg
the convoluted features. The outcomes were analyzed, and further ex
mentation. Numerous advancements have surfaced, either through ad
periments were conducted using an ensemble comprising the top four
justments to the original U-Net structure or its integration with
individual machine learning models, applied to both sets of features.
alternative architectures.
ce = Fs ⋅Fs ⋅w⋅αFm + w⋅ρ⋅αFm ⋅αFm (10) ChatGPT Initially, individual models are employed to analyze both
In the MobileNet framework, the multiplier variable w assumes the original features and convoluted features, with the outcomes pre
values contextual to the study. When focusing on brain tumor classifi sented in Table 3 and Table 4. Among the nine machine learning models
cation experiments, this multiplier typically ranges between 1 and n. considered, it is observed that the RF, ETC, LR, and SVM models
Meanwhile, the resolution multiplier r remains constant at 1. The consistently yield the most favorable outcomes across both feature sets.
equation following calculates the computational workload, represented In this phase of the study, ensembles comprising these machine learning
as cost e .
Table 2
coste = Fs ⋅Fs ⋅w⋅ρ⋅Fm ⋅Fm (11) Experimental setup for the proposed system.
Libraries Scikitlearn, TensorFlow
3.8. Evaluation metrics Language Python 3.8
RAM 8GB
The evaluation of the machine learning models in this study relies on OS 64-bit window 10
metrics like accuracy, precision, recall, and F1 score, computed from the CPU Core i7, 7th Gen with 2.8 GHz processor
GPU Nvidia, 1060, 8GB
values in the confusion matrix. These matrices range between 0 and 1,
7
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
Table 3 Table 6
Accuracy of models with original features. Ensemble model results using convoluted Features set.
Model Accuracy with original features Model Accuracy Precision Recall F1 score
models are evaluated using the original features. The findings from the
5. Discussion
ensemble learning models indicate that the RF + SVM ensemble model
exhibits superior performance in terms of accuracy, achieving a rate of
5.1. Deep learning model performance with original features
95%. This accuracy rate is almost 2% greater than that of the other
ensemble learning models. The subsequent model, SVM + ETC, attains a
The application of deep learning models involves utilizing both
classification accuracy of 92%. The RF + SVM model demonstrates a
original and convoluted features, and the results from these models are
precision of 95%, recall of 98%, and an F1 score of 96% in the context of
outlined in Table 8. Among the three deep learning models, it is evident
breast cancer diagnosis. Table 5 displays the outcomes of the ensemble
that CNN exhibits superior performance on both sets of features. In this
learning models on the original feature set.
phase of the experiments, the proposed model’s validity is assessed by
comparing it to state-of-the-art deep learning models. However, the
4.4. Performance of ensemble model on convoluted features ensemble model outperforms the deep learning models in terms of ac
curacy. The accuracy of MLP decreases when using CNN features, while
The performance of the ensemble models is also evaluated using the LSTM accuracy improves. This improvement can be attributed to LSTM’s
features extracted by the customized CNN model. The results of these capacity to extract more meaningful features, facilitating sequence
experiments are presented in Table 6. The findings indicate that the RF generation. CNN’s accuracy remains nearly the same because it receives
+ SVM model outperforms alternative models, achieving an accuracy of the same convoluted features and an additional layer for predictions.
99% as well as precision, recall, and F1 scores of 99% each. The ETC +
LR model has demonstrated the least favorable outcomes, achieving an 5.1.1. Results comparison of transfer learning models
accuracy rate of 94%. The utilization of features from the CNN model in This section will undertake an analysis of the outcomes derived from
conjunction with ensemble learning models yields superior outcomes two transfer learning models, namely MobileNet and UNET, in relation
compared to the utilization of the original features alone. to the utilization of image data. Table 9 demonstrates the complete
classification report of the two models and clearly states that the UNET
4.5. Results of K-fold cross-validation model exhibits superior accuracy, precision, recall, and F1 score
compared to the MobileNet model.
K-fold cross-validation is implemented to evaluate the effectiveness
of the proposed model. The objective of cross-validation is to validate 5.1.2. Comparative evaluation of image and feature-based datasets
the results of the presented model and assess its robustness. This tech In order to assess the efficacy of the U-Net transfer learning model, a
nique is employed to evaluate the model’s performance across various comparative analysis was undertaken to analyze the performance of
subsets of the data. In this study, a 5-fold cross-validation approach is both the machine learning model and transfer learning models using
utilized, and the corresponding results are detailed in Table 7. The cross- numerical and image data. The findings indicate a significant
enhancement in the efficacy of the transfer learning models. The outputs
Table 5
Ensemble model results using Original Features set. Table 8
Model Accuracy Precision Recall F1 score Accuracy of deep learning models with original and convoluted features.
RF + SVM [31] 95.89% 95.91% 98.54% 96.99% Model Accuracy
RF + ETC [31] 93.34% 93.45% 95.11% 94.37%
Original features Convoluted features
RF + LR [31] 89.55% 90.65% 88.25% 89.17%
ETC + SVM [31] 94.14% 93.78% 95.64% 94.24% MLP 87.69% 84.41%
ETC + LR [31] 90.34% 91.45% 91.67% 91.55% CNN 90.22% 90.70%
SVM + LR [31] 91.73% 92.64% 96.98% 95.74% LSTM 85.95% 88.34%
8
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
Table 9 Table 11
Results comparison of transfer learning models. Performance comparison with state-of-the-art studies.
Model Accuracy Precision Recall F1 score Ref. Technique Accuracy
‘Breast Cancer Survival’, which contains 330 patient records with the In this scenario, we denote the original learning model as (a), while
feature Patient_ID, Age, Gender, and expression levels of four proteins the more straightforward explanatory model is referred to as g(a’ ). Here,
(Protein1, Protein2, Protein3, Protein4). The dataset also includes the a’j , where j represents a simplified seismic attribute number, represents
Breast cancer stage of the patient (Tumor_Stage), Histology (type of
these attributes. To calculate SHAP values, labeled as “j,” we consider all
cancer), ER, PR, and HER2 status, Surgery_type, Date of Surgery, Date of
potential input orderings. In the estimation process, a’j indicates the
Last Visit, and Patient Status (Alive/Dead). The second dataset [57]
presence or absence of a specific seismic attribute, using an input vector.
contains 10 Quantitative features to show the presence or absence of
Finally, ϕ0 represents the model prediction when none of the attributes
breast cancer in a patient. The features are Age (years), BMI (kg/m2),
are taken into account during the estimation. The comprehensive
Glucose (mg/dL), Insulin (μU/mL), HOMA, Leptin (ng/mL), Adiponectin
feature importance, calculated using SHAPly, is presented in descending
(μg/mL), Resistin (ng/mL), MCP-1(pg/dL), and Labels (absence or
order in Table 12.
presence). Fig. 4 presents the accuracy comparison of the classifiers used
The SHAP analysis underscores the significance of features in the
in experiments. The proposed model obtained 97.34% accuracy on the
prediction of breast cancer. While SHAP feature importance surpasses
first dataset and 96.67% accuracy on the dataset which greatly shows
traditional methods, relying solely on it provides somewhat constrained
the stability of the proposed model on all kinds of datasets. Important
additional insights. SHAPly XAI provides a more comprehensive and
terms are explained in Table 13.
information-laden portrayal of SHAP values, unveiling not just the
relative importance of features but also their intricate connections with
5.4. Shapley additive explanations the predicted outcome. SHAP explanations furnish valuable insights into
how features contribute to a specific instance. The cumulative sum of
Deciphering the connections between inputs and outputs in learning feature contributions, in conjunction with the bias term, equals the
models is often intricate due to the perceived complexity of these models model’s initial prediction, representing the forecast before applying the
as mysterious, black-box algorithms. This lack of transparency poses a inverse link function.
significant challenge in grasping the significance of features, especially
5.5. Limitations
Table 10
Comparison analysis of image and feature-based dataset techniques. The present study contributes significantly by proposing a machine
Features Based Data Image Data learning-based approach leveraging XAI for breast cancer detection,
aiming for high accuracy. However, certain limitations are evident:
Best ML Performing Models’ Results Best TL Models’ Results
Models Original Features CNN Features Models Accuracy • The study’s effectiveness might be influenced by the size of the
VC (RF + SVM) 95.8% 99.9% U-Net 99.99% dataset used for training and testing the models. A larger and more
VC (ETC + SVM) 94.1% 97.7% MobileNet 98.5%
9
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
Table 12 Table 13
Feature importance calculated using SHAPly. Acronym Table.
Weight Feature Abbreviation Description
10
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
utilisation of the U-NET transfer learning model is employed for image- References
based prediction. Additionally, an ensemble model is constructed by
integrating characteristics from a customised convolutional neural [1] WHO. World Health Organization, Cancer: Key Facts. 2022. https://www.who.
int/news-room/fact-sheets/detail/cancer, July 2022. Online; accessed 10 January
network (CNN) model with an ensemble comprising random forest (RF) 2023.
and support vector machine (SVM). The experiments aim to evaluate the [2] Yi-Sheng Sun, Zhao Zhao, Zhang-Nv Yang, Xu Fang, Lu Hang-Jing, Zhi-Yong Zhu,
influence of original features compared to convoluted features. A Wen Shi, Jianmin Jiang, Ping-Ping Yao, Han-Ping Zhu, Risk factors and
preventions of breast cancer, Int. J. Biol. Sci. 13 (11) (2017) 1387.
comparative analysis is carried out to assess the efficacy of various [3] WHO. World Health Organization, Breast Cancer. https://www.who.int/ne
classifiers in accurately detecting BC, utilising the Wisconsin dataset. ws-room/fact-sheets/detail/breastcancer, July 2022. Online; accessed 10 January
The model under consideration exhibits promising capabilities in 2023.
[4] WHO. World Health Organization, Breast Cancer. https://www.who.int/new
enhancing BC diagnosis, with a remarkable accuracy rate of 99.99\%. s-room/fact-sheets/detail/breast-cancer, March 2021. Online; accessed 10 January
The present study contributes to the advancement of BC diagnosis by 2023.
introducing a novel strategy based on machine learning and discussing [5] Freddie Bray, Jacques Ferlay, Isabelle Soerjomataram, Rebecca L. Siegel, Lindsey
A. Torre, Ahmedin Jemal, et al., Globocan estimates of incidence and mortality
the interpretation of the variables using XAI. The primary objective of
worldwide for 36 cancers in 185 countries, CA Cancer J. Clin. 68 (6) (2018)
this approach is to get a notable level of precision, hence facilitating the 394–424.
early and reliable identification of BC. Ultimately, the implementation [6] Amrita Ray Chaudhury, Ranjani Iyer, Kaveri K. Iychettira, A. Sreedevi, Diagnosis of
of this approach is expected to enhance patient outcomes.This study invasive ductal carcinoma using image processing techniques, in: In 2011
International Conference on Image Information Processing, IEEE, 2011, pp. 1–6.
aims to establish a comprehensive framework for effectively identifying [7] Fredika M. Robertson, Melissa Bondy, Wei Yang, Hideko Yamauchi,
BC using explainable Artificial Intelligence (XAI) coupled with image Shannon Wiggins, Samira Kamrudin, Savitri Krishnamurthy, Huong Le-Petross,
and numerical data features, thereby mitigating the associated mortality Luc Bidaut, Audrey N. Player, et al., Inflammatory breast cancer: the disease, the
biology, the treatment, CA Cancer J. Clin. 60 (6) (2010) 351–375.
risk linked to this prevalent reason of death among women. We [8] Serena Masciari, Nina Larsson, Janine Senz, Niki Boyd, Pardeep Kaurah, Michaela
employed U-NET transfer learning model for image-based prediction. In J. Kandel, Lyndsay N. Harris, Hugo C. Pinheiro, Armelle Troussard,
addition, We proposed an ensemble model, leveraging its reported su Penelope Miron, et al., Germline e-cadherin mutations in familial lobular breast
cancer, J. Med. Genet. 44 (11) (2007) 726–731.
perior performance in the existing literature. Furthermore, our training [9] American Cancer Society. American Cancer Society, Breast Cancer. https://www.
process employs features derived from the CNN model. The accuracy of cancer.org/content/dam/cancer-org/research/cancer-facts-and-statistics/breast-ca
the suggested model in distinguishing patients from normal ones is an ncer-facts-and-figures/breast-cancer-facts-and-figures-2015-2016.pdf, July 2022.
Online; accessed 10 January 2023.
impressive 0.999. Moreover, the use of CNN-based features in [10] A. Pasumpon, Pandian., Identification and classification of cancer cells using
conjunction with models consistently leads to more optimal outcomes. capsule network with pathological images, J. Artif. Intell. 1 (01) (2019) 37–44.
We demonstrate the efficiency and durability of the provided model by [11] Andrei Chekkoury, Parmeshwar Khurd, Jie Ni, Claus Bahlmann, Ali Kamen,
Amar Patel, Leo Grady, Maneesh Singh, Martin Groher, Nassir Navab, et al.,
doing K-fold cross-validation and comparing its performance to estab
Automated malignancy detection in breast histopathological images, in: Medical
lished state-of-the-art models. In future studies, we plan to apply our Imaging 2012: Computer-Aided Diagnosis 8315, SPIE, 2012, pp. 332–344.
model to a variety of datasets spanning different domains, including BC [12] Furqan Rustam, Abid Ishaq, Kashif Munir, Mubarak Almutairi, Naila Aslam,
pictures, and analyse the corresponding numerical values produced Imran Ashraf, Incorporating cnn features for optimizing performance of ensemble
classifier for cardiovascular disease prediction, Diagnostics 12 (6) (2022) 1474.
from the microscopic characteristics in these images. [13] Vaibhav Rupapara, Furqan Rustam, Abid Ishaq, Ernesto Lee, Imran Ashraf, Chi-
square and pca based feature selection for diabetes detection with ensemble
CRediT authorship contribution statement classifier, Intell. Automat. Soft Comput. 36 (2) (2023).
[14] Muhammad Idrees, Ayesha Sohail, Explainable machine learning of the breast
cancer staging for designing smart biomarker sensors, Sensors Int. 3 (2022)
Raafat M. Munshi: Data curation, Resources, Writing – original 100202.
draft. Lucia Cascone: Project administration, Writing – review & edit [15] M. Rodriguez-Sampaio, Mariano Rincón, Sonia Valladares-Rodríguez, M. Bachiller-
Mayoral, Explainable artificial intelligence to detect breast cancer: A qualitative
ing. Nazik Alturki: Conceptualization, Methodology, Funding, Project case-based visual interpretability approach, in: International Work-Conference on
administration, Writing – review & editing. Oumaima Saidani: the Interplay between Natural and Artificial Computation, Springer, 2022,
Conceptualization, Methodology, Writing – original draft. Amal pp. 557–566.
[16] Fabián Silva-Aravena, Hugo Núñez Delafuente, Jimmy H. Gutiérrez-Bahamondes,
Alshardan: Conceptualization, Methodology, Software, Visualization. Jenny Morales, A hybrid algorithm of ml and xai to prevent breast cancer: a
Muhammad Umer: Methodology, Software, Writing – original draft. strategy to support decision making, Cancers 15 (9) (2023) 2443.
[17] Samir S. Yadav, Shivajirao M. Jadhav, Thermal infrared imaging based breast
cancer diagnosis using machine learning techniques, Multimed. Tools Appl. (2022)
Declaration of competing interest
1–19.
[18] Habib Dhahri, Eslam Al Maghayreh, Awais Mahmood, Wail Elkilani, Mohammed
The authors declare no conflict of interest. Faisal Nagi, et al., Automated breast cancer diagnosis based on machine learning
algorithms, J. Healthcare Eng. 2019 (2019).
[19] Z. Wang, M. Li, H. Wang, H. Jiang, Y. Yao, H. Zhang, J. Xin, Breast cancer detection
Data availability using extreme learning machine based on feature fusion with CNN deep features,
IEEE Access 7 (2019) 105146–105158.
The datasets generated during and/or analyzed during the current [20] Saad Awadh Alanazi, M.M. Kamruzzaman, Md Nazirul Islam Sarker,
Madallah Alruwaili, Yousef Alhwaiti, Nasser Alshammari, Muhammad
study can be downloaded from the following link. https://github. Hameed Siddiqi, Boosting breast cancer detection using convolutional neural
com/MUmerSabir/MDPIDiagnostic network, J. Healthcare Eng. 2021 (2021).
[21] Muhammad Umer, Mahum Naveed, Fadwa Alrowais, Abid Ishaq, Abdullah Al
Hejaili, Shtwai Alsubai, Ala’Abdulmajid Eshmawi, Abdullah Mohamed, and Imran
Acknowledgments Ashraf., Breast cancer detection using convoluted features and ensemble machine
learning algorithm, Cancers 14 (23) (2022) 6015.
This research work was funded by Institutional Fund Projects under [22] Muhammet Fatih Ak, A comparative analysis of breast cancer detection and
diagnosis using data visualization and machine learning applications, Healthcare 8
grant no. (IFPIP: 1328-415-1443) and by Princess Nourah bint Abdul (2) (2020) 111.
rahman University Researchers Supporting Project number [23] Yong Joon Suh, Jaewon Jung, Bum-Joo Cho, Automated breast cancer detection in
(PNURSP2024R333), Princess Nourah bint Abdulrahman University, digital mammograms of various densities via deep learning, J. Personal. Med. 10
(4) (2020) 211.
Riyadh, Saudi Arabia. The authors gratefully acknowledge the technical
[24] Jing Zheng, Denan Lin, Zhongjun Gao, Shuang Wang, Mingjie He, Jipeng Fan, Deep
and financial support provided by the Ministry of Education and King learning assisted efficient adaboost algorithm for breast cancer detection and early
Abdulaziz University, DSR. Jeddah, Saudi Arabia. diagnosis, IEEE Access 8 (2020) 96946–96954.
[25] Hanan Aljuaid, Nazik Alturki, Najah Alsubaie, Lucia Cavallaro, Antonio Liotta,
Computer-aided diagnosis for breast cancer classification using deep neural
networks and transfer learning, Comput. Methods Prog. Biomed. 223 (2022)
106951.
11
R.M. Munshi et al. Image and Vision Computing 142 (2024) 104910
[26] Manav Mangukiya, Anuj Vaghani, Meet Savani, Breast cancer detection with citation analysis using textual features and smote resampling techniques, Pattern
machine learning, Int. J. Res. Appl. Sci. Eng. Technol. 10 (2) (2022) 141–145. Recogn. Lett. 150 (2021) 250–257.
[27] Xiaomei Wang, Ijaz Ahmad, Danish Javeed, Syeda Armana Zaidi, Fahad [43] Lucia Cascone, Saima Sadiq, Saleem Ullah, Seyedali Mirjalili, Hafeez Ur
M. Alotaibi, Mohamed E. Ghoneim, Yousef Ibrahim Daradkeh, Junaid Asghar, Rehman Siddiqui, Muhammad Umer, Predicting household electric power
Elsayed Tag Eldin, Intelligent hybrid deep learning model for breast cancer consumption using multi-step time series with convolutional lstm, Big Data Res. 31
detection, Electronics 11 (17) (2022) 2767. (2023) 100360.
[28] Liang Tang, Jianming Li, Ju Meihua Bao, Yiwei Chen Xiang, Yan Wang, Genetic [44] Ahmad Hameed, Muhammad Umer, Umair Hafeez, Hassan Mustafa,
association between her2 and esr2 polymorphisms and ovarian cancer: a meta- Ahmed Sohaib, Muhammad Abubakar Siddique, Hamza Ahmad Madni, Skin lesion
analysis, Onco Targets Ther (2018) 1055–1066. classification in dermoscopic images using stacked convolutional neural network,
[29] Xiaoliang Xie, Xulin Wang, Yuebin Liang, Jingya Yang, Wu Yan, Li Li, Xin Sun, J. Ambient. Intell. Humaniz. Comput. (2021) 1–15.
Pingping Bing, Binsheng He, Geng Tian, et al., Evaluating cancer-related [45] Yuewu Hou, Zhaoying Liu, Ting Zhang, Yujian Li, C-unet: complement unet for
biomarkers based on pathological images: a systematic review, Front. Oncol. 11 remote sensing road extraction, Sensors 21 (6) (2021) 2153.
(2021) 763527. [46] Parvathaneni Naga Srinivasu, Jalluri Gnana SivaSai, Muhammad Fazal Ijaz, Akash
[30] Sihe Chen, Yongheng Chen, Yu Lanting, Hu. Xiangmei, Overexpression of socs4 Kumar Bhoi, Wonjoon Kim, James Jin Kang, Classification of skin disease using
inhibits proliferation and migration of cervical cancer cells by regulating jak1/ deep learning neural networks with mobilenet v2 and lstm, Sensors 21 (8) (2021)
stat3 signaling pathway, Eur. J. Gynaecol. Oncol. 42 (3) (2021) 554–560. 2852.
[31] H. Karamti, R. Alharthi, M. Umer, H. Shaiba, A. Ishaq, N. Abuzinadah, S. Alsubai, [47] D. Lavanya, Dr K. Usha, Rani., Analysis of feature selection with classification:
I. Ashraf, Breast cancer detection employing stacked ensemble model with breast cancer datasets, Indian J. Comput. Sci. Eng. (IJCSE) 2 (5) (2011) 756–763.
convolutional features, Cancer Biomarkers (2023) 1–16 (Preprint). [48] Sonam Jawahar Singh, Ramanathan Rajaraman, Tanmay Tulsidas Verlekar, Breast
[32] UCI Repository. Uci machine learning repository. URL https://archive.ics.uci.edu/ cancer prediction using auto-encoders, in: International Conference on Data
ml/datasets/Breast+Cancer+Wisconsin+%28Diagnostic%29. Management, Analytics & Innovation, Springer, 2023, pp. 121–132.
[33] Leo Breiman, Bagging predictors, Mach. Learn. 24 (2) (1996) 123–140. [49] O. Ibrahim Obaid, Mazin Abed Mohammed, M.K.A. Ghani, A. Mostafa, Fahad Taha,
[34] Gérard Biau, Erwan Scornet, A random forest guided tour, Test 25 (2) (2016) et al., Evaluating the performance of machine learning techniques in the
197–227. classification of Wisconsin breast cancer, Int. J. Eng. Technol. 7 (4.36) (2018)
[35] Mubariz Manzoor, Muhammad Umer, Saima Sadiq, Abid Ishaq, Saleem Ullah, 160–166.
Hamza Ahmad Madni, Carmen Bisogni, Rfcnn: traffic accident severity prediction [50] Prithwish Ghosh. Breast cancer wisconsin (diagnostic) prediction.
based on decision level fusion of machine and deep learning model, IEEE Access 9 [51] Ravi Kumar Sachdeva, Priyanka Bathla, A machine learning-based framework for
(2021) 128359–128371. diagnosis of breast cancer, International Journal of Software Innovation (IJSI) 10
[36] Afaq Juna, Muhammad Umer, Saima Sadiq, Hanen Karamti, (1) (2022) 1–11.
Ala’Abdulmajid Eshmawi, Abdullah Mohamed, Imran Ashraf, Water quality [52] Muhammet Fatih Ak, A comparative analysis of breast cancer detection and
prediction using knn imputer and multilayer perceptron, Water 14 (17) (2022) diagnosis using data visualization and machine learning applications, Healthcare 8
2592. (2) (2020), https://doi.org/10.3390/healthcare8020111. ISSN 2227-9032. URL,
[37] Elham Besharati, Marjan Naderan, Ehsan Namjoo, Lr-hids: logistic regression host- https://www.mdpi.com/2227-9032/8/2/111.
based intrusion detection system for cloud environments, J. Ambient. Intell. [53] Ashutosh Kumar Dubey, Umesh Gupta, Sonal Jain, Analysis of k-means clustering
Humaniz. Comput. 10 (9) (2019) 3669–3692. approach on the breast cancer Wisconsin dataset, Int. J. Comput. Assist. Radiol.
[38] Samina Sarwat, Naeem Ullah, Saima Sadiq, Robina Saleem, Muhammad Umer, Surg. 11 (11) (2016) 2033–2047.
Ala’Abdulmajid Eshmawi, Abdullah Mohamed, Imran Ashraf, Predicting students’ [54] Allison Murphy, Breast cancer Wisconsin (diagnostic) data analysis using gfs-tsk,
academic performance with conditional generative adversarial network and deep in: North American Fuzzy Information Processing Society Annual Conference,
svm, Sensors 22 (13) (2022) 4834. Springer, 2021, pp. 302–308.
[39] Imran Ashraf, Manideep Narra, Muhammad Umer, Rizwan Majeed, Saima Sadiq, [55] Sami Akbulut, Ipek Balikci Cicek, Cemil Colak, Classification of breast cancer on
Fawad Javaid, Nouman Rasool, A deep learning-based smart framework for cyber- the strength of potential risk factors with boosting models: a public health
physical and satellite system security threats detection, Electronics 11 (4) (2022) informatics application, Med. Bull. Haseki/Haseki Tip Bulteni 60 (3) (2022).
667. [56] Kreesh Rajani. Breast cancer survival dataset. URL https://www.kaggle.com/datas
[40] Muhammad Umer, Saima Sadiq, Michele Nappi, Muhammad Usman Sana, ets/kreeshrajani/breast-cancer-survival-dataset/code.
Imran Ashraf, et al., Etcnn: extra tree and convolutional neural network-based [57] Ankit Khushal Barai. Uci machine learning repository. URL https://www.kaggle.
ensemble model for covid-19 tweets sentiment classification, Pattern Recogn. Lett. com/datasets/ankitbarai507/breast-cancer-dataset/code.
164 (2022) 224–231. [58] Muhammad Aurangzeb Ahmad, Carly Eckert, Ankur Teredesai, Interpretable
[41] Rizwan Majeed, Nurul Azma Abdullah, Muhammad Faheem Mushtaq, machine learning in healthcare, in: Proceedings of the 2018 ACM International
Muhammad Umer, Michele Nappi, Intelligent cyber-security system for iot-aided Conference on Bioinformatics, Computational Biology, and health INFORMATICS,
drones using voting classifier, Electronics 10 (23) (2021) 2926. 2018, pp. 559–560.
[42] Muhammad Umer, Saima Sadiq, Malik Muhammad Saad Missen, Zahid Hameed, [59] Scott M. Lundberg, Su-In Lee, A unified approach to interpreting model
Zahid Aslam, Muhammad Abubakar Siddique, Michele Nappi, Scientific papers predictions, Adv. Neural Inf. Proces. Syst. 30 (2017).
12