3
3
https://doi.org/10.1007/s00521-023-09005-x (0123456789().,-volV)(0123456789().,-volV)
ORIGINAL ARTICLE
Received: 27 April 2023 / Accepted: 22 August 2023 / Published online: 19 September 2023
The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2023
Abstract
Breast cancer is a significant global health concern, with millions of cases and deaths each year. Accurate diagnosis is
critical for timely treatment and medication. Machine learning techniques have shown promising results in detecting breast
cancer. Previous studies have primarily used single-modality data for breast cancer diagnosis. Hence, this work aims to
mobilize the benefits of multimodal data over unimodality samples. This study proposes a custom deep learning-based
model pipeline that works over this multimodal data. This work has been separated into three phases. Phase 1 and Phase 2
under the unimodal category examine gene expression data and histopathological images separately. The Cancer Genome
Atlas makes these datasets available. In Phase 3, the proposed pipeline operates on both data types’ samples for each
patient in the multimodal category. This study investigates how data pre-processing (cleaning, transformation, reduction)
and cascaded filtering affect model performance. Precision, recall, f1-score, and accuracy assessed the models, whereas L2
regularization, exponentially weighted moving average, and transfer learning minimized over-fitting. A custom deep neural
network and support vector machine obtained 86% accuracy in Phase 1, whereas the VGG16 model reached 80.21%
accuracy in Phase 2. In Phase 3, the curated multimodal dataset was applied to a custom deep learning pipeline (VGG16
backbone with hyper-tuned machine learning models as head classifiers) to achieve 94% accuracy, demonstrating the
importance of multimodal data over unimodal in breast cancer subtype classification. These findings highlight the
importance of multimodal data for breast cancer diagnosis and subtype prediction.
Keywords Machine learning Unimodal data Multimodal data Breast cancer molecular subtypes Deep neural network
123
24110 Neural Computing and Applications (2023) 35:24109–24121
Using artificial intelligence (AI) in the current cancer 3. For an effective analysis, we use textual or quantitative
screening procedure makes acquiring results much more gene expression data in the first phase, histopatholog-
straightforward and convenient. Among the benefits of AI ical image data in the second phase, and integrated
approaches in breast cancer screening include quicker and data, or multimodal data, in the third phase. For each of
more accurate outcomes [8]. By enabling early identifica- the three phases, we use different ML models and also
tion and treatment of breast cancer, the use of AI and fine-tune these models to overcome the overfitting
machine learning (ML) has boosted the patient’s chance of issue.
survival [9, 10]. 4. Finally, the deep learning based pipeline applied over
Histopathological images (biopsy report), text data, CT the multimodal data samples used in this work
scans (computerized tomography scan), X-rays, gene generates a fine-grained classification output as non-
expressions (genetic activity), copy number variations cancerous cells as Normal and cancer subtypes based
(genetic alterations), and signals are a few of the different on molecules as Basal, Her2-enriched, Luminal-A,
data modalities used to detect breast cancer. Each modality Luminal-B.
can be used to identify, track, and treat the illness while
also offering particular insights into the biology of breast
cancer. Histopathological images, for example, allow us to
2 Related work
examine the structure and behaviour of breast cancer cells
at a microscopic molecular level. CT scans and X-rays
This section provides a summary of relevant research that
generate detailed images of the breast tissue, while gene
has utilized unimodal data (e.g. imaging data or gene
expressions and copy number variations provide informa-
expression data only) and multimodal data (e.g. combining
tion about the genetic makeup of the cancer cells. Magnetic
two or more data related to breast cancer) for the diagnosis
resonance imaging (MRI) signals, for example, can be used
of breast cancer. In Table 1, we also present a brief sum-
to point out the variations in the breast tissue that may
mary of recent research in this area.
signify the presence of cancer. With the integration of
different modalities, machine learning (ML) models can be
2.1 Previous research using unimodal data
trained to make diagnostic and prognostic models more
reliable and accurate. This brings AI closer to clinical
Studies using unimodal data have applied various machine
practice.
learning techniques to images or gene expression data to
The majority of existing research uses ML and deep
detect anomalies indicative of breast cancer. Kurt et al.
learning models that rely on unimodal data, which can only
[12] analyze unimodal gene expression data (METABRIC)
project a limited number of features, overlooking the larger
based on clinical and RNA-seq collected from The Cancer
clinical environment, unavoidably limiting their potential.
Genome Atlas (TCGA)1 and the Molecular Taxonomy of
Therefore, it becomes challenging to consistently link
Breast Cancer International Collaboration. In the frame-
patient characteristics to specific cancer subtypes. Inte-
work of ensemble modeling, the authors assessed six
gration of numerous data modalities offers chances to
diagnostic techniques and identified an mRNA diagnostic
improve diagnostic and prognostic models’ accuracy and
panel that can distinguish patients from healthy individuals
resilience, bringing AI closer to clinical practice [11].
and classify breast cancer into genetic subgroups. Kim
Therefore, we propose an approach that leverages mul-
et al. [13] examine the reproducibility and accuracy of
timodal data to improve the performance of machine
machine learning models for a biomedical classification
learning-based models to predict breast cancer molecular
task (including breast cancer). The authors train multiple
subtypes. This work aims to achieve the following
machine learning models (logistic regression, random for-
objectives:
est, SVM, neural nets) on gene expression data from
1. To show the efficacy of multimodal data samples over TGCA. The paper also suggests using ensemble methods
unimodal samples for breast cancer detection, a dataset and evaluating reproducibility. Phan et al. [14] employs
was curated from TCGA-BRCA such that the benefits convolutional neural networks (CNNs) to predict breast
of multimodalities can be highlighted. cancer gene expression signatures from histopathological
2. This work also proposes a custom DNN pipeline for images. The authors use four pre-trained models (VGG16,
processing unimodal and multimodal data samples ResNet50, ResNet101, and Xception) using pathological
consisting of gene expression data and histological images from a recurring breast cancer patient in the first
images for the same patient IDs. This includes phase of transfer learning and the ‘‘The Cancer Genome
comprehensive pre-processing for sample
enhancement. 1
https://portal.gdc.cancer.gov/projects/TCGA-BRCA.
123
Neural Computing and Applications (2023) 35:24109–24121 24111
Atlas Breast Invasive Carcinoma (TCGA-BRCA) dataset’’ [19] compile 4.5 million images into a single dataset.
in the second phase of the transfer learning. The study ResNet-50, ResNet-18, and VGG16 train to classify these
shows that ResNet101 can achieve reasonably high accu- images by the imaging modalities used to record them (CT,
racy (up to 0.91) in predicting breast cancer subtypes from MRI, PET, and X-ray) across several body locations. This
images without annotations. Xie et al. [15] propose using study demonstrates that it is possible to train Deep
CNNs to analyze breast cancer histopathology images to Learning (DL) Convolutional Neural Networks (CNNs)
classify them into cancer subtypes (e.g. Luminal-A, Basal). with multimodal datasets containing millions of images.
The study shows an accuracy of 97.6% using CNN Such models are applicable to real-world applications with
(Inception_ResNet_V2). hyper-scale volumes of image data. Li et al. [20] propose a
unique end-to-end multi-omics GNN framework that
2.2 Previous research using multimodal data includes inter-omics and intra-omics relationships derived
from recognized biological knowledge. For precise cate-
Recent years have witnessed major growth advancements gorization, the proposed approach integrates both learned
in the evolution of technology based on deep learning for graph features and global genome features. With the inte-
fusing multimodal data. As an example, Liu et al. [16] gration of multimodal data, Lipkova et al. [11] establish
suggest a hybrid model based on deep learning that uses approaches for AI interpretability and exploratory paths. In
multiple types of data to predict breast cancer subtypes. oncology, a variety of techniques, including radiography,
The authors combine patient imaging and gene data histology, genetics, and electronic medical records, are
modalities to create the multimodal fusion framework. used to describe the patient’s condition. They use radiog-
They use the TCGA-BRCA dataset to predict four breast raphy, histology, genetics, and oncology-specific electronic
cancer subtypes based on their genetic characteristics. Ash medical records to characterize the diagnosis. Their find-
et al. [17] utilize histopathology images to characterize ings support the use of multimodal AI to address problems
complex features such as tumor stage, to link these features including classifying cancer subtypes, predicting survival
to high-dimensional genetic markers. For analyzing com- rates, and recommending personalized therapies. Popovici
parable histology images and gene expression data, the et al. [21] present a novel approach for a joint analysis of
authors apply autoencoder and correlation analysis meth- histopathology images and gene expression data. They
ods. They intend to identify subsets of genes for whom jointly cluster the image features and gene expression data
levels of expressions are associated with subsets of mor- to find patterns in the integrated data. The clustering results
phological characteristics from the accompanying image. identify patient categories with distinct imaging and
Sun et al. [18] propose a model based on multiple kernel genetic characteristics. The use of full-color data and the
learning which employs genetic data and histopathological inclusion of other types of attributes can easily be added to
image data effectively. The model executes feature fusion, their approach. A recent survey [22] describes a variety of
which is ingrained in breast cancer classification, using the deep learning techniques for cancer diagnosis that use
aforementioned heterogeneous characteristics. By inte- several data modalities, including images, gene expression
grating 102 medical imaging datasets, Macfadyen et al. data, and clinical data. The authors present a thorough
123
24112 Neural Computing and Applications (2023) 35:24109–24121
review of contemporary deep learning approaches for integration of data obtained from phases one and two to
multimodal data fusion (early, late, and join) in cancer construct a multimodal dataset. We examine various clas-
diagnosis, including CNNs, autoencoders, and recurrent sification models to identify or categorize the subtypes of
neural networks (RNNs). breast cancer based on molecules. Furthermore, using both
The accumulation of intratumor is one of the primary unimodal and multimodal datasets, this study presents a
causes of treatment and recovery failure in breast cancer, complete analysis of the basic aspects of learning.
which is a heterogeneous disease constituted of diverse
genetic subtypes. Thus, the identification of cancer sub- 3.1 Phase 1: Gene expression data analysis
types becomes essential for accurate diagnosis and treat- and ML model exploration
ment. In particular, we emphasize the breast cancer
subtypes based on molecules in our work. In the first stage of the unimodal data, the gene expression
It may be observed from the review presented, that numeric dataset is utilized as the foundation. This section
almost all of the works, largely revolve around exploring gives a summary of the description of the dataset, the
custom machine learning models for some specific uni- workflow, and the models, as well as a proposed method-
modal sample type only and hence are limited by the ology for identifying breast cancer subtypes.
model’s performance over multimodal scenarios. The
• Workflow
favour of multimodal data based on the model’s perfor-
Figure 2 outlines the workflow used for the deploy-
mances is clearly shown in Table 1. As a result, multi-
ment of classification models trained for breast cancer
modal data not only improve prediction accuracy but also
subtype prediction using gene expression data. The data
enable cross-validation and measurement of complex
preprocessing stage involves three distinct steps, which
characteristics that are difficult (or even impossible) to
are as follows:
accurately represent with only a single data source. Com-
Data Cleaning: Handle missing data by ignoring
pared to multimodal data models, unimodal data models
tuples and removing samples with the majority of 0
provide fewer traits and information for precise disease
values to robust and highly accurate model.
detection. Even in the real-world scenario, the task of
Data Transformation: The dataset values already
cancer detection usually surrounds exploring inputs from
use Log2 normalization as primitive processing in view
multiple modalities for reaching any definite conclusions.
of the fact that log transformation reduces data
Hence in this work, we also aim to explore avenues of
variability and makes data conform precisely to the
working with multi-modalities for breast cancer molecular
normal distribution.
subtype detection.
Data Reduction: We explored various dimension-
ality reduction methods (Principal Component Analy-
sis, Chi-Square, Linear Discriminant Analysis, Singular
3 Material and methods
Value Decomposition) and it was observed that Prin-
cipal Component Analysis (PCA) outperformed com-
In this paper, we consider two publicly available breast
pared to the others methods. By selecting the ideal
cancer datasets on the TCGA website to diagnose its 4
number of features (components), PCA can compress
molecular subtypes: Basal, Her2-enriched, Luminal-A, and
your data while keeping the majority of the information
Luminal-B with 1 non-cancerous type: Normal [23, 24].
content.
The first dataset is related to the gene expression levels of
The 831 remaining samples are partitioned into
patients, whereas the second dataset consists of
training, testing, and validation sets, and it is observed
histopathological images of patients. Unimodal datasets are
that Support Vector Machine (SVM), Stochastic Gra-
analyzed independently, while multimodal datasets involve
dient Descent (SGD), K-Nearest Neighbor (KNN), and
the integration of both gene expression data and
the suggested Deep Neural Network (DNN) exhibit
histopathological images. We present our work structured
superior performance compared to other classification
into three distinctive phases, which are depicted in Fig. 1.
models in terms of precision, recall, F1-score, and
The initial phase entails exploring different machine
accuracy.
learning models alongside the proposed model for classi-
• Gene Expression Dataset
fying breast cancer molecular subtypes through gene
The initial step involved collecting a dataset of 1,098
expression data. The second phase outlines the preparation
samples, each with 20,211 genes, from the National
of the image dataset for analysis, which involves imple-
Cancer Institute website (TCGA-BRCA project).2
menting deep learning classification techniques and pre-
processing methods for the pathological images of breast
2
cancer. The third phase of this study involves the https://portal.gdc.cancer.gov/projects/TCGA-BRCA.
123
Neural Computing and Applications (2023) 35:24109–24121 24113
Fig. 1 Distribution of work overview in three phases (Phase 1: Based on gene expression data, Phase 2: Based on histopathological image data,
Phase 3: Based on multimodal data)
Normal 32
Basal 148
Her2-enriched 71
Luminal-A 412
Luminal-B 168
Total 831
123
24114 Neural Computing and Applications (2023) 35:24109–24121
3.2 Phase 2: Histopathological image data samples. These patches or samples are then separated
analysis and ML model exploration into sets for training, testing, and validation.
Step 6: In the last step, we use a transfer learning-
Using histopathology scans of patients with the same based VGG16 model to diagnose or classify the
patient-id as in phase 1, we create an image dataset for molecular subtypes of breast cancer.
testing a deep learning model that classifies the dataset into
five classes(4 classes based on molecular subtypes and 1
class for Normal samples). The description of this phase is • Preparing the Dataset
explained in the following subsections: We apply the following actions step by step to
prepare the image dataset (five histopathological
• Workflow images from TCGA-BRCA) for exploration.
Figure 4 illustrates the process flow of phase 2,
which employs a deep learning model and a distinct
data type to aid in diagnosing breast cancer subtypes. 1. Cut each image into patches of size 1024 by 1024 and
Each step is explained below: give their labels the same as those belonging to the
• Step 1: Initially, we collect five large-sized histopatho- original large image.
logical images, each of which belongs to one class. 2. Then applies the first filter, the canny filter, to all the
Step 2: We have partitioned these five images into image patches to reduce the effect of noise. This filter
equally sized (1024 1024) image patches. Table 3 is based on a Gaussian derivative, which is then used to
provides an account of the number of image patches in figure out how strong the gradients are [28].
each class. 3. Apply the convolutional filter to all image patches
Step 3: During the pre-processing stage, we use two obtained after the second step. During the convolu-
pairs of cascaded filters to eliminate noise and analyze tional filter, the brightness of a pixel is calculated based
their influence on the performance of the model for the on a weighted average of the luminance of its
image dataset. The first pair consists of Sobel and neighboring pixels [29].
Hough filters, whereas the second pair consists of 4. In the last step, a projective image transformation
Canny and Convolutional filters. approach is used to enhance the filtered image patches.
Step 4: After subjecting the image patches to After ensuing each step, Table 3 displays the configuration
filtering, we apply two transformation methods, Affine of image patches. Step 1 generates a collection of 1465
and Projective [27], to enhance the number of image image patches, which is subsequently reduced to 863 image
patches. By utilizing the projective transformation patches after applying both filters. To increase the number
method, we achieve noteworthy outcomes. Figure 5 of image patches, a projective transformation is applied.
depicts the fundamental transformation style for the Finally, we obtained a total of 1726 samples in our curated
projective method, which demonstrates how the visual image dataset.
objects alter as the observer’s viewpoint shifts. • Model Applied
Step 5: After applying the transformation methods, Image classification heavily relies on deep learning
now there are 1726 image patches obtained as total methods [30]. CNN-based models, in particular, have
Fig. 4 Process flow for exploring and deploying model on pathological image data
123
Neural Computing and Applications (2023) 35:24109–24121 24115
Table 3 Configuration of
Breast Cancer subtype Before Image Filtration After Image Filtration Projective Enhancement
histopathological image patches
Normal 378 235 470
Basal 441 264 528
Her2-enriched 266 165 330
Luminal-A 240 83 166
Luminal-B 140 116 232
Total 1465 863 1726
Normal 470
3.3 Phase 3: Multimodal data analysis and ML Basal 528
model exploration Her2-enriched 330
Luminal-A 166
By integrating the datasets from the first and second pha- Luminal-B 232
ses, we curated a multimodal dataset for ML model Total 1726
exploration. This dataset contains both gene expression and
histopathological image features [34], for a single patient
123
24116 Neural Computing and Applications (2023) 35:24109–24121
Fig. 6 Process flow to prepare multimodal dataset using gene expression data and histopathological image data
123
Neural Computing and Applications (2023) 35:24109–24121 24117
perform well when no data pre-processing steps are used, image pre-processing. Table 6 presents the accuracy values
with accuracy ranging from 61 to 75%. for both sets: the first set (Sobel and Hough filters with
Both macro and weighted averages are considered for Affine Transformation) and the second set (Canny and
analyzing all classes. The macro average computes preci- Convolution filters with Projective Transformation) based
sion, recall, and F1-score for each label, without account- on different epoch values. The first set achieves a maxi-
ing for the proportion of each label in the dataset, and mum accuracy of 73:37% over 150 epochs, while the
returns the average. Conversely, the weighted average second set attains a maximum accuracy of 80:21% over
computes precision, recall, and F1-score for each label in 300 epochs. Notably, accuracy does not increase beyond
the dataset, and returns the average [38]. 300 epochs for both sets, remaining constant in the range of
Machine learning curves aid in comparing different 100 to 300 epochs. These results are far superior to those
methods, selecting model parameters during design, opti- obtained without employing any filters or enhancing
mizing convergence by modifying optimization, and esti- methods on image samples as described in section 3.2 of
mating the training volume [39]. Figure 8 depicts the this work, with the accuracy ranging between 57% and
learning curve [40], for the support vector machine model, 70%.
showcasing the relationship between the training score and Figure 9 represents the graph between accuracy and
cross-validation score. Notably, this model attains the epochs for two sets of pre-processing methods. We can
highest accuracy. The learning curve employs training easily interpret that, using the second set, the HPT-VGG16
examples on the x-axis and score values on the y-axis. The model perform well in comparison to the first set across all
SVM classifier utilizes RBF (radial basis function) as the epochs.
kernel value while discussing the hyperparameters.
4.4 Performance of models based on multimodal
4.3 Performance of models based on unimodal dataset
histopathological image dataset
In this phase, we evaluate different classification models
Based on epochs [41], in this phase, we compare two dif- using the multimodal dataset. Table 7 presents the perfor-
ferent sets of methods using the HPT-VGG16 model during mance metric values for the DNN, KNN, SGD, MLP, and
SVM models, with precision, recall, F1-score, and accu-
racy utilized as performance metrics along with their macro
and weighted average values. The accuracy of all other
models ranges from 81% to 90%, while the multi-layered
perceptron model achieved the highest accuracy of 94%.
Not only accuracy but also the MLP model achieved the
highest macro and average recall values of 0.92 and 0.93
respectively which signifies very low false negatives.
To represent the performance of the multi-layered per-
ceptron model in this phase, accuracy and loss curves are
employed. These curves provide insight into how the
learning performance varies throughout the number of
epochs and are crucial in estimating training and validation
loss [40]. Figure 10 displays the accuracy and loss curves
for phase 3.
On the y-axis are displayed the accuracy and loss values,
Fig. 8 Learning curve for support vector machine model on gene
expression data
while the x-axis displays the number of epochs. In fine-
123
24118 Neural Computing and Applications (2023) 35:24109–24121
Table 6 Performance of
Epochs Sobel?Hough ?Affine Canny?Convolutional ?Projective
VGG16 model using
histopathological image data in 100 63.81% 72.18%
terms of accuracy
150 73.37% 75.83%
200 72.82% 78.01%
250 73.01% 79.41%
300 71.64% 80.21%
350 73.00% 78.99%
Bold font represent the highest values
123
Neural Computing and Applications (2023) 35:24109–24121 24119
5 Conclusion
Unimodal (Gene Expression DNN, SVM Weight of L2 regularization (0.05), Decay Coefficient of EWMA 86
Data) (0.99)
Unimodal (Image Data) HPT-VGG16 Canny & Convolutional Filters ? Projective Transformation 80.21
Multimodal Data MLP Solver(Adam), initial learning rate(0.01) 94
123
24120 Neural Computing and Applications (2023) 35:24109–24121
Table 9 Overall comparison of results: Type1, Type2, Type3 are the types of data and Acc1 is the accuracy corresponds to Model1, Acc2 is the
accuracy corresponds to Model2, Acc3 is the accuracy corresponds to Model3
Paper Data Type Model Result
Type1 Type2 Type3 Model1 Model2 Model3 Acc1 (%) Acc2 Acc3
one of its main limitations, which may make the overfitting Current and future burden of breast cancer: global statistics for
problem worse. However, to overcome this issue as much 2020 and 2040. Breast 66:15–23
6. Kuhl CK (2023) What the future holds for the screening, diag-
as feasible, this study employs the Exponential Weighted nosis, and treatment of breast cancer. Radiological Society of
Moving Average model (EWMA) in addition to the L2 North America
regularization, and transfer learning approaches. This lim- 7. Unger-Saldaña K (2014) Challenges to the early diagnosis and
itation can also be resolved using oversampling techniques. treatment of breast cancer in developing countries. World J Clin
Oncol 5(3):465
As an extension to the present work, we plan to gather 8. Dileep G, Gyani SGG (2022) Artificial intelligence in breast
diverse medical information (CT scans, X-rays, copy cancer screening and diagnosis. Cureus 14(10):e30318
number variation, signals etc.) beyond numeric and 9. Nassif AB, Talib MA, Nasir Q, Afadar Y, Elgendy O (2022)
histopathological images for breast cancer diagnosis to Breast cancer detection using artificial intelligence techniques: a
systematic literature review. Artif Intell Med 102276
validate various machine learning and deep learning 10. Yersal O, Barutca S (2014) Biological subtypes of breast cancer:
models. prognostic and therapeutic implications. World J Clin Oncol
5(3):412
11. Lipkova J, Chen RJ, Chen B, Lu MY, Barbieri M, Shao D,
Funding There is no funding involved in writing this paper. Vaidya AJ, Chen C, Zhuang L, Williamson DF et al (2022)
Artificial intelligence for multimodal data integration in oncol-
Data availability The dataset analyzed during this study can be found ogy. Cancer Cell 40(10):1095–1110
in (https://portal.gdc.cancer.gov/projects/TCGA-BRCA). 12. Kurt F, Agaoglu M, Arga KY (2022) Precision oncology: an
ensembled machine learning approach to identify a candidate
MRNA panel for stratification of patients with breast cancer.
Declarations OMICS 26(9):504–511
13. Kim AA, Zaim SR, Subbian V (2020) Assessing reproducibility
and veracity across machine learning techniques in biomedicine:
Conflict of interest There is no potential conflict of interest in our
a case study using TCGA data. Int J Med Inform 141:104148
paper. And all authors have seen the manuscript and approved it to
14. Phan NN, Huang C-C, Tseng L-M, Chuang EY (2021) Predicting
submit to your journal.
breast cancer gene expression signature by applying deep con-
volutional neural networks from unannotated pathological ima-
ges. Front Oncol 11:769447
References 15. Xie J, Liu R, Luttrell J IV, Zhang C (2019) Deep learning based
analysis of histopathological images of breast cancer. Front Genet
1. DeSantis CE, Bray F, Ferlay J, Lortet-Tieulent J, Anderson BO, 10:80
Jemal A (2015) International variation in female breast cancer 16. Liu T, Huang J, Liao T, Pu R, Liu S, Peng Y (2022) A hybrid
incidence and mortality rates international variation in female deep learning model for predicting molecular subtypes of human
breast cancer rates. Cancer Epidemiol Biomark Prevent breast cancer using multimodal data. Irbm 43(1):62–74
24(10):1495–1506 17. Ash JT, Darnell G, Munro D, Engelhardt BE (2021) Joint analysis
2. Goldin A, Venditti JM, Macdonald JS, Muggia FM, Henney JE, of expression levels and histological images identifies genes
Devita Jr VT (1981) Current results of the screening program at associated with tissue morphology. Nat Commun 12(1):1609
the division of cancer treatment, national cancer institute. Eur J 18. Sun D, Li A, Tang B, Wang M (2018) Integrating genomic data
Cancer (1965) 17(2), 129–142 and pathological images to effectively predict breast cancer
3. Momenimovahed Z, Salehiniya H (2019) Epidemiological char- clinical outcome. Comput Methods Programs Biomed 161:45–53
acteristics of and risk factors for breast cancer in the world. 19. MacFadyen C, Duraiswamy A, Harris-Birtill D (2023) Classifi-
Breast Cancer Targets Ther 151–164 cation of hyper-scale multimodal imaging datasets. medRxiv
4. Siegel RL, Miller KD, Wagle NS, Jemal A (2023) Cancer 2023-01
statistics. CA Cancer J Clin 73(1):17–48 20. Li B, Nabavi S (2023) A multimodal graph neural network
5. Arnold M, Morgan E, Rumgay H, Mafra A, Singh D, Laversanne framework for cancer molecular subtype classification. arXiv
M, Vignat J, Gralow JR, Cardoso F, Siesling S et al (2022) preprint arXiv:2302.12838
123
Neural Computing and Applications (2023) 35:24109–24121 24121
21. Popovici V, Budinska E, Čápková L, Schwarz D, Dušek L, Feit J, 33. Tammina S (2019) Transfer learning using VGG-16 with deep
Jaggi R (2016) Joint analysis of histopathology image features convolutional neural network for classifying images. Int J Sci Res
and gene expression in breast cancer. BMC Bioinform 17(1):1–9 Publ (IJSRP) 9(10):143–150
22. M’Sabah CEL, Bouziane A, Ferdi Y (2021) A survey on deep 34. Khalid S, Khalil T, Nasreen S (2014) A survey of feature
learning methods for cancer diagnosis using multimodal data selection and feature extraction techniques in machine learning.
fusion. In: 2021 international conference on e-health and bio- In: 2014 science and information conference. IEEE, pp 372–378
engineering (EHB), pp 1–4. IEEE 35. Ross AA, Govindarajan R (2005) Feature level fusion of hand
23. Hou Y (2020) Breast cancer pathological image classification and face biometrics. In: Biometric technology for human iden-
based on deep learning. J Xray Sci Technol 28(4):727–738 tification II, vol 5779. SPIE, pp 196–204
24. Zhu Z, Albadawy E, Saha A, Zhang J, Harowicz MR, Mazur- 36. Taud H, Mas J (2018) Multilayer perceptron (MLP). Geomatic
owski MA (2018) Breast cancer molecular subtype classification approaches for modeling land change scenarios, pp 451–455
using deep features: preliminary results. In: Medical imaging 37. Tohka J, Van Gils M (2021) Evaluation of machine learning
2018: computer-aided diagnosis, vol 10575. SPIE, pp 651–656 algorithms for health and wellness applications: a tutorial.
25. Kotsiantis SB, Zaharakis I, Pintelas P et al (2007) Supervised Comput Biol Med 132:104324
machine learning: a review of classification techniques. Emerging 38. Takahashi K, Yamamoto K, Kuchiba A, Koyama T (2022)
Artif Intell Appl Comput Eng 160(1):3–24 Confidence interval for micro-averaged f 1 and macro-averaged f
26. Dwivedi AK (2018) Artificial neural network model for effective 1 scores. Appl Intell 52(5):4961–4972
cancer classification using microarray gene expression data. 39. Mohr F, van Rijn JN (2022) Learning curves for decision making
Neural Comput Appl 29:1545–1554 in supervised machine learning—a survey. arXiv preprint arXiv:
27. Redzuwan R, Radzi NAM, Din NM, Mustafa I (2015) Affine 2201.12150
versus projective transformation for sift and ransac image 40. Viering T, Loog M (2022) The shape of learning curves: a
matching methods. In: 2015 IEEE international conference on review. IEEE Trans Pattern Anal Mach Intell
signal and image processing applications (ICSIPA). IEEE, 41. Brownlee J (2018) What is the difference between a batch and an
pp 447–451 epoch in a neural network. Mach Learn Mastery 20
28. Rana R, Verma A (2014) Comparison and enhancement of digital 42. Sun D, Wang M, Li A (2018) A multimodal deep neural network
image by using canny filter and sobel filter. IOSR J Comput Eng for human breast cancer prognosis prediction by integrating
16(1):06–10 multi-dimensional data. IEEE/ACM Trans Comput Biol Bioinf
29. Jeong W-K, Pfister H, Fatica M (2011) Medical image processing 16(3):841–850
using GPU-accelerated ITK image filters. In: GPU computing 43. Liu H, Dai Z, So D, Le QV (2021) Pay attention to MLPS. Adv
gems emerald edition. Elsevier, New York, pp. 737–749 Neural Inf Process Syst 34:9204–9215
30. Suzuki K (2017) Overview of deep learning in medical imaging.
Radiol Phys Technol 10(3):257–273 Publisher’s Note Springer Nature remains neutral with regard to
31. Alzubaidi L, Fadhel MA, Al-Shamma O, Zhang J, Santamarı́a J, jurisdictional claims in published maps and institutional affiliations.
Duan Y, Oleiwi SR (2020) Towards a better understanding of
transfer learning for medical imaging: a case study. Appl Sci
Springer Nature or its licensor (e.g. a society or other partner) holds
10(13):4523
exclusive rights to this article under a publishing agreement with the
32. Boumaraf S, Liu X, Zheng Z, Ma X, Ferkous C (2021) A new
author(s) or other rightsholder(s); author self-archiving of the
transfer learning based approach to magnification dependent and
accepted manuscript version of this article is solely governed by the
independent classification of breast cancer in histopathological
terms of such publishing agreement and applicable law.
images. Biomed Signal Process Control 63:102192
123