Cancers Pyhtonlibrary
Cancers Pyhtonlibrary
Article
MR-Class: A Python Tool for Brain MR Image Classification
Utilizing One-vs-All DCNNs to Deal with the Open-Set
Recognition Problem
Patrick Salome 1,2,3,4, * , Francesco Sforazzini 1,2,3 , Gianluca Grugnara 5 , Andreas Kudak 4,6,7 ,
Matthias Dostal 4,6,7 , Christel Herold-Mende 8,9 , Sabine Heiland 5 , Jürgen Debus 3,4,6 ,
Amir Abdollahi 1,3,4,6 and Maximilian Knoll 1,3,4,6, *
1 Clinical Cooperation Unit Radiation Oncology, German Cancer Research Center, 69120 Heidelberg, Germany
2 Heidelberg Medical Faculty, Heidelberg University, 69117 Heidelberg, Germany
3 German Cancer Consortium Core Center Heidelberg, 69120 Heidelberg, Germany
4 Heidelberg Ion-Beam Therapy Center, 69120 Heidelberg, Germany
5 Department of Neuroradiology, Heidelberg University Hospital, 69120 Heidelberg, Germany
6 Department of Radiation Oncology, Heidelberg University Hospital, 69120 Heidelberg, Germany
7 Clinical Cooperation Unit Radiation Therapy, German Cancer Research Center, 69120 Heidelberg, Germany
8 Brain Tumour Group, European Organization for Research and Treatment of Cancer, 1200 Brussels, Belgium
9 Division of Neurosurgical Research, Department of Neurosurgery, University of Heidelberg,
69117 Heidelberg, Germany
* Correspondence: [email protected] (P.S.); [email protected] (M.K.)
Simple Summary: MR-Class is a deep learning-based MR image classification tool for brain images
that facilitates and speeds up the initialization of big data MR-based studies by providing fast, robust
and quality-assured imaging sequence classifications. Our studies observed up to 10% misclassifi-
cation rates due to corrupt and misleading DICOM metadata. This highlights the need for a tool
such as MR-Class to help with data curation. MR-Class can be integrated into workflows for DICOM
inconsistency checks and flagging or completing missing DICOM metadata and thus contribute to
Citation: Salome, P.; Sforazzini, F.;
the faster deployment of clinical artificial intelligence applications.
Grugnara, G.; Kudak, A.; Dostal, M.;
Herold-Mende, C.; Heiland, S.;
Debus, J.; Abdollahi, A.; Knoll, M.
Abstract: Background: MR image classification in datasets collected from multiple sources is compli-
MR-Class: A Python Tool for Brain cated by inconsistent and missing DICOM metadata. Therefore, we aimed to establish a method for
MR Image Classification Utilizing the efficient automatic classification of MR brain sequences. Methods: Deep convolutional neural net-
One-vs-All DCNNs to Deal with the works (DCNN) were trained as one-vs-all classifiers to differentiate between six classes: T1 weighted
Open-Set Recognition Problem. (w), contrast-enhanced T1w, T2w, T2w-FLAIR, ADC, and SWI. Each classifier yields a probability,
Cancers 2023, 15, 1820. https:// allowing threshold-based and relative probability assignment while excluding images with low
doi.org/10.3390/cancers15061820 probability (label: unknown, open-set recognition problem). Data from three high-grade glioma
Academic Editors: Jean-Emmanuel (HGG) cohorts was assessed; C1 (320 patients, 20,101 MRI images) was used for training, while
Bibault and Lei Xing C2 (197, 11,333) and C3 (256, 3522) were for testing. Two raters manually checked images through
an interactive labeling tool. Finally, MR-Class’ added value was evaluated via radiomics model
Received: 6 January 2023
performance for progression-free survival (PFS) prediction in C2, utilizing the concordance index
Revised: 11 March 2023
(C-I). Results: Approximately 10% of annotation errors were observed in each cohort between the
Accepted: 15 March 2023
DICOM series descriptions and the derived labels. MR-Class accuracy was 96.7% [95% Cl: 95.8, 97.3]
Published: 17 March 2023
for C2 and 94.4% [93.6, 96.1] for C3. A total of 620 images were misclassified; manual assessment of
those frequently showed motion artifacts or alterations of anatomy by large tumors. Implementation
of MR-Class increased the PFS model C-I by 14.6% on average, compared to a model trained without
Copyright: © 2023 by the authors. MR-Class. Conclusions: We provide a DCNN-based method for the sequence classification of brain
Licensee MDPI, Basel, Switzerland.
MR images and demonstrate its usability in two independent HGG datasets.
This article is an open access article
distributed under the terms and
Keywords: content-based image classification; data curation and preparation; convolutional neural
conditions of the Creative Commons
networks (CNN); deep learning; artificial intelligence (AI)
Attribution (CC BY) license (https://
creativecommons.org/licenses/by/
4.0/).
1. Introduction
Magnetic resonance imaging (MRI) has become a crucial imaging modality in the
detection and staging of various types of cancer, including brain tumors [1]. With its
high-resolution images, MRI can distinguish between healthy and diseased tissues, pro-
viding high soft-tissue contrast and overcoming limitations of other imaging techniques,
specifically in disease monitoring [1]. Therefore, there has been an increase in interest in
MRI-based artificial intelligence (AI) applications and studies in recent years [2]. An essen-
tial step in the data preparation phase of such studies is accurately classifying MR images,
since each image communicates specific anatomical or physiological information [1]. An
example is brain tumor segmentation algorithms requiring information from multiple MR
modalities, as distinguishing between healthy brain tissue and tumors is often challenging.
Ensuring that the right sequences are used for analysis, i.e., the classification of sequences,
can be a demanding and time-intensive task. This is particularly true when dealing with
a large amount of data from different sources, such as multiple scanners and treatment
centers, which may have inconsistent naming schemes. In particular, retrospective data
collection introduces additional challenges (non-prespecified protocols and sequences).
Gueld et al. found that relying on image metadata (i.e., information stored in the
DICOM header) to classify medical images can often lead to unreliable results [3]. This is
because there may not always be a consistent match between the DICOM tags and the actual
examination protocols used. These inconsistencies are often introduced in order to improve
the quality of the imaging, such as implementing different imaging protocols for different
regions of the body to account for variabilities in patients’ anatomies [3]. Harvey et al.
report data labeling as the costliest part of radiomics studies [4] and that consistent and
unbiased labeling should be performed across the entire dataset to yield robust machine
learning models [4]. However, this can be challenging when large amounts of data are
considered. Therefore, automating the retrieval of medical images and classifying data
based on its content would bring several advantages, including improved time efficiency,
accuracy, and ultimately, reproducibility.
Compared to text-based image classification, content-based image classification (CBIC)
is independent of inconsistencies between different image sources, is unaffected by human
error, and is less labor-intensive [5]. CBIC methods for medical images include the use of
traditional classification machine learning techniques such as K-nearest neighbor (kNN) [6],
support vector machine [7] (SVM), as well as deep learning methods [8]. After the success
of the deep convolutional neural network (DCNN) AlexNet [9] in the ImageNet [10]
classification challenge, an increase in interest in DCNN has been seen when dealing with
image classification tasks [11–13]. Regarding the utilization of DCNNs for medical image
retrieval and classification, four studies have been identified that achieved high levels
of accuracy (>90%) in classifying body organs and MR images [14–17]. A summary of
these models can be seen in Supplementary Table S1. One limitation of these methods is
their inability to address the open-set recognition problem, which refers to the failure of
a network trained to classify a specific number of classes to handle unknown classes [18].
This problem is frequently encountered in clinical cohorts since datasets exported from
hospitals’ picture archiving and communication systems (PACS) usually comprise all
available medical images and data, leading to various medical image modalities and
sequences.
This work addresses the open-set recognition problem by training a DCNN-based MR
image classifier using a one-vs-all approach. Following a comparison study of published
DCNNs for medical image classification to determine the chosen DCNN model, one-vs-all
binary class-specific DCNN classifiers were trained to recognize a particular MR image,
thus forming MR-Class.
Cancers 2023, 15, 1820 3 of 16
2.2. MR Scans
Multiparametric MRIs (mpMRI) were collected from multiple scanners in all three
datasets, resulting in heterogeneous modalities and MR sequence protocols (Supplementary
Table S3). Both conventional multislice (2D) scans acquired in the axial, sagittal, or coronal
planes, as well as 3D scans, are available. The MR sequences found in the cohorts are the
widely used sequences for brain tumor imaging [20] in clinical routines and trials [21–23].
All MR images discovered in the training cohort were utilized for training. However, one-
vs-all DCNN classifiers were only trained for T1w, contrast-enhanced T1w (T1wce), T2w,
T2w fluid-attenuated inversion recovery (FLAIR), apparent diffusion coefficient (ADC), and
susceptibility-weighted imaging (SWI). It should be noted that no SWI scans were found
in C3. The in-plane resolution ranged from 0.33 × 0.33 to 2 × 2 mm for C1, 0.45 × 0.45
to 1.40 × 1.40 mm for C2, and 0.45 × 0.45 to 1.14 × 1.14 mm for C3. The slice thickness
of all MR scans ranged from 0.9 to 7.5 mm. Each MR image was manually labeled by
human experts using an in-house interactive labeling tool. The “Series Description” (SD)
and “Contrast/Bolus Agent” DICOM attributes were then extracted and compared to the
derived labels to evaluate the consistency of the metadata. Sample images found in the
training and testing cohorts are shown in Figure 1.
Figure 1. Cont.
Cancers 2023, 15, 1820 4 of 16
Figure 1. Sample images of the different MR images present in the three datasets C1–C3.
Φ-Net and DeepDicomSort. As for the 2D and 3D ResNet-18 DCNNs, magnetic field
inhomogeneities of the T1w images were first corrected using the N4ITK algorithm [26].
After reorienting to a common orientation, in-plane cropping was performed to remove
background voxels. Then, to account for resolution variability, all MR scans were resampled
to a uniform pixel spacing of 2 × 2 mm2 , and volumes were interpolated to a 2 mm slice
thickness. Images were then cropped around the brain into a digital grid of 224 × 224 × 224.
Padding was performed when the image shape was smaller than the target grid. Lastly, a
z-score normalization of the brain voxels was applied to bring all MR images to the same
intensity scale. The formula of the Z-score normalization is as follows:
x−µ
=z (1)
σ
where x is the voxel intensity, µ is the mean of the intensity distribution, and σ is the
standard deviation.
Table 1. Number (%) of MR images from the training cohort, C1, considered for each one-vs-all
DCNN classifier. T2w-FL: T2-FLAIR.
Training Validation
DCNN Targeted Remaining Targeted Remaining
Classifier Class Images Class Images
T1w-vs-all 3152 (15.7) 12,929 (64.3) 788 (3.9) 3232 (16.1)
T2w-vs-all 1576 (7.9) 14,505 (72.1) 394 (2.0) 3626 (18.0)
T2w-FL-vs-all 1535 (7.6) 14,546 (72.4) 384 (1.9) 3636 (18.1)
ADC-vs-all 1550 (7.7) 14,530 (72.3) 388 (1.9) 3633 (18.1)
SWI-vs-all 1183 (5.9) 14,898 (74.1) 296 (1.5) 3724 (18.5)
Figure 2. MR-Class training workflow. MR-Class comprises five one-vs-all DCNNs, one for each
class, and the T1w-vs-T1wce binary DCNN. After MR image preprocessing, each DCNN was trained
with an 80%/20% training/validation split, with class 1 representing the DCNNs’ target class and
0 for the rest. For the T1w-vs-T1wce DCNN, class 0 was assigned to T1w and 1 to T1wce. T2w-FL:
T2w-FLAIR, T1wce: T1w contrast-enhanced.
from around the middle slice of the corresponding MR acquisition plane. Figure 3 provides
an overview of the inference workflow.
Figure 3. MR-Class inference workflow. C2 and C3 were used for testing. After preprocessing, MR
images are passed to the 5 one-vs-all DCNN classifiers. A classification probability threshold of 0.95
was used. If none of the classifiers labels an image, it is rendered as other. If more than one classifier
labels a specific image, then the image is labeled by the classifier with the highest probability.
The classifications were compared to the ground truth labels, and accuracy was
calculated as the number of correct predictions divided by the total number of images.
A 95% confidence interval (CI) was determined using the Wilson interval method [29].
Classification sensitivity and specificity were also calculated to assess the performance of
each classifier. Additionally, misclassified images were examined to determine the causes
of misclassifications.
3. Results
3.1. Metadata Consistency
Between all three datasets, 2704 different DICOM SDs were found (an overview of
the number of SDs found for each MR scan is shown in Supplementary Table S4). A total
of 11.4%, 10.6%, and 10.7% of the SDs for C1, C2, and C3, respectively, had misleading
Cancers 2023, 15, 1820 8 of 16
or inconsistent entries, not allowing for the proper identification of the MR image class
(Table 2).
Table 2. Percentage of labeling errors for each class considered in the cohorts. T2w-FL: T2w-FLAIR.
C1 C2 C3
n % Error n % Error n % Error
T1w 2023 15.1 1189 11.2 433 13.4
T1wce 1917 13.9 4315 13.4 1096 9.9
T2w 1970 9.3 630 11.7 347 10.3
T2w-FL 1919 7.2 811 10.5 389 8.2
ADC 1938 7.6 895 8.4 122 5.5
SWI 1479 6.3 486 6.6 - -
Other 8855 13.1 3007 7.3 1135 12.1
All 20,101 11.4 11,333 10.6 3522 10.7
Table 3. Classification accuracy of the different DCNN architectures in the study. T2w-FL: T2w-
FLAIR.
All classifiers achieved a high comparable accuracy, with the 2D ResNet-18 having the
highest overall accuracy of 98.6%. The training took 18–20 h for the 3D DCNN (Φ-Net and
3D ResNet-18) and 8–10 h for the 2D DCNN (DeepDicomSort and 2D ResNet-18) on an
Intel Xeon processor with 8 cores and 32 Gb of RAM and a graphics card NVIDIA GeForce
GTX 1060 (6 Gb). The average inference time was 0.15 s for a single 2D slice and 4.92 s for a
3D image. Thus, the DCNN one-vs-all architecture implemented in MR-Class was that of
the 2D ResNet-18 (a schematic representation is shown in Figure 4).
Table 4. Validation classification accuracies of all six binary DCNN classifiers on C1. T2wFL: T2w-
FLAIR.
All six classifiers demonstrate high validation accuracies, with the lowest being 97.7%
for the T1w-vs-T1wce and the highest being 99.7% for the SWI-vs-all, with a value of
99.6% for the ADC-vs-all tasks. Upon passing back the training set dataset I to MR-Class
during inference mode, an accuracy of 97.4% [95% CI: 96.2, 98.4] was achieved, meaning
that out of 20101 MR scans, MR-Class was unable to learn 519. In the multi-class versus
multiple binary one-vs-all classification experiment, where only the image volumes of the
six considered MR sequences were considered, the validation accuracy was comparable at
98.6% and 98.1%, respectively.
Figure 5 illustrates the distributions of the classification probabilities obtained by the
MR-Class for all three cohorts. Based on C1, a probability cutoff threshold of 0.95 was set
for testing MR-Class on C2 and C3.
Figure 5. Distribution of the probabilities of correct and wrong labeled images for all three cohorts
in the study when inferred to MR-Class. Based on the distributions of C1, a cutoff classification
threshold probability of 0.95 was used. Histogram bin width = 0.01.
and T1w-vs-all had the lowest sensitivity at 91.9% and 96.6%, respectively, while all other
DCNNs had high sensitivity (>99%) (Figure 6A, upper panel). In the multi-class normalized
confusion matrix (Figure 6A, lower panel), T1w classification was found to be the least
reliable, with an accuracy of 91.17%. When tested against the independent C3 dataset,
MR-Class achieved an accuracy of 94.4% [95% CI: 93.6, 96.1], with 196 out of 3522 scans
being misclassified. T1w-vs-T1wce exhibited the lowest sensitivity at 97.4%, while all
other DCNNs had a sensitivity greater than 98%. Specificity ranged from 91.3% (T2w-
vs-all) to 98.8% (T1w-vs-T1wce) (Figure 6B, upper panel). In the multi-class confusion
matrix (Figure 6B, lower panel), T2w classification was found to be the least reliable,
with an accuracy of 91.35%, with 8.65% classified as “other”. The next section includes
investigations on the misclassified images.
Figure 6. Confusion matrices of the six DCNNs for C2 (A) and C3 (B). The upper panels in (A,B) show
the confusion matrices for datasets C2 and C3. The lower panels in (A,B) show MR-Class normalized
confusion matrices for datasets C2 and C3, i.e., the percentages (%) of correct classification results
per class. SE: sensitivity; SP: specificity. Class ‘Other’: when none of the DCNNs labels an image; n:
number of scans per class, T2w-FL: T2w-FLAIR.
Cancers 2023, 15, 1820 11 of 16
Figure 7. Examples of misclassified images. The first two images are examples of a misclassified MR,
possibly due to blurry images (left) and alterations in expected anatomy (displaced ventricles, large
tumor, right). The next three MR images show incorrect predictions due to different MR artifacts
(shading, motion, aliasing). All of these images are falsely classified as “other”. The last image is a
diffusion-weighted image (DWI), specifically a trace DWI, misclassified as T2w.
Category n %
MR artifact-other 146 26.84
MR artifact-middle slice blurring 127 23.35
Tumor/GTV displacing ventricles 122 22.43
Similar content-different sequence 80 14.71
DWI as T2w 76 13.97
DICOM corrupted images 69 12.68
The MR-Class-curated dataset excluded 4 misclassified T1wce images as they were labeled
“other”.
Figure 8. Box plots of the 1st–99th percentile C-Is attained by the MR-class and DICOM series
description (SD) curated dataset models fitted by the respective signatures after three resampling
approaches. MCCV: Monte Carlo cross-validation, BStrap: bootstrapping, CV: cross-validation.
4. Discussion
In this manuscript, we present an MRI image sequence classifier, MR-Class, which
differentiates between T1w, contrast-enhanced T1w, T2w, T2w-FLAIR, ADC, and SWI
while handling unknown classes. Testing was performed on two independent cohorts,
where classification accuracies of 96.7% [95% CI: 95.8, 97.3] and 94.4% [95% CI: 93.6,
96.1] were observed. MR-Class consists of five one-vs-all DCNNs (one for each class),
followed by a binary classifier for T1w images, to determine whether a contrast agent was
administrated. This design enables MR-Class to handle unknown classes since each DCNN
only classifies an image if it belongs to its respective class, and thus an image not labeled
by any of the DCNNs is rendered as unknown. In order to compare the effectiveness of
the multiple one-vs-all binary classifier approach with the basic multi-class classification
method, we conducted a multi-class vs multiple dual-class classification experiment. The
results showed that both methods achieved a similar classification accuracy for MR brain
image classification, with the multi-class method achieving a 98.6% accuracy and the
multiple one-vs-all method achieving a 98.1% accuracy. However, the latter approach has
an advantage in that it can handle the open-set recognition problem often encountered
when dealing with data from clinical cohorts, which can ultimately help reduce MRI study
design times.
MR image DICOM series description (SD) entries usually follow the MR sequence
protocol applied. However, they are MR model specific and are sometimes edited by clinical
staff. We observed around 10% discrepancies in each cohort when the SD was compared to
the manually derived labels. Typical SDs that do not allow for clear MR scan classifications
are SDs with only the sequence name, e.g., spin echo (SE), or the scan direction, e.g., axial,
which can stand for any MR sequence. Typographical errors and empty SD attributes were
also observed.
Overall, high accuracies were obtained for all DCNNs in the comparison study. In
conjunction with the high performance achieved in the literature on medical image classi-
fication [13,14,16,17], it is apparent that DCNNs can learn the intricacies behind different
medical image modalities. In the DCNN architecture comparison study, 2D ResNet-18 had
the best overall accuracy and thus was the architecture chosen for MR-Class. Furthermore,
it was seen that the 2D DCNNs outperformed their 3D counterparts in MR sequence clas-
sification. MR scans correctly classified by the 2D DCNNs, while misclassified by the 3D
DCNNs, are mainly conventional 2D axial, sagittal, or coronal scans with slice thicknesses
ranging between 5 and 9 mm. Scans with a field of view that only encompassed the tumor
area were misclassified by both 3D DCNNs (representative images can be seen in Supple-
Cancers 2023, 15, 1820 13 of 16
mentary Figure S2). It is important to note that no data augmentation was performed in
the comparison study.
All six one-vs-all classifiers had high validation accuracies, with the lowest being
97.7% for the T1w-vs-T1wce. After inferring back the training cohort C1 to MR-Class, it
was observed that 519 images could not be learned, out of which 336 belonged to class
“other”, representing 3.8% of the other images used for training. This low error percentage
demonstrates that MR-Class can learn to handle different sequences indirectly.
The testing of MR-Class against C2 and C3 yielded an average accuracy of 96.1%,
where 620 images (4.2%) were classified incorrectly. Overall, T2w-vs-all had the worst
performance, with a specificity of 93.5% and 91.3% in C2 and C3. This is mostly due to
the presence of diffusion-weighted imaging (DWI) sequences (frequently encountered
in the datasets), which are inherently a series of T2w sequences. Similarly, C3 included
T1w-FLAIR images falsely misclassified as T1w or T2w-FLAIR. Thus, different sequences
with similar content are prone to misclassification by MR-Class. A solution could be to
train a subsequent classifier to distinguish between similar sequences, as performed for
the T1wce images. The majority of misclassified images had severe blurring or other MR
artifacts. A higher prevalence of misclassifications was observed in C3 compared to C2, and
many of these misclassifications were false negatives, resulting in images being labeled as
unclassifiable by MR-Class. This could be beneficial for radiomics models as any corrupted
images would be automatically disregarded, and all images labeled as a specific class
would have analogous content. Another subset of misclassified images showed tumor
volumes overlapping with the ventricles, and statistical analysis confirmed that altered
anatomy (specifically, ventricle displacement by large tumors) could be a possible reason
for misclassification. However, further analysis is needed to assess the impact of surgery
on alterations in overall anatomy, such as biopsy, partial resection, total resection, as well
as the effect of treatments such as chemotherapy and radiotherapy on tumor patterns and
contrast enhancement.
To improve the accuracy of MR-Class and reduce misclassifications, additional data
including more diverse MR images should be considered. Additionally, for DWI images
misclassified as T2w, a possible solution would be to train a subsequent classifier to
distinguish between different MR image protocols, as was done for T1wce images. Another
approach to reducing misclassification is to design a workflow that incorporates both
text-based and content-based classification (MR-Class). In cases of mismatch between
the two methods, manual intervention can be introduced to classify the image correctly.
Misclassifications can be corrected through manual label adjustments, and upon the user’s
acceptance, these images can be used to retrain the models using a transfer learning
approach to improve MR-Class accuracy.
An essential step in building a radiomics application is to verify the input data labels
before training the machine learning model, as inconsistent data can lead to the model
drastically failing [32]. However, this was not performed while building the different
survival models to demonstrate the applicability of MR-Class in MR-based radiomics
applications. CPHs models were built with the T1wce MR sequences of cohort C2 to
predict the patients’ PFS after performing a text-based curation using the DICOM SDs
and a content-based curation using MR-Class. The MR-Class-curated model achieved an
average C-I increase of 14.6%. This is mainly due to the content dissimilarity between the
different images in the DICOM SD-curated dataset compared to the MR-Class-curated
dataset.
MR-Class can facilitate the preparation of longitudinal studies for RT treatment assess-
ment as MR data from the three cohorts include scans taken before, after, and throughout
the delivery of the RT fractions, which resulted in different tumor volume masses between
the different scans, as well as apparent radiation scaring in some of the MR images. Fur-
thermore, the data cohort includes images taken directly after the surgical resection of the
tumor, resulting in visible surgical holes and void tumor beds.
Cancers 2023, 15, 1820 14 of 16
5. Conclusions
MR-Class is a helpful tool for automating the classification of MR images, thus elimi-
nating the need for manual sorting and saving time for researchers. The tool is particularly
useful for studies involving large amounts of data and different naming schemes, as it
classifies images based on their content rather than metadata, and can automatically dis-
regard corrupted images. Future work includes expanding the tool to include additional
modalities and sequences for different anatomical sites.
Supplementary Materials: The following supporting information can be downloaded at: https:
//www.mdpi.com/article/10.3390/cancers15061820/s1. Figure S1: An example of ResNet and VGG
architectures with 18 and 16 layers and two output neurons; Figure S2: Images classified correctly
and wrongly by 3D and 2D classifiers; Figure S3: T1wce preprocessing diagram applied before
survival prediction modeling; Table S1: State-of-the-art methods specifications; Table S2: Patient
demographics; Table S3: MR models found in the cohorts; Table S4: Number of series descriptions
found for each class; Table S5: The number of shape, first- and second-order statistics derived per
sequence and calculated on both the original and derived images.
Author Contributions: Conceptualization, P.S., F.S., A.A. and M.K.; Data curation, P.S. and F.S.;
Formal analysis, P.S., F.S. and M.K.; Funding acquisition, J.D. and A.A.; Investigation, G.G., C.H.-M.
and S.H.; Methodology, P.S. and M.K.; Resources, A.K. and M.D.; Software, P.S. and F.S.; Supervision,
J.D., A.A. and M.K.; Visualization, P.S.; Writing—original draft, P.S.; Writing—review and editing,
F.S., G.G., C.H.-M., S.H., A.A. and M.K. All authors have read and agreed to the published version of
the manuscript.
Funding: This study was supported by the European Union’s Horizon 2020 research and innova-
tion programme under the Marie Skłodowska-Curie Innovative Training Network grant agreement
(MSCA-ITN-PREDICT, No 766276), the collaborative research center of the German Research Founda-
tion (DFG, Unite, SFB-1389, Project 404521405), Zentrum für Personalisierte Medizin (ZPM-Network
BW, Project PROMI), the intramural funds of the National Center for Tumor Diseases (NCT) and the
German Cancer Consortium (DKTK) Radiation Oncology programs.
Institutional Review Board Statement: The study was conducted according to the guidelines of the
Declaration of Helsinki and approved by the Institutional Review Board of the Medical Faculty of
Heidelberg University (approval number S-540/2010, date of last updated approval: 20 July 2020).
Informed Consent Statement: Informed consent was obtained from all subjects involved in the S-540
study.
Data Availability Statement: MR-Class is available on our Github page, https://github.com/TRO-
HIT/MR-Class, and is integrated into our big data curation tool for radiotherapy applications, Py-
CuRT [33] https://github.com/TRO-HIT/PyCRT. The public C3 used for testing can be downloaded
at https://wiki.cancerimagingarchive.net/display/Public/TCGA-GBM. C1 and C2 are available
from the corresponding authors on reasonable request.
Acknowledgments: This publication uses some data collected and results obtained within the
framework of the PhD thesis of Patrick Salome published at the medical faculty of the University of
Heidelberg.
Cancers 2023, 15, 1820 15 of 16
Conflicts of Interest: P.S. No relevant relationships. F.S. No relevant relationships. A.K. No relevant
relationships. N.B. No relevant relationships. J.D. Grants/contracts from/with Viewray, CRI—The
Clinical Research Institute, Accuray International Sarl, RaySearch Laboratories, Vision RT, Merck
Serono, Astellas Pharma, AstraZeneca, Siemens Healthcare, Solution Akademie, Ergomed PLC
Surrey Research Park, Quintiles, Pharmaceutical Research Associates, Boehringer Ingelheim Pharma
& CoKG, PTW-Freiburg Dr. Pychlau, Nanobiotix, Accuray, Varian; participation on a data safety
monitoring board or advisory board for Merck Serono. A.A. Predict MarieCurie innovative training
network (ITN), in frame of Horizon 2020 from the European Union, Marie Skłodowska-Curie grant
agreement No 766276. M.K. No relevant relationships.
References
1. Westbrook, C.; Talbot, J. MRI in Practice; John Wiley & Sons: Hoboken, NJ, USA, 2018.
2. Kickingereder, P.; Isensee, F.; Tursunova, I.; Petersen, J.; Neuberger, U.; Bonekamp, D.; Brugnara, G.; Schell, M.; Kessler, T.; Foltyn,
M.; et al. Automated quantitative tumour response assessment of MRI in neuro-oncology with artificial neural networks: A
multicentre, retrospective study. Lancet Oncol. 2019, 20, 728–740. [CrossRef] [PubMed]
3. Gueld, M.O.; Kohnen, M.; Keysers, D.; Schubert, H.; Wein, B.B.; Bredno, J.; Lehmann, T.M. Quality of DICOM header information
for image categorization. In Proceedings of the Medical Imaging 2002: PACS and Integrated Medical Information Systems:
Design and Evaluation, San Diego, CA, USA, 16 May 2002; pp. 280–287.
4. Harvey, H.; Glocker, B. A standardized approach for preparing imaging data for machine learning tasks in radiology. In Artificial
Intelligence in Medical Imaging; Springer: Berlin, Germany, 2019; pp. 61–72.
5. Névéol, A.; Deserno, T.M.; Darmoni, S.J.; Güld, M.O.; Aronson, A.R. Natural language processing versus content-based image
analysis for medical document retrieval. J. Am. Soc. Inf. Sci. Technol. 2009, 60, 123–134. [CrossRef]
6. Wagle, S.; Mangai, J.A.; Kumar, V.S. An improved medical image classification model using data mining techniques. In
Proceedings of the 2013 7th IEEE GCC Conference and Exhibition (GCC), Doha, Qatar, 17–20 November 2013; pp. 114–118.
7. Varol, E.; Gaonkar, B.; Erus, G.; Schultz, R.; Davatzikos, C. Feature ranking based nested support vector machine ensemble
for medical image classification. In Proceedings of the 2012 9th IEEE International Symposium on Biomedical Imaging (ISBI),
Barcelona, Spain, 2–5 May 2012; pp. 146–149.
8. Shen, D.; Wu, G.; Suk, H.-I. Deep Learning in Medical Image Analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [CrossRef]
[PubMed]
9. Alom, M.Z.; Taha, T.M.; Yakopcic, C.; Westberg, S.; Sidike, P.; Nasrin, M.S.; Van Esesn, B.C.; Awwal, A.A.; Asari, V.K. The history
began from alexnet: A comprehensive survey on deep learning approaches. arXiv 2018, arXiv:180301164.
10. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Li, F.-F. ImageNet: A large-scale hierarchical image database. In Proceedings of the
2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255.
11. Liang, M.; Tang, W.; Xu, D.M.; Jirapatnakul, A.C.; Reeves, A.P.; Henschke, C.I.; Yankelevitz, D. Low-Dose CT Screening for Lung
Cancer: Computer-aided Detection of Missed Lung Cancers. Radiology 2016, 281, 279–288. [CrossRef] [PubMed]
12. Setio, A.A.; Ciompi, F.; Litjens, G.; Gerke, P.; Jacobs, C.; van Riel, S.J.; Wille, M.M.; Naqibullah, M.; Sanchez, C.I.; van Ginneken, B.
Pulmonary Nodule Detection in CT Images: False Positive Reduction Using Multi-View Convolutional Networks. IEEE Trans.
Med. Imaging 2016, 35, 1160–1169. [CrossRef]
13. Kang, G.; Liu, K.; Hou, B.; Zhang, N. 3D multi-view convolutional neural networks for lung nodule classification. PLoS ONE
2017, 12, e0188290. [CrossRef]
14. Qayyum, A.; Anwar, S.M.; Awais, M.; Majid, M. Medical image retrieval using deep convolutional neural network. Neurocomputing
2017, 266, 8–20. [CrossRef]
15. Ayyachamy, S.; Alex, V.; Khened, M.; Krishnamurthi, G. Medical image retrieval using Resnet-18. In Proceedings of the SPIE
Medical Imaging, San Diego, CA, USA, 16–21 February 2019; Volume 1095410, pp. 233–241.
16. Remedios, S.; Roy, S.; Pham, D.L.; Butman, J.A. Classifying magnetic resonance image modalities with convolutional neural
networks. In Proceedings of the SPIE Medical Imaging, Houston, TX, USA, 10–15 February 2018; Volume 10575, pp. 558–563.
[CrossRef]
17. van der Voort, S.R.; Smits, M.; Klein, S. DeepDicomSort: An Automatic Sorting Algorithm for Brain Magnetic Resonance Imaging
Data. Neuroinformatics 2021, 19, 159–184. [CrossRef]
18. Scheirer, W.J.; de Rezende Rocha, A.; Sapkota, A.; Boult, T.E. Toward open set recognition. IEEE Trans. Pattern Anal. Mach. Intell.
2012, 35, 1757–1772. [CrossRef]
19. Scarpace, L.; Mikkelsen, L.; Cha, T.; Rao, S.; Tekchandani, S.; Gutman, S.; Pierce, D. Radiology Data from the Cancer Genome
Atlas Glioblastoma Multiforme [TCGA-GBM] Collection. The Cancer Imaging Archive. 2016. Available online: https://www.
cancerimagingarchive.net (accessed on 2 December 2019).
20. Ellingson, B.M.; Bendszus, M.; Boxerman, J.; Barboriak, D.; Erickson, B.J.; Smits, M.; Nelson, S.J.; Gerstner, E.; Alexander, B.;
Goldmacher, G.; et al. Consensus recommendations for a standardized Brain Tumor Imaging Protocol in clinical trials. Neuro.
Oncol. 2015, 17, 1188–1198.
Cancers 2023, 15, 1820 16 of 16
21. Combs, S.E.; Kieser, M.; Rieken, S.; Habermehl, D.; Jäkel, O.; Haberer, T.; Nikoghosyan, A.; Haselmann, R.; Unterberg, A.;
Wick, W.; et al. Randomized phase II study evaluating a carbon ion boost applied after combined radiochemotherapy with
te-mozolomide versus a proton boost after radiochemotherapy with temozolomide in patients with primary glioblastoma: The
CLEOPATRA Trial. BMC Cancer 2010, 10, 478. [CrossRef]
22. Combs, S.E.; Burkholder, I.; Edler, L.; Rieken, S.; Habermehl, D.; Jäkel, O.; Haberer, T.; Haselmann, R.; Unterberg, A.; Wick, W.;
et al. Randomized phase I/II study to evaluate carbon ion radiotherapy versus fractionated stereotactic radiotherapy in patients
with recurrent or progressive gliomas: The CINDERELLA trial. BMC Cancer 2010, 10, 533. [CrossRef]
23. Niyazi, M.; Adeberg, S.; Kaul, D.; Boulesteix, A.-L.; Bougatf, N.; Fleischmann, D.F.; Grün, A.; Krämer, A.; Rödel, C.; Eckert, F.;
et al. Independent validation of a new reirradiation risk score (RRRS) for glioma patients predicting post-recurrence survival: A
multicenter DKTK/ROG analysis. Radiother. Oncol. 2018, 127, 121–127. [CrossRef]
24. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:14091556.
25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society
Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778.
26. Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 Bias Correction.
IEEE Trans. Med. Imaging 2010, 29, 1310–1320. [CrossRef]
27. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch:
An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8026–8037.
28. Pérez-García, F.; Sparks, R.; Ourselin, S. TorchIO: A Python library for efficient loading, preprocessing, augmentation and
patch-based sampling of medical images in deep learning. Comput. Methods Programs Biomed. 2021, 208, 106236. [CrossRef]
29. Wallis, S. Binomial Confidence Intervals and Contingency Tests: Mathematical Fundamentals and the Evaluation of Alter-native
Methods. J. Quant. Linguist. 2013, 20, 178–208. [CrossRef]
30. Lin, D.Y.; Wei, L.J. The robust inference for the Cox proportional hazards model. J. Am. Stat. Assoc. 1989, 84, 1074–1078. [CrossRef]
31. Van Griethuysen, J.J.; Fedorov, A.; Parmar, C.; Hosny, A.; Aucoin, N.; Narayan, V.; Beets-Tan, R.G.; Fillion-Robin, J.C.; Pieper, S.;
Aerts, H.J. Computational radiomics system to decode the radiographic phenotype. Cancer Res. 2017, 77, e104–e107. [CrossRef]
32. Sanders, H.; Saxe, J. Garbage in, Garbage out: How Purportedly Great ML Models can be Screwed up by Bad Data. In Proceedings
of the Black Hat, Las Vegas, NV, USA, 22–27 July 2017. Available online: https://www.blackhat.com/us-17/call-for-papers.
html#review (accessed on 14 March 2023).
33. Sforazzini, F.; Salome, P.; Kudak, A.; Ulrich, M.; Bougatf, N.; Debus, J.; Knoll, M.; Abdollahi, A. pyCuRT: An Automated Data
Cura-tion Workflow for Radiotherapy Big Data Analysis using Pythons’ NyPipe. Int. J. Radiat. Oncol. Biol. Phys. 2020, 108, e772.
[CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.