0% found this document useful (0 votes)
12 views14 pages

Fracture Detection in Pediatric Wrist Trauma X-Ray

Uploaded by

paulcedricramos
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
12 views14 pages

Fracture Detection in Pediatric Wrist Trauma X-Ray

Uploaded by

paulcedricramos
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 14

www.nature.

com/scientificreports

OPEN Fracture detection in pediatric wrist


trauma X‑ray images using YOLOv8
algorithm
Rui‑Yang Ju 1 & Weiming Cai 2*

Hospital emergency departments frequently receive lots of bone fracture cases, with pediatric wrist
trauma fracture accounting for the majority of them. Before pediatric surgeons perform surgery, they
need to ask patients how the fracture occurred and analyze the fracture situation by interpreting
X-ray images. The interpretation of X-ray images often requires a combination of techniques from
radiologists and surgeons, which requires time-consuming specialized training. With the rise of deep
learning in the field of computer vision, network models applying for fracture detection has become an
important research topic. In this paper, we use data augmentation to improve the model performance
of YOLOv8 algorithm (the latest version of You Only Look Once) on a pediatric wrist trauma X-ray
dataset (GRAZPEDWRI-DX), which is a public dataset. The experimental results show that our model
has reached the state-of-the-art (SOTA) mean average precision (mAP 50). Specifically, mAP 50 of our
model is 0.638, which is significantly higher than the 0.634 and 0.636 of the improved YOLOv7 and
original YOLOv8 models. To enable surgeons to use our model for fracture detection on pediatric wrist
trauma X-ray images, we have designed the application “Fracture Detection Using YOLOv8 App” to
assist surgeons in diagnosing fractures, reducing the probability of error analysis, and providing more
useful information for surgery.

In hospital emergency rooms, radiologists are often asked to examine patients with fractures in various parts of
the body, such as the wrist and arm. Fractures can generally be classified as open or closed, with open fractures
occurring when the bone pierces the skin, and closed fractures occurring when the skin remains intact despite
the broken bone. Before performing surgery, the surgeon must inquire about the medical history of the patients
and conduct a thorough examination to diagnose fracture. In recent medical imaging, three types of devices,
including X-ray, Magnetic Resonance Imaging (MRI), and Computed Tomography (CT), are commonly used
to diagnose f­ racture1. And X-ray is the most widely used device due to its cost-effectiveness.
Fractures of the distal radius and ulna account for the majority of wrist trauma in pediatric p ­ atients2,3. In
prestigious hospitals of developed areas, there are many experienced radiologists who are capable of correctly
analyzing X-ray images; while in some small hospitals of underdeveloped regions, there are only young and inex-
perienced surgeons who may be unable to correctly interpret X-ray images. Therefore, a shortage of radiologists
would seriously jeopardize timely patient c­ are4,5. Specifically, some hospitals in Africa have even limited access
to specialist r­ eports6, which badly affects the probaility of the sucess of surgery. According to the s­ urvey7,8, the
percentage of X-ray images misinterpreted have reached 26%.
With the advancement of deep learning, neural network models have been introduced in medical image
­processing9–12. In recent years, researchers have started to apply object detection ­models13–15 to fracture
­detection16–19, which is a popular research topic in computer vision (CV).
Deep learning methods in the field of object detection are divided into two-stage and one-stage algorithms.
Two-stage algorithm models such as R-CNN13 and its improved m ­ odels20–26 generate location and class prob-
abilities in two stages. Whereas one-stage algorithm models directly produce the location and class probabilities
of objects, resulting in the improvement of the model inference speed. In addition to the classical one-stage
algorithm models, such as S­ SD27, ­RetinaNet28, ­CornerNet29, ­CenterNet30, and C ­ entripetalNet31, You Only Look
Once (YOLO) series algorithm m ­ odels32–34 are preferred for real-time a­ pplications35 due to the good balance
between the model accuracy and inference speed.
In this paper, we first use YOLOv8 ­algorithm36 to train models of different sizes on the GRAZPEDWRI-DX37
dataset. After evaluation of the model performances of YOLOv8, we train the models by using data augmentation

1
Graduate Institute of Networking and Multimedia, National Taiwan University, Taipei City 106335,
Taiwan. 2Department of Hand and Foot Surgery, Jingjiang People’s Hospital, Jingjiang City 214500, China. *email:
[email protected]

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 1


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol.:(0123456789)
www.nature.com/scientificreports/

to detect wrist fractures in children. We compare YOLOv8 models using our training method with YOLOv7
and its improved models, and the experimental results demonstrate that our models have the highest the mean
average precision (mAP 50) value.
The contributions of this paper are summarized as follows:

• We use data augmentation to improve the model performance of YOLOv8 model. The experimental results
show that the mean average precision of YOLOv8 model using our training method for fracture detection
on the GRAZPEDWRI-DX dataset reaches SOTA value.
• This work develops an application to detect wrist fracture in children, which aims to help pediatric surgeons
interpret X-ray images without the assistance of the radiologist, and reduce the probability of X-ray image
analysis errors.

This paper is structured as follows: Section "Related work" describes the deep learning methods for detecting frac-
ture, and describes the application of YOLOv5 model in medical image processing. Section "Proposed method"
introduces the whole process of training and the architecture of our model. Section "Experiments" presents the
improved performance of YOLOv8 model using our training method compared with YOLOv7 and its improved
models. Section "Application" describes our proposed application to assist pediatric surgeons in analyzing X-ray
images. Finally, Sect. "Conclusions and future work" discusses the conclusions and future work of this paper.

Related Work
In recent years, neural networks have been widely utilized in image data for fracture detection. Guan et al.38
achieved the average precision of 82.1% on 3,842 thigh fracture X-ray images using the Dilated Convolutional
Feature Pyramid Network (DCFPN). Wang et al.39 employed a novel R-CNN13 network ParalleNet as the back-
bone network for fracture detection on 3842 thigh fracture X-ray images. In addition to thigh fracture detection,
about arm fracture detection, Guan et al.40 used R-CNN for detection on Musculoskeletal-Radiograph (MURA)
­dataset41 and obtained an average precision of 62.04%. Ma and L ­ uo43 used Faster R-CNN21 for fracture detec-
tion on a part of 1052 bone images of the dataset and the proposed CrackNet model for fracture classification
on the whole dataset. Wu et al.42 proposed Feature Ambiguity Mitigate Operator (FAMO) model based on
­ResNeXt10147 and F ­ PN48 for bone fracture detection on 9040 radiographs of various body parts. Qi et al.49 uti-
lized Fast R-CNN20 with ­ResNet5050 as the backbone network to detect nine different types of fractures on 2,333
fracture X-ray images. Xue et al.44 utilized the Faster R-CNN model for hand fracture detection on 3067 hand
trauma X-ray images, achieving an average precision of 70.0%. Sha et al.45,46 used ­YOLOv251 and Faster R-CNN21
models for fracture detection on 5134 CT images of spine fractures respectively. Experiments showed that the
average precision of YOLOv2 reached 75.3%, which was higher than 73.3% of Faster R-CNN, and inference
time of YOLOv2 for each CT image is 27 ms, which is much faster than 381 ms of Faster R-CNN. From Table 1,
it can be seen that even though most of the works using R-CNN series models have shown excellent results, the
inference speed is not satisfactory.
YOLO series ­models32–34 offer a balance of performance in terms of the model accuracy and inference speed,
which is suitable for mobile devices in real-time X-ray images detection. Hržić et al.52 proposed a machine
learning model based on YOLOv4 method to help radiologists diagnose fractures and demonstrated that the
AUC-ROC (area under the receiver operator characteristic curve) value of YOLO 512 Anchor model-AI was
significantly higher than that of radiologists. YOLOv5 m ­ odel53, which was proposed by Ultralytics in 2021, has
been deployed on mobile phones as the “iDetection” application. On this basis, Yuan et al.54 employed external
attention and 3D feature fusion techniques in YOLOv5 model to detect skull fractures in CT images. Warin et al.55
used YOLOv5 model to detect maxillofacial fractures in 3407 maxillofacial bone CT images, and classified the
fracture conditions into frontal, midfacial, mandibular fractures and no fracture. Rib fractures are a precursor
injury to physical abuse in children, and chest X-ray (CXR) images are preferred for effective diagnosis of rib
fracture conditions because of their convenience and low radiation dose. Tsai et al.56 used data augmentation
with YOLOv5 model to detect rib fractures in CXR images. And Burkow et al.57 applied YOLOv5 model to detect

Author Task Model Dataset mAPval 50


38
Guan et al. Thigh fracture detection DCFPN 3842 thigh fracture X-ray radiographs 0.821
Wang et al.39 Thigh fracture detection R-CNN 3842 thigh fracture X-ray radiographs 0.878
Guan et al.40 Arm fracture detection R-CNN Musculoskeletal-Radiograph (MURA)41 0.620
Wu et al.42 Bone fracture detection FAMO 9040 radiographs of various body parts 0.774
­ uo43
Ma and L Bone fracture detection Faster R-CNN 1052 bone x-ray images 0.884
Xue et al.44 Hand fracture detection Faster R-CNN 3067 hand trauma x-ray images 0.700
Sha et al.45 Spine fracture detection Faster R-CNN 5134 spine fractures CT images 0.733
Sha et al.46 Spine fracture detection YOLOv2 5134 spine fractures CT images 0.753

Table 1.  Experimental results of other studies on fracture detection in various parts of the body based on deep
learning method.

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 2


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol:.(1234567890)
www.nature.com/scientificreports/

rib fractures in 704 pediatric CXR images, the model obtained the F2 score value of 0.58. To identify and detect
mandibular fractures in panoramic radiographs, Warin et al.58 used convolutional neural networks (CNNs) and
YOLOv5 model to implement it. Fatima et al.59 used YOLOv5 model to localize vertebrae, which is important
for detecting spinal deformities and fractures, and obtained an average precision of 0.94 at an IoU (Intersection
over Union) threshold of 0.5. Moreover, Mushtaq et al.60 applied YOLOv5 model to localize the lumbar spine
and obtained an average precision value of 0.975. Nevertheless, relatively few researches have been reported on
pediatric wrist fracture detection using YOLOv5 model. While YOLOv8 was proposed by Ultralytics in 2023,
we use this algorithm to train the model for the first time in pediatric wrist fracture detection.

Proposed method
In this section, we introduce the process of the model training, validation and testing on the dataset, the archi-
tecture of YOLOv8 model, and the data augmentation technique employed during training. Figure 1 illustrates
the flowchart depicting the model training process and performance evaluation. We randomly divide the 20,327
X-ray images of the GRAZPEDWRI-DX dataset into the training, validation, and test set, where the training set
is expanded to 28,408 X-ray images by data augmentation from the original 14,204 X-ray images. We design our
model according to YOLOv8 algorithm, and the architecture of YOLOv8 algorithm is shown in Fig. 2.

Data augmentation
During the model training process, data augmentation is employed in this work to extend the dataset. Specifically,
we adjust the contrast and brightness of the original X-ray image to enhance the visibility of bone-anomaly. This
is achieved using the addWeighted function available in OpenCV (Open Source Computer Vision Library). The
equation is presented below:
Output = Input1 × α + Input2 × β + γ , (1)
where Input1 and Input2 are the two input images of the same size respectively, α represents the weight assigned
to the first input image, β denotes the weight assigned to the second input image, and γ represents the scalar
value added to each sum. Since our purpose is to adjust the contrast and brightness of the original input image,
we take the same image as Input1 and Input2 respectively and set β to 0. The value of α and γ represent the pro-
portion of the contrast and the brightness of the image respectively. The image after adjusting the contrast and
brightness is shown in Fig. 3. After comparing different settings, we finally decided to set α to 1.2 and γ to 30 to
avoid the output image being too bright.

Figure 1.  Flowchart of the model training, validation and testing on the dataset. The extended training set is
used to double the number of X-ray images by data augmentation.

Figure 2.  The architecture of YOLOv8 algorithm, which is divided into four parts, including backbone, neck,
head, and loss.

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 3


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol.:(0123456789)
www.nature.com/scientificreports/

Figure 3.  Examples of pediatric wrist X-ray images using data augmentation. (a) the original images, (b) the
adjusted images.

Model architecture
Our model architecture consists of backbone, neck, and head, as shown in Fig. 4. In the following subsections,
we introduce the design concepts of each part of the model architecture, and the modules of different parts.

Backbone
The backbone of the model uses Cross Stage Partial (CSP)61 architecture to split the feature map into two parts.
The first part uses convolution operations, and the second part is concatenated with the output of the previous
part. The CSP architecture improves the learning ability of the CNNs and reduces the computational cost of the
model.
YOLOv836 introduces C2f module by combining the C3 module and the concept of ELAN from ­YOLOv732,
which allows the model to obtain richer gradient flow information. The C3 module consists of 3 ConvModule
and n DarknetBottleNeck, and the C2f module consists of 2 ConvModule and n DarknetBottleNeck connected

Figure 4.  Detailed illustration of YOLOv8 model architecture. The Backbone, Neck, and Head are the three
parts of our model, and C2f, ConvModule, DarknetBottleneck, and SPPF are modules.

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 4


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol:.(1234567890)
www.nature.com/scientificreports/

through Split and Concat, as illustrated in Figure 4, where the ConvModule consists of Conv-BN-SiLU, and n is
the number of the bottleneck. Unlike ­YOLOv553, we use the C2f module instead of the C3 module.
Furthermore, we reduce the number of blocks in each stage compared to YOLOv5 to further reduce the com-
putational cost. Specifically, our model reduces the number of blocks to 3,6,6,3 in Stage 1 to Stage 4, respectively.
Additionally, we adopt the Spatial Pyramid Pooling - Fast (SPPF) module in Stage 4, which is an improvement
from Spatial Pyramid Pooling (SPP)62 to improve the inference speed of the model. These modifications lead to
our model with a better learning ability and shorter inference time.

Neck
Generally, deeper networks obtain more feature information, resulting in better dense prediction. However,
excessively deep networks reduce the location information of the object, and too many convolution operations
will lead to information loss for small objects. Therefore, it is necessary to use Feature Pyramid Network (FPN)48
and Path Aggregation Network (PAN)63 architectures for multi-scale feature fusion. As illustrated in Fig. 4, the
Neck part of our model architecture uses multi-scale feature fusion to combine features from different layers of
the network. The upper layers acquire more information due to the additional network layers, whereas the lower
layers preserve location information due to fewer convolution layers.
Inspired by YOLOv5, where FPN upsamples from top to bottom to increase the amount of feature informa-
tion in the bottom feature map; and PAN downsamples from bottom to top to obtain more the top feature map
information. These two feature outputs are merged to ensure precise predictions for images of various sizes. We
adopt FP-PAN (Feature Pyramid-Path Aggregation Network) in our model, and delete convolution operations
in upsampling to reduce the computational cost.

Head
Different from YOLOv5 model utilizing a coupled head, we use a decoupled ­head33, where the classification and
detection heads are separated. Figure 4 illustrates that our model deletes the objectness branch and only retains
the classification and regression branches. Anchor-Base employes a large number of anchors in the image to
determine the four offsets of the regression object from the anchors. It adjusts the precise object location using
the corresponding anchors and offsets. In contrast, we adopt Anchor-Free64, which identifies the center of the
object and estimates the distance between the center and the bounding box.

Loss
For positive and negative sample assignment, the Task Aligned Assigner of Task-aligned One-stage Object
Detection (TOOD)65 is used in our model training to select positive samples based on the weighted scores of
classification and regression, as shown in Eq. 2 below:

t = s α × uβ , (2)
where s is the predicted score corresponding to the labeled class, and u is the IoU of the prediction and the
ground truth bounding box.
In addition, our model has classification and regression branches, where the classification branch uses Binary
Cross-Entropy (BCE) Loss, and the equation is shown below:

(3)
   
Lossn = −w yn log xn + 1 − yn log (1 − xn ) ,

where w is the weight, yn is the labeled value, and xn is the predicted value of the model.
The regression branch uses Distribute Focal Loss (DFL)66 and Complete IoU (CIoU) L ­ oss67, where DFL is
used to expand the probability of the value around the object y. Its equation is shown as follows:
DFL(Sn , Sn+1 ) = −((yn+1 − y) log(Sn ) + (y − yn ) log(Sn+1 )), (4)
where the equations of Sn and Sn+1 are shown below:
yn+1 − y y − yn
Sn =
yn+1 − yn
, Sn+1 =
yn+1 − yn
. (5)

CIoU Loss adds an influence factor to Distance IoU (DIoU) ­Loss68 by considering the aspect ratio of the predic-
tion and the ground truth bounding box. The equation is shown below:
Distance22 v2
CIoULoss = 1 − IoU + + , (6)
DistanceC2 (1 − IoU) + ν

where ν is the parameter that measures the consistency of the aspect ratio, defined as follows:
4 w gt wp 2
ν= (arctan − arctan ) , (7)
π2 hgt hp
where w is the weight of the bounding box, and h is the height of the bounding box.

Ethics approval
This research does not involve human participants and/or animals.

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 5


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol.:(0123456789)
www.nature.com/scientificreports/

Experiments
Dataset
Medical University of Graz provides a public dataset named GRAZPEDWRI-DX37, which consists of 20,327
X-ray images of wrist trauma in children. These images were collected from 6,091 patients between 2008 and
2018 by multiple pediatric radiologists at the Department of Pediatric Surgery of the University Hospital Graz.
The images are annotated in 9 different classes by placing bounding boxes on them.
To perform the experiments shown in Table 5 and Table 6, we divide the GRAZPEDWRI-DX dataset ran-
domly into three sets: training set, validation set, and test set. The sizes of these sets are approximately 70%, 20%,
and 10% of the original dataset, respectively. Specifically, our training set consists of 14,204 images (69.88%), our
validation set consists of 4,094 images (20.14%), and our test set consists of 2029 images (9.98%). The code for
splitting the dataset can be found on our GitHub. We also provide csv files of training, validation and test data
on our GitHub, but it should be noted that each split is random and therefore not reproducible.

Evaluation metric
Intersection over Union (IoU)
Intersection over Union (IoU) is a classical metric for evaluating the performance of the model for object detec-
tion. It calculates the ratio of the overlap and union between the generated candidate bounding box and the
ground truth bounding box, which measures the intersection of these two bounding boxes. The IoU is repre-
sented by the following equation:
area(C) ∩ area(G)
IoU = , (8)
area(C) ∪ area(G)
where C represents the generated candidate bounding box, and G represents the ground truth bounding box
containing the object. The performance of the model improves as the IoU value increases, with higher IoU values
indicating less difference between the generated candidate and ground truth bounding boxes.

Precision‑recall curve
Precision-Recall Curve (P-R Curve)69 is a curve with recall as the x-axis and precision as the y-axis. Each point
represents a different threshold value, and all points are connected as a curve. The recall (R) and precision (P)
are calculated according to the following equations:
TP TP
Recall = , Precision = , (9)
TP + FN TP + FP
where True Positive (TP ) denotes the prediction result as a positive class and is judged to be true; False Positive
( FP ) denotes the prediction result as a positive class but is judged to be false, and False Negative ( FN ) denotes
the prediction result as a negative class but is judged to be false.

F1‑score
The F-score is a commonly used metric to evaluate the model accuracy, providing a balanced measure of per-
formance by incorporating both precision and recall. The F-score equation is as follows:
1 + β 2 × Precision × Recall
 
F-score = (10)
β 2 × Precision + Recall
When β = 1, the F1-score is determined by the harmonic mean of precision and recall, and its equation is as
follows:
2 × Precision × Recall 2TP
F1 = = (11)
Precision + Recall 2TP + FP + FN

Experiment setup
During the model training process, we utilize pre-trained YOLOv8 model from the MS COCO (Microsoft Com-
mon Objects in Context) val2017 d ­ ataset72. The research reports provided by U
­ ltralytics36,53 suggests that YOLOv5
training requires 300 epochs, while training YOLOv8 requires 500 epochs. Since we use pre-trained model, we
initially set the total number of epochs to 200 with a patience of 50, which indicate that the training would end
early if no observable improvement is noticed after waiting for 50 epochs. In the experiment comparing the
effect of the optimizer on the model performance, we notice that the best epcoh of all the models is within 100,
as shown in Table 4, mostly concentrated between 50 and 70 epochs. Therefore, to save computing resources,
we adjust the number of epochs for our model training to 100.
As the ­suggestion36 of Glenn, for model training hyperparameters, the ­Adam73 optimizer is more suitable for
small custom datasets, while the ­SGD74 optimizer perform better on larger datasets. To prove the above conclu-
sion, we train YOLOv8 algorithm models using the Adam and SGD optimizers, respectively, and compare the
effects on the model performance. The comparison results are shown in Table 4.
For the experiments, we choose the SGD optimizer with an initial learning rate of 1×10−2, a weight decay of
5×10−4 , and a momentum of 0.937 during our model training. We set the input image size to 640 and 1024 for
training on a single GPU GeForce RTX 3080Ti 12GB with a batch size of 16. We train the model using Python 3.8

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 6


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol:.(1234567890)
www.nature.com/scientificreports/

and PyTorch 1.8.2, and recommend readers to use Python 3.7 or higher and PyTorch 1.7 or higher for training.
It is noteworthy that due to GPU memory limitations, we choose 3 worker threads to load data on GPU GeForce
RTX 3080Ti 12GB when training our model. Therefore, using GPUs with larger memory and more computing
power can effectively increase the speed of model training.

Ablation study
In order to demonstrate the positive effect of our training method on the performance of YOLOv8 model, we
conduct an ablation study on YOLOv8s model by calculating each evaluation metric for each class, as shown in
Table 2. Among all classes, YOLOv8s model has good accuracy in detecting fracture, metal and text, with mAP 50
of each above 0.9. On the opposite, the detection ability of bone-anomaly is poor, with mAP 50 of 0.11. Therefore,
we increase the contrast and brightness of X-ray images to make bone-anomaly easier to detect. Table 3 presents
the predictions of YOLOv8s model using our training method for each class. Compared with YOLOv8s model,
the mAP value predicted by the model using our training method for bone-anomaly increased from 0.11 to 0.169,

mAPval mAPval
Class Boxes Instances Precision Recall 50 50-95
All 47435 9613 0.674 0.605 0.623 0.395
Boneanomaly 276 53 0.505 0.094 0.110 0.035
Bonelesion 45 8 0.629 0.250 0.416 0.212
Fracture 18090 3740 0.885 0.903 0.947 0.572
Metal 818 168 0.878 0.899 0.920 0.768
Periostealreaction 3453 697 0.645 0.684 0.689 0.357
Pronatorsign 567 104 0.561 0.713 0.611 0.338
Softtissue 464 89 0.324 0.315 0.251 0.125
text 23722 4754 0.961 0.984 0.991 0.750

Table 2.  Validation results of YOLOv8 for each class on the GRAZPEDWRI-DX dataset when the input image
size is 1024.

mAPval mAPval
Class Boxes Instances Precision Recall 50 50-95
All 47435 9613 0.694 0.592 0.631 0.402
Boneanomaly 276 53 0.510 0.151 0.169 0.076
Bonelesion 45 8 0.658 0.243 0.414 0.213
Fracture 18090 3740 0.899 0.896 0.947 0.569
Metal 818 168 0.898 0.890 0.924 0.780
Periostealreaction 3453 697 0.721 0.654 0.700 0.359
Pronatorsign 567 104 0.534 0.683 0.611 0.342
Softtissue 464 89 0.367 0.236 0.241 0.120
text 23722 4754 0.961 0.981 0.991 0.754

Table 3.  Validation results of our model for each class on the GRAZPEDWRI-DX dataset when the input
image size is 1024.

mAPval mAPval Speed GPU


Model Size Optimizer Best Epoch 50 50-95 RTX 3080Ti (ms)
YOLOv8s 640 SGD 56 0.611 0.389 4.4
YOLOv8s 640 Adam 57 0.604 0.383 4.3
YOLOv8s 1024 SGD 36 0.623 0.395 5.4
YOLOv8s 1024 Adam 47 0.625 0.399 4.9
YOLOv8m 640 SGD 52 0.621 0.396 4.9
YOLOv8m 640 Adam 62 0.621 0.403 5.5
YOLOv8m 1024 SGD 35 0.624 0.402 9.9
YOLOv8m 1024 Adam 70 0.626 0.401 10.0

Table 4.  Model performance comparison of YOLOv8 models using SGD and Adam optimizers. For training
with the SGD optimizer, the initial learning rate is 1×10−2; for training with the Adam optimizer, the initial
learning rate is 1×10−3.

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 7


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol.:(0123456789)
www.nature.com/scientificreports/

mAPval mAPval Speed CPU Speed GPU


Model 50 50–95 Intel Core i5 (ms) RTX 3080Ti (ms) PARAMS (M) FLOPs (B)
YOLOv5n 0.589 0.339 \ 2.8 1.77 4.2
YOLOv8n 0.601 0.374 67.4 2.9 3.01 8.1
Ours 0.605 0.379 111.3 3.4 3.01 8.2
YOLOv5s 0.601 0.357 \ 3.3 7.03 15.8
YOLOv8s 0.604 0.383 191.5 4.3 11.13 28.5
Ours 0.612 0.392 285.1 4.9 11.13 28.7
YOLOv5m 0.613 0.371 \ 4.0 20.89 48.0
YOLOv8m 0.621 0.403 536.4 5.5 25.84 78.7
Ours 0.629 0.404 685.9 5.1 25.84 78.7
YOLOv5l 0.620 0.379 \ 5.6 46.15 107.8
YOLOv8l 0.624 0.403 1006.3 7.4 43.61 164.9
Ours 0.637 0.406 1370.8 7.2 43.61 164.9

Table 5.  Quantitative comparison of fracture detection when the input image size is 640. Speed means the
total time of validate per image, and the total time includes the preprocessing, inference, and post-processing
time.

mAPval mAPval Speed CPU Speed GPU


Model 50 50–95 Intel Core i5 (ms) RTX 3080Ti (ms) PARAMS (M) FLOPs (B)
YOLOv5n 0.600 0.347 \ 3.2 1.77 4.2
YOLOv8n 0.605 0.387 212.1 3.3 3.01 8.1
Ours 0.608 0.391 260.4 4.4 3.01 8.1
YOLOv5s 0.622 0.371 \ 4.4 7.03 15.8
YOLOv8s 0.625 0.399 519.5 4.9 11.13 28.5
Ours 0.631 0.402 717.1 6.2 11.13 28.5
YOLOv5m 0.624 0.380 \ 7.1 20.89 48.0
YOLOv8m 0.626 0.401 1521.5 10.0 25.84 78.7
Ours 0.635 0.411 1724.4 9.4 25.85 78.7
YOLOv5l 0.626 0.378 \ 11.3 46.15 107.8
YOLOv8l 0.636 0.404 2671.1 15.1 43.61 164.9
Ours 0.638 0.415 3864.5 13.6 43.61 164.9

Table 6.  Quantitative comparison of fracture detection when the input image size is 1024. Speed means the
total time of validate per image, and the total time includes the preprocessing, inference, and post-processing
time.

an increase of 53.6%. Figure 5 also shows that our model has a better performance in detecting bone-anomaly,
which enables the improvement of the overall model performance. From the ablation study presented above, we
demonstrate that the model performance can be improved by using our training method (data augmentation).
In addition to the data enhancement, researchers can also improve model performance by adding modules such
as the Convolutional Block Attention Module (CBAM)70.

Experimental results
Before training our model, in order to choose an optimizer that has a more positive effect on the model perfor-
mance, we compare the performance of models trained with the S­ GD74 optimizer and the A ­ dam73 optimizer. As
shown in Table 4, using the SGD optimizer to train the model requires less epochs of weight updates. Specifically,
for YOLOv8m model with an input image size of 1024, the model trained with the SGD optimizer achieves the
best performance at the 35th epoch, while the best performance of the model trained with the Adam optimizer
is at the 70th epoch. In terms of mAP and inference time, there is not much difference in the performance of
the models trained with the two optimizers. Specifically, when the input image size is 640, the mAP value of
YOLOv8s model trained with the SGD optimizer is 0.007 higher than that of the model trained with the Adam
optimizer, while the inference time is 0.1ms slower. Therefore, according to the above experimental results and
the suggestion by G­ lenn36,53, for YOLOv8 model training on a training set of 14,204 X-ray images, we choose
the Adam optimizer. However, after using data augmentation, the number of X-ray images in the training set
extend to 28,408, so we switch to the SGD optimizer to train our model.
After using data augmentation, our models have a better mAP value than that of YOLOv8 model, as shown
in Table 5 and Table 6. Specifically, when the input image size is 640, compared with YOLOv8m model and

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 8


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol:.(1234567890)
www.nature.com/scientificreports/

Figure 5.  Detailed illustration of the validation at the input image size of 1024, (a) is our model, and (b) is
YOLOv8 model.

YOLOv8l model, the mAP 50 of our model improves from 0.621 to 0.629, and from 0.623 to 0.637, respectively.
Although the inference time on the CPU is increased from 536.4 ms and 1006.3 ms to 685.9 ms and 1370.8 ms,
respectively, the number of parameters and FLOPs are the same, which means that our model can be deployed
on the same computing power platform. In addition, we compare the performance of our model with that of
YOLOv7 and its improved models. As shown in Table 7, the mAP value of our model is higher than those of
­YOLOv732, YOLOv7 with Convolutional Block Attention Module (CBAM)70 and YOLOv7 model with Global
Attention Mechanism (GAM)71, which demonstrates that our model has obtained SOTA performance.
This paper aims to design a pediatric wrist fracture detection application, so we use our model for fracture
detection. Figure 6 shows the results of manual annotation by the radiologist and the results predicted using our
model. These results demonstrate that our model has a good ability to detect fractures in single fracture cases,
but metal puncture and dense multiple fracture situations badly affects the accuracy of prediction.

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 9


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol.:(0123456789)
www.nature.com/scientificreports/

Figure 6.  Examples of pediatric wrist fracture detection on X-ray images. (a) manually labeled images, (b)
predicted images.

mAPval
Model Precision Recall F1 50
YOLOv553 0.682 0.581 0.607 0.626
YOLOv732 0.556 0.582 0.569 0.628
YOLOv732 + ­CBAM70 0.709 0.593 0.646 0.633
YOLOv732 + ­GAM71 0.745 0.574 0.646 0.634
YOLOv836 0.694 0.679 0.623 0.636
Ours 0.734 0.592 0.635 0.638

Table 7.  Evaluation of wrist fracture detection with other state-of-the-art (SOTA) models on the
GRAZPEDWRI-DX dataset.

Application
After completing model training, we utilize a Python library that includes the Qt toolkit, PySide6, to develop
a Graphical User Interface (GUI) application. Specifically, PySide6 is the Qt6-based version of the PySide GUI
library from the Qt Company.
According to the model performance evaluation results in Tables 5 and 6, we choose our model with YOLOv8s
algorithm and the input image size of 1024, to perform fracture detection. Our model is exported to onnx format,
and is applied to the GUI application. Figure 7 depicts the flowchart of the GUI application operation on macOS.
As can be seen from the illustration, our application is named “Fracture Detection Using YOLOv8 App”. Users
can open and predict the images, and save the predictions in this application. In summary, our application is
designed to assist pediatric surgeons in analyzing fractures on pediatric wrist trauma X-ray images.

Conclusions and future work


Ultralytics proposed the latest version of YOLO series (YOLOv8) in 2023. Although there are relatively few
research works on YOLOv8 model for medical image processing, we apply it to fracture detection and use data
augmentation to improve the model performance. We randomly divide the dataset, consisting of 20,327 pediatric
wrist trauma X-ray images from 6091 patients, into training, test, and validation sets to train the model and
evaluate the performance.
Furthermore, we develop an application named “Fracture Detection Using YOLOv8 App” to analyze pediatric
wrist trauma X-ray images for fracture detection. Our application aims to assist pediatric surgeons in interpreting
X-ray images, reduce the probability of misclassification, and provide a better information base for surgery. The
application is currently available for macOS, and in the future, we plan to deploy different sizes of our model in
the application, and extend the application to iOS and Android. This will enable inexperienced pediatric surgeons
in hospitals located in underdeveloped areas to use their mobile devices to analyze pediatric wrist X-ray images.
In addition, we provide the specific steps for training the model and the trained model in our GitHub. If
readers wish to use YOLOv8 model to detect fracture in other parts of the body except the pediatric wrist, they
can use our trained model as the pre-training model, which can greatly improve the performance of the model.

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 10


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol:.(1234567890)
www.nature.com/scientificreports/

Figure 7.  Example of using the application “Fracture Detection with YOLOv8 Application” on macOS
operating system .

Data availability
The datasets analysed during the current study are available at Figshare under https://​doi.​org/​10.​6084/​m9.​figsh​
are.​14825​193.​v2. The implementation code and the trained model for this study can be found on GitHub at
https://​github.​com/​Ruiya​ngJu/​Bone_​Fract​ure_​Detec​tion_​YOLOv8.

Received: 16 May 2023; Accepted: 14 November 2023

References
1. Fractures, health, hopkins medicine. https://​www.​hopki​nsmed​icine.​org/​health/​condi​tions-​and-​disea​ses/​ fract​ures (2021).
2. Hedström, E. M., Svensson, O., Bergström, U. & Michno, P. Epidemiology of fractures in children and adolescents: Increased
incidence over the past decade: A population-based study from northern Sweden. Acta Orthop. 81, 148–153 (2010).
3. Randsborg, P.-H. et al. Fractures in children: Epidemiology and activity-specific fracture rates. JBJS 95, e42 (2013).
4. Burki, T. K. Shortfall of consultant clinical radiologists in the UK. Lancet Oncol. 19, e518 (2018).
5. Rimmer, A. Radiologist shortage leaves patient care at risk, warns royal college. BMJ Br. Med. J. (Online) 359 (2017).
6. Rosman, D. et al. Imaging in the land of 1000 hills: Rwanda radiology country report. J. Glob. Radiol. 1 (2015).
7. Mounts, J., Clingenpeel, J., McGuire, E., Byers, E. & Kireeva, Y. Most frequently missed fractures in the emergency department.
Clin. Pediatr. 50, 183–186 (2011).
8. Erhan, E., Kara, P., Oyar, O. & Unluer, E. Overlooked extremity fractures in the emergency department. Ulus Travma Acil Cerrahi
Derg 19, 25–8 (2013).
9. Adams, S. J., Henderson, R. D., Yi, X. & Babyn, P. Artificial intelligence solutions for analysis of x-ray images. Can. Assoc. Radiol.
J. 72, 60–72 (2021).
10. Tanzi, L. et al. Hierarchical fracture classification of proximal femur x-ray images using a multistage deep learning approach. Eur.
J. Radiol. 133, 109373 (2020).
11. Chung, S. W. et al. Automated detection and classification of the proximal humerus fracture by using deep learning algorithm.
Acta Orthop. 89, 468–473 (2018).
12. Choi, J. W. et al. Using a dual-input convolutional neural network for automated detection of pediatric supracondylar fracture on
conventional radiography. Invest. Radiol. 55, 101–110 (2020).
13. Girshick, R., Donahue, J., Darrell, T. & Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation.
Proc. of the IEEE conference on computer vision and pattern recognition, 580–587 (2014).
14. Ju, R.-Y., Lin, T.-Y., Jian, J.-H., Chiang, J.-S. & Yang, W.-B. Threshnet: An efficient densenet using threshold mechanism to reduce
connections. IEEE Access 10, 82834–82843 (2022).
15. Ju, R.-Y., Lin, T.-Y., Jian, J.-H. & Chiang, J.-S. Efficient convolutional neural networks on raspberry pi for image classification. J.
Real-Time Image Proc. 20, 21 (2023).
16. Gan, K. et al. Artificial intelligence detection of distal radius fractures: A comparison between the convolutional neural network
and professional assessments. Acta Orthop. 90, 394–400 (2019).
17. Kim, D. & MacKinnon, T. Artificial intelligence in fracture detection: Transfer learning from deep convolutional neural networks.
Clin. Radiol. 73, 439–445 (2018).
18. Lindsey, R. et al. Deep neural network improves fracture detection by clinicians. Proc. Natl. Acad. Sci. 115, 11591–11596 (2018).
19. Blüthgen, C. et al. Detection and localization of distal radius fractures: Deep learning system versus radiologists. Eur. J. Radiol.
126, 108925 (2020).
20. Girshick, R. Fast r-cnn. Proc. of the IEEE international conference on computer vision 1440–1448 (2015).

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 11


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol.:(0123456789)
www.nature.com/scientificreports/

21. Ren, S., He, K., Girshick, R. & Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Proc. Advances
in neural information processing systems 28 (2015).
22. Cai, Z. & Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. Proc. of the IEEE conference on computer vision
and pattern recognition 6154–6162 (2018).
23. Lu, X., Li, B., Yue, Y., Li, Q. & Yan, J. Grid r-cnn. Proc. of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
7363–7372 (2019).
24. Pang, J. et al. Libra r-cnn: Towards balanced learning for object detection. Proc. of the IEEE/CVF conference on computer vision
and pattern recognition 821–830 (2019).
25. Zhang, H., Chang, H., Ma, B., Wang, N. & Chen, X. Dynamic r-cnn: Towards high quality object detection via dynamic training.
Proc. Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, Aug 23–28, 2020, Part XV 16, 260–275 (Springer,
2020).
26. He, K., Gkioxari, G., Dollár, P. & Girshick, R. Mask r-cnn. Proc. of the IEEE international conference on computer vision 2961–2969
(2017).
27. Liu, W. et al. Ssd: Single shot multibox detector. Computer Vision–ECCV 2016: 14th European Conference Oct 11–14, 2016, Part I
14, 21–37 (Amsterdam, The Netherlands, Springer, 2016).
28. Lin, T.-Y., Goyal, P., Girshick, R., He, K. & Dollár, P. Focal loss for dense object detection. Proc. of the IEEE international conference
on computer vision 2980–2988 (2017).
29. Law, H. & Deng, J. Cornernet: Detecting objects as paired keypoints. Proc. of the European conference on computer vision (ECCV),
734–750 (2018).
30. Duan, K. et al. Centernet: Keypoint triplets for object detection. Proc. of the IEEE/CVF international conference on computer vision
6569–6578 (2019).
31. Dong, Z. et al. Centripetalnet: Pursuing high-quality keypoint pairs for object detection. Proc. of the IEEE/CVF conference on
computer vision and pattern recognition 10519–10528 (2020).
32. Wang, C.-Y., Bochkovskiy, A. & Liao, H.-Y. M. Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object
detectors. arXiv preprint arXiv:​2207.​02696 (2022).
33. Ge, Z., Liu, S., Wang, F., Li, Z. & Sun, J. Yolox: Exceeding yolo series in 2021. arXiv preprint arXiv:​2107.​08430 (2021).
34. Wang, C.-Y., Yeh, I.-H. & Liao, H.-Y. M. You only learn one representation: Unified network for multiple tasks. arXiv preprint
arXiv:​2105.​04206 (2021).
35. Ju, R.-Y., Chen, C.-C., Chiang, J.-S., Lin, Y.-S. & Chen, W.-H. Resolution enhancement processing on low quality images using
swin transformer based on interval dense connection strategy. Multimed. Tools Appl. 1–17 (2023).
36. Glenn, J. Ultralytics yolov8. https://​github.​com/​ultra​lytics/​ultra​lytics (2023).
37. Nagy, E., Janisch, M., Hržić, F., Sorantin, E. & Tschauner, S. A pediatric wrist trauma x-ray dataset (grazpedwri-dx) for machine
learning. Sci. Data 9, 222 (2022).
38. Guan, B., Yao, J., Zhang, G. & Wang, X. Thigh fracture detection using deep learning method based on new dilated convolutional
feature pyramid network. Pattern Recogn. Lett. 125, 521–526 (2019).
39. Wang, M. et al. Parallelnet: Multiple backbone network for detection tasks on thigh bone fracture. Multimed. Syst. 27, 1091–1100
(2021).
40. Guan, B., Zhang, G., Yao, J., Wang, X. & Wang, M. Arm fracture detection in x-rays based on improved deep convolutional neural
network. Comput. Electric. Eng. 81, 106530 (2020).
41. Rajpurkar, P. et al. Mura dataset: Towards radiologist-level abnormality detection in musculoskeletal radiographs. In Medical
Imaging with Deep Learning (2018).
42. Wu, H.-Z. et al. The feature ambiguity mitigate operator model helps improve bone fracture detection on x-ray radiograph. Sci.
Rep. 11, 1–10 (2021).
43. Ma, Y. & Luo, Y. Bone fracture detection through the two-stage system of crack-sensitive convolutional neural network. Inf. Med.
Unlocked 22, 100452 (2021).
44. Xue, L. et al. Detection and localization of hand fractures based on ga_faster r-cnn. Alex. Eng. J. 60, 4555–4562 (2021).
45. Sha, G., Wu, J. & Yu, B. Detection of spinal fracture lesions based on improved yolov2. Proc. 2020 IEEE International Conference
on Artificial Intelligence and Computer Applications (ICAICA) 235–238 (IEEE, 2020).
46. Sha, G., Wu, J. & Yu, B. Detection of spinal fracture lesions based on improved faster-rcnn. 2020 IEEE International Conference on
Artificial Intelligence and Information Systems (ICAIIS) 29–32 (IEEE, 2020).
47. Xie, S., Girshick, R., Dollár, P., Tu, Z. & He, K. Aggregated residual transformations for deep neural networks. Proc. of the IEEE
conference on computer vision and pattern recognition 1492–1500 (2017).
48. Lin, T.-Y. et al. Feature pyramid networks for object detection. Proc. of the IEEE conference on computer vision and pattern recogni-
tion 2117–2125 (2017).
49. Qi, Y. et al. Ground truth annotated femoral x-ray image dataset and object detection based method for fracture types classifica-
tion. IEEE Access 8, 189436–189444 (2020).
50. He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. Proc. of the IEEE conference on computer vision
and pattern recognition 770–778 (2016).
51. Redmon, J. & Farhadi, A. Yolo9000: better, faster, stronger. Proc. of the IEEE conference on computer vision and pattern recognition
7263–7271 (2017).
52. Hržić, F., Tschauner, S., Sorantin, E. & Štajduhar, I. Fracture recognition in paediatric wrist radiographs: An object detection
approach. Mathematics 10, 2939 (2022).
53. Glenn, J. Ultralytics yolov5. https://​github.​com/​ultra​lytics/​yolov5 (2022).
54. Yuan, G., Liu, G., Wu, X. & Jiang, R. An improved yolov5 for skull fracture detection. Exploration of Novel Intelligent Optimization
Algorithms: 12th International Symposium, ISICA 2021 Nov 20–21, 2021, Revised Selected Papers, 175–188 (Guangzhou, China,
Springer, 2022).
55. Warin, K. et al. Maxillofacial fracture detection and classification in computed tomography images using convolutional neural
network-based models. Sci. Rep. 13, 3434 (2023).
56. Tsai, H.-C. et al. Automatic rib fracture detection and localization from frontal and oblique chest x-rays. Proc. 2022 10th Interna-
tional Conference on Orange Technology (ICOT) 1–4 (IEEE, 2022).
57. Burkow, J. et al. Avalanche decision schemes to improve pediatric rib fracture detection. Proc. Medical Imaging 2022: Computer-
Aided Diagnosis vol. 12033, 597–604 (SPIE, 2022).
58. Warin, K. et al. Assessment of deep convolutional neural network models for mandibular fracture detection in panoramic radio-
graphs. Int. J. Oral Maxillofac. Surg. 51, 1488–1494 (2022).
59. Fatima, J., Mohsan, M., Jameel, A., Akram, M. U. & Muzaffar Syed, A. Vertebrae localization and spine segmentation on radio-
graphic images for feature-based curvature classification for scoliosis. Concurr. Comput. Pract. Exp. 34, e7300 (2022).
60. Mushtaq, M., Akram, M. U., Alghamdi, N. S., Fatima, J. & Masood, R. F. Localization and edge-based segmentation of lumbar
spine vertebrae to identify the deformities using deep learning models. Sensors 22, 1547 (2022).
61. Wang, C.-Y. et al. Cspnet: A new backbone that can enhance learning capability of cnn. Proc. of the IEEE/CVF conference on
computer vision and pattern recognition workshops 390–391 (2020).

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 12


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol:.(1234567890)
www.nature.com/scientificreports/

62. He, K., Zhang, X., Ren, S. & Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans.
Pattern Anal. Mach. Intell. 37, 1904–1916 (2015).
63. Liu, S., Qi, L., Qin, H., Shi, J. & Jia, J. Path aggregation network for instance segmentation. Proc. of the IEEE conference on computer
vision and pattern recognition 8759–8768 (2018).
64. Tian, Z., Shen, C., Chen, H. & He, T. Fcos: A simple and strong anchor-free object detector. IEEE Trans. Pattern Anal. Mach. Intell.
44, 1922–1933 (2020).
65. Feng, C., Zhong, Y., Gao, Y., Scott, M. R. & Huang, W. Tood: Task-aligned one-stage object detection. Proc. 2021 IEEE/CVF Inter-
national Conference on Computer Vision (ICCV) 3490–3499 (IEEE Computer Society, 2021).
66. Li, X. et al. Generalized focal loss: Learning qualified and distributed bounding boxes for dense object detection. Adv. Neural. Inf.
Process. Syst. 33, 21002–21012 (2020).
67. Zheng, Z. et al. Enhancing geometric factors in model learning and inference for object detection and instance segmentation. IEEE
Trans. Cybern. 52, 8574–8586 (2021).
68. Zheng, Z. et al. Distance-iou loss: Faster and better learning for bounding box regression. Proc. of the AAAI conference on artificial
intelligence vol. 34, 12993–13000 (2020).
69. Boyd, K., Eng, K. H. & Page, C. D. Area under the precision-recall curve: point estimates and confidence intervals. Proc. Machine
Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2013 Sept 23–27, 2013, Part III 13, 451–466
(Prague, Czech Republic, Springer, 2013).
70. Woo, S., Park, J., Lee, J.-Y. & Kweon, I. S. Cbam: Convolutional block attention module. Proc. of the European conference on computer
vision (ECCV) 3–19 (2018).
71. Liu, Y., Shao, Z. & Hoffmann, N. Global attention mechanism: Retain information to enhance channel-spatial interactions. arXiv
preprint arXiv:​2112.​05561 (2021).
72. Lin, T.-Y. et al. Microsoft coco: Common objects in context. Proc. Computer Vision–ECCV 2014: 13th European Conference Sept
6–12, 2014, Part V 13, 740–755 (Zurich, Switzerland, Springer, 2014).
73. Kingma, D. P. & Ba, J. Adam: A method for stochastic optimization. arXiv preprint arXiv:​1412.​6980 (2014).
74. Ruder, S. An overview of gradient descent optimization algorithms. arXiv preprint arXiv:​1609.​04747 (2016).

Author contributions
R.J. conceived and conducted the experiments, and W.C. provided knowledge about the fracture and analyzed
the results. All authors reviewed the manuscript.

Funding
The authors did not receive support from any organization for the submitted work.

Competing interests
The authors declare no competing interests.

Additional information
Correspondence and requests for materials should be addressed to W.C.
Reprints and permissions information is available at www.nature.com/reprints.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and
institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International
License, which permits use, sharing, adaptation, distribution and reproduction in any medium or
format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the
Creative Commons licence, and indicate if changes were made. The images or other third party material in this
article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the
material. If material is not included in the article’s Creative Commons licence and your intended use is not
permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder. To view a copy of this licence, visit http://​creat​iveco​mmons.​org/​licen​ses/​by/4.​0/.

© The Author(s) 2023

Scientific Reports | (2023) 13:20077 | https://doi.org/10.1038/s41598-023-47460-7 13


Content courtesy of Springer Nature, terms of use apply. Rights reserved
Vol.:(0123456789)
Terms and Conditions
Springer Nature journal content, brought to you courtesy of Springer Nature Customer Service Center GmbH (“Springer Nature”).
Springer Nature supports a reasonable amount of sharing of research papers by authors, subscribers and authorised users (“Users”), for small-
scale personal, non-commercial use provided that all copyright, trade and service marks and other proprietary notices are maintained. By
accessing, sharing, receiving or otherwise using the Springer Nature journal content you agree to these terms of use (“Terms”). For these
purposes, Springer Nature considers academic use (by researchers and students) to be non-commercial.
These Terms are supplementary and will apply in addition to any applicable website terms and conditions, a relevant site licence or a personal
subscription. These Terms will prevail over any conflict or ambiguity with regards to the relevant terms, a site licence or a personal subscription
(to the extent of the conflict or ambiguity only). For Creative Commons-licensed articles, the terms of the Creative Commons license used will
apply.
We collect and use personal data to provide access to the Springer Nature journal content. We may also use these personal data internally within
ResearchGate and Springer Nature and as agreed share it, in an anonymised way, for purposes of tracking, analysis and reporting. We will not
otherwise disclose your personal data outside the ResearchGate or the Springer Nature group of companies unless we have your permission as
detailed in the Privacy Policy.
While Users may use the Springer Nature journal content for small scale, personal non-commercial use, it is important to note that Users may
not:

1. use such content for the purpose of providing other users with access on a regular or large scale basis or as a means to circumvent access
control;
2. use such content where to do so would be considered a criminal or statutory offence in any jurisdiction, or gives rise to civil liability, or is
otherwise unlawful;
3. falsely or misleadingly imply or suggest endorsement, approval , sponsorship, or association unless explicitly agreed to by Springer Nature in
writing;
4. use bots or other automated methods to access the content or redirect messages
5. override any security feature or exclusionary protocol; or
6. share the content in order to create substitute for Springer Nature products or services or a systematic database of Springer Nature journal
content.
In line with the restriction against commercial use, Springer Nature does not permit the creation of a product or service that creates revenue,
royalties, rent or income from our content or its inclusion as part of a paid for service or for other commercial gain. Springer Nature journal
content cannot be used for inter-library loans and librarians may not upload Springer Nature journal content on a large scale into their, or any
other, institutional repository.
These terms of use are reviewed regularly and may be amended at any time. Springer Nature is not obligated to publish any information or
content on this website and may remove it or features or functionality at our sole discretion, at any time with or without notice. Springer Nature
may revoke this licence to you at any time and remove access to any copies of the Springer Nature journal content which have been saved.
To the fullest extent permitted by law, Springer Nature makes no warranties, representations or guarantees to Users, either express or implied
with respect to the Springer nature journal content and all parties disclaim and waive any implied warranties or warranties imposed by law,
including merchantability or fitness for any particular purpose.
Please note that these rights do not automatically extend to content, data or other material published by Springer Nature that may be licensed
from third parties.
If you would like to use or distribute our Springer Nature journal content to a wider audience or on a regular basis or in any other manner not
expressly permitted by these Terms, please contact Springer Nature at

[email protected]

You might also like