A Comparative Study of Deep Learning Architectures On Melanoma Detection
A Comparative Study of Deep Learning Architectures On Melanoma Detection
A R T I C LE I N FO A B S T R A C T
Keywords: Melanoma is the most aggressive type of skin cancer, which significantly reduces the life expectancy. Early
Cancer classification detection of melanoma can reduce the morbidity and mortality associated with skin cancer. Dermoscopic images
Computational diagnosis acquired by dermoscopic instruments are used in computational analysis for skin cancer detection. However,
Convolutional neural networks some image quality limitations such as noises, shadows, artefacts exist that could compromise the robustness of
Deep learning
the skin image analysis. Hence, developing an automatic intelligent system for skin cancer diagnosis with ac-
Melanoma detection
curate detection rate is crucial. In this paper, we evaluate the performance of several state-of-the-art convolu-
tional neural networks in dermoscopic images of skin lesions. Our experiment is conducted on a graphics pro-
cessing unit (GPU) to speed up the training and deployment process. To enhance the quality of images, we
employ different pre-processing steps. We also apply data augmentation methodology such as horizontal and
vertical flipping techniques to address the class skewness problem. Both pre-processing and data augmentation
could help to improve the final accuracy.
1. Introduction clinical images - since such images captured using conventional digital
cameras are inexpensive and non-invasive - used in computational
Computational analysis for inferring the skin lesion data is an area analysis for patients of different ages and ethnicities (Oliveira et al.,
of increased research interest because of its importance in cancer di- 2016a). Dermoscopic images acquired by dermoscope, a special non-
agnosis and treatment planning. Melanocytes are melanin-producing invasive imaging instrument, generally have better invariance to illu-
cells whose main purpose is to give color to skin, hair and eyes (Al- mination and contrast (Bi et al., 2018). However, different image ac-
masni et al., 2018; Flores and Scharcanski, 2016). Skin cancer, the most quisition conditions and parameters such as variable distances, non-
common type of cancer (Xu et al., 2018), can be classified into two ideal illumination, poor resolution, and also image quality limitations
categories: melanoma and non-melanoma (Okur and Turkan, 2018). such as noises, shadows, artefacts, hairs and reflections decrease the
The high morbidity and considerable healthcare cost associated with image quality that could compromise the robustness of the skin lesion
the malignant type of lesion (Fig. 1a) has motivated researchers to analysis (Jafari et al., 2016). Therefore, a computer aided diagnosis
develop more accurate and flexible algorithms for early melanoma system that is able to analyze dermoscopic-based images may be take
detection (Pathan et al., 2018) since, abnormal melanocyte cells divide into consideration to aid dermatologist’s in melanoma diagnosis
out of control and spread to other tissues of the body and become highly (Oliveira et al., 2016a).
metastasis and increase the number of deaths. According to the data For skin lesion image analysis in this paper, the image pre-proces-
provided by American Cancer Society (American Cancer Society, 2018), sing steps are fundamental in skin lesion detection. Different image pre-
the rates of melanoma have been increasing during the last 30 years. In processing step including color space correction, contrast enhancement
2018, about 91,270 new melanoma cases will be diagnosed and about method and noise reduction can effectively improve the prediction of
9320 people are expected to die of melanoma in the United States. skin cancer (Okur and Turkan, 2018). The objective of this study is to
However, melanoma is highly curable if diagnosed in its early stage. apply the state-of-the-art deep learning architectures on dermoscopic
Fig. 1 illustrates some examples of skin lesions. images to classify skin lesions.
Employing effective non-invasive diagnostic techniques aid derma- The rest of the paper is organized as follows: A review of previous
tologists to collect data, providing insights to melanoma structure and related works, motivations and contributions of this study are presented
shape. Macroscopic images, better known as non-dermoscopic or in Section 2. A detailed description of data augmentation and
⁎
Corresponding author.
E-mail address: [email protected] (S. Hosseinzadeh Kassani).
https://doi.org/10.1016/j.tice.2019.04.009
Received 3 March 2019; Accepted 19 April 2019
Available online 22 April 2019
0040-8166/ © 2019 Elsevier Ltd. All rights reserved.
S. Hosseinzadeh Kassani and P. Hosseinzadeh Kassani Tissue and Cell 58 (2019) 76–83
preprocessing steps, candidate Deep Convolutional Neural Networks In a study done by Ayan, (2018), the performance of a CNN archi-
(DCNN) and evaluation metrics is proposed in Section 3. The experi- tecture is compared between a non-augmented dataset and augmented
mental data, results and discussion of comparative study are reported in dataset for classification of skin lesions. They proposed that the data
Section 4, and finally Section 5 draws the conclusion. augmentation methods could be useful for building powerful classifiers
with insufficient data. Results showed that the network using the aug-
2. Related works mented dataset has achieved better accuracy rate than non-augmented
data.
During the last few years, advances in deep convolutional neural Erçal et al. (1999), introduced a skin cancer classification using
networks have shown promising results in object recognition tasks and diagnostic-tree based hierarchical neural networks and fuzzy logic
became a demanding research domain for classification in medical based on morphological features. There were four classes in their study.
image processing (Naylor et al., 2017). Most of the studies in melanoma These classes are: malignant melanoma, atypical mole, basal cell car-
detection involve only machine learning algorithms, however, there are cinoma or actinic keratosis, and intradermal nevus or seborrheic kera-
a few thoroughly skin lesions classification studies using deep learning tosis.
algorithms. Fatima et al. (2012) proposed a methodology based on a Multi-
Instead of training a CNN from scratch with randomly initialized Parameter Extraction and Classification System (MPECS) of twenty-one
parameters, Kawahara et al. (2016) utilizeed a pretrained CNN to features using six phases and then analyzed them based on statistical
classify skin images on the entire dataset. This pre-training highly re- methods for early detection of skin cancer melanoma. Some of the
duceed the training time of CNN and achieves 85.8% accuracy over 5- features are lesion borders, color, symmetry, area, perimeter and the
classes. eccentricity. They used Sobel edge detection for detecting lesion edges.
Liao's work (Liao, 2016) provided an universal skin disease classi- The symmetry feature is computed by horizontal and vertical axis of
fication by using transfer learning strategy and fine-tuning ImageNet lesions. The color spreading factor of the lesions is computed based on
pretrained weights on the Dermnet dataset to improve the training the similarity of neighborhood pixels. Their study demonstrated sin-
performance. They evaluate the model performance on both the gular statistical analysis from extracted features is not adequately suf-
Dermnet dataset and skin images from New York State Department of ficient to accurately classify the skin lesions. Therefore, advanced
Health skin disease dataset. Their method achieved 73.1% Top-1accu- classification methods such as support vector machines and decision
racy and 91.0% Top-5 accuracy on the Dermnet dataset. and 69.5% tree algorithms could be considerate for classification.
Top-5 accuracies on the OLE dataset. We tabulate some of the recent studies on Melanoma detection at
Ercal et al. (1994) developed a method to detect skin tumors using Table 1.
color features. The accuracy reported by this approach is about 82%.
The highest success rate is achieved for “intradermal nevus” at 100% 2.1. Motivations and contributions
and the lowest success rate is for “melanoma” images at 77%.
Menegola et al. (2017) proposed knowledge transfer method to With above review in place, we found that accurate detection of skin
enhance performance of deep learning for melanoma screening. In their cancer is crucial for providing necessary treatment for patients. Several
study, a pre-trained model trained on the Kaggle Challenge for Diabetic studies show that developing accurate methods by employing deep
Retinopathy Detection dataset. They expected that the deeper models learning algorithm plays an emerging role in the diagnosis of various
and transfer learning from a related dataset improve the performance diseases such as cancer (Hu et al., 2018), pneumonia (Kermany et al.,
and leads to better results. However, their work suggest that the ex- 2018), and Alzheimer (Amoroso et al., 2018). Interestingly, study at
perimental design is sensitive to the type of skin lesions (benign or Dorj et al. (2018), shows high accuracies obtained for skin cancer
malignant). classification using AlexNet (Krizhevsky et al., 2012). Studies at Pereira
Lopez et al. (2017) emolyed a deep-learning based approach for et al. (2016), Vasconcelos and Vasconcelos, (2017) and Wahab et al.
early detection of melanoma. Their solution is based on a modified (2017) show that data augmentation and pre-processing may further
VGGNet architecture and transfer learning technique to solve the skin help to get a more accurate and unbiased detector. According to the
lesion classification task. The proposed method achieveed a sensitivity observations of the study at Lei et al. (2018), the highest accuracy,
value of 78.66% on the ISIC Archive dataset. 93.46% is gained by a modified ResNet50 model with a deeper model,
Table 1
List of recent studies on Melanoma detection.
Dataset Methods (Pre-processing, augmentation) Accuracy Year of publication
77
S. Hosseinzadeh Kassani and P. Hosseinzadeh Kassani Tissue and Cell 58 (2019) 76–83
i.e. using more hidden layers in the structure of DCNN. So, all these augmentation techniques by the number of random images transfor-
studies motivated us to examine the performance of five state of the arts mation such as horizontal and vertical flips, random contrast and
DCNNs, namely AlexNet (Krizhevsky et al., 2012), VGGNet16 random brightness. These techniques help to avoid the negative effect
(Simonyan and Zisserman, 2014), VGGNet19 (Simonyan and of strong class imbalances and hence increase the performance of
Zisserman, 2014), ResNet50 (He et al., 2016), and Xception (Chollet, DCNNs.
2017) with the aim of verifying the effectiveness of various pre-
processing and data augmentation techniques. 3.3. Deep convolutional neural networks
In the following, we list contributions of this study:
DCNNs have achieved the state-of-the-art performance in many
1 We added two fully connected layers to combines the feature maps applications such as image and speech recognition (Fayek et al., 2017),
of various intermediate layers of DCNNs. For both layers, dropout automatic video classification (Shao et al., 2014), object detection
method is employed which avoids the overfitting problem. (Pathak et al., 2018) and natural language processing (Sun et al., 2017).
Moreover, the strategy of transferring features learned from A Convolution Neural Network (CNN), known as ConvNet, is a specific
ImageNet dataset (which is different form medical datasets) is used. type of feed-forward neural network with a stack of convolutional
To minimize the training loss, L2 regularization (ridge regression) layers, each followed by pooling layers in order to extract features from
technique is also employed for classification task. These changes the input data and produce a set of high-level feature maps at each level
could bring higher classification performance to our learners. of convolution. The feature maps information is summarized using
2 To reduce heterogeneity of data, and hence, improvement of the pooling layers in order to reduce the number of parameters and further
classification performance, the pre-processing steps such as contrast is followed by a fully connected layer to produce the final classification
enhancement, color space transformation, and a specific type of il- (Sainath et al., 2013; Krizhevsky et al., 2012). Fig. 2 illustrates a gra-
lumination correction proposed by Cavalcanti et al. at (Cavalcanti phical representation of a typical convolution neural network.
et al., 2010) are used. These preprocessing steps also help to reduce
the sensitivity of the DCNNs to contrast and intensity bias. There- 3.4. Xception architecture
fore, the DCNNs are able to learn high-level structure information
more efficiently and ensure to gain a better prediction. The Xception deep neural network which stands for extreme in-
3 The dataset provided for this study is severely skewed. To balance ception, was made by François Chollet Chollet (2017). Xception ar-
the dataset, an oversampling method for minority classes examples is chitecture has depth-wise separable convolutions. Xception has 36
proposed. This method consists of different data augmentation convolutional layers to extract important features and is inspired by
techniques such as horizontal and vertical flips, random contrast Inception Chollet (2017) wherein the Inception modules are replaced
and random brightness that aid to avoid the negative effect of class with depth-wise separable convolutions consisting of a depth-wise
imbalances. convolution.
In this section we briefly explain the pre-processing and data aug- AlexNet (Krizhevsky et al., 2012) architecture won the first prize in
mentation methods, DCNNs, and evaluation metrics of learners. All the ImageNet Large-Scale Visual Recognition Challenge (ILSVRC) 2012
these networks are compared with each other and also with and competition. AlexNet is the first breakthrough in the architecture of
without preprocessing and augmentation steps. CNNs applied to large datasets. The basic architecture of AlexNet
consists of five convolutional layers, two normalization layers, three
3.1. Pre-processing max-pooling layers, three fully-connected layers, and a linear layer with
Softmax activation function in the output to ensure each neuron pre-
The key point of pre-processing is to reduce the effect of noise and dicts class probabilities of a particular image.
unbalanced patterns in the dataset that tend to decrease the ability of
DCNNs in learning important features. Common pre-processing steps 3.6. VGGNet architecture
are contrast enhancement, color space transformation, and illumination
correction (Oliveira et al., 2016b). Also, non-skin noises such as hair, VGGNet was introduced in 2014 by Karen Simonyan and Andrew
and rulers should be eliminated from skin images. There are also noisy Zisserman (Simonyan and Zisserman, 2014) from Visual Geometry
representations such as uneven illumination patterns, reflections, poor Group (VGG) of the University of Oxford. It achieves one of the top
background, different lightening and shadows that adversely affect performances in the ImageNet Large Scale Visual Recognition Chal-
image quality. lenge (ILSVRC) 2014. VGGNet utilizes smaller filters of 3 × 3, com-
pared to AlexNet 11 × 11 filter, in order to provide better features
3.2. Data augmentation extraction from images. The study also verifies that using much smaller
filters in order to increase the depth of network instead of its width
Oversampling and undersampling are common methods to fix im- plays a critical role for gaining higher performance. There are two
balanced datasets. Oversampling is utilized to generate either exact versions of this architecture: VGG16 and VGG19 with different depths
copies, or modified copies of the original samples that belongs to the and layers. VGG19 is deeper than VGG16. However, the number of
minority class to adjust the class distribution. There are different data parameters for VGG19 is larger and is hence more expensive to train the
78
S. Hosseinzadeh Kassani and P. Hosseinzadeh Kassani Tissue and Cell 58 (2019) 76–83
4. Result and discussion The results obtained from candidate DCNN architectures are pre-
sented in Table 3 for Melanoma classification. The results show that
4.1. Dataset and experimental setup ResNet50 architecture with data augmentation and pre-processing steps
achieves the best accuracy rate in the majority of the validation pro-
The dataset used for this research is skin lesion analysis towards cesses and it has the highest average F-score of 92.74%, and accuracy of
melanoma detection provided by International Skin Imaging 92.08%.
79
S. Hosseinzadeh Kassani and P. Hosseinzadeh Kassani Tissue and Cell 58 (2019) 76–83
Fig. 3. pre-processed and augmented data. (a) Original dermoscopy image, (b) Image after application of illumination correction, (c) Image after application of
contrast enhancement space and horizontal flip, (d) Image after transformation into HSV color space and horizontal flip.
80
S. Hosseinzadeh Kassani and P. Hosseinzadeh Kassani Tissue and Cell 58 (2019) 76–83
Table 4
The classification performance of different methods based on different classes.
Methods AKIEC BCC BKL DF MEL NV VASC Average ± Std
AlexNet 0.9111 0.8492 0.7053 0.9836 0.80 0.6816 0.9605 0.84 ± .11
VGGNet16 0.9726 0.8826 0.7535 0.9933 0.9438 0.7871 1 0.90 ± .0.09
VGGNet19 0.8749 0.961 0.8262 1 0.8764 0.7625 1 0.89 ± 0.08
ResNet50 0.9743 0.929 0.7717 0.9905 0.9465 0.9232 1 0.92 ± 0.07
Xception 0.9646 0.914 0.7652 1 0.8911 0.8427 1 0.90 ± 0.08
The best values under the same conditions are shown boldface.
and vertical flip, random contrast and random brightness. Augmenting obtained results. Our experiments show that ResNet50 outperforms the
data enabled us to increase the size of training set and reduce the counterparts AlexNet, Xception, VGGNet16 and VGGNet19 archi-
overfitting problem. We employed various standard evaluation metrics tectures with classification accuracy as high as 92.08% and F-score
such as specificity, sensitivity, accuracy, F-measure to evaluate the equal to 92.74%. For future research direction, we aim to boost
Fig. 5. Confusion matrix of skin lesion classification using different DCNN models. The values on the main diagonal represent all correctly classified instances. The
row under each confusion matrix shows the rate of accuracy achieved for each predicted class and misclass predictions.
81
S. Hosseinzadeh Kassani and P. Hosseinzadeh Kassani Tissue and Cell 58 (2019) 76–83
Table 5 also thankful to the reviewers in advance for their comments and
The number of misclassified samples from each model. suggestions.
Xception AlexNet VGG16 VGG19 ResNet50
References
Number of misclassified test 1139 2092 1305 1272 890
instances (Total: 11,233 Al-masni, M.A., Al-antari, M.A., Choi, M.T., Han, S.M., Kim, T.S., 2018. Skin lesion seg-
instances) mentation in dermoscopy images via deep full resolution convolutional networks.
Comput. Methods Programs Biomed.
“American Cancer Society,” 2018. [Online]. Available: https://www.cancer.org/cancer/
accuracy using deep learning-based ensemble models for melanoma melanoma-skin-cancer/about/key-statistics.html.
Amoroso, N., et al., 2018. Deep learning reveals Alzheimer’s disease onset in MCI sub-
detection. jects: Results from an international challenge. J. Neurosci. Methods 302, 3–9.
Ayan, E.H.M.Ü., 2018. Data augmentation importance for classification of skin lesions via
deep learning. Electric Electronics, Computer Science, Biomedical Engineerings’
Disclosure of potential conflict of interest Meeting (EBBT).
Barata, C., Marques, J.S., Celebi, M.E., 2014. Improving dermoscopy image analysis using
Non-Declared. color constancy. Image Processing (ICIP), 2014 IEEE International Conference on. pp.
3527–3531.
Bi, L., Kim, J., Ahn, E., Feng, D., Fulham, M., 2016. Automatic melanoma detection via
Acknowledgments multi-scale lesion-biased representation and joint reverse classification. 2016 IEEE
13th International Symposium on Biomedical Imaging (ISBI). pp. 1055–1058.
Bi, L., Kim, J., Ahn, E., Kumar, A., Feng, D., Fulham, M., 2019. Step-wise integration of
The authors would like to express our gratitude to Dr. Ralph Deters deep class-specific learning for dermoscopic image segmentation. Pattern Recognit.
of the Department of Computer Science from University of 85, 78–89.
Cavalcanti, P.G., Scharcanski, J., Lopes, C.B.O., 2010. Shading attenuation in human skin
Saskatchewan, Canada for providing his support of this study. We are
82
S. Hosseinzadeh Kassani and P. Hosseinzadeh Kassani Tissue and Cell 58 (2019) 76–83
color images. International Symposium on Visual Computing. pp. 190–198. Naylor, P., Laé, M., Reyal, F., Walter, T., 2017. Nuclei segmentation in histopathology
Chollet, F., 2017. Xception: deep learning with depthwise separable convolutions. images using deep neural networks. Biomedical Imaging (ISBI 2017), 2017 IEEE 14th
Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, International Symposium on. pp. 933–936.
CVPR 2017. Okur, E., Turkan, M., 2018. A survey on automated melanoma detection. Eng. Appl. Artif.
Codella, N.C.F., et al., 2018. Skin lesion analysis toward melanoma detection: a challenge Intell. 73, 50–67.
at the 2017 International symposium on biomedical imaging (ISBI), hosted by the Oliveira, R.B., Filho, M.E., Ma, Z., Papa, J.P., Pereira, A.S., Tavares, J.M.R.S., 2016a.
international skin imaging collaboration (ISIC). Proceedings - International Computational methods for the image segmentation of pigmented skin lesions: a
Symposium on Biomedical Imaging. review. Comput. Methods Programs Biomed.
Dorj, U.-O., Lee, K.-K., Choi, J.-Y., Lee, M., 2018. The skin cancer classification using deep Oliveira, R.B., Mercedes Filho, E., Ma, Z., Papa, J.P., Pereira, A.S., Tavares, J.M.R.S.,
convolutional neural network. Multimed. Tools Appl. 77 (Apr. (8)), 9909–9924. 2016b. Computational methods for the image segmentation of pigmented skin le-
Ercal, F., Chawla, A., Stoecker, W.V., Lee, H.-C., Moss, R.H., 1994. Neural network di- sions: a review. Comput. Methods Programs Biomed. 131, 127–141.
agnosis of malignant melanoma from color images. IEEE Trans. Biomed. Eng. 41 (no. Pathak, A.R., Pandey, M., Rautaray, S., 2018. Application of deep learning for object
9), 837–845. detection. Procedia Comput. Sci. 132, 1706–1717.
Erçal, F., Lee, H.-C., Stoecker, W.V., Moss, R.H., 1, 1999. Skin Cancer Classification Using Pathan, S., Prabhu, K.G., Siddalingaswamy, P.C., 2018. Techniques and algorithms for
Hierarchical Neural Networks and Fuzzy Systems. computer aided diagnosis of pigmented skin lesions—a review. Biomed. Signal
Fatima, R., Khan, M.Z.A., Dhruve, K.P., 2012. Computer aided multi-parameter extraction Process. Control. 39, 237–262.
system to aid early detection of skin cancer melanoma. Int. J. Comput. Sci. Netw. Pereira, S., Pinto, A., Alves, V., Silva, C.A., 2016. Brain tumor segmentation using con-
Secur. 12 (10), 74–86. volutional neural networks in MRI images. IEEE Trans. Med. Imaging 35 (5),
Fayek, H.M., Lech, M., Cavedon, L., 2017. Evaluating deep learning architectures for 1240–1251.
speech emotion recognition. Neural Netw. 92, 60–68. Sainath, T.N., Mohamed, A., Kingsbury, B., Ramabhadran, B., 2013. Deep convolutional
Flores, E., Scharcanski, J., 2016. Segmentation of melanocytic skin lesions using feature neural networks for LVCSR. 2013 IEEE International Conference on Acoustics, Speech
learning and dictionaries. Expert Syst. Appl. 56, 300–309. and Signal Processing (ICASSP). pp. 8614–8618.
He, K., Zhang, X., Ren, S., Sun, J., 2016. Deep residual learning for image recognition. Satheesha, T.Y., Satyanarayana, D., Prasad, M.N.G., Dhruve, K.D., 2017. Melanoma is skin
2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). deep: a 3D reconstruction technique for computerized dermoscopic skin lesion clas-
Hu, Z., Tang, J., Wang, Z., Zhang, K., Zhang, L., Sun, Q., 2018. Deep learning for image- sification. IEEE J. Transl. Eng. Heal. Med. 5, 1–17.
based cancer detection and diagnosis − a survey. Pattern Recognit. 83, 134–149. Shao, L., Cai, Z., Liu, L., Lu, K., 2017. Performance evaluation of deep feature learning for
“ISIC 2018 [Online]. Available: https://challenge2018.isic-archive.com/. RGB-D image/video classification. Ann. Pure Appl. Log. 385-386, 266–283.
Jafari, M.H., et al., 2016. Skin lesion segmentation in clinical images using deep learning. Simonyan, K., Zisserman, A., 2014. Very deep convolutional networks for large-scale
2016 23rd International Conference on Pattern Recognition (ICPR). pp. 337–342. image recognition. CoRR vol. abs/1409.1.
Kawahara, J., BenTaieb, A., Hamarneh, G., 2016. Deep features to classify skin lesions. Sun, S., Luo, C., Chen, J., 2017. A review of natural language processing techniques for
2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI). pp. opinion mining systems. Inf. Fusion 36, 10–25.
1397–1400. Vasconcelos, C.N., Vasconcelos, B.N., 2017. Experiments using deep learning for der-
Kermany, D.S., et al., 2018. Identifying medical diagnoses and treatable diseases by moscopy image analysis. Pattern Recognit. Lett.
image-based deep learning. Cell 172 (5), 1122–1131. Wahab, N., Khan, A., Lee, Y.S., 2017. Two-phase deep convolutional neural network for
Krizhevsky, A., Sutskever, I., Hinton, G.E., 2012. 1 ImageNet classification with deep reducing class skewness in histopathological images based breast cancer detection.
convolutional neural networks. Adv. Neural Inf. Process. Syst. Comput. Biol. Med. 85, 86–97.
Lei, H., et al., 2018. A deeply supervised residual network for HEp-2 cell classification via Waheed, Z., Waheed, A., Zafar, M., Riaz, F., 2017. An efficient machine learning approach
cross-modal transfer learning. Pattern Recognit. 79, 290–302. for the detection of melanoma using dermoscopic images. International Conference
Liao, H., 2016. A deep learning approach to universal skin disease classification. Univ. on Communication, Computing and Digital Systems (C-CODE). pp. 316–319.
Rochester Dep. Comput. Sci. CSC. Xu, H., Lu, C., Berendt, R., Jha, N., Mandal, M., 2018. Automated analysis and classifi-
Lopez, A.R., Giro-i-Nieto, X., Burdick, J., Marques, O., 2017. Skin lesion classification cation of melanocytic tumor on skin whole slide images. Comput. Comput. Med.
from dermoscopic images using deep learning techniques. 2017 13th IASTED Imaging Graph 66, 124–134.
International Conference on Biomedical Engineering (BioMed). pp. 49–54. Yang, S., Oh, B., Hahm, S., Chung, K.-Y., Lee, B.-U., 2017. Ridge and furrow pattern
Matsunaga, K., Hamada, A., Minagawa, A., Koga, H., 2017. Image classification of mel- classification for acral lentiginous melanoma using dermoscopic images. Biomed.
anoma, nevus and seborrheic keratosis by deep neural network ensemble. arXiv Signal Process. Control 32, 90–96.
Prepr. arXiv1703.03108. Yu, L., Chen, H., Dou, Q., Qin, J., Heng, P.-A., 2017. Automated melanoma recognition in
Menegola, A., Fornaciali, M., Pires, R., Bittencourt, F.V., Avila, S., Valle, E., 2017. dermoscopy images via very deep residual networks. IEEE Trans. Med. Imaging 36
Knowledge transfer for melanoma screening with deep learning. Proceedings - (4), 994–1004.
International Symposium on Biomedical Imaging.
83