Abstract
Background
Melanoma has become more widespread over the past 30 years and early detection is a major factor in reducing mortality rates associated with this type of skin cancer. Therefore, having access to an automatic, reliable system that is able to detect the presence of melanoma via a dermatoscopic image of lesions and/or skin pigmentation can be a very useful tool in the area of medical diagnosis.
Methods
Among state-of-the-art methods used for automated or computer assisted medical diagnosis, attention should be drawn to Deep Learning based on Convolutional Neural Networks, wherewith segmentation, classification and detection systems for several diseases have been implemented. The method proposed in this paper involves an initial stage that automatically crops the region of interest within a dermatoscopic image using the Mask and Region-based Convolutional Neural Network technique, and a second stage based on a ResNet152 structure, which classifies lesions as either “benign” or “malignant”.
Results
Training, validation and testing of the proposed model was carried out using the database associated to the challenge set out at the 2017 International Symposium on Biomedical Imaging. On the test data set, the proposed model achieves an increase in accuracy and balanced accuracy of 3.66% and 9.96%, respectively, with respect to the best accuracy and the best sensitivity/specificity ratio reported to date for melanoma detection in this challenge. Additionally, unlike previous models, the specificity and sensitivity achieve a high score (greater than 0.8) simultaneously, which indicates that the model is good for accurate discrimination between benign and malignant lesion, not biased towards any of those classes.
Conclusions
The results achieved with the proposed model suggest a significant improvement over the results obtained in the state of the art as far as performance of skin lesion classifiers (malignant/benign) is concerned.
Keywords: Mask R_CNN, Deep learning, Transfer learning, Convolutional neural network, Object detection, Object classification
Background
By the year 2019, the American Cancer Society (ACS) has calculated that approximately 96,480 new cases of melanomas will be diagnosed and that approximately 7230 people will die from melanoma in the United States [1]. Melanoma is considered to be the most serious type of skin cancer, meaning it is particularly important to accurately diagnose it as soon as possible, to allow treatment before any dissemination and/or metastasis, as this will significantly increase the chance of recovery [1–4]. Generally speaking, melanoma can be recognized via a visual inspection that focuses on the region of the cutaneous lesion. However, there is great similarity between melanomas and other skin lesions such as nevus, increasing the difficulty in performing cancer classification and diagnostic tasks in skin lesions [5]. Therefore, ensuring access to a reliable, practical system that is able to conduct an automated evaluation of a skin lesion would have a positive influence on patient care and increases the chance of early detection of cancer, thus reducing mortality rates associated with this disease.
In dermatology, both diagnosis and monitoring of skin lesions have relied mainly on visual inspection and other non-invasive evaluations, invasive procedures are avoided because they can destroy the lesions and make it impossible to carry out a clinical monitoring of its evolution [6]. In highly selected patient populations, computer assisted diagnosis methods have demonstrated high sensitivity and may be useful as a back-up diagnostic aid for specialists to reduce the risk of missing melanomas [7]. One of the non-invasive methods employed in diagnosis and monitoring is the use of confocal microscopes available for clinical use, which provide sharp images because they capture only the light from the plane of the sample in focus. In vivo reflectance confocal microscopy captures high resolution images in real time and is used in the evaluation of various dermatological conditions [6]. Reflectance confocal microscopy may reduce unnecessary excisions without missing melanoma cases [8]. For its part, research results in [9] show that dermoscopy is more accurate than visual inspection alone; the results reported in [9, 10] indicate that melanomas would not be detected using visual inspection. In our proposed model, the conclusive analysis of dermatoscopic images is automated through artificial intelligence algorithms, obtaining a preliminary characterization that serves as support for dermatologists in order to obtain faster and more precise diagnoses.
Deep automatic models base their learning on the samples used for their training, this work uses the ISIC 2017 database, which has more than 2000 high resolution dermatoscopic images grouped into 3 main categories: Melanoma, nevus and keratosis. Testing determines if these data is enough to achieve high performance detecting and rejecting benign and malignant dysplasic lesions.
From a technical standpoint of the problem, it has been found that many cutaneous lesion detectors have been proposed using Deep Learning models based on Convolutional Neural Networks (CNN), such as GoogLeNet-AlexNet-ResNet-VGGNet ensembles [11], R_CNN [12], the Mask and Region-based Convolutional Neural Network (Mask R_CNN) and DeeplabV3+ method ensemble [13] and in general many CNN structures, making this approach one of the most powerful for effective feature extraction and classification [14]. The method proposed in this work exploits the potentialities of Mask R_CNN for detection of objects, patterns and/or figures within an image together with potentiality of the ResNet152 model for classification purposes. The ResNet152 model was selected after testing with other ResNet, Inception and VGG structures, all of them with demonstrated high-performance on several image classification tasks [15, 16].
The main factors that make classification of dermatoscopic images difficult are: the presence of hairs, inks, ruler markings, colored patches, glimmers of light, drops, oil bubbles, blood vessels, hypopigmentation areas and/or inflammation around the lesion, among others [17]. Extraction features in CNN is affected when this undesired noise is present in the image, directly affecting performance of the classifier. Therefore, accurate extraction of the region of interest (ROI) is considered an essential step towards improving performance of the system [18, 19]. To solve the melanoma detection problem, taking into account the issues mentioned above, we propose a two-stage classification method:
Stage 1: Use Mask R_CNN to create a bounding box around the skin lesion, with as few visual noise as possible. The output of this stage is the cropped bounding box.
Stage 2: Classification of the cropped area using ResNet152. Although Mask R_CNN can also be used for image classification, not only for object detection as in Stage 1, we decided to separate the two tasks in order to be able to use higher performing classification models.
Something that should be stressed in this work is the constant search for high-performance models using different balancing ratios for the training data used by the algorithms. This is of great interest for the scientific community, given that unbalanced repositories are very common in the medical field.
The most recently published dermatoscopic image classification work (malignant/benign) for the ISIC (International Skin Imaging Collaboration) 2017 challenge, using the data set from the International Symposium on Biomedical Imaging (ISBI) 2017 challenge, is taken as a reference within the reviewed state-of-the-art. An objective comparison of results is made with our proposed method using the evaluation metrics adopted by the challenge such as specificity, sensitivity and accuracy. However, we added another metric—balance accuracy—suggested for this particular case in which there is an unbalanced database [20–22]. A Receiver Operating Characteristic (ROC) space is provided in order to compare and display performance of various classifier models in a single graph, and this is a very useful metric for decision-making in the medical field, the use of which has been increasing in machine learning research [23].
The rest of this paper is organized as follows: "Methods" section explains in detail the proposed method; "Results" section shows the results obtained from the experiments carried out to select the best performing model for the proposed method and compares the best results obtained from the ISIC 2017 challenge with the proposed method; "Discussion" section provides a discussion and finally, "Conclusions" section explain the conclusions.
Methods
We propose an automated classification method for cutaneous lesion in digital dermatoscopic images, in order to detect the presence of melanoma. This method comprises two fundamental stages, which are: Stage 1: Cropping a bounding box around only the skin lesion in the input image, using Mask R_CNN; and Stage 2: Classification of the cropped bounding box using ResNet152, as described in Fig. 1 with a functional block diagram of the system.
In the first stage, Mask R_CNN makes a selection of candidates, i.e. it filters regions within the image that may contain a skin lesion in order to be classified. The main purpose of this step is to crop a bounding box of the image around the region of interest (ROI). We then train, validate and test the model in this stage using the dataset from the ISIC 2017 challenge [24], which contains 1995 dermatoscopic images with their respective masks for training, 149 dermatoscopic images for validation and 598 dermatoscopic images for testing [25].
Mask R_CNN trains the system with masks provided by a clinical expert, so as to be able to identify those pixels that correspond to a skin lesion within the image, i.e. our region of interest (ROI). Although there are many candidates in an image, calculating the probability that one candidate belongs to a skin lesion plays a major role when rejecting other elements from the image such as marks, patches or bubbles, among others. Mask R_CNN takes candidates and classifies them by generating probabilities as shown in Fig. 2a, in which the area corresponding to the yellow patch present in the image has a lower probability of being a skin lesion (0.505 probability) than the probability calculated in a second region that genuinely corresponds to a skin lesion (0.982 probability). Lastly, we crop the area defined by the refined bounding box and create a new image Fig. 2b that is then classified on Stage 2.
In Stage 2, in order to classify the cropped images identified as skins lesions from Stage 1, we decided to include a ResNet152 classifier that outputs a benign or malignant label. A “malignant” class being understood as referring to those skin lesions identified as melanoma and “benign” class referring to all those lesions not identified as melanoma.
Figure 1 also shows the generic internal structure of ResNet152, which comprises two blocks: one containing the feature extraction layers known as convolution layers, and the other performing the classification via a fully connected network. In overall terms, this convolutional network is made up of over 60 million parameters that need to be adjusted in a process known as training, which usually requires a very large volume of images. The relatively small amount of data available in the ISBI 2017 database (1995 items for training and 149 for validation) makes it unsuitable to train such a large network from scratch. Therefore, we decided to use the technique known as Transfer Learning, which uses a network pre-trained using a very large general-purpose database, and subsequently retrains it using an smaller, specific-purpose database. In our case, this was undertaken following the steps shown in Fig. 3.
An iterative adjustment of the learning rate [27] is undertaken following the procedure described in Fig. 3 for the five training scenarios listed in Table 1, in which the ratio between the number of malignant and benign dermatoscopic images is modified.
Table 1.
Model | Technique to modify the malignant/benign training ratio | Malignant/Benign training ratio |
---|---|---|
Model 1 | None | 0.231 |
Model 2 | Data augmentation applied to “malignant” class images, no changes to “benign” class images | 0.440 |
Model 3 | Data augmentation applied to “malignant” class images, down sampling applied to “benign” class images | 0.463 |
Model 4 | Data augmentation applied to “malignant” class images, down sampling applied in “benign” class images | 1 |
Model 5 | Data augmentation and addition of PH2 images applied to “malignant” class images, no changes to “benign” class images | 0.256 |
The original dataset from the ISBI 2017 challenge was used to obtain model 1, unlike models 2, 3, 4 and 5, in which we expanded the size of the “malignant” class training data set by 90% by generating modified versions of the dermatoscopic images and applying the following data augmentation techniques: rotation augmentation (image rotation by 180 degrees) and vertical flip augmentation (pixel reordering by row reversal). In models 3 and 4, we also reduced the number of dermatoscopic images from the “benign” class by different ratios. Lastly, in model 5 we included dermatoscopic images with “malignant” skin lesions extracted from the PH2 database [28].
The only hyper-parameters adjusted were the learning rate and the number of epochs, with the best performing values for each training scenario shown in Table 2. The momentum and batch size are set at 0.9 and 32 respectively.
Table 2.
Model/Hyper-parameters | Number of epochs with learning rate 0.001 | Number of epochs with learning rate 0.0001 | |
---|---|---|---|
Model 1 | 79 | 87 | |
Model 2 | 85 | 95 | |
Model 3 | 67 | 90 | |
Model 4 | 93 | 74 | |
Model 5 | 56 | 59 |
For our application, the ResNet152 classifies between “malignant” or “benign” classes, thus resulting in a binary classifier whose performance is evaluated using specificity, sensitivity, accuracy and balance accuracy metrics, in which the reference class is the “malignant” class. Sensitivity and specificity are the standard metrics used to evaluate performance of classifiers in the field of medicine [13]. In the case of our skin lesion classifier, a specificity (Eq. (1)) close to 1.0 indicates good performance when correctly classifying the “benign” class. Conversely, a sensitivity (Eq. (2)) close to 1.0 indicates good performance when correctly classifying the “malignant” class (melanoma lesion). Accuracy (Eq. (3)) is another very commonly-used evaluation metric in the area of medical applications that measures the overall performance of a classifier [11–13, 17]; however, this metric is not suitable when the data set is highly unbalanced, as a classifier that is biased towards the most frequently occurring class leads to an optimistic estimate that would provide misleading results. This problem can be addressed by replacing this metric with a different one known as balanced accuracy for the specific case of a binary-type classifier [20, 21]. Specificity, sensitivity, accuracy and balanced accuracy of classifiers are calculated from the resulting confusion matrices, and also compared in a Receiver Operating Characteristic (ROC) space, as shown in Fig. 4.
A singular point in the ROC space is better than another if the latter is located closer to the northeast corner than the first, i.e. the coordinate (0,1) represents a perfect classifier. The ROC space is commonly used for decision-making in the field of medicine, as it shows the relative compensation that may exist between benefits (true positives) and costs (false positives) [23].
Results
This section shows the results obtained when applying proposed method to the five training scenarios described in Table 1, in which a range of data transform techniques are used to reduce the degree of data imbalance during training.
Model 1 was obtained by training the classifier with the data set from the ISBI 2017 challenge, which contains 1995 dermatoscopic images for training, of which 1620 correspond to benign lesions and 375 to malignant lesions; as well as 149 dermatoscopic images for validation purposes, 119 corresponding to benign lesions and 30 to malignant lesions. Model 1 was trained without applying any augmentation technique or reduction data, obtaining the confusion matrix shown in Fig. 5 when tested with the test data from ISBI 2017 challenge, which comprises 598 dermatoscopic images, 481 corresponding to benign lesions and 117 to malignant lesions. All testing results we provide in this section for the five models were obtained over this same test data set.
Despite the fact that excellent performance is noted when classifying the “benign” class, i.e. high specificity, the same behaviour is not observed when classifying the “malignant” class, i.e. low specificity. A possible fundamental reason for this is the clear bias in the training data set towards the “benign” class. This reason, justifies the use of data augmentation/reduction techniques in order to reduce the training imbalance ratio and achieve better overall results, as it is explored with models 2–5.
To obtain model 2, a 90% transform augmentation is applied on the dermatoscopic images from the “malignant” class, thus increasing the balance ratio between the two classes from 0.231 to 0.440. The test data same as before is used to evaluate model 2, obtaining the confusion matrix shown in Fig. 6.
The number of “malignant” test samples that are correctly classified improves from 43 to 84 with respect to model 1. While there is also a drop in the number of “benign” samples correctly classified, it is not nearly as significant, therefore model 2 is postulated as a better model since it achieves a better balance in specificity versus sensitivity than model 1.
In model 3, in addition to the same data augmentation procedure used in model 2 for the “malign” class, data down sampling is applied to the “benign” class images in order to increase the training balance ratio to 0.463. The confusion matrix shown in Fig. 7 was obtained from the evaluation of model 3.
We can see that despite the fact that the degree of data imbalance is lower, the number of “malignant” class images that had been correctly classified failed to improve with respect to model 2—rather, the opposite was true, it decreased.
Model 4 is obtained in the same way as model 3, but the down sampling and augmentation transforms are adjusted to a 1.0 training balance ratio, i.e. until we get the same number of malignant and benign training images. The confusion matrix shown in Fig. 8 was obtained from the test data set.
Despite the fact that the training data is perfectly balanced, the number of correctly classified images from the “benign” class was reduced considerably.
Lastly, for model 5 data augmentation for the “malign” class was performed using data from the PH2 data set [28], while no changes were made to the “benign” class data. This resulted in a 0.256 training balance ratio, i.e. a minimum increase with respect to the original situation used for model 1. The confusion matrix obtained is shown in Fig. 9.
Despite the fact that new images belonging to the “malignant” class from a new dataset are added, no improvement in classification of “malignant” class images is noted.
Lastly, the best model from the previous tests (i.e. model 2) was retrained for 60 more epochs and a lower learning rate of 0.0001. The resulting model is called model 6 and its confusion matrix is shown in Fig. 10. Meanwhile, Table 3 summarizes the results obtained for all 6 tested models, where it is clear that, overall, model 6 shows the best performance.
Table 3.
Evaluation metric | Model 1 | Model 2 | Model 3 | Model 4 | Model 5 | Model 6 |
---|---|---|---|---|---|---|
Specificity | 0.968 | 0.906 | 0.938 | 0.701 | 0.946 | 0.925 |
Sensitivity | 0.367 | 0.718 | 0.556 | 0.752 | 0.393 | 0.820 |
Accuracy | 0.851 | 0.870 | 0.863 | 0.711 | 0.838 | 0.904 |
Balanced accuracy | 0.668 | 0.812 | 0.747 | 0.726 | 0.670 | 0.872 |
Bold values indicate the most relevant and representatives in this rearch work
Table 4 compares the results of the proposed model (model 6 from the previous section, now called eVida) against the best performing models reported for the ISIC challenge 2017 for melanoma classification. For the purposes of comparison, the table has been sorted from highest to lowest according to the reported Area Under the RC Curve (AUC) value, since AUC is the metric chosen by the organizers to define the ranking in this challenge. The challenge not only considers the AUC for melanoma detection, but computes an overall score that averages the AUC for melanoma detection with the AUC for keratosis detection. In this research work, we only consider the AUC for melanoma detection, since no keratosis tests were performed.
Table 4.
Organization | ACC_M | AUC_M | SE_M | SP_M | Balance accuracy | Overall score |
---|---|---|---|---|---|---|
eVida (proposed method M6) | 0.904 | 0.872 | 0.820 | 0.925 | 0.872 | 0.848 |
RECOD Titans | 0.872 | 0.874 | 0.547 | 0.950 | 0.749 | 0.792 |
Popleyi | 0.858 | 0.870 | 0.427 | 0.963 | 0.695 | 0.762 |
Kazuhisa Matsunaga | 0.828 | 0.868 | 0.735 | 0.851 | 0.793 | 0.798 |
Monty python | 0.823 | 0.856 | 0.103 | 0.998 | 0.551 | 0.687 |
T D | 0.845 | 0.836 | 0.350 | 0.965 | 0.658 | 0.726 |
Xulei Yang | 0.830 | 0.830 | 0.436 | 0.925 | 0.681 | 0.708 |
Rafael Sousa | 0.827 | 0.805 | 0.521 | 0.901 | 0.711 | 0.727 |
x j | 0.843 | 0.804 | 0.376 | 0.957 | 0.667 | 0.710 |
Cristina Vasconcelos | 0.830 | 0.791 | 0.171 | 0.990 | 0.581 | 0.660 |
Cristina Vasconcelos | 0.825 | 0.789 | 0.171 | 0.983 | 0.577 | 0.658 |
Euijoon Ahn | 0.805 | 0.786 | 0.009 | 0.998 | 0.504 | 0.614 |
Balázs Harangi | 0.828 | 0.783 | 0.470 | 0.915 | 0.693 | 0.701 |
Matt Berseth | 0.822 | 0.782 | 0.222 | 0.967 | 0.595 | 0.652 |
INESC Tecnalia | 0.480 | 0.765 | 0.906 | 0.377 | 0.642 | 0.601 |
Dylan Shen | 0.832 | 0.759 | 0.308 | 0.959 | 0.634 | 0.663 |
Vic Lee | 0.832 | 0.757 | 0.308 | 0.959 | 0.634 | 0.665 |
Masih Mahbod | 0.732 | 0.715 | 0.402 | 0.812 | 0.607 | 0.610 |
Dennis Murphree | 0.760 | 0.684 | 0.231 | 0.888 | 0.560 | 0.574 |
Hao Chang | 0.770 | 0.636 | 0.103 | 0.932 | 0.518 | 0.541 |
Jaisakthi S.M | 0.748 | 0.623 | 0.419 | 0.828 | 0.624 | 0.614 |
Wenhao Zhang | 0.805 | 0.500 | 0.000 | 1.000 | 0.500 | 0.581 |
Wiselin Jiji | 0.503 | 0.495 | 0.470 | 0.511 | 0.491 | 0.433 |
Yanzhi Song | 0.723 | 0.475 | 0.068 | 0.882 | 0.475 | 0.467 |
Bold values indicate the most relevant and representatives in this rearch work
ACC accuracy, AUC area under the RC curve, SE_M sensitivity, SP_M specificity for melanoma detection
The comparison between the proposed eVida model and the results reported by the RECOD Titans team (participant with the best previously reported AUC and ACC scores for melanoma detection) suggests an advantage for our proposed model since there is an increase in accuracy and balanced accuracy of 3.66% and 9.96% respectively for the test data. Meanwhile the reported AUC for both models is not significantly different. The proposed model also shows improvement over what is reported by Kazuhisa Matsunaga (Rank 1 in overall Lesion Classification), whose model presents the best balance between sensitivity and specificity among the previously reported results in melanoma detection for this challenge. Unlike the other reported models, in the proposed eVida model the specificity and sensitivity are both improved simultaneously to a high value (greater than 0.8), which indicates that the model is good for the detection of melanoma and non-melanoma (fewer overall errors). The metric called overall score in Table 4, which is defined as the average of the ACC, AUC, SE_M, and SP_M metrics, evidences the advantage of the eVida model over the previously reported models.
Figure 11 shows a ROC space in which the five models tested in this work are located with labels M1, M2, M3, M4, M5 and M6. The models proposed by Recod Titans and Kazuhisa Matsunaga are also located in the figure.
Bearing in mind that the best model is the one whose singular point in the ROC space is found closest to the coordinate (0, 1), where the perfect classifier is located, we then measured these Euclidean distances and the results listed in Table 5 were obtained, arranged from lower to higher. From those values it can be concluded that the best classifier is eVida M6.
Table 5.
Rank | Model | Distance between model and coordinate (0, 1) | |
---|---|---|---|
1 | Model 6 | 0.195 | |
2 | Model 2 | 0.297 | |
3 | Kazuhisa Matsunaga | 0.304 | |
4 | Model 4 | 0.388 | |
5 | Model 3 | 0.448 | |
6 | Recod Titans | 0.456 | |
7 | Model 5 | 0.609 | |
8 | Model 1 | 0.634 |
Discussion
According to the results from the different tests described in these work and by contrasting them with the appraisals highlighted in [11], an increase in the number of original images does not fully guarantee a better result in terms of classification metrics. This was the case when we added “malignant” skin lesion images from the PH2 data set and obtained lower sensitivity than that what was obtained without PH2 images. Therefore, we are of the opinion that to get an improvement in classifier performance, if we include new images then these need to be of the same nature, i.e. two databases that have not been subject to the same pre-processing are not necessarily complementary to each other.
The performance of image classification systems can be affected by common transformations or alterations in the images, such as: rotation, flipping, zooming, contrast / brightness adjustment, etc. [30]. By introducing these transformations on the training data set, it is possible to increase the number of training samples while improving the robustness of the model against such transformations [31]. This data augmentation can act as a regularizer to prevent or reduce overfitting in neural networks [31–34] and improve performance in unbalanced data set [35]. In our work, the use of data augmentation techniques from the transformation of the original images, allowed to increase the training balance index (Malignant/Benign training ratio) and produced a positive impact that is used in the training stage, as shown in models 2–4.
Artificial intelligence (AI) algorithms focus primarily on the injury image for which they were designed. Although the evaluation of the surrounding tissue is not necessarily a predictor of the disease, it is key when making a diagnosis and in some cases, establishing a prognosis. In the case of actinic keratosis, chronic exposure to ultraviolet light causes damage to the tissue surrounding the main lesion, however, a human inspector still focuses on the main lesion, leaving aside this peripheral damage, which would improve the frequency of correct diagnoses by 32.5% [36]. In the case of melanoma, there may be involvement of the surrounding tissue, manifesting with metastatic lesions in transit, which are of poor prognosis or with satellite lesions that are locoregional cutaneous manifestations of dissemination as a consequence of embolization of tumor cells between the primary tumor and the regional lymph node; therefore, these lesions are highly predictive of lymphatic invasion and predict the development of disseminated disease [37, 38].
Morevoer, Tschandl et al. [36] showed that in Keratoses actinic detection, the surrounding area have clues for diagnosis, however each disease is described with different techniques and characteristics [39]. Thus, although the importance of perilesional skin involvement is clear, the objective of our study is to identify a lesion corresponding to melanoma and does not seek to establish a prognosis of the disease. Based on ABCDE [40] rule we can assume the study is focused only on the probable Melanoma lesion area because the relevant data to differentiate from other diseases is concentrated here, we make our model focus in this region by cropping the background before performing the classification task.
Conclusions
The comparative analysis shown in "Discussion" section suggests a better performance for the proposed eVida M6 model over the methods previously proposed for the ISIC 2017 challenge. Automatic extraction of the region of interest within a dermatoscopic image suggests a significant improvement in classifier performance by eliminating pixels from the image that do not provide the classifier with lesion information.
The experiments in data augmentation described in "Results" section show that reducing data imbalance may be helpful to improve classification performance, but careful fine tuning is required, e.g. making training data perfectly balanced not necessarily results on a better model.
Finally, the eVida M6 model is a reliable predictor with an excellent balance between overall accuracy (0.904), sensitivity (0.820) and specificity (0.925). Within the set of tested models and also comparing with the best results reported in melanoma detection for the ISIC 2017 challenge, eVida M6 is an improvement in the state-of-the-art for automatic diagnosis of melanoma from dermatoscopic images.
Acknowledgements
Not applicable.
Abbreviations
- Mask R_CNN
Mask and Region-based Convolutional Neural Network
- CNN
Convolutional Neural Network
- ACS
American Cancer Society
- ROI
Region Of Interest
- ROC
Receiver Operating Characteristic
- ISIC
International Skin Imaging Collaboration
- ISBI
International Symposium on Biomedical Imaging
- AUC_M
Area Under the RC Curve for Melanoma detection
- ACC_M
Accuracy for Melanoma detection
- SE_M
Sensitivity for Melanoma detection
- SP_M
Specificity for Melanoma detection
Authors’ contributions
MJ programmed and built the proposed algorithms to the classification task of melanoma and no-melanoma images and he helped to design the structure of the proposed solution Mask RCNN to focus the interest area in the image. LC collected the data and programmed the algorithms to organize it and she helped to design the structure of the proposed solution Mask RCNN+ CNN. BG designed the general solution of the problem and she contributed much to get the deep learning final structure of the solution. And WP contributed to build the general structure of the solution of the problem and helped to adjust the hyperparameters to improve the performance. All authors read and approved the final manuscript.
Funding
Not applicable.
Availability of data and materials
The datasets analysed during the current study are available in the ISIC repository, in https://www.isic-archive.com/#!/topWithHeader/onlyHeaderTop/gallery [24]/“ISIC Archive”/Accessed: 04-Jul-2020. The other dataset is available in the PH2 repository, in https://www.fc.up.pt/addi/ph2 database.html [28]/“ADDI—Automatic computer-based Diagnosis system for Dermoscopy Images” Accessed: 24-Jul-2019. Public access to the databases is open.
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Mario Fernando Jojoa Acosta, Email: mario.jojoa@opendeusto.es.
Liesle Yail Caballero Tovar, Email: lieslec@uninorte.edu.co.
Maria Begonya Garcia-Zapirain, Email: mbgarciazapi@deusto.es.
Winston Spencer Percybrooks, Email: wpercyb@uninorte.edu.co.
References
- 1.Mayo Foundation for Medical Education and Research. Melanoma: descripción general. 2019. https://www.mayoclinic.org/es-es/diseases-conditions/melanoma/symptoms-causes/syc-20374884. Accessed 15 July 2019.
- 2.Skin Cancer (Including Melanoma)—Patient Version—National Cancer Institute. 2019. https://www.cancer.gov/types/skin. Accessed 15 July 2019.
- 3.Detect skin cancer|American Academy of Dermatology. 2019. https://www.aad.org/public/spot-skin-cancer/learn-about-skin-cancer/detect. Accessed 15 July 2019.
- 4.Puede detectarse temprano el cáncer de piel tipo melanoma. 2019. https://www.cancer.org/es/cancer/cancer-de-piel-tipo-melanoma/deteccion-diagnostico-clasificacion-por-etapas/deteccion.html. Accessed 15 July 2019.
- 5.Lunares comunes, nevos displásicos y el riesgo de melanoma-Instituto Nacional Del Cáncer. 2019. https://www.cancer.gov/espanol/tipos/piel/hoja-informativa-lunares. Accessed 15 July 2019.
- 6.Maarouf M, Costello CM, Gonzalez S, Angulo I, Curiel-Lewandrowski CN, Shi VY. In vivo reflectance confocal microscopy: emerging role in noninvasive diagnosis and monitoring of eczematous dermatoses. Actas Dermo Sifiliog (Engl Ed) 2019;110(8):626–636. doi: 10.1016/j.adengl.2019.06.005. [DOI] [PubMed] [Google Scholar]
- 7.Ferrante di Ruffano L, et al. Computer-assisted diagnosis techniques (dermoscopy and spectroscopy-based) for diagnosing skin cancer in adults. Cochrane Database Syst Rev. 2018 doi: 10.1002/14651858.CD013186. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Dinnes J, et al. Reflectance confocal microscopy for diagnosing cutaneous melanoma in adults. Cochrane Database Syst Rev. 2018 doi: 10.1002/14651858.CD013190. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.di Ruano FL. Dermoscopy, with and without visual inspection, for diagnosing melanoma in adults (review) Cochrane Libr Cochrane Database Syst Rev. 2018 doi: 10.1002/14651858.CD011902.pub2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Dinnes J, et al. Visual inspection for diagnosing cutaneous melanoma in adults. Cochrane Database Syst Rev. 2018 doi: 10.1002/14651858.CD013194. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Harangi B. Skin lesion classification with ensembles of deep convolutional neural networks. J Biomed Inform. 2018;86:25–32. doi: 10.1016/j.jbi.2018.08.006. [DOI] [PubMed] [Google Scholar]
- 12.Nida N, Irtaza A, Javed A, Yousaf MH, Mahmood MT. Melanoma lesion detection and segmentation using deep region based convolutional neural network and fuzzy C-means clustering. Int J Med Inform. 2019 doi: 10.1016/j.ijmedinf.2019.01.005. [DOI] [PubMed] [Google Scholar]
- 13.Goyal M, Yap MH. Automatic lesion boundary segmentation in dermoscopic images with ensemble deep learning methods. 2019. http://arxiv.org/abs/1902.00809. Accessed 16 July 2019.
- 14.Jafari MH, Nasr-Esfahani E, Karimi N, Soroushmehr SMR, Samavi S, Najarian K. Extraction of skin lesions from non-dermoscopic images using deep learning. Int J Comput Assist Radiol Surg. 2016 doi: 10.1007/s11548-017-1567-8. [DOI] [PubMed] [Google Scholar]
- 15.CNN Architectures: LeNet, AlexNet, VGG, GoogLeNet, ResNet and more…. 2019. https://medium.com/@sidereal/cnns-architectures-lenet-alexnet-vgg-googlenet-resnet-and-more-666091488df5. Accessed 20 July 2019.
- 16.Guo K, Xu T, Kui X, Zhang R, Chi T. iFusion: towards efficient intelligence fusion for deep learning from real-time and heterogeneous data. Inf Fusion. 2019;51:215–223. doi: 10.1016/j.inffus.2019.02.008. [DOI] [Google Scholar]
- 17.Garcia-Arroyo JL, Garcia-Zapirain B. Segmentation of skin lesions in dermoscopy images using fuzzy classification of pixels and histogram thresholding. Comput Methods Programs Biomed. 2019 doi: 10.1016/j.cmpb.2018.11.001. [DOI] [PubMed] [Google Scholar]
- 18.Garnavi R, Aldeen M, Celebi ME, Bhuiyan A, Dolianitis C, Varigos G. Automatic segmentation of dermoscopy images using histogram thresholding on optimal color channels. Int J Med Med Sci. 2011;1:126–134. [Google Scholar]
- 19.Schaefer G, Krawczyk B, Celebi ME, Iyatomi H. An ensemble classification approach for melanoma diagnosis. Memetic Comput. 2014 doi: 10.1007/s12293-014-0144-8. [DOI] [Google Scholar]
- 20.Brodersen KH, Ong CS, Stephan KE, Buhmann JM. The balanced accuracy and its posterior distribution. In: 20th International conference on pattern recognition. 2010; p. 3121–24. 10.1109/ICPR.2010.764.
- 21.R. J. Urbanowicz, J. H. Moore (2015) ExSTraCS 2.0: description and evaluation of a scalable learning classifier system. Evol Intell 8(2–3): 89–116. DOI: 10.1007/s12065-015-0128-8. [DOI] [PMC free article] [PubMed]
- 22.Velez DR, et al. A balanced accuracy function for epistasis modeling in imbalanced datasets using multifactor dimensionality reduction. Genet Epidemiol. 2007;31(4):306–315. doi: 10.1002/gepi.20211. [DOI] [PubMed] [Google Scholar]
- 23.Fawcett T. An introduction to ROC analysis. Pattern Recognit Lett. 2006 doi: 10.1016/j.patrec.2005.10.010. [DOI] [Google Scholar]
- 24.ISIC Archive. https://www.isic-archive.com/#!/topWithHeader/onlyHeaderTop/gallery. Accessed 04 July 2020.
- 25.ISBI 2017. https://biomedicalimaging.org/2017/challenges/. Accessed 22 July 2019.
- 26.ImageNet. http://www.image-net.org/. Accessed 22 July 2019.
- 27.Smith LN. A disciplined approach to neural network hyper-parameters: part 1—learning rate, batch size, momentum, and weight decay. 2018. http://arxiv.org/abs/1803.09820. Accessed 19 July 2019.
- 28.ADDI—Automatic computer-based Diagnosis system for Dermoscopy Images. https://www.fc.up.pt/addi/ph2 database.html. Accessed 24 July 2019.
- 29.“Covalic,” ISIC 2017: skin lesion analysis towards melanoma detection. International Skin Imaging Collaboration. 2017. https://challenge.kitware.com/#phase/584b0afccad3a51cc66c8e38. Accessed 12 May 2020.
- 30.Navarrete-Dechent C, Dusza SW, Liopyris K, Marghoob AA, Halpern AC, Marchetti MA. Automated dermatological diagnosis: hype or reality? J Invest Dermatol. 2018;138(10):2277–2279. doi: 10.1016/j.jid.2018.04.040. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Wong SC, Gatt A, Stamatescu V, McDonnell MD. Understanding data augmentation for classification: when to warp? 2016. 10.1109/DICTA.2016.7797091.
- 32.Xiang A, Wang F. Towards interpretable skin lesion classification with deep learning models. In: AMIA annual symposium proceedings, vol. 2019, p. 1246–55. /pmc/articles/PMC7153112/?report=abstract. Accessed 05 Nov 2020 [PMC free article] [PubMed]
- 33.Simard PY, Steinkraus D, Platt JC. Best practices for convolutional neural networks applied to visual document analysis. 2003. 10.1109/ICDAR.2003.1227801.
- 34.Cireşan DC, Meier U, Gambardella LM, Schmidhuber J. Deep, big, simple neural nets for handwritten digit recognition. Neural Comput. 2010 doi: 10.1162/NECO_a_00052. [DOI] [PubMed] [Google Scholar]
- 35.Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE: synthetic minority over-sampling technique. J Artif Intell Res. 2002 doi: 10.1613/jair.953. [DOI] [Google Scholar]
- 36.Tschandl P, et al. Human–computer collaboration for skin cancer recognition. Nat Med. 2020 doi: 10.1038/s41591-020-0942-0. [DOI] [PubMed] [Google Scholar]
- 37.Borgstein PJ, Meijer S, Van Diest PJ. Are locoregional cutaneous metastases in melanoma predictable. Ann Surg Oncol. 1999;6(3):315–321. doi: 10.1007/s10434-999-0315-x. [DOI] [PubMed] [Google Scholar]
- 38.Bann DV, Chaikhoutdinov I, Zhu J, Andrews G. Satellite and in-transit metastatic disease in melanoma skin cancer: a retrospective review of disease presentation, treatment, and outcomes. Dermatol Surg. 2019 doi: 10.1097/DSS.0000000000001643. [DOI] [PubMed] [Google Scholar]
- 39.Dearborn FM. Enfermedades De La Piel - Frederik M. Dearborn - Google Libros, Jain Publishers, 1999. https://books.google.com.co/books/about/Enfermedades_De_La_Piel.html?id=Db22rHD5nGsC&redir_esc=y. Accessed 04 Nov 2020.
- 40.Tsao H, et al. Early detection of melanoma: reviewing the ABCDEs. J Am Acad Dermatol. 2015 doi: 10.1016/j.jaad.2015.01.025. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The datasets analysed during the current study are available in the ISIC repository, in https://www.isic-archive.com/#!/topWithHeader/onlyHeaderTop/gallery [24]/“ISIC Archive”/Accessed: 04-Jul-2020. The other dataset is available in the PH2 repository, in https://www.fc.up.pt/addi/ph2 database.html [28]/“ADDI—Automatic computer-based Diagnosis system for Dermoscopy Images” Accessed: 24-Jul-2019. Public access to the databases is open.