Skip to Content
DiagnosticsDiagnostics
  • Article
  • Open Access

22 May 2023

Diagnosing Melanomas in Dermoscopy Images Using Deep Learning

,
,
and
1
Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakakah 72341, Saudi Arabia
2
Department of Electrical Engineering, Shoubra Faculty of Engineering, Benha University, Cairo 11672, Egypt
3
Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakakah 72341, Saudi Arabia
4
School of Computer Science (SCS), Taylor’s University, Subang Jaya 47500, Malaysia
This article belongs to the Special Issue Imaging Diagnosis for Melanoma 2.0

Abstract

When it comes to skin tumors and cancers, melanoma ranks among the most prevalent and deadly. With the advancement of deep learning and computer vision, it is now possible to quickly and accurately determine whether or not a patient has malignancy. This is significant since a prompt identification greatly decreases the likelihood of a fatal outcome. Artificial intelligence has the potential to improve healthcare in many ways, including melanoma diagnosis. In a nutshell, this research employed an Inception-V3 and InceptionResnet-V2 strategy for melanoma recognition. The feature extraction layers that were previously frozen were fine-tuned after the newly added top layers were trained. This study used data from the HAM10000 dataset, which included an unrepresentative sample of seven different forms of skin cancer. To fix the discrepancy, we utilized data augmentation. The proposed models outperformed the results of the previous investigation with an effectiveness of 0.89 for Inception-V3 and 0.91 for InceptionResnet-V2.

1. Introduction

Among all cancers, skin tumors have the highest potential for malignancy [1,2]. Melanoma (MEL), basal cell carcinoma (BCC), squamous cell carcinoma (SCC), and non-melanoma comprise the most common types of skin cancer. However, actinic keratosis (AK), Kaposi sarcoma (KS), lymphoma, and keratoacanthoma [3,4] are exceptionally rare forms of cancer. Figure 1 depicts an upward trend in the incidence of all types of skin cancer.
Figure 1. The prevalence of several forms of skin cancer (derived from Ref. [5]).
The majority of cases of skin cancer fall into two categories: melanoma and non-melanoma [1,6,7]. Cancer mortality and medical expenditures are both increased by the presence of malignant lesions, so scientists have been working on techniques that can identify pre-malignant skin lesions with a high degree of accuracy and flexibility. Due to their rapid proliferation, invasion, and dissemination, malignant melanocyte cells must be identified at an early stage [8]. Dermoscopy is commonly used by specialists to identify the malignancy or benignity of a skin lesion.
The term “dermoscopy” refers to a method that makes use of a magnifying lens and a light source to enhance the visibility of various medical features [9,10]. It reveals the hidden morphologies to the naked eye. The ABCD rule [11], a seven-point checklist [12], and pattern analysis [13] are just some of the methods that have been established to enhance the precision of skin melanoma identification. However, dermoscopy images taken by laypeople have a prognostic validity of 75% to 84% for melanoma; the processing is time-consuming and subjective, depending on the dermatologist’s skill [14]. Professionals have created computer-aided diagnosis (CAD) methods to help them work around these problems [10,14]. Advances in computer-aided cancer diagnosis can be largely attributed to deep learning (DL)-based AI [15,16].
Applying DL methods in skin lesion classification helps automate the screening and early diagnosis of skin cancer, even in areas without easy access to dermatologists or laboratories [17]. Traditional classifiers [18,19] require feature extraction performed by humans before being supplied into computer-aided dermoscopy image processing.
In this research, various models were used (including InceptionV3 and Inception-Resnet). InceptionResnet-V2 incorporates the residual connections within the Inception design. For this reason, Inception-Resnet is the most reliable option. Computational effectiveness is poor and fewer parameters are realized in the Inception-Resnet model. Additionally, it provides a high-performance boost, efficient use of computing resources, and a slightly higher computation load, all of which contribute to the Inception-Resnet network’s high-performance output.
Our examination of deep neural network (DNN) classification performance on the HAM10000 dataset led us to the conclusion that the employed InceptionResnet-V2 model was more accurate at classification than competing DNNs. This means that individual investigations may be needed to determine which network works best with specific medical imaging datasets.
As a consequence, the following are the paper’s primary contributions:
  • An enhanced super-resolution generative adversarial network (ESRGAN) was employed using 10,000 training photos to generate high-quality images for the Human against Machine dataset (HAM10000 dataset [20]).
  • Because the HAM10000 dataset contains uneven data, we employed augmentation to balance the data throughout all classes.
  • A deep comparative evaluation employing several assessment parameters such as accuracy, specificity, sensitivity, a confusion matrix, and the F1-score established whether the proposed system was feasible.
  • Together with Inception-V3 and InceptionResnet-V2, it was used to fine-tune the weights of HAM10000-trained networks.
  • To boost the recommended method’s scalability and safeguard against overfitting, we used a supplementary training procedure assisted by several training strategy variations (e.g., data augmentation, learning rate, batch size, and validation patience).
For the identification of numerous skin lesions, this research proposed an optimization technique involving transfer learning models. For this, we used Inception-V3 and InceptionResNet-V2 architectures to pre-train the weights of each model. The HAM10000 dataset, which contains images of skin lesions, was utilized to compare the models’ results. The dataset’s class imbalance necessitated an oversampling technique. The rest of the paper is organized as follows. Section 2 displays the cited references. In Section 3, we detail the dataset, the analysis methods employed, and the proposed methodology. The findings of the experiments are presented in Section 4, and their analysis follows in Section 5. In Section 6, we provide a summary.

3. Proposed Methodology

Throughout this section, we go into painstaking depth regarding our methodology. As illustrated in Figure 2, there were three primary steps to this method. Pre-processing skin photos improved their quality for categorization by removing minute noise. The skin image’s edges were then delineated using ground truth, and the backdrop was then removed. Then, we applied many distinct segmented images in the transfer learning models. Automated learning of informative representations from skin images was accomplished using transfer learning models. In the following sections, we will outline the details of the proposed method.
Figure 2. Proposed Approach for Identifying Skin Cancer.

3.1. HAM10000 Dataset

The identification performance of the proposed models was assessed using the HAM10000 [20] dataset of lesions with pigmentation. The authors compiled dermatoscopic images from various demographics that were captured and stored employing various techniques. As a training set for scholarly use, it was anticipated to contain 10015 dermatoscopic images divided among seven classes, as seen in Figure 3.
Figure 3. Classes of HAM10000 dataset.
Histopathology was employed to verify the diagnosis in over 50 percent of the cases; in the other half, the truth was established through a subsequent examination and the agreement of experts. The HAM10000 metadata file contains a lesion-id column that may be utilized to track the position of the lesion across the many photos that comprise the dataset.
The HAM10000 dataset was separated into training, validation, and testing data as illustrated in Table 2, where 327 were Akiec, 514 were Bcc, 1113 were Mel, 142 were Vasc, 6705 were Nv, 115 were Df, and 1099 were bkl.
Table 2. Inadequate sample size prior to using augmentation techniques.

3.2. Image Pre-Processing Step

Maximizing dermoscopy image resolution and removing various types of noise from images of skin lesions were also essential steps in bringing the proposed strategy into practice. A high-quality image is essential for developing a trustworthy model for the classification of skin lesions. Image segmentation, enhancement, data supplementation, resizing, and normalization are all a part of this process. Overfitting occurs when there are more variables to learn than can reasonably be accounted for due to an increase in network complexity in the model. Overfitting due to a small and uneven sample size of training photos was addressed by splitting the HAM10000 dataset into three distinct parts (e.g., training, verification, and assessment), and then data augmentation was employed to even out the forecasting capability across the board. Masks for rotation, reflection, shifting, and scaling are provided alongside the enhanced images for each image in the dataset.

3.3. ESRGAN

Ledig et al. [33] developed an Enhanced Super-Resolution Generative Adversarial Network (ESRGAN). By training generators and discriminators against each other, such technique, which is based on adversarial learning, generates texture features that are compatible with the distribution of genuine images. By extracting visual features at a single scale, this method achieves super-resolution. An image super-resolution rebuilding may be employed to determine the corresponding image of high quality from one or more photographs of different resolutions as shown in Figure 4. Examples of effective applications of this technology include remote sensing, medical imaging, picture compression, video monitoring, and military applications. Many academics have focused on image super-resolution since it is an important image-processing technology, and there have been numerous effective ways of achieving image super-resolution proposed [33].
Figure 4. Results of the suggested image-enhancement technology: (a) unprocessed image; (b) processed image.

3.4. Segmentation

The dermoscopy images were processed according to the image preparation technique in order to extract the ROI. The original images were multiplied by the images that reflected their respective ground truths in order to create the ROIs displayed in Figure 5.
Figure 5. Main image, ground truth, and segmented ROI examples.

3.5. Data Augmentation

Prior to actually presenting the DNN with the original dataset images, we added more data to the training set as a preliminary step. Data augmentation is the most common way to avoid overfitting when training models. This is done by adding new images to the dataset in a way that retains the class information [8]. The main idea is that changes that can be made to replicable data do not change what the image means, so new samples can be made. Most of the time, increasing the amount of training data for DL models improves their performance. When it comes to dermatological images, we can make a lot of adjustments to each one by using what makes it distinctive. For these images, fading, flipping either vertically or horizontally, or spinning the images by a specific angle does not affect the performance of deep NNs. As ways to improve the data, horizontal flips; random rotations between 90 and 270 degrees; and changes to the saturation, exposure, and hue values of the original images were chosen. In particular, these parameters were set to have values of 1.5, 1.5, and 0.1, respectively. Each image in the training set was subjected to the changes listed above, giving the network a new sample. In this study, we used the ImageDataGenerator interface of Tensorflow2.0 to flip, rotate, and shift the input image data. Adding significantly modified versions of existing data or new synthetic data derived from existing data is how data augmentation works to increase the total amount of available data, as seen in Figure 6.
Figure 6. Image augmentation examples for the exact same image.
It was crucial that the DL models be trained on a large number of photos that were fairly distributed because there was a noticeable disparity in the dataset (see Table 2). As demonstrated in Table 3, if we followed this reasoning and applied augmentation (oversampling) processes to the relevant classes after the dataset had been balanced, we obtained well-balanced data with about equal numbers of images in each category.
Table 3. Equitable dataset using augmentation (oversampling) with segmented images.

3.6. Transfer Models for Learning

To solve a new problem that is similar to the original problem but unique, ML practitioners can engage in a process known as transfer learning. In comparison to the conventional neural network, it simply requires a small amount of training data and a quick training period to attain a high accuracy. A discussion of the transfer models employed is presented in the following subsections.

3.7. Model Training Utilizing Inception-V3

The theoretical underpinnings of the method are laid out and described in this subsection. Among these transfer learning pre-trained deep models is Inception-v3 [11,12], which builds on the design of its predecessors Inception-v1 [34,35] and Inception-v2 [36,37]. ImageNet datasets [38,39] have been used to train the Inception-v3 model, which then has been utilized to recognize a hundred distinct classes. Error rates have dropped from 17.3% for the best system to 3.5% for the top five in ImageNet.
In particular, the technique developed by Serre et al. [40]—which may involve multiple levels of processing—served as an inspiration for Inception. Using the method proposed by Lin et al. [41], the creators of Inception increased the accuracy of the neural networks used in the system. They were immune to computational limitations because the dimensions had been shrunk to 11 convolutions. Using Inception [42], scholars have significantly reduced the period and effort needed for DL picture classification. They aimed to strike a compromise between the two commonly utilized techniques for boosting performance; namely, expanding in depth and width and separating data into distinct layers without resorting to any empirical evidence or empirical analysis. The Inception DL system’s 22-layer architecture, in which every filter is a learned one, was designed with this specific end in mind. Input into the next layer was generated from highly correlated categories using a correlation statistical analysis based on the work of Arora et al. [43]. Eventually, after dimension reduction, every one of these layers is reduced to a series of 1x1 convolutions [40].

3.8. Model Training Utilizing Inception-Resnet

State-of-the-art performance was achieved in the 2015 ILSVRC challenge by including residual connections in a more conventional architecture; the resulting network was competitive with the most recent generation of the Inception-V3 network. In light of this, one may reasonably speculate as to whether combining the Inception architecture with residual blocks might yield any beneficial results. Regarding Inception network training using residual links, the training time for InceptioResnet-V2 [43] was proven to be drastically reduced. Even though Inception networks with residual connections are more expensive, there is some evidence that they outperform those with no residual connections. Szegedy et al. [44] introduced various novel, simplified topologies for Inception networks (both residual and non-residual). The ILSVRC 2012 classification challenge benefited greatly from these improvements when applied to single-frame identification.

4. Experiments and Results

In this section, we begin by giving an overview of the experimental setup by describing the models and parameters used, the datasets we utilized, and the hardware specifications of the machines used to conduct the experiments. The outcomes for each dataset and model comparison follow.

4.1. Instruction and Deployment of Inception-v3 and Inception-Resnet-V2

Transfer learning models were tested on the HAM10000 dataset and evaluated by comparing them to the best practices; 90% of the data (9016 photos) was for training and 10% was for testing (984 images). The validation procedure employed 10% of the training set (992). Each image was reduced to 227 by 227 by 3 pixels and magnified to a size of 39,430 throughout the training phase. An RTX3060-equipped Linux PC with 8 GB of RAM was used to test the TensorFlow Keras. Eighty percent of all TL models were trained on an uninformed collection of photos. After training, a validation set comprising 10% of the data was used to ensure that only the most accurate weight combinations were kept. Approved models were pre-trained on the HAM10000 dataset using the Adam optimizer and a learning rate approach that reduced the learning rate while the model was inactive for the validation patience. The following hyperparameters were passed to Adam during training: 0.90 momentum, 10 patience, and 50 epochs. The range of batch sizes ranged from 2 to 64, and each increment was double the preceding value. As part of our arsenal of methods to stop the spread of infectious forms, we also used a method called “batching”.

4.2. Criteria for Assessment

Throughout this subsection, we describe the study’s evaluation measures and the outcomes. The classifier effectiveness is a widely used measure of classification performance (Ac). Equation (1) depicts the formula, which was calculated by dividing the total number of examples by the proportion of correct identifications. Typically, the performance of image-classification systems is measured by their sensitivity and specificity. Equation (2) presents a formula for specificity that becomes more accurate as more photos are accurately labeled. The number of pictures in the dataset that had a linear relationship was determined according to Equation (3). More accurate predictions can be expected from a system with a higher F-score. Accuracy and sensitivity are not enough to determine a system’s value. The formula for determining the F-score (Fsc) is stated in Equation (4). The top N accuracy was the fourth measurement, and it refers to how well the model N’s highest likelihood responses fit the anticipated softmax distribution. If one of the N predictions was the correct label, then the classification was valid.
A c = T p + T n T p + T n + F p   + F n
S p e c i f i c i t y = T n T n + F p  
S e n s i t i v i t y = T p T p + F n  
F s c = 2 P r e R e c P r e + R e c
The abbreviation Tp indicates “true positive”, Tn indicates a true negative, Fp indicates a false positive, and Fn indicates a false negative.

4.3. Effectiveness among Several DNN Models

Several TL classification methods were trained and validated using the HAM10000 skin lesion identification challenge dataset (including Inception-V3 and InceptionResnet-V2). The HAM10000 dataset was evaluated numerous times using a 90/10 crossover among training and testing, and the findings are given below. This split was decided upon to reduce the overall time required to finish the job. Classifiers were trained with a batch size between 2 and 64 and learning rates of 1E^4, 1E^5, and 1E^6 for Inception-V3 and InceptionResnet-V2. Freezing different numbers of layers allowed Inception-V3 and InceptionResnet-V2 to be fine-tuned for optimal accuracy. An ensemble of models was generated by repeatedly running the same model with the same set of parameters. The accuracy varied from run to run because the weights were generated randomly. For Inception-V3 and InceptionResnet-V2 training on the HAM10000 dataset, only the maximum run results were maintained; these are reported in Table 4 and Table 5, respectively. The tables demonstrate that the highest success rates for Inception-V3 and InceptionResnet-V2 were 89.7% and 90.1%, respectively. The confusion matrices created by Inception-V3 and InceptionResnet-V2 are depicted in Figure 7 and Figure 8, respectively.
Table 4. Highest Effectiveness via Finest Tuning with Inception-V3.
Table 5. Highest Effectiveness via Finest Tuning with InceptionResnet-V2.
Figure 7. The Inception-V3 Confusion Matrix Top Performer.
Figure 8. The InceptionResnet-V2 Confusion Matrix Top Performer.
Table 6 and Table 7 indicate the total amount of test photos for every class in the HAM10000 dataset. With 95% specificity, 98% sensitivity, and 96% Fsc when using the Inception-V3 learning model and 94% specificity, 99% sensitivity, and 97% Fsc when using the modified InceptionResnet-V2 learning model, it was clear that the Nv class had the most data points out of all the classes tested (795 total).
Table 6. Detailed findings produced for every category by the Inception-V3 learning model.
Table 7. Detailed findings produced for every category by the InceptionResnet-V2 learning model.
It was demonstrated that dermatologists can benefit from using lesion photos to improve the accuracy of infection diagnoses and decrease their burden.

4.4. Comparison to Alternative Approaches

Comparisons of effectiveness with respect to competing methods are shown in Table 8. According to Table 8, our method was more productive and efficient than competing methods. Overall, the accuracy rate of the suggested InceptionResnet-V2 model was higher than that of state-of-the-art methods by 91.26%. ESRGAN’s greater overall resolution constituted the main reason for the improvement’s success.
Table 8. Comparison with other methods.

5. Discussion

The results of our analysis demonstrated that the alternative approaches were significantly less accurate. We ascribe the enhancement to a trifecta of factors, one of which was ESRGAN’s higher overall resolution. Moreover, we employed a variety of architectures, some of which were more effective than others at generalizing and adapting to various types of data. Medical image categorization could not be enhanced by transfer learning architectures due to the lack of distinguishing features. While Inception-V3 excelled at image detection, it underperformed the proposed InceptionResnet-V2 when applied to medical images. Due to their lack of semantic meaning for real images, InceptionResnet-V2 features are more flexible and generally applicable in medical imaging (compared to Inception-V2). As a result, fine-tuning enhanced the precision of the two models. Deep networks were found to be superior to their shallow counterparts when it came to distinguishing crucial elements when trained on a smaller sample. The successful outcomes of these methods are shown in Figure 7 and Figure 8.

6. Conclusions

It is now possible to diagnose seven different types of cancer quickly and accurately through the study of skin blemishes. To improve the contrast of the lesion image and get rid of noise, the proposed method made use of image-enhancing techniques. To avoid overfitting and increase the capabilities of the suggested DL approaches, Inception-V3 and InceptionResnet-V2 were trained on the forefront of pre-processed lesions using augmentation procedures. The presented approach was evaluated by employing images of lesions from the HAM10000 dataset. It was speculated that the Inception-V3 and InceptionResnet-V2 versions of the conception model had accuracy rates similar to those of dermatology specialists with board certification. The study’s originality and novelty also reside in the fact that it used ESRGAN as a pre-processing step for the numerous models it employed (Inception-V3 and InceptionResnet-V2). Our freshly trained model performed as well as, if not better than, the baseline. Tests comparing the suggested system to others showed that it outperformed the competition with 91.26% accuracy. The technique’s viability can only be established by putting it through its paces on a huge dataset basically including a sizable number of future cancer predictions. DenseNet, VGG, and AlexNet are three techniques that show promise for future analysis of the cancer dataset.

Author Contributions

Conceptualization, M.H.; Data curation, G.A.; Formal analysis, W.G.; Funding acquisition, G.A.; Investigation, W.G.; Methodology, G.A.; Project administration, N.Z.J.; Supervision, M.H. and N.Z.J.; Writing—original draft, W.G.; Writing—review and editing, M.H. and N.Z.J. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deputyship for Research and Innovation, Ministry of Education in Saudi Arabia for funding this research work through project number 223202.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Will be furnished on request.

Acknowledgments

The authors extend their appreciation to the Deputyship for Research and Innovation, Ministry of Education in Saudi Arabia for funding this research work through project number 223202.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Saeed, J.; Zeebaree, S. Skin Lesion Classification Based on Deep Convolutional Neural Networks Architectures. J. Appl. Sci. Technol. Trends 2021, 2, 41–51. [Google Scholar] [CrossRef]
  2. Albahar, M.A. Skin Lesion Classification Using Convolutional Neural Network With Novel Regularizer. IEEE Access 2019, 7, 38306–38313. [Google Scholar] [CrossRef]
  3. Khan, I.U.; Aslam, N.; Anwar, T.; Aljameel, S.S.; Ullah, M.; Khan, R.; Rehman, A.; Akhtar, N. Remote Diagnosis and Triaging Model for Skin Cancer Using EfficientNet and Extreme Gradient Boosting. Complexity 2021, 2021, 5591614. [Google Scholar] [CrossRef]
  4. Gouda, W.; Sama, N.U.; Al-Waakid, G.; Humayun, M.; Jhanjhi, N.Z. Detection of Skin Cancer Based on Skin Lesion Images Using Deep Learning. Healthcare 2022, 10, 1183. [Google Scholar] [CrossRef] [PubMed]
  5. Nikitkina, A.I.; Bikmulina, P.Y.; Gafarova, E.R.; Kosheleva, N.V.; Efremov, Y.M.; Bezrukov, E.A.; Butnaru, D.V.; Dolganova, I.N.; Chernomyrdin, N.V.; Cherkasova, O.P.; et al. Terahertz radiation and the skin: A review. J. Biomed. Opt. 2021, 26, 043005. [Google Scholar] [CrossRef]
  6. Xu, H.; Lu, C.; Berendt, R.; Jha, N.; Mandal, M. Automated analysis and classification of melanocytic tumor on skin whole slide images. Comput. Med. Imaging Graph. 2018, 66, 124–134. [Google Scholar] [CrossRef]
  7. Alwakid, G.; Gouda, W.; Humayun, M.; Sama, N.U. Melanoma Detection Using Deep Learning-Based Classifications. Healthcare 2022, 10, 2481. [Google Scholar] [CrossRef]
  8. Namozov, A.; Cho, Y.I. Convolutional neural network algorithm with parameterized activation function for melanoma classification. In Proceedings of the 2018 International Conference on Information and Communication Technology Convergence (ICTC), Jeju, Republic of Korea, 17–19 October 2018. [Google Scholar]
  9. Ozkan, I.A.; Koklu, M. Skin lesion classification using machine learning algorithms. Int. J. Intell. Syst. Appl. Eng. 2017, 5, 285–289. [Google Scholar] [CrossRef]
  10. Thamizhamuthu, R.; Manjula, D. Skin Melanoma Classification System Using Deep Learning. Comput. Mater. Contin. 2021, 68, 1147–1160. [Google Scholar] [CrossRef]
  11. Stolz, W. ABCD rule of dermatoscopy: A new practical method for early recognition of malignant melanoma. Eur. J. Dermatol. 1994, 4, 521–527. [Google Scholar]
  12. Argenziano, G.; Fabbrocini, G.; Garli, P.; De Gorge, V.; Sammarco, E.; Delfino, M. Epiluminescence microscopy for the diagnosis of doubtful melanocytic skin lesions: Comparison of the ABCD rule of dermatoscopy and a new 7-point checklist based on pattern analysis. Arch. Dermatol. 1998, 134, 1563–1570. [Google Scholar] [CrossRef] [PubMed]
  13. Pehamberger, H.; Steiner, A.; Wolff, K. In vivo epiluminescence microscopy of pigmented skin lesions. I. Pattern analysis of pigmented skin lesions. J. Am. Acad. Dermatol. 1987, 17, 571–583. [Google Scholar] [CrossRef] [PubMed]
  14. Reshma, G.; Al-Atroshi, C.; Nassa, V.K.; Geetha, B.; Sunitha, G.; Galety, M.G.; Neelakandan, S. Deep Learning-Based Skin Lesion Diagnosis Model Using Dermoscopic Images. Intell. Autom. Soft Comput. 2022, 31, 621–634. [Google Scholar] [CrossRef]
  15. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef]
  16. Yao, P.; Shen, S.; Xu, M.; Liu, P.; Zhang, F.; Xing, J.; Shao, P.; Kaffenberger, B.; Xu, R.X. Single model deep learning on imbalanced small datasets for skin lesion classification. arXiv 2021, arXiv:2102.01284. [Google Scholar] [CrossRef]
  17. Adegun, A.; Viriri, S. Deep learning techniques for skin lesion analysis and melanoma cancer detection: A survey of state-of-the-art. Artif. Intell. Rev. 2020, 54, 811–841. [Google Scholar] [CrossRef]
  18. Yang, J.; Sun, X.; Liang, J.; Rosin, P.L. Clinical skin lesion diagnosis using representations inspired by dermatologist criteria. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  19. Satheesha, T.; Satyanarayana, D.; Prasad, M.N.G.; Dhruve, D.K. Melanoma is skin deep: A 3D reconstruction technique for computerized dermoscopic skin lesion classifi-cation. IEEE J. Transl. Eng. Health Med. 2017, 5, 1–17. [Google Scholar] [CrossRef]
  20. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 180161. [Google Scholar] [CrossRef]
  21. Lembhe, A.; Motarwar, P.; Patil, R.; Elias, S. Enhancement in Skin Cancer Detection using Image Super Resolution and Convolutional Neural Network. Procedia Comput. Sci. 2023, 218, 164–173. [Google Scholar] [CrossRef]
  22. Balaha, H.M.; Hassan, A.E.-S. Skin cancer diagnosis based on deep transfer learning and sparrow search algorithm. Neural Comput. Appl. 2022, 35, 815–853. [Google Scholar] [CrossRef]
  23. Brinker, T.J.; Hekler, A.; Enk, A.H.; Klode, J.; Hauschild, A.; Berking, C.; Schilling, B.; Haferkamp, S.; Schadendorf, D.; Fröhling, S.; et al. A convolutional neural network trained with dermoscopic images performed on par with 145 dermatologists in a clinical melanoma image classification task. Eur. J. Cancer 2019, 111, 148–154. [Google Scholar] [CrossRef] [PubMed]
  24. Tembhurne, J.V.; Hebbar, N.; Patil, H.Y.; Diwan, T. Skin cancer detection using ensemble of machine learning and deep learning techniques. Multimedia Tools Appl. 2023, 1–24. [Google Scholar] [CrossRef]
  25. Mazhar, T.; Haq, I.; Ditta, A.; Mohsan, S.A.H.; Rehman, F.; Zafar, I.; Gansau, J.A.; Goh, L.P.W. The Role of Machine Learning and Deep Learning Approaches for the Detection of Skin Cancer. Healthcare 2023, 11, 415. [Google Scholar] [CrossRef] [PubMed]
  26. Haenssle, H.A.; Fink, C.; Schneiderbauer, R.; Toberer, F.; Buhl, T.; Blum, A.; Kalloo, A.; Hassen, A.B.H.; Thomas, L.; Enk, A.; et al. Man against machine: Diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Ann. Oncol. Off. J. Eur. Soc. Med. Oncol. 2018, 29, 1836–1842. [Google Scholar] [CrossRef] [PubMed]
  27. Alenezi, F.; Armghan, A.; Polat, K. A multi-stage melanoma recognition framework with deep residual neural network and hyperparameter optimization-based decision support in dermoscopy images. Expert Syst. Appl. 2023, 215, 119352. [Google Scholar] [CrossRef]
  28. Inthiyaz, S.; Altahan, B.R.; Ahammad, S.H.; Rajesh, V.; Kalangi, R.R.; Smirani, L.K.; Hossain, A.; Rashed, A.N.Z. Skin disease detection using deep learning. Adv. Eng. Softw. 2023, 175, 103361. [Google Scholar] [CrossRef]
  29. Mendes, D.B.; da Silva, N.C. Skin lesions classification using convolutional neural networks in clinical images. arXiv 2018, arXiv:1812.02316. [Google Scholar]
  30. Aijaz, S.F.; Khan, S.J.; Azim, F.; Shakeel, C.S.; Hassan, U. Deep Learning Application for Effective Classification of Different Types of Psoriasis. J. Healthc. Eng. 2022, 2022, 7541583. [Google Scholar] [CrossRef]
  31. Khan, M.A.; Akram, T.; Zhang, Y.-D.; Sharif, M. Attributes based skin lesion detection and recognition: A mask RCNN and transfer learning-based deep learning framework. Pattern Recognit. Lett. 2021, 143, 58–66. [Google Scholar] [CrossRef]
  32. Zare, R.; Pourkazemi, A. DenseNet approach to segmentation and classification of dermatoscopic skin lesions images. arXiv 2021, arXiv:2110.04632. [Google Scholar]
  33. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.P.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  34. Li, S.; Zhao, M.; Fang, Z.; Zhang, Y.; Li, H. Image Super-Resolution Using Lightweight Multiscale Residual Dense Network. Int. J. Opt. 2020, 2020, 2852865. [Google Scholar] [CrossRef]
  35. Xia, X.; Xu, C.; Nan, B. Inception-v3 for flower classification. In Proceedings of the 2017 2nd International Conference on Image, Vision and Computing (ICIVC), Chengdu, China, 2–4 June 2017. [Google Scholar]
  36. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  37. Krause, J.; Sapp, B.; Howard, A.; Zhou, H.; Toshev, A.; Duerig, T.; Philbin, J.; Fei-Fei, L. The unreasonable effectiveness of noisy data for fine-grained recognition. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  38. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems (NIPS’12), Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  39. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  40. Serre, T.; Wolf, L.; Bileschi, S.; Riesenhuber, M.; Poggio, T. Robust object recognition with cortex-like mechanisms. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 411–426. [Google Scholar] [CrossRef] [PubMed]
  41. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  42. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  43. Arora, S.; Bhaskara, A.; Ge, R.; Ma, T. Provable bounds for learning some deep representations. In Proceedings of the International Conference on Machine Learning, Beijing, China, 21–26 June 2014. [Google Scholar]
  44. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  45. Ameri, A. A Deep Learning Approach to Skin Cancer Detection in Dermoscopy Images. J. Biomed. Phys. Eng. 2020, 10, 801–806. [Google Scholar] [CrossRef] [PubMed]
  46. Sae-Lim, W.; Wettayaprasit, W.; Aiyarak, P. Convolutional neural networks using MobileNet for skin lesion classification. In Proceedings of the 2019 16th International Joint Conference on Computer Science and Software Engineering (JCSSE), Chonburi, Thailand, 10–12 July 2019. [Google Scholar]
  47. Salian, A.C.; Vaze, S.; Singh, P.; Shaikh, G.N.; Chapaneri, S.; Jayaswal, D. Skin lesion classification using deep learning architectures. In Proceedings of the 2020 3rd International Conference on Com-Munication System, Computing and IT Applications (CSCITA), Mumbai, India, 3–4 April 2020. [Google Scholar]
  48. Pham, T.C.; Tran, G.S.; Nghiem, T.P.; Doucet, A.; Luong, C.M.; Hoang, V.D. A comparative study for classification of skin cancer. In Proceedings of the 2019 International Conference on System Science and Engineering (ICSSE), Dong Hoi, Vietnam, 19–21 July 2019. [Google Scholar]
  49. Rahman, Z.; Ami, A.M. A transfer learning based approach for skin lesion classification from imbalanced data. In Proceedings of the 2020 11th International Conference on Electrical and Computer Engineering (ICECE), Dhaka, Bangladesh, 17–19 December 2020. [Google Scholar]
  50. Polat, K.; Koc, K.O. Detection of Skin Diseases from Dermoscopy Image Using the combination of Convolutional Neural Network and One-versus-All. J. Artif. Intell. Syst. 2020, 2, 80–97. [Google Scholar] [CrossRef]
  51. Srinivasu, P.N.; SivaSai, J.G.; Ijaz, M.F.; Bhoi, A.K.; Kim, W.; Kang, J.J. Classification of skin disease using deep learning neural networks with MobileNet V2 and LSTM. Sensors 2021, 21, 2852. [Google Scholar] [CrossRef]
  52. Kausar, N.; Hameed, A.; Sattar, M.; Ashraf, R.; Imran, A.S.; Abidin, M.Z.U.; Ali, A. Multiclass Skin Cancer Classification Using Ensemble of Fine-Tuned Deep Learning Models. Appl. Sci. 2021, 11, 10593. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.